frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

The Case for the Return of Fine-Tuning

https://welovesota.com/article/the-case-for-the-return-of-fine-tuning
35•nanark•3h ago

Comments

oli5679•1h ago
The OpenAI fine-tuning api is pretty good - you need to label an evaluation benchmark anyway to systematically iterate on prompts and context, and it’s often creates good results if you give it a 50-100 examples, either beating frontier models or allowing a far cheaper and faster model to catch up.

It requires no local gpus, just creating a json and posting to OpenAI

https://platform.openai.com/docs/guides/model-optimization

deaux•43m ago
They don't offer it for GPT-5 series, as a result much of the time fine-tuning Gemini 2.5-Flash is a better deal.
melpomene•1h ago
This website loads at impressive speeds (from Europe)! Rarely seen anything more snappy. Dynamic loading of content as you scroll, small compressed images without looking like it (webp). Well crafted!
hshdhdhehd•1h ago
Magic of a CDN? Plus avoiding JS probably. Haven't checked source though.
CuriouslyC•1h ago
Fine tuning by pretraining over a RL tuned model is dumb AF. RL task tuning works quite well.
HarHarVeryFunny•48m ago
You may have no choice in how the model you are fine tuning was trained, and may have no interest in verticals it was RL tuned for.

In any case, platforms like tinker.ai support both SFT and RL.

CuriouslyC•17m ago
Why would you choose a model where the trained in priors don't match your use case? Also, keep in mind that RL'd in behavior includes things like reasoning and how to answer questions correctly, so you're literally taking smart models and making them dumber by doing SFT. To top it off, SFT only produces really good results when you have traces that closely model the actual behavior you're trying to get the model to display. If you're just trying to fine tune in a knowledge base, a well tuned RAG setup + better prompts win every time.
imcritic•12m ago
Because you need a solution for your problem and the available tools are what they are and nothing else and you don't have enough resources to train your own model.
empiko•52m ago
Fine-tuning is a good technique to have in a toolbox, but in reality, it is feasible only in some use cases. On one hand, many NLP tasks are already easy enough for LLMs to have near perfect accuracy and fine tuning is not needed. On the other hand, really complex tasks are really difficult to fine-tune and clevem data collection might be pretty expensive. Fine-tuning can help with the use cases somewhere in the middle, not too simple, not too complex, feasible for data collection, etc.
meander_water•32m ago
A couple of examples I have seen recently which makes me agree with OP:

- PaddleOCR, a 0.9B model that reaches SOTA accuracy across text, tables, formulas, charts & handwriting. [0]

- A 3B and 8B model which performs HTML to json extraction at GPT-5 level accuracy at 40-80x less cost, and faster inference. [1]

I think it makes sense to fine tune when you're optimizing for a specific task.

[0] https://huggingface.co/papers/2510.14528

[1] https://www.reddit.com/r/LocalLLaMA/comments/1o8m0ti/we_buil...

Show HN: Duck-UI – Browser-Based SQL IDE for DuckDB

https://demo.duckui.com
36•caioricciuti•1h ago•16 comments

What Happened in 2007?

https://whathappenedin2007.com/
29•AJRF•1h ago•6 comments

The Case for the Return of Fine-Tuning

https://welovesota.com/article/the-case-for-the-return-of-fine-tuning
35•nanark•3h ago•10 comments

EQ: A video about all forms of equalizers

https://www.youtube.com/watch?v=CLAt95PrwL4
188•robinhouston•1d ago•45 comments

OpenAI researcher announced GPT-5 math breakthrough that never happened

https://the-decoder.com/leading-openai-researcher-announced-a-gpt-5-math-breakthrough-that-never-...
20•Topfi•1h ago•1 comments

Jupyter Collaboration has a history slider

https://blog.jupyter.org/exploring-a-documents-timeline-in-jupyterlab-6084f96db263
30•fghorow•6d ago•5 comments

Titan submersible’s $62 SanDisk memory card found undamaged at wreckage site

https://www.tomshardware.com/pc-components/microsd-cards/tragic-oceangate-titan-submersibles-usd6...
362•WithinReason•2d ago•179 comments

Root System Drawings

https://images.wur.nl/digital/collection/coll13/search
371•bookofjoe•22h ago•74 comments

The Accountability Problem

https://www.jamesshore.com/v2/blog/2025/the-accountability-problem
83•FrancoisBosun•10h ago•28 comments

Pebble is officially back on iOS and Android

https://twitter.com/ericmigi/status/1979576965494710564
12•vlod•48m ago•1 comments

Chen-Ning Yang, Nobel laureate, dies at 103

https://www.chinadaily.com.cn/a/202510/18/WS68f3170ea310f735438b5bf2.html
247•nhatcher•1d ago•57 comments

How to sequence your DNA for <$2k

https://maxlangenkamp.substack.com/p/how-to-sequence-your-dna-for-2k
190•yichab0d•16h ago•81 comments

When you opened a screen shot of a video in Paint, the video was playing in it

https://devblogs.microsoft.com/oldnewthing/20251014-00/?p=111681
332•birdculture•2d ago•55 comments

How does Turbo listen for Turbo Streams

https://ducktypelabs.com/how-does-turbo-listen-for-turbo-streams/
65•sidk_•5d ago•8 comments

How one of the longest dinosaur trackways in the world was uncovered in the UK

https://www.bbc.co.uk/news/resources/idt-5f8c77b0-92bc-40f2-bf21-6793abbe5ffe
22•6LLvveMx2koXfwn•5d ago•0 comments

Why the open social web matters now

https://werd.io/why-the-open-social-web-matters-now/
178•benwerd•4d ago•109 comments

Flowistry: An IDE plugin for Rust that focuses on relevant code

https://github.com/willcrichton/flowistry
243•Bogdanp•22h ago•32 comments

Uber will offer gig work like AI data labeling to drivers while not on the road

https://www.cnbc.com/2025/10/16/uber-will-offer-us-drivers-more-gig-work-including-ai-data-labeli...
8•bobertdowney•43m ago•0 comments

./watch

https://dotslashwatch.com/
364•shrx•1d ago•101 comments

K8s with 1M nodes

https://bchess.github.io/k8s-1m/
238•denysvitali•2d ago•55 comments

Secret diplomatic message deciphered after 350 years

https://www.nationalarchives.gov.uk/explore-the-collection/the-collection-blog/secret-diplomatic-...
156•robin_reala•2d ago•32 comments

Tinnitus Neuromodulator

https://mynoise.net/NoiseMachines/neuromodulationTonesGenerator.php
322•gjvc•20h ago•210 comments

The optimistic case for protein foundation model companies

https://www.owlposting.com/p/the-optimistic-case-for-protein-foundation-193
11•crescit_eundo•1w ago•0 comments

IDEs we had 30 years ago and lost (2023)

https://blogsystem5.substack.com/p/the-ides-we-had-30-years-ago-and
524•AlexeyBrin•1d ago•469 comments

GoGoGrandparent (YC S16) Is Hiring Back End and Full-Stack Engineers

1•davidchl•11h ago

Coral NPU: A full-stack platform for Edge AI

https://research.google/blog/coral-npu-a-full-stack-platform-for-edge-ai/
138•LER0ever•3d ago•24 comments

Adding Breadcrumbs to a Rails Application

https://avohq.io/blog/breadcrumbs-rails
57•flow-flow•5d ago•7 comments

Lego Theft Ring

https://www.nytimes.com/2025/10/18/us/lego-theft-california-arrest.html
6•sanj•1h ago•0 comments

BQN "Macros" with •Decompose (2023)

https://saltysylvi.github.io/blog/bqn-macros.html
8•ofalkaed•1w ago•0 comments

Space junk falls on Western Australian minesite

https://www.abc.net.au/news/2025-10-19/wa-space-debris-reentry-investigation/105909612
39•dabiged•4h ago•7 comments