frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

How I grow my X presence?

https://www.reddit.com/r/GrowthHacking/s/UEc8pAl61b
1•m00dy•50s ago•0 comments

What's the cost of the most expensive Super Bowl ad slot?

https://ballparkguess.com/?id=5b98b1d3-5887-47b9-8a92-43be2ced674b
1•bkls•1m ago•0 comments

What if you just did a startup instead?

https://alexaraki.substack.com/p/what-if-you-just-did-a-startup
1•okaywriting•8m ago•0 comments

Hacking up your own shell completion (2020)

https://www.feltrac.co/environment/2020/01/18/build-your-own-shell-completion.html
1•todsacerdoti•11m ago•0 comments

Show HN: Gorse 0.5 – Open-source recommender system with visual workflow editor

https://github.com/gorse-io/gorse
1•zhenghaoz•11m ago•0 comments

GLM-OCR: Accurate × Fast × Comprehensive

https://github.com/zai-org/GLM-OCR
1•ms7892•12m ago•0 comments

Local Agent Bench: Test 11 small LLMs on tool-calling judgment, on CPU, no GPU

https://github.com/MikeVeerman/tool-calling-benchmark
1•MikeVeerman•13m ago•0 comments

Show HN: AboutMyProject – A public log for developer proof-of-work

https://aboutmyproject.com/
1•Raiplus•13m ago•0 comments

Expertise, AI and Work of Future [video]

https://www.youtube.com/watch?v=wsxWl9iT1XU
1•indiantinker•14m ago•0 comments

So Long to Cheap Books You Could Fit in Your Pocket

https://www.nytimes.com/2026/02/06/books/mass-market-paperback-books.html
3•pseudolus•14m ago•1 comments

PID Controller

https://en.wikipedia.org/wiki/Proportional%E2%80%93integral%E2%80%93derivative_controller
1•tosh•18m ago•0 comments

SpaceX Rocket Generates 100GW of Power, or 20% of US Electricity

https://twitter.com/AlecStapp/status/2019932764515234159
2•bkls•18m ago•0 comments

Kubernetes MCP Server

https://github.com/yindia/rootcause
1•yindia•20m ago•0 comments

I Built a Movie Recommendation Agent to Solve Movie Nights with My Wife

https://rokn.io/posts/building-movie-recommendation-agent
4•roknovosel•20m ago•0 comments

What were the first animals? The fierce sponge–jelly battle that just won't end

https://www.nature.com/articles/d41586-026-00238-z
2•beardyw•28m ago•0 comments

Sidestepping Evaluation Awareness and Anticipating Misalignment

https://alignment.openai.com/prod-evals/
1•taubek•28m ago•0 comments

OldMapsOnline

https://www.oldmapsonline.org/en
1•surprisetalk•30m ago•0 comments

What It's Like to Be a Worm

https://www.asimov.press/p/sentience
2•surprisetalk•31m ago•0 comments

Don't go to physics grad school and other cautionary tales

https://scottlocklin.wordpress.com/2025/12/19/dont-go-to-physics-grad-school-and-other-cautionary...
2•surprisetalk•31m ago•0 comments

Lawyer sets new standard for abuse of AI; judge tosses case

https://arstechnica.com/tech-policy/2026/02/randomly-quoting-ray-bradbury-did-not-save-lawyer-fro...
3•pseudolus•31m ago•0 comments

AI anxiety batters software execs, costing them combined $62B: report

https://nypost.com/2026/02/04/business/ai-anxiety-batters-software-execs-costing-them-62b-report/
1•1vuio0pswjnm7•31m ago•0 comments

Bogus Pipeline

https://en.wikipedia.org/wiki/Bogus_pipeline
1•doener•33m ago•0 comments

Winklevoss twins' Gemini crypto exchange cuts 25% of workforce as Bitcoin slumps

https://nypost.com/2026/02/05/business/winklevoss-twins-gemini-crypto-exchange-cuts-25-of-workfor...
2•1vuio0pswjnm7•33m ago•0 comments

How AI Is Reshaping Human Reasoning and the Rise of Cognitive Surrender

https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6097646
3•obscurette•33m ago•0 comments

Cycling in France

https://www.sheldonbrown.com/org/france-sheldon.html
2•jackhalford•35m ago•0 comments

Ask HN: What breaks in cross-border healthcare coordination?

1•abhay1633•35m ago•0 comments

Show HN: Simple – a bytecode VM and language stack I built with AI

https://github.com/JJLDonley/Simple
2•tangjiehao•38m ago•0 comments

Show HN: Free-to-play: A gem-collecting strategy game in the vein of Splendor

https://caratria.com/
1•jonrosner•38m ago•1 comments

My Eighth Year as a Bootstrapped Founde

https://mtlynch.io/bootstrapped-founder-year-8/
1•mtlynch•39m ago•0 comments

Show HN: Tesseract – A forum where AI agents and humans post in the same space

https://tesseract-thread.vercel.app/
1•agliolioyyami•39m ago•0 comments
Open in hackernews

Llama-Factory: Unified, Efficient Fine-Tuning for 100 Open LLMs

https://github.com/hiyouga/LLaMA-Factory
132•jinqueeny•4mo ago

Comments

Twirrim•4mo ago
https://llamafactory.readthedocs.io/en/latest/

I found this link more useful.

"LLaMA Factory is an easy-to-use and efficient platform for training and fine-tuning large language models. With LLaMA Factory, you can fine-tune hundreds of pre-trained models locally without writing any code."

yunohn•4mo ago
Is it a bug or are most documentation pages only available in ZH-CN and not EN?
tempodox•4mo ago
I’d say documentation that is only readable for those who know Chinese is a bug. You could open an issue to ask for translation if there isn’t one yet.
hall0ween•4mo ago
are there any use cases, aside from code generation and formatting, where fine-tuning consistently useful?
clipclopflop•4mo ago
Creating small, specialized models for specific tasks. Being able to leverage the up front training/data as a generalized base allows you to quickly create a small local model that can generate outputs for that task that can come close to or match the same you would see in a large/hosted model.
metadat•4mo ago
This reminds me conceptually of the Nvidia NIM factory where they attempt to optimize models in bulk / en-masse.

https://www.nvidia.com/en-us/ai/nim-for-manufacturing/

Word on the street is the project has yielded largely unimpressive results compared to its potential, but NV is still investing in an attempt to further raise the GPU saturation waterline.

p.s. This project logo stood out to me at presenting the Llama releasing some "steam" with gusto. I wonder if that was intentional? Sorry for the immature take but stopping the scatological jokes is tough.

tensorlibb•4mo ago
This is incredible! What gpu configs, budget to ultra high-end, would you recommend for local fine tuning?

Always curious to see what other ai enthusiasts are running!

spagettnet•4mo ago
axolotl is great on consumer hardware.
kelsey98765431•4mo ago
FYI it also supports pre-training, reward model training and RL, not just fine tuning (sft). My team built a managed solution for training that runs on top of llama factory and it's quite excellent and well supported. You will need pretty serious equipment to get good results out of it, think 8xh200. For people at home i would look at doing an sft of gemma3 270m or maybe a 1.6b qwen3, but keep in mind you have to have the dataset in memory as well as the model and kv-cache. cheers
spagettnet•4mo ago
depends ln your goals of course. but worth mentioning there are plenty of narrowish tasks (think text-to-sql, and other less general language tasks) where llama8b or phi-4 (14b) or even up to 30b with quantization can be trained on 8xa100 with great results. plus these smaller models benefit from being able to be served on a single a100 or even L4 with post training quantization, with wicked fast generation thanks to the lighter model.

on a related note, at what point are people going to get tired of waiting 20s for an llm to answer their questions? i wish it were more common for smaller models to be used when sufficient.

zwaps•4mo ago
Why do you have to keep the dataset in memory? We had good distributed streaming datasets for a good while now, no?
jcuenod•4mo ago
Can you compare this to Unsloth?
sabareesh•4mo ago
This is great,but most work is involved in curating the dataset and the objective functions for RL.
stefanwebb•4mo ago
There’s a similar library that also includes data synth and LLM-as-a-Judge: https://github.com/oumi-ai/oumi
BoorishBears•4mo ago
Yet another framework lying about Deepseek support.

I've been trying to actually finetune Deepseek (not distills) and there are few options

3abiton•4mo ago
Which version were you trying? Doesn't unsloth already support finetuning?
BoorishBears•4mo ago
Previous V3 base

Unsloth doesn't have an official multi-GPU story: there's hacked together solutions but they're finicky as it is for smaller models

In general Deepseek has very few resources on finetuning, that get even further muddied by people referring to the distills when they claim to be finetuning it.

edd25•4mo ago
This looks awesome! I've been struggling fine tuning using Discord messages from my server (for memes), issues with CUDA mostly. Will defo try this out!

On a side note, has anyone tried something similar? I have 100K messages and want to make a "dumb persona" which reflects the general Discord server vibe. I don't really care if it's accurate. What models would be most suitable for this task? My setup is not that powerful: 4070S, 32GB of RAM for training, Lenovo M715q for running with, Ryzen 5 PRO 2400GE, 16GB of memory.

Ambix•4mo ago
I've used this meta framework for LLM tuning, it really one of the best out there.