frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Al Lowe on model trains, funny deaths and working with Disney

https://spillhistorie.no/2026/02/06/interview-with-sierra-veteran-al-lowe/
50•thelok•3h ago•6 comments

Hoot: Scheme on WebAssembly

https://www.spritely.institute/hoot/
115•AlexeyBrin•6h ago•20 comments

Stories from 25 Years of Software Development

https://susam.net/twenty-five-years-of-computing.html
49•vinhnx•4h ago•7 comments

OpenCiv3: Open-source, cross-platform reimagining of Civilization III

https://openciv3.org/
811•klaussilveira•21h ago•246 comments

The AI boom is causing shortages everywhere else

https://www.washingtonpost.com/technology/2026/02/07/ai-spending-economy-shortages/
91•1vuio0pswjnm7•7h ago•102 comments

Reinforcement Learning from Human Feedback

https://rlhfbook.com/
72•onurkanbkrc•6h ago•5 comments

The Waymo World Model

https://waymo.com/blog/2026/02/the-waymo-world-model-a-new-frontier-for-autonomous-driving-simula...
1053•xnx•1d ago•600 comments

Start all of your commands with a comma (2009)

https://rhodesmill.org/brandon/2009/commands-with-comma/
470•theblazehen•2d ago•173 comments

Vocal Guide – belt sing without killing yourself

https://jesperordrup.github.io/vocal-guide/
196•jesperordrup•11h ago•67 comments

Selection Rather Than Prediction

https://voratiq.com/blog/selection-rather-than-prediction/
8•languid-photic•3d ago•1 comments

Speed up responses with fast mode

https://code.claude.com/docs/en/fast-mode
9•surprisetalk•1h ago•2 comments

U.S. Jobs Disappear at Fastest January Pace Since Great Recession

https://www.forbes.com/sites/mikestunson/2026/02/05/us-jobs-disappear-at-fastest-january-pace-sin...
44•alephnerd•1h ago•14 comments

France's homegrown open source online office suite

https://github.com/suitenumerique
536•nar001•5h ago•248 comments

Coding agents have replaced every framework I used

https://blog.alaindichiappari.dev/p/software-engineering-is-back
204•alainrk•6h ago•310 comments

A Fresh Look at IBM 3270 Information Display System

https://www.rs-online.com/designspark/a-fresh-look-at-ibm-3270-information-display-system
33•rbanffy•4d ago•6 comments

72M Points of Interest

https://tech.marksblogg.com/overture-places-pois.html
26•marklit•5d ago•1 comments

Unseen Footage of Atari Battlezone Arcade Cabinet Production

https://arcadeblogger.com/2026/02/02/unseen-footage-of-atari-battlezone-cabinet-production/
110•videotopia•4d ago•30 comments

Software factories and the agentic moment

https://factory.strongdm.ai/
63•mellosouls•4h ago•68 comments

Where did all the starships go?

https://www.datawrapper.de/blog/science-fiction-decline
67•speckx•4d ago•71 comments

Show HN: Kappal – CLI to Run Docker Compose YML on Kubernetes for Local Dev

https://github.com/sandys/kappal
21•sandGorgon•2d ago•11 comments

Show HN: Look Ma, No Linux: Shell, App Installer, Vi, Cc on ESP32-S3 / BreezyBox

https://github.com/valdanylchuk/breezydemo
271•isitcontent•21h ago•36 comments

Learning from context is harder than we thought

https://hy.tencent.com/research/100025?langVersion=en
199•limoce•4d ago•110 comments

Monty: A minimal, secure Python interpreter written in Rust for use by AI

https://github.com/pydantic/monty
284•dmpetrov•21h ago•151 comments

Making geo joins faster with H3 indexes

https://floedb.ai/blog/how-we-made-geo-joins-400-faster-with-h3-indexes
155•matheusalmeida•2d ago•48 comments

Hackers (1995) Animated Experience

https://hackers-1995.vercel.app/
553•todsacerdoti•1d ago•267 comments

Sheldon Brown's Bicycle Technical Info

https://www.sheldonbrown.com/
424•ostacke•1d ago•110 comments

Ga68, a GNU Algol 68 Compiler

https://fosdem.org/2026/schedule/event/PEXRTN-ga68-intro/
41•matt_d•4d ago•16 comments

Show HN: If you lose your memory, how to regain access to your computer?

https://eljojo.github.io/rememory/
348•eljojo•1d ago•214 comments

An Update on Heroku

https://www.heroku.com/blog/an-update-on-heroku/
466•lstoll•1d ago•308 comments

Show HN: I spent 4 years building a UI design tool with only the features I use

https://vecti.com
367•vecti•23h ago•167 comments
Open in hackernews

Phi-4 Reasoning Models

https://azure.microsoft.com/en-us/blog/one-year-of-phi-small-language-models-making-big-leaps-in-ai/
131•meetpateltech•9mo ago

Comments

refulgentis•9mo ago
These look quite incredible. I work on a llama.cpp GUI wrapper and its quite surprising to see how well Microsoft's Phi-4 releases set it apart as the only competition below ~7B, it'll probably take a year for the FOSS community to implement and digest it completely (it can do multimodal! TTS! STT! Conversation!)
wd776g5•9mo ago
The linked article says 14B parameters. edit and I guess the "plus" model is 21B?
refulgentis•9mo ago
grep "As seen above, Phi-4-mini-reasoning with 3.8B parameters outperforms models of over twice its size."

re: reasoning plus, "Phi-4-reasoning-plus builds upon Phi-4-reasoning capabilities, further trained with reinforcement learning to utilize more inference-time compute, using 1.5x more tokens than Phi-4-reasoning, to deliver higher accuracy.", presumably also 14B

thenameless7741•9mo ago
> it'll probably take a year for the FOSS community to implement and digest it completely

The local community seems to have converged on a few wrappers: Open WebUI (general-purpose), LM Studio (proprietary), and SillyTavern (for role-playing). Now that llama.cpp has an OpenAI-compatible server (llama-server), there's a lot more options to choose from.

I've noticed there really aren't many active FOSS wrappers these days - most of them have either been abandoned or aren't being released with the frequency we saw when OpenAI API first launched. So it would be awesome if you could share your wrapper with us at some point.

trc001•9mo ago
Oobabooga is still good as a Swiss Army knife sort of wrapper for a single user trying out new models
pzo•9mo ago
I think OP means that FOSS didn't digest many multimodals of phi4-mini-multimodal such as Audio Input (STT) and Audio Output (TTS), also Image Input also not much supported in many FOSS.
thenameless7741•9mo ago
AFAIK, Phi-4-multimodal doesn't support TTS, but I understand OP's point.

The recent Qwen's release is an excellent example of model providers collaborating with the local community (which include inference engine developers and model quantizers?). It would be nice if this collaboration extended to wrapper developers as well, so that end-users can enjoy a great UX from day one of any model release.

refulgentis•9mo ago
Hah, ty, I badly misunderstood the release materials
loufe•9mo ago
Kobold.cpp is still my preference for a gui. Single portable exe with good flexibility in configuration if you want it, no need if not.
joshstrange•9mo ago
I've been happier with LibreChat over Open WebUI. Mostly because I wasn't a fan of the `pipelines` stuff in Open WebUI and lack of MCP support (probably has changed now?). But then I don't love how LibreChat wants to push its (expensive) code runner service.
gthompson512•9mo ago
Sorry if this comment is outdated or ill-informed, but it is hard to follow the current news. Do the Phi models still have issues with training on the test set, or have they fixed that?
csdvrx•9mo ago
Is anyone here using phi-4 multimodal for image-to-text tasks?

The phi models often punch above their weight, and I got curious about the vision models after reading https://unsloth.ai/blog/phi4 stories of finetuning

Since lmarena.ai only has the phi-4 text model, I've tried "phi-4 multimodal instruct" from openrouter.ai.

However, the results I get are far below what I would have expected.

Is there any "Microsoft validated" source (like https://chat.qwen.ai/c/guest for qwen) to easily try phi4 vision?

adt•9mo ago
https://lifearchitect.ai/models-table/
danielhanchen•9mo ago
We uploaded GGUFs for anyone who wants to run them locally.

[EDIT] - I fixed all chat templates so no need for --jinja as at 10:00PM SF time.

Phi-4-mini-reasoning GGUF: https://huggingface.co/unsloth/Phi-4-mini-reasoning-GGUF

Phi-4-reasoning-plus-GGUF: https://huggingface.co/unsloth/Phi-4-reasoning-plus-GGUF

danielhanchen•9mo ago
Final edit 11:40PM PT/SF time: I noticed not using --jinja doesn't work - so unfortunately you'll have to use it.

Also it looks very very very verbose. The recommended settings are --temp 0.8 --top-p 0.95

justanotheratom•9mo ago
is there a well-established tool-chain for finetuning these models?
idonotknowwhy•9mo ago
Unsloth. Check their colab notebooks
behnamoh•9mo ago
what's Microsoft's plan here? on one hand they have a huge share in OpenAI and Mistral, and on the other hand they release open weight models.
changoplatanero•9mo ago
Building their own models gives them more leverage for negotiating favorable terms with openai. The key to a good negotiation is having a strong BATNA, best alternative to a negotiated agreement.
onlyrealcuzzo•9mo ago
I think OP poses a strong question.

If you destroy $500B in market cap in your share in said companies to get a better deal in a small segment of your core business...

That might not end up being a very bad deal overall.

The reality is, if MSFT wasn't delivering this SOTA SLMs now, someone else would be doing it within a few months.

OneDeuxTriSeiGo•9mo ago
Microsoft's push long term seems to be for local-first SLMs. That's the end goal.
csdvrx•9mo ago
Yes: they are building both the software and the hardware for that: https://blogs.windows.com/windowsexperience/2024/12/06/phi-s...
gitroom•9mo ago
Honestly the Phi-4 stuff is starting to get real interesting for me. Im still confused about Microsofts whole play here, but thats kind of what makes it fun to watch.
csdvrx•9mo ago
Read https://blogs.windows.com/windowsexperience/2024/12/06/phi-s...

I submitted it, as it gives a better picture of what Microsoft is trying to do: both the hardware, and the software.

Phi is small, not just for shows, but also to be able to run locally on the hardware they are planning for it to run on the copilot branded devices.

mrbonner•9mo ago
The example prompt for reasoning model that never fails to amuse me: "How amy letter 'r's in the word 'strrawberrry'?"

Phi-4-mini-reasoning: thought for 2 min 3 sec

<think> Okay, let's see here. The user wants to know how many times the letter 'r' appears in the word 'strrawberrry'. Hmm, first I need to make sure I'm spelling the word correctly. Let me write it out slowly: S-T-R-A-W-B-E-R-R-Y. ...

But I'm overcomplicating this. Given that the user explicitly writes 'strawberrry' with multiple r's (as indicated by \texttt{rrrr}), I think they mean there are three R's in that word. Therefore, counting them would give an answer of 3. </think>

The word "strrawberrry" contains *5* occurrences of the letter 'r'.

To determine this: 1. Break down the spelling: S-T-R-A-W-B-E-R-R-Y. 2. Identify each 'R': - The first R is in the "straw" part (position 3). - Additional Rs appear at positions corresponding to three instances of 'r' due to repetition.

*Answer:* \boxed{5}

I have no idea how it came to that conclusion with those thoughts lol.

The full transcript: https://pastebin.com/tyF5aSb9

T-A•9mo ago
> I have no idea how it came to that conclusion with those thoughts lol.

"There’s no specific reason why the reported Chain-of-Thought must accurately reflect the true reasoning process; there might even be circumstances where a model actively hides aspects of its thought process from the user." [1]

[1] https://www.anthropic.com/research/reasoning-models-dont-say...