frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Start all of your commands with a comma (2009)

https://rhodesmill.org/brandon/2009/commands-with-comma/
261•theblazehen•2d ago•88 comments

Hoot: Scheme on WebAssembly

https://www.spritely.institute/hoot/
27•AlexeyBrin•1h ago•3 comments

OpenCiv3: Open-source, cross-platform reimagining of Civilization III

https://openciv3.org/
707•klaussilveira•15h ago•206 comments

The Waymo World Model

https://waymo.com/blog/2026/02/the-waymo-world-model-a-new-frontier-for-autonomous-driving-simula...
970•xnx•21h ago•558 comments

Reinforcement Learning from Human Feedback

https://arxiv.org/abs/2504.12501
9•onurkanbkrc•51m ago•0 comments

Vocal Guide – belt sing without killing yourself

https://jesperordrup.github.io/vocal-guide/
73•jesperordrup•6h ago•32 comments

Making geo joins faster with H3 indexes

https://floedb.ai/blog/how-we-made-geo-joins-400-faster-with-h3-indexes
135•matheusalmeida•2d ago•35 comments

Where did all the starships go?

https://www.datawrapper.de/blog/science-fiction-decline
46•speckx•4d ago•38 comments

Unseen Footage of Atari Battlezone Arcade Cabinet Production

https://arcadeblogger.com/2026/02/02/unseen-footage-of-atari-battlezone-cabinet-production/
68•videotopia•4d ago•7 comments

Welcome to the Room – A lesson in leadership by Satya Nadella

https://www.jsnover.com/blog/2026/02/01/welcome-to-the-room/
39•kaonwarb•3d ago•30 comments

Ga68, a GNU Algol 68 Compiler

https://fosdem.org/2026/schedule/event/PEXRTN-ga68-intro/
13•matt_d•3d ago•2 comments

What Is Ruliology?

https://writings.stephenwolfram.com/2026/01/what-is-ruliology/
45•helloplanets•4d ago•46 comments

Show HN: Look Ma, No Linux: Shell, App Installer, Vi, Cc on ESP32-S3 / BreezyBox

https://github.com/valdanylchuk/breezydemo
240•isitcontent•16h ago•26 comments

Monty: A minimal, secure Python interpreter written in Rust for use by AI

https://github.com/pydantic/monty
238•dmpetrov•16h ago•128 comments

Show HN: I spent 4 years building a UI design tool with only the features I use

https://vecti.com
340•vecti•18h ago•150 comments

Hackers (1995) Animated Experience

https://hackers-1995.vercel.app/
506•todsacerdoti•23h ago•248 comments

Sheldon Brown's Bicycle Technical Info

https://www.sheldonbrown.com/
390•ostacke•22h ago•99 comments

Show HN: If you lose your memory, how to regain access to your computer?

https://eljojo.github.io/rememory/
306•eljojo•18h ago•189 comments

Microsoft open-sources LiteBox, a security-focused library OS

https://github.com/microsoft/litebox
361•aktau•22h ago•186 comments

An Update on Heroku

https://www.heroku.com/blog/an-update-on-heroku/
430•lstoll•22h ago•284 comments

Cross-Region MSK Replication: K2K vs. MirrorMaker2

https://medium.com/lensesio/cross-region-msk-replication-a-comprehensive-performance-comparison-o...
3•andmarios•4d ago•1 comments

Was Benoit Mandelbrot a hedgehog or a fox?

https://arxiv.org/abs/2602.01122
25•bikenaga•3d ago•11 comments

PC Floppy Copy Protection: Vault Prolok

https://martypc.blogspot.com/2024/09/pc-floppy-copy-protection-vault-prolok.html
71•kmm•5d ago•10 comments

Dark Alley Mathematics

https://blog.szczepan.org/blog/three-points/
96•quibono•4d ago•22 comments

The AI boom is causing shortages everywhere else

https://www.washingtonpost.com/technology/2026/02/07/ai-spending-economy-shortages/
26•1vuio0pswjnm7•2h ago•17 comments

How to effectively write quality code with AI

https://heidenstedt.org/posts/2026/how-to-effectively-write-quality-code-with-ai/
271•i5heu•18h ago•219 comments

Delimited Continuations vs. Lwt for Threads

https://mirageos.org/blog/delimcc-vs-lwt
34•romes•4d ago•3 comments

I now assume that all ads on Apple news are scams

https://kirkville.com/i-now-assume-that-all-ads-on-apple-news-are-scams/
1079•cdrnsf•1d ago•463 comments

Introducing the Developer Knowledge API and MCP Server

https://developers.googleblog.com/introducing-the-developer-knowledge-api-and-mcp-server/
64•gfortaine•13h ago•30 comments

Understanding Neural Network, Visually

https://visualrambling.space/neural-network/
306•surprisetalk•3d ago•45 comments
Open in hackernews

Ultrathink is a Claude Code magic word

https://simonwillison.net/2025/Apr/19/claude-code-best-practices/
113•ghuntley•9mo ago

Comments

andrewfromx•9mo ago
I asked Claude if this was true, and Claude confirmed.
canadiantim•9mo ago
That very very quickly moved from blog to twitter to blog to HN. Gotta love the velocity of information these days
doubled112•9mo ago
Link first, ask questions later
sauravt•9mo ago
megathink sounds better
benatkin•9mo ago
And doublemegathink if you want it to do two megathinks in parallel
Terr_•9mo ago
Not to be confused with doublethink, a mode that is always active for LLMs.
dghlsakjg•9mo ago
But paradoxically only allocates 1/3 the tokens according to the code.

Perhaps they should switch to the metric thinking system.

Gigathinking, and Terathinking should be on the menu as well.

fragmede•9mo ago
Crazy that it's a key word that's implemented in the code that expands the context window, and that a light touch of reverse engineering was required to find it.
dickfickling•9mo ago
It’s described here: https://www.anthropic.com/engineering/claude-code-best-pract...
fragmede•9mo ago
Ah yes, the documentation. If everyone read documentation, we wouldn't need LLMs to read it for us!
zenkey•9mo ago
It would be cool if these "secret keywords" were more directly exposed in the UI somehow, perhaps as a toggleable developer/experimental mode? I would have a lot of fun tinkering with them.
refulgentis•9mo ago
It's for Claude Code FWIW, just leaving a sigil here for fellow API implementers who are confused: your general point stands (though I wonder about UI affordances other than text given it's a CLI tool)
m1keil•9mo ago
I hope we will exit this stage of magic spells and incantations sooner rather than later.
Frummy•9mo ago
I hope we delve deeper into pentacles and rites in candlelit basements to appease black boxes of neural mimicries of canaanite archetypes
timschmidt•9mo ago
So long as we get to play weird science[1] during the montage.

1: https://www.youtube.com/watch?v=Jm-upHSP9KU

patcon•9mo ago
Sincerely, I respect your response to how arbitrary it seems in this form.

But... I'd like you to take a moment and think really hard about whether this is truly novel behavior for LLMs, or rather something that has always been part of the interplay between inter-agent communication and intra-agent thought :)

bee_rider•9mo ago
It sounds like it is a “specific phrase mapped directly” based on another comment here? I guess that means hardcoded? Not completely sure, though.
simonw•9mo ago
It's hard-coded - this isn't a weird model thing, Claude Code detects the exact string "ultrathink" and sets the thinking token budget to 31999.

I included that de-obfuscated code in my post: https://simonwillison.net/2025/Apr/19/claude-code-best-pract...

4b11b4•9mo ago
I thought that earlier on, I don't think we will though
ffsm8•9mo ago
Maybe we will get to Babylon 5s reality of having literal tech wizards or whatever they were called - which summon "demons" etc
slowmovintarget•9mo ago
Technomages. They were humans (and some others) bonded to Shadow-manipulated creatures that could affect spacetime. Technomages learned to "cast spells" in a private language they developed with their symbiote. These were essentially commands to "do the thing I'm talking about."

One of the challenges Technomages faced after acquiring their symbiote is gaining control over both the personal language and the impulses for destruction embedded by the Shadows (bias? bias).

So yes, learning the tricks of getting our AI agents to do the thing we actually meant, while aiming to overcome bias built in to the models when we can see it does sound an awful lot like Technomages dealing with their symbiote.

vunderba•9mo ago
@dickfickling beat me to it, but ultrathink is already explicitly called out in the public Anthropic documentation:

"Ask Claude to make a plan for how to approach a specific problem. We recommend using the word "think" to trigger extended thinking mode, which gives Claude additional computation time to evaluate alternatives more thoroughly. These specific phrases are mapped directly to increasing levels of thinking budget in the system: "think" < "think hard" < "think harder" < "ultrathink." Each level allocates progressively more thinking budget for Claude to use."

https://www.anthropic.com/engineering/claude-code-best-pract...

I don't know what the max allowable "budget_tokens" is for Claude 3.7 Thinking mode, but the SDK shows an example of 32k which matches up with the article's findings.

simonw•9mo ago
Looks like that documentation is incorrect. It suggests there are four levels - "think" < "think hard" < "think harder" < "ultrathink." - but if you look in the code there are actually only three.
techjamie•9mo ago
Pointed out elsewhere in this thread, but the docs call this phrase and others out as specific words to allocate a higher thinking budget.*

I'm assuming it's to unify the experience between chat and API, and avoid switching models for it ala OAI.

* https://www.anthropic.com/engineering/claude-code-best-pract...

simonw•9mo ago
I quoted the documentation in my article.
wpollock•9mo ago
Nice to know, although I was taught that the magic word is "please".
replwoacause•9mo ago
This would be helpful information if I hadn’t already switched to Gemini 2.5 because it’s 96% cheaper
user3939382•9mo ago
After stunts like Amp and Web Integrity (among others) I don’t care what they charge, I want nothing to do with Google.
bn-l•9mo ago
It does feel like a Faustian bargain using it.
esperent•9mo ago
I've had a frustrating time over the last couple of days with Gemini 2.5 Pro.

First I asked it to help me reverse the direction of text on a circle in Photoshop. It gave me very specific instructions which don't work and continued to argue with me that I was doing something wrong - I did my own research and found it's not actually possible to do this in Photoshop, and the instructions it was giving me were for Illustrator. 30 minutes of my time wasted.

This morning I asked it how to remove the axis lines from the orthographic view in Blender 4.3. I explained carefully that I know how to remove them in perspective view but that wasn't working for orthographic views. It over and over told me how to remove them from perspective views, telling me to use non existent UI elements, even drawing ASCII diagrams of how to find the nonexistent icons. When I said they didn't exist, it would circle back to telling me how to turn them off in the perspective view.

It turns out, again, it's not possible to remove grid lines from orthographic views in Blender (at least without messing around with the theme settings, or turning off the grid entirely).

In both cases it was incredibly persistent in stating the wrong way to do things, even when I was saying that it didn't work. I felt like it was gaslighting me, moreso than with any previous model I've used.

I haven't yet used it for writing code but these two experiences don't make me feel hopeful. The worst part about dealing with AI is when they are confidently incorrect.

simonw•9mo ago
These are good examples of things that I wouldn't expect an LLM to get right, based purely on my own intuition.

I don't believe they have much training material on the UI for tools at the moment - it may well come in the future as these new "computer use" models get fed vast amounts of screen capture videos, but to date my hunch is that there hasn't been much focus on that, especially for tools like Photoshop and Blender (training them to use a web browser is a whole lot more useful for the moment).

I'd encourage you not to assume they suck at code just because they suck at answering questions about Photoshop and Blender. I wrote about that a while back: "Don’t fall into the trap of anthropomorphizing LLMs and assuming that failures which would discredit a human should discredit the machine in the same way." - https://simonwillison.net/2025/Mar/11/using-llms-for-code/#s...

esperent•9mo ago
Have you actually extensively tried using any LLMs for help with Blender/Photoshop/other apps, or are you just speaking based on your intuition?

I use Claude for help with Blender all the time and it's amazing, for the most part. It does have an in depth knowledge of the UI and also of many specific technical ways of doing things. The main thing that it gets stuck on is UI changes between versions, and to be fair, I gets stuck on this too.

I will try this same query on Claude tomorrow when I'm in my office. I suspect it'll get it wrong as well, but it's not so much the getting wrong that I had an issue with, but how persistent Gemini was in refusing to admit error and making it seem like I was the one at fault.

simonw•9mo ago
I've tried it for a few different GUI things (and "how do I do X on website Y" things) with very mixed results. I've not used it for Blender.

Really interesting to hear that Claude does well at this kind of problem! Maybe that's thanks to training they did relative to their Claude Computer Use research last year. https://simonwillison.net/2024/Oct/22/computer-use/

layer8•9mo ago
I think I'll wait for Hyperthink.
namaria•9mo ago
I'm personally holding out for Ludicrousthink
sn9•9mo ago
Tengu think? As in Japanese Tengu?
xrd•9mo ago
I'm scratching my head a bit at this one.

I already assume that the models are shifting underneath me. It's very frustrating that most non-developers just think you can ask an LLM a question and it will respond accurately each time. They are designed to make creative output and even if you dial down the temperature they still can hallucinate.

Why not be explicit about the thinking budget instead of aliasing it to a number with a term like ultrathink?

It's a cute word, and fun to know is managed on the client side, but isn't it again more imprecision to tools the are already suffering from that?

j_bum•9mo ago
> shifting underneath me

I love that language. This is something that’s constantly bothered me from the get go. Maybe I’m just wearing a tinfoil hat, but I swear I’ve noticed variations in behavior and performance within models across time.

This has been super annoying to me.

I always use the chat interfaces (mostly Claude atm), so I guess that just puts me at the whim of which sub-version of model Anthropic is serving for the day.

Or again, maybe I’m just hallucinating.

seanhunter•9mo ago
It’s fair to point out this is probably already evidence for AGI. Most days I personally am only able to infrathink at best.