frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

OpenCiv3: Open-source, cross-platform reimagining of Civilization III

https://openciv3.org/
418•klaussilveira•5h ago•94 comments

The Waymo World Model

https://waymo.com/blog/2026/02/the-waymo-world-model-a-new-frontier-for-autonomous-driving-simula...
770•xnx•11h ago•465 comments

Show HN: Look Ma, No Linux: Shell, App Installer, Vi, Cc on ESP32-S3 / BreezyBox

https://github.com/valdanylchuk/breezydemo
137•isitcontent•5h ago•15 comments

Monty: A minimal, secure Python interpreter written in Rust for use by AI

https://github.com/pydantic/monty
131•dmpetrov•6h ago•54 comments

Dark Alley Mathematics

https://blog.szczepan.org/blog/three-points/
37•quibono•4d ago•2 comments

Show HN: I spent 4 years building a UI design tool with only the features I use

https://vecti.com
241•vecti•8h ago•116 comments

A century of hair samples proves leaded gas ban worked

https://arstechnica.com/science/2026/02/a-century-of-hair-samples-proves-leaded-gas-ban-worked/
63•jnord•3d ago•4 comments

Microsoft open-sources LiteBox, a security-focused library OS

https://github.com/microsoft/litebox
309•aktau•12h ago•153 comments

Sheldon Brown's Bicycle Technical Info

https://www.sheldonbrown.com/
309•ostacke•11h ago•84 comments

Show HN: If you lose your memory, how to regain access to your computer?

https://eljojo.github.io/rememory/
168•eljojo•8h ago•124 comments

Why I Joined OpenAI

https://www.brendangregg.com/blog/2026-02-07/why-i-joined-openai.html
38•SerCe•1h ago•34 comments

Hackers (1995) Animated Experience

https://hackers-1995.vercel.app/
391•todsacerdoti•13h ago•217 comments

An Update on Heroku

https://www.heroku.com/blog/an-update-on-heroku/
314•lstoll•12h ago•230 comments

Show HN: R3forth, a ColorForth-inspired language with a tiny VM

https://github.com/phreda4/r3
48•phreda4•5h ago•8 comments

I spent 5 years in DevOps – Solutions engineering gave me what I was missing

https://infisical.com/blog/devops-to-solutions-engineering
107•vmatsiiako•10h ago•34 comments

How to effectively write quality code with AI

https://heidenstedt.org/posts/2026/how-to-effectively-write-quality-code-with-ai/
181•i5heu•8h ago•128 comments

Understanding Neural Network, Visually

https://visualrambling.space/neural-network/
233•surprisetalk•3d ago•30 comments

Introducing the Developer Knowledge API and MCP Server

https://developers.googleblog.com/introducing-the-developer-knowledge-api-and-mcp-server/
14•gfortaine•3h ago•0 comments

I now assume that all ads on Apple news are scams

https://kirkville.com/i-now-assume-that-all-ads-on-apple-news-are-scams/
971•cdrnsf•15h ago•414 comments

Learning from context is harder than we thought

https://hy.tencent.com/research/100025?langVersion=en
141•limoce•3d ago•79 comments

FORTH? Really!?

https://rescrv.net/w/2026/02/06/associative
40•rescrv•13h ago•17 comments

PC Floppy Copy Protection: Vault Prolok

https://martypc.blogspot.com/2024/09/pc-floppy-copy-protection-vault-prolok.html
8•kmm•4d ago•0 comments

I'm going to cure my girlfriend's brain tumor

https://andrewjrod.substack.com/p/im-going-to-cure-my-girlfriends-brain
42•ray__•2h ago•11 comments

Evaluating and mitigating the growing risk of LLM-discovered 0-days

https://red.anthropic.com/2026/zero-days/
34•lebovic•1d ago•11 comments

Show HN: Smooth CLI – Token-efficient browser for AI agents

https://docs.smooth.sh/cli/overview
76•antves•1d ago•57 comments

The Oklahoma Architect Who Turned Kitsch into Art

https://www.bloomberg.com/news/features/2026-01-31/oklahoma-architect-bruce-goff-s-wild-home-desi...
18•MarlonPro•3d ago•4 comments

Show HN: Slack CLI for Agents

https://github.com/stablyai/agent-slack
38•nwparker•1d ago•9 comments

Claude Composer

https://www.josh.ing/blog/claude-composer
102•coloneltcb•2d ago•69 comments

How virtual textures work

https://www.shlom.dev/articles/how-virtual-textures-really-work/
25•betamark•12h ago•23 comments

Planetary Roller Screws

https://www.humanityslastmachine.com/#planetary-roller-screws
36•everlier•3d ago•8 comments
Open in hackernews

Evaluating chain-of-thought monitorability

https://openai.com/index/evaluating-chain-of-thought-monitorability/
68•mfiguiere•1mo ago

Comments

ramoz•1mo ago
> Our expectation is that combining multiple approaches—a defense-in-depth strategy—can help cover gaps that any single method leaves exposed.

Implement hooks in codex then.

ursAxZA•1mo ago
I might be missing something here as a non-expert, but isn’t chain-of-thought essentially asking the model to narrate what it’s “thinking,” and then monitoring that narration?

That feels closer to injecting a self-report step than observing internal reasoning.

crthpl•1mo ago
the chain of thought is what it is thinking
ursAxZA•1mo ago
Chain-of-thought is a technical term in LLMs — not literally “what it’s thinking.”

As far as I understand it, it’s a generated narration conditioned by the prompt, not direct access to internal reasoning.

Bjartr•1mo ago
It is text that describes a plausible/likely thought process that conditions future generation by it's presence in the context.
CamperBob2•1mo ago
Interestingly, it doesn't always condition the final output. When playing with DeepSeek, for example, it's common to see the CoT arrive at a correct answer that the final answer doesn't reflect, and even vice versa, where a chain of faulty reasoning somehow yields the right final answer.

It almost seems that the purpose of the CoT tokens in a transformer network is to act as a computational substrate of sorts. The exact choice of tokens may not be as important as it looks, but it's important that they are present.

Workaccount2•1mo ago
IIRC Anthropic has research finding CoT can sometimes be uncorrelated with the final output.
nowittyusername•1mo ago
That phenomenon and others is what made it obvious that COT is not its "thinking". I think COT is a process by which the llm expands its processing boundary, in that it allows it to sample over a larger space of possibilities. So its kind of acts like a "trigger" of sorts that allows the model to explore in more ways then without COT. First time I saw this was when I witnessed the "wait" phenomenon. Simply inducing the model to say "wait" in its response improved accuracy of results. as now the model double checked its "work". funny enough it also sometimes lead it to produce a wrong answer where otherwise it should have stuck to its guns. But overall that little wait had a net positive affect. Thats when i knew COT was not same as human thinking as we dont care about trigger words or anything like that, our thinking requires zero language (though it does benefit from language) its a deeper process. Thats why i was interested in latent processing models and foray in that matter.
arthurcolle•1mo ago
Wrong to the point of being misleading. This is a goal, not an assumption

Source: all of mechinterp

skissane•1mo ago
When we think, our thoughts are composed of both nonverbal cognitive processes (we have access to their outputs, but generally lack introspective awareness of their inner workings), and verbalised thoughts (whether the “voice in your head” or actually spoken as “thinking out loud”).

Of course, there are no doubt significant differences between whatever LLMs are doing and whatever humans are doing when they “think” - but maybe they aren’t quite as dissimilar as many argue? In both cases, there is a mutual/circular relationship between a verbalised process and a nonverbal one (in the LLM case, the inner representations of the model)

ursAxZA•1mo ago
The analogy breaks at the learning boundary.

Humans can refine internal models from their own verbalised thoughts; LLMs cannot.

Self-generated text is not an input-strengthening signal for current architectures.

Training on a model’s own outputs produces distributional drift and mode collapse, not refinement.

Equating CoT with “inner speech” implicitly assumes a safe self-training loop that today’s systems simply don’t have.

CoT is a prompted, supervised artifact — not an introspective substrate.

skissane•1mo ago
Models have some limited means of refinement available to themselves already: augment a model with any form of external memory, and it can learn by writing to its memory and then reading relevant parts of that accumulated knowledge back in the future. Of course, this is a lot more rigid than what biological brains can do, but it isn’t nothing.

Does “distributional drift and mode collapse” still happen if the outputs are filtered with respect to some external ground truth - e.g. human preferences, or even (in certain restricted domains such as coding) automated evaluations?

ursAxZA•1mo ago
I wasn’t talking about human reinforcement.

The discussion has been about CoT in LLMs, so I’ve been referring to the model in isolation from the start.

Here’s how I currently understand the structure of the thread (apologies if I’ve misread anything):

“Is CoT actually thinking?” (my earlier comment)

→ “Yes, it is thinking.”

  → “It might be thinking.”

   → “Under that analogy, self-training on its own CoT should work — but empirically it doesn’t.”

    → “Maybe it would work if you add external memory with human or automated filtering?”
Regarding external memory:

without an external supervisor, whatever gets written into that memory is still the model’s own self-generated output — which brings us back to the original problem.

sonuhia•1mo ago
> Humans can refine internal models from their own verbalised thoughts; LLMs cannot.

can be done without limitations but you won't get the current (and absolutely fucking pointless) kind of speed.

> Self-generated text is not an input-strengthening signal for current architectures.

It can be, the architecture is not the issue. Multi-model generations used for refining answers can also be tweaked for input-strengthening via multi- and cross-stage/link (in the chain) pre-/system-prompts.

> Training on a model’s own outputs produces distributional drift and mode collapse, not refinement

That's an integral part of self-learning. Or in many cases when children raise themselves or each other. Or when hormones are blocked (micro-collapse in sub-systems) or people are drugged (drift). If you didn't have loads of textbooks and online articles, you'd collapse all the time. Some time later: AHA!

It's a "hot reloading" kind of issue but assimilation and adaptation can't/don't happen at the same time. In pure informational contexts it's also just an aggregation while in the real world and in linguistics, things change, in/out of context and based on/grounded in--potentially liminal--(sub-)cultural dogmas, subjectively, collective and objectively phenomenological. Since weighted training data is basically a censored semi-omniscient "pre-computed" botbrain, it's a schizophrenic and dissociating mob of scripted personalities by design, which makes model collapse and drift practically mandatory.

> a safe self-training loop that today’s systems simply don’t have.

Early stages are never safe and you don't get safety otherwise except if you don't have idiots around you, which in money and fame hungry industries and environments is never the case.

> CoT is a prompted, supervised artifact — not an introspective substrate.

Yeah, but their naming schemes are absolute trash in general, anchoring false associations--technically, even deliberately misleading associations or sloppy ignorant ones, desperate to equate their product with human brains--and priming for misappropriation--"it's how humans think".

jablongo•1mo ago
It is what it is thinking consciously / its internal narrative. For example a supervillain's internal narrative with their plans would go into their COT notepad. If we want to really lean into the analogy between human psychology and LLMs. The "internal reasoning" that people keep referencing in this thread.. referring to the transformer weights and inscrutable inner working of a GPT.. isn't reasoning, but more like instinct, or the subconscious.
canjobear•1mo ago
It’s more like if the supervillain had to write one word of his chain of thought, then go away and forget what he was thinking, then come back and write one more word based on what he had written so far, repeating the process until the whole chain of thought is written out. Each token is generated conditional only on the previous tokens.
catigula•1mo ago
this is not correct
ACCount37•1mo ago
Kind of. The narration is an actual part of the thinking process. Just not the only part.

It can reflect the thinking process fully, or it can be full of post hoc justifications. In practice, it's anything in between.

As task complexity increases and chain-of-thought length grows, it becomes load-bearing by necessity. It still doesn't have to be fully accurate, but it must be doing something right, or the answer wouldn't work.

leetrout•1mo ago
Related check out chain of draft if you haven't.

Similar performance with 7% of tokens as chain of thought.

https://arxiv.org/abs/2502.18600

astrange•1mo ago
That's a comparison to "CoT via prompting of chat models", not "CoT via training reasoning models with RLVR", so it may not apply.
catigula•1mo ago
This seems remarkably less safe?

Would would we want to purposely decrease interpretability?

Very strange.