frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Tuna-2: Pixel Embeddings Beat Vision Encoders for Understanding, Generation

https://github.com/facebookresearch/tuna-2
1•gmays•50s ago•0 comments

ClickFix Campaign: I Searched for Tailscale and Found a Phishing Operation

https://perezbox.com/articles/clickfix-campaign-tailscale-nine-month-phishing-operation
1•sucuri2•2m ago•0 comments

Local LLM Proxy: Turn Idle LLM Compute into Universal Credits

https://github.com/wink-run/local-llm-proxy
1•zwling8•2m ago•0 comments

Kettle: Attested Builds for Verifiable Software Provenance

https://arxiv.org/abs/2605.08363
1•badcryptobitch•4m ago•0 comments

Sam Altman's Business Dealings Under GOP Scrutiny Ahead of OpenAI's IPO

https://www.wsj.com/tech/ai/sam-altmans-business-dealings-under-gop-scrutiny-ahead-of-openais-ipo...
3•JumpCrisscross•10m ago•1 comments

Productivity Isn't About Going Faster

https://humanpro.co/articles/productivity-isnt-about-going-faster/
2•gx•12m ago•0 comments

End-to-end encrypted RCS messaging rolling out for Android and iPhone users

https://blog.google/products-and-platforms/platforms/android/android-ios-end-to-end-encrypted-rcs...
3•thunderbong•14m ago•0 comments

Elon Musk's Grok Is Losing Ground in AI Race

https://www.wsj.com/tech/ai/anthropic-spacex-ai-deal-elon-musk-f86ea369
2•JumpCrisscross•15m ago•1 comments

MCP-Guardrail

https://github.com/kaarhsg-cpu/mcp-guardrail
1•karthikhermes•17m ago•0 comments

Skill1: Unified Evolution of Skill-Augmented Agents via Reinforcement Learning

https://arxiv.org/abs/2605.06130
1•lexandstuff•21m ago•0 comments

“Something rather unusual is happening in Russia” – My BBC Radio Dispatch

https://www.youtube.com/watch?v=K9su3iLw9hU
2•consumer451•27m ago•0 comments

Tofersen, a New Treatment for ALS, Reverses Symptoms for Some

https://www.nytimes.com/2026/05/06/well/als-treatment-tofersen-qalsody.html
1•MaysonL•32m ago•0 comments

Ask HN: Can NPM postinstall script be run in a sandbox? how to do that?

1•imWildCat•33m ago•0 comments

Cloudflare Is Down

6•sammy2255•33m ago•4 comments

Instructure Pays Ransom to Canvas Hackers

https://www.insidehighered.com/news/tech-innovation/administrative-tech/2026/05/11/instructure-pa...
1•Cider9986•45m ago•0 comments

Trying San Francisco's most experimental depression treatments

https://www.theguardian.com/society/2026/may/11/silicon-valley-depression-wellness-boom
2•andsoitis•46m ago•0 comments

Earthworms Don't Bio-Accumulate Microplastics, So There May Be Hope for Us

https://hackaday.com/2026/05/05/earthworms-dont-bio-accumulate-microplastics-so-there-may-be-hope...
1•lxm•49m ago•0 comments

Tesla's New AI Tech Could Save Lives, Thanks to Quicker Airbags

https://www.caranddriver.com/news/a71269172/tesla-ai-tech-airbag-deployment/
1•RickJWagner•49m ago•0 comments

RegexPSPACE: Regex LLM Benchmark

https://arxiv.org/abs/2510.09227
1•thatxliner•59m ago•0 comments

It has a few rough edges but it works and it's free

https://github.com/slucerodev/ExoArmur-Core
1•slucerodev•1h ago•0 comments

Show HN: Sigmashake Desktop – AI Coding Agent Guardrails

https://sigmashake.com/desktop
1•cavalrytactics•1h ago•1 comments

Why Everyone's Picking Up a PSP Again in 2026

https://gardinerbryant.com/psp-in-2026/
2•Kate0CoolLibby•1h ago•0 comments

America is experiencing a productivity miracle

https://www.economist.com/finance-and-economics/2026/05/11/america-is-experiencing-a-productivity...
2•simonpure•1h ago•0 comments

Software Internals Book Club

https://eatonphil.com/bookclub.html
8•aragonite•1h ago•0 comments

Fake building: Claude wrote 3k lines instead of import pywikibot

https://fireflysentinel.github.io/posts/fake-building-claude-3000-lines/
30•firef1y1203•1h ago•12 comments

Notes from inside China's AI labs

https://www.interconnects.ai/p/notes-from-inside-chinas-ai-labs
2•gmays•1h ago•1 comments

Mona Lisa's Identity Established Beyond Doubt (2008)

https://www.uni-heidelberg.de/press/news/press552e.html
1•thunderbong•1h ago•0 comments

Microsoft fires head of Israeli subsidiary over surveillance of Palestinians

https://www.pcgamer.com/gaming-industry/microsoft-fires-head-of-israeli-subsidiary-and-other-mana...
17•donohoe•1h ago•2 comments

Floci: Light, fluffy, and always free – The AWS Local Emulator alternative

https://github.com/floci-io/floci
1•linkdd•1h ago•0 comments

A consistent pattern of lying': trial exposes what insiders think of Sam Altman

https://www.theguardian.com/us-news/2026/may/11/musk-v-openai-altman-trial
9•andsoitis•1h ago•0 comments
Open in hackernews

The Unreasonable Effectiveness of Reasonless Intermediate Tokens

https://arxiv.org/abs/2505.13775
4•YeGoblynQueenne•11mo ago

Comments

tocs3•11mo ago
I asked ChatGPT to restate this in more laymen's terms (posted below) and I am not to surprised at the answer.

"Lately, some AI models have shown impressive abilities to solve complex problems, and many people credit this to a method called Chain of Thought (CoT), where the model is trained to think through steps like a human might. In this paper, we take a closer look at that idea to see if it's really what's driving better performance.

We focus on the model’s step-by-step thinking (the words it generates along the way) — often treated like human "thoughts" — and examine whether these actually help the model solve problems more accurately. To test this, we train AI models using clean, correct step-by-step reasoning paths and final answers, all based on a known solving method (A* search). This lets us check both the final answers and the reasoning steps to see how they relate.

Interestingly, we find that even when a model gives the right answer, its reasoning steps can still be wrong or messy. To go further, we even train models using completely random and incorrect reasoning steps — and surprisingly, they still perform about the same, and sometimes even better, than those trained on correct steps.

This suggests that the step-by-step "thoughts" the model shows aren’t as meaningful or reliable as many assume. In short, just because a model looks like it’s reasoning through a problem doesn’t mean it actually is — and we should be careful not to treat its outputs as if it thinks like a human or follows strict logic."