frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

You Are Insignificant. That's a Good Thing

https://www.joanwestenberg.com/p/you-are-insignificant-that-s-a-good-thing
1•gpi•4m ago•0 comments

The High Price of Environmental Responsibility

https://rodgercuddington.substack.com/p/the-high-price-of-environmental-responsibility
2•freespirt•6m ago•1 comments

An LED panel that shows the aviation around you

https://github.com/AxisNimble/TheFlightWall_OSS
1•yzydserd•8m ago•0 comments

Artificial muscles, or Robotics 2.0 (RU) (Anthropomorphic robotics)

https://habr.com/ru/articles/969722/
1•chromoblob•10m ago•1 comments

iPhone Fold Will Be Creaseless and Cost $2,400, Report Says

https://www.cnet.com/tech/mobile/iphone-fold-will-be-creaseless-and-cost-2400-report-says/
1•thunderbong•11m ago•0 comments

Generative UI: A rich, custom, visual interactive user experience for any prompt

https://research.google/blog/generative-ui-a-rich-custom-visual-interactive-user-experience-for-a...
1•pramodbiligiri•11m ago•0 comments

Ask HN: Ever done large contract work?

3•ripped_britches•11m ago•0 comments

Meditation as Wakeful Relaxation: Unclenching Smooth Muscle

https://psychotechnology.substack.com/p/meditation-as-wakeful-relaxation
1•eatitraw•12m ago•0 comments

Turn Claude threads into Notion-grade assets you can trust

https://claudeai2notion.aluo.app
1•chatgpt2notion•16m ago•0 comments

Immigrations: Gestiona Tu Visa LLC exposed 67GB of immigrant's data

https://medium.com/@newschu.substack.com/immigrations-gestiona-tu-visa-llc-exposed-67gb-of-immigr...
1•khyy_•20m ago•0 comments

A Deep Dive into MCP and the Future of AI Tooling

https://a16z.com/a-deep-dive-into-mcp-and-the-future-of-ai-tooling/
1•stosssik•20m ago•0 comments

Claude Is Broken in Armenian

https://twitter.com/dyushag/status/1993143599286886525
1•ag8•20m ago•0 comments

California prosecutors used AI to file inaccurate motion in criminal case

https://www.theguardian.com/us-news/2025/nov/26/prosecutor-ai-inaccurate-motion
1•trusche•25m ago•0 comments

The Emoji Layer

https://github.com/jrecyclebin/emojilayer
1•jrecyclebin•26m ago•0 comments

Palo Alto Networks to Acquire Chronosphere (Creators of M3DB)

https://www.paloaltonetworks.com/company/press/2025/palo-alto-networks-to-acquire-chronosphere--n...
3•leo_e•27m ago•0 comments

Mixpanel Security Breach

https://mixpanel.com/blog/sms-security-incident/
3•jaredwiener•28m ago•1 comments

The Nerd Reich – Silicon Valley Fascism and the War on Democracy

https://www.simonandschuster.com/books/The-Nerd-Reich/Gil-Duran/9781668221402
4•brunohaid•37m ago•0 comments

Intellect-3: A 100B+ MoE trained with large-scale RL

https://www.primeintellect.ai/blog/intellect-3
2•meetpateltech•39m ago•0 comments

3D visualization of audio latent spaces (AI Vector Map of Audio)

https://github.com/lyramakesmusic/latent-musicvis
2•caust1c•52m ago•1 comments

Families tracked victims and survivors in real time in Hong Kong tower fire

https://www.abc.net.au/news/2025-11-27/how-families-tracked-hong-kong-fatal-fire-in-real-time-onl...
3•charlieyu1•53m ago•0 comments

Elevating Intelligence via Efficient Model and Tool Orchestration

https://arxiv.org/abs/2511.21689
2•georgehe9•54m ago•0 comments

Ask HN: How many screens do you usually work with?

1•vpaulus•57m ago•4 comments

Show HN: GitHub Activity Analytics Powered by ClickHouse

https://velocity.clickhouse.com/
2•saisrirampur•58m ago•0 comments

Agentic Learner with Grow-and-Refine Multimodal Semantic Memory

https://arxiv.org/abs/2511.21678
1•badmonster•1h ago•0 comments

Artificial 'nose' tells people when certain smells are present

https://www.science.org/content/article/artificial-nose-tells-people-when-certain-smells-are-present
3•ashishgupta2209•1h ago•0 comments

Turning old bread into flour and then into tasty tortillas

https://www.rnz.co.nz/life/food/turning-old-bread-into-flour-and-then-into-tasty-tortillas
2•colinprince•1h ago•0 comments

I run a personal IPv6 BGP network. Netflix is blocking a /64 of our /36. Why?

https://www.neelc.org/posts/netflix-blocks-our-ipv6/
5•ericdiao•1h ago•1 comments

Building a Low-Cost Satellite Tracker

https://hackaday.com/2025/11/26/building-a-low-cost-satellite-tracker/
3•toomuchtodo•1h ago•1 comments

Tesla FSD software may not be approved by EU regulator after all

https://techcrunch.com/2025/11/24/tesla-fsd-software-may-not-be-approved-by-eu-regulator-after-all/
3•gochuks•1h ago•0 comments

BazelCon 2025 Recap

https://blogsystem5.substack.com/p/bazelcon-2025-recap
1•jmmv•1h ago•0 comments
Open in hackernews

The Unreasonable Effectiveness of Reasonless Intermediate Tokens

https://arxiv.org/abs/2505.13775
4•YeGoblynQueenne•6mo ago

Comments

tocs3•6mo ago
I asked ChatGPT to restate this in more laymen's terms (posted below) and I am not to surprised at the answer.

"Lately, some AI models have shown impressive abilities to solve complex problems, and many people credit this to a method called Chain of Thought (CoT), where the model is trained to think through steps like a human might. In this paper, we take a closer look at that idea to see if it's really what's driving better performance.

We focus on the model’s step-by-step thinking (the words it generates along the way) — often treated like human "thoughts" — and examine whether these actually help the model solve problems more accurately. To test this, we train AI models using clean, correct step-by-step reasoning paths and final answers, all based on a known solving method (A* search). This lets us check both the final answers and the reasoning steps to see how they relate.

Interestingly, we find that even when a model gives the right answer, its reasoning steps can still be wrong or messy. To go further, we even train models using completely random and incorrect reasoning steps — and surprisingly, they still perform about the same, and sometimes even better, than those trained on correct steps.

This suggests that the step-by-step "thoughts" the model shows aren’t as meaningful or reliable as many assume. In short, just because a model looks like it’s reasoning through a problem doesn’t mean it actually is — and we should be careful not to treat its outputs as if it thinks like a human or follows strict logic."