frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Portable C Compiler

https://en.wikipedia.org/wiki/Portable_C_Compiler
1•guerrilla•2m ago•0 comments

Show HN: Kokki – A "Dual-Core" System Prompt to Reduce LLM Hallucinations

1•Ginsabo•2m ago•0 comments

Software Engineering Transformation 2026

https://mfranc.com/blog/ai-2026/
1•michal-franc•3m ago•0 comments

Microsoft purges Win11 printer drivers, devices on borrowed time

https://www.tomshardware.com/peripherals/printers/microsoft-stops-distrubitng-legacy-v3-and-v4-pr...
2•rolph•4m ago•0 comments

Lunch with the FT: Tarek Mansour

https://www.ft.com/content/a4cebf4c-c26c-48bb-82c8-5701d8256282
1•hhs•7m ago•0 comments

Old Mexico and her lost provinces (1883)

https://www.gutenberg.org/cache/epub/77881/pg77881-images.html
1•petethomas•10m ago•0 comments

'AI' is a dick move, redux

https://www.baldurbjarnason.com/notes/2026/note-on-debating-llm-fans/
2•cratermoon•12m ago•0 comments

The source code was the moat. But not anymore

https://philipotoole.com/the-source-code-was-the-moat-no-longer/
1•otoolep•12m ago•0 comments

Does anyone else feel like their inbox has become their job?

1•cfata•12m ago•0 comments

An AI model that can read and diagnose a brain MRI in seconds

https://www.michiganmedicine.org/health-lab/ai-model-can-read-and-diagnose-brain-mri-seconds
1•hhs•15m ago•0 comments

Dev with 5 of experience switched to Rails, what should I be careful about?

1•vampiregrey•17m ago•0 comments

AlphaFace: High Fidelity and Real-Time Face Swapper Robust to Facial Pose

https://arxiv.org/abs/2601.16429
1•PaulHoule•18m ago•0 comments

Scientists discover “levitating” time crystals that you can hold in your hand

https://www.nyu.edu/about/news-publications/news/2026/february/scientists-discover--levitating--t...
1•hhs•20m ago•0 comments

Rammstein – Deutschland (C64 Cover, Real SID, 8-bit – 2019) [video]

https://www.youtube.com/watch?v=3VReIuv1GFo
1•erickhill•21m ago•0 comments

Tell HN: Yet Another Round of Zendesk Spam

1•Philpax•21m ago•0 comments

Postgres Message Queue (PGMQ)

https://github.com/pgmq/pgmq
1•Lwrless•25m ago•0 comments

Show HN: Django-rclone: Database and media backups for Django, powered by rclone

https://github.com/kjnez/django-rclone
1•cui•28m ago•1 comments

NY lawmakers proposed statewide data center moratorium

https://www.niagara-gazette.com/news/local_news/ny-lawmakers-proposed-statewide-data-center-morat...
1•geox•29m ago•0 comments

OpenClaw AI chatbots are running amok – these scientists are listening in

https://www.nature.com/articles/d41586-026-00370-w
2•EA-3167•29m ago•0 comments

Show HN: AI agent forgets user preferences every session. This fixes it

https://www.pref0.com/
6•fliellerjulian•31m ago•0 comments

Introduce the Vouch/Denouncement Contribution Model

https://github.com/ghostty-org/ghostty/pull/10559
2•DustinEchoes•33m ago•0 comments

Show HN: SSHcode – Always-On Claude Code/OpenCode over Tailscale and Hetzner

https://github.com/sultanvaliyev/sshcode
1•sultanvaliyev•34m ago•0 comments

Microsoft appointed a quality czar. He has no direct reports and no budget

https://jpcaparas.medium.com/microsoft-appointed-a-quality-czar-he-has-no-direct-reports-and-no-b...
2•RickJWagner•35m ago•0 comments

Multi-agent coordination on Claude Code: 8 production pain points and patterns

https://gist.github.com/sigalovskinick/6cc1cef061f76b7edd198e0ebc863397
1•nikolasi•36m ago•0 comments

Washington Post CEO Will Lewis Steps Down After Stormy Tenure

https://www.nytimes.com/2026/02/07/technology/washington-post-will-lewis.html
13•jbegley•37m ago•3 comments

DevXT – Building the Future with AI That Acts

https://devxt.com
2•superpecmuscles•37m ago•4 comments

A Minimal OpenClaw Built with the OpenCode SDK

https://github.com/CefBoud/MonClaw
1•cefboud•38m ago•0 comments

The silent death of Good Code

https://amit.prasad.me/blog/rip-good-code
3•amitprasad•38m ago•0 comments

The Internal Negotiation You Have When Your Heart Rate Gets Uncomfortable

https://www.vo2maxpro.com/blog/internal-negotiation-heart-rate
1•GoodluckH•39m ago•0 comments

Show HN: Glance – Fast CSV inspection for the terminal (SIMD-accelerated)

https://github.com/AveryClapp/glance
2•AveryClapp•40m ago•0 comments
Open in hackernews

Transformers are like great eyes, while Recurrent models are like a stomach

1•MrPan•2w ago
I’ve been training two small models on a classic long novel (Hongloumeng, 2.6M bytes) to see how they "learn" a story. After a few days of watching the logs, I noticed something interesting about where Transformers struggle.

The "Goldfish" Problem The Transformer is incredibly fast at learning how to finish a sentence. Because it uses "Attention," it’s like a student with perfect short-term memory. But it is "blind" to the long run. It only sees 128 characters at a time. It has no way to remember the beginning of the book while it's reading the end.

The Crossover My "Infinite Brain" model (a recurrent architecture) started out much worse. It was confused and the output was garbage. But around the 5th time reading the book, it "crossed over" and started beating the Transformer.

Because the Brain carries a small "memory state" forward forever, it eventually builds a "vibe" of the whole book that the Transformer just can't see.

What I learned:

Transformers are like great eyes: They see the immediate details perfectly.

Recurrent models are like a stomach: They digest the whole thing slowly, but they keep the "nutrients" of the story for much longer.

It’s a small toy experiment, but it reminded me that while Attention is powerful, having a persistent "soul" or memory state still matters for long-form data.

Loaded 2634700 bytes. Each batch chunk: 82334 Total steps to read book once: 643 Step 20 | Brain: 5.563 | Trans: 4.369 SAMPLE: 黛玉|)'Q:��t�д��T*��䎰�"��-��H� Step 40 | Brain: 5.397 | Trans: 3.469 SAMPLE: 黛玉C9j������%�����H�)���IF� Step 60 | Brain: 5.256 | Trans: 3.181 SAMPLE: 黛玉���f�XݪsʇGa���7��K��|)[o Step 80 | Brain: 5.107 | Trans: 3.015 � ����EU) 玉�����.��-n�������: Step 100 | Brain: 4.925 | Trans: 2.891 SAMPLE: 黛玉���5��X3�"��䜍��r��V:ی��;(�

......

Step 5540 | Brain: 2.057 | Trans: 2.327 SAMPLE: 黛玉,只賈王夫白說.這毫 Step 5560 | Brain: 2.079 | Trans: 2.354 SAMPLE: 黛玉了一大人,然時眾不要 Step 5580 | Brain: 2.112 | Trans: 2.381 SAMPLE: 黛玉姐的個 坐叫來.不著� Step 5600 | Brain: 2.115 | Trans: 2.394 SAMPLE: 黛玉天豥太人說䟥這撆, � Step 5620 | Brain: 2.026 | Trans: 2.290 SAMPLE: 黛玉奶歉打.命盞嚄看梅聴 Step 5640 | Brain: 2.134 | Trans: 2.414 SAMPLE: 黛玉圉起不又尚不什了,政 Step 5660 | Brain: 2.071 | Trans: 2.331 SAMPLE: 黛玉怎我古搬親就徴上一就 Step 5680 | Brain: 2.127 | Trans: 2.361 SAMPLE: 黛玉,,越于頭姐姒實眒頓 Step 5700 | Brain: 2.175 | Trans: 2.436 SAMPLE: 黛玉的你 又罜家中還太歉� Step 5720 | Brain: 2.127 | Trans: 2.348 SAMPLE: 黛玉半,他面是以. "釳有� Step 5740 | Brain: 2.141 | Trans: 2.394 SAMPLE: 黛玉,是難隄日既.阯要听 Step 5760 | Brain: 2.170 | Trans: 2.429 SAMPLE: 黛玉寶便吃,有了。”凌的 Step 5780 | Brain: 2.145 | Trans: 2.378 SAMPLE: 黛玉那好,乴日不拍說賈拿 Finished Read #9 Step 5800 | Brain: 2.092 | Trans: 2.376 SAMPLE: 黛玉人來,姐气有你兄話. Step 5820 | Brain: 2.058 | Trans: 2.321 SAMPLE: 黛玉了.政遚的�,著嬑 家

https://github.com/MrPan2048/GeometricTransformer