frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

LLM Neuroanatomy II: Modern LLM Hacking and Hints of a Universal Language?

https://dnhkng.github.io/posts/rys-ii/
38•realberkeaslan•5h ago

Comments

JPLeRouzic•1h ago
Has anyone started to implement this technique in Llama.cpp or similar inference tool?
dnhkng•1h ago
There was some work done on this a while back, during the FrankenMerge craze of 23'

I am working with TurboDerp to integrate this into the Exllama v3 format.

_lex•1h ago
We've discovered the language. It changes the economics of computing.

As in, this entire cloud buildout is unnecessary because it becomes like using a calculator.

Reach out to chat.

cjameskeller•1h ago
Would you be willing to elaborate? I would be curious to hear more.
lostmsu•1h ago
How's the reproducibility of the results? Like avg score of 10 runs vs original.
dnhkng•46m ago
Author here: The code is up on GitHub.

The probes I used seem to help identify good configurations, but are quite noisey. A small probe set was initially used to make the scan tractable, and then the higher ranked models were retested on a set ~10x larger.

yodon•1h ago
If you look at convolutional neural nets used in image processing, it's super common for the first layer or so to learn a family of wavelet basis functions. Later layers then do recognition in wavelet space, without that space ever being explained or communicated to the training algorithm.

This work here is obviously more complex than that, but suggests something similar is going on with early layers transforming to some sort of generalized basis functions defining a universal language representation.

yodon•1h ago
Apologies if I missed this in the article (or in the first article in the series) - what happens if you add two copies of the layer set? Does performance improve over adding one copy of the layer set?
dnhkng•48m ago
Author here: That was done in this blog post, in the beam search. I started with the best re-layer configs, and iteratively added more blocks, including the same multiple times, during a long beam search.

It turns out this does not help (somewhat surprisingly).

skyde•5m ago
Actually not surprised. I guess this is for the same reason “say it twice” [1] is working. Because LLm are trained as causal language model, past token cannot attend to future token. One copy of the layer set solve this. [1]https://arxiv.org/html/2512.14982v1
saidnooneever•58m ago
it sometimes makes me think of a video at some point of a guy (Daniel Tammet) who had some brain difference,which caused him to be extremely fast at language learning. He said all language carries the same patterns for him, which he sees through synestesia or whatever.

he learnt icelandic in week and had a fluent conversation on their national TV to prove it. (this is nuts, that language is extremely difficult to pickup with nasal sounds etc.)

ofcourse i guess its not even close to average to have such a abilities as a human, but i wonder if at some point LLMs and AI algorithms and models might shed light on such kind of abstractions (like some mentioned in comments also about image recognition algos) that might help humans actually learn these things themselves, train on them and perhaps even get taught such a thing as a skill.

LiteLLM Python package compromised by supply-chain attack

https://github.com/BerriAI/litellm/issues/24512
397•theanonymousone•3h ago•175 comments

The bridge to wealth is being pulled up with AI

https://danielhomola.com/m%20&%20e/ai/your-bridge-to-wealth-is-being-pulled-up/
128•dankai•1h ago•56 comments

Major insider trading on oil detected ahead of Iran talks

https://www.wsj.com/livecoverage/stock-market-today-dow-sp-500-nasdaq-03-24-2026/card/mystery-jum...
46•psim1•16m ago•2 comments

Microsoft's "Fix" for Windows 11: Flowers After the Beating

https://www.sambent.com/microsofts-plan-to-fix-windows-11-is-gaslighting/
644•h0ek•6h ago•477 comments

Nanobrew: The fastest macOS package manager compatible with brew

https://nanobrew.trilok.ai/
53•syrusakbary•4h ago•29 comments

Debunking Zswap and Zram Myths

https://chrisdown.name/2026/03/24/zswap-vs-zram-when-to-use-what.html
97•javierhonduco•5h ago•22 comments

Secure Domain Name System (DNS) Deployment 2026 Guide [pdf]

https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-81r3.pdf
42•XzetaU8•3h ago•1 comments

Ripgrep is faster than grep, ag, git grep, ucg, pt, sift (2016)

https://burntsushi.net/ripgrep/
212•jxmorris12•9h ago•86 comments

curl > /dev/sda: How I made a Linux distro that runs wget | dd

https://astrid.tech/2026/03/24/0/curl-to-dev-sda/
98•astralbijection•5h ago•39 comments

Opera: Rewind The Web to 1996 (Opera at 30)

https://www.web-rewind.com
142•thushanfernando•7h ago•81 comments

Hypothesis, Antithesis, Synthesis

https://antithesis.com/blog/2026/hegel/
6•alpaylan•14m ago•0 comments

So where are all the AI apps?

https://www.answer.ai/posts/2026-03-12-so-where-are-all-the-ai-apps.html
146•tanelpoder•1h ago•170 comments

Box of Secrets: Discreetly modding an apartment intercom to work with Apple Home

https://www.jackhogan.me/blog/box-of-secrets/
230•jackhogan11•1d ago•79 comments

Log File Viewer for the Terminal

https://lnav.org/
246•wiradikusuma•10h ago•38 comments

io_uring, libaio performance across Linux kernels and an unexpected IOMMU trap

https://blog.ydb.tech/how-io-uring-overtook-libaio-performance-across-linux-kernels-and-an-unexpe...
15•tanelpoder•2h ago•8 comments

LLM Neuroanatomy II: Modern LLM Hacking and Hints of a Universal Language?

https://dnhkng.github.io/posts/rys-ii/
39•realberkeaslan•5h ago•11 comments

LaGuardia pilots raised safety alarms months before deadly runway crash

https://www.theguardian.com/us-news/2026/mar/24/laguardia-airplane-pilots-safety-concerns-crash
18•m_fayer•23m ago•3 comments

MSA: Memory Sparse Attention

https://github.com/EverMind-AI/MSA
61•chaosprint•3d ago•5 comments

NanoClaw Adopts OneCLI Agent Vault

https://nanoclaw.dev/blog/nanoclaw-agent-vault/
72•turntable_pride•2h ago•11 comments

iPhone 17 Pro Demonstrated Running a 400B LLM

https://twitter.com/anemll/status/2035901335984611412
677•anemll•1d ago•305 comments

Autoresearch on an old research idea

https://ykumar.me/blog/eclip-autoresearch/
399•ykumards•21h ago•87 comments

No-build, no-NPM, SSR-first JavaScript framework if you hate React, love HTML

https://qitejs.qount25.dev
94•usrbinenv•5d ago•81 comments

BIO – The Bao I/O Co-Processor

https://www.crowdsupply.com/baochip/dabao/updates/bio-the-bao-i-o-co-processor
70•hasheddan•2d ago•18 comments

A 6502 disassembler with a TUI: A modern take on Regenerator

https://github.com/ricardoquesada/regenerator2000
71•wslh•3d ago•7 comments

Missile Defense Is NP-Complete

https://smu160.github.io/posts/missile-defense-is-np-complete/
147•O3marchnative•2h ago•148 comments

The Jellies That Evolved a Different Way to Keep Time

https://www.quantamagazine.org/the-jellies-that-evolved-a-different-way-to-keep-time-20260320/
8•jyunwai•3d ago•5 comments

Dune3d: A parametric 3D CAD application

https://github.com/dune3d/dune3d
208•luu•2d ago•85 comments

Claude Code Cheat Sheet

https://cc.storyfox.cz
552•phasE89•17h ago•177 comments

FCC updates covered list to include foreign-made consumer routers

https://www.fcc.gov/document/fcc-updates-covered-list-include-foreign-made-consumer-routers
404•moonka•18h ago•275 comments

Show HN: Cq – Stack Overflow for AI coding agents

https://blog.mozilla.ai/cq-stack-overflow-for-agents/
188•peteski22•23h ago•80 comments