frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Pretraining Language Models via Neural Cellular Automata

https://hanseungwook.github.io/blog/nca-pre-pre-training/
29•shmublu•3d ago

Comments

voxleone•1h ago
Neural cellular automata are interesting because they shift learning from “predict tokens” to “model state evolution.” That feels much closer to a transition-based view of systems, where structure emerges from repeated local updates (transitions) rather than being encoded explicitly.

I'm working on a theoretical/computational framework, the Functional Universe, intended for modeling physical reality as functional state evolution. i would say it could be used to replicate your CA process. Won't link it here to signal my good faith discussing this issue - it's on my GH.

dzink•1h ago
“The long-term vision is: foundation models that acquire reasoning from fully synthetic data, then learn semantics from a small, curated corpus of natural language. This would help us build models that reason without inheriting human biases from inception.”
qsera•38m ago
I think this is a bit risky, because it assumes that all knowledge that a human posses about nature is acquired after birth.

But is that correct? I think organisms also come with a partial built in understanding of nature at birth.

jamilton•5m ago
I don’t think that assumption is being made, why do you think that? In terms of metaphor, training a model could be considered both knowledge acquired after birth and its evolution. But I don’t think it’s particularly useful to stay thinking in metaphors.
throw-qqqqq•1m ago
> I think organisms also come with a partial built in understanding of nature at birth

I agree. Most organisms are quite pre-trained: they have “instincts” and natural behaviors.

E.g. newly hatched turtles know to crawl towards the ocean immediately when they hatch. They don’t learn that on their way.

It seems to me that most lifeforms come into this world pre-trained.

benob•1m ago
Reminds me of "Universal pre-training by iterated random computation" https://arxiv.org/pdf/2506.20057, with bit less formal approach.

I wonder if there is a closed-form solution for those kinds of initialization methods (call them pre-training if you wish). A solution that would allow attention heads to detect a variety of diverse patterns, yet more structured than random init.

Astral to Join OpenAI

https://astral.sh/blog/openai
276•ibraheemdev•1h ago•107 comments

OpenBSD: PF queues break the 4 Gbps barrier

https://undeadly.org/cgi?action=article;sid=20260319125859
26•defrost•38m ago•6 comments

Juggalo Makeup Blocks Facial Recognition Technology (2019)

https://consequence.net/2019/07/juggalo-makeup-facial-recognition/
59•speckx•1h ago•18 comments

Afroman found not liable in defamation case brought by Ohio cops who raided home

https://nypost.com/2026/03/18/us-news/afroman-found-not-liable-in-bizarre-ohio-defamation-case/
520•antonymoose•4h ago•151 comments

Consensus Board Game

https://matklad.github.io/2026/03/19/consensus-board-game.html
5•surprisetalk•14m ago•0 comments

Conway's Game of Life, in real life

https://lcamtuf.substack.com/p/conways-game-of-life-in-real-life
226•surprisetalk•10h ago•59 comments

'Your Frustration Is the Product'

https://daringfireball.net/2026/03/your_frustration_is_the_product
116•llm_nerd•2h ago•44 comments

Afroman Wins Civil Trial over Use of Police Raid Footage in His Music Videos

https://www.nytimes.com/2026/03/19/us/afroman-trial-lemon-cake-verdict.html
137•pseudolus•2h ago•17 comments

How many branches can your CPU predict?

https://lemire.me/blog/2026/03/18/how-many-branches-can-your-cpu-predict/
38•ibobev•1h ago•3 comments

Pretraining Language Models via Neural Cellular Automata

https://hanseungwook.github.io/blog/nca-pre-pre-training/
29•shmublu•3d ago•6 comments

OpenAI to Acquire Astral

https://openai.com/index/openai-to-acquire-astral/
107•meetpateltech•1h ago•51 comments

Nvidia greenboost: transparently extend GPU VRAM using system RAM/NVMe

https://gitlab.com/IsolatedOctopi/nvidia_greenboost
411•mmastrac•3d ago•111 comments

Warranty Void If Regenerated

https://nearzero.software/p/warranty-void-if-regenerated
435•Stwerner•17h ago•261 comments

OpenRocket

https://openrocket.info/
648•zeristor•4d ago•112 comments

Iran war energy shock sparks global push to reduce fossil fuel dependence

https://www.reuters.com/business/energy/iran-war-energy-shock-sparks-global-push-reduce-fossil-fu...
159•geox•3h ago•164 comments

Eniac, the First General-Purpose Digital Computer, Turns 80

https://spectrum.ieee.org/eniac-80-ieee-milestone
56•baruchel•8h ago•24 comments

Stdwin: Standard window interface by Guido Van Rossum [pdf]

https://ir.cwi.nl/pub/5998/5998D.pdf
55•ivanbelenky•1d ago•30 comments

Austin’s surge of new housing construction drove down rents

https://www.pew.org/en/research-and-analysis/articles/2026/03/18/austins-surge-of-new-housing-con...
643•matthest•14h ago•764 comments

Gluon: Explicit Performance

https://www.lei.chat/posts/gluon-explicit-performance/
3•matt_d•2d ago•0 comments

The next fight over the use of facial recognition could be in the supermarkets

https://www.politico.com/newsletters/digital-future-daily/2026/03/16/the-facial-recognition-groce...
16•speckx•1h ago•2 comments

LotusNotes

https://computer.rip/2026-03-14-lotusnotes.html
127•TMWNN•4d ago•68 comments

A sufficiently detailed spec is code

https://haskellforall.com/2026/03/a-sufficiently-detailed-spec-is-code
482•signa11•11h ago•250 comments

Show HN: Duplicate 3 layers in a 24B LLM, logical deduction .22→.76. No training

https://github.com/alainnothere/llm-circuit-finder
179•xlayn•16h ago•66 comments

2% of ICML papers desk rejected because the authors used LLM in their reviews

https://blog.icml.cc/2026/03/18/on-violations-of-llm-review-policies/
132•sergdigon•4h ago•125 comments

Wander – A tiny, decentralised tool to explore the small web

https://susam.net/wander/
311•susam•1d ago•77 comments

Autoresearch for SAT Solvers

https://github.com/iliazintchenko/agent-sat
147•chaisan•13h ago•29 comments

Nvidia NemoClaw

https://github.com/NVIDIA/NemoClaw
343•hmokiguess•22h ago•225 comments

The math that explains why bell curves are everywhere

https://www.quantamagazine.org/the-math-that-explains-why-bell-curves-are-everywhere-20260316/
166•ibobev•2d ago•94 comments

Show HN: I built 48 lightweight SVG backgrounds you can copy/paste

https://www.svgbackgrounds.com/set/free-svg-backgrounds-and-patterns/
325•visiwig•22h ago•63 comments

Cook: A simple CLI for orchestrating Claude Code

https://rjcorwin.github.io/cook/
253•staticvar•12h ago•73 comments