frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

The Scourge of 'Spot-Fixing' Is Coming for American Sports

https://www.wsj.com/sports/baseball/spot-fixing-mlb-cricket-soccer-a573caaa
1•PaulHoule•1m ago•0 comments

Scientists hide messages in papers to game AI peer review

https://www.nature.com/articles/d41586-025-02172-y
1•signa11•1m ago•0 comments

Striatal astrocytes modulate behavioral flexibility and whole-body metabolism

https://www.nature.com/articles/s41467-025-60968-y
1•PaulHoule•4m ago•0 comments

Found a good prayer website to pray for others or request prayers

https://sendtheprayer.net
1•victor_cl•7m ago•0 comments

Boeing fuel switches safe, regulator says after Air India crash

https://www.bbc.com/news/articles/ce9xpgnx3vdo
1•maxbond•13m ago•1 comments

Building a Simple Router with OpenBSD

https://btxx.org/posts/openbsd-router/
1•transpute•20m ago•0 comments

A Minimal DDPM

https://github.com/metalwhale/minimal-ddpm
1•metalwhale•21m ago•0 comments

NixOS: Declarative Management, Imperative Privilege Escalation

https://labs.snyk.io/resources/nixos-deep-dive/
3•Bogdanp•23m ago•0 comments

Razor, Gun, Fence – Kieranhealy.org

https://kieranhealy.org/blog/archives/2025/06/26/razor-gun-fence/
3•MaysonL•29m ago•0 comments

Why the Upper Middle Class Isn't Special Anymore

https://ofdollarsanddata.com/the-death-of-the-amex-lounge/
3•deanmoriarty•32m ago•0 comments

Show HN: KVM backup script with telegram notifications

https://gist.github.com/tuxx/e1c896007b536490b98d2b261d46cf70
1•tuxxness•36m ago•0 comments

Show HN: Clippy – a better pbcopy for macOS that handles files properly

https://github.com/neilberkman/clippy
2•nberkman•39m ago•0 comments

Show HN: I built a dream interpreter in JavaScript, no AI, no server, just logic

https://github.com/Dino-Nuggies45/Dream-Interpreter
9•DinoNuggies456•47m ago•4 comments

Traceable Randomness

https://random.colorado.edu/concepts/traceable-randomness
1•owl_vision•50m ago•0 comments

Tracking Protestware Spread: 28 NPM Packages Affected by Payload Targeting

https://socket.dev/blog/protestware-update-28-npm-packages-affected-by-payload-targeting-russian-language-users
1•feross•50m ago•0 comments

China's new digital ID system raises surveillance, censorship concerns

https://www.washingtonpost.com/world/2025/07/15/china-digital-id-internet-surveillance/
3•bookofjoe•52m ago•1 comments

How Elon Musk's X is fueling the MAGA-Trump split

https://www.politico.com/news/2025/07/15/elon-musk-x-maga-00455128
4•c420•53m ago•1 comments

Ask HN: A project isn't dead just because it's quiet – how to tell people that?

3•fernvenue•58m ago•4 comments

Show HN: Turing Test Game – but with Social Deception

https://amonghumans.io
1•Kehvinbehvin•59m ago•1 comments

Asymmetry of Verification and Verifier's Law

https://www.jasonwei.net/blog/asymmetry-of-verification-and-verifiers-law
1•hasheddan•59m ago•0 comments

Grok's new porn companion is rated for kids 12 and older in the App Store

https://www.platformer.news/grok-ani-app-store-rating-nsfw-avatar-apple/
2•spenvo•1h ago•0 comments

GenAI-Powered Inference

https://arxiv.org/abs/2507.03897
1•JackeJR•1h ago•1 comments

UK fintech Curve in talks to be acquired by Lloyds

https://www.headforpoints.com/2025/07/13/lloyds-bank-in-talks-to-buy-curve/
1•gregorvand•1h ago•0 comments

AWS announced support for clusters with up to 100k nodes

https://aws.amazon.com/blogs/containers/under-the-hood-amazon-eks-ultra-scale-clusters/
4•dropbox_miner•1h ago•3 comments

World's 'oldest' marathon runner dies at 114 in hit-and-run

https://www.bbc.com/news/articles/cpqnppnx0z1o
1•layer8•1h ago•0 comments

Show HN: Tlsinfo.me – check your JA3/JA4 TLS fingerprints

https://tlsinfo.me/json
2•elpy1•1h ago•0 comments

Some Australian dolphins use sponges to hunt fish, but it's harder than it looks

https://apnews.com/article/dolphins-australia-sponge-noses-9ba412c3d0184ee84a66ec8b5a5b5319
1•c420•1h ago•1 comments

Sexting with Gemini

https://www.theatlantic.com/magazine/archive/2025/08/google-gemini-ai-sexting/683248/
1•JumpCrisscross•1h ago•2 comments

The AI That Broke the Internet's Back

https://medium.com/@th71852/the-ai-that-broke-the-internets-back-24c1bd2e825e
1•antiochIst•1h ago•0 comments

Retrieval Embedding Benchmark

https://huggingface.co/spaces/embedding-benchmark/RTEB
1•fzliu•1h ago•0 comments
Open in hackernews

Hierarchical Modeling (H-Nets)

https://cartesia.ai/blog/hierarchical-modeling
59•marviel•6h ago

Comments

marviel•6h ago
> H-Net demonstrates three important results on language modeling:

> 1. H-Nets scale better with data than state-of-the-art Transformers with BPE tokenization, while learning directly from raw bytes. This improved scaling is even more pronounced on domains without natural tokenization boundaries, like Chinese, code, and DNA.

> 2. H-Nets can be stacked together to learn from deeper hierarchies, which further improves performance.

> 3. H-Nets are significantly more robust to small perturbations in input data like casing, showing an avenue for creating models that are more robust and aligned with human reasoning.

marviel•5h ago
https://arxiv.org/pdf/2507.07955

paper

modeless•5h ago
I don't know if this is the one but something like this is clearly the future IMO. We need more levels of hierarchy to efficiently generalize to longer sequences with high level structure. Back when Byte Latent Transformers came out I thought extending the idea to more levels of hierarchy was the way to go, and this seems to be basically that?

Another article about H-Nets: https://main-horse.github.io/posts/hnet-inf/

macawfish•2h ago
Yes... This seems like a generalization of "large concept models" in a certain way
cs702•5h ago
I've only skimmed the paper, but it looks interesting and credible, so I've added it to my reading list.

Thank you for sharing on HN!

---

EDIT: The hierarchical composition and routing aspects of this work vaguely remind me of https://github.com/glassroom/heinsen_routing/ but it has been a while since I played with that. UPDATE: After spending a bit more time on the OP, it's different, but the ideas are related, like routing based on similarity.

marviel•5h ago
No problem! I'm still parsing it myself, but it seems promising in theory, and the result curves are impressive.
gdiamos•5h ago
How does it handle images?
marviel•5h ago
it mentions native multimodality somewhere in either the Arxiv or post -- seems like it might handle it well?
miven•4h ago
As far as I understand the "chunking" of input bytes is learned completely end to end, so it's basically up to the model to figure out how to most efficiently delineate and aggregate the information from the inputs according to the patterns provided to it during training.

Since it's end to end this allows them to apply this process not only to raw byte encodings but basically representations of any level, such as stacking two stages of aggregation one after another.

So in principle they could either let the model do its thing on raw bytes of an image or alternatively maybe cut it up into tiny patches ViT-style and feed that to their H-Net.

I wonder how hard would it be to adapt chunking to work in 2D and what would that even look like.

Some other notes on how multimodal inputs could be handled using this architecture are mentioned in Albert Gu's (one of the author's) blog, although only briefly, there's still much to figure out it would seem: https://goombalab.github.io/blog/2025/hnet-future/#alternati...

marviel•4h ago
Thanks for sharing this blog post is a great speculative deep-dive.
aeon_ai•5h ago
Seems likely to be relevant for memory formation/consolidation/management.

Big, if so.

cubefox•4h ago
As Mamba didn't make it, will H-Nets replace Transformers?
marviel•4h ago
It's meant to replace the BPE tokenizer piece, so it isn't a full Language Model by itself.

In fact in Gu's blog post (linked in a post below) it's mentioned that they created a Mamba model that used this in place of the tokenizer.

vannevar•3h ago
>The best AI architectures in use today treat all inputs equally.

Doesn't this architecture also treat all inputs equally? It seems like an encoder that preprocesses the input by inferring hierarchy. But don't all models essentially do that while training?

modeless•3h ago
If I understand correctly, each level of the hierarchy divides its input into chunks of variable size, but outputs a fixed amount for each chunk. The chunking is learned. The model can choose to compress data by making its input chunks bigger, depending on their content.
macawfish•1h ago
Hand wavy idea: I wonder if we couldn't take this to another level and have some kind of general graph representation along with hierarchical reductions of it.

I sort of disagree with the assertion that "language is fundamentally hierarchical" in that it supposes there is a single abstraction hierarchy that's universally preferable or correct. That's just not true. It doesn't hurt anybody and it's definitely simpler to choose just one useful one (a hierarchy) but why learn only one? Why not learn multiple and also learn how to modulate between them?