frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

Firefox moves to GitHub

https://github.com/mozilla-firefox/firefox
388•thefilmore•4h ago•228 comments

FastVLM: Efficient vision encoding for vision language models

https://github.com/apple/ml-fastvlm
259•nhod•8h ago•46 comments

A programming language made for me

https://zylinski.se/posts/a-programming-language-for-me/
16•gingerBill•1h ago•4 comments

Ask HN: How are you acquiring your first hundred users?

13•amanchanda•57m ago•12 comments

TransMLA: Multi-head latent attention is all you need

https://arxiv.org/abs/2502.07864
67•ocean_moist•6h ago•13 comments

Open Hardware Ethernet Switch project, part 1

https://serd.es/2025/05/08/Switch-project-pt1.html
140•luu•4d ago•17 comments

Anti-Personnel Computing (2023)

https://erratique.ch/writings/anti-personnel-computing
11•transpute•1h ago•3 comments

15 Years of Shader Minification

https://www.ctrl-alt-test.fr/2025/15-years-of-shader-minification/
85•laurentlb•3d ago•16 comments

The Barbican

https://arslan.io/2025/05/12/barbican-estate/
544•farslan•18h ago•201 comments

Air Traffic Control

https://computer.rip/2025-05-11-air-traffic-control.html
174•1317•1d ago•58 comments

Alephic Writing Style Guide

https://www.alephic.com/company/writing
29•otoolep•3d ago•13 comments

Revisiting Image Maps

https://css-tricks.com/revisiting-image-maps/
26•thm•3d ago•12 comments

We Fixed 2k+ Security Issues (2023)

https://www.sudhishkr.com/posts/20230609_how-we-fixed-2kplus-security-issues/
3•mooreds•1d ago•0 comments

Can you trust that permission pop-up on macOS?

https://wts.dev/posts/tcc-who/
289•nmgycombinator•15h ago•199 comments

A conversation about AI for science with Jason Pruet

https://www.lanl.gov/media/publications/1663/0125-qa-jason-pruet
148•LAsteNERD•13h ago•126 comments

Understanding LucasArts' iMUSE System

https://github.com/meshula/LabMidi/blob/main/LabMuse/imuse-technical.md
124•todsacerdoti•10h ago•26 comments

RIP Usenix ATC

https://bcantrill.dtrace.org/2025/05/11/rip-usenix-atc/
167•joecobb•17h ago•37 comments

How to avoid P hacking

https://www.nature.com/articles/d41586-025-01246-1
59•benocodes•4d ago•50 comments

A community-led fork of Organic Maps

https://www.comaps.app/news/2025-05-12/3/
301•maelito•21h ago•195 comments

HealthBench – An evaluation for AI systems and human health

https://openai.com/index/healthbench/
155•mfiguiere•15h ago•129 comments

University of Texas-led team solves a big problem for fusion energy

https://news.utexas.edu/2025/05/05/university-of-texas-led-team-solves-a-big-problem-for-fusion-energy/
245•signa11•21h ago•168 comments

Launch HN: ParaQuery (YC X25) – GPU Accelerated Spark/SQL

118•winwang•17h ago•71 comments

Reviving a modular cargo bike design from the 1930s

https://www.core77.com/posts/136773/Reviving-a-Modular-Cargo-Bike-Design-from-the-1930s
184•surprisetalk•19h ago•146 comments

Writing N-body gravity simulations code in Python

https://alvinng4.github.io/grav_sim/5_steps_to_n_body_simulation/
126•dargscisyhp•3d ago•23 comments

Trial by Fire: The crash of Aeroflot flight 1492

https://admiralcloudberg.medium.com/trial-by-fire-the-crash-of-aeroflot-flight-1492-ee61cebcf6ec
6•shmeeed•2h ago•2 comments

Ruby 3.5 Feature: Namespace on read

https://bugs.ruby-lang.org/issues/21311
206•ksec•19h ago•97 comments

Offline vs. online ML pipelines

https://decodingml.substack.com/p/offline-vs-online-ml-pipelines
4•rbanffy•3d ago•0 comments

NASA study reveals Venus crust surprise

https://science.nasa.gov/science-research/astromaterials/nasa-study-reveals-venus-crust-surprise/
79•mnem•3d ago•77 comments

FedRAMP 20x – One Month in and Moving Fast

https://www.fedramp.gov/2025-04-24-fedramp-20x-one-month-in-and-moving-fast/
80•transpute•9h ago•53 comments

Wtfis: Passive hostname, domain and IP lookup tool for non-robots

https://github.com/pirxthepilot/wtfis
96•todsacerdoti•11h ago•9 comments
Open in hackernews

TransMLA: Multi-head latent attention is all you need

https://arxiv.org/abs/2502.07864
67•ocean_moist•6h ago

Comments

olq_plo•3h ago
Very cool idea. Can't wait for converted models on HF.
kavalg•3h ago
My (possibly wrong) TLDR: TransMLA is a method to "compress" an already trained GQA model, with the additional option to further fine tune it. Shall make inference faster.
freeqaz•2h ago
Also makes models smarter ("expressive")
yorwba•2h ago
It is not a method to compress a Grouped-Query Attention model, but to expand it into an equivalent Multi-head Latent Attention model with the same key-value cache size but larger effective key/value vectors and a correspondingly larger number of trainable parameters. With additional training, you can then obtain a better model that only uses a little bit more memory.
kristel100•1h ago
Still wrapping my head around this architecture, but the idea of reducing headcount while maintaining performance is compelling. Would love to see a benchmark against something like FlashAttention.
octocop•1h ago
These titles need to stop, we've seen that in fact it is not all you need.
tankenmate•1h ago
The title of this paper is a reference to a previous paper titled "Attention Is All You Need"[0][1]. This seminal work described the transformer model that is the basis for almost all LLMs, and is almost certainly the most cited paper on AI even though it was only published in 2017.

[0] https://arxiv.org/abs/1706.03762 [1] https://en.wikipedia.org/wiki/Attention_Is_All_You_Need

kristopolous•27m ago
Right, it's an 8 year old reference that's been made hundreds of times.

People seem to love going to the references graveyard, digging up tired and dead ones and drag them around town hoping everyone thinks they're clever.

Also this was from 3 months ago.

seeknotfind•36m ago
All you need titles stopping is all you need.
EGreg•22m ago
We need more than that, and all you need to stop saying that!!
Etheryte•6m ago
All you need is love, and for these titles to stop. (But they won't do that.)
wiz21c•42m ago
Not quite related, but do the mamba models gain ground ?

Answering my own question: https://www.reddit.com/r/MachineLearning/comments/1hpg91o/d_...

EGreg•21m ago
All you need to stop posting titles like that !