frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Goto Considered Awesome [video]

https://www.youtube.com/watch?v=1UKVEUGEk6Y
1•linkdd•1m ago•0 comments

Show HN: I Built a Free AI LinkedIn Carousel Generator

https://carousel-ai.intellisell.ai/
1•troyethaniel•2m ago•0 comments

Implementing Auto Tiling with Just 5 Tiles

https://www.kyledunbar.dev/2026/02/05/Implementing-auto-tiling-with-just-5-tiles.html
1•todsacerdoti•3m ago•0 comments

Open Challange (Get all Universities involved

https://x.com/i/grok/share/3513b9001b8445e49e4795c93bcb1855
1•rwilliamspbgops•4m ago•0 comments

Apple Tried to Tamper Proof AirTag 2 Speakers – I Broke It [video]

https://www.youtube.com/watch?v=QLK6ixQpQsQ
2•gnabgib•6m ago•0 comments

Show HN: Vibe as a Code / VaaC – new approach to vibe coding

https://www.npmjs.com/package/@gace/vaac
1•bstrama•7m ago•0 comments

Show HN: More beautiful and usable Hacker News

https://twitter.com/shivamhwp/status/2020125417995436090
3•shivamhwp•8m ago•0 comments

Toledo Derailment Rescue [video]

https://www.youtube.com/watch?v=wPHh5yHxkfU
1•samsolomon•10m ago•0 comments

War Department Cuts Ties with Harvard University

https://www.war.gov/News/News-Stories/Article/Article/4399812/war-department-cuts-ties-with-harva...
5•geox•14m ago•0 comments

Show HN: LocalGPT – A local-first AI assistant in Rust with persistent memory

https://github.com/localgpt-app/localgpt
1•yi_wang•14m ago•0 comments

A Bid-Based NFT Advertising Grid

https://bidsabillion.com/
1•chainbuilder•18m ago•1 comments

AI readability score for your documentation

https://docsalot.dev/tools/docsagent-score
1•fazkan•25m ago•0 comments

NASA Study: Non-Biologic Processes Don't Explain Mars Organics

https://science.nasa.gov/blogs/science-news/2026/02/06/nasa-study-non-biologic-processes-dont-ful...
2•bediger4000•29m ago•2 comments

I inhaled traffic fumes to find out where air pollution goes in my body

https://www.bbc.com/news/articles/c74w48d8epgo
2•dabinat•29m ago•0 comments

X said it would give $1M to a user who had previously shared racist posts

https://www.nbcnews.com/tech/internet/x-pays-1-million-prize-creator-history-racist-posts-rcna257768
4•doener•32m ago•1 comments

155M US land parcel boundaries

https://www.kaggle.com/datasets/landrecordsus/us-parcel-layer
2•tjwebbnorfolk•36m ago•0 comments

Private Inference

https://confer.to/blog/2026/01/private-inference/
2•jbegley•39m ago•1 comments

Font Rendering from First Principles

https://mccloskeybr.com/articles/font_rendering.html
1•krapp•42m ago•0 comments

Show HN: Seedance 2.0 AI video generator for creators and ecommerce

https://seedance-2.net
1•dallen97•47m ago•0 comments

Wally: A fun, reliable voice assistant in the shape of a penguin

https://github.com/JLW-7/Wally
2•PaulHoule•48m ago•0 comments

Rewriting Pycparser with the Help of an LLM

https://eli.thegreenplace.net/2026/rewriting-pycparser-with-the-help-of-an-llm/
2•y1n0•50m ago•0 comments

Lobsters Vibecoding Challenge

https://gist.github.com/MostAwesomeDude/bb8cbfd005a33f5dd262d1f20a63a693
2•tolerance•50m ago•0 comments

E-Commerce vs. Social Commerce

https://moondala.one/
1•HamoodBahzar•50m ago•1 comments

Avoiding Modern C++ – Anton Mikhailov [video]

https://www.youtube.com/watch?v=ShSGHb65f3M
2•linkdd•52m ago•0 comments

Show HN: AegisMind–AI system with 12 brain regions modeled on human neuroscience

https://www.aegismind.app
2•aegismind_app•56m ago•1 comments

Zig – Package Management Workflow Enhancements

https://ziglang.org/devlog/2026/#2026-02-06
1•Retro_Dev•57m ago•0 comments

AI-powered text correction for macOS

https://taipo.app/
1•neuling•1h ago•1 comments

AppSecMaster – Learn Application Security with hands on challenges

https://www.appsecmaster.net/en
1•aqeisi•1h ago•1 comments

Fibonacci Number Certificates

https://www.johndcook.com/blog/2026/02/05/fibonacci-certificate/
2•y1n0•1h ago•0 comments

AI Overviews are killing the web search, and there's nothing we can do about it

https://www.neowin.net/editorials/ai-overviews-are-killing-the-web-search-and-theres-nothing-we-c...
5•bundie•1h ago•1 comments
Open in hackernews

The Fractured Entangled Representation Hypothesis

https://github.com/akarshkumar0101/fer
59•akarshkumar0101•8mo ago

Comments

akarshkumar0101•8mo ago
Tweet: https://x.com/kenneth0stanley/status/1924650124829196370 Arxiv: https://arxiv.org/abs/2505.11581
pvg•8mo ago
Sounds like you're one of the co-authors? Probably worth mentioning if the case so people know they can discuss the work with one of the work-doers.
akarshkumar0101•8mo ago
I mentioned that in the original post, but I don't see that text here anymore (thats why I added links via comment)... I am new to hackernews
messe•8mo ago
I believe they just mean that you should edit the comment where you added the links to mention that you are the author, to add that additional context.
pvg•8mo ago
I just meant 'it's good for people to know one of the authors is in the thread because it makes for more interesting conversation'. Clearly did not figure out how to do that without starting a bunch of meta!
macintux•8mo ago
I believe this could (or should) have been a Show HN, which would have allowed you to include explanatory text. See the top of this page for the rules.

https://news.ycombinator.com/show

Welcome to the site. There are a lot of features which are less obvious, which you’ll discover over time.

pvg•8mo ago
Reading material usually can't be a Show HN but you can just post your work without that and say you're involved.
macintux•8mo ago
The repo includes runnable code.

> Show HN is for something you've made that other people can play with… On topic: things people can run on their computers or hold in their hands

pvg•8mo ago
A lot of writing includes runnable code and isn't a Show HN. It's a comparatively narrow category.
ipunchghosts•8mo ago
I am interested in doing research like this. Is there any way I can be a part of it or a similar group? I have been fighting for funding from DoD for many years but to no avail so I largely have to do this research on my own time or solve my current grant's problems so that i can work on this. In my mind, this kind of research is the most interesting and important right now in the deep learning field. I am a hard worker and a high-throughput thinking... how can i get connected to otherwise with a similar mindset?
scarmig•8mo ago
Did you investigate other search processes besides SGD? I'm thinking of those often termed "biologically plausible" (e.g. forward-forward, FA). Are their internal representations closer to the fractured or unified representations?
timewizard•8mo ago
> Much of the excitement in modern AI is driven by the observation that scaling up existing systems leads to better performance.

Scaling up almost always leads to better performance. If you're only getting linear gains though then there is absolutely nothing to be excited about. You are in a dead end.

goldemerald•8mo ago
This is an interesting line of research but missing a key aspect: there's (almost) no references to the linear representation hypothesis. Much work on neural network interpretability lately has shown individual neurons are polysemantic, and therefore practically useless for explainability. My hypothesis is fitting linear probes (or a sparse autoencoder) would reveal linearly semantic attributes.

It is unfortunate because they briefly mention Neel Nanda's Othello experiments, but not the wide array of experiments like the NeurIPS Oral "Linear Representation Hypothesis in Language Models" or even golden gate Claude.

ipunchghosts•8mo ago
Is what your saying imply that there is a rotation matrix you can apply to each activation output to make it less entangled?
goldemerald•8mo ago
Not quite. For an underlying semantic concept (e.g., smiling face), you can go from a basis vector [0,1,0,...,0] to the original latent space via a single rotation. You could then induce said concept by manipulating the original latent point by traversing along that linear direction.
ipunchghosts•8mo ago
I think we are saying the same thing. Please correct me though where I am wrong. You could look at the maps in some way but instead of the basis being one hot dimensions (the standard basis), it could be rotated.
akarshkumar0101•8mo ago
We mention this issue exactly in the fourth paragraph in Section 4 and in Appendix F!
akarshkumar0101•8mo ago
We mention this issue exactly in the fourth paragraph in Section 4 and in Appendix F!
goldemerald•8mo ago
That is addressing the incomprehensibility of PCA and applying a transformation to the entire latent space. I've never found PCA to be meaningful for deep learning. As far as I can tell, polysemous issue with neurons cannot be addressed with a single linear transformation. There is no sparse analysis (via linear probes or SAEs) and hence the unaddressed issue.
ipunchghosts•8mo ago
I am glad they evaluated this hypothesis using weight decay which is primarily thought of to induce a structured representation. My first thought was that the entire paper was useless if they didn't do this experiment.

I find it rather interesting that the structured representations go from sparse to full to sparse as a function of layer depth. I have noticed that applying weight decay penalty as an exponential function of layer depth gives improved results over using a global weight decay.

cwmoore•8mo ago
Isn't this simply mirroronic gravitation?
light_hue_1•8mo ago
"I looked at the representations of a network and I don't like them".

Great! There's no mathematical definition of what a fractured representation is. It's whatever art preferences you have.

Our personal preferences aren't a good predictor of which network will work well. We wasted decades with classical AI and graphical models encoding our aesthetic into models. Just to find out that the results are totally worthless.

Can we stop please? I get it. I too like beautiful things. But we can't hold on to things that don't work. Entire fields like linguistics are dying because they refuse to abandon this nonsense.