frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Do you have a mathematically attractive face?

https://www.doimog.com
1•a_n•2m ago•1 comments

Code only says what it does

https://brooker.co.za/blog/2020/06/23/code.html
1•logicprog•8m ago•0 comments

The success of 'natural language programming'

https://brooker.co.za/blog/2025/12/16/natural-language.html
1•logicprog•8m ago•0 comments

The Scriptovision Super Micro Script video titler is almost a home computer

http://oldvcr.blogspot.com/2026/02/the-scriptovision-super-micro-script.html
2•todsacerdoti•8m ago•0 comments

Discovering the "original" iPhone from 1995 [video]

https://www.youtube.com/watch?v=7cip9w-UxIc
1•fortran77•10m ago•0 comments

Psychometric Comparability of LLM-Based Digital Twins

https://arxiv.org/abs/2601.14264
1•PaulHoule•11m ago•0 comments

SidePop – track revenue, costs, and overall business health in one place

https://www.sidepop.io
1•ecaglar•14m ago•1 comments

The Other Markov's Inequality

https://www.ethanepperly.com/index.php/2026/01/16/the-other-markovs-inequality/
1•tzury•15m ago•0 comments

The Cascading Effects of Repackaged APIs [pdf]

https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6055034
1•Tejas_dmg•17m ago•0 comments

Lightweight and extensible compatibility layer between dataframe libraries

https://narwhals-dev.github.io/narwhals/
1•kermatt•20m ago•0 comments

Haskell for all: Beyond agentic coding

https://haskellforall.com/2026/02/beyond-agentic-coding
2•RebelPotato•24m ago•0 comments

Dorsey's Block cutting up to 10% of staff

https://www.reuters.com/business/dorseys-block-cutting-up-10-staff-bloomberg-news-reports-2026-02...
2•dev_tty01•26m ago•0 comments

Show HN: Freenet Lives – Real-Time Decentralized Apps at Scale [video]

https://www.youtube.com/watch?v=3SxNBz1VTE0
1•sanity•28m ago•1 comments

In the AI age, 'slow and steady' doesn't win

https://www.semafor.com/article/01/30/2026/in-the-ai-age-slow-and-steady-is-on-the-outs
1•mooreds•35m ago•1 comments

Administration won't let student deported to Honduras return

https://www.reuters.com/world/us/trump-administration-wont-let-student-deported-honduras-return-2...
1•petethomas•35m ago•0 comments

How were the NIST ECDSA curve parameters generated? (2023)

https://saweis.net/posts/nist-curve-seed-origins.html
2•mooreds•36m ago•0 comments

AI, networks and Mechanical Turks (2025)

https://www.ben-evans.com/benedictevans/2025/11/23/ai-networks-and-mechanical-turks
1•mooreds•36m ago•0 comments

Goto Considered Awesome [video]

https://www.youtube.com/watch?v=1UKVEUGEk6Y
1•linkdd•39m ago•0 comments

Show HN: I Built a Free AI LinkedIn Carousel Generator

https://carousel-ai.intellisell.ai/
1•troyethaniel•40m ago•0 comments

Implementing Auto Tiling with Just 5 Tiles

https://www.kyledunbar.dev/2026/02/05/Implementing-auto-tiling-with-just-5-tiles.html
1•todsacerdoti•41m ago•0 comments

Open Challange (Get all Universities involved

https://x.com/i/grok/share/3513b9001b8445e49e4795c93bcb1855
1•rwilliamspbgops•42m ago•0 comments

Apple Tried to Tamper Proof AirTag 2 Speakers – I Broke It [video]

https://www.youtube.com/watch?v=QLK6ixQpQsQ
2•gnabgib•44m ago•0 comments

Show HN: Isolating AI-generated code from human code | Vibe as a Code

https://www.npmjs.com/package/@gace/vaac
1•bstrama•45m ago•0 comments

Show HN: More beautiful and usable Hacker News

https://twitter.com/shivamhwp/status/2020125417995436090
3•shivamhwp•46m ago•0 comments

Toledo Derailment Rescue [video]

https://www.youtube.com/watch?v=wPHh5yHxkfU
1•samsolomon•48m ago•0 comments

War Department Cuts Ties with Harvard University

https://www.war.gov/News/News-Stories/Article/Article/4399812/war-department-cuts-ties-with-harva...
9•geox•51m ago•1 comments

Show HN: LocalGPT – A local-first AI assistant in Rust with persistent memory

https://github.com/localgpt-app/localgpt
5•yi_wang•52m ago•0 comments

A Bid-Based NFT Advertising Grid

https://bidsabillion.com/
1•chainbuilder•56m ago•1 comments

AI readability score for your documentation

https://docsalot.dev/tools/docsagent-score
1•fazkan•1h ago•0 comments

NASA Study: Non-Biologic Processes Don't Explain Mars Organics

https://science.nasa.gov/blogs/science-news/2026/02/06/nasa-study-non-biologic-processes-dont-ful...
3•bediger4000•1h ago•2 comments
Open in hackernews

Ask HN: Will AI models over time converge into the same system?

7•ThinkBeat•6mo ago
I probably am not using the correct terms here so sorry about that.

If all general LLM are eventually exposed to the same data, and a lot of the same use cases will they over time converge in responses?

Even if they are of different arcitecture? or are the current architecture companies use for their big LLM close enough to each other?

Comments

allears•6mo ago
Not an expert, but I believe it's just the opposite. Even with the same LLM and the same training data, responses diverge. And that can be a problem.
drooby•6mo ago
Id think yes.

Intelligence is a model of reality and the future. They'll converge into the same system as a reflection of the laws of physics and human psychology.

And then when they are used as weapons they'll perhaps try to diverge and it will become an arms race to create models of the adversaries models.

_

Another way to look at it is our own history. Intelligent apes all "converged" into our one homo sapien.

Buttons840•6mo ago
I wonder how much of the AI depends on its initial weights? If in coming decades we understand better how neural networks work, it would be funny to look back and realize that Google beat OpenAI because they got lucky with their initial weights or something.
joules77•6mo ago
At a basic level it generates a probability distribution of what the next token should be.

There are a zillion questions that can be asked where you can get a prob dist where multiple tokens have the same probability (flat probability distributions). Then it has to randomly pick one and you can get large variation.

l33tbro•6mo ago
I'd guess no. While they have similar training data, there is plenty of novelty and unique data entering each model due to how each user is using it. This is why ideas like model collapse are fun in theory, but don't really play out due to the irregular ways LLMs are used in the real world.

I could be wrong, but I have not heard a convincing argument for what you propose.

ijk•6mo ago
In aggregate? Signs point to yes. For the general purpose SFT base models. We see some evidence even with RNNs vs Transformers. You're essentially finding a function that models language. Use the same optimization function, get a similar result.

However, the RL and especially the RLHF does a lot to reshape the responses, and that's potentially a lot more varied. For the training that wasn't just cribbed from ChatGPT, anyway.

Lastly, it's unlikely that you'll get the _exact same_ responses; there's too many variables at inference time alone. And as for training, we can fingerprint models by their vocabulary to a certain extent. So in practical terms there's probably always going to be some differences.

This assumes our current training approaches don't change too drastically, of course.

UltraSane•6mo ago
This is called the The Platonic Representation Hypothesis

https://arxiv.org/abs/2405.07987

We argue that representations in AI models, particularly deep networks, are converging. First, we survey many examples of convergence in the literature: over time and across multiple domains, the ways by which different neural networks represent data are becoming more aligned. Next, we demonstrate convergence across data modalities: as vision models and language models get larger, they measure distance between datapoints in a more and more alike way. We hypothesize that this convergence is driving toward a shared statistical model of reality, akin to Plato's concept of an ideal reality. We term such a representation the platonic representation and discuss several possible selective pressures toward it. Finally, we discuss the implications of these trends, their limitations, and counterexamples to our analysis.

moomoo11•6mo ago
There are like maybe <100 people who actually contribute actively to LLMs.

Just treat it like a commodity (like cloud infrastructure) and build cool shit using it.

If the provider can roll that feature into their offerings then you’re not actually adding any value to the world.

mikewarot•6mo ago
I'm fairly certain that wouldn't happen. Unless you were to overfit the models until the error were to drop to zero, which would likely take almost infinite time. If you did get that point, you've managed to achieve lossless compression of the training data into the weights of the model.

Given that AI models are randomly initialized with noise, and the goal of training is to avoid overfit, there will always be variance between the weights of models, even if trained from the same data, due to those initial conditions, and chaos theory.

And all of the above, is for the same model architecture. I expect you could do some principle component analysis and come up with a transform to work between models, again if they were overfit to zero error. (After all, that would be a compression engine instead of an AI at that point)

Upon reflection, it seems to me that free Stanford AI course I took a decade ago actually stuck. 8)