frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

Open in hackernews

Ask HN: Will AI models over time converge into the same system?

6•ThinkBeat•6h ago
I probably am not using the correct terms here so sorry about that.

If all general LLM are eventually exposed to the same data, and a lot of the same use cases will they over time converge in responses?

Even if they are of different arcitecture? or are the current architecture companies use for their big LLM close enough to each other?

Comments

allears•6h ago
Not an expert, but I believe it's just the opposite. Even with the same LLM and the same training data, responses diverge. And that can be a problem.
drooby•5h ago
Id think yes.

Intelligence is a model of reality and the future. They'll converge into the same system as a reflection of the laws of physics and human psychology.

And then when they are used as weapons they'll perhaps try to diverge and it will become an arms race to create models of the adversaries models.

_

Another way to look at it is our own history. Intelligent apes all "converged" into our one homo sapien.

Buttons840•5h ago
I wonder how much of the AI depends on its initial weights? If in coming decades we understand better how neural networks work, it would be funny to look back and realize that Google beat OpenAI because they got lucky with their initial weights or something.
joules77•5h ago
At a basic level it generates a probability distribution of what the next token should be.

There are a zillion questions that can be asked where you can get a prob dist where multiple tokens have the same probability (flat probability distributions). Then it has to randomly pick one and you can get large variation.

l33tbro•4h ago
I'd guess no. While they have similar training data, there is plenty of novelty and unique data entering each model due to how each user is using it. This is why ideas like model collapse are fun in theory, but don't really play out due to the irregular ways LLMs are used in the real world.

I could be wrong, but I have not heard a convincing argument for what you propose.

ijk•4h ago
In aggregate? Signs point to yes. For the general purpose SFT base models. We see some evidence even with RNNs vs Transformers. You're essentially finding a function that models language. Use the same optimization function, get a similar result.

However, the RL and especially the RLHF does a lot to reshape the responses, and that's potentially a lot more varied. For the training that wasn't just cribbed from ChatGPT, anyway.

Lastly, it's unlikely that you'll get the _exact same_ responses; there's too many variables at inference time alone. And as for training, we can fingerprint models by their vocabulary to a certain extent. So in practical terms there's probably always going to be some differences.

This assumes our current training approaches don't change too drastically, of course.

UltraSane•1h ago
This is called the The Platonic Representation Hypothesis

https://arxiv.org/abs/2405.07987

We argue that representations in AI models, particularly deep networks, are converging. First, we survey many examples of convergence in the literature: over time and across multiple domains, the ways by which different neural networks represent data are becoming more aligned. Next, we demonstrate convergence across data modalities: as vision models and language models get larger, they measure distance between datapoints in a more and more alike way. We hypothesize that this convergence is driving toward a shared statistical model of reality, akin to Plato's concept of an ideal reality. We term such a representation the platonic representation and discuss several possible selective pressures toward it. Finally, we discuss the implications of these trends, their limitations, and counterexamples to our analysis.

YouTube No Translation

https://addons.mozilla.org/en-US/firefox/addon/youtube-no-translation/
1•thefox•3m ago•0 comments

Pimping My Casio: Part Deux

https://blog.jgc.org/2025/07/pimping-my-casio-part-deux.html
1•r4um•3m ago•0 comments

Deriving Muon

https://jeremybernste.in/writing/deriving-muon
1•pizza•6m ago•0 comments

Federal Analogue Act

https://en.wikipedia.org/wiki/Federal_Analogue_Act
4•Bluestein•12m ago•0 comments

Sdcsdsfvvsf [pdf]

https://knightadrc.wustl.edu/app/uploads/formidable/124/Como-llamar-a-Avianca-desde-Barranquilla.pdf
1•graceevans•13m ago•0 comments

Show HN: Compass CNC – Open-Source Handheld CNC Router

https://www.compassrouter.com
1•camchaney•13m ago•0 comments

Run 15 Niche Accounts Without Flags, Boost Monthly Pledges by 2.5x

1•yt1314•15m ago•0 comments

Every part on a bicycle is safety critical

https://escapecollective.com/threaded-43-every-part-on-a-bike-is-safety-critical/
1•spooky_deep•18m ago•0 comments

Pelias: A modular open-source geocoder using Elasticsearch

https://github.com/pelias/pelias
1•saikatsg•18m ago•0 comments

Exhausted man defeats AI model in world coding championship

https://arstechnica.com/civis/threads/exhausted-man-defeats-ai-model-in-world-coding-championship.1508421/page-3
4•Bluestein•20m ago•0 comments

You Can't Buy Integration (2021)

https://martinfowler.com/articles/cant-buy-integration.html
1•mpweiher•23m ago•0 comments

Ship Your Directory in 1-Hour with DirEasy Boilerplate

https://www.direasy.com/en
1•allentown521•25m ago•1 comments

My Stroke of Insight [video]

https://www.ted.com/talks/jill_bolte_taylor_my_stroke_of_insight
4•Bluestein•26m ago•0 comments

A brief history of Mac primary coding languages

https://eclecticlight.co/2025/07/19/a-brief-history-of-primary-coding-languages/
3•ingve•28m ago•0 comments

Chip Benchmark: Hardware-Centric Performance Insights for AI Workloads

https://www.herdora.com/blog/chip-benchmark
1•technoabsurdist•31m ago•0 comments

Just Published a Blog: How to Add a Gmail OAuth2 Account with RustMailer

https://www.indiehackers.com/post/just-published-a-blog-how-to-add-a-gmail-oauth2-account-with-rustmailer-TMSpJURy2VrUSOkCqYp9
1•rustmailer•34m ago•0 comments

I built a memoryless AI persona system at 18

https://medium.com/@jasonlyu3007/lra-core01-how-i-built-a-memoryless-ai-persona-system-at-18-fde1c82cbb99
1•Lra_core•35m ago•1 comments

Optimize Next.js Performance with Code Splitting: What to Load, When, and Why

https://medium.com/@aalbertini95_90842/optimize-next-js-performance-with-smart-code-splitting-what-to-load-when-and-why-485dac08cd24
1•BoopyKiki•35m ago•0 comments

Async Isn't Real and Cannot Hurt You [video]

https://www.youtube.com/watch?v=AiSl4vf40WU
1•noisem4ker•41m ago•1 comments

Inside a click farm (2024)

https://www.scmp.com/magazines/post-magazine/article/3266034/inside-click-farm-photographers-expose-how-they-work-and-thrive-plain-sight
2•phaser•48m ago•0 comments

My Favourite German Word

https://vurt.org/articles/my-favourite-german-word/
2•taubek•50m ago•1 comments

Ask HN: Should I Pivot to Data Engineering/MLOps from Senior SDE?

1•febed•52m ago•0 comments

Made a Web app for devs to share projects / find contributors

https://opensourceradar.org
2•ReddBird•1h ago•7 comments

Swipe Clean – Clean up storage iOS app

https://apps.apple.com/us/app/swipe-clean-free-up-storage/id6475195875
1•bubbly_snow•1h ago•1 comments

Scientists achieve first experimental observation of transverse Thomson effect

https://phys.org/news/2025-07-scientists-experimental-transverse-thomson-effect.html
2•IdealeZahlen•1h ago•0 comments

AI: How can we control an alien intelligence? [video]

https://www.youtube.com/watch?v=0BnZMeFtoAM
1•austinallegro•1h ago•0 comments

The Trial of Roman Storm and Tornado Cash: The United States vs. Digital Privacy

https://www.therage.co/tornado-cash-trial-intro/
1•janandonly•1h ago•0 comments

Stealth Macintosh Portable Case Mod

https://biosrhythm.com/?p=2956
2•Bogdanp•1h ago•0 comments

US revokes visas of Brazilian judges after crack down on ex-president Bolsonaro

https://nypost.com/2025/07/18/world-news/us-revokes-visas-of-brazilian-justices-after-its-top-court-cracks-down-on-ex-president-bolsonaro/
9•matheusmoreira•1h ago•0 comments

We're Speeding Toward Skynet

6•cranberryturkey•1h ago•4 comments