frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

US tech rules the European market

https://proton.me/blog/us-tech-rules-europe
1•devonnull•7s ago•0 comments

The first appearance of "I use arch btw"

https://twojepc.pl/boardArchPytanie2005_103352.html
1•Jotalea•5m ago•1 comments

Early test stage of scripts to extract ntlm and kerberos hashes from pcaps

https://gist.github.com/dleto614/5663b9de7e7449d217e6e38a5e5386c2
1•grepcat•11m ago•0 comments

Why Are Tech Workers So Dissatisfied? [video]

https://www.youtube.com/watch?v=28SuvTE5xNE
1•mgh2•16m ago•0 comments

Safari's disadvantage is OS updates (2024)

https://www.alvar.dev/blog/safari-disadvantage-os-updates
2•mooreds•24m ago•0 comments

Show HN: Metis Agent v0.6.1 – Run OpenAI's GPT OSS Locally with No API Keys

https://github.com/metisos/metisos_agentV1
1•cjohnsonpr•25m ago•0 comments

No point in fighting Drivers who appeal speed cameras almost guaranteed to lose

https://www.abcactionnews.com/news/state/theres-no-point-in-fighting-drivers-who-appeal-school-speed-zone-camera-fines-almost-guaranteed-to-lose
1•josephcsible•26m ago•1 comments

OpenAI in Talks for Share Sale Valuing Startup at $500B

https://www.bloomberg.com/news/articles/2025-08-06/openai-in-talks-for-share-sale-valuing-startup-at-500-billion
1•mfiguiere•30m ago•0 comments

Trump admin warns states: Don't try to lower broadband prices

https://arstechnica.com/tech-policy/2025/08/trump-admin-warns-states-dont-try-to-lower-broadband-prices/
6•duxup•32m ago•1 comments

Show HN: Minimal terminal-style portfolio template (~13KB) with GitHub API data

https://github.com/Cod-e-Codes/lilweb-template
1•Cod-e-Codes•33m ago•0 comments

I Was Reincarnated as the 7th Prince Season 1 Anime Review

https://www.animenewsnetwork.com/review/i-was-reincarnated-as-the-7th-prince/.226628
1•PaulHoule•36m ago•0 comments

Billions of starfish have died in a decade-long epidemic

https://www.cbsnews.com/news/starfish-sea-star-died-epidemic-scientists-know-why/
1•Brajeshwar•37m ago•0 comments

Kids shoes with a hidden AirTag compartment

https://techcrunch.com/2025/07/30/skechers-is-making-kids-shoes-with-a-hidden-airtag-compartment/
1•walterbell•38m ago•1 comments

Arenas in Rust

https://russellw.github.io/arenas
2•rwallace•42m ago•0 comments

If you're a direct employee of HP/Compaq you're not allowed to look at this code

https://www.ukcert.org.uk/repository/exploits/NETSYS_COM%20-%20The%20Intelligent%20Hacker%27s%20Choice%20-%20http--www_netsys_com-library-alerts-2002-08-05-dxchpwd.txt
2•dijksterhuis•43m ago•0 comments

The Mistake That Killed Excite: The HomeNetwork

https://en.wikipedia.org/wiki/@Home_Network
2•sans_souse•47m ago•1 comments

Show HN: I built a browser extension to add comment threads on any website

https://medium.com/@oencab2/why-im-building-a-browser-that-lets-you-leave-comments-on-the-internet-itself-9d4c2404d4b8
2•itzoen•49m ago•0 comments

Electric motor runs without metal coils

https://newatlas.com/technology/kist-cnt-cscec-carbon-nanotube-wire/
3•westurner•52m ago•1 comments

How to build realistic AI companions

https://www.emotionmachine.ai/blog/realistic-ai-companions
1•sarbak•1h ago•0 comments

What if technology is our weakness?

3•morpheos137•1h ago•1 comments

Spacebar Clicker – Ultimate Auto Clicker Game Online

https://spacebarclickers.online/
2•nico_nico•1h ago•1 comments

Show HN: PinpoinTodays – Daily Answers and History for LinkedIn's Pinpoint Game

https://pinpointodays.com
1•qinggeng•1h ago•0 comments

Science Titan sub firm used intimidation tactics and flawed safety practices

https://www.bbc.com/news/live/cdjxp4n2371t
10•teleforce•1h ago•1 comments

Cerebras now supports OpenAI GPT-OSS-120B at 3k Tokens Per SEC

https://www.cerebras.ai/news/cerebras-helps-power-openai-s-open-model-at-world-record-inference-speeds-gpt-oss-120b-delivers
3•me551ah•1h ago•0 comments

Man who lit cigarette from French war memorial flame faces legal action

https://www.theguardian.com/world/2025/aug/05/french-minister-legal-action-against-man-lit-cigarette-memorial
4•wslh•1h ago•2 comments

Never miss a conversation that matters

https://socialystener.com/
1•usamak23•1h ago•0 comments

Show HN: Virtual Ontologies with Claude Code

https://medium.com/@michael.craig.fitzgerald/whither-ontologies-d871bd3a8098
2•mcfitzgerald•1h ago•0 comments

RFK Halts mRNA vaccine research

https://www.pbs.org/newshour/health/rfk-jr-pulls-funding-for-vaccines-being-developed-to-fight-respiratory-viruses
10•worik•1h ago•1 comments

Kitten TTS: 25MB CPU-Only, Open-Source Voice Model

https://algogist.com/kitten-tts-the-25mb-ai-voice-model-thats-about-to-change-everything-runs-on-a-potato/
72•jainilprajapati•1h ago•26 comments

Engineer restores pay phones for free public use

https://www.npr.org/2025/08/04/nx-s1-5484013/engineer-restores-pay-phones-for-free-public-use
5•andsoitis•1h ago•1 comments
Open in hackernews

Model Collapse and the Need for Human-Generated Training Data

https://glthr.com/model-collapse-and-the-need-for-human-generated-training-data
7•glth•2h ago

Comments

sans_souse•12m ago
I think a point often missed is that it's not just what the substance and quality of those sources and their associated decline but also the overall decline of sources, period. The first phases involved training the models with a massive backlog of raw knowlege, communicated over thousands of years, and for the majority of that span, this was a world much different from our today; in short, all of our knowledge was "boots on the ground" type, and all of it served to aid our growth, and our record of this tells such a story.

But our knowledge and growth today is so narrow in scope (in a sense) and there's an ever looming scenario ready to present itself where our perceived growth is actually a recursion and the answer to "what is the purpose" becomes "there is none"

ipython•12m ago
So I’ve heard of this model collapse theory. But I’ve also heard of model providers who are intentionally training with synthetically generated data (as a result of insufficient “real” data).

So I’m curious where the line is? Are there phases in the training/continued pre training/alignment/rlhf pipeline where synthetic data isn’t just harmless but actually beneficial? Is it a question of quantity or a question of how much novelty is in the training data?