frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Show HN: Mirror Parliament where users vote on top of politicians and draft laws

https://github.com/fokdelafons/lustra
1•fokdelafons•27s ago•0 comments

Ask HN: Opus 4.6 ignoring instructions, how to use 4.5 in Claude Code instead?

1•Chance-Device•1m ago•0 comments

We Mourn Our Craft

https://nolanlawson.com/2026/02/07/we-mourn-our-craft/
1•ColinWright•4m ago•0 comments

Jim Fan calls pixels the ultimate motor controller

https://robotsandstartups.substack.com/p/humanoids-platform-urdf-kitchen-nvidias
1•robotlaunch•8m ago•0 comments

Exploring a Modern SMTPE 2110 Broadcast Truck with My Dad

https://www.jeffgeerling.com/blog/2026/exploring-a-modern-smpte-2110-broadcast-truck-with-my-dad/
1•HotGarbage•8m ago•0 comments

AI UX Playground: Real-world examples of AI interaction design

https://www.aiuxplayground.com/
1•javiercr•9m ago•0 comments

The Field Guide to Design Futures

https://designfutures.guide/
1•andyjohnson0•9m ago•0 comments

The Other Leverage in Software and AI

https://tomtunguz.com/the-other-leverage-in-software-and-ai/
1•gmays•11m ago•0 comments

AUR malware scanner written in Rust

https://github.com/Sohimaster/traur
3•sohimaster•13m ago•1 comments

Free FFmpeg API [video]

https://www.youtube.com/watch?v=6RAuSVa4MLI
3•harshalone•13m ago•1 comments

Are AI agents ready for the workplace? A new benchmark raises doubts

https://techcrunch.com/2026/01/22/are-ai-agents-ready-for-the-workplace-a-new-benchmark-raises-do...
2•PaulHoule•18m ago•0 comments

Show HN: AI Watermark and Stego Scanner

https://ulrischa.github.io/AIWatermarkDetector/
1•ulrischa•19m ago•0 comments

Clarity vs. complexity: the invisible work of subtraction

https://www.alexscamp.com/p/clarity-vs-complexity-the-invisible
1•dovhyi•20m ago•0 comments

Solid-State Freezer Needs No Refrigerants

https://spectrum.ieee.org/subzero-elastocaloric-cooling
2•Brajeshwar•20m ago•0 comments

Ask HN: Will LLMs/AI Decrease Human Intelligence and Make Expertise a Commodity?

1•mc-0•21m ago•1 comments

From Zero to Hero: A Brief Introduction to Spring Boot

https://jcob-sikorski.github.io/me/writing/from-zero-to-hello-world-spring-boot
1•jcob_sikorski•22m ago•1 comments

NSA detected phone call between foreign intelligence and person close to Trump

https://www.theguardian.com/us-news/2026/feb/07/nsa-foreign-intelligence-trump-whistleblower
8•c420•22m ago•1 comments

How to Fake a Robotics Result

https://itcanthink.substack.com/p/how-to-fake-a-robotics-result
1•ai_critic•23m ago•0 comments

It's time for the world to boycott the US

https://www.aljazeera.com/opinions/2026/2/5/its-time-for-the-world-to-boycott-the-us
3•HotGarbage•23m ago•0 comments

Show HN: Semantic Search for terminal commands in the Browser (No Back end)

https://jslambda.github.io/tldr-vsearch/
1•jslambda•23m ago•1 comments

The AI CEO Experiment

https://yukicapital.com/blog/the-ai-ceo-experiment/
2•romainsimon•25m ago•0 comments

Speed up responses with fast mode

https://code.claude.com/docs/en/fast-mode
4•surprisetalk•28m ago•0 comments

MS-DOS game copy protection and cracks

https://www.dosdays.co.uk/topics/game_cracks.php
4•TheCraiggers•29m ago•0 comments

Updates on GNU/Hurd progress [video]

https://fosdem.org/2026/schedule/event/7FZXHF-updates_on_gnuhurd_progress_rump_drivers_64bit_smp_...
2•birdculture•30m ago•0 comments

Epstein took a photo of his 2015 dinner with Zuckerberg and Musk

https://xcancel.com/search?f=tweets&q=davenewworld_2%2Fstatus%2F2020128223850316274
14•doener•30m ago•2 comments

MyFlames: View MySQL execution plans as interactive FlameGraphs and BarCharts

https://github.com/vgrippa/myflames
1•tanelpoder•32m ago•0 comments

Show HN: LLM of Babel

https://clairefro.github.io/llm-of-babel/
1•marjipan200•32m ago•0 comments

A modern iperf3 alternative with a live TUI, multi-client server, QUIC support

https://github.com/lance0/xfr
3•tanelpoder•33m ago•0 comments

Famfamfam Silk icons – also with CSS spritesheet

https://github.com/legacy-icons/famfamfam-silk
1•thunderbong•34m ago•0 comments

Apple is the only Big Tech company whose capex declined last quarter

https://sherwood.news/tech/apple-is-the-only-big-tech-company-whose-capex-declined-last-quarter/
4•elsewhen•37m ago•0 comments
Open in hackernews

Vintage Large Language Models

https://owainevans.github.io/talk-transcript.html
83•pr337h4m•2mo ago

Comments

mountainriver•2mo ago
Very cool! I’ve been wanting to do this do a long time!
nxobject•2mo ago
I love the ideas about how we might use historical LLMs to inquire into the past!

I imagine that (the author hints at this), to do this rigorously, spelling out assumptions etc, you’d have to build off theoretical frameworks used to inductively synthesize/qualify interviews and texts, currently around in history and the social sciences.

abeppu•2mo ago
The talk focuses for a bit on having pure data from before the given date. But it doesn't consider that the data available from before that time may be subject to strong selection bias, based on what's interesting to people doing scholarship or archival work after that date. E.g. have we disproportionately digitized the notes/letters/journals of figures whose ideas have gained traction after their death?

The article makes a comparison to financial backtesting. If you form a dataset of historical prices of stocks which are _currently_ in the S&P500, even if you only use price data before time t, models trained against your data will expect that prices go up and companies never die, because they've only seen the price history of successful firms.

alalv•2mo ago
It mentions that problem in the first section
malkia•2mo ago
Not a financial person by any means, but doesn't the Black Swan Theory basically disproves such methods due to rarity of an event that might have huge impact without something to predict (in the past) that it might happen, or even if it can be predicted - the impact cannot?

For example: Chernobyl, COVID, 2008 financial crisis and even 9/11

ACCount37•2mo ago
All models are wrong, but some are useful.

If you had a financial model that somehow predicted everything but black swan events, that would still be enough to make yourself rich beyond belief.

dboon•2mo ago
The talk explicitly addresses this exact issue.
ideashower•2mo ago
I like the idea of using vintage LLMs to study explicit and implicit bias. e.g. text before mid-19th century believing in racial superiority, gender discrimination, imperial authority or slavery. Comparing that to text since then. I'm sure there are more ideas when you use temporal constraints on training data.
digdugdirk•2mo ago
I've been wanting to do this on historical court records - building upon the existing cases, one by one, using llms as the "Judge". It'd be interesting to see which cases branch off from the established precedent, and how that cascades into the present.

Any thoughts how one could get started with this?

UltraSane•2mo ago
Over the long term LLMs are going to become very interesting snapshots of history. Imagine prompting an LLM from 2025 in 2125.
lukan•2mo ago
I would probably prefer wikipedia snapshots (including debate) as a future historian.
selfhoster11•2mo ago
The more options you have, the better IMO.
i80and•2mo ago
Maybe in the sense that a CueCat is interesting to us today.
nxobject•2mo ago
You're right: I wish OpenAI could find a way to "donate" GPT-2 or GPT-3 to the CHM, or some open archive.

I feel like that generation of models was around the point where we were getting pleasantly surprised by the behaviors of models. (I think people were having fun translating things into sonnets back then?)

unleaded•2mo ago
Someone has sort of done this:

https://www.reddit.com/r/LocalLLaMA/comments/1mvnmjo/my_llm_...

I doubt a better one would cost $200,000,000.

ijk•2mo ago
I was hoping that this would be about Llama 1 and comparison with GPT-contaminated models.
kingkongjaffa•2mo ago
This would be a good way to verify emergent model capability to synthesize new knowledge.

You give an LLM all the information from right before a topic was discovered or invented, and then you see if it can independently generate the new knowledge or not.

It would be hard to know for sure if a discovery was genuine or accidentally included in the training data though.

qingcharles•2mo ago
I saw Musk repost a boast that Grok created a whole new ("superior") element design for a incandescent bulb using Edison's patent. The implication was that Grok was superior to Edison's team. I was just sat there thinking about the 100+ years of incandescent bulb research that Grok has sucked up from various science papers and random Internet archives. Surely none of that was any help at all /s.
carsoon•2mo ago
Using old models is a good way to received less biased information about an active event. Once a major event occurs information wars happen that try and change narratives and erase old information. But because models were trained before this the bias that the event causes is not yet present.
lukev•2mo ago
I’m sorry I don’t quite follow… how can a model provide information at all about events it was trained before?
pixl97•2mo ago
Overspecialization of models is a thing.

>Overspecialization of models, often referred to as overfitting in machine learning, is a condition where a model learns the details and noise in the training data so well that it negatively impacts its performance on new, unseen data. This prevents the model from being able to generalize its knowledge effectively.

phs318u•2mo ago
You provide the info... and the bias.
carsoon•2mo ago
Everyone introduces bias. But for instance getting a model trained pre war vs after a war starts is super different. If I want to get raw information about 2 nations then models are in some ways a good source. Because most other parts of the internet can get changed or wiped. A model is "stuck" with the information it had exactly at that point so cannot be directly affected by new information attacks.

It is crucial to have a good framework in how you ask your questions though to avoid bias when using these systems and to try and focus on raw facts. To test ideas I like to make it fight for both opposite extreme sides of an argument then I can make up my own mind.

carsoon•2mo ago
For instance I want information about 2 countries currently at war. By asking about these countries from an older model then we get more factual information about the countries. If we ask about them and the information is seeded from news articles etc AFTER the war started then they will be biasedly influenced and often have disclaimers like "But it should be noted that x y z" showing that there is some MAJOR bias that occurred from the training on the news.

If I want an unbiased reason for what happened before a war started i would want all the information about 2 countries at different points before the war. Because after a military war starts an INFORMATION war also starts. Propaganda will be spread from both sides as wars are just as much about global support as they are about military dominance.

carsoon•2mo ago
We need a library of Alexandria for primary sources. If we had source transparency then referencing back to original sources would be more clear. We could do cool things like these vintage models to reduce bias from current events. Also books in every language and books for teaching each language would help with multimodality. Copyright makes it difficult to achieve the best results for LLM creation and usage though.
rootnod3•2mo ago
As if the language models currently would give a damn about copyright...
carsoon•2mo ago
The problem is they have to hide their sources due to copyright. So they train on copyright data but must obscure it in the output. Thus they must hide the sources of truth making it impossible to fact check them directly and the reason that hallucinations are so common and unavoidable in the current pattern.
nxobject•2mo ago
Ironically enough, that would be practical for "vintage LLM" - perhaps (morally) obligatory?
ontouchstart•2mo ago
Cool idea. This might be a interesting literary project along this line ;-)

https://www.gutenberg.org/cache/epub/86/pg86-images.html