frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Talkie: a 13B vintage language model from 1930

https://talkie-lm.com/introducing-talkie
63•jekude•3h ago

Comments

aftbit•1h ago
Darn I've only got ~20 GB of VRAM. I really need to get a stronger machine for this sort of stuff.
Wowfunhappy•39m ago
How much system memory do you have? Llama.cpp can split layers across cpu and gpu. Speeds will be slower of course but it's not unusable at all.
MerrimanInd•37m ago
20GB isn't enough for a 13B parameter model? I thought the 29-31B models could run on a 24GB GTX x090 card?

I'm currently shopping for a local LLM setup and between something like the Framework Desktop with 64-128GB of shared RAM or just adding a 3090 or 4090 to my homelab so I'm very curious what hardware is working well for others.

zamadatix•7m ago
> 20GB isn't enough for a 13B parameter model? I thought the 29-31B models could run on a 24GB GTX x090 card?

Parameters are like Hertz - they don't really tell you much until you know the rest anyways. In this case, a parameter is a bfloat16 (2 bytes). I'm sure someone will bother to makes quants at some point.

> I'm currently shopping for a local LLM setup and between something like the Framework Desktop with 64-128GB of shared RAM or just adding a 3090 or 4090 to my homelab so I'm very curious what hardware is working well for others.

I grabbed a 395 laptop w/ 128 GB to be a personal travel workstation. Great for that purpose. Not exactly a speed demon with LLMs but it can load large ones (which run even slower as a result) and that wasn't really my intent. I've found GPUs make more usable local LLMs, particularly in the speed department, but I suppose that depends more on how you really use them and how much you're willing to pay to have enough total VRAM.

It's next to impossible to make your money back on local (regardless what you buy) so I'd just say "go for whatever amount of best you're willing to put money down for" and enjoy it.

pizzalife•1h ago
This is cool. Is it possible to easily install with ollama?
walrus01•1h ago
I think that one could also take a much larger model (35B or 122B sized) and give it a thorough system prompt to only speak in the manner of a well educated Victorian/Edwardian era gentleman, if you want an "old timey" LLM.
zellyn•52m ago
As we learn how to train smarter models on less data, it’ll become more and more interesting to see whether models like this can invent post-1930 math, science, etc. and make predictions.

[Edit: serves me right for not reading tfa. My points are well-covered]

yesitcan•55m ago
Vintage is a funny thing to call this. Is it running on vacuum tube hardware?
teleforce•50m ago
>Have you ever daydreamed about talking to someone from the past?

Fun facts, LLM was once envisioned by Steve Jobs in one of his interviews [1].

Essentially one of his main wish in life is to meet and interract with Aristotle, in which according to him at the time, computer in the future can make it possible.

[1] In 1985 Steve Jobs described a machine that would help people get answers from Aristotle–modern LLM [video]:

https://youtu.be/yolkEfuUaGs

cedilla•32m ago
The idea of talking to a machine that has all of humanities knowledge and gives answers is older than electronic computing. It certainly wasn't a novel idea when Jobs gave that speech. At that time, the field of artificial intelligence was old enough to become US president.
jcgrillo•25m ago
Except... not at all? The vast majority of the training data required to create an artificial Aristotle has been lost forever. Smash your coffee cup on the ground. Now reassemble it and put the coffee back in. Once you can repeatably do that I'll begin to believe you can train an artificial Aristotle.
freetanga•15m ago
Imagine aiming for Aristotle and landing on Siri…
twoodfin•48m ago
The Python example is fascinating, and a good rejoinder to anyone still dismissing LLM’s as stochastic parrots.
sega_sai•43m ago
It is cool. I find the idea of trying to understand whether these types of models can come up with things like General relativity, or maybe some results really interesting.
pmw•35m ago
Related: https://github.com/haykgrigo3/TimeCapsuleLLM

> A language model trained from scratch exclusively on data from certain places and time periods to reduce modern bias and emulate the voice, vocabulary, and worldview of the era.

Discussed here: https://news.ycombinator.com/item?id=46590280

simonw•20m ago
Whoa, Alec Radford is on the list of authors! He was instrumental in building the original GPT models at OpenAI.
teraflop•17m ago
I have no real quibble with the blog post itself, but I take issue with the title that calls it a "vintage model".

The blog post defines a "vintage model" as one that is trained only on data before a particular cutoff point:

> Vintage LMs are contamination-free by construction, enabling unique generalization experiments [...] The most important objective when training vintage language models is that no data leaks into the training corpus from after the intended knowledge cutoff

But as they acknowledge later, there are multiple major data leakage issues in their training pipeline, and their model does in fact have quite a bit of anachronistic knowledge. So it fails at what they call the most important objective. It's fair to say that they are working toward something that meets their definition of "vintage", but they're not there yet.

____tom____•12m ago
>Have you ever daydreamed about talking to someone from the past?

It's going to be more like corresponding with someone from the past. We don't have much in the way of recorded speech from that area, so this will be built from written records. Much more than now, the written records are going to be formal and edited, reflecting a different pattern than casual speech or writing.

Having said that, this is cool. I recently had to OCR a two-hundred year old book with the usual garish fonts from that era. It was remarkably easy to do, and accurate.

alexpotato•2m ago
I was reading Nate Silver's book "On The Edge" and there is an interesting part where he takes predictions on the usage of nuclear weapons taken from just after World War 2 and compares them to what the Bayesian prediction would be given what actually happened.

Post World War 2, some people had the odds per year at 10%. Some of that is probably a mix of recency bias + not understanding how to use new weapons etc etc but as Silver points out, the odds were much lower.

I mention this only b/c the "could something trained on LLMs of the time predict the future" always makes me think of it.

To My Students

http://ozark.hendrix.edu/~yorgey/forest/00FD/index.xml
195•marvinborner•1h ago•79 comments

New Integrated by Design FreeBSD Book

https://vivianvoss.net/blog/integrated-by-design-launch
46•vermaden•1h ago•10 comments

Microsoft and OpenAI end their exclusive and revenue-sharing deal

https://www.bloomberg.com/news/articles/2026-04-27/microsoft-to-stop-sharing-revenue-with-main-ai...
738•helsinkiandrew•11h ago•649 comments

Talkie: a 13B vintage language model from 1930

https://talkie-lm.com/introducing-talkie
64•jekude•3h ago•19 comments

Meetings Are Forcing Functions

https://www.mooreds.com/wordpress/archives/3734
27•zdw•1d ago•9 comments

Three men are facing charges in Toronto SMS Blaster arrests

https://www.tps.ca/media-centre/stories/unprecedented-sms-blaster-arrests/
78•gnabgib•4h ago•42 comments

Generative AI Vegetarianism

https://sboots.ca/2026/03/11/generative-ai-vegetarianism/
14•marvinborner•1h ago•12 comments

Is my blue your blue?

https://ismy.blue/
311•theogravity•4h ago•214 comments

Easyduino: Open Source PCB Devboards for KiCad

https://github.com/Hanqaqa/Easyduino
168•Hanqaqa•7h ago•24 comments

The quiet resurgence of RF engineering

https://atempleton.bearblog.dev/quiet-resurgence-of-rf-engineering/
125•merlinq•2d ago•68 comments

Spanish archaeologists discover trove of ancient shipwrecks in Bay of Gibraltar

https://www.theguardian.com/science/2026/apr/15/hidden-treasures-spanish-archaeologists-discover-...
79•1659447091•1d ago•9 comments

4TB of voice samples just stolen from 40k AI contractors at Mercor

https://app.oravys.com/blog/mercor-breach-2026
432•Oravys•15h ago•160 comments

Men who stare at walls

https://www.alexselimov.com/posts/men_who_stare_at_walls/
435•aselimov3•14h ago•201 comments

Networking changes coming in macOS 27

https://eclecticlight.co/2026/04/23/networking-changes-coming-in-macos-27/
201•pvtmert•9h ago•169 comments

Claude Pro: Opus model will only be available if extra usage is enabled

https://support.claude.com/en/articles/11940350-claude-code-model-configuration
45•yrds96•2h ago•20 comments

China blocks Meta's acquisition of AI startup Manus

https://www.cnbc.com/2026/04/27/meta-manus-china-blocks-acquisition-ai-startup.html
294•yakkomajuri•13h ago•195 comments

The woes of sanitizing SVGs

https://muffin.ink/blog/scratch-svg-sanitization/
177•varun_ch•9h ago•70 comments

Radar Laboratory – Interactive Radar Phenomenology

https://radarlaboratory.com/
21•jonbaer•2d ago•0 comments

How I leared what a decoupling capacitor is for, the hard way

https://nbelakovski.substack.com/p/how-i-learned-what-a-decoupling-capacitor
16•actinium226•2d ago•1 comments

The Secret Life of NaN

https://anniecherkaev.com/the-secret-life-of-nan
27•prakashqwerty•1d ago•12 comments

Super ZSNES – GPU Powered SNES Emulator

https://zsnes.com/
224•haunter•7h ago•64 comments

GitHub Copilot is moving to usage-based billing

https://github.blog/news-insights/company-news/github-copilot-is-moving-to-usage-based-billing/
532•frizlab•9h ago•409 comments

Fully Featured Audio DSP Firmware for the Raspberry Pi Pico

https://github.com/WeebLabs/DSPi
260•BoingBoomTschak•2d ago•75 comments

Pgbackrest is no longer being maintained

https://github.com/pgbackrest/pgbackrest
392•c0l0•14h ago•204 comments

Show HN: OSS Agent I built topped the TerminalBench on Gemini-3-flash-preview

https://github.com/dirac-run/dirac
293•GodelNumbering•12h ago•118 comments

FDA approves first gene therapy for treatment of genetic hearing loss

https://www.fda.gov/news-events/press-announcements/fda-approves-first-ever-gene-therapy-treatmen...
211•JeanKage•14h ago•81 comments

“Why not just use Lean?”

https://lawrencecpaulson.github.io//2026/04/23/Why_not_Lean.html
246•ibobev•10h ago•175 comments

NPM website was down

https://status.npmjs.org
113•18nleung•3h ago•53 comments

Magic by return of post: How mail order delivered the occult

https://publicdomainreview.org/essay/magic-by-return-of-post/
38•Vigier•2d ago•4 comments

Quarkdown – Markdown with Superpowers

https://quarkdown.com/
250•amai•16h ago•91 comments