I'm a news avoider, so I usually feel some smug self-satisfaction in those instances, but I wondered if there was a way to satisfy the urge to doomscroll without the anxiety.
My hypothesis: Apply a 40-year latency buffer. You get the intellectual stimulation of "Big Events" without the fog of war, because you know the world didn't end.
40 years creates a mirror between the Reagan Era and today. The parallels include celebrity populism, Cold War tensions (Soviets vs. Russia), and inflation economics.
The system ingests raw newspaper scans and uses a multi-step LLM pipeline to generate the daily edition:
OCR & Ingestion: Converts raw pixels to text.
Scoring: Grades events on metrics like Dramatic Irony and Name Recognition to surface stories that are interesting with hindsight. For example, a dry business blurb about Steve Jobs leaving Apple scores highly because the future context creates a narrative arc.
Objective Fact Extraction: Extracts a list of discrete, verifiable facts from the raw text.
Generation: Uses those extracted facts as the ground truth to write new headlines and story summaries.
I expected a zen experience. Instead, I got an entertaining docudrama. Historical events are surprisingly compelling when serialized over weeks.
For example, on Oct 7, 1985, Palestinian hijackers took over the cruise ship Achille Lauro. Reading this on a delay in 2025, the story unfolded over weeks: first they threw an American in a wheelchair overboard, then US fighter jets forced the escape plane to land, leading to a military standoff between US Navy SEALs and the Italian Air Force. Unbelievably, the US backed down, but the later diplomatic fallout led the Italian Prime Minister to resign.
It hits the dopamine receptors of the news cycle, but with the comfort of a known outcome.
Stack: React, Node.js (Caskada for the LLM pipeline orchestration), Gemini for OCR/Scoring.
Link: https://forty.news (No signup required, it's only if you want the stories emailed to you daily/weekly)
hu3•19m ago
This is neat! But I wonder about longevity of the project if it relies on scanning newspapers.
Do you have an endless suply? Perhaps there is some digital archive you could use?
culi•5m ago
- Arcanum is the largest and continuously expanding digital periodical database from Eastern Europe, which contains scientific and specialized journals, encyclopaedias, weekly and daily newspapers and more
- NewspaperARCHIVE.com is an online database of digitized newspapers, with over 2 billion news articles; coverage extends from 1607 to the present from US, Canada, the UK, and 20 other countries.
- Newspapers.com includes more than 800 million pages from 20,000+ newspapers. The collection includes some major newspapers for limited periods (e.g., first 72 years of the New York Times), but mostly consists of US regional papers from the 1700s to the late 1980s. Free accounts through the Wikipedia Library include access to Newspapers.com Publisher Extra content.
- ProQuest is a multidisciplinary research provider. This access includes ProQuest Central, which includes a large collection of journals and newspapers, Literature Online, the HNP Chinese Newspaper Collections, and the Historical New York Times.
- Wikilala is a digital repository consisting of more than 109,000 documents in printed form, including 45,000 newspapers, 32,000 journals, 4,000 books and 26,000 articles concerning the history of the Ottoman Empire from its founding to the modern times.
Also most newspapers maintain their own archives, usually accessible online. Here's some I get access to: The Corriere della Sera (one of Italy's oldest and most read newspapers); The Corriere della Sera (a century of historical archives); The Times of Malta (Founded in 1935, it is the oldest daily newspaper still in circulation in Malta); ZEIT ONLINE (online version of Die Zeit, a German weekly newspaper) — and quite a few more