frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

OpenCiv3: Open-source, cross-platform reimagining of Civilization III

https://openciv3.org/
612•klaussilveira•12h ago•180 comments

The Waymo World Model

https://waymo.com/blog/2026/02/the-waymo-world-model-a-new-frontier-for-autonomous-driving-simula...
915•xnx•17h ago•545 comments

What Is Ruliology?

https://writings.stephenwolfram.com/2026/01/what-is-ruliology/
29•helloplanets•4d ago•22 comments

How we made geo joins 400× faster with H3 indexes

https://floedb.ai/blog/how-we-made-geo-joins-400-faster-with-h3-indexes
102•matheusalmeida•1d ago•24 comments

Unseen Footage of Atari Battlezone Arcade Cabinet Production

https://arcadeblogger.com/2026/02/02/unseen-footage-of-atari-battlezone-cabinet-production/
36•videotopia•4d ago•1 comments

Show HN: Look Ma, No Linux: Shell, App Installer, Vi, Cc on ESP32-S3 / BreezyBox

https://github.com/valdanylchuk/breezydemo
212•isitcontent•12h ago•25 comments

Jeffrey Snover: "Welcome to the Room"

https://www.jsnover.com/blog/2026/02/01/welcome-to-the-room/
5•kaonwarb•3d ago•1 comments

Monty: A minimal, secure Python interpreter written in Rust for use by AI

https://github.com/pydantic/monty
206•dmpetrov•12h ago•101 comments

Show HN: I spent 4 years building a UI design tool with only the features I use

https://vecti.com
316•vecti•14h ago•140 comments

Microsoft open-sources LiteBox, a security-focused library OS

https://github.com/microsoft/litebox
355•aktau•18h ago•181 comments

Sheldon Brown's Bicycle Technical Info

https://www.sheldonbrown.com/
361•ostacke•18h ago•94 comments

Hackers (1995) Animated Experience

https://hackers-1995.vercel.app/
471•todsacerdoti•20h ago•232 comments

Show HN: If you lose your memory, how to regain access to your computer?

https://eljojo.github.io/rememory/
267•eljojo•15h ago•157 comments

An Update on Heroku

https://www.heroku.com/blog/an-update-on-heroku/
400•lstoll•18h ago•271 comments

Delimited Continuations vs. Lwt for Threads

https://mirageos.org/blog/delimcc-vs-lwt
25•romes•4d ago•3 comments

Dark Alley Mathematics

https://blog.szczepan.org/blog/three-points/
82•quibono•4d ago•20 comments

PC Floppy Copy Protection: Vault Prolok

https://martypc.blogspot.com/2024/09/pc-floppy-copy-protection-vault-prolok.html
54•kmm•4d ago•3 comments

Was Benoit Mandelbrot a hedgehog or a fox?

https://arxiv.org/abs/2602.01122
9•bikenaga•3d ago•2 comments

How to effectively write quality code with AI

https://heidenstedt.org/posts/2026/how-to-effectively-write-quality-code-with-ai/
242•i5heu•15h ago•183 comments

Introducing the Developer Knowledge API and MCP Server

https://developers.googleblog.com/introducing-the-developer-knowledge-api-and-mcp-server/
51•gfortaine•10h ago•16 comments

I spent 5 years in DevOps – Solutions engineering gave me what I was missing

https://infisical.com/blog/devops-to-solutions-engineering
138•vmatsiiako•17h ago•60 comments

Understanding Neural Network, Visually

https://visualrambling.space/neural-network/
275•surprisetalk•3d ago•37 comments

Show HN: R3forth, a ColorForth-inspired language with a tiny VM

https://github.com/phreda4/r3
68•phreda4•11h ago•13 comments

I now assume that all ads on Apple news are scams

https://kirkville.com/i-now-assume-that-all-ads-on-apple-news-are-scams/
1052•cdrnsf•21h ago•433 comments

Why I Joined OpenAI

https://www.brendangregg.com/blog/2026-02-07/why-i-joined-openai.html
127•SerCe•8h ago•111 comments

Female Asian Elephant Calf Born at the Smithsonian National Zoo

https://www.si.edu/newsdesk/releases/female-asian-elephant-calf-born-smithsonians-national-zoo-an...
28•gmays•7h ago•10 comments

Learning from context is harder than we thought

https://hy.tencent.com/research/100025?langVersion=en
173•limoce•3d ago•93 comments

Vocal Guide – belt sing without killing yourself

https://jesperordrup.github.io/vocal-guide/
7•jesperordrup•2h ago•4 comments

FORTH? Really!?

https://rescrv.net/w/2026/02/06/associative
61•rescrv•20h ago•22 comments

Zlob.h 100% POSIX and glibc compatible globbing lib that is faste and better

https://github.com/dmtrKovalenko/zlob
17•neogoose•4h ago•9 comments
Open in hackernews

Springer Nature book on machine learning is full of made-up citations

https://retractionwatch.com/2025/06/30/springer-nature-book-on-machine-learning-is-full-of-made-up-citations/
139•ArmageddonIt•7mo ago

Comments

veltas•7mo ago
Unfortunately not surprising, the quality of a lot of textbooks has been bad for a long time. Students aren't discerning and lecturers often don't try the book out themselves.
gammalost•7mo ago
I agree. I feel that Springer is not doing enough to uphold their reputation. One example of this being a book on RL that I found[1]. It is clear that no one seriously reviewed the content of this book. They are, despite its clear flaws charging 50+ euro.

https://link.springer.com/book/10.1007/978-3-031-37345-9

WillAdams•7mo ago
Yeah, ages ago, when I was doing typesetting, it was disheartening how unaware authors were of the state of things in the fields which they were writing about --- I'm still annoyed that when I pointed out that an article in an "encyclopedia" on the history of spreadsheets failed to mention Javelin or Lotus Improv it was not updated to include those notable examples.

Magazines are even worse --- David Pogue claimed Steve Jobs used Windows 95 on a ThinkPad in one of his columns, when a moment's reflection, and a check of the approved models list at NeXT would have made it obvious it was running NeXTstep.

Even books aren't immune, a recent book on a tool cabinet held up as an example of perfection:

https://lostartpress.com/products/virtuoso

mis-spells H.O. Studley's name on the inside front cover "Henery" as well as making many other typos, myriad bad breaks, pedestrian typesetting which poorly presents numbers and dimensions (failing to use the multiplication symbol or primes) and that they are unwilling to fix a duplicated photo is enshrined in the excerpt which they publish online:

https://blog.lostartpress.com/wp-content/uploads/2016/10/vir...

where what should be photo of an iconic pair of jewelers pliers on pg. 70 is replaced with that of a pair of flat pliers from pg. 142. (any reputable publisher would have done a cancel and fixed that)

Sturgeon's Law, 90% of everything is crap, and I would be a far less grey, and far younger person if I had back all the time and energy I spent fixing files mangled by Adobe Illustrator, or where the wrong typesetting tool was used for the job (the six-weeks re-setting the book re-set by the vendor in Quark XPress when it needed to be in LaTeX was the longest of my life).

EDIT: by extension, I guess it's now 90% of everything is AI-generated crap, 90% of what's left is traditional crap, leaving 1% of worthwhile stuff.

cess11•7mo ago
What reputation would that be?

It was, in part, Springer that enabled Robert Maxwell.

antegamisou•7mo ago
Understandably I'm becoming a bit dogmatic but I'll say it again, AIMA/PRML/ESL are still the best reference textbooks for foundational AI/ML and will be for a long time.

AIMA is Artificial Intelligence: A Modern Approach by Stuart Russell and Peter Norvig

PRML is Pattern Recognition and Machine Learning by Christopher Bishop.

ESL is Elements of Statistical Learning by Trevor Hastie, Robert Tibshirani and Jerome Friedman.

ludicrousdispla•7mo ago
>> LLM-generated citations might look legitimate, but the content of the citations might be fabricated.

Friendy reminder that the entire output from an LLM is fabricated.

amelius•7mo ago
Fabricate is a word with ambiguous meaning. It can mean both "make up", but also simply "produce".
dwayne_dibley•7mo ago
I fabricated this reply out of my brain.
ktallett•7mo ago
I think in this situation both meanings are needing to be used. It produced made up content.
xg15•7mo ago
Technically yes, but not all of it has lost grounding with reality?
mapleoin•7mo ago
You could say that about Alice in Wonderland.
bryanrasmussen•7mo ago
probably a better word here would be fabulated.

on edit: that is to say the content of the citations might be fabulated, while the rest is merely fabricated.

leereeves•7mo ago
I didn't realize "fabulated" was a word. TIL, thank you. But in this case it doesn't sound like the right word; it means: "To tell invented stories, often those that involve fantasy, such as fables."

I think "confabulated" is more appropriate: "To fill in gaps in one's memory with fabrications that one believes to be facts."

bryanrasmussen•7mo ago
there are of course a number of related words from the same roots, but whereas fabulate does have a connection to fantasy, since fabulation is something a fabulist or fabulator does, it also has wider applications, as can be easily seen in confabulate as the con prefix means "With" or "together" so Confabulation is essentially "with fabulation"

That said I guess the version of fabulation I was using is pretty antiquated, probably due to my reading too much 19th century fiction where describing a passel of lies as a pure fabulation would be something people would do on the regular.

techas•7mo ago
I saw this recently on some congress abstracts. I think it is just AI generated content. References look real and don’t exist.
PicassoCTs•7mo ago
To imagine this driving a singularity, meanwhile its putting the final nail in science, together with paper-spam and research-reward decline. They are going to hang us tech-priests from the lamp-posts when the consequences of this bullshit artistry hit home.
haffi112•7mo ago
You would think that Springer did the due diligence here, but what is the value of a brand such as Springer if they let these AI slops through their cracks?

This is an opportunity for brands to sell verifiability, i.e., that the content they are selling has been properly vetted, which was obviously not the case here.

cess11•7mo ago
Why would one think that? All of the big journal publishers have had paper millers and fraudsters and endless amounts of "tortured phrases" under their names for a long, long time.
WillAdams•7mo ago
Back when I was doing academic publishing I'd use a regex to find all the hyperlinks, then a script (written by a co-worker, thanks again Dan!) to determine if they were working or no.

A similar approach should work w/ a DOI.

RossBencina•7mo ago
In the past I've had GPT4 output references with valid DOIs. Problem was the DOIs were for completely different (and unrelated) works. So you'd need to retrieve the canonical title and authors for the DOI and cross check it.
thoroughburro•7mo ago
And then make sure the arguments and evidence it presents are as the LLM represented them to be.
ofjcihen•7mo ago
At which point it’s more of a hassle to use an LLM than not.
rixed•7mo ago
And then check that the cited article was not itself an AI piece that managed to get published.
cyclecycle•7mo ago
A classic case.

I work on Veracity https://groundedai.company/veracity/ which does citation checking for academic publishers. I see stuff like this all the time in paper submissions. Publishers are inundated

rbanffy•7mo ago
Don’t publishers ban authors who attempt such shenanigans?
bumby•7mo ago
Not all journals require a DOI link for each reference. Most good ones do seem to have a system to verify the reference exists and is complete; I assume there’s some automation to that process but I’d love to hear from journal editorial staff if that’s really the case.
Vinayak_A_B•7mo ago
If I make a citation verifier, will conference/journal guys pay for it? First verify if the citation is legit, like the paper actually exists, after that another LLM that reads the paper cited and gives a rating out of 10, whether it fits the context or not. [ONLY FOR LIT SURVEY]
SiempreViernes•7mo ago
No, they aren't paying the reviewers in the first place.
codewench•7mo ago
So given that the output of an LLM is unreliable at best, your plan is to verify that a LLM didn't bullshit you by asking another LLM?

That sounds... counterproductive

thoroughburro•7mo ago
You’re offering to doublecheck measurements made with a bad ruler by using that same ruler.
PeterStuer•7mo ago
Given that the existence of a reference is fairly trivial to check, I'd wager the authors would not care enough to pay for this. As for 'fit', this is very much in the eye of the beholder and a paper can be cited for the most trivial part. Overcitation is usually not seen as a problem. Omitting citations the reviewer considers 'essential', often from their own lab or circles, is seen as non-negotiability.

So the better 'idea' would be to produce a CYA citation assistant that for a given paper adds all the remotely plausible references for all the known potential reviewers of a journal or conference. I honestly think this is not a hard problem, but doubt even that can be commercialized beyond Google Ads monetization.

amelius•7mo ago
So was the entire text machine-generated?

Or did they take a human-written text and asked a machine to generate references/citations for it?

passwordoops•7mo ago
Why would anyone write a book then ask for citations?
amelius•7mo ago
Because collecting/formatting citations is not the most fun part of the writing process (?)

And maybe the authors were over-confident in the capabilities of current AI.

ktallett•7mo ago
Many write and then find citations to fit what they said rather than write based on what citable sources suggest
zero_k•7mo ago
Springer? You mean the publisher we are currently fighting so they won't mess up our peer-reviewed research paper that we wrote and paid for the privilege for them to mess up (ehm, sorry "publish")? Colour me surprised.
dandanua•7mo ago
We are approaching publishers' heaven, where AI reviewers review AI written books and articles (with AI editors fixing their style), allowing publishers to keep collecting billions from essentially mandatory subscriptions from institutions.
flohofwoe•7mo ago
It's fine, because human readers will also be replaced with AI which produce a quick summary ;)
rbanffy•7mo ago
Or answer specific questions when needed.
dwayne_dibley•7mo ago
'Based on a tip from a reader, we checked 18 of the 46 citations in the book.' Why not just check them all?
maweki•7mo ago
They didn't just click a link. They contacted the supposed authors for comment. That would be a reason for not checking all of them.
b00ty4breakfast•7mo ago
This seems like the very thing that AI advocates would want to avoid. It certainly doesn't fill me, as an outsider to the whole thing, with much confidence for the future of AI-generated content but maybe I'm not the target sucker....err, I mean target demographic
Isamu•7mo ago
One of the potential uses of AI that I have most wanted is automated citation lookup and validation.

First check if the citation references a real thing. Then actually read and summarize the referenced text and give a confidence level that it says what was claimed.

But no, we have AI that are compounding the problem. That says something about unaligned incentives.

pyrale•7mo ago
> One of the potential uses of AI that I have most wanted is automated citation lookup and validation.

Also one of the things AI is likely the least suited for.

best I could imagine an AI can do is offer sources for you to check for a given citation.

Isamu•7mo ago
>Also one of the things AI is likely the least suited for.

I agree, if we are using the current idea of AI as language models.

But that’s very limiting. I’m old enough to remember when AI meant everything a human could do. Not just some subset that is being deceptively marketed as potentially the whole thing.

pyrale•7mo ago
The thing is, we have plenty of examples where systems that are not labeled as ai vastly outperform humans at searching knowledge bases.

A google-like search tool would probably be all you need for citations if 1) Google hadn’t killed accuracy in favor of ad placement and 2) the rest of the world hadn’t poisoned the dataset in hopes of getting picked up by Google as a first page result.

Now the hard problem isn’t to search so much as to curate an un-polluted dataset.

misja111•7mo ago
"He (the author) did not answer our questions asking if he used an LLM to generate text for the book. However, he told us, “reliably determining whether content (or an issue) is AI generated remains a challenge, as even human-written text can appear ‘AI-like.’ This challenge is only expected to grow, as LLMs … continue to advance in fluency and sophistication.”

Lol, that answer sounds suspiciously much like LLM generated as well ..

DebtDeflation•7mo ago
It's true that "AI detection algorithms" are not particularly reliable.

It's also true that if you have fake CITATIONS in your works that such algorithms aren't necessary to know the work is trash - either it was written by AI or you knowingly faked your research and it doesn't really matter which.

MengerSponge•7mo ago
My "Plagiarism Machine #1 Fan" shirt has people asking a lot of questions already answered by my shirt.
PeterStuer•7mo ago
Would it be possible to 'squat' the non existent references and turbo boost oneself into 'most cited author' territory? :)
rbanffy•7mo ago
Use AI to do it, then write a paper about what you did.
MarlonPro•7mo ago
Bad news for old-school people who still love books as a learning resource.
alok-g•7mo ago
A next development would be people developing and using citation checkers. That would fix just that problem. The deeper underlying quality problem with statements in the text often remaining unverified/incorrect would remain unfixed.

If the authors are to manually and genuinely put some citations, chances would be higher that they are familiar with the cited work and the statement for which the work is cited is actually corroborated by the citation.