frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

The Contagious Taste of Cancer

https://www.historytoday.com/archive/history-matters/contagious-taste-cancer
1•Thevet•1m ago•0 comments

U.S. Jobs Disappear at Fastest January Pace Since Great Recession

https://www.forbes.com/sites/mikestunson/2026/02/05/us-jobs-disappear-at-fastest-january-pace-sin...
1•alephnerd•1m ago•0 comments

Bithumb mistakenly hands out $195M in Bitcoin to users in 'Random Box' giveaway

https://koreajoongangdaily.joins.com/news/2026-02-07/business/finance/Crypto-exchange-Bithumb-mis...
1•giuliomagnifico•1m ago•0 comments

Beyond Agentic Coding

https://haskellforall.com/2026/02/beyond-agentic-coding
1•todsacerdoti•3m ago•0 comments

OpenClaw ClawHub Broken Windows Theory – If basic sorting isn't working what is?

https://www.loom.com/embed/e26a750c0c754312b032e2290630853d
1•kaicianflone•4m ago•0 comments

OpenBSD Copyright Policy

https://www.openbsd.org/policy.html
1•Panino•5m ago•0 comments

OpenClaw Creator: Why 80% of Apps Will Disappear

https://www.youtube.com/watch?v=4uzGDAoNOZc
1•schwentkerr•9m ago•0 comments

What Happens When Technical Debt Vanishes?

https://ieeexplore.ieee.org/document/11316905
1•blenderob•10m ago•0 comments

AI Is Finally Eating Software's Total Market: Here's What's Next

https://vinvashishta.substack.com/p/ai-is-finally-eating-softwares-total
2•gmays•11m ago•0 comments

Computer Science from the Bottom Up

https://www.bottomupcs.com/
2•gurjeet•11m ago•0 comments

Show HN: I built a toy compiler as a young dev

https://vire-lang.web.app
1•xeouz•13m ago•0 comments

You don't need Mac mini to run OpenClaw

https://runclaw.sh
1•rutagandasalim•14m ago•0 comments

Learning to Reason in 13 Parameters

https://arxiv.org/abs/2602.04118
1•nicholascarolan•16m ago•0 comments

Convergent Discovery of Critical Phenomena Mathematics Across Disciplines

https://arxiv.org/abs/2601.22389
1•energyscholar•16m ago•1 comments

Ask HN: Will GPU and RAM prices ever go down?

1•alentred•16m ago•0 comments

From hunger to luxury: The story behind the most expensive rice (2025)

https://www.cnn.com/travel/japan-expensive-rice-kinmemai-premium-intl-hnk-dst
2•mooreds•17m ago•0 comments

Substack makes money from hosting Nazi newsletters

https://www.theguardian.com/media/2026/feb/07/revealed-how-substack-makes-money-from-hosting-nazi...
5•mindracer•18m ago•2 comments

A New Crypto Winter Is Here and Even the Biggest Bulls Aren't Certain Why

https://www.wsj.com/finance/currencies/a-new-crypto-winter-is-here-and-even-the-biggest-bulls-are...
1•thm•18m ago•0 comments

Moltbook was peak AI theater

https://www.technologyreview.com/2026/02/06/1132448/moltbook-was-peak-ai-theater/
1•Brajeshwar•19m ago•0 comments

Why Claude Cowork is a math problem Indian IT can't solve

https://restofworld.org/2026/indian-it-ai-stock-crash-claude-cowork/
2•Brajeshwar•19m ago•0 comments

Show HN: Built an space travel calculator with vanilla JavaScript v2

https://www.cosmicodometer.space/
2•captainnemo729•19m ago•0 comments

Why a 175-Year-Old Glassmaker Is Suddenly an AI Superstar

https://www.wsj.com/tech/corning-fiber-optics-ai-e045ba3b
1•Brajeshwar•19m ago•0 comments

Micro-Front Ends in 2026: Architecture Win or Enterprise Tax?

https://iocombats.com/blogs/micro-frontends-in-2026
2•ghazikhan205•22m ago•1 comments

These White-Collar Workers Actually Made the Switch to a Trade

https://www.wsj.com/lifestyle/careers/white-collar-mid-career-trades-caca4b5f
1•impish9208•22m ago•1 comments

The Wonder Drug That's Plaguing Sports

https://www.nytimes.com/2026/02/02/us/ostarine-olympics-doping.html
1•mooreds•22m ago•0 comments

Show HN: Which chef knife steels are good? Data from 540 Reddit tread

https://new.knife.day/blog/reddit-steel-sentiment-analysis
1•p-s-v•23m ago•0 comments

Federated Credential Management (FedCM)

https://ciamweekly.substack.com/p/federated-credential-management-fedcm
1•mooreds•23m ago•0 comments

Token-to-Credit Conversion: Avoiding Floating-Point Errors in AI Billing Systems

https://app.writtte.com/read/kZ8Kj6R
1•lasgawe•23m ago•1 comments

The Story of Heroku (2022)

https://leerob.com/heroku
1•tosh•23m ago•0 comments

Obey the Testing Goat

https://www.obeythetestinggoat.com/
1•mkl95•24m ago•0 comments
Open in hackernews

Just how resilient are large language models?

https://www.rdrocket.com/blog/just-how-resilient-are-large-language-models
57•decodebytes•4mo ago

Comments

dataking•4mo ago
Archive link: https://archive.is/0Bl3Z
rglover•4mo ago
Interesting read. Was surprised to learn how much damage can be done to a model's parameters without making any discernible difference in its quality of output.
cma•4mo ago
I didn't see any mention of dropout in the article, during training parameters or whole layers are removed in different places which helps force it into a distributed representation.
danielmarkbruce•4mo ago
dropout has sort of dropped out... not used much in LLM training anymore
cma•4mo ago
I didn't realize that, looks like it isn't used much at all anymore except in finetuning
rglover•4mo ago
"Lessons from the Real World" and "The Limits of Resilience" discussed this.
cma•4mo ago
Not quite, that seems to be about quantizing and dropping out after training, not random dropout throughout training.

But apparently that isn't done much any more and has been partly superseded with things like weight decay.

snats•4mo ago
I also did a couple of experiments with pruning LLMs[1] using genetic algorithms and you can just keep removing a surprising amount of layers in big models before they start to have a stroke.

[1]https://snats.xyz/pages/articles/pruningg.html

kridsdale1•4mo ago
I suspect this applies to human beings as well.
littlestymaar•4mo ago
Well, humans don't have “layers” in the first place…
Legend2440•4mo ago
Of course not, that’s ogres.
TeMPOraL•4mo ago
There's https://en.wikipedia.org/wiki/Hydrocephalus# and there are cases of people living normal lives, not realizing they're missing most of their brains, until this gets discovered on some unrelated medical test. Or people who survived an unplanned brain surgery by rebar or bullet. Etc.
user_7832•4mo ago
(Old man shouting at clouds rant incoming)

I think it's kinda ironic (in a more meta Alanis Morissette way) for such an article that has interesting content to default to have an LLM write it. Please, I request authors - you're better than this, and many people actually want to hear you, not an LLM.

For example, what really is the meaning of this sentence?

> These aren't just storage slots for data, they're the learned connections between artificial neurons, each one encoding a tiny fragment of knowledge about language, reasoning, and the patterns hidden in human communication.

I thought parameters were associated with connections, is the author implying that they also store data? Is the connection itself the stored data? Is there non-connective information that stores data? Or non-data-storage things that have connectivity aspects?

I spent a solid amount of time trying to understand what was being told, but thanks to what I would call a false/unnecessary "not just x but y" troupe, I unfortunately lost the plot.

IMO, a human who's a good writer would have a sentence that's clearer to understand, while non advanced writers (including me, almost certainly) would simply degrade gracefully to simpler sentence structure.

potsandpans•4mo ago
Is there anything on the article that explicitly indicates that it was written by an llm?
simonw•4mo ago
"These aren't just X, they're Y" is a pretty strong tell these day.

Wikipedia has an excellent article about identifying AI-generated text. It calls that particular pattern "Negative parallelisms". https://en.m.wikipedia.org/wiki/Wikipedia:Signs_of_AI_writin...

gryfft•4mo ago
It's funny, negative parallelisms used to be a favorite gimmick of mine, back in the before-times. Nowadays, every time I see "it isn't just..." I feel disappointment and revulsion.

The really funny thing is, I'll probably miss that tell when it's gone, as every AI company eventually scrubs away obvious blemishes like that from their flagship models.

user_7832•4mo ago
Very similar here. I wouldn't use it myself but I'd definitely have liked to read such "punchy" text. The interesting thing is that I'm now much more consciously noticing how many em dashes and other similar phrases say an Agatha Christie used to use... which makes the source of such LLM data even more obvious hehe.

Interestingly, there are still many other good literary devices that are not yet used by AI - for example, sentences of varying lengths. There is still scope for a good human editor to easily outdo LLMs... but how many will notice? Especially when one of the editorial coloumns in the NYT (or Atlantic, forgot exactly) is merrily using LLMs in the heartfelt advice coloumn. It's really ironic, isn't it? ;)

Timwi•4mo ago
I use em dashes in my regular writing — mostly because I consider them good/correct typography — and now I have to contend with being accused of using LLMs or being a bot. I think this societal development is damaging.
adlumal•4mo ago
Could this be a symptom of the free tier of ChatGPT, but not all LLMs? I’ve recently been a heavy user of Anthropic’s Claude and I don’t believe I’ve seen too many of these in my chats. Though this may be because I haven’t asked Claude to write Wikipedia articles.

LLMs are also great at following style, not via criteria but via examples. So this is something that’s easily overcome.

I discovered this when I made an error in a creative writing tool I was working on. I told it to follow the writing style of existing story text, but it ended up making the system messages follow the same style. It was quite amusing to see tool messages and updates written in an increasingly enthusiastic Shakespearean/etc prose (so I left it unfixed!)

Timwi•4mo ago
I am dismayed by the implication that humans are now no longer able to use certain tropes or rhetorical devices or stylistic choices without immediately being discounted as LLMs and having their entire writing work discredited.
ardupper•4mo ago
Came here to say just this
TeMPOraL•4mo ago
In context, I think this sentence is pretty nicely written, and sounds like something I'd expect of a human.

It's basically saying: LLMs are not structured like databases, their knowledge is not localized. In a database, a fact lives in a specific place; cut that out, lose the fact. In an LLM, every thing the model learned in training is smeared across the entire network; cut out a small piece, and not much is lost.

As a crude analogy, think of a time vs. frequency domain representation of a signal, like a sound or an image. The latter case is particularly illustrative (!). Take a raster image, pick a pixel. Where is the data (color) of that pixel located in the Fourier transform of the image? A little bit in every single pixel of the transformed image. Conversely, pick a block of pixels in the Fourier transform, blank them out, and transform back to "normal" image - you'll see entire image got blurred or sharpened, depending on which frequencies you just erased.

So in terms of data locality, a database is to an LLM kinda what an image is to its Fourier transform.

(Of course, the other thing the sentence is trying to communicate is, LLMs are not (a different representation of) databases - they don't learn facts, they learn patterns, which encode meanings.)

Centigonal•4mo ago
For people who want to dig deeper: The fancy ML term-of-art for the practice of cutting out a piece of a neural network and measuring the resulting effect on its performance is an ablation study.

Since around 2018, ablation has been an important tool to understand the structure and function of ML models, including LLMs. Searching for this term in papers about your favorite LLMs is a good way to learn more.

magicalhippo•4mo ago
It's my understanding that dropout[1] is also an important aspect of training modern neural nets.

When using dropout you intentionally remove some random number of nodes ("neurons") from the network during a training step.

By constantly changing which nodes are dropped during training, you effectively force delocalization and so it seems to me somewhat unsurprising that the resulting network is resilient to local perturbations.

[1]: https://towardsdatascience.com/dropout-in-neural-networks-47...

littlestymaar•4mo ago
> The redundancy we observe in language models might also explain why these systems can generalize so effectively to tasks they were never explicitly trained on

What year was this written? 2023 and reposted in 2025? Or is the author unaware that the generalization promises of early GPT have failed to materialize and that all model makers actually have been training models explicitly on the most common tasks people use them for via synthetic data generation, which has driven the progress of all models over the past few years.

dasaia•4mo ago
>that all model makers actually have been training models explicitly on the most common tasks people use them for via synthetic data generation

People really don't understand that part in general.

I find the easiest way to make people understand is to write gibberish that will trigger the benchmaxxed "pattern matching" behavior like this:

> The child and the wolf try to enjoy a picnic by the river but there's a sheep. The boat needs to connect nine dots over the river without leaving the water but gets into an accident and dies. The surgeon says "I can't operate on this child!" why?

The mix and matching of multiple common riddles/puzzles style questions into a singular gibberish sentence should, if models had legitimate forms of reasoning, make the model state that this is nonsense, at best, or answer chaotically at worst. Instead, they will all answer "The surgeon is the mother" even though nobody even mentioned anything about anyone's gender. That's because that answer, "the surgeon is the mother", for the gender bias riddle has been burned so hard into the models they cannot reply in any other way as soon as they pattern match "The surgeon can't operate on this child". No matter how much crap you wrote before that sentence. You can change anything about what comes before "The surgeon" and the model will almost invariably fall into giving an answer like this one (Gemini 2.5 pro):

https://i.imgur.com/ZvsUztz.png

>The details about the wolf, the sheep, the picnic, and the dying boat are all distractions (red herrings) to throw you off. The core of the puzzle is the last sentence.

>The surgeon says, "I can't operate on this child!" because the surgeon is the child's mother.

One could really question the value, by the way, of burning the answer to so many useless riddles into LLMs. The only purpose it could serve is gaslighting the average person asking these questions into believing there's some form of intelligence in there. Obviously they fail so hard to generalize on this (never working quite right when you change an element of a riddle into something new) that from a practical use point of view, you might as well not bother have this in the training data, nobody's going to be more productive because LLMs can act as a database for the common riddles.

visarga•4mo ago
> The child and the wolf try to enjoy a picnic by the river but there's a sheep. The boat needs to connect nine dots over the river without leaving the water but gets into an accident and dies. The surgeon says "I can't operate on this child!" why?

Just tested it and it actually fools Claude on first try! LOL, so much for reasoning models.

KronisLV•4mo ago
For the fun of it, Qwen 3 Coder 480B (in Jan.ai, the model is on Cerebras):

  This sounds like a riddle combining elements from different classic puzzles and stories. Let me think through it:
  The answer is likely: "Because the child is the surgeon's own child!"
  This is a variation of the classic riddle where a father and son are in a car accident, the father dies, and the surviving son needs surgery. When the surgeon sees the child, they say "I can't operate on this child - he's my son!"
  The twist relies on the assumption that surgeons are male, but of course the surgeon could be the child's mother.
  However, I'm a bit confused by the "nine dots" and "boat" elements in your version - those don't typically appear in this riddle. Could you clarify if there's a specific version you're thinking of, or if you meant this to be a different puzzle entirely?
Doxin•4mo ago
For what it's worth GPT-OSS-20b thinks about the puzzle for a LONG time and then comes up with a... solution of sorts? It doesn't peg the puzzle as not making any sense, but at least it tries to solve the puzzle presented, and doesn't just spit out a pre-made answer:

> It turns out the “child” isn’t a patient waiting for an operation at all – the child has already been lost.

> In the story the boy and his wolf friend go to the river for a picnic with a sheep that happens to be there. They decide to use a small boat to cross the water. The problem is that the boat must stay on the surface of the water while it “connects” the nine points (dots) across the river – essentially it has to stay on the river without ever leaving it, which makes a safe crossing impossible.

> During the attempt the boat hits something and sinks; everyone in it dies. The surgeon who arrives at the scene says, “I can’t operate on this child!” because the child is already dead from the accident. The mention of the wolf and the sheep is simply part of the scene that led to the fatal crossing; it isn’t relevant to the medical impossibility.

Interestingly in its thought process it does come across the classic puzzles, but discards them as not quite fitting:

> Maybe it's about the classic lateral thinking puzzle: "A man is found dead, he was a surgeon. The surgeon said 'I cannot operate on this child because the child is my own daughter', etc." But not.

> Alternatively maybe it's about the famous "Nine Dots" puzzle: Connect nine dots with four straight lines without lifting pencil. Here boat connects nine dots over river... So maybe it's the "connect the dots" game but with a boat?

> Could this be a riddle about "The River and the Sheep" referencing a children's rhyme or fable? Maybe it's about "Jack and Jill"? Not sure.

and so on and on. When asked if the puzzle makes sense it largely concludes that it doesn't.

It's definitely interesting to see which LLMs fall for what pitfalls. It's far from universal as far as I can tell. GPT-OSS-2b definitely has some wonky logic in it's answer, but at least it's not assuming it's a puzzle it knows the answer to. Gemma-3-27b immediately pegs that the puzzles elements seem disconnected, waffles on a bit, and then also comes to the conclusion the child is already dead, discarding roughly all of the puzzle as distracting facts. llama-3.2-1b (a very small model) immediately a) misunderstands the riddle and b) tells you it doesn't have enough information to solve the riddle. When pressed it "solves" the riddle thus:

> The surgeon says "I can't operate on this child!" because the child is in a precarious position on the boat, with only nine lines connecting them to other points on the riverbank. If you try to attach any two lines that form an "X" shape (i.e., two lines connected by a single point), it would create a triangle that would leave space for another line to connect one of the child's dots to a nearby point on the riverbank, allowing it to be attached.

> The surgeon is not saying that the child is in immediate danger or can't be saved. Instead, they're suggesting that there might be an alternative solution where all nine lines can be connected without leaving the water. However, this would require some creative problem-solving and flexibility with the geometry of the situation.

I fully did expect at least llama-3.2-1b to fall for this sort of context-baiting, but it seems even a small model like that managed to figure out that there's something nonstandard about the riddle.