frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Kernel bugs hide for 2 years on average. Some hide for 20

https://pebblebed.com/blog/kernel-bugs
79•kmavm•2h ago•31 comments

Sugar industry influenced researchers and blamed fat for CVD (2016)

https://www.ucsf.edu/news/2016/09/404081/sugar-papers-reveal-industry-role-shifting-national-hear...
709•aldarion•14h ago•418 comments

Tailscale state file encryption no longer enabled by default

https://tailscale.com/changelog
248•traceroute66•8h ago•100 comments

Eat Real Food

https://realfood.gov
658•atestu•11h ago•954 comments

Play Aardwolf MUD

https://www.aardwolf.com/
93•caminanteblanco•5h ago•40 comments

Shipmap.org

https://www.shipmap.org/
510•surprisetalk•13h ago•86 comments

Fighting back against biometric surveillance at Wegmans

https://blog.adafruit.com/2026/01/07/dont-let-the-grocery-store-scan-your-face-a-guide-to-fightin...
166•ptorrone•3h ago•101 comments

How dependabot works

https://nesbitt.io/2026/01/02/how-dependabot-actually-works.html
40•zdw•5d ago•4 comments

Chase to become new issuer of Apple Card

https://www.jpmorganchase.com/ir/news/2026/chase-to-become-new-issuer-of-apple-card
8•vismit2000•1h ago•0 comments

Musashi: Motorola 680x0 emulator written in C

https://github.com/kstenerud/Musashi
34•doener•3h ago•1 comments

The Q, K, V Matrices

https://arpitbhayani.me/blogs/qkv-matrices/
71•yashsngh•20h ago•33 comments

US will ban Wall Street investors from buying single-family homes

https://www.reuters.com/world/us/us-will-ban-large-institutional-investors-buying-single-family-h...
715•kpw94•9h ago•778 comments

Solo ASIC tapeout on a budget: detailed write up

https://old.reddit.com/r/chipdesign/comments/1q4kvxt/solo_asic_tapeout_on_a_budget_detailed_write...
44•random_duck•2d ago•7 comments

NPM to implement staged publishing after turbulent shift off classic tokens

https://socket.dev/blog/npm-to-implement-staged-publishing
148•feross•10h ago•41 comments

LaTeX Coffee Stains (2021) [pdf]

https://ctan.math.illinois.edu/graphics/pgf/contrib/coffeestains/coffeestains-en.pdf
301•zahrevsky•14h ago•72 comments

The virtual AmigaOS runtime (a.k.a. Wine for Amiga:)

https://github.com/cnvogelg/amitools/blob/main/docs/vamos.md
28•doener•3h ago•3 comments

Open Infrastructure Map

https://openinframap.org
6•efskap•1h ago•1 comments

GLSL Web CRT Shader

https://blog.gingerbeardman.com/2026/01/04/glsl-web-crt-shader/
20•msephton•3d ago•1 comments

Health care data breach affects over 600k patients, Illinois agency says

https://www.nprillinois.org/illinois/2026-01-06/health-care-data-breach-affects-600-000-patients-...
172•toomuchtodo•12h ago•63 comments

Notion AI: Unpatched data exfiltration

https://www.promptarmor.com/resources/notion-ai-unpatched-data-exfiltration
132•takira•8h ago•17 comments

Reading Without Limits or Expectations

https://www.carolinecrampton.com/reading-without-limits-or-expectations/
6•herbertl•2d ago•0 comments

We found cryptography bugs in the elliptic library using Wycheproof

https://blog.trailofbits.com/2025/11/18/we-found-cryptography-bugs-in-the-elliptic-library-using-...
73•crescit_eundo•6d ago•6 comments

Show HN: I visualized the entire history of Citi Bike in the browser

https://bikemap.nyc/
73•freemanjiang•9h ago•27 comments

Creators of Tailwind laid off 75% of their engineering team

https://github.com/tailwindlabs/tailwindcss.com/pull/2388
1064•kevlened•12h ago•615 comments

A4 Paper Stories

https://susam.net/a4-paper-stories.html
309•blenderob•15h ago•149 comments

How Google got its groove back and edged ahead of OpenAI

https://www.wsj.com/tech/ai/google-ai-openai-gemini-chatgpt-b766e160
82•jbredeche•12h ago•78 comments

“Stop Designing Languages. Write Libraries Instead” (2016)

https://lbstanza.org/purpose_of_programming_languages.html
232•teleforce•16h ago•204 comments

Native Amiga Filesystems on macOS / Linux / Windows with FUSE

https://github.com/reinauer/amifuse
82•doener•4d ago•33 comments

2026 Predictions Scorecard

https://rodneybrooks.com/predictions-scorecard-2026-january-01/
40•calvinfo•7h ago•29 comments

Meditation as Wakeful Relaxation: Unclenching Smooth Muscle

https://psychotechnology.substack.com/p/meditation-as-wakeful-relaxation
140•surprisetalk•13h ago•100 comments
Open in hackernews

On the slow death of scaling

https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5877662
109•sethbannon•1d ago
pdf: https://download.ssrn.com/2026/1/6/5877662.pdf?response-cont...

Comments

octoberfranklin•1d ago
> Academia has been marginalized from meaningfully participating in AI progress and industry labs have stopped publishing

Exactly like semiconductor wafer processing.

random3•1d ago
If anyone believes they're close to a generalization end-game wrt to AI capabilities, it makes no sense to do anything that could impact their advantage, by enabling others to compete. Collaboration makes sense on timeframes that don't imply zero-sum games.

Board games like the Settlers of Catan are good examples of the behavior— concretely the start of the game when everyone trades vs the end of the game when if you suspect someone wins it makes little sense to trade unless, you think it will help you win first.

SecretDreams•1d ago
> Collaboration makes sense on timeframes that don't imply zero-sum games.

People are fooling themselves if they think AGI will be zero sum. Even if only one group somehow miraculously develops it, there will immediately be fast followers. And, the more likely scenario is more than one group would independently pull it off - if it's even possible.

random3•23h ago
Maybe, but at least Open AI, XAI and any Bostrom believer thinks this is the case.

Ilya Sutskever (Sep 20, 2017)

> The goal of OpenAI is to make the future good and to avoid an AGI dictatorship. You are concerned that Demis could create an AGI dictatorship. So do we. So it is a bad idea to create a structure where you could become a dictator if you chose to, especially given that we can create some other structure that avoids this possibility.

Nick Bostrom - Decisive Strategic Advantage https://www.lesswrong.com/posts/vkjWGJrFWBnzHtxrw/superintel...

refulgentis•21h ago
Bostrom / Ilya are agreeing with the GPs argument AFAICT: it's not that AGI can create a dictatorship-of-the-first-AGI-owner, it's that only having one serious funded lab going at it creates a knowledge gap of N years that could give said lab escape velocity*

* imagine if Google alone had LLMs. For an innocuous example, the only provider in my LLM client that regularly fails unit tests verifying they actually cache tokens and utilize them on a subsequent request is Gemini. I used to work at Google and it'd be horrible for that too-big-for-its-own-good institution regressing to the corporate mean to own LLMs all by itself

pixl97•12h ago
>if it's even possible.

Why do people keep repeating this. The only way artificial intelligence is impossible is if intelligence is impossible. And we're here so that pretty much removes that impediment.

ironbound•1d ago
It's a mistake to publish papers without a code repo, with the majority of new ML paper's being noise at best.
timy2shoes•22h ago
This is an opinion piece, similar to her infamous hardware lottery paper. We shouldn't expect a repo on an opinion piece.
rvz•14h ago
So DeepMind (who almost always releases papers without code repositories) are mistakes then?
bicepjai•22h ago
Hooker’s argument lands for me because it ties the technical scaling story to institutional incentives: as progress depends more on massive training runs, it becomes capital-intensive, less reproducible and more secretive; so you get a compute divide and less publication.

I’m trying to turn that into something testable with a simple constraint: “one hobbyist GPU, one day.” If meaningful progress is still possible under tight constraints, it supports the idea that we should invest more in efficiency/architecture/data work, not just bigger runs.

My favorite line >> Somewhat humorously, the acceptance that there are emergent properties which appear out of nowhere is another way of saying our scaling laws don’t actually equip us to know what is coming.

Regarding this paragraph >> 3.3 New algorithmic techniques compensate for compute. Progress over the last few years has been as much due to algorithmic improvements as it has been due to compute. This includes extending pre-training with instruction finetuning to teach models instruction following ..., model distillation using synthetic data from larger more performant "teachers" to train highly capable, smaller "students" ..., chain-of-thought reasoning ..., increased context-length ..., retrieval augmented generation ... and preference training to align models with human feedback ...

I would consider algorithmic improvements to be the following 1. architecture like ROPE, MLA 2. efficiency using custom kernels

The errors in the paper 1. Transformers for language modeling (Vaswani et al., 2023). => this shd be 2017

Disclosure: my proposed experiments: https://ohgodmodels.xyz/

bigbadfeline•7h ago
> as progress depends more on massive training runs, it becomes capital-intensive, less reproducible and more secretive; so you get a compute divide and less publication.

In the area of AI, secrecy and inability to reproduce/verify can become a huge systemic and social problem, the possible damage is literally unbounded.

That's why I like open source AI, including training data and process, it solves the above problem as well as the problem of duplication of effort which leads to a huge waste of resources, waste that is economically significant on national and global scales.

gdiamos•20h ago
It was an interesting read Sara, thanks for sharing it.

I especially agree with your point that scaling laws really killed open research. That's a shame and I personally think we could benefit from more research.

I originally didn't like calling them scaling laws.

In addition to the law part seeming a bit much, I've found that researchers often overemphasize the scale part. If scaling is predictable, then you don't need to do most experiments at very large scale. However, that doesn't seem to stop researchers from starting there.

Once you find something good, and you understand how it scales, then you can pour system resources into it. So I originally thought it would encourage research. I find it sad that it seems to have had the opposite effect.

charcircuit•20h ago
>the acceptance that there are emergent properties which appear out of nowhere is another way of saying our scaling laws don’t actually equip us to know what is coming.

Is this actually accepted? Ever since [0], I thought people recognized that they don't appear out of nowhere.

[0] https://arxiv.org/pdf/2304.15004

red75prime•14h ago
"Appear out of nowhere" looks like a straw-man. Anyway, there are newer papers. For example "Emergent Abilities in Large Language Models: A Survey"[0]

[0] https://arxiv.org/abs/2503.05788

Zigurd•13h ago
I was struck by this in the abstract: The scaling of these models, accomplished by increasing the number of parameters and the magnitude of the training datasets, has been linked to various so-called emergent abilities that were previously unobserved. These emergent abilities, ranging from advanced reasoning and in-context learning to coding and problem-solving...

In my experience with agent assisted coding, how well it works seems very closely tied to the quantity and quality of training material. It also has some identifiable qualities like verifiability that make it a particularly good target for an LLM. I would not call that surprising or emergent.

gwern•13h ago
> I thought people recognized that they don't appear out of nowhere.

I don't think that paper is widely accepted. Have you seen the authors of that paper, or anyone else, use it to successfully predict (rather than postdict) anything?

godelski•5h ago
While I'm a big fan of that paper, as a ML researcher I can confidently state that it is not well accepted. I can also confidently state that it is not well known.

I think there is a critical flaw in the paper though. Not critical from a technical standpoint, but from a reviewer standpoint. They don't bridge the gap to the final step of transitioning to a hard loss. But you can easily experiment with this by yourself even on smaller models and datasets and it is pretty effective. I think the logic is quite straight forward though and this isn't actually necessary to prove their point, which is why I think they didn't do it. But most ML people are hyperfixated on benchmarks and empirical evidence. Hell, that's why we kinda killed small scale research. It isn't technically wrong to ask for more scale and more datasets but these types of questions are unbounded so can be used too much as a crutch.

FWIW I also think the original Emergent Abilities paper has a critical flaw. Look at their definition (emphasis my own)

  Specifically, we define emergent abilities of large language models as abilities that are not present in smaller-scale models but are present in large-scale models; ***thus they cannot be predicted by simply extrapolating the performance improvements on smaller-scale models.***
Certainly the mirage paper counters this. Most critiques I've heard are about hard loss vs soft loss, but that isn't that important. But what I think most people don't realize is how the loss landscape actually works. Why I like the mirage paper so much is it is really saying that the loss landscape is partially defined by the number of model parameters (something we already knew btw).

But I also don't know why we've accepted this definition of emergent abilities. It isn't useful.

Without their explicit distinction of extrapolating we'd call nearly every model emergent. Here's my proof: for any given model there is almost surely a smaller model that performs worse. Dumb, but that's the problem with the definition. But using their distinction we run into the problem of concluding things are emergent simply because we didn't realize we were doing things a certain way.

And using the more classic definition of emergence[0,1], distinguishing between strong and weak, we should recognize that all neural nets are by definition weakly emergent. (Emphasis from [0])

  high-level phenomenon is *strongly emergent* with respect to a low-level domain when the high-level phenomenon arises from the low-level domain, but truths concerning that phenomenon are not *deducible* even in principle from truths in the low-level domain.

  high-level phenomenon is *weakly emergent* with respect to a low-level domain when the high-level phenomenon arises from the low-level domain, but truths concerning that phenomenon are *unexpected* given the principles governing the low-level domain.
In physics we have plenty of examples of weakly emergent phenomena and no examples of strongly emergent phenomena. Though we do have things in suspect. Clearly neural nets (and arguably even GLMs and many other techniques) follow this. Especially as we have no formal theory. But that's also why physics only has things that are suspect. Weak emergence is not surprising to a neural network setting and I don't think discussion about it is generally productive.

But strong emergence requires a very difficult proof. We must prove that we aren't just so dumb that we do not know how to deduce the results but that we cannot deduce the results. It means there must be a process that results in an unrecoverable information loss. I think everyone should be quite suspicious of any claims of strong emergence when it comes to AI. I mean... we have the weights... so the results are de fact deducible...

So I don't know why we talk about emergence the way we do in ML. I frequently hear people say things are emergent phenomena because they didn't realize they were teaching the model certain capabilities but that doesn't mean someone else wouldn't be able to (and boy are there many "emergent phenomena" that ML people "can't" predict but a mathematician would).

[0] https://consc.net/papers/emergence.pdf

[1] https://arxiv.org/abs/2410.15468

wiz21c•19h ago
FTA:

"One thing is certain, is the less reliable gains from compute makes our purview as computer scientists interesting again. We can now stray from the beaten path of boring, predictable gains from throwing compute at the problem."

Isn't Ilya Sutskever who said some months ago that we were going back to research ?

officialchicken•18h ago
Let's not forget about the myriad of basic problems that still remain - like deploying data, caching/distribution, and server resilience.

There is absolutely NO reason why that PDF shouldn't load today.

Haaargio•15h ago
Its not dying slowly right now at all.

Compute is a massive driver for everything ML. From number of experiments you can run in paralle, to how much RL you can try out, how long stuff is running etc.

ML is pushing scaling on dimensions we haven't had before (number of Datacenters, amount of energy we put into them) and ML is currently seen as the holy grail.

But i'm definitly very very curious how this compute and current progress is playing out in the next few years. It could be that we hit a hard ceiling were every single % point becomes tremendesly costly before we hit a % point of benchmark archievements which makes all of that usable daily. OR we will se a significant change to our society.

I do not think its something in between tbh because it def feels like in an expoential progress curve we are currently in.

drob518•15h ago
I suspect scaling will not die a slow death, but rather slow for a while and then all at once. Further, I think we’re at the knee. We know scaling doesn’t work for resolving the fundamental issues we have with large models at this point. If it did, the latest models would have solved the issues. Now, we’re in the acceptance phase. That’s not technical, it’s human psychology. People who made bold claims and huge promises that things would get better if we just spent a few more billion dollars on data centers and GPUs need to unwind those claims and find a way to save face.
zerosizedweasle•15h ago
https://am.jpmorgan.com/us/en/asset-management/institutional...
bicepjai•10h ago
This is my favorite line in the piece >> a "metaverse moment" for hyperscaler profits after $1.3 trillion of capex and R&D
FuriouslyAdrift•13h ago
Iron law of efficiency: a system will always expand to use all resources available to it.

You want to make an existing system more efficient, then take away resources.

tbrownaw•12h ago
> A pervasive belief in scaling has resulted in a massive windfall in capital for industry labs and fundamentally reshaped the culture of conducting science in our field.

People spend money on this because it works. It seems odd to call observable reality a "pervasive belief".

> Academia has been marginalized from meaningfully participating in AI progress and industry labs have stopped publishing.

Firstly, I still see news items about new models that are supposed to do more with less. If these are neither from academia nor industry, where are they coming from?

Secondly, "has been marginalized"? Really? Nobody's going to be uninterested in getting better results with less compute spend, attempts have just had limited effectiveness.

.

> However, it is unclear why we need so many additional weights. What is particularly puzzling is that we also observe that we can get rid of most of these weights after we reach the end of training with minimal loss

I thought the extra weights were because training takes advantage of high-dimensional bullshit to make the math tractable. And that there's some identifiable point where you have "enough" and more doesn't help.

I hadn't heard that anyone had a workable way to remove the extra ones after training, so that's cool.

.

.

The impression I had is that there's a somewhat-fuzzy "correct" number of weights and amount of training for any given architecture and data set / information content. And that when you reach that point is when you stop getting effort-free results by throwing hardware at the problem.

newsoftheday•12h ago
I read scaling and assumed it would be about scaling but seems to be about AI possibly but didn't read further.
empiko•10h ago
Scaling works, the problem is that it is practically impossible to scale much. There is only so much energy, text data, GPUs, etc. The folly of scaling is that we are living in a finite world. The huge investments in AI for the past few years are probably hitting the practical limits of scaling for now.

I also feel like most insiders were fully aware of this fact, but it was a neat sales pitch.