frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Root System Drawings

https://images.wur.nl/digital/collection/coll13/search
202•bookofjoe•7h ago•34 comments

Is Postgres read heavy or write heavy?

https://www.crunchydata.com/blog/is-postgres-read-heavy-or-write-heavy-and-why-should-you-care
37•soheilpro•1d ago•0 comments

Tinnitus Neuromodulator

https://mynoise.net/NoiseMachines/neuromodulationTonesGenerator.php
173•gjvc•5h ago•118 comments

Flowistry: An IDE plugin for Rust that focuses on relevant code

https://github.com/willcrichton/flowistry
106•Bogdanp•6h ago•17 comments

How to sequence your DNA for <$2k

https://maxlangenkamp.substack.com/p/how-to-sequence-your-dna-for-2k
13•yichab0d•1h ago•3 comments

What Dynamic Typing Is For

https://unplannedobsolescence.com/blog/what-dynamic-typing-is-for/
47•hit8run•4d ago•32 comments

Who invented deep residual learning?

https://people.idsia.ch/~juergen/who-invented-residual-neural-networks.html
53•timlod•5d ago•16 comments

./watch

https://dotslashwatch.com/
273•shrx•11h ago•74 comments

Solution to CIA’s kryptos sculpture is found in Smithsonian vault

https://www.nytimes.com/2025/10/16/science/kryptos-cia-solution-sanborn-auction.html
64•elahieh•2d ago•14 comments

Chen-Ning Yang, Nobel laureate, dies at 103

https://www.chinadaily.com.cn/a/202510/18/WS68f3170ea310f735438b5bf2.html
36•nhatcher•15h ago•12 comments

Using CUE to unify IoT sensor data

https://aran.dev/posts/cue/using-cue-to-unify-iot-sensor-data/
19•mvdan•8h ago•1 comments

Secret diplomatic message deciphered after 350 years

https://www.nationalarchives.gov.uk/explore-the-collection/the-collection-blog/secret-diplomatic-...
48•robin_reala•2d ago•4 comments

Titan submersible’s $62 SanDisk memory card found undamaged at wreckage site

https://www.tomshardware.com/pc-components/microsd-cards/tragic-oceangate-titan-submersibles-usd6...
68•WithinReason•1d ago•30 comments

Liva AI (YC S25) Is Hiring

https://www.ycombinator.com/companies/liva-ai/jobs/inrUYH9-founding-engineer
1•ashlleymo•4h ago

K8s with 1M nodes

https://bchess.github.io/k8s-1m/
51•denysvitali•1d ago•11 comments

Why the open social web matters now

https://werd.io/why-the-open-social-web-matters-now/
41•benwerd•4d ago•4 comments

Ripgrep 15.0

https://github.com/BurntSushi/ripgrep/releases/tag/15.0.0
277•robin_reala•7h ago•65 comments

New Work by Gary Larson

https://www.thefarside.com/new-stuff
466•jkestner•23h ago•122 comments

Coral NPU: A full-stack platform for Edge AI

https://research.google/blog/coral-npu-a-full-stack-platform-for-edge-ai/
72•LER0ever•2d ago•9 comments

When you opened a screen shot of a video in Paint, the video was playing in it

https://devblogs.microsoft.com/oldnewthing/20251014-00/?p=111681
89•birdculture•2d ago•9 comments

Ruby Blocks

https://tech.stonecharioteer.com/posts/2025/ruby-blocks/
163•stonecharioteer•4d ago•96 comments

Show HN: The Shape of YouTube

https://soy.leg.ovh/
14•hide_on_bush•6d ago•6 comments

SQL Anti-Patterns

https://datamethods.substack.com/p/sql-anti-patterns-you-should-avoid
187•zekrom•8h ago•134 comments

Picturing Mathematics

https://mathenchant.wordpress.com/2025/10/18/picturing-mathematics/
25•jamespropp•5h ago•0 comments

Lux: A luxurious package manager for Lua

https://github.com/lumen-oss/lux
46•Lyngbakr•8h ago•12 comments

Attention is a luxury good

https://seths.blog/2025/10/attention-is-a-luxury-good/
127•herbertl•5h ago•75 comments

Fast calculation of the distance to cubic Bezier curves on the GPU

https://blog.pkh.me/p/46-fast-calculation-of-the-distance-to-cubic-bezier-curves-on-the-gpu.html
103•ux•11h ago•22 comments

Our Paint – a featureless but programmable painting program

https://www.WellObserve.com/OurPaint/index_en.html
31•ksymph•6d ago•5 comments

Carbonized 1,300-Year-Old Bread Loaves Unearthed in Turkey

https://ancientist.com/1300-year-old-communion-bread-unearthed-in-karaman-a-loaf-for-the-farmer-c...
5•ilamont•5d ago•1 comments

AMD's Chiplet APU: An Overview of Strix Halo

https://chipsandcheese.com/p/amds-chiplet-apu-an-overview-of-strix
148•zdw•16h ago•55 comments
Open in hackernews

Who invented deep residual learning?

https://people.idsia.ch/~juergen/who-invented-residual-neural-networks.html
53•timlod•5d ago

Comments

aDyslecticCrow•3h ago
I thought it was ResNet that invented the technique, but it's interesting to see it rooted back through LSTM which feels like a very architecture. ResNet really made massive waves in the field, and it was hard finding a paper that didn't reference it for a while.
scarmig•3h ago
From the domain, I'm guessing the answer is Schmidhuber.
ansk•2h ago
Of all Schmidhuber's credit-attribution grievances, this is the one I am most sympathetic to. I think if he spent less time remarking on how other people didn't actually invent things (e.g. Hinton and backprop, LeCun and CNNs, etc.) or making tenuous arguments about how modern techniques are really just instances of some idea he briefly explored decades ago (GANs, attention), and instead just focused on how this single line of research (namely, gradient flow and training dynamics in deep neural networks) laid the foundation for modern deep learning, he'd have a much better reputation and probably a Turing award. That said, I do respect the extent to which he continues his credit-attribution crusade even to his own reputational detriment.
godelski•31m ago
I think one of the best things to learn from Schmidhuber is that progress involves a lot of players and over a lot of time. Attribution is actually a difficult game and usually we are only assigning credit to those at the end of some milestone. It's like giving a gold medal to the runner in the last leg of a relay race or focusing only on the lead singer of a band. It's never one person that does it alone. Shoulders of giants, but those giants are just a couple of dudes in a really big trenchcoat.

Another important lesson is that often good ideas get passed over because of hype or politics. We often like to pretend that science is all about the merit and what is correct. Unfortunately this isn't true. It is that way in the long run, but in the short run there's a lot of politics and humans still get in their own way. This is a solvable problem, but we need to acknowledge it and create systematic changes. Unfortunately a lot of that is coupled to the aforementioned one.

  > I do respect the extent to which he continues his credit-attribution crusade even to his own reputational detriment.

As should we all. Clearly he was upset that others got credit for his contributions. But what I do appreciate is that he has recognized that it is a problem bigger than him, and is trying to combat the problem at large and not just his own little battlefield. That's respectable.
dchftcs•10m ago
It's a bit of an aside but I believe this is one reason Zuckerberg's vision for establishing the superintelligence lab is misguided. Including VCs, too many people get distracted by rock stars in this gold rush.
ekjhgkejhgk•1h ago
I spent some time in the academia.

The person with whom an idea ends up associated often isn't the first person to have the idea. Most often is the person who explains why the idea is important, or find a killer application for the idea, or otherwise popularizes the idea.

That said, you can open what Schmidhuber would say is the paper which invented residual NNs. Try and see if you notice anything about the paper that perhaps would hinder the adoption of its ideas [1].

[1] https://people.idsia.ch/~juergen/SeppHochreiter1991ThesisAdv...

seanmcdirmid•1h ago
Surely they wrote some papers in English even if they wrote their dissertation in German? Most people don’t go straight to dissertations anyway, it’s more of a place to go after you read a much shorter paper.
ekjhgkejhgk•1h ago
Correct, that's [2]. In [2] they even say "[we] derive de main result using the approach first proposed in " and cite [1]. So the paper that everyone knows, in English (and with Bengio), explictly say that the original idea is in a paper in German, and still the scientific community chose not to cite the German original.

[1] https://people.idsia.ch/~juergen/SeppHochreiter1991ThesisAdv...

[2] https://sferics.idsia.ch/pub/juergen/gradientflow.pdf

MurizS•48m ago
I think what you're referring to is also known as Stigler's law of eponymy [1], which is interestingly self-referential and ironic in its own naming. There's also the related "Matthew effect" [2] in the sciences.

[1] https://en.wikipedia.org/wiki/Stigler's_law_of_eponymy

[2] https://en.wikipedia.org/wiki/Matthew_effect

dchftcs•4m ago
Einstein published his relativity papers originally in German.
alyxya•1h ago
The notion of inventing or creating something in ML doesn't seem very important as many people can independently come up with the same idea. Conversely, you can create novel results just by reviewing old literature and demonstrating it in a project.
ekjhgkejhgk•1h ago
That's how all/most science normally works.

Conversely, a huge amount of science is just scientists going "here's something I found interesting" but no one can figure out what to do with it. Then 30 or 100 years go by and it's a useful in a field that didn't even exist at the time.

alyxya•22m ago
It doesn’t apply to empirical science because there’s a lot more variation in observations. The variation of ideas in ML model architecture is limited by being theoretical.
ekjhgkejhgk•1h ago
To comment on the substance.

It seems that these two people Schimidhuber and Hochreiter were perhaps solving the right problem for the wrong reasons. They thought this was important because they expected that RNNs could hold memory indefinitely. Because of BPTT, you can think of that as a NN with infinitely many layers. At the time I believe nobody worries about vanishing gradient for deep NNs, because the compute power for networks that deep just didn't exist. But nowadays that's exactly how their solution is applied.

That's science for you.

gwern•1h ago
> Note again that a residual connection is not just an arbitrary shortcut connection or skip connection (e.g., 1988)[LA88][SEG1-3] from one layer to another! No, its weight must be 1.0, like in the 1997 LSTM, or in the 1999 initialized LSTM, or the initialized Highway Net, or the ResNet. If the weight had some other arbitrary real value far from 1.0, then the vanishing/exploding gradient problem[VAN1] would raise its ugly head, unless it was under control by an initially open gate that learns when to keep or temporarily remove the connection's residual property, like in the 1999 initialized LSTM, or the initialized Highway Net.

After reading Lang & Witbrock 1988 https://gwern.net/doc/ai/nn/fully-connected/1988-lang.pdf I'm not sure how convincing I find this explanation.

HarHarVeryFunny•38m ago
How about Schmidhuber actually invents the next big thing rather than waiting for it to come along then claim credit for it?