frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Al Lowe on model trains, funny deaths and working with Disney

https://spillhistorie.no/2026/02/06/interview-with-sierra-veteran-al-lowe/
38•thelok•2h ago•3 comments

Hoot: Scheme on WebAssembly

https://www.spritely.institute/hoot/
101•AlexeyBrin•6h ago•18 comments

First Proof

https://arxiv.org/abs/2602.05192
51•samasblack•3h ago•37 comments

OpenCiv3: Open-source, cross-platform reimagining of Civilization III

https://openciv3.org/
789•klaussilveira•20h ago•242 comments

Stories from 25 Years of Software Development

https://susam.net/twenty-five-years-of-computing.html
38•vinhnx•3h ago•5 comments

Reinforcement Learning from Human Feedback

https://rlhfbook.com/
62•onurkanbkrc•5h ago•5 comments

Start all of your commands with a comma (2009)

https://rhodesmill.org/brandon/2009/commands-with-comma/
462•theblazehen•2d ago•165 comments

The Waymo World Model

https://waymo.com/blog/2026/02/the-waymo-world-model-a-new-frontier-for-autonomous-driving-simula...
1040•xnx•1d ago•587 comments

France's homegrown open source online office suite

https://github.com/suitenumerique
506•nar001•4h ago•234 comments

Software factories and the agentic moment

https://factory.strongdm.ai/
48•mellosouls•3h ago•49 comments

Vocal Guide – belt sing without killing yourself

https://jesperordrup.github.io/vocal-guide/
183•jesperordrup•10h ago•65 comments

The AI boom is causing shortages everywhere else

https://www.washingtonpost.com/technology/2026/02/07/ai-spending-economy-shortages/
63•1vuio0pswjnm7•7h ago•59 comments

Coding agents have replaced every framework I used

https://blog.alaindichiappari.dev/p/software-engineering-is-back
186•alainrk•5h ago•280 comments

A Fresh Look at IBM 3270 Information Display System

https://www.rs-online.com/designspark/a-fresh-look-at-ibm-3270-information-display-system
27•rbanffy•4d ago•5 comments

What Is Stoicism?

https://stoacentral.com/guides/what-is-stoicism
15•0xmattf•2h ago•7 comments

72M Points of Interest

https://tech.marksblogg.com/overture-places-pois.html
19•marklit•5d ago•0 comments

Unseen Footage of Atari Battlezone Arcade Cabinet Production

https://arcadeblogger.com/2026/02/02/unseen-footage-of-atari-battlezone-cabinet-production/
108•videotopia•4d ago•27 comments

Where did all the starships go?

https://www.datawrapper.de/blog/science-fiction-decline
58•speckx•4d ago•62 comments

Show HN: Look Ma, No Linux: Shell, App Installer, Vi, Cc on ESP32-S3 / BreezyBox

https://github.com/valdanylchuk/breezydemo
268•isitcontent•20h ago•34 comments

British drivers over 70 to face eye tests every three years

https://www.bbc.com/news/articles/c205nxy0p31o
169•bookofjoe•2h ago•152 comments

Learning from context is harder than we thought

https://hy.tencent.com/research/100025?langVersion=en
197•limoce•4d ago•107 comments

Monty: A minimal, secure Python interpreter written in Rust for use by AI

https://github.com/pydantic/monty
281•dmpetrov•21h ago•150 comments

Making geo joins faster with H3 indexes

https://floedb.ai/blog/how-we-made-geo-joins-400-faster-with-h3-indexes
152•matheusalmeida•2d ago•47 comments

Hackers (1995) Animated Experience

https://hackers-1995.vercel.app/
548•todsacerdoti•1d ago•266 comments

Sheldon Brown's Bicycle Technical Info

https://www.sheldonbrown.com/
422•ostacke•1d ago•110 comments

Ga68, a GNU Algol 68 Compiler

https://fosdem.org/2026/schedule/event/PEXRTN-ga68-intro/
37•matt_d•4d ago•13 comments

Show HN: I spent 4 years building a UI design tool with only the features I use

https://vecti.com
365•vecti•23h ago•167 comments

An Update on Heroku

https://www.heroku.com/blog/an-update-on-heroku/
465•lstoll•1d ago•305 comments

Show HN: If you lose your memory, how to regain access to your computer?

https://eljojo.github.io/rememory/
341•eljojo•23h ago•209 comments

What Is Ruliology?

https://writings.stephenwolfram.com/2026/01/what-is-ruliology/
66•helloplanets•4d ago•70 comments
Open in hackernews

A short statistical reasoning test

https://emiruz.com/post/2025-08-17-statistical-reasoning/
56•usgroup•5mo ago

Comments

jldugger•5mo ago
( replying to a now deleted post)

>> the uncertainty in the number of trials > Has no meaning to me.

What the author is trying to get at in the admittedly poorly worded question is that the trials are noisy measures of an underlying effect. Your job is to sort by effect size, while accounting for the random chance that a low sample size trial just got unlucky.

You might argue that the question is much harder than the author assumes, since your best guess at the actual effect size seems like it should still just be the success rate, even if the low sample size trials have wider error bars. You'd need to come up with some sort of heuristic that says why 7/9 deserves a lower rank than 50/70 using binomial confidence intervals.

Probably that heuristic is intended to be a bayesian approach? Like, if you add just two successes and two failures to each scenario as a prior, thats enough to put the 50/70 option ahead.

jldugger•5mo ago
And I guess since they answer the questions at the bottom, it seems their intent is indeed the simplistic approach

> The lower bound of which can be used to order the fractions, and so control the risk of over-estimation.

It not clear to me from the question whether the cost of a mistake is in the over-estimating the underlying effect or in misranking the effects, and that seems like it would drive your heuristic selection.

usgroup•5mo ago
From the question:

“However, it is very important that the uncertainty in the number of trials is taken into account because over-estimating a fraction is a costly mistake.“

Seems fairly clear to me that you’re supposed to use a lower bound estimate to take into account variance on the fraction due to the number of trials in a way to bounds the chance of over estimation.

Further, there is no need for a heuristic when there a several statistical models for this exact problem with clear properties. Some are given in the answer.

thekoma•5mo ago
Out of context, the expression "the uncertainty in the number of trials" would refer to missing knowledge in terms of how many trials actually ran.

In the context of the post this doesn't make sense, so the reader is left to hypothesize what the writer actually meant.

usgroup•5mo ago
I agree it could be clearer but as a general rule, if you find an interpretation under which the question doesn’t make sense, try considering another interpretation.
taylorius•5mo ago
I think "uncertainty due to the number of trials" would be clearer.
tomsmeding•5mo ago
Uncertainty in the number of successes due to the number of trials, would be even better.
jdhwosnhw•5mo ago
That would also be an incorrect phrasing. This entire thread is a good illustration of the difficulty of speaking precisely about probabilistic concepts.

(The number of successes has zero uncertainty. If you flip a coin 10 times and get 5 heads, there is no uncertainty on the number of heads. In general, for any statistical model the uncertainty is only with respect to an underlying model parameter - in this example, while your number of successes is perfectly known, it can be used to infer a probability of success, p, of 0.5, and there is uncertainty associated with that inferred probability.)

jldugger•5mo ago
I guess I struggle to understand why under estimation is worse than over estimation, when the final result is a ranking. It seems like they're equally likely to produce an incorrect ranking!
kruffalon•5mo ago
I wrote the deleted comment you are replying to.

The essence of my comment was that this text/test is not for me (one person of the general public) but more like a few leetcode-style questions for statisticians.

Your attempt to explain what I didn't understand just proves my point as I don't really understand what you are saying either.

And that's ok: this is just not for me! (And that's why I deleted my original comment)

dmurray•5mo ago
The problems are really underspecified for statisticians, too. Leetcode is normally very clear on the requirements.

> it is very important that the uncertainty in the number of trials is taken into account because over-estimating a fraction is a costly mistake.

This is not some precise jargon that is meaningless to the layman but completely clearly specified to a professional statistician. It's more like the specification written by your non-technical product manager for how some technical feature should work. A skilled data scientist will have the experience and the context to figure out what it's probably asking for, but he might write down a few more clarifying details before giving it to a junior on his team to implement.

If testing these kind of guess-what-the-stakeholder probably-means skills is the point of this test, it's quite good at it. But that's not what leetcode is for.

MontyCarloHall•5mo ago
Put more simply: suppose I have a coin that might be biased. I decide to asses this via repeatedly flipping it—if it’s biased, it will disproportionately land on heads or tails. If I flip 10 times and get 4 heads/6 tails, I don’t have the power to make a confident assessment of any bias. On the other hand, if I flip it 100 times and get 40 heads/60 tails, I am a bit more certain. At 1000 flips, with 400 heads/600 tails, I am extremely confident the coin is biased. Even though the fraction of heads is identically 40% across all three sets of flips, the underlying counts yield very different amounts of confidence on how close to 40% the coin’s bias is. The first question is a way of rigorously quantifying this confidence.

I don’t think this is “leetcode for statisticians.” This question (and the other two) are all examples of concrete, real-world problems that people across a variety of quantitative disciplines frequently encounter.

In fact, the first question is directly relevant to voting on this site. When sorting replies by fraction of upvotes, how should the forum software rank a new reply with 1 upvote/0 downvotes, versus an older reply with 4 upvotes/1 downvote? What about an older, more controversial reply with 20 upvotes/7 downvotes? 15 upvotes/2 downvotes?

jldugger•5mo ago
> This question (and the other two) are all examples of concrete, real-world problems that people across a variety of quantitative disciplines frequently encounter.

Indeed, I use this technique to sort search results in Splunk, as an extension of TF-IDF. Consider a scenario where us-east-2 is broken but us-east-1 is fine (clearly just a hypothetical!). Split the logs along that good/bad dimension, and then break down by some other pattern; log class, punct, etc. Usually I use a prior of 50:50 to help sort out the "happened once in bad cluster" events.

jldugger•5mo ago
No worries, I wasn't really trying to explain it anyways, as much as seeking confirmation from the rest of HN that this question is ill specified. Judging from responses, yes, it is.

If "binomial distribution" and "confidence interval" are unfamiliar terms then you probably are not prepared to pass OP's "statistical reasoning test" regardless. I think most engineers wouldn't, and I only understood the intent of question 1 because my pandemic lockdown project was reading a stats textbook cover to cover.

yobbo•5mo ago
I think the formula should be (n+1)/(n+m+1) which should correspond to the mean of a binomial distribution with a uniform prior. So it's adding 1 to each count of observations.

This is probably the formula to memorise and check against.

kqr•5mo ago
You mean (n+1)/(n+m+2) and yes, this is Laplace's rule of succession. It won't give you a confidence interval, but it gives you a posterior point estimate.

If you want a rough 95 % confidence interval without complicated maths, the Agresti–Coull interval is useful. It's computed as if the distribution was normal, but pretending there were two more successes and failures instead.

yobbo•5mo ago
Yep, you're correct. It should be (n+1)/(n+m+2).

If you have access to a machine or lookup tables, you might as well plug in the values for the distribution Beta(1+n,1+m) which should correspond to the joint density.

(The formula above corresponds to the mean of this distribution, so it's probably right but I haven't work it through myself now ...)

usgroup•5mo ago
See the Jeffrey posterior section here:

https://en.m.wikipedia.org/wiki/Binomial_proportion_confiden...

The blog post uses a non informative Jeffrey prior.

kqr•5mo ago
The first question seems a little unfair because it does not say how much more expensive overestimation is compared to underestimation. It implicitly assumes 19:1 given that it's ordering by the 0.05th quantile of the posterior distribution, but that's information not contained in the question.
energy123•5mo ago
The general framework would be to sort by `U(dist(ratio))`. The choice of `U` (utility function) being a separate question to the estimation of `dist(ratio)`.
usgroup•5mo ago
Bingo.
bjornsing•5mo ago
Is there a better principaled approach to #1 than Monte Carlo sampling from beta distributions?
usgroup•5mo ago
The Jeffrey posterior in the answer is closed form and Bayesian. The other answer is a profile likelihood.

Neither involve Monte Carlo sampling. Both are general and principled.

bjornsing•5mo ago
The answer looks a bit simplistic compared to the question (as I interpreted it at least). In order to estimate the risk of incorrect ordering you have to calculate P(p2 > p1) where p1 and p2 are drawn from different beta distributions. AFAIK there’s no closed form expression for that probability (so Monte Carlo is one possible approach).
usgroup•5mo ago
The question doesn’t ask for that —- it explicitly asks us to control for over estimation of the fraction — although I rather like your interpretation as an extension.
bjornsing•5mo ago
Ok. I may have read a bit too much into this paragraph:

> Order the items above, smallest fraction first, whilst taking into account the uncertainty in the number of trials to bound the probability of over-estimating each fraction.

But why mention ordering if you’re not looking for statistical reasoning around the ordering in particular?

thorum•5mo ago
I don’t know enough about statistics to answer these with math, but I’ve been on quite a few buses and it’s common at some stops for bus arrivals to cluster around specific times. If you always leave after first you see, and most of your random observations are before the first bus, won’t you (almost) always miss the others?