frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

AIVO Standard Operational AI Reliance Observation Protocol

https://zenodo.org/records/18286718
1•businessmate•11m ago•1 comments

Is the World Random?

https://mantrna.com/astrobench
2•prabhatkr•21m ago•0 comments

Show HN: 30min video analysis for $0.003 via frame-tiling and Vision API

https://github.com/unhaya/vam-seek-ai
3•haasiy•24m ago•0 comments

300X fast clustering with rust-louvain for nodes

https://github.com/FastBuilderAI/rust-louvain
2•prabhatkr•24m ago•0 comments

Quantum Name Service (QNS)- Path to Web5

https://github.com/aevov/qns
2•cr8oscloud•26m ago•1 comments

Show HN: vr.dev – simple 3D/VR/XR portfolio and links (Meta hit hard this week)

https://www.vr.dev/
2•vrdev•27m ago•0 comments

Shackleton and the Endurance Expedition: Photos from the 1915 Disastrous Journey

https://www.utterlyinteresting.com/post/the-amazing-survival-story-of-ernest-shackleton-and-his-e...
4•nomagicbullet•28m ago•1 comments

Show HN: Task Orchestrator – Production Safety for Claude Code Agents

https://github.com/TC407-api/task-orchestrator
2•Travis_Cole•28m ago•1 comments

Model is intended for use particularly for language learning

https://huggingface.co/EnversonAI/DeepSeek-R1-FineTuned-AdaptiveQGen-COT
2•AslanMammadli•43m ago•1 comments

Ask HN: Is repalcing an enterprise product with LLMs a realistic strategy?

2•chandmk•43m ago•0 comments

Pushing the smallest possible change to production

https://ankursethi.com/blog/smallest-possible-change/
2•GeneralMaximus•44m ago•0 comments

Why Xcode's AI Writes Better SwiftUI Than Claude Code, Codex

https://www.ameyalambat.com/blog/swiftui-skills
3•ameyalambat128•46m ago•0 comments

Show HN: Open-Source DLP for LLMs

https://github.com/dorcha-inc/ceil-dlp
2•unclecolm•49m ago•0 comments

Cursor AI refusing $20 refund after 3 days of broken service

2•Waldopro•49m ago•1 comments

Show HN: Monitor Claude/Codex usage on Linux via browser cookies (no API keys)

https://github.com/NihilDigit/waybar-ai-usage
4•NihilDigit•52m ago•1 comments

Spectrum Brings NBA Games in Apple Immersive to Apple Vision Pro

https://www.apple.com/newsroom/2025/10/spectrum-brings-nba-games-in-apple-immersive-to-apple-visi...
1•Austin_Conlon•55m ago•0 comments

Crypto holder loses $283M to scammer impersonating wallet support

https://bsky.app/profile/web3isgoinggreat.com/post/3mcn26h32wp2q
5•unforgivenpasta•59m ago•1 comments

AI-Powered Diabetes Analysis with GitHub Copilot and Claude Skills [video]

https://www.youtube.com/watch?v=on5R6PWj8Wg
4•shanselman•1h ago•0 comments

No Chess on a Dead Planet

https://indianexpress.com/article/sports/chess/climate-activists-protests-hold-up-tata-steel-ches...
1•akbarnama•1h ago•0 comments

Show HN: Vanslist – Craigslist for tech freelancers, no fees

https://vanslist.com
1•netgeniuskid•1h ago•0 comments

Show HN: Turkish Sieve Engine – GPU-Accelerated Prime Number Generator

https://github.com/bilgisofttr/turkishsieve
1•bilgisoft•1h ago•0 comments

Tell HN: Google Trust and Safety is a joke

2•tokyobreakfast•1h ago•1 comments

The relentless rule of my fitness tracker

https://timharford.com/2025/10/the-relentless-rule-of-my-fitness-tracker/
9•Arnt•1h ago•2 comments

Aldrich Ames built a career on betraying trust

https://www.economist.com/obituary/2026/01/15/aldrich-ames-built-a-career-on-betraying-trust
1•petethomas•1h ago•0 comments

Show HN: macOS Screenshot Organizer

https://www.shotsnap.ai/
2•libiny•1h ago•0 comments

'We'll Sue': White House's Warning to CBS Is Sign of a New Media Status Quo

https://www.nytimes.com/2026/01/17/business/media/cbs-news-trump-interview.html
3•stopbulying•1h ago•2 comments

jQuery 4.0.0 Released

https://blog.jquery.com/2026/01/17/jquery-4-0-0/
57•OuterVale•1h ago•11 comments

SkillHub – NPM for AI agent rules, share team standards across 13 AI tools

https://github.com/cloudvalley-tech/skillhub
1•zxh•1h ago•1 comments

Show HN: StarFetch – A lightweight, modern system fetch tool in Rust

https://github.com/Linus-Shyu/StarFetch_Core
1•LinusShyu•1h ago•1 comments

Show HN: Intuitive TUI for Ghostty Terminal Configuration

https://github.com/intaek-h/ghofig
1•intaek•1h ago•0 comments
Open in hackernews

Erdos 281 solved with ChatGPT 5.2 Pro

https://twitter.com/neelsomani/status/2012695714187325745
95•nl•2h ago

Comments

carbocation•1h ago
The erdosproblems thread itself contains comments from Terence Tao: https://www.erdosproblems.com/forum/thread/281
redbluered•1h ago
Has anyone verified this?

I've "solved" many math problems with LLMs, with LLMs giving full confidence in subtly or significantly incorrect solutions.

I'm very curious here. The Open AI memory orders and claims about capacity limits restricting access to better models are interesting too.

bpodgursky•56m ago
Terence Tao gave it the thumbs up. I don't think you're going to do better than that.
bparsons•13m ago
It's already been walked back.
pessimist•57m ago
From Terry Tao's comments in the thread:

"Very nice! ... actually the thing that impresses me more than the proof method is the avoidance of errors, such as making mistakes with interchanges of limits or quantifiers (which is the main pitfall to avoid here). Previous generations of LLMs would almost certainly have fumbled these delicate issues.

...

I am going ahead and placing this result on the wiki as a Section 1 result (perhaps the most unambiguous instance of such, to date)"

The pace of change in math is going to be something to watch closely. Many minor theorems will fall. Next major milestone: Can LLMs generate useful abstractions?

radioactivist•40m ago
Seems like the someone dug something up from the literature on this problem (see top comment on the erdosproblems.com thread)

"On following the references, it seems that the result in fact follows (after applying Rogers' theorem) from a 1936 paper of Davenport and Erdos (!), which proves the second result you mention. ... In the meantime, I am moving this problem to Section 2 on the wiki (though the new proof is still rather different from the literature proof)."

ares623•56m ago
This must be what it feels like to be a CEO and someone tells me they solved coding.
mikert89•48m ago
I have 15 years of software engineering experience across some top companies. I truly believe that ai will far surpass human beings at coding, and more broadly logic work. We are very close
daxfohl•36m ago
They already do. What they suck at is common sense. Unfortunately good software requires both.
marktl•27m ago
Or is it fortunate (for a short period at least).
anonzzzies•25m ago
Most people also suck at common sense, including most programmers, hence most programmers do not write good software to begin with.
anonzzzies•26m ago
HN will be the last place to admit it; people here seem to be holding out with the vague 'I tried it and it came up with crap'. While many of us are shipping software without touching (much) code anymore. I have written code for over 40 years and this is nothing like no-code or whatever 'replacing programmers' before, this is clearly different judging from the people who cannot code with a gun to their heads but still are shipping apps: it does not really matter if anyone believes me or not. I am making more money than ever with fewer people than ever delivering more than ever.

We are very close.

(by the way; I like writing code and I still do for fun)

user3939382•11m ago
They can only code to specification which is where even teams of humans get lost. Without much smarter architecture for AI (LLMs as is are a joke) that needle isn’t going to move.
a_tartaruga•47m ago
Out of curiosity why has the LLM math solving community been focused on the Erdos problems over other open problems? Are they of a certain nature where we would expect LLMs to be especially good at solving them?
krackers•35m ago
I guess they are at a difficulty where it's not too hard (unlike millennium prize problems), is fairly tightly scoped (unlike open ended research), and has some gravitas (so it's not some obscure theorem that's only unproven because of it's lack of noteworthiness).
dernett•45m ago
This is crazy. It's clear that these models don't have human intelligence, but it's undeniable at this point that they have _some_ form of intelligence.
qudat•43m ago
My take is that a huge part of human intelligence is pattern matching. We just didn’t understand how much multidimensional geometry influenced our matches
sdwr•37m ago
I don't think it's accurate to describe LLMs as pattern matching. Prediction is the mechanism they use to ingest and output information, and they end up with a (relatively) deep model of the world under the hood.
DrewADesign•22m ago
Which is even more true for humans.
keeda•16m ago
Yes, the world model building is achieved via pattern matching and happens during ingestion and training, but that is also part of the intelligence.
keeda•23m ago
Yes, it could be that intelligence is essentially a sophisticated form of recursive, brute force pattern matching.

I'm beginning to think the Bitter Lesson applies to organic intelligence as well, because basic pattern matching can be implemented relatively simply using very basic mathematical operations like multiply and accumulate, and so it can scale with massive parallelization of relatively simple building blocks.

ekianjo•34m ago
It's pattern matching. Which is actually what we measure in IQ tests, just saying.
jadenpeterson•30m ago
There's some nuance. IQ tests measure pattern matching and, in an underlying way, other facets of intelligence - memory, for example. How well can an LLM 'remember' a thing? Sometimes Claude will perform compaction when its context window reaches 200k "tokens" then it seems a little colder to me, but maybe that's just my imagination. I'm kind of a "power user".
rurban•27m ago
I call it matching. Pattern matching had a different meaning.
altmanaltman•30m ago
Depends on what you mean by intelligence, human intelligence and human
brendyn•25m ago
If LLMs weren't created by us but where something discovered in another species' behaviour it would be 100% labelled intelligence
sequin•43m ago
FWIW, I just gave Deepseek the same prompt and it solved it too (much faster than the 41m of ChatGPT). I then gave both proofs to Opus and it confirmed their equivalence.

The answer is yes. Assume, for the sake of contradiction, that there exists an \(\epsilon > 0\) such that for every \(k\), there exists a choice of congruence classes \(a_1^{(k)}, \dots, a_k^{(k)}\) for which the set of integers not covered by the first \(k\) congruences has density at least \(\epsilon\).

For each \(k\), let \(F_k\) be the set of all infinite sequences of residues \((a_i)_{i=1}^\infty\) such that the uncovered set from the first \(k\) congruences has density at least \(\epsilon\). Each \(F_k\) is nonempty (by assumption) and closed in the product topology (since it depends only on the first \(k\) coordinates). Moreover, \(F_{k+1} \subseteq F_k\) because adding a congruence can only reduce the uncovered set. By the compactness of the product of finite sets, \(\bigcap_{k \ge 1} F_k\) is nonempty.

Choose an infinite sequence \((a_i) \in \bigcap_{k \ge 1} F_k\). For this sequence, let \(U_k\) be the set of integers not covered by the first \(k\) congruences, and let \(d_k\) be the density of \(U_k\). Then \(d_k \ge \epsilon\) for all \(k\). Since \(U_{k+1} \subseteq U_k\), the sets \(U_k\) are decreasing and periodic, and their intersection \(U = \bigcap_{k \ge 1} U_k\) has density \(d = \lim_{k \to \infty} d_k \ge \epsilon\). However, by hypothesis, for any choice of residues, the uncovered set has density \(0\), a contradiction.

Therefore, for every \(\epsilon > 0\), there exists a \(k\) such that for every choice of congruence classes \(a_i\), the density of integers not covered by the first \(k\) congruences is less than \(\epsilon\).

\boxed{\text{Yes}}

amluto•28m ago
I find it interesting that, as someone utterly unfamiliar with ergodic theory, Dini’s theorem, etc, I find Deepseek’s proof somewhat comprehensible, whereas I do not find GPT-5.2’s proof comprehensible at all. I suspect that I’d need to delve into the terminology in the GPT proof if I tried to verify Deepseek’s, so maybe GPT’s is being more straightforward about the underlying theory it relies on?
ashleyn•35m ago
I guess the first question I have is if these problems solved by LLMs are just low-hanging fruit that human researchers either didn't get around to or show much interest in - or if there's some actual beef here to the idea that LLMs can independently conduct original research and solve hard problems.
dyauspitr•33m ago
There is still value on letting these LLMs loose on the periphery and knocking out all the low hanging fruit humanity hasn’t had the time to get around to. Also, I don’t know this, but if it is a problem on Erdos I presume people have tried to solve it atleast a little bit before it makes it to the list.
doctoboggan•34m ago
Can anyone give a little more color on the nature of Erdos problems? Are these problems that many mathematicians have spend years tackling with no result? Or do some of the problems evade scrutiny and go un-attempted for most of the time?

EDIT: After reading a link someone else posted to Terrance Tao's wiki page, he has a paragraph that somewhat answers this question:

> Erdős problems vary widely in difficulty (by several orders of magnitude), with a core of very interesting, but extremely difficult problems at one end of the spectrum, and a "long tail" of under-explored problems at the other, many of which are "low hanging fruit" that are very suitable for being attacked by current AI tools. Unfortunately, it is hard to tell in advance which category a given problem falls into, short of an expert literature review. (However, if an Erdős problem is only stated once in the literature, and there is scant record of any followup work on the problem, this suggests that the problem may be of the second category.)

from here: https://github.com/teorth/erdosproblems/wiki/AI-contribution...

xeeeeeeeeeeenu•34m ago
> no prior solutions found.

This is no longer true, a prior solution has just been found[1], so the LLM proof has been moved to the Section 2 of Terence Tao's wiki[2].

[1] - https://www.erdosproblems.com/forum/thread/281#post-3325

[2] - https://github.com/teorth/erdosproblems/wiki/AI-contribution...