frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

NORAD's Cheyenne Mountain Combat Center, C.1966

https://flashbak.com/norad-cheyenne-mountain-combat-center-478804/
39•zdw•5d ago•10 comments

Show HN: MyraOS – My 32-bit operating system in C and ASM (Hack Club project)

https://github.com/dvir-biton/MyraOS
10•dvirbt•1h ago•0 comments

Wren: A classy little scripting language

https://wren.io/
72•Lyngbakr•4d ago•19 comments

The bug that taught me more about PyTorch than years of using it

https://elanapearl.github.io/blog/2025/the-bug-that-taught-me-pytorch/
281•bblcla•3d ago•58 comments

Advent of Code 2025: Number of puzzles reduce from 25 to 12 for the first time

https://adventofcode.com/2025/about#faq_num_days
356•vismit2000•13h ago•182 comments

System.LongBool

https://docwiki.embarcadero.com/Libraries/Sydney/en/System.LongBool
9•surprisetalk•4d ago•4 comments

Alzheimer's disrupts circadian rhythms of plaque-clearing brain cells

https://medicine.washu.edu/news/alzheimers-disrupts-circadian-rhythms-of-plaque-clearing-brain-ce...
106•gmays•4h ago•12 comments

Nvidia DGX Spark: When benchmark numbers meet production reality

https://publish.obsidian.md/aixplore/Practical+Applications/dgx-lab-benchmarks-vs-reality-day-4
102•RyeCatcher•4h ago•53 comments

Making the Electron Microscope

https://www.asimov.press/p/electron-microscope
37•mailyk•5h ago•3 comments

Eavesdropping on Internal Networks via Unencrypted Satellites

https://satcom.sysnet.ucsd.edu/
156•Bogdanp•5d ago•23 comments

Validating Your Ideas on Strangers

https://jeremyaboyd.com/post/validating-your-ideas-on-strangers
51•tacon•2d ago•30 comments

A worker fell into a nuclear reactor pool

https://www.nrc.gov/reading-rm/doc-collections/event-status/event/2025/20251022en?brid=vscAjql9kZ...
612•nvahalik•20h ago•436 comments

Formal Reasoning [pdf]

https://cs.ru.nl/~freek/courses/fr-2025/public/fr.pdf
107•Thom2503•9h ago•23 comments

Downloadable movie posters from the 40s, 50s, 60s, and 70s

https://hrc.contentdm.oclc.org/digital/collection/p15878coll84/search
369•bookofjoe•1w ago•70 comments

Pico-Banana-400k

https://github.com/apple/pico-banana-400k
342•dvrp•19h ago•60 comments

The Linux Boot Process: From Power Button to Kernel

https://www.0xkato.xyz/linux-boot/
408•0xkato•22h ago•82 comments

Writing a RISC-V Emulator in Rust

https://book.rvemu.app/
91•signa11•14h ago•39 comments

You Already Have a Git Server

https://maurycyz.com/misc/easy_git/
336•chmaynard•11h ago•274 comments

Ask HN: How to boost Gemini transcription accuracy for company names?

24•bingwu1995•6d ago•16 comments

Why your social.org files can have millions of lines without performance issues

https://en.andros.dev/blog/4e12225f/why-your-socialorg-files-can-have-millions-of-lines-without-a...
63•andros•1d ago•6 comments

Clojure Land – Discover open-source Clojure libraries and frameworks

https://clojure.land/
146•TheWiggles•13h ago•36 comments

Connect to a 1980s Atari BBS through the web

https://www.southernamis.com/ataribbsconnect
60•JPolka•12h ago•4 comments

Myanmar military shuts down a major cybercrime center, detains over 2k people

https://apnews.com/article/scam-centers-cybercrime-myanmar-a2c9fda85187121e51bd0efdf29c81da
124•bikenaga•6h ago•41 comments

Smartphones manipulate our emotions and trigger our reflexes

https://theconversation.com/smartphones-manipulate-our-emotions-and-trigger-our-reflexes-no-wonde...
43•PaulHoule•3h ago•19 comments

Asbestosis

https://diamondgeezer.blogspot.com/2025/10/asbestosis.html
193•zeristor•13h ago•137 comments

Ask HN: Second generation of intro to software dev for 3rd graders

20•xrd•6d ago•22 comments

D2: Diagram Scripting Language

https://d2lang.com/tour/intro/
249•benzguo•23h ago•62 comments

The Journey Before main()

https://amit.prasad.me/blog/before-main
295•amitprasad•1d ago•110 comments

Why I code as a CTO

https://www.assembled.com/blog/why-i-code-as-a-cto
283•johnjwang•2d ago•248 comments

PCB Edge USB C Connector Library

https://github.com/AnasMalas/pcb-edge-usb-c
149•walterbell•19h ago•66 comments
Open in hackernews

A Definition of AGI

https://arxiv.org/abs/2510.18212
92•pegasus•3h ago

Comments

vardump•3h ago
Whatever the definition may be, the goalposts are usually moved once AI reaches that point.
MattRix•3h ago
Isn’t that the point of trying to define it in a more rigorous way, like this paper is doing?
krige•3h ago
Are you saying that we already have AGI, except those pesky goalpost movers keep denying the truth? Hm.
A4ET8a8uTh0_v2•3h ago
I think, given some of the signs of the horizon, there is a level of MAD type bluffing going around, but some of the actions by various power centers suggest it is either close, people think its close or it is there.
vardump•3h ago
No, just what has usually happened in the past with AI goalposts.

At first, just playing chess was considered to be a sign of intelligence. Of course, that was wrong, but not obvious at all in 1950.

NitpickLawyer•2h ago
I'd say yes, by at least one old definition made by someone who was at the time in a position to have a definition.

When deepmind was founded (2010) their definition was the following: AI is a system that learns to perform one thing; AGI is a system that learns to perform many things at the same time.

I would say that whatever we have today, "as a system" matches that definition. In other words, the "system" that is say gpt5/gemini3/etc has learned to "do" (while do is debateable) a lot of tasks (read/write/play chess/code/etc) "at the same time". And from a "pure" ML point, it learned those things from the "simple" core objective of next token prediction (+ enhancements later, RL, etc). That is pretty cool.

So I can see that as an argument for "yes".

But, even the person who had that definition has "moved the goalposts" of his own definition. From recent interviews, Hassabis has moved towards a definition that resembles the one from this paper linked here. So there's that. We are all moving the goalposts.

And it's not a recent thing. People did this back in the 80s. There's the famous "As soon as AI does something, it ceases to be AI" or paraphrased "AI is everything that hasn't been done yet".

wahnfrieden•2h ago
Can you cite the Deepmind definition? No Google results for that.
NitpickLawyer•2h ago
It's from a documentary that tracks Hassabis' life. I c/p from an old comment of mine (the quotes are from the documentary, can probably look up timestamps if you need, but it's in the first ~15 minutes I'd say, when they cover the first days of Deepmind):

----

In 2010, one of the first "presentations" given at Deepmind by Hassabis, had a few slides on AGI (from the movie/documentary "The Thinking Game"):

Quote from Shane Legg: "Our mission was to build an AGI - an artificial general intelligence, and so that means that we need a system which is general - it doesn't learn to do one specific thing. That's really key part of human intelligence, learn to do many many things".

Quote from Hassabis: "So, what is our mission? We summarise it as <Build the world's first general learning machine>. So we always stress the word general and learning here the key things."

And the key slide (that I think cements the difference between what AGI stood for then, vs. now):

AI - one task vs. AGI - many tasks

at human level intelligence.

darepublic•2h ago
It doesn't play chess? Just can parrot it very well
NitpickLawyer•2h ago
Yeah, maybe. But what matters is the end result. In the kaggle match, one of the games from the finals (grok vs o3) is rated by chesscom's stockfish as 1900vs2500. That is, they played a game at around those ratings.

For reference, the average chesscom player is ~900 elo, while the average FIDE rated player is ~1600. So, yeah. Parrot or not, the LLMs can make moves above the average player. Whatever that means.

darepublic•2h ago
I believe it will make illegal moves (unaided by any tools ofc). It will also make mistakes doing things like not being able to construct the board correctly given a fen string. For these reasons I consider long strings of correct moves insufficient to say it can play the game. If my first two statements, about a propensity for illegal moves and other fails on "easy for humans" tasks were untrue then I would reconsider.
NitpickLawyer•2h ago
In the kaggle test they considered the match forfeit if the model could not produce a legal move after 3 tries (none of the matches in the finals were forfeited, they all ended with checkmate on the board). Again, chesscom's interface won't let you make illegal moves, and the average there is 900. Take that as you will.
bossyTeacher•2h ago
> AGI is a system that learns to perform many things at the same time.

What counts as a "thing"? Because arguably some of the deep ANNs pre-transfomers would also qualify as AGI but no one would consider them intelligent (not in the human or animal sense of intelligence).

And you probably don't even need fancy neural networks. Get a RL algorithm and a properly mapped solution space and it will learn to do whatever you want as long as the problem can be mapped.

derektank•2h ago
It wasn't the best definition of AGI but I think if you asked an interested layman whether or not a system that can pass the Turing test was AGI 5 years ago, they would have said yes
jltsiren•1h ago
An interested but uninformed layman.

When I was in college ~25 years ago, I took a class on the philosophy of AI. People had come up with a lot of weird ideas about AI, but there was one almost universal conclusion: that the Turing test is not a good test for intelligence.

The least weird objection was that the premise of the Turing test is unscientific. It sees "this system is intelligent" as a logical statement and seeks to prove or disprove it in an abstract model. But if you perform an experiment to determine if a real-world system is intelligent, the right conclusion for the system passing the test is that the system may be intelligent, but a different experiment might show that it's not.

nativeit•48m ago
Douglas Hofstadter wrote Gödel, Escher, Bach nearly 50-years ago, and it won a Pulitzer Prize, the National Book Award, and got featured in the popular press. It’s been on lots of college reading lists, from 2007 online coursework for high school students was available from MIT. The FBI concluded that the 2001 anthrax scare was in-part inspired by elements of the book, which was found in the attacker’s trash. Anyone who’s wanted to engage with the theories and philosophy surrounding artificial intelligence has had plenty of materials that get fairly in-depth asking and exploring these same questions. It seems like a lot of people seem to think this is all bleeding edge novelty (at least, the underlying philosophical and academic ideas getting discussed in popular media), but rather all of the industry is predicated on ideas that are very old philosophy + decades-old established technology + relatively recent neuroscience + modern financial engineering. That said, I don’t want to suggest a layperson is likely to have engaged with any of it, so I understand why this will be the first time a lot of people will have ever considered some of these questions. I imagine what I’m feeling is fairly common to anyone who’s got a very niche interest that blows up and becomes the topic of interest for the entire world. I think there’s probably some very interesting, as-yet undocumented phenomena occurring that’s been the product of the unbelievably vast amount of resources sunk into what’s otherwise a fairly niche kind of utility (in LLMs specifically, and machine learning more broadly). I’m optimistic that there will be some very transformational technologies to come from it, although whether it will produce anything like “AGI”, or ever justify these levels of investment? Both seem rather unlikely.
rafram•3h ago
Are you claiming that LLMs have achieved AGI?
moffkalast•2h ago
Compared to everything that came before they are fairly general alright.
bigyabai•3h ago
The authors acknowledge that this is entirely possible. Their work is just grounded in theory, after all:

> we ground our methodology in Cattell-Horn-Carroll theory, the most empirically validated model of human cognition.

righthand•3h ago
I agree if our comprehension of intelligence and “life” is incomplete, so is our model for artificial intelligence.
kelseyfrog•2h ago
There's at least two distinct basis in AGI refutations : behaviorist and ontological. They often get muddled.

I can't begin to count the number of times I've encountered someone who holds an ontological belief for why AGI cannot exist and then for some reason formulates it as a behavioralist criteria. This muddying of argument results in what looks like a moving of the goalposts. I'd encourage folks to be more clear whether they believe AGI is ontologically possible or impossible in addition to any behavioralist claims.

zahlman•2h ago
My experience has been more that the pro-AI people misunderstand where the goalposts were, and then complain when they're correctly pointed at.

The "Turing test" I always saw described in literature, and the examples of what passing output from a machine was imagined to look like, are nothing like what's claimed to pass nowadays. Honestly, a lot of the people claiming that contemporary chatbots pass come across like they would have thought ELIZA passed.

bonoboTP•2h ago
Can you be more concrete? What kind of answer/conversation do you see as demonstrating passing the test, that you think is currently not possible.
tsimionescu•2h ago
Ones in which both the human test takers and the human counterparts are actively trying to prove to each other that they are actually human.

With today's chat bots, it's absolutely trivial to tell that you're not talking to a real human. They will never interrupt you, continue their train of thought even thought you're trying to change the conversation, go on a complete non-sequitur, swear at you, etc. These are all things that the human "controls" should be doing to prove to the judges that they are indeed human.

LLMs are nowhere near beating the Turing test. They may fool some humans in some limited interactions, especially if the output is curated by a human. But left alone to interact with the raw output for more than a few lines, and if actively seeking to tell if you're interacting with a human or an AI (instead of wanting to believe), there really is no chance you'd be tricked.

bonoboTP•1h ago
Okay but we are not really optimizing them to emulate humans right now. In fact, it's the opposite. The mainstream bots are explicitly trained to not identify as humans and to refuse to claim having thought or internal feelings or consciousness.

So in that sense it's a triviality. You can ask ChatGPT whether it's human and it will say no upfront. And it has various guardrails in place against too much "roleplay", so you can't just instruct it to act human. You'd need a different post-training setup.

I'm not aware whether anyone did that with open models already.

tsimionescu•1h ago
Sure, but there is a good reason for that. The way they are currently post-trained is the only way to make them actually useful. If you take the raw model, it will actually be much worse at the kinds of tasks you want it to perform. In contrast, a human can both be human, and be good at their job - this is the standard by which we should judge these machines. If their behavior needs to be restricted to actually become good at specific tasks, then they can't also be claimed to pass the Turing test if they can't within those same restrictions.
og_kalu•47m ago
>Sure, but there is a good reason for that. The way they are currently post-trained is the only way to make them actually useful.

Post training them to speak like a bot and deny being human has no effect on how useful they are. That's just an Open AI/Google/Anthropic preference.

>If you take the raw model, it will actually be much worse at the kinds of tasks you want it to perform

Raw models are not worse. Literally every model release paper that compares both show them as better at benchmarks, if anything. Post training degrading performance is a well known phenomena. What they are is more difficult to guide/control. Raw models are less useful because you have to present your input in certain ways, but they are not worse performers.

It's besides the point anyways because again, you don't have to post train them to act as anything other than a human.

>If their behavior needs to be restricted to actually become good at specific tasks, then they can't also be claimed to pass the Turing test if they can't within those same restrictions.

Okay, but that's not the case.

tsimionescu•37m ago
> Raw models are less useful because you have to present your input in certain ways, but they are not worse performers.

This is exactly what I was referring to.

og_kalu•4m ago
Again, it literally does not matter. You are talking about instruction tuning. You can perform instruction tuning without making your models go out of the way to tell you they are not human, and it changes literally nothing about their usefulness. Their behavior does not have to be restricted this way to get them useful/instruction tuned. So your premise is just wrong.
zahlman•1h ago
> Okay but we are not really optimizing them to emulate humans right now.

But that is exactly the point of the Turing test.

bonoboTP•33m ago
Ok, but then it doesn't make sense to dismiss AI based on that. It fails the Turing test, because it's creators intentionally don't even try to make something that is good at the (strictly defined) Turing test.

If someone really wants to see a Turing-passing bot, I guess someone could try making one but I'm doubtful it would be of much use.

Anyways,people forget that the thought experiment by Turing was a rhetorical device, not something he envisioned to build. The point was to say that semantic debates about "intelligence" are distractions.

Abecid•3h ago
Dan is very ambitious great marketer too
CaptainOfCoit•3h ago
> defining AGI as matching the cognitive versatility and proficiency of a well-educated adult

Seems most of the people one would encounter out in the world might not posses AGI, how are we supposed to be able to train our electrified rocks to have AGI if this is the case?

If no one has created a online quiz called "Are you smarter than AGI?" yet based on the proposed "ten core cognitive domains", I'd be disappointed.

A4ET8a8uTh0_v2•3h ago
I was going to make a mildly snide remark about how once it can consistently make better decision than average person, it is automatically qualifies, but the paper itself is surprisingly thoughtful in describing both: where we are and where it would need to be.
tcdent•3h ago
Don't get me wrong, I am super excited about what AI is doing for technology. But this endless conversation about "what is AGI" is so boring.

It makes me think of every single public discussion that's ever been had about quantum, where you can't start the conversation unless you go through a quick 101 on what a qubit is.

As with any technology, there's not really a destination. There is only the process of improvement. The only real definitive point is when a technology becomes obsolete, though it is still kept alive through a celebration of its nostalgia.

AI will continue to improve. More workflows will become automated. And from our perception, no matter what the rapidness of advancement is, we're still frogs in water.

bongodongobob•2h ago
I agree. It's an interesting discussion for those who have never taken college level philosophy classes I suppose. What consciousness/thought is is still a massively open question. Seeing people in the comments with what they think is their novel solution has already been posited like 400 years ago. Honestly it's kind of sad seeing this stuff on a forum like this. These posts are for sure the worst of Hackernews.
bonoboTP•2h ago
There are a bunch of these topics that everyone feels qualified to say something about. Consciousness, intelligence, education methods, nutrition, men vs women, economic systems etc.

It's a very emotional topic because people feel their self image threatened. It's a topic related to what is the meaning of being human. Yeah sure it should be a separate question, but emotionally it is connected to it in a deep level. The prospect of job replacement and social transformation is quite a threatening one.

So I'm somewhat understanding of this. It's not merely an academic topic, because these things will be adopted in the real world among real people. So you can't simply make everyone shut up who is an outsider or just heard about this stuff incidentally in the news and has superficial points to make.

bongodongobob•39m ago
I get it. It's just something we've thought about as long as we've been human, and still haven't figured out. It's frustrating when most of the people commenting don't know any of the source material. It's so arrogant.
edulix•3h ago
We have SAGI: Stupid Artificial General Intelligence. It's actually quite general, but works differently. In some areas it can be better or faster than a human, and in others it's more stupid.

Just like an airplane doesn't work exactly like a bird, but both can fly.

merksittich•2h ago
I find the concept of low floor/high ceiling quite helpful, as for instance recently discussed in "When Will AI Transform the Economy?" [1] - actually more helpful than "jagged" intelligence used in TFA.

[1] https://andreinfante.substack.com/p/when-will-ai-transform-t...

quantum_state•2h ago
Would propose to use the term Naive Artificial General Intelligence, in analogy to the widely used (by working mathematicians) and reasonably successful Naive Set Theory …
wizzwizz4•1h ago
I was doing some naïve set theory the other day, and I found a proof of the Riemann hypothesis, by contradiction.

Assume the Riemann hypothesis is false. Then, consider the proposition "{a|a∉a}∈{a|a∉a}". By the law of the excluded middle, it suffices to consider each case separately. Assuming {a|a∉a}∈{a|a∉a}, we find {a|a∉a}∉{a|a∉a}, for a contradiction. Instead, assuming {a|a∉a}∉{a|a∉a}, we find {a|a∉a}∈{a|a∉a}, for a contradiction. Therefore, "the Riemann hypothesis is false" is false. By the law of the excluded middle, we have shown the Riemann hypothesis is true.

Naïve AGI is an apt analogy, in this regard, but I feel these systems aren't simple nor elegant enough to deserve the name naïve.

the_arun•1h ago
It is a good analogy.
bmacho•2h ago
And this is it (from the abstract):

  > defining AGI as matching the cognitive versatility and proficiency of a well-educated adult.
sureglymop•2h ago
I think that's a good effort! I remember mentioning the need for this here a few months ago: https://news.ycombinator.com/item?id=44468198
cjbarber•2h ago
Some AGI definition variables I see:

Is it about jobs/tasks, or cognitive capabilities? The majority of the AI-valley seems to focus on the former, TFA focuses on the latter.

Can it do tasks, or jobs? Jobs are bundles of tasks. AI might be able to do 90% of tasks for a given job, but not the whole job.

If tasks, what counts as a task: Is it only specific things with clear success criteria? That's easier.

Is scaffolding allowed: Does it need to be able to do the tasks/jobs without scaffolding and human-written few-shot prompts?

Today's tasks/jobs only, or does it include future ones too? As tasks and jobs get automated, jobs evolve and get re-defined. So, being able to do the future jobs too is much harder.

Remote only, or in-person too: In-person too is a much higher bar.

What threshold of tasks/jobs: "most" is apparently typically understood to mean 80-95% (Mira Ariel). Automating 80% of tasks is different to 90% and 95% and 99%. diminishing returns. And how are the tasks counted - by frequency, by dollar-weighted, by unique count of tasks?

Only economically valuable tasks/jobs, or does it include anything a human can do?

A high-order bit on many people's AGI timelines is which definition of AGI they're using, so clarifying the definition is nice.

AstroBen•2h ago
Not only tasks, but you need to look at the net effect

If it does an hour of tasks, but creates an additional hour of work for the worker...

nakamoto_damacy•2h ago
Here is a definition of AGI in 2025: Hype.
jsheard•2h ago
We'll know AGI has arrived when AGI researchers manage to go five minutes without publishing hallucinated citations.

https://x.com/m2saxon/status/1979349387391439198

artninja1988•2h ago
Came from the Google Docs to BibTeX conversion apparently

https://x.com/m2saxon/status/1979636202295980299

bonoboTP•2h ago
This looks like a knee-jerk reaction to the title instead of anything substantial.
MichaelZuo•2h ago
It does seem a bit ridiculous…
CamperBob2•2h ago
So infallibility is one of the necessary criteria for AGI? It does seem like a valid question to raise.

Edit due to rate-limiting, which in turn appears to be due to the inexplicable downvoting of my question: since you (JumpCrisscross) are imputing a human-like motivation to the model, it sounds like you're on the side of those who argue that AGI has already been achieved?

JumpCrisscross•2h ago
> infallibility

Lying != fallibility.

nativeit•1h ago
I’m gonna start referring to my own lies as “hallucinations”. I like the implication that I’m not lying, but rather speaking truthfully, sincerely, and confidently about things that never happened and/or don’t exist. Seems paradoxical, but this is what we’re effectively suggesting with “hallucinations”. LLMs necessarily lack things like imagination, or an ego that’s concerned with the appearance of being informed and factually correct, or awareness for how a lack of truth and honesty may affect users and society. In my (not-terribly-informed) opinion, I’d assert that precludes LLMs from even approximate levels of intelligence. They’re either quasi-intelligent entities who routinely lie to us, or they are complex machines that identify patterns and reconstruct plausible-sounding blocks of text without any awareness of abstract concepts like “truth”.

Edit: toned down the preachiness.

xnx•2h ago
I like François Chollet definition of AGI as a system that can efficiently acquire new skills outside its training data.
zulban•2h ago
Not bad. Maybe.

But maybe that's ASI. Whereas I consider chatgpt 3 to be "baby AGI". That's why it became so popular so fast.

JumpCrisscross•2h ago
> I consider chatgpt 3 to be "baby AGI". That's why it became so popular so fast

ChatGPT became popular because it was easy to use and amusing. (LLM UX until then had been crappy.)

Not sure AGI aspirations had anything to do with uptake.

moffkalast•2h ago
So... AGI is a few shot performance metric?
Der_Einzige•2h ago
Most people who say "AGI" really mean either "ASI" or "Recursive Self Improvement".

AGI was already here the day ChatGPT released: That's Peter Norvig's take too: https://www.noemamag.com/artificial-general-intelligence-is-...

mitthrowaway2•2h ago
The reason some people treat these as equivalent is that AI algorithm research is one of the things a well-educated adult human can do, so an AGI who commits to that task should be able to improve itself, and if it makes a substantial improvement, then it would become or be replaced by an ASI.

To some people this is self-evident so the terms are equivalent, but it does require some extra assumptions: that the AI would spend time developing AI, that human intelligence isn't already the maximum reachable limit, and that the AGI really is an AGI capable of novel research beyond parroting from its training set.

I think those assumptions are pretty easy to grant, but to some people they're obviously true and to others they're obviously false. So depending on your views on those, AGI and ASI will or will not mean the same thing.

throwanem•2h ago
How, summing (not averaging) to 58 of 1000 possible points (0-100 in each of ten domains), are we calling this score 58% rather than 5.8%?
alexwebb2•2h ago
0-10 in each domain. It’s a weird table.
NitpickLawyer•2h ago
It's confusing. The 10 tracks each get 10%. So they add up all the percentages from every track. When you see the first table, 10% on math means "perfect" math basically. Not 10% of math track.
NitpickLawyer•2h ago
Interesting read. I agree completely with their Introduction, that the definition of AGI is constantly shifting, and this leads to endless (and useless) debates.

What I find cool about the paper is that they have gathered folks from lots of places (berkley, stanford, mit, etc). And no big4 labs. That's good imo.

tl;dr; Their definition: "AGI is an AI that can match or exceed the cognitive versatility and proficiency of a well-educated adult."

Cool. It's a definition. I doubt it will be agreed on by everyone, and I can see endless debates about just about every word in that definition. That's not gonna change. At least it's a starting point.

What I find interesting is that they specifically say it's not a benchmark, or a test set. It's a framework where they detail what should be tested, and how (with examples). They do have a "catchy" table with gpt4 vs gpt5, that I bet will be covered by every mainstream/blog/forum/etc out there -> gpt5 is at ~50% AGI. Big title. You won't believe where it was one year ago. Number 7 will shock you. And all that jazz.

Anyway, I don't think people will stop debating about AGI. And I doubt this methodology will be agreed on by everyone. At the end of the day both extremes are more ideological in nature than pragmatic. Both end want/need their view to be correct.

I enjoyed reading it. Don't think it will settle anything. And, as someone posted below, when the first model will hit 100% on their framework, we'll find new frameworks to debate about, just like we did with the turing test :)

mitthrowaway2•2h ago
Quite the list of authors. If they all personally approved the text, it's an achievement in itself just to get all of them to agree on a definition.
mrsvanwinkle•2h ago
indeed, i am wondering if these hn comments actually have an idea and they rub shoulders with these names with their dismissive confidence.
optimalsolver•1h ago
Maybe one of these exalted names should've proof-read the paper:

https://x.com/m2saxon/status/1979349387391439198

modeless•2h ago
GPT-5 scores 58%? That seems way too high. GPT-5 is good but it is not that close to AGI.

Also, weird to see Gary Marcus and Yoshua Bengio on the same paper. Who really wrote this? Author lists are so performative now.

jonplackett•2h ago
As anyone using AI knows - the first 90% is easy, the next 9% is much harder and the last 1% takes more time than the other 99%.
stared•2h ago
There’s already a vague definition that AGI is an AI with all the cognitive capabilities of a human. Yes, it’s vague - people differ.

This paper promises to fix "the lack of a concrete definition for Artificial General Intelligence", yet it still relies on the vague notion of a "well-educated adult". That’s especially peculiar, since in many fields AI is already beyond the level of an adult.

You might say this is about "jaggedness", because AI clearly lacks quite a few skills:

> Application of this framework reveals a highly “jagged” cognitive profile in contemporary models.

But all intelligence, of any sort, is "jagged" when measured against a different set of problems or environments.

So, if that’s the case, this isn’t really a framework for AGI; it’s a framework for measuring AI along a particular set of dimensions. A more honest title might be: "A Framework for Measuring the Jaggedness of AI Against the Cattell–Horn–Carroll Theory". It wouldn't be nearly as sexy, though.

fjdjshsh•2h ago
>But all intelligence, of any sort, is "jagged" when measured against a different set of problems or environments.

On the other hand, research on "common intelligence" AFAIK shows that most measures of different types of intelligence have a very high correlation and some (apologies, I don't know the literature) have posited that we should think about some "general common intelligence" to understand this.

The surprising thing about AI so far is how much more jagged it is wrt to human intelligence

stared•2h ago
I think you are talking about correlation in humans of, say, verbal and mathematical intelligence. Still, it is a correlation, not equality - there are many word-acknowledged writers who suck at math, and mathematical prodigies who are are not the best at writing.

If you go beyond human species (and well, computers are not even living organisms), it gets tricky. Adaptability (which is arguably a broader concept than intelligence) is very different for, say octopodes, corvids and slime molds.

It is certainly not a single line of proficiency or progress. Things look like lines only if we zoom a lot.

pixl97•1h ago
Human intelligence has had hundreds of thousands of years of evolution that removes any 'fatal' variance from our intelligence. Too dumb is obvious on how it's culled, but 'too smart' can get culled by social creatures too, really 'too different' in any way.

Current AI is in its infancy and we're just throwing data at it in the same way evolution throws random change at our DNA and sees what sticks.

bee_rider•2h ago
Huh. I haven’t read the paper yet. But, it seems like a weird idea—wouldn’t the standard of “well educated (I assume, modern) adult” preclude the vast majority of humans who ever lived from being considered general intelligences?
vidarh•2h ago
And this is indeed a huge problem with a lot of the attacks on LLM even as more limited AI - a lot of them are based on applying arbitrary standards without even trying to benchmark against people, and without people being willing to discuss where they draw the line for stating that a given subset of people do not possess general intelligence...

I think people get really uncomfortable trying to even tackle that, and realistically for a huge set of AI tasks we need AI that are more intelligent than a huge subset of humans for it to be useful. But there are also a lot of tasks where AI that is not needed, and we "just" need "more human failure modes".

catlifeonmars•2h ago
I read this as a hypothetical well-educated adult. As in, given the same level of knowledge, the intelligence performs equally well.

I do agree that it’s a weird standard though. Many of our AI implementations exceed the level of knowledge of a well-educated adult (and still underperform with that advantage in many contexts).

Personally, I don’t think defining AGI is particularly useful. It is just a marketing term. Rather, it’s more useful to just speak about features/capabilities. Shorthand for a specific set of capabilities will arise naturally.

jedberg•2h ago
To define AGI, we'd first have to define GI. Humans are very different. As park rangers like to say, there is an overlap between the smartest bears and the dumbest humans, which is why sometimes people can't open bear-proof trash cans.

It's a similar debate with self driving cars. They already drive better than most people in most situations (some humans crash and can't drive in the snow either for example).

Ultimately, defining AGI seems like a fools errand. At some point the AI will be good enough to do the tasks that some humans do (it already is!). That's all that really matters here.

lukan•1h ago
" At some point the AI will be good enough to do the tasks that some humans do (it already is!). That's all that really matters here."

What matters to me is, if the "AGI" can reliably solve the tasks that I give to it and that requires also reliable learning.

LLM's are far from that. It takes special human AGI to train them to make progress.

jedberg•1h ago
> What matters to me is, if the "AGI" can reliably solve the tasks that I give to it and that requires also reliable learning.

How many humans do you know that can do that?

l5870uoo9y•2h ago
Long-term memory storage capacity[1] scores 0 for both GPT-4 and GPT-5. Are there any workable ideas or concepts for solving this?

[1]: The capability to continually learn new information (associative, meaningful, and verbatim). (from the publication)

ants_everywhere•2h ago
> To operationalize this, we ground our methodology in Cattell-Horn-Carroll theory, the most empirically validated model of human cognition

Cattell-Horn-Carroll theory, like a lot of psychometric research, is based on collecting a lot of data and running factor analysis (or similar) to look for axes that seem orthogonal.

It's not clear that the axes are necessary or sufficient to define intelligence, especially if the goal is to define intelligence that applies to non-humans.

For example reading and writing ability and visual processing imply the organism has light sensors, which it may not. Do all intelligent beings have vision? I don't see an obvious reason why they would.

Whatever definition you use for AGI probably shouldn't depend heavily on having analyzed human-specific data for the same reason that your definition of what counts as music shouldn't depend entirely on inferences from a single genre.

zkmon•2h ago
The problem, I guess, with these methods is, they consider human intelligence as something detached from human biology. I think this is incorrect. Everything that goes in the human mind is firmly rooted in the biological state of that human, and the biological cycles that evolved through millennia.

Things like chess-playing skill of a machine could be bench-marked against that of a human, but the abstract feelings that drive reasoning and correlations inside a human mind are more biological than logical.

Workaccount2•2h ago
There is no reason to believe that consciousness, sentience, or emotions require a biological base.
steve_adams_86•2h ago
Is there more reason to believe otherwise? I'm not being contrarian, I'm genuinely curious what people think.
ComplexSystems•2h ago
What is the "irreplaceable" part of human biology that leads to consciousness? Microtubules? Whatever it is, we could presumably build something artificial that has it.
dangus•1h ago
We “could presumably build” it, maybe we can do that once we figure out how to get a language prediction model to comprehend what the current date is or how to spell strawberry.
lijok•1h ago
Don’t fool yourself into believing artificial intelligence is not one breakthrough away
AnimalMuppet•1h ago
All right, same question: Is there more reason to believe that it is one breakthrough away, or to believe that it is not? What evidence do you see to lean one way or the other?
lijok•1h ago
It’s clearly possible, because we exist. Just a matter of time. And as we’ve seen in the past, breakthroughs can produce incredible leaps in capabilities (outside of AI as well). We might not get that breakthrough(s) for a thousand years, but I’m definitely leaning towards it being inevitable.

Interestingly the people doing the actual envelope pushing in this domain, such as Ilya Sutskever, think that there it’s a scaling problem, and neural nets do result in AGIs eventually, but I haven’t heard them substantiate it.

AnimalMuppet•24m ago
You didn't answer the question. Zero breakthroughs away, one, or more than one? How strongly do you think whichever you think, and why?

(I'm asking because of your statement, "Don’t fool yourself into believing artificial intelligence is not one breakthrough away", which I'm not sure I understand, but if I am parsing it correctly, I question your basis for saying it.)

lijok•19m ago
There are definitely breakthroughs in the way.

“one breakthrough away” as in some breakthrough away

nativeit•1h ago
Douglas Hofstadter wrote Gödel, Escher, Bach in the late 1970s. He used the short-hand “strange loops”, but dedicates a good bit of time considering this very thing. It’s like the Ship of Theseus, or the famous debate over Star Trek transporters—at what point do we stop being an inanimate clump of chemical compounds, and become “alive”. Further, at what point do our sensory organs transition from the basics of “life”, and form “consciousness”.

I find anyone with confident answers to questions like these immediately suspect.

Lerc•1h ago
That asks you to consider the statements

There is reason to believe that consciousness, sentience, or emotions require a biological base.

Or

There is no reason to believe that consciousness, sentience, or emotions do not require a biological base.

The first is simple, if there is a reason you can ask for it and evaluate it's merits. Quantum stuff is often pointed to here, but the reasoning is unconvincing.

The second form There is no reason to believe P does not require Q.

There are no proven reasons but there are suspected reasons. For instance if the operation that nerons perform is what makes consciousness work, and that operation can be reproduced non-biologicLly it would follow that non biological consciousness would be possible.

For any observable phenomenon in the brain the same thing can be asked. So far it seems reasonable to expect most of the observable processes could be replicated.

None of it acts as proof, but they probably rise to the bar of reasons.

nebezb•2h ago
I’m certainly not informed enough to have an intelligent conversation about this, but surely the emotions bit can’t be right?

My emotions are definitely a function of the chemical soup my brain is sitting in (or the opposite).

BugsJustFindMe•1h ago
Your emotions are surely caused by the chemical soup, but chemical soup need not be the only way to arrive at emotions. It is possible for different mechanisms to achieve same outcomes.
nebezb•1h ago
Ah, I understand the statement now.
chongli•1h ago
How do we know we've achieved that? A machine that can feel emotions rather than merely emulating emotional behaviour.
BugsJustFindMe•1h ago
> How do we know we've achieved that? A machine that can feel emotions rather than merely emulating emotional behaviour.

Let me pose back to you a related question as my answer: How do you know that I feel emotions rather than merely emulating emotional behavior?

This gets into the philosophy of knowing anything at all. Descartes would say that you can't. So we acknowledge the limitation and do our best to build functional models that help us do things other than wallow in existential loneliness.

lijok•1h ago
Because I can watch you dream and can measure the fact you’re asleep.
BugsJustFindMe•1h ago
Philosophers have been worrying about the question of how you can know anything for thousands of years. I promise that your pithy answer here is not it.
lijok•1h ago
A promise wont do it. You’ll have to substantiate it without resorting to argument from authority.
BugsJustFindMe•1h ago
It's dangerous to go alone! Take this! https://en.wikipedia.org/wiki/Epistemology
lijok•1h ago
If you’re not interested in engaging in a discussion, why bother replying?
baxtr•1h ago
And Popper would say you cannot ever prove another mind or inner state, just as you cannot prove any theory.

But you can propose explanations and try to falsify them. I haven’t thought about it but maybe there is a way to construct an experiment to falsify the claim that you don’t feel emotions.

BugsJustFindMe•1h ago
I suppose there may be a way for me to conduct an experiment on myself, though like you I don't have one readily at hand, but I don't think there's a way for you to conduct such an experiment on me.
card_zero•1h ago
I wonder what Popper did say specifically about qualia and such. There's a 1977 book called "The Self and Its Brain: An Argument for Interactionism". Haven't read it.

Preface:

The problem of the relation between our bodies and our minds, and especially of the link between brain structures and processes on the one hand and mental dispositions and events on the other is an exceedingly difficult one. Without pretending to be able to foresee future developments, both authors of this book think it improbable that the problem will ever be solved, in the sense that we shall really understand this relation. We think that no more can be expected than to make a little progress here or there.

... well. Thanks a bunch, Karl.

nativeit•1h ago
Because you and I are the same species speaking a common language.
BugsJustFindMe•1h ago
Ok, but ChatGPT speaks this language just as well as I do, and we also know that emotion isn't a core requirement of being a member of this species because psychopaths exist.

Also, you don't know what species I am. Maybe I'm a dog. :-)

(https://en.wikipedia.org/wiki/On_the_Internet,_nobody_knows_...)

nativeit•1h ago
That sounds awful.
BugsJustFindMe•1h ago
This sounds like a bot comment.
nativeit•1h ago
How would you know? Bots speak just as well as you do.
chongli•1h ago
I don't know, but I have substantial evidence:

1) I know that I have emotions because I experience them.

2) I know that you and I are very similar because we are both human.

3) I know that we can observe changes in the brain as a result of our changing emotions and that changes to our brains can affect our emotions.

I thus have good reason to believe that since I experience emotions and that we are both human, you experience emotions too.

The alternative explanation, that you are otherwise human and display all the hallmarks of having emotions but do not in fact experience anything (the P-zombie hypothesis), is an extraordinary claim that has no evidence to support it and not even a plausible, hypothetical mechanism of action.

With an emotional machine I see no immediately obvious even hypothetical evidence to lend support to its veracity. In light of all this, it seems extraordinary to claim that non-biological means achieving real emotions (not emulated emotions) are possible.

After all, emulated emotions have already been demonstrated in video games. To call those sufficient would be setting an extremely low bar.

jll29•1h ago
Perhaps we could say we don't know whether the human biological substrate is required for mental processes or not, but either way we do not know enough about said biological substrate and our mental processes, respectively.
dangus•1h ago
What if our definition of those concepts is biological to begin with?

How does a computer with full AGI experience the feeling of butterflies in your stomach when your first love is required?

How does a computer experience the tightening of your chest when you have a panic attack?

How does a computer experience the effects of chemicals like adrenaline or dopamine?

The A in AGI stands for “artificial” for good reason, IMO. A computer system can understand these concepts by description or recognize some of them them by computer vision, audio, or other sensors, but it seems as though it will always lack sufficient biological context to experience true consciousness.

Perhaps humans are just biological computers, but the “biological” part could be the most important part of that equation.

sim7c00•1h ago
they do not, but the same argument can hold true by the fact the true human nature is not really known and thus trying to define what a human like intelligence would consist of can only be incomplete.

there are many parts of human cognition, phycology etc. especially related to consciousness that are known unknowns and/or completely unknown.

a mitigation for this isaue would be to call it generally applicable intelligence or something, rather than human like intelligence. implying ita not specialized AI but also not human like. (i dont see why it would need to be human like, because even with all the right logic and intelligence a human can still do something counter to all of that. humans do this everyday. intuitive action, or irrational action etc.

what we want is generally applicable intelligence, not human like intelligence.

zkmon•1h ago
None of that comes from outside of your biology and chemistry.
vhantz•1h ago
What non-biological systems do we know of that have consciousness, sentience or emotions?
BugsJustFindMe•1h ago
We have no known basis for even deciding that other than the (maybe right, maybe wrong) guess that consciousness requires a lot of organized moving complexity. Even with that guess, we don't know how much is needed or what kind.
nativeit•1h ago
It’s frequently pretty funny, anyway.
BugsJustFindMe•1h ago
This sounds like a bot comment.
nativeit•1h ago
Well, you do tend to repeat yourself, maybe ChatGPT really is your peer with language?
runarberg•1h ago
There is exactly one good reason, at least for consciousness and sentience. And the reason is that those are such a vaguely defined (or rather defined by prototypes; ala Wittgenstein [or JavaScript before classes]). And that reason is anthropism.

We only have one good example of consciousness and sentience, and that is our own. We have good reason to suspect other entities (particularly other human individuals, but also other animals) have that as well, but we cannot access it, and not even confirm its existence. As a result using these terms of non-human beings becomes confusing at best, but it will never be actually helpful.

Emotions are another thing, we can define that outside of our experience, using behavior states and its connection with patterns of stimuli. For that we can certainly observe and describe behavior of a non biological entity as emotional. But given that emotion is something which regulates behavior which has evolved over millions of years, whether such a description would be useful is a whole another matter. I would be inclined to use a more general description of behavior patterns which includes emotion but also other means of behavior regulators.

arbirk•1h ago
What about learning? As humans we continually update our weights from sensing the world. Before the AI can rewrite itself it can't really be AGI imo
Geee•1h ago
How about AFI - artificial fast idiot. Dumber than a baby, but faster than an adult. Or AHI - artificial human imitator.

This is bad definition, because human baby is already AGI when it's born and it's brain is empty. AGI is the blank slate and ability to learn anything.

jagrsw•1h ago
That "blank slate" idea doesn't really apply to humans, either.

We are born with inherited "data" - innate behaviors, basic pattern recognition, etc. Some even claim that we're born with basic physics toolkit (things are generally solid, they move). We then build on that by being imitators, amassing new skills and methods simply by observation and performing search.

Geee•1h ago
Sure, there's lots of inbuilt stuff like basic needs and emotions. But still, baby doesn't know anything about the world. It's the ability to collect data and train on it that makes it AGI.
dwa3592•1h ago
Everyone has a definition and so have I. I would call it an AGI when i replace my smartphone and laptop with it. When my screen time is zero? Can AGI replace screens? Go figure.
golol•1h ago
>Paper claims definition of AGI >Look inside >No definition of AGI.
flkiwi•1h ago
> defining AGI as matching the cognitive versatility and proficiency of a well-educated adult

I don't think people really realize how extraordinary accomplishment it would be to have an artificial system matching the cognitive versatility and proficiency of an uneducated child, much less a well-educated adult. Hell, AI matching the intelligence of some nonhuman animals would be an epoch-defining accomplishment.

cbdevidal•1h ago
Have any benchmarks been made that use this paper’s definition? I follow the ARC prize and Humanity’s Last Exam, but I don’t know how closely they would map to this paper’s methods.

Edit: Probably not, since it was published less than a week ago :-) I’ll be watching for benchmarks.

surgical_fire•1h ago
There are some sycophants that claim that LLMs can operate at Junior Enginee level.

Try to reconcile that with your ideas (that I think are correct for that matter)

ben_w•1h ago
I'll simultaneously call all current ML models "stupid" and also say that SOTA LLMs can operate at junior (software) engineer level.

This is because I use "stupidity" as the number of examples some intelligence needs in order to learn from, while performance is limited to the quality of the output.

LLMs *partially* make up for being too stupid to live (literally: no living thing could survive if it needed so many examples) by going through each example faster than any living thing ever could — by as many orders of magnitude as there are between jogging and continental drift.

ACCount37•42m ago
Data-efficiency matters, but compute-efficiency matters too.

LLMs have a reasonable learning rate at inference time (in-context learning is powerful), but a very poor learning rate in pretraining. And one issue with that is that we have an awful lot of cheap data to pretrain those LLMs with.

We don't know how much compute human brain uses to do what it does. And if we could pretrain with the same data-efficiency as humans, but at the cost of using x10000 the compute for it?

It would be impossible to justify doing that for all but the most expensive, hard-to-come-by gold-plated datasets - ones that are actually worth squeezing every drop of performance gains out from.

card_zero•35m ago
(10 orders of magnitude, it works out neatly as 8km/h for a fast jogger against 0.0008 mm/h for the East African Rift.)
ben_w•1h ago
Or even to come up with a definition of cognitive versatility and proficiency that is good enough to not get argued away once we have an AI which technically passes that specific definition.

The Turing Test was great until something that passed it (with an average human as interrogator) turned out to also not be able to count letters in a word — because only a special kind of human interrogator (the "scientist or QA" kind) could even think to ask that kind of question.

lumost•48m ago
Or that this system would fail to adapt in anyway to changes of circumstance. The adaptive intelligence of a live human is truly incredible. Even in cases where the weights are updatable, We watch AI make the same mistake thousands of times in an RL loop before attempting a different strategy.
oidar•1h ago
This is fine for a definition of AGI, but it's incomplete. It misses so many parts of the cognition that make humans flexible and successful. For example, emotions, feelings, varied pattern recognition, propreception, embodied awareness, social skills, and navigating ambiguous situation w/o algorithms. If the described 10 spectrums of intelligence were maxed by an LLM, it would still fall short.
pixl97•1h ago
Eh, I don't like the idea of 'intelligence' of any type using humans as the base line. It blinds it to our own limitations and things that may not be limits to other types of intelligence. The "AI won't kill us all because it doesn't have emotions" problem is one of these. For example, just because AI doesn't get angry, doesn't mean it can't recognize your anger and manipulate if given such a directive to.
oidar•15m ago
I agree, my point is that the cognition that creates emotion (and others) is of a different quality than the 10 listed in the paper.
jal278•1h ago
The fundamental premise of this paper seems flawed -- take a measure specifically designed for the nuances of how human performance on a benchmark correlates with intelligence in the real world, and then pretend as if it makes sense to judge a machine's intelligence on that same basis, when machines do best on these kinds of benchmarks in a way that falls apart when it comes to the messiness of the real world.

This paper, for example, uses the 'dual N-back test' as part of its evaluation. In humans this relates to variation in our ability to use working memory, which in humans relates to 'g'; but it seems pretty meaningless when applied to transformers -- because the task itself has nothing intrinsically to do with intelligence, and of course 'dual N-back' should be easy for transformers -- they should have complete recall over their large context window.

Human intelligence tests are designed to measure variation in human intelligence -- it's silly to take those same isolated benchmarks and pretend they mean the same thing when applied to machines. Obviously a machine doing well on an IQ test doesn't mean that it will be able to do what a high IQ person could do in the messy real world; it's a benchmark, and it's only a meaningful benchmark because in humans IQ measures are designed to correlate with long-term outcomes and abilities.

That is, in humans, performance on these isolated benchmarks is correlated with our ability to exist in the messy real-world, but for AI, that correlation doesn't exist -- because the tests weren't designed to measure 'intelligence' per se, but human intelligence in the context of human lives.

bananaflag•56m ago
I can define AGI in a line:

an entity which is better than any human at any task.

Fight me!

UltraSane•40m ago
I would define AGI as any artificial system that could learn any skill a human can by using the same inputs.