frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

AGI Is Here

https://www.robinsloan.com/winter-garden/agi-is-here/
15•cmod•1d ago

Comments

akagusu•1d ago
Who cares about AGI? If it happens some day, it will not good for me and I can do nothing about it, so, who cares?
rvz•1d ago
Spoiler: It isn’t.

Maybe AGI is here for the author and mediocre web developers, otherwise the big AI labs would have replaced their AI researchers already and commercial airliners would have already replaced their pilots with GPTs.

This is exactly why “AGI” is meaningless.

shmerl•1d ago
No, it's not here. Sophisticated automatic parrot is not intelligent.
petermcneeley•1d ago
AGI is when the last human is terminated.
lostmsu•1d ago
What if AIs move that goal post further?
gedy•1d ago
I'll say AGI is "here" when they are agents in our Slack/chat, emails, and planning tools doing that strategic work that management does. As a dev, "AI" is currently just a handy English language interface to web search.
denkmoon•1d ago
You could argue LLMs already achieve the same work that management does.
gedy•1d ago
Yeah I think a lot of the mediocre PM type work of getting from this tool, input, etc to summarizing requirements into that tool, etc is exactly what LLMs do well.
ares623•1d ago
I heard OpenAI are going to put together a panel of experts to declare AGI. I wonder how that's going.
danpalmer•1d ago
> If you appeared in a puff of smoke before the authors of that paper, just after publication — a few months before half of them cleaved from OpenAI to form Anthropic — and carried with you a laptop linked through time to the big models of 2026, what would their appraisal be ? There’s no doubt in my mind they would say: Wow, we really did it ! This is obviously AGI!

I really don't think this would be the reaction. I'd say they would (or should) look at the systems we have now and see a very clear path between where they were then and where we are now, with all the positives _and negatives_. We still get hallucinations. We still get misalignment, if anything as capabilities have improved so has the potential for damage when things go wrong. It's pretty clear to me that late 2025 models are just better versions of what we had in 2021.

That's not to say they're not more useful, more valuable, they absolutely are! But that's all about product integrations, speed, and turning up the dial on inference compute. They're still fundamentally the same things.

The next big step forward, the thing that LLMs are obviously missing, is memory. The fact we're messing around with context windows, attention across the context space, chat lookup and fact saving features, etc, are all patches over the fact that LLMs can't remember anything in the way that humans (or pretty much any animal) can. It's clear that we need a paradigm shift on memory to unlock the next level of performance.

rvz•1d ago
> It's clear that we need a paradigm shift on memory to unlock the next level of performance.

I think this is on point to the next phase of LLMs or a different neural network architecture that improves on top of them, alongside continual learning.

Adding memory capabilities would mostly benefit local "reasoning" models than online ones as you would be saving tokens to do more tasks, than generating more tokens to use more "skills" or tools. (Unless you pay more for memory capabilities to Anthropic or OpenAI).

It's kind of why you see LLMs being unable to play certain games or doing hundreds of visual tasks very quickly without adding lots of harnesses and tools or giving it a pre-defined map to help it understand the visual setting.

As I said before [0], the easiest way to understand the memory limitations with LLMs is Claude Playing Pokemon with it struggling with basic tasks that a 5 year old can learn continuously.

[0] https://news.ycombinator.com/item?id=43291895

danpalmer•1d ago
Continual learning is definitely part of it. Perhaps part of it (or something else) is learning much faster from many fewer examples.
fragmede•1d ago
with beads, or shoving it in git, or .MD files, it's not clear that we do.
danpalmer•1d ago
These are all very much in the same category of hacks that I mentioned.

A cat doesn't know its way around a house when it's born, but it also doesn't have to flick through markdown files to find its way around. A child can touch a hot stove once and be neurotic about touching hot things for the rest of their life, without having to read flash cards each morning or think for a few minutes about "what do I know about stoves" every time they're in the kitchen.

fragmede•1d ago
Call them a "hack" all you want, they seem to work. What's particularly intesting is how claude has been trained on skills, so it doesn't need to be taught how to use a skill, so that's been baked into it.
danpalmer•1d ago
I'm not claiming they don't work in some sense, but as a user you have to be fairly deeply aware of how they work, context engineering is A Thing, you have to tell LLMs to remember stuff, etc.

We're hacking around the fact that the models don't learn in normal use. That's in no way controversial.

A model that continuously learnt would not need the same sort of context engineering, external memory databases, etc.

fragmede•12h ago
You speak the truth but looking back, what I reacted to is

> It's clear that we need a paradigm shift on memory to unlock the next level of performance.

and my take is that we might not need to get there to get the next level of performance, based on how well the latest models are able to utilize these hacks of a memory feature. On top of that, Claude was specifically RLHF'd to have the skills concept, so it's good with those. We disagree. Let's let time see who ends up being right.

Yizahi•1d ago
We have LLM memory, it's a training data from which the model was initially programmed. To allow adding or changing LLM memory, we would need to retrain model completely or partially. And that is not realistic any time soon. All other attempts at LLM memory would be just an obscure hack of splitting context window into parts and feeding input from different files. Literally nothing would change if you input half of the query from one file, half from another called "memory.txt" or if you just input whole query from a single file twice as big.
tra3•1d ago
I feel like it's kinda maybe here. Stochastic parrot or not, I can ask for "tea, earl grey, hot" and get an orange juice. It's way better than this time last year.

It's not perfect, but it doesn't need to be, to be useful.

tbrownaw•1d ago
I want a thing that has a fuzzy capability <X>.

To make this easier to discuss and think about, here's a short name <Y> for that thing, and a longer definition <Z> that probably describes that thing.

.

Oh look, if I take the literal definitions of some words in <Y>, there's a thing that fits!

Therefore <X> is solved!

palmotea•1d ago
Typical software engineer "reasoning."
albatross79•1d ago
It's not even AI, let alone AGI. It's a high dimensional statistical map of language, that's all.
CamperBob2•23h ago
"It's a mindless next-token predictor," says the human, as he skillfully emulates a mindless next-token predictor.
albatross79•22h ago
"the apparent is the same as the real", "this two dimensional photo is the same as the world it was taken in" says the tech bro, as he unskilfully attempts to emulate a philosopher.
smitty1e•1d ago
Let me know when AGI has fixed the tax code and streamlined the Code of Federal Regulations.
Terr_•1d ago
[Recycled from a dupe submission]

> This is why I pro­pose uni­lat­eral dec­la­ra­tion as a strategic countermove [... tearing] away the veil of any-minute-now mil­lenar­i­anism to reveal deployed tech­nology

I think that in an ideal world, this would thoroughly embarrass the over-promisers by forcing them to put-up-or-shut-up, and it's fun to imagine... however I worry that it won't work out that way. Instead of deflating the nonsense in its tracks, it'll just give it more momentum and worsen the eventual mess.

> What do I mean by AGI ?

Can we fight it with a better term? Something like... Oh, I dunno, maybe "Artificial Narrative Intelligence", in the same sense that we could say A* is a kind of pathfinding intelligence.

I say "narrative" because we've got these machines that grow "fitting" documents, and are often used with stories to "decide" what happens next. For example, the story setting is a Support Page, the Customer Character says X, and the Very Helpful Robot Character then does Y and says Z in response, etc.

However just because these stories fit surprisingly well doesn't mean it's doing the kind of "thinking" we really dreamed of.

> You some­times read about employees of AI com­pa­nies absorbed by their own products. Nobody on Earth has spent more hours talking to YakGPT than Katie Echo! Nobody can pump more code out of Shan­non­Soft than Johnny Narcissus! Recalling my Twitter expe­ri­ence, I think boasts (and posts) of this kind should inspire caution.

To me a lot of that feels like just the thing-of-the-day LinkedIn Lunacy, albeit running at an unusual intensity.

andsoitis•1d ago
> AGI is here!

Can AGI not speak for itself? Does it need humans to speak and act on its behalf? Who are the high-priests and what are the sects?

Spherical Snake

https://kevinalbs.com/spherical_snake/
258•subset•5d ago•61 comments

Oral microbiome sequencing after taking probiotics

https://blog.booleanbiotech.com/oral-microbiome-biogaia
48•sethbannon•1h ago•10 comments

Calling All Hackers: How money works (2024)

https://phrack.org/issues/71/17
58•krrishd•2h ago•3 comments

Comparing AI agents to cybersecurity professionals in real-world pen testing

https://arxiv.org/abs/2512.09882
34•littlexsparkee•1h ago•12 comments

Vienam bans unskippable ads

https://saigoneer.com/vietnam-news/28652-vienam-bans-unskippable-ads,-requires-skip-button-to-app...
904•hoherd•5h ago•484 comments

CES 2026: Taking the Lids Off AMD's Venice and MI400 SoCs

https://chipsandcheese.com/p/ces-2026-taking-the-lids-off-amds
13•rbanffy•42m ago•2 comments

Laylo (YC S20) – Head of Growth (Organic and Partners and Loops and AI) – Remote US

https://www.ycombinator.com/companies/laylo/jobs/ZtLHRXe-head-of-growth
1•amellin794•43m ago

Video Game Websites in the early 00s

https://www.webdesignmuseum.org/exhibitions/video-game-websites-in-the-early-00s
89•klaussilveira•3h ago•50 comments

Locating a Photo of a Vehicle in 30 Seconds with GeoSpy

https://geospy.ai/blog/locating-a-photo-of-a-vehicle-in-30-seconds-with-geospy
72•kachapopopow•4h ago•70 comments

Launch HN: Tamarind Bio (YC W24) – AI Inference Provider for Drug Discovery

42•denizkavi•4h ago•13 comments

High-Performance DBMSs with io_uring: When and How to use it

https://arxiv.org/abs/2512.04859
59•matt_d•2h ago•12 comments

Opus 4.5 is not the normal AI agent experience that I have had thus far

https://burkeholland.github.io/posts/opus-4-5-change-everything/
153•tbassetto•4h ago•218 comments

Show HN: Foundertrace – chain of YC startups founded by its employees

https://foundertrace.com/
17•loondri•2d ago•1 comments

Show HN: Mantic.sh – A structural code search engine for AI agents

https://github.com/marcoaapfortes/Mantic.sh
35•marcoaapfortes•8h ago•11 comments

Show HN: Finding similarities in New Yorker covers

https://shoplurker.com/labs/newyorker-covers/
10•tkp-415•2h ago•0 comments

Passing of Joe Mancuso author of Masonite (Python web framework)

https://github.com/MasoniteFramework/masonite/discussions/853
104•wilsonfiifi•4h ago•7 comments

Show HN: Prism.Tools – Free and privacy-focused developer utilities

https://blgardner.github.io/prism.tools/
301•BLGardner•9h ago•88 comments

Show HN: 48-digit prime numbers every git commit

https://textonly.github.io/git-prime/
32•keepamovin•5d ago•10 comments

Stop Doom Scrolling, Start Doom Coding: Build via the terminal from your phone

https://github.com/rberg27/doom-coding
145•rbergamini27•2h ago•144 comments

A 30B Qwen Model Walks into a Raspberry Pi and Runs in Real Time

https://byteshape.com/blogs/Qwen3-30B-A3B-Instruct-2507/
11•dataminer•1h ago•0 comments

How HTML changes in ePub

https://www.htmhell.dev/adventcalendar/2025/11/
46•raybb•3d ago•11 comments

Loongarch Improvements with Box64

https://box86.org/2026/01/new-box64-v0-4-0-released/
52•aaronday•5h ago•2 comments

I wanted a camera that doesn't exist – so I built it

https://medium.com/@cristi.baluta/i-wanted-a-camera-that-doesnt-exist-so-i-built-it-5f9864533eb7
212•cyrc•4d ago•64 comments

Self hosting my media library with Jellyfin and Wireguard on Hetzner

https://layandreas.github.io/personal-blog/posts/how-spotify-made-me-self-host/
81•wismwasm•2h ago•105 comments

Hierarchical Autoregressive Modeling for Memory-Efficient Language Generation

https://arxiv.org/abs/2512.20687
35•PaulHoule•4h ago•0 comments

Why is the Gmail app 700 MB?

https://akr.am/blog/posts/why-is-the-gmail-app-700-mb
310•thefilmore•5h ago•277 comments

Why Big Companies Keep Failing: The Stack Fallacy (2016)

https://techcrunch.com/2016/01/18/why-big-companies-keep-failing-the-stack-fallacy/
55•bobbiechen•5h ago•38 comments

Dude, where's my supersonic jet?

https://rationaloptimistsociety.substack.com/p/dude-wheres-my-supersonic-jet
75•noleary•4h ago•177 comments

Volkswagen Brings Back Physical Buttons

https://www.caranddriver.com/news/a69916699/volkswagen-interior-physical-buttons-return/
261•stephc_int13•5h ago•6 comments

Writing an eigenvalue solver in Rust for WebAssembly

https://abstractnonsense.xyz/blog/2025-12-31-eigenvalue-solver-in-rust-for-webassembly/
10•subset•5d ago•0 comments