frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Replacement.ai

https://replacement.ai
139•wh313•48m ago•18 comments

Abandoned land drives dangerous heat in Houston, Texas A&M study finds

https://stories.tamu.edu/news/2025/10/07/abandoned-land-drives-dangerous-heat-in-houston-texas-am...
29•PaulHoule•1h ago•13 comments

Show HN: Duck-UI – Browser-Based SQL IDE for DuckDB

https://demo.duckui.com
79•caioricciuti•3h ago•25 comments

How to Assemble an Electric Heating Element from Scratch

https://solar.lowtechmagazine.com/2025/10/how-to-build-an-electric-heating-element-from-scratch/
9•surprisetalk•1h ago•1 comments

Improving PixelMelt's Kindle Web Deobfuscator

https://shkspr.mobi/blog/2025/10/improving-pixelmelts-kindle-web-deobfuscator/
31•ColinWright•2h ago•1 comments

The case for the return of fine-tuning

https://welovesota.com/article/the-case-for-the-return-of-fine-tuning
69•nanark•4h ago•23 comments

OpenAI researcher announced GPT-5 math breakthrough that never happened

https://the-decoder.com/leading-openai-researcher-announced-a-gpt-5-math-breakthrough-that-never-...
169•Topfi•3h ago•98 comments

EQ: A video about all forms of equalizers

https://www.youtube.com/watch?v=CLAt95PrwL4
211•robinhouston•1d ago•57 comments

Lost Jack Kerouac story found among assassinated mafia boss' belongings

https://www.sfgate.com/sf-culture/article/lost-jack-kerouac-chapter-found-mafia-boss-estate-21098...
26•rmason•4d ago•11 comments

A Tower on Billionaires' Row Is Full of Cracks. Who's to Blame?

https://www.nytimes.com/2025/10/19/nyregion/432-park-avenue-condo-tower.html
32•danso•1h ago•16 comments

Pebble is officially back on iOS and Android

https://twitter.com/ericmigi/status/1979576965494710564
56•vlod•2h ago•6 comments

Titan submersible’s $62 SanDisk memory card found undamaged at wreckage site

https://www.tomshardware.com/pc-components/microsd-cards/tragic-oceangate-titan-submersibles-usd6...
384•WithinReason•2d ago•187 comments

Jupyter Collaboration has a history slider

https://blog.jupyter.org/exploring-a-documents-timeline-in-jupyterlab-6084f96db263
38•fghorow•6d ago•10 comments

Chen-Ning Yang, Nobel laureate, dies at 103

https://www.chinadaily.com.cn/a/202510/18/WS68f3170ea310f735438b5bf2.html
257•nhatcher•1d ago•62 comments

The Accountability Problem

https://www.jamesshore.com/v2/blog/2025/the-accountability-problem
98•FrancoisBosun•12h ago•39 comments

Root System Drawings

https://images.wur.nl/digital/collection/coll13/search
378•bookofjoe•1d ago•74 comments

How one of the longest dinosaur trackways in the world was uncovered in the UK

https://www.bbc.co.uk/news/resources/idt-5f8c77b0-92bc-40f2-bf21-6793abbe5ffe
32•6LLvveMx2koXfwn•5d ago•2 comments

How to sequence your DNA for <$2k

https://maxlangenkamp.substack.com/p/how-to-sequence-your-dna-for-2k
207•yichab0d•18h ago•89 comments

When you opened a screen shot of a video in Paint, the video was playing in it

https://devblogs.microsoft.com/oldnewthing/20251014-00/?p=111681
354•birdculture•2d ago•59 comments

BQN "Macros" with •Decompose (2023)

https://saltysylvi.github.io/blog/bqn-macros.html
16•ofalkaed•1w ago•3 comments

Xubuntu.org Might Be Compromised

https://old.reddit.com/r/Ubuntu/comments/1oa4549/xubuntuorg_might_be_compromised/
4•kekqqq•10m ago•0 comments

Immix: A Mark-Region Garbage Collector (2008) [pdf]

https://www.steveblackburn.org/pubs/papers/immix-pldi-2008.pdf
4•djoldman•5d ago•0 comments

Why the open social web matters now

https://werd.io/why-the-open-social-web-matters-now/
192•benwerd•4d ago•116 comments

How does Turbo listen for Turbo Streams

https://ducktypelabs.com/how-does-turbo-listen-for-turbo-streams/
72•sidk_•5d ago•8 comments

GoGoGrandparent (YC S16) Is Hiring Back End and Full-Stack Engineers

1•davidchl•13h ago

Flowistry: An IDE plugin for Rust that focuses on relevant code

https://github.com/willcrichton/flowistry
254•Bogdanp•1d ago•33 comments

Tinnitus Neuromodulator

https://mynoise.net/NoiseMachines/neuromodulationTonesGenerator.php
342•gjvc•22h ago•214 comments

./watch

https://dotslashwatch.com/
368•shrx•1d ago•105 comments

K8s with 1M nodes

https://bchess.github.io/k8s-1m/
248•denysvitali•2d ago•61 comments

Secret diplomatic message deciphered after 350 years

https://www.nationalarchives.gov.uk/explore-the-collection/the-collection-blog/secret-diplomatic-...
168•robin_reala•2d ago•34 comments
Open in hackernews

OpenAI researcher announced GPT-5 math breakthrough that never happened

https://the-decoder.com/leading-openai-researcher-announced-a-gpt-5-math-breakthrough-that-never-happened/
168•Topfi•3h ago

Comments

amelius•1h ago
> Summary (from the article)

* OpenAI researchers claimed or suggested that GPT-5 had solved unsolved math problems, but in reality, the model only found known results that were unfamiliar to the operator of erdosproblems.com.

* Mathematician Thomas Bloom and Deepmind CEO Demis Hassabis criticized the announcement as misleading, leading the researchers to retract or amend their original claims.

* According to mathematician Terence Tao, AI models like GPT-5 are currently most helpful for speeding up basic research tasks such as literature review, rather than independently solving complex mathematical problems.

HarHarVeryFunny•33m ago
> GPT-5 had only surfaced existing research that Bloom had missed

So GPT-5 didn't derive anything itself - it was just an effective search engine for prior research, which is useful, but not any sort of breakthough whatsoever.

jgalt212•1h ago
After the circular financing schemes involving hundreds of billions of dollars were uncovered, nothing I read about the AI business and its artificial hype machine surprises me anymore.
bbor•1h ago
This is just tit-for-tat clickbait. The researcher’s wording was a bit unclear for sure, but far from incorrect.
resoluteteeth•1h ago
I disagree. There is no way to interpret "GPT-5 just found solutions to 10 (!) previously unsolved Erdos problems" as saying something other than GPT-5 having solved them.

If it just found existing solutions then they obviously weren't "previously unsolved" so the tweet is wrong.

He clearly misunderstood the situation and jumped to the conclusion that GPT-5 had actually solved the problems because that's what he wanted to believe.

That said, the misunderstanding is understandable because the tweet he was responding to said they had been listed as "open", but solving unsolved erdos problems by itself would be such a big deal that he probably should have double checked it.

andrewstuart•1h ago
Humans hallucinating about AI.
MattGaiser•1h ago
Humans "hallucinate" in the AI way constantly, which is why I don't see them as a barrier to LLMs replacing humans in many contexts. It really isn't unusual for a human to make stuff up or be unaware of stuff.
pas•1h ago
it's the same thing with self-driving, if you can make it safer than a good human driver it's enough. but the bar is pretty low with driving (as evidenced by the hundreds of thousands of collisions and deaths and permanent disabilities each year). and rather high in scientific publishing.
zeknife•1h ago
A human being informed of a mistake will usually be able to resolve it and learn something in the process, whereas an LLM is more likely to spiral into nonsense
alimw•59m ago
You must know better humans than I do.
MattGaiser•44m ago
You must know people without egos. Humans are better at correcting their mistakes, but far worse at admitting them.

But yes, as an edge case handler humans still have an edge.

topaz0•37m ago
LLMs by contrast love to admit their mistakes and self-flagellate, and then go on to not correct them. Seems like a worse tradeoff.
thaumasiotes•31m ago
It's true that the big public-facing chatbots love to admit to mistakes.

It's not obvious to me that they're better at admitting their mistakes. Part of being good at admitting mistakes is recognizing when you haven't made one. That humans tend to lean too far in that direction shouldn't suggest that the right amount of that behavior is... less than zero.

skeeter2020•22m ago
Not when your goal is to create ASI: Artificial Sycophant Intelligence
tonyhart7•34m ago
and this is why LLM is getting cooked

they feed an internet data into that shit, they basically "told" LLM to behave because surprise surprise, human sometimes can be more nasty

zeroonetwothree•43m ago
Humans are a bit better at knowing which things are important and doing more research. Also better at being honest when directly pressed. And infinitely better at learning from errors.

(Yes, not everyone, but we do have some mechanisms to judge or encourage)

lapcat•42m ago
> Humans "hallucinate" in the AI way constantly

This claim is ambiguous. The use of the word "Humans" here obscures rather than clarifies the issue. Individual humans typically do not "hallucinate" constantly, especially not on the job. Any individual human who is as bad at their job as an LLM should indeed be replaced, by a more competent individual human, not by an equally incompetent LLM. This was true long before LLMs were invented.

In the movie "Bill and Ted's Excellent Adventure," the titular characters attempt to write a history report by asking questions of random strangers in a convenience store parking lot. This of course is ridiculous and more a reflection of the extreme laziness of Bill and Ted than anything else. Today, the lazy Bill and Ted would ask ChatGPT instead. It's equally ridiculous to defend the wild inaccuracy and hallucinations of LLMs by comparing them to average humans. It's not the job of humans to answer random questions on any subject.

Human subject matter experts are not perfect, but they’re much better than average and don’t hallucinate on their subjects. They also have accountability and paper trails, can be individually discounted for gross misconduct, unlike LLMs.

skeeter2020•23m ago
Do you think the OpenAI human, when informed of their "oopsie" replied "You're right, there is existing evidence that this problem has already been solved. Blah Blah Blah ... and that's why our new model has made a huge breakthrough against previously unsolved math problems!"
random9749832•1h ago
Best case: Hallucination

Worst case (more probable): Lying

MPSimmons•1h ago
Hanlon's Razor
forgetfulness•1h ago
Lying is a stupid way of selling something and making money
reaperducer•26m ago
Lying is a stupid way of selling something and making money

Works for Elon.

random9749832•26m ago
They are expanding into the adult market because they are running out of ideas. I think common sense is enough to decide what is what here.
JKCalhoun•1h ago
"OpenAI Researcher Hallucinates GPT-5 Math Breakthrough" could be a headline from The Onion.
antegamisou•53m ago
Seriously those headlines are getting DailyMail sensationalism levels of ridiculous.
nicce•41m ago
I the old world we would just use the word bullshit.
reaperducer•28m ago
"OpenAI Researcher Hallucinates GPT-5 Math Breakthrough" could be a headline from The Onion.

Off topic, but I saw The Onion on sale in the magazine rack of Barnes and Noble last month.

For those who miss when it was a free rag in sidewalk newsstands, and don't want to pony up for a full subscription, this is an option.

pera•1h ago
Heh stockholders are not hallucinating: They know very well what they are doing.
skeeter2020•25m ago
retail investors? no way. The fever-dream may continue for a while but eventually it will end. Meanwhile we don't even know our full exposure to AI. It's going to be ugly and beyond burying gold in my backyard I can't even figure out how to hedge against this monster.
rixed•1h ago
These days AI just obsequiously praise whatever stupid ideas the human throw at them, which encourage humans into hallucinating breakthroughs.

But it's only a matter of time before AI gets better at prompt engineering.

/s?

Palmik•43m ago
More like humans hallucinating about humans hallucinating about AI, see here: https://news.ycombinator.com/item?id=45634120
alkyon•7m ago
They started believing the very lies they invented.
moffkalast•3m ago
"The truth is usually just an excuse for a lack of imagination."
mentalgear•1h ago
Another instance of openAI manipulating results to prolong their unsustainable circular hype bubble.

The inevitable collapse could be even more devastating than the 2008 financial crisis.

All while so vast resources are being wasted on non-verifiable gen AI slob, while real approaches (neuro-symbolic like DeepMind's AlphaFold) are mostly ignored financially because they don't generate the quick stock market increases that hype does.

the_duke•1h ago
People keep spouting this, but I don't see how the AI bubble bursting would be all that devastating.

2008 was a systemic breakdown rippling through the foundations of the financial system.

It would lead to a market crash (80% of gains this year were big tech/AI) and likely a full recession in the US, but nothing nearly as dramatic as a global systemic crisis.

In contrast to the dot com bubble, the huge AI spending is also concentrated on relatively few companies, many with deep pockets from other revenue sources (Google, Meta, Microsoft, Oracle), and the others are mostly private companies that won't have massive impact on the stock market.

A sudden stop in AI craze would be hard for hardware companies and a few big AI only startups , but the financial fallout would be much more contained than either dot com or 2008.

Topfi•1h ago
Isn't the dot com bubble a far better proxy? Notably, todays spending is both higher and more concentrated in a few companies that a large part of the population has exposure to (most dot com companies weren't publicly traded and far smaller vs MSFT, Alphabet, Meta, Oracle, NVDA making up most investment today) by way of pension funds, ETFs, etc.?
the_duke•1h ago
Sure, but all of the above have solid businesses that rake in lots of money, revenue based on AI is a small percentage for them.

An AI bust would take the stock price down a good deal, but the stock gains have been relatively moderate. Year on year: Microsoft +14%, Meta +24%, Google +40, Oracle +60%, ... And a notable chunk of those gains have indirectly come from the dollar devaluing.

Nvidia would be hit much harder of course.

There is a good amount of smaller AI startups, but a lot of the AI development is concentrated on the big dogs, it's not nearly as systemic as in dot com, where a lot of businesses went under completely.

And even with an AI freeze, there is plenty of value and usage there already that will not go away, but will keep expanding (AI chat, AI coding, etc) which will mitigate things.

MattGaiser•1h ago
> People keep spouting this, but I don't see how the AI bubble bursting would be all that devastating.

Well, an enormous amount of debt is being raised and issued for AI and US economic growth is nearly entirely AI. Crypto bros showed the other day that they were leveraged to the hilt on coins and it wouldn't surprise me if people are the same way on AI. It is pretty heavily tied to the financial system at this point.

techblueberry•52m ago
I think my theory into contagion would be that There’s been lots of talk about these companies starting to rack up debt, and I think AI is so tied into the US GDP that things like -

If the stock market crashes, there’s lots of talk about how wealth and debt are interlinked. Could the crash be general enough to start calls on debt backed by stocks.

My recollection in 2008 was that we didn’t learn about how bad it was until after. The tech companies have been so desperate for a win, I wonder if some of them are over their skis in some way, and if there are banks that are risking it all on AI. (We know for some tech bros think the bet on AI is a longtermist like bet; closer to religion than reason and that it’s worth risking everything because the payback could be in the hundreds of trillions)

Combine this with the fact that AI is like what - 30% of the US economy? Magnificent 7 are 60%?

What happens if sustainable PE ratios in tech collapse. Does it take out Tesla?

Maybe the contagion is just the impact on the US economy which, classically anyways has been intermingled with everything.

I would bet almost everything that there is some lie at the center of this thing that we aren’t really aware of yet.

the_duke•33m ago
It may well be that an AI bubble burst is the tipping point, but I think that tipping point was coming either way.

The US admin has been (almost desperately) trying to prop up markets and an already struggling economy. If it wasn't AI, it could have been another industry.

I think AI is more of a sideshow in this context. The bigger story is the dollar losing its dominant position , money draining out into Gold/Silver/other stock markets, India buying oil from Russia in Yen, a global economy that has for years been propped up by government spending (US/China/Europe/...), large and lasting geopolitical power balance shifts, ...

These things don't happen over night, but the effects are compounding.

Some of the above (dollar devaluation) is actually what the current admin wanted, which I would see as an admission of global shifts. We might see much larger changes to the whole financial system in the coming decades, which will have a lot of effects.

SJC_Hacker•24m ago
> Combine this with the fact that AI is like what - 30% of the US economy? Magnificent 7 are 60%?

Nowhere close. US GDP is like $30 trillion. Open AI revenue is ~$4 billion. All the other AI companies revenue might amount to $10 billion at most, and that is being generous. $10 billion/ $30 trillaion is not even 1%.

You are forgetting all those "boring" sectors that form the basis of economies like agriculture and energy. They have always been bigger than the tech sector at any point, but they aren't "sexy" because there isn't the potential "exponential growth" that tech companies

Theodores•42m ago
When America sneezes, the rest of the world catches a cold. This was said after the OG 1929 crash and I can remember it said in the 80s. Nobody says it anymore.

Due to exorbitant privilege, with the dollar as the only currency that matters, every country that trades with America is swapping goods and services for 'bits of green paper'. Unless buying oil from Russia, these bits of green paper are needed to buy oil. National currencies and the Euro might as well be casino chips, mere proxies for dollars.

Just last week the IMF issued a warning regarding AI stocks and the risk they pose to the global economy if promises are not delivered.

With every hiccup, whether that be the dot com boom, 2008 or the pandemic, the way out is to print more money, with this money going in at the top, for the banks, not the masses. This amounts to devaluation.

When the Ukraine crisis started, the Russian President stopped politely going along with Western capitalism and called the West out for printing too much money during the pandemic. Cut off from SWIFT and with many sanctions, Russia started trading in other currencies with BRICS partners. We are now at a stage of the game where the BRICS countries, of which there are many, already have a backup plan for when the next US financial catastrophe happens. They just won't use the dollar anymore. Note that currently, China doesn't want any dollars making it back to its own economy, since that would cause inflation. So they invest their dollars in Belt and Road initiatives, keeping those green bits of paper safely away from China. They don't even need exports to the USA or Europe since they have a vast home market to develop.

Note that Russia's reserve of dollars and euros was confiscated. They have nothing to lose so they aren't going to come back into the Western financial system.

Hence, you are right. A market crash won't be a global systematic crisis, it just means that Shanghai becomes the financial capital of the world, with no money printing unless it is backed up by mineral, energy or other resources that have tangible value. This won't be great for the collective West, but pretty good for the rest of the world.

the_duke•30m ago
I have similar views on many points, see my response to a sibling comment.

I just think that effects of the AI bubble bursting would be at most a symptom or trigger of much larger geopolitical and financial shifts that would happen anyway.

jcranmer•23m ago
There's a few variables which can make it much worse.

The first is how much of the capital expenditures are being fueled by debt that won't be repaid, and how much that unpaid debt harms lending institutions. This is fundamentally how a few bad debts in 2008 broke the entire financial system: bad loans felled Lehman Brothers, which caused one money market fund to break the buck, which spurred a massive exodus from the money markets rather literally overnight.

The second issue is the psychological impact of 40% of market value just evaporating. A lot of people have indirect exposure to the stock market and these stocks in particular (via 401(k)s or pensions), and seeing that much of their wealth evaporate will definitely have some repercussions on consumer confidence.

kif•1h ago
This honestly doesn’t surprise me. We have reached a point where it’s becoming clearer and clearer that AGI is nowhere to be seen, whereas advances in LLM ability to ‘reason’ have slowed down to (almost?) a halt.
dawnerd•1h ago
But if you ask an AI hype person they’ll say we’re almost there we just need a bit more gigawatts of compute!
vbezhenar•1h ago
In my book, chat-based AGI has been reached years ago, when I couldn't reliably distinguish computer from human.

Solving problems that humanity couldn't solve is super-AGI or something like that. It's not there indeed.

3836293648•1h ago
Beating the Turing Test is not AGI, but it is beating the Turing Test and that was impressive enough when it happened
jdiff•25m ago
We're not even solving problems that humanity can solve. There's been several times where I've posed to models a geometry problem that was novel but possible for me to solve on my own, but LLMs have fallen flat on executing them every time. I'm no mathematician, these are not complex problems, but they're well beyond any AI, even when guided. Instead, they're left to me, my trusty whiteboard, and a non-negligible amount of manual brute force shuffling of terms until it comes out right.

They're good at the Turing test. But that only marks them as indistinguishable from humans in casual conversation. They are fantastic at that. And a few other things, to be clear. Quick comprehension of an entire codebase for fast queries is horribly useful. But they are a long way from human-level general intelligence.

steveBK123•1h ago
Hence the pivot into ads, shop-in-chat and umm.. adult content.
Analemma_•1h ago
“AGI achieved internally”

Another case of culture flowing from the top I guess.

strangescript•1h ago
This entire thing has been pretty disingenuous on both sides of the fence. All the anti-AI (or anti OpenAI) people are doing victory laps, but what GPT-5 Pro did is still very valuable.

1) What good is your open problem set if really its a trivial "google search" away from being solved. Why are they not catching any blame here?

2) These answers still weren't perfectly laid out for the most part. GPT-5 was still doing some cognitive lifting to piece it together.

If a human would have done this by hand it would have made news and instead the narrative would have been inverted to ask serious questions about the validity of some these style problem sets and/or ask the question how many other solutions are out there that just need pieced together from pre-existing research.

But, you know, AI Bad.

nurettin•1h ago
AI great, but AI not creative, yet.
puttycat•1h ago
This is a strawman argument. No anti-AI sentiment was involved here. Simply the fact that finding and matching text on the Internet is several orders of magnitude easier than finding novel solutions to hard math problems.
matsemann•1h ago
You're moving the goal post.
Topfi•1h ago
> What good is your open problem set if really its a trivial "google search" away from being solved. Why are they not catching any blame here?

They are a community run database, not the sole arbiter and source of this information. We learned the most basic research back in highschool, I'd hope researchers from top institutions now working for one of the biggest frontier labs can do the same prior to making a claim, but microblogging has and continues to be a blight on any accurate information so nothing new there.

> GPT-5 was still doing some cognitive lifting to piece it together.

Cognitive lifting? It's a model, not a person, but besides that fact, this was already published literature. Handy that a LLM can be a slightly better search, but calling claims of "solving maths problems" out as irresponsible and inaccurate is the only right choice in this case.

> If a human would have done this by hand it would have made news [...]

"Researcher does basic literature review" isn't news in this or any other scenario. If we did a press release every journal club, there wouldn't be enough time to print a single page advert.

> [...] how many other solutions are out there that just need pieced together from pre-existing research [...]

I am not certain you actually looked into the model output or why this was such an embarrassment.

> But, you know, AI Bad.

AI hype very bad. AI anthropomorphism even worse.

andrepd•1h ago
> 1) What good is your open problem set if really its a trivial "google search" away from being solved. Why are they not catching any blame here?

Please explain how this is in any way related to the matter at hand. What is the relation between the incompleteness of an math problem database, and AI hypesters lying about the capabilities of GPT5? I fail to see the relevance.

> If a human would have done this by hand it would have made news

If someone updated information on an obscure math problem aggregator database this would be news?? Again, I fail to see your point here.

lukev•54m ago
Framing this question as "AI good" OR "AI bad" is culture-war thinking.

The real problem here is that there's clearly a strong incentive for the big labs to deceive the public (and/or themselves) about the actual scientific and technical capabilities of LLMs. As Karpathy pointed out on the recent Dwarkesh podcast, LLMs are quite terrible at novel problems, but this has become sort of an "Emperor's new clothes" situation where nobody with a financial stake will actually admit that, even though it's common knowledge if you actually work with these things.

And this directly leads to the misallocation of billions of dollars and potentially trillions in economic damage as companies align their 5-year strategies towards capabilities that are (right now) still science fiction.

The truth is at stake.

random9749832•1h ago
You are telling me a language model trained on Reddit can't solve novel problems? Shocking.

Edit: we are in peak damage control phase of the hype cycle.

phplovesong•1h ago
How fing obvious was it that AI slop did not do anything other than scarpe some websites.
Jweb_Guru•1h ago
I felt like I was going crazy when people uncritically accepted the original claim from OpenAI. Have people actually used these models?
amirhirsch•1h ago
The sad truth about this incident is that it reveals that OpenAI does not have a serious effort to actually work on unsolved math problems.
rowanG077•53m ago
How so? I wouldn't put much stock into a roque employee announcing something wrong.
mrbungie•47m ago
That's not any employee, its their VP of Science.
amirhirsch•37m ago
The people involved are very smart and must know that AI doing novel math is a canary for AGI. A serious effort around solving open problems would not fuck up this kind of announcement.
coldtea•38m ago
"rogue employee"
jebarker•51m ago
That’s a non sequitur. They’re a fairly large organization, I’d be amazed if they don’t have multiple research sub-teams pursuing all sorts of different directions.
grafmax•51m ago
I realized they jumped the shark when they announced the pivots to ads and porn. Markets haven’t caught on yet.
zeroonetwothree•45m ago
They know where the money is.
grafmax•34m ago
It’s standard practice for VC companies to enshittify after building a moat, relying on user lock-in. What’s remarkable is how quickly they’ve had to shift gears. And with this rapid pivot it’s questionable how large that moat really is.
HarHarVeryFunny•39m ago
The porn / sex-chat one is really disappointing. It seems they've given up even pretending that they are trying to do something beneficial for society. This is just a pure society-be-damned money grab.
disgruntledphd2•8m ago
They've raised far too much money for those kinda ethics, unfortunately.
goalieca•12m ago
The porn pivot makes perfect sense. Porn is already quite fake and unconvincing and none of that matters.
throwacct•7m ago
Unfortunately, the porn pivot might be their path to "profitability".
mrbombastic•1m ago
It might not matter as far as profitability is concerned, ethically the second order effects will be very problematic. I am no puritan but the widespread availability of porn has already affected peoples sexual expectations greatly. AI generated porn is going to remove even more guardrails for behavior previously considered deviant, people will view and bring those expectations back to real life.
JKCalhoun•1h ago
I try not to lose sight of the first time that I heard (some years back) that people were using this new LLM thing for DM'ing ("dungeon mastering", leading) a game of Dungeons and Dragons. I thought, this must be bullshit or some kind of witchcraft.

Definitely not anti-AI here. I think I have been disappointed though, since then, to slowly learn that they're (still) little beyond that.

Still amazing though. And better than a Google search (IMHO).

llm_nerd•59m ago
Yann LeCun's "Hoisted by their own GPTards" is fantastic.
frays•41m ago
I might be missing context here, but I'm surprised to see Yann using language that plays on 'retard.'

That seems out of character for him - more like something I'd expect from Elon Musk. What's the context I'm missing?

nova22033•38m ago
It's a play on the word petard
microtonal•31m ago
I found this background useful as a non-native speaker: https://en.wikipedia.org/wiki/Hoist_with_his_own_petard
grey-area•33m ago
Hoist (thrown in the air) by your own petard (bomb) is a common phrase.
znkr•32m ago
I don’t think it’s a wordplay with the r-word, but rather a reference to the famous Shakespeare quote: “Hoist with his own petard”. It’s become an English proverb. (A petard is a smallish bomb)
NitpickLawyer•30m ago
While Yann is clearly brilliant, and has a deeper understanding of the roots of the filed than many of us mortals, I think he's been on a debbie downer trend lately, and more importantly, some of his public stances have been proven wrong in mere months / years after he made them.

I remember a public talk, where he was on the stage with some young researcher from MS. (I think it was one of the authors of the "sparks of brilliance in gpt4" paper, but not sure).

Anyway, throughout that talk he kept talking above the guy, and didn't seem to listen, even though he obviously didn't try the "raw", "unaligned" model that the folks at MS were talking about.

And he made 2 big claims:

1) LLMs can't do math. He went on to "argue" that LLMs trick you with poetry that sounds good, but is highly subjective, and when tested on hard verifiable problems like math, they fail.

2) LLMs can't plan.

Well, merely one year later, here we are. AIME is saturated (with tool use), gold at IMO, and current agentic uses clearly can plan (and follow up with the plan, re-write parts, finish tasks, etc etc).

So, yeah, I'd take everything any one singular person says with a huge grain of salt. No matter how brilliant said individual is.

mrbungie•14m ago
Pretty sure you can fill a room with serious researchers that at the very least will doubt about 2) being solved with LLMs, especially when talking about formal planning with pure LLMs and without a planning framwork.

PS: So just we're clear: formal planning in AI </> making a coding plan in Cursor.

Palmik•45m ago
The original tweet was clearly misunderstood...

https://x.com/SebastienBubeck/status/1977181716457701775:

> gpt5-pro is superhuman at literature search:

> it just solved Erdos Problem #339 (listed as open in the official database https://erdosproblems.com/forum/thread/339) by realizing that it had actually been solved 20 years ago

https://x.com/MarkSellke/status/1979226538059931886:

> Update: Mehtaab and I pushed further on this. Using thousands of GPT5 queries, we found solutions to 10 Erdős problems that were listed as open: 223, 339, 494, 515, 621, 822, 883 (part 2/2), 903, 1043, 1079.

It's clearly talking about finding existing solutions to "open" problems.

The main mistake is by Kevin Weil, OpenAI CTO, who misunderstood the tweet:

https://x.com/kevinweil/status/1979270343941591525:

> you are totally right—I actually misunderstood @MarkSellke's original post, embarrassingly enough. Still very cool, but not the right words. Will delete this since I can't edit it any longer I think.

Obviously embarassing, but completely overblown reaction. Just another way for people to dunk on OpenAI :)

Topfi•28m ago
If holding the CTO of OpenAI accountable for his wildly inaccurate statement constitutes "dunking on OpenAI", then I'd say dunk away.

He, more than anyone else, should be able to for one parse the original statements correctly and for another maybe realize that if one of their models had accomplished what he seemed to think GPT-5 had, that may require some more scrutiny and research before posting it. That would have, after all, been a clear and incredibly massive development for the space, something the CTO of OpenAI should recognize instantly.

The amount of people that told me this is clear and indisputable proof that AGI/ASI/whatever is either around the corner or already here is far more than zero and arguing against their misunderstanding was made all the more challenging because "the CTO of OpenAI knows more than you" is quite a solid appeal to authority.

I'd recommend maybe a waiting period of 48h before any authority in any field can send a tweet, that might resolve some of the inaccuracies and the incredibly annoying need to just jump on wild bandwagons...

zozbot234•27m ago
"you are totally right—I actually misunderstood" ...like, seriously? Did an AI come up with this retraction, or are humans actually talking like robots now?
Topfi•20m ago
Guess even the CTO of OpenAI relies on Anthropic models in a pinch...
827a•40m ago
This happening the same week as DeepMind’s seemingly legitimate AI-assisted cancer treatment breakthrough is a startlingly bad look for OpenAI.

My boss always used to say “our only policy is, don’t be the reason we need to create a new policy”. I suspect OpenAI is going to have some new public communication policies going forward.

gpjt•38m ago
To be fair to the OpenAI team, if read in context the situation is at worst ambiguous.

The deleted tweet that the article is about said "GPT-5 just found solutions to 10 (!) previously unsolved Erdös problems, and made progress on 11 others. These have all been open for decades." If it had been posted stand-alone then I would certainly agree that it was misleading, but it was not.

It was a quote-tweet of this: https://x.com/MarkSellke/status/1979226538059931886?t=OigN6t..., where the author is saying he's "pushing further on this".

The "this" in question is what this second tweet is in turn quote-tweeting: https://x.com/SebastienBubeck/status/1977181716457701775?t=T... -- where the author says "gpt5-pro is superhuman at literature search: [...] it just solved Erdos Problem #339 (listed as open in the official database erdosproblems.com/forum/thread/3…) by realizing that it had actually been solved 20 years ago"

So, reading the thread in order, you get

  * SebastienBubeck: "GPT-5 is really good at literature search, it 'solved' an apparently-open problem by finding an existing solution"
  * MarkSellke: "Now it's done ten more"
  * kevinweil: "Look at this cool stuff we've done!"
I think the problem here is the way quote-tweets work -- you only see the quoted post and not anything that it in turn is quoting. Kevin Weil had the two previous quotes in his context when he did his post and didn't consider the fact that readers would only see the first level, so wouldn't have Sebastien Bubek's post in mind when they read his.

That seems like an easy mistake to entirely honestly make, and I think the pile-on is a little unfair.

card_zero•15m ago
So the first guy said "solved [...] by realizing that it had actually been solved 20 years ago", and the second guy said "found solutions to 10 (!) previously unsolved Erdös problems".

Previously unsolved. The context doesn't make that true, does it?

moefh•14m ago
> Kevin Weil had the two previous quotes in his context when he did his post and didn't consider the fact that readers would only see the first level, so wouldn't have Sebastien Bubek's post in mind when they read his.

No, Weil said he himself misunderstood Sellke's post[1].

Note Weil's wording (10 previously unsolved Erdos problems) vs. Sellke's wording (10 Erdos problems that were listed as open).

[1] https://x.com/kevinweil/status/1979270343941591525

Frieren•3m ago
> "GPT-5 is really good at literature search, it 'solved' an apparently-open problem by finding an existing solution"

Survivor bias.

I can assure you that GPT-5 fucks up even relatively easy searches. I need to have a very good idea how the results looks like and the ability to test it to be able to use any result from GPT-5.

If I throw the dice 1000 times and post about it each time that I got a double six. Am I the best dice thrower that there is?

ripped_britches•28m ago
I make mistakes all the time. This seems like a genuine mistake, not malice.

Imagine if you were talking about your own work online, you make an honest mistake, then the whole industry roasts you for it.

I’m so tired of hearing everyone take stabs at people at OpenAI just because they don’t personally like sama or something.

strongbond•21m ago
Maybe Open AI shouldn't be so stabable?
Timsky•26m ago
> GPT-5 is proving useful as a literature review assistant

No, it does not. It only produces a highly convincing counterfeit. I am honestly happy for people who are satisfied with its output: life is way easier for them than for me. Obviously, the machine discriminates me personally. When I spend hours in the library looking for some engineering-related math made in the 70s-80s, as a last resort measure, I can try to play this gambling with chat, hoping for any tiny clue to answer my question. And then for the following hours, I am trying to understand what is wrong with the chat output. Most often, I experience the "it simply can't be" feeling, and I know I am not the only one having it.

cedws•26m ago
Making such a claim should at the very least require proof that the information wasn’t in the training data.
d_tr•14m ago
Whatever happened to "don't get high on your own supply"?