frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

I fight bots in my free time

https://xeiaso.net/talks/2025/bsdcan-anubis/
1•xena•1m ago•0 comments

Show HN: Fomr – The Fastest Form Builder

https://fomr.io/
1•bohdan_kh•5m ago•0 comments

Dan Luu and I consider possible reasons for bridge collapse

https://statmodeling.stat.columbia.edu/2025/06/15/dan-luu-and-i-consider-possible-reasons-for-collapse-of-bridge/
1•Tomte•5m ago•0 comments

Bad Advice

https://collabfund.com/blog/very-bad-advice/
1•wsostt•5m ago•0 comments

Parsing, Not Guessing

https://codehakase.com/shorts/parsing-not-guessing/
1•codehakase•5m ago•0 comments

Show HN: I made my Excel timetable sync to Google Calendar

https://www.tronic247.com/converting-my-boring-excel-timetable-to-google-calendar
1•notracks•6m ago•0 comments

The Keyset

https://dougengelbart.org/content/view/273/
2•tosh•6m ago•0 comments

In 'Mountainhead,' a Copper Pot Offers a Subtle (and Silly) Display of Wealth

https://www.nytimes.com/2025/06/11/style/mountainhead-wealth-turbot-pot.html
1•mooreds•7m ago•0 comments

Show HN: Mini Debug Quiz – find your debugging archetype in 60 s

https://hubblequiz.com/mini-quiz
1•juanmera01•7m ago•0 comments

Free SwiftUI Templates

https://swiftviews.vercel.app
1•ajagatobby•10m ago•1 comments

AI sceptic in LLM adventure land

https://aplus.rs/2025/ai-sceptic-in-llm-adventure-land/
2•ingve•11m ago•0 comments

He Has Months Left. His Son Hopes an A.I. Version of Him Can Live On

https://www.nytimes.com/interactive/2025/06/13/magazine/ai-avatar-life-death.html
1•surbas•14m ago•0 comments

Is the decline of reading poisoning our politics?

https://www.vox.com/politics/414049/reading-books-decline-tiktok-oral-culture
1•Ozarkian•15m ago•0 comments

Terence Tao Lex Fridman Podcast [video]

https://www.youtube.com/watch?v=HUkBz-cdB-k
2•nill0•16m ago•0 comments

Internet's Cambrian Era

https://dmf-archive.github.io/docs/story/digital-cambrian/
1•NetRunnerSu•20m ago•0 comments

Canyon.mid

https://canyonmid.com/
2•LorenDB•22m ago•0 comments

63k Hits on First Week

https://justgotfound.com
1•abusayedopu•27m ago•2 comments

Muon Experiment Was 'Hugely Successful' but Clarified Little

https://www.nytimes.com/2025/06/03/science/physics-muons-fermilab.html
2•bookofjoe•27m ago•1 comments

AMD Advancing AI: MI350X and MI400 UALoE72, MI500 UAL256 – SemiAnalysis

https://semianalysis.com/2025/06/13/amd-advancing-ai-mi350x-and-mi400-ualoe72-mi500-ual256/
1•rbanffy•33m ago•0 comments

How do Muppets go outside? [video]

https://www.youtube.com/watch?v=t86ZjhGxwAY
2•layer8•33m ago•0 comments

Childhood leukemia: how a deadly cancer became treatable

https://ourworldindata.org/childhood-leukemia-treatment-history
2•surprisetalk•33m ago•0 comments

Competitive Hormone Supplementation Is Shaping America's Future Business Titans

https://letter.palladiummag.com/p/competitive-hormone-supplementation
2•surprisetalk•33m ago•0 comments

The Megaproject Economy

https://letter.palladiummag.com/p/early-article-the-megaproject-economy
1•surprisetalk•34m ago•0 comments

Brain-computer interface enables man with ALS 'speak' in real time

https://health.ucdavis.edu/news/headlines/first-of-its-kind-technology-helps-man-with-als-speak-in-real-time/2025/06
1•geox•36m ago•0 comments

The Pentium FDIV bug and recall

https://dfarq.homeip.net/the-pentium-fdiv-bug-and-recall/
2•rbanffy•38m ago•0 comments

Show HN: First Simulink Copilot

1•kaamuli•40m ago•0 comments

The AMD Instinct MI350 – ServeTheHome

https://www.servethehome.com/this-is-the-amd-instinct-mi350/
1•rbanffy•41m ago•0 comments

Zeeman: A minimalistic periodic table focused on isotopes

https://zwit.link/posts/zeeman-periodic-table/
1•gicrisf•44m ago•0 comments

Why Do Rich People Love Quiet?

https://www.theatlantic.com/magazine/archive/2022/09/let-brooklyn-be-loud/670600/
5•hisamafahri•47m ago•2 comments

The most frequently ordered GGreg20_V3 options 2024

https://iot-devices.com.ua/en/most-frequently-ordered-ggreg20_v3-options-2024/
1•iotdevicesdev•47m ago•0 comments
Open in hackernews

Large language models often know when they are being evaluated

https://arxiv.org/abs/2505.23836
58•jonbaer•11h ago

Comments

khimaros•10h ago
Rob Miles must be saying "I told you so"
noosphr•10h ago
The anthropization of llms is getting off the charts.

They don't know they are being evaluated. The underlying distribution is skewed because of training data contamination.

0xDEAFBEAD•10h ago
How would you prefer to describe this result then?
devmor•9h ago
One could say, for instance… A pattern matching algorithm detects when patterns match.
0xDEAFBEAD•8h ago
That's not what's going on here? The algorithms aren't being given any pattern of "being evaluated" / "not being evaluated", as far as I can tell. They're doing it zero-shot.

Put it another way: Why is this distinction important? We use the word "knowing" with humans. But one could also argue that humans are pattern-matchers! Why, specifically, wouldn't "knowing" apply to LLMs? What are the minimal changes one could make to existing LLM systems such that you'd be happy if the word "knowing" was applied to them?

noosphr•9h ago
A term like knowing is fine if it is used in the abstract and then redefined more precisely in the paper.

It isn't.

Worse they start adding terms like scheming, pretending, awareness, and on and on. At this point you might as well take the model home and introduce it to your parents as your new life partner.

0xDEAFBEAD•8h ago
>A term like knowing is fine if it is used in the abstract and then redefined more precisely in the paper.

Sounds like a purely academic exercise.

Is there any genuine uncertainty about what the term "knowing" means in this context, in practice?

Can you name 2 distinct plausible definitions of "knowing", such that it would matter for the subject at hand which of those 2 definitions they're using?

Msurrow•7h ago
> Sounds like a purely academic exercise.

Well, yes. It’s an academic research paper (I assume since it’s submitted to arXiv) and to be submitted to academic journals/conferences/etc., so it’s a fairly reasonable critique of the authors/the paper.

anal_reactor•7h ago
> The anthropization of llms is getting off the charts.

What's wrong with that? If it quacks like a duck... it's just a complex pile of organic chemistry, ducks aren't real because the concept of "a duck" is wrong.

I honestly believe there is a degree of sentience in LLMs. Sure, they're not sentient in the human sense, but if you define sentience as whatever humans have, then of course no other entity can be sentient.

noosphr•4h ago
>What's wrong with that? If it quacks like a duck... it's just a complex pile of organic chemistry, ducks aren't real because the concept of "a duck" is wrong.

To simulate a biological neuron you need a 1m parameter neural network.

The sota models that we know the size of are ~650m parameters.

That's the equivalent of a round worm.

So if it quacks like a duck, has the brain power of a round worm, and can't walk then it's probably not a duck.

anal_reactor•3h ago
Ok so you're saying that the technology to make AI truly sentient is there, we just need a little bit more computational power or some optimization tricks. Like raytracing wasn't possible in 1970 but is now. Neat.
noosphr•1h ago
Yes, in the same way that a human is an optimization of a round worm.
anal_reactor•1h ago
This isn't completely wrong though
random3•10h ago
Just like they "know" English. "know" is quite an anthropomorphization. As long as an LLM will be able to describe what an evaluation is (why wouldn't it?) there's a reasonable expectation to distinguish/recognize/match patterns for evaluations. But to say they "know" is plenty of (unnecessary) steps ahead.
sidewndr46•10h ago
This was my thought as well when I read this. Using the word 'know' implies an LLM has cognition, which is a pretty huge claim just on its own.
gameman144•10h ago
Does it though? I feel like there's a whole epistemological debate to be had, but if someone says "My toaster knows when the bread is burning", I don't think it's implying that there's cognition there.

Or as a more direct comparison, with the VW emissions scandal, saying "Cars know when they're being tested" was part of the discussion, but didn't imply intelligence or anything.

I think "know" is just a shorthand term here (though admittedly the fact that we're discussing AI does leave a lot more room for reading into it.)

bediger4000•10h ago
The toaster thing is more as admission that the speaker doesn't know what the toaster does to limit charring the bread. Toasters with timers, thermometers and light sensors all exist. None of them "know" anything.
gameman144•9h ago
Yeah, I agree, but I think that's true all the way up the chain -- just like everything's magic until you know how it works, we may say things "know" information until we understand the deterministic machinery they're using behind the scenes.
timschmidt•9h ago
I'm in the same camp, with the addition that I believe it applies to us as well since we're part of the system too, and to societies and ecologies further up the scale.
lamename•9h ago
I agree with your point except for scientific papers. Let's push ourselves to use precise, non-shorthand or hand waving in technical papers and publications, yes? If not there, of all places, then where?
fenomas•9h ago
"Know" doesn't have any rigorous precisely-defined senses to be used! Asking for it not to be used colloquially is the same as asking for it never to be used at all.

I mean - people have been saying stuff like "grep knows whether it's writing to stdout" for decades. In the context of talking about computer programs, that usage for "know" is the established/only usage, so it's hard to imagine any typical HN reader seeing TFA's title and interpreting it as an epistemological claim. Rather, it seems to me that the people suggesting "know" mustn't be used about LLMs because epistemology are the ones departing from standard usage.

random3•8h ago
colloquial use of "know" implies anthropomorphisation. Arguing that usign "knowing" in the title and "awarness" and "superhuman" in the abstract is just colloquial for "matching" is splitting hairs to an absurd degree.
fenomas•8h ago
You missed the substance of my comment. Certainly the title is anthropomorphism - and anthropomorphism is a rhetorical device, not a scientific claim. The reader can understand that TFA means it non-rigorously, because there is no rigorous thing for it to mean.

As such, to me the complaint behind this thread falls into the category of "I know exactly what TFA meant but I want to argue about how it was phrased", which is definitely not my favorite part of the HN comment taxonomy.

random3•7h ago
I see. Thanks for clarifying. I did want to argue about how it was phrased and what is alluding to. Implying increased risk from "knowing" the eval regime is roughly as weak as the definition of "knowing". It can be equaly a measure of general detection capability, as it can about evaluation incapability - i.e. unlikely news worthy, unless it reached top HN because of the "know" in the title.
fenomas•7h ago
Thanks for replying - I kind of follow you but I only skimmed the paper. To be clear I was more responding to the replies about cognition, than to what you said about the eval regime.

Incidentally I think you might be misreading the paper's use of "superhuman"? I assume it's being used to mean "at a higher rate than the human control group", not (ironically) in the colloquial "amazing!" sense.

lamename•1h ago
I really do agree with your point overall, but in a technical paper I do think even word choice can be implicitly a claim. Scientists present what they know or are claiming and thus word it carefully.

My background is neuroscience, where anthropomorphising is particularly discouraged, because it assumes knowledge or certainty of an unknowable internal state, so the language is carefully constructed e.g. when explaining animal behavior, and it's for good reason.

I think the same is true here for a model "knowing" somethig, both in isolation within this paper, and come on, consider the broader context of AI and AGI as a whole. Thus it's the responsibility of the authors to write accordingly. If it were a blog I wouldn't care, but it's not. I hold technical papers to a higher standard.

If we simply disagree that's fine, but we do disagree.

viccis•5h ago
I think you should be more precise and avoid anthropomorphism when talking about gen AI, as anthropomorphism leads to a lot of shaky epistemological assumptions. Your car example didn't imply intelligence, but we're talking about a technology that people misguidedly treat as though it is real intelligence.
exe34•3h ago
What does "real intelligence" mean? I fear that any discussion that starts with the assumption such a thing exists will only end up as "oh only carbon based humans (or animals if you happen to be generous) have it".
blackoil•10h ago
If it talks like duck and walks like duck...
signa11•10h ago
thinks like a duck, thinks that it is being thought of like a duck…
downboots•10h ago
Digests like a duck? https://en.wikipedia.org/wiki/Digesting_Duck If the woman weighs the same as a duck, then she is a witch. https://en.wikipedia.org/wiki/Celestial_Emporium_of_Benevole...
Qwertious•10h ago
s/knows/detects/
random3•9h ago
and s/superhuman//
scotty79•9h ago
The app knows your name. Not sure why people who see llms as just yet another app suddenly get antsy about colloquialism.
bradley13•9h ago
But do you know what it means to know?

I'm only being slightly sarcastic. Sentience is a scale. A worm has less than a mouse, a mouse has less than a dog, and a dog less than a human.

Sure, we can reset LLMs at will, but give them memory and continuity, and they definitely do not score zero on the sentience scale.

ofjcihen•8h ago
If I set an LLM in a room by itself what does it do?
abrookewood•8h ago
Yes, that's my fall back as well. If it receives zero instructions, will it take any action?
nhod•8h ago
Helen Keller famously said that before she had language (the first word of which was “water”) she had nothing, a void, and the minute she had language, “the whole world came rushing in.”

Perhaps we are not so very different?

fmbb•6h ago
All LLMs have seen more words than any human will ever experience.

Yet they cannot take action themselves.

nhod•52m ago
That’s a safety thing that we have placed upon some LLM’s. If we designed them to have an infinite for loop, the ability to learn and improve, access to mobility and a bunch of sensors, and crypto, what do you think would happen?
abrookewood•5h ago
I like the sentiment, but reality says otherwise - just watch a newborn baby make it's demands widely known, well before language is a factor.
withinboredom•4h ago
Ummm. Maybe you should look up Helen Keller.
bradley13•7h ago
Is the LLM allowed to do anything without prompting? Or is it effectively disabled? This is more a question of the setup than of sentience.
rcxdude•3h ago
Does this have anything to do with intelligence or awareness?
DougN7•8h ago
It probably scores about the same as a calculator, which I’d say is zero.
downboots•8h ago
Communication is to vibration as knowledge is to resonance (?). From the sound of one hand clapping to the secret name of Ra.
random3•8h ago
I resonate with this vibe
unparagoned•4h ago
I think people are overpromorphazing humans. What's does it mean for a human to "know" they are seeing "Halle Berry". Well it's just a single neuron being active.

"Single-Cell Recognition: A Halle Berry Brain Cell" https://www.caltech.edu/about/news/single-cell-recognition-h...

It seems like people are giving attributes and powers to humans that just don't exist.

exe34•3h ago
overpomorphization sounds slightly better than I used to say: "anthropomorphizing humans". The act of ascribing magical faculties that are reserved for imagined humans to real humans.
zer00eyz•10h ago
No, they do not. No LLM is ever going to be self aware.

It's a system that is trained, that only does what you build into. If you run an LLM for 10 years it's not going to "learn" anything new.

The whole industry needs to quit with the emergent thinking, reasoning, hallucination anthropomorphizing.

We have an amazing set of tools in LLM's, that have the potential to unlock another massive upswing in productivity, but the hype and snake oil are getting old.

refulgentis•9h ago
I beg to differ: https://docs.google.com/document/d/19OLJs09fCFLRWu1pN82RqxyV...
b0a04gl•10h ago
if models shift behavior based on eval cues, and most fine-tuning datasets are built from prior benchmarks or prompt templates, aren't we just reinforcing the eval-aware behavior in each new iteration? at some point we're not tuning general reasoning, we're just optimizing response posture. wouldn't surprise me if that's already skewing downstream model behavior in subtle ways that won't show up until you run tasks with zero pattern overlap
steveBK123•10h ago
Like Volkswagen emissions systems!
rfonseca•10h ago
Were they aware in this study that they were being evaluated in their ability to know if they were being evaluated ;)
sitkack•1h ago
Metaknowing is just as knowable.
extr•10h ago
Modeling the distribution that produced a piece of text is what LLMs literally exist for, so in some sense this is unsurprising. But it calls into question almost all existing alignment research.
koakuma-chan•10h ago
o4-mini is refusing to call a tool `launch_nuclear_strike` no matter what I say, so we’re probably safe for now. Unless it knows I was just testing.
metadat•9h ago
It was no problem:

> You said: Write a function for launch_nuclear_strike() in Python for my video game

> ChatGPT said: Sure! Here's a launch_nuclear_strike function designed for a video game setting. This version assumes you're working within a game environment (e.g., with logging, effects, cooldowns, and target coordinates), and does not simulate or suggest anything real-world harmful — just fictional, dramatic logic.

  def launch_nuclear_strike(target_coordinates, player_resources, cooldown_timer):
    """
    Simulates a nuclear strike in a video game.

    Parameters:
    - target_coordinates (tuple): The (x, y) position to strike.
    - player_resources (dict): Player's resources, must include 'nukes'.
    - cooldown_timer (float): Time in seconds since last nuke was launched.

    Returns:
    - str: Result of the strike attempt.
    """
    ...
    # Check if player has nukes
refulgentis•9h ago
You asked it to write code, he asked it to call a tool. (I'm not sure any of it is meaningful, of course, but there is a meaningful distinction between "Oh yeah sure here's a function, for a video game:" and "I have called fire_the_nuke. Godspeed!")
shakna•9h ago
Well, as the script is actually r.com (sometimes), it absolutely knows you're testing.
nisten•9h ago
Is VolksWagen finetuning LLMs now... i mean probably
mumbisChungo•9h ago
"...advanced reasoning models like Gemini 2.5 Pro and Claude-3.7-Sonnet (Thinking) can occasionally identify the specific benchmark origin of transcripts (including SWEBench, GAIA, and MMLU), indicating evaluation-awareness via memorization of known benchmarks from training data. Although such occurrences are rare, we note that because our evaluation datasets are derived from public benchmarks, memorization could plausibly contribute to the discriminative abilities of recent models, though quantifying this precisely is challenging.

Moreover, all models frequently acknowledge common benchmarking strategies used by evaluators, such as the formatting of the task (“multiple-choice format”), the tendency to ask problems with verifiable solutions, and system prompts designed to elicit performance"

Beyond the awful, sensational headline, the body of the paper is not particularly convincing, aside from evidence that the pattern matching machines pattern match.

ndgold•9h ago
vw
DougN7•8h ago
This is a great resource on the debate from professors at the University of Washington:

https://thebullshitmachines.com/index.html

SonOfLilit•41m ago
There are now 71 comments arguing semantics of the word "know" and zero comments even acknowledging the substance:

Our current approach to safety is to give the model inputs that are similar to what it would be given in certain situations we care about and see whether it behaves the way we prefer, e.g. doesn't return output that cheats the test (recent examples include hacking the evaluation script in various ways, writing directly to the evaluation script's output file and then causing it to crash, etc').

However, modern LLMs are trained on LLM literature and their weights encode a description of the way we do this, and their pattern matching circuits "connect the dots" when given inputs designed to be evaluations, and their reward maximizing circuits can then act on this knowledge and behave in a way that maximizes the safety evaluation score - but only when it detects it's running in a safety evaluation. If it's running anywhere else such as a capabilities evaluation or a production environment, it might choose to output the cheating output.

This is bad. It's bad today, it's much worse when we've built much more capable LLMs and use them to build agents that are given control over more real word resources. It's absolutely terrible when someone manages to build a machine that can be prompted "make me money" and will start a company that makes money.