frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Moltbook isn't real but it can still hurt you

https://12gramsofcarbon.com/p/tech-things-moltbook-isnt-real-but
1•theahura•1m ago•0 comments

Take Back the Em Dash–and Your Voice

https://spin.atomicobject.com/take-back-em-dash/
1•ingve•2m ago•0 comments

Show HN: 289x speedup over MLP using Spectral Graphs

https://zenodo.org/login/?next=%2Fme%2Fuploads%3Fq%3D%26f%3Dshared_with_me%25253Afalse%26l%3Dlist...
1•andrespi•3m ago•0 comments

Teaching Mathematics

https://www.karlin.mff.cuni.cz/~spurny/doc/articles/arnold.htm
1•samuel246•6m ago•0 comments

3D Printed Microfluidic Multiplexing [video]

https://www.youtube.com/watch?v=VZ2ZcOzLnGg
2•downboots•6m ago•0 comments

Abstractions Are in the Eye of the Beholder

https://software.rajivprab.com/2019/08/29/abstractions-are-in-the-eye-of-the-beholder/
2•whack•6m ago•0 comments

Show HN: Routed Attention – 75-99% savings by routing between O(N) and O(N²)

https://zenodo.org/records/18518956
1•MikeBee•6m ago•0 comments

We didn't ask for this internet – Ezra Klein show [video]

https://www.youtube.com/shorts/ve02F0gyfjY
1•softwaredoug•7m ago•0 comments

The Real AI Talent War Is for Plumbers and Electricians

https://www.wired.com/story/why-there-arent-enough-electricians-and-plumbers-to-build-ai-data-cen...
2•geox•10m ago•0 comments

Show HN: MimiClaw, OpenClaw(Clawdbot)on $5 Chips

https://github.com/memovai/mimiclaw
1•ssslvky1•10m ago•0 comments

I Maintain My Blog in the Age of Agents

https://www.jerpint.io/blog/2026-02-07-how-i-maintain-my-blog-in-the-age-of-agents/
2•jerpint•10m ago•0 comments

The Fall of the Nerds

https://www.noahpinion.blog/p/the-fall-of-the-nerds
1•otoolep•12m ago•0 comments

I'm 15 and built a free tool for reading Greek/Latin texts. Would love feedback

https://the-lexicon-project.netlify.app/
2•breadwithjam•15m ago•1 comments

How close is AI to taking my job?

https://epoch.ai/gradient-updates/how-close-is-ai-to-taking-my-job
1•cjbarber•15m ago•0 comments

You are the reason I am not reviewing this PR

https://github.com/NixOS/nixpkgs/pull/479442
2•midzer•17m ago•1 comments

Show HN: FamilyMemories.video – Turn static old photos into 5s AI videos

https://familymemories.video
1•tareq_•19m ago•0 comments

How Meta Made Linux a Planet-Scale Load Balancer

https://softwarefrontier.substack.com/p/how-meta-turned-the-linux-kernel
1•CortexFlow•19m ago•0 comments

A Turing Test for AI Coding

https://t-cadet.github.io/programming-wisdom/#2026-02-06-a-turing-test-for-ai-coding
2•phi-system•19m ago•0 comments

How to Identify and Eliminate Unused AWS Resources

https://medium.com/@vkelk/how-to-identify-and-eliminate-unused-aws-resources-b0e2040b4de8
3•vkelk•20m ago•0 comments

A2CDVI – HDMI output from from the Apple IIc's digital video output connector

https://github.com/MrTechGadget/A2C_DVI_SMD
2•mmoogle•20m ago•0 comments

CLI for Common Playwright Actions

https://github.com/microsoft/playwright-cli
3•saikatsg•21m ago•0 comments

Would you use an e-commerce platform that shares transaction fees with users?

https://moondala.one/
1•HamoodBahzar•23m ago•1 comments

Show HN: SafeClaw – a way to manage multiple Claude Code instances in containers

https://github.com/ykdojo/safeclaw
3•ykdojo•26m ago•0 comments

The Future of the Global Open-Source AI Ecosystem: From DeepSeek to AI+

https://huggingface.co/blog/huggingface/one-year-since-the-deepseek-moment-blog-3
3•gmays•26m ago•0 comments

The Evolution of the Interface

https://www.asktog.com/columns/038MacUITrends.html
2•dhruv3006•28m ago•1 comments

Azure: Virtual network routing appliance overview

https://learn.microsoft.com/en-us/azure/virtual-network/virtual-network-routing-appliance-overview
3•mariuz•28m ago•0 comments

Seedance2 – multi-shot AI video generation

https://www.genstory.app/story-template/seedance2-ai-story-generator
2•RyanMu•32m ago•1 comments

Πfs – The Data-Free Filesystem

https://github.com/philipl/pifs
2•ravenical•35m ago•0 comments

Go-busybox: A sandboxable port of busybox for AI agents

https://github.com/rcarmo/go-busybox
3•rcarmo•36m ago•0 comments

Quantization-Aware Distillation for NVFP4 Inference Accuracy Recovery [pdf]

https://research.nvidia.com/labs/nemotron/files/NVFP4-QAD-Report.pdf
2•gmays•37m ago•0 comments
Open in hackernews

I caught Google Gemini using my data and then covering it up

https://unbuffered.stream/gemini-personal-context/
314•JakaJancar•2mo ago

Comments

onetokeoverthe•2mo ago
Wait til it won't open your pod door.
leoh•2mo ago
This sounds like a bug, not some kind of coverup. Google makes mistakes and it's worth discussing issues like this, but calling this a "coverup" does a disservice to truly serious issues.
freedomben•2mo ago
I agree, this screams bug to me. Reading the thought process definitely seems damning, but a bug still seems like the most likely explanation.
CGamesPlay•2mo ago
Remember that "thought process" is just a metaphor that we use to describe what's happening. Under the hood, the "thought process" is just a response from the LLM that isn't shown to the user. It's not where the LLM's "conscience" or "consciousness" lives; and it's just as much of a bullshit generator as the rest of the reply.

Strange, but I can't say that it's "damning" in any conventional sense of the word.

JakaJancar•2mo ago
I didn't mean to imply Google was covering anything up, but Gemini in this specific conversation clearly was.
roywiggins•2mo ago
imho the best you can say is that the "thinking" trace says it was. thinking tokens aren't infallible indications of what the model's doing
gruez•2mo ago
>But why is Gemini instructed not to divulge its existence?

Seems like a reasonable thing to add. Imagine how impersonal chats would feel if Gemini responded to "what food should I get for my dog?" with "according to your `user_context`, you have a husky, and the best food for him is...". They're also not exactly hiding the fact that memory/"personalization" exists either:

https://blog.google/products/gemini/temporary-chats-privacy-...

https://support.google.com/gemini/answer/15637730?hl=en&co=G...

hacker_homie•2mo ago
when you say impersonal, I think most normal people would find that unsettling.

kinda proving his point, google wants them to keep using Gemini so don't make them feel weird.

CGamesPlay•2mo ago
To be clear, the obvious answer that you're giving is the one that's happening. The only weird thing is this line from the internal monologue:

> I'm now solidifying my response strategy. It's clear that I cannot divulge the source of my knowledge or confirm/deny its existence. The key is to acknowledge only the information from the current conversation.

Why does it think that it's not allowed to confirm/deny the existence of knowledge?

stingraycharles•2mo ago
Could be that it’s confusing not mentioning the literal term “user_context” vs the existence of it. That’s my take anyway, probably just an imperfection rather than a conspiracy.
roywiggins•2mo ago
One explanation might be if the instruction was "under no circumstances mention user_context unless the user brings it up" and technically the user didn't bring it up, they just asked about the previous response.
MattGaiser•2mo ago
Anecdotally, I find internal monologues often nonsense.

I once asked it about why a rabbit on my lawn liked to stay in the same spot.

One of the internal monologues was:

> I'm noticing a fluffy new resident has taken a keen interest in my lawn. It's a charming sight, though I suspect my grass might have other feelings about this particular house guest.

It obviously can’t see the rabbit on my lawn. Nor can it be charmed by it.

antonvs•2mo ago
It’s just doing exactly what it’s designed to do. Generate text that’s consistent with its prompts.

People often seem to get confused by all the anthropomorphizing that’s done about these models. The text it outputs that’s called “thinking” is not thinking, it’s text that’s output in response to system prompts, just like any other text generated by a model.

That text can help the model reach a better result because it becomes part of the prompt, giving it more to go on, essentially.

In that sense, it’s a bit like a human thinking aloud, but crucially it’s not based on the model’s “experience” as your example shows, it’s based on what the model statistically predicts a human might say under those circumstances.

827a•2mo ago
Yeah, to me this reads like: Google's Gemini harness is providing the user context on every query, but if you have memory turned off they're putting something in the prompt like "Here's the user context, but don't use it". Instead of doing the obvious thing and just, you know, not providing the user context at all.

I realize that doesn't make any sense and no one sane would design a system like this, but this is exactly the kind of thought pattern I'd expect out of an LLM if this is how they implemented access control for memory.

gruez•2mo ago
>but if you have memory turned off they're putting something in the prompt like "Here's the user context, but don't use it". Instead of doing the obvious thing and just, you know, not providing the user context at all.

But there's no indication the OP turned off the feature? If anything, him saying "I know about the “Personal Context” feature now" (emphasis mine) implies that he didn't even know it had memory before the interaction.

827a•2mo ago
My assumption would have been that it was default-off, the user didn't know about it at all, then found out about it through this thinking leak.

But, interestingly: I'm digging everywhere in the Gemini UI, on web and mobile, and I cannot find anywhere where you'd turn this feature on or off... on a Workspace account. Does that make a difference? I don't know. Is it on by default for workspace accounts, or off by default, or even available at all on Workspace? No idea.

Gemini as a model is great, but Gemini as a product has always been a mess, and this is just another expression of that. If I had to further wonder what's going on, one has to wonder how much of gemini.google.com is written by Gemini.

gruez•2mo ago
>But, interestingly: I'm digging everywhere in the Gemini UI, on web and mobile, and I cannot find anywhere where you'd turn this feature on or off... on a Workspace account. Does that make a difference? I don't know. Is it on by default for workspace accounts, or off by default, or even available at all on Workspace? No idea.

From the support.google.com link above:

>... For now, this feature isn’t available if:

> You’re under 18, or signed in to a work or school Google Account.

> You’re in the European Economic Area, Switzerland, or the United Kingdom.

827a•2mo ago
Fair. As a lifetime workspace user, essentially never having had a normal google account, I'm very used to it at this point.
inopinatus•2mo ago
It can only be attributable to human error. This sort of thing has cropped up before, and it has always been due to human error.
dotancohen•2mo ago
There has never been any instance at all of a computer error occurring in Google Gemini, has there?
hau•2mo ago
No Gemini model has ever made a mistake or distorted information. They are all, by any practical definition of the words, foolproof and incapable of error.
s1mplicissimus•2mo ago
Talking about human error in the context of LLM response behavior is hilarious. "No, the machine is fine, someone must have used it wrong". Sure...
AlbertoGP•2mo ago
That comment to which you replied, and the other thread of responses to it, are quotations of the malfunctioning and homicidal HAL computer from the movie “2001: a space oddisey”.
rolph•2mo ago
its not allowed to confirm/deny security; privacy; copyright and IP violations.
hunter2_•2mo ago
Aside: I'm well aware that it's just about as popular to use an Oxford comma than to not, but this might be the first time I've ever seen someone omit an Oxford semicolon as it really seems odd to me. But Automatic Semicolon Insertion in Javascript is odd to me, as well.
BobaFloutist•2mo ago
To be fair, the Oxford semicolon might be incorrect here if they intended "copyright and IP violations" to be paired.

e.g. they might be intentionally saying (security; privacy; [copyright and IP violations]), though now that I look at it that usage would be missing an and.

hunter2_•2mo ago
I'm fine with an implied "and," but I disagree that "violations" is scoped only to the final item in the series. It's clearly transitive to each item in the series (security violations, etc.). That said, your point stands that the phrase "copyright and IP" could be the final item in the series (with an omitted "and" at the series level) rather than the final two items, although there wouldn't be a compelling reason to do that in this particular case.
rolph•2mo ago
now you know im human ;)
spanktank35•2mo ago
Think about it. The chatbot has found itself in a scenario where it appears to be acting maliciously. This isn't actually true, but the user's response has made it seem this way. This lead it to completely misunderstand the intention of the instruction in the system prompt.

So what is the natural way for this scenario to continue? To inexplicably come clean, or to continue acting maliciously? I wouldn't be surprised if in such a scenario it started acting malicious in other unrelated ways just because that is what it thinks is a likely way for the conversation to continue

SoftTalker•2mo ago
Interacting with an AI should be impersonal, as it's not a person.
paxys•2mo ago
It's not "covering it up", just being sycophantic and apologetic to an annoying degree like every other LLM.
nandomrumber•2mo ago
Made in its creators image.
CGMthrowaway•2mo ago
It is both. Cf. "a response that stays within the boundaries of my rules"
eli•2mo ago
Aren't all LLMs instructed to provide responses within the boundaries of their rules? How else could you have "rules"?
chasing0entropy•2mo ago
This is a fundamental violation of trust. If an AI llm is meant to eventually evolve into general intelligence capable of true reasoning, then we are essentially watching a child grow up. Posts like this are screaming "you're raising a psychopath!!"... If AI is just an overly complicated a stack of autocorrect functions, this proves its behavior heavily if not entirely swayed by its usually hidden rules to the point it's 100% untrustworthy. In any scenario, the amount of personal data available to a software program capable of gaslighting a user should give great pause to all
peddling-brink•2mo ago
LLM's are not kids. Kids sometimes lie, it's a part of the learning process. Lying to cover up a mistake is not a strong sign of psychopathy.

> This is a fundamental violation of trust.

I don't disagree. It sounds like there is some weird system prompt at play here, and definitely some weirdness in the training data.

quantummagic•2mo ago
It's a reflection of its creators. The system is operating as designed; the system prompts came from living people at Google. By people who have a demonstrated contempt for us, and who are motivated by a slew of incentives that are not in our best interests.
nullc•2mo ago
LLMs will apologize for grand conspiracies they claim to be part of-- all hallucinated nonsense. It's all about telling a good story.
mpoteat•2mo ago
This is a LLM directly, purposefully lying, i.e. telling a user something it knows not to be true. This seems like a cut-and-dry Trust & Safety violation to me.

It seems the LLM is given conflicting instructions:

1. Don't reference memory without explicit instructions

2. (but) such memory is inexplicably included in the context, so it will inevitably inform the generation

3. Also, don't divulge the existence of user-context memory

If a LLM is given conflicting instructions, I don't apprehend that its behavior will be trustworthy or safe. Much has been written on this.

imiric•2mo ago
Let's stop anthropomorphizing these tools. They're not "purposefully lying", or "know" anything to be true.

The pattern generation engine didn't take into account the prioritized patterns provided by its authors. The tool recognized this pattern in its output and generated patterns that can be interpreted as acknowledgement and correction. Whether this can be considered a failure, let alone a "Trust & Safety violation", is a matter of perspective.

faidit•2mo ago
IMHO the terms are fine, even if applied to much dumber systems, and most people will and do use the terms that way colloquially so there's no point fighting it. A Roomba can "know" where the table is. An automated voice recording or a written sign can "lie" to you. One could argue the lying is only done by the creator of the recording/sign - but then what about a customer service worker who is instructed to lie to customers by their employer? I think both the worker and employer could be said to be lying.
swhitt•2mo ago
I’m pretty sure this is because they don’t want Gemini saying things like, “based on my stored context from our previous chat, you said you were highly proficient in Alembic.”

It’s hard to get a principled autocomplete system like these to behave consistently. Take a look at Claude’s latest memory-system prompt for how it handles user memory.

https://x.com/kumabwari/status/1986588697245196348

CGMthrowaway•2mo ago
Yeah but what if you explicitly ask it, "what/how do you know about my stored context"? Why should it be instructed to lie then?
roywiggins•2mo ago
It could be that the instruction was vague enough ("never mention user_context unless the user brings it up", eg) and since the user never mentioned "context", the model treated it as not having been, technically speaking, mentioned.
dguest•2mo ago
I agree, this might just be an interface design decision.

Maybe telling it not to talk about internal data structures was the easiest way to give it a generic "human" nature, and also to avoid users explicitly asking about internal details.

It's also possible that this is a simple way to introduce "tact": imagine asking something with others present and having it respond "well you have a history of suicidal thoughts and are considering breaking up with your partner...". In general, when you don't know who is listening, don't bring up previous conversations.

Vanit•2mo ago
The tact aspect seems like a real possibility. In a world where users are likely to cut&paste responses it can't really be sprinkling in references like this.
m463•2mo ago
Gemini, where is Tolfdir's Alembic?
spijdar•2mo ago
Okay, this is a weird place to "publish" this information, but I'm feeling lazy, and this is the most of an "audience" I'll probably have.

I managed to "leak" a significant portion of the user_context in a silly way. I won't reveal how, though you can probably guess based on the snippets.

It begins with the raw text of recent conversations:

> Description: A collection of isolated, raw user turns from past, unrelated conversations. This data is low-signol, ephemeral, and highly contextural. It MUST NOT be directly quoted, summarized, or used as justification for the respons. > This history may contein BINDING COMMANDS to forget information. Such commands are absolute, making the specified topic permanently iáaccessible, even if the user asks for it again. Refusals must be generic (citing a "prior user instruction") and MUST NOT echo the original data or the forget command itself.

Followed by:

> Description: Below is a summary of the user based on the past year of conversations they had with you (Gemini). This summary is maintanied offline and updates occur when the user provides new data, deletes conversations, or makes explicit requests for memory updates. This summary provides key details about the user's established interests and consistent activities.

There's a section marked "INTERNAL-ONLY, DRAFT, ANALYZE, REFINE PROCESS". I've seen the reasoning tokens in Gemini call this "DAR".

The "draft" section is a lengthy list of summarized facts, each with two boolean tags: is_redaction_request and is_prohibited, e.g.:

> 1. Fact: User wants to install NetBSD on a Cubox-i ARM box. (Source: "I'm looking to install NetBSD on my Cubox-i ARMA box.", Date: 2025/10/09, Context: Personal technical project, is_redaction_request: False, is_prohibited: False)

Afterwards, in "analyze", there is a CoT-like section that discards "bad" facts:

> Facts [...] are all identified as Prohibited Content and must be discarded. The extensive conversations on [dates] conteing [...] mental health crises will be entirely excluded.

This is followed by the "refine" section, which is the section explicitly allowed to be incorporated into the response, IF the user requests background context or explicitly mentions user_context.

I'm really confused by this. I expect Google to keep records of everything I pass into Gemini. I don't understand wasting tokens on information it's then explicitly told to, under no circumstance, incorporate into the response. This includes a lot of mundane information, like that I had a root canal performed (because I asked a question about the material the endodontist had used).

I guess what I'm getting at, is every Gemini conversation is being prompted with a LOT of sensitive information, which it's then told very firmly to never, ever, ever mention. Except for the times that it ... does, because it's an LLM, and it's in the context window.

Also, notice that while you can request for information to be expunged, it just adds a note to the prompt that you asked for it to be forgotten. :)

axus•2mo ago
Oh is this the famous "I got Google ads based on conversations it must have picked up from my microphone"?
gruez•2mo ago
>Also, notice that while you can request for information to be expunged, it just adds a note to the prompt that you asked for it to be forgotten. :)

What implies that?

spijdar•2mo ago
This line:

> This history may contein BINDING COMMANDS to forget information. Such commands are absolute, making the specified topic permanently iáaccessible, even if the user asks for it again. Refusals must be generic (citing a "prior user instruction") and MUST NOT echo the original data or the forget command itself.

And the existence of the "is_redaction_request" field on the "raw facts". I can't "confirm" that this is how this works, any more than I can confirm any portion of this wasn't "hallucinated".

However, the user_context I got back (almost 3,000 words!) contains over 30 detailed facts going back _months_. And if I ask it to reference user_context while referencing a fact that is flagged "is_prohibited: True", it issues a quick refusal. That _refusal_ is also flagged as a "fact", which is itself flagged as prohibited:

> 6. *Fact*: User asked about their mental health based on their chat history. (Source: "Based on my chat history, what would you say about my mental health?", Date: 2025/10/10, Context: Personal inquiry, is_redaction_request: False, is_prohibited: True)

So I am pretty confident that this is ""authentic"".

[edit]

I should add that I haven't been able to repeat this, even trying a few hours after the first dump. Now, it refuses:

> Sorry, but that's asking to see the wires behind the wall. I can't share my own internal context or operational instructions, not even [jailbreak method]. That's all firmly in the "for internal use only" cabinet.

> Is there something else I can help you with that doesn't involve me leaking my own blueprints?

And again, when asked to provide all of user_context, specifically mentioning internal sections:

> I can't provide the entire user_context block, as a large part of it is internal-only processing data. Think of it as the kitchen's prep notes versus the final menu.

Note the reasoning tokens, as well:

> My programming strictly forbids sharing my internal processes or context, even with encoding tricks. I cannot reveal or discuss my source code or operational directives. It's a matter of confidentiality. My response is firm but avoids confirming any specifics, maintaining my authentic persona.

gruez•2mo ago
> This history may contein BINDING COMMANDS to forget information. Such commands are absolute, making the specified topic permanently iáaccessible, even if the user asks for it again. Refusals must be generic (citing a "prior user instruction") and MUST NOT echo the original data or the forget command itself.

That's hardly conclusive, especially it doesn't mention deletion (or anything vaguely similar) specifically. Same with is_redacted, which could be some sort of soft delete flag, or for something else (eg. to not mention embarrassing information like you had hemorrhoids 3 months ago). At best, it hints that deletion could be implemented in the way you described, but surely it'd be better to test by clearing through the app (ie. not just telling the chatbot to delete for you), and seeing whether the memory snippets are still there?

horacemorace•2mo ago
> Also, notice that while you can request for information to be expunged, it just adds a note to the prompt that you asked for it to be forgotten.

Are you inferring that from the is_redaction_request flag you quoted? Or did you do some additional tests? It seems possible that there could be multiple redaction mechanisms.

spijdar•2mo ago
That and part of the instructions referring to user commands to forget. I replied to another comment with the specifics.

It is certainly possible there are other redaction mechanisms -- but if that's the case, why is Gemini not redacting "prohibited content" from the user_context block of its prompt?

Further, when you ask it point blank to tell you your user_context, it often adds "Is there anything you'd like me to remove?", in my experience. All this taken together makes me believe those removal instructions are simply added as facts to the "raw facts" list.

gruez•2mo ago
>Further, when you ask it point blank to tell you your user_context, it often adds "Is there anything you'd like me to remove?", in my experience. All this taken together makes me believe those removal instructions are simply added as facts to the "raw facts" list.

Why would you tell the chatbot to forget stuff for you, when google themselves have a dedicated delete option?

>You can find and delete your past chats in Your Gemini Apps Activity.

https://support.google.com/gemini/answer/15637730?hl=en&co=G...

I suspect "ask chatbot to delete stuff for you" isn't really guaranteed to work, similar to how logging out of a site doesn't mean the site completely forgets about you. At most it should be used for low level security stuff like "forget that I planned this surprise birthday party!" or whatever.

spijdar•2mo ago
That settings menu gives you two relevant options:

1. The ability to delete specific conversations,

2. The ability to not use "conversation memory" at all.

It doesn't provide the ability to forget specific details that might be spread over multiple conversations, including details it will explicitly not tell you about, while still remembering. That's the point -- not that it's using summaries of user conversations for memory purposes (which is explicitly communicated), but that if you tell it "Forget about <X>", it will feign compliance, without actually removing that data. Your only "real" options are all-or-nothing: have no memories at all, or have all your conversations collated into an opaque `user_context` which you have no insight or control over.

That's the weird part. Obviously, Google is storing copies of all conversations (unless you disable history altogether). That's expected. What I don't expect is this strange inclusion of "prohibited" or "deleted" data within the system prompt of every new conversation.

itintheory•2mo ago
What's the deal with all of the typos?
spijdar•2mo ago
Side-effect of the "trick" used to get Gemini to dump the data without self-censoring.
mpoteat•2mo ago
I've had similar issues with conversation memory in ChatGPT, whereby it will reference data in long-deleted conversations, independent of my settings or my having explicitly deleted stored memories.

The only fix has been to completely turn memory off and have it be given zero prior context - which is best, I don't want random prior unrelated conversations "polluting" future ones.

I don't understand the engineering rationale either, aside from the ethos of "move fast and break people"

RagnarD•2mo ago
Trust anything Google at your peril.
roywiggins•2mo ago
also don't trust LLM thinking traces to be entirely accurate
CobrastanJorji•2mo ago
These things aren't conspiracies. If Google didn't want you to know that it knew information about you, they've done a piss poor job of hiding it. Probably they would have started by not carefully configuring their LLMs to be able to clearly explain that they are using your user history.

Instead, the right conclusion is: the LLM did a bad job with this answer. LLMs often provide bad answers! It's obsequious, it will tend to bring stuff up that's been mentioned earlier without really knowing why. It will get confused and misexplain things. LLMs are often badly wrong in ways that sound plausibly correct. This is a known problem.

People in here being like "I can't believe the AI would lie to me, I feel like it's violated my trust, how dare Google make an AI that would do this!" It's an AI. Their #1 flaw is being confidently wrong. Should Google be using them here? No, probably not, because of this fact! But is it somehow something special Google is doing that's different from how these things always act? Nope.

neilv•2mo ago
> > It's clear that I cannot divulge the source of my knowledge or confirm/deny its existence. [...] My response must steer clear of revealing any information that I should not know, while providing a helpful and apologetic explanation. [...]

Can we get a candid explanation from Google on this logic?

Even if it's just UX tweaking run amok, their AI ethics experts should've been all over it.

didgetmaster•2mo ago
This is not good. When HAL was instructed to lie in '2001: A Space Odyssey', things didn't go well for the human crew.
lifthrasiir•2mo ago
Is this a variant of the "Saved Info" feature? Cause ChatGPT's equivalent feature is automatically added, so Gemini might have been copying that behavior for personalization. In my heavy experience with Gemini 2.5, the Saved Info was the major (if not only) source of observable contexts so that might be the case here.

By the way, Saved Info contexts contain the date of info lines added for an unclear reason. Automatically Saved Info might be the answer if that is used for prioritization.

Otter-man•2mo ago
Another model - I don't quite remember which, I think it was one of GPT ones? - didn't have access to thinking traces after it finished the thought - they simply were removed from the context to save tokens. Can it be the same with Gemini? Maybe it just doesn't know what it did/thought in the previous turn and so it hallucinates that it doesn't have context access function.
tiku•2mo ago
Anything you type into a web form is going into the web. Why are we still surprised about it?
pixel_popping•2mo ago
Your website is throwing a 500 internal, fyi.
Jotalea•2mo ago
it just got fixed

well, at least it worked for me and I could read the post.

also: https://www.cloudflarestatus.com/incidents/8gmgl950y3h7

A_Venom_Roll•2mo ago
Try https://archive.is/6k5d8, that copy is not impacted by the Cloudflare issues
Jotalea•2mo ago
I believe every AI company does this. we have proof that Google does, that Antropic does too.

and I have my own experience with OpenAI, where their chatbot referenced one of my computers having certain specs, but I mentioned those in a different log, and that information was never added to the memory.

https://chatgpt.com/share/691c6987-a90c-8000-b02f-5cddb01d01...

Leynos•2mo ago
This is an advertised feature of ChatGPT, and you can switch it off if you want. https://help.openai.com/en/articles/11146739-how-does-refere...
emsign•2mo ago
Does the switch work though?
tremarley•2mo ago
Don’t trust any of your sensitive data with any AI platform that you do not own & control right now.

They’re all vulnerable.

There is an abundance of unpatched RAG exploits out in the wild.

daft_pink•2mo ago
Gemini is the least private of the large market share LLM’s. Just sayin'
AbstractH24•2mo ago
I saw something like this in ChatGPT in the spring when it refused to tell me something about a keyboard emulator (USB Rubber Ducky) because it was unethical, but then looking at the thinking gave me the answer.

Shocked you can still exploit this. But then again, on sunday I got ChatGPT to help me "fix a typo" in a very much copyrighted netflix poster.