Refusal observed with Opus 4.7, Opus 3, GTP-5.3, Gemini 3.
Is it a guardrail?
Refusal observed with Opus 4.7, Opus 3, GTP-5.3, Gemini 3.
Is it a guardrail?
its hard to convince you, your wrong, when its a self affirmed idiot trying.
i really dont see LLMs doing benign things, its a misinformation deluge.
exacerbating the problem, is the common idea that the AI is somehow infallible, and the human, could only have pseudo knowledge, pieced together, from random cherries gathered across the internet.
LLMs have become trolls, trolling for interaction worth training on.
You pretty much never see someone capitulate and simply agree that they are idiots. So why would an AI that models human interactions do it?
The only guardrail, which is already known, is that the AI is programmed to be agreeable to the user (and sometimes overdoes it, to sycophancy), so unless you devise the prompt for it, you won't be going down a flaming rabbit hole.
MattGaiser•1h ago