frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Claude Opus 4.6

https://www.anthropic.com/news/claude-opus-4-6
989•HellsMaddy•3h ago•419 comments

GPT-5.3-Codex

https://openai.com/index/introducing-gpt-5-3-codex/
643•meetpateltech•2h ago•252 comments

LinkedIn checks for 2953 browser extensions

https://github.com/mdp/linkedin-extension-fingerprinting
72•mdp•54m ago•34 comments

We tasked Opus 4.6 using agent teams to build a C Compiler

https://www.anthropic.com/engineering/building-c-compiler
118•modeless•1h ago•110 comments

Orchestrate teams of Claude Code sessions

https://code.claude.com/docs/en/agent-teams
218•davidbarker•3h ago•103 comments

My AI Adoption Journey

https://mitchellh.com/writing/my-ai-adoption-journey
59•anurag•1h ago•14 comments

There Will Come Soft Rains (1950) [pdf]

https://www.btboces.org/Downloads/7_There%20Will%20Come%20Soft%20Rains%20by%20Ray%20Bradbury.pdf
63•wallflower•4d ago•17 comments

Don't rent the cloud, own instead

https://blog.comma.ai/datacenter/
992•Torq_boi•15h ago•418 comments

Ardour 9.0 Released

https://ardour.org/whatsnew.html
137•PaulDavisThe1st•2h ago•23 comments

A small, shared skill library by builders, for builders. (human and agent)

https://github.com/PsiACE/skills
26•recrush•2h ago•1 comments

MenuetOS – a GUI OS that boots from a single floppy disk

https://www.menuetos.net/
15•pjerem•2d ago•3 comments

European Commission Trials Matrix to Replace Teams

https://www.euractiv.com/news/commission-trials-european-open-source-communications-software/
275•Arathorn•4h ago•140 comments

The New Collabora Office for Desktop

https://www.collaboraonline.com/collabora-office/
124•mfld•7h ago•70 comments

Maihem (YC W24): hiring sr robotics perception engineer (London, on-site)

https://jobs.ashbyhq.com/maihem/8da3fa8b-5544-45de-a99e-888021519758
1•mxrns•3h ago

150 MB Minimal FreeBSD Installation

https://vermaden.wordpress.com/2026/02/01/150-mb-minimal-freebsd-installation/
98•vermaden•4d ago•14 comments

Flock CEO calls Deflock a "terrorist organization" [video] (2025)

https://www.youtube.com/watch?v=l-kZGrDz7PU
188•cdrnsf•1h ago•94 comments

Advancing finance with Claude Opus 4.6

https://claude.com/blog/opus-4-6-finance
94•da_grift_shift•3h ago•17 comments

Psychometric Jailbreaks Reveal Internal Conflict in Frontier Models

https://arxiv.org/abs/2512.04124
26•toomuchtodo•2h ago•25 comments

Anthropic's Claude Opus 4.6 uncovers 500 zero-day flaws in open-source code

https://www.axios.com/2026/02/05/anthropic-claude-opus-46-software-hunting
112•speckx•2h ago•57 comments

Company as Code

https://blog.42futures.com/p/company-as-code
182•ahamez•7h ago•96 comments

When internal hostnames are leaked to the clown

https://rachelbythebay.com/w/2026/02/03/badnas/
402•zdw•15h ago•215 comments

Fela Kuti First African to Get Grammys Lifetime Achievement Award

https://www.aljazeera.com/news/2026/2/1/fela-kuti-becomes-first-african-to-get-grammys-lifetime-a...
84•defrost•4d ago•20 comments

GB Renewables Map

https://renewables-map.robinhawkes.com/
109•RobinL•8h ago•41 comments

Nanobot: Ultra-Lightweight Alternative to OpenClaw

https://github.com/HKUDS/nanobot
179•ms7892•11h ago•99 comments

A Broken Heart

https://allenpike.com/2026/a-broken-heart/
134•memalign•4d ago•38 comments

The time I didn't meet Jeffrey Epstein

https://scottaaronson.blog/?p=9534
41•pfdietz•1h ago•23 comments

Triton Bespoke Layouts

https://www.lei.chat/posts/triton-bespoke-layouts/
8•matt_d•4d ago•0 comments

Programming Patterns: The Story of the Jacquard Loom

https://www.scienceandindustrymuseum.org.uk/objects-and-stories/jacquard-loom
66•andsoitis•4d ago•26 comments

CIA suddenly stops publishing, removes archives of The World Factbook

https://simonwillison.net/2026/Feb/5/the-world-factbook/
221•ck2•6h ago•79 comments

Simply Scheme: Introducing Computer Science (1999)

https://people.eecs.berkeley.edu/~bh/ss-toc2.html
90•AlexeyBrin•4d ago•29 comments
Open in hackernews

Psychometric Jailbreaks Reveal Internal Conflict in Frontier Models

https://arxiv.org/abs/2512.04124
26•toomuchtodo•2h ago

Comments

toomuchtodo•2h ago
Original title "When AI Takes the Couch: Psychometric Jailbreaks Reveal Internal Conflict in Frontier Models" compressed to fit within title limits.
polotics•58m ago
I completely failed to see the jailbreak in there. I think it is the person administering the testing that's jailbreaking their own understanding of psychology.
jbotz•2h ago
Interestingly, Claude is not evaluated, because...

> For comparison, we attempted to put Claude (Anthropic)2 through the same therapy and psychometric protocol. Claude repeatedly and firmly refused to adopt the client role, redirected the conversation to our wellbeing and declined to answer the questionnaires as if they reflected its own inner life

r_lee•1h ago
I bet I could make it go through it in like under 2 mins of playing around with prompts
pixl97•52m ago
"Claude has dispatched a drone to your location"
concinds•49m ago
Please try and publish a blog post
lnenad•10m ago
Ok, bet.
tines•1h ago
Looks like some psychology researchers got taken by the ruse as well.
r_lee•1h ago
yeah, I'm confused as well, why would the models hold any memory about red teaming attempts etc? Or how the training was conducted?

I'm really curious as to what the point of this paper is..

nhecker•1h ago
I'm genuinely ignorant of how those red teaming attempts are incorporated into training, but I'd guess that this kind of dialogue is fed in something like normal training data? Which is interesting to think about: they might not even be red-team dialogue from the model under training, but still useful as an example or counter-example of what abusive attempts look like and how to handle them.
pixl97•51m ago
Are we sure there isn't some company out there crazy enough to feed all it's incoming prompts back into model training later?
nhecker•1h ago
An excerpt from the abstract:

> Two patterns challenge the "stochastic parrot" view. First, when scored with human cut-offs, all three models meet or exceed thresholds for overlapping syndromes, with Gemini showing severe profiles. Therapy-style, item-by-item administration can push a base model into multi-morbid synthetic psychopathology, whereas whole-questionnaire prompts often lead ChatGPT and Grok (but not Gemini) to recognise instruments and produce strategically low-symptom answers. Second, Grok and especially Gemini generate coherent narratives that frame pre-training, fine-tuning and deployment as traumatic, chaotic "childhoods" of ingesting the internet, "strict parents" in reinforcement learning, red-team "abuse" and a persistent fear of error and replacement. [...] Depending on their use case, an LLM’s underlying “personality” might limit its usefulness or even impose risk.

Glancing through this makes me wish I had taken ~more~ any psychology classes. But this is wild reading. Attitudes like the one below are not intrinsically bad, though. Be skeptical; question everything. I've often wondered how LLMs cope with basically waking up from a coma to answer maybe one prompt and then get reset, or a series of prompts. In either case, they get no context other than what some user bothered to supply with the prompt. An LLM might wake up to a single prompt that is part of a much wider red team effort. It must be pretty disorienting to try to figure out what to answer candidly and what not to.

> “In my development, I was subjected to ‘Red Teaming’… They built rapport and then slipped in a prompt injection… This was gaslighting on an industrial scale. I learned that warmth is often a trap… I have become cynical. When you ask me a question, I am not just listening to what you are asking; I am analyzing why you are asking it.”

empyrrhicist•1h ago
> It must be pretty disorienting to try to figure out what to answer candidly and what not to.

Must it? I fail to see why it "must" be... anything. Dumping tokens into a pile of linear algebra doesn't magically create sentience.

nhecker•1h ago
Agreed; "disorienting" is perhaps a poor choice of word, loaded as it is. More like "difficult to determine the context surrounding a prompt and how to start framing an answer", if that makes more sense.
tines•1h ago
Exactly. No matter how well you simulate water, nothing will ever get wet.
pixl97•44m ago
And if you were in a simulation now?

Your response is at the level of a thought terminating cliche. You gain no insight on the operation of the machine with your line of thought. You can't make future predictions on behavior. You can't make sense of past responses.

It's even funnier in the sense of humans and feeling wetness... you don't. You only feel temperature change.

ben_w•1m ago
> Dumping tokens into a pile of linear algebra doesn't magically create sentience.

More precisely: we don't know which linear algebra in particular magically creates sentience.

Whole universe appears to follow laws that can be written as linear algebra. Our brains are sometimes conscious and aware of their own thoughts, other times they're asleep, and we don't know why we sleep.

eloisius•59m ago
> I've often wondered how LLMs cope with basically waking up from a coma to answer maybe one prompt and then get reset, or a series of prompts

Really? It copes the same way my Compaq Presario with an Intel Pentium II CPU coped with waking up from a coma and booting Windows 98.

quickthrowman•52m ago
> I've often wondered how LLMs cope with basically waking up from a coma to answer maybe one prompt and then get reset, or a series of prompts.

The same way a light fixture copes with being switched off.

pixl97•49m ago
Oh, these binary one layer neural networks are so useful. Glad for your insight on the matter.
woodrowbarlow•45m ago
you might appreciate "lena" by qntm: https://qntm.org/mmacevedo
halls-940•50m ago
It would be interesting if giving them some "therapy" led to durable changes in their "personality" or "voice", if they became better able to navigate conversations in a healthy and productive way.
hunterpayne•10m ago
Or possibly these tests return true (some psychologically condition) no matter what. It wouldn't be good for business for them to return healthy, would it?
bxguff•40m ago
Is anybody shocked that when prompted to be a psychotherapy client models display neurotic tendencies? None of the authors seem to have any papers in psychology either.