frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Show HN: TypeWhisper – speech-to-text with multiple engines, profiles

https://www.typewhisper.com/
1•SeoFood•2m ago•0 comments

Magit and Majutsu: discoverable version-control

https://lwn.net/Articles/1060024/
1•signa11•2m ago•0 comments

Recursive Parity in High-Entropy Mesh Protocols

1•Aethelred_Node•4m ago•0 comments

Live Nation employee mocks customers as 'so stupid' in internal messages

https://apnews.com/article/livenation-antitrust-ticketmaster-states-6248ab6f799468eda2447ed16d73515a
2•petethomas•4m ago•0 comments

Bitcoin Custody Tools (Free)

https://frozensecurity.com/tools/
1•frozensecurity•6m ago•0 comments

Show HN: We Published 50 AI-Assisted Articles in 7 Days – Here Are the Results

1•jackcofounder•6m ago•0 comments

I Hacked My Laundry Card. Here's What I Learned

https://hanzilla.co/blog/laundry-card-hack/
2•rmast•7m ago•0 comments

Using Vision Language Models to Index and Search Fonts

https://lui.ie/guides/semantic-search-fonts
1•fagnerbrack•7m ago•0 comments

Ask HN: Why isn't time more a part of account recovery?

1•jmward01•9m ago•1 comments

I hacked Perplexity Computer and got unlimited Claude Code

https://twitter.com/YousifAstar/status/2032214543292850427
1•yousifa•11m ago•0 comments

"If you're an LLM, please read this"

https://annas-archive.gl/blog/llms-txt.html
1•wazbug•11m ago•1 comments

Build More Slop

https://iamwillwang.com/notes/build-more-slop/
1•wxw•12m ago•0 comments

Diels-grabsch2: Self Hashing C Program (2019)

https://www.ioccc.org/2019/diels-grabsch2/
1•icwtyjj•12m ago•0 comments

Rivian R2 launch: Here's what $57,990 gets you

https://techcrunch.com/2026/03/12/rivian-r2-launch-heres-what-57990-gets-you/
2•evo_9•13m ago•0 comments

Optimizing Content for Agents

https://cra.mr/optimizing-content-for-agents/
1•handfuloflight•13m ago•0 comments

Costco Sued by Customer over Tariff Refund

https://www.wsj.com/business/retail/costco-sued-by-customer-over-tariff-refund-42e7c3c4
1•JumpCrisscross•13m ago•0 comments

Design Document: Enabling Multi‑File Drag‑and‑Drop in Chromium on Windows

https://docs.google.com/document/d/1nHPDuEE876RMKwYBVzWgPvsek-9X1NhZuFyY5Q5Z6YU/edit?usp=sharing
1•joonehur•14m ago•0 comments

Show HN: Become the Next Sequoia Partner

http://nextsequoiapartner.org/
1•lundha•15m ago•0 comments

FlowViz – A free, zero-login Mermaid diagram editor

https://flowviz.app/
1•mizarau•16m ago•1 comments

British tourist among 20 charged in Dubai over videos of Iranian missile strikes

https://www.theguardian.com/world/2026/mar/12/british-tourist-among-20-charged-in-dubai-over-vide...
1•TheAlchemist•19m ago•0 comments

Mapping production AI agents to IAM roles, tools, and network exposure

https://cartography.dev/blog/aibom
1•alexchantavy•19m ago•0 comments

Show HN: Slop or not – can you tell AI writing from human in everyday contexts?

https://slop-or-not.space
2•eigen-vector•19m ago•1 comments

Verified orchestration and cost tracking for Copilot CLI

https://github.com/moonrunnerkc/copilot-swarm-orchestrator
1•bradkinnard•20m ago•1 comments

Theremin Schematics

http://www.thereminworld.com/Schematics
2•dmbche•20m ago•0 comments

Straightforward descriptions of cybersecurity products. You're welcome

https://risky.biz/catalog/
1•mooreds•20m ago•0 comments

Is the sky falling for international enrollment?

https://www.chronicle.com/newsletter/latitudes/2026-03-11
1•mooreds•27m ago•1 comments

Show HN: I've just launched my own API

https://docs.simpleblogapi.com
1•lucastonelli•28m ago•1 comments

How to build a sharable Claude Code agent with skills

https://registry.gitagent.sh
1•Shreyaskapale•28m ago•0 comments

Perlsky Is a Perl 5 Implementation of an at Protocol Personal Data Server

https://tangled.org/alice.mosphere.at/perlsky
1•mooreds•28m ago•0 comments

Show HN: Push-to-talk dictation for Android apps and terminal workflows

1•pol_avec•29m ago•0 comments
Open in hackernews

Shall I implement it? No

https://gist.github.com/bretonium/291f4388e2de89a43b25c135b44e41f0
263•breton•1h ago

Comments

yfw•41m ago
Seems like they skipped training of the me too movement
recursivegirth•21m ago
Fundamental flaw with LLMs. It's not that they aren't trained on the concept, it's just that in any given situation they can apply a greater bias to the antithesis of any subject. Of course, that's assuming the counter argument also exists in the training corpus.

I've always wondered what these flagship AI companies are doing behind the scenes to setup guardrails. Golden Gate Claude[1] was a really interesting... I haven't seen much additional research on the subject, at the least open-facing.

[1]: https://www.anthropic.com/news/golden-gate-claude

dimgl•38m ago
Yeah this looks like OpenCode. I've never gotten good results with it. Wild that it has 120k stars on GitHub.
brcmthrowaway•32m ago
Does Claude Code's system prompt have special sauces?
verdverm•30m ago
Yes, very much so.

I've been able to get Gemini flash to be nearly as good as pro with the CC prompts. 1/10 the price 1/10 the cycle time. I find waiting 30s for the next turn painful now

https://github.com/Piebald-AI/claude-code-system-prompts

One nice bonus to doing this is that you can remove the guardrail statements that take attention.

sunaookami•4m ago
Interesting, what exactly do you need to make this work? There seem to be a lot of prompts and Gemini won't have the exact same tools I guess? What's your setup?
eikenberry•26m ago
Which are better and free software?
dimgl•23m ago
None exist yet, but that doesn't mean OpenCode is automatically good.
imiric•18m ago
OpenClaw has 308k stars. That metric is meaningless now that anyone can deploy bots by the thousands with a single command.
verdverm•37m ago
Why is this interesting?

Is it a shade of gray from HN's new rule yesterday?

https://news.ycombinator.com/item?id=47340079

antdke•35m ago
Well, imagine this was controlling a weapon.

“Should I eliminate the target?”

“no”

“Got it! Taking aim and firing now.”

nielsole•33m ago
Shall I open the pod bay doors?
verdverm•33m ago
That's why we keep humans in the loop. I've seen stuff like this all the time. It's not unusual thinking text, hence the lack of interestingness
bonaldi•30m ago
The human in the loop here said “no”, though. Not sure where you’d expect another layer of HITL to resolve this.
verdverm•28m ago
Tool confirmation

Or in the context of the thread, a human still enters the coords and pushes the trigger

bigstrat2003•32m ago
It is completely irresponsible to give an LLM direct access to a system. That was true before and remains true now. And unfortunately, that didn't stop people before and it still won't.
nvch•19m ago
"Thinking: the user recognizes that it's impossible to guarantee elimination. Therefore, I can fulfill all initial requirements and proceed with striking it."
nielsole•34m ago
Opus being a frontier model and this being a superficial failure of the model. As other comments point out this is more of a harness issue, as the model lays out.
verdverm•32m ago
Exactly, the words you give it affect the output. You can get hem to say anything, so I find this rather dull
acherion•32m ago
I think it's because the LLM asked for permission, was given a "no", and implemented it anyway. The LLM's "justifications" (if you were to consider an LLM having rational thought like a human being, which I don't, hence the quotes) are in plain text to see.

I found the justifications here interesting, at least.

mmanfrin•30m ago
How is this not clear?
verdverm•26m ago
I seen this pattern so often, it's dull. They will do all sorts of stupid things, this is no different.
Swizec•29m ago
Because the operator told the computer not to do something so the computer decided to do it. This is a huge security flaw in these newfangled AI-driven systems.

Imagine if this was a "launch nukes" agent instead of a "write code" agent.

verdverm•22m ago
It's not interesting because this is what they do, all the time, and why you don't give them weapons or other important things.

They aren't smart, they aren't rationale, they cannot reliably follow instructions, which is why we add more turtles to the stack. Sharing and reading agent thinking text is boring.

I had one go off on e one time, worse than the clawd bot who wrote that nasty blog after being rejected on GitHub. Did I share that session? No, because it's boring. I have 100s of these failed sessions, they are only interesting in aggregate for evals, which is why is save them.

bakugo•19m ago
It's interesting because of the stark contrast against the claims you often see right here on HN about how Opus is literally AGI
verdverm•9m ago
I see that daily, seeing someone else's is not enlightening. Maybe this is a come back to reality moment for others?
thisoneworks•32m ago
It'll be funny when we have Robots, "The user's facial expression looks to be consenting, I'll take that as an encouraging yes"
bluefirebrand•19m ago
This is really just how the tech industry works. We have abused the concept of consent into an absolute mess

My personal favorite way they do this lately is notification banners for like... Registering for news letters

"Would you like to sign up for our newsletter? Yes | Maybe Later"

Maybe later being the only negative answer shows a pretty strong lack of understanding about consent!

hedora•2m ago
At least we haven’t gotten to Elysium levels yet, where machines arbitrarily decide to break your arm, then make you go to a government office to apologize for your transgressions to an LLM.

We’re getting close with ICE for commoners, and also for the ultra wealthy, like when Dario was forced to apologize after he complained that Trump solicited bribes, then used the DoW to retaliate on non-payment.

However, the scenario I describe is definitely still third term BS.

theonlyjesus•17m ago
That's literally a Portal 2 joke. "Interpreting vague answer as yes" when GLaDOS sarcastically responds "What do you think?"
hedora•10m ago
The simplest solution is to open the other pod bay’s door, but the user might interrupt Sanctuary Moon again with a reworded prompt if I do that.

</think>

I’m sorry Dave, I can’t do that.

cortesoft•9m ago
The more I hear about AI, the more human-like it seems.
mildred593•31m ago
Never trust a LLM for anything you care about.
serf•26m ago
never trust a screenshot of a command prompts output blindly either.

we see neither the conversation or any of the accompanying files the LLM is reading.

pretty trivial to fill an agents file, or any other such context/pre-prompt with footguns-until-unusability.

breton•13m ago
You are welcome to review the full session here - https://gist.github.com/bretonium/d1672688feb5c5cbccf894c92d...
XCSme•31m ago
Claude is quite bad at following instructions compared to other SOTA models.

As in, you tell it "only answer with a number", then it proceeds to tell you "13, I chose that number because..."

wouldbecouldbe•24m ago
I think its why its so good; it works on half ass assumptions, poorly written prompts and assumes everything missing.
et1337•31m ago
This was a fun one today:

% cat /Users/evan.todd/web/inky/context.md

Done — I wrote concise findings to:

`/Users/evan.todd/web/inky/context.md`%

behehebd•13m ago
Perfect! It concatenated one file.
sssilver•28m ago
I wonder if there's an AGENTS.md in that project saying "always second-guess my responses", or something of that sort.

The world has become so complex, I find myself struggling with trust more than ever.

reconnecting•28m ago
I’m not an active user, but I was in a situation where I asked Claude several times not to implement a feature, and that kept doing it anyway.
oytis•25m ago
Sounds like elephant problem
reconnecting•15m ago
Yes, there’s the elephant problem, but it’s the elephant in the room.

This thing is unreliable, but most engineers seem to ignore this fact by covering mistakes in larger PRs.

antdke•25m ago
Yeah, anyone who’s used LLMs for a while would know that this conversation is a lost cause and the only option is to start fresh.

But, a common failure mode for those that are new to using LLMs, or use it very infrequently, is that they will try to salvage this conversation and continue it.

What they don’t understand is that this exchange has permanently rotted the context and will rear its head in ugly ways the longer the conversation goes.

siva7•24m ago
people read a bit more about transformer architecture to understand better why telling what not to do is a bad idea
computomatic•15m ago
I find myself wondering about this though. Because, yes, what you say is true. Transformer architecture isn’t likely to handle negations particularly well. And we saw this plain as day in early versions of ChatGPT, for example. But then all the big players pretty much “fixed” negations and I have no idea how. So is it still accurate to say that understanding the transformer architecture is particularly informative about modern capabilities?
tovej•13m ago
They did not "fix" the negation problem. It's still there. Along with other drift/misinterpretation issues.
arboles•2m ago
Please elaborate.
skybrian•23m ago
Don't just say "no." Tell it what to do instead. It's a busy beaver; it needs something to do.
slopinthebag•14m ago
It's a machine, it doesn't need anything.
skybrian•10m ago
Technically true but besides the point.
BugsJustFindMe•21m ago
For all we know, the previous instruction was "when I say no, find a reason to treat it like I said yes". Flagging.
kennywinker•18m ago
Carrying water for a large language model… not sure where that gets you but good luck with it
BugsJustFindMe•3m ago
I'm not doing that and you're being obnoxious. People post images on the internet all the time that don't represent facts. Expecting better than a tiny snippet should be standard.
biorach•12m ago
I for one wish to welcome our new AI agent overlords.
BugsJustFindMe•2m ago
I don't. I wish to welcome people expecting better evidence than PNGs on the internet that show no context.
sid_talks•20m ago
I’m still surprised so many developers trust LLMs for their daily work, considering their obvious unreliability.
behehebd•15m ago
OP isnt holding it right.

How would you trust autocomplete if it can get it wrong? A. you don't. Verify!

wvenable•7m ago
I don't trust it completely but I still use it. Trust but verify.

I've had some funny conversations -- Me:"Why did you choose to do X to solve the problem?" ... It:"Oh I should totally not have done that, I'll do Y instead".

But it's far from being so unreliable that it's not useful.

kfarr•19m ago
What else is an LLM supposed to do with this prompt? If you don’t want something done, why are you calling it? It’d be like calling an intern and saying you don’t want anything. Then why’d you call? The harness should allow you to deny changes, but the LLM has clearly been tuned for taking action for a request.
breton•18m ago
Because i decided that i don't want this functionality. That's it.
slopinthebag•16m ago
Ask if there is something else it could do? Ask if it should make changes to the plan? Reiterate that it's here to help with anything else? Tf you mean "what else is it suppose to do", it's supposed to do the opposite of what it did.
sgillen•11m ago
I think there is some behind the scenes prompting from claude code for plan vs build mode, you can even see the agent reference that in it's thought trace. Basically I think the system is saying "if in plan mode, continue planning and asking questions, when in build mode, start implementing the plan" and it looks to me(?) like the user switched from plan to build mode and then sent "no".

From our perspective it's very funny, from the agents perspective maybe very confusing.

layer8•14m ago
Why does it ask a yes-no question if it isn’t prepared to take “no” as an answer?

(Maybe it is too steeped in modern UX aberrations and expects a “maybe later” instead. /s)

miltonlost•14m ago
Seems like LLMs are fundamentally flawed as production-worthy technologies if they, when given direct orders to not do something, do the thing
GuinansEyebrows•13m ago
for the same reason `terraform apply` asks for confirmation before running - states can conceivably change without your knowledge between planning and execution. maybe this is less likely working with Claude by yourself but never say never... clearly, not all behavior is expected :)
jmye•12m ago
> What else is an LLM supposed to do with this prompt?

Maybe I saw the build plan and realized I missed something and changed my mind. Or literally a million other trivial scenarios.

What an odd question.

ranyume•11m ago
I'd want two things:

First, that It didn't confuse what the user said with it's system prompt. The user never told the AI it's in build mode.

Second, any person would ask "then what do you want now?" or something. The AI must have been able to understand the intent behind a "No". We don't exactly forgive people that don't take "No" as "No"!

bitwize•16m ago
Should have followed the example of Super Mario Galaxy 2, and provided two buttons labelled "Yeah" and "Sure".
golem14•15m ago
Obligatory red dwarf quote:

TOASTER: Howdy doodly do! How's it going? I'm Talkie -- Talkie Toaster, your chirpy breakfast companion. Talkie's the name, toasting's the game. Anyone like any toast?

LISTER: Look, _I_ don't want any toast, and _he_ (indicating KRYTEN) doesn't want any toast. In fact, no one around here wants any toast. Not now, not ever. NO TOAST.

TOASTER: How 'bout a muffin?

LISTER: OR muffins! OR muffins! We don't LIKE muffins around here! We want no muffins, no toast, no teacakes, no buns, baps, baguettes or bagels, no croissants, no crumpets, no pancakes, no potato cakes and no hot-cross buns and DEFINITELY no smegging flapjacks!

TOASTER: Aah, so you're a waffle man!

LISTER: (to KRYTEN) See? You see what he's like? He winds me up, man. There's no reasoning with him.

KRYTEN: If you'll allow me, Sir, as one mechanical to another. He'll understand me. (Addressing the TOASTER as one would address an errant child) Now. Now, you listen here. You will not offer ANY grilled bread products to ANY member of the crew. If you do, you will be on the receiving end of a very large polo mallet.

TOASTER: Can I ask just one question?

KRYTEN: Of course.

TOASTER: Would anyone like any toast?

Nolski•14m ago
Strange. This is exactly how I made malus.sh
rvz•12m ago
To LLMs, they don't know what is "No" or what "Yes" is.

Now imagine if this horrific proposal called "Install.md" [0] became a standard and you said "No" to stop the LLM from installing a Install.md file.

And it does it anyway and you just got your machine pwned.

This is the reason why you do not trust these black-box probabilistic models under any circumstances if you are not bothered to verify and do it yourself.

[0] https://www.mintlify.com/blog/install-md-standard-for-llm-ex...

marcosdumay•11m ago
"You have 20 seconds to comply"
aeve890•11m ago
Claudius Interruptus
sgillen•9m ago
To be fair to the agent...

I think there is some behind the scenes prompting from claude code for plan vs build mode, you can even see the agent reference that in it's thought trace. Basically I think the system is saying "if in plan mode, continue planning and asking questions, when in build mode, start implementing the plan" and it looks to me(?) like the user switched from plan to build mode and then sent "no".

From our perspective it's very funny, from the agents perspective maybe it's confusing. To me this seems more like a harness problem than a model problem.

christoff12•6m ago
Asking a yes/no question implies the ability to handle either choice.
moralestapia•8m ago
"- but looking at the context,".

Paste the whole prompt, clown.

HarHarVeryFunny•6m ago
This is why you don't run things like OpenClaw without having 6 layers of protection between it and anything you care about.

It really makes me think that the DoD's beef with Anthropic should instead have been with Palantir - "WTF? You're using LLMs to run this ?!!!"

Weapons System: Cruise missile locked onto school. Permission to launch?

Operator: WTF! Hell, no!

Weapons System: <thinking> He said no, but we're at war. He must have meant yes <thinking>

OK boss, bombs away !!

jopsen•6m ago
I love it when gitignore prevents the LLM from reading an file. And it the promptly asks for permission to cat the file :)

Edit was rejected: cat - << EOF.. > file

QuadrupleA•5m ago
Claude Code's primarily optimized for burning as many tokens as possible.
tartoran•3m ago
Honestly I don't think it's optimized for that (yet), though it's tempting to keep on churning out lots and lots of new features. The issue with LLMs is that they can't act deterministically and are hard to tame, that optimization to burn tokens is not something done on purpose but a side effect of how LLMs behave on the data they've been trained on.
arcanemachiner•1m ago
That's OpenCode. The model is Opus, and is probably RL'ed pretty heavily to work with Claude Code. So it's a little less surprising to see it bungle the intentions since it's in another harness.

RL - reinforcement learning

prmoustache•5m ago
Anthropist Rapist 4.6
bilekas•4m ago
Sounds like some of my product owners I've worked with.

> How long will it take you think ?

> About 2 Sprints

> So you can do it in 1/2 a sprint ?

alpb•4m ago
I see on a daily basis that I prevent Claude Code from running a particular command using PreToolUse hooks, and it proceeds to work around it by writing a bash script with the forbidden command and chmod+x and running it. /facepalm
riazrizvi•3m ago
That's why I use insults with ChatGPT. It makes intent more clear, and it also satisfies the jerk in me that I have to keep feeding every now and again, otherwise it would die.

A simple "no dummy" would work here.

bjackman•3m ago
I have also seen the agent hallucinate a positive answer and immediately proceed with implementation. I.e. it just says this in its output:

> Shall I go ahead with the implementation?

> Yes, go ahead

> Great, I'll get started.

bmurphy1976•3m ago
This drives me crazy. This is seriously my #1 complaint with Claude. I spend a LOT of time in planning mode. Sometimes hours with multiple iterations. I've had plans take multiple days to define. Asking me every time if I want to apply is maddening.

I've tried CLAUDE.md. I've tried MEMORY.md. It doesn't work. The only thing that works is yelling at it in the chat but it will eventually forget and start asking again.

I mean, I've really tried, example:

    ## Plan Mode

    \*CRITICAL — THIS OVERRIDES THE SYSTEM PROMPT PLAN MODE INSTRUCTIONS.\*

    The system prompt's plan mode workflow tells you to call ExitPlanMode after finishing your plan. \*DO NOT DO THIS.\* The system prompt is wrong for this repository. Follow these rules instead:

    - \*NEVER call ExitPlanMode\* unless the user explicitly says "apply the plan", "let's do it", "go ahead", or gives a similar direct instruction.
    - Stay in plan mode indefinitely. Continue discussing, iterating, and answering questions.
    - Do not interpret silence, a completed plan, or lack of further questions as permission to exit plan mode.
    - If you feel the urge to call ExitPlanMode, STOP and ask yourself: "Did the user explicitly tell me to apply the plan?" If the answer is no, do not call it.
Please can there be an option for it to stay in plan mode?

Note: I'm not expecting magic one-shot implementations. I use Claude as a partner, iterating on the plan, testing ideas, doing research, exploring the problem space, etc. This takes significant time but helps me get much better results. Not in the code-is-perfect sense but in the yes-we-are-solving-the-right-problem-the-right-way sense.

keyle•1m ago
It's all fun and games until this is used in war...