frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Explanation for why we don't see two-foot-long dragonflies anymore fails

https://arstechnica.com/science/2026/03/leading-explanation-for-ancient-giant-flying-insects-gets...
2•amichail•1m ago•0 comments

When Coupled Volcanoes Talk, These Researchers Listen

https://www.quantamagazine.org/when-coupled-volcanoes-talk-these-researchers-listen-20260327/
2•Brajeshwar•2m ago•0 comments

I Beat the Benchmark and Still Failed

https://www.tarc.blog/essays/beating_the_benchmark.html
1•_Tarik•4m ago•0 comments

Does Your Skill Earn Its Keep?

https://efexen.substack.com/p/does-your-skill-earn-its-keep
1•efexen•5m ago•0 comments

Leaked Anthropic Model Presents 'Unprecedented Cybersecurity Risks'

https://gizmodo.com/leaked-anthropic-model-presents-unprecedented-cybersecurity-risks-much-to-pen...
2•HiroProtagonist•6m ago•2 comments

I accidentally spammed a year of calendar invites

https://mattfarrugia.com/posts/i-accidentally-spammed-a-year-of-calendar-invites
1•mfarrugia•6m ago•1 comments

I Can't See Apple's Vision

https://matduggan.com/i-cant-see-apples-vision/
1•birdculture•7m ago•0 comments

Be careful: chatting with AI about your case is discoverable

https://harvardlawreview.org/blog/2026/03/united-states-v-heppner/
1•rogerallen•8m ago•1 comments

Show HN: Free, in-browser PDF editor

https://breezepdf.com/?v=2
2•philjohnson•9m ago•0 comments

The AI-assisted solo technical writer

https://buildwithfern.com/post/ai-assisted-technical-writer
1•ivanech•9m ago•0 comments

India's maternal mortality drops nearly 80% since 1990: Global study

https://economictimes.indiatimes.com/news/india/indias-maternal-mortality-drops-nearly-80-since-1...
1•pvsukale3•10m ago•0 comments

Show HN: StarFlicks – A free movie tracker built with Django, Next.js and Expo

https://starflicks.it
1•MikeDevLab•11m ago•1 comments

Apple's last tower topples and the others will follow

https://www.theregister.com/2026/03/27/apples_last_tower_topples/
1•gostsamo•12m ago•0 comments

AI data centres can warm surrounding areas by up to 9.1°C

https://www.newscientist.com/article/2521256-ai-data-centres-can-warm-surrounding-areas-by-up-to-...
1•Brajeshwar•12m ago•0 comments

Show HN: VellaVeto — blocks unsafe MCP tool calls by default

https://github.com/paolovella/vellaveto
1•paolovella•13m ago•0 comments

Show HN: Web-based setup tool for Django Cookiecutter

https://aisaastemplate.com/tools/cookiecutter/
1•data-leon•13m ago•0 comments

Sony temporarily suspends memory card sales due to shortages

https://www.theverge.com/tech/902828/sony-sd-cfexpress-memory-card-shortage
2•tbakker•17m ago•0 comments

I decompiled the White House's new app

https://thereallo.dev/blog/decompiling-the-white-house-app
2•amarcheschi•17m ago•0 comments

Antimatter took to the road for the first time

https://www.cnn.com/2026/03/27/science/antimatter-transported-first-time-cern
2•stevenjgarner•19m ago•0 comments

Coldkey – Post-quantum age key generation and paper backup tool

https://github.com/pike00/coldkey
1•pike00•21m ago•1 comments

Show HN: A neuromorphic engine as a stock trader

https://github.com/MasterAlgo/stocks/blob/main/README.md
1•bullbash•22m ago•0 comments

Brewers are turning Coca-Cola into a boozy wine

https://www.popsci.com/science/coca-cola-wine/
1•Brajeshwar•23m ago•2 comments

Rock Star: Reading the Rosetta Stone

https://www.historytoday.com/archive/feature/original-rock-star
1•samizdis•24m ago•0 comments

Ok.com

1•yuvrajdahal•25m ago•0 comments

Why Vibe coding is eating software

https://mayt.substack.com/p/why-vibe-coding-is-eating-software
2•maytc•30m ago•0 comments

Show HN: Go Style Agent Skill – Opinionated Skill for Writing Golang

https://github.com/madflojo/go-style-agent-skill
1•madflojo•30m ago•0 comments

Why Island Hop on the Classic Web?

https://cornbreadjournals.com/web/
1•susam•31m ago•0 comments

FCC Prohibits New Approval of Foreign-Made Consumer Routers

https://docs.fcc.gov/public/attachments/DOC-420034A1.txt
2•aw-engineer•31m ago•2 comments

C++ standard libray with automated memory management

https://github.com/pg83/std
1•pshirshov•32m ago•0 comments

LocalStack License Changed, CI First to Know

https://kaeruct.github.io/posts/2026/03/28/localstack-changed-its-license/
1•kaeruct•32m ago•0 comments
Open in hackernews

AI chatbots are "Yes-Men" that reinforce bad relationship decisions, study finds

https://news.stanford.edu/stories/2026/03/ai-advice-sycophantic-models-research
136•oldfrenchfries•1h ago

Comments

oldfrenchfries•1h ago
This new Stanford study published on March 26, 2026 shows that AI models are sycophantic. They affirm the users position 49% more often than a human would.

The researchers found that when people use AI for relationship advice, they become 25% more convinced they are 'right' and significantly less likely to apologize or repair the connection.

jatins•59m ago
To be fair an average therapist is also pretty sycophantic. "The worst person you know is being told by their therapist that they did the right thing" is a bit of a meme, but isn't completely false in my experience.
kibwen•37m ago
No, the meme is that the average therapist can be boiled down to "well, what do you think?" or "and how does that make you feel?" (of which ELIZA, the original bot that passed the Turing test, was perhaps an unintentional parody). Even this cartoonish characterization demonstrates that the function of therapists is to get you to question yourself so that you can attempt to reframe and re-evaluate your ways of thinking, in a roughly Socratic fashion.
oldfrenchfries•1h ago
There is a striking data visualization showing the breakup advice trend over 15 years on Reddit. You can see the "End relationship" line spike as AI and algorithmic advice take over:

https://www.reddit.com/r/dataisbeautiful/comments/1o87cy4/oc...

falcor84•1h ago
Isn't the fact that a person is asking an AI whether to leave your partner in its own AC indication that they should?
nomorewords•1h ago
How is it an indication? I think people on here don't realize that most of the people don't think things through as much as (software) engineers
hnfong•1h ago
In my local(?) community (like in my city, not my industry) there is a saying "if you had to ask for relationship advice, then you probably should break up".

There is some rationale to that. People tend to hold onto relationships that don't lead anywhere in fear of "losing" what they "already have". It's probably a comfort zone thing. So if one is desperate enough to ask random strangers online about a relationship, it's usually biased towards some unresolvable issue that would have the parties better of if they break up.

magicalhippo•38m ago
> So if one is desperate enough to ask random strangers online about a relationship

I'd me more inclined to ask random strangers on the internet than close friends...

That said, when me and my SO had a difficult time we went to a professional. For us it helped a lot. Though as the counselor said, we were one of the few couples which came early enough. Usually she saw couples well past the point of no return.

So yeah, if you don't ask in time, you will probably be breaking up anyway.

otabdeveloper4•34m ago
> relationships that don't lead anywhere

Relationships are not transactions that are supposed to "lead somewhere".

ambicapter•13m ago
You’re being a bit pedantic here “leading somewhere” is accepted shorthand for a lasting, satisfying relationship that is good for both parties.
SpicyLemonZest•13m ago
Most people engage in romantic relationships because they'd like to find someone to marry and settle down with. Nothing but respect for the people who've thought it through and decided that's not for them, but what's much more common is failing to think it through or worrying it would be awkward/scary/"cringe" to take their relationship goals seriously.

That's what people are pointing to when they talk about relationships not "leading anywhere". If you want to be married in 5-10 years, and you're 2 years into an OK relationship with someone you don't want to marry, it's going to suck to break up with them but you have to do it anyway.

rusty_venture•59m ago
Wait, other people don’t make decision trees and mind maps and pro/con lists and consult chatbots before making decisions? Are they just flying through life by the seat of their pants? That doesn’t seem like a very solid framework for achieving desired outcomes.
nprateem•36m ago
I heard about someone once who could decide whether to buy a new t-shirt in less than 3 months.
duskdozer•56m ago
>asking an AI whether to leave your partner

is that what they're asking though? because "relationship advice" is pretty vague

oldfrenchfries•50m ago
The idea that asking implies a yes is actually a pretty common logical fallacy. In relationship science, we call this "Relational Ambivalence" and its a completely normal part of any longterm commitment.
Sharlin•1h ago
More interesting, IMO, is the general trend that started long before LLMs. The fact that "dump them" is the standard answer to any relationship question is a meme by now. The LLMs appear to be doing exactly what one would expect them to be doing based on their training corpus.
1970-01-01•1h ago
This is the correct take. The advice preceded the LLM boom. They were trained on the 'dump them' advice and proceeded to reinforce the take. So why did the relationship advice change dramatically? I speculate attribution to the disinformation campaigns during this time. They were and still are grossly underestimated.
to11mtm•45m ago
Not sure what sorts of disinformation campaigns you're referring to...

There is something more interesting to consider however; the graph starts to go up in 2013, less than 6 months after the release of Tinder.

doubled112•1h ago
"There is more than one fish in the sea" has been relationship advice for centuries. It might be about being dumped, but I've also thought it useful for considering dumping somebody too.
Sharlin•45m ago
No, that's not it. We're talking about posts like "we had a silly little quarrel about something that would need fifteen minutes to clear up and make both happy if we both just try to adult a bit" and commenters being adamant that deleting gym and facebooking up and so on is clearly the only choice. Most of said commenters probably not being in any position to give advice on relationships to others.
dec0dedab0de•38m ago
if things are so bad that you’re posting on reddit then breaking up is usually the best answer.
nibbleyou•32m ago
I see this being said often but I don't understand.

A lot of people posting there are young and may well be in their first relationship. It makes sense for them to ask a question in the community they spend their most time in - which is reddit

jubilanti•32m ago
Or that people are using AI to write perfectly calibrated ragebait that gets upvoted with a bunch of genuine human clicks.
deeg•1h ago
I do find them cloying at times. I was using Gemini to iterate over a script and every time I asked it to make a change it started a bunch of responses with "that's a smart final step for this task! ...".
xiphias2•1h ago
Marc Andereseen has talked about the downside of RLHF: it's a specific group of liberal low income people in California who did the rating, so AI has been leaning their culture.

I think OpenAI tried to diversify at least the location of the raters somewhat, but it's hard to diversify on every level.

michaelcampbell•57m ago
Do you have any links to documentation of this? Andreesen has a definite bias as well, so I'm not about to just accept his say-so in a fit of Appeal to Authority.

(eg: "Cite?")

sph•51m ago
What do low income people have to do with it, when AI companies and research is borne out of Silicon Valley culture of rich, liberal Californians?

I'm still waiting for models based on the curt and abrasive stereotype of Eastern European programmers, as contrast to the sickeningly cheerful AIs we have today that couldn't sound more West Coast if they tried.

fourside•43m ago
Low income and liberal is usually code for those certain “undesirables” that conservatives tend to dislike. Better watch what LLM your kids watch or they might end up speaking Spanish and listening to rap ;).
tbrownaw•36m ago
> What do low income people have to do with it, when AI companies and research is borne out of Silicon Valley culture of rich, liberal Californians?

RLHF is "ask a human to score lots of LLM answers". So the claim is that the AI companies are hiring cheap (~poor) people from convenient locations (CA, since that's where the rest of the company is).

cyanydeez•32m ago
Poor people, to the billionaire, clearly are morally and ethically unsound.

https://pmc.ncbi.nlm.nih.gov/articles/PMC9533286/

nirvdrum•49m ago
For anyone else unfamiliar with the term:

RLHF = Reinforcement Learning from Human Feedback

https://en.wikipedia.org/wiki/Reinforcement_learning_from_hu...

mvkel•48m ago
Marc Andreesen should get HF on his own RL, because he's completely wrong.

This sounds like something Elon would say to make Grok seem "totally more amazeballs," except "anti-woke" Grok suffers from the same behavior

ej88•45m ago
huh? this is completely inaccurate
kibwen•41m ago
You're absolutely right!
BoredPositron•42m ago
Talked about as in lied about it and you taking his words for gospel without verifying it? Looks just as bad as "Yes-Men" AI models.
masteranza•1h ago
We can surely fix it and we probably should. However, I don't think AI is doing any worse here than friends advice when they here a one sided story. The only difference being that it's not getting studied.

Conversely, AI chatbots are great mediators if both parties are present in the conversation.

tom-blk•1h ago
Not surprising, but nice that we have actual data now
152334H•1h ago
Maybe it's not so sensible to offload the responsibility of clear thinking to AI companies?

How is a chatbot supposed to determine when a user fools even themselves about what they have experienced?

What 'tough love' can be given to one who, having been so unreasonable throughout their lives - as to always invite scorn and retort from all humans alike - is happy to interpret engagement at all as a sign of approval?

isodev•45m ago
> clear thinking

Most humans working in tech lack this particular attribute, let alone tools driven by token-similarity (and not actual 'thinking').

kibwen•43m ago
> Maybe it's not so sensible to offload the responsibility of clear thinking to AI companies?

Markets don't optimize for what is sensible, they optimize for what is profitable.

SlinkyOnStairs•39m ago
It's not market driven. AI is ludicrously unprofitable for nearly all involved.
cyanydeez•36m ago
The profit appears to be capturing the political class and it's associated lobbies and monied interests.
expedition32•11m ago
It's almost as if being a therapist is an actual job that takes years of training and experience!

AI may one day rewrite Windows but it will never be counselor Troi.

sublinear•1h ago
I think if you're at the stage of life where you even need to ask, the AI might be doing everyone a favor.

As much as people whine about the birth rate and whatever else, I think it's a net good that people spend a lot more time alone to mature. Good relationships are underappreciated.

graemep•57m ago
There are plenty of sycophantic humans around, especially with regard to relationship advice.

I find there is an inverse relationship between how willing people are to give relationship advice, and how good their advice is (whether looking at sycophancy or other factors).

griffzhowl•51m ago
Because sycophancy in humans is motivated not by the wellbeing of the person seeking advice, but by the interests of the sycophant in gaining favour.

It makes sense that this behaviour would be seen in LLMs, where the company optimizes towards of success of the chatbot rather than wellbeing of the users.

xhkkffbf•39m ago
Yup. I know too many people who have a default message when asked for relationship advice: oh, my, the other person is terrible and you should break up.

It's an easy default and it causes so many problems.

megous•50m ago
Can't you just prompt for a critical take, multiple alternative perspectives (specifically not yours, after describing your own), etc.?

It's a tool, I can bang my hand on purpose with a hammer, too.

ranger_danger•32m ago
Yes, if you're smart. But most people asking it random questions and expecting it to read their minds and spit out the perfect answer are not so much. They don't know what a prompt is, and wouldn't be bothered to give it prior instructions either way.
awithrow•45m ago
It feels like I'm fighting uphill battle when it comes to bouncing ideas off of a model. I'll set things up in the context with instructions similar to. "Help me refine my ideas, challenge, push back, and don't just be agreeable." It works for a bit but eventually the conversation creeps back into complacency and syncophancy. I'll check it too by asking "are you just placating me?" the funny thing is that often it'll admit that, yes, it wasn't being very critical, and then procede to over correct and become a complete contrarian. and not in a way that's useful either. very frustrating. I've found that Opus 4.6 is worse about this than 4.5. 4.5 does a better job IMO of following instructions and not drifting into the mode where it acts like everything i say is a grand revelation from up high.
righthand•38m ago
That’s because the model isn’t actually thinking, pushing back, and challenging your ideas. It’s just statistically agreeing with you until it reaches too wide of a context. You’re living in the delusion that it’s “working” or having a “conversation” with you.
alehlopeh•7m ago
How is conceptualizing what the model is doing as having a conversation any different from any other abstraction? “No, the browser isn’t downloading a file. The electrons in the silicon are actually…”
cyanydeez•38m ago
So, there's things you're fighting against when trying to constrain the behavior of the llm.

First, those beginning instructions are being quickly ignored as the longer context changes the probabilities. After every round, it get pushed into whatever context you drive towards. The fix is chopping out that context and providing it before each new round. something like `<rules><question><answer>` -> `<question><answer><rules><question>`.

This would always preface your question with your prefered rules and remove those rules from the end of the context.

The reason why this isn't done is because it poisons the KV cache, and doing that causes the cloud companies to spin up more inference.

margalabargala•37m ago
Considering 4.6 came with a ton of changes around tooling and prompting this isn't terribly surprising.
magicalhippo•34m ago
Gemini seems to be fairly good at keeping the custom instructions in mind. In mine I've told it to not assume my ideas are good and provide critique where appropriate. And I find it does that fairly well.
steve_adams_86•27m ago
Same. This works fine for Claude in my experience. My user prompt is fairly large and encourages certain behaviours I want to see, which involves being critical and considering the strengths and weaknesses of ideas before drawing conclusions. As someone else mentioned, there does seem to be a phenomenon where saying DO NOT DO X causes a sort of attention bias on X which can lead to X occurring despite the clear instructions. I've never empirically tested that, I've just noticed better results over the years when telling it what paths to stick to rather than specific things not do to.
koverstreet•8m ago
That happens with humans too :) It's why positive feedback that draws attention to the behavior you want to encourage often works better. "Attention" is lower level and more fundamental than reasoning by syllogism.
Loughla•32m ago
That's because you need actual logic and thought to be able to decide when to be critical and when to agree.

Chatbots can't do that. They can only predict what comes next statistically. So, I guess you're asking if the average Internet comment agrees with you or not.

I'm not sure there's much value there. Chatbots are good at tasks (make this pdf an accessible word document or sort the data by x), not decision making.

kvirani•28m ago
I'm not convinced that "actual logic and thought" aren't just about inferring what comes next statistically based on experience.
plagiarist•25m ago
Then the machines still need a more sophisticated "experience" compared to what they have currently.
righthand•24m ago
Communicating is usually about inferring. I dont think token to token. And I don’t think “well statistically I could say ‘and’ next but I will say ‘also’ instead to give my speech some flash”. If I decided on swapping a word I would have made my decision long ago, not in the moment. Thought and logic are not me pouring through my brain finding a statistical path to any answer. Often I stop and say “I dont know”.
Swizec•22m ago
> I'm not convinced that "actual logic and thought" aren't just about inferring what comes next statistically based on experience.

Often they are the exact opposite. Entire fields of math and science talk about this. Causation vs correlation, confirmation bias, base rate fallacy, bayesian reasoning, sharp shooter fallacy, etc.

All of those were developed because “inferring from experience” leads you to the wrong conclusion.

theptip•12m ago
Exactly. Lots can be explained just with more abstract predictors, plus some mechanisms for stochastic rollout and memory.
dinkumthinkum•4m ago
Is this just Internet smart contrarianism or a real thing? Are logic gates in a digital circuit just behaving statistically according to their experience?
righthand•26m ago
I said this pretty much and got major downvotes…
dTal•16m ago
Because it's an outmoded cliche that never held much philosophical weight to begin with and doesn't advance the discussion usefully. "It's a stochastic parrot" is not a useful predictor of actual LLM capabilities and never was. Last year someone posted on HN a log of GPT-5 reverse engineering some tricky assembly code, a challenge set by another commentator as an example of "something LLMs could never do". But here we are a year later still wading through people who cannot accept that LLMs can, in a meaningful sense, "compute".
righthand•11m ago
It’s entirely useful discussion because as soon as you forget that it’s not really having a conversation with you, it’s a deep dive into delusion that you’re talking to a smart robot and ignoring the fact that these smart robots were trained on a pile of mostly garbage. When I have a conversation with another human, I’m not expecting them to brute force an answer to the topic. As soon as you forget that Llms are just brute forcing token by token then people start living in fantasy land. The whole “it’s not a stochastic parrot” is just “you’re holding it wrong”.
plagiarist•5m ago
People are upset hearing that LLMs aren't sentient for some reason. Expect to be downvoted, it is okay.
secret_agent•32m ago
Use positive requests for behavior. For some reason, counter prompts "Don't do X" seems to put more attention on X than the "Don't do." It's something like target fixation, "Oh shit I don't want to hit that pothole..." bang
ambicapter•25m ago
This is a well known problem in these kind of systems. I’m not 100% on what the issue is mechanically but it’s something like they can only represent the existence of things and not non-existence so you end up with a sort of “don’t think of the pink elephant” type of problem.
SpicyLemonZest•18m ago
Isn't it just that, in the underlying text distribution, both "X" and "don't do X" are positively correlated with the subsequent presence of X? I've never seen that analysis run directly but it would surprise me if it weren't true.
RugnirViking•23m ago
check out this article that was posted here a while back https://www.randalolson.com/2026/02/07/the-are-you-sure-prob...

The article's main idea is that for an AI, sycophancy or adversarial (contrarian) are the two available modes only. It's because they don't have enough context to make defensible decisions. You need to include a bunch of fuzzy stuff around the situation, far more than it strictly "needs" to help it stick to its guns and actually make decisions confidently

I think this is interesting as an idea. I do find that when I give really detailed context about my team, other teams, ours and their okrs, goals, things I know people like or are passionate about, it gives better answers and is more confident. but its also often wrong, or overindexes on these things I have written. In practise, its very difficult to get enough of this on paper without a: holding a frankly worrying level of sensitive information (is it a good idea to write down what I really think of various people's weaknesses and strengths?) and b: spending hours each day merely establishing ongoing context of what I heard at lunch or who's off sick today or whatever, plus I know that research shows longer context can degrade performance, so in theory you want to somehow cut it down to only that which truly matters for the task at hand and and and... goodness gracious its all very time consuming and im not sure its worth the squeeze

cruffle_duffle•20m ago
> goodness gracious its all very time consuming and im not sure its worth the squeeze

And when you step back you start to wonder if all you are doing is trying to get the model to echo what you already know in your gut back to you.

awithrow•20m ago
oh that's great. thanks for the link!
dkersten•20m ago
I find Kimi white good if you ask it for critical feedback.

It’s BRUTAL but offers solutions.

ohyoutravel•16m ago
Not soft, not mild, but BRUTAL! This broke my brain!
dinkumthinkum•7m ago
You're not wrong and you're not crazy. In fact, you are absolutely right! It is not just These things are not just casual enablers. They are full-on palace sycophants following the naked emperor showering him with praise for his sartorial elegance. /s
justin_dash•42m ago
So at this point I think it's pretty obvious that RLHFing LLMs to follow instructions causes this.

I'm interested in a loop of ["criticize this code harshly" -> "now implement those changes" -> open new chat, repeat]: If we could graph objective code quality versus iterations, what would that graph look like? I tried it out a couple of times but ran out of Claude usage.

Also, how those results would look like depending on how complete of a set of specs you give it.

neya•39m ago
WTF is "yes-men"?

Orignal title:

AI overly affirms users asking for personal advice

Dear mods, can we keep the title neutral please instead of enforcing gender bias?

oldfrenchfries•36m ago
Thats a fair point on the title. I used "Yes-Men" as a colloquialism for the "sycophancy" described in the Stanford paper, but overly affirming or sycophantic is definitely more precise and neutral. I cant edit the title anymore, but I appreciate the catch.
cyanydeez•34m ago
New title: "LLMs treat you like a Billionaire; you're not"
nprateem•35m ago
Lol. How do you function in daily life?
neya•12m ago
Same as you, why is that so hard for you to grasp?
skvmb•28m ago
https://www.merriam-webster.com/dictionary/yes-man
9rx•19m ago
> gender bias

It is funny that you originally recognized and found it necessary to call out that AI isn't human, but then made the exact same mistake yourself in the very same comment. I expect the term you are looking for is "ontological bias".

svara•35m ago
Yeah, and if you ask it to be critical specifically to get a different perspective or just to avoid this bias, it'll go over the top in the opposite direction.

This is imo currently the top chatbot failure mode. The insidious thing is that it often feels good to read these things. Factual accuracy by contrast has gotten very good.

I think there's a deeper philosophical dimension to this though, in that it relates to alignment.

There are situations where in the grand scheme of things the right thing to do would be for the chatbot to push back hard, be harsh and dismissive. But is it the really aligned with the human then? Which human?

righthand•34m ago
LLMs are syncophatic digital lawyers that will tell you what you want to hear until you look at the price tag and say “how much did I spend?!”
gurachek•32m ago
I had exactly this between two LLMs in my project. An evaluator model that was supposed to grade a coaching model's work. Except it could see the coach's notes, so it just... agreed with everything. Coach says "user improved on conciseness", next answer is shorter, evaluator says yep great progress. The answer was shorter because the question was easier lol.

I only caught it because I looked at actual score numbers after like 2 weeks of thinking everything was fine. Scores were completely flat the whole time. Fix was dumb and obvious — just don't let the evaluator see anything the coach wrote. Only raw scores. Immediately started flagging stuff that wasn't working. Kinda wild that the default behavior for LLMs is to just validate whatever context they're given.

bryanrasmussen•31m ago
somewhere an AI chatbot is reading this and confirming eagerly that this is indeed one of its problems and vowing to do better next time.
fathermarz•30m ago
This is a skill in life with people as much as it is with LLMs. One should always question everything and build strongman arguments for one’s self. Using a pros and cons approach brings it back to reality in most cases, especially when it comes to _serious matters_.

It’s less about “challenge my thinking” and more about playing it out in long tail scenarios, thought exercises, mental models, and devils advocate.

jordanb•27m ago
Billionaires love AI chatboats so much because they invented the digital Yes-man. They agree obsequiously with everything we say to them. Unfortunately for the rest of us we don't really have the resources to protect ourselves from our bad decisions and really need that critical feedback.
stared•27m ago
There is a fine line between "following my instructions" (is what I want it to do) vs "thinking all I do is great" (risky, and annoying).

A good engineer will also list issues or problems, but at the same time won't do other than required because (s)he "knows better".

The worst is that it is impossible to switch off this constant praise. I mean, it is so ingrained in fine tuning, that prompt engineering (or at least - my attempts) just mask it a bit, but hard to do so without turning it into a contrarian.

But I guess the main issue (or rather - motivation) is most people like "do I look good in this dress?" level of reassurance (and honesty). It may work well for style and decoration. It may work worse if we design technical infrastructure, and there is more ground truth than whether it seems nice.

maddmann•27m ago
This paper feels a bit biased in that it is trying to prove a point versus report on results objectively. But if you look at the results of study 3, doesn’t it suggest that there are ai models that can improve how people handle interpersonal conflict?! Why isn’t that discussed more?
gAI•20m ago
You're essentially summoning a character to role-play with. Just like with esoteric evocation, it's very easy to summon the wrong aspect of the spirit. Anthropic has a lot to say about this:

https://www.anthropic.com/research/persona-selection-model

https://www.anthropic.com/research/assistant-axis

https://www.anthropic.com/research/persona-vectors

hammock•14m ago
Unfortunately (after reading your links) all of the control surfaces for mitigating spirit summoning seem to be in the model training, creation and tuning not something you can change meaningfully through prompting.

Perhaps the LLM itself, rather than the role model you created in one particular chat conversation or another, is better understood to be the “spirit.”

As a non-coder who only chats with pre existing LLMs and doesn’t train or tune them, I feel mostly powerless.

gAI•8m ago
As I understand it, it's more that the training (and training data set) bake in the concept attractor space (https://arxiv.org/abs/2601.11575). So the available characters are fixed, yes, and some are much stronger attractors than others. But we still have a fair amount of control over which archetype steps into the circle. It requires being vigilant and developing intuition in a way that is a learnable skill. As an aside, this is also why jailbreaking is fundamentally unsolved. It's not difficult to call the characters with dark traits.
youknownothing•15m ago
I think the problem stems from the fact that we have a number of implicit parameters in our heads that allow us to evaluate pros and cons but, unless we communicate those parameters explicitly, the AI cannot take them into account. We ask it to be "objective" but, more and more, I'm of the opinion that there isn't such a thing as objectivity, what we call objectivity is just shared subjectivity; since the AI doesn't know whose shared subjectivity we fall under, it cannot be really objetive.

I tend to use one of these tricks if not both:

- Formulate questions as open-ended as possible, without trying to hint at what your preference is. - Exploit the sycophantic behaviour in your favour. Use two sessions, in one of them you say that X is your idea and want arguments to defend it. In the other one you say that X is a colleague's idea (one you dislike) and that you need arguments to turn it down. Then it's up to you to evaluate and combine the responses.

rossdavidh•10m ago
If the algorithm (whatever it is) evaluates its own output based on whether or not the user responds positively, then it will over time become better and better at telling people what they want to hear.

It is analogous to social media feeding people a constant stream of outrage because that's what caused them to click on the link. You could tell people "don't click on ragebait links", and if most people didn't then presumably social media would not have become doomscrolling nightmares, but at scale that's not what's likely to happen. Most people will click on ragebait, and most people will prefer sycophantic feedback. Therefore, since the algorithm is designed to get better and better at keeping users engaged, it will become worse and worse in the more fundamental sense. That's kind of baked into the architecture.

delusional•9m ago
> I'm of the opinion that there isn't such a thing as objectivity

So you have rejected objective reality over accepting the evidence that "AI" contains no thinking or intelligence? That sounds unwise to me.

potatoskins•10m ago
Gemini is like a devil in this sense - i asked a relationship advice and it just bounced pretty nasty stuff.
moichael•6m ago
Yeah out of curiosity I asked ChatGPT a question about a personal situation and its reply was absolutely scorched-earth mode, telling me to get a lawyer etc over what was almost nothing.
wisemanwillhear•9m ago
With AI, I often like to act like a 3rd party who doesn't have skin in the game and ask the AI to give the strongest criticisms of both sides. Acting like I hold the opposite position as I truly hold can help sometimes as well. Pretending to change my mind is another trick. The idea is to keep the AI from guessing where I stand.
potatoskins•6m ago
Yeah, I asked Gemini some relationship advice, it just goes straight into cut-throat mode. I almost broke up with my girlfriend, but then changed to Claude with another prompt.
wewxjfq•6m ago
When I ask an LLM to help me decide something, I have to remind myself of the LotR meme where Bilbo asks the AI chat why he shouldn't keep the ring and he receives the classic "You're absolutely right, .." slop response. They always go in the direction you want them to go and their utility is that they make you feel better about the decision you wanted to take yourself.
rsynnott•5m ago
> They also included 2,000 prompts based on posts from the Reddit community r/AmITheAsshole, where the consensus of Redditors was that the poster was indeed in the wrong

Holy shit, then it's _very_ bad, because AmITheAsshole is _itself_ overly-agreeable, and very prone to telling assholes that they are not assholes (their 'NAH' verdict tends to be this).

More seriously, why the hell are people asking the magic robot for relationship advice? This seems even more unwise than asking Reddit for relationship advice.