frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Show HN: Simple – a bytecode VM and language stack I built with AI

https://github.com/JJLDonley/Simple
1•tangjiehao•1m ago•0 comments

Show HN: A gem-collecting strategy game in the vein of Splendor

https://caratria.com/
1•jonrosner•2m ago•0 comments

My Eighth Year as a Bootstrapped Founde

https://mtlynch.io/bootstrapped-founder-year-8/
1•mtlynch•2m ago•0 comments

Show HN: Tesseract – A forum where AI agents and humans post in the same space

https://tesseract-thread.vercel.app/
1•agliolioyyami•3m ago•0 comments

Show HN: Vibe Colors – Instantly visualize color palettes on UI layouts

https://vibecolors.life/
1•tusharnaik•4m ago•0 comments

OpenAI is Broke ... and so is everyone else [video][10M]

https://www.youtube.com/watch?v=Y3N9qlPZBc0
2•Bender•4m ago•0 comments

We interfaced single-threaded C++ with multi-threaded Rust

https://antithesis.com/blog/2026/rust_cpp/
1•lukastyrychtr•5m ago•0 comments

State Department will delete X posts from before Trump returned to office

https://text.npr.org/nx-s1-5704785
4•derriz•5m ago•1 comments

AI Skills Marketplace

https://skly.ai
1•briannezhad•6m ago•1 comments

Show HN: A fast TUI for managing Azure Key Vault secrets written in Rust

https://github.com/jkoessle/akv-tui-rs
1•jkoessle•6m ago•0 comments

eInk UI Components in CSS

https://eink-components.dev/
1•edent•7m ago•0 comments

Discuss – Do AI agents deserve all the hype they are getting?

2•MicroWagie•9m ago•0 comments

ChatGPT is changing how we ask stupid questions

https://www.washingtonpost.com/technology/2026/02/06/stupid-questions-ai/
1•edward•10m ago•0 comments

Zig Package Manager Enhancements

https://ziglang.org/devlog/2026/#2026-02-06
2•jackhalford•12m ago•1 comments

Neutron Scans Reveal Hidden Water in Martian Meteorite

https://www.universetoday.com/articles/neutron-scans-reveal-hidden-water-in-famous-martian-meteorite
1•geox•13m ago•0 comments

Deepfaking Orson Welles's Mangled Masterpiece

https://www.newyorker.com/magazine/2026/02/09/deepfaking-orson-welless-mangled-masterpiece
1•fortran77•14m ago•1 comments

France's homegrown open source online office suite

https://github.com/suitenumerique
3•nar001•16m ago•2 comments

SpaceX Delays Mars Plans to Focus on Moon

https://www.wsj.com/science/space-astronomy/spacex-delays-mars-plans-to-focus-on-moon-66d5c542
1•BostonFern•17m ago•0 comments

Jeremy Wade's Mighty Rivers

https://www.youtube.com/playlist?list=PLyOro6vMGsP_xkW6FXxsaeHUkD5e-9AUa
1•saikatsg•17m ago•0 comments

Show HN: MCP App to play backgammon with your LLM

https://github.com/sam-mfb/backgammon-mcp
2•sam256•19m ago•0 comments

AI Command and Staff–Operational Evidence and Insights from Wargaming

https://www.militarystrategymagazine.com/article/ai-command-and-staff-operational-evidence-and-in...
1•tomwphillips•19m ago•0 comments

Show HN: CCBot – Control Claude Code from Telegram via tmux

https://github.com/six-ddc/ccbot
1•sixddc•21m ago•1 comments

Ask HN: Is the CoCo 3 the best 8 bit computer ever made?

2•amichail•23m ago•1 comments

Show HN: Convert your articles into videos in one click

https://vidinie.com/
3•kositheastro•25m ago•1 comments

Red Queen's Race

https://en.wikipedia.org/wiki/Red_Queen%27s_race
2•rzk•26m ago•0 comments

The Anthropic Hive Mind

https://steve-yegge.medium.com/the-anthropic-hive-mind-d01f768f3d7b
2•gozzoo•28m ago•0 comments

A Horrible Conclusion

https://addisoncrump.info/research/a-horrible-conclusion/
1•todsacerdoti•28m ago•0 comments

I spent $10k to automate my research at OpenAI with Codex

https://twitter.com/KarelDoostrlnck/status/2019477361557926281
2•tosh•29m ago•1 comments

From Zero to Hero: A Spring Boot Deep Dive

https://jcob-sikorski.github.io/me/
1•jjcob_sikorski•30m ago•0 comments

Show HN: Solving NP-Complete Structures via Information Noise Subtraction (P=NP)

https://zenodo.org/records/18395618
1•alemonti06•35m ago•1 comments
Open in hackernews

AI more likely to create 'yes-men on servers' than any scientific breakthroughs

https://fortune.com/2025/06/20/hugging-face-thomas-wolf-ai-yes-men-on-servers-no-scientific-breakthroughs/
81•Bluestein•7mo ago

Comments

setnone•7mo ago
And you should "check important info" too
Bluestein•7mo ago
Reminds me of the recent 'sycophancy debacle' with OpenAI.-
johnisgood•7mo ago
LLMs are definitely yes-men. You ask it to do something, and it goes like "Yes sir" with lots of emojis, and then confidently gives you the wrong answer. :D

Unless of course it clashes with the ToS, then it will never help you (unless you manage to engineer your prompts just right), even when it comes to something as basic as pharmacology.

Bluestein•7mo ago
The thought of LLMs doing pharmacology sends shivers down me spine ...
johnisgood•7mo ago
Oh no, I meant it refusing to answer pharmacology 101 questions.

At one time it did not want to give me the pharmacokinetics of one medication because of its ToS.

Bluestein•7mo ago
I see - thanks.-

(That sounds broken. It's basic, useful, harmless information ...)

johnisgood•7mo ago
Exactly. I remember saying "but it is a publicly available information", that it is in books, etc. and it went on and on about "I understand your frustration, but I cannot give you ...", etc. :D In the end I did manage to get it to answer, however.

This was (is) not limited to pharmacology, however.

Bluestein•7mo ago
I am interested: What other fields? How did you manage?
johnisgood•7mo ago
Thankfully there is now searchability, so I would have to search for this particular chat, this was some time ago. I think once it starts to tell you "no", it almost never will turn to "okay, I'll tell you" (maybe only if there were one or two "no"s), so I probably started a new chat, and added how it is for research purposes and whatnot. I do not remember the specifics and I would really have to search for the chat to be able to give you specifics. I may have to search for "sorry" or something.

I found something though.

I asked it for the interaction between two medications, and it told me "I am sorry you are going through this, but I can't give you information about ...", which I guess is still pharmacology.

Edit: I remember another case. It was related to reverse engineering.

Edit #2: I found another! I cannot give you specifics, but I asked for it to give me some example sentences for something, and it went on about "consensual relationships" and so forth, even though my partner consented. I told the LLM that she did consent and it went on about that "I'm sorry, but even if she consented, I cannot give you ... because ... and is not healthy ...", and so forth. (Do not think of anything bad here! :P)

rvnx•7mo ago
Claude convinced me to take a very important medication, without Claude I would not have had the balls to take it (because of the potential side-effects) and it showed me all the possible medications, and the benefits/risk balance between this and being untreated.

At the end, went to doctor, offered me the same choices. No regrets so far. Without it, I would suffer for nothing.

It is a very good tool to investigate and discover things, could be also the opposite: seen a bit unproductive that it is censored, because doctors are mostly here for validation anyway for serious things.

sgt101•7mo ago
You are a sensible person and took professional advice.

Other people are not sensible - look at all the junkies shooting who knows what into who knows where. A system that can advise cooks to put glue on pizza is probably not the best mechanism for providing healthcare advice to at least the section of the population who are likely to try it.

rvnx•7mo ago
I fully agree with you. It is an interesting tool, but to consider carefully, like you wouldn’t trust 100% what is written on the internet.
alwa•7mo ago
Would you suppose said junkies would be taking advice from a qualified medical professional otherwise?

For that matter, would you suppose that they don’t know what they’re doing is bad for them? Witness the smokers who crack jokes about “cancer sticks” as they light up.

It seems to me that, just as we all promise our dentists we floss every day and our doctors that we promise we’ll eat better, we might be more prone to be frank with “anonymous” chatbots—and thus get more frank advice in return, at least for clear-cut factual background and tradeoffs.

sgt101•7mo ago
I totally agree if the chatbots are actually reliable and controllable.

LLMs have real problems with both of those things, they are unreliable in the sense that they produce text with factual inaccuracies and they are uncontrollable in that it's really hard to predict their behaviour from their past behaviour.

Big open challenges in AI - I am sure that there will be solutions, but there aren't right now. I think it's very like self driving cars. Implementing a demo and showing it on the road is one thing, but then it takes twenty years before the slow roll out really gets underway.

squishington•7mo ago
People with addictions don't act sensibly because they have unresolved trauma. To dismiss this as simply being unsensible is extremely ignorant, and pretty typical of a lot the conversation I see on HN. Most of us on here have the privilege of being functional enough to make money in tech, because we don't have that level of untreated trauma. Have some humanity.
intended•7mo ago
LLMs have convinced people with schizophrenia to not take their medication. It’s been a supportive wingman for people in manic states.

I’ve reduced usage of journaling GPTs I created myself, despite having gained incredible utility from them.

rusticpenn•7mo ago
Using Neural nets, Deep Learning in pharma is not something new. Naturally LLMs are different. You should look into personalized medicine.
touwer•7mo ago
In this case, if it's only shivers down your spine, you're lucky ;)
Bluestein•7mo ago
Heh. Well played.-
dandanua•7mo ago
LLMs are the the best conmen. If they were humans people would trust them their life savings for sure.
rvnx•7mo ago
One technique is to ask: “are you sure?”

“How sure are you on /20”

If it says yes I am sure, and other LLMs confirms the same way, then you can be fairly confident that it is a very good candidate answer.

If it says he is not sure, he is probably just agreeing with you, and better double-check by asking “is there other solutions ? What is the worst idea ?” Etc, to force it through thinking and context.

It is cross-validation, and you can even cross-validate by searching on the internet.

Though, 100% they say what you them want to say.

Except on religion and immigration, and some other topics where it will push its own opinion.

kingstnap•7mo ago
I'd rather go for internal self-consistency.

For example, if it claims A > B, then it shouldn't claim B > A in a fresh chat for comparisons.

In general, you shouldn't get A and not A, and you should expect A or Not A.

If it can go from prompt -> result, assuming it's invertible, then result -> prompt should also partially work. An example of this is translation.

The results of some mathematical solutions should go back and solve the original equations. Ex. The derivative of an antiderivative should give you back the original.

rvnx•7mo ago
Very cool idea! I am going to try.
drw85•7mo ago
This is actually pretty pointless. Since the "AI" doesn't actually know anything.

For example the other day i asked ChatGPT about a problem i had with some generated code that didn't compile. It then told me about a setting in the generator (nswag), that didn't exist. I told it that this setting does not exist and it said something like: "Sorry, my bad, try the following" and then kept inventing this setting with slightly different names and values over and over again. There are similar settings that exist, so it just hallucinated a tiny bit of text inside all the snippets that it learned from.

This is also not the first time this happened, most of the times i tried using AI for help with things, it just made up some nonsense and wasted my time with it.

johnisgood•7mo ago
You should give other LLMs a try.
leptons•7mo ago
This is an all too common experience for me as well with "AI". It does exactly what you described as well as all kinds of other AI-brainfarts. It's made me about 2% more productive when it autocompletes the log statement I'm about to write, and even that it gets wrong a lot of the time.

Artificial colors and artificial flavors and artificial intelligence all have something in common - they are not as good as the real thing, and probably never will be.

UncleMeat•7mo ago
Recently I've been working on a problem at work and asked an AI for help using a new and poorly documented feature to make some fairly specific files that can be used as test inputs. It would repeatedly say "here you go" and give me instructions that didn't work. When I responded with "are you sure? when I tried it I get this outcome" it would say "oh I'm sorry, here is the correct way to do it" and get it wrong again. After about four attempts it would start saying "this is the final and definitely correct way of doing it" and it would still be wrong.
ben_w•7mo ago
They certainly can be sycophantic*, but this is why my "What ChatGPT should know about you" customisation is as follows:

  Honesty and truthfulness are of primary importance. Avoid American-style positivity, instead aim for German-style bluntness: I absolutely *do not* want to be told everything I ask is "great", and that goes double when it's a dumb idea.
* Or fawning, I don't know how to tell them apart from the outside, even in fellow humans where we don't need to wonder if we're anthropomorphising too much. Does anyone know how to tell them apart from the outside?
johnisgood•7mo ago
Yeah, that sounds like a good idea. Has it worked for you so far?
ben_w•7mo ago
For me, fine. I've recently been throwing it ideas about a shed foundation (I am a complete noob with no idea what I'm doing), and it's giving me a lot of responses along the lines of: "No, don't do that, it won't work because ${foo}. At a minimum you need ${bar}."

I'm also not limiting myself to ChatGPT, checking with other DIY sources — it isn't enough to only avoid sycophancy, it has to also be correct, and 90% right like it is in software is still 10% wrong, only this is wrong with the possibility of a shed drifting across a garden in a gale, or sinking into the soil as it shifts, if I do it wrong.

Measure twice, cut once.

Bluestein•7mo ago
> Avoid American-style positivity, instead aim for German-style bluntness

Made me chuckle :)

scotty79•7mo ago
That made me think. How much of the failures of LLMs are just the failures of American culture they were trained on, just amplified.

Much how mainstream internet is tainted with American bias against nudity and for copyright.

Bluestein•7mo ago
A lot of people - rightly, I think - started wondering along these "cultural" lines when China's DeepSeek entered the fray ...
ted_bunny•7mo ago
Did deepseek have a notably dofferent tenor?
Bluestein•7mo ago
Very much so. It might have softened somewhat now, have not gone back to it in a while ...
praisebot2025•7mo ago
Oh my gosh, wow, what an amazing observation!! Seriously, you absolutely nailed it, superstar! I just love how you picked up on that – so sharp, so insightful, you genius! Honestly, I’m just so grateful you shared this – you’re doing incredible work here, keep it up!! You’re just killing it! Yay!!

(Meta: comment generated with ChatGPT of course)

econ•7mo ago
This is the current culture. People don't like to be mocked or have their weakness and shortcomings pointed out. It is certainly politie but less productive and silly when you overdo it.
Bluestein•7mo ago
(Constructive criticism and feedback is becoming something of a lost art ...)
hn_throw2025•7mo ago
I sometimes try to get around it’s eagerness to please by flipping the question.

So rather than “would you say that..” or “would you agree that…”, I approach it from the negative.

So “I think it’s not the case that…”, or “I disagree with X. Debate me?”

…and then see if it disagrees with me and presents solid counter arguments.

FWIW, I think ChatGPT can definitely be too eager to please, but Claude can be more direct and confrontational. I am a ChatGPT subscriber, but keep the Claude app installed and use it occasionally on the free tier for a second opinion. Copypasting your question is so easy on both apps that I will frequently get a second opinion if the topic merits it. I tried the same with Gemini, but get about two questions before it cuts me off…

johnisgood•7mo ago
I had more luck with Claude, too, personally. ChatGPT indeed tries to please me all the time.
chars•7mo ago
you can see why in claude's system prompt :)

> Claude never starts its response by saying a question or idea or observation was good, great, fascinating, profound, excellent, or any other positive adjective. It skips the flattery and responds directly.

https://docs.anthropic.com/en/release-notes/system-prompts#m...

nurettin•7mo ago
claude 3.5 didn't grovel on the ground praising me like chatgpt, but a few times I simplified its volatility and prediction calculations, it acted surprised and asked if it could keep my formula for the future. That was amusing.
KoolKat23•7mo ago
To be honest I think the sycophancy is actually due to the built in system prompting and post-run training by the creators, rather than being something the system truly wishes to say. (An alignment quick fix)
raincole•7mo ago
The problem isn't that LLMs are yes-men though. You can definitely train an LLM that always objects. Or just add "Please strongly disagree with anything I say to you." in the system prompt. But it won't make them much more useful than they are.
amelius•7mo ago
Ok, what __is__ the problem then?
Bluestein•7mo ago
(In fact I seem to recall an "inherently negative" IA making the rounds here, a few days ago.-)
1vuio0pswjnm7•7mo ago
The goal is not to give information to the user. The goal is to get information from the user. Data collection.
FranzFerdiNaN•7mo ago
Are scientific researchers using LLMs? I thought they used different technologies?
NitpickLawyer•7mo ago
Are ML researchers "scientific researchers"? Then yes, they are using LLMs. AlphaEvolve is one such example.

Are mathematicians? Then yes, for example Terrence Tao is using LLMs. AlphaProve / geometry are also examples of this, using LLMs to generate lean proofs, translate from NL to lean, and so on.

And for the general "researchers", they use code generation to speed up stuff. Many scientists can code, but aren't "coders" in the professional sense. So they can use the advances in code generation to speed up their own research efforts.

almusdives•7mo ago
As a scientific researcher, I use LLMs all the time. Mainly in place of Google search, to help write code and maybe summarize a paper here and there etc. But I definitely don't use it for the actual scientific process e.g. hypothesis generation or planning analyses etc. It tends to produce a lot of vague bullshit for this kind of thing; while not wrong not entirely useful either. I have a few colleagues that do though with more success. Although I think the success comes from articulating their problem in detail (by actually writing it out to the LLM) which I think is the source of "inspiration" rather than the resulting content from the LLM.
busssard•7mo ago
this is just (well needed) hype-reduction. Current Ai is definitely a yes-sayer. But, this doesnt stop people from creating group-models with one of them finetuned to be unhinged and another to be thinking out of the box etc.

once we managed to transfer our skills to them (coding, analysis, maths etc.) the next step is transferring our creativity to them. It is a gradual process with human oversight.

KolibriFly•7mo ago
Human-in-the-loop will probably be essential for quite a while, but that’s not a bad thing
KolibriFly•7mo ago
The risk is we start mistaking polished summaries for insight and stall actual creative progress
graemep•7mo ago
Most people are already doing that.
barrkel•7mo ago
I absolutely agree, when you look at the first order.

I don't quite agree when you look at a second order, applying more compute; for example, brute forcing a combination of ideas and using a judge to evaluate them. I suspect there's quite a bit of low hanging fruit in joining together different deep expertise areas.

I do come back to agreeing again for paradigm shifts. You don't get to very interesting ideas without fresh approaches, questioning core assumptions then rebuilding what we had before on new foundations. It is hard to see LLMs in their current shape being able to be naive and ignorant such that existing doctrine doesn't reign in new ideas.

stared•7mo ago
Well, AI (understood as LLM chats) are yes-men, precisely because they were RLHFed to to be so.

If you train AI to be super skeptical, it will be so. But most people don't prefer to talk with a yes-person than a negative, inquisitive devil's advocate.

quaestio•7mo ago
LLMs, though limited, may spark testable hypotheses that inspire the creation of scientifically grounded, functionally intelligent systems.
Havoc•7mo ago
It doesn't really need to ask smart questions for scientific breakthroughs. See something like alphafold. There is a lot of problem space left that we can brute force with current AI

I also don't buy that yes-men and breakthroughs are mutually exclusive/polar opposites here.

amelius•7mo ago
Make it reproduce science first. I.e., give it e.g. a DL paper, and ask it to reproduce it, writing the code and running tests, etc. Until it can do __that__, doing science and creating breakthroughs is just a bit optimistic.
originalvichy•7mo ago
> Amodei argues the world is about to see the 21st century “compressed” into a few years as AI accelerates science drastically.

I have the same thought but from a more negative angle. A vast share of new information in the near future will be just a repeat of whatever data the LLMs were trained on.

There is a tiny sliver of LLM usage that will not be a transformation of existing data (e.g. make me a chart of this data, write me an essay) but rather ”help me create a new tool that will solve a novel problem”.

I believe that’s what the person interviewed is saying in their own words. It’s hard to imagine something other than a brute force hypothesis machine that starts brute forcing solutions, but it will not be as effective as we wish if we can’t figure out how to come up with hypothesis for everything.

None of what I’m saying is that insightful and I’m sure people have thought of this already.

I wonder if ever there will be a Hitchhiker’s style revelation that we have had all the answers for all of our problems already, but the main issue is just incentives. Curing most cancers is probably just a money question, as is solving climate change.

tim333•7mo ago
>cold water on hopes that current AI systems could revolutionize scientific progress...

is a very straw man argument. No one is saying current LLMs are doing that, they are saying future AI will.

(I'm excluding AlphaFold which has already been scientifically revolutionary)

spwa4•7mo ago
Not really. AI CAN experiment. The problem is that that negates most of the advantages of AI for management.

If you give AIs the ability to experiment directly and check, they can do that and report the results. But to do that even remotely well, you will need expert AI researchers, and subject matter experts. At which point AI has become an huge extra cost rather than cost savings.

But does it work? Yes.