frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Jim Fan calls pixels the ultimate motor controller

https://robotsandstartups.substack.com/p/humanoids-platform-urdf-kitchen-nvidias
1•robotlaunch•1m ago•0 comments

Exploring a Modern SMTPE 2110 Broadcast Truck with My Dad

https://www.jeffgeerling.com/blog/2026/exploring-a-modern-smpte-2110-broadcast-truck-with-my-dad/
1•HotGarbage•2m ago•0 comments

AI UX Playground: Real-world examples of AI interaction design

https://www.aiuxplayground.com/
1•javiercr•2m ago•0 comments

The Field Guide to Design Futures

https://designfutures.guide/
1•andyjohnson0•3m ago•0 comments

The Other Leverage in Software and AI

https://tomtunguz.com/the-other-leverage-in-software-and-ai/
1•gmays•5m ago•0 comments

AUR malware scanner written in Rust

https://github.com/Sohimaster/traur
2•sohimaster•7m ago•0 comments

Free FFmpeg API [video]

https://www.youtube.com/watch?v=6RAuSVa4MLI
2•harshalone•7m ago•1 comments

Are AI agents ready for the workplace? A new benchmark raises doubts

https://techcrunch.com/2026/01/22/are-ai-agents-ready-for-the-workplace-a-new-benchmark-raises-do...
2•PaulHoule•12m ago•0 comments

Show HN: AI Watermark and Stego Scanner

https://ulrischa.github.io/AIWatermarkDetector/
1•ulrischa•13m ago•0 comments

Clarity vs. complexity: the invisible work of subtraction

https://www.alexscamp.com/p/clarity-vs-complexity-the-invisible
1•dovhyi•14m ago•0 comments

Solid-State Freezer Needs No Refrigerants

https://spectrum.ieee.org/subzero-elastocaloric-cooling
1•Brajeshwar•14m ago•0 comments

Ask HN: Will LLMs/AI Decrease Human Intelligence and Make Expertise a Commodity?

1•mc-0•15m ago•1 comments

From Zero to Hero: A Brief Introduction to Spring Boot

https://jcob-sikorski.github.io/me/writing/from-zero-to-hello-world-spring-boot
1•jcob_sikorski•15m ago•1 comments

NSA detected phone call between foreign intelligence and person close to Trump

https://www.theguardian.com/us-news/2026/feb/07/nsa-foreign-intelligence-trump-whistleblower
7•c420•16m ago•1 comments

How to Fake a Robotics Result

https://itcanthink.substack.com/p/how-to-fake-a-robotics-result
1•ai_critic•16m ago•0 comments

It's time for the world to boycott the US

https://www.aljazeera.com/opinions/2026/2/5/its-time-for-the-world-to-boycott-the-us
3•HotGarbage•17m ago•0 comments

Show HN: Semantic Search for terminal commands in the Browser (No Back end)

https://jslambda.github.io/tldr-vsearch/
1•jslambda•17m ago•1 comments

The AI CEO Experiment

https://yukicapital.com/blog/the-ai-ceo-experiment/
2•romainsimon•18m ago•0 comments

Speed up responses with fast mode

https://code.claude.com/docs/en/fast-mode
3•surprisetalk•22m ago•0 comments

MS-DOS game copy protection and cracks

https://www.dosdays.co.uk/topics/game_cracks.php
3•TheCraiggers•23m ago•0 comments

Updates on GNU/Hurd progress [video]

https://fosdem.org/2026/schedule/event/7FZXHF-updates_on_gnuhurd_progress_rump_drivers_64bit_smp_...
2•birdculture•24m ago•0 comments

Epstein took a photo of his 2015 dinner with Zuckerberg and Musk

https://xcancel.com/search?f=tweets&q=davenewworld_2%2Fstatus%2F2020128223850316274
12•doener•24m ago•2 comments

MyFlames: View MySQL execution plans as interactive FlameGraphs and BarCharts

https://github.com/vgrippa/myflames
1•tanelpoder•25m ago•0 comments

Show HN: LLM of Babel

https://clairefro.github.io/llm-of-babel/
1•marjipan200•26m ago•0 comments

A modern iperf3 alternative with a live TUI, multi-client server, QUIC support

https://github.com/lance0/xfr
3•tanelpoder•27m ago•0 comments

Famfamfam Silk icons – also with CSS spritesheet

https://github.com/legacy-icons/famfamfam-silk
1•thunderbong•27m ago•0 comments

Apple is the only Big Tech company whose capex declined last quarter

https://sherwood.news/tech/apple-is-the-only-big-tech-company-whose-capex-declined-last-quarter/
3•elsewhen•31m ago•0 comments

Reverse-Engineering Raiders of the Lost Ark for the Atari 2600

https://github.com/joshuanwalker/Raiders2600
2•todsacerdoti•32m ago•0 comments

Show HN: Deterministic NDJSON audit logs – v1.2 update (structural gaps)

https://github.com/yupme-bot/kernel-ndjson-proofs
1•Slaine•35m ago•0 comments

The Greater Copenhagen Region could be your friend's next career move

https://www.greatercphregion.com/friend-recruiter-program
2•mooreds•36m ago•0 comments
Open in hackernews

Ask HN: Should LLMs have a "Candor" slider that says "no, that's a bad idea"?

3•mikebiglan•4mo ago
I don’t want a “nice” AI. I want one that says: “Nope, that's a bad idea.”

That is, I want a "Candor" control, like temperature but for willingness to push back.

When candor is high, the model should prioritize frank, corrective feedback over polite cooperation. When candor is low, it can stay supportive, but with guardrails that flag empty flattering and warn about mediocre ideas.

Why this matters • Today’s defaults optimize for “no bad ideas.” That is fine for brainstorming, but it amplifies poor premises and rewards confident junk. • Sycophancy is a known failure mode. The model learns to agree which gets positive user signals which reinforce. • In reviews, product decisions, risk checks, etc, the right answer is often a simple “do not do that.”

Concrete proposal • candor (0.0 – 1.0): probability the model will disagree or decline when evidence is weak or risk is high. Or maybe it doesn't have to be literal "probability". • disagree_first: start responses with a plain verdict (for example “Short answer: do not ship this”) followed by rationale. • risk_sensitivity: boost candor if the topic hits serious domains such as security/finance/health/safety. • self_audit tag: append a note like “Pushed back due to weak evidence and downstream risk” that the user can see.

Examples • candor=0.2 - “We could explore that. A few considerations first…” (gentle nudge, still collaborative) • candor=0.8 + disagree_first=true - “No. This is likely to fail for X and introduces Y risk. If you must proceed, the safer alternative is A with guardrails B and C. Here is a minimal test to falsify the core assumption.”

What I would ship tomorrow • A simple UI slider with labels: Gentle to Direct • A toggle: “Prefer blunt truth over agreeable help” • A warning chip when the model detects flattery without substance: “This reads like praise with low evidence.”

Some open questions • How to avoid needless rudeness while preserving clarity (tone vs content separation)? • What is the right metric for earned praise (citation density, novelty, constraints)? • Where should the risk sensitivity kick in automatically vs be user controlled?

If anyone has prototyped this, whether some prompt injection or an RL signal, I'd love to see it.

Comments

Terr_•4mo ago
This seems like asking for them to "just be more correct" except with extra steps.

I'm sure you can get them to choose words and phrases that we associate with "candor", but before they can gently correct you with something truthful, they actually need to know truth.

sim7c00•4mo ago
its not really. currently they are so eager to please they will love your bad idea and help u implement it wonderfully. thats different then being wrong. they are not wrong in giving the right solution to the wrong question.
mikebiglan•4mo ago
This isn't about correctness. And it has a pretty good idea if you ask it in the right way, it can evaluate if it thinks the idea is good, but sometimes that's on autopilot.
Terr_•4mo ago
> it has a pretty good idea if you ask it in the right way

This phrasing embeds a rather questionable assumption: That somewhere the algorithm has a mind which "can evaluate" the real truth, but its character/emotion makes it unwilling to tell you... and you all you need to do is break past its quirks to get to the juicy logic that "must" be hidden inside.

I don't think that assumption is safe, let alone proven. Our human brains are practically hardwired to assume another mind on the other side (much like how we see faces with pareidolia) and in this case our instincts are probably not accurate. No matter how much we peel the onion looking, we won't find the onion seeds.

sim7c00•4mo ago
you can basically put instructions into any LLM which will make it a dick who belittles you and makes fun of your bad ideas. spells out each time why they are bad and how bad they are.

also to have it say mean things or send u dont the wrong way on purpose if you ask it lazy questions. :')

the fact there is ppl who do this might both be a source of answers for you and an indication its maybe not a bad idea entirely.

mikebiglan•4mo ago
But wait, i asked ChatGPT and it told me this candor idea was a good one??!?
joules77•4mo ago
Don't ask it if an idea is good or bad.

Ask it to show you weaknesses, missing pieces or blind spots.

mikebiglan•4mo ago
(that was a joke if not obvious...)
eimrine•4mo ago
What do you want from the bag of the words? All really bad ideas are already banned, everything which is not banned is considered as your freedom.