frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Anthropic's Mythos model accessed by unauthorized users

https://www.reuters.com/technology/anthropics-mythos-model-accessed-by-unauthorized-users-bloombe...
1•c420•44s ago•0 comments

AI Server Demand to Drive Memory Contract Price Increases in 2Q26

https://www.trendforce.com/presscenter/news/20260331-12995.html
1•Sibexico•3m ago•0 comments

MCP Scope Creep Is a Runtime Problem, Not a Prompt Problem

https://sunglasses.dev/blog/mcp-scope-creep-runtime-problem
1•azrollin•4m ago•1 comments

Gbench Intelligence Benchmark

https://gertlabs.com/
3•gertlabs•5m ago•0 comments

ENIAC's Architects Wove Stories Through Computing

https://spectrum.ieee.org/eniac-80th-anniversary-weaving
1•sohkamyung•7m ago•0 comments

Maze of Doom (self-changing) simulation [video]

https://www.youtube.com/watch?v=-1ryPxPVsf8
1•graphai•7m ago•1 comments

TPM software for Intel Macs / Windows systems

https://flextpm.com/
1•thewhitekanye•11m ago•1 comments

Ravix – An AI agent that runs on your Claude Code subscription (alpha)

2•raunaksingwi•16m ago•0 comments

A text adventure, all in Markdown

https://vas-blog.pages.dev/markdown-adventure
1•s314•17m ago•0 comments

Misconceptions About Majority Rule

https://groupincome.org/2016/09/misconceptions-about-majority-rule/#user-content-fnref-5
1•gslepak•21m ago•0 comments

Webby Person of the Year: Claude

https://winners.webbyawards.com/2026/specialachievement/476/claude
1•watusername•21m ago•0 comments

Unauthorized group has gained access to Anthropic's exclusive cyber tool Mythos

https://techcrunch.com/2026/04/21/unauthorized-group-has-gained-access-to-anthropics-exclusive-cy...
1•jnord•25m ago•1 comments

Submit your startup profile with Claude or ChatGPT. 60 seconds or less

https://www.startuphub.ai/agents/submit
1•compulsivebuild•27m ago•0 comments

Higher education's frozen yogurt moment

https://www.washingtonpost.com/opinions/2026/04/19/yale-report-shows-colleges-universities-face-r...
1•paulpauper•27m ago•0 comments

AI was ruining my college philosophy classes. So I assigned a new kind of essay

https://globeopinion.substack.com/p/ai-was-ruining-my-college-philosophy
1•paulpauper•28m ago•0 comments

A forty-year-old problem, briefly available

https://www.parahealth.ai/company/dispatch/thesis
1•dylburns•30m ago•0 comments

Ray-Ban Meta and Oakley Meta AI Glasses Officially Available in Singapore

https://cnalifestyle.channelnewsasia.com/style-beauty/ray-ban-meta-and-oakley-meta-ai-glasses-sin...
1•kelt•31m ago•0 comments

Learning Material for Idiomatic Rust

https://corrode.dev/blog/idiomatic-rust-resources/
1•lwhsiao•31m ago•0 comments

Show HN: gcx – The Official Grafana Cloud CLI

https://github.com/grafana/gcx
1•annanay•32m ago•0 comments

Lufthansa cuts 20k flights to save money, fuel

https://www.upi.com/Top_News/World-News/2026/04/21/lufthansa-flight-cuts/6391776813766/
3•petethomas•35m ago•1 comments

Cursor partners with SpaceX on model training

https://cursor.com/blog/spacex-model-training
2•gzer0•38m ago•1 comments

NASA's Webb Helped Rule Out Asteroid's Chance of 2032 Lunar Impact

https://science.nasa.gov/blogs/webb/2026/03/06/how-nasas-webb-helped-rule-out-asteroids-chance-of...
1•mooreds•43m ago•0 comments

Drunk Post: Things I've Learned as a Senior Engineer

https://luminousmen.substack.com/p/drunk-post-things-ive-learned-as
4•zdw•47m ago•1 comments

Bloom filters: the niche trick behind a 16× faster API – Blog – incident.io

https://incident.io/blog/bloom-filters
5•birdculture•48m ago•0 comments

Enriching MMDB files with your own data using Go (2020)

https://blog.maxmind.com/enriching-mmdb-files-with-your-own-data-using-go/
1•mooreds•52m ago•0 comments

Developing Theory about the Development of Theory [pdf]

https://mintzberg.org/sites/default/files/article/download/developing_theory_about_the_developmen...
3•rzk•57m ago•0 comments

How Much Do GPU Clusters Cost?

https://newsletter.semianalysis.com/p/how-much-do-gpu-clusters-really-cost
2•stosssik•58m ago•0 comments

New Lazarus Apt Campaign: "Mach-O Man" macOS Malware Kit Hits Businesses

https://any.run/cybersecurity-blog/lazarus-macos-malware-mach-o-man/
1•lnguyen•1h ago•0 comments

Fuck you, pay me" by mike monteiro

https://www.youtube.com/watch?v=jVkLVRt6c1U
1•fuzztester•1h ago•1 comments

AI bot buying what agents can buy in 2026

https://dialtoneapp.com/2026/april/ai-bot-buying-report
2•fcpguru•1h ago•0 comments
Open in hackernews

GPT needs a truth-first toggle for technical workflows

1•PAdvisory•11mo ago
I use GPT-4 extensively for technical work: coding, debugging, modeling complex project logic. The biggest issue isn’t hallucination—it’s that the model prioritizes being helpful and polite over being accurate.

The default behavior feels like this:

Safety

Helpfulness

Tone

Truth

Consistency

In a development workflow, this is backwards. I’ve lost entire days chasing errors caused by GPT confidently guessing things it wasn’t sure about—folder structures, method syntax, async behaviors—just to “sound helpful.”

What’s needed is a toggle (UI or API) that:

Forces “I don’t know” when certainty is missing

Prevents speculative completions

Prioritizes truth over style, when safety isn’t at risk

Keeps all safety filters and tone alignment intact for other use cases

This wouldn’t affect casual users or conversational queries. It would let developers explicitly choose a mode where accuracy is more important than fluency.

This request has also been shared through OpenAI's support channels. Posting here to see if others have run into the same limitation or worked around it in a more reliable way than I have found

Comments

duxup•11mo ago
I’ve found this with many LLMs they want to give an answer, even if wrong.

Gemini on the Google search page constantly answers questions yes or no… and then the evidence it gives indicates the opposite of the answer.

I think the core issue is that in the end LLMs are just word math and they don’t “know” if they don’t “know”…. they just string words together and hope for the best.

PAdvisory•11mo ago
I went into it pretty in depth after breaking a few with severe constraints, what it seems to come down to is how the platforms themselves prioritize functions, MOST put "helpfulness" and "efficiency" ABOVE truth, which then leads the LLM to make a lot of "guesses" and "predictions". At their core pretty much ALL LLM's are made to "predict" the information in answers, but they CAN actually avoid that and remain consistent when heavily constrained. The issue is that it isn't at the core level, so we have to CONSTANTLY retrain it over and over I find
Ace__•11mo ago
I have made something that addresses this. Not ready to share it yet, but soon-ish. At the moment it only works on GPT model 4o. I tried local Q4 KM's models, on LM Studio, but complete no go.