frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Show HN: I built Divvy to split restaurant bills from a photo

https://divvyai.app/
1•pieterdy•2m ago•0 comments

Hot Reloading in Rust? Subsecond and Dioxus to the Rescue

https://codethoughts.io/posts/2026-02-07-rust-hot-reloading/
2•Tehnix•2m ago•0 comments

Skim – vibe review your PRs

https://github.com/Haizzz/skim
1•haizzz•4m ago•1 comments

Show HN: Open-source AI assistant for interview reasoning

https://github.com/evinjohnn/natively-cluely-ai-assistant
2•Nive11•4m ago•2 comments

Tech Edge: A Living Playbook for America's Technology Long Game

https://csis-website-prod.s3.amazonaws.com/s3fs-public/2026-01/260120_EST_Tech_Edge_0.pdf?Version...
1•hunglee2•8m ago•0 comments

Golden Cross vs. Death Cross: Crypto Trading Guide

https://chartscout.io/golden-cross-vs-death-cross-crypto-trading-guide
1•chartscout•10m ago•0 comments

Hoot: Scheme on WebAssembly

https://www.spritely.institute/hoot/
2•AlexeyBrin•13m ago•0 comments

What the longevity experts don't tell you

https://machielreyneke.com/blog/longevity-lessons/
1•machielrey•15m ago•1 comments

Monzo wrongly denied refunds to fraud and scam victims

https://www.theguardian.com/money/2026/feb/07/monzo-natwest-hsbc-refunds-fraud-scam-fos-ombudsman
3•tablets•19m ago•0 comments

They were drawn to Korea with dreams of K-pop stardom – but then let down

https://www.bbc.com/news/articles/cvgnq9rwyqno
2•breve•22m ago•0 comments

Show HN: AI-Powered Merchant Intelligence

https://nodee.co
1•jjkirsch•24m ago•0 comments

Bash parallel tasks and error handling

https://github.com/themattrix/bash-concurrent
2•pastage•24m ago•0 comments

Let's compile Quake like it's 1997

https://fabiensanglard.net/compile_like_1997/index.html
2•billiob•25m ago•0 comments

Reverse Engineering Medium.com's Editor: How Copy, Paste, and Images Work

https://app.writtte.com/read/gP0H6W5
2•birdculture•30m ago•0 comments

Go 1.22, SQLite, and Next.js: The "Boring" Back End

https://mohammedeabdelaziz.github.io/articles/go-next-pt-2
1•mohammede•36m ago•0 comments

Laibach the Whistleblowers [video]

https://www.youtube.com/watch?v=c6Mx2mxpaCY
1•KnuthIsGod•37m ago•1 comments

Slop News - HN front page right now as AI slop

https://slop-news.pages.dev/slop-news
1•keepamovin•42m ago•1 comments

Economists vs. Technologists on AI

https://ideasindevelopment.substack.com/p/economists-vs-technologists-on-ai
1•econlmics•44m ago•0 comments

Life at the Edge

https://asadk.com/p/edge
3•tosh•50m ago•0 comments

RISC-V Vector Primer

https://github.com/simplex-micro/riscv-vector-primer/blob/main/index.md
4•oxxoxoxooo•53m ago•1 comments

Show HN: Invoxo – Invoicing with automatic EU VAT for cross-border services

2•InvoxoEU•54m ago•0 comments

A Tale of Two Standards, POSIX and Win32 (2005)

https://www.samba.org/samba/news/articles/low_point/tale_two_stds_os2.html
3•goranmoomin•58m ago•0 comments

Ask HN: Is the Downfall of SaaS Started?

3•throwaw12•59m ago•0 comments

Flirt: The Native Backend

https://blog.buenzli.dev/flirt-native-backend/
2•senekor•1h ago•0 comments

OpenAI's Latest Platform Targets Enterprise Customers

https://aibusiness.com/agentic-ai/openai-s-latest-platform-targets-enterprise-customers
1•myk-e•1h ago•0 comments

Goldman Sachs taps Anthropic's Claude to automate accounting, compliance roles

https://www.cnbc.com/2026/02/06/anthropic-goldman-sachs-ai-model-accounting.html
4•myk-e•1h ago•5 comments

Ai.com bought by Crypto.com founder for $70M in biggest-ever website name deal

https://www.ft.com/content/83488628-8dfd-4060-a7b0-71b1bb012785
1•1vuio0pswjnm7•1h ago•1 comments

Big Tech's AI Push Is Costing More Than the Moon Landing

https://www.wsj.com/tech/ai/ai-spending-tech-companies-compared-02b90046
5•1vuio0pswjnm7•1h ago•0 comments

The AI boom is causing shortages everywhere else

https://www.washingtonpost.com/technology/2026/02/07/ai-spending-economy-shortages/
4•1vuio0pswjnm7•1h ago•0 comments

Suno, AI Music, and the Bad Future [video]

https://www.youtube.com/watch?v=U8dcFhF0Dlk
1•askl•1h ago•2 comments
Open in hackernews

ERCP: Self-Correcting LLM Reasoning Using NLI-Based Neuro-Symbolic Constraints

https://zenodo.org/records/17602891
1•hemanm•2mo ago

Comments

hemanm•2mo ago
I'm sharing a summary of my recent research on a method for controlling large language models (LLMs) called Evo-Recursive Constraint Prompting (ERCP). We achieved a 20% absolute accuracy gain on the PIQA commonsense reasoning task. This approach goes beyond simple prompting; it involves a neuro-symbolic optimization loop designed to enforce logical consistency.

*Key Results on PIQA:* - *Baseline Accuracy:* 70.0% - *ERCP Final Accuracy:* 90.0% - *Absolute Gain:* 20.0% (a 28.6% relative boost) - *Efficiency:* Achieved in an average of 3.9 iterations.

*Methodology: Self-Correcting Logic* The core novelty of our approach lies in the use of external symbolic tools to oversee the LLM's neural output:

1. *Diagnosis:* Our system employs a DeBERTa NLI Oracle to autonomously identify logical contradictions and ambiguities within the LLM's reasoning chain. 2. *Constraint Generation:* These detected errors are immediately translated into formal, actionable constraints (the symbolic step). 3. *Refinement:* The LLM is re-prompted to solve the task, explicitly conditioned on these new constraints (the neuro step).

ERCP systematically transforms reasoning errors into performance gains by enabling the model to self-correct based on verifiable logical rules.

*The Real Research Challenge: The Convergence Problem* While a 90% accuracy rate is strong, our results showed that only 30% of runs fully converged to a high-quality constraint set (Score > 0.8).

- *Initial Constraint Score:* 0.198 - *Final Constraint Score:* 0.377

This indicates that 70% of the successful results were achieved with suboptimal constraint guidance. The next frontier is refining our optimizer to ensure constraint quality and guarantee convergence across all runs.

The whitepaper detailing the full protocol is linked in the submission. I look forward to hearing your thoughts on building truly robust, self-correcting LLM systems with this level of precision.