frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

An educational fork of OpenClaw in which ALL safety guardrails have been removed

https://github.com/kenm47/yoloclaw
2•mooreds•1m ago•0 comments

How One Rock Poisoned (Almost) The Entire Planet

https://www.youtube.com/watch?v=cMx139eTxoc
1•lnenad•3m ago•0 comments

Speed Is the Moat: Inference Performance on AMD GPUs

https://www.amd.com/en/developer/resources/technical-articles/2026/inference-performance-on-amd-g...
1•latchkey•5m ago•0 comments

The Future of VSP Scrolling on the C64

https://kodiak64.co.uk/blog/future-of-VSP-scrolling
1•amichail•5m ago•0 comments

Show HN: Writing a C++20M:N Scheduler from Scratch (EBR, Work-Stealing)

https://github.com/lixiasky-back/tiny_coro-build_your_own_MN_scheduler
1•lixiasky•5m ago•0 comments

Yellow Journalism

https://en.wikipedia.org/wiki/Yellow_journalism
2•qwfqwef•8m ago•0 comments

You Want It Darker?

https://cinemasojourns.com/2026/02/17/you-want-it-darker/
2•jjgreen•8m ago•0 comments

Add schedule pause periods to triggers page

1•nishiohiroshi•8m ago•0 comments

The mysterious symptom popping up in some GLP-1 users

https://www.vox.com/explain-it-to-me/479202/glp-1-flatness-apathy-symptom
1•Hooke•9m ago•0 comments

Show HN: Air – Open-source black box for AI agents (tamper-evident audit trails)

https://github.com/nostalgicskinco/air-blackbox-gateway
2•shotwellj•9m ago•1 comments

LayerV – Senior Software Engineer – Remote (US) – $125k–150k equity and benefits

https://layerv.ai/
1•joeollis•9m ago•0 comments

pg_background: Make Postgres do the long work (while your session stays light)

https://vibhorkumar.wordpress.com/2026/02/16/pg_background-make-postgres-do-the-long-work-while-y...
2•tanelpoder•11m ago•0 comments

Skunk mating season becoming a headache for Bay Area residents

https://www.sfgate.com/bayarea/article/skunks-spray-valentine-21350189.php
2•turtlegrids•12m ago•0 comments

The Great Reboot

1•security1011015•13m ago•0 comments

Show HN: Privatiser – Redact secrets, IPs, and PII before sharing with AI

https://privatiser.net
1•r0otie•14m ago•0 comments

Local memory for any LLM agent

https://github.com/jmuncor/mumpu
1•jmuncor•14m ago•1 comments

Gravity Basins (2024) [video]

https://www.youtube.com/watch?v=LavXSS5Xtbg
1•dgellow•15m ago•0 comments

Magic Words Need Measuring Sticks

https://jotter.jonathankingston.co.uk/blog/2026/02/17/magic-words-need-measuring-sticks/
1•kingstonTime•16m ago•1 comments

Godot veteran says 'AI slop' pull requests have become overwhelming

https://www.gamedeveloper.com/programming/godot-co-founder-says-ai-slop-pull-requests-have-become...
3•haunter•19m ago•0 comments

I Use Obsidian

https://stephango.com/vault
1•hisamafahri•21m ago•0 comments

Ask HN: Are compiler errors for unused code necessary?

2•qwool•22m ago•2 comments

Memories Family

https://familymemories.video
1•tareq_•23m ago•3 comments

Book a Meeting with a YC Founder

https://y-cal.vercel.app/
1•abrarmurad416•27m ago•0 comments

Ask HN: Can AI replace apps, or will economics keep the app market alive?

1•maccraft•27m ago•1 comments

Show HN: Preference-aware routing for OpenClaw via Plano

https://github.com/katanemo/plano/tree/main/demos/llm_routing/openclaw_routing
1•sparacha•31m ago•0 comments

The Servo project and its impact on the web platform ecosystem

https://servo.org/slides/2026-02-fosdem-servo-web-platform/
2•mmphosis•31m ago•0 comments

Mira: An agent that never forgets anything. Persistent, shared memory

https://www.co-span.com/
2•dvt•33m ago•0 comments

Python HTTP server using Erlang and BEAM

https://hornbeam.dev/
1•polyrand•33m ago•0 comments

Dual nationals face scramble for UK passports as new rules come into force

https://www.bbc.com/news/articles/cx2d9yk2kpjo
5•tartoran•34m ago•0 comments

GraphQLite: SQLite graph extension supporting Cypher

https://colliery-io.github.io/graphqlite/latest/
2•dude01•35m ago•0 comments
Open in hackernews

Show HN: I built a thinking framework for Claude

https://bengiaventures.github.io/effective-thinking-skill/
1•bengia•1h ago
I built an open-source Claude Code skill called /think that applies a structured 5-element analysis framework (ground in facts, stress-test for failure, reframe the question, trace implications, audit your own reasoning) before synthesizing a recommendation. The obvious question: does it actually produce better output than just asking Claude directly? To test this, I ran blind A/B comparisons. Two isolated Claude Opus 4.6 agents get the same question — one runs /think, one responds naturally. Both responses are anonymized (framework markers stripped, sections retitled by content) and presented blind. The test covers 5 topics any professional would recognize: scaling a team post-fundraise, build vs buy decisions, when to pivot a product, SaaS pricing strategy, and the remote/hybrid/office debate. An AI judge scored /think winning all 5 pairs. But AI judging AI is circular — which is why the blind test is live for humans to judge. What I found so far (~21 comparisons across calibration + blind tests):

/think wins ~69% of comparisons overall Risk coverage is the clearest advantage (17-2 across all tests) — it consistently surfaces failure modes the organic response misses Decision impact is nearly even — organic Claude is often more actionable for practical problems Novel insight is mostly a wash — both find similar core insights, just different ones No decisive gaps in either direction. The advantage is depth and rigor, not dramatic superiority

Honest limitations:

All judges so far are AI. The whole point of publishing the blind test is to get human validation. ~21 comparisons is a pattern, not statistical significance Anonymization isn't perfect — /think responses have stylistic tells (confidence assessments, "what would change this conclusion" sections) The framework costs significantly more tokens

The skill itself is a recursive learning agent — it persists what it learns to a .think/ directory and loads that context in future sessions. Over time it builds project-specific knowledge. It also used its own framework to diagnose and fix its own weaknesses after the first round of testing. Everything is open source: https://github.com/bengiaventures/effective-thinking-skill I'd genuinely like to know if the blind test matches what the AI judges found, or if humans see something different. Takes about 15 minutes.