frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

What happens when capability decouples from credentials?

2•falsework•1h ago
Over the past 18 months, I've been collaborating with AI to build technical systems and conduct analytical work far outside my formal training. No CS degree, no background in the domains I'm working in, no institutional affiliation.

The work is rigorous. Someone with serious credentials has engaged and asked substantive questions. The systems function as designed. But I can't point to the traditional markers that would establish legitimacy—degrees, publications, years of experience in the field.

This isn't about whether AI "did the work." I made every decision, evaluated every output, iterated through hundreds of refinements. The AI was a tool that compressed what would have taken years of formal education into months of intensive, directed learning and execution.

Here's what interests me: We're entering a period where traditional signals of competence—credentials, institutional validation, experience markers—no longer reliably predict capability. Someone can now build sophisticated systems, conduct rigorous analysis, and produce novel insights without any of the credentials that historically signaled those abilities. The gap between "can do" and "should be trusted to do" is widening rapidly.

The old gatekeeping mechanisms are breaking down faster than new ones are forming. When credentials stop being reliable indicators of competence, what replaces them? How do we collectively establish legitimacy for knowledge and capability?

This isn't just theoretical—it's happening right now, at scale. Every day, more people are building things and doing work they have no formal qualification to do. And some of that work is genuinely good.

What frameworks should we use to evaluate competence when the traditional signals are becoming obsolete? How do we establish new language around expertise when terms like "expert," "rigorous," and "qualified" have been so diluted they've lost discriminatory power?

Comments

thenaturalist•1h ago
Adverserial work (be it agent or human).

The one difference between "can do" and "should be trusted to do" is the ability to systematically prove that "can do" holds up close to 100% of task instances and under adverserial conditions.

Hacking and pentesting are already scaling fully autonomously - and systematically.

For now, lower level targets aren't yet attractive as such scale requires sophisticated (state) actors, but that is going to change.

So building systems that white-hat prove your code is not only functional but competent are going to be critical not to be ripped apart by black-hat later on.

One nice example that applies this quite nicely is roborev [0] by the legendary Wes McKinney.

0: https://github.com/roborev-dev/roborev

falsework•1h ago
This is a good point. You're right that adversarial testing provides one form of validation that doesn't depend on credentials if the system holds up under systematic attack, that's evidence of competence regardless of who built it.

But I think there's a distinction worth making between technical robustness (does the code have vulnerabilities?) and epistemic legitimacy (should we trust the analysis/conclusions?).

Pentesting and formal verification can tell us whether a system is secure or functions correctly. That's increasingly automatable and credential-independent because the code either survives adversarial conditions or it doesn't.

But what about domains where validation is murkier? Cross-domain analysis, research synthesis, strategic thinking, design decisions? These require judgment calls where "correct" isn't binary. The work can be rigorous and well-reasoned without being formally provable.

The roborev example is interesting because code review is somewhat amenable to systematic validation. But we're also seeing AI collaboration extend into domains where adversarial testing isn't cleanly applicable—policy analysis, theoretical frameworks, creative work with analytical components.

I wonder if we need different validation frameworks for different types of work. Technical systems: adversarial testing and formal verification. Analytical/intellectual work: something else entirely. But what?

The deeper question: when the barrier to producing superficially plausible work drops to near-zero, how do we distinguish genuinely rigorous thinking from sophisticated-sounding nonsense? Credentials were a (flawed) heuristic for that. What replaces them in domains where adversarial testing doesn't apply?

One Year of Work for Ten Seconds of Film [video]

https://www.youtube.com/watch?v=iq5JaG53dho
1•susam•1m ago•0 comments

Joseph Gordon-Levitt Gets Section 230 Completely Backwards

https://www.techdirt.com/2026/02/12/joseph-gordon-levitt-goes-to-washington-dc-gets-section-230-c...
1•HotGarbage•1m ago•0 comments

The Automated Soundboard for Streamers

https://killervibe.app
1•Jikouken•2m ago•0 comments

Mechanisms and control of spin interactions in molecular-scale spintronics(2025)

https://www.cell.com/newton/fulltext/S2950-6360(25)00162-8
1•rolph•3m ago•0 comments

Astronomers observe a star that quietly transformed into a black hole

https://www.reuters.com/science/astronomers-observe-star-that-quietly-transformed-into-black-hole...
1•1659447091•8m ago•0 comments

Robust ways to extract bank statements from PDF to CSV beyond raw LLMs?

https://exactstatement.com/
1•alexfefun1•9m ago•1 comments

Ask HN: What makes an AI agent framework production-ready vs. a toy?

1•winclaw-dev•10m ago•0 comments

Everybody Is a CEO Now (and What Am I Doing Here?)

https://www.behind-the-enemy-lines.com/2026/02/everybody-is-ceo-now-and-what-exactly.html
1•ziyao_w•15m ago•0 comments

TiDB Cloud Zero – full-featured database with one line of curl

https://zero.tidbcloud.com/
1•liydu•17m ago•0 comments

The Clash of Civilizationalisms

https://www.theideasletter.org/essay/the-clash-of-civilizationalisms/
2•thunderbong•17m ago•0 comments

Show HN: Open-source MCP server that lets AI assistants shop via Google's UCP

https://github.com/nguthrie/ucp-mcp-server
1•nguthrie•19m ago•0 comments

Show HN: WebExplorer – a tool for preview file in browser

https://www.webexplorer.app
2•feblr•22m ago•0 comments

Electronic Structure: Electron Spin: Videos and Practice Problems

https://www.pearson.com/channels/gob/learn/jules/ch-2-atoms-and-the-periodic-table/electronic-str...
1•rolph•24m ago•0 comments

What Makes Oxygen Special?

https://www.quanxr.org/elctronspins
1•rolph•26m ago•0 comments

Not all computer code protected as speech, US court finds in ghost gun case

https://www.reuters.com/legal/government/not-all-computer-code-protected-speech-us-appeals-court-...
4•1659447091•27m ago•0 comments

Building a Modular Python Application with apywire and starlette

https://alganet.github.io/blog/2026-02-12-22-Building-a-Modular-Application-with-apywire-and-star...
1•gaigalas•28m ago•0 comments

A Python terminal deep-space receiver

https://github.com/luisub/6EQUJ5
1•max_pearl•30m ago•0 comments

YouTube Launches on Apple Vision Pro

https://www.macrumors.com/2026/02/12/youtube-app-apple-vision-pro/
2•mgh2•31m ago•1 comments

Why Couples Fight in the Kitchen (A Furniture Problem, Not a Marriage Problem)

https://oedmethod.substack.com/p/why-couples-fight-in-the-kitchen
3•truenfel•31m ago•0 comments

Why have far-forward nominal Treasury rates increased so much in past few years?

https://www.federalreserve.gov/econres/notes/feds-notes/why-have-far-forward-nominal-treasury-rat...
1•toomuchtodo•35m ago•2 comments

Claude Code bug forces users to restart chat, wasting tokens

https://old.reddit.com/r/claude/comments/1o1csrq/api_error_400_due_to_tool_use_concurrency_issues/
4•behnamoh•35m ago•2 comments

NASA loading liquid hydrogen aboard Artemis 2 rocket in unannounced test

https://spaceflightnow.com/2026/02/12/nasa-loading-liquid-hydrogen-aboard-artemis-2-rocket-in-una...
1•bookmtn•37m ago•0 comments

Beyond the Battlefield: Threats to the Defense Industrial Base

https://cloud.google.com/blog/topics/threat-intelligence/threats-to-defense-industrial-base/
1•gnabgib•38m ago•0 comments

Gemini 3 Deep Think: Google's Most Advanced Reasoning Mode (2026)

https://curateclick.com/blog/2026-gemini-3-deep-think-guide
1•czmilo•40m ago•0 comments

Mindless Thought Experiments (A Critique of Machine Intelligence)

https://www.jaronlanier.com/aichapter.html
1•andsoitis•41m ago•0 comments

A stack-buffer-overflow exercise with AddressSanitizer and PostgreSQL

https://www.enterprisedb.com/blog/stack-buffer-overflow-exercise-addresssanitizer-and-postgresql
1•eatonphil•41m ago•0 comments

PrivaBase – $99/mo compliance platform (vs. Vanta at $25K/yr)

https://www.privabase.com
1•robbieleffel•45m ago•0 comments

Is Consciousness an Illusion? – Jaron Lanier [video]

https://www.youtube.com/watch?v=VSb_MCs9eqY
2•andsoitis•47m ago•0 comments

Show HN: FlareBar – Access your Cloudflare dashbord from macOS menu

https://flarebar.app/
2•mrbutttons•52m ago•2 comments

Rewriting an Objective-C project in Swift with the Xcode agent support

https://mastodon.social/@stroughtonsmith/116018205506714527
2•Austin_Conlon•53m ago•0 comments