frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Making frontier cybersecurity capabilities available to defenders

https://www.anthropic.com/news/claude-code-security
52•surprisetalk•2h ago

Comments

upghost•1h ago
Anakin: I'm going to save the world with my AI vulnerability scanner, Padme.

Padme: You're scanning for vulnerabilities so you can fix them, Anakin?

Anakin: ...

Padme: You're scanning for vulnerabilities so you can FIX THEM, right, Annie?

czbond•1h ago
Definitely will be a fight against bad actors pulling bulk open source software projects, npm packages, etc and running this for their own 0 days.

I hope Anthropic can place alerts for their team to look for accounts with abnormal usage pre-emptively.

tptacek•1h ago
You want frontier models to actively prevent people from using them to do vulnerability research because you're worried bad people will do vulnerability research?
czbond•1h ago
Not at all. I was suggesting if an account is performing source code level request scanning of "numerous" codebases - that it could be an account of interest. A sign of mis-use.

This is different than someones "npm audit" suggesting issues with packages in a build and updating to new revisions. Also different than iterating deeply on source code for a project (eg: nginx web server).

tptacek•1h ago
I don't understand the joke here.
ukuina•11m ago
A vuln scanner is dual-use.
nikcub•2m ago
I assume that's why this is gated behind a request for access from teams / enterprise users rather than being GA
drcongo•1h ago
I thought they'd noticed how many of my Claude tokens I've been burning trying to build defences against the AI bot swarms. Sadly not.
reconnecting•38m ago
Is it only crawlers or bots that abuse your product?

We have been developing our own system (1) for several years, and it's built by engineers, not Claude. Take a look — maybe it could be helpful for your case.

1. https://github.com/tirrenotechnologies/tirreno

deadbabe•1h ago
Solve a problem and everyone praises you.

No one knows you also caused that problem.

nadis•1h ago
> "Rather than scanning for known patterns, Claude Code Security reads and reasons about your code the way a human security researcher would: understanding how components interact, tracing how data moves through your application, and catching complex vulnerabilities that rule-based tools miss."

Fascinating! Our team has been blending static code analysis and AI for a while and think it's a clever approach for the security use case the Anthropic team's targeting here.

bink•1h ago
I hope this is better than their competitors products. So far I've been underwhelmed. They basically just find stuff that's already identified by static analysis tooling and toss in a bunch of false positives from the AI scans.
david_shaw•1h ago
There's a lot of skepticism in the security world about whether AI agents can "think outside the box" enough to replicate or augment senior-level security engineers.

I don't yet have access to Claude Code Security, but I think that line of reasoning misses the point. Maybe even the real benefit.

Just like architectural thinking is still important when developing software with AI, creative security assessments will probably always be a key component of security evaluation.

But you don't need highly paid security engineers to tell you that you forgot to sanitize input, or you're using a vulnerable component, or to identify any of the myriad issues we currently use "dumb" scanners for.

My hope is that tools like this can help automate away the "busywork" of security. We'll see how well it really works.

tptacek•1h ago
I am seeing something closer to the opposite of skepticism among vulnerability researchers. It's not my place to name names, but for every Halvar Flake talking publicly about this stuff, there are 4 more people of similar stature talking privately about it.
awestroke•1h ago
Claude Opus 4.6 has been amazing at identifying security vulnerabilities for us. Less than 50% falae positives.
samuelknight•4m ago
LLMs and particularly Claude are very capable security engineers. My startup builds offensive pentesting agents (so more like red teaming), and if you give it a few hours to churn on an endpoint it will find all sorts of wacky things a human won't bother to check.
ievans•1h ago
Not super surprising that Anthropic is shipping a vulnerability detection feature -- OpenAI announced Aardvark back in October (https://openai.com/index/introducing-aardvark/) and Google announced BigSleep in Nov 2024 (https://cloud.google.com/blog/products/identity-security/clo...).

The impact question is really around scale; a few weeks ago Anthropic claimed 500 "high-severity" vulnerabilities discovered by Opus 4.6 (https://red.anthropic.com/2026/zero-days/). There's been some skepticism about whether they are truly high severity, but it's a much larger number than what BigSleep found (~20) and Aardvark hasn't released public numbers.

As someone who founded a company in the space (Semgrep), I really appreciated that the DARPA AIxCC competition required players using LLMs for vulnerability discovery to disclose $cost/vuln and the confusion matrix of false positives along with it. It's clear that LLMs are super valuable for vulnerability discovery, but without that information it's difficult to know which foundation model is really leading.

What we've found is that giving LLM security agents access to good tools (Semgrep, CodeQL, etc.) makes them significantly better esp. when it comes to false positives. We think the future is more "virtual security engineer" agents using tools with humans acting as the appsec manager. Would be very interested to hear from other people on HN who have been trying this approach!

nikcub•8m ago
> What we've found is that giving LLM security agents access to good tools (Semgrep, CodeQL, etc.) makes them significantly better

100% agree - I spun out an internal tool I've been using to close the loop with website audits (more focus on website sec + perf + seo etc. rather than appsec) in agents and the results so far have been remarkable:

https://squirrelscan.com/

Human written rules with an agent step that dynamically updates config to squash false positives (with human verification) and find new issues in a loop.

Keep Android Open

https://f-droid.org/2026/02/20/twif.html
470•LorenDB•2h ago•172 comments

Ggml.ai joins Hugging Face to ensure the long-term progress of Local AI

https://github.com/ggml-org/llama.cpp/discussions/19759
539•lairv•6h ago•121 comments

I found a useful Git one liner buried in leaked CIA developer docs

https://spencer.wtf/2026/02/20/cleaning-up-merged-git-branches-a-one-liner-from-the-cias-leaked-d...
493•spencerldixon•6h ago•186 comments

I found a Vulnerability. They found a Lawyer

https://dixken.de/blog/i-found-a-vulnerability-they-found-a-lawyer
45•toomuchtodo•59m ago•19 comments

Lil' Fun Langs

https://taylor.town/scrapscript-000
59•surprisetalk•2h ago•4 comments

Making frontier cybersecurity capabilities available to defenders

https://www.anthropic.com/news/claude-code-security
52•surprisetalk•2h ago•19 comments

Trump's global tariffs struck down by US Supreme Court

https://www.bbc.com/news/live/c0l9r67drg7t
922•blackguardx•4h ago•746 comments

Testing Super Mario Using a Behavior Model Autonomously

https://testflows.com/blog/testing-super-mario-using-a-behavior-model-autonomously-part1/
18•Naulius•1h ago•4 comments

Facebook is absolutely cooked

https://pilk.website/3/facebook-is-absolutely-cooked
233•npilk•1h ago•181 comments

The path to ubiquitous AI (17k tokens/sec)

https://taalas.com/the-path-to-ubiquitous-ai/
591•sidnarsipur•9h ago•346 comments

Child's Play: Tech's new generation and the end of thinking

https://harpers.org/archive/2026/03/childs-play-sam-kriss-ai-startup-roy-lee/
250•ramimac•5h ago•154 comments

Untapped Way to Learn a Codebase: Build a Visualizer

https://jimmyhmiller.com/learn-codebase-visualizer
163•andreabergia•11h ago•27 comments

Legion Health (YC) Is Hiring Cracked SWEs for Autonomous Mental Health

https://jobs.ashbyhq.com/legionhealth/ffdd2b52-eb21-489e-b124-3c0804231424
1•ympatel•3h ago

PayPal discloses data breach that exposed user info for 6 months

https://www.bleepingcomputer.com/news/security/paypal-discloses-data-breach-exposing-users-person...
205•el_duderino•6h ago•60 comments

The Popper Principle

https://theamericanscholar.org/the-popper-principle/
42•lermontov•1d ago•23 comments

Do you want to build a community where users search or hang? (2021)

https://www.mooreds.com/wordpress/archives/3486
9•mooreds•3d ago•3 comments

No Skill. No Taste

https://blog.kinglycrow.com/no-skill-no-taste/
132•ianbutler•4h ago•132 comments

How to Review an AUR Package

https://bertptrs.nl/2026/01/30/how-to-review-an-aur-package.html
13•exploraz•3d ago•0 comments

Wikipedia bans Archive.today after site executed DDoS and altered web captures

https://arstechnica.com/tech-policy/2026/02/wikipedia-bans-archive-today-after-site-executed-ddos...
29•nobody9999•1h ago•15 comments

Raspberry Pi Pico 2 at 873.5MHz with 3.05V Core Abuse

https://learn.pimoroni.com/article/overclocking-the-pico-2
115•Lwrless•11h ago•35 comments

The Rediscovery of 103 Hokusai Lost Sketches (2021)

https://japan-forward.com/eternal-hokusai-the-rediscovery-of-103-hokusai-lost-sketches/
51•debo_•4d ago•5 comments

Consistency diffusion language models: Up to 14x faster, no quality loss

https://www.together.ai/blog/consistency-diffusion-language-models
192•zagwdt•16h ago•83 comments

AI is not a coworker, it's an exoskeleton

https://www.kasava.dev/blog/ai-as-exoskeleton
459•benbeingbin•1d ago•470 comments

Infrastructure decisions I endorse or regret after 4 years at a startup (2024)

https://cep.dev/posts/every-infrastructure-decision-i-endorse-or-regret-after-4-years-running-inf...
447•Meetvelde•3d ago•199 comments

Web Components: The Framework-Free Renaissance

https://www.caimito.net/en/blog/2026/02/17/web-components-the-framework-free-renaissance.html
166•mpweiher•11h ago•111 comments

Visible Spectra of the Elements

https://atomic-spectra.net/
31•djoldman•3d ago•2 comments

Lessons learned from `oapi-codegen`'s time in the GitHub Secure Open Source Fund

https://www.jvt.me/posts/2026/02/17/oapi-codegen-github-secure/
9•zdw•2d ago•0 comments

KFC, Nando's, and others ditch chicken welfare pledge

https://www.bbc.co.uk/news/articles/cm2r6jqm042o
28•penguin_booze•1h ago•7 comments

Notes on Clarifying Man Pages

https://jvns.ca/blog/2026/02/18/man-pages/
58•surprisetalk•1d ago•38 comments

Reading the undocumented MEMS accelerometer on Apple Silicon MacBooks via iokit

https://github.com/olvvier/apple-silicon-accelerometer
128•todsacerdoti•15h ago•55 comments