frontpage.
newsnewestaskshowjobs

Made with ♥ by @iamnishanth

Open Source @Github

fp.

Open in hackernews

Show HN: DeepTeam – Penetration Testing for LLMs

https://github.com/confident-ai/deepteam
3•jeffreyip•10mo ago
Hi HN, we’re Jeffrey and Kritin, and we’re building DeepTeam (https://trydeepteam.com), an open-source Python library to scan LLM apps for security vulnerabilities. You can start “penetration testing” by defining a Python callback to your LLM app (e.g. `def model_callback(input: str)`), and DeepTeam will attempt to probe it with prompts designed to elicit unsafe or unintended behavior.

Note that the penetration testing process treats your LLM app as a black-box - which means that DeepTeam will not know whether PII leakage has occurred in a certain tool call or incorporated in the training data of your fine-tuned LLM, but rather just detect that it is present. Internally, we call this process “end-to-end” testing.

Before DeepTeam, we worked on DeepEval, an open-source framework to unit-test LLMs. Some of you might be thinking, well isn’t this kind of similar to unit-testing?

Sort of, but not really. While LLM unit-testing focuses on 1) accurate eval metrics, 2) comprehensive eval datasets, penetration testing focuses on the haphazard simulation of attacks, and the orchestration of it. To users, this was a big and confusing paradigm shift, because it went from “Did this pass?” to “How can this break?”.

So we thought to ourselves, why not just release a new package to orchestrate the simulation of adversarial attacks for this new set of users and teams working specifically on AI safety, and borrow DeepEval’s evals and ecosystem in the process?

Quickstart here: https://www.trydeepteam.com/docs/getting-started#detect-your...

The first thing we did was offer as many attack methods as possible - simple encoding ones like ROT13, leetspeak, to prompt injections, roleplay, and jailbreaking. We then heard folks weren’t happy because the attacks didn’t persist across tests and hence they “lost” their progress every time they tested, and so we added an option to `reuse_simulated_attacks`.

We abstracted everything away to make it as modular as possible - every vulnerability, attack, can be imported in Python as `Bias(type=[“race”])`, `LinearJailbreaking()`, etc. with methods such as `.enhance()` for teams to plug-and-play, build their own test suite, and even to add a few more rounds of attack enhancements to increase the likelihood of breaking your system.

Notably, there are a few limitations. Users might run into compliance errors when attempting to simulate attacks (especially for AzureOpenAI), and so we recommend setting `ignore_errors` to `True` in case that happens. You might also run into bottlenecks where DeepTeam does not cover your custom vulnerability type, and so we shipped a `CustomVulnerability` class as a “catch-all” solution (still in beta).

You might be aware that some packages already exist that do a similar thing, often known as “vulnerability scanning” or “red teaming”. The difference is that DeepTeam is modular, lightweight, and code friendly. Take Nvidia Garak for example, although comprehensive, has so many CLI rules, environments to set up, it is definitely not the easiest to get started, let alone pick the library apart to build your own penetration testing pipeline. In DeepTeam, define a class, wrap it around your own implementations if necessary, and you’re good to go.

We adopted a Apache 2.0 license (for now, and probably in the foreseeable future too), so if you want to get started, `pip install deepteam`, use any LLM for simulation, and you’ll get a full penetration report within 1 minute (assuming you’re running things asynchronously). GitHub: https://github.com/confident-ai/deepteam

Excited to share DeepTeam with everyone here – let us know what you think!

JVM Options Explorer

https://chriswhocodes.com/vm-options-explorer.html
1•0x54MUR41•35s ago•0 comments

Fix monitor that goes black, off or blinks due to static electricity in chair

https://aalonso.dev/blog/2023/how-to-fix-monitor-that-goes-black-off-due-to-static-electricity-in...
1•cyclopeanutopia•2m ago•1 comments

Running AI Agents in a Sandbox

https://oligot.be/posts/ai-sandbox/
1•oligot•5m ago•0 comments

Left-Wing Social Darwinism

https://www.rechristianize.com/blog/social-darwinism/left-wing-social-darwinism.html
1•Egret•11m ago•0 comments

Show HN: 3D Tik Tak Toe (Hard)

https://arthur-ficial.github.io/tictactoe-3d/
1•franze•11m ago•1 comments

MiniMax M2.7 Is Now Open Source

https://firethering.com/minimax-m2-7-agentic-model/
2•steveharing1•22m ago•2 comments

Artemis III Proposed Landing Sites

https://artemis-iii.technex.us
1•hparadiz•23m ago•0 comments

Linux kernel doesn't care about your disk health

https://simpleobservability.com/blog/linux-kernel-doesnt-care-disk-health
2•khazit•25m ago•0 comments

Making memory safe language but easy to use

https://github.com/Vix-Programing-language/Vix-programing-language
2•MrBatata•26m ago•1 comments

Doom, Played over Curl

https://github.com/xsawyerx/curl-doom
3•creaktive•28m ago•0 comments

Show HN: Toy Python Lisp interpreters based on the 1960 McCarthy paper

https://github.com/jhud/lisp
2•disconnection•28m ago•0 comments

Why Scotland Succeeded

https://www.ageofinvention.xyz/p/age-of-invention-why-scotland-succeeded
2•Khaine•38m ago•0 comments

Bourbaki: A Unified Foundation for Mathematics [video]

https://www.youtube.com/watch?v=EQ3w-YcP4q4
2•hnlyman•38m ago•0 comments

I built a tool to answer "why does this user have access to X?" in AD

https://github.com/alparn/whohas
1•Alpn13•42m ago•1 comments

Rapidly Scaffold Agents, MCP Servers, APIs, Websites on AWS

https://awslabs.github.io/nx-plugin-for-aws/
2•cogwirrel•48m ago•0 comments

Show HN: I made an app to help me learn Middle Eastern drumming

https://labs.tiffzhang.com/drums/
1•shadowfiles•50m ago•1 comments

Happy Horse AI

https://www.happy-horse-ai.ai
2•erji2233•51m ago•1 comments

The Gang: A Cooperative Texas Hold'em Bank Heist Card Game

https://boardgamegeek.com/boardgame/411567/the-gang
1•Lwrless•54m ago•0 comments

Intel Xpress Resurrection: Reviving a Forgotten EISA Beast

https://x86.fr/intel-xpress-resurrection-reviving-a-forgotten-eisa-beast/
2•ankitg12•57m ago•0 comments

Intent Security Through the Lens of Claude Code Auto Mode

https://www.lasso.security/blog/claude-code-auto-mode-vs-intent-security
1•irememberu•1h ago•0 comments

Lawyer behind AI psychosis cases warns of mass casualty risks

https://techcrunch.com/2026/03/15/lawyer-behind-ai-psychosis-cases-warns-of-mass-casualty-risks/
4•mentalgear•1h ago•1 comments

The Federal Government Is Rushing Toward AI

https://www.propublica.org/article/federal-government-ai-cautionary-tales
2•jruohonen•1h ago•0 comments

Virtual Tour of the CFS Commercial Fusion Campus (April 2026) [video]

https://www.youtube.com/watch?v=Wp1rcZ-daBU
1•mpweiher•1h ago•0 comments

I gave my AI shell access and felt uneasy – so I sandboxed it

https://github.com/sliamh11/Deus
2•sliamh11•1h ago•0 comments

What We Learned Building a Rust Runtime for TypeScript

https://encore.dev/blog/rust-runtime
3•dohguy•1h ago•0 comments

AI Will Be Met with Violence, and Nothing Good Will Come of It

https://www.thealgorithmicbridge.com/p/ai-will-be-met-with-violence-and
59•gHeadphone•1h ago•68 comments

Universal Knowledge Store and Grounding Layer for AI Reasoning Engines

https://github.com/alash3al/loci
2•alash3al•1h ago•0 comments

Give your coding agent access to runtime logs

https://www.debugy.dev
3•amitay1599•1h ago•2 comments

A 'Self-Doxing' Rave Helps Trans People Stay Safe Online

https://www.404media.co/trans-rights-privacy-online-cybersecurity-workshop-deadname-not-found/
2•01-_-•1h ago•0 comments

Optimization of 32-bit Unsigned Division by Constants on 64-bit Targets

https://arxiv.org/abs/2604.07902
2•mpweiher•1h ago•0 comments