Colorado AI Act Jun 30, 2026 | EU AI Act Aug 2, 2026 | California ADMT Jan 1, 2026
Back to Blog
Open Source

Why auto-redteam.com Is an Open-Source Commitment

Joe Braidwood
Joe Braidwood
Co-founder & CEO
· March 26, 2026 · 6 min read

auto-redteam.com exists because developers need direct leverage. If teams are being asked to ship AI that is safe, reliable, and secure in production, the baseline hardening loop cannot stay trapped behind enterprise sales motions.

Developers are being asked to ship AI systems that are not just useful, but trustworthy in production. That is a much bigger ask than getting a model call to work. It means thinking about prompt injection, data leakage, jailbreaks, unsafe tool use, brittle guardrails, and the long tail of behavior that only shows up once a system is live.

That is why we built auto-redteam. And that is why we launched auto-redteam.com as the public home of that open-source commitment.

This Is an Open-Source Commitment

We do not think the right answer to AI security is to keep the foundational hardening tooling behind a budget cycle. If developers everywhere are being asked to take responsibility for production AI, they should have access to tools that help them do that work well right now.

auto-redteam.com is our way of making that capability easy to find, easy to try, and easy to build on. The point is not theater. The point is to help teams pressure-test real AI applications, surface meaningful failure modes, and strengthen systems before those weaknesses become incidents, procurement friction, or governance problems.

Inspired by Karpathy’s Developer-First Instinct

Part of the inspiration here is the developer-first instinct Andrej Karpathy represents: if something matters for the future of software, builders need direct leverage, not just commentary.

AI security cannot become a discipline that only large companies can afford to practice well. The baseline has to rise for everyone. Open tools are one of the few mechanisms that actually make that possible.

Open
Free entry point for developers
Real
Application-boundary testing
Repeatable
Behavioral probes that can run again
Practical
A hardening loop teams can actually use

Why This Matters

Most teams deploying AI are stuck in the same middle ground. They know production behavior is what matters. They know static checklists are not enough. They know a one-time eval is not the same as operational confidence. They still do not have a simple, developer-native way to continuously harden what they are shipping.

That gap matters for startups trying to move fast without being reckless, for platform teams standardizing safe deployment, and for regulated industries where “we thought it was fine” is not a satisfying answer after the fact.

What auto-redteam Is For

auto-redteam is designed to help developers auto-harden production AI applications. In practical terms, that means:

  • pointing the tool at a real application boundary, not just an abstract model
  • running repeatable behavioral probes that look for realistic weakness patterns
  • making it easy to see where a system is brittle, leaky, or too easy to bypass
  • creating a feedback loop teams can use to improve the system, then test it again

This is not about replacing engineering judgment. It is about making good engineering judgment easier to exercise, faster to operationalize, and harder to postpone.

Why We’re Launching It as Part of the Governance and AI Security Community

We do not see open-source hardening and governance infrastructure as separate conversations. They are the same stack viewed from different heights.

At the ground level, developers need tools that help them make systems stronger. At the organizational level, security teams need controls that enforce policy consistently. At the governance level, leaders need evidence that the system actually behaved the way they said it would.

Those are not competing needs. They are a continuum. Open-source hardening helps the whole ecosystem. Verifiable infrastructure helps that hardening stand up at scale.

The Vision: Auto-Hardening Plus GLACIS Infrastructure

The safe way to deploy AI at scale is not a one-time eval. It is not a dashboard. It is not a policy PDF. It is a stack.

  1. First, you auto-harden the system. You probe it, pressure-test it, learn how it fails, and close obvious gaps as the system evolves.
  2. Then you enforce runtime controls. You decide what is allowed, what is blocked, what is escalated, and what must be recorded in production.
  3. Then you generate evidence. Not just that controls were configured, but that they actually executed on the interactions that mattered.

That is the connection: auto-redteam helps teams harden production AI before failure becomes incident. GLACIS infrastructure helps them enforce policy and prove that the controls actually ran once the system is live.

What We Hope Happens Next

We want developers to use auto-redteam because it is useful on day one. We want teams to contribute to it because the failure modes of production AI are broader than any one company can model alone. And we want security leaders to treat auto-hardening as a default engineering practice, not a special event.

If you are serious about deploying AI at scale, you should be able to show three things: you pressure-test systems continuously, you enforce meaningful controls in production, and you can prove those controls actually ran.

Start Here

If you are building or operating production AI, start with the hardening loop. Install auto-redteam. Point it at a real system. See what breaks before somebody else does.

Launch
pip install glacis-autoredteam
auto-redteam.com
Pango waving

Ready to harden before you scale?

Start with behavioral hardening using auto-redteam, then layer GLACIS infrastructure on top when you need runtime enforcement and proof.

Visit auto-redteam.com