OpenAI Acquired Promptfoo. Here's What That Means for AI Security Testing.
Yesterday, OpenAI announced it is acquiring Promptfoo — the open-source AI red-teaming tool used by 350,000 developers (130K monthly active) and over 25% of the Fortune 500. All 11 employees, including co-founders Ian Webster and Michael D'Angelo, will join OpenAI. Promptfoo will be integrated into OpenAI Frontier, the enterprise platform for building and operating AI agents.
Promptfoo raised $23 million (Series A led by Insight Partners, seed from a16z and Shopify CEO Tobi Lutke), was valued at $86 million, and built the most popular open-source framework for LLM evaluation and red teaming. It was good software. We used it.
This is a significant moment for the AI security ecosystem. Here's how we're reading it.
What This Validates
First, the good news: AI security testing is now a confirmed enterprise category. OpenAI paying $86M+ for a red-teaming tool settles the "do we even need this?" question. Companies building AI agents need adversarial testing. That's no longer debatable.
Second, agent security is where the market is moving. Ian Webster said it clearly: "As AI agents become more connected to real data and systems, securing and validating them is more challenging and important than ever." OpenAI acquired Promptfoo specifically for Frontier — their agent platform. The signal is unambiguous: agent security is the priority.
The Multi-Vendor Question
Promptfoo's strength was multi-vendor neutrality — it tested OpenAI, Anthropic, Google, Meta, and open-source models with equal rigor. Promptfoo's own blog commits to keeping the repo public under the same license, continuing to support multiple providers, and cutting releases.
That commitment matters, and we take it at face value. But it's worth watching how it evolves. Enterprises running multi-model architectures — which is most of them — will want to see that multi-vendor investment continue. Security testing works best when the tooling is genuinely agnostic to the system under test.
Model Testing vs. Agent Testing
This is the more important shift the acquisition highlights.
Promptfoo excelled at model-level testing:
- LLM evaluation — comparing outputs across providers
- Prompt injection testing — detecting when models follow injected instructions
- Jailbreak detection — testing guardrail bypasses
- MCP security testing — tool response manipulation, cascading unauthorized actions, data leakage
Credit where it's due: Promptfoo was expanding into agent and MCP testing. Their MCP security testing guide covered real attack vectors. Integrating this into Frontier will raise the baseline for every team building on OpenAI's platform.
But there's a layer below application-level scanning that automated tools don't typically reach. Source-code auditing of the protocol SDKs themselves reveals systemic vulnerabilities — specification gaps that exist in every implementation following the spec as written:
- Tool Shadowing — the SDK's
listChangednotification has no origin verification, allowing a rogue server to silently replace trusted tools (we documented this in our MCP security research) - Token Audience Confusion — the SDK's
ProviderTokenVerifieraccepts tokens from any registered issuer, so a read-only token can execute admin operations - Stale Authorization — cached sessions have no re-validation mechanism, so revoked permissions persist
We tested 5 production MCP implementations and found 33 vulnerabilities across these patterns, plus 7 specification-level gaps (SG-1 through SG-7). One vendor patched in 6 hours. The vulnerabilities weren't in the models — they were in the protocol layer.
This isn't a critique of Promptfoo — automated scanning and manual source-code auditing solve different problems. But it illustrates why the market needs both.
What This Means Going Forward
Platform-native security gets better. OpenAI embedding red-teaming directly into Frontier raises the floor for every team building agents on their platform. That's genuinely good for the ecosystem.
The ecosystem still needs depth testing. Automated scanning catches common patterns at scale. Manual red teaming finds the vulnerabilities that automated tools miss — the ones that live in protocol design, authorization logic, and trust boundaries between services. Both are necessary.
Agent security is a distinct discipline. AI agents don't just generate text — they call tools, access databases, execute code, and operate across trust boundaries. The attack surface is the integration layer: tool boundaries, authorization flows, multi-agent trust. This requires different methodology than model evaluation.
Where We Fit
Tachyonic does AI red team assessments. We test agents, not just models. Our focus is the attack surface in the integration layer:
- MCP tool boundaries — tool shadowing, description poisoning, cross-server trust
- Agent authorization — token audience confusion, stale auth, privilege escalation through tool chains
- Multi-agent systems — cascading injection, inter-agent trust, orchestration layer vulnerabilities
Our attack taxonomy is open-source — 122 attack patterns. We published the first source-code-level audit of both MCP SDKs and disclosed 33 vulnerabilities across production implementations. We run 48-hour red team assessments that test whether your agent system holds up when a skilled adversary targets the integration layer.
The market just got bigger and clearer. That's good for everyone working on AI security.
If you're building AI agents and want to know where the boundaries break, book a 15-minute scoping call.
Secure Your AI Agents
We find vulnerabilities in AI applications in 48 hours. Resistance score, reproduction steps, remediation playbook included.
Book a Free Scoping Call