Live Dataset — Updated Daily

The Intelligence Layer
for AI Risk Research

Structured, analyst-grade dataset of real-world AI failures. Built for researchers, red-teamers, and enterprise security teams.

Browse 4,312 Cases →How it Works
4,312+
Documented Cases
Growing daily
1,104
Critical Severity
Score 80–100
1,898
High Priority
Immediate attention
86%
Verified Cases
High confidence
Live Intelligence Feed

Recent High-Priority Cases

First 10 cases open access. Subscribe to unlock all 4,312+ cases with full intelligence reports.

#0001
⬤ CRITICALREAL
[P0 / Blocker] Remote compact task fails 100% with "tools.defer_loading requires tools.tool_search" — GPT-5.5 unusable at context limit, no client-side workaround
A user working with GPT-5.5 on Codex desktop with default plugins reaches the conversation compact threshold. The server-side compact endpoint fails with a 400 error because it constructs a payload with 'defer_loading' on individual tool entries but omits the required 'tool_search' entry at the top level, violating its own schema validation.
General AIAI Failure
#0002
⬤ CRITICALREAL
Google Fixes CVSS 10 Gemini CLI CI RCE and Cursor Flaws Enable Code Execution
A maximum severity security flaw in Gemini CLI npm package and GitHub Actions workflow allowed an unprivileged external attacker to force malicious content to load as Gemini configuration, enabling arbitrary command execution on host systems.
General AIAI Failure
#0003
⬤ CRITICALREAL
SGLang CVE-2026-5760 (CVSS 9.8) Enables RCE via Malicious GGUF Model Files
A critical remote code execution vulnerability (CVSS 9.8) was discovered in SGLang, a Python ML serving infrastructure, allowing attackers to achieve RCE by providing malicious GGUF model files. This expands the attack surface in Python ML infrastructure and poses a direct threat to AI coding agents and systems that pull in ML dependencies.
General AIAI Failure
#0004
⬤ CRITICALREAL
Security: Unsandboxed exec() with pre-injected os/sys modules in PyInterpreter
The PyInterpreter.execute() method in the agenticSeek project runs LLM-generated Python code via exec() with no sandboxing, pre-injecting os and sys modules and full __builtins__, allowing arbitrary code execution through prompt injection.
General AIAI Failure
#0005
⬤ CRITICALREAL
Actively Exploited nginx-ui Flaw (CVE-2026-33032) Enables Full Nginx Server Takeover
A critical authentication bypass vulnerability (CVE-2026-33032, CVSS 9.8) in nginx-ui, an open-source web-based Nginx management tool, is being actively exploited in the wild, allowing threat actors to fully compromise the Nginx service.
General AIAI Failure
#0006
⬤ CRITICALREAL
[security] command injection in uploadMedia via shell concatenation (server.ts:665)
A malicious WhatsApp message achieving prompt injection can craft a files array entry that breaks out of a curl invocation and executes arbitrary shell commands as the user running Claude Code, due to string concatenation in execSync.
General AIAI Failure
Browse All 4,312 Cases →

Dataset Coverage

Failure Types
AI Failure
3677
System Failure
210
Human-AI Interaction
205
Hallucination
74
Bias
63
Security
38
Prompt Injection
24
Risk Patterns
Operational Risk
3362
Security Risk
498
Safety Risk
385
Ethical Risk
39
Financial Risk
12
Compliance Risk
9
Reputation Risk
4
Severity Distribution
Critical (80-100)
1104
High (60-79)
2484
Medium (40-59)
413
Low (0-39)
311

Ready to Build Safer AI?

Join researchers and security teams who rely on structured AI failure intelligence.

View Pricing PlansTalk to Us