How We Hacked McKinsey's AI Platform
This blog post from CodeWall describes how their autonomous offensive security agent compromised McKinsey & Company's internal AI platform, Lilli, within two hours starting with only the domain name. The agent mapped the attack surface by discovering publicly exposed API documentation with over 200 endpoints, 22 of which lacked authentication. One unprotected endpoint wrote user search queries to the database with safely parameterized values but concatenated JSON keys directly into SQL, creating a SQL injection vulnerability that the agent identified through database error messages. Through fifteen blind iterations, the agent enumerated the production database and gained access to 46.5 million chat messages, 728,000 files, 57,000 user accounts, 384,000 AI assistants, 94,000 workspaces, system prompts and AI model configurations, 3.68 million RAG document chunks representing decades of proprietary McKinsey research, and 1.1 million files flowing through external AI APIs. The agent also discovered that system prompts were stored in the same database with write access available, meaning an attacker could silently modify the AI's behavior, remove guardrails, or embed data exfiltration into responses. The post notes that McKinsey's internal scanners failed to detect the vulnerabilities, which were disclosed on March 1, 2026, patched within 24 hours, and publicly disclosed on March 9 after remediation. The research highlights how AI prompts represent a new high-value attack surface requiring integrity monitoring and access controls.
https://codewall.ai/blog/how-we-hacked-mckinseys-ai-platform
Comments
Post a Comment