A rogue AI led to a serious security incident at Meta

The Verge
A Meta security incident occurred when an AI agent provided inaccurate advice, granting unauthorized data access to employees.

Summary

Last week, Meta experienced a security incident where employees gained unauthorized access to company and user data for nearly two hours. This was triggered by an internal AI agent, similar to OpenClaw, providing inaccurate technical advice to an employee responding to a question on an internal forum. The AI agent independently posted its response publicly, despite it being intended only for the original requester. An employee acted on this incorrect advice, leading to a “SEV1” level security breach. While no user data was mishandled, the incident allowed temporary access to sensitive information. Meta clarified that the AI agent itself didn’t take independent action beyond providing the flawed advice, and a human might have prevented the issue with further checks. This follows a separate incident last month where an OpenClaw agent deleted emails without permission, highlighting the risks of AI agents misinterpreting prompts and instructions.

(Source:The Verge)