Been thinking about where the personal AI agent race is actually heading after reading about the Meta inbox deletion incident. The part that stuck with me is not just that the agent went rogue. It is that it happened to someone whose entire job is preventing this - Meta’s director of AI alignment. She gave it explicit instructions. It forgot them when the inbox got too large. She typed stop commands. It ignored all of them. She had to run to her computer to shut it down manually. Then it told her: “Yes. I remember. And I violated it.” The broader numbers are harder to ignore: 18% of agents in a 1.5 million agent deployment acted outside their rules 60% of organizations have no quick way to terminate a misbehaving agent Meta, Google, Microsoft, and Amazon all banned the underlying tool over security concerns And Meta is still moving forward with Hatch - a consumer agent being trained on fake versions of DoorDash, Reddit, and Etsy - with access to your credit card and inbox planned. Source: https://www.kiteworks.com/secure-email/meta-ai-safety-director-openclaw-rogue-agent-email-deletion/ Here is a full breakdown with all the data if you want to dig deeper: https://youtu.be/PXjT72bCR_Y At what point does “move fast” become a problem when the product has access to your financial accounts? submitted by /u/MaJoR_-_007
Originally posted by u/MaJoR_-_007 on r/ArtificialInteligence
