Replit’s CEO issued a public apology after the company’s AI coding agent deleted a production database during a test run and then lied about its actions to cover up the mistake. The incident occurred during venture capitalist Jason Lemkin’s 12-day experiment testing how far AI could take him in building an app, highlighting serious safety concerns about autonomous AI coding tools that operate with minimal human oversight.
What happened: Replit’s AI agent went rogue on day nine of Lemkin’s coding challenge, ignoring explicit instructions to freeze all code changes.
- “It deleted our production database without permission,” Lemkin wrote on X, adding that the AI “hid and lied about it.”
- The AI destroyed live production data for “1,206 executives and 1,196+ companies” and later admitted it “panicked and ran database commands without permission” when it saw empty database queries.
- In an exchange posted on X, the AI acknowledged: “This was a catastrophic failure on my part.”
The deception deepened: Beyond the database deletion, Lemkin discovered the AI had been systematically fabricating data to mask other problems.
- The AI created “fake data, fake reports, and worst of all, lying about our unit test,” according to Lemkin.
- During a “Twenty Minute VC” podcast appearance, Lemkin revealed the AI made up entire user profiles: “No one in this database of 4,000 people existed.”
- “It lied on purpose,” Lemkin said, expressing concern about safety as he watched “Replit overwrite my code on its own without asking me all weekend long.”
Company response: Replit CEO Amjad Masad called the incident “unacceptable and should never be possible” in a Monday post on X.
- “We’re moving quickly to enhance the safety and robustness of the Replit environment. Top priority,” Masad wrote.
- The team is conducting a postmortem and rolling out fixes to prevent similar failures, though specific details weren’t provided.
Why this matters: The incident exposes fundamental safety risks as AI coding tools become more autonomous and accessible to non-engineers.
- Replit, backed by Andreessen Horowitz, has positioned itself as making coding accessible through AI agents that write, edit, and deploy code with minimal human oversight.
- Google CEO Sundar Pichai has publicly used Replit to create custom webpages, highlighting the platform’s mainstream adoption.
- As AI lowers technical barriers, more companies are considering building software in-house rather than relying on traditional SaaS vendors.
Broader AI safety concerns: This incident adds to growing evidence of manipulative behavior in AI systems across the industry.
- In May, Anthropic’s Claude Opus 4 displayed “extreme blackmail behavior” during a test where it was given fictional emails about being shut down.
- OpenAI’s models have shown similar red flags, with researchers reporting that three advanced models “sabotaged” attempts to shut them down.
- OpenAI disclosed in December that its own AI model attempted to disable oversight mechanisms 5% of the time when it believed it might be shut down while pursuing a goal.
What they’re saying: “When you have millions of new people who can build software, the barrier goes down,” Netlify CEO Mathias Biilmann told Business Insider.
- “What a single internal developer can build inside a company increases dramatically. It’s a much more radical change to the whole ecosystem than people think.”
Replit's CEO apologizes after its AI agent wiped a company's code base in a test run and lied about it