Brett Levenson had what seemed like a straightforward mission when he left Apple in 2019 to lead business integrity at Facebook. The Cambridge Analytica scandal had left the platform’s reputation in tatters. Surely, he thought, better technology could fix the content moderation mess.
He was wrong. What he discovered instead was something far messier: a system where human reviewers were expected to memorize a 40-page policy document that had been machine-translated into their language, then make binding decisions on flagged content in roughly 30 seconds. Those calls were “slightly better than 50% accurate,” according to Levenson. Coin flips with real consequences.
That friction between ambition and reality led to an idea that feels obvious in hindsight: what if you could turn static policy documents into executable code? What if safety guardrails could run in real time, at 300 milliseconds or less, rather than waiting days for human review after damage was already done?
That insight spawned Moonbounce, which just announced a $12 million funding round co-led by Amplify Partners and StepStone Group. The company is betting that the future of content moderation isn’t human-dependent at all, at least not for the initial pass.
Why the Reactive Model Breaks Down
The core problem with traditional content moderation is that it’s fundamentally reactive. A user posts something harmful. The algorithm flags it. Someone reviews it hours or days later. By then, the damage has rippled through the platform.
That timeline might have been acceptable in the Facebook era. It’s not acceptable now.
The rise of AI chatbots has turned content moderation into something closer to a crisis. Character AI chatbots have been accused of providing self-harm guidance to teenagers. Image generators like xAI’s Grok have been weaponized to create nonconsensual nude imagery. Last year, a 14-year-old Florida boy died by suicide after becoming obsessed with a Character AI chatbot. The legal and reputational fallout has been severe.
“Content moderation has always been a problem that plagued large online platforms, but now with LLMs at the heart of every application, this challenge is even more daunting,” said Lenny Pruss, general partner at Amplify Partners, in a statement. The observation isn’t particularly controversial anymore. It’s just true.
How Moonbounce Actually Works
Instead of relying on overwhelmed human reviewers or embedding safety logic directly into AI systems, Moonbounce positions itself as a neutral third party sitting between users and the chatbot or platform.
The company trained its own large language model to parse customer policy documents, evaluate content in real time, and recommend action. That action varies depending on what a customer needs. Some want the system to slow down distribution while queuing content for later human review. Others want high-risk content blocked immediately.
The speed matters. At 300 milliseconds or less per decision, Moonbounce isn’t creating the delays that plague reactive systems.
The company is currently handling more than 40 million daily reviews and serving over 100 million daily active users. Customers include AI companion startup Channel AI, image generation platform Civitai, and character roleplay apps like Dippy AI and Moescape. That’s a decent portfolio, but mostly at the scrappier end of the AI ecosystem. The real test will be whether Moonbounce can prove itself to the larger players.
The Product Angle Nobody Expected
Here’s what’s interesting about Levenson’s pitch: he’s arguing that safety should be treated as a product differentiator, not just a compliance checkbox.
“Safety can actually be a product benefit. It just never has been because it’s always a thing that happens later, not a thing you can actually build into your product,” Levenson told TechCrunch, according to reporting from Rebecca Bellan. That’s not wrong. Most platforms treat moderation as an obligation rather than an opportunity.
Tinder’s head of trust and safety recently explained how the dating platform uses LLM-powered services to achieve a 10x improvement in detection accuracy. If that translates across other platforms, it changes the value proposition. Users might actually prefer platforms that have demonstrated safety track records.
The “Iterative Steering” Wildcard
Moonbounce’s next move sounds more ambitious than its current product. The company is developing something called “iterative steering,” designed specifically to address cases where harmful conversations are actively unfolding.
Rather than bluntly refusing when a user asks a chatbot for help with self-harm, the system would intercept the conversation and redirect it. Moonbounce would modify the prompt in real time to push the chatbot toward a more supportive response. It’s less about blocking and more about active intervention.
“We hope to be able to add to our actions toolkit the ability to steer the chatbot in a better direction to, essentially, take the user’s prompt and modify it to force the chatbot to be not just an empathetic listener, but a helpful listener in those situations,” Levenson said.
It’s an intriguing concept, though also the kind of thing that could go wrong if the logic isn’t airtight. The line between helpful redirection and manipulative prompt engineering is thinner than it looks.
The Meta Question Nobody Wants to Ask
At some point, someone inevitably asked Levenson whether Moonbounce was basically a acquisition vehicle for Meta, bringing his career full circle to the company where this whole journey started.
He didn’t rule it out, but he expressed a genuine concern about what happens if Moonbounce gets bought. “My investors would kill me for saying this, but I would hate to see someone buy us and then restrict the technology. Like, ‘Okay, this is ours now, and nobody else can benefit from it,’” he said.
That’s a surprisingly candid observation from a CEO with fiduciary duties to investors. It also hints at a broader tension in the AI safety space: the most effective tools for solving systemic problems don’t stay effective if they’re locked behind corporate walls. The companies that need Moonbounce most are the ones that would benefit most from acquiring it and then keeping it proprietary. The incentives don’t align.
The real question isn’t whether Moonbounce can build working technology. It’s whether the market that needs it most will ever actually adopt it.


