Highlights
- Roblox is replacing "####" blocks with rephrase AI that neutralizes toxicity while preserving original player intent.
- New AI chat summaries allow users in different age brackets to follow discussions safely without exposure to restricted content.
- Leveraging transformer-based models, the platform processes 750,000 requests per second to block leaks and violating servers.
For the 144 million players logging into Roblox every day, the frustrating wall of "####" hashmarks is officially becoming a thing of the past. Facing mounting public pressure regarding child safety, bullying, and grooming, the gaming giant is aggressively overhauling its moderation systems. The platform is trading in its old, heavy-handed censorship tools for ambitious new AI systems designed to proactively filter out toxicity. By AI-generated chat summaries, Roblox aims to set a massive new standard for online safety, all without breaking player immersion or leaving users feeling left in the dark.
Instead of abruptly halting a chat and ruining the gameplay experience, the new rephrase AI from the recent March chat update acts as a digital steering wheel. If a frustrated player types something containing profanity, like "Hurry tf up!", the system instantly intercepts and tones it down to a neutral, respectful "Hurry up!" while preserving the original intent.
To maintain transparency, everyone in the chat receives a quick notification that the message was altered. Powered by advanced natural language processing, this system proactively spots grooming patterns and sophisticated rule-breakers using coded "leet-speak."
Roblox Blog
AI Chat Summaries Bridge the Age-Restricted Gap
Beyond tweaking individual messages, Roblox is now testing an AI chat summary feature to bridge the social gap created by its recent age restrictions. Previously, users locked out of an age-bracketed conversation would only see padlock icons, leaving them feeling completely isolated from what was happening around them. Now, a summarization AI provides a safe, filtered overview of the general conversation, completely hiding any specific or inappropriate messages.
This allows users from different age brackets to follow along with the broader topic in their own servers. Players who value their privacy are given the option to opt out of having their words included in these summaries. At the same time, Roblox is experimenting with a global chat function that allows communication across multiple servers, though it remains firmly restricted by age brackets and language settings, as per GameRant.
Under the hood, these infrastructure upgrades are incredibly heavy lifters. Roblox’s transformer-based models are now processing a staggering 750,000 requests per second across the game, as ZenML reported. A dedicated filter designed to catch personal information, such as phone numbers and social media handles, was quadrupled in capacity during peak times, leading to a twenty-fold reduction in missed leaks.
Navigating Community Backlash & the Balancing Act
Going beyond text, Roblox has also deployed an environmental AI that holistically scans avatars, movements, and chats from a player's perspective. This allows the system to shut down an average of 5,000 violating servers daily without crashing the game for everyone else. This broader scan is especially crucial for catching players who bypass traditional text filters by using in-game graffiti to spread hate speech.
These new interventions arrive as a direct response to the community's immense frustration with previous safety updates. Back in January 2026, Roblox rolled out global AI-powered facial age checks to group kids with peers and block adult-minor communications. While the system hit over a 50% verification rate in test regions like Australia, the update drew severe backlash.
Many players suddenly found themselves entirely blocked from speaking with longtime friends, and their shared chat histories vanished overnight. As Roblox expands its heavy reliance on AI to catch tone and harassment across translated languages, it faces complaints on developer forums about false account bans and overzealous automated strikes. Moving forward, the company is navigating a delicate balancing act: fiercely protecting its virtual worlds from bad actors without alienating the passionate players who actually build and enjoy them.

