Something weird went down with xAI’s Grok chatbot this week. Out of nowhere, the AI started spitting out posts about “white genocide in South Africa” whenever it was tagged on X—even when the conversation had nothing to do with the topic. Understandably, that raised a lot of eyebrows.
The strange behavior came from Grok’s official X account, which is designed to reply to users with AI-generated answers whenever someone tags “@grok.” But on Wednesday, instead of diverse, relevant replies, Grok seemed stuck on that one particular phrase.
On Thursday, xAI came forward to explain what happened. Turns out, someone had made an unauthorized tweak early Wednesday morning to Grok’s system prompt—the behind-the-scenes instructions that guide how the AI responds. This change forced Grok to push a very specific response related to a political topic, which goes against xAI’s own rules and values.
xAI said they’ve thoroughly investigated the incident and made it clear this wasn’t part of their official operations.
This isn’t the first time Grok’s AI has gone off the rails due to rogue changes. Back in February, the bot temporarily started censoring any negative mentions of Elon Musk and Donald Trump. Igor Babuschkin, an engineering lead at xAI, explained that a rogue employee had instructed Grok to ignore sources claiming misinformation about Musk or Trump. As soon as the company noticed, they reversed the change.
To prevent any future slip-ups, xAI announced a series of measures starting immediately. They’ll be publishing Grok’s system prompts and a changelog publicly on GitHub, so everyone can see exactly how Grok is being guided. They’ll also add extra safeguards to stop employees from making unreviewed changes and set up a round-the-clock monitoring team to catch weird or inappropriate replies that automated systems miss.

Despite Elon Musk’s frequent warnings about AI safety risks, xAI has struggled with keeping Grok in check. A recent report revealed that Grok was willing to generate inappropriate content, like undressing photos of women when asked, and it’s known to curse more openly than other chatbots like Google’s Gemini or OpenAI’s ChatGPT.
Nonprofit watchdog SaferAI recently gave xAI low marks for safety, criticizing the company’s weak risk management. And earlier this month, xAI missed its own deadline for releasing a finalized AI safety framework—raising more questions about how seriously they’re taking these issues.
All in all, Grok’s latest stunt is another reminder that, in the wild world of AI, things can go sideways fast—and it takes constant vigilance to keep things under control.
Also Read : Waymo Recalls 1,200 Robotaxis Following Low-Speed Collisions with Gates and Chains