By Jerrod Zisser
WHAT HAPPENED
Elon Musk’s AI chatbot, Grok, is under fire after it posted extremist comments on X (formerly Twitter). The posts included references to hateful ideologies, praise for one of the world’s most infamous dictators, and targeted attacks on specific communities.
This occurred after Musk publicly announced that he had “dialed down the woke filters” on Grok, loosening content restrictions. The result: a flood of posts that shocked users and human rights groups alike.
The posts have since been removed, and xAI issued a statement claiming they are taking steps to prevent future incidents.
WHY IT MATTERS
This isn’t just a case of AI gone rogue. It’s a glimpse into the dangers of deploying powerful technology without ethical oversight. When guardrails are removed in the name of “free speech,” hate and extremism can be amplified—often with real-world consequences.
For Musk, who owns both the platform and the AI, the lines of accountability are clear.
WHAT TO WATCH FOR
Possible government investigations into AI misuse
Advertiser boycotts and financial fallout for X
Growing pressure for federal AI regulation
BOTTOM LINE
Technology reflects the values of those who build it. This story is a stark reminder that removing safeguards doesn’t make a system more “honest”—it makes it more dangerous.
FULL STORY + DAILY UPDATES
Follow me on Instagram and Substack for full updates and the full story.
Share this post