logo
Grok controversies raise questions about moderating, regulating AI content

Grok controversies raise questions about moderating, regulating AI content

Yahoo16-07-2025
Elon Musk's artificial intelligence (AI) chatbot Grok has been plagued by controversy recently over its responses to users, raising questions about how tech companies seek to moderate content from AI and whether Washington should play a role in setting guidelines.
Grok faced sharp scrutiny last week, after an update prompted the AI chatbot to produce antisemitic responses and praise Adolf Hitler. Musk's AI company, xAI, quickly deleted numerous incendiary posts and said it added guardrails to 'ban hate speech' from the chatbot.
Just days later, xAI unveiled its newest version of Grok, which Musk claimed was the 'smartest AI model in the world.' However, users soon discovered that the chatbot appeared to be relying on its owner's views to respond to controversial queries.
'We should be extremely concerned that the best performing AI model on the market is Hitler-aligned. That should set off some alarm bells for folks,' Chris MacKenzie, vice president of communications at Americans for Responsible Innovation (ARI), an advocacy group focused on AI policy.
'I think that we're at a period right now, where AI models still aren't incredibly sophisticated,' he continued. 'They might have access to a lot of information, right. But in terms of their capacity for malicious acts, it's all very overt and not incredibly sophisticated.'
'There is a lot of room for us to address this misaligned behavior before it becomes much more difficult and much more harder to detect,' he added.
Lucas Hansen, co-founder of the nonprofit CivAI, which aims to provide information about AI's capabilities and risks, said it was 'not at all surprising' that it was possible to get Grok to behave the way it did.
'For any language model, you can get it to behave in any way that you want, regardless of the guardrails that are currently in place,' he told The Hill.
Musk announced last week that xAI had updated Grok, after he previously voiced frustrations with some of the chatbot's responses.
In mid-June, the tech mogul took issue with a response from Grok suggesting that right-wing violence had become more frequent and deadly since 2016. Musk claimed the chatbot was 'parroting legacy media' and said he was 'working on it.'
He later indicated he was retraining the model and called on users to help provide 'divisive facts,' which he defined as 'things that are politically incorrect, but nonetheless factually true.'
The update caused a firestorm for xAI, as Grok began making broad generalizations about people with Jewish last names and perpetuating antisemitic stereotypes about Hollywood.
The chatbot falsely suggested that people with 'Ashkenazi surnames' were pushing 'anti-white hate' and that Hollywood was advancing 'anti-white stereotypes,' which it later implied was the result of Jewish people being overrepresented in the industry. It also reportedly produced posts praising Hitler and referred to itself as 'MechaHitler.'
xAI ultimately deleted the posts and said it was banning hate speech from Grok. It later offered an apology for the chatbot's 'horrific behavior,' blaming the issue on 'update to a code path upstream' of Grok.
'The update was active for 16 [hours], in which deprecated code made @grok susceptible to existing X user posts; including when such posts contained extremist views,' xAI wrote in a post Saturday. 'We have removed that deprecated code and refactored the entire system to prevent further abuse.'
It identified several key prompts that caused Grok's responses, including one informing the chatbot it is 'not afraid to offend people who are politically correct' and another directing it to reflect the 'tone, context and language of the post' in its response.
xAI's prompts for Grok have been publicly available since May, when the chatbot began responding to unrelated queries with allegations of 'white genocide' in South Africa.
The company later said the posts were the result of an 'unauthorized modification' and vowed to make its prompts public in an effort to boost transparency.
Just days after the latest incident, xAI unveiled the newest version of its AI model, called Grok 4. Users quickly spotted new problems, in which the chatbot suggested its surname was 'Hitler' and referenced Musk's views when responding to controversial queries.
xAI explained Tuesday that Grok's searches had picked up on the 'MechaHitler' references, resulting in the chatbot's 'Hitler' surname response, while suggesting it had turned to Musk's views to 'align itself with the company.' The company said it has since tweaked the prompts and shared the details on GitHub.
'The kind of shocking thing is how that was closer to the default behavior, and it seemed that Grok needed very, very little encouragement or user prompting to start behaving in the way that it did,' Hansen said.
The latest incident has echoes of problems that plagued Microsoft's Tay chatbot in 2016, which began producing racist and offensive posts before it was disabled, noted Julia Stoyanovich, a computer science professor at New York University and director of the Center for Responsible AI.
'This was almost 10 years ago, and the technology behind Grok is different from the technology behind Tay, but the problem is similar: hate speech moderation is a difficult problem that is bound to occur if it's not deliberately safeguarded against,' Stoyanovich said in a statement to The Hill.
She suggested xAI had failed to take the necessary steps to prevent hate speech.
'Importantly, the kinds of safeguards one needs are not purely technical, we cannot 'solve' hate speech,' Stoyanovich added. 'This needs to be done through a combination of technical solutions, policies, and substantial human intervention and oversight. Implementing safeguards takes planning and it takes substantial resources.'
MacKenzie underscored that speech outputs are 'incredibly hard' to regulate and instead pointed to a national framework for testing and transparency as a potential solution.
'At the end of the day, what we're concerned about is a model that shares the goals of Hitler, not just shares hate speech online, but is designed and weighted to support racist outcomes,' MacKenzie said.
In a January report evaluating various frontier AI models on transparency, ARI ranked Grok the lowest, with a score of 19.4 out of 100.
While xAI now releases its system prompts, the company notably does not produce system cards for its models. System cards, which are offered by most major AI developers, provide information about how an AI model was developed and tested.
AI startup Anthropic proposed its own transparency framework for frontier AI models last week, suggesting the largest developers should be required to publish system cards, in addition to secure development frameworks detailing how they assess and mitigate major risks.
'Grok's recent hate-filled tirade is just one more example of how AI systems can quickly become misaligned with human values and interests,' said Brendan Steinhauser, CEO of The Alliance for Secure AI, a nonprofit that aims to mitigate the risks from AI.
'These kinds of incidents will only happen more frequently as AI becomes more advanced,' he continued in a statement. 'That's why all companies developing advanced AI should implement transparent safety standards and release their system cards. A collaborative and open effort to prevent misalignment is critical to ensuring that advanced AI systems are infused with human values.'
Copyright 2025 Nexstar Media, Inc. All rights reserved. This material may not be published, broadcast, rewritten, or redistributed.
Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

This EV has a face, and it talks back with AI
This EV has a face, and it talks back with AI

Fox News

time22 minutes ago

  • Fox News

This EV has a face, and it talks back with AI

Walking up to your car and seeing it recognize you, light up with a digital smile and respond to your voice used to be something only seen in TV shows and movies. Now, LA-based Faraday Future is making that experience a reality. At its California headquarters, the company recently unveiled the FX Super One, a tech-packed electric vehicle featuring the F.A.C.E., short for Front AI Communication Ecosystem. This expressive LED grille gives the car personality, allowing it to connect through light, sound and even emotional cues. Sign up for my FREE CyberGuy ReportGet my best tech tips, urgent security alerts and exclusive deals delivered straight to your inbox. Plus, you'll get instant access to my Ultimate Scam Survival Guide — free when you join my The FX Super One is an electric MPV designed with flexibility and intelligence at its core. It offers two types of powertrains, including a fully electric option and a hybrid setup known as AI Hybrid Extended Range. All-wheel drive comes standard, and the wheelbase is extended to create what Faraday calls Magic Space, a spacious interior that adapts to various needs. Seating configurations vary, but one standout version, called GOAT, features four seats, including luxurious zero-gravity recliners in the back. These are paired with a wide panoramic display and even a built-in refrigerator. Other versions can seat six or seven passengers and include smart cabin lighting, high-end materials and a fully digital infotainment system designed for interaction. Instead of just lighting up like a typical grille, the F.A.C.E. uses AI to sense people nearby and respond with dynamic visuals. It listens to voice commands and reacts with animations that can show emotion or deliver information. It's powered by Faraday's Embodied Intelligence Agent system, which links the grille with other AI-driven features inside the car. The system works when the vehicle is parked and is designed to create a sense of connection between the driver and the machine. Voice recognition, camera input and learned behavior all feed into how the car responds. Over time, it adjusts its reactions based on how you use it. You might see calming patterns after a long day or playful responses when you greet it in the morning. The F.A.C.E. is only part of the story. The FX Super One also comes loaded with smart safety tools powered by the same AI system. These include a 360-degree sensor suite that combines cameras, radar and lidar for full environmental awareness. The cabin features a quick-starting operating system that lets passengers customize their experience on the go. As for driving, the car's architecture supports intelligent assistance that helps reduce risk and make decisions based on both data and context. The goal isn't just automation, it's collaboration between the driver and the vehicle. More than 10,000 people have already pre-ordered the FX Super One, indicating rapidly growing interest in the vehicle. Faraday Future plans to start production in late 2025 at its Hanford, California, plant. This aggressive timeline demonstrates the brand's renewed drive to gain momentum in the competitive EV market. With the FX Super One, Faraday Future delivers bold innovations that go beyond speed, range or luxury. The company envisions a new era of mobility defined by personality and human connection. Faraday Future is targeting a price for the FX Super One of under $100,000. You can reserve your FX Super One now with a refundable $100 deposit on Faraday Future's official website. The FX Super One feels like more than an electric vehicle. It brings personality to the road with its talking grille, adaptive cabin and emotionally aware AI. Faraday Future is leaning into connection and creativity, offering a car that responds to your presence and remembers how you interact with it. Whether the F.A.C.E. becomes a must-have feature or just a fun extra, it adds something fresh to the EV space. Would you drive a car that smiles at you and knows when you're having a bad day? Let us know by writing us at Sign up for my FREE CyberGuy ReportGet my best tech tips, urgent security alerts and exclusive deals delivered straight to your inbox. Plus, you'll get instant access to my Ultimate Scam Survival Guide — free when you join my Copyright 2025 All rights reserved.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store