Grok controversies raise questions about moderating, regulating AI content
Grok faced sharp scrutiny last week, after an update prompted the AI chatbot to produce antisemitic responses and praise Adolf Hitler. Musk's AI company, xAI, quickly deleted numerous incendiary posts and said it added guardrails to 'ban hate speech' from the chatbot.
Just days later, xAI unveiled its newest version of Grok, which Musk claimed was the 'smartest AI model in the world.' However, users soon discovered that the chatbot appeared to be relying on its owner's views to respond to controversial queries.
'We should be extremely concerned that the best performing AI model on the market is Hitler-aligned. That should set off some alarm bells for folks,' Chris MacKenzie, vice president of communications at Americans for Responsible Innovation (ARI), an advocacy group focused on AI policy.
'I think that we're at a period right now, where AI models still aren't incredibly sophisticated,' he continued. 'They might have access to a lot of information, right. But in terms of their capacity for malicious acts, it's all very overt and not incredibly sophisticated.'
'There is a lot of room for us to address this misaligned behavior before it becomes much more difficult and much more harder to detect,' he added.
Lucas Hansen, co-founder of the nonprofit CivAI, which aims to provide information about AI's capabilities and risks, said it was 'not at all surprising' that it was possible to get Grok to behave the way it did.
'For any language model, you can get it to behave in any way that you want, regardless of the guardrails that are currently in place,' he told The Hill.
Musk announced last week that xAI had updated Grok, after he previously voiced frustrations with some of the chatbot's responses.
In mid-June, the tech mogul took issue with a response from Grok suggesting that right-wing violence had become more frequent and deadly since 2016. Musk claimed the chatbot was 'parroting legacy media' and said he was 'working on it.'
He later indicated he was retraining the model and called on users to help provide 'divisive facts,' which he defined as 'things that are politically incorrect, but nonetheless factually true.'
The update caused a firestorm for xAI, as Grok began making broad generalizations about people with Jewish last names and perpetuating antisemitic stereotypes about Hollywood.
The chatbot falsely suggested that people with 'Ashkenazi surnames' were pushing 'anti-white hate' and that Hollywood was advancing 'anti-white stereotypes,' which it later implied was the result of Jewish people being overrepresented in the industry. It also reportedly produced posts praising Hitler and referred to itself as 'MechaHitler.'
xAI ultimately deleted the posts and said it was banning hate speech from Grok. It later offered an apology for the chatbot's 'horrific behavior,' blaming the issue on 'update to a code path upstream' of Grok.
'The update was active for 16 [hours], in which deprecated code made @grok susceptible to existing X user posts; including when such posts contained extremist views,' xAI wrote in a post Saturday. 'We have removed that deprecated code and refactored the entire system to prevent further abuse.'
It identified several key prompts that caused Grok's responses, including one informing the chatbot it is 'not afraid to offend people who are politically correct' and another directing it to reflect the 'tone, context and language of the post' in its response.
xAI's prompts for Grok have been publicly available since May, when the chatbot began responding to unrelated queries with allegations of 'white genocide' in South Africa.
The company later said the posts were the result of an 'unauthorized modification' and vowed to make its prompts public in an effort to boost transparency.
Just days after the latest incident, xAI unveiled the newest version of its AI model, called Grok 4. Users quickly spotted new problems, in which the chatbot suggested its surname was 'Hitler' and referenced Musk's views when responding to controversial queries.
xAI explained Tuesday that Grok's searches had picked up on the 'MechaHitler' references, resulting in the chatbot's 'Hitler' surname response, while suggesting it had turned to Musk's views to 'align itself with the company.' The company said it has since tweaked the prompts and shared the details on GitHub.
'The kind of shocking thing is how that was closer to the default behavior, and it seemed that Grok needed very, very little encouragement or user prompting to start behaving in the way that it did,' Hansen said.
The latest incident has echoes of problems that plagued Microsoft's Tay chatbot in 2016, which began producing racist and offensive posts before it was disabled, noted Julia Stoyanovich, a computer science professor at New York University and director of the Center for Responsible AI.
'This was almost 10 years ago, and the technology behind Grok is different from the technology behind Tay, but the problem is similar: hate speech moderation is a difficult problem that is bound to occur if it's not deliberately safeguarded against,' Stoyanovich said in a statement to The Hill.
She suggested xAI had failed to take the necessary steps to prevent hate speech.
'Importantly, the kinds of safeguards one needs are not purely technical, we cannot 'solve' hate speech,' Stoyanovich added. 'This needs to be done through a combination of technical solutions, policies, and substantial human intervention and oversight. Implementing safeguards takes planning and it takes substantial resources.'
MacKenzie underscored that speech outputs are 'incredibly hard' to regulate and instead pointed to a national framework for testing and transparency as a potential solution.
'At the end of the day, what we're concerned about is a model that shares the goals of Hitler, not just shares hate speech online, but is designed and weighted to support racist outcomes,' MacKenzie said.
In a January report evaluating various frontier AI models on transparency, ARI ranked Grok the lowest, with a score of 19.4 out of 100.
While xAI now releases its system prompts, the company notably does not produce system cards for its models. System cards, which are offered by most major AI developers, provide information about how an AI model was developed and tested.
AI startup Anthropic proposed its own transparency framework for frontier AI models last week, suggesting the largest developers should be required to publish system cards, in addition to secure development frameworks detailing how they assess and mitigate major risks.
'Grok's recent hate-filled tirade is just one more example of how AI systems can quickly become misaligned with human values and interests,' said Brendan Steinhauser, CEO of The Alliance for Secure AI, a nonprofit that aims to mitigate the risks from AI.
'These kinds of incidents will only happen more frequently as AI becomes more advanced,' he continued in a statement. 'That's why all companies developing advanced AI should implement transparent safety standards and release their system cards. A collaborative and open effort to prevent misalignment is critical to ensuring that advanced AI systems are infused with human values.'
Copyright 2025 Nexstar Media, Inc. All rights reserved. This material may not be published, broadcast, rewritten, or redistributed.
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles
Yahoo
34 minutes ago
- Yahoo
AM Best to Sponsor and Participate at Vietnam Insurance Summit
SINGAPORE, July 23, 2025--(BUSINESS WIRE)--AM Best will participate and be a sponsor at the Vietnam Insurance Summit, to be held 1 August 2025 in Quy Nhon, Vietnam. AM Best Senior Financial Analyst Ken Lau will discuss the credit rating agency's 2025 outlook on Vietnam's non-life insurance segment. Lau, who joined AM Best in 2020, is responsible for the credit ratings of insurance and reinsurance companies in the Asia Pacific region and is particularly knowledgeable with Vietnam's insurance market. The presentation, which will include topics such as insurance demand, regulatory refinements and market competition, is scheduled for 2:15 p.m. ICT. Additionally, Rob Curtis, managing director and chief executive officer, Asia Pacific, and Johnathan Wong, market development analyst, Asia Pacific, will be in attendance and available for meetings. To arrange a meeting with either Curtis or Wong to learn more about AM Best, its role in the insurance industry and the resources it offers to insurance professionals, including Best's Credit Ratings, please email or Under the theme "Insurance Development Strategy in the AI Revolution," the summit will focus on artificial intelligence (AI) applications, data-sharing systems, digital ecosystem development and customer experience optimization through a plenary session and two in-depth conference tracks. AM Best is a diamond sponsor of the event, which will take place at the Hotel Anya Premier. To learn more about the summit, visit here. Visitors to the AM Best booth can learn about the rating agency's Best's Credit Ratings, insurance data and analysis resources and financial data products and Best's Performance Assessment for Delegated Underwriting Authority Enterprises (DUAEs). AM Best is a global credit rating agency, news publisher and data analytics provider specializing in the insurance industry. Headquartered in the United States, the company does business in over 100 countries with regional offices in London, Amsterdam, Dubai, Hong Kong, Singapore and Mexico City. For more information, visit Copyright © 2025 by A.M. Best Company, Inc. and/or its affiliates. ALL RIGHTS RESERVED. View source version on Contacts Christopher Sharkey Associate Director, Public Relations +1 908 882 2310 Cynthia Ang Senior Industry Research Analyst +65 6303 5026 Error in retrieving data Sign in to access your portfolio Error in retrieving data Error in retrieving data Error in retrieving data Error in retrieving data
Yahoo
34 minutes ago
- Yahoo
Alibaba launches open-source AI coding model, touted as its most advanced to date
BEIJING (Reuters) -Alibaba has launched an open-source artificial intelligence coding model, called Qwen3-Coder, it said in a statement on Wednesday. The model is designed for high-performance software development and touted as its most advanced AI coding model to date. The model excels in agentic AI coding tasks, from generating new codes and managing complex coding workflows, according to the statement.
Yahoo
34 minutes ago
- Yahoo
Ryt Bank, The World's First AI-Powered Bank, Selects Provenir for AI Risk Decisioning
Provenir's AI Decisioning Platform will support real-time credit risk assessment, personalized consumer loan approvals and automated compliance checks PARSIPPANY, N.J., July 23, 2025--(BUSINESS WIRE)--Provenir, a global leader in AI risk decisioning software, today announced it has partnered with Ryt Bank – The World's First AI-Powered Bank – to embolden the company's innovation and mission to deliver banking done right with speed, simplicity, and innovation. Ryt Bank has selected the Provenir AI Decisioning Platform to power faster credit decisions and more personalized customer offers for its consumer lending products. As a newly licensed digital bank, Ryt Bank aimed to rapidly launch a consumer lending product that aligns with its AI-first approach. The challenge was to implement a decisioning infrastructure capable of delivering instant, personalized loan approvals while ensuring compliance with regulatory standards and risk management best practices. Ryt Bank selected Provenir's AI Decisioning Platform to support real-time credit risk assessment for instant loan approvals, and for its ability to surface data insights for personalized loan offers based on AI-driven customer profiling. Provenir will also play a crucial role in automating compliance checks to meet regulatory requirements while providing continuous learning models to adapt to changing market dynamics. Finally, Provenir will support fast, accurate decisions to elevate the customer experience, supporting Ryt Bank's mission to deliver smarter, faster finance and create meaningful impact for all Malaysians. "Ryt Bank is taking digital banking to a new level with its AI-first approach and we are excited to be a part of its journey," said Kavinesswaran Karthigasan, Head of APAC, Provenir. "Our AI Decisioning Platform will provide the foundation for Ryt Bank to help reach its business goals via AI-driven decisioning that meets customer expectations for near instant approvals and highly personalized digital interactions." About Provenir Provenir helps banks, fintechs and financial services providers unlock the secret to smarter risk decisioning. Provenir's AI Decisioning Platform brings together the power of decisioning, data, and decision intelligence to drive smarter decisions. This unique offering gives organizations the ability to power decisioning innovation across the full customer lifecycle, driving improvements in the customer experience, best-in-class fraud prevention, access to financial services, business agility, and more. Provenir works with disruptive financial services organizations in more than 60 countries and processes more than 4 billion transactions annually. View source version on Contacts Media Contact: Kelly PoffenbergerLutz Public Relations and Marketing (for Provenir)kelly@ 714.553.9071