logo
#

Latest news with #Innovation

Grok controversies raise questions about moderating, regulating AI content
Grok controversies raise questions about moderating, regulating AI content

Yahoo

time8 hours ago

  • Business
  • Yahoo

Grok controversies raise questions about moderating, regulating AI content

Elon Musk's artificial intelligence (AI) chatbot Grok has been plagued by controversy recently over its responses to users, raising questions about how tech companies seek to moderate content from AI and whether Washington should play a role in setting guidelines. Grok faced sharp scrutiny last week, after an update prompted the AI chatbot to produce antisemitic responses and praise Adolf Hitler. Musk's AI company, xAI, quickly deleted numerous incendiary posts and said it added guardrails to 'ban hate speech' from the chatbot. Just days later, xAI unveiled its newest version of Grok, which Musk claimed was the 'smartest AI model in the world.' However, users soon discovered that the chatbot appeared to be relying on its owner's views to respond to controversial queries. 'We should be extremely concerned that the best performing AI model on the market is Hitler-aligned. That should set off some alarm bells for folks,' Chris MacKenzie, vice president of communications at Americans for Responsible Innovation (ARI), an advocacy group focused on AI policy. 'I think that we're at a period right now, where AI models still aren't incredibly sophisticated,' he continued. 'They might have access to a lot of information, right. But in terms of their capacity for malicious acts, it's all very overt and not incredibly sophisticated.' 'There is a lot of room for us to address this misaligned behavior before it becomes much more difficult and much more harder to detect,' he added. Lucas Hansen, co-founder of the nonprofit CivAI, which aims to provide information about AI's capabilities and risks, said it was 'not at all surprising' that it was possible to get Grok to behave the way it did. 'For any language model, you can get it to behave in any way that you want, regardless of the guardrails that are currently in place,' he told The Hill. Musk announced last week that xAI had updated Grok, after he previously voiced frustrations with some of the chatbot's responses. In mid-June, the tech mogul took issue with a response from Grok suggesting that right-wing violence had become more frequent and deadly since 2016. Musk claimed the chatbot was 'parroting legacy media' and said he was 'working on it.' He later indicated he was retraining the model and called on users to help provide 'divisive facts,' which he defined as 'things that are politically incorrect, but nonetheless factually true.' The update caused a firestorm for xAI, as Grok began making broad generalizations about people with Jewish last names and perpetuating antisemitic stereotypes about Hollywood. The chatbot falsely suggested that people with 'Ashkenazi surnames' were pushing 'anti-white hate' and that Hollywood was advancing 'anti-white stereotypes,' which it later implied was the result of Jewish people being overrepresented in the industry. It also reportedly produced posts praising Hitler and referred to itself as 'MechaHitler.' xAI ultimately deleted the posts and said it was banning hate speech from Grok. It later offered an apology for the chatbot's 'horrific behavior,' blaming the issue on 'update to a code path upstream' of Grok. 'The update was active for 16 [hours], in which deprecated code made @grok susceptible to existing X user posts; including when such posts contained extremist views,' xAI wrote in a post Saturday. 'We have removed that deprecated code and refactored the entire system to prevent further abuse.' It identified several key prompts that caused Grok's responses, including one informing the chatbot it is 'not afraid to offend people who are politically correct' and another directing it to reflect the 'tone, context and language of the post' in its response. xAI's prompts for Grok have been publicly available since May, when the chatbot began responding to unrelated queries with allegations of 'white genocide' in South Africa. The company later said the posts were the result of an 'unauthorized modification' and vowed to make its prompts public in an effort to boost transparency. Just days after the latest incident, xAI unveiled the newest version of its AI model, called Grok 4. Users quickly spotted new problems, in which the chatbot suggested its surname was 'Hitler' and referenced Musk's views when responding to controversial queries. xAI explained Tuesday that Grok's searches had picked up on the 'MechaHitler' references, resulting in the chatbot's 'Hitler' surname response, while suggesting it had turned to Musk's views to 'align itself with the company.' The company said it has since tweaked the prompts and shared the details on GitHub. 'The kind of shocking thing is how that was closer to the default behavior, and it seemed that Grok needed very, very little encouragement or user prompting to start behaving in the way that it did,' Hansen said. The latest incident has echoes of problems that plagued Microsoft's Tay chatbot in 2016, which began producing racist and offensive posts before it was disabled, noted Julia Stoyanovich, a computer science professor at New York University and director of the Center for Responsible AI. 'This was almost 10 years ago, and the technology behind Grok is different from the technology behind Tay, but the problem is similar: hate speech moderation is a difficult problem that is bound to occur if it's not deliberately safeguarded against,' Stoyanovich said in a statement to The Hill. She suggested xAI had failed to take the necessary steps to prevent hate speech. 'Importantly, the kinds of safeguards one needs are not purely technical, we cannot 'solve' hate speech,' Stoyanovich added. 'This needs to be done through a combination of technical solutions, policies, and substantial human intervention and oversight. Implementing safeguards takes planning and it takes substantial resources.' MacKenzie underscored that speech outputs are 'incredibly hard' to regulate and instead pointed to a national framework for testing and transparency as a potential solution. 'At the end of the day, what we're concerned about is a model that shares the goals of Hitler, not just shares hate speech online, but is designed and weighted to support racist outcomes,' MacKenzie said. In a January report evaluating various frontier AI models on transparency, ARI ranked Grok the lowest, with a score of 19.4 out of 100. While xAI now releases its system prompts, the company notably does not produce system cards for its models. System cards, which are offered by most major AI developers, provide information about how an AI model was developed and tested. AI startup Anthropic proposed its own transparency framework for frontier AI models last week, suggesting the largest developers should be required to publish system cards, in addition to secure development frameworks detailing how they assess and mitigate major risks. 'Grok's recent hate-filled tirade is just one more example of how AI systems can quickly become misaligned with human values and interests,' said Brendan Steinhauser, CEO of The Alliance for Secure AI, a nonprofit that aims to mitigate the risks from AI. 'These kinds of incidents will only happen more frequently as AI becomes more advanced,' he continued in a statement. 'That's why all companies developing advanced AI should implement transparent safety standards and release their system cards. A collaborative and open effort to prevent misalignment is critical to ensuring that advanced AI systems are infused with human values.' Copyright 2025 Nexstar Media, Inc. All rights reserved. This material may not be published, broadcast, rewritten, or redistributed.

Engineering Transformation Through Applied Innovation: How Gnanendra M Reddy Advances DevOps, AI, and Cloud-Native Compliance
Engineering Transformation Through Applied Innovation: How Gnanendra M Reddy Advances DevOps, AI, and Cloud-Native Compliance

India.com

time9 hours ago

  • Business
  • India.com

Engineering Transformation Through Applied Innovation: How Gnanendra M Reddy Advances DevOps, AI, and Cloud-Native Compliance

In the evolving world of enterprise-scale cloud computing and DevOps transformation, few professionals exhibit the same practical command and foresight as Gnanendra M Reddy. With a career grounded in cloud migration, containerization, CI/CD automation, Cyber Securityand governance frameworks, Gnanendra's contributions address the critical needs for scalability, compliance, and operational efficiency. His expertise combines hands-on technical insight with strategic system design, a balance that sets him apart as a thought leader in cloud-native and DevOps spaces. Gnanendra's research explores key themes in DevOps enablement and cloud-native modernization. His work bridges production-level insights with sustainable system models, creating tested patterns for infrastructure modernization, all aligned with domain-specific needs and system behaviour. Enabling CI/CD Efficiency in R&D-Driven DevOps Environments Published in the Essex Journal of AI Ethics and Responsible Innovation, vol. 3, 2023, Gnanendra's paper, 'Agile and DevOps Transformation in Large-Scale R&D Centres: A Case Study on CI/CD Efficiency Gains', explores the operational challenges faced by R&D organizations. The study addresses deployment bottlenecks in environments with high experimental churn and asynchronous delivery timelines. By applying his expertise in Terraform, GitLab automation, and Kubernetes orchestration, Gnanendra designed modular CI/CD blueprints that supported isolated experimentation without compromising pipeline stability. 'By introducing controlled stages and approval checkpoints based on artifact type and criticality, we preserved delivery agility while enhancing governance,' Gnanendra states in the paper. His solution unified delivery rhythms across teams, maintaining research integrity while incorporating compliance checkpoints, reusable templates, and container scanning stages—all aligned with pipeline maturity and integration frequency. His work demonstrates the ability to scale DevOps philosophy without rigidity, merging innovation with structured automation. Automating Scalable Data Engineering with AI in Cloud Ecosystems In The Newark Journal of Human-Centric AI and Robotics Interaction, vol. 3, pp. 182–223, 2023, Gnanendra co-authored 'AI-Powered Data Engineering: Automating ETL Pipelines for Scalable Cloud Analytics'. The paper discusses how AI can optimize and automate ETL processes across complex cloud environments. Addressing challenges such as diverse datasets and fluctuating data volumes, Gnanendra designed a dynamic orchestration framework combining pipeline inference, metadata tagging, and adaptive scheduling. 'Automation in data engineering must respect the fluidity of data behaviour while maintaining consistency in lineage and validation,' Gnanendra explains in the article. His solution used reinforcement logic to determine pipeline run triggers, reducing idle cycles and improving throughput. The AI-based classifiers predicted data readiness based on historical trends, ensuring timely processing. The system's strength lies in its dynamic Directed Acyclic Graph (DAG) optimization, customized for domain-specific workload patterns. This work reflects Gnanendra's ability to integrate AI into orchestration, optimizing both data engineering constraints and intelligent pipeline mechanics. His approach resulted in measurable improvements in data readiness and orchestration efficiency. Securing Cloud-Native Compliance Through Containerized Migration In the American Journal of Data Science and Artificial Intelligence Innovations, vol. 2, pp. 147–186, 2022, Gnanendra's research titled 'Ensuring Compliance in Cloud-Native Deployments: Migrating VMware Tanzu Workloads to Azure Kubernetes Service (AKS)' addresses compliance enforcement during containerized workload migrations. Focused on large-scale migration projects, his work strikes a balance between achieving cloud-native scalability and maintaining regulatory traceability. 'Ensuring traceable compliance must begin at the deployment plan, not post-deployment,' Gnanendra emphasizes. His research outlines a layered migration process where workloads are containerized and migrated in waves based on classification and dependency mapping. By integrating policy-as-code modules into Kubernetes deployment processes, Gnanendra ensured both runtime security validation and post-deployment compliance reporting. The paper also highlights his use of GitOps workflows for controlled change propagation and audit logging. This architecture demonstrates his mastery of cloud security, operational scale, and compliance fidelity, solidifying Gnanendra's capability to lead secure and scalable migration efforts. Certifications: Deep Expertise in Cloud-Native and DevSecOps Domains Gnanendra's certifications underscore his expertise in cloud-native technologies, automation frameworks, and security compliance. At the core of his portfolio is his recognition as a Kubestronaut, a title earned through his advanced capabilities in Kubernetes architecture, security, and multi-cluster orchestration. This distinction sets him apart as an industry leader in orchestrating containerized applications at scale. His Kubernetes certifications include: Certified Kubernetes Security Specialist (CKS) Certified Kubernetes Administrator (CKA) Certified Kubernetes Developer (CKAD) Kubernetes and Cloud Native Associate (KCNA) Kubernetes and Cloud Security Associate (KCSA) Additionally, Gnanendra holds the HashiCorp Terraform Certified Associate (003) certification, demonstrating his ability to manage infrastructure as code across multi-cloud environments. Complementing his DevOps expertise, Gnanendra has earned credentials in compliance and platform governance: Certified CIS – Risk and Compliance Implementer (ServiceNow) Certified System Administrator – ServiceNow ITIL V3 Certified, confirming his expertise in structured service management across IT operations. These certifications strengthen Gnanendra's ability to lead end-to-end transformation initiatives. His qualifications cover everything from design and automation to security and compliance, ensuring comprehensive, scalable solutions. Transforming Challenges into Scalable Solutions Gnanendra's contributions, reflected in his research and certifications, showcase his ongoing ability to address enterprise-scale challenges and deliver secure, scalable solutions. His work combines domain expertise with an execution-first mindset, continuously shaping infrastructure strategies that are practical, reproducible, and ready for regulation. As cloud-native ecosystems evolve, Gnanendra's decisions, governance models, and automation frameworks remain relevant, impactful, and aligned with the demands of high-compliance, high-performance environments.

Apple's iPhone 17 Secret Is Out: Event Date & Huge Upgrades Leaked
Apple's iPhone 17 Secret Is Out: Event Date & Huge Upgrades Leaked

Geeky Gadgets

time10 hours ago

  • Geeky Gadgets

Apple's iPhone 17 Secret Is Out: Event Date & Huge Upgrades Leaked

Apple's much-anticipated iPhone 17 event is rumored to take place in early September 2025, with speculation pointing to September 9 or 10. This timing aligns with Apple's established tradition of unveiling its flagship products shortly after Labor Day. The event is expected to introduce significant updates to the iPhone lineup, alongside enhancements to other key products in Apple's ecosystem. For users, this means a range of exciting innovations in design, performance, and functionality, all aimed at delivering a more seamless and enhanced experience. The video below from Saran Byte gives us more details about the rumored iPhone 17 event. Watch this video on YouTube. iPhone 17 Series: Redefining Innovation The iPhone 17 series is set to introduce four models: – iPhone 17 – iPhone 17 Air – iPhone 17 Pro – iPhone 17 Pro Max These devices promise a combination of aesthetic refinements and functional advancements. Here's what you can expect: Design Enhancements: The iPhone 17 series will feature thinner bezels and larger screen sizes, ranging from 6.3 to 6.9 inches. These changes aim to create a more immersive display experience. Pro models will come with aluminum frames, offering a premium look and improved durability. The iPhone 17 series will feature thinner bezels and larger screen sizes, ranging from 6.3 to 6.9 inches. These changes aim to create a more immersive display experience. Pro models will come with aluminum frames, offering a premium look and improved durability. Camera Upgrades: A redesigned camera module will include a 48MP telephoto lens for sharper zoom capabilities and a 24MP front camera for enhanced selfies and video calls. These improvements cater to both casual users and photography enthusiasts. A redesigned camera module will include a 48MP telephoto lens for sharper zoom capabilities and a 24MP front camera for enhanced selfies and video calls. These improvements cater to both casual users and photography enthusiasts. Performance Boost: The A19 Pro chipset, exclusive to Pro models, promises faster processing speeds and improved energy efficiency. Additionally, vapor chamber cooling in Pro models will address overheating issues, making them ideal for demanding tasks such as gaming and video editing. These upgrades reflect Apple's commitment to delivering devices that balance innovative technology with user-centric design. Dynamic Island and iOS 26: A Smarter Interface Apple continues to refine its user experience with updates to the Dynamic Island feature, now enhanced through iOS 26. This latest iteration introduces several improvements designed to make interactions more intuitive and efficient: Improved Multitasking: Users can effortlessly manage notifications and switch between apps with greater ease, enhancing productivity. Users can effortlessly manage notifications and switch between apps with greater ease, enhancing productivity. Customization Options: iOS 26 offers expanded customization features, allowing users to tailor their device's interface to better suit their preferences. iOS 26 offers expanded customization features, allowing users to tailor their device's interface to better suit their preferences. Smoother Navigation: Streamlined interactions make everyday tasks more intuitive, making sure a seamless user experience. These refinements aim to make the iPhone 17 series not just a smartphone but an integral part of your daily life, offering personalized functionality and enhanced usability. Expanding the Apple Ecosystem The event is expected to go beyond the iPhone, with Apple unveiling updates across its broader product lineup. These announcements will likely include: Apple Watch: The SE 3, Series 11, and Ultra 3 models are anticipated to feature improved health tracking, longer battery life, and enhanced durability, catering to both casual users and fitness enthusiasts. The SE 3, Series 11, and Ultra 3 models are anticipated to feature improved health tracking, longer battery life, and enhanced durability, catering to both casual users and fitness enthusiasts. AirPods Pro 3: Enhanced sound quality and smarter integration with Apple devices promise a superior audio experience, making them a must-have for audiophiles. Enhanced sound quality and smarter integration with Apple devices promise a superior audio experience, making them a must-have for audiophiles. HomePod Mini 2: Upgraded sound performance and smarter features aim to create a more connected and immersive home environment. Upgraded sound performance and smarter features aim to create a more connected and immersive home environment. AirTags: Potential updates to Apple's tracking devices could further strengthen their functionality and integration within the Apple ecosystem. These updates underscore Apple's commitment to creating a seamless, interconnected experience across all its devices, making sure that users can enjoy a cohesive and efficient digital lifestyle. Pricing Adjustments and Production Shifts With the introduction of new features and technologies, reports suggest a $100 price increase across all iPhone 17 models. This adjustment is attributed to rising production costs and tariffs. While the higher price point may impact affordability for some, Apple aims to justify the increase through significant upgrades in design, performance, and functionality. In addition, Apple is diversifying its supply chain by shifting some iPhone production to India. This strategic move reflects the company's efforts to reduce dependency on a single region while maintaining quality. However, Pro models will continue to be manufactured primarily in China, ensuring consistent standards for Apple's premium offerings. AI Integration: The Future of Apple Intelligence Apple is quietly advancing its AI capabilities, integrating intelligent features across its devices to enhance usability and personalization. These innovations include: Smarter Siri: Enhanced voice recognition and contextual understanding enable more natural and efficient interactions. Enhanced voice recognition and contextual understanding enable more natural and efficient interactions. Advanced Photo Processing: AI-driven tools improve image quality and offer more sophisticated editing options, catering to photography enthusiasts. AI-driven tools improve image quality and offer more sophisticated editing options, catering to photography enthusiasts. Personalization: AI-powered features adapt to individual usage patterns, delivering tailored suggestions and functionality. While these AI-driven enhancements may not take center stage at the event, they represent a significant step forward in making Apple's devices more intuitive and user-friendly, aligning with the company's vision of seamless technology integration. What to Expect from Apple's 2025 Lineup The iPhone 17 series represents a bold step forward for Apple, addressing past criticisms while introducing meaningful upgrades in design, performance, and functionality. Beyond the iPhone, updates to the Apple Watch, AirPods, and HomePod Mini further strengthen Apple's ecosystem, offering users a more connected and seamless experience. As the event approaches, consider how these advancements align with your needs and preferences. Whether you're drawn to the improved cameras, faster processors, or refined user interface, Apple's 2025 lineup promises to deliver innovations that could redefine how you interact with technology. Here are additional guides from our expansive article library that you may find useful on iPhone 17. Source & Image Credit: SaranByte Filed Under: Apple, Apple iPhone, Top News Latest Geeky Gadgets Deals Disclosure: Some of our articles include affiliate links. If you buy something through one of these links, Geeky Gadgets may earn an affiliate commission. Learn about our Disclosure Policy.

BUiD organises 9th Doctoral Research Conference (BDRC)2025
BUiD organises 9th Doctoral Research Conference (BDRC)2025

Al Bawaba

time16 hours ago

  • Business
  • Al Bawaba

BUiD organises 9th Doctoral Research Conference (BDRC)2025

The British University in Dubai (BUiD) organised the 9th Doctoral Research Conference (BDRC 2025), continuing its commitment to supporting doctoral research and fostering collaboration across BUiD Doctoral Research Conference 2025 was held in a hybrid format, welcoming participation from both in-person and virtual attendees. Open to BUiD students and those from partner universities, the conference provided an opportunity for doctoral students to present and publish their work, engage in scholarly dialogue, and receive feedback from academic peers and Khalid Al Marri, Dean of Research commented, 'The Doctoral Research Conference is more than just a platform for academic exchange—it's a celebration of innovation, interdisciplinary collaboration, and the future of research. At BUiD, we are proud to support emerging scholars who are shaping knowledge and practice across diverse fields.'The Conference included several tracks such as Education, Business, Management & Law, Engineering, and IT. Featuring research in Artificial Intelligence, Civil and Electrical Engineering, Cybersecurity, Renewable Energy, Big Data, Robotics, Smart Infrastructure, Water & Sustainability, and BDRC 2025 emphasised interdisciplinary collaboration and innovation, providing a platform to explore the latest research trends, practical challenges, and breakthrough solutions across sectors. Participants gained valuable insights from keynote speeches, academic talks, and opportunities to connect with fellow researchers and Conference reinforces BUiD's commitment to fostering a dynamic research culture that both contributes to academic advancement and tackles regional and global challenges. The Conference confirms BUiD's mission to build a strong research culture that helps grow knowledge locally and worldwide.

Grok controversies raise questions about moderating, regulating AI content
Grok controversies raise questions about moderating, regulating AI content

The Hill

timea day ago

  • Business
  • The Hill

Grok controversies raise questions about moderating, regulating AI content

Elon Musk's artificial intelligence (AI) chatbot Grok has been plagued by controversy recently over its responses to users, raising questions about how tech companies seek to moderate content from AI and whether Washington should play a role in setting guidelines. Grok faced sharp scrutiny last week, after an update prompted the AI chatbot to produce antisemitic responses and praise Adolf Hitler. Musk's AI company, xAI, quickly deleted numerous incendiary posts and said it added guardrails to 'ban hate speech' from the chatbot. Just days later, xAI unveiled its newest version of Grok, which Musk claimed was the 'smartest AI model in the world.' However, users soon discovered that the chatbot appeared to be relying on its owner's views to respond to controversial queries. 'We should be extremely concerned that the best performing AI model on the market is Hitler-aligned. That should set off some alarm bells for folks,' Chris MacKenzie, vice president of communications at Americans for Responsible Innovation (ARI), an advocacy group focused on AI policy. 'I think that we're at a period right now, where AI models still aren't incredibly sophisticated,' he continued. 'They might have access to a lot of information, right. But in terms of their capacity for malicious acts, it's all very overt and not incredibly sophisticated.' 'There is a lot of room for us to address this misaligned behavior before it becomes much more difficult and much more harder to detect,' he added. Lucas Hansen, co-founder of the nonprofit CivAI, which aims to provide information about AI's capabilities and risks, said it was 'not at all surprising' that it was possible to get Grok to behave the way it did. 'For any language model, you can get it to behave in any way that you want, regardless of the guardrails that are currently in place,' he told The Hill. Musk announced last week that xAI had updated Grok, after he previously voiced frustrations with some of the chatbot's responses. In mid-June, the tech mogul took issue with a response from Grok suggesting that right-wing violence had become more frequent and deadly since 2016. Musk claimed the chatbot was 'parroting legacy media' and said he was 'working on it.' He later indicated he was retraining the model and called on users to help provide 'divisive facts,' which he defined as 'things that are politically incorrect, but nonetheless factually true.' The update caused a firestorm for xAI, as Grok began making broad generalizations about people with Jewish last names and perpetuating antisemitic stereotypes about Hollywood. The chatbot falsely suggested that people with 'Ashkenazi surnames' were pushing 'anti-white hate' and that Hollywood was advancing 'anti-white stereotypes,' which it later implied was the result of Jewish people being overrepresented in the industry. It also reportedly produced posts praising Hitler and referred to itself as 'MechaHitler.' xAI ultimately deleted the posts and said it was banning hate speech from Grok. It later offered an apology for the chatbot's 'horrific behavior,' blaming the issue on 'update to a code path upstream' of Grok. 'The update was active for 16 [hours], in which deprecated code made @grok susceptible to existing X user posts; including when such posts contained extremist views,' xAI wrote in a post Saturday. 'We have removed that deprecated code and refactored the entire system to prevent further abuse.' It identified several key prompts that caused Grok's responses, including one informing the chatbot it is 'not afraid to offend people who are politically correct' and another directing it to reflect the 'tone, context and language of the post' in its response. xAI's prompts for Grok have been publicly available since May, when the chatbot began responding to unrelated queries with allegations of 'white genocide' in South Africa. The company later said the posts were the result of an 'unauthorized modification' and vowed to make its prompts public in an effort to boost transparency. Just days after the latest incident, xAI unveiled the newest version of its AI model, called Grok 4. Users quickly spotted new problems, in which the chatbot suggested its surname was 'Hitler' and referenced Musk's views when responding to controversial queries. xAI explained Tuesday that Grok's searches had picked up on the 'MechaHitler' references, resulting in the chatbot's 'Hitler' surname response, while suggesting it had turned to Musk's views to 'align itself with the company.' The company said it has since tweaked the prompts and shared the details on GitHub. 'The kind of shocking thing is how that was closer to the default behavior, and it seemed that Grok needed very, very little encouragement or user prompting to start behaving in the way that it did,' Hansen said. The latest incident has echoes of problems that plagued Microsoft's Tay chatbot in 2016, which began producing racist and offensive posts before it was disabled, noted Julia Stoyanovich, a computer science professor at New York University and director of the Center for Responsible AI. 'This was almost 10 years ago, and the technology behind Grok is different from the technology behind Tay, but the problem is similar: hate speech moderation is a difficult problem that is bound to occur if it's not deliberately safeguarded against,' Stoyanovich said in a statement to The Hill. She suggested xAI had failed to take the necessary steps to prevent hate speech. 'Importantly, the kinds of safeguards one needs are not purely technical, we cannot 'solve' hate speech,' Stoyanovich added. 'This needs to be done through a combination of technical solutions, policies, and substantial human intervention and oversight. Implementing safeguards takes planning and it takes substantial resources.' MacKenzie underscored that speech outputs are 'incredibly hard' to regulate and instead pointed to a national framework for testing and transparency as a potential solution. 'At the end of the day, what we're concerned about is a model that shares the goals of Hitler, not just shares hate speech online, but is designed and weighted to support racist outcomes,' MacKenzie said. In a January report evaluating various frontier AI models on transparency, ARI ranked Grok the lowest, with a score of 19.4 out of 100. While xAI now releases its system prompts, the company notably does not produce system cards for its models. System cards, which are offered by most major AI developers, provide information about how an AI model was developed and tested. AI startup Anthropic proposed its own transparency framework for frontier AI models last week, suggesting the largest developers should be required to publish system cards, in addition to secure development frameworks detailing how they assess and mitigate major risks. 'Grok's recent hate-filled tirade is just one more example of how AI systems can quickly become misaligned with human values and interests,' said Brendan Steinhauser, CEO of The Alliance for Secure AI, a nonprofit that aims to mitigate the risks from AI. 'These kinds of incidents will only happen more frequently as AI becomes more advanced,' he continued in a statement. 'That's why all companies developing advanced AI should implement transparent safety standards and release their system cards. A collaborative and open effort to prevent misalignment is critical to ensuring that advanced AI systems are infused with human values.'

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store