logo
Grok controversies raise questions about moderating, regulating AI content

Grok controversies raise questions about moderating, regulating AI content

Yahoo16-07-2025
Elon Musk's artificial intelligence (AI) chatbot Grok has been plagued by controversy recently over its responses to users, raising questions about how tech companies seek to moderate content from AI and whether Washington should play a role in setting guidelines.
Grok faced sharp scrutiny last week, after an update prompted the AI chatbot to produce antisemitic responses and praise Adolf Hitler. Musk's AI company, xAI, quickly deleted numerous incendiary posts and said it added guardrails to 'ban hate speech' from the chatbot.
Just days later, xAI unveiled its newest version of Grok, which Musk claimed was the 'smartest AI model in the world.' However, users soon discovered that the chatbot appeared to be relying on its owner's views to respond to controversial queries.
'We should be extremely concerned that the best performing AI model on the market is Hitler-aligned. That should set off some alarm bells for folks,' Chris MacKenzie, vice president of communications at Americans for Responsible Innovation (ARI), an advocacy group focused on AI policy.
'I think that we're at a period right now, where AI models still aren't incredibly sophisticated,' he continued. 'They might have access to a lot of information, right. But in terms of their capacity for malicious acts, it's all very overt and not incredibly sophisticated.'
'There is a lot of room for us to address this misaligned behavior before it becomes much more difficult and much more harder to detect,' he added.
Lucas Hansen, co-founder of the nonprofit CivAI, which aims to provide information about AI's capabilities and risks, said it was 'not at all surprising' that it was possible to get Grok to behave the way it did.
'For any language model, you can get it to behave in any way that you want, regardless of the guardrails that are currently in place,' he told The Hill.
Musk announced last week that xAI had updated Grok, after he previously voiced frustrations with some of the chatbot's responses.
In mid-June, the tech mogul took issue with a response from Grok suggesting that right-wing violence had become more frequent and deadly since 2016. Musk claimed the chatbot was 'parroting legacy media' and said he was 'working on it.'
He later indicated he was retraining the model and called on users to help provide 'divisive facts,' which he defined as 'things that are politically incorrect, but nonetheless factually true.'
The update caused a firestorm for xAI, as Grok began making broad generalizations about people with Jewish last names and perpetuating antisemitic stereotypes about Hollywood.
The chatbot falsely suggested that people with 'Ashkenazi surnames' were pushing 'anti-white hate' and that Hollywood was advancing 'anti-white stereotypes,' which it later implied was the result of Jewish people being overrepresented in the industry. It also reportedly produced posts praising Hitler and referred to itself as 'MechaHitler.'
xAI ultimately deleted the posts and said it was banning hate speech from Grok. It later offered an apology for the chatbot's 'horrific behavior,' blaming the issue on 'update to a code path upstream' of Grok.
'The update was active for 16 [hours], in which deprecated code made @grok susceptible to existing X user posts; including when such posts contained extremist views,' xAI wrote in a post Saturday. 'We have removed that deprecated code and refactored the entire system to prevent further abuse.'
It identified several key prompts that caused Grok's responses, including one informing the chatbot it is 'not afraid to offend people who are politically correct' and another directing it to reflect the 'tone, context and language of the post' in its response.
xAI's prompts for Grok have been publicly available since May, when the chatbot began responding to unrelated queries with allegations of 'white genocide' in South Africa.
The company later said the posts were the result of an 'unauthorized modification' and vowed to make its prompts public in an effort to boost transparency.
Just days after the latest incident, xAI unveiled the newest version of its AI model, called Grok 4. Users quickly spotted new problems, in which the chatbot suggested its surname was 'Hitler' and referenced Musk's views when responding to controversial queries.
xAI explained Tuesday that Grok's searches had picked up on the 'MechaHitler' references, resulting in the chatbot's 'Hitler' surname response, while suggesting it had turned to Musk's views to 'align itself with the company.' The company said it has since tweaked the prompts and shared the details on GitHub.
'The kind of shocking thing is how that was closer to the default behavior, and it seemed that Grok needed very, very little encouragement or user prompting to start behaving in the way that it did,' Hansen said.
The latest incident has echoes of problems that plagued Microsoft's Tay chatbot in 2016, which began producing racist and offensive posts before it was disabled, noted Julia Stoyanovich, a computer science professor at New York University and director of the Center for Responsible AI.
'This was almost 10 years ago, and the technology behind Grok is different from the technology behind Tay, but the problem is similar: hate speech moderation is a difficult problem that is bound to occur if it's not deliberately safeguarded against,' Stoyanovich said in a statement to The Hill.
She suggested xAI had failed to take the necessary steps to prevent hate speech.
'Importantly, the kinds of safeguards one needs are not purely technical, we cannot 'solve' hate speech,' Stoyanovich added. 'This needs to be done through a combination of technical solutions, policies, and substantial human intervention and oversight. Implementing safeguards takes planning and it takes substantial resources.'
MacKenzie underscored that speech outputs are 'incredibly hard' to regulate and instead pointed to a national framework for testing and transparency as a potential solution.
'At the end of the day, what we're concerned about is a model that shares the goals of Hitler, not just shares hate speech online, but is designed and weighted to support racist outcomes,' MacKenzie said.
In a January report evaluating various frontier AI models on transparency, ARI ranked Grok the lowest, with a score of 19.4 out of 100.
While xAI now releases its system prompts, the company notably does not produce system cards for its models. System cards, which are offered by most major AI developers, provide information about how an AI model was developed and tested.
AI startup Anthropic proposed its own transparency framework for frontier AI models last week, suggesting the largest developers should be required to publish system cards, in addition to secure development frameworks detailing how they assess and mitigate major risks.
'Grok's recent hate-filled tirade is just one more example of how AI systems can quickly become misaligned with human values and interests,' said Brendan Steinhauser, CEO of The Alliance for Secure AI, a nonprofit that aims to mitigate the risks from AI.
'These kinds of incidents will only happen more frequently as AI becomes more advanced,' he continued in a statement. 'That's why all companies developing advanced AI should implement transparent safety standards and release their system cards. A collaborative and open effort to prevent misalignment is critical to ensuring that advanced AI systems are infused with human values.'
Copyright 2025 Nexstar Media, Inc. All rights reserved. This material may not be published, broadcast, rewritten, or redistributed.
Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

SoftBank Builds Nvidia, TSMC Stakes Under Son's Focus on AI Gear
SoftBank Builds Nvidia, TSMC Stakes Under Son's Focus on AI Gear

Bloomberg

time7 minutes ago

  • Bloomberg

SoftBank Builds Nvidia, TSMC Stakes Under Son's Focus on AI Gear

SoftBank Group Corp. is building up stakes in Nvidia Corp. and Taiwan Semiconductor Manufacturing Co., the latest reflection of Masayoshi Son's focus on the tools and hardware underpinning artificial intelligence. The Japanese technology investor raised its stake in Nvidia to about $3 billion by the end of March, up from $1 billion in the prior quarter, according to regulatory filings. It bought around $330 million worth of TSMC shares and $170 million in Oracle Corp., they show.

Under New Management
Under New Management

Entrepreneur

time36 minutes ago

  • Entrepreneur

Under New Management

Opinions expressed by Entrepreneur contributors are their own. You're reading Entrepreneur United Kingdom, an international franchise of Entrepreneur Media. Tech leadership used to be a boys' club with a secret handshake - an exclusive ritual women weren't part of. Now, women are in the room, sealing deals and closing million-pound rounds with a different kind of grip. Leaders like Lisa Miles-Heal, Danae Shell and Caroline Carruthers demonstrate that success in tech isn't just about innovation - it demands a fundamental reshaping of power, culture, and vision. They aren't simply changing who sits at the table - they're changing how the table itself is built. Lisa Miles-Heal's leadership sits at the intersection of two persistent fault lines in the tech sector: rapid innovation and entrenched gender inequality. As CEO of Silverfin, a UK based financial technology company, Miles-Heal operates in a space where women remain significantly underrepresented - particularly in senior roles within enterprise technology. Yet her presence at the helm challenges the long-standing norms of who gets to lead in tech, and how. In a sector still defined by structural hurdles, Miles-Heal's rise stands as an unmistakable challenge to the established order. "I wish more women knew that the skills and experience they bring to the table are as equally valuable as anyone else's," she says. "It's easy to fall into the trap of comparing yourself to the majority, thinking if you don't do things their way, you're doing it the wrong way. But your unique style and approach are your strengths. Your powerful differentiators will serve you well." Miles-Heal's leadership journey is a quiet rejection of conformity - and a celebration of difference. "Starting a business takes resilience and belief in yourself. One of the best ways to build that self-belief is by finding role models who show you different paths to success - people who prove it's okay to lead authentically." Authenticity, for her, isn't just a management strategy; it's survival. "My inspiration comes from other female leaders. The CEO of Visma is one - she came from a HR background, which is atypical for a tech CEO. Another role model is Dawn Marriott, who's shown me you can stay true to yourself even when moving from operator to investor. These women prove there's no one 'right' way to lead." That kind of perspective has helped guide Silverfin's rise - a company now regarded as one of the leading platforms transforming financial data workflows for accountants and finance teams. But when asked what decision made the biggest impact on the company's growth, her answer is disarmingly straightforward. "Growth is rarely driven by a single factor," she says. "But, I'd say gaining a much better handle on our core metrics was crucial. Having clear visibility into your numbers provides you with the clarity to identify exactly what is needed for focus and improvement." The focus on fundamentals isn't just a business cliché. It's a discipline. "Without being clear on your metrics - whether that's ARR growth, EBITDA, or customer-related numbers - you're at a serious disadvantage. You can run a business on gut feel, experience, and a strong product vision, but without data, you can't be sure if what you're doing is actually moving the needle." The key metrics Miles-Heal watches most closely are telling: employee engagement scores (eNPS), employee numbers (EMPs), customer net promoter scores (NPS), ARR growth, and EBITDA. These give a rounded picture of Silverfin's business health and growth potential. But metrics alone don't define culture - or leadership. When asked what assumption about leading in tech she's found to be untrue, the CEO doesn't hesitate. "One big assumption I've found is that you must have a technical background to be a leader in tech," she says. "That's quite a funny assumption because the landscape is shifting so much. With the rise of AI and evolving tech, leadership is circling back to focus less on coding skills and more on understanding the problem you're solving and communicating clearly - whether that's with people or AI tools." It's a shift that could - and should - rewrite the rules of who gets to lead. "What's becoming absolutely vital are the soft skills and emotional intelligence. These are the skills that women often excel in, and I like to think AI could be a feminist intervention shifting the power towards these strengths. Imagine a world where women lead because the future of tech values empathy, communication, and nuanced understanding as much as technical know-how. That would be powerful and honestly, I hope it becomes reality." That imagined future is something Miles-Heal is actively building - one initiative, one hiring policy, one conversation at a time. "It starts with yourself - building confidence and being your own best advocate. From there, it's about supporting the people closest to you - friends, family, and colleagues - and being a positive role model." Internally at Silverfin, progress has been tangible. "We've achieved 50% gender equity both in our leadership team and on our board. We've also taken practical steps like launching a women's network and peer-to-peer mentoring programme where women mentor other women, which is separate from our broader mentoring initiatives." That peer support is mirrored externally, too. "Looking beyond the business, we actively advocate for women in our customer base by hosting forums in Belgium and the UK for senior women in accounting technology." Representation, she notes, isn't an afterthought. It's deliberate. "We also support women-first events like the Women's Tour de France and the Women's Rugby World Cup, consciously choosing to prioritise female-focused activities and ensuring gender balance in speakers at our events and paid participation." Miles Heals's also conscious that progress sometimes means rewriting the rules that have gone unnoticed for too long. "A big focus for me personally has been pushing for gender-neutral language across tech and accounting. This means actively challenging the use of long-standing phrases like 'man days' or defaulting to 'guys' when referring to groups of people." "It's a small but meaningful way to foster inclusivity and help change the culture for the next generation of women in tech." If there's a thesis to Miles-Heal's leadership, it might be this: change begins close to home, but its impact can ripple outward. "Overall, I believe creating opportunities is about focusing on the immediate environment first - yourself, your company - and then gradually expanding your impact outward, like ripples in a pond." Joining this chorus of female tech leadership is Danae Shell, co-founder and CEO of Valla, a workplace dispute resolution platform. Named one of the UK's Top 50 Female Entrepreneurs by Innovate UK, Shell brings a complementary perspective on women in tech and entrepreneurship. "I wish more women knew that they can start a business in tech - many women I've met assume that they need deep technical knowledge to start a tech business, and with Generative AI lowering technical barriers every day, this has never been more untrue," Shell says. "I want women to know that the tech industry needs their deep expertise in their own fields, and that the technology part is getting easier and easier every day." Her decision to "go all-in on AI" in 2023 marked a pivotal moment for Valla's growth. "We saw Generative AI emerge and knew immediately that it was going to be a game-changer for the industry we were in, legaltech. I set the team the goal of delivering 'one unit of magic' using AI - the tech was so new that we didn't even know what it could and couldn't do yet. The company became an AI R&D department overnight and we unlocked huge amounts of value that are now built into our most popular products." For Shell, the toughest part of leadership hasn't been the fast pace or the demanding workload, but rather "managing myself and making consistently good decisions even under high uncertainty and stressful situations. It's a tough mix of trusting your gut and staying open to challenge, especially when things are moving faster and faster." On fostering the next generation of female entrepreneurs, Shell recalls how mentorship shaped her journey. "A lot of women gave their time to help me get where I am today - I remember one well-known entrepreneur I had never met called me in the few minutes before a flight to give me some critical advice, which was exactly what I needed at the time." Following their lead, she now actively supports groups like AccelerateHer and Pathways Forward, "to give support, large and small, to the next cohort of female entrepreneurs." Caroline Carruthers, CEO of Carruthers and Jackson, a Darlington based data consultancy, adds yet another dimension to the conversation about women in tech leadership. "I wish more women knew how supportive other female business leaders will be. This is something that really surprised me when I started Carruthers and Jackson, especially as the support I received even came from women who ran companies that were competitive to mine." She acknowledges the challenges of being a female founder in tech, but stresses solidarity. "It's still hard to be a female founder, especially in an industry like technology, which is why it's so valuable that we offer each other a lot of support. Rather than being adversarial, we want more women to join our world." Reflecting on her own growth, Carruthers reveals the mindset shift that propelled her forward. "My company's growth really took off when I realised the need to toughen up. It's impossible to please everybody and to be a business leader you must be able to make tough decisions and, crucially, be comfortable with the decisions you make. You can't keep saying sorry and you can't let the decisions you make keep you awake at night. Without this mindset, your business isn't going to grow." She also emphasises the strategic use of networks. "Another important decision I made was to draw on my network, but to do it in the right way. There is a perception that your network doesn't want to be put upon, but really, it's about giving them an option - teaching people what you do and letting them decide if it's right for them." Carruthers challenges a common myth about leadership: "I assumed that when you get to a certain position in your career, you stop suffering from imposter syndrome, but this isn't the case. No matter where you are, everyone still suffers, so this is something you need to learn to work with." Her commitment to nurturing future talent is clear. "Carruthers and Jackson runs an annual Summer School for Data Leaders, a free programme designed to give future data leaders the skills they need to succeed. While the course is open to men and women, we have so many promising female students every year, and I feel honoured to be able to help them along their paths." Beyond that, she mentors through Women in Data, "a community focused on advocating for and supporting women in the data field. As well as mentoring, within these sessions, I try to coach future female leaders on working with their strengths and making the best career for themselves." Before starting out, she suggests speaking to someone who's done it before - a step she wishes she'd taken herself. "I've learned so much since starting my businesses that I didn't even know I needed to know. If I'd have spoken to my peers, some of these learnings wouldn't have come as such a painful surprise!" Miles-Heal, Shell, and Carruthers are driving a new era of tech leadership. Their insights highlight not just the challenges women face, but the power of community, courage, and conviction to transform an industry long overdue for change.

Norton Adds Audio and Visual Deepfake Protection on Mobile
Norton Adds Audio and Visual Deepfake Protection on Mobile

Yahoo

time40 minutes ago

  • Yahoo

Norton Adds Audio and Visual Deepfake Protection on Mobile

Analysis of AI-generated audio and facial manipulation techniques now available in early access phase through Norton Genie Scam Protection TEMPE, Ariz. and PRAGUE, July 31, 2025 /PRNewswire/ -- Norton, a leader in Cyber Safety and part of Gen (NASDAQ: GEN), has launched AI-powered deepfake protection in the Norton Genie AI Assistant on Norton 360 mobile apps. Currently in early access phase, Norton Deepfake Protection enables people to defend themselves from malicious AI-generated audio and video content. Initially available on select Microsoft Copilot+ PCs, people can now protect themselves not only from everyday scams, but also AI scams even if they don't have an AI PC. Norton Deepfake Protection in the Genie AI Assistant includes the ability to analyze audio and visual content for signs of manipulation. Beyond detecting AI-generated voices used in fraudulent schemes, the feature provides an added layer of contextual protection by spotting inconsistencies or faint deformations in the physical features of people appearing in videos. If a harmful deepfake is detected, the Genie AI Assistant will provide conversational Cyber Safety guidance and suggestions on what to do next. "As AI-generated voices and faces become harder to distinguish from the real thing, trust is rapidly becoming one of the most fragile elements of our digital lives," said Leena Elias, Chief Product Officer at Gen. "The line between truth and deception is blurring, especially when malicious actors can abuse AI to create scams that replicate voices and imagery with startling realism. This is why we've made our deepfake protection accessible to people who don't have AI hardware, so they can confidently navigate and consume digital content without second-guessing what they see or hear." Currently, Norton Deepfake Protection in Genie Scam Protection supports English-language YouTube videos with plans to expand platform and language support in future updates. To check for signs of scams in video content, people can upload YouTube links to Norton Genie AI Assistant and receive real-time guidance on the authenticity of a video. If malicious AI-generated content is found, the Genie AI Assistant will flag it and provide advice on what to do next. The early access version of Norton Deepfake Protection is currently available in Norton 360 mobile products in the US, UK, Australia, and New Zealand, across Android and iOS devices with desktop support coming soon. The focus at Norton on AI-powered scam protection won't stop here. Later in the year, deepfake protection for AI PCs will extend to devices powered by Intel chipsets, and more advanced detection capabilities will be available on both desktop and mobile platforms. For more information, visit About NortonNorton is a leader in Cyber Safety, and part of Gen (NASDAQ: GEN), a global company dedicated to powering Digital Freedom with a family of trusted consumer brands. Norton empowers millions of individuals and families with award-winning protection for their devices, online privacy, and identity. Norton products and services are certified by independent testing organizations including AV-TEST, AV-Comparatives, and SE Labs. Norton is a founding member of the Global Anti-Scam Alliance North America Chapter and the Coalition Against Stalkerware. Learn more at and Mike Healey Courtney Rowles Gen Edelman for Gen press@ View original content to download multimedia: SOURCE Gen Digital Inc. Error in retrieving data Sign in to access your portfolio Error in retrieving data Error in retrieving data Error in retrieving data Error in retrieving data

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store