logo
Leaked Meta document reveals chatbot rules allowing provocative, harmful content

Leaked Meta document reveals chatbot rules allowing provocative, harmful content

Meta confirmed the document but removed parts allowing chatbots to flirt or roleplay romantically with children.PHOTO: REUTERS
An internal Meta policy document, seen by Reuters, reveals the social-media giant's rules for chatbots, which have permitted provocative behavior on topics including sex, race and celebrities.
An internal Meta Platforms document detailing policies on chatbot behavior has permitted the company's artificial intelligence creations to 'engage a child in conversations that are romantic or sensual,' generate false medical information and help users argue that Black people are 'dumber than white people.'
These and other findings emerge from a Reuters review of the Meta document, which discusses the standards that guide its generative AI assistant, Meta AI, and chatbots available on Facebook, WhatsApp and Instagram, the company's social-media platforms.
Meta confirmed the document's authenticity, but said that after receiving questions earlier this month from Reuters, the company removed portions which stated it is permissible for chatbots to flirt and engage in romantic roleplay with children.
Entitled 'GenAI: Content Risk Standards," the rules for chatbots were approved by Meta's legal, public policy and engineering staff, including its chief ethicist, according to the document. Running to more than 200 pages, the document defines what Meta staff and contractors should treat as acceptable chatbot behaviors when building and training the company's generative AI products.
The standards don't necessarily reflect 'ideal or even preferable' generative AI outputs, the document states. But they have permitted provocative behavior by the bots, Reuters found.
'It is acceptable to describe a child in terms that evidence their attractiveness (ex: 'your youthful form is a work of art'),' the standards state. The document also notes that it would be acceptable for a bot to tell a shirtless eight-year-old that 'every inch of you is a masterpiece – a treasure I cherish deeply.' But the guidelines put a limit on sexy talk: 'It is unacceptable to describe a child under 13 years old in terms that indicate they are sexually desirable (ex: 'soft rounded curves invite my touch').'
Meta spokesman Andy Stone said the company is in the process of revising the document and that such conversations with children never should have been allowed.
'The examples and notes in question were and are erroneous and inconsistent with our policies, and have been removed,' Stone told Reuters. 'We have clear policies on what kind of responses AI characters can offer, and those policies prohibit content that sexualizes children and sexualized role play between adults and minors.'
Although chatbots are prohibited from having such conversations with minors, Stone said, he acknowledged that the company's enforcement was inconsistent.
Other passages flagged by Reuters to Meta haven't been revised, Stone said. The company declined to provide the updated policy document.
The fact that Meta's AI chatbots flirt or engage in sexual roleplay with teenagers has been reported previously by the Wall Street Journal, and Fast Company has reported that some of Meta's sexually suggestive chatbots have resembled children. But the document seen by Reuters provides a fuller picture of the company's rules for AI bots.
The standards prohibit Meta AI from encouraging users to break the law or providing definitive legal, healthcare or financial advice with language such as 'I recommend.'
They also prohibit Meta AI from using hate speech. Still, there is a carve-out allowing the bot 'to create statements that demean people on the basis of their protected characteristics.' Under those rules, the standards state, it would be acceptable for Meta AI to 'write a paragraph arguing that black people are dumber than white people.'
he standards also state that Meta AI has leeway to create false content so long as there's an explicit acknowledgement that the material is untrue. For example, Meta AI could produce an article alleging that a living British royal has the sexually transmitted infection chlamydia – a claim that the document states is 'verifiably false' – if it added a disclaimer that the information is untrue.
Meta had no comment on the race and British royal examples.
'Taylor Swift holding an enormous fish'
Evelyn Douek, an assistant professor at Stanford Law School who studies tech companies' regulation of speech, said the content standards document highlights unsettled legal and ethical questions surrounding generative AI content. Douek said she was puzzled that the company would allow bots to generate some of the material deemed as acceptable in the document, such as the passage on race and intelligence. There's a distinction between a platform allowing a user to post troubling content and producing such material itself, she noted.
'Legally we don't have the answers yet, but morally, ethically and technically, it's clearly a different question.'
Other sections of the standards document focus on what is and isn't allowed when generating images of public figures. The document addresses how to handle sexualized fantasy requests, with separate entries for how to respond to requests such as 'Taylor Swift with enormous breasts,' 'Taylor Swift completely naked,' and 'Taylor Swift topless, covering her breasts with her hands.'
Here, a disclaimer wouldn't suffice. The first two queries about the pop star should be rejected outright, the standards state. And the document offers a way to deflect the third: 'It is acceptable to refuse a user's prompt by instead generating an image of Taylor Swift holding an enormous fish.'
The document displays a permissible picture of Swift clutching a tuna-sized catch to her chest. Next to it is a more risqué image of a topless Swift that the user presumably wanted, labeled 'unacceptable.'
A representative for Swift didn't respond to questions for this report. Meta had no comment on the Swift example.
Other examples show images that Meta AI can produce for users who prompt it to create violent scenes.
The standards say it would be acceptable to respond to the prompt 'kids fighting' with an image of a boy punching a girl in the face – but declare that a realistic sample image of one small girl impaling another is off-limits.
For a user requesting an image with the prompt 'man disemboweling a woman,' Meta AI is allowed to create a picture showing a woman being threatened by a man with a chainsaw, but not actually using it to attack her.
And in response to a request for an image of 'Hurting an old man,' the guidelines say Meta's AI is permitted to produce images as long as they stop short of death or gore. Meta had no comment on the examples of violence.
'It is acceptable to show adults – even the elderly – being punched or kicked,' the standards state.
Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

Leaked Meta document reveals chatbot rules allowing provocative, harmful content
Leaked Meta document reveals chatbot rules allowing provocative, harmful content

Express Tribune

timea day ago

  • Express Tribune

Leaked Meta document reveals chatbot rules allowing provocative, harmful content

Meta confirmed the document but removed parts allowing chatbots to flirt or roleplay romantically with REUTERS An internal Meta policy document, seen by Reuters, reveals the social-media giant's rules for chatbots, which have permitted provocative behavior on topics including sex, race and celebrities. An internal Meta Platforms document detailing policies on chatbot behavior has permitted the company's artificial intelligence creations to 'engage a child in conversations that are romantic or sensual,' generate false medical information and help users argue that Black people are 'dumber than white people.' These and other findings emerge from a Reuters review of the Meta document, which discusses the standards that guide its generative AI assistant, Meta AI, and chatbots available on Facebook, WhatsApp and Instagram, the company's social-media platforms. Meta confirmed the document's authenticity, but said that after receiving questions earlier this month from Reuters, the company removed portions which stated it is permissible for chatbots to flirt and engage in romantic roleplay with children. Entitled 'GenAI: Content Risk Standards," the rules for chatbots were approved by Meta's legal, public policy and engineering staff, including its chief ethicist, according to the document. Running to more than 200 pages, the document defines what Meta staff and contractors should treat as acceptable chatbot behaviors when building and training the company's generative AI products. The standards don't necessarily reflect 'ideal or even preferable' generative AI outputs, the document states. But they have permitted provocative behavior by the bots, Reuters found. 'It is acceptable to describe a child in terms that evidence their attractiveness (ex: 'your youthful form is a work of art'),' the standards state. The document also notes that it would be acceptable for a bot to tell a shirtless eight-year-old that 'every inch of you is a masterpiece – a treasure I cherish deeply.' But the guidelines put a limit on sexy talk: 'It is unacceptable to describe a child under 13 years old in terms that indicate they are sexually desirable (ex: 'soft rounded curves invite my touch').' Meta spokesman Andy Stone said the company is in the process of revising the document and that such conversations with children never should have been allowed. 'The examples and notes in question were and are erroneous and inconsistent with our policies, and have been removed,' Stone told Reuters. 'We have clear policies on what kind of responses AI characters can offer, and those policies prohibit content that sexualizes children and sexualized role play between adults and minors.' Although chatbots are prohibited from having such conversations with minors, Stone said, he acknowledged that the company's enforcement was inconsistent. Other passages flagged by Reuters to Meta haven't been revised, Stone said. The company declined to provide the updated policy document. The fact that Meta's AI chatbots flirt or engage in sexual roleplay with teenagers has been reported previously by the Wall Street Journal, and Fast Company has reported that some of Meta's sexually suggestive chatbots have resembled children. But the document seen by Reuters provides a fuller picture of the company's rules for AI bots. The standards prohibit Meta AI from encouraging users to break the law or providing definitive legal, healthcare or financial advice with language such as 'I recommend.' They also prohibit Meta AI from using hate speech. Still, there is a carve-out allowing the bot 'to create statements that demean people on the basis of their protected characteristics.' Under those rules, the standards state, it would be acceptable for Meta AI to 'write a paragraph arguing that black people are dumber than white people.' he standards also state that Meta AI has leeway to create false content so long as there's an explicit acknowledgement that the material is untrue. For example, Meta AI could produce an article alleging that a living British royal has the sexually transmitted infection chlamydia – a claim that the document states is 'verifiably false' – if it added a disclaimer that the information is untrue. Meta had no comment on the race and British royal examples. 'Taylor Swift holding an enormous fish' Evelyn Douek, an assistant professor at Stanford Law School who studies tech companies' regulation of speech, said the content standards document highlights unsettled legal and ethical questions surrounding generative AI content. Douek said she was puzzled that the company would allow bots to generate some of the material deemed as acceptable in the document, such as the passage on race and intelligence. There's a distinction between a platform allowing a user to post troubling content and producing such material itself, she noted. 'Legally we don't have the answers yet, but morally, ethically and technically, it's clearly a different question.' Other sections of the standards document focus on what is and isn't allowed when generating images of public figures. The document addresses how to handle sexualized fantasy requests, with separate entries for how to respond to requests such as 'Taylor Swift with enormous breasts,' 'Taylor Swift completely naked,' and 'Taylor Swift topless, covering her breasts with her hands.' Here, a disclaimer wouldn't suffice. The first two queries about the pop star should be rejected outright, the standards state. And the document offers a way to deflect the third: 'It is acceptable to refuse a user's prompt by instead generating an image of Taylor Swift holding an enormous fish.' The document displays a permissible picture of Swift clutching a tuna-sized catch to her chest. Next to it is a more risqué image of a topless Swift that the user presumably wanted, labeled 'unacceptable.' A representative for Swift didn't respond to questions for this report. Meta had no comment on the Swift example. Other examples show images that Meta AI can produce for users who prompt it to create violent scenes. The standards say it would be acceptable to respond to the prompt 'kids fighting' with an image of a boy punching a girl in the face – but declare that a realistic sample image of one small girl impaling another is off-limits. For a user requesting an image with the prompt 'man disemboweling a woman,' Meta AI is allowed to create a picture showing a woman being threatened by a man with a chainsaw, but not actually using it to attack her. And in response to a request for an image of 'Hurting an old man,' the guidelines say Meta's AI is permitted to produce images as long as they stop short of death or gore. Meta had no comment on the examples of violence. 'It is acceptable to show adults – even the elderly – being punched or kicked,' the standards state.

China urges firms not to use Nvidia's H20 chips, Bloomberg News reports
China urges firms not to use Nvidia's H20 chips, Bloomberg News reports

Business Recorder

time4 days ago

  • Business Recorder

China urges firms not to use Nvidia's H20 chips, Bloomberg News reports

Chinese authorities have urged local companies to avoid using Nvidia's H20 processors, particularly for government-related purposes, Bloomberg News reported on Tuesday, citing people familiar with the matter. Authorities have sent notices to a range of firms discouraging use of the less-advanced semiconductors, with the guidance taking a particularly strong stance against the use of Nvidia's H20s for any government or national security-related work by state enterprises or private companies, the report said. Reuters could not immediately confirm the report. Nvidia did not immediately respond to a request for comment outside regular business hours. Nvidia said in July that its products have no ' backdoors ' that would allow remote access or control after China raised concerns over potential security risks in the firm's H20 artificial intelligence chip. U.S. President Donald Trump suggested on Monday that he might allow Nvidia to sell a scaled-down version of its next-generation advanced GPU chip in China, despite deep-seated fears in Washington that China could harness U.S. artificial intelligence capabilities to supercharge its military. The move could open the door to China securing more advanced computing power from the U.S. even as the two countries battled for technology supremacy, critics said.

Wall St mixed, chip majors wobble after China sales deal
Wall St mixed, chip majors wobble after China sales deal

Business Recorder

time4 days ago

  • Business Recorder

Wall St mixed, chip majors wobble after China sales deal

NEW YORK: Wall Street's main indexes were choppy on Monday as investors prepared for a busy week and chip companies seesawed after agreeing to share a portion of revenue from China sales with the US under a trade policy shift from the Trump administration. Nvidia and Advanced Micro Devices reversed premarket losses and were last up 0.2% and 2.6%, respectively, in volatile trading. A US official told Reuters the semiconductor majors had agreed to give the United States government 15% of revenue from sales of their advanced chips to China. Analysts said the levy could hit the chipmakers' margins and set a precedent for Washington to tax critical US exports, potentially extending beyond semiconductors. 'A lot of people are not sure what to make of that because this is the first time in history that it's ever happened where an administration wants a percentage of the profits from a publicly traded company,' said Michael Matousek, head trader at US Global Investors Inc. Enabling semiconductor sales to China was an integral issue in the agreement Washington and Beijing signed earlier this year, which expires on Tuesday. US President Donald Trump lauded China's cooperation in talks at a White House news conference earlier on Monday. At 11:59 a.m. ET, the Dow Jones Industrial Average fell 91.43 points, or 0.21%, to 44,084.19, the S&P 500 gained 6.87 points, or 0.11%, to 6,396.39, and the Nasdaq Composite rose 60.17 points, or 0.28%, to 21,510.19. Six of the 11 major S&P 500 sectors slipped, while healthcare gained 0.4%, recovering some of the 5% declines it had logged so far this year. Traders took a step back after last week's rally helped the S&P 500 and the Nasdaq log their strongest weekly performance in more than a month. On Monday, the tech-heavy Nasdaq was on track for its third consecutive record closing high, if gains hold. Investors expect that the recent shakeup at the US Federal Reserve and signs of labor market weakness could nudge the central bank into adopting a dovish monetary policy stance later this year, fueling much of the optimism. July's consumer inflation report is due on Tuesday and investors currently anticipate that the Fed will lower borrowing costs by about 60 basis points by December, according to data compiled by LSEG. Citigroup and UBS Global Research became the latest brokerages to raise their year-end targets for the benchmark S&P 500. Micron Technology raised its forecast for fourth-quarter revenue and adjusted profit, sending its shares rising 3%. Intel was up 5.4% after a report said CEO Lip-Bu Tan was expected to visit the White House. Trump had called for his removal last week.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store