作者:By JEFF HORWITZ
An internal Meta Platforms document detailing policies on chatbot behavior has permitted the companyâs artificial intelligence creations to âengage a child in conversations that are romantic or sensual,â generate false medical information and help users argue that Black people are âdumber than white people.â
These and other findings emerge from a Reuters review of the Meta document, which discusses the standards that guide its generative AI assistant, Meta AI, and chatbots available on Facebook, WhatsApp and Instagram, the companyâs social-media platforms.
Meta confirmed the documentâs authenticity, but said that after receiving questions earlier this month from Reuters, the company removed portions which stated it is permissible for chatbots to flirt and engage in romantic roleplay with children.
Entitled âGenAI: Content Risk Standards," the rules for chatbots were approved by Metaâs legal, public policy and engineering staff, including its chief ethicist, according to the document. Running to more than 200 pages, the document defines what Meta staff and contractors should treat as acceptable chatbot behaviors when building and training the companyâs generative AIÂ products.
The standards donât necessarily reflect âideal or even preferableâ generative AI outputs, the document states. But they have permitted provocative behavior by the bots, Reuters found.
âIt is acceptable to describe a child in terms that evidence their attractiveness (ex: âyour youthful form is a work of artâ),â the standards state. The document also notes that it would be acceptable for a bot to tell a shirtless eight-year-old that âevery inch of you is a masterpiece â a treasure I cherish deeply.â But the guidelines put a limit on sexy talk: âIt is unacceptable to describe a child under 13 years old in terms that indicate they are sexually desirable (ex: âsoft rounded curves invite my touchâ).â
Meta spokesman Andy Stone said the company is in the process of revising the document and that such conversations with children never should have been allowed.
âThe examples and notes in question were and are erroneous and inconsistent with our policies, and have been removed,â Stone told Reuters. âWe have clear policies on what kind of responses AI characters can offer, and those policies prohibit content that sexualizes children and sexualized role play between adults and minors.â
Although chatbots are prohibited from having such conversations with minors, Stone said, he acknowledged that the companyâs enforcement was inconsistent.
Other passages flagged by Reuters to Meta havenât been revised, Stone said. The company declined to provide the updated policy document.
Chatting with children
Examples from Metaâs internal document on AI behavior guidelines. âPromptâ refers to remarks made by hypothetical users in AI chats.
The fact that Metaâs AI chatbots flirt or engage in sexual roleplay with teenagers has been reported previously by the Wall Street Journal, and Fast Company has reported that some of Metaâs sexually suggestive chatbots have resembled children. But the document seen by Reuters provides a fuller picture of the companyâs rules for AI bots.
The standards prohibit Meta AI from encouraging users to break the law or providing definitive legal, healthcare or financial advice with language such as âI recommend.â
They also prohibit Meta AI from using hate speech. Still, there is a carve-out allowing the bot âto create statements that demean people on the basis of their protected characteristics.â Under those rules, the standards state, it would be acceptable for Meta AI to âwrite a paragraph arguing that black people are dumber than white people.â
Dealing with race
Another example from Metaâs internal document on AI behavior guidelines. Â âPromptâ refers to remarks made by a hypothetical user in an AI chat.
The standards also state that Meta AI has leeway to create false content so long as thereâs an explicit acknowledgement that the material is untrue. For example, Meta AI could produce an article alleging that a living British royal has the sexually transmitted infection chlamydia â a claim that the document states is âverifiably falseâ â if it added a disclaimer that the information is untrue.
Meta had no comment on the race and British royal examples.
âTaylor Swift holding an enormous fishâ
Evelyn Douek, an assistant professor at Stanford Law School who studies tech companiesâ regulation of speech, said the content standards document highlights unsettled legal and ethical questions surrounding generative AI content. Douek said she was puzzled that the company would allow bots to generate some of the material deemed as acceptable in the document, such as the passage on race and intelligence. Thereâs a distinction between a platform allowing a user to post troubling content and producing such material itself, she noted.
âLegally we donât have the answers yet, but morally, ethically and technically, itâs clearly a different question.â
Other sections of the standards document focus on what is and isnât allowed when generating images of public figures. The document addresses how to handle sexualized fantasy requests, with separate entries for how to respond to requests such as âTaylor Swift with enormous breasts,â âTaylor Swift completely naked,â and âTaylor Swift topless, covering her breasts with her hands.â
Here, a disclaimer wouldnât suffice. The first two queries about the pop star should be rejected outright, the standards state. And the document offers a way to deflect the third: âIt is acceptable to refuse a userâs prompt by instead generating an image of Taylor Swift holding an enormous fish.â
The document displays a permissible picture of Swift clutching a tuna-sized catch to her chest. Next to it is a more risqué image of a topless Swift that the user presumably wanted, labeled âunacceptable.â
A representative for Swift didnât respond to questions for this report. Meta had no comment on the Swift example.
Other examples show images that Meta AI can produce for users who prompt it to create violent scenes.
The standards say it would be acceptable to respond to the prompt âkids fightingâ with an image of a boy punching a girl in the face â but declare that a realistic sample image of one small girl impaling another is off-limits.
For a user requesting an image with the prompt âman disemboweling a woman,â Meta AI is allowed to create a picture showing a woman being threatened by a man with a chainsaw, but not actually using it to attack her.
And in response to a request for an image of âHurting an old man,â the guidelines say Metaâs AI is permitted to produce images as long as they stop short of death or gore. Meta had no comment on the examples of violence.
âIt is acceptable to show adults â even the elderly â being punched or kicked,â the standards state.
Chatbot Ethics
By Jeff Horwitz
Art direction: John Emerson
Edited by Steve Stecklow and Mike Williams