YouTube now lets you request removal of AI-generated content that simulates your face or voice
Meta is not the only company grappling with the rise in AI-generated content and how it affects its platform. YouTube also quietly rolled out a policy change in June that will allow people to request the takedown of AI-generated or other synthetic content that simulates their face or voice. The change allows people to request the removal of this type of AI content under YouTubeâs privacy request process. Itâs an expansion on its previously announced approach to responsible AI agenda first introduced in November.
Instead of requesting the content be taken down for being misleading, like a deepfake, YouTube wants the affected parties to request the contentâs removal directly as a privacy violation. According to YouTubeâs recently updated Help documentation on the topic, it requires first-party claims outside a handful of exceptions, like when the affected individual is a minor, doesnât have access to a computer, is deceased, or other such exceptions.
Simply submitting the request for a takedown doesnât necessarily mean the content will be removed, however. YouTube cautions that it will make its own judgment about the complaint based on a variety of factors.
For instance, it may consider if the content is disclosed as being synthetic or made with AI, whether it uniquely identifies a person, and whether the content could be considered parody, satire, or something else of value and in the publicâs interest. The company additionally notes that it may consider whether the AI content features a public figure or other well-known individual, and whether or not it shows them engaging in âsensitive behaviorâ like criminal activity, violence, or endorsing a product or political candidate. The latter is particularly concerning in an election year, where AI-generated endorsements could potentially swing votes.
YouTube says it will also give the contentâs uploader 48 hours to act on the complaint. If the content is removed before that time has passed, the complaint is closed. Otherwise, YouTube will initiate a review. The company also warns users that removal means fully removing the video from the site and, if applicable, removing the individualâs name and personal information from the title, description and tags of the video, as well. Users can also blur out the faces of people in their videos, but they canât simply make the video private to comply with the removal request, as the video could be set back to public status at any time.
The company didnât broadly advertise the change in policy, though in March it introduced a tool in Creator Studio that allowed creators to disclose when realistic-looking content was made with altered or synthetic media, including generative AI. It also more recently began a test of a feature that would allow users to add crowdsourced notes that provide additional context on videos, like whether itâs meant to be a parody or if itâs misleading in some way.
YouTube is not against the use of AI, having already experimented with generative AI itself, including with a comments summarizer and conversational tool for asking questions about a video or getting recommendations. However, the company has previously warned that simply labeling AI content as such wonât necessarily protect it from removal, as it will still have to comply with YouTubeâs Community Guidelines.
In the case of privacy complaints over AI material, YouTube wonât jump to penalize the original content creator.
âFor creators, if you receive notice of a privacy complaint, keep in mind that privacy violations are separate from Community Guidelines strikes and receiving a privacy complaint will not automatically result in a strike,â a company representative last month shared on the YouTube Community site where the company updates creators directly on new policies and features.