Facebook and Instagram will require political ads running on their platforms to disclose if they were created using artificial intelligence, their parent company announced on Wednesday.
Under the new policy by Meta, labels acknowledging the use of AI will appear on users' screens when they click on ads. The rule takes effect Jan 1 and will be applied worldwide.
The development of new AI programmes has made it easier than ever to quickly generate lifelike audio, images and video. In the wrong hands, the technology could be used to create fake videos of a candidate or frightening images of election fraud or polling place violence. When strapped to the powerful algorithms of social media, these fakes could mislead and confuse voters on a scale never seen.
Meta Platforms Inc and other tech platforms have been criticised for not doing more to address this risk. Wednesday's announcement which comes on the day House lawmakers hold a hearing on deepfakes isn't likely to assuage those concerns.
While officials in Europe are working on comprehensive regulations for the use of AI, time is running out for lawmakers in the United States to pass regulations ahead of the 2024 election.
Earlier this year, the Federal Election Commission began a process to potentially regulate AI-generated deepfakes in political ads before the 2024 election. President Joe Biden's administration last week issued an executive order intended to encourage responsible development of AI. Among other provisions, it will require AI developers to provide safety data and other information about their programs with the government.
The US isn't the only nation holding a high-profile vote next year: National elections are also scheduled in countries including Mexico, South Africa, Ukraine, Taiwan and Pakistan.
More From This Section
AI-generated political ads have already made an appearance in the US. In April, the Republican National Committee released an entirely AI-generated ad meant to show the future of the United States if Biden, a Democrat, is reelected. It employed fake but realistic photos showing boarded-up storefronts, armoured military patrols in the streets, and waves of immigrants creating panic. The ad was labeled to inform viewers that AI was used.
In June, Florida Gov Ron DeSantis' presidential campaign shared an attack ad against his GOP primary opponent Donald Trump that used AI-generated images of the former president hugging infectious disease expert Dr Anthony Fauci.
It's gotten to be a very difficult job for the casual observer to figure out: What do I believe here? said Vince Lynch, an AI developer and CEO of the AI company IV.AI. Lynch said some combination of federal regulation and voluntary policies by tech companies is needed to protect the public. The companies need to take responsibility, Lynch said.
Meta's new policy will cover any advertisement for a social issue, election or political candidate that includes a realistic image of a person or event that has been altered using AI. More modest use of the technology to resize or sharpen an image, for instance, would be allowed with no disclosure.
Besides labels informing a viewer when an ad contains AI-generated imagery, information about the ad's use of AI will be included in Facebook's online ad library. Meta, which is based in Menlo Park, California, says content that violates the rule will be removed.
Google announced a similar AI labeling policy for political ads in September. Under that rule, political ads that play on YouTube or other Google platforms will have to disclose the use of AI-altered voices or imagery.