Outrage over deepfake porn images of Taylor Swift

Outrage over deepfake porn images of Taylor Swift

US singer-songwriter Taylor Swift performs during her Eras Tour in Inglewood, California on August 7, 2023
US singer-songwriter Taylor Swift performs during her Eras Tour in Inglewood, California on August 7, 2023. Photo: Michael Tran / AFP/File
Source: AFP

Fans of Taylor Swift and politicians expressed outrage on Friday at AI-generated fake porn images that went viral on X and were still available on other platforms.

One image of the US megastar was seen 47 million times on X, the former Twitter, before it was removed Thursday. According to US media, the post was live on the platform for around 17 hours.

Deepfake porn images of celebrities are not new but activists and regulators are worried that easy-to-use tools employing generative artificial intelligence (AI) will create an uncontrollable flood of toxic or harmful content.

But the targeting of Swift, the second most listened-to artist in the world on Spotify (after Canadian rapper Drake), could shine a new light on the phenomenon with her legions of fans outraged at the development.

Read also

From Japan to the world: how to translate a game

"The only 'silver lining' about it happening to Taylor Swift is that she likely has enough power to get legislation passed to eliminate it. You people are sick," wrote influencer Danisha Carter on X.

X is one of the biggest platforms for porn content in the world, analysts say, as its policies on nudity are looser than Meta-owned platforms Facebook or Instagram.

This has been tolerated by Apple and Google, the gatekeepers for online content through the guidelines they set for their app stores on iPhones and Android smartphones.

In a statement, X said that "posting Non-Consensual Nudity (NCN) images is strictly prohibited on X and we have a zero-tolerance policy towards such content."

The Elon Musk-owned platform said that it was "actively removing all identified images and taking appropriate actions against the accounts responsible for posting them."

Read also

AI audience row at Sundance sparks walkout, highlights division

It was also "closely monitoring the situation to ensure that any further violations are immediately addressed, and the content is removed."

Swift's representatives did not immediately respond to a request for comment.

'Easier and cheaper'

"What’s happened to Taylor Swift is nothing new. For years, women have been targets of deepfakes without their consent," said Yvette Clarke, a Democratic congresswoman from New York who has backed legislation to fight deepfake porn.

"And with advancements in AI, creating deepfakes is easier & cheaper," she added.

Tom Keane, a Republican congressman, warned that "AI technology is advancing faster than the necessary guardrails. Whether the victim is Taylor Swift or any young person across our country, we need to establish safeguards to combat this alarming trend."

Many well-publicized cases of deepfake audio and video have targeted politicians or celebrities, with women by far the biggest targets through graphic, sexually explicit images found easily on the internet.

Read also

LA Times slashes a fifth of its newsroom jobs

Software to create the images is widely available on the web.

According to research cited by Wired magazine, 113,000 deepfake videos were uploaded to the most popular porn websites in the first nine months of 2023.

And research in 2019 from a startup found that 96 percent of deepfake videos on the internet were pornographic.

Source: AFP

Authors:
AFP avatar

AFP AFP text, photo, graphic, audio or video material shall not be published, broadcast, rewritten for broadcast or publication or redistributed directly or indirectly in any medium. AFP news material may not be stored in whole or in part in a computer or otherwise except for personal and non-commercial use. AFP will not be held liable for any delays, inaccuracies, errors or omissions in any AFP news material or in transmission or delivery of all or any part thereof or for any damages whatsoever. As a newswire service, AFP does not obtain releases from subjects, individuals, groups or entities contained in its photographs, videos, graphics or quoted in its texts. Further, no clearance is obtained from the owners of any trademarks or copyrighted materials whose marks and materials are included in AFP material. Therefore you will be solely responsible for obtaining any and all necessary releases from whatever individuals and/or entities necessary for any uses of AFP material.