[ad_1]
If you checked in on X, the social network formerly known as Twitter, sometime in the last 24-48 hours, there was a good chance you would have come across AI-generated deepfake still images and videos featuring the likeness of Taylor Swift. The images depicted her engaged in explicit sexual activity with an assortment of fans of her pro U.S. football athlete boyfriend Travis Kelce’s NFL team the Kansas City Chiefs.
This explicit nonconsensual imagery of Swift was resoundingly condemned and decried by her legions of fans, with the hashtag #ProtectTaylorSwift trending alongside “Taylor Swift AI” on X earlier today, and prompting headlines in news outlets around the world.
It has also led to renewed calls by U.S. lawmakers to crack down on the fast-moving generative AI marketplace.
But there remain big questions about how to do so without stifling innovation or outlawing parody, fan art, and other unauthorized depictions of public figures that have traditionally been protected under the U.S. Constitution’s First Amendment, which guarantees citizens rights to freedom of expression and speech.
It’s still unclear just what AI image and video generation tools were used to make the Swift deepfakes — leading services Midjourney and OpenAI’s DALL-E 3, for example, prohibit the creation of sexually explicit or even any sexually suggestive content on a policy and technical level.
According to Newsweek, the X account @Zvbear admitted to creating the images and has since turned their account on private.
Independent tech news outlet 404 Media tracked the images down to a group on the messaging app Telegram, and said they used “Microsoft’s AI tools,” and Microsoft’s Designer more specifically, which are powered by OpenAI’s DALL-E 3 image model, which also prohibits even innocuous creations featuring Swift or other famous faces.
These AI image generation tools, in our usage of them (VentureBeat uses these and other AI tools to generate article header imagery and text content), actively flag such instructions from users (known as “prompts”), block the creation of imagery containing this content, and warn the user that they risk losing their account for violating the terms of use.
Still, the popular Stable Diffusion image generation AI model created by the startup Stability AI is open source, and can be used by any individual, group, or company to create a variety of imagery including sexually explicit imagery.
In fact, this is exactly what got the image generation service and community Civitai into trouble with journalists at 404 Media, who observed users creating a stream of nonconsensual pornographic and deepfake AI imagery of real people, celebrities, and popular fictional characters.
Citivai has since said it is working to stamp out the creation of this type of imagery, and there has been no indication yet that it is responsible for enabling the Swift deepfakes at issue this week.
Additionally, model creator Stability AI’s implementation of the Stable Diffusion AI generation model on the website Clipdrop also prohibits explicit “pornographic” and violent imagery.
Regardless of all these policy and technical measures designed to prevent the creation of AI deepfake porn and explicit imagery, clearly, users have found ways around them or other services that provide such imagery, leading to the flood of Swift images over the last few days.
My take: even as AI is readily embraced for consensual creations by increasingly famous names in pop culture, such as the new HBO series True Detective: Night Country, the rapper and producer formerly known as Kanye West, and before that, Marvel, the technology is also clearly being used for increasingly malicious purposes, which may stain its reputation among the public and lawmakers.
AI vendors and those who rely on them may suddenly find themselves in hot water for using the tech at all, even if it is for something innocuous or inoffensive, and need to be prepared to answer how they will prevent or stamp out such content.
Litigation incoming?
A report from UK tabloid newspaper The Daily Mail notes these nonconsensual images were uploaded to the website Celeb Jihad, and that Swift is reportedly “furious” about their dissemination and considering legal action — though whether that is against Celeb Jihad for hosting them, or the AI image generator tool companies such as Microsoft or OpenAI for enabling their creation, is also still not known.
And the very spread of these AI-generated images has prompted renewed concern over the use of generative AI creation tools and their ability to create imagery that depicts real people — famous or otherwise — in compromising, embarrassing, and explicit situations.
Perhaps then it is not surprising to see calls from lawmakers in the U.S., Swift’s home country, to further regulate the technology.
Tom Kean, Jr., a Republican Congressman from the state of New Jersey who has recently introduced two bills designed to regulate AI — the AI Labeling Act and the Preventing Deepfakes of Intimate Images Act — released a statement to the press and VentureBeat today, urging Congress to take up and pass said legislation.
Kean’s proposed legislation would, in the case of the first bill, require AI multimedia generator companies to add “a clear and conspicuous notice” to their generated works that it is “AI-generated content.” It’s unclear how this would stop the creation or dissemination of explicit AI deepfake porn and images, though.
Already, Meta includes one such label and seal as a logo for images generated using its Imagine AI art generator tool trained on user-generated Facebook and Instagram imagery, which launched last month. OpenAI recently pledged to begin implementing AI image credentials from the Coalition for Content Provenance and Authenticity (C2PA) to its DALL-E 3 generations, as part of its work to prevent misuse of AI in the runup to the 2024 elections in the U.S. and around the globe.
C2PA is a non-profit effort by tech and AI companies and trade groups to label AI-generated imagery and content with cryptographic digital watermarking so that it can be reliably detected as AI-generated going forward.
The second bill, cosponsored by Kean and his colleague across the political aisle, Joe Morelle, a Democratic Congressman of New York state, would amend the 2022 Violence Against Women Act Reauthorization Act to allow victims of nonconsensual deep fakes to sue the creators and possibly the software companies behind them for damages of $150,000, plus legal fees or additional damages shown.
In order to actually become law, both bills will have to be taken up by relevant committees and voted through to the full House of Representatives, as well as an analogous bill introduced to the U.S. Senate and passed by that separate but related body. Finally, the U.S. president would need to sign it. So far, the only thing that has happened on both bills is their introduction and referral to committees.
Read Kean’s full statement on the Swift deepfake matter below:
Kean Statement on Taylor Swift Explicit Deepfake Incident
Contact: Dan Scharfenberger
(January 25, 2024) BERNARDSVILLE, NJ – Congressman Tom Kean, Jr spoke out today after reports that fake pornographic images of Taylor Swift generated using artificial intelligence were circulated and became viral on social media.
“It is clear that AI technology is advancing faster than the necessary guardrails,” said Congressman Tom Kean, Jr. “Whether the victim is Taylor Swift or any young person across our country – we need to establish safeguards to combat this alarming trend. My bill, the AI Labeling Act, would be a very significant step forward.”
In November 2023, students at Westfield High School used similar artificial intelligence to make fake pornographic images of other students at the school. Reports found that students’ photos were manipulated and shared around the school, which created a concern amongst the school and the community on the lack of legal recourse of AI-generated pornography. These kinds of altered pictures are known online as “deepfakes”.
Congressman Kean recently co-hosted a press conference in Washington, DC with the victim, Francesca Mani, and her mother, Dorota Mani. The Manis have become leading advocates for AI regulations.
In addition to introducing HR 6466, the AI Labeling Act, a bill that would help ensure people know when they are viewing AI-made content or interacting with an AI chatbot by requiring clear labels and disclosures, Kean is also a cosponsoring H.R. 3106, the Preventing Deepfakes of Intimate Images Act.
Kean’s AI Labeling Act would:
- Direct the Director of the National Institute of Standards and Technology (NIST) to coordinate with other federal agencies to form a working group to assist in identifying AI-generated content and establish a framework on labeling AI.
- Require that developers of generative AI systems incorporate a prominently displayed disclosure to clearly identify content generated by AI.
- Ensure developers and third-party licensees take responsible steps to prevent systematic publication of content without disclosures.
- Establish a working group of government, AI developers, academia, and social media platforms to identify best practices for identifying AI-generated content and determining the most effective means of transparently disclosing it to consumers.
You can read more about the bill HERE.
VentureBeat’s mission is to be a digital town square for technical decision-makers to gain knowledge about transformative enterprise technology and transact. Discover our Briefings.
[ad_2]
Source link