Raising Concerns and Showcasing Potential
A recent deepfake video featuring First Lady of the United States Jill Biden has sparked discussions about the power and challenges of advanced synthetic media technologies. The video, created by filmmaker Kenneth Lurt, shows Jill Biden criticizing her husband President Biden’s political policies in relation to the Israeli-Palestine and Hamas conflict. With the help of machine learning techniques and an AI voice generator, Lurt was able to create a realistic-sounding speech that garnered attention on social media platforms like X and Reddit.
“The goal of using AI Jill Biden was to create something absurd and cinematic enough to get folks to actually engage with the reality of what’s happening in Palestine. The drama of a radical first-lady calling out her own husband and standing up to the American empire — it’s too juicy to look away,”
– Kenneth Lurt, Filmmaker and Producer
Creating synthetic voices like these involves training AI models to clone existing voices using vast amounts of natural speech data. By analyzing interviews and appearances of Jill Biden, Lurt’s AI tool was able to generate new speech in her voice pattern and cadence. In addition to the synthetic audio, Lurt edited together various video clips related to the conflict in Gaza, creating a superficially plausible narrative. This video is just one example of the increasing use of AI and deepfake technology in political advertising.
The Rise of Synthetic Media in Political Advertising
Political campaigns have already started leveraging synthetic media to promote or attack candidates. Earlier this year, the RNC released an ad featuring generative imagery of a potential future Biden victory in 2024. The Never Back Down PAC followed suit with a million-dollar ad buy featuring an AI-generated version of Trump criticizing Gov. Reynolds of Iowa. These examples demonstrate how synthetic media can be used to shape public opinion and influence elections.
However, there are concerns about the proliferation of misleading information through manipulated media. Satirical content creator C3PMeme even posted a fake video in September 2023, depicting Ron DeSantis announcing his withdrawal from the 2024 presidential race. While intended as satire, this video highlighted the ease with which deepfakes can be created and the potential for misinformation.
“Most AI anything is boring and useless because it’s used as a cheap cheat code for creativity, talent, experience, and human passion. If I took away the script, the post production, the real conflict, and just left the voice saying random things, the project would be nothing.”
– Kenneth Lurt, Filmmaker and Producer
Filmmaker Kenneth Lurt believes that human skills and creativity are still essential in creating convincing synthetic media. While AI tools offer convenience, they lack the depth and quality that can be achieved through human filmmaking techniques. Despite the concerns surrounding deepfakes, Lurt sees the potential to use synthetic media for storytelling and raising awareness about real-world issues.
The Ethical Challenges and Future Implications
Lurt’s deepfake video aimed to draw attention to the ongoing suffering in Palestine and present an alternate scenario where a powerful figure like Jill Biden publicly condemns her husband’s policies. The use of advanced generative technologies allowed Lurt to create provocative content that forces viewers to confront the harsh realities on the ground.
However, the rise of synthetic media also raises concerns about truth, trust, and accountability. Regulators and advocates have proposed various strategies to address the threats posed by deepfakes, but comprehensive legislation and oversight remain uncertain. Companies and organizations are faced with the challenge of determining content policies that balance protected speech and the potential for misinformation.
Instead of outright bans, targeted mitigation strategies that prioritize media literacy education and responsible use of technology could be more effective. By promoting awareness of manipulation signs and fact-checking disputed claims, individuals can develop the skills needed to analyze and navigate emerging synthetics without sacrificing freedom of expression.
Coordinated efforts between stakeholders, including regulators, organizations, and the public, are necessary to mitigate the risks of synthetic media without stifling innovation or infringing on political expression.
“I think that the concept of a shared reality is pretty much dead…I’m sure there are plenty of bad actors out there.”
– Kenneth Lurt, Filmmaker and Producer
While some may question the appropriateness of Kenneth Lurt’s tactics, his project serves as a case study highlighting both the promise and ethical dilemmas associated with advanced generative technologies. It emphasizes the need for ongoing discussions and responsible approaches to navigate the evolving landscape of synthetic media.