AI Makes It Harder to Spot Deep Fakes Than Ever Before, but Awareness Is Key
By
Cayce Myers

As artificial intelligence programs continue to develop and access is easier than ever, it’s making it harder to separate fact from fiction. Just this week, an AI-generated image of an explosion near the Pentagon made headlines online and even slightly impacted the stock market until it was quickly deemed a hoax.

Cayce Myers, a professor in Virginia Tech’s School of Communication, has been studying this ever evolving technology and shares his take on the future of deep fakes and how to spot them.

detecting deepfakes“It is becoming increasingly difficult to identify disinformation, particularly sophisticated AI generated deep fake,” says Myers. “The cost barrier for generative AI is also so low that now almost anyone with a computer and internet has access to AI.”

Myers believes because of this we will see a lot more disinformation – both visual and written – over the next few years. “Spotting this disinformation is going to require users to have more media literacy and savvy in examining the truth of any claim.”

While photoshop programs have been used for years, Myers says the difference between that and disinformation created with AI is one of sophistication and scope. “Photoshop allows for fake images, but AI can create altered videos that are very compelling. Given that disinformation is now a widespread source of content online this type of fake news content can reach a much wider audience, especially if the content goes viral.”

When it comes to combating disinformation, Myers says there are two main sources – ourselves and the AI companies.

“Examining sources, understanding warning signs of disinformation, and being diligent in what we share online is one personal way to combat the spread of disinformation,” he says. “However,  that is not going to be enough. Companies that produce AI content and social media companies where disinformation is spread will need to implement some level of guardrails to prevent the widespread disinformation from being spread.”

Myers explains the problem is that the technology of AI has developed so fast that it’s likely that any mechanism to prevent the spread of AI generated disinformation will not be full proof.

Attempts to regulate AI are going on in the U.S. at the federal, state, and even local level.  Lawmakers are considering a variety of issues including disinformation, discrimination, intellectual property infringement, and privacy.

“The issue is that lawmakers do not want to create a new law regulating AI before we know where the technology is going. Creating a law too fast can stifle AI’s development and growth, creating one too slow may open the door for a lot of potential problems. Striking a balance will be a challenge,” says Myers.


Cayce Myers [@CayceMyers]is a professor of public relations and director of graduate studies at the School of Communication at Virginia Tech. His work focuses on media history, political communication, and laws that affect public relations practice. He is the author of Public Relations History: Theory Practice and Profession and Money in Politics: Campaign Fundraising in the 2020 Presidential Election.