What are deepfakes, their risks and how to spot them (2024)

Ever since the term Deepfake appeared online in 2017, it has become more popular because of its innovative way of creating artificial videos and the dangers it poses. More recently, the term came into the mainstream after fake nude photographs of American singer Taylor Swift proliferated on X (formerly known as Twitter), which led to calls in Congress for new legislation.

These AI-generated images of real people, and which appear authentic, have garnered significant attention in light of Swift’s targeting. Some states have already enacted laws targeting deepfakes, while others are considering measures to combat their proliferation. Efforts include deepfake detection algorithms and embedding codes in content to identify misuse. Model legislation proposed by the American Legislative Exchange Council focuses on criminalizing possession and distribution of deepfakes depicting minors and allowing victims to sue for nonconsensual distribution of sexual content.

However, ensuring effective enforcement and navigating free speech concerns remain significant challenges. Federal legislation has also been introduced to provide individuals with property rights over their likeness and voice, allowing them to sue for misleading deepfakes. States such as Indiana and Missouri are pushing for legislation criminalizing the creation and distribution of sexually explicit deepfakes without consent.

But deepfake p*rnography is just the tip of the iceberg when we talk about the risk this type of technology poses. Deepfakes have several potential uses that could represent different harms, such as fake news, hoaxes, financial fraud and other types of p*rnography like revenge p*rn or child sexual abuse material.

What are deepfakes?

Deepfakes are videos, photos, or audio recordings of real-life people that seem authentic, but have been manipulated with Artificial Intelligence, according to the U.S. Government Accountability Office (GAO). The name comes from the type of machine learning used to generate this type of media, deep learning.

The GAO says deepfakes are tools that can be used for exploitation and disinformation. They could influence elections and cause damage to public and private figures, “but so far have mainly been used for non-consensual p*rnography”, as was the case with Taylor Swift.

How do deepfakes work?

Deepfakes utilize advanced AI techniques like autoencoders and generative adversarial networks (GANs) to create realistic synthetic media —both examples of deep learning, which can take certain types of data and learn to produce a similar media that resembles the example.

An autoencoder is an artificial neural network — designed to replicate how the human brain learns information — trained to recreate input from a simple representation, which means that it can reconstruct an image or a video taking a basic file. GANs consists of two competing artificial neural networks, one is trying to produce a fake version and the other tries to detect it. They work constantly, resulting in a more “realistic” or “accurate” portrayal. According to GAO, “GANs create more convincing deepfakes, but are more difficult to use”.

Improvements on these types of technology are making deepfakes harder to detect. In the past, viewers could easily detect fraudulent content, but this may no longer be the case, considering how realistic some images, videos and audios seem.

Risks of deepfakes

As mentioned, deepfake technology could be used to create several types of content, such as p*rnography using a celebrity or any person’s face without their consent, or fake news with altered videos of politicians saying things they never said or doing things they never did.

A report made by the Department of Homeland Security states that “the threat of Deepfakes and synthetic media comes not from the technology used to create it, but from people’s natural inclination to believe what they see”, and highlight that deepfakes and synthetic medicare effective in spreading misinformation or disinformation despite not being “advanced or believable”.

The Department also highlights how divided the opinion of experts on the urgency of the threat synthetic media and deepfakes pose is. It says that the spectrum of concerns range from “an urgent threat” to “don’t panic just be prepared”.

How to detect deepfakes

Technological detection of deepfakes relies on extensive and diverse datasets for training detection tools, but current datasets are insufficient and require constant updates to effectively detect manipulated media. Automated detection tools are still under development, with ongoing research aiming to automatically identify deepfakes and assess the integrity of digital content. However, detection techniques often spur the development of more sophisticated deepfake methods, so regular updates to detection tools are necessary.

Even with effective detection, disinformation spread through deepfake videos may still be impactful due to audience unawareness or lack of verification. Social media platforms have inconsistent standards for moderating deepfakes, and proposed legal regulations raise concerns about freedom of speech, privacy rights, and enforcement challenges.

As for human detection, while in the past it could’ve been easy to spot a fake video — there were common visual mistakes like inconsistent eye blinking or lack of definition in certain areas of the media —, currently, with its advances it is becoming even harder to spot fake content.

Sign up for our weekly newsletter to get more English-language news coverage from EL PAÍS USA Edition

What are deepfakes, their risks and how to spot them (2024)

FAQs

What are deepfakes, their risks and how to spot them? ›

A “deepfake” is fabricated hyper-realistic digital media, including video, image, and audio content. Not only has this technology created confusion, skepticism, and the spread of misinformation, deepfakes also pose a threat to privacy and security.

What are the risks of deepfakes? ›

Even scarier are the AI-generated deepfakes that can mimic a person's voice, face and gestures. New cyber attack tools can deliver disinformation and fraudulent messages at a scale and sophistication not seen before. Simply put, AI-generated fraud is harder than ever to detect and stop.

What are some examples of deepfakes? ›

Another example involving celebrities include a TikTok account that posts video of a person who bears the likeness of actor Tom Cruise. The article on MIT's website also provided an example of deepfake: a video featuring the late Former President Richard Nixon announcing a failed moon landing.

What are the potential problems that deepfakes can cause? ›

Deepfakes can use images or videos of individuals without their consent, potentially violating their privacy and autonomy. Non-consensual deepfake videos can cause significant harm to individuals by exploiting and manipulating their likeness for explicit or damaging content.

Will we be able to detect deepfakes? ›

They've been used to try to influence elections and to create non-consensual p*rnography. Researchers are developing new ways to detect deepfakes, such as AI models that spot color abnormalities. And authentication methods, like digital watermarks, can help prove that a video or image has been altered.

Why are deepfakes hard to detect? ›

Deepfake detection systems work very differently from how human beings listen. They analyze audio samples for artifacts like missing frequencies that are often left behind when audio is programmatically generated.

What are the negatives of deepfakes? ›

Not only has this technology created confusion, skepticism, and the spread of misinformation, deepfakes also pose a threat to privacy and security. With the ability to convincingly impersonate anyone, cybercriminals can orchestrate phishing scams or identity theft operations with alarming precision.

Why should we be worried about deepfakes? ›

Deepfakes are creating havoc across the globe, spreading fake news and p*rnography, being used to steal identities, exploiting celebrities, scamming ordinary people and even influencing elections.

Are deepfakes illegal? ›

Deepfakes of child sexual abuse or “revenge p*rn” — creating an image of a real person with the intent to harm their reputation — could be illegal obscenity. Perjury. Deepfakes could be used to manufacture false testimony in a video deposition, court testimony or other statements made under oath.

How to spot AI fakes? ›

How to identify AI-generated videos
  1. Look out for strange shadows, blurs, or light flickers. In some AI-generated videos, shadows or light may appear to flicker only on the face of the person speaking or possibly only in the background. ...
  2. Unnatural body language. This is another AI giveaway. ...
  3. Take a closer listen.

How do I know if a photo is AI or not? ›

Because artificial intelligence is piecing together its creations from the original work of others, it can show some inconsistencies close up. When you examine an image for signs of AI, zoom in as much as possible on every part of it. Stray pixels, odd outlines, and misplaced shapes will be easier to see this way.

Can deepfakes be tracked? ›

As these generative artificial intelligence (AI) technologies become more common, researchers are now tracking their proliferation through a database of political deepfakes.

What are the malicious uses of deepfakes? ›

Beyond politics, deepfakes present a threat to personal security. The technology can be used for malicious purposes, such as blackmail, fraud, and cyberbullying. As a result, the need for vigilance and sophisticated detection methods has never been more critical.

Is deepfake a cyber threat? ›

In the context of cybersecurity, deepfakes are categorized as social engineering attacks that can be used to breach an organization's systems and compromise internal data.

Top Articles
Latest Posts
Article information

Author: Kimberely Baumbach CPA

Last Updated:

Views: 5627

Rating: 4 / 5 (41 voted)

Reviews: 88% of readers found this page helpful

Author information

Name: Kimberely Baumbach CPA

Birthday: 1996-01-14

Address: 8381 Boyce Course, Imeldachester, ND 74681

Phone: +3571286597580

Job: Product Banking Analyst

Hobby: Cosplaying, Inline skating, Amateur radio, Baton twirling, Mountaineering, Flying, Archery

Introduction: My name is Kimberely Baumbach CPA, I am a gorgeous, bright, charming, encouraging, zealous, lively, good person who loves writing and wants to share my knowledge and understanding with you.