Science & Tech Spotlight: Deepfakes
A deepfake is a video, photo, or audio recording that seems real but has been manipulated with artificial intelligence technologies. Deepfakes are usually pornographic and disproportionately victimize women. However, deepfakes can also be used to influence elections or incite civil unrest.
Anyone with basic computing skills and a home computer can create a deepfake. Some methods to detect deepfakes use AI to analyze videos for details that deepfakes fail to imitate realistically—such as blinking or facial tics.
However, detection is challenging because deepfake technologies evolve rapidly.
Potential errors in deepfakes
Illustration showing errors in inconsistent eye blinking, mismatched earrings, and features lacking definition
Why This Matters
Deepfakes are powerful tools that can be used for exploitation and disinformation. Deepfakes could influence elections and erode trust but so far have mainly been used for non-consensual pornography. The underlying artificial intelligence (AI) technologies are widely available at low cost, and improvements are making deepfakes harder to detect.
What is it? A deepfake is a video, photo, or audio recording that seems real but has been manipulated with AI. The underlying technology can replace faces, manipulate facial expressions, synthesize faces, and synthesize speech. Deepfakes can depict someone appearing to say or do something that they in fact never said or did.
Figure 1. Deepfake videos commonly swap faces or manipulate facial expressions. In face swapping, the face on the left is placed on another person’s body. In facial manipulation, the expressions of the face on the left are imitated by the face on the right.
While deepfakes have benign and legitimate applications in areas such as entertainment and commerce, they are commonly used for exploitation. According to a recent report from the company Deeptrace, much of deepfake content online is pornographic, and deepfake pornography disproportionately victimizes women. Further, there is concern about potential growth in the use of deepfakes for other purposes, particularly disinformation. Deepfakes could be used to influence elections or incite civil unrest, or as a weapon of psychological warfare. They could also lead to disregard of legitimate evidence of wrongdoing and, more generally, undermine public trust in audiovisual content.
How does it work? Deepfakes rely on artificial neural networks, which are computer systems modeled loosely on the human brain that recognize patterns in data. Developing a deepfake photo or video typically involves feeding hundreds or thousands of images into the artificial neural network, “training” it to identify and reconstruct patterns—usually faces.
Deepfakes use different underlying AI technologies—notably autoencoders and generative adversarial networks (GANs). An autoencoder is an artificial neural network trained to reconstruct input from a simpler representation. A GAN is made up of two competing artificial neural networks, one trying to produce a fake, the other trying to detect it. This competition continues over many cycles, resulting in a more plausible rendering of, for example, faces in video. GANs generally produce more convincing deepfakes but are more difficult to use.
Researchers and internet companies have experimented with several methods to detect deepfakes. These methods typically also use AI to analyze videos for digital artifacts or details that deepfakes fail to imitate realistically, such as blinking or facial tics.
How mature is it? Anyone with basic computer skills and a home computer can create a deepfake. Computer applications are openly available on the internet with tutorials on how to create deepfake videos. However, to develop a somewhat realistic deepfake, these applications generally still require hundreds or thousands of training images of the faces to be swapped or manipulated, making celebrities and government leaders the most common subjects. More convincing deepfakes created with GANs require more advanced technical skills and resources. As artificial neural network technologies have advanced rapidly in parallel with more powerful and abundant computing, so has the ability to produce realistic deepfakes.
There are some potentially benign or beneficial uses of this technology:
■ Entertainment. Voices and likenesses can be used in a movie to achieve a creative effect or maintain a cohesive story when the entertainers themselves are not available.
■ E-commerce. Retailers could let customers use their likenesses to virtually try on clothing.
■ Communication. Speech synthesis and facial manipulation can make it appear that a person is authentically speaking another language.
■ Data needs for detection. Deepfake detection tools must generally be trained with large and diverse data sets to reliably detect deepfakes. Technology companies and researchers have released data sets to help train detection tools, but the current data sets are not sufficient by themselves. Detection tools must be constantly updated with data of increasing sophistication to ensure that they continue to be effective at detecting manipulated media.
■ Detection is not yet automated. Current tools cannot perform a complete and automated analysis that reliably detects deepfakes. Research programs are currently working on means to automatically detect deepfakes, provide information on how they were created, and assess the overall integrity of digital content.
■ Adaptation to detection. Techniques used to identify deepfakes tend to lead to the development of more sophisticated deepfake techniques. This “cat and mouse” situation means detection tools must be regularly updated to keep pace.
■ Detection may not be enough. Even a perfect detection technology may not prevent a fake video from being effective as disinformation, because many viewers may be unaware of deepfakes or may not take the time to check the reliability of the videos they see.
■ Inconsistent social media standards. The major social media companies have different standards for moderating deepfakes.
■ Legal issues. Proposed laws or regulations addressing deepfake media may raise questions regarding an individual’s freedom of speech and expression and the privacy rights of individuals falsely portrayed in deepfakes. Moreover, potential federal legislation aimed at combating deepfakes could face enforcement challenges.
Policy Context and Questions
Any policy response seeking to address deepfakes would likely face constitutional and other legal challenges along with the technical challenges of detection. Key policy questions include:
■ What is the maturity of deepfake detection technology? How much progress have federal programs and public-private partnerships made in developing such technology? What expertise will be required to ensure detection keeps pace with deepfake technology?
■ What rights do individuals have to their privacy and likeness? What rights do creators of deepfakes have under the First Amendment? What policy options exist regarding election interference? What options exist regarding exploitation and image abuse, such as non-consensual pornography?
■ What can be done to educate the public about deepfakes? Should manipulated media be marked or labeled? Should media be traceable to its origin to determine authenticity?
■ What should the roles of media outlets and social media companies be in detecting and moderating content that has been altered or falsified?
For more information, contact Karen Howard at 202-512-6888 or HowardK@gao.gov.