Deepfake Technology: What Are Deepfakes? How Do They Make Deepfakes? (2024)

Table of Contents (click to expand)
  • What Is Deepfake?
  • Deepfake Text
  • Deepfake Video
  • Deepfake Audio
  • The Road Ahead With Deepfakes

Deepfake is a new media technology wherein a person simply takes existing text, picture, video, or audio and then manipulates, i.e., ‘fakes’ it to look like someone else using advanced artificial intelligence (AI) and neural network (NN) technology

After its first appearance a few years back, deepfake technology has evolved from an innocuous tech geek’s chicanery to a malicious slandering weapon. In this article, we’ll see what exactly this dreaded deepfake tech is, how it works, what different forms it comes in, and how we can detect or bust a deepfake.

Recommended Video for you:


What Is Deepfake?

Deepfake is one of the buzzwords in media technology wherein a person simply takes existing text, picture, video, or audio and then manipulates, i.e., ‘fakes’ it to look like someone else using advanced artificial intelligence (AI) and neural network (NN) technology.

Want to put abusive words in the mouth of your nemesis? Or swap the movie protagonist with your favorite Hollywood superstar? Or do you just want to make yourself dance like Michael Jackson? Then deepfake is what you need!

Deepfake content is growing exponentially. Unfortunately, deepfake tech has already been repeatedly used to gain political mileage, to tarnish the image of a rival, or to commit financial fraud.

Let’s now look into the three main types of deepfakes and explore the data science that allows them to work. We’ll also focus on deepfake detection technologies that researchers and security consultants are working on to curb the malicious use of deepfakes.

Also Read: How Does FaceApp Work?

Deepfake Text

In the early days of artificial intelligence (AI) and natural language processing (NLP), it was posited that it would be a challenge for a machine to do a creative activity like painting or writing. Fast-forward to 2021; with the powerful language models and libraries that have been built over the years by the incremental work of researchers and data science professionals, top-rated AI-generated prose can now write with humanlike pith and coherence.

Deepfake Technology: What Are Deepfakes? How Do They Make Deepfakes? (1)

GPT-2

Take, for example, GPT-2—the latest breed of the text-generation system released by research lab OpenAI from Silicon Valley. This tech has impressed both the layman and domain experts with its ability to churn out coherent text with minimal prompts.

OpenAI engineers used over 8 million textual documents that were scrapped (method of relevant data extraction from webpages) and combined with a billion parameters for modeling and training of GPT-2 AI.

The essence of deepfake and other such technologies like deeplearning, which leverage artificial technologies, lies in training the software to think and adapt itself using past data it is fed through data sets. You can read more about artificial intelligence in this article.

Using GPT-2, you can just punch in the headline and the deepfake text algorithm will generate a fictitious news story around that headline. Or simply supply it a first line of a poem and it will return the whole verse.

Many media houses are using deepfake algorithms coupled with web scrapping to generate stories or blogs that are written by the software themselves.

Researchers at the Middlebury Institute of International Studies’ Center on Terrorism, Extremism, and Counter-terrorism (CTEC) warns that tools like GPT-2 can be misused to propagate racial supremacy or disseminate radical messages by extremist organizations.

Deepfakes On Social Media

In tandem with writing stories or blogs, deepfake technology can also be leveraged to create a fake online profile that would be hard for a normal user to discern. For example, a Bloomberg (non-existing) journalist with the name Maisy Kinsley on social networking sites like LinkedIn and Twitter was plausibly a deepfake. Her profile picture appeared strange, perhaps computer generated. The profile was probably created for financial benefit, as the profile of Maisy Kinsley repeatedly tried to connect with short-sellers of Tesla stock on social media. Short-sellers are people who are bearish on the stock market and they short, i.e, sell the stock with the conviction that the stock price will fall and then buy stock at a lower price, effectively generating a handsome profit.

Another profile with the name Katie Jones, which supposedly mentioned working at the Center for Strategic and International Studies, was found to be a deepfake created with the mala fide intention of spying.

Detecting Textual Deepfakes

Researchers from the Allen Institute for Artificial Intelligence have developed a software tool called Grover to detect synthetic content floating online. Researchers claim that this software is able to detect deepfake-written essays 92% of the time. Grover works on a test set compiled from Common Crawl, an open-source web archive and crawler. Similarly, a team of scientists from Harvard and the MIT-IBM Watson laboratory have come together to design Giant Language Model Test Room, a web tool that seeks to discern whether the text inputted is generated by AI.

Also Read: Bot Or Not: How To Tell A Bot From A Human

Deepfake Video

Making fake photos and videos is the main arsenal of deepfakes. It is the most used form of deepfake, given that we are living in the ubiquitous world of social media, wherein pictures and videos elucidate events and stories better than plain text.

Modern video-generating AI is more capable, and perhaps more dangerous, than its natural language counterpart. Seoul-based tech company Hyperconnect recently developed a tool called MarioNETte that can generate deepfake videos of historical figures, celebrities, and politicians. This is done through facial reenactment by another person, whose facial expressions are then superimposed on the targeted personality whose deepfake is to be created.

How Deepfake Video Is Produced?

This video trickery employs a technique called generative adversarial network (GAN). GAN is a part of a machine learning branch called neural networks. These networks are designed to emulate the neuronal processes of the human brain. Programmers can train neural networks to recognize or manipulate a specific task.

In GAN used for deepfake generation, two neural networks are pitted against each other to generate a realistic output. The purpose of doing so is to ensure that the deepfakes are created to look as real as possible. The essence of GAN lies in the rivalry between the two neural networks. In GAN, the picture forger and the forgery detector repeatedly attempt to outsmart one another. Both the neural networks are trained using the same data set.

The first net is called the generator, whose job it is to generate a forged image using noise vectors (a list of random numbers) that look as realistic as possible. The second net, called the discriminator, determines the veracity of the generated images. It compares the forged image generated by the generator with the genuine images in the data set to determine which images are real and which are fake. On the basis of those results, the generator varies the parameter for generating images. This cycle goes on until the discriminator fails to ascertain that a generated image is bogus, which is then used in the final output. This is why deepfakes look so eerily real.

Deepfake Technology: What Are Deepfakes? How Do They Make Deepfakes? (2)

Detecting Deepfake Videos

Forensic experts across the globe are toiling hard to come up with ways and tools to identify deepfakes, as they are becoming more and more convincing every day.

For example, consider this deepfake demonstration video of Obama released by Buzzfeed in 2018, which stupefied viewers across the globe. You can check it out here:

As machine-learning tools are reaching the masses, it has become much easier to create convincing fake videos that could be used to disseminate propaganda-driven news or to simply harass a targeted individual.

The US Defense Department (DARPA) has released a tool for detecting deepfakes called Media Forensics. Originally, the program was developed to automate existing forensic tools, but with the rise of deepfakes, they have used AI to counter AI-driven deepfakes. Let’s see how it works.

The resultant video generated using deepfake technically has discernible differences in the way the video’s metadata is distributed, as compared to the original one. These differences are referred to as glimpses in matrix, which is what DARPA’s deepfake detection tool tries to leverage when detecting deepfake media.

Siwei Lyu, professor from the computer science department at the State University of New York, has noted that faces created using deepfake technology seldom blink. Even if they do, it seems unnatural. He posits that this is because most of the deepfake-driven videos are trained using still images. Still, photographs of a person are generally taken when their eyes are open. Besides eye blinking, other data points on facial movements, such as when they raise upper lip while conversing, how they shake their heads etc. can also provide clues as to whether the streamed video is fake.

Deepfake Audio

The power of artificial intelligence and neural networks isn’t just limited to text, pictures, and video. They can clone a person’s voice with the same ease. All that is required is a data set of the audio recording of a person whose voice needs to be emulated. Deepfake algorithms will learn from that data set and becomes empowered to recreate the prosody of a targeted person’s speech.

Commercial software is being released in the market like Lyrebird and Deep Voice, wherein you need to speak only a few sentences before the AI has grown accustomed to your voice and intonation. As you feed in more audio of yourself, this software becomes powerful enough to clone your voice. After feeding in the dataset of your own audio samples, you can just give a sentence or a paragraph and this deepfake software will narrate the text in your voice!

Detecting Deepfake Audio

Right now, there are not many dedicated deepfake audio tools available, but developers and cybersecurity companies are working in this domain to come up with better protective solutions in this regard.

For example, last year, developers at tech startup Resemble developed an open-source tool called Resemblyzer for the detection of deepfake audio clips. Resemblyzer uses advanced machine-learning algorithms for deriving computation representations of voice samples to predict whether they are real or fake. Whenever a user submits an audio file for evaluation, it generates a mathematical representation summarizing the unique characteristics of the submitted voice sample. Through this conversion, it becomes possible for the machine to detect whether the voice is real or artificially produced by deepfake tools.

Also Read: What Is Artificial Intelligence And How Is It Powering Our Lives?

The Road Ahead With Deepfakes

An investigation done by Deeptrace labs last year found that over 14,000 deepfake videos are lurking online. They have also noted a jump of 84% in their production in a span of just seven months. Interestingly, more than 90% of deepfake videos are p*rnographic material, wherein famous women are face-swapped in p*rn.

As deepfake is getting serious traction, it is posing a serious problem of intruding not just on the privacy, but also the dignity of individuals. Ironically, to counter AI-powered deepfakes, artificial intelligence itself is being used. Although a ‘good’ AI is helping to identify deepfakes, this detection system relies highly upon the dataset it consumes for training. This means they can work well to detect deepfake videos of celebrities, as a vast amount of data is available about them. But to detect the deepfake of a person who has a low profile would be challenging for such detection systems.

Social media tech giants are working on deepfake detection systems as well. Facebook recently announced that it is working on an automated system to identify deepfake content on its platform and weed it out. On similar lines, Twitter proposed flagging deepfakes and eliminating them if they’re found to be provocative.

Although we acknowledge and appreciate efforts by these tech companies, only time will tell how successful they are at keeping malicious deepfakes at bay!

References (click to expand)
  1. (2019) Deep fakes – an emerging risk to individuals and societies alike. Tilburg University
  2. Solaiman, I., Brundage, M., Clark, J., Askell, A., Herbert-Voss, A., Wu, J., … Wang, J. (2019). Release Strategies and the Social Impacts of Language Models (Version 2). arXiv.
  3. Politically linked deepfake LinkedIn profile sparks spy fears .... The Register
  4. Computer Science and Telecommunications Board, Intelligence Community Studies Board, Division on Engineering and Physical Sciences, & National Academies of Sciences, Engineering, and Medicine. (2019). Implications of Artificial Intelligence for Cybersecurity. (A. Johnson & E. Grumbling, Eds.), []. National Academies Press.
  5. CorentinJ/Real-Time-Voice-Cloning - GitHub. GitHub, Inc.
Deepfake Technology: What Are Deepfakes? How Do They Make Deepfakes? (2024)

FAQs

Deepfake Technology: What Are Deepfakes? How Do They Make Deepfakes? ›

Using artificial intelligence, deepfakes can mimic a person's voice and facial features. The technology uses an audio recording of someone's voice to make it say things that the person might never have said. It can mimic someone's facial movements from videos of them, or even just a picture of their face.

How are deepfakes created? ›

While the act of creating fake content is not new, deepfakes leverage tools and techniques from machine learning and artificial intelligence, including facial recognition algorithms and artificial neural networks such as variational autoencoders (VAEs) and generative adversarial networks (GANs).

What technology is used to make deepfakes? ›

The artificial intelligence and deep-learning technology currently used for deepfakes typically involve generative adversarial networks, or GANs, and autoencoders.

Is it illegal to make deepfakes? ›

There are no federal laws banning deepfake p*rn, but several bills have been introduced in Congress, including the AI Labeling Act of 2023 and the DEFIANCE Act of 2024.

What are people using to make deepfakes? ›

Most deepfake tools and apps utilize machine learning, APIs, programming language, neural network algorithms, computer vision, and generative adversarial networks (GANs).

Can deepfakes be tracked? ›

As these generative artificial intelligence (AI) technologies become more common, researchers are now tracking their proliferation through a database of political deepfakes.

How hard are deepfakes to make? ›

Creating a deepfake used to be a time-consuming, expensive process that required a high level of technical expertise. However, with the advent of machine learning algorithms, it's now possible to make a deepfake in just a few minutes using off-the-shelf software and a few dollars.

Do you need permission to deepfake someone? ›

The Copyright Act provides copyright protection for works, including films, music, and creative content. Individuals infringing upon copyrights by creating deepfakes using copyrighted works without permission can face legal action under this act.

What is the unethical use of deepfakes? ›

Deepfakes are a subset of AI outputs, utilizing deep learning techniques like generative adversarial networks (GANs) to generate highly realistic but fabricated content, often raising ethical and legal concerns related to privacy, intellectual property, consent and the spread of misinformation.

Can software detect deepfakes? ›

Deepware is advanced software that uses artificial intelligence and machine learning technologies to detect and mitigate deepfakes. It identifies videos, images, and audio files and determines if they are fake or not.

What are the bad uses of deepfakes? ›

Not only has this technology created confusion, skepticism, and the spread of misinformation, deepfakes also pose a threat to privacy and security. With the ability to convincingly impersonate anyone, cybercriminals can orchestrate phishing scams or identity theft operations with alarming precision.

Is deep fake easy? ›

Initially, understanding what is a deepfake was simpler as early deepfakes were relatively easy to spot due to their low quality and visible flaws. However, as AI algorithms have become more sophisticated and computing power has increased, deepfakes have become incredibly realistic and harder to detect.

What techniques are used in deepfake creation? ›

Deepfakes are made using deep learning, a type of artificial intelligence. They specifically use techniques like Generative Adversarial Networks (GANs). In this process, two neural networks, 'generator' and 'discriminator' work in tandem. The realism of deepfakes has been increasing due to advancements in AI.

What is the algorithm for deep fake? ›

Deepfake content is created by using two algorithms that compete with one another. One is called a generator and the other one is called a discriminator. The generator creates the fake digital content and asks the discriminator to find out if the content is real or artificial.

What are the system requirements for deepfakes? ›

As with all machine learning techniques, deepfakes can be created on any PC with a minimum of 4 GB of RAM. However, a machine with 8 GB of RAM or higher and a GPU (a graphics card) is strongly recommended. Training a model on a CPU is likely to take months to complete, which does not make it a realistic endeavor.

How is deepfake detected? ›

Blurring or misalignment: If the edges of images are blurry or visuals are misaligned — for example, where someone's face and neck meet their body — you'll know something is amiss. Inconsistent audio and noise: Deepfake creators usually spend more time on the video images rather than the audio.

Top Articles
Latest Posts
Article information

Author: Rueben Jacobs

Last Updated:

Views: 6339

Rating: 4.7 / 5 (77 voted)

Reviews: 92% of readers found this page helpful

Author information

Name: Rueben Jacobs

Birthday: 1999-03-14

Address: 951 Caterina Walk, Schambergerside, CA 67667-0896

Phone: +6881806848632

Job: Internal Education Planner

Hobby: Candle making, Cabaret, Poi, Gambling, Rock climbing, Wood carving, Computer programming

Introduction: My name is Rueben Jacobs, I am a cooperative, beautiful, kind, comfortable, glamorous, open, magnificent person who loves writing and wants to share my knowledge and understanding with you.