deep fakes explained|| Introduction To Deepfake Technology by Waqas Toor

1 year ago
327

The fake in Deep fake is transparent: deep fakes are not real; they're fake. The deep-in-deep fake is a bit murkier. Deep has long-established use to describe what is difficult, complicated, intense, etc., and that use applies here, but deep in deep fake is also likely related to its use in deep learning, another term that has yet to meet our criteria for entry. The meaning of deep learning is still settling, but most often it refers to a kind of machine learning (that is, a complex process by which a computer can improve its performance) that uses layered neural networks (that is, computer architecture in which processors are interconnected in a way that suggests neural connections in the human brain) to enhance the accuracy of the machine learning algorithms.

Deepfakes employ two separate sets of algorithms acting in conjunction: the first algorithm creates a video, and the second one tries to determine if the video is real or not. If the second algorithm can tell that the video is fake, the first algorithm tries again, having learned from the second algorithm what not to do. And the pair of algorithms go around and around until they spit out a result that the programmers feel is sufficiently real-looking.

who are made
University researchers and special effects studios have long pushed the boundaries of what’s possible with video and image manipulation. But deepfakes themselves were born in 2017 when a Reddit user of the same name posted doctored porn clips on the site. The videos swapped the faces of celebrities – Gal Gadot, Taylor Swift, Scarlett Johansson, and others – onto porn performers.

It takes a few steps to make a face-swap video. First, you run thousands of face shots of the two people through an AI algorithm called an encoder. The encoder finds and learns similarities between the two faces, and reduces them to their shared common features, compressing the images in the process. A second AI algorithm called a decoder is then taught to recover the faces from the compressed images. Because the faces are different, you train one decoder to recover the first person’s face, and another decoder to recover the second person’s face. To perform the face swap, you simply feed encoded images into the “wrong” decoder. For example, a compressed image of person A’s face is fed into the decoder trained on person B. The decoder then reconstructs the face of person B with the expressions and orientation of face A. For a convincing video, this has to be done on every frame.

Loading comments...