Eerily Realistic Fake Videos Can generates By A new computer program | The Tech Blog World


About The Tech Blog World


The Tech Blog World

|The Tech Blog World| Admin/Author

The Tech Blog World is a online publication Media/News Website mainly focused on technology, with editorial team in the India. The important things is that, it provides tech news and tech reviews of technology and it was first launched in 2016. To Know More About The Tech Blog World, Visit This Link Here. To Know More About Us.





SUBSCRIBE TO OUR DAILY EMAIL NEWSLETTER & UPDATES


Make Sure You Don't Miss Interesting Happenings By Joining Our Newsletter Program.

Enter your email address here:



No spam, we promise. You can unsubscribe at any time and we'll never share your details without your permission.


"Do You Think, This Post Is Useful To Others. If You Like This Post". Please Do Share This Post.

SHARE THIS POST ON SOCIAL MEDIA:

Eerily Realistic Fake Videos Can generates By A new computer program

A new computer program generates eerily realistic fake videos

A new computer program generates eerily realistic fake videos



A new computer program can manipulate a video such that the person on-screen mirrors the movements and expressions of someone in a different video. Unlike other film-fudging software, this program can tamper with far more than facial expressions. The algorithm, to be presented August 16 at the 2018 SIGGRAPH meeting in Vancouver, also tweaks head and torso poses, eye movements and background details to create more lifelike fakes.

These video forgeries are “astonishingly realistic,” says Adam Finkelstein, a computer scientist at Princeton University not involved in the work. This system could help produce dubbed films where the actors’ lip movements match the voiceover, or even movies that star dead actors reanimated through old footage, he says. But giving internet users the power to create ultrarealistic phony videos of public figures could also take fake news to the next level (SN: 8/4/18, p. 22).
The algorithm starts by scanning two videos frame by frame, tracking 66 facial “landmarks” — like points along the eyes, nose and mouth — to map a person’s features, expression, head tilt and line of sight. For example, those videos might show former President Barack Obama and Russian President Vladimir Putin. Then, to make Putin mimic Obama’s behavior, the program distorts Putin’s image to adopt Obama’s head pose, facial expression and eye line in each frame. The program can also tweak shadows, change Putin’s hair or adjust the height of his shoulders to match his new head pose. The result is a video of Putin doing an eerily on-point imitation of Obama’s exact motions and expressions.

Copycat

A new computer program analyzes the appearance of someone in one video (the “input”) and transfers that person’s facial expression, head pose and line of sight onto a person in another video (the “output”). That can generate footage of the second person doing and saying things they never actually did.


A new computer program generates eerily realistic fake videos
Computer scientist Christian Theobalt of the Max Planck Institute for Informatics in Saarbrücken, Germany, and colleagues tested their program on 135 volunteers, who watched five-second clips of real and forged videos and reported whether they thought each clip was authentic. The dummy videos fooled, on average, 50 percent of viewers. But people may have been more critical of doctored footage during the study than they would have been if they naturally encountered these clips online because they were primed to anticipate fakes. Even when study participants were watching genuine clips, 20 percent, on average, still believed the clips were not real.
The new software still has some limits: The program can only fiddle with videos shot by a stationary camera, framed to show someone’s head and shoulders in front of an unchanging background. And the algorithm can’t shift a person’s pose too much from their original video. That is, a clip of Putin speaking directly into the camera couldn’t be edited to make him turn around, because the software wouldn’t know what the back of Putin’s head looks like.
Still, it’s easy to imagine how this kind of digital puppetry could be used to spread potentially dangerous misinformation. “The researchers who are developing this stuff are getting ahead of the curve — in the sense that now [this algorithm] is out there, and people are more aware of the types of manipulations that are possible,” says Kyle Olszewski, a computer scientist at the University of Southern California in Los Angeles.
TheTechBlogWorld. Thanks For Reading This Post. If You Like This Post, Leave Your Comment And Please Do Like & Share This Post To Others. Keep Reading "TheTechBlogWorld".
Staff Writer    Updated On : Tuesday, September 04, 2018
POST TAGS:

Post a Comment

[disqus][facebook]

Contact Form

Name

Email *

Message *

Powered by Blogger.
Javascript DisablePlease Enable Javascript To See All Widget