NVIDIA’s AI: Amazing DeepFakes And Virtual Avatars!
5:29

NVIDIA’s AI: Amazing DeepFakes And Virtual Avatars!

Two Minute Papers 18.09.2022 98 777 просмотров 4 326 лайков

Machine-readable: Markdown · JSON API · Site index

Поделиться Telegram VK Бот
Транскрипт Скачать .md
Анализ с AI
Описание видео
❤️ Check out Weights & Biases and say hi in their community forum here: https://wandb.me/paperforum 📝 The paper "One-Shot Free-View Neural Talking-Head Synthesis for Video Conferencing" is available here: https://nvlabs.github.io/face-vid2vid/ Try it out: http://imaginaire.cc/vid2vid-cameo/ ❤️ Watch these videos in early access on our Patreon page or join us here on YouTube: - https://www.patreon.com/TwoMinutePapers - https://www.youtube.com/channel/UCbfYPyITQ-7l4upoX8nvctg/join 🙏 We would like to thank our generous Patreon supporters who make Two Minute Papers possible: Aleksandr Mashrabov, Alex Balfanz, Alex Haro, Andrew Melnychuk, Benji Rabhan, Bryan Learn, B Shang, Christian Ahlin, Eric Martel, Geronimo Moralez, Gordon Child, Ivo Galic, Jace O'Brien, Jack Lukic, John Le, Jonas, Jonathan, Kenneth Davis, Klaus Busse, Kyle Davis, Lorin Atzberger, Lukas Biewald, Matthew Allen Fisher, Michael Albrecht, Michael Tedder, Nevin Spoljaric, Nikhil Velpanur, Owen Campbell-Moore, Owen Skarpness, Rajarshi Nigam, Ramsey Elbasheer, Steef, Taras Bobrovytsky, Ted Johnson, Thomas Krcmar, Timothy Sum Hon Mun, Torsten Reil, Tybie Fitzhugh, Ueli Gallizzi. If you wish to appear here or pick up other perks, click here: https://www.patreon.com/TwoMinutePapers Thumbnail background design: Felícia Zsolnai-Fehér - http://felicia.hu Károly Zsolnai-Fehér's links: Instagram: https://www.instagram.com/twominutepapers/ Twitter: https://twitter.com/twominutepapers Web: https://cg.tuwien.ac.at/~zsolnai/ #nvidia #deepfake

Оглавление (2 сегментов)

Segment 1 (00:00 - 05:00)

Dear Fellow Scholars, this is Two Minute  Papers with Dr. Károly Zsolnai-Fehér. Today we are going to see that the  research papers that you see here   in this series are real. Here you see NVIDIA’s  game changing videoconferencing AI. So what does   this do? Why is this so interesting? How does  this transfer a video of us over the internet?    Well, here is a crazy idea. It doesn’t! What?   Transmitting video without transmitting video?    How is that even possible?   Well, now, it is possible! What they do in this work, is take only the  first image from the video, and they throw away   the entire video afterwards! But before that,  it stores a tiny bit of information from it,   which is, how our head is moving over time,  and how our expressions change. That is an   absolutely outrageous idea, except the fact that  it works. It not only works, but it works really   well. And because this is an amazing paper,  it does not stop there - it can do even more. Look at these two previous methods  trying to frontalize the input video.    This means that we look to the side a little,  and the algorithm synthesizes a new image of   us as if the camera was right in front of us.   That sounds like a science fiction movie, except   that it seems absolutely impossible given how  much these techniques are struggling with the   task…until we look at the new method. My goodness.   There is some jumpiness in the neck movement   in the output video here, and some warping issues  here, but otherwise, very impressive results.    Now if you have been holding on to your papers  so far, now, squeeze that paper, because these   previous methods are not some ancient papers  that were published a long time ago. Not at   all! Both of them were published within the same  year as the new paper. How amazing is that. Wow. And it could also perform these deepfakes too.   Look! We only need one image of the target person,   and we can transfer all of our gestures to  them, in a way that is significantly better   than most previous methods. Now, of course,  not even this technique was perfect,   it still struggled a great deal in the presence  of occluder objects, but still, just the fact   that this is now possible feels like we are  living in the future. What a time to be alive! Now, I said that today we are going to  see that this paper is as real as it gets,   so, what does that mean? Well, today,   anyone can try this technique. Yes, just  one year after publishing this paper,   and it is now available as a demo. The link to  it is available in the video description below. And, get this, some people at NVIDIA are already  using it for their virtual meetings. And by “it”,   I mean both the compression engine, where  the previous industry standard compression   algorithm could do this when given very little  data. That is not much. But the new technique,   with the same amount of data, can now  do this. That is insanity. Loving it. And, they also use their own gestures to create  these deepfakes and make virtual characters   come alive, or to frontalize their videos when  talking to each other. It almost feels like we   are living in a science fiction movie.   And all this is out there for us to use. Also, these technologies will soon  be part of the NVIDIA Video Codec SDK   as the AI Face Codec, which means that it will be  soon deployed to an even wider audience to use.    These companies are already using it. So,  this is one more amazing example that shows   that the papers that you see here in Two Minute  Papers are real. Sometimes so real that we can go   from a research paper to a product in just  a year. That is absolutely miraculous. So, what do you think? What would you use  this for? Let me know in the comments below!

Segment 2 (05:00 - 05:00)

Thanks for watching and for your generous  support, and I'll see you next time!

Другие видео автора — Two Minute Papers

Ctrl+V

Экстракт Знаний в Telegram

Экстракты и дистилляты из лучших YouTube-каналов — сразу после публикации.

Подписаться

Дайджест Экстрактов

Лучшие методички за неделю — каждый понедельник