Stable Diffusion: 8 New Amazing Results!
6:25

Stable Diffusion: 8 New Amazing Results!

Two Minute Papers 04.07.2023 118 436 просмотров 4 769 лайков

Machine-readable: Markdown · JSON API · Site index

Поделиться Telegram VK Бот
Транскрипт Скачать .md
Анализ с AI
Описание видео
❤️ Check out Lambda here and sign up for their GPU Cloud: https://lambdalabs.com/papers 📝 The paper "Diffusion Self-Guidance for Controllable Image Generation" is available here: https://arxiv.org/abs/2306.00986 Try it out! ControlNet - https://github.com/lllyasviel/ControlNet ControlNet guide - how install and use it: https://stable-diffusion-art.com/controlnet/ SD Webui: https://github.com/AUTOMATIC1111/stable-diffusion-webui ControlNet extension for SD Webui: https://github.com/Mikubill/sd-webui-controlnet Sources: https://www.reddit.com/r/StableDiffusion/comments/13jhaih/i_used_controlnet_tile_on_an_old_space_drawing_i/ https://www.reddit.com/r/StableDiffusion/comments/13tyl2c/1536x1536_60_fps_stable_diffusion_animation_made/ https://www.reddit.com/r/StableDiffusion/comments/13ty6hb/rock_stars_as_various_characters_in_the_1800s/ https://www.reddit.com/r/StableDiffusion/comments/13uerov/some_dnd_inspired_watercolor_style_portraits/ https://www.reddit.com/r/StableDiffusion/comments/13y11oo/google_creates_diffusion_selfguidance/ https://www.reddit.com/r/StableDiffusion/comments/13y9dtm/reconstructing_baby_faces_from_4d_ultrasound/ https://twitter.com/ben_ferns/status/1665907480600391682 https://twitter.com/ben_ferns/status/1665907480600391682 https://mp.weixin.qq.com/s/i4WR5ULH1ZZYl8Watf3EPw https://www.youtube.com/watch?v=1RvZWHtFXuY My latest paper on simulations that look almost like reality is available for free here: https://rdcu.be/cWPfD Or this is the orig. Nature Physics link with clickable citations: https://www.nature.com/articles/s41567-022-01788-5 Thumbnail background image: sutrik 🙏 We would like to thank our generous Patreon supporters who make Two Minute Papers possible: Aleksandr Mashrabov, Alex Balfanz, Alex Haro, Andrew Melnychuk, Benji Rabhan, Bret Brizzee, Bryan Learn, B Shang, Christian Ahlin, Eric Martel, Geronimo Moralez, Gordon Child, Jace O'Brien, Jack Lukic, John Le, Kenneth Davis, Klaus Busse, Kyle Davis, Lorin Atzberger, Lukas Biewald, Martin, Matthew Valle, Michael Albrecht, Michael Tedder, Nikhil Velpanur, Owen Campbell-Moore, Owen Skarpness, Rajarshi Nigam, Ramsey Elbasheer, Richard Sundvall, Steef, Taras Bobrovytsky, Ted Johnson, Thomas Krcmar, Timothy Sum Hon Mun, Torsten Reil, Tybie Fitzhugh, Ueli Gallizzi. If you wish to appear here or pick up other perks, click here: https://www.patreon.com/TwoMinutePapers Thumbnail background design: Felícia Zsolnai-Fehér - http://felicia.hu Károly Zsolnai-Fehér's links: Twitter: https://twitter.com/twominutepapers Web: https://cg.tuwien.ac.at/~zsolnai/

Оглавление (7 сегментов)

Intro

Dear Fellow Scholars, this is Two Minute  Papers with Dr. Károly Zsolnai-Fehér. Today I will show you some papers  and results with Stable Diffusion,   a free and open source image generator AI. And  I cannot believe how good these results are. First, we had an earlier example with ControlNet  and dancing. This means that in goes something,

Handdrawn anime

for instance, a video of us, and out  comes a hand drawn anime version of   it. Now, the advantage of this is  that we don’t need to look perfect,   and we don’t even need to be at this exact  place in the background, because all of this   can be synthesized. And today, someone took  this idea to the next level. Look. And now,

Blue haired woman

if you think this is a blue haired woman converted  into to a more painterly blue haired woman… well,   have a look at this. Oh yes, we don’t even need  to be the character we want to be, we just need   to kind of look like the character we want to  be. This is going to be an amazing tool for the   digital artists of the future. And remember,  this needs a tiny bit of time and expertise,   but it is something that you can run for  free on your own machine, and some of the   other examples, even in your pocket, on  your smartphone. What a time to be alive! Two, we can even use Stable Diffusion’s  ControlNet to provide it input images for   it to redraw. For instance, we can give it this  childhood drawing, and we can ask this little AI

Childhood drawing

to create a super high quality version of  it. My goodness. This is incredibly good. Three, now here is a crazy idea. Why not give  it a QR-code as an input image, and see what   happens? Well, these look great…so this was  a good thought experiment, but of course,   these won’t work as QR codes. Just as a piece  of art. Wait a second. Hold on to your papers,   Fellow Scholars, do they actually work as QR-codes  and are they readable from a smartphone camera?    Yes they are! Make sure to play with the distance  a bit, perhaps increase the distance a little,   but every single one of these work for me. I  can’t believe it. Artificial creativity. Wow.

Video to video

Four, and these video to video workflows  are becoming so incredibly good that…are   you seeing what I am seeing? Can that  really be? These can now be generated   almost immediately. If we wish to create a  smoother video, we would have to wait a bit,   but not in the order of hours. Just in the  order of seconds, and super quick previews   can be done like this. This would not have been  possible just a few months ago. Unbelievable. Five, with the right prompts, Stable  Diffusion can also create an 1800s

Custom portraits

version of our favorite rockstars. Elton John,  Freddie Mercury, Dave Grohl, Hendrix or Jagger,   you name it. I love how creative some of  these solutions are and how they retain   their looks and personalities while  putting them into an 1800s setting. Six, we can also create custom portraits for our   dungeons and dragons characters.   Just give the AI the description,   and there we go. The prompts and the workflow  for some of these images are available,   which means that you can likely reproduce  and tweak them yourself at home. So good. Seven, in the meantime, Google just published  a really cool paper on how to guide these image

Natural language

generator AIs with natural language. First,  we can ask for an image of a giant macaron   and croissant in front of the Eiffel Tower. Well,  yes, I hear you saying, Doc, with today’s papers,   that is not surprising. What is surprising  is that without generating a new image,   we can swap these objects, enlarge  the macaron, and if we don’t like it,   replace it with a different one, or even better  use our own photo and ask it to recapture its   appearance. Which it had done really well as  the lighting now also has an effect on these   objects as well. Not just the background  is different, everything is different. So,   interacting with text to image AIs via  natural language. How cool is that? Eight, creating zoom videos has also improved  a great deal. Now we can choose a few images,   and stitch them together to  create these zoom videos,   and they are now really smooth. Previous  ones had some flickering issues,   but this one is now so much better. And all this  improvement in just a few months! Unbelievable. I put a link to all of these  works in the video description,   and for now, let the experiments begin! Thanks for watching and for your generous  support, and I'll see you next time!

Другие видео автора — Two Minute Papers

Ctrl+V

Экстракт Знаний в Telegram

Экстракты и дистилляты из лучших YouTube-каналов — сразу после публикации.

Подписаться

Дайджест Экстрактов

Лучшие методички за неделю — каждый понедельник