# DeepMind's AI Learns The Piano From The Masters of The Past

## Метаданные

- **Канал:** Two Minute Papers
- **YouTube:** https://www.youtube.com/watch?v=8GUYAVXmhsI
- **Дата:** 29.07.2018
- **Длительность:** 4:00
- **Просмотры:** 58,293

## Описание

The paper "The challenge of realistic music generation: modelling raw audio at scale" is available here:
https://arxiv.org/abs/1806.10474
https://drive.google.com/drive/folders/1fvS-DU8AcK078-5k6WGudiBn0XSeE0_D

Pick up cool perks on our Patreon page: https://www.patreon.com/TwoMinutePapers

We would like to thank our generous Patreon supporters who make Two Minute Papers possible:
313V, Andrew Melnychuk, Angelos Evripiotis, Brian Gilman, Christian Ahlin, Christoph Jadanowski, Dennis Abts, Emmanuel, Eric Haddad, Esa Turkulainen, Geronimo Moralez, Kjartan Olason, Lorin Atzberger, Marten Rauschenberg, Michael Albrecht, Michael Jensen, Morten Punnerud Engelstad, Nader Shakerin, Owen Skarpness, Rafael Harutyuynyan, Raul Araújo da Silva, Rob Rowe, Robin Graham, Ryan Monsurate, Shawn Azman, Steef, Steve Messina, Sunil Kim, Thomas Krcmar, Torsten Reil, Zach Boldyga.
https://www.patreon.com/TwoMinutePapers

Two Minute Papers Merch:
US: http://twominutepapers.com/
EU/Worldwide: https://shop.spreadshirt.net/TwoMinutePapers/

Thumbnail background image credit: https://pixabay.com/photo-1839406/
Score image credit: https://pixabay.com/en/piano-music-score-music-sheet-1655558/
Splash screen/thumbnail design: Felícia Fehér - http://felicia.hu

Károly Zsolnai-Fehér's links:
Facebook: https://www.facebook.com/TwoMinutePapers/
Twitter: https://twitter.com/karoly_zsolnai
Web: https://cg.tuwien.ac.at/~zsolnai/

## Содержание

### [0:00](https://www.youtube.com/watch?v=8GUYAVXmhsI) Segment 1 (00:00 - 04:00)

Dear Fellow Scholars, this is Two Minute Papers with Károly Zsolnai-Fehér. Today, we will listen to a new AI from DeepMind that is capable of creating beautiful piano music. Because there are many other algorithms that do that, to put things into perspective, let's talk about the two key differentiating factors that set this method apart from previously existing techniques. One, music is typically learned from high-level representations, such as the score or MIDI data. This is a precise representation of what needs to be played, but they don't tell us how to play them. These small nuances are what makes the music come alive, and this is exactly what is missing from most of the synthesis techniques. This new method is able to learn these structures and generates not midi signals but raw audio waveforms. And two, it is better at retaining stylistic consistency. Most previous techniques create music that is consistent on a shorter time-scale, but do not take into consideration what was played 30 seconds ago, and therefore they lack the high-level structure that is the hallmark of quality songwriting. However, this new method shows stylistic consistency over longer time periods. Let's give it a quick listen and talk about the architecture of this learning algorithm after that. While we listen, I'll show you the composers it has learned from to produce this. I have never heard any AI-generated music before with such articulation and the harmonies are also absolutely amazing. Truly stunning results. It uses an architecture that goes by the name autoregressive discrete autoencoder. This contains an encoder module that takes a raw audio waveform and compresses it down into an internal representation, where the decoder part is responsible for reconstructing the raw audio from this internal representation. Both of them are neural networks. The autoregressive part means that the algorithm looks at previous time steps in the learned audio signals when producing new notes, and is implemented in the encoder module. Essentially, this is what gives the algorithm longer-term memory to remember what it played earlier. As you have seen the dataset the algorithm learned from as the music was playing, I am also really curious how we can exert artistic control over the output by changing the dataset. Essentially, you can likely change what the student learns by changing the textbooks used to teach them. For now, let's marvel at one more sound sample. This is already incredible, and I can only imagine what we will be able to do not ten years from now, just a year from now. Thanks for watching and for your generous support, and I'll see you next time!

---
*Источник: https://ekstraktznaniy.ru/video/14435*