Generating Sound with Neural Networks: Difference between revisions
Jump to navigation
Jump to search
(3 intermediate revisions by the same user not shown) | |||
Line 9: | Line 9: | ||
= Sound Generation with Deep Learning - Approaches and Challenges = | = Sound Generation with Deep Learning - Approaches and Challenges = | ||
<evlplayer id="player1" w="480" h="360" service="youtube" defaultid="pwV8K9wXY2E" /> | <evlplayer id="player1" w="480" h="360" service="youtube" defaultid="pwV8K9wXY2E" /> | ||
0:00 Intro | |||
0:33 Defining the sound generation task | |||
1:17 Classification of sound generation systems | |||
2:14 Types of generated sounds | |||
3:41 Sound representations | |||
4:07 Generation from raw audio | |||
7:40 Challenges of raw audio generation | |||
10:21 Generation from spectrograms | |||
16:12 Advantages of generation from spectrograms | |||
18:07 Challenges of generation from spectrograms | |||
20:26 Can we generate sound with MFCCs? | |||
21:26 DL architectures for sound generation | |||
22:13 Inputs for generation | |||
24:03 Details about the sound generative system we'll build | |||
24:44 What's next? | |||
Mentioned papers: | |||
* Wavenet: A Generative Model for Raw Audio: | |||
https://arxiv.org/pdf/1609.03499.pdf | |||
* Jukebox: A Generative Model for Music | |||
https://arxiv.org/pdf/2005.00341 | |||
* DrumGAN: Synthesis of Drum Sounds with Timbral Feature Conditioning Using Generative Adversarial Networks | |||
https://arxiv.org/pdf/2008.12073 | |||
* Melnet: A generative model for audio in the frequency domain | |||
https://arxiv.org/pdf/1906.01083.pdf | |||
= | = Autoencoders Explained Easily= | ||
<evlplayer id="player1" w="480" h="360" service="youtube" defaultid=" | <evlplayer id="player1" w="480" h="360" service="youtube" defaultid="xwrzh4e8DLs" /> | ||
= How to Implement Autoencoders in Python and Keras || The Encoder = | |||
<evlplayer id="player1" w="480" h="360" service="youtube" defaultid="TtyoFTyJuEY" /> | |||
= How to Implement Autoencoders in Python and Keras - The Decoder = | = How to Implement Autoencoders in Python and Keras - The Decoder = |
Latest revision as of 15:37, 12 December 2021
Generating Sound with Neural Networks
Learn how to generate sound from audio files 🎧 🎧 using Variational Autoencoders.
Sound Generation with Neural Networks - INTRO[edit]
Sound Generation with Deep Learning - Approaches and Challenges[edit]
0:00 Intro 0:33 Defining the sound generation task 1:17 Classification of sound generation systems 2:14 Types of generated sounds 3:41 Sound representations 4:07 Generation from raw audio 7:40 Challenges of raw audio generation 10:21 Generation from spectrograms 16:12 Advantages of generation from spectrograms 18:07 Challenges of generation from spectrograms 20:26 Can we generate sound with MFCCs? 21:26 DL architectures for sound generation 22:13 Inputs for generation 24:03 Details about the sound generative system we'll build 24:44 What's next?
Mentioned papers:
- Wavenet: A Generative Model for Raw Audio:
https://arxiv.org/pdf/1609.03499.pdf
- Jukebox: A Generative Model for Music
https://arxiv.org/pdf/2005.00341
- DrumGAN: Synthesis of Drum Sounds with Timbral Feature Conditioning Using Generative Adversarial Networks
https://arxiv.org/pdf/2008.12073
- Melnet: A generative model for audio in the frequency domain
https://arxiv.org/pdf/1906.01083.pdf
Autoencoders Explained Easily[edit]
How to Implement Autoencoders in Python and Keras || The Encoder[edit]
How to Implement Autoencoders in Python and Keras - The Decoder[edit]
Building and Training an Autoencoder in Keras + TensorFlow + Python[edit]
Saving the Autoencoder in Keras[edit]
Generation with AutoEncoders: Results and Limitations[edit]
From Autoencoders to Variational Autoencoders: Improving the Encoder[edit]
From Autoencoders to Variational Autoencoders: Improving the Loss Function[edit]
How to implement a Variational AutoEncoder in Python and Keras[edit]
Preprocessing Audio Datasets for Machine Learning[edit]
Training a VAE with Speech Data in Keras[edit]