Overview
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the groundbreaking MusicLM model in this comprehensive video breakdown. Delve into the world of text-based music generation as the presenter analyzes Google's innovative approach to creating convincing short music clips with high audio fidelity. Learn about the model's architecture, including key components like SoundStream, w2v-BERT, and MuLan. Understand the training and inference processes, examine experimental results, and discuss limitations. Gain insights into the research procedure behind this cutting-edge technology that combines deep learning base models to revolutionize the Music AI community. Compare MusicLM with other text-to-music models like Riffusion and Mubert AI, and witness demonstrations of its capabilities.
Syllabus
Intro
Text-to-music
MusicLM demo
Riffusion and Mubert AI
MusicLM architecture
Components overview
SoundStream
w2v-BERT
MuLan
Training
Inference
Experiments
Limitations
Thoughts on research procedure
Taught by
Valerio Velardo - The Sound of AI