Chapter 15. Music and Deep Learning
The other chapters in this book are all about processing of images or texts. Those chapters represent the balance of media in deep learning research, but that is not to say that sound processing isnât interesting and that we havenât seen some great developments in this area in the last few years. Speech recognition and speech synthesis are what made home assistants like Amazon Alexa and Google Home a possibility. The old sitcom joke where the phone dials the wrong number hasnât really been current since Siri came out.
It is easy to start experimenting with these systems; there are APIs out there that let you get a simple voice app up and running in a few hours. The voice processing, however, is done in Amazon, Google, or Appleâs data center, so we canât really count these as deep learning experiments. Building state-of-the-art voice recognition systems is hard, although Mozillaâs Deep Speech is making some impressive progress.
This chapter focuses on music. Weâll start out with training a music classification model that can tell us what music weâre listening to. Weâll then use the results of this model to index local MP3s, making it possible to find songs similar in style. After that weâll use the Spotify API to create a corpus of public playlists that weâll use to train a music recommender.
The notebooks for this chapter are:
15.1 Song Classification 15.2 Index Local MP3s 15.3 Spotify Playlists 15.4 Train a Music Recommender ...
Get Deep Learning Cookbook now with the O’Reilly learning platform.
O’Reilly members experience books, live events, courses curated by job role, and more from O’Reilly and nearly 200 top publishers.