This book is a survey and analysis of how deep learning can be used to generate musical content. The authors offer a comprehensive presentation of the foundations of deep learning techniques for music generation. They also develop a conceptual framework used to classify and analyze various types of architecture, encoding models, generation strategies, and ways to control the generation. The five dimensions of this framework are: objective (the kind of musical content to be generated, e.g., melody, accompaniment); representation (the musical elements to be considered and how to encode them, e.g., chord, silence, piano roll, one-hot encoding); architecture (the structure organizing neurons, their connexions, and the flow of their activations, e.g., feedforward, recurrent, variational autoencoder); challenge (the desired properties and issues, e.g., variability, incrementality, adaptability); and strategy (the way to model and control the process of generation, e.g., single-step feedforward, iterative feedforward, decoder feedforward, sampling). To illustrate the possible design decisions and to allow comparison and correlation analysis they analyze and classify more than 40 systems, and they discuss important open challenges such as interactivity, originality, and structure.
The authors have extensive knowledge and experience in all related research, technical, performance, and business aspects. The book is suitable for students, practitioners, and researchers in the artificial intelligence, machine learning, and music creation domains. The reader does not require any prior knowledge about artificial neural networks, deep learning, or computer music. The text is fully supported with a comprehensive table of acronyms, bibliography, glossary, and index, and supplementary material is available from the authors' website.