ProGress: Structured Music Generation via Graph Diffusion and Hierarchical Music Analysis

Demo Page
Anonymous Authors
[Paper] | [Code Repo]


Abstract

Artificial Intelligence (AI) for music generation is undergoing rapid developments, with recent symbolic models leveraging sophisticated deep learning and diffusion-model algorithms. One drawback of existing models is that they lack structural cohesion, particularly with respect to harmonic–melodic structure. Furthermore, such existing models are largely “black-box” in nature and are not musically interpretable. This paper addresses these limitations via a novel generative music framework that incorporates concepts of Schenkerian analysis (SchA) in concert with a diffusion modeling framework. This framework, which we call ProGress (Prolongation-enhanced DiGress), adapts state-of-the-art deep models for discrete diffusion (in particular, the DiGress model of Vignac et al., 2023) for interpretable and structured music generation.

Concretely, our contributions include:

  1. Novel adaptations of the DiGress model for music generation;
  2. A novel SchA-inspired phrase fusion methodology; and
  3. A framework allowing users to control various aspects of the generation process to create coherent musical compositions.

Results from human experiments suggest superior performance to existing state-of-the-art methods.

ProGress Survey Examples


A Little Tango


We used our model trained on Bach to generate music with tango rhythms. Simply changing the timbre made some nice results!