Passer au contenu

/ Department of Computer Science and Operations Research

Je donne

Rechercher

Soutenance de thèse de Yusong Wu

Dear all / Bonjour à tous,

We are happy to invite you to the Predoc III Evaluation of Yusong Wu on Friday, August 23rd at 1 pm (hybrid mode).

Vous êtes cordialement invité.e.s à la soutenance de thèse de Yusong Wu , le vendredi, 23 Août à 1pm (mode hybride).


Title: Synchronous Music Generative Models via Reinforcement Learning

Date: August 23rd, 1 pm - 3h30 pm

Location: Agora

Link: Lien zoom

 

Jury

President
Aishwarya Agrawal
Director
Aaron Courville
Co-Director
Anna Huang
Regular Member
Pablo Samuel Castro

 

Abstract

 

While existing generative models are adept at creating expressive
compositions and accompaniments, they are not designed for live,
interactive scenarios that require real-time collaboration, adaptation, and
anticipation among multiple participants. Synchronous music interaction,
such as jamming, accompaniment, and improvisation, involves simultaneous
musical actions that necessitate a high degree of coordination and
collaborative creativity.

To address this gap, in the proposal, we aim to develop a framework of
building models for real-time synchronous music interaction. We first
explore the cognitive and physiological foundations of such interactions
and review the current state of music generation models and datasets.
Building on this background, we propose a formulation of the synchronous
music interaction problem as a multi-agent general-sum game. We then
outline a design approach that conceptualizes music interaction models as
generative agents, which are first pretrained using maximum likelihood
estimation and then fine-tuned with reinforcement learning.

We also present ReaLchords, an online generative model developed as part of
this research, which improvises chord accompaniments in real-time to a
user’s melody. Through experiments and listening tests, ReaLchords has
demonstrated effective adaptation to unfamiliar inputs, producing
harmonically and temporally coherent accompaniments. We will also present a
real-time interactive music system, ReaLJam, built upon ReaLchords,
showcasing how ReaLchords enables dynamic and responsive real-time music
interaction.

Finally, the proposal outlines ongoing and future work, including the
extension of ReaLchords to multi-agent co-adaptive settings, the
development of models that interact via audio inputs and outputs, and the
exploration of learning from human interactions and feedback.