Audiocraft is a library for audio processing and generation with deep learning. It features the state-of-the-art EnCodec audio compressor / tokenizer, along with MusicGen, a simple and controllable music generation LM with textual and melodic conditioning.
Conditioners are defined in config/conditioner like chroma2music.yaml. The chroma model's parameters are populated under self_wav:.
I have defined another conditioner, and would like to train MusicGen conditioned using two WaveformConditioners, namely the ChromaStemConditioner and the new conditioner that I defined. What is the best way to do that?
Is there a syntax for doing this in the conditioner YAML or is it better to create a new WaveformConditioner that simply returns a stacked tensor formed from both conditions?
Conditioners are defined in
config/conditioner
like chroma2music.yaml. Thechroma
model's parameters are populated underself_wav:
.I have defined another conditioner, and would like to train MusicGen conditioned using two
WaveformConditioner
s, namely theChromaStemConditioner
and the new conditioner that I defined. What is the best way to do that?Is there a syntax for doing this in the conditioner YAML or is it better to create a new
WaveformConditioner
that simply returns a stacked tensor formed from both conditions?