This is a trainer for MusicGen model. Currently it's very basic but I'll add more features soon.
More information on the current training quality on the experiments section
Create a folder, in it, place your audio and caption files. They must be WAV and TXT format respectively.
In this example, segment_000.txt contains the caption "jazz music, jobim" for wav file segment_000.wav
Run python3 run.py --dataset /home/ubuntu/dataset
, replace /home/ubuntu/dataset
with the path to your dataset. Make sure to use the full path, not a relative path.
dataset_path
: String, path to your dataset with WAV and TXT pairs.model_id
: String, MusicGen model to use. Can besmall
/medium
/large
. Default:medium
lr
: Float, learning rate. Default:0.0001
/1e-4
epochs
: Integer, epoch count. Default:5
use_wandb
: Integer,1
to enable wandb,0
to disable it. Default:0
= Disabledsave_step
: Integer, amount of steps to save a checkpoint. Default: None
You can set these options like this: python3 run.py --use_wandb=1
Once training finishes, the model (and checkpoints) will be available under the models
folder in the same path you ran the trainer on.
To load them, simply run the following on your generation script:
model.lm.load_state_dict(torch.load('models/lm_final.pt'))
Where model
is the MusicGen Object and models/lm_final.pt
is the path to your model (or checkpoint).
Encodec seems to struggle with electronic music. Even just Encoding->Decoding has many problems.
4:00 - 4:30 - Moe Shop - WONDER POP
Original: https://voca.ro/1jbsor6BAyLY
Encode -> Decode: https://voca.ro/1kF2yyGyRn0y
Overfit -> Generate -> Decode: https://voca.ro/1f6ru5ieejJY
Softer and less aggressive melodies seem to play best with encodec and musicgen. One of these are bossa nova, which to me sounds great:
1:20 - 1:50 - Tom Jobim - Children's Games
Original: https://voca.ro/1dm9QpRqa5rj (last 5 seconds are ignored)
Encode -> Decode: https://voca.ro/19LpwVE44si7
Overfit -> Generate -> Decode: https://voca.ro/1hJGVdxsvBOG
@article{copet2023simple,
title={Simple and Controllable Music Generation},
author={Jade Copet and Felix Kreuk and Itai Gat and Tal Remez and David Kant and Gabriel Synnaeve and Yossi Adi and Alexandre Défossez},
year={2023},
journal={arXiv preprint arXiv:2306.05284},
}
Special thanks to elyxlz (223864514326560768@discord) for helping me with the masks.