Training

#2
by Politrees - opened

Hi!
I didn't find the Issue tab in your GitHub repository, so I'll ask here.

Is there a guide on how to start a workout, how exactly to train, and which dataset is needed? I would also like to try to start a workout.

You will need to prepare a large training dataset consisting of:

  • WAV files containing audio data
  • PV files containing MIDI at 10 ms, which corresponds to a hop length of 160

You can use the MIR-1K dataset provided here:
https://huggingface.co/datasets/AnhP/Mir-1k-use-DJCM-training/resolve/main/dataset-10ms.zip

After downloading and extracting the dataset:
Replace the dataset path in the code from "hybrid" to your own dataset path.

Run train.py to start training.
If you want to experiment, you can modify the default training parameters to suit your needs.

image

AnhP changed discussion status to closed
AnhP changed discussion status to open

Thanks! Is there a benchmark code that you used to create your comparison tables?

Thanks! Is there a benchmark code that you used to create your comparison tables?
https://github.com/lars76/pitch-benchmark

If it's not a secret, did you train on this same dataset or a different one, and approximately how many hours of data are there?
No matter how much I try, my results are far from what you achieved. I even added 2 hours of my own data, and it got better, but it's still far off.

If it's not a secret, did you train on this same dataset or a different one, and approximately how many hours of data are there?
No matter how much I try, my results are far from what you achieved. I even added 2 hours of my own data, and it got better, but it's still far off.

If you want, you can try a mix of
Mir-1K + PTDB TUG + Vocadito
or
Mir-1K + PTDB TUG + 3,903 files from M4Singer, with half extracted using PM and the other half using RMVPE, like I did.

Note: you may not achieve the expected results, as I continuously adjusted the parameters during training.

Sign up or log in to comment