william590y's picture
Upload folder using huggingface_hub
151b875 verified
# Human Evaluation for the Anticipatory Music Transformer
## Generating clips for the qualification round
First we select five clips with melodic content from the Lakh MIDI test set: split `f`. Selected clips are stored to the `qualify` directory.
```
python melody-select.py $DATAPATH/lmd_full/f/ -o qualify -c 5 -s 1 -v
```
Then we generate accompaniments to these clips. We specify the reference midis (`-d` option) for the retrieval baseline.
```
python accompany.py qualify -r -d $DATAPATH/lmd_full/f/
```
## Generating clips for the prompted completion round
We generate prompted completions using an autoregressive model (or an anticipatory autoregressive model) checkpoint stored at $MODELPATH.
First, we randomly select 50 prompts and completions from a collection of completions generated using the FIGARO Music Transformer (stored at $FIGARO). Store these prompts at $PROMPTPATH:
```
python figaro-select.py $FIGARO -o $PROMPTPATH -c 50 -s 999 -v
```
Generate completions using a model stored at $MODELPATH and store the results to $PROMPTPATH/$OUTPUT:
```
python prompt.py $PROMPTPATH $MODELPATH -o $OUTPUT -c 50 -v
```
Generate completions using an interarrival-time model:
```
python prompt-interarrival.py $PROMPTPATH $MODELPATH $OUTPUT -c 50 -v
```
## Generating clips for the accompaniment round
We generate accompaniments using an anticipatory autoregressive model checkpoint stored at $MODELPATH.
First, select 50 clips with melodic content:
```
python melody-select.py $DATAPATH/lmd_full/f/ -o accompany -c 50 -v
```
Generate anticipatory accompaniments (`-a` flag):
```
python accompany.py accompany --model $MODELPATH -av -c 50
```
Generate the autoregressive baseline (`-b` flag):
```
python accompany.py accompany --model $MODELPATH -bv -c 50
```
Generate the retrieval baseline (`-r` flag):
```
python accompany.py accompany -d $DATAPATH/lmd_full/f/ -rv -c 50
```