william590y's picture
Upload folder using huggingface_hub
151b875 verified

Human Evaluation for the Anticipatory Music Transformer

Generating clips for the qualification round

First we select five clips with melodic content from the Lakh MIDI test set: split f. Selected clips are stored to the qualify directory.

python melody-select.py $DATAPATH/lmd_full/f/ -o qualify -c 5 -s 1 -v

Then we generate accompaniments to these clips. We specify the reference midis (-d option) for the retrieval baseline.

python accompany.py qualify -r -d $DATAPATH/lmd_full/f/

Generating clips for the prompted completion round

We generate prompted completions using an autoregressive model (or an anticipatory autoregressive model) checkpoint stored at $MODELPATH.

First, we randomly select 50 prompts and completions from a collection of completions generated using the FIGARO Music Transformer (stored at $FIGARO). Store these prompts at $PROMPTPATH:

python figaro-select.py $FIGARO -o $PROMPTPATH -c 50 -s 999 -v

Generate completions using a model stored at $MODELPATH and store the results to $PROMPTPATH/$OUTPUT:

python prompt.py $PROMPTPATH $MODELPATH -o $OUTPUT -c 50 -v

Generate completions using an interarrival-time model:

python prompt-interarrival.py $PROMPTPATH $MODELPATH $OUTPUT -c 50 -v

Generating clips for the accompaniment round

We generate accompaniments using an anticipatory autoregressive model checkpoint stored at $MODELPATH.

First, select 50 clips with melodic content:

python melody-select.py $DATAPATH/lmd_full/f/ -o accompany -c 50 -v

Generate anticipatory accompaniments (-a flag):

python accompany.py accompany --model $MODELPATH -av -c 50

Generate the autoregressive baseline (-b flag):

python accompany.py accompany --model $MODELPATH -bv -c 50

Generate the retrieval baseline (-r flag):

python accompany.py accompany -d $DATAPATH/lmd_full/f/ -rv -c 50