MLM-base / README.md
ilya16's picture
Create README.md
b97befc verified
metadata
license: cc-by-nc-sa-4.0
datasets:
  - SyMuPe/PERiScoPe
tags:
  - music
  - piano
  - midi
  - expressive-performance
  - transformer
  - MLM

SyMuPe: MLM baseline

MLM-base is a Transformer-based masked language modeling baseline for expressive piano performance rendering.

Introduced in the paper: SyMuPe: Affective and Controllable Symbolic Music Performance.

Architecture

  • Type: Transformer Encoder
  • Objective: Masked Performance Modeling (MLM)
  • Inputs:
    • Score features (y): Pitch, Position, PositionShift, Duration
    • Performance features (x): Velocity, TimeShift, TimeDuration, TimeDurationSustain
    • Conditioning (c_s): Velocity and Tempo score tokens for tempo and dynamics.
  • Outputs: Categorical distributions for unmasked performance tokens.
  • Training: Trained for 300,000 iterations on the PERiScoPe v1.0 dataset as described in the paper.

Quick Start

To use this model, ensure you have the symupe library installed (refer to the GitHub repo for installation instructions).

import torch
from symusic import Score

from symupe.data.tokenizers import SyMuPe
from symupe.inference import AutoGenerator, perform_score, save_performances
from symupe.models import AutoModel

device = torch.device("cuda" if torch.cuda.is_available() else "cpu")

# Load the model and tokenizer directly from the Hub
model = AutoModel.from_pretrained("SyMuPe/MLM-base").to(device)
tokenizer = SyMuPe.from_pretrained("SyMuPe/MLM-base")

# Prepare generator for the model
generator = AutoGenerator.from_model(model, tokenizer, device=device)

# Load score MIDI
score_midi = Score("score.mid")

# Perform score MIDI (tokenization is handled inside)
gen_results = perform_score(
    generator=generator,
    score=score_midi,
    use_score_context=True,
    num_samples=8,
    seed=23
)
# gen_results[i] is PerformanceRenderingResult(...) containing:
# - score_midi, score_seq, gen_seq, perf_seq, perf_midi, perf_midi_sus

# Save performed MIDI files in a single directory
save_performances(gen_results, out_dir="samples/mlm", save_midi=True)

License

The model weights are distributed under the CC-BY-NC-SA 4.0 license.

Citation

If you use the dataset, please cite the paper:

@inproceedings{borovik2025symupe,
  title = {{SyMuPe: Affective and Controllable Symbolic Music Performance}},
  author = {Borovik, Ilya and Gavrilev, Dmitrii and Viro, Vladimir},
  year = {2025},
  booktitle = {Proceedings of the 33rd ACM International Conference on Multimedia},
  pages = {10699--10708},
  doi = {10.1145/3746027.3755871}
}