| AdamW.lr: 0.0001 | |
| AdamW.weight_decay: 0.0 | |
| DecoderTransformer.depth: 8 | |
| DecoderTransformer.dim: 512 | |
| DecoderTransformer.dropout: 0.1 | |
| DecoderTransformer.heads: 8 | |
| DecoderTransformer.max_seq_len: 514 | |
| accelerator: auto | |
| args.debug: 0 | |
| args.load: configs/gen_models/decoder_only_online_chord.yml | |
| args.save: null | |
| args.unknown: [] | |
| batch_size: 64 | |
| cache_dir: data/cache | |
| checkpoint_interval: 1000 | |
| checkpoint_metric: val/loss | |
| checkpoint_mode: min | |
| checkpoint_top_k: -1 | |
| compile: true | |
| devices: auto | |
| gradient_clip_val: 1.0 | |
| log_every_n_steps: 1 | |
| max_len: 512 | |
| model_part: chord | |
| model_type: decoder_only | |
| num_nodes: 1 | |
| num_workers: 8 | |
| overfit_batches: 0 | |
| precision: bf16-mixed | |
| sample_interval: 5000 | |
| save_dir: logs/decoder_only_online_chord | |
| seed: 42 | |
| strategy: auto | |
| train_steps: 30000 | |
| use_fabric: false | |
| val_interval: 1000 | |
| wandb_project: realchords | |