LoRA_Emoji / train_state.json
mcuo's picture
Upload 5 files
ade06db verified
{
"step": 3000,
"base_model_path": "E:\\Emoji-TTS\\checkpoints\\Aratako_Irodori-TTS-500M-v2\\model.safetensors",
"base_model_sha256": "e5add885303babe328eae2c426475e41bc2f2aca2bcb010a2ceb6b6ddd1b8d9c",
"base_model_config": {
"latent_dim": 32,
"latent_patch_size": 1,
"model_dim": 1280,
"num_layers": 12,
"num_heads": 20,
"mlp_ratio": 2.875,
"text_mlp_ratio": 2.6,
"speaker_mlp_ratio": 2.6,
"dropout": 0.0,
"text_vocab_size": 99574,
"text_tokenizer_repo": "llm-jp/llm-jp-3-150m",
"text_add_bos": true,
"text_dim": 512,
"text_layers": 10,
"text_heads": 8,
"speaker_dim": 768,
"speaker_layers": 8,
"speaker_heads": 12,
"speaker_patch_size": 1,
"timestep_embed_dim": 512,
"adaln_rank": 192,
"norm_eps": 1e-05
},
"lora_config": {
"r": 16,
"lora_alpha": 32.0,
"lora_dropout": 0.05,
"target_modules": [
"wq",
"wk",
"wv",
"wo"
]
},
"train_config": {
"manifest_path": "E:\\Emoji-TTS\\data\\train_manifest.jsonl",
"output_dir": "E:\\Emoji-TTS\\lora\\my_lora",
"batch_size": 4,
"num_workers": 2,
"dataloader_persistent_workers": false,
"dataloader_prefetch_factor": 2,
"allow_tf32": false,
"compile_model": false,
"learning_rate": 0.0001,
"weight_decay": 0.01,
"optimizer": "adamw",
"adam_beta1": 0.9,
"adam_beta2": 0.999,
"adam_eps": 1e-08,
"muon_momentum": 0.95,
"muon_adjust_lr_fn": "match_rms_adamw",
"lr_scheduler": "none",
"warmup_steps": 0,
"caption_warmup": false,
"caption_warmup_steps": 0,
"stable_steps": 0,
"min_lr_scale": 0.1,
"max_steps": 3000,
"log_every": 10,
"save_every": 500,
"checkpoint_best_n": 0,
"valid_ratio": 0.1,
"valid_every": 100,
"progress": true,
"progress_all_ranks": false,
"precision": "bf16",
"grad_clip_norm": 1.0,
"gradient_accumulation_steps": 1,
"max_text_len": 256,
"max_caption_len": null,
"text_condition_dropout": 0.1,
"caption_condition_dropout": 0.1,
"speaker_condition_dropout": 0.1,
"max_latent_steps": 750,
"fixed_target_latent_steps": null,
"fixed_target_full_mask": false,
"timestep_logit_mean": 0.0,
"timestep_logit_std": 1.0,
"timestep_stratified": false,
"timestep_min": 0.001,
"timestep_max": 0.999,
"wandb_enabled": false,
"wandb_project": null,
"wandb_entity": null,
"wandb_run_name": null,
"wandb_mode": "online",
"ddp_find_unused_parameters": false,
"lora_enabled": false,
"lora_r": 16,
"lora_alpha": 32,
"lora_dropout": 0.0,
"lora_bias": "none",
"lora_target_modules": "diffusion_attn",
"seed": 0
},
"ema_decay": 0.9999
}