| | name: final |
| |
|
| | seed: 998244353 |
| | device_specific_seed: true |
| | workder_specific_seed: true |
| |
|
| |
|
| | data: |
| | data_path: train_data/chenyuan/final.yml |
| | image_size: 512 |
| | apply_chat_template: true |
| | max_token_len: 4000 |
| | img_scale_num: 16 |
| | use_text_loss: false |
| | interleave: true |
| | use_input_image_for_vlm: true |
| | |
| | |
| | model: |
| | |
| | pretrained_vae_model_name_or_path: /share_2/luoxin/modelscope/hub/models/FLUX.1-dev |
| | pretrained_text_encoder_model_name_or_path: /share/shitao/models/Qwen2.5-VL-3B-Instruct |
| | pretrained_diffusion_model_name_or_path: /share_2/luoxin/projects/Ominigenv2/experiments/0605_mix5/checkpoint-1000/pytorch_model_fsdp.bin |
| | use_vae_input: false |
| | use_all_condition: true |
| | use_image_tokens: false |
| | use_hw_tokens: false |
| | use_only_text_token_from_mllm: true |
| | use_only_prev_text_token_from_mllm: false |
| |
|
| | arch_type: Lumina2Transformer_separate_ref_process |
| | arch_opt: |
| | patch_size: 2 |
| | in_channels: 16 |
| | hidden_size: 2520 |
| | num_layers: 32 |
| | num_refiner_layers: 2 |
| | num_attention_heads: 21 |
| | num_kv_heads: 7 |
| | multiple_of: 256 |
| | norm_eps: !!float 1e-05 |
| | axes_dim_rope: [40, 40, 40] |
| | axes_lens: [12000, 12000, 12000] |
| | text_feat_dim: 2048 |
| | timestep_scale: !!float 1000 |
| | position_id_version: ominicontrol |
| | separate_ref_image_patch_embed: true |
| | separate_ref_image_refiner: true |
| | use_fused_rms_norm: true |
| | use_fused_swiglu: true |
| | proj_feat_depth: 8 |
| | projector_dim: 2048 |
| | proj_feat_dim: 1024 |
| |
|
| | transport: |
| | snr_type: lognorm |
| | do_shift: true |
| | dynamic_time_shift: true |
| |
|
| | train: |
| | |
| | global_batch_size: 128 |
| | batch_size: 2 |
| | gradient_accumulation_steps: 1 |
| |
|
| | num_train_epochs: 5 |
| | |
| | dataloader_num_workers: 6 |
| |
|
| | |
| | learning_rate: !!float 1e-6 |
| | mllm_lr: !!float 3e-7 |
| | mllm_projector_lr: !!float 1e-7 |
| | mllm_vision_lr: !!float 1e-7 |
| | mllm_image_token_lr: !!float 1e-4 |
| |
|
| | scale_lr: false |
| | lr_scheduler: timm_cosine |
| | t_initial: 200000 |
| | lr_min: 0 |
| | cycle_decay: 0.5 |
| | warmup_t: 500 |
| | warmup_lr_init: 0 |
| | warmup_prefix: true |
| | t_in_epochs: false |
| |
|
| | |
| |
|
| | use_8bit_adam: false |
| | adam_beta1: 0.9 |
| | adam_beta2: 0.95 |
| | adam_weight_decay: !!float 0.01 |
| | adam_epsilon: !!float 1e-08 |
| | max_grad_norm: 1 |
| |
|
| | gradient_checkpointing: true |
| | |
| | set_grads_to_none: true |
| |
|
| | |
| | allow_tf32: false |
| | mixed_precision: 'bf16' |
| |
|
| | ema_decay: 0.0 |
| |
|
| | repa_loss_weight: !!float 0.0 |
| | repa_encoder: dinov2_vitl14_reg |
| | repa_batch_infer: true |
| |
|
| | text_loss_weight: !!float 1.0 |
| |
|
| | train_diffusion: true |
| | train_mllm: true |
| | train_diffusion_context_refiner: true |
| | |
| | val: |
| | validation_steps: 200 |
| | train_visualization_steps: 200 |
| | |
| | logger: |
| | log_with: [wandb, tensorboard] |
| | |
| |
|
| | checkpointing_steps: 1000 |
| | checkpoints_total_limit: ~ |
| |
|
| | cache_dir: |
| | resume_from_checkpoint: latest |
| | previous_stage_ckpt: |