ThatDustyGuy commited on
Commit
280efaa
·
1 Parent(s): 8f1b284

Upload folder using huggingface_hub

Browse files
step5500/adapter_config.json ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": null,
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": false,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 128,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.0,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 128,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "to_k",
29
+ "to_v",
30
+ "to_add_out",
31
+ "add_k_proj",
32
+ "to_out.0",
33
+ "add_v_proj",
34
+ "proj",
35
+ "net.2",
36
+ "txt_mod.1",
37
+ "add_q_proj",
38
+ "img_mod.1",
39
+ "to_q"
40
+ ],
41
+ "target_parameters": null,
42
+ "task_type": null,
43
+ "trainable_token_indices": null,
44
+ "use_dora": false,
45
+ "use_qalora": false,
46
+ "use_rslora": false
47
+ }
step5500/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1508bc39cecdd9aa5816e4c4d21a143e2c2c1ece330749eafc9f92b81eefd8e8
3
+ size 2359534496
step5500/qwen_lora_config.toml ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Main training settings
2
+ output_dir = '/workspace/qwen-lora-output'
3
+ dataset = '/workspace/qwen_dataset.toml'
4
+ epochs = 10000
5
+ max_steps = 6000
6
+ micro_batch_size_per_gpu = 1
7
+ pipeline_stages = 1
8
+ gradient_accumulation_steps = 1
9
+ gradient_clipping = 1.0
10
+ warmup_steps = 200
11
+ save_every_n_steps = 250
12
+ checkpoint_every_n_minutes = 120
13
+ activation_checkpointing = true
14
+
15
+ eval_every_n_steps = 500
16
+ eval_before_first_step = true
17
+ eval_micro_batch_size_per_gpu = 1
18
+ eval_gradient_accumulation_steps = 1
19
+
20
+
21
+ [model]
22
+ type = 'qwen_image'
23
+ # This path is used to find the tokenizer config files.
24
+ diffusers_path = '/workspace/models/Qwen-Image'
25
+
26
+ # Explicitly define paths to each core model component using the correct key names.
27
+ transformer_path = '/workspace/models/qwen_transformer/split_files/diffusion_models/qwen_image_bf16.safetensors'
28
+ text_encoder_path = '/workspace/models/qwen_text_encoder/split_files/text_encoders/qwen_2.5_vl_7b.safetensors'
29
+ vae_path = '/workspace/models/Qwen-Image/vae/diffusion_pytorch_model.safetensors'
30
+
31
+ # BF16 is great for L40S, and fp8 for the transformer saves VRAM with minimal quality loss.
32
+ dtype = 'bfloat16'
33
+ transformer_dtype = 'bfloat16'
34
+ timestep_sample_method = 'logit_normal'
35
+
36
+ [adapter]
37
+ type = 'lora'
38
+ rank = 128
39
+ dtype = 'bfloat16'
40
+
41
+ [optimizer]
42
+ type = 'adamw_optimi'
43
+ lr = 5e-5
44
+ betas = [0.9, 0.99]
45
+ weight_decay = 0.01
46
+ eps = 1e-8
47
+
48
+ [monitoring]
49
+ enable_wandb = true
50
+ wandb_api_key = 'cc3f7d9c61eecb958c5b52f48087ecef56cc32b4' # Paste your key here
51
+ wandb_tracker_name = 'Qwen-Image-Dlay_LoRA'
52
+ wandb_run_name = 'L40S-Run-1'