Sri2901 commited on
Commit
bb7b452
·
verified ·
1 Parent(s): 4bf7473

Upload model via CLI

Browse files
.job_config.json ADDED
@@ -0,0 +1,112 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "job": "extension",
3
+ "config": {
4
+ "name": "my_first_lora_v1",
5
+ "process": [
6
+ {
7
+ "type": "ui_trainer",
8
+ "training_folder": "/workspace/ai-toolkit/output",
9
+ "sqlite_db_path": "/workspace/ai-toolkit/aitk_db.db",
10
+ "device": "cuda",
11
+ "trigger_word": "studi@",
12
+ "performance_log_every": 10,
13
+ "network": {
14
+ "type": "lora",
15
+ "linear": 32,
16
+ "linear_alpha": 32,
17
+ "conv": 16,
18
+ "conv_alpha": 16,
19
+ "lokr_full_rank": true,
20
+ "lokr_factor": -1,
21
+ "network_kwargs": {
22
+ "ignore_if_contains": []
23
+ }
24
+ },
25
+ "save": {
26
+ "dtype": "bf16",
27
+ "save_every": 250,
28
+ "max_step_saves_to_keep": 6,
29
+ "save_format": "diffusers",
30
+ "push_to_hub": false
31
+ },
32
+ "datasets": [
33
+ {
34
+ "folder_path": "/workspace/ai-toolkit/datasets/studio_women",
35
+ "control_path": null,
36
+ "mask_path": null,
37
+ "mask_min_value": 0.1,
38
+ "default_caption": "",
39
+ "caption_ext": "txt",
40
+ "caption_dropout_rate": 0.1,
41
+ "cache_latents_to_disk": false,
42
+ "is_reg": false,
43
+ "network_weight": 1,
44
+ "resolution": [
45
+ 512,
46
+ 768,
47
+ 1024,
48
+ 1280
49
+ ],
50
+ "controls": []
51
+ }
52
+ ],
53
+ "train": {
54
+ "batch_size": 1,
55
+ "bypass_guidance_embedding": false,
56
+ "steps": 1500,
57
+ "gradient_accumulation": 1,
58
+ "train_unet": true,
59
+ "train_text_encoder": false,
60
+ "gradient_checkpointing": true,
61
+ "noise_scheduler": "flowmatch",
62
+ "optimizer": "adamw8bit",
63
+ "timestep_type": "sigmoid",
64
+ "content_or_style": "balanced",
65
+ "optimizer_params": {
66
+ "weight_decay": 0.0001
67
+ },
68
+ "unload_text_encoder": false,
69
+ "lr": 0.0004,
70
+ "ema_config": {
71
+ "use_ema": true,
72
+ "ema_decay": 0.99
73
+ },
74
+ "dtype": "bf16",
75
+ "diff_output_preservation": false,
76
+ "diff_output_preservation_multiplier": 1,
77
+ "diff_output_preservation_class": "person"
78
+ },
79
+ "model": {
80
+ "name_or_path": "black-forest-labs/FLUX.1-dev",
81
+ "quantize": false,
82
+ "quantize_te": true,
83
+ "arch": "flux",
84
+ "low_vram": false,
85
+ "model_kwargs": {}
86
+ },
87
+ "sample": {
88
+ "sampler": "flowmatch",
89
+ "sample_every": 250,
90
+ "width": 1024,
91
+ "height": 1024,
92
+ "prompts": [
93
+ " studi@ young women wearing a black leather jacket and jeans, studio bg",
94
+ "studi@ young women sitting at a table, natural lighting, candid portrait, studio bg",
95
+ "close-up portrait of young women, studi@ , soft natural lighting, looking at camera, beige studio bg"
96
+ ],
97
+ "neg": "",
98
+ "seed": 42,
99
+ "walk_seed": true,
100
+ "guidance_scale": 3,
101
+ "sample_steps": 25,
102
+ "num_frames": 1,
103
+ "fps": 1
104
+ }
105
+ }
106
+ ]
107
+ },
108
+ "meta": {
109
+ "name": "[name]",
110
+ "version": "1.0"
111
+ }
112
+ }
README.md ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - text-to-image
4
+ - flux
5
+ - lora
6
+ - diffusers
7
+ - template:sd-lora
8
+ - ai-toolkit
9
+ base_model: black-forest-labs/FLUX.1-dev
10
+ instance_prompt: studi@
11
+ license: other
12
+ license_name: flux-1-dev-non-commercial-license
13
+ license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md
14
+ widget:
15
+ - text: ' studi@ young women wearing a black leather jacket and jeans, studio bg'
16
+ output:
17
+ url: samples/1753786454930__000001500_2.jpg
18
+ - text: studi@ young women sitting at a table, natural lighting, candid portrait,
19
+ studio bg
20
+ output:
21
+ url: samples/1753786440442__000001500_1.jpg
22
+ - text: close-up portrait of young women, studi@ , soft natural lighting, looking
23
+ at camera, beige studio bg
24
+ output:
25
+ url: samples/1753786426009__000001500_0.jpg
26
+ ---
27
+
28
+ # my_first_lora_v1
29
+
30
+ Model trained with AI Toolkit by Ostris
31
+
32
+ <Gallery />
33
+
34
+ ## Trigger words
35
+
36
+ You should use `studi@` to trigger the image generation.
37
+
38
+ ## Download model and use it with ComfyUI, AUTOMATIC1111, SD.Next, Invoke AI, etc.
39
+
40
+ Weights for this model are available in Safetensors format.
41
+
42
+ [Download](/username/my_first_lora_v1/tree/main) them in the Files & versions tab.
43
+
44
+ ## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers)
45
+
46
+ ```py
47
+ from diffusers import AutoPipelineForText2Image
48
+ import torch
49
+
50
+ pipeline = AutoPipelineForText2Image.from_pretrained('black-forest-labs/FLUX.1-dev', torch_dtype=torch.bfloat16).to('cuda')
51
+ pipeline.load_lora_weights('username/my_first_lora_v1', weight_name='my_first_lora_v1_000001250.safetensors')
52
+ image = pipeline(' studi@ young women wearing a black leather jacket and jeans, studio bg').images[0]
53
+ image.save("my_image.png")
54
+ ```
55
+
56
+ For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters)
config.yaml ADDED
@@ -0,0 +1,94 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ job: extension
2
+ config:
3
+ name: my_first_lora_v1
4
+ process:
5
+ - type: ui_trainer
6
+ training_folder: /workspace/ai-toolkit/output
7
+ sqlite_db_path: /workspace/ai-toolkit/aitk_db.db
8
+ device: cuda
9
+ trigger_word: studi@
10
+ performance_log_every: 10
11
+ network:
12
+ type: lora
13
+ linear: 32
14
+ linear_alpha: 32
15
+ conv: 16
16
+ conv_alpha: 16
17
+ lokr_full_rank: true
18
+ lokr_factor: -1
19
+ network_kwargs:
20
+ ignore_if_contains: []
21
+ save:
22
+ dtype: bf16
23
+ save_every: 250
24
+ max_step_saves_to_keep: 6
25
+ save_format: diffusers
26
+ push_to_hub: false
27
+ datasets:
28
+ - folder_path: /workspace/ai-toolkit/datasets/studio_women
29
+ control_path: null
30
+ mask_path: null
31
+ mask_min_value: 0.1
32
+ default_caption: ''
33
+ caption_ext: txt
34
+ caption_dropout_rate: 0.1
35
+ cache_latents_to_disk: false
36
+ is_reg: false
37
+ network_weight: 1
38
+ resolution:
39
+ - 512
40
+ - 768
41
+ - 1024
42
+ - 1280
43
+ controls: []
44
+ train:
45
+ batch_size: 1
46
+ bypass_guidance_embedding: false
47
+ steps: 1500
48
+ gradient_accumulation: 1
49
+ train_unet: true
50
+ train_text_encoder: false
51
+ gradient_checkpointing: true
52
+ noise_scheduler: flowmatch
53
+ optimizer: adamw8bit
54
+ timestep_type: sigmoid
55
+ content_or_style: balanced
56
+ optimizer_params:
57
+ weight_decay: 0.0001
58
+ unload_text_encoder: false
59
+ lr: 0.0004
60
+ ema_config:
61
+ use_ema: true
62
+ ema_decay: 0.99
63
+ dtype: bf16
64
+ diff_output_preservation: false
65
+ diff_output_preservation_multiplier: 1
66
+ diff_output_preservation_class: person
67
+ model:
68
+ name_or_path: black-forest-labs/FLUX.1-dev
69
+ quantize: false
70
+ quantize_te: true
71
+ arch: flux
72
+ low_vram: false
73
+ model_kwargs: {}
74
+ sample:
75
+ sampler: flowmatch
76
+ sample_every: 250
77
+ width: 1024
78
+ height: 1024
79
+ prompts:
80
+ - ' studi@ young women wearing a black leather jacket and jeans, studio bg'
81
+ - studi@ young women sitting at a table, natural lighting, candid portrait,
82
+ studio bg
83
+ - close-up portrait of young women, studi@ , soft natural lighting, looking
84
+ at camera, beige studio bg
85
+ neg: ''
86
+ seed: 42
87
+ walk_seed: true
88
+ guidance_scale: 3
89
+ sample_steps: 25
90
+ num_frames: 1
91
+ fps: 1
92
+ meta:
93
+ name: my_first_lora_v1
94
+ version: '1.0'
log.txt ADDED
The diff for this file is too large to render. See raw diff
 
logs/0_log.txt ADDED
The diff for this file is too large to render. See raw diff
 
logs/1_log.txt ADDED
The diff for this file is too large to render. See raw diff
 
my_first_lora_v1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc64ef9c323a88df7eb1bb5edf566aaffb1c100be440d35a8e28c499f7a043e6
3
+ size 343806440
my_first_lora_v1_000000250.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6aa6c785790f2be988b626308f35f91f561776c9b54004c4d092e5b6e72c01a
3
+ size 343806440
my_first_lora_v1_000000500.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b47f9428815b69b0e8949566d465064dd08528dd8614d69ceff3b8df221675af
3
+ size 343806440
my_first_lora_v1_000000750.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50d6b8753405120e24a066c38322ef444c580c7c03310febfba0cfdfb05b2175
3
+ size 343806440
my_first_lora_v1_000001250.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:722cb94b8229c33a3aee287f027155ba47c5cb516f0bb81fbd20dfda0acec5bd
3
+ size 343806440
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe49c32c75459b34d0215598be1c10da3eb770c8369a4b45a56578da2c779596
3
+ size 350203012
samples/1753781646163__000000000_0.jpg ADDED
samples/1753781660626__000000000_1.jpg ADDED
samples/1753781675027__000000000_2.jpg ADDED
samples/1753782377494__000000250_0.jpg ADDED
samples/1753782391922__000000250_1.jpg ADDED
samples/1753782406424__000000250_2.jpg ADDED
samples/1753783279638__000000500_0.jpg ADDED
samples/1753783294043__000000500_1.jpg ADDED
samples/1753783308784__000000500_2.jpg ADDED
samples/1753784008223__000000750_0.jpg ADDED
samples/1753784022683__000000750_1.jpg ADDED
samples/1753784037134__000000750_2.jpg ADDED
samples/1753784732167__000001000_0.jpg ADDED
samples/1753784746559__000001000_1.jpg ADDED
samples/1753784760991__000001000_2.jpg ADDED
samples/1753785696767__000001250_0.jpg ADDED
samples/1753785711182__000001250_1.jpg ADDED
samples/1753785725596__000001250_2.jpg ADDED
samples/1753786426009__000001500_0.jpg ADDED
samples/1753786440442__000001500_1.jpg ADDED
samples/1753786454930__000001500_2.jpg ADDED