tommycik commited on
Commit
b0b4ea7
·
verified ·
1 Parent(s): 3d65e14

End of training

Browse files
Files changed (46) hide show
  1. .gitattributes +1 -0
  2. README.md +47 -0
  3. checkpoint-1000/flux_controlnet/config.json +36 -0
  4. checkpoint-1000/flux_controlnet/diffusion_pytorch_model.safetensors +3 -0
  5. checkpoint-1000/optimizer.bin +3 -0
  6. checkpoint-1000/random_states_0.pkl +3 -0
  7. checkpoint-1000/scheduler.bin +3 -0
  8. checkpoint-1250/flux_controlnet/config.json +36 -0
  9. checkpoint-1250/flux_controlnet/diffusion_pytorch_model.safetensors +3 -0
  10. checkpoint-1250/optimizer.bin +3 -0
  11. checkpoint-1250/random_states_0.pkl +3 -0
  12. checkpoint-1250/scheduler.bin +3 -0
  13. checkpoint-1500/flux_controlnet/config.json +36 -0
  14. checkpoint-1500/flux_controlnet/diffusion_pytorch_model.safetensors +3 -0
  15. checkpoint-1500/optimizer.bin +3 -0
  16. checkpoint-1500/random_states_0.pkl +3 -0
  17. checkpoint-1500/scheduler.bin +3 -0
  18. checkpoint-1750/flux_controlnet/config.json +36 -0
  19. checkpoint-1750/flux_controlnet/diffusion_pytorch_model.safetensors +3 -0
  20. checkpoint-1750/optimizer.bin +3 -0
  21. checkpoint-1750/random_states_0.pkl +3 -0
  22. checkpoint-1750/scheduler.bin +3 -0
  23. checkpoint-2000/flux_controlnet/config.json +36 -0
  24. checkpoint-2000/flux_controlnet/diffusion_pytorch_model.safetensors +3 -0
  25. checkpoint-2000/optimizer.bin +3 -0
  26. checkpoint-2000/random_states_0.pkl +3 -0
  27. checkpoint-2000/scheduler.bin +3 -0
  28. checkpoint-250/flux_controlnet/config.json +36 -0
  29. checkpoint-250/flux_controlnet/diffusion_pytorch_model.safetensors +3 -0
  30. checkpoint-250/optimizer.bin +3 -0
  31. checkpoint-250/random_states_0.pkl +3 -0
  32. checkpoint-250/scheduler.bin +3 -0
  33. checkpoint-500/flux_controlnet/config.json +36 -0
  34. checkpoint-500/flux_controlnet/diffusion_pytorch_model.safetensors +3 -0
  35. checkpoint-500/optimizer.bin +3 -0
  36. checkpoint-500/random_states_0.pkl +3 -0
  37. checkpoint-500/scheduler.bin +3 -0
  38. checkpoint-750/flux_controlnet/config.json +36 -0
  39. checkpoint-750/flux_controlnet/diffusion_pytorch_model.safetensors +3 -0
  40. checkpoint-750/optimizer.bin +3 -0
  41. checkpoint-750/random_states_0.pkl +3 -0
  42. checkpoint-750/scheduler.bin +3 -0
  43. config.json +36 -0
  44. diffusion_pytorch_model.safetensors +3 -0
  45. image_control.png +0 -0
  46. images_0.png +3 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ images_0.png filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: black-forest-labs/FLUX.1-dev
3
+ library_name: diffusers
4
+ license: other
5
+ inference: true
6
+ tags:
7
+ - flux
8
+ - flux-diffusers
9
+ - text-to-image
10
+ - diffusers
11
+ - controlnet
12
+ - diffusers-training
13
+ ---
14
+
15
+ <!-- This model card has been generated automatically according to the information the training script had access to. You
16
+ should probably proofread and complete it, then remove this comment. -->
17
+
18
+
19
+ # controlnet-tommycik/ControlNetCannyReducedImproved
20
+
21
+ These are controlnet weights trained on black-forest-labs/FLUX.1-dev with new type of conditioning.
22
+ You can find some example images below.
23
+
24
+ prompt: transparent cocktail galss with elegant stem and a double curved bowl on a white background
25
+ ![images_0)](./images_0.png)
26
+
27
+
28
+ ## License
29
+
30
+ Please adhere to the licensing terms as described [here](https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md)
31
+
32
+
33
+ ## Intended uses & limitations
34
+
35
+ #### How to use
36
+
37
+ ```python
38
+ # TODO: add an example code snippet for running this diffusion pipeline
39
+ ```
40
+
41
+ #### Limitations and bias
42
+
43
+ [TODO: provide examples of latent issues and potential remediations]
44
+
45
+ ## Training details
46
+
47
+ [TODO: describe the data used to train the model]
checkpoint-1000/flux_controlnet/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "FluxControlNetModel",
3
+ "_diffusers_version": "0.36.0.dev0",
4
+ "_name_or_path": "InstantX/FLUX.1-dev-Controlnet-Canny",
5
+ "attention_head_dim": 128,
6
+ "axes_dims_rope": [
7
+ 16,
8
+ 56,
9
+ 56
10
+ ],
11
+ "conditioning_embedding_channels": null,
12
+ "guidance_embeds": true,
13
+ "in_channels": 64,
14
+ "joint_attention_dim": 4096,
15
+ "num_attention_heads": 24,
16
+ "num_layers": 5,
17
+ "num_mode": null,
18
+ "num_single_layers": 0,
19
+ "patch_size": 1,
20
+ "pooled_projection_dim": 768,
21
+ "quantization_config": {
22
+ "_load_in_4bit": true,
23
+ "_load_in_8bit": false,
24
+ "bnb_4bit_compute_dtype": "bfloat16",
25
+ "bnb_4bit_quant_storage": "uint8",
26
+ "bnb_4bit_quant_type": "fp4",
27
+ "bnb_4bit_use_double_quant": false,
28
+ "llm_int8_enable_fp32_cpu_offload": false,
29
+ "llm_int8_has_fp16_weight": false,
30
+ "llm_int8_skip_modules": null,
31
+ "llm_int8_threshold": 6.0,
32
+ "load_in_4bit": true,
33
+ "load_in_8bit": false,
34
+ "quant_method": "bitsandbytes"
35
+ }
36
+ }
checkpoint-1000/flux_controlnet/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d25b6731b7ed5816f37c5eb18a7fbb851dc9a3492eb1cfdeb791584821ce799e
3
+ size 1008570723
checkpoint-1000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62ddb4a8d8dfd4cbe1c11a0bca4a5262090944a87d0b0629a0c4133d9ef4e398
3
+ size 33159
checkpoint-1000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9768169cebb25bc45f38abedf881836984ce5d28ba7cb869db9ea8c4376f7732
3
+ size 14757
checkpoint-1000/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96fd5d03b06f8704e17c61a7ba9ba335390d542429abc5a72867fb02c1888823
3
+ size 1401
checkpoint-1250/flux_controlnet/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "FluxControlNetModel",
3
+ "_diffusers_version": "0.36.0.dev0",
4
+ "_name_or_path": "InstantX/FLUX.1-dev-Controlnet-Canny",
5
+ "attention_head_dim": 128,
6
+ "axes_dims_rope": [
7
+ 16,
8
+ 56,
9
+ 56
10
+ ],
11
+ "conditioning_embedding_channels": null,
12
+ "guidance_embeds": true,
13
+ "in_channels": 64,
14
+ "joint_attention_dim": 4096,
15
+ "num_attention_heads": 24,
16
+ "num_layers": 5,
17
+ "num_mode": null,
18
+ "num_single_layers": 0,
19
+ "patch_size": 1,
20
+ "pooled_projection_dim": 768,
21
+ "quantization_config": {
22
+ "_load_in_4bit": true,
23
+ "_load_in_8bit": false,
24
+ "bnb_4bit_compute_dtype": "bfloat16",
25
+ "bnb_4bit_quant_storage": "uint8",
26
+ "bnb_4bit_quant_type": "fp4",
27
+ "bnb_4bit_use_double_quant": false,
28
+ "llm_int8_enable_fp32_cpu_offload": false,
29
+ "llm_int8_has_fp16_weight": false,
30
+ "llm_int8_skip_modules": null,
31
+ "llm_int8_threshold": 6.0,
32
+ "load_in_4bit": true,
33
+ "load_in_8bit": false,
34
+ "quant_method": "bitsandbytes"
35
+ }
36
+ }
checkpoint-1250/flux_controlnet/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1c2aa5a7a73a9c4c9d67395d949bd1b9296703cab594e9308d899e9311d6895
3
+ size 1008570723
checkpoint-1250/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:793fcccd007a5c9a29d77de182d3c409cf89fdaad2e494720c16dcf94ce425d5
3
+ size 33159
checkpoint-1250/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ec2ddbecf67731a0da9e6078326bf352f92425d881db406a8ef5ad17ea4e913
3
+ size 14757
checkpoint-1250/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30651fb15a8970551ae553f929fc860efe4e810d3bd1e182aab843cf197c2c5f
3
+ size 1401
checkpoint-1500/flux_controlnet/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "FluxControlNetModel",
3
+ "_diffusers_version": "0.36.0.dev0",
4
+ "_name_or_path": "InstantX/FLUX.1-dev-Controlnet-Canny",
5
+ "attention_head_dim": 128,
6
+ "axes_dims_rope": [
7
+ 16,
8
+ 56,
9
+ 56
10
+ ],
11
+ "conditioning_embedding_channels": null,
12
+ "guidance_embeds": true,
13
+ "in_channels": 64,
14
+ "joint_attention_dim": 4096,
15
+ "num_attention_heads": 24,
16
+ "num_layers": 5,
17
+ "num_mode": null,
18
+ "num_single_layers": 0,
19
+ "patch_size": 1,
20
+ "pooled_projection_dim": 768,
21
+ "quantization_config": {
22
+ "_load_in_4bit": true,
23
+ "_load_in_8bit": false,
24
+ "bnb_4bit_compute_dtype": "bfloat16",
25
+ "bnb_4bit_quant_storage": "uint8",
26
+ "bnb_4bit_quant_type": "fp4",
27
+ "bnb_4bit_use_double_quant": false,
28
+ "llm_int8_enable_fp32_cpu_offload": false,
29
+ "llm_int8_has_fp16_weight": false,
30
+ "llm_int8_skip_modules": null,
31
+ "llm_int8_threshold": 6.0,
32
+ "load_in_4bit": true,
33
+ "load_in_8bit": false,
34
+ "quant_method": "bitsandbytes"
35
+ }
36
+ }
checkpoint-1500/flux_controlnet/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3934c1d5a26fc742090558346a0cf84327806f70bcabe29d631c32b6ce83a31
3
+ size 1008570723
checkpoint-1500/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d50ee435f84c801c294888897a9ad7cb23e9d303f360393bdafc5117d52fc5c
3
+ size 33159
checkpoint-1500/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17cde62129d58313313cb7c9ad89d5781b7e88af3afbf81490a51a404f8ad8dd
3
+ size 14757
checkpoint-1500/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3bab69489908fe43afc88716a6e486d0fe2f8ae562be8ea4721c9ab2af7ea2d
3
+ size 1401
checkpoint-1750/flux_controlnet/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "FluxControlNetModel",
3
+ "_diffusers_version": "0.36.0.dev0",
4
+ "_name_or_path": "InstantX/FLUX.1-dev-Controlnet-Canny",
5
+ "attention_head_dim": 128,
6
+ "axes_dims_rope": [
7
+ 16,
8
+ 56,
9
+ 56
10
+ ],
11
+ "conditioning_embedding_channels": null,
12
+ "guidance_embeds": true,
13
+ "in_channels": 64,
14
+ "joint_attention_dim": 4096,
15
+ "num_attention_heads": 24,
16
+ "num_layers": 5,
17
+ "num_mode": null,
18
+ "num_single_layers": 0,
19
+ "patch_size": 1,
20
+ "pooled_projection_dim": 768,
21
+ "quantization_config": {
22
+ "_load_in_4bit": true,
23
+ "_load_in_8bit": false,
24
+ "bnb_4bit_compute_dtype": "bfloat16",
25
+ "bnb_4bit_quant_storage": "uint8",
26
+ "bnb_4bit_quant_type": "fp4",
27
+ "bnb_4bit_use_double_quant": false,
28
+ "llm_int8_enable_fp32_cpu_offload": false,
29
+ "llm_int8_has_fp16_weight": false,
30
+ "llm_int8_skip_modules": null,
31
+ "llm_int8_threshold": 6.0,
32
+ "load_in_4bit": true,
33
+ "load_in_8bit": false,
34
+ "quant_method": "bitsandbytes"
35
+ }
36
+ }
checkpoint-1750/flux_controlnet/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e99354338eb577012e432f9f27c31ca2ebb88e67cece70505d0709a3972a9d20
3
+ size 1008570723
checkpoint-1750/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17b2c0cfccd12847de278cb921070a3e00ff442091ed292cdc5bda0317903f20
3
+ size 33159
checkpoint-1750/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff61e78c88da4aab0b7eea10335526d8f2207db9fb18d479d47e4393211d1713
3
+ size 14757
checkpoint-1750/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a647ef4dd3a8d1e0c3c5ce896bcc84db69f68821974b9cb412a8b02e0f40df0
3
+ size 1401
checkpoint-2000/flux_controlnet/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "FluxControlNetModel",
3
+ "_diffusers_version": "0.36.0.dev0",
4
+ "_name_or_path": "InstantX/FLUX.1-dev-Controlnet-Canny",
5
+ "attention_head_dim": 128,
6
+ "axes_dims_rope": [
7
+ 16,
8
+ 56,
9
+ 56
10
+ ],
11
+ "conditioning_embedding_channels": null,
12
+ "guidance_embeds": true,
13
+ "in_channels": 64,
14
+ "joint_attention_dim": 4096,
15
+ "num_attention_heads": 24,
16
+ "num_layers": 5,
17
+ "num_mode": null,
18
+ "num_single_layers": 0,
19
+ "patch_size": 1,
20
+ "pooled_projection_dim": 768,
21
+ "quantization_config": {
22
+ "_load_in_4bit": true,
23
+ "_load_in_8bit": false,
24
+ "bnb_4bit_compute_dtype": "bfloat16",
25
+ "bnb_4bit_quant_storage": "uint8",
26
+ "bnb_4bit_quant_type": "fp4",
27
+ "bnb_4bit_use_double_quant": false,
28
+ "llm_int8_enable_fp32_cpu_offload": false,
29
+ "llm_int8_has_fp16_weight": false,
30
+ "llm_int8_skip_modules": null,
31
+ "llm_int8_threshold": 6.0,
32
+ "load_in_4bit": true,
33
+ "load_in_8bit": false,
34
+ "quant_method": "bitsandbytes"
35
+ }
36
+ }
checkpoint-2000/flux_controlnet/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85a9086f139eef2e96e45ce4bc967ea6a626ab01c7dea3960616aa431e175109
3
+ size 1008570723
checkpoint-2000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1986a8e61b5529b5577c11f01d26b376b6eaf190b39eb51103f45a9d4f13d13
3
+ size 33159
checkpoint-2000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36fc8bff5a36460a19f575cad28dd05a5d3757424636c6469e4d9640f775c79a
3
+ size 14757
checkpoint-2000/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:620b1f766e932f519b62240af863a9cd84d9cdaac1d1a297a1b5614281c2b46f
3
+ size 1401
checkpoint-250/flux_controlnet/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "FluxControlNetModel",
3
+ "_diffusers_version": "0.36.0.dev0",
4
+ "_name_or_path": "InstantX/FLUX.1-dev-Controlnet-Canny",
5
+ "attention_head_dim": 128,
6
+ "axes_dims_rope": [
7
+ 16,
8
+ 56,
9
+ 56
10
+ ],
11
+ "conditioning_embedding_channels": null,
12
+ "guidance_embeds": true,
13
+ "in_channels": 64,
14
+ "joint_attention_dim": 4096,
15
+ "num_attention_heads": 24,
16
+ "num_layers": 5,
17
+ "num_mode": null,
18
+ "num_single_layers": 0,
19
+ "patch_size": 1,
20
+ "pooled_projection_dim": 768,
21
+ "quantization_config": {
22
+ "_load_in_4bit": true,
23
+ "_load_in_8bit": false,
24
+ "bnb_4bit_compute_dtype": "bfloat16",
25
+ "bnb_4bit_quant_storage": "uint8",
26
+ "bnb_4bit_quant_type": "fp4",
27
+ "bnb_4bit_use_double_quant": false,
28
+ "llm_int8_enable_fp32_cpu_offload": false,
29
+ "llm_int8_has_fp16_weight": false,
30
+ "llm_int8_skip_modules": null,
31
+ "llm_int8_threshold": 6.0,
32
+ "load_in_4bit": true,
33
+ "load_in_8bit": false,
34
+ "quant_method": "bitsandbytes"
35
+ }
36
+ }
checkpoint-250/flux_controlnet/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31eb3063b26c33e506b2bd4bacdfda12abbcb15dbe4420dd775407600afede73
3
+ size 1008570723
checkpoint-250/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c9e218d549fce8962313cbc5e500cf9535e2dc8b2a6efc5e49f0ed0103f29dc
3
+ size 33159
checkpoint-250/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:483b234b87dd3086624ec2f0e53d7770a9bd91a890c6737f4b4e0cbfd8b01868
3
+ size 14757
checkpoint-250/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8b8c19e212c816e6ae6a68c9d3993f89829e0225f2363b08614133d806bf87c
3
+ size 1401
checkpoint-500/flux_controlnet/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "FluxControlNetModel",
3
+ "_diffusers_version": "0.36.0.dev0",
4
+ "_name_or_path": "InstantX/FLUX.1-dev-Controlnet-Canny",
5
+ "attention_head_dim": 128,
6
+ "axes_dims_rope": [
7
+ 16,
8
+ 56,
9
+ 56
10
+ ],
11
+ "conditioning_embedding_channels": null,
12
+ "guidance_embeds": true,
13
+ "in_channels": 64,
14
+ "joint_attention_dim": 4096,
15
+ "num_attention_heads": 24,
16
+ "num_layers": 5,
17
+ "num_mode": null,
18
+ "num_single_layers": 0,
19
+ "patch_size": 1,
20
+ "pooled_projection_dim": 768,
21
+ "quantization_config": {
22
+ "_load_in_4bit": true,
23
+ "_load_in_8bit": false,
24
+ "bnb_4bit_compute_dtype": "bfloat16",
25
+ "bnb_4bit_quant_storage": "uint8",
26
+ "bnb_4bit_quant_type": "fp4",
27
+ "bnb_4bit_use_double_quant": false,
28
+ "llm_int8_enable_fp32_cpu_offload": false,
29
+ "llm_int8_has_fp16_weight": false,
30
+ "llm_int8_skip_modules": null,
31
+ "llm_int8_threshold": 6.0,
32
+ "load_in_4bit": true,
33
+ "load_in_8bit": false,
34
+ "quant_method": "bitsandbytes"
35
+ }
36
+ }
checkpoint-500/flux_controlnet/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b84b141612a77cf56b99302cdab7df75dae0660480dfae116482aed196320403
3
+ size 1008570723
checkpoint-500/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48cc7fef9f2cf5303cf360bbdc2e3d57304095da4eab740bb642afb791370ef3
3
+ size 33159
checkpoint-500/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34a88794ecddc721de627e1eeb071460fe30cfb1471187ac1cf7dc6fd00d94af
3
+ size 14757
checkpoint-500/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e313c5da3fa01ea79bb9c28dac3eb2036c47353f16d8b188949ad9ffd1b0319
3
+ size 1401
checkpoint-750/flux_controlnet/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "FluxControlNetModel",
3
+ "_diffusers_version": "0.36.0.dev0",
4
+ "_name_or_path": "InstantX/FLUX.1-dev-Controlnet-Canny",
5
+ "attention_head_dim": 128,
6
+ "axes_dims_rope": [
7
+ 16,
8
+ 56,
9
+ 56
10
+ ],
11
+ "conditioning_embedding_channels": null,
12
+ "guidance_embeds": true,
13
+ "in_channels": 64,
14
+ "joint_attention_dim": 4096,
15
+ "num_attention_heads": 24,
16
+ "num_layers": 5,
17
+ "num_mode": null,
18
+ "num_single_layers": 0,
19
+ "patch_size": 1,
20
+ "pooled_projection_dim": 768,
21
+ "quantization_config": {
22
+ "_load_in_4bit": true,
23
+ "_load_in_8bit": false,
24
+ "bnb_4bit_compute_dtype": "bfloat16",
25
+ "bnb_4bit_quant_storage": "uint8",
26
+ "bnb_4bit_quant_type": "fp4",
27
+ "bnb_4bit_use_double_quant": false,
28
+ "llm_int8_enable_fp32_cpu_offload": false,
29
+ "llm_int8_has_fp16_weight": false,
30
+ "llm_int8_skip_modules": null,
31
+ "llm_int8_threshold": 6.0,
32
+ "load_in_4bit": true,
33
+ "load_in_8bit": false,
34
+ "quant_method": "bitsandbytes"
35
+ }
36
+ }
checkpoint-750/flux_controlnet/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58224dc63796d5c5bedbc8a6d61d484784cf1b9654771e319b978462809ecde5
3
+ size 1008570723
checkpoint-750/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0af2720096a4f4c9e5ea67c72c59f0884833e7fb643a59fef34dae7ee1cdf3e0
3
+ size 33159
checkpoint-750/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1213c40a99a7d682bc2634e7c15c4dfb072f72ff971e006f45d08e93e3ef3364
3
+ size 14757
checkpoint-750/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbb209ec03db3d6bfce21a13d8495b94eeffe5267259af8c5465bec102b81f82
3
+ size 1401
config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "FluxControlNetModel",
3
+ "_diffusers_version": "0.36.0.dev0",
4
+ "_name_or_path": "InstantX/FLUX.1-dev-Controlnet-Canny",
5
+ "attention_head_dim": 128,
6
+ "axes_dims_rope": [
7
+ 16,
8
+ 56,
9
+ 56
10
+ ],
11
+ "conditioning_embedding_channels": null,
12
+ "guidance_embeds": true,
13
+ "in_channels": 64,
14
+ "joint_attention_dim": 4096,
15
+ "num_attention_heads": 24,
16
+ "num_layers": 5,
17
+ "num_mode": null,
18
+ "num_single_layers": 0,
19
+ "patch_size": 1,
20
+ "pooled_projection_dim": 768,
21
+ "quantization_config": {
22
+ "_load_in_4bit": true,
23
+ "_load_in_8bit": false,
24
+ "bnb_4bit_compute_dtype": "bfloat16",
25
+ "bnb_4bit_quant_storage": "uint8",
26
+ "bnb_4bit_quant_type": "fp4",
27
+ "bnb_4bit_use_double_quant": false,
28
+ "llm_int8_enable_fp32_cpu_offload": false,
29
+ "llm_int8_has_fp16_weight": false,
30
+ "llm_int8_skip_modules": null,
31
+ "llm_int8_threshold": 6.0,
32
+ "load_in_4bit": true,
33
+ "load_in_8bit": false,
34
+ "quant_method": "bitsandbytes"
35
+ }
36
+ }
diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fea131fdf9ea739efb1c078afca481222bfeb2be938dfad3c2288fc72fc7c025
3
+ size 1009166467
image_control.png ADDED
images_0.png ADDED

Git LFS Details

  • SHA256: 927fffe962408b1a2e3ed966353fad5d086e6bc63133e3d2ba5b6d43ef24a650
  • Pointer size: 132 Bytes
  • Size of remote file: 1.03 MB