Disty0 commited on
Commit
1d1e9c5
·
verified ·
1 Parent(s): 9462803

Upload a better quant

Browse files
model_index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "FluxPipeline",
3
- "_diffusers_version": "0.36.0.dev0",
4
  "_name_or_path": "black-forest-labs/FLUX.1-dev",
5
  "feature_extractor": [
6
  null,
 
1
  {
2
  "_class_name": "FluxPipeline",
3
+ "_diffusers_version": "0.35.1",
4
  "_name_or_path": "black-forest-labs/FLUX.1-dev",
5
  "feature_extractor": [
6
  null,
scheduler/scheduler_config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "FlowMatchEulerDiscreteScheduler",
3
- "_diffusers_version": "0.36.0.dev0",
4
  "base_image_seq_len": 256,
5
  "base_shift": 0.5,
6
  "invert_sigmas": false,
 
1
  {
2
  "_class_name": "FlowMatchEulerDiscreteScheduler",
3
+ "_diffusers_version": "0.35.1",
4
  "base_image_seq_len": 256,
5
  "base_shift": 0.5,
6
  "invert_sigmas": false,
text_encoder_2/config.json CHANGED
@@ -23,21 +23,21 @@
23
  "output_past": true,
24
  "pad_token_id": 0,
25
  "quantization_config": {
26
- "add_skip_keys": false,
27
  "dequantize_fp32": false,
28
  "group_size": 0,
29
  "is_integer": true,
30
  "modules_dtype_dict": {},
31
  "modules_to_not_convert": [
32
- "wo",
33
  "wo"
34
  ],
35
  "non_blocking": false,
36
  "quant_conv": false,
37
  "quant_method": "sdnq",
38
- "quantization_device": null,
39
- "return_device": null,
40
  "svd_rank": 32,
 
41
  "use_quantized_matmul": false,
42
  "use_quantized_matmul_conv": false,
43
  "use_svd": true,
 
23
  "output_past": true,
24
  "pad_token_id": 0,
25
  "quantization_config": {
26
+ "add_skip_keys": true,
27
  "dequantize_fp32": false,
28
  "group_size": 0,
29
  "is_integer": true,
30
  "modules_dtype_dict": {},
31
  "modules_to_not_convert": [
 
32
  "wo"
33
  ],
34
  "non_blocking": false,
35
  "quant_conv": false,
36
  "quant_method": "sdnq",
37
+ "quantization_device": "xpu",
38
+ "return_device": "cpu",
39
  "svd_rank": 32,
40
+ "svd_steps": 8,
41
  "use_quantized_matmul": false,
42
  "use_quantized_matmul_conv": false,
43
  "use_svd": true,
text_encoder_2/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cfc8e0a4a7d2b29c410fe8a25059e0c154d0acca214d8571f0af671dc1c8fe96
3
  size 6309787408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0301d1a2d789a1b4e6570566ee114343b790aeeb36ba0c8a59e866b262385510
3
  size 6309787408
text_encoder_2/quantization_config.json CHANGED
@@ -1,19 +1,19 @@
1
  {
2
- "add_skip_keys": false,
3
  "dequantize_fp32": false,
4
  "group_size": 0,
5
  "is_integer": true,
6
  "modules_dtype_dict": {},
7
  "modules_to_not_convert": [
8
- "wo",
9
  "wo"
10
  ],
11
  "non_blocking": false,
12
  "quant_conv": false,
13
  "quant_method": "sdnq",
14
- "quantization_device": null,
15
- "return_device": null,
16
  "svd_rank": 32,
 
17
  "use_quantized_matmul": false,
18
  "use_quantized_matmul_conv": false,
19
  "use_svd": true,
 
1
  {
2
+ "add_skip_keys": true,
3
  "dequantize_fp32": false,
4
  "group_size": 0,
5
  "is_integer": true,
6
  "modules_dtype_dict": {},
7
  "modules_to_not_convert": [
 
8
  "wo"
9
  ],
10
  "non_blocking": false,
11
  "quant_conv": false,
12
  "quant_method": "sdnq",
13
+ "quantization_device": "xpu",
14
+ "return_device": "cpu",
15
  "svd_rank": 32,
16
+ "svd_steps": 8,
17
  "use_quantized_matmul": false,
18
  "use_quantized_matmul_conv": false,
19
  "use_svd": true,
tokenizer_2/tokenizer.json CHANGED
@@ -1,7 +1,14 @@
1
  {
2
  "version": "1.0",
3
  "truncation": null,
4
- "padding": null,
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
  "truncation": null,
4
+ "padding": {
5
+ "strategy": "BatchLongest",
6
+ "direction": "Right",
7
+ "pad_to_multiple_of": null,
8
+ "pad_id": 0,
9
+ "pad_type_id": 0,
10
+ "pad_token": "<pad>"
11
+ },
12
  "added_tokens": [
13
  {
14
  "id": 0,
transformer/config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "FluxTransformer2DModel",
3
- "_diffusers_version": "0.36.0.dev0",
4
  "_name_or_path": "black-forest-labs/FLUX.1-dev",
5
  "attention_head_dim": 128,
6
  "axes_dims_rope": [
@@ -18,21 +18,27 @@
18
  "patch_size": 1,
19
  "pooled_projection_dim": 768,
20
  "quantization_config": {
21
- "add_skip_keys": false,
22
  "dequantize_fp32": false,
23
  "group_size": 0,
24
  "is_integer": true,
25
  "modules_dtype_dict": {},
26
  "modules_to_not_convert": [
 
27
  "pos_embed",
28
- "norm"
 
 
 
 
29
  ],
30
  "non_blocking": false,
31
  "quant_conv": false,
32
  "quant_method": "sdnq",
33
- "quantization_device": null,
34
- "return_device": null,
35
  "svd_rank": 32,
 
36
  "use_quantized_matmul": false,
37
  "use_quantized_matmul_conv": false,
38
  "use_svd": true,
 
1
  {
2
  "_class_name": "FluxTransformer2DModel",
3
+ "_diffusers_version": "0.35.1",
4
  "_name_or_path": "black-forest-labs/FLUX.1-dev",
5
  "attention_head_dim": 128,
6
  "axes_dims_rope": [
 
18
  "patch_size": 1,
19
  "pooled_projection_dim": 768,
20
  "quantization_config": {
21
+ "add_skip_keys": true,
22
  "dequantize_fp32": false,
23
  "group_size": 0,
24
  "is_integer": true,
25
  "modules_dtype_dict": {},
26
  "modules_to_not_convert": [
27
+ ".proj_out",
28
  "pos_embed",
29
+ ".x_embedder",
30
+ ".norm_out",
31
+ "single_transformer_blocks.0.norm.linear.weight",
32
+ ".context_embedder",
33
+ ".time_text_embed"
34
  ],
35
  "non_blocking": false,
36
  "quant_conv": false,
37
  "quant_method": "sdnq",
38
+ "quantization_device": "xpu",
39
+ "return_device": "cpu",
40
  "svd_rank": 32,
41
+ "svd_steps": 8,
42
  "use_quantized_matmul": false,
43
  "use_quantized_matmul_conv": false,
44
  "use_svd": true,
transformer/diffusion_pytorch_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9b1de96674e1330a1c9a30429a099d84e9eac2db08a8692290207d11d95bb098
3
- size 8224750072
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77d0d617e815df520a5aee5c8e77fd3e973e54b5a47d54cdeb9c1db8c73b8598
3
+ size 6806132768
transformer/quantization_config.json CHANGED
@@ -1,19 +1,25 @@
1
  {
2
- "add_skip_keys": false,
3
  "dequantize_fp32": false,
4
  "group_size": 0,
5
  "is_integer": true,
6
  "modules_dtype_dict": {},
7
  "modules_to_not_convert": [
 
8
  "pos_embed",
9
- "norm"
 
 
 
 
10
  ],
11
  "non_blocking": false,
12
  "quant_conv": false,
13
  "quant_method": "sdnq",
14
- "quantization_device": null,
15
- "return_device": null,
16
  "svd_rank": 32,
 
17
  "use_quantized_matmul": false,
18
  "use_quantized_matmul_conv": false,
19
  "use_svd": true,
 
1
  {
2
+ "add_skip_keys": true,
3
  "dequantize_fp32": false,
4
  "group_size": 0,
5
  "is_integer": true,
6
  "modules_dtype_dict": {},
7
  "modules_to_not_convert": [
8
+ ".proj_out",
9
  "pos_embed",
10
+ ".x_embedder",
11
+ ".norm_out",
12
+ "single_transformer_blocks.0.norm.linear.weight",
13
+ ".context_embedder",
14
+ ".time_text_embed"
15
  ],
16
  "non_blocking": false,
17
  "quant_conv": false,
18
  "quant_method": "sdnq",
19
+ "quantization_device": "xpu",
20
+ "return_device": "cpu",
21
  "svd_rank": 32,
22
+ "svd_steps": 8,
23
  "use_quantized_matmul": false,
24
  "use_quantized_matmul_conv": false,
25
  "use_svd": true,
vae/config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "_class_name": "AutoencoderKL",
3
- "_diffusers_version": "0.36.0.dev0",
4
- "_name_or_path": "/mnt/DataSSD/AI/models/sd-webui/Diffusers/models--black-forest-labs--FLUX.1-dev/snapshots/3de623fc3c33e44ffbe2bad470d0f45bccf2eb21/vae",
5
  "act_fn": "silu",
6
  "block_out_channels": [
7
  128,
 
1
  {
2
  "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.35.1",
4
+ "_name_or_path": "/home/disty/.cache/huggingface/hub/models--black-forest-labs--FLUX.1-dev/snapshots/3de623fc3c33e44ffbe2bad470d0f45bccf2eb21/vae",
5
  "act_fn": "silu",
6
  "block_out_channels": [
7
  128,