File size: 1,609 Bytes
8622ebe
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
{
  "_class_name": "ZImageTransformer2DModel",
  "_diffusers_version": "0.36.0",
  "_name_or_path": "Tongyi-MAI/Z-Image-Turbo",
  "all_f_patch_size": [
    1
  ],
  "all_patch_size": [
    2
  ],
  "axes_dims": [
    32,
    48,
    48
  ],
  "axes_lens": [
    1536,
    512,
    512
  ],
  "cap_feat_dim": 2560,
  "dim": 3840,
  "in_channels": 16,
  "n_heads": 30,
  "n_kv_heads": 30,
  "n_layers": 30,
  "n_refiner_layers": 2,
  "norm_eps": 1e-05,
  "qk_norm": true,
  "quantization_config": {
    "add_skip_keys": true,
    "dequantize_fp32": false,
    "dynamic_loss_threshold": 0.01,
    "group_size": 0,
    "is_integer": true,
    "is_training": false,
    "modules_dtype_dict": {},
    "modules_to_not_convert": [
      "prediction_coefs",
      "siglip_embedder",
      "cap_embedder",
      "norm_out",
      "correction_coefs",
      "lm_head",
      "embed_tokens",
      "layers.0.adaLN_modulation.0.weight",
      "all_final_layer",
      "embedding_projection",
      "t_embedder",
      "proj_out",
      "all_x_embedder"
    ],
    "non_blocking": false,
    "quant_conv": false,
    "quant_method": "sdnq",
    "quantization_device": "cuda",
    "quantized_matmul_dtype": null,
    "return_device": "cpu",
    "sdnq_version": "0.1.3",
    "svd_rank": 32,
    "svd_steps": 8,
    "use_dynamic_quantization": false,
    "use_grad_ckpt": true,
    "use_quantized_matmul": true,
    "use_quantized_matmul_conv": false,
    "use_static_quantization": true,
    "use_stochastic_rounding": false,
    "use_svd": false,
    "weights_dtype": "int8"
  },
  "rope_theta": 256.0,
  "t_scale": 1000.0
}