XCogni commited on
Commit
b735213
·
verified ·
1 Parent(s): 68d3e97

Delete lyco.toml

Browse files
Files changed (1) hide show
  1. lyco.toml +0 -64
lyco.toml DELETED
@@ -1,64 +0,0 @@
1
- enable_conv = true
2
-
3
- # An example for use different algo/settings in "full" preset
4
- unet_target_module = [
5
- "Transformer2DModel",
6
- "ResnetBlock2D",
7
- "Downsample2D",
8
- "Upsample2D",
9
- "HunYuanDiTBlock",
10
- "DoubleStreamBlock",
11
- "SingleStreamBlock",
12
- "SingleDiTBlock",
13
- "MMDoubleStreamBlock", #HunYuanVideo
14
- "MMSingleStreamBlock", #HunYuanVideo
15
- ]
16
- unet_target_name = [
17
- "conv_in",
18
- "conv_out",
19
- "time_embedding.linear_1",
20
- "time_embedding.linear_2",
21
- ]
22
- text_encoder_target_module = [
23
- "CLIPAttention",
24
- "CLIPSdpaAttention",
25
- "CLIPMLP",
26
- "MT5Block",
27
- "BertLayer",
28
- ]
29
- text_encoder_target_name = [
30
- # "token_embedding", # not supported, Embedding module in CLIP
31
- ]
32
-
33
-
34
- [module_algo_map]
35
- [module_algo_map.CrossAttention]
36
- algo = "lokr"
37
- dim = 10000
38
- alpha = 10000
39
- factor = 4
40
-
41
- [module_algo_map.FeedForward] # MLP Layer in UNet
42
- algo = "lokr"
43
- dim = 10000
44
- alpha = 10000
45
- factor = 2
46
-
47
- [module_algo_map.ResnetBlock2D] # ResBlock in UNet
48
- algo = "lokr"
49
- dim = 10000
50
- alpha = 10000
51
- use_tucker = true # Use tucker decomposition for convolution
52
- factor = 8
53
-
54
- [module_algo_map.CLIPAttention] # Attention Layer in TE
55
- algo = "lokr"
56
- dim = 10000
57
- alpha = 10000
58
- factor = 8
59
-
60
- [module_algo_map.CLIPMLP] # MLP Layer in TE
61
- algo = "lokr"
62
- dim = 10000
63
- alpha = 10000
64
- factor = 8