CodeGoat24 commited on
Commit
4b10970
·
verified ·
1 Parent(s): 43a416f

Upload folder using huggingface_hub

Browse files
lora_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "step": 160,
3
+ "lora_params": {
4
+ "lora_rank": 64,
5
+ "lora_alpha": 128,
6
+ "target_modules": [
7
+ "attn.to_q",
8
+ "attn.to_k",
9
+ "attn.to_v",
10
+ "attn.to_out.0",
11
+ "attn.add_q_proj",
12
+ "attn.add_k_proj",
13
+ "attn.add_v_proj",
14
+ "attn.to_add_out",
15
+ "ff.linear_in",
16
+ "ff.linear_out",
17
+ "ff_context.linear_in",
18
+ "ff_context.linear_out",
19
+ "attn.to_qkv_mlp_proj"
20
+ ]
21
+ }
22
+ }
lora_optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c564b995c57f86fea7aa7819c34346b008ed8cb4bb2cb154e70f4ce2a0ac354f
3
+ size 1141056542
pytorch_lora_weights.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e72bc5058fe28f64b6e80f90640dfad1677d1b37f55487cfc793df163bf4e6f9
3
+ size 570458416