Devops-hestabit commited on
Commit
c0f1b1e
·
1 Parent(s): 12e45bf

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +12 -0
  2. checkpoint/config.json +28 -0
  3. checkpoint/pytorch_model.bin/key_to_filename.json +3 -0
  4. checkpoint/pytorch_model.bin/p0.model.embed_tokens.weight +3 -0
  5. checkpoint/pytorch_model.bin/p1.model.layers.0.self_attn.q_proj.weight +3 -0
  6. checkpoint/pytorch_model.bin/p10.model.layers.1.self_attn.q_proj.weight +3 -0
  7. checkpoint/pytorch_model.bin/p100.model.layers.11.self_attn.q_proj.weight +3 -0
  8. checkpoint/pytorch_model.bin/p101.model.layers.11.self_attn.k_proj.weight +3 -0
  9. checkpoint/pytorch_model.bin/p102.model.layers.11.self_attn.v_proj.weight +3 -0
  10. checkpoint/pytorch_model.bin/p103.model.layers.11.self_attn.o_proj.weight +3 -0
  11. checkpoint/pytorch_model.bin/p104.model.layers.11.mlp.gate_proj.weight +3 -0
  12. checkpoint/pytorch_model.bin/p105.model.layers.11.mlp.up_proj.weight +3 -0
  13. checkpoint/pytorch_model.bin/p106.model.layers.11.mlp.down_proj.weight +3 -0
  14. checkpoint/pytorch_model.bin/p107.model.layers.11.input_layernorm.weight +3 -0
  15. checkpoint/pytorch_model.bin/p108.model.layers.11.post_attention_layernorm.weight +3 -0
  16. checkpoint/pytorch_model.bin/p109.model.layers.12.self_attn.q_proj.weight +3 -0
  17. checkpoint/pytorch_model.bin/p11.model.layers.1.self_attn.k_proj.weight +3 -0
  18. checkpoint/pytorch_model.bin/p110.model.layers.12.self_attn.k_proj.weight +3 -0
  19. checkpoint/pytorch_model.bin/p111.model.layers.12.self_attn.v_proj.weight +3 -0
  20. checkpoint/pytorch_model.bin/p112.model.layers.12.self_attn.o_proj.weight +3 -0
  21. checkpoint/pytorch_model.bin/p113.model.layers.12.mlp.gate_proj.weight +3 -0
  22. checkpoint/pytorch_model.bin/p114.model.layers.12.mlp.up_proj.weight +3 -0
  23. checkpoint/pytorch_model.bin/p115.model.layers.12.mlp.down_proj.weight +3 -0
  24. checkpoint/pytorch_model.bin/p116.model.layers.12.input_layernorm.weight +3 -0
  25. checkpoint/pytorch_model.bin/p117.model.layers.12.post_attention_layernorm.weight +3 -0
  26. checkpoint/pytorch_model.bin/p118.model.layers.13.self_attn.q_proj.weight +3 -0
  27. checkpoint/pytorch_model.bin/p119.model.layers.13.self_attn.k_proj.weight +3 -0
  28. checkpoint/pytorch_model.bin/p12.model.layers.1.self_attn.v_proj.weight +3 -0
  29. checkpoint/pytorch_model.bin/p120.model.layers.13.self_attn.v_proj.weight +3 -0
  30. checkpoint/pytorch_model.bin/p121.model.layers.13.self_attn.o_proj.weight +3 -0
  31. checkpoint/pytorch_model.bin/p122.model.layers.13.mlp.gate_proj.weight +3 -0
  32. checkpoint/pytorch_model.bin/p123.model.layers.13.mlp.up_proj.weight +3 -0
  33. checkpoint/pytorch_model.bin/p124.model.layers.13.mlp.down_proj.weight +3 -0
  34. checkpoint/pytorch_model.bin/p125.model.layers.13.input_layernorm.weight +3 -0
  35. checkpoint/pytorch_model.bin/p126.model.layers.13.post_attention_layernorm.weight +3 -0
  36. checkpoint/pytorch_model.bin/p127.model.layers.14.self_attn.q_proj.weight +3 -0
  37. checkpoint/pytorch_model.bin/p128.model.layers.14.self_attn.k_proj.weight +3 -0
  38. checkpoint/pytorch_model.bin/p129.model.layers.14.self_attn.v_proj.weight +3 -0
  39. checkpoint/pytorch_model.bin/p13.model.layers.1.self_attn.o_proj.weight +3 -0
  40. checkpoint/pytorch_model.bin/p130.model.layers.14.self_attn.o_proj.weight +3 -0
  41. checkpoint/pytorch_model.bin/p131.model.layers.14.mlp.gate_proj.weight +3 -0
  42. checkpoint/pytorch_model.bin/p132.model.layers.14.mlp.up_proj.weight +3 -0
  43. checkpoint/pytorch_model.bin/p133.model.layers.14.mlp.down_proj.weight +3 -0
  44. checkpoint/pytorch_model.bin/p134.model.layers.14.input_layernorm.weight +3 -0
  45. checkpoint/pytorch_model.bin/p135.model.layers.14.post_attention_layernorm.weight +3 -0
  46. checkpoint/pytorch_model.bin/p136.model.layers.15.self_attn.q_proj.weight +3 -0
  47. checkpoint/pytorch_model.bin/p137.model.layers.15.self_attn.k_proj.weight +3 -0
  48. checkpoint/pytorch_model.bin/p138.model.layers.15.self_attn.v_proj.weight +3 -0
  49. checkpoint/pytorch_model.bin/p139.model.layers.15.self_attn.o_proj.weight +3 -0
  50. checkpoint/pytorch_model.bin/p14.model.layers.1.mlp.gate_proj.weight +3 -0
.gitattributes CHANGED
@@ -33,3 +33,15 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ compiled/06d8d79b2e6cc1226b11.neff filter=lfs diff=lfs merge=lfs -text
37
+ compiled/155d761f0f04c0dfafa4.neff filter=lfs diff=lfs merge=lfs -text
38
+ compiled/1e48d25d17534153c49b.neff filter=lfs diff=lfs merge=lfs -text
39
+ compiled/3acba0b9c2b6ed6cef4e.neff filter=lfs diff=lfs merge=lfs -text
40
+ compiled/4037bfa8570c7a4c8263.neff filter=lfs diff=lfs merge=lfs -text
41
+ compiled/593719fb1e5bc92c6930.neff filter=lfs diff=lfs merge=lfs -text
42
+ compiled/94788500d449002e920a.neff filter=lfs diff=lfs merge=lfs -text
43
+ compiled/b986fb8e44ec2c49abde.neff filter=lfs diff=lfs merge=lfs -text
44
+ compiled/c3a9ecdf6193d51e3d68.neff filter=lfs diff=lfs merge=lfs -text
45
+ compiled/d3a82db7b71be48af62f.neff filter=lfs diff=lfs merge=lfs -text
46
+ compiled/e236958f63e42d975382.neff filter=lfs diff=lfs merge=lfs -text
47
+ compiled/fcb85c009c403f3ca0bd.neff filter=lfs diff=lfs merge=lfs -text
checkpoint/config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "jondurbin/airoboros-l2-13b-gpt4-m2.0",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "bos_token_id": 1,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 5120,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 13824,
13
+ "max_position_embeddings": 4096,
14
+ "model_type": "llama",
15
+ "num_attention_heads": 40,
16
+ "num_hidden_layers": 40,
17
+ "num_key_value_heads": 40,
18
+ "pad_token_id": 0,
19
+ "pretraining_tp": 1,
20
+ "rms_norm_eps": 1e-05,
21
+ "rope_scaling": null,
22
+ "rope_theta": 10000.0,
23
+ "tie_word_embeddings": false,
24
+ "torch_dtype": "float32",
25
+ "transformers_version": "4.34.1",
26
+ "use_cache": true,
27
+ "vocab_size": 32000
28
+ }
checkpoint/pytorch_model.bin/key_to_filename.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:accc6d0b7e8ac18844510c952c638af18d0928c1f979af789d331fe83a76804b
3
+ size 33029
checkpoint/pytorch_model.bin/p0.model.embed_tokens.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9aaaa24241120eac161cd1f69a4e8547550d9871f7e3820f01f35741b519cd8
3
+ size 655360789
checkpoint/pytorch_model.bin/p1.model.layers.0.self_attn.q_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a67d64b0505b633e24993def23cabe2edc261f0318b3d22a4e24331091344616
3
+ size 104858492
checkpoint/pytorch_model.bin/p10.model.layers.1.self_attn.q_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc4772667e6f4c1349c73d7fa746cb0f454f0133098ca5bd56a54e9c685235cb
3
+ size 104858495
checkpoint/pytorch_model.bin/p100.model.layers.11.self_attn.q_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de323e203436284774bc08a23bd5a25d6f1981143abba373e37b0e4eda453874
3
+ size 104858501
checkpoint/pytorch_model.bin/p101.model.layers.11.self_attn.k_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3244744166d7bba0d564b9ccf2f2593a86c69c02d405133635635dca5517738
3
+ size 104858501
checkpoint/pytorch_model.bin/p102.model.layers.11.self_attn.v_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe642bf3ed15a254a144767ba9d4163045504b410b0cd1c1732476962eb53128
3
+ size 104858501
checkpoint/pytorch_model.bin/p103.model.layers.11.self_attn.o_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25a2ea749c4576fd9ca498cc1179346f6e8769878423723d4478869df8bd0607
3
+ size 104858501
checkpoint/pytorch_model.bin/p104.model.layers.11.mlp.gate_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0ca1e9e827e6bb2aaa971f9639df6bfc8958604bf2409ece5ab889bc30ca1c0
3
+ size 283116412
checkpoint/pytorch_model.bin/p105.model.layers.11.mlp.up_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a8900c2154490f7917b6493cfb8756aac376560e65f5ad1afa208b9e7828001
3
+ size 283116406
checkpoint/pytorch_model.bin/p106.model.layers.11.mlp.down_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36327bf78adca1d9765618daa20b6f6f02f548352a171a0ef87c6406a02f6720
3
+ size 283116412
checkpoint/pytorch_model.bin/p107.model.layers.11.input_layernorm.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af1d88f954e875ea901da53f07c9327a0a3635ec89a53292e8a81a7f5f537410
3
+ size 21378
checkpoint/pytorch_model.bin/p108.model.layers.11.post_attention_layernorm.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00ff8ce3e38bb504133efa54d872f34f75595b75fe553d6bc91a76f3948b8c7c
3
+ size 21405
checkpoint/pytorch_model.bin/p109.model.layers.12.self_attn.q_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5556eb4eb855c0dc69bd5fa1f060fbcab3918d30ed7765b831c34fa34fa8eb8
3
+ size 104858501
checkpoint/pytorch_model.bin/p11.model.layers.1.self_attn.k_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:750ded4345faa78568625d1e2ce658448b3945bf782e2f7c378a934aa64219ec
3
+ size 104858495
checkpoint/pytorch_model.bin/p110.model.layers.12.self_attn.k_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:972c9ce15513aaab55514de49dd816185c37f5f23c202575a70d8576abb96d7e
3
+ size 104858501
checkpoint/pytorch_model.bin/p111.model.layers.12.self_attn.v_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afa8b149b43cec42f3d44af50c11475f98c919f0b88e74ba7cd4fcac468ab7aa
3
+ size 104858501
checkpoint/pytorch_model.bin/p112.model.layers.12.self_attn.o_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:826bf9e894fb8f85d63c2cddb353daae7524fa66c29542b4cc61eaefc694fc15
3
+ size 104858501
checkpoint/pytorch_model.bin/p113.model.layers.12.mlp.gate_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2309c340af613796a648e95c12bde3a1611d5d8db66ffb237eca3197e2fbe4f
3
+ size 283116412
checkpoint/pytorch_model.bin/p114.model.layers.12.mlp.up_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05e3aaa99aeb69ec96ddb38973af3171de9a07b502aadf524890a4ece9eb23c8
3
+ size 283116406
checkpoint/pytorch_model.bin/p115.model.layers.12.mlp.down_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d15ee223be501396c9ce93eda8bed997b1712a83568d01823c70356d94ca889
3
+ size 283116412
checkpoint/pytorch_model.bin/p116.model.layers.12.input_layernorm.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ce9facf56b3eeb1dd522ac904bf0dcb2d09912c1f528c7e0da63d0bedfb0ed6
3
+ size 21378
checkpoint/pytorch_model.bin/p117.model.layers.12.post_attention_layernorm.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6abedc7e80884cb9af91e27dbff994b3c2cd802e26c2e4c55784c2b21fd0de6
3
+ size 21405
checkpoint/pytorch_model.bin/p118.model.layers.13.self_attn.q_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58d610ada643504269f5ff2e9e8e910988b0f886578793d8cfe808998ec7cdd6
3
+ size 104858501
checkpoint/pytorch_model.bin/p119.model.layers.13.self_attn.k_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c062661b53267e7ec1b35673d4cea27258f041e1e1112272222368fe796ef64
3
+ size 104858501
checkpoint/pytorch_model.bin/p12.model.layers.1.self_attn.v_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8387c0b3987149153e4be965f1613fbf6301d5787d1a393ad22f2ab0eb83a90
3
+ size 104858495
checkpoint/pytorch_model.bin/p120.model.layers.13.self_attn.v_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b74d2f60a84849cc63eb1482203498ea6d9a2053b75ecb9a5221d41c03da058
3
+ size 104858501
checkpoint/pytorch_model.bin/p121.model.layers.13.self_attn.o_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa62eb23ecdcb02f7afef52c5cb1b251ce199f7c076c55df609349a392be402d
3
+ size 104858501
checkpoint/pytorch_model.bin/p122.model.layers.13.mlp.gate_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:757651bae7fcfb48439b94b38539d0ca6efd534d5a4ac9dc7a61e6bc9a944fed
3
+ size 283116412
checkpoint/pytorch_model.bin/p123.model.layers.13.mlp.up_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2aef687b6723268dfb66b429e6d92b2b9a23b53d4032f8d320cf3e66f884b561
3
+ size 283116406
checkpoint/pytorch_model.bin/p124.model.layers.13.mlp.down_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c454735c471485d6ab00be92b3924a6f197fabd4e5f92ce097ac66bf944dd438
3
+ size 283116412
checkpoint/pytorch_model.bin/p125.model.layers.13.input_layernorm.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ad6fda8f21dcbdb6a95c4f2bedb7c939885ead00958468176899c7ee315f910
3
+ size 21378
checkpoint/pytorch_model.bin/p126.model.layers.13.post_attention_layernorm.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71548670929dc74b2cbf57ad2dbe25e25f1e7fbf6b0feff33a8c89833a5f2e10
3
+ size 21405
checkpoint/pytorch_model.bin/p127.model.layers.14.self_attn.q_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5305ff60b80af77dcbde725555c6f4ab7d8481cca6c3ee5e58fd4f6965c940b
3
+ size 104858501
checkpoint/pytorch_model.bin/p128.model.layers.14.self_attn.k_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:753fd9362116e82b1fe6227b5fa088720811c8caf2f0621297a651cb7e72924b
3
+ size 104858501
checkpoint/pytorch_model.bin/p129.model.layers.14.self_attn.v_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16a666e6b0e8bc0327d8a2e77726e163ade96b1d812ce2a9b81a5d5082965470
3
+ size 104858501
checkpoint/pytorch_model.bin/p13.model.layers.1.self_attn.o_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cde8648de1cd77df822dfb837caaa2908a573064adaaa960401465376d10da05
3
+ size 104858495
checkpoint/pytorch_model.bin/p130.model.layers.14.self_attn.o_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f67313942c68e5e11ce22ebc70c527940316aaa11fb8078aa13e48ddf39c5e3c
3
+ size 104858501
checkpoint/pytorch_model.bin/p131.model.layers.14.mlp.gate_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e59d4fddd888758ffdb96e2c91c2093b6cd7dc75e765418ee60d026a16555fa3
3
+ size 283116412
checkpoint/pytorch_model.bin/p132.model.layers.14.mlp.up_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a26abc86bcc06b1560f392e7666dbdcc17f2dddabf6fcd0fe5a1b57bd5ef4c42
3
+ size 283116406
checkpoint/pytorch_model.bin/p133.model.layers.14.mlp.down_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d1b55bc7c6f21f836053154a6537e2961350b7c25147dfdb80b85a2dddcd1a9
3
+ size 283116412
checkpoint/pytorch_model.bin/p134.model.layers.14.input_layernorm.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:911b3fe118a541cf8a1cf2cc905c18f0a8a35ad81e0f5d714ef8a3bdd5a82b8c
3
+ size 21378
checkpoint/pytorch_model.bin/p135.model.layers.14.post_attention_layernorm.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db57c17c2e30306111784a7c575cbf8f79e37cb15b86ed90bf093dfbd1e280ec
3
+ size 21405
checkpoint/pytorch_model.bin/p136.model.layers.15.self_attn.q_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4c61f8c63fb39a0369346183845c15e5553fa255d5207b0c6b97660c65aab56
3
+ size 104858501
checkpoint/pytorch_model.bin/p137.model.layers.15.self_attn.k_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:916c6d916d05a16d5647d14caadeb319f9713a8cdba5a12e58f2774a48f1da75
3
+ size 104858501
checkpoint/pytorch_model.bin/p138.model.layers.15.self_attn.v_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f38a6ee4c2ba67819b9b7041e98d68e7f0badcdd5bb9808b645484d7326429a0
3
+ size 104858501
checkpoint/pytorch_model.bin/p139.model.layers.15.self_attn.o_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:526ec3bdb9fa597e5d339e53af57c3b056d5cc90a85ca615e3243be202e54c2f
3
+ size 104858501
checkpoint/pytorch_model.bin/p14.model.layers.1.mlp.gate_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb1bda4f4159080587b22c1b33d997d5639439f62e8419d8ba1ddbfbd4f81284
3
+ size 283116406