gate369 commited on
Commit
62d8c04
·
verified ·
1 Parent(s): 31fbe9b

Upload folder using huggingface_hub

Browse files
config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "vocab_size": 61,
3
+ "max_context_length": 128000,
4
+ "embed_dim": 1020,
5
+ "num_heads": 12,
6
+ "num_layers": 5,
7
+ "ff_dim": 2048,
8
+ "dropout": 0.01952712192382383
9
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43259921130c8c809870360cfa92de190dec0115e6163d11bd9bed0f9dd2b367
3
+ size 689706152
model.safetensors.index.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 689699944
4
+ },
5
+ "weight_map": {
6
+ "token_embedding.weight": "model.safetensors",
7
+ "position_embedding.weight": "model.safetensors",
8
+ "blocks.0.attention.context_scale": "model.safetensors",
9
+ "blocks.0.attention.qkv.weight": "model.safetensors",
10
+ "blocks.0.attention.out_proj.weight": "model.safetensors",
11
+ "blocks.0.attention.out_proj.bias": "model.safetensors",
12
+ "blocks.0.norm1.weight": "model.safetensors",
13
+ "blocks.0.norm1.bias": "model.safetensors",
14
+ "blocks.0.norm2.weight": "model.safetensors",
15
+ "blocks.0.norm2.bias": "model.safetensors",
16
+ "blocks.0.ff.0.weight": "model.safetensors",
17
+ "blocks.0.ff.0.bias": "model.safetensors",
18
+ "blocks.0.ff.2.weight": "model.safetensors",
19
+ "blocks.0.ff.2.bias": "model.safetensors",
20
+ "blocks.1.attention.context_scale": "model.safetensors",
21
+ "blocks.1.attention.qkv.weight": "model.safetensors",
22
+ "blocks.1.attention.out_proj.weight": "model.safetensors",
23
+ "blocks.1.attention.out_proj.bias": "model.safetensors",
24
+ "blocks.1.norm1.weight": "model.safetensors",
25
+ "blocks.1.norm1.bias": "model.safetensors",
26
+ "blocks.1.norm2.weight": "model.safetensors",
27
+ "blocks.1.norm2.bias": "model.safetensors",
28
+ "blocks.1.ff.0.weight": "model.safetensors",
29
+ "blocks.1.ff.0.bias": "model.safetensors",
30
+ "blocks.1.ff.2.weight": "model.safetensors",
31
+ "blocks.1.ff.2.bias": "model.safetensors",
32
+ "blocks.2.attention.context_scale": "model.safetensors",
33
+ "blocks.2.attention.qkv.weight": "model.safetensors",
34
+ "blocks.2.attention.out_proj.weight": "model.safetensors",
35
+ "blocks.2.attention.out_proj.bias": "model.safetensors",
36
+ "blocks.2.norm1.weight": "model.safetensors",
37
+ "blocks.2.norm1.bias": "model.safetensors",
38
+ "blocks.2.norm2.weight": "model.safetensors",
39
+ "blocks.2.norm2.bias": "model.safetensors",
40
+ "blocks.2.ff.0.weight": "model.safetensors",
41
+ "blocks.2.ff.0.bias": "model.safetensors",
42
+ "blocks.2.ff.2.weight": "model.safetensors",
43
+ "blocks.2.ff.2.bias": "model.safetensors",
44
+ "blocks.3.attention.context_scale": "model.safetensors",
45
+ "blocks.3.attention.qkv.weight": "model.safetensors",
46
+ "blocks.3.attention.out_proj.weight": "model.safetensors",
47
+ "blocks.3.attention.out_proj.bias": "model.safetensors",
48
+ "blocks.3.norm1.weight": "model.safetensors",
49
+ "blocks.3.norm1.bias": "model.safetensors",
50
+ "blocks.3.norm2.weight": "model.safetensors",
51
+ "blocks.3.norm2.bias": "model.safetensors",
52
+ "blocks.3.ff.0.weight": "model.safetensors",
53
+ "blocks.3.ff.0.bias": "model.safetensors",
54
+ "blocks.3.ff.2.weight": "model.safetensors",
55
+ "blocks.3.ff.2.bias": "model.safetensors",
56
+ "blocks.4.attention.context_scale": "model.safetensors",
57
+ "blocks.4.attention.qkv.weight": "model.safetensors",
58
+ "blocks.4.attention.out_proj.weight": "model.safetensors",
59
+ "blocks.4.attention.out_proj.bias": "model.safetensors",
60
+ "blocks.4.norm1.weight": "model.safetensors",
61
+ "blocks.4.norm1.bias": "model.safetensors",
62
+ "blocks.4.norm2.weight": "model.safetensors",
63
+ "blocks.4.norm2.bias": "model.safetensors",
64
+ "blocks.4.ff.0.weight": "model.safetensors",
65
+ "blocks.4.ff.0.bias": "model.safetensors",
66
+ "blocks.4.ff.2.weight": "model.safetensors",
67
+ "blocks.4.ff.2.bias": "model.safetensors",
68
+ "norm.weight": "model.safetensors",
69
+ "norm.bias": "model.safetensors",
70
+ "output.weight": "model.safetensors",
71
+ "output.bias": "model.safetensors"
72
+ }
73
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "unk_token": "[UNK]"
3
+ }
tokenizer.json ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "\n": 0,
3
+ " ": 1,
4
+ "!": 2,
5
+ "\"": 3,
6
+ "'": 4,
7
+ ",": 5,
8
+ "-": 6,
9
+ ".": 7,
10
+ ":": 8,
11
+ "<": 9,
12
+ ">": 10,
13
+ "?": 11,
14
+ "A": 12,
15
+ "B": 13,
16
+ "C": 14,
17
+ "D": 15,
18
+ "E": 16,
19
+ "F": 17,
20
+ "G": 18,
21
+ "H": 19,
22
+ "I": 20,
23
+ "K": 21,
24
+ "L": 22,
25
+ "M": 23,
26
+ "N": 24,
27
+ "O": 25,
28
+ "P": 26,
29
+ "R": 27,
30
+ "S": 28,
31
+ "T": 29,
32
+ "V": 30,
33
+ "W": 31,
34
+ "Y": 32,
35
+ "Z": 33,
36
+ "a": 34,
37
+ "b": 35,
38
+ "c": 36,
39
+ "d": 37,
40
+ "e": 38,
41
+ "f": 39,
42
+ "g": 40,
43
+ "h": 41,
44
+ "i": 42,
45
+ "j": 43,
46
+ "k": 44,
47
+ "l": 45,
48
+ "m": 46,
49
+ "n": 47,
50
+ "o": 48,
51
+ "p": 49,
52
+ "q": 50,
53
+ "r": 51,
54
+ "s": 52,
55
+ "t": 53,
56
+ "u": 54,
57
+ "v": 55,
58
+ "w": 56,
59
+ "x": 57,
60
+ "y": 58,
61
+ "z": 59,
62
+ "|": 60
63
+ }
tokenizer.model ADDED
File without changes
tokenizer_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "tokenizer_class": "CharLevelTokenizer",
3
+ "unk_token": "[UNK]"
4
+ }