rkazants commited on
Commit
f3f69af
·
verified ·
1 Parent(s): 2ce2f39

Upload 6 files

Browse files

Add a tiny-model

config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "MambaForCausalLM"
4
+ ],
5
+ "bos_token_id": 0,
6
+ "conv_kernel": 2,
7
+ "d_model": 6,
8
+ "d_state": 3,
9
+ "eos_token_id": 0,
10
+ "expand": 1,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 768,
13
+ "initializer_range": 0.1,
14
+ "intermediate_size": 768,
15
+ "layer_norm_epsilon": 1e-05,
16
+ "model_type": "mamba",
17
+ "n_layer": 1,
18
+ "num_hidden_layers": 32,
19
+ "pad_token_id": 0,
20
+ "rescale_prenorm_residual": false,
21
+ "residual_in_fp32": true,
22
+ "state_size": 16,
23
+ "time_step_floor": 0.0001,
24
+ "time_step_init_scheme": "random",
25
+ "time_step_max": 0.1,
26
+ "time_step_min": 0.001,
27
+ "time_step_rank": 48,
28
+ "time_step_scale": 1.0,
29
+ "torch_dtype": "float32",
30
+ "transformers_version": "4.49.0",
31
+ "use_bias": false,
32
+ "use_cache": true,
33
+ "use_conv_bias": true,
34
+ "use_mambapy": false,
35
+ "vocab_size": 16
36
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 0,
4
+ "eos_token_id": 0,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.49.0"
7
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5251e63c7f5894eb3eb7c706beed5ba4b306bbe85f5fe92f4b63e07b750c7610
3
+ size 241324440
special_tokens_map.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<bos>",
3
+ "eos_token": "<eos>",
4
+ "pad_token": "<pad>"
5
+ }
tokenizer.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
5
+ "added_tokens": [],
6
+ "normalizer": null,
7
+ "pre_tokenizer": {
8
+ "type": "Split",
9
+ "pattern": {
10
+ "String": ""
11
+ },
12
+ "behavior": "Isolated",
13
+ "invert": false
14
+ },
15
+ "post_processor": null,
16
+ "decoder": null,
17
+ "model": {
18
+ "type": "WordLevel",
19
+ "vocab": {
20
+ "<pad>": 0,
21
+ "<bos>": 1,
22
+ "<eos>": 2,
23
+ "a": 3,
24
+ "b": 4,
25
+ "c": 5,
26
+ "d": 6,
27
+ "e": 7,
28
+ "f": 8,
29
+ "g": 9,
30
+ "h": 10,
31
+ "i": 11,
32
+ "j": 12,
33
+ "k": 13,
34
+ "l": 14,
35
+ "m": 15,
36
+ "n": 16,
37
+ "o": 17,
38
+ "p": 18,
39
+ "q": 19,
40
+ "r": 20,
41
+ "s": 21,
42
+ "t": 22,
43
+ "u": 23,
44
+ "v": 24,
45
+ "w": 25,
46
+ "x": 26,
47
+ "y": 27,
48
+ "z": 28
49
+ },
50
+ "unk_token": "<pad>"
51
+ }
52
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {},
3
+ "bos_token": "<bos>",
4
+ "clean_up_tokenization_spaces": false,
5
+ "eos_token": "<eos>",
6
+ "extra_special_tokens": {},
7
+ "model_max_length": 1000000000000000019884624838656,
8
+ "pad_token": "<pad>",
9
+ "tokenizer_class": "PreTrainedTokenizer"
10
+ }