Harras111 commited on
Commit
5920679
·
verified ·
1 Parent(s): 3a056e7

Upload folder using huggingface_hub

Browse files
config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "EOUModelForHF"
4
+ ],
5
+ "base_model_name": "HuggingFaceTB/SmolLM2-135M",
6
+ "dropout_rate": 0.1,
7
+ "hidden_size": 576,
8
+ "model_type": "eou_model",
9
+ "torch_dtype": "float32",
10
+ "transformers_version": "4.53.2"
11
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dafd872bd7c14c91bcc21ea4fd2052d1580822074a3dad2c07f1a4d4bce59281
3
+ size 538759636
special_tokens_map.json CHANGED
@@ -32,7 +32,13 @@
32
  "rstrip": false,
33
  "single_word": false
34
  },
35
- "pad_token": "<|endoftext|>",
 
 
 
 
 
 
36
  "unk_token": {
37
  "content": "<|endoftext|>",
38
  "lstrip": false,
 
32
  "rstrip": false,
33
  "single_word": false
34
  },
35
+ "pad_token": {
36
+ "content": "<|endoftext|>",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false
41
+ },
42
  "unk_token": {
43
  "content": "<|endoftext|>",
44
  "lstrip": false,
tokenizer_config.json CHANGED
@@ -161,9 +161,16 @@
161
  "clean_up_tokenization_spaces": false,
162
  "eos_token": "<|endoftext|>",
163
  "extra_special_tokens": {},
 
164
  "model_max_length": 8192,
 
165
  "pad_token": "<|endoftext|>",
 
 
 
166
  "tokenizer_class": "GPT2Tokenizer",
 
 
167
  "unk_token": "<|endoftext|>",
168
  "vocab_size": 49152
169
  }
 
161
  "clean_up_tokenization_spaces": false,
162
  "eos_token": "<|endoftext|>",
163
  "extra_special_tokens": {},
164
+ "max_length": 512,
165
  "model_max_length": 8192,
166
+ "pad_to_multiple_of": null,
167
  "pad_token": "<|endoftext|>",
168
+ "pad_token_type_id": 0,
169
+ "padding_side": "right",
170
+ "stride": 0,
171
  "tokenizer_class": "GPT2Tokenizer",
172
+ "truncation_side": "right",
173
+ "truncation_strategy": "longest_first",
174
  "unk_token": "<|endoftext|>",
175
  "vocab_size": 49152
176
  }