sgugger commited on
Commit
a69f6d7
·
1 Parent(s): bedf3a4

Upload tiny models for PegasusForCausalLM

Browse files
config.json CHANGED
@@ -6,7 +6,6 @@
6
  ],
7
  "attention_dropout": 0.1,
8
  "bos_token_id": 0,
9
- "classifier_dropout": 0.0,
10
  "d_model": 16,
11
  "decoder_attention_heads": 4,
12
  "decoder_ffn_dim": 4,
@@ -29,7 +28,7 @@
29
  "pad_token_id": 0,
30
  "scale_embedding": false,
31
  "torch_dtype": "float32",
32
- "transformers_version": "4.25.0.dev0",
33
  "use_cache": true,
34
- "vocab_size": 1305
35
  }
 
6
  ],
7
  "attention_dropout": 0.1,
8
  "bos_token_id": 0,
 
9
  "d_model": 16,
10
  "decoder_attention_heads": 4,
11
  "decoder_ffn_dim": 4,
 
28
  "pad_token_id": 0,
29
  "scale_embedding": false,
30
  "torch_dtype": "float32",
31
+ "transformers_version": "4.28.0.dev0",
32
  "use_cache": true,
33
+ "vocab_size": 96103
34
  }
generation_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 0,
4
+ "decoder_start_token_id": 0,
5
+ "eos_token_id": 1,
6
+ "forced_eos_token_id": 1,
7
+ "pad_token_id": 0,
8
+ "transformers_version": "4.28.0.dev0"
9
+ }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9bafe071e293138636ffafdbfb862ebb82fc342785cb60e0c966f844d4695c54
3
- size 133657
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c72d59aa7198d32ea1ef56b5fc209db2d2fc950aa234398f50208dc26123fd7f
3
+ size 6200729
spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0015189ef36359283fec8b93cf6d9ce51bca37eb1101defc68a53b394913b96c
3
+ size 1912529
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -108,7 +108,6 @@
108
  "mask_token": "<mask_2>",
109
  "mask_token_sent": "<mask_1>",
110
  "model_max_length": 200,
111
- "name_or_path": "google/pegasus-large",
112
  "offset": 103,
113
  "pad_token": "<pad>",
114
  "sp_model_kwargs": {},
 
108
  "mask_token": "<mask_2>",
109
  "mask_token_sent": "<mask_1>",
110
  "model_max_length": 200,
 
111
  "offset": 103,
112
  "pad_token": "<pad>",
113
  "sp_model_kwargs": {},