Mojtabazarrin commited on
Commit
e065af4
·
1 Parent(s): f83f7b0

Upload WhisperForConditionalGeneration

Browse files
Files changed (3) hide show
  1. config.json +2 -2
  2. generation_config.json +2 -2
  3. pytorch_model.bin +1 -1
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "Mojtabazarrin/whisper-base-fa",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
@@ -45,7 +45,7 @@
45
  "scale_embedding": false,
46
  "suppress_tokens": [],
47
  "torch_dtype": "float32",
48
- "transformers_version": "4.34.0",
49
  "use_cache": true,
50
  "use_weighted_layer_sum": false,
51
  "vocab_size": 51865
 
1
  {
2
+ "_name_or_path": "/kaggle/input/whisper-base-fa-train/whisper-base-fa/checkpoint-4000",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
 
45
  "scale_embedding": false,
46
  "suppress_tokens": [],
47
  "torch_dtype": "float32",
48
+ "transformers_version": "4.33.0",
49
  "use_cache": true,
50
  "use_weighted_layer_sum": false,
51
  "vocab_size": 51865
generation_config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "_from_model_config": true,
3
  "begin_suppress_tokens": [
4
  220,
5
  50257
@@ -9,6 +8,7 @@
9
  "eos_token_id": 50257,
10
  "max_length": 448,
11
  "pad_token_id": 50257,
 
12
  "suppress_tokens": [],
13
- "transformers_version": "4.34.0"
14
  }
 
1
  {
 
2
  "begin_suppress_tokens": [
3
  220,
4
  50257
 
8
  "eos_token_id": 50257,
9
  "max_length": 448,
10
  "pad_token_id": 50257,
11
+ "return_timestamps": false,
12
  "suppress_tokens": [],
13
+ "transformers_version": "4.33.0"
14
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0813531dceaccbe5e1da8e1b04515fb8da872d7c57f7925b6c362bfc905c382e
3
  size 290455393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1d4c77517f4d94828693f5c978919fdf1c0221afb8fc4abe0f80b07b45934b2
3
  size 290455393