shiimi commited on
Commit
4b5eca6
·
verified ·
1 Parent(s): dcc7712

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "facebook/wav2vec2-xls-r-300m",
3
  "activation_dropout": 0.0,
4
  "adapter_kernel_size": 3,
5
  "adapter_stride": 2,
 
1
  {
2
+ "_name_or_path": "wav2vec2-large-xlsr-dv-demo",
3
  "activation_dropout": 0.0,
4
  "adapter_kernel_size": 3,
5
  "adapter_stride": 2,
preprocessor_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "feature_extractor_type": "Wav2Vec2FeatureExtractor",
4
+ "feature_size": 1,
5
+ "padding_side": "right",
6
+ "padding_value": 0.0,
7
+ "return_attention_mask": true,
8
+ "sampling_rate": 16000
9
+ }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fa22d306e84155b7402171fb798b4bd364bf757f0ff134a476dab6cd8cd6d291
3
- size 1262181098
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a560c2d9dec62f482b62e595b58dedfd707e3e96f40d75e880d18b9c80b4ed61
3
+ size 1262175210
special_tokens_map.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "bos_token": "<s>",
3
  "eos_token": "</s>",
4
- "pad_token": "[PAD]",
5
- "unk_token": "[UNK]"
6
  }
 
1
  {
2
  "bos_token": "<s>",
3
  "eos_token": "</s>",
4
+ "pad_token": "<pad>",
5
+ "unk_token": "<unk>"
6
  }
tokenizer_config.json CHANGED
@@ -2,11 +2,12 @@
2
  "bos_token": "<s>",
3
  "clean_up_tokenization_spaces": true,
4
  "do_lower_case": false,
 
5
  "eos_token": "</s>",
6
  "model_max_length": 1000000000000000019884624838656,
7
- "pad_token": "[PAD]",
8
- "replace_word_delimiter_char": " ",
9
- "tokenizer_class": "Wav2Vec2CTCTokenizer",
10
- "unk_token": "[UNK]",
11
  "word_delimiter_token": "|"
12
  }
 
2
  "bos_token": "<s>",
3
  "clean_up_tokenization_spaces": true,
4
  "do_lower_case": false,
5
+ "do_normalize": false,
6
  "eos_token": "</s>",
7
  "model_max_length": 1000000000000000019884624838656,
8
+ "pad_token": "<pad>",
9
+ "return_attention_mask": false,
10
+ "tokenizer_class": "Wav2Vec2Tokenizer",
11
+ "unk_token": "<unk>",
12
  "word_delimiter_token": "|"
13
  }