Update with commit d2cdefb9ec5080b78be302740a9cbaf48241b5c6
Browse filesSee: https://github.com/huggingface/transformers/commit/d2cdefb9ec5080b78be302740a9cbaf48241b5c6
- frameworks.json +1 -0
- pipeline_tags.json +5 -0
frameworks.json
CHANGED
|
@@ -208,6 +208,7 @@
|
|
| 208 |
{"model_type":"vits","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoTokenizer"}
|
| 209 |
{"model_type":"vivit","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoTokenizer"}
|
| 210 |
{"model_type":"wav2vec2","pytorch":true,"tensorflow":true,"flax":true,"processor":"AutoProcessor"}
|
|
|
|
| 211 |
{"model_type":"wav2vec2-conformer","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoProcessor"}
|
| 212 |
{"model_type":"wavlm","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoProcessor"}
|
| 213 |
{"model_type":"whisper","pytorch":true,"tensorflow":true,"flax":true,"processor":"AutoProcessor"}
|
|
|
|
| 208 |
{"model_type":"vits","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoTokenizer"}
|
| 209 |
{"model_type":"vivit","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoTokenizer"}
|
| 210 |
{"model_type":"wav2vec2","pytorch":true,"tensorflow":true,"flax":true,"processor":"AutoProcessor"}
|
| 211 |
+
{"model_type":"wav2vec2-bert","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoProcessor"}
|
| 212 |
{"model_type":"wav2vec2-conformer","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoProcessor"}
|
| 213 |
{"model_type":"wavlm","pytorch":true,"tensorflow":false,"flax":false,"processor":"AutoProcessor"}
|
| 214 |
{"model_type":"whisper","pytorch":true,"tensorflow":true,"flax":true,"processor":"AutoProcessor"}
|
pipeline_tags.json
CHANGED
|
@@ -986,6 +986,11 @@
|
|
| 986 |
{"model_class":"VitsModel","pipeline_tag":"text-to-audio","auto_class":"AutoModelForTextToWaveform"}
|
| 987 |
{"model_class":"VivitForVideoClassification","pipeline_tag":"video-classification","auto_class":"AutoModelForVideoClassification"}
|
| 988 |
{"model_class":"VivitModel","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 989 |
{"model_class":"Wav2Vec2ConformerForAudioFrameClassification","pipeline_tag":"audio-frame-classification","auto_class":"AutoModelForAudioFrameClassification"}
|
| 990 |
{"model_class":"Wav2Vec2ConformerForCTC","pipeline_tag":"automatic-speech-recognition","auto_class":"AutoModelForCTC"}
|
| 991 |
{"model_class":"Wav2Vec2ConformerForPreTraining","pipeline_tag":"pretraining","auto_class":"AutoModelForPreTraining"}
|
|
|
|
| 986 |
{"model_class":"VitsModel","pipeline_tag":"text-to-audio","auto_class":"AutoModelForTextToWaveform"}
|
| 987 |
{"model_class":"VivitForVideoClassification","pipeline_tag":"video-classification","auto_class":"AutoModelForVideoClassification"}
|
| 988 |
{"model_class":"VivitModel","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
|
| 989 |
+
{"model_class":"Wav2Vec2BertForAudioFrameClassification","pipeline_tag":"audio-frame-classification","auto_class":"AutoModelForAudioFrameClassification"}
|
| 990 |
+
{"model_class":"Wav2Vec2BertForCTC","pipeline_tag":"automatic-speech-recognition","auto_class":"AutoModelForCTC"}
|
| 991 |
+
{"model_class":"Wav2Vec2BertForSequenceClassification","pipeline_tag":"audio-classification","auto_class":"AutoModelForAudioClassification"}
|
| 992 |
+
{"model_class":"Wav2Vec2BertForXVector","pipeline_tag":"audio-xvector","auto_class":"AutoModelForAudioXVector"}
|
| 993 |
+
{"model_class":"Wav2Vec2BertModel","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
|
| 994 |
{"model_class":"Wav2Vec2ConformerForAudioFrameClassification","pipeline_tag":"audio-frame-classification","auto_class":"AutoModelForAudioFrameClassification"}
|
| 995 |
{"model_class":"Wav2Vec2ConformerForCTC","pipeline_tag":"automatic-speech-recognition","auto_class":"AutoModelForCTC"}
|
| 996 |
{"model_class":"Wav2Vec2ConformerForPreTraining","pipeline_tag":"pretraining","auto_class":"AutoModelForPreTraining"}
|