Update with commit b768d8b157e227c7c84e57014f5183d461556579
Browse filesSee: https://github.com/huggingface/transformers/commit/b768d8b157e227c7c84e57014f5183d461556579
- frameworks.json +1 -0
- pipeline_tags.json +4 -1
frameworks.json
CHANGED
|
@@ -411,6 +411,7 @@
|
|
| 411 |
{"model_type":"vivit","pytorch":true,"processor":"AutoTokenizer"}
|
| 412 |
{"model_type":"vjepa2","pytorch":true,"processor":"AutoTokenizer"}
|
| 413 |
{"model_type":"voxtral","pytorch":true,"processor":"AutoProcessor"}
|
|
|
|
| 414 |
{"model_type":"wav2vec2","pytorch":true,"processor":"AutoProcessor"}
|
| 415 |
{"model_type":"wav2vec2-bert","pytorch":true,"processor":"AutoProcessor"}
|
| 416 |
{"model_type":"wav2vec2-conformer","pytorch":true,"processor":"AutoProcessor"}
|
|
|
|
| 411 |
{"model_type":"vivit","pytorch":true,"processor":"AutoTokenizer"}
|
| 412 |
{"model_type":"vjepa2","pytorch":true,"processor":"AutoTokenizer"}
|
| 413 |
{"model_type":"voxtral","pytorch":true,"processor":"AutoProcessor"}
|
| 414 |
+
{"model_type":"voxtral_realtime","pytorch":true,"processor":"AutoProcessor"}
|
| 415 |
{"model_type":"wav2vec2","pytorch":true,"processor":"AutoProcessor"}
|
| 416 |
{"model_type":"wav2vec2-bert","pytorch":true,"processor":"AutoProcessor"}
|
| 417 |
{"model_type":"wav2vec2-conformer","pytorch":true,"processor":"AutoProcessor"}
|
pipeline_tags.json
CHANGED
|
@@ -1503,7 +1503,10 @@
|
|
| 1503 |
{"model_class":"VivitForVideoClassification","pipeline_tag":"video-classification","auto_class":"AutoModelForVideoClassification"}
|
| 1504 |
{"model_class":"VivitModel","pipeline_tag":"image-feature-extraction","auto_class":"AutoModel"}
|
| 1505 |
{"model_class":"VoxtralEncoder","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
|
| 1506 |
-
{"model_class":"VoxtralForConditionalGeneration","pipeline_tag":"
|
|
|
|
|
|
|
|
|
|
| 1507 |
{"model_class":"Wav2Vec2BertForAudioFrameClassification","pipeline_tag":"audio-frame-classification","auto_class":"AutoModelForAudioFrameClassification"}
|
| 1508 |
{"model_class":"Wav2Vec2BertForCTC","pipeline_tag":"automatic-speech-recognition","auto_class":"AutoModelForCTC"}
|
| 1509 |
{"model_class":"Wav2Vec2BertForSequenceClassification","pipeline_tag":"audio-classification","auto_class":"AutoModelForAudioClassification"}
|
|
|
|
| 1503 |
{"model_class":"VivitForVideoClassification","pipeline_tag":"video-classification","auto_class":"AutoModelForVideoClassification"}
|
| 1504 |
{"model_class":"VivitModel","pipeline_tag":"image-feature-extraction","auto_class":"AutoModel"}
|
| 1505 |
{"model_class":"VoxtralEncoder","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
|
| 1506 |
+
{"model_class":"VoxtralForConditionalGeneration","pipeline_tag":"automatic-speech-recognition","auto_class":"AutoModelForSpeechSeq2Seq"}
|
| 1507 |
+
{"model_class":"VoxtralRealtimeEncoder","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
|
| 1508 |
+
{"model_class":"VoxtralRealtimeForConditionalGeneration","pipeline_tag":"automatic-speech-recognition","auto_class":"AutoModelForSpeechSeq2Seq"}
|
| 1509 |
+
{"model_class":"VoxtralRealtimeTextModel","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
|
| 1510 |
{"model_class":"Wav2Vec2BertForAudioFrameClassification","pipeline_tag":"audio-frame-classification","auto_class":"AutoModelForAudioFrameClassification"}
|
| 1511 |
{"model_class":"Wav2Vec2BertForCTC","pipeline_tag":"automatic-speech-recognition","auto_class":"AutoModelForCTC"}
|
| 1512 |
{"model_class":"Wav2Vec2BertForSequenceClassification","pipeline_tag":"audio-classification","auto_class":"AutoModelForAudioClassification"}
|