Update with commit 93dd4fb9faad8c3b0f181e581f6e4ac061c0edfe
Browse filesSee: https://github.com/huggingface/transformers/commit/93dd4fb9faad8c3b0f181e581f6e4ac061c0edfe
- frameworks.json +1 -0
- pipeline_tags.json +2 -0
frameworks.json
CHANGED
|
@@ -345,6 +345,7 @@
|
|
| 345 |
{"model_type":"siglip_vision_model","pytorch":true,"processor":"AutoTokenizer"}
|
| 346 |
{"model_type":"smollm3","pytorch":true,"processor":"AutoTokenizer"}
|
| 347 |
{"model_type":"smolvlm","pytorch":true,"processor":"AutoProcessor"}
|
|
|
|
| 348 |
{"model_type":"speech-encoder-decoder","pytorch":true,"processor":"AutoTokenizer"}
|
| 349 |
{"model_type":"speech_to_text","pytorch":true,"processor":"AutoProcessor"}
|
| 350 |
{"model_type":"speecht5","pytorch":true,"processor":"AutoProcessor"}
|
|
|
|
| 345 |
{"model_type":"siglip_vision_model","pytorch":true,"processor":"AutoTokenizer"}
|
| 346 |
{"model_type":"smollm3","pytorch":true,"processor":"AutoTokenizer"}
|
| 347 |
{"model_type":"smolvlm","pytorch":true,"processor":"AutoProcessor"}
|
| 348 |
+
{"model_type":"solar_open","pytorch":true,"processor":"AutoTokenizer"}
|
| 349 |
{"model_type":"speech-encoder-decoder","pytorch":true,"processor":"AutoTokenizer"}
|
| 350 |
{"model_type":"speech_to_text","pytorch":true,"processor":"AutoProcessor"}
|
| 351 |
{"model_type":"speecht5","pytorch":true,"processor":"AutoProcessor"}
|
pipeline_tags.json
CHANGED
|
@@ -1110,6 +1110,8 @@
|
|
| 1110 |
{"model_class":"SmolVLMForConditionalGeneration","pipeline_tag":"image-to-text","auto_class":"AutoModelForImageTextToText"}
|
| 1111 |
{"model_class":"SmolVLMModel","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
|
| 1112 |
{"model_class":"SmolVLMVisionTransformer","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
|
|
|
|
|
|
|
| 1113 |
{"model_class":"Speech2Text2ForCausalLM","pipeline_tag":"text-generation","auto_class":"AutoModelForCausalLM"}
|
| 1114 |
{"model_class":"Speech2TextForConditionalGeneration","pipeline_tag":"automatic-speech-recognition","auto_class":"AutoModelForSpeechSeq2Seq"}
|
| 1115 |
{"model_class":"Speech2TextModel","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
|
|
|
|
| 1110 |
{"model_class":"SmolVLMForConditionalGeneration","pipeline_tag":"image-to-text","auto_class":"AutoModelForImageTextToText"}
|
| 1111 |
{"model_class":"SmolVLMModel","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
|
| 1112 |
{"model_class":"SmolVLMVisionTransformer","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
|
| 1113 |
+
{"model_class":"SolarOpenForCausalLM","pipeline_tag":"text-generation","auto_class":"AutoModelForCausalLM"}
|
| 1114 |
+
{"model_class":"SolarOpenModel","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
|
| 1115 |
{"model_class":"Speech2Text2ForCausalLM","pipeline_tag":"text-generation","auto_class":"AutoModelForCausalLM"}
|
| 1116 |
{"model_class":"Speech2TextForConditionalGeneration","pipeline_tag":"automatic-speech-recognition","auto_class":"AutoModelForSpeechSeq2Seq"}
|
| 1117 |
{"model_class":"Speech2TextModel","pipeline_tag":"feature-extraction","auto_class":"AutoModel"}
|