diff --git "a/processor_classes.json" "b/processor_classes.json" --- "a/processor_classes.json" +++ "b/processor_classes.json" @@ -1,1854 +1,3087 @@ { "AlbertModel": { - "tokenizer_class": "AlbertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "AlbertTokenizerFast" + ], + "processor_classes": [] }, "BartModel": { - "tokenizer_class": "BartTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BartTokenizerFast", + "BartTokenizer" + ], + "processor_classes": [] }, "BeitModel": { - "tokenizer_class": null, - "processor_class": "BeitImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "BeitImageProcessor" + ] }, "BertLMHeadModel": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "BertModel": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "BigBirdModel": { - "tokenizer_class": "BigBirdTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BigBirdTokenizerFast" + ], + "processor_classes": [] }, "BigBirdPegasusModel": { - "tokenizer_class": "PegasusTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "PegasusTokenizerFast" + ], + "processor_classes": [] }, "BlenderbotSmallModel": { - "tokenizer_class": null, - "processor_class": null + "tokenizer_classes": [], + "processor_classes": [] }, "BlenderbotModel": { - "tokenizer_class": "BlenderbotTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BlenderbotTokenizerFast", + "BlenderbotTokenizer" + ], + "processor_classes": [] }, "BloomModel": { - "tokenizer_class": "BloomTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BloomTokenizerFast" + ], + "processor_classes": [] }, "CanineModel": { - "tokenizer_class": "CanineTokenizer", - "processor_class": null + "tokenizer_classes": [ + "CanineTokenizer" + ], + "processor_classes": [] }, "CLIPModel": { - "tokenizer_class": "CLIPTokenizerFast", - "processor_class": "CLIPImageProcessor" + "tokenizer_classes": [ + "CLIPTokenizerFast", + "CLIPTokenizer" + ], + "processor_classes": [ + "CLIPImageProcessor" + ] }, "CodeGenModel": { - "tokenizer_class": "CodeGenTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "CodeGenTokenizerFast", + "CodeGenTokenizer" + ], + "processor_classes": [] }, "ConditionalDetrModel": { - "tokenizer_class": null, - "processor_class": "ConditionalDetrFeatureExtractor" + "tokenizer_classes": [], + "processor_classes": [ + "ConditionalDetrFeatureExtractor" + ] }, "ConvBertModel": { - "tokenizer_class": "ConvBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "ConvBertTokenizerFast", + "ConvBertTokenizer" + ], + "processor_classes": [] }, "ConvNextModel": { - "tokenizer_class": null, - "processor_class": "ConvNextImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ConvNextImageProcessor" + ] }, "CTRLLMHeadModel": { - "tokenizer_class": "CTRLTokenizer", - "processor_class": null + "tokenizer_classes": [ + "CTRLTokenizer" + ], + "processor_classes": [] }, "CTRLModel": { - "tokenizer_class": "CTRLTokenizer", - "processor_class": null + "tokenizer_classes": [ + "CTRLTokenizer" + ], + "processor_classes": [] }, "CvtModel": { - "tokenizer_class": null, - "processor_class": "ConvNextImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ConvNextImageProcessor" + ] }, "Data2VecAudioModel": { - "tokenizer_class": null, - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "Data2VecTextModel": { - "tokenizer_class": "RobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RobertaTokenizerFast", + "RobertaTokenizer" + ], + "processor_classes": [] }, "Data2VecVisionModel": { - "tokenizer_class": null, - "processor_class": "BeitImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "BeitImageProcessor" + ] }, "DebertaV2Model": { - "tokenizer_class": "DebertaV2TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "DebertaV2TokenizerFast" + ], + "processor_classes": [] }, "DebertaModel": { - "tokenizer_class": "DebertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "DebertaTokenizerFast", + "DebertaTokenizer" + ], + "processor_classes": [] }, "DeformableDetrModel": { - "tokenizer_class": null, - "processor_class": "DeformableDetrFeatureExtractor" + "tokenizer_classes": [], + "processor_classes": [ + "DeformableDetrFeatureExtractor" + ] }, "DeiTModel": { - "tokenizer_class": null, - "processor_class": "DeiTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "DeiTImageProcessor" + ] }, "DetrModel": { - "tokenizer_class": null, - "processor_class": "DetrFeatureExtractor" + "tokenizer_classes": [], + "processor_classes": [ + "DetrFeatureExtractor" + ] }, "DistilBertModel": { - "tokenizer_class": "DistilBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "DistilBertTokenizerFast", + "DistilBertTokenizer" + ], + "processor_classes": [] }, "DonutSwinModel": { - "tokenizer_class": null, - "processor_class": "DonutFeatureExtractor" + "tokenizer_classes": [], + "processor_classes": [ + "DonutFeatureExtractor" + ] }, "DPTModel": { - "tokenizer_class": null, - "processor_class": "DPTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "DPTImageProcessor" + ] }, "ElectraModel": { - "tokenizer_class": "ElectraTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "ElectraTokenizerFast", + "ElectraTokenizer" + ], + "processor_classes": [] }, "ErnieModel": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "EsmModel": { - "tokenizer_class": "EsmTokenizer", - "processor_class": null + "tokenizer_classes": [ + "EsmTokenizer" + ], + "processor_classes": [] }, "FlaubertModel": { - "tokenizer_class": "FlaubertTokenizer", - "processor_class": null + "tokenizer_classes": [ + "FlaubertTokenizer" + ], + "processor_classes": [] }, "FlaubertWithLMHeadModel": { - "tokenizer_class": "FlaubertTokenizer", - "processor_class": null + "tokenizer_classes": [ + "FlaubertTokenizer" + ], + "processor_classes": [] }, "FlavaModel": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": "FlavaImageProcessor" + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [ + "FlavaImageProcessor" + ] }, "FNetModel": { - "tokenizer_class": "FNetTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "FNetTokenizerFast" + ], + "processor_classes": [] }, "FSMTModel": { - "tokenizer_class": "FSMTTokenizer", - "processor_class": null + "tokenizer_classes": [ + "FSMTTokenizer" + ], + "processor_classes": [] }, "FunnelBaseModel": { - "tokenizer_class": "FunnelTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "FunnelTokenizerFast", + "FunnelTokenizer" + ], + "processor_classes": [] }, "FunnelModel": { - "tokenizer_class": "FunnelTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "FunnelTokenizerFast", + "FunnelTokenizer" + ], + "processor_classes": [] }, "GLPNModel": { - "tokenizer_class": null, - "processor_class": "GLPNImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "GLPNImageProcessor" + ] }, "GPT2LMHeadModel": { - "tokenizer_class": "GPT2TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "GPT2TokenizerFast", + "GPT2Tokenizer" + ], + "processor_classes": [] }, "GPT2Model": { - "tokenizer_class": "GPT2TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "GPT2TokenizerFast", + "GPT2Tokenizer" + ], + "processor_classes": [] }, "GPTNeoModel": { - "tokenizer_class": "GPT2TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "GPT2TokenizerFast", + "GPT2Tokenizer" + ], + "processor_classes": [] }, "GPTNeoXModel": { - "tokenizer_class": "GPTNeoXTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "GPTNeoXTokenizerFast" + ], + "processor_classes": [] }, "GPTNeoXJapaneseModel": { - "tokenizer_class": "GPTNeoXJapaneseTokenizer", - "processor_class": null + "tokenizer_classes": [ + "GPTNeoXJapaneseTokenizer" + ], + "processor_classes": [] }, "GPTJModel": { - "tokenizer_class": "GPT2TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "GPT2TokenizerFast", + "GPT2Tokenizer" + ], + "processor_classes": [] }, "GroupViTModel": { - "tokenizer_class": "CLIPTokenizerFast", - "processor_class": "CLIPImageProcessor" + "tokenizer_classes": [ + "CLIPTokenizerFast", + "CLIPTokenizer" + ], + "processor_classes": [ + "CLIPImageProcessor" + ] }, "HubertModel": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": null + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [] }, "IBertModel": { - "tokenizer_class": "RobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RobertaTokenizerFast", + "RobertaTokenizer" + ], + "processor_classes": [] }, "ImageGPTModel": { - "tokenizer_class": null, - "processor_class": "ImageGPTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ImageGPTImageProcessor" + ] }, "LayoutLMModel": { - "tokenizer_class": "LayoutLMTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LayoutLMTokenizerFast", + "LayoutLMTokenizer" + ], + "processor_classes": [] }, "LayoutLMv2Model": { - "tokenizer_class": "LayoutLMv2TokenizerFast", - "processor_class": "LayoutLMv2ImageProcessor" + "tokenizer_classes": [ + "LayoutLMv2TokenizerFast", + "LayoutLMv2Tokenizer" + ], + "processor_classes": [ + "LayoutLMv2ImageProcessor" + ] }, "LayoutLMv3Model": { - "tokenizer_class": "LayoutLMv3TokenizerFast", - "processor_class": "LayoutLMv3ImageProcessor" + "tokenizer_classes": [ + "LayoutLMv3TokenizerFast", + "LayoutLMv3Tokenizer" + ], + "processor_classes": [ + "LayoutLMv3ImageProcessor" + ] }, "LEDModel": { - "tokenizer_class": "LEDTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LEDTokenizerFast", + "LEDTokenizer" + ], + "processor_classes": [] }, "LevitModel": { - "tokenizer_class": null, - "processor_class": "LevitImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "LevitImageProcessor" + ] }, "LiltModel": { - "tokenizer_class": "LayoutLMv3TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LayoutLMv3TokenizerFast", + "LayoutLMv3Tokenizer" + ], + "processor_classes": [] }, "LongformerModel": { - "tokenizer_class": "LongformerTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LongformerTokenizerFast", + "LongformerTokenizer" + ], + "processor_classes": [] }, "LongT5Model": { - "tokenizer_class": "T5TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "T5TokenizerFast" + ], + "processor_classes": [] }, "LukeModel": { - "tokenizer_class": "LukeTokenizer", - "processor_class": null + "tokenizer_classes": [ + "LukeTokenizer" + ], + "processor_classes": [] }, "LxmertModel": { - "tokenizer_class": "LxmertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LxmertTokenizerFast", + "LxmertTokenizer" + ], + "processor_classes": [] }, "M2M100Model": { - "tokenizer_class": "M2M100Tokenizer", - "processor_class": null + "tokenizer_classes": [ + "M2M100Tokenizer" + ], + "processor_classes": [] }, "MarianMTModel": { - "tokenizer_class": "MarianTokenizer", - "processor_class": null + "tokenizer_classes": [ + "MarianTokenizer" + ], + "processor_classes": [] }, "MarianModel": { - "tokenizer_class": "MarianTokenizer", - "processor_class": null + "tokenizer_classes": [ + "MarianTokenizer" + ], + "processor_classes": [] }, "MarkupLMModel": { - "tokenizer_class": "MarkupLMTokenizerFast", - "processor_class": "MarkupLMFeatureExtractor" + "tokenizer_classes": [ + "MarkupLMTokenizerFast", + "MarkupLMTokenizer" + ], + "processor_classes": [ + "MarkupLMFeatureExtractor" + ] }, "MaskFormerModel": { - "tokenizer_class": null, - "processor_class": "MaskFormerFeatureExtractor" + "tokenizer_classes": [], + "processor_classes": [ + "MaskFormerFeatureExtractor" + ] }, "MBartModel": { - "tokenizer_class": "MBartTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MBartTokenizerFast", + "MBartTokenizer" + ], + "processor_classes": [] }, "MCTCTModel": { - "tokenizer_class": null, - "processor_class": "MCTCTFeatureExtractor" + "tokenizer_classes": [], + "processor_classes": [ + "MCTCTFeatureExtractor" + ] }, "MegatronBertModel": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "MobileBertModel": { - "tokenizer_class": "MobileBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MobileBertTokenizerFast", + "MobileBertTokenizer" + ], + "processor_classes": [] }, "MobileViTModel": { - "tokenizer_class": null, - "processor_class": "MobileViTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "MobileViTImageProcessor" + ] }, "MPNetModel": { - "tokenizer_class": "MPNetTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MPNetTokenizerFast", + "MPNetTokenizer" + ], + "processor_classes": [] }, "MvpModel": { - "tokenizer_class": "MvpTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MvpTokenizerFast", + "MvpTokenizer" + ], + "processor_classes": [] }, "NezhaModel": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "NystromformerModel": { - "tokenizer_class": "AlbertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "AlbertTokenizerFast" + ], + "processor_classes": [] }, "OpenAIGPTLMHeadModel": { - "tokenizer_class": "OpenAIGPTTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "OpenAIGPTTokenizerFast", + "OpenAIGPTTokenizer" + ], + "processor_classes": [] }, "OpenAIGPTModel": { - "tokenizer_class": "OpenAIGPTTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "OpenAIGPTTokenizerFast", + "OpenAIGPTTokenizer" + ], + "processor_classes": [] }, "OPTModel": { - "tokenizer_class": "GPT2TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "GPT2TokenizerFast", + "GPT2Tokenizer" + ], + "processor_classes": [] }, "OwlViTModel": { - "tokenizer_class": "CLIPTokenizerFast", - "processor_class": "OwlViTFeatureExtractor" + "tokenizer_classes": [ + "CLIPTokenizerFast", + "CLIPTokenizer" + ], + "processor_classes": [ + "OwlViTFeatureExtractor" + ] }, "PegasusModel": { - "tokenizer_class": "PegasusTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "PegasusTokenizerFast" + ], + "processor_classes": [] }, "PegasusXModel": { - "tokenizer_class": "PegasusTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "PegasusTokenizerFast" + ], + "processor_classes": [] }, "PerceiverModel": { - "tokenizer_class": "PerceiverTokenizer", - "processor_class": null + "tokenizer_classes": [ + "PerceiverTokenizer" + ], + "processor_classes": [] }, "PLBartModel": { - "tokenizer_class": "PLBartTokenizer", - "processor_class": null + "tokenizer_classes": [ + "PLBartTokenizer" + ], + "processor_classes": [] }, "PoolFormerModel": { - "tokenizer_class": null, - "processor_class": "PoolFormerImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "PoolFormerImageProcessor" + ] }, "ProphetNetModel": { - "tokenizer_class": "ProphetNetTokenizer", - "processor_class": null + "tokenizer_classes": [ + "ProphetNetTokenizer" + ], + "processor_classes": [] }, "ReformerModel": { - "tokenizer_class": "ReformerTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "ReformerTokenizerFast", + "ReformerTokenizer" + ], + "processor_classes": [] }, "RegNetModel": { - "tokenizer_class": null, - "processor_class": "ConvNextImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ConvNextImageProcessor" + ] }, "RemBertModel": { - "tokenizer_class": "RemBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RemBertTokenizerFast" + ], + "processor_classes": [] }, "ResNetModel": { - "tokenizer_class": null, - "processor_class": "ConvNextImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ConvNextImageProcessor" + ] }, "RobertaModel": { - "tokenizer_class": "RobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RobertaTokenizerFast", + "RobertaTokenizer" + ], + "processor_classes": [] }, "RoFormerModel": { - "tokenizer_class": "RoFormerTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RoFormerTokenizerFast", + "RoFormerTokenizer" + ], + "processor_classes": [] }, "SegformerModel": { - "tokenizer_class": null, - "processor_class": "SegformerImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "SegformerImageProcessor" + ] }, "SEWDModel": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "SEWModel": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "Speech2TextModel": { - "tokenizer_class": "Speech2TextTokenizer", - "processor_class": "Speech2TextFeatureExtractor" + "tokenizer_classes": [ + "Speech2TextTokenizer" + ], + "processor_classes": [ + "Speech2TextFeatureExtractor" + ] }, "SplinterModel": { - "tokenizer_class": null, - "processor_class": null + "tokenizer_classes": [], + "processor_classes": [] }, "SqueezeBertModel": { - "tokenizer_class": "SqueezeBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "SqueezeBertTokenizerFast", + "SqueezeBertTokenizer" + ], + "processor_classes": [] }, "SwinModel": { - "tokenizer_class": null, - "processor_class": "ViTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ViTImageProcessor" + ] }, "Swinv2Model": { - "tokenizer_class": null, - "processor_class": "ViTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ViTImageProcessor" + ] }, "T5Model": { - "tokenizer_class": "T5TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "T5TokenizerFast" + ], + "processor_classes": [] }, "TableTransformerModel": { - "tokenizer_class": null, - "processor_class": "DetrFeatureExtractor" + "tokenizer_classes": [], + "processor_classes": [ + "DetrFeatureExtractor" + ] }, "TapasModel": { - "tokenizer_class": "TapasTokenizer", - "processor_class": null + "tokenizer_classes": [ + "TapasTokenizer" + ], + "processor_classes": [] }, "TransfoXLLMHeadModel": { - "tokenizer_class": "TransfoXLTokenizer", - "processor_class": null + "tokenizer_classes": [ + "TransfoXLTokenizer" + ], + "processor_classes": [] }, "TransfoXLModel": { - "tokenizer_class": "TransfoXLTokenizer", - "processor_class": null + "tokenizer_classes": [ + "TransfoXLTokenizer" + ], + "processor_classes": [] }, "UniSpeechSatModel": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "UniSpeechModel": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "VanModel": { - "tokenizer_class": null, - "processor_class": "ConvNextImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ConvNextImageProcessor" + ] }, "VideoMAEModel": { - "tokenizer_class": null, - "processor_class": "VideoMAEImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "VideoMAEImageProcessor" + ] }, "ViltModel": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": "ViltImageProcessor" + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [ + "ViltImageProcessor" + ] }, "VisualBertModel": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "ViTModel": { - "tokenizer_class": null, - "processor_class": "ViTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ViTImageProcessor" + ] }, "ViTMAEModel": { - "tokenizer_class": null, - "processor_class": "ViTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ViTImageProcessor" + ] }, "ViTMSNModel": { - "tokenizer_class": null, - "processor_class": "ViTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ViTImageProcessor" + ] }, "Wav2Vec2ConformerModel": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "Wav2Vec2Model": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "WavLMModel": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "WhisperModel": { - "tokenizer_class": "WhisperTokenizer", - "processor_class": "WhisperFeatureExtractor" + "tokenizer_classes": [ + "WhisperTokenizer" + ], + "processor_classes": [ + "WhisperFeatureExtractor" + ] }, "XCLIPModel": { - "tokenizer_class": "CLIPTokenizerFast", - "processor_class": "VideoMAEImageProcessor" + "tokenizer_classes": [ + "CLIPTokenizerFast", + "CLIPTokenizer" + ], + "processor_classes": [ + "VideoMAEImageProcessor", + "CLIPImageProcessor" + ] }, "XGLMModel": { - "tokenizer_class": "XGLMTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "XGLMTokenizerFast" + ], + "processor_classes": [] }, "XLMRobertaXLModel": { - "tokenizer_class": "XLMRobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "XLMRobertaTokenizerFast" + ], + "processor_classes": [] }, "XLMModel": { - "tokenizer_class": "XLMTokenizer", - "processor_class": null + "tokenizer_classes": [ + "XLMTokenizer" + ], + "processor_classes": [] }, "XLMWithLMHeadModel": { - "tokenizer_class": "XLMTokenizer", - "processor_class": null + "tokenizer_classes": [ + "XLMTokenizer" + ], + "processor_classes": [] }, "XLNetLMHeadModel": { - "tokenizer_class": "XLNetTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "XLNetTokenizerFast", + "XLNetTokenizer" + ], + "processor_classes": [] }, "XLNetModel": { - "tokenizer_class": "XLNetTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "XLNetTokenizerFast", + "XLNetTokenizer" + ], + "processor_classes": [] }, "YolosModel": { - "tokenizer_class": null, - "processor_class": "YolosFeatureExtractor" + "tokenizer_classes": [], + "processor_classes": [ + "YolosFeatureExtractor" + ] }, "YosoModel": { - "tokenizer_class": "AlbertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "AlbertTokenizerFast" + ], + "processor_classes": [] }, "AlbertForMaskedLM": { - "tokenizer_class": "AlbertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "AlbertTokenizerFast" + ], + "processor_classes": [] }, "AlbertForMultipleChoice": { - "tokenizer_class": "AlbertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "AlbertTokenizerFast" + ], + "processor_classes": [] }, "AlbertForPreTraining": { - "tokenizer_class": "AlbertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "AlbertTokenizerFast" + ], + "processor_classes": [] }, "AlbertForQuestionAnswering": { - "tokenizer_class": "AlbertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "AlbertTokenizerFast" + ], + "processor_classes": [] }, "AlbertForSequenceClassification": { - "tokenizer_class": "AlbertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "AlbertTokenizerFast" + ], + "processor_classes": [] }, "AlbertForTokenClassification": { - "tokenizer_class": "AlbertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "AlbertTokenizerFast" + ], + "processor_classes": [] }, "BartForCausalLM": { - "tokenizer_class": "BartTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BartTokenizerFast", + "BartTokenizer" + ], + "processor_classes": [] }, "BartForConditionalGeneration": { - "tokenizer_class": "BartTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BartTokenizerFast", + "BartTokenizer" + ], + "processor_classes": [] }, "BartForQuestionAnswering": { - "tokenizer_class": "BartTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BartTokenizerFast", + "BartTokenizer" + ], + "processor_classes": [] }, "BartForSequenceClassification": { - "tokenizer_class": "BartTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BartTokenizerFast", + "BartTokenizer" + ], + "processor_classes": [] }, "BeitForImageClassification": { - "tokenizer_class": null, - "processor_class": "BeitImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "BeitImageProcessor" + ] }, "BeitForSemanticSegmentation": { - "tokenizer_class": null, - "processor_class": "BeitImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "BeitImageProcessor" + ] }, "BertForMaskedLM": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "BertForMultipleChoice": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "BertForNextSentencePrediction": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "BertForPreTraining": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "BertForQuestionAnswering": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "BertForSequenceClassification": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "BertForTokenClassification": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "BigBirdForCausalLM": { - "tokenizer_class": "BigBirdTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BigBirdTokenizerFast" + ], + "processor_classes": [] }, "BigBirdForMaskedLM": { - "tokenizer_class": "BigBirdTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BigBirdTokenizerFast" + ], + "processor_classes": [] }, "BigBirdForMultipleChoice": { - "tokenizer_class": "BigBirdTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BigBirdTokenizerFast" + ], + "processor_classes": [] }, "BigBirdForPreTraining": { - "tokenizer_class": "BigBirdTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BigBirdTokenizerFast" + ], + "processor_classes": [] }, "BigBirdForQuestionAnswering": { - "tokenizer_class": "BigBirdTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BigBirdTokenizerFast" + ], + "processor_classes": [] }, "BigBirdForSequenceClassification": { - "tokenizer_class": "BigBirdTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BigBirdTokenizerFast" + ], + "processor_classes": [] }, "BigBirdForTokenClassification": { - "tokenizer_class": "BigBirdTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BigBirdTokenizerFast" + ], + "processor_classes": [] }, "BigBirdPegasusForCausalLM": { - "tokenizer_class": "PegasusTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "PegasusTokenizerFast" + ], + "processor_classes": [] }, "BigBirdPegasusForConditionalGeneration": { - "tokenizer_class": "PegasusTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "PegasusTokenizerFast" + ], + "processor_classes": [] }, "BigBirdPegasusForQuestionAnswering": { - "tokenizer_class": "PegasusTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "PegasusTokenizerFast" + ], + "processor_classes": [] }, "BigBirdPegasusForSequenceClassification": { - "tokenizer_class": "PegasusTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "PegasusTokenizerFast" + ], + "processor_classes": [] }, "BlenderbotForCausalLM": { - "tokenizer_class": "BlenderbotTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BlenderbotTokenizerFast", + "BlenderbotTokenizer" + ], + "processor_classes": [] }, "BlenderbotForConditionalGeneration": { - "tokenizer_class": "BlenderbotTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BlenderbotTokenizerFast", + "BlenderbotTokenizer" + ], + "processor_classes": [] }, "BlenderbotSmallForCausalLM": { - "tokenizer_class": null, - "processor_class": null + "tokenizer_classes": [], + "processor_classes": [] }, "BlenderbotSmallForConditionalGeneration": { - "tokenizer_class": null, - "processor_class": null + "tokenizer_classes": [], + "processor_classes": [] }, "BloomForCausalLM": { - "tokenizer_class": "BloomTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BloomTokenizerFast" + ], + "processor_classes": [] }, "BloomForQuestionAnswering": { - "tokenizer_class": "BloomTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BloomTokenizerFast" + ], + "processor_classes": [] }, "BloomForSequenceClassification": { - "tokenizer_class": "BloomTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BloomTokenizerFast" + ], + "processor_classes": [] }, "BloomForTokenClassification": { - "tokenizer_class": "BloomTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BloomTokenizerFast" + ], + "processor_classes": [] }, "CTRLForSequenceClassification": { - "tokenizer_class": "CTRLTokenizer", - "processor_class": null + "tokenizer_classes": [ + "CTRLTokenizer" + ], + "processor_classes": [] }, "CanineForMultipleChoice": { - "tokenizer_class": "CanineTokenizer", - "processor_class": null + "tokenizer_classes": [ + "CanineTokenizer" + ], + "processor_classes": [] }, "CanineForQuestionAnswering": { - "tokenizer_class": "CanineTokenizer", - "processor_class": null + "tokenizer_classes": [ + "CanineTokenizer" + ], + "processor_classes": [] }, "CanineForSequenceClassification": { - "tokenizer_class": "CanineTokenizer", - "processor_class": null + "tokenizer_classes": [ + "CanineTokenizer" + ], + "processor_classes": [] }, "CanineForTokenClassification": { - "tokenizer_class": "CanineTokenizer", - "processor_class": null + "tokenizer_classes": [ + "CanineTokenizer" + ], + "processor_classes": [] }, "CodeGenForCausalLM": { - "tokenizer_class": "CodeGenTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "CodeGenTokenizerFast", + "CodeGenTokenizer" + ], + "processor_classes": [] }, "ConditionalDetrForObjectDetection": { - "tokenizer_class": null, - "processor_class": "ConditionalDetrFeatureExtractor" + "tokenizer_classes": [], + "processor_classes": [ + "ConditionalDetrFeatureExtractor" + ] }, "ConvBertForMaskedLM": { - "tokenizer_class": "ConvBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "ConvBertTokenizerFast", + "ConvBertTokenizer" + ], + "processor_classes": [] }, "ConvBertForMultipleChoice": { - "tokenizer_class": "ConvBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "ConvBertTokenizerFast", + "ConvBertTokenizer" + ], + "processor_classes": [] }, "ConvBertForQuestionAnswering": { - "tokenizer_class": "ConvBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "ConvBertTokenizerFast", + "ConvBertTokenizer" + ], + "processor_classes": [] }, "ConvBertForSequenceClassification": { - "tokenizer_class": "ConvBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "ConvBertTokenizerFast", + "ConvBertTokenizer" + ], + "processor_classes": [] }, "ConvBertForTokenClassification": { - "tokenizer_class": "ConvBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "ConvBertTokenizerFast", + "ConvBertTokenizer" + ], + "processor_classes": [] }, "ConvNextForImageClassification": { - "tokenizer_class": null, - "processor_class": "ConvNextImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ConvNextImageProcessor" + ] }, "CvtForImageClassification": { - "tokenizer_class": null, - "processor_class": "ConvNextImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ConvNextImageProcessor" + ] }, "DPTForDepthEstimation": { - "tokenizer_class": null, - "processor_class": "DPTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "DPTImageProcessor" + ] }, "DPTForSemanticSegmentation": { - "tokenizer_class": null, - "processor_class": "DPTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "DPTImageProcessor" + ] }, "Data2VecAudioForCTC": { - "tokenizer_class": null, - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "Data2VecAudioForSequenceClassification": { - "tokenizer_class": null, - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "Data2VecAudioForXVector": { - "tokenizer_class": null, - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "Data2VecTextForCausalLM": { - "tokenizer_class": "RobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RobertaTokenizerFast", + "RobertaTokenizer" + ], + "processor_classes": [] }, "Data2VecTextForMaskedLM": { - "tokenizer_class": "RobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RobertaTokenizerFast", + "RobertaTokenizer" + ], + "processor_classes": [] }, "Data2VecTextForMultipleChoice": { - "tokenizer_class": "RobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RobertaTokenizerFast", + "RobertaTokenizer" + ], + "processor_classes": [] }, "Data2VecTextForQuestionAnswering": { - "tokenizer_class": "RobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RobertaTokenizerFast", + "RobertaTokenizer" + ], + "processor_classes": [] }, "Data2VecTextForSequenceClassification": { - "tokenizer_class": "RobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RobertaTokenizerFast", + "RobertaTokenizer" + ], + "processor_classes": [] }, "Data2VecTextForTokenClassification": { - "tokenizer_class": "RobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RobertaTokenizerFast", + "RobertaTokenizer" + ], + "processor_classes": [] }, "Data2VecVisionForImageClassification": { - "tokenizer_class": null, - "processor_class": "BeitImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "BeitImageProcessor" + ] }, "Data2VecVisionForSemanticSegmentation": { - "tokenizer_class": null, - "processor_class": "BeitImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "BeitImageProcessor" + ] }, "DebertaForMaskedLM": { - "tokenizer_class": "DebertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "DebertaTokenizerFast", + "DebertaTokenizer" + ], + "processor_classes": [] }, "DebertaForQuestionAnswering": { - "tokenizer_class": "DebertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "DebertaTokenizerFast", + "DebertaTokenizer" + ], + "processor_classes": [] }, "DebertaForSequenceClassification": { - "tokenizer_class": "DebertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "DebertaTokenizerFast", + "DebertaTokenizer" + ], + "processor_classes": [] }, "DebertaForTokenClassification": { - "tokenizer_class": "DebertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "DebertaTokenizerFast", + "DebertaTokenizer" + ], + "processor_classes": [] }, "DebertaV2ForMaskedLM": { - "tokenizer_class": "DebertaV2TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "DebertaV2TokenizerFast" + ], + "processor_classes": [] }, "DebertaV2ForMultipleChoice": { - "tokenizer_class": "DebertaV2TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "DebertaV2TokenizerFast" + ], + "processor_classes": [] }, "DebertaV2ForQuestionAnswering": { - "tokenizer_class": "DebertaV2TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "DebertaV2TokenizerFast" + ], + "processor_classes": [] }, "DebertaV2ForSequenceClassification": { - "tokenizer_class": "DebertaV2TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "DebertaV2TokenizerFast" + ], + "processor_classes": [] }, "DebertaV2ForTokenClassification": { - "tokenizer_class": "DebertaV2TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "DebertaV2TokenizerFast" + ], + "processor_classes": [] }, "DeformableDetrForObjectDetection": { - "tokenizer_class": null, - "processor_class": "DeformableDetrFeatureExtractor" + "tokenizer_classes": [], + "processor_classes": [ + "DeformableDetrFeatureExtractor" + ] }, "DeiTForImageClassification": { - "tokenizer_class": null, - "processor_class": "DeiTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "DeiTImageProcessor" + ] }, "DeiTForImageClassificationWithTeacher": { - "tokenizer_class": null, - "processor_class": "DeiTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "DeiTImageProcessor" + ] }, "DeiTForMaskedImageModeling": { - "tokenizer_class": null, - "processor_class": "DeiTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "DeiTImageProcessor" + ] }, "DetrForObjectDetection": { - "tokenizer_class": null, - "processor_class": "DetrFeatureExtractor" + "tokenizer_classes": [], + "processor_classes": [ + "DetrFeatureExtractor" + ] }, "DetrForSegmentation": { - "tokenizer_class": null, - "processor_class": "DetrFeatureExtractor" + "tokenizer_classes": [], + "processor_classes": [ + "DetrFeatureExtractor" + ] }, "DistilBertForMaskedLM": { - "tokenizer_class": "DistilBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "DistilBertTokenizerFast", + "DistilBertTokenizer" + ], + "processor_classes": [] }, "DistilBertForMultipleChoice": { - "tokenizer_class": "DistilBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "DistilBertTokenizerFast", + "DistilBertTokenizer" + ], + "processor_classes": [] }, "DistilBertForQuestionAnswering": { - "tokenizer_class": "DistilBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "DistilBertTokenizerFast", + "DistilBertTokenizer" + ], + "processor_classes": [] }, "DistilBertForSequenceClassification": { - "tokenizer_class": "DistilBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "DistilBertTokenizerFast", + "DistilBertTokenizer" + ], + "processor_classes": [] }, "DistilBertForTokenClassification": { - "tokenizer_class": "DistilBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "DistilBertTokenizerFast", + "DistilBertTokenizer" + ], + "processor_classes": [] }, "ElectraForCausalLM": { - "tokenizer_class": "ElectraTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "ElectraTokenizerFast", + "ElectraTokenizer" + ], + "processor_classes": [] }, "ElectraForMaskedLM": { - "tokenizer_class": "ElectraTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "ElectraTokenizerFast", + "ElectraTokenizer" + ], + "processor_classes": [] }, "ElectraForMultipleChoice": { - "tokenizer_class": "ElectraTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "ElectraTokenizerFast", + "ElectraTokenizer" + ], + "processor_classes": [] }, "ElectraForPreTraining": { - "tokenizer_class": "ElectraTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "ElectraTokenizerFast", + "ElectraTokenizer" + ], + "processor_classes": [] }, "ElectraForQuestionAnswering": { - "tokenizer_class": "ElectraTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "ElectraTokenizerFast", + "ElectraTokenizer" + ], + "processor_classes": [] }, "ElectraForSequenceClassification": { - "tokenizer_class": "ElectraTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "ElectraTokenizerFast", + "ElectraTokenizer" + ], + "processor_classes": [] }, "ElectraForTokenClassification": { - "tokenizer_class": "ElectraTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "ElectraTokenizerFast", + "ElectraTokenizer" + ], + "processor_classes": [] }, "ErnieForCausalLM": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "ErnieForMaskedLM": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "ErnieForMultipleChoice": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "ErnieForNextSentencePrediction": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "ErnieForPreTraining": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "ErnieForQuestionAnswering": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "ErnieForSequenceClassification": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "ErnieForTokenClassification": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "EsmForMaskedLM": { - "tokenizer_class": "EsmTokenizer", - "processor_class": null + "tokenizer_classes": [ + "EsmTokenizer" + ], + "processor_classes": [] }, "EsmForSequenceClassification": { - "tokenizer_class": "EsmTokenizer", - "processor_class": null + "tokenizer_classes": [ + "EsmTokenizer" + ], + "processor_classes": [] }, "EsmForTokenClassification": { - "tokenizer_class": "EsmTokenizer", - "processor_class": null + "tokenizer_classes": [ + "EsmTokenizer" + ], + "processor_classes": [] }, "FNetForMaskedLM": { - "tokenizer_class": "FNetTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "FNetTokenizerFast" + ], + "processor_classes": [] }, "FNetForMultipleChoice": { - "tokenizer_class": "FNetTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "FNetTokenizerFast" + ], + "processor_classes": [] }, "FNetForNextSentencePrediction": { - "tokenizer_class": "FNetTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "FNetTokenizerFast" + ], + "processor_classes": [] }, "FNetForPreTraining": { - "tokenizer_class": "FNetTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "FNetTokenizerFast" + ], + "processor_classes": [] }, "FNetForQuestionAnswering": { - "tokenizer_class": "FNetTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "FNetTokenizerFast" + ], + "processor_classes": [] }, "FNetForSequenceClassification": { - "tokenizer_class": "FNetTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "FNetTokenizerFast" + ], + "processor_classes": [] }, "FNetForTokenClassification": { - "tokenizer_class": "FNetTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "FNetTokenizerFast" + ], + "processor_classes": [] }, "FSMTForConditionalGeneration": { - "tokenizer_class": "FSMTTokenizer", - "processor_class": null + "tokenizer_classes": [ + "FSMTTokenizer" + ], + "processor_classes": [] }, "FlaubertForMultipleChoice": { - "tokenizer_class": "FlaubertTokenizer", - "processor_class": null + "tokenizer_classes": [ + "FlaubertTokenizer" + ], + "processor_classes": [] }, "FlaubertForQuestionAnsweringSimple": { - "tokenizer_class": "FlaubertTokenizer", - "processor_class": null + "tokenizer_classes": [ + "FlaubertTokenizer" + ], + "processor_classes": [] }, "FlaubertForSequenceClassification": { - "tokenizer_class": "FlaubertTokenizer", - "processor_class": null + "tokenizer_classes": [ + "FlaubertTokenizer" + ], + "processor_classes": [] }, "FlaubertForTokenClassification": { - "tokenizer_class": "FlaubertTokenizer", - "processor_class": null + "tokenizer_classes": [ + "FlaubertTokenizer" + ], + "processor_classes": [] }, "FlavaForPreTraining": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": "FlavaImageProcessor" + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [ + "FlavaImageProcessor" + ] }, "FunnelForMaskedLM": { - "tokenizer_class": "FunnelTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "FunnelTokenizerFast", + "FunnelTokenizer" + ], + "processor_classes": [] }, "FunnelForMultipleChoice": { - "tokenizer_class": "FunnelTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "FunnelTokenizerFast", + "FunnelTokenizer" + ], + "processor_classes": [] }, "FunnelForPreTraining": { - "tokenizer_class": "FunnelTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "FunnelTokenizerFast", + "FunnelTokenizer" + ], + "processor_classes": [] }, "FunnelForQuestionAnswering": { - "tokenizer_class": "FunnelTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "FunnelTokenizerFast", + "FunnelTokenizer" + ], + "processor_classes": [] }, "FunnelForSequenceClassification": { - "tokenizer_class": "FunnelTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "FunnelTokenizerFast", + "FunnelTokenizer" + ], + "processor_classes": [] }, "FunnelForTokenClassification": { - "tokenizer_class": "FunnelTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "FunnelTokenizerFast", + "FunnelTokenizer" + ], + "processor_classes": [] }, "GLPNForDepthEstimation": { - "tokenizer_class": null, - "processor_class": "GLPNImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "GLPNImageProcessor" + ] }, "GPT2ForSequenceClassification": { - "tokenizer_class": "GPT2TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "GPT2TokenizerFast", + "GPT2Tokenizer" + ], + "processor_classes": [] }, "GPT2ForTokenClassification": { - "tokenizer_class": "GPT2TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "GPT2TokenizerFast", + "GPT2Tokenizer" + ], + "processor_classes": [] }, "GPTJForCausalLM": { - "tokenizer_class": "GPT2TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "GPT2TokenizerFast", + "GPT2Tokenizer" + ], + "processor_classes": [] }, "GPTJForQuestionAnswering": { - "tokenizer_class": "GPT2TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "GPT2TokenizerFast", + "GPT2Tokenizer" + ], + "processor_classes": [] }, "GPTJForSequenceClassification": { - "tokenizer_class": "GPT2TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "GPT2TokenizerFast", + "GPT2Tokenizer" + ], + "processor_classes": [] }, "GPTNeoForCausalLM": { - "tokenizer_class": "GPT2TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "GPT2TokenizerFast", + "GPT2Tokenizer" + ], + "processor_classes": [] }, "GPTNeoForSequenceClassification": { - "tokenizer_class": "GPT2TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "GPT2TokenizerFast", + "GPT2Tokenizer" + ], + "processor_classes": [] }, "GPTNeoXForCausalLM": { - "tokenizer_class": "GPTNeoXTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "GPTNeoXTokenizerFast" + ], + "processor_classes": [] }, "GPTNeoXJapaneseForCausalLM": { - "tokenizer_class": "GPTNeoXJapaneseTokenizer", - "processor_class": null + "tokenizer_classes": [ + "GPTNeoXJapaneseTokenizer" + ], + "processor_classes": [] }, "HubertForCTC": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": null + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [] }, "HubertForSequenceClassification": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": null + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [] }, "IBertForMaskedLM": { - "tokenizer_class": "RobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RobertaTokenizerFast", + "RobertaTokenizer" + ], + "processor_classes": [] }, "IBertForMultipleChoice": { - "tokenizer_class": "RobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RobertaTokenizerFast", + "RobertaTokenizer" + ], + "processor_classes": [] }, "IBertForQuestionAnswering": { - "tokenizer_class": "RobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RobertaTokenizerFast", + "RobertaTokenizer" + ], + "processor_classes": [] }, "IBertForSequenceClassification": { - "tokenizer_class": "RobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RobertaTokenizerFast", + "RobertaTokenizer" + ], + "processor_classes": [] }, "IBertForTokenClassification": { - "tokenizer_class": "RobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RobertaTokenizerFast", + "RobertaTokenizer" + ], + "processor_classes": [] }, "ImageGPTForCausalImageModeling": { - "tokenizer_class": null, - "processor_class": "ImageGPTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ImageGPTImageProcessor" + ] }, "ImageGPTForImageClassification": { - "tokenizer_class": null, - "processor_class": "ImageGPTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ImageGPTImageProcessor" + ] }, "LEDForConditionalGeneration": { - "tokenizer_class": "LEDTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LEDTokenizerFast", + "LEDTokenizer" + ], + "processor_classes": [] }, "LEDForQuestionAnswering": { - "tokenizer_class": "LEDTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LEDTokenizerFast", + "LEDTokenizer" + ], + "processor_classes": [] }, "LEDForSequenceClassification": { - "tokenizer_class": "LEDTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LEDTokenizerFast", + "LEDTokenizer" + ], + "processor_classes": [] }, "LayoutLMForMaskedLM": { - "tokenizer_class": "LayoutLMTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LayoutLMTokenizerFast", + "LayoutLMTokenizer" + ], + "processor_classes": [] }, "LayoutLMForQuestionAnswering": { - "tokenizer_class": "LayoutLMTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LayoutLMTokenizerFast", + "LayoutLMTokenizer" + ], + "processor_classes": [] }, "LayoutLMForSequenceClassification": { - "tokenizer_class": "LayoutLMTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LayoutLMTokenizerFast", + "LayoutLMTokenizer" + ], + "processor_classes": [] }, "LayoutLMForTokenClassification": { - "tokenizer_class": "LayoutLMTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LayoutLMTokenizerFast", + "LayoutLMTokenizer" + ], + "processor_classes": [] }, "LayoutLMv2ForQuestionAnswering": { - "tokenizer_class": "LayoutLMv2TokenizerFast", - "processor_class": "LayoutLMv2ImageProcessor" + "tokenizer_classes": [ + "LayoutLMv2TokenizerFast", + "LayoutLMv2Tokenizer" + ], + "processor_classes": [ + "LayoutLMv2ImageProcessor" + ] }, "LayoutLMv2ForSequenceClassification": { - "tokenizer_class": "LayoutLMv2TokenizerFast", - "processor_class": "LayoutLMv2ImageProcessor" + "tokenizer_classes": [ + "LayoutLMv2TokenizerFast", + "LayoutLMv2Tokenizer" + ], + "processor_classes": [ + "LayoutLMv2ImageProcessor" + ] }, "LayoutLMv2ForTokenClassification": { - "tokenizer_class": "LayoutLMv2TokenizerFast", - "processor_class": "LayoutLMv2ImageProcessor" + "tokenizer_classes": [ + "LayoutLMv2TokenizerFast", + "LayoutLMv2Tokenizer" + ], + "processor_classes": [ + "LayoutLMv2ImageProcessor" + ] }, "LayoutLMv3ForQuestionAnswering": { - "tokenizer_class": "LayoutLMv3TokenizerFast", - "processor_class": "LayoutLMv3ImageProcessor" + "tokenizer_classes": [ + "LayoutLMv3TokenizerFast", + "LayoutLMv3Tokenizer" + ], + "processor_classes": [ + "LayoutLMv3ImageProcessor" + ] }, "LayoutLMv3ForSequenceClassification": { - "tokenizer_class": "LayoutLMv3TokenizerFast", - "processor_class": "LayoutLMv3ImageProcessor" + "tokenizer_classes": [ + "LayoutLMv3TokenizerFast", + "LayoutLMv3Tokenizer" + ], + "processor_classes": [ + "LayoutLMv3ImageProcessor" + ] }, "LayoutLMv3ForTokenClassification": { - "tokenizer_class": "LayoutLMv3TokenizerFast", - "processor_class": "LayoutLMv3ImageProcessor" + "tokenizer_classes": [ + "LayoutLMv3TokenizerFast", + "LayoutLMv3Tokenizer" + ], + "processor_classes": [ + "LayoutLMv3ImageProcessor" + ] }, "LevitForImageClassification": { - "tokenizer_class": null, - "processor_class": "LevitImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "LevitImageProcessor" + ] }, "LevitForImageClassificationWithTeacher": { - "tokenizer_class": null, - "processor_class": "LevitImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "LevitImageProcessor" + ] }, "LiltForQuestionAnswering": { - "tokenizer_class": "LayoutLMv3TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LayoutLMv3TokenizerFast", + "LayoutLMv3Tokenizer" + ], + "processor_classes": [] }, "LiltForSequenceClassification": { - "tokenizer_class": "LayoutLMv3TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LayoutLMv3TokenizerFast", + "LayoutLMv3Tokenizer" + ], + "processor_classes": [] }, "LiltForTokenClassification": { - "tokenizer_class": "LayoutLMv3TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LayoutLMv3TokenizerFast", + "LayoutLMv3Tokenizer" + ], + "processor_classes": [] }, "LongT5ForConditionalGeneration": { - "tokenizer_class": "T5TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "T5TokenizerFast" + ], + "processor_classes": [] }, "LongformerForMaskedLM": { - "tokenizer_class": "LongformerTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LongformerTokenizerFast", + "LongformerTokenizer" + ], + "processor_classes": [] }, "LongformerForMultipleChoice": { - "tokenizer_class": "LongformerTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LongformerTokenizerFast", + "LongformerTokenizer" + ], + "processor_classes": [] }, "LongformerForQuestionAnswering": { - "tokenizer_class": "LongformerTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LongformerTokenizerFast", + "LongformerTokenizer" + ], + "processor_classes": [] }, "LongformerForSequenceClassification": { - "tokenizer_class": "LongformerTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LongformerTokenizerFast", + "LongformerTokenizer" + ], + "processor_classes": [] }, "LongformerForTokenClassification": { - "tokenizer_class": "LongformerTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LongformerTokenizerFast", + "LongformerTokenizer" + ], + "processor_classes": [] }, "LukeForMaskedLM": { - "tokenizer_class": "LukeTokenizer", - "processor_class": null + "tokenizer_classes": [ + "LukeTokenizer" + ], + "processor_classes": [] }, "LukeForMultipleChoice": { - "tokenizer_class": "LukeTokenizer", - "processor_class": null + "tokenizer_classes": [ + "LukeTokenizer" + ], + "processor_classes": [] }, "LukeForQuestionAnswering": { - "tokenizer_class": "LukeTokenizer", - "processor_class": null + "tokenizer_classes": [ + "LukeTokenizer" + ], + "processor_classes": [] }, "LukeForSequenceClassification": { - "tokenizer_class": "LukeTokenizer", - "processor_class": null + "tokenizer_classes": [ + "LukeTokenizer" + ], + "processor_classes": [] }, "LukeForTokenClassification": { - "tokenizer_class": "LukeTokenizer", - "processor_class": null + "tokenizer_classes": [ + "LukeTokenizer" + ], + "processor_classes": [] }, "LxmertForPreTraining": { - "tokenizer_class": "LxmertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LxmertTokenizerFast", + "LxmertTokenizer" + ], + "processor_classes": [] }, "LxmertForQuestionAnswering": { - "tokenizer_class": "LxmertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "LxmertTokenizerFast", + "LxmertTokenizer" + ], + "processor_classes": [] }, "M2M100ForConditionalGeneration": { - "tokenizer_class": "M2M100Tokenizer", - "processor_class": null + "tokenizer_classes": [ + "M2M100Tokenizer" + ], + "processor_classes": [] }, "MBartForCausalLM": { - "tokenizer_class": "MBartTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MBartTokenizerFast", + "MBartTokenizer" + ], + "processor_classes": [] }, "MBartForConditionalGeneration": { - "tokenizer_class": "MBartTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MBartTokenizerFast", + "MBartTokenizer" + ], + "processor_classes": [] }, "MBartForQuestionAnswering": { - "tokenizer_class": "MBartTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MBartTokenizerFast", + "MBartTokenizer" + ], + "processor_classes": [] }, "MBartForSequenceClassification": { - "tokenizer_class": "MBartTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MBartTokenizerFast", + "MBartTokenizer" + ], + "processor_classes": [] }, "MCTCTForCTC": { - "tokenizer_class": null, - "processor_class": "MCTCTFeatureExtractor" + "tokenizer_classes": [], + "processor_classes": [ + "MCTCTFeatureExtractor" + ] }, "MPNetForMaskedLM": { - "tokenizer_class": "MPNetTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MPNetTokenizerFast", + "MPNetTokenizer" + ], + "processor_classes": [] }, "MPNetForMultipleChoice": { - "tokenizer_class": "MPNetTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MPNetTokenizerFast", + "MPNetTokenizer" + ], + "processor_classes": [] }, "MPNetForQuestionAnswering": { - "tokenizer_class": "MPNetTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MPNetTokenizerFast", + "MPNetTokenizer" + ], + "processor_classes": [] }, "MPNetForSequenceClassification": { - "tokenizer_class": "MPNetTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MPNetTokenizerFast", + "MPNetTokenizer" + ], + "processor_classes": [] }, "MPNetForTokenClassification": { - "tokenizer_class": "MPNetTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MPNetTokenizerFast", + "MPNetTokenizer" + ], + "processor_classes": [] }, "MarianForCausalLM": { - "tokenizer_class": "MarianTokenizer", - "processor_class": null + "tokenizer_classes": [ + "MarianTokenizer" + ], + "processor_classes": [] }, "MarkupLMForQuestionAnswering": { - "tokenizer_class": "MarkupLMTokenizerFast", - "processor_class": "MarkupLMFeatureExtractor" + "tokenizer_classes": [ + "MarkupLMTokenizerFast", + "MarkupLMTokenizer" + ], + "processor_classes": [ + "MarkupLMFeatureExtractor" + ] }, "MarkupLMForSequenceClassification": { - "tokenizer_class": "MarkupLMTokenizerFast", - "processor_class": "MarkupLMFeatureExtractor" + "tokenizer_classes": [ + "MarkupLMTokenizerFast", + "MarkupLMTokenizer" + ], + "processor_classes": [ + "MarkupLMFeatureExtractor" + ] }, "MarkupLMForTokenClassification": { - "tokenizer_class": "MarkupLMTokenizerFast", - "processor_class": "MarkupLMFeatureExtractor" + "tokenizer_classes": [ + "MarkupLMTokenizerFast", + "MarkupLMTokenizer" + ], + "processor_classes": [ + "MarkupLMFeatureExtractor" + ] }, "MaskFormerForInstanceSegmentation": { - "tokenizer_class": null, - "processor_class": "MaskFormerFeatureExtractor" + "tokenizer_classes": [], + "processor_classes": [ + "MaskFormerFeatureExtractor" + ] }, "MegatronBertForCausalLM": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "MegatronBertForMaskedLM": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "MegatronBertForMultipleChoice": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "MegatronBertForNextSentencePrediction": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "MegatronBertForPreTraining": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "MegatronBertForQuestionAnswering": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "MegatronBertForSequenceClassification": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "MegatronBertForTokenClassification": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "MobileBertForMaskedLM": { - "tokenizer_class": "MobileBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MobileBertTokenizerFast", + "MobileBertTokenizer" + ], + "processor_classes": [] }, "MobileBertForMultipleChoice": { - "tokenizer_class": "MobileBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MobileBertTokenizerFast", + "MobileBertTokenizer" + ], + "processor_classes": [] }, "MobileBertForNextSentencePrediction": { - "tokenizer_class": "MobileBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MobileBertTokenizerFast", + "MobileBertTokenizer" + ], + "processor_classes": [] }, "MobileBertForPreTraining": { - "tokenizer_class": "MobileBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MobileBertTokenizerFast", + "MobileBertTokenizer" + ], + "processor_classes": [] }, "MobileBertForQuestionAnswering": { - "tokenizer_class": "MobileBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MobileBertTokenizerFast", + "MobileBertTokenizer" + ], + "processor_classes": [] }, "MobileBertForSequenceClassification": { - "tokenizer_class": "MobileBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MobileBertTokenizerFast", + "MobileBertTokenizer" + ], + "processor_classes": [] }, "MobileBertForTokenClassification": { - "tokenizer_class": "MobileBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MobileBertTokenizerFast", + "MobileBertTokenizer" + ], + "processor_classes": [] }, "MobileViTForImageClassification": { - "tokenizer_class": null, - "processor_class": "MobileViTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "MobileViTImageProcessor" + ] }, "MobileViTForSemanticSegmentation": { - "tokenizer_class": null, - "processor_class": "MobileViTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "MobileViTImageProcessor" + ] }, "MvpForCausalLM": { - "tokenizer_class": "MvpTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MvpTokenizerFast", + "MvpTokenizer" + ], + "processor_classes": [] }, "MvpForConditionalGeneration": { - "tokenizer_class": "MvpTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MvpTokenizerFast", + "MvpTokenizer" + ], + "processor_classes": [] }, "MvpForQuestionAnswering": { - "tokenizer_class": "MvpTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MvpTokenizerFast", + "MvpTokenizer" + ], + "processor_classes": [] }, "MvpForSequenceClassification": { - "tokenizer_class": "MvpTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "MvpTokenizerFast", + "MvpTokenizer" + ], + "processor_classes": [] }, "NezhaForMaskedLM": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "NezhaForMultipleChoice": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "NezhaForNextSentencePrediction": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "NezhaForPreTraining": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "NezhaForQuestionAnswering": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "NezhaForSequenceClassification": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "NezhaForTokenClassification": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "NystromformerForMaskedLM": { - "tokenizer_class": "AlbertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "AlbertTokenizerFast" + ], + "processor_classes": [] }, "NystromformerForMultipleChoice": { - "tokenizer_class": "AlbertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "AlbertTokenizerFast" + ], + "processor_classes": [] }, "NystromformerForQuestionAnswering": { - "tokenizer_class": "AlbertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "AlbertTokenizerFast" + ], + "processor_classes": [] }, "NystromformerForSequenceClassification": { - "tokenizer_class": "AlbertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "AlbertTokenizerFast" + ], + "processor_classes": [] }, "NystromformerForTokenClassification": { - "tokenizer_class": "AlbertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "AlbertTokenizerFast" + ], + "processor_classes": [] }, "OPTForCausalLM": { - "tokenizer_class": "GPT2TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "GPT2TokenizerFast", + "GPT2Tokenizer" + ], + "processor_classes": [] }, "OPTForQuestionAnswering": { - "tokenizer_class": "GPT2TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "GPT2TokenizerFast", + "GPT2Tokenizer" + ], + "processor_classes": [] }, "OPTForSequenceClassification": { - "tokenizer_class": "GPT2TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "GPT2TokenizerFast", + "GPT2Tokenizer" + ], + "processor_classes": [] }, "OpenAIGPTForSequenceClassification": { - "tokenizer_class": "OpenAIGPTTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "OpenAIGPTTokenizerFast", + "OpenAIGPTTokenizer" + ], + "processor_classes": [] }, "OwlViTForObjectDetection": { - "tokenizer_class": "CLIPTokenizerFast", - "processor_class": "OwlViTFeatureExtractor" + "tokenizer_classes": [ + "CLIPTokenizerFast", + "CLIPTokenizer" + ], + "processor_classes": [ + "OwlViTFeatureExtractor" + ] }, "PLBartForCausalLM": { - "tokenizer_class": "PLBartTokenizer", - "processor_class": null + "tokenizer_classes": [ + "PLBartTokenizer" + ], + "processor_classes": [] }, "PLBartForConditionalGeneration": { - "tokenizer_class": "PLBartTokenizer", - "processor_class": null + "tokenizer_classes": [ + "PLBartTokenizer" + ], + "processor_classes": [] }, "PLBartForSequenceClassification": { - "tokenizer_class": "PLBartTokenizer", - "processor_class": null + "tokenizer_classes": [ + "PLBartTokenizer" + ], + "processor_classes": [] }, "PegasusForCausalLM": { - "tokenizer_class": "PegasusTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "PegasusTokenizerFast" + ], + "processor_classes": [] }, "PegasusForConditionalGeneration": { - "tokenizer_class": "PegasusTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "PegasusTokenizerFast" + ], + "processor_classes": [] }, "PegasusXForConditionalGeneration": { - "tokenizer_class": "PegasusTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "PegasusTokenizerFast" + ], + "processor_classes": [] }, "PerceiverForImageClassificationConvProcessing": { - "tokenizer_class": "PerceiverTokenizer", - "processor_class": null + "tokenizer_classes": [ + "PerceiverTokenizer" + ], + "processor_classes": [] }, "PerceiverForImageClassificationFourier": { - "tokenizer_class": "PerceiverTokenizer", - "processor_class": null + "tokenizer_classes": [ + "PerceiverTokenizer" + ], + "processor_classes": [] }, "PerceiverForImageClassificationLearned": { - "tokenizer_class": "PerceiverTokenizer", - "processor_class": null + "tokenizer_classes": [ + "PerceiverTokenizer" + ], + "processor_classes": [] }, "PerceiverForMaskedLM": { - "tokenizer_class": "PerceiverTokenizer", - "processor_class": null + "tokenizer_classes": [ + "PerceiverTokenizer" + ], + "processor_classes": [] }, "PerceiverForSequenceClassification": { - "tokenizer_class": "PerceiverTokenizer", - "processor_class": null + "tokenizer_classes": [ + "PerceiverTokenizer" + ], + "processor_classes": [] }, "PoolFormerForImageClassification": { - "tokenizer_class": null, - "processor_class": "PoolFormerImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "PoolFormerImageProcessor" + ] }, "ProphetNetForCausalLM": { - "tokenizer_class": "ProphetNetTokenizer", - "processor_class": null + "tokenizer_classes": [ + "ProphetNetTokenizer" + ], + "processor_classes": [] }, "ProphetNetForConditionalGeneration": { - "tokenizer_class": "ProphetNetTokenizer", - "processor_class": null + "tokenizer_classes": [ + "ProphetNetTokenizer" + ], + "processor_classes": [] }, "ReformerForMaskedLM": { - "tokenizer_class": "ReformerTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "ReformerTokenizerFast", + "ReformerTokenizer" + ], + "processor_classes": [] }, "ReformerForQuestionAnswering": { - "tokenizer_class": "ReformerTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "ReformerTokenizerFast", + "ReformerTokenizer" + ], + "processor_classes": [] }, "ReformerForSequenceClassification": { - "tokenizer_class": "ReformerTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "ReformerTokenizerFast", + "ReformerTokenizer" + ], + "processor_classes": [] }, "RegNetForImageClassification": { - "tokenizer_class": null, - "processor_class": "ConvNextImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ConvNextImageProcessor" + ] }, "RemBertForCausalLM": { - "tokenizer_class": "RemBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RemBertTokenizerFast" + ], + "processor_classes": [] }, "RemBertForMaskedLM": { - "tokenizer_class": "RemBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RemBertTokenizerFast" + ], + "processor_classes": [] }, "RemBertForMultipleChoice": { - "tokenizer_class": "RemBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RemBertTokenizerFast" + ], + "processor_classes": [] }, "RemBertForQuestionAnswering": { - "tokenizer_class": "RemBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RemBertTokenizerFast" + ], + "processor_classes": [] }, "RemBertForSequenceClassification": { - "tokenizer_class": "RemBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RemBertTokenizerFast" + ], + "processor_classes": [] }, "RemBertForTokenClassification": { - "tokenizer_class": "RemBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RemBertTokenizerFast" + ], + "processor_classes": [] }, "ResNetForImageClassification": { - "tokenizer_class": null, - "processor_class": "ConvNextImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ConvNextImageProcessor" + ] }, "RoFormerForCausalLM": { - "tokenizer_class": "RoFormerTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RoFormerTokenizerFast", + "RoFormerTokenizer" + ], + "processor_classes": [] }, "RoFormerForMaskedLM": { - "tokenizer_class": "RoFormerTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RoFormerTokenizerFast", + "RoFormerTokenizer" + ], + "processor_classes": [] }, "RoFormerForMultipleChoice": { - "tokenizer_class": "RoFormerTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RoFormerTokenizerFast", + "RoFormerTokenizer" + ], + "processor_classes": [] }, "RoFormerForQuestionAnswering": { - "tokenizer_class": "RoFormerTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RoFormerTokenizerFast", + "RoFormerTokenizer" + ], + "processor_classes": [] }, "RoFormerForSequenceClassification": { - "tokenizer_class": "RoFormerTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RoFormerTokenizerFast", + "RoFormerTokenizer" + ], + "processor_classes": [] }, "RoFormerForTokenClassification": { - "tokenizer_class": "RoFormerTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RoFormerTokenizerFast", + "RoFormerTokenizer" + ], + "processor_classes": [] }, "RobertaForCausalLM": { - "tokenizer_class": "RobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RobertaTokenizerFast", + "RobertaTokenizer" + ], + "processor_classes": [] }, "RobertaForMaskedLM": { - "tokenizer_class": "RobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RobertaTokenizerFast", + "RobertaTokenizer" + ], + "processor_classes": [] }, "RobertaForMultipleChoice": { - "tokenizer_class": "RobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RobertaTokenizerFast", + "RobertaTokenizer" + ], + "processor_classes": [] }, "RobertaForQuestionAnswering": { - "tokenizer_class": "RobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RobertaTokenizerFast", + "RobertaTokenizer" + ], + "processor_classes": [] }, "RobertaForSequenceClassification": { - "tokenizer_class": "RobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RobertaTokenizerFast", + "RobertaTokenizer" + ], + "processor_classes": [] }, "RobertaForTokenClassification": { - "tokenizer_class": "RobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "RobertaTokenizerFast", + "RobertaTokenizer" + ], + "processor_classes": [] }, "SEWDForCTC": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "SEWDForSequenceClassification": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "SEWForCTC": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "SEWForSequenceClassification": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "SegformerForImageClassification": { - "tokenizer_class": null, - "processor_class": "SegformerImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "SegformerImageProcessor" + ] }, "SegformerForSemanticSegmentation": { - "tokenizer_class": null, - "processor_class": "SegformerImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "SegformerImageProcessor" + ] }, "Speech2TextForConditionalGeneration": { - "tokenizer_class": "Speech2TextTokenizer", - "processor_class": "Speech2TextFeatureExtractor" + "tokenizer_classes": [ + "Speech2TextTokenizer" + ], + "processor_classes": [ + "Speech2TextFeatureExtractor" + ] }, "SplinterForPreTraining": { - "tokenizer_class": null, - "processor_class": null + "tokenizer_classes": [], + "processor_classes": [] }, "SplinterForQuestionAnswering": { - "tokenizer_class": null, - "processor_class": null + "tokenizer_classes": [], + "processor_classes": [] }, "SqueezeBertForMaskedLM": { - "tokenizer_class": "SqueezeBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "SqueezeBertTokenizerFast", + "SqueezeBertTokenizer" + ], + "processor_classes": [] }, "SqueezeBertForMultipleChoice": { - "tokenizer_class": "SqueezeBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "SqueezeBertTokenizerFast", + "SqueezeBertTokenizer" + ], + "processor_classes": [] }, "SqueezeBertForQuestionAnswering": { - "tokenizer_class": "SqueezeBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "SqueezeBertTokenizerFast", + "SqueezeBertTokenizer" + ], + "processor_classes": [] }, "SqueezeBertForSequenceClassification": { - "tokenizer_class": "SqueezeBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "SqueezeBertTokenizerFast", + "SqueezeBertTokenizer" + ], + "processor_classes": [] }, "SqueezeBertForTokenClassification": { - "tokenizer_class": "SqueezeBertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "SqueezeBertTokenizerFast", + "SqueezeBertTokenizer" + ], + "processor_classes": [] }, "SwinForImageClassification": { - "tokenizer_class": null, - "processor_class": "ViTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ViTImageProcessor" + ] }, "SwinForMaskedImageModeling": { - "tokenizer_class": null, - "processor_class": "ViTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ViTImageProcessor" + ] }, "Swinv2ForImageClassification": { - "tokenizer_class": null, - "processor_class": "ViTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ViTImageProcessor" + ] }, "Swinv2ForMaskedImageModeling": { - "tokenizer_class": null, - "processor_class": "ViTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ViTImageProcessor" + ] }, "T5ForConditionalGeneration": { - "tokenizer_class": "T5TokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "T5TokenizerFast" + ], + "processor_classes": [] }, "TableTransformerForObjectDetection": { - "tokenizer_class": null, - "processor_class": "DetrFeatureExtractor" + "tokenizer_classes": [], + "processor_classes": [ + "DetrFeatureExtractor" + ] }, "TapasForMaskedLM": { - "tokenizer_class": "TapasTokenizer", - "processor_class": null + "tokenizer_classes": [ + "TapasTokenizer" + ], + "processor_classes": [] }, "TapasForQuestionAnswering": { - "tokenizer_class": "TapasTokenizer", - "processor_class": null + "tokenizer_classes": [ + "TapasTokenizer" + ], + "processor_classes": [] }, "TapasForSequenceClassification": { - "tokenizer_class": "TapasTokenizer", - "processor_class": null + "tokenizer_classes": [ + "TapasTokenizer" + ], + "processor_classes": [] }, "TransfoXLForSequenceClassification": { - "tokenizer_class": "TransfoXLTokenizer", - "processor_class": null + "tokenizer_classes": [ + "TransfoXLTokenizer" + ], + "processor_classes": [] }, "UniSpeechForCTC": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "UniSpeechForPreTraining": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "UniSpeechForSequenceClassification": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "UniSpeechSatForCTC": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "UniSpeechSatForPreTraining": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "UniSpeechSatForSequenceClassification": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "UniSpeechSatForXVector": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "VanForImageClassification": { - "tokenizer_class": null, - "processor_class": "ConvNextImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ConvNextImageProcessor" + ] }, "ViTForImageClassification": { - "tokenizer_class": null, - "processor_class": "ViTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ViTImageProcessor" + ] }, "ViTForMaskedImageModeling": { - "tokenizer_class": null, - "processor_class": "ViTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ViTImageProcessor" + ] }, "ViTMAEForPreTraining": { - "tokenizer_class": null, - "processor_class": "ViTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ViTImageProcessor" + ] }, "ViTMSNForImageClassification": { - "tokenizer_class": null, - "processor_class": "ViTImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "ViTImageProcessor" + ] }, "VideoMAEForPreTraining": { - "tokenizer_class": null, - "processor_class": "VideoMAEImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "VideoMAEImageProcessor" + ] }, "VideoMAEForVideoClassification": { - "tokenizer_class": null, - "processor_class": "VideoMAEImageProcessor" + "tokenizer_classes": [], + "processor_classes": [ + "VideoMAEImageProcessor" + ] }, "ViltForQuestionAnswering": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": "ViltImageProcessor" + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [ + "ViltImageProcessor" + ] }, "VisualBertForPreTraining": { - "tokenizer_class": "BertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "BertTokenizerFast", + "BertTokenizer" + ], + "processor_classes": [] }, "Wav2Vec2ConformerForCTC": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "Wav2Vec2ConformerForPreTraining": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "Wav2Vec2ConformerForSequenceClassification": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "Wav2Vec2ConformerForXVector": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "Wav2Vec2ForCTC": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "Wav2Vec2ForMaskedLM": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "Wav2Vec2ForPreTraining": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "Wav2Vec2ForSequenceClassification": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "Wav2Vec2ForXVector": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "WavLMForCTC": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "WavLMForSequenceClassification": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "WavLMForXVector": { - "tokenizer_class": "Wav2Vec2CTCTokenizer", - "processor_class": "Wav2Vec2FeatureExtractor" + "tokenizer_classes": [ + "Wav2Vec2CTCTokenizer" + ], + "processor_classes": [ + "Wav2Vec2FeatureExtractor" + ] }, "WhisperForConditionalGeneration": { - "tokenizer_class": "WhisperTokenizer", - "processor_class": "WhisperFeatureExtractor" + "tokenizer_classes": [ + "WhisperTokenizer" + ], + "processor_classes": [ + "WhisperFeatureExtractor" + ] }, "XGLMForCausalLM": { - "tokenizer_class": "XGLMTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "XGLMTokenizerFast" + ], + "processor_classes": [] }, "XLMForMultipleChoice": { - "tokenizer_class": "XLMTokenizer", - "processor_class": null + "tokenizer_classes": [ + "XLMTokenizer" + ], + "processor_classes": [] }, "XLMForQuestionAnsweringSimple": { - "tokenizer_class": "XLMTokenizer", - "processor_class": null + "tokenizer_classes": [ + "XLMTokenizer" + ], + "processor_classes": [] }, "XLMForSequenceClassification": { - "tokenizer_class": "XLMTokenizer", - "processor_class": null + "tokenizer_classes": [ + "XLMTokenizer" + ], + "processor_classes": [] }, "XLMForTokenClassification": { - "tokenizer_class": "XLMTokenizer", - "processor_class": null + "tokenizer_classes": [ + "XLMTokenizer" + ], + "processor_classes": [] }, "XLMRobertaXLForCausalLM": { - "tokenizer_class": "XLMRobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "XLMRobertaTokenizerFast" + ], + "processor_classes": [] }, "XLMRobertaXLForMaskedLM": { - "tokenizer_class": "XLMRobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "XLMRobertaTokenizerFast" + ], + "processor_classes": [] }, "XLMRobertaXLForMultipleChoice": { - "tokenizer_class": "XLMRobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "XLMRobertaTokenizerFast" + ], + "processor_classes": [] }, "XLMRobertaXLForQuestionAnswering": { - "tokenizer_class": "XLMRobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "XLMRobertaTokenizerFast" + ], + "processor_classes": [] }, "XLMRobertaXLForSequenceClassification": { - "tokenizer_class": "XLMRobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "XLMRobertaTokenizerFast" + ], + "processor_classes": [] }, "XLMRobertaXLForTokenClassification": { - "tokenizer_class": "XLMRobertaTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "XLMRobertaTokenizerFast" + ], + "processor_classes": [] }, "XLNetForMultipleChoice": { - "tokenizer_class": "XLNetTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "XLNetTokenizerFast", + "XLNetTokenizer" + ], + "processor_classes": [] }, "XLNetForQuestionAnsweringSimple": { - "tokenizer_class": "XLNetTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "XLNetTokenizerFast", + "XLNetTokenizer" + ], + "processor_classes": [] }, "XLNetForSequenceClassification": { - "tokenizer_class": "XLNetTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "XLNetTokenizerFast", + "XLNetTokenizer" + ], + "processor_classes": [] }, "XLNetForTokenClassification": { - "tokenizer_class": "XLNetTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "XLNetTokenizerFast", + "XLNetTokenizer" + ], + "processor_classes": [] }, "YolosForObjectDetection": { - "tokenizer_class": null, - "processor_class": "YolosFeatureExtractor" + "tokenizer_classes": [], + "processor_classes": [ + "YolosFeatureExtractor" + ] }, "YosoForMaskedLM": { - "tokenizer_class": "AlbertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "AlbertTokenizerFast" + ], + "processor_classes": [] }, "YosoForMultipleChoice": { - "tokenizer_class": "AlbertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "AlbertTokenizerFast" + ], + "processor_classes": [] }, "YosoForQuestionAnswering": { - "tokenizer_class": "AlbertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "AlbertTokenizerFast" + ], + "processor_classes": [] }, "YosoForSequenceClassification": { - "tokenizer_class": "AlbertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "AlbertTokenizerFast" + ], + "processor_classes": [] }, "YosoForTokenClassification": { - "tokenizer_class": "AlbertTokenizerFast", - "processor_class": null + "tokenizer_classes": [ + "AlbertTokenizerFast" + ], + "processor_classes": [] } } \ No newline at end of file