Unnamed: 0
int64
0
245k
repo_id
stringlengths
4
122
author
stringlengths
2
42
model_type
stringlengths
2
34
files_per_repo
int64
0
77k
downloads_30d
int64
0
55.9M
library
stringlengths
2
37
likes
int64
0
8.48k
pipeline
stringlengths
5
30
pytorch
bool
2 classes
tensorflow
bool
2 classes
jax
bool
2 classes
license
stringlengths
2
33
languages
stringlengths
2
1.63k
datasets
stringlengths
2
5.05k
co2
stringlengths
3
342
prs_count
int64
0
168
prs_open
int64
0
121
prs_merged
int64
0
167
prs_closed
int64
0
35
discussions_count
int64
0
226
discussions_open
int64
0
155
discussions_closed
int64
0
76
tags
stringlengths
2
7.26k
has_model_index
bool
2 classes
has_metadata
bool
2 classes
has_text
bool
2 classes
text_length
int64
0
849k
10,700
anuragshas/wav2vec2-large-xls-r-300m-pa-in
anuragshas
wav2vec2
19
6
transformers
0
automatic-speech-recognition
true
false
false
apache-2.0
['pa']
['mozilla-foundation/common_voice_7_0']
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 'wav2vec2', 'automatic-speech-recognition', 'pa', 'dataset:mozilla-foundation/common_voice_7_0', 'transformers', 'generated_from_trainer', 'robust-speech-event', 'hf-asr-leaderboard', 'license:apache-2.0', 'model-index']
true
true
true
2,939
10,701
anuragshas/wav2vec2-large-xls-r-300m-ur-cv8
anuragshas
wav2vec2
19
8
transformers
0
automatic-speech-recognition
true
false
false
apache-2.0
['ur']
['mozilla-foundation/common_voice_8_0']
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 'wav2vec2', 'automatic-speech-recognition', 'ur', 'dataset:mozilla-foundation/common_voice_8_0', 'transformers', 'generated_from_trainer', 'robust-speech-event', 'hf-asr-leaderboard', 'license:apache-2.0', 'model-index']
true
true
true
3,191
10,702
anuragshas/wav2vec2-large-xls-r-300m-ur
anuragshas
wav2vec2
15
8
transformers
0
automatic-speech-recognition
true
false
false
apache-2.0
null
['common_voice']
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 'wav2vec2', 'automatic-speech-recognition', 'dataset:common_voice', 'transformers', 'generated_from_trainer', 'license:apache-2.0']
true
true
true
1,471
10,703
anuragshas/wav2vec2-large-xlsr-53-dv
anuragshas
wav2vec2
9
9
transformers
1
automatic-speech-recognition
true
false
true
apache-2.0
['dv']
['common_voice']
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'wav2vec2', 'automatic-speech-recognition', 'dv', 'dataset:common_voice', 'transformers', 'audio', 'speech', 'xlsr-fine-tuning-week', 'license:apache-2.0', 'model-index']
true
true
true
3,271
10,704
anuragshas/wav2vec2-large-xlsr-53-hsb
anuragshas
wav2vec2
9
8
transformers
0
automatic-speech-recognition
true
false
true
apache-2.0
['hsb']
['common_voice']
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'wav2vec2', 'automatic-speech-recognition', 'hsb', 'dataset:common_voice', 'transformers', 'audio', 'speech', 'xlsr-fine-tuning-week', 'license:apache-2.0', 'model-index']
true
true
true
3,294
10,705
anuragshas/wav2vec2-large-xlsr-53-ia
anuragshas
wav2vec2
9
7
transformers
0
automatic-speech-recognition
true
false
true
apache-2.0
['ia']
['common_voice']
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'wav2vec2', 'automatic-speech-recognition', 'ia', 'dataset:common_voice', 'transformers', 'audio', 'speech', 'xlsr-fine-tuning-week', 'license:apache-2.0', 'model-index']
true
true
true
3,269
10,706
anuragshas/wav2vec2-large-xlsr-53-odia
anuragshas
wav2vec2
10
7
transformers
0
automatic-speech-recognition
true
false
true
apache-2.0
['or']
['common_voice']
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'wav2vec2', 'automatic-speech-recognition', 'or', 'dataset:common_voice', 'transformers', 'audio', 'speech', 'xlsr-fine-tuning-week', 'license:apache-2.0', 'model-index']
true
true
true
3,252
10,707
anuragshas/wav2vec2-large-xlsr-53-rm-sursilv
anuragshas
wav2vec2
9
7
transformers
0
automatic-speech-recognition
true
false
true
apache-2.0
['rm-sursilv']
['common_voice']
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'wav2vec2', 'automatic-speech-recognition', 'rm-sursilv', 'dataset:common_voice', 'transformers', 'audio', 'speech', 'xlsr-fine-tuning-week', 'license:apache-2.0', 'model-index']
true
true
true
3,339
10,708
anuragshas/wav2vec2-large-xlsr-53-rm-vallader
anuragshas
wav2vec2
9
9
transformers
0
automatic-speech-recognition
true
false
true
apache-2.0
['rm-vallader']
['common_voice']
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'wav2vec2', 'automatic-speech-recognition', 'rm-vallader', 'dataset:common_voice', 'transformers', 'audio', 'speech', 'xlsr-fine-tuning-week', 'license:apache-2.0', 'model-index']
true
true
true
3,527
10,709
anuragshas/wav2vec2-large-xlsr-53-sah
anuragshas
wav2vec2
9
7
transformers
0
automatic-speech-recognition
true
false
true
apache-2.0
['sah']
['common_voice']
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'wav2vec2', 'automatic-speech-recognition', 'sah', 'dataset:common_voice', 'transformers', 'audio', 'speech', 'xlsr-fine-tuning-week', 'license:apache-2.0', 'model-index']
true
true
true
3,267
10,710
anuragshas/wav2vec2-large-xlsr-53-telugu
anuragshas
wav2vec2
9
30
transformers
1
automatic-speech-recognition
true
false
true
apache-2.0
['te']
['openslr']
null
1
1
0
0
0
0
0
['pytorch', 'jax', 'wav2vec2', 'automatic-speech-recognition', 'te', 'dataset:openslr', 'transformers', 'audio', 'speech', 'xlsr-fine-tuning-week', 'license:apache-2.0', 'model-index']
true
true
true
4,434
10,711
anuragshas/wav2vec2-large-xlsr-53-vietnamese
anuragshas
wav2vec2
9
34
transformers
0
automatic-speech-recognition
true
false
true
apache-2.0
['vi']
['common_voice']
null
1
1
0
0
0
0
0
['pytorch', 'jax', 'wav2vec2', 'automatic-speech-recognition', 'vi', 'dataset:common_voice', 'transformers', 'audio', 'speech', 'xlsr-fine-tuning-week', 'license:apache-2.0', 'model-index']
true
true
true
3,294
10,712
anuragshas/wav2vec2-large-xlsr-as
anuragshas
wav2vec2
10
6
transformers
0
automatic-speech-recognition
true
false
true
apache-2.0
['as']
['common_voice']
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'wav2vec2', 'automatic-speech-recognition', 'as', 'dataset:common_voice', 'transformers', 'audio', 'speech', 'xlsr-fine-tuning-week', 'license:apache-2.0', 'model-index']
true
true
true
3,444
10,713
anuragshas/wav2vec2-xls-r-1b-hi-cv8
anuragshas
wav2vec2
24
6
transformers
0
automatic-speech-recognition
true
false
false
apache-2.0
['hi']
['common_voice']
null
0
0
0
0
0
0
0
['pytorch', 'wav2vec2', 'automatic-speech-recognition', 'hi', 'dataset:common_voice', 'transformers', 'mozilla-foundation/common_voice_8_0', 'generated_from_trainer', 'license:apache-2.0']
true
true
true
2,748
10,714
anuragshas/wav2vec2-xls-r-1b-hi-with-lm
anuragshas
wav2vec2
23
8
transformers
1
automatic-speech-recognition
true
false
false
apache-2.0
['hi']
['mozilla-foundation/common_voice_8_0']
null
0
0
0
0
0
0
0
['pytorch', 'wav2vec2', 'automatic-speech-recognition', 'hi', 'dataset:mozilla-foundation/common_voice_8_0', 'transformers', 'generated_from_trainer', 'hf-asr-leaderboard', 'mozilla-foundation/common_voice_8_0', 'robust-speech-event', 'license:apache-2.0', 'model-index', 'has_space']
true
true
true
3,954
10,715
anuragshas/wav2vec2-xls-r-1b-hi
anuragshas
wav2vec2
24
9
transformers
0
automatic-speech-recognition
true
false
false
apache-2.0
['hi']
['mozilla-foundation/common_voice_7_0']
null
0
0
0
0
0
0
0
['pytorch', 'wav2vec2', 'automatic-speech-recognition', 'hi', 'dataset:mozilla-foundation/common_voice_7_0', 'transformers', 'generated_from_trainer', 'hf-asr-leaderboard', 'mozilla-foundation/common_voice_7_0', 'robust-speech-event', 'license:apache-2.0', 'model-index']
true
true
true
4,707
10,716
anuragshas/wav2vec2-xls-r-300m-lv-cv8-with-lm
anuragshas
wav2vec2
24
6
transformers
0
automatic-speech-recognition
true
false
false
apache-2.0
['lv']
['mozilla-foundation/common_voice_8_0']
null
0
0
0
0
0
0
0
['pytorch', 'wav2vec2', 'automatic-speech-recognition', 'lv', 'dataset:mozilla-foundation/common_voice_8_0', 'transformers', 'mozilla-foundation/common_voice_8_0', 'generated_from_trainer', 'robust-speech-event', 'hf-asr-leaderboard', 'license:apache-2.0', 'model-index']
true
true
true
3,880
10,717
anuragshas/wav2vec2-xls-r-300m-mr-cv8-with-lm
anuragshas
wav2vec2
22
8
transformers
0
automatic-speech-recognition
true
false
false
apache-2.0
['mr']
['common_voice']
null
0
0
0
0
0
0
0
['pytorch', 'wav2vec2', 'automatic-speech-recognition', 'mr', 'dataset:common_voice', 'transformers', 'mozilla-foundation/common_voice_8_0', 'generated_from_trainer', 'license:apache-2.0']
true
true
true
2,932
10,718
anuragshas/wav2vec2-xls-r-300m-mt-cv8-with-lm
anuragshas
wav2vec2
24
7
transformers
0
automatic-speech-recognition
true
false
false
apache-2.0
['mt']
['mozilla-foundation/common_voice_8_0']
null
0
0
0
0
0
0
0
['pytorch', 'wav2vec2', 'automatic-speech-recognition', 'mt', 'dataset:mozilla-foundation/common_voice_8_0', 'transformers', 'mozilla-foundation/common_voice_8_0', 'generated_from_trainer', 'robust-speech-event', 'hf-asr-leaderboard', 'license:apache-2.0', 'model-index']
true
true
true
3,438
10,719
anuragshas/wav2vec2-xls-r-300m-pa-IN-cv8-with-lm
anuragshas
wav2vec2
19
12
transformers
0
automatic-speech-recognition
true
false
false
apache-2.0
['pa-IN']
['common_voice']
null
1
1
0
0
0
0
0
['pytorch', 'wav2vec2', 'automatic-speech-recognition', 'pa-IN', 'dataset:common_voice', 'transformers', 'mozilla-foundation/common_voice_8_0', 'generated_from_trainer', 'license:apache-2.0']
true
true
true
2,052
10,720
anuragshas/wav2vec2-xls-r-300m-sk-cv8-with-lm
anuragshas
wav2vec2
24
8
transformers
0
automatic-speech-recognition
true
false
false
apache-2.0
['sk']
['mozilla-foundation/common_voice_8_0']
null
0
0
0
0
0
0
0
['pytorch', 'wav2vec2', 'automatic-speech-recognition', 'sk', 'dataset:mozilla-foundation/common_voice_8_0', 'transformers', 'generated_from_trainer', 'hf-asr-leaderboard', 'mozilla-foundation/common_voice_8_0', 'robust-speech-event', 'license:apache-2.0', 'model-index']
true
true
true
4,154
10,721
anuragshas/wav2vec2-xls-r-300m-sl-cv8-with-lm
anuragshas
wav2vec2
25
6
transformers
0
automatic-speech-recognition
true
false
false
apache-2.0
['sl']
['mozilla-foundation/common_voice_8_0']
null
0
0
0
0
0
0
0
['pytorch', 'wav2vec2', 'automatic-speech-recognition', 'sl', 'dataset:mozilla-foundation/common_voice_8_0', 'transformers', 'generated_from_trainer', 'hf-asr-leaderboard', 'mozilla-foundation/common_voice_8_0', 'robust-speech-event', 'license:apache-2.0', 'model-index']
true
true
true
3,489
10,722
anuragshas/wav2vec2-xls-r-300m-ta-cv8
anuragshas
wav2vec2
13
6
transformers
0
automatic-speech-recognition
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'wav2vec2', 'automatic-speech-recognition', 'transformers']
false
false
false
0
10,723
anuragshas/wav2vec2-xlsr-53-pa-in
anuragshas
wav2vec2
10
7
transformers
0
automatic-speech-recognition
true
false
true
apache-2.0
['pa-IN']
['common_voice']
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'wav2vec2', 'automatic-speech-recognition', 'pa-IN', 'dataset:common_voice', 'transformers', 'audio', 'speech', 'xlsr-fine-tuning-week', 'license:apache-2.0', 'model-index']
true
true
true
3,261
10,724
anuragshas/wav2vec2-xlsr-53-rm-vallader-with-lm
anuragshas
wav2vec2
13
7
transformers
0
automatic-speech-recognition
true
false
false
apache-2.0
null
['common_voice']
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 'wav2vec2', 'automatic-speech-recognition', 'dataset:common_voice', 'transformers', 'generated_from_trainer', 'license:apache-2.0']
true
true
true
1,846
10,725
anuragshas/wav2vec2-xlsr-53-tamil
anuragshas
wav2vec2
10
36
transformers
0
automatic-speech-recognition
true
false
true
apache-2.0
['ta']
['common_voice']
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'wav2vec2', 'automatic-speech-recognition', 'ta', 'dataset:common_voice', 'transformers', 'audio', 'speech', 'xlsr-fine-tuning-week', 'license:apache-2.0', 'model-index']
true
true
true
3,247
10,726
anusha/t5-base-finetuned-wikiSQL-sql-to-en
anusha
t5
15
16
transformers
0
text2text-generation
true
false
true
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'jax', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
false
0
10,727
anusha/t5-base-finetuned-wikiSQL-sql-to-en_1
anusha
t5
8
9
transformers
0
text2text-generation
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
false
0
10,728
anusha/t5-base-finetuned-wikiSQL-sql-to-en_15i
anusha
t5
8
29
transformers
1
text2text-generation
true
false
true
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'jax', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
false
0
10,729
anushakamath/wav2vec2-xls-r-300m-hindi
anushakamath
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
10,730
anushakamath/wav2vec2-xls-r-300m-punjabi-in
anushakamath
wav2vec2
8
7
transformers
0
automatic-speech-recognition
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'wav2vec2', 'automatic-speech-recognition', 'transformers']
false
false
false
0
10,731
anushakamath/wav2vec2-xls-r-300m-punjabi
anushakamath
null
4
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
10,732
anvaldesg/distilbert-base-uncased-finetuned-squad
anvaldesg
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
10,733
anweasha/DialoGPT-small-Chandler
anweasha
gpt2
21
12
transformers
0
conversational
true
false
false
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'gpt2', 'text-generation', 'transformers', 'conversational']
false
true
true
26
10,734
anweasha/DialoGPT-small-Jake
anweasha
gpt2
11
12
transformers
0
conversational
true
false
false
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'gpt2', 'text-generation', 'transformers', 'conversational']
false
true
true
30
10,735
any0019/text_style_classifier
any0019
bert
3
29
transformers
0
text-classification
true
false
false
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'bert', 'text-classification', 'transformers', 'has_space']
false
false
false
0
10,736
any0019/text_style_mlm_negative
any0019
bert
3
5
transformers
0
fill-mask
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'fill-mask', 'transformers', 'autotrain_compatible', 'has_space']
false
false
false
0
10,737
any0019/text_style_mlm_positive
any0019
bert
3
7
transformers
0
fill-mask
true
false
false
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'bert', 'fill-mask', 'transformers', 'autotrain_compatible', 'has_space']
false
false
false
0
10,738
anzorq/t5-v1_1-small-ru_kbd-cased
anzorq
t5
7
19
transformers
0
translation
true
false
false
null
['ru', 'kbd']
['anzorq/kbd-ru-1.67M-temp', '17753 Russian-Kabardian pairs of text']
null
0
0
0
0
0
0
0
['pytorch', 't5', 'text2text-generation', 'ru', 'kbd', 'dataset:anzorq/kbd-ru-1.67M-temp', 'dataset:17753 Russian-Kabardian pairs of text', 'transformers', 'translation', 'autotrain_compatible']
false
true
true
385
10,739
aodiniz/bert_uncased_L-10_H-512_A-8_cord19-200616
aodiniz
bert
9
7
transformers
0
fill-mask
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'fill-mask', 'arxiv:1908.08962', 'transformers', 'autotrain_compatible']
false
false
true
820
10,740
aodiniz/bert_uncased_L-10_H-512_A-8_cord19-200616_squad2
aodiniz
bert
13
18
transformers
0
question-answering
true
false
true
null
null
['squad_v2']
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'dataset:squad_v2', 'arxiv:1908.08962', 'transformers', 'autotrain_compatible']
false
true
true
984
10,741
aodiniz/bert_uncased_L-10_H-512_A-8_cord19-200616_squad2_covid-qna
aodiniz
bert
12
11
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,742
aodiniz/bert_uncased_L-10_H-512_A-8_squad2
aodiniz
bert
12
11
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,743
aodiniz/bert_uncased_L-10_H-512_A-8_squad2_covid-qna
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,744
aodiniz/bert_uncased_L-2_H-128_A-2_cord19-200616
aodiniz
bert
8
7
transformers
0
fill-mask
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'fill-mask', 'transformers', 'autotrain_compatible']
false
false
false
0
10,745
aodiniz/bert_uncased_L-2_H-128_A-2_cord19-200616_squad2
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,746
aodiniz/bert_uncased_L-2_H-128_A-2_cord19-200616_squad2_covid-qna
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,747
aodiniz/bert_uncased_L-2_H-128_A-2_squad2
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,748
aodiniz/bert_uncased_L-2_H-128_A-2_squad2_covid-qna
aodiniz
bert
12
11
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,749
aodiniz/bert_uncased_L-2_H-512_A-8_cord19-200616
aodiniz
bert
9
7
transformers
0
fill-mask
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'fill-mask', 'arxiv:1908.08962', 'transformers', 'autotrain_compatible']
false
false
true
815
10,750
aodiniz/bert_uncased_L-2_H-512_A-8_cord19-200616_squad2
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,751
aodiniz/bert_uncased_L-2_H-512_A-8_cord19-200616_squad2_covid-qna
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,752
aodiniz/bert_uncased_L-2_H-512_A-8_squad2
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,753
aodiniz/bert_uncased_L-2_H-512_A-8_squad2_covid-qna
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,754
aodiniz/bert_uncased_L-4_H-256_A-4_cord19-200616
aodiniz
bert
9
7
transformers
0
fill-mask
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'fill-mask', 'arxiv:1908.08962', 'transformers', 'autotrain_compatible']
false
false
true
815
10,755
aodiniz/bert_uncased_L-4_H-256_A-4_cord19-200616_squad2
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,756
aodiniz/bert_uncased_L-4_H-256_A-4_cord19-200616_squad2_covid-qna
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,757
aodiniz/bert_uncased_L-4_H-256_A-4_squad2
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,758
aodiniz/bert_uncased_L-4_H-256_A-4_squad2_covid-qna
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,759
aodiniz/bert_uncased_L-4_H-512_A-8_cord19-200616
aodiniz
bert
8
7
transformers
0
fill-mask
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'fill-mask', 'transformers', 'autotrain_compatible']
false
false
false
0
10,760
aodiniz/bert_uncased_L-4_H-512_A-8_cord19-200616_squad2
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,761
aodiniz/bert_uncased_L-4_H-512_A-8_cord19-200616_squad2_covid-qna
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,762
aodiniz/bert_uncased_L-4_H-512_A-8_squad2
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,763
aodiniz/bert_uncased_L-4_H-512_A-8_squad2_covid-qna
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,764
aodiniz/bert_uncased_L-4_H-768_A-12_cord19-200616
aodiniz
bert
8
7
transformers
0
fill-mask
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'fill-mask', 'transformers', 'autotrain_compatible']
false
false
false
0
10,765
aodiniz/bert_uncased_L-4_H-768_A-12_cord19-200616_squad2
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,766
aodiniz/bert_uncased_L-4_H-768_A-12_cord19-200616_squad2_covid-qna
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,767
aodiniz/bert_uncased_L-4_H-768_A-12_squad2
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,768
aodiniz/bert_uncased_L-4_H-768_A-12_squad2_covid-qna
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,769
aodiniz/bert_uncased_L-6_H-128_A-2_cord19-200616
aodiniz
bert
8
7
transformers
0
fill-mask
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'fill-mask', 'transformers', 'autotrain_compatible']
false
false
false
0
10,770
aodiniz/bert_uncased_L-6_H-128_A-2_cord19-200616_squad2
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,771
aodiniz/bert_uncased_L-6_H-128_A-2_cord19-200616_squad2_covid-qna
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,772
aodiniz/bert_uncased_L-6_H-128_A-2_squad2
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,773
aodiniz/bert_uncased_L-6_H-128_A-2_squad2_covid-qna
aodiniz
bert
12
9
transformers
0
question-answering
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
10,774
aogara/slai_transformer
aogara
null
2
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
true
76
10,775
aorona/dickens
aorona
gpt_neo
9
11
transformers
0
text-generation
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'gpt_neo', 'text-generation', 'transformers']
false
false
false
0
10,776
aoryabinin/aoryabinin_gpt_ai_dungeon_ru
aoryabinin
gpt2
7
11
transformers
0
text-generation
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'gpt2', 'text-generation', 'transformers']
false
false
false
0
10,777
aozorahime/my-new-model
aozorahime
bert
10
8
transformers
0
question-answering
true
false
false
apache-2.0
null
['xsum']
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'question-answering', 'dataset:xsum', 'transformers', 'generated_from_trainer', 'license:apache-2.0', 'autotrain_compatible']
true
true
true
921
10,778
apeguero/wav2vec2-large-xls-r-300m-tr-colab-3
apeguero
wav2vec2
10
6
transformers
0
automatic-speech-recognition
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'wav2vec2', 'automatic-speech-recognition', 'transformers']
false
false
false
0
10,779
apeguero/wav2vec2-large-xls-r-300m-tr-colab2
apeguero
null
5
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
10,780
aphuongle95/xlnet_effect_partial_new
aphuongle95
xlnet
7
13
transformers
0
text-generation
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'xlnet', 'text-generation', 'transformers']
false
false
false
0
10,781
aplnestrella/Aladdin-Bot
aplnestrella
gpt2
11
15
transformers
0
conversational
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'gpt2', 'text-generation', 'transformers', 'conversational']
false
true
true
14
10,782
apol/dalle-mini
apol
bart
12
22
transformers
7
text-to-image
false
false
true
null
['en']
null
null
1
0
0
1
1
1
0
['jax', 'bart', 'text2text-generation', 'en', 'arxiv:1910.13461', 'transformers', 'text-to-image', 'autotrain_compatible']
false
true
true
2,999
10,783
apoorvumang/kgt5-test
apoorvumang
null
2
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
true
6
10,784
apoorvumang/kgt5-wikikg90mv2
apoorvumang
t5
9
45
transformers
1
text2text-generation
true
true
false
mit
null
null
null
1
1
0
0
0
0
0
['pytorch', 'tf', 't5', 'text2text-generation', 'transformers', 'license:mit', 'autotrain_compatible']
false
true
true
5,827
10,785
app-test-user/test-tensorboard
app-test-user
null
39
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
['tensorboard']
false
false
true
1
10,786
appleternity/bert-base-uncased-finetuned-coda19
appleternity
bert
7
11
transformers
0
text-classification
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'text-classification', 'transformers']
false
false
false
0
10,787
appleternity/scibert-uncased-finetuned-coda19
appleternity
bert
7
10
transformers
0
text-classification
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'text-classification', 'transformers']
false
false
false
0
10,788
aqj213/t5-base-customised-1k-tokens-pisa-state-only-finetuned
aqj213
t5
13
4
transformers
0
text2text-generation
false
false
true
null
null
null
null
0
0
0
0
0
0
0
['jax', 'tensorboard', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
false
0
10,789
aqj213/t5-base-pisa-state-only-finetuned
aqj213
t5
14
4
transformers
0
text2text-generation
false
false
true
null
null
null
null
0
0
0
0
0
0
0
['jax', 'tensorboard', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
false
0
10,790
aqj213/t5-small-pisa-state-only-finetuned
aqj213
t5
17
4
transformers
0
text2text-generation
false
false
true
null
null
null
null
0
0
0
0
0
0
0
['jax', 'tensorboard', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
false
0
10,791
aqj213/t5-v1_1-large-last-1-step-pisa-state-only-finetuned
aqj213
t5
16
5
transformers
0
text2text-generation
false
false
true
null
null
null
null
0
0
0
0
0
0
0
['jax', 'tensorboard', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
false
0
10,792
aqj213/t5-v1_1-large-pisa-state-only-finetuned
aqj213
t5
18
4
transformers
0
text2text-generation
false
false
true
null
null
null
null
0
0
0
0
0
0
0
['jax', 'tensorboard', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
false
0
10,793
arampacha/DialoGPT-medium-simpsons
arampacha
gpt2
9
21
transformers
1
conversational
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'gpt2', 'text-generation', 'transformers', 'conversational', 'has_space']
false
true
true
153
10,794
arampacha/clip-rsicd-v5
arampacha
clip
9
32
transformers
0
zero-shot-image-classification
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'clip', 'zero-shot-image-classification', 'transformers']
false
false
false
0
10,795
arampacha/wav2vec2-large-xlsr-czech
arampacha
wav2vec2
9
41
transformers
1
automatic-speech-recognition
true
false
true
apache-2.0
['cs']
null
null
1
1
0
0
0
0
0
['pytorch', 'jax', 'wav2vec2', 'automatic-speech-recognition', 'cs', 'transformers', 'audio', 'speech', 'xlsr-fine-tuning-week', 'license:apache-2.0', 'model-index']
true
true
true
3,996
10,796
arampacha/wav2vec2-large-xlsr-ukrainian
arampacha
wav2vec2
9
9
transformers
1
automatic-speech-recognition
true
false
true
apache-2.0
['uk']
null
null
1
1
0
0
0
0
0
['pytorch', 'jax', 'wav2vec2', 'automatic-speech-recognition', 'uk', 'transformers', 'audio', 'speech', 'xlsr-fine-tuning-week', 'license:apache-2.0', 'model-index']
true
true
true
4,145
10,797
arampacha/wav2vec2-xls-r-1b-hy-cv
arampacha
wav2vec2
25
6
transformers
0
automatic-speech-recognition
true
false
false
apache-2.0
['hy']
['mozilla-foundation/common_voice_8_0']
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 'wav2vec2', 'automatic-speech-recognition', 'hy', 'dataset:mozilla-foundation/common_voice_8_0', 'transformers', 'mozilla-foundation/common_voice_8_0', 'generated_from_trainer', 'robust-speech-event', 'hf-asr-leaderboard', 'license:apache-2.0', 'model-index']
true
true
true
2,392
10,798
arampacha/wav2vec2-xls-r-1b-hy
arampacha
wav2vec2
48
8
transformers
1
automatic-speech-recognition
true
false
false
apache-2.0
['hy']
['common_voice']
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 'wav2vec2', 'automatic-speech-recognition', 'hy', 'dataset:common_voice', 'transformers', 'generated_from_trainer', 'hf-asr-leaderboard', 'mozilla-foundation/common_voice_8_0', 'robust-speech-event', 'license:apache-2.0', 'model-index']
true
true
true
2,029
10,799
arampacha/wav2vec2-xls-r-1b-ka
arampacha
wav2vec2
33
87
transformers
2
automatic-speech-recognition
true
false
false
apache-2.0
['ka']
['common_voice']
null
1
1
0
0
1
1
0
['pytorch', 'tensorboard', 'wav2vec2', 'automatic-speech-recognition', 'ka', 'dataset:common_voice', 'transformers', 'mozilla-foundation/common_voice_8_0', 'generated_from_trainer', 'robust-speech-event', 'hf-asr-leaderboard', 'license:apache-2.0', 'model-index']
true
true
true
2,048