Unnamed: 0
int64
0
245k
repo_id
stringlengths
4
122
author
stringlengths
2
42
model_type
stringlengths
2
34
files_per_repo
int64
0
77k
downloads_30d
int64
0
55.9M
library
stringlengths
2
37
likes
int64
0
8.48k
pipeline
stringlengths
5
30
pytorch
bool
2 classes
tensorflow
bool
2 classes
jax
bool
2 classes
license
stringlengths
2
33
languages
stringlengths
2
1.63k
datasets
stringlengths
2
5.05k
co2
stringlengths
3
342
prs_count
int64
0
168
prs_open
int64
0
121
prs_merged
int64
0
167
prs_closed
int64
0
35
discussions_count
int64
0
226
discussions_open
int64
0
155
discussions_closed
int64
0
76
tags
stringlengths
2
7.26k
has_model_index
bool
2 classes
has_metadata
bool
2 classes
has_text
bool
2 classes
text_length
int64
0
849k
12,900
d8oss/gamio-small
d8oss
gpt2
9
10
transformers
0
text-generation
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'gpt2', 'text-generation', 'transformers']
false
false
false
0
12,901
d8oss/giw-medium
d8oss
gpt2
9
10
transformers
0
text-generation
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'gpt2', 'text-generation', 'transformers']
false
false
false
0
12,902
dadada/mbart-finetuned-few-task
dadada
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,903
dadada/opus-mt-zh-en-ep1-renri-zh-to-en
dadada
marian
13
21
transformers
0
text2text-generation
true
false
false
apache-2.0
null
null
null
1
1
0
0
0
0
0
['pytorch', 'tensorboard', 'marian', 'text2text-generation', 'transformers', 'generated_from_trainer', 'license:apache-2.0', 'autotrain_compatible']
false
true
true
1,312
12,904
dadangheksa/finetuning-sentiment-model-3000-samples
dadangheksa
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,905
daekeun-ml/koelectra-small-v3-korsts
daekeun-ml
electra
8
13
transformers
0
sentence-similarity
true
false
false
cc-by-4.0
['ko']
['korsts']
null
0
0
0
0
0
0
0
['pytorch', 'electra', 'text-classification', 'ko', 'dataset:korsts', 'transformers', 'sentence-similarity', 'license:cc-by-4.0']
false
true
true
3,989
12,906
daekeun-ml/koelectra-small-v3-nsmc
daekeun-ml
electra
9
108
transformers
1
text-classification
true
false
false
mit
['ko']
['nsmc']
null
1
1
0
0
0
0
0
['pytorch', 'electra', 'text-classification', 'ko', 'dataset:nsmc', 'transformers', 'classification', 'license:mit']
false
true
true
4,575
12,907
dahele/tip
dahele
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,908
daisymzhang/bert-base-uncased-finetuned-swag
daisymzhang
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,909
daito/test
daito
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,910
dakshvar22/LaBSE
dakshvar22
bert
8
9
transformers
0
feature-extraction
true
true
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'tf', 'jax', 'bert', 'feature-extraction', 'transformers']
false
false
false
0
12,911
dalchob/distilbert-base-uncased-finetuned-ner
dalchob
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,912
dalle-mini/dalle-mini
dalle-mini
dallebart
10
89
transformers
300
text-to-image
false
false
true
apache-2.0
['en']
null
{'emissions': 7540, 'source': 'MLCo2 Machine Learning Impact calculator', 'geographical_location': 'East USA', 'hardware_used': 'TPU v3-8'}
41
5
6
29
3
3
0
['jax', 'dallebart', 'en', 'arxiv:2102.08981', 'arxiv:2012.09841', 'arxiv:1910.13461', 'arxiv:1910.09700', 'transformers', 'text-to-image', 'license:apache-2.0', 'co2_eq_emissions', 'has_space']
true
true
true
13,937
12,913
dalle-mini/vqgan_imagenet_f16_16384
dalle-mini
null
4
5,420
transformers
39
null
false
false
true
null
null
null
null
2
1
0
1
3
3
0
['jax', 'transformers', 'has_space']
false
false
true
1,458
12,914
damien-ir/ko-rest-electra-discriminator
damien-ir
electra
5
9
transformers
0
null
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'electra', 'pretraining', 'transformers']
false
false
false
0
12,915
damien-ir/ko-rest-electra-generator
damien-ir
electra
5
9
transformers
0
fill-mask
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'electra', 'fill-mask', 'transformers', 'autotrain_compatible']
false
false
false
0
12,916
damien-ir/kosentelectra-discriminator-v1
damien-ir
electra
5
7
transformers
0
null
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'electra', 'pretraining', 'transformers']
false
false
false
0
12,917
damien-ir/kosentelectra-discriminator-v2-mixed
damien-ir
electra
11
11
transformers
0
text-classification
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'electra', 'text-classification', 'transformers']
false
false
false
0
12,918
damien-ir/kosentelectra-discriminator-v2-small
damien-ir
electra
5
9
transformers
0
null
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'electra', 'pretraining', 'transformers']
false
false
false
0
12,919
damien-ir/kosentelectra-discriminator-v2
damien-ir
electra
5
7
transformers
0
null
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'electra', 'pretraining', 'transformers']
false
false
false
0
12,920
damien-ir/kosentelectra-discriminator-v3
damien-ir
electra
5
7
transformers
0
null
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'electra', 'pretraining', 'transformers']
false
false
false
0
12,921
damien-ir/kosentelectra-discriminator-v4
damien-ir
electra
5
8
transformers
0
null
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'electra', 'pretraining', 'transformers']
false
false
false
0
12,922
damien-ir/kosentelectra-discriminator-v5
damien-ir
electra
5
7
transformers
0
null
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'electra', 'pretraining', 'transformers']
false
false
false
0
12,923
damien-ir/kosentelectra-generator-v1
damien-ir
electra
5
9
transformers
0
fill-mask
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'electra', 'fill-mask', 'transformers', 'autotrain_compatible']
false
false
false
0
12,924
damien-ir/kosentelectra-generator-v2
damien-ir
electra
5
9
transformers
0
fill-mask
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'electra', 'fill-mask', 'transformers', 'autotrain_compatible']
false
false
false
0
12,925
damien-ir/kosentelectra-generator-v3
damien-ir
electra
5
9
transformers
0
fill-mask
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'electra', 'fill-mask', 'transformers', 'autotrain_compatible']
false
false
false
0
12,926
damien-ir/kosentelectra-generator-v4
damien-ir
electra
5
9
transformers
0
fill-mask
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'electra', 'fill-mask', 'transformers', 'autotrain_compatible']
false
false
false
0
12,927
damien-ir/kosentelectra-generator-v5
damien-ir
electra
5
9
transformers
0
fill-mask
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'electra', 'fill-mask', 'transformers', 'autotrain_compatible']
false
false
false
0
12,928
damlab/HIV_BERT
damlab
bert
8
8
transformers
0
fill-mask
true
false
false
mit
null
['damlab/HIV_FLT']
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'fill-mask', 'dataset:damlab/HIV_FLT', 'transformers', 'license:mit', 'autotrain_compatible']
false
true
true
4,259
12,929
damlab/HIV_PR_resist
damlab
bert
8
13
transformers
0
text-classification
true
false
false
mit
null
null
null
3
3
0
0
0
0
0
['pytorch', 'bert', 'text-classification', 'transformers', 'license:mit']
false
true
true
3,405
12,930
damlab/HIV_V3_Coreceptor
damlab
bert
8
10
transformers
0
text-classification
true
false
false
mit
null
null
null
3
3
0
0
0
0
0
['pytorch', 'bert', 'text-classification', 'transformers', 'license:mit']
false
true
true
4,024
12,931
damlab/HIV_V3_bodysite
damlab
bert
8
19
transformers
0
text-classification
true
false
false
null
null
['damlab/HIV_V3_bodysite']
null
1
1
0
0
0
0
0
['pytorch', 'bert', 'text-classification', 'dataset:damlab/HIV_V3_bodysite', 'transformers']
false
true
true
4,898
12,932
damondanieli/tweettest
damondanieli
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,933
danasone/bart-small-ru-en
danasone
bart
8
15
transformers
1
text2text-generation
true
false
false
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'bart', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
false
0
12,934
danasone/rubert-tiny-essay
danasone
bert
6
7
transformers
0
token-classification
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'token-classification', 'transformers', 'autotrain_compatible']
false
false
false
0
12,935
danasone/rubert-tiny-speech
danasone
bert
6
7
transformers
0
token-classification
true
false
false
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'bert', 'token-classification', 'transformers', 'autotrain_compatible']
false
false
false
0
12,936
danasone/testpush
danasone
vision-encoder-decoder
4
4
transformers
0
null
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'vision-encoder-decoder', 'transformers']
false
false
false
0
12,937
danbri/HelloWorld
danbri
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,938
danchang11/GPT2-TraditionalChat
danchang11
gpt2
9
10
transformers
0
text-generation
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'gpt2', 'transformers', 'text-generation']
false
true
true
11
12,939
dandelin/vilt-b32-finetuned-coco
dandelin
vilt
9
1,251
transformers
0
null
true
false
false
apache-2.0
null
null
null
0
0
0
0
1
1
0
['pytorch', 'vilt', 'arxiv:2102.03334', 'transformers', 'license:apache-2.0']
false
true
true
1,900
12,940
dandelin/vilt-b32-finetuned-flickr30k
dandelin
vilt
9
16
transformers
1
null
true
false
false
apache-2.0
null
null
null
0
0
0
0
1
1
0
['pytorch', 'vilt', 'arxiv:1505.04870', 'arxiv:2102.03334', 'transformers', 'license:apache-2.0']
false
true
true
2,103
12,941
dandelin/vilt-b32-finetuned-nlvr2
dandelin
vilt
9
566
transformers
1
null
true
false
false
apache-2.0
null
null
null
0
0
0
0
0
0
0
['pytorch', 'vilt', 'arxiv:2102.03334', 'transformers', 'license:apache-2.0', 'has_space']
false
true
true
2,071
12,942
dandelin/vilt-b32-finetuned-vqa
dandelin
vilt
9
43,302
transformers
172
visual-question-answering
true
false
false
apache-2.0
null
null
null
3
1
2
0
4
2
2
['pytorch', 'vilt', 'arxiv:2102.03334', 'transformers', 'visual-question-answering', 'license:apache-2.0', 'has_space']
false
true
true
1,834
12,943
dandelin/vilt-b32-mlm-itm
dandelin
vilt
4
63
transformers
2
null
true
false
false
apache-2.0
null
null
null
0
0
0
0
0
0
0
['pytorch', 'vilt', 'arxiv:2102.03334', 'transformers', 'license:apache-2.0']
false
true
true
1,140
12,944
dandelin/vilt-b32-mlm
dandelin
vilt
9
8,701
transformers
5
fill-mask
true
false
false
apache-2.0
null
null
null
2
1
0
1
0
0
0
['pytorch', 'vilt', 'fill-mask', 'arxiv:2102.03334', 'transformers', 'license:apache-2.0', 'autotrain_compatible', 'has_space']
false
true
true
2,822
12,945
danf0/MADE
danf0
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,946
danghuy1999/gpt2-viwiki
danghuy1999
gpt2
7
10
transformers
4
null
true
true
false
mit
['vi']
null
null
0
0
0
0
0
0
0
['pytorch', 'tf', 'gpt2', 'vi', 'transformers', 'gpt2-viwiki', 'license:mit']
false
true
true
3,121
12,947
dangvantuan/sentence-camembert-large
dangvantuan
camembert
9
11,421
transformers
22
sentence-similarity
true
true
false
apache-2.0
['fr']
['stsb_multi_mt']
null
4
3
1
0
5
5
0
['pytorch', 'tf', 'camembert', 'feature-extraction', 'fr', 'dataset:stsb_multi_mt', 'arxiv:1908.10084', 'transformers', 'Text', 'Sentence Similarity', 'Sentence-Embedding', 'camembert-large', 'license:apache-2.0', 'sentence-similarity', 'model-index', 'has_space']
true
true
true
4,704
12,948
danhsf/mt5-small-finetuned-en-to-ro-fp_false
danhsf
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,949
danhsf/mt5-small-finetuned-en-to-ro
danhsf
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,950
danhsf/mt5-small-finetuned-hi-to-en
danhsf
mt5
14
8
transformers
0
text2text-generation
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 'mt5', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
false
0
12,951
danhsf/t5-base-finetuned-en-to-hi
danhsf
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,952
danhsf/t5-small-finetuned-en-to-hi-maxlength-256
danhsf
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,953
danhsf/t5-small-finetuned-en-to-hi
danhsf
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,954
danhsf/t5-small-finetuned-en-to-pt
danhsf
t5
15
8
transformers
0
text2text-generation
true
false
false
apache-2.0
null
null
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 't5', 'text2text-generation', 'transformers', 'generated_from_trainer', 'license:apache-2.0', 'autotrain_compatible']
true
true
true
1,869
12,955
danhsf/t5-small-finetuned-en-to-ro-fp_false
danhsf
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,956
danhsf/t5-small-finetuned-en-to-ro-lr_2e-3-fp_false
danhsf
t5
12
9
transformers
0
text2text-generation
true
false
false
apache-2.0
null
['wmt16']
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 't5', 'text2text-generation', 'dataset:wmt16', 'transformers', 'generated_from_trainer', 'license:apache-2.0', 'model-index', 'autotrain_compatible']
true
true
true
2,525
12,957
danhsf/t5-small-finetuned-en-to-ro-lr_2e-3
danhsf
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,958
danhsf/t5-small-finetuned-en-to-ro-weight_decay
danhsf
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,959
danhsf/t5-small-finetuned-en-to-ro
danhsf
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,960
danhsf/t5-small-finetuned-hi-to-en
danhsf
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,961
danhsf/t5-small-finetuned-ro-to-en
danhsf
t5
14
8
transformers
0
text2text-generation
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
false
0
12,962
danicodes/autonlp-legal-text-summary-457311749
danicodes
pegasus
9
26
transformers
0
text2text-generation
true
false
false
null
['unk']
['danicodes/autonlp-data-legal-text-summary']
10.148805588432941
1
1
0
0
0
0
0
['pytorch', 'pegasus', 'text2text-generation', 'unk', 'dataset:danicodes/autonlp-data-legal-text-summary', 'transformers', 'autonlp', 'co2_eq_emissions', 'autotrain_compatible']
false
true
true
554
12,963
danie08/distilbert-base-uncased-finetuned-cola
danie08
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,964
danielbispov/t5-small-finetuned-fi-to-en
danielbispov
t5
14
8
transformers
0
text2text-generation
true
false
false
apache-2.0
null
['wmt19']
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 't5', 'text2text-generation', 'dataset:wmt19', 'transformers', 'generated_from_trainer', 'license:apache-2.0', 'model-index', 'autotrain_compatible']
true
true
true
1,254
12,965
danielbubiola/bangla_asr
danielbubiola
wav2vec2
28
6
transformers
0
automatic-speech-recognition
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 'wav2vec2', 'automatic-speech-recognition', 'transformers', 'generated_from_trainer']
true
true
true
1,733
12,966
danielbubiola/daniel_asr
danielbubiola
wav2vec2
12
6
transformers
0
automatic-speech-recognition
true
false
false
apache-2.0
null
null
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 'wav2vec2', 'automatic-speech-recognition', 'transformers', 'generated_from_trainer', 'license:apache-2.0']
true
true
true
1,621
12,967
danielf0123/distilbert-base-uncased-finetuned-ner
danielf0123
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,968
danielferreira/emotion-text-classification
danielferreira
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,969
daniellee93/text_summarization
daniellee93
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,970
danielo200/model_danielo
danielo200
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,971
danielrama/QuestionAnswering
danielrama
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,972
danielvasic/SemCorBert
danielvasic
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,973
danielvasic/distilbert-wordnet-uncased
danielvasic
distilbert
6
38
transformers
0
fill-mask
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'distilbert', 'fill-mask', 'transformers', 'autotrain_compatible']
false
false
false
0
12,974
danielvasic/en_acnl_electra_pipeline
danielvasic
null
22
0
spacy
0
token-classification
false
false
false
null
null
null
null
0
0
0
0
0
0
0
['en', 'spacy', 'token-classification', 'model-index']
false
false
false
0
12,975
danielvasic/en_acnl_roberta_pipeline
danielvasic
null
23
67
spacy
0
text-classification
false
false
false
null
null
null
null
0
0
0
0
0
0
0
['en', 'dataset:conll2012_ontonotesv5', 'spacy', 'token-classification', 'license:cc-by-4.0', 'text-classification', 'model-index']
false
false
false
0
12,976
danielvasic/hr_bertic_pipeline
danielvasic
null
19
1
spacy
0
token-classification
false
false
false
null
null
null
null
0
0
0
0
0
0
0
['hr', 'spacy', 'token-classification', 'model-index']
false
false
false
0
12,977
danielvasic/hr_hroberta_pipeline
danielvasic
null
19
1
spacy
1
token-classification
false
false
false
null
null
null
null
0
0
0
0
0
0
0
['hr', 'dataset:classla/hr500k', 'spacy', 'token-classification', 'license:cc', 'model-index']
false
false
false
0
12,978
danielvasic/hr_sroberta_pipeline
danielvasic
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,979
danildany/DialoGPT-small-MichaelScott
danildany
gpt2
11
12
transformers
0
conversational
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'gpt2', 'text-generation', 'transformers', 'conversational']
false
true
true
31
12,980
danimlk93/bert-finetuned-squad-accelerate
danimlk93
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,981
danish-col3/model_name
danish-col3
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,982
danlou/albert-xxlarge-v2-finetuned-csqa-ih
danlou
albert
9
5
transformers
1
multiple-choice
true
false
false
apache-2.0
null
null
null
0
0
0
0
0
0
0
['pytorch', 'albert', 'multiple-choice', 'transformers', 'generated_from_trainer', 'license:apache-2.0']
false
true
true
1,493
12,983
danlou/albert-xxlarge-v2-finetuned-csqa
danlou
albert
9
7
transformers
2
multiple-choice
true
false
false
apache-2.0
null
['commonsense_qa']
null
0
0
0
0
0
0
0
['pytorch', 'albert', 'multiple-choice', 'dataset:commonsense_qa', 'transformers', 'generated_from_trainer', 'license:apache-2.0']
false
true
true
1,499
12,984
danlou/aristo-roberta-finetuned-csqa
danlou
roberta
11
31
transformers
1
multiple-choice
true
false
false
mit
null
['commonsense_qa']
null
0
0
0
0
0
0
0
['pytorch', 'roberta', 'multiple-choice', 'dataset:commonsense_qa', 'transformers', 'generated_from_trainer', 'license:mit']
false
true
true
1,510
12,985
danlou/distilbert-base-uncased-finetuned-cola
danlou
distilbert
9
8
transformers
0
text-classification
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'distilbert', 'text-classification', 'transformers']
false
false
false
0
12,986
danlou/distilbert-base-uncased-finetuned-rte
danlou
distilbert
10
10
transformers
0
text-classification
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'distilbert', 'text-classification', 'transformers']
false
false
true
7
12,987
danlou/roberta-large-finetuned-csqa-ih
danlou
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,988
danlou/roberta-large-finetuned-csqa
danlou
roberta
11
5
transformers
0
multiple-choice
true
false
false
mit
null
['commonsense_qa']
null
0
0
0
0
0
0
0
['pytorch', 'roberta', 'multiple-choice', 'dataset:commonsense_qa', 'transformers', 'generated_from_trainer', 'license:mit']
false
true
true
1,487
12,989
dannersm/wav2vec2-large-xlsr-53-chilean-lessons
dannersm
wav2vec2
11
6
transformers
0
automatic-speech-recognition
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'wav2vec2', 'automatic-speech-recognition', 'transformers']
false
false
false
0
12,990
danny48/DialoGPT-small-harrypotter
danny48
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,991
danny481/DialoGPT-small-datnguyenchatbot
danny481
gpt2
10
12
transformers
0
conversational
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'gpt2', 'text-generation', 'transformers', 'conversational']
false
true
true
10
12,992
danny481/DialoGPT-small-harrypotter
danny481
gpt2
11
12
transformers
0
conversational
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'gpt2', 'text-generation', 'transformers', 'conversational']
false
true
true
23
12,993
danny481/Final_ChatBot
danny481
gpt2
11
17
transformers
0
conversational
true
false
false
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'gpt2', 'text-generation', 'transformers', 'conversational']
false
true
true
25
12,994
danny911kr/calm-base
danny911kr
t5
7
8
transformers
0
text2text-generation
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
true
518
12,995
danny911kr/calm-large
danny911kr
t5
7
11
transformers
0
text2text-generation
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
true
518
12,996
danny911kr/calm-mix-base
danny911kr
t5
7
8
transformers
0
text2text-generation
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
true
518
12,997
danny911kr/calm-mix-large
danny911kr
t5
7
8
transformers
0
text2text-generation
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
true
518
12,998
danny911kr/tapas_simsiam_mlm_1
danny911kr
tapas
6
8
transformers
0
feature-extraction
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'tapas', 'feature-extraction', 'transformers']
false
false
false
0
12,999
danny911kr/tapas_simsiam_mlm_2
danny911kr
tapas
6
8
transformers
0
feature-extraction
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'tapas', 'feature-extraction', 'transformers']
false
false
false
0