Unnamed: 0
int64
0
245k
repo_id
stringlengths
4
122
author
stringlengths
2
42
model_type
stringlengths
2
34
files_per_repo
int64
0
77k
downloads_30d
int64
0
55.9M
library
stringlengths
2
37
likes
int64
0
8.48k
pipeline
stringlengths
5
30
pytorch
bool
2 classes
tensorflow
bool
2 classes
jax
bool
2 classes
license
stringlengths
2
33
languages
stringlengths
2
1.63k
datasets
stringlengths
2
5.05k
co2
stringlengths
3
342
prs_count
int64
0
168
prs_open
int64
0
121
prs_merged
int64
0
167
prs_closed
int64
0
35
discussions_count
int64
0
226
discussions_open
int64
0
155
discussions_closed
int64
0
76
tags
stringlengths
2
7.26k
has_model_index
bool
2 classes
has_metadata
bool
2 classes
has_text
bool
2 classes
text_length
int64
0
849k
12,200
castorini/afriberta_base
castorini
xlm-roberta
8
27
transformers
0
fill-mask
true
true
false
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'tf', 'xlm-roberta', 'fill-mask', 'transformers', 'autotrain_compatible']
false
false
true
2,640
12,201
castorini/afriberta_large
castorini
xlm-roberta
8
283
transformers
4
fill-mask
true
true
false
mit
['om', 'am', 'rw', 'rn', 'ha', 'ig', 'so', 'sw', 'ti', 'yo', 'pcm', 'multilingual']
['castorini/afriberta-corpus']
null
3
1
2
0
0
0
0
['pytorch', 'tf', 'xlm-roberta', 'fill-mask', 'om', 'am', 'rw', 'rn', 'ha', 'ig', 'so', 'sw', 'ti', 'yo', 'pcm', 'multilingual', 'dataset:castorini/afriberta-corpus', 'transformers', 'license:mit', 'autotrain_compatible']
false
true
true
2,533
12,202
castorini/afriberta_small
castorini
xlm-roberta
8
30
transformers
0
fill-mask
true
true
false
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'tf', 'xlm-roberta', 'fill-mask', 'transformers', 'autotrain_compatible']
false
false
true
2,651
12,203
castorini/ance-dpr-context-multi
castorini
dpr
7
4
transformers
1
null
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'dpr', 'arxiv:2007.00808', 'transformers']
false
false
true
497
12,204
castorini/ance-dpr-question-multi
castorini
dpr
7
343
transformers
0
feature-extraction
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'dpr', 'feature-extraction', 'arxiv:2007.00808', 'transformers']
false
false
true
495
12,205
castorini/ance-msmarco-doc-firstp
castorini
roberta
8
6
transformers
0
null
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'roberta', 'arxiv:2007.00808', 'transformers']
false
false
true
495
12,206
castorini/ance-msmarco-doc-maxp
castorini
roberta
8
12
transformers
1
null
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'roberta', 'arxiv:2007.00808', 'transformers']
false
false
true
495
12,207
castorini/ance-msmarco-passage
castorini
roberta
8
1,179
transformers
0
null
true
false
false
null
['en']
null
null
2
0
1
1
0
0
0
['pytorch', 'roberta', 'en', 'arxiv:1910.09700', 'transformers']
false
true
true
4,251
12,208
castorini/azbert-base
castorini
bert
16
9
transformers
0
fill-mask
true
false
false
mit
['en']
null
null
1
1
0
0
0
0
0
['pytorch', 'tensorboard', 'bert', 'pretraining', 'en', 'transformers', 'azbert', 'fill-mask', 'license:mit']
false
true
true
1,813
12,209
castorini/bpr-nq-ctx-encoder
castorini
dpr
7
4
transformers
0
null
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'dpr', 'transformers']
false
false
true
270
12,210
castorini/bpr-nq-question-encoder
castorini
dpr
7
8
transformers
0
feature-extraction
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'dpr', 'feature-extraction', 'transformers']
false
false
true
270
12,211
castorini/dkrr-dpr-nq-retriever
castorini
bert
4
22
transformers
0
feature-extraction
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'feature-extraction', 'arxiv:2012.04584', 'transformers']
false
false
true
397
12,212
castorini/dkrr-dpr-tqa-retriever
castorini
bert
4
3
transformers
0
null
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'arxiv:2012.04584', 'transformers']
false
false
true
397
12,213
castorini/doc2query-t5-base-msmarco
castorini
t5
9
982
transformers
11
text2text-generation
true
false
true
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'jax', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
true
63
12,214
castorini/doc2query-t5-large-msmarco
castorini
t5
8
385
transformers
0
text2text-generation
true
false
true
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'jax', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
true
63
12,215
castorini/duot5-3b-med-msmarco
castorini
t5
8
9
transformers
0
feature-extraction
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 't5', 'feature-extraction', 'arxiv:2101.05667', 'transformers']
false
false
true
504
12,216
castorini/duot5-3b-msmarco
castorini
t5
8
56
transformers
0
feature-extraction
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 't5', 'feature-extraction', 'arxiv:2101.05667', 'transformers']
false
false
true
491
12,217
castorini/duot5-base-msmarco-10k
castorini
t5
6
24
transformers
0
text2text-generation
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
false
0
12,218
castorini/duot5-base-msmarco
castorini
t5
8
314
transformers
0
text2text-generation
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 't5', 'text2text-generation', 'arxiv:2101.05667', 'transformers', 'autotrain_compatible']
false
false
true
344
12,219
castorini/mdpr-tied-pft-msmarco
castorini
bert
6
486
transformers
0
feature-extraction
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'feature-extraction', 'transformers']
false
false
false
0
12,220
castorini/mdpr-passage-nq
castorini
dpr
6
16
transformers
1
null
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'dpr', 'transformers']
false
false
false
0
12,221
castorini/mdpr-question-nq
castorini
dpr
6
40
transformers
0
feature-extraction
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'dpr', 'feature-extraction', 'transformers']
false
false
false
0
12,222
castorini/monobert-large-msmarco-finetune-only
castorini
bert
8
32
transformers
0
text-classification
true
false
true
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'jax', 'bert', 'text-classification', 'transformers']
false
false
true
455
12,223
castorini/monobert-large-msmarco
castorini
null
7
745
transformers
2
null
true
false
false
null
null
null
null
1
0
0
1
0
0
0
['pytorch', 'transformers']
false
false
false
0
12,224
castorini/monot5-3b-med-msmarco
castorini
t5
7
11
transformers
2
feature-extraction
true
false
false
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 't5', 'feature-extraction', 'transformers']
false
false
true
455
12,225
castorini/monot5-3b-msmarco
castorini
t5
7
38
transformers
0
feature-extraction
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 't5', 'feature-extraction', 'transformers']
false
false
true
324
12,226
castorini/monot5-base-med-msmarco
castorini
t5
8
25
transformers
1
feature-extraction
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 't5', 'feature-extraction', 'transformers']
false
false
true
455
12,227
castorini/monot5-base-msmarco-10k
castorini
t5
8
23,399
transformers
9
text2text-generation
true
false
true
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'jax', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
true
803
12,228
castorini/monot5-base-msmarco
castorini
t5
8
1,674
transformers
4
text2text-generation
true
false
true
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'jax', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
true
788
12,229
castorini/monot5-large-msmarco-10k
castorini
t5
8
87
transformers
0
text2text-generation
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
true
805
12,230
castorini/monot5-large-msmarco
castorini
t5
8
392
transformers
0
feature-extraction
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 't5', 'feature-extraction', 'transformers']
false
false
true
662
12,231
castorini/t5-base-canard
castorini
t5
8
439
transformers
0
text2text-generation
true
false
true
null
null
null
null
1
1
0
0
1
0
1
['pytorch', 'jax', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
true
565
12,232
castorini/tct_colbert-msmarco
castorini
null
5
2,727
transformers
0
null
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'arxiv:2010.11386', 'transformers']
false
false
true
446
12,233
castorini/tct_colbert-v2-hn-msmarco
castorini
bert
7
8
transformers
0
feature-extraction
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'feature-extraction', 'transformers']
false
false
true
498
12,234
castorini/tct_colbert-v2-hnp-msmarco-r2
castorini
bert
8
8
transformers
0
feature-extraction
true
false
false
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'bert', 'feature-extraction', 'transformers']
false
false
true
923
12,235
castorini/tct_colbert-v2-hnp-msmarco
castorini
bert
7
2,530
transformers
3
feature-extraction
true
false
false
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'bert', 'feature-extraction', 'transformers']
false
false
true
498
12,236
castorini/tct_colbert-v2-msmarco-cqe
castorini
bert
5
56
transformers
2
feature-extraction
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'feature-extraction', 'transformers']
false
false
true
738
12,237
castorini/tct_colbert-v2-msmarco
castorini
bert
7
3,157
transformers
0
feature-extraction
true
false
false
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'bert', 'feature-extraction', 'transformers']
false
false
true
498
12,238
castorini/unicoil-msmarco-passage
castorini
bert
6
2,519
transformers
2
feature-extraction
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'feature-extraction', 'transformers']
false
false
false
0
12,239
castorini/unicoil-noexp-msmarco-passage
castorini
bert
6
1,862
transformers
0
feature-extraction
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'feature-extraction', 'transformers']
false
false
false
0
12,240
catalina-melo/distilBERT-severity-rating
catalina-melo
distilbert
7
3
transformers
0
text-classification
false
true
false
null
null
null
null
0
0
0
0
0
0
0
['tf', 'distilbert', 'text-classification', 'transformers']
false
false
false
0
12,241
catalina-melo/distilbert-base-uncased-finetuned-cola
catalina-melo
distilbert
7
3
transformers
0
text-classification
false
true
false
null
null
null
null
0
0
0
0
0
0
0
['tf', 'distilbert', 'text-classification', 'transformers']
false
false
false
0
12,242
cataremix15/distilbert-tiln-proj
cataremix15
distilbert
7
10
transformers
0
text-classification
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'distilbert', 'text-classification', 'transformers']
false
false
false
0
12,243
catharina/arcane
catharina
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,244
cathyccc/112
cathyccc
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,245
cathyhou/roberta-base-finetuned-cola
cathyhou
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,246
catlovehotpot/bert-base-uncased-finetuned-swag
catlovehotpot
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,247
catluvr/v
catluvr
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,248
potaycat/vinanews-gpt2-kinda
potaycat
gpt2
6
10
transformers
0
text-generation
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'gpt2', 'text-generation', 'transformers']
false
false
false
0
12,249
cb-insights-team/news_ner
cb-insights-team
null
2
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
true
67
12,250
cbakos/bert-base-uncased-finetuned-ner
cbakos
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,251
cbrew475/mpnet-metric
cbrew475
mpnet
96
8
transformers
0
text-classification
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'mpnet', 'text-classification', 'transformers']
false
false
false
0
12,252
cc4/bert-tok
cc4
null
6
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,253
cc4/my-new-shiny-tokenizer-2
cc4
null
6
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,254
cc4/my-new-shiny-tokenizer-3
cc4
null
6
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,255
cc4/my-new-shiny-tokenizer
cc4
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,256
ccdv/lsg-legal-base-uncased-4096
ccdv
bert
10
42
transformers
2
fill-mask
true
false
false
null
['en']
null
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'pretraining', 'custom_code', 'en', 'arxiv:2210.15497', 'transformers', 'long context', 'legal', 'fill-mask', 'has_space']
false
true
true
6,323
12,257
ccdv/lsg-legal-small-uncased-4096
ccdv
bert
10
5
transformers
0
fill-mask
true
false
false
null
['en']
null
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'pretraining', 'custom_code', 'en', 'arxiv:2210.15497', 'transformers', 'long context', 'legal', 'fill-mask']
false
true
true
6,354
12,258
ccdv/lsg-bart-base-4096
ccdv
bart
11
449
transformers
1
fill-mask
true
false
false
null
['en']
null
null
0
0
0
0
1
1
0
['pytorch', 'bart', 'text2text-generation', 'custom_code', 'en', 'arxiv:2210.15497', 'arxiv:1910.13461', 'transformers', 'summarization', 'long context', 'fill-mask', 'autotrain_compatible']
false
true
true
5,853
12,259
ccdv/lsg-bart-large-4096
ccdv
bart
11
38
transformers
0
fill-mask
true
false
false
null
['en']
null
null
0
0
0
0
0
0
0
['pytorch', 'bart', 'text2text-generation', 'custom_code', 'en', 'arxiv:2210.15497', 'arxiv:1910.13461', 'transformers', 'summarization', 'long context', 'fill-mask', 'autotrain_compatible']
false
true
true
5,865
12,260
ccdv/lsg-barthez-4096
ccdv
mbart
10
41
transformers
0
fill-mask
true
false
false
null
['fr']
null
null
0
0
0
0
0
0
0
['pytorch', 'mbart', 'text2text-generation', 'custom_code', 'fr', 'arxiv:2210.15497', 'transformers', 'summarization', 'bart', 'long context', 'fill-mask', 'autotrain_compatible']
false
true
true
5,410
12,261
ccdv/lsg-camembert-base-4096
ccdv
camembert
10
15
transformers
2
fill-mask
true
false
false
null
['fr']
null
null
0
0
0
0
0
0
0
['pytorch', 'camembert', 'fill-mask', 'custom_code', 'fr', 'arxiv:2210.15497', 'transformers', 'long context', 'autotrain_compatible']
false
true
true
6,151
12,262
ccdv/lsg-base-4096
ccdv
roberta
11
11
transformers
1
fill-mask
true
false
false
null
['en']
null
null
0
0
0
0
0
0
0
['pytorch', 'roberta', 'fill-mask', 'custom_code', 'en', 'arxiv:2210.15497', 'transformers', 'long context', 'autotrain_compatible']
false
true
true
5,615
12,263
ccdv/lsg-pegasus-large-4096
ccdv
pegasus
10
45
transformers
0
fill-mask
true
false
false
null
['en']
null
null
0
0
0
0
0
0
0
['pytorch', 'pegasus', 'text2text-generation', 'custom_code', 'en', 'arxiv:2210.15497', 'arxiv:1912.08777', 'transformers', 'summarization', 'long context', 'fill-mask', 'autotrain_compatible']
false
true
true
5,268
12,264
ccoreilly/wav2vec2-large-100k-voxpopuli-catala
ccoreilly
wav2vec2
10
52
transformers
1
automatic-speech-recognition
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'wav2vec2', 'automatic-speech-recognition', 'ca', 'dataset:common_voice', 'dataset:parlament_parla', 'transformers', 'audio', 'speech', 'speech-to-text', 'license:apache-2.0', 'model-index']
false
false
false
0
12,265
ccoreilly/wav2vec2-large-xlsr-catala
ccoreilly
wav2vec2
10
18
transformers
0
automatic-speech-recognition
true
false
true
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'jax', 'wav2vec2', 'automatic-speech-recognition', 'ca', 'dataset:common_voice', 'dataset:parlament_parla', 'transformers', 'audio', 'speech', 'xlsr-fine-tuning-week', 'license:apache-2.0', 'model-index']
false
false
false
0
12,266
cd-dvd/testmodel2
cd-dvd
gpt_neo
9
10
transformers
0
text-generation
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'gpt_neo', 'text-generation', 'transformers', 'Text Generation']
false
true
true
104
12,267
cdiamana/distilbert-base-uncased-finetuned-squad
cdiamana
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,268
cdwyer1bod/first_model
cdwyer1bod
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,269
cecilia/q
cecilia
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,270
cedpsam/chatbot_fr
cedpsam
gpt2
9
610
transformers
0
conversational
true
false
true
null
['fr']
null
null
1
1
0
0
0
0
0
['pytorch', 'jax', 'gpt2', 'text-generation', 'fr', 'transformers', 'conversational', 'has_space']
false
true
true
1,673
12,271
celential/erc
celential
albert
5
13
transformers
0
text-classification
true
false
false
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'albert', 'text-classification', 'transformers']
false
false
false
0
12,272
celine/emotion-detection_indobenchmark-indobert-lite-base-p1
celine
albert
7
7
transformers
0
text-classification
false
true
false
null
null
null
null
0
0
0
0
0
0
0
['tf', 'albert', 'text-classification', 'transformers']
false
false
false
0
12,273
celine/hate-speech_indobenchmark-indobert-lite-base-p1
celine
albert
7
7
transformers
0
text-classification
false
true
false
null
null
null
null
0
0
0
0
0
0
0
['tf', 'albert', 'text-classification', 'transformers']
false
false
false
0
12,274
celineguan/wav2vec2-base-timit-demo-colab
celineguan
null
4
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,275
celinelee/answer-extraction
celinelee
distilbert
3
16
transformers
0
question-answering
true
false
false
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'distilbert', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
12,276
celtics1863/env-bert-chinese
celtics1863
bert
8
10
transformers
2
fill-mask
true
false
false
null
['zh']
null
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'fill-mask', 'zh', 'transformers', 'pretrain', 'environment', 'autotrain_compatible']
false
true
true
195
12,277
celtics1863/env-bert-cls-chinese
celtics1863
bert
10
13
transformers
0
text-classification
true
false
false
null
['zh']
null
null
1
1
0
0
0
0
0
['pytorch', 'bert', 'text-classification', 'zh', 'transformers', 'environment', 'multi-class', 'classification']
false
true
true
200
12,278
celtics1863/env-bert-large-chinese
celtics1863
bert
7
5
transformers
0
fill-mask
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'fill-mask', 'transformers', 'autotrain_compatible']
false
false
false
0
12,279
celtics1863/env-bert-topic
celtics1863
bert
10
22
transformers
0
text-classification
true
false
false
null
['zh']
null
null
1
1
0
0
0
0
0
['pytorch', 'bert', 'text-classification', 'zh', 'transformers', 'pretrain', 'environment', 'classification', 'topic classification']
false
true
true
646
12,280
cemdenizsel/10k-finetuned-bert-model
cemdenizsel
bert
3
9
transformers
0
text-classification
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'text-classification', 'transformers']
false
false
false
0
12,281
cemdenizsel/51k-finetuned-bert-model
cemdenizsel
bert
3
9
transformers
0
text-classification
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'text-classification', 'transformers']
false
false
false
0
12,282
cemdenizsel/51k-pretrained-bert-model
cemdenizsel
bert
3
9
transformers
0
text-classification
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'text-classification', 'transformers']
false
false
false
0
12,283
cemigo/cemigo-test-model
cemigo
null
2
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
true
65
12,284
cemigo/test-bart
cemigo
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,285
centon21/DialoGPT-small-harrypotter
centon21
null
11
4
transformers
0
conversational
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'transformers', 'conversational']
false
true
true
29
12,286
ceostroff/harry-potter-gpt2-fanfiction
ceostroff
gpt2
12
95
transformers
0
text-generation
true
true
true
mit
['en']
null
null
1
1
0
0
0
0
0
['pytorch', 'tf', 'jax', 'gpt2', 'text-generation', 'en', 'transformers', 'harry-potter', 'license:mit']
false
true
true
307
12,287
cerulianx/dall-e
cerulianx
null
6
0
null
0
null
false
false
false
afl-3.0
null
null
null
0
0
0
0
0
0
0
['license:afl-3.0']
false
true
false
0
12,288
ceshine/TinyBERT_L-4_H-312_v2-distill-AllNLI
ceshine
bert
9
10
transformers
0
feature-extraction
true
false
true
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'bert', 'feature-extraction', 'transformers']
false
false
true
1,043
12,289
ceshine/t5-paraphrase-paws-msrp-opinosis
ceshine
t5
9
22
transformers
3
text2text-generation
true
false
true
apache-2.0
['en']
null
null
1
0
1
0
0
0
0
['pytorch', 'jax', 'safetensors', 't5', 'text2text-generation', 'en', 'transformers', 'paraphrasing', 'paraphrase', 'license:apache-2.0', 'autotrain_compatible', 'has_space']
false
true
true
187
12,290
ceshine/t5-paraphrase-quora-paws
ceshine
t5
9
127
transformers
1
text2text-generation
true
false
true
apache-2.0
['en']
null
null
1
0
1
0
0
0
0
['pytorch', 'jax', 'safetensors', 't5', 'text2text-generation', 'en', 'transformers', 'paraphrasing', 'paraphrase', 'license:apache-2.0', 'autotrain_compatible', 'has_space']
false
true
true
177
12,291
cestwc/bart-base-concise-baseline
cestwc
bart
7
9
transformers
0
text2text-generation
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'bart', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
false
0
12,292
cestwc/bart-paranmt5m
cestwc
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
12,293
cestwc/roberta-base-bigram-binary
cestwc
roberta
9
11
transformers
0
token-classification
true
false
false
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'roberta', 'token-classification', 'transformers', 'autotrain_compatible']
false
false
false
0
12,294
cestwc/roberta-base-unigram-quaternary
cestwc
roberta
9
8
transformers
0
token-classification
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'roberta', 'token-classification', 'transformers', 'autotrain_compatible']
false
false
false
0
12,295
cestwc/roberta-base-unigram-ternary
cestwc
roberta
8
11
transformers
0
token-classification
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'roberta', 'token-classification', 'transformers', 'autotrain_compatible']
false
false
false
0
12,296
ceyda/clip-reply
ceyda
hybrid-clip
3
7
transformers
0
null
false
false
true
null
null
null
null
0
0
0
0
0
0
0
['jax', 'hybrid-clip', 'transformers', 'has_space']
false
false
false
0
12,297
ceyda/wav2vec2-base-760-turkish
ceyda
wav2vec2
9
16
transformers
1
automatic-speech-recognition
true
false
false
apache-2.0
['tr']
['common_voice']
null
1
0
1
0
1
1
0
['pytorch', 'safetensors', 'wav2vec2', 'automatic-speech-recognition', 'tr', 'dataset:common_voice', 'transformers', 'audio', 'speech', 'xlsr-fine-tuning-week', 'license:apache-2.0', 'model-index']
true
true
true
3,614
12,298
ceyda/wav2vec2-base-760
ceyda
wav2vec2
4
11
transformers
1
feature-extraction
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'wav2vec2', 'feature-extraction', 'transformers']
false
false
true
46
12,299
ceyda/wav2vec2-large-xlsr-53-turkish
ceyda
wav2vec2
9
17
transformers
1
automatic-speech-recognition
true
false
true
apache-2.0
['tr']
['common_voice']
null
0
0
0
0
0
0
0
['pytorch', 'jax', 'wav2vec2', 'automatic-speech-recognition', 'tr', 'dataset:common_voice', 'transformers', 'audio', 'speech', 'xlsr-fine-tuning-week', 'license:apache-2.0', 'model-index']
true
true
true
3,383