Unnamed: 0
int64
0
245k
repo_id
stringlengths
4
122
author
stringlengths
2
42
model_type
stringlengths
2
34
files_per_repo
int64
0
77k
downloads_30d
int64
0
55.9M
library
stringlengths
2
37
likes
int64
0
8.48k
pipeline
stringlengths
5
30
pytorch
bool
2 classes
tensorflow
bool
2 classes
jax
bool
2 classes
license
stringlengths
2
33
languages
stringlengths
2
1.63k
datasets
stringlengths
2
5.05k
co2
stringlengths
3
342
prs_count
int64
0
168
prs_open
int64
0
121
prs_merged
int64
0
167
prs_closed
int64
0
35
discussions_count
int64
0
226
discussions_open
int64
0
155
discussions_closed
int64
0
76
tags
stringlengths
2
7.26k
has_model_index
bool
2 classes
has_metadata
bool
2 classes
has_text
bool
2 classes
text_length
int64
0
849k
22,500
mateocolina/xlm-roberta-base-finetuned-marc-en
mateocolina
xlm-roberta
12
8
transformers
0
text-classification
true
false
false
mit
null
['amazon_reviews_multi']
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 'xlm-roberta', 'text-classification', 'dataset:amazon_reviews_multi', 'transformers', 'generated_from_trainer', 'license:mit']
true
true
true
1,275
22,501
matheusntg/character-bert-pt-normal
matheusntg
character_bert
3
5
transformers
0
null
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'character_bert', 'transformers']
false
false
false
0
22,502
matheusntg/character-bert-pt-small
matheusntg
character_bert
3
5
transformers
0
null
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'character_bert', 'transformers']
false
false
false
0
22,503
mathew/layoutlmv2-finetuned-funsd-1024
mathew
layoutlmv2
8
11
transformers
0
token-classification
true
false
false
cc-by-sa-4.0
null
null
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 'layoutlmv2', 'token-classification', 'transformers', 'generated_from_trainer', 'license:cc-by-sa-4.0', 'autotrain_compatible']
true
true
true
1,047
22,504
matprado/DialoGPT-small-rick-sanchez
matprado
gpt2
11
45
transformers
0
conversational
true
false
false
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'gpt2', 'text-generation', 'transformers', 'conversational']
false
true
true
6
22,505
matrix/test
matrix
null
2
0
null
0
null
false
false
false
apache-2.0
['en']
['bookcorpus', 'wikipedia']
null
0
0
0
0
0
0
0
['en', 'dataset:bookcorpus', 'dataset:wikipedia', 'arxiv:1810.04805', 'exbert', 'license:apache-2.0']
false
true
true
486
22,506
matrix/test1
matrix
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,507
matt-gm/gremon
matt-gm
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,508
mattchurgin/bert-finetuned-ner
mattchurgin
bert
9
9
transformers
0
token-classification
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'token-classification', 'transformers', 'autotrain_compatible']
false
false
false
0
22,509
mattchurgin/distilbert-mrpc
mattchurgin
distilbert
10
8
transformers
0
text-classification
true
false
false
apache-2.0
null
['glue']
null
1
1
0
0
0
0
0
['pytorch', 'distilbert', 'text-classification', 'dataset:glue', 'transformers', 'generated_from_trainer', 'license:apache-2.0', 'model-index']
true
true
true
2,091
22,510
mattchurgin/distilbert-sst2
mattchurgin
distilbert
10
10
transformers
0
text-classification
true
false
false
apache-2.0
null
['glue']
null
1
1
0
0
0
0
0
['pytorch', 'distilbert', 'text-classification', 'dataset:glue', 'transformers', 'generated_from_trainer', 'license:apache-2.0']
true
true
true
1,115
22,511
mattchurgin/wav2vec2-large-xls-r-300m-tr-mchurgin
mattchurgin
null
6
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,512
mattchurgin/xls-r-eng
mattchurgin
wav2vec2
19
11
transformers
0
automatic-speech-recognition
true
false
false
apache-2.0
['ab']
['common_voice']
null
0
0
0
0
0
0
0
['pytorch', 'wav2vec2', 'automatic-speech-recognition', 'ab', 'dataset:common_voice', 'transformers', 'mozilla-foundation/common_voice_7_0', 'generated_from_trainer', 'license:apache-2.0']
true
true
true
1,115
22,513
matthartman/matt_test_model
matthartman
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,514
matthewburke/korean_sentiment
matthewburke
electra
9
381
transformers
3
text-classification
true
false
false
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'electra', 'text-classification', 'transformers', 'has_space']
false
false
true
249
22,515
matthias-wright/gpt2
matthias-wright
null
12
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
true
897
22,516
matthias-wright/resnet
matthias-wright
null
7
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
['arxiv:1512.03385']
false
false
true
652
22,517
matthias-wright/stylegan2
matthias-wright
null
24
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
['arxiv:1912.04958']
false
false
true
680
22,518
matthias-wright/vgg
matthias-wright
null
4
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
['arxiv:1409.1556']
false
false
true
663
22,519
mattkuenzel/my-new-shiny-tokenizer
mattkuenzel
null
6
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,520
mattmcclean/distilbert-base-uncased-finetuned-emotion
mattmcclean
distilbert
12
8
transformers
0
text-classification
true
false
false
apache-2.0
null
['emotion']
null
1
1
0
0
0
0
0
['pytorch', 'tensorboard', 'distilbert', 'text-classification', 'dataset:emotion', 'transformers', 'generated_from_trainer', 'license:apache-2.0', 'model-index']
true
true
true
1,338
22,521
mattymoopers/distilbert-base-uncased-finetuned-ner
mattymoopers
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,522
maureengatu/maize_leaf_disease_classifier
maureengatu
null
6
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,523
maurice/PolitBERT
maurice
bert
12
15
transformers
0
fill-mask
true
false
true
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'jax', 'bert', 'fill-mask', 'transformers', 'autotrain_compatible']
false
false
true
1,435
22,524
maxdavish/test-model
maxdavish
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,525
maxidl/iML-distilbert-base-uncased-predict
maxidl
distilbert
3
7
transformers
0
text-classification
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'distilbert', 'text-classification', 'transformers']
false
false
false
0
22,526
maxidl/iML-distilbert-base-uncased-select
maxidl
distilbert
3
8
transformers
0
token-classification
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'distilbert', 'token-classification', 'transformers', 'autotrain_compatible']
false
false
false
0
22,527
maxidl/wav2vec2-large-xlsr-german
maxidl
wav2vec2
9
222
transformers
0
automatic-speech-recognition
true
false
true
apache-2.0
['de']
['common_voice']
null
2
2
0
0
0
0
0
['pytorch', 'jax', 'wav2vec2', 'automatic-speech-recognition', 'de', 'dataset:common_voice', 'transformers', 'audio', 'speech', 'xlsr-fine-tuning-week', 'license:apache-2.0']
true
true
true
6,647
22,528
maximedb/autonlp-vaccinchat-22134694
maximedb
roberta
11
10
transformers
0
text-classification
true
true
false
null
['nl']
['maximedb/autonlp-data-vaccinchat']
14.525955245648218
0
0
0
0
0
0
0
['pytorch', 'tf', 'roberta', 'text-classification', 'nl', 'dataset:maximedb/autonlp-data-vaccinchat', 'transformers', 'autonlp', 'co2_eq_emissions']
false
true
true
1,221
22,529
maximedb/drclips
maximedb
bert
18
5
transformers
0
fill-mask
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 'bert', 'fill-mask', 'transformers', 'autotrain_compatible']
false
false
false
0
22,530
maximedb/drclips128
maximedb
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,531
maximedb/drclips224
maximedb
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,532
maximedb/drclipsbis
maximedb
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,533
maximedb/mfaq-bert
maximedb
bert
9
7
transformers
0
feature-extraction
true
true
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'tf', 'bert', 'feature-extraction', 'transformers']
false
false
false
0
22,534
maximedb/mfaq-mbert
maximedb
bert
3
7
transformers
0
feature-extraction
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'feature-extraction', 'transformers']
false
false
false
0
22,535
maximedb/mqa-cross-encoder
maximedb
xlm-roberta
13
8
transformers
0
text-classification
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'xlm-roberta', 'text-classification', 'transformers']
false
false
true
6
22,536
maximedb/output
maximedb
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,537
maximedb/paws-x-all-x-en
maximedb
xlm-roberta
18
12
transformers
0
text-classification
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 'xlm-roberta', 'text-classification', 'transformers']
false
false
false
0
22,538
maximedb/paws-x-all
maximedb
xlm-roberta
15
7
transformers
0
text-classification
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 'xlm-roberta', 'text-classification', 'transformers']
false
false
false
0
22,539
maximedb/polyfaq_cross
maximedb
xlm-roberta
8
7
transformers
0
text-classification
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'xlm-roberta', 'text-classification', 'transformers']
false
false
false
0
22,540
maximedb/splade-roberta
maximedb
roberta
8
12
transformers
0
fill-mask
true
false
false
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'roberta', 'fill-mask', 'transformers', 'autotrain_compatible']
false
false
false
0
22,541
maximedb/test-2
maximedb
xlm-roberta
9
8
transformers
0
feature-extraction
true
true
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'tf', 'xlm-roberta', 'feature-extraction', 'transformers']
false
false
false
0
22,542
maximedb/wav2vec2-large-xls-r-300m-turkish-colab
maximedb
null
5
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,543
maxliaops/lasertagger-chinese
maxliaops
null
2
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
true
492
22,544
maxpe/bertin-roberta-base-spanish_semeval18_emodetection
maxpe
roberta
5
20
transformers
0
text-classification
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'roberta', 'text-classification', 'transformers']
false
false
true
3,516
22,545
maxpe/twitter-roberta-base_semeval18_emodetection
maxpe
roberta
9
8
transformers
0
text-classification
true
false
false
null
null
null
null
0
0
0
0
1
1
0
['pytorch', 'roberta', 'text-classification', 'transformers']
false
false
true
3,482
22,546
maxspaziani/bert-base-italian-uncased-finetuned-ComunaliRoma
maxspaziani
bert
9
6
transformers
0
fill-mask
true
false
false
mit
null
null
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 'bert', 'fill-mask', 'transformers', 'generated_from_trainer', 'license:mit', 'autotrain_compatible']
true
true
true
1,303
22,547
maxspaziani/bert-base-italian-xxl-uncased-finetuned-ComunaliRoma
maxspaziani
bert
8
5
transformers
0
fill-mask
true
false
false
mit
null
null
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 'bert', 'fill-mask', 'transformers', 'generated_from_trainer', 'license:mit', 'autotrain_compatible']
true
true
true
1,315
22,548
maxspaziani/distilroberta-base-finetuned-wikitext2
maxspaziani
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,549
maxwatcha/test
maxwatcha
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,550
maxxx2021/DialGPT-small-harrypotter
maxxx2021
gpt2
11
18
transformers
0
conversational
true
false
false
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'gpt2', 'text-generation', 'transformers', 'conversational']
false
true
true
28
22,551
mayu0007/pegasus_large_covid
mayu0007
pegasus
9
24
transformers
1
summarization
true
false
false
null
['en']
['CORD-19']
null
1
1
0
0
0
0
0
['pytorch', 'pegasus', 'text2text-generation', 'en', 'dataset:CORD-19', 'arxiv:1912.08777', 'transformers', 'summarization', 'autotrain_compatible']
false
true
true
17,929
22,552
mazancourt/politics-sentence-classifier
mazancourt
camembert
10
23
transformers
2
text-classification
true
false
false
null
['fr']
['mazancourt/autonlp-data-politics-sentence-classifier']
1.06099358268878
1
0
1
0
0
0
0
['pytorch', 'safetensors', 'camembert', 'text-classification', 'fr', 'dataset:mazancourt/autonlp-data-politics-sentence-classifier', 'transformers', 'autonlp', 'Text Classification', 'Politics', 'co2_eq_emissions']
false
true
true
2,188
22,553
maze/FastStyleTransfer
maze
null
7
0
null
0
null
false
false
false
mit
null
null
null
0
0
0
0
0
0
0
['arxiv:1603.08155', 'arxiv:1512.03385', 'license:mit', 'has_space']
false
true
true
3,732
22,554
mazerte/distilbert-base-uncased-finetuned-cola
mazerte
distilbert
3
3
transformers
0
text-classification
false
true
false
null
null
null
null
0
0
0
0
0
0
0
['tf', 'distilbert', 'text-classification', 'transformers']
false
false
false
0
22,555
mazula/test
mazula
null
2
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
true
11
22,556
mbartolo/electra-large-synqa
mbartolo
electra
8
20
transformers
3
question-answering
true
false
false
apache-2.0
['en']
['adversarial_qa', 'mbartolo/synQA', 'squad']
null
4
3
1
0
0
0
0
['pytorch', 'electra', 'question-answering', 'en', 'dataset:adversarial_qa', 'dataset:mbartolo/synQA', 'dataset:squad', 'arxiv:2002.00293', 'arxiv:2104.08678', 'transformers', 'license:apache-2.0', 'model-index', 'autotrain_compatible']
true
true
true
736
22,557
mbartolo/roberta-large-synqa-ext
mbartolo
roberta
9
1,016
transformers
1
question-answering
true
false
false
apache-2.0
['en']
['adversarial_qa', 'mbartolo/synQA', 'squad']
null
3
2
1
0
0
0
0
['pytorch', 'roberta', 'question-answering', 'en', 'dataset:adversarial_qa', 'dataset:mbartolo/synQA', 'dataset:squad', 'arxiv:2002.00293', 'arxiv:2104.08678', 'transformers', 'license:apache-2.0', 'model-index', 'autotrain_compatible']
true
true
true
711
22,558
mbartolo/roberta-large-synqa
mbartolo
roberta
9
75
transformers
1
question-answering
true
false
false
apache-2.0
['en']
['adversarial_qa', 'mbartolo/synQA', 'squad']
null
4
2
2
0
0
0
0
['pytorch', 'roberta', 'question-answering', 'en', 'dataset:adversarial_qa', 'dataset:mbartolo/synQA', 'dataset:squad', 'arxiv:2002.00293', 'arxiv:2104.08678', 'transformers', 'license:apache-2.0', 'model-index', 'autotrain_compatible']
true
true
true
663
22,559
mbateman/bert-finetuned-ner-accelerate
mbateman
bert
7
7
transformers
0
token-classification
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'token-classification', 'transformers', 'autotrain_compatible']
false
false
false
0
22,560
mbateman/bert-finetuned-ner
mbateman
bert
10
11
transformers
0
token-classification
true
false
false
apache-2.0
null
['conll2003']
null
0
0
0
0
0
0
0
['pytorch', 'bert', 'token-classification', 'dataset:conll2003', 'transformers', 'generated_from_trainer', 'license:apache-2.0', 'model-index', 'autotrain_compatible']
true
true
true
1,519
22,561
mbateman/bert-finetuned-squad-accelerate
mbateman
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,562
mbateman/bert-finetuned-squad
mbateman
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,563
mbateman/code-search-net-tokenizer
mbateman
null
6
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,564
mbateman/codeparrot-ds-accelerate
mbateman
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,565
mbateman/codeparrot-ds
mbateman
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,566
mbateman/distialbert-base-uncased-finetuned-squad-d5716d28
mbateman
distilbert
2
7
transformers
0
fill-mask
false
false
false
null
null
null
null
0
0
0
0
0
0
0
['distilbert', 'fill-mask', 'transformers', 'autotrain_compatible']
false
false
false
0
22,567
mbateman/distilbert-base-uncased-finetuned-imdb-accelerate
mbateman
distilbert
7
6
transformers
0
fill-mask
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'distilbert', 'fill-mask', 'transformers', 'autotrain_compatible']
false
false
false
0
22,568
mbateman/distilbert-base-uncased-finetuned-imdb
mbateman
distilbert
6
6
transformers
0
fill-mask
true
false
false
apache-2.0
null
['imdb']
null
0
0
0
0
0
0
0
['pytorch', 'distilbert', 'fill-mask', 'dataset:imdb', 'transformers', 'generated_from_trainer', 'license:apache-2.0', 'autotrain_compatible']
true
true
true
1,319
22,569
mbateman/distilbert-base-uncased-finetuned-squad-d5716d28
mbateman
distilbert
8
13
transformers
0
question-answering
true
false
false
apache-2.0
['en']
['squad']
null
0
0
0
0
0
0
0
['pytorch', 'distilbert', 'fill-mask', 'en', 'dataset:squad', 'arxiv:1910.01108', 'transformers', 'question-answering', 'license:apache-2.0', 'autotrain_compatible']
false
true
true
1,392
22,570
mbateman/dummy-model
mbateman
camembert
7
6
transformers
0
fill-mask
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'camembert', 'fill-mask', 'transformers', 'autotrain_compatible']
false
false
false
0
22,571
mbateman/dummy
mbateman
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,572
mbateman/marian-finetuned-kde4-en-to-fr-accelerate
mbateman
marian
8
10
transformers
0
text2text-generation
true
false
false
null
null
null
null
0
0
0
0
0
0
0
['pytorch', 'marian', 'text2text-generation', 'transformers', 'autotrain_compatible']
false
false
false
0
22,573
mbateman/marian-finetuned-kde4-en-to-fr
mbateman
marian
13
10
transformers
0
translation
true
false
false
apache-2.0
null
['kde4']
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 'marian', 'text2text-generation', 'dataset:kde4', 'transformers', 'translation', 'generated_from_trainer', 'license:apache-2.0', 'autotrain_compatible']
true
true
true
988
22,574
mbateman/mt5-small-finetuned-amazon-en-es
mbateman
mt5
13
12
transformers
0
summarization
true
false
false
apache-2.0
null
null
null
0
0
0
0
0
0
0
['pytorch', 'tensorboard', 'mt5', 'text2text-generation', 'transformers', 'summarization', 'generated_from_trainer', 'license:apache-2.0', 'autotrain_compatible']
true
true
true
1,997
22,575
mbateman/test-bert-finetuned-squad-accelerate
mbateman
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,576
mbeck/roberta-base-squad2
mbeck
roberta
12
145
transformers
0
question-answering
true
false
true
null
null
null
null
1
1
0
0
0
0
0
['pytorch', 'jax', 'roberta', 'question-answering', 'transformers', 'autotrain_compatible']
false
false
false
0
22,577
mbertoni/signaturizer
mbertoni
null
1
0
null
0
null
false
false
false
null
null
null
null
0
0
0
0
0
0
0
[]
false
false
false
0
22,578
mbeukman/xlm-roberta-base-finetuned-amharic-finetuned-ner-amharic
mbeukman
xlm-roberta
12
8
transformers
0
token-classification
true
false
false
null
['am']
['masakhaner']
null
1
1
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'am', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
9,154
22,579
mbeukman/xlm-roberta-base-finetuned-amharic-finetuned-ner-swahili
mbeukman
xlm-roberta
12
10
transformers
0
token-classification
true
false
false
null
['sw']
['masakhaner']
null
1
1
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'sw', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
12,109
22,580
mbeukman/xlm-roberta-base-finetuned-hausa-finetuned-ner-hausa
mbeukman
xlm-roberta
12
8
transformers
0
token-classification
true
false
false
null
['ha']
['masakhaner']
null
0
0
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'ha', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
9,111
22,581
mbeukman/xlm-roberta-base-finetuned-hausa-finetuned-ner-swahili
mbeukman
xlm-roberta
12
8
transformers
0
token-classification
true
false
false
null
['sw']
['masakhaner']
null
0
0
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'sw', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
11,732
22,582
mbeukman/xlm-roberta-base-finetuned-igbo-finetuned-ner-igbo
mbeukman
xlm-roberta
12
8
transformers
0
token-classification
true
false
false
null
['ig']
['masakhaner']
null
0
0
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'ig', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
9,112
22,583
mbeukman/xlm-roberta-base-finetuned-igbo-finetuned-ner-swahili
mbeukman
xlm-roberta
12
8
transformers
0
token-classification
true
false
false
null
['sw']
['masakhaner']
null
1
1
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'sw', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
11,728
22,584
mbeukman/xlm-roberta-base-finetuned-kinyarwanda-finetuned-ner-kinyarwanda
mbeukman
xlm-roberta
12
10
transformers
1
token-classification
true
false
false
null
['rw']
['masakhaner']
null
1
1
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'rw', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
9,256
22,585
mbeukman/xlm-roberta-base-finetuned-kinyarwanda-finetuned-ner-swahili
mbeukman
xlm-roberta
12
8
transformers
0
token-classification
true
false
false
null
['sw']
['masakhaner']
null
0
0
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'sw', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
11,756
22,586
mbeukman/xlm-roberta-base-finetuned-luganda-finetuned-ner-luganda
mbeukman
xlm-roberta
12
8
transformers
0
token-classification
true
false
false
null
['lug']
['masakhaner']
null
1
1
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'lug', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
9,152
22,587
mbeukman/xlm-roberta-base-finetuned-luganda-finetuned-ner-swahili
mbeukman
xlm-roberta
12
12
transformers
0
token-classification
true
false
false
null
['sw']
['masakhaner']
null
0
0
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'sw', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
11,740
22,588
mbeukman/xlm-roberta-base-finetuned-luo-finetuned-ner-luo
mbeukman
xlm-roberta
12
8
transformers
0
token-classification
true
false
false
null
['luo']
['masakhaner']
null
0
0
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'luo', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
9,233
22,589
mbeukman/xlm-roberta-base-finetuned-luo-finetuned-ner-swahili
mbeukman
xlm-roberta
12
8
transformers
0
token-classification
true
false
false
null
['sw']
['masakhaner']
null
1
1
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'sw', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
11,724
22,590
mbeukman/xlm-roberta-base-finetuned-naija-finetuned-ner-naija
mbeukman
xlm-roberta
12
8
transformers
0
token-classification
true
false
false
null
['pcm']
['masakhaner']
null
0
0
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'pcm', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
9,214
22,591
mbeukman/xlm-roberta-base-finetuned-naija-finetuned-ner-swahili
mbeukman
xlm-roberta
12
10
transformers
0
token-classification
true
false
false
null
['sw']
['masakhaner']
null
0
0
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'sw', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
11,732
22,592
mbeukman/xlm-roberta-base-finetuned-ner-amharic
mbeukman
xlm-roberta
12
27
transformers
1
token-classification
true
false
false
null
['am']
['masakhaner']
null
2
2
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'am', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
9,075
22,593
mbeukman/xlm-roberta-base-finetuned-ner-hausa
mbeukman
xlm-roberta
12
48
transformers
0
token-classification
true
false
false
null
['ha']
['masakhaner']
null
0
0
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'ha', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
9,040
22,594
mbeukman/xlm-roberta-base-finetuned-ner-igbo
mbeukman
xlm-roberta
12
11
transformers
0
token-classification
true
false
false
null
['ig']
['masakhaner']
null
0
0
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'ig', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
9,045
22,595
mbeukman/xlm-roberta-base-finetuned-ner-kinyarwanda
mbeukman
xlm-roberta
12
8
transformers
0
token-classification
true
false
false
null
['rw']
['masakhaner']
null
1
1
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'rw', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
9,161
22,596
mbeukman/xlm-roberta-base-finetuned-ner-luganda
mbeukman
xlm-roberta
12
11
transformers
0
token-classification
true
false
false
null
['lug']
['masakhaner']
null
1
1
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'lug', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
9,073
22,597
mbeukman/xlm-roberta-base-finetuned-ner-luo
mbeukman
xlm-roberta
12
11
transformers
0
token-classification
true
false
false
null
['luo']
['masakhaner']
null
0
0
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'luo', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
9,170
22,598
mbeukman/xlm-roberta-base-finetuned-ner-naija
mbeukman
xlm-roberta
12
59
transformers
0
token-classification
true
false
false
null
['pcm']
['masakhaner']
null
0
0
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'pcm', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
9,143
22,599
mbeukman/xlm-roberta-base-finetuned-ner-swahili
mbeukman
xlm-roberta
12
8
transformers
1
token-classification
true
false
false
null
['sw']
['masakhaner']
null
1
1
0
0
0
0
0
['pytorch', 'xlm-roberta', 'token-classification', 'sw', 'dataset:masakhaner', 'arxiv:2103.11811', 'transformers', 'NER', 'autotrain_compatible']
false
true
true
11,661