Unnamed: 0
int64 0
245k
| repo_id
stringlengths 4
122
| author
stringlengths 2
42
⌀ | model_type
stringlengths 2
34
⌀ | files_per_repo
int64 0
77k
| downloads_30d
int64 0
55.9M
| library
stringlengths 2
37
⌀ | likes
int64 0
8.48k
| pipeline
stringlengths 5
30
⌀ | pytorch
bool 2
classes | tensorflow
bool 2
classes | jax
bool 2
classes | license
stringlengths 2
33
⌀ | languages
stringlengths 2
1.63k
⌀ | datasets
stringlengths 2
5.05k
⌀ | co2
stringlengths 3
342
⌀ | prs_count
int64 0
168
| prs_open
int64 0
121
| prs_merged
int64 0
167
| prs_closed
int64 0
35
| discussions_count
int64 0
226
| discussions_open
int64 0
155
| discussions_closed
int64 0
76
| tags
stringlengths 2
7.26k
| has_model_index
bool 2
classes | has_metadata
bool 2
classes | has_text
bool 2
classes | text_length
int64 0
849k
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
21,200 |
keras-io/low-light-image-enhancement
|
keras-io
| null | 7 | 5 |
keras
| 13 |
image-to-image
| false | false | false |
apache-2.0
| null | null | null | 1 | 0 | 0 | 1 | 1 | 1 | 0 |
['keras', 'image-to-image', 'license:apache-2.0', 'has_space']
| false | true | true | 2,385 |
21,201 |
keras-io/lowlight-enhance-mirnet
|
keras-io
| null | 11 | 52 |
keras
| 15 |
image-to-image
| false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['tensorboard', 'keras', 'image-to-image', 'has_space']
| false | true | true | 1,616 |
21,202 |
keras-io/mobile-vit-xxs
|
keras-io
| null | 6 | 31 |
keras
| 0 |
image-classification
| false | false | false |
['cc0-1.0']
| null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['keras', 'computer-vision', 'image-classification', 'license:cc0-1.0', 'has_space']
| false | true | true | 1,044 |
21,203 |
keras-io/monocular-depth-estimation
|
keras-io
| null | 11 | 42 |
keras
| 11 |
image-segmentation
| false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 1 | 1 | 0 |
['tensorboard', 'keras', 'image-segmentation', 'has_space']
| false | true | true | 2,022 |
21,204 |
keras-io/multimodal-entailment
|
keras-io
| null | 7 | 5 |
keras
| 3 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['keras', 'multimodal-entailment', 'generic', 'has_space']
| false | true | true | 1,331 |
21,205 |
keras-io/ner-with-transformers
|
keras-io
| null | 7 | 2 |
keras
| 1 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['keras', 'multimodal-entailment', 'generic', 'has_space']
| false | true | true | 1,197 |
21,206 |
keras-io/ocr-for-captcha
|
keras-io
| null | 8 | 48 |
keras
| 14 |
image-to-text
| false | false | false |
['cc0-1.0']
| null | null | null | 1 | 0 | 1 | 0 | 0 | 0 | 0 |
['keras', 'ocr', 'computer vision', 'object detection', 'image-to-text', 'license:cc0-1.0', 'has_space']
| false | true | true | 765 |
21,207 |
keras-io/pixel-cnn-mnist
|
keras-io
| null | 6 | 2 |
keras
| 0 | null | false | false | false |
['cc0-1.0']
| null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['keras', 'convnet', 'mnist', 'generative', 'license:cc0-1.0', 'has_space']
| false | true | true | 555 |
21,208 |
keras-io/pointnet_segmentation
|
keras-io
| null | 6 | 0 |
keras
| 3 | null | false | false | false |
cc0-1.0
| null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['keras', 'arxiv:1612.00593', 'arxiv:1506.02025', 'pointnet', 'segmentation', '3d', 'image', 'license:cc0-1.0', 'has_space']
| false | true | true | 1,990 |
21,209 |
keras-io/ppo-cartpole
|
keras-io
| null | 7 | 1 |
keras
| 0 | null | false | false | false |
['cc0-1.0']
| null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['keras', 'reinforcement learning', 'proximal policy optimization', 'license:cc0-1.0']
| false | true | true | 1,719 |
21,210 |
keras-io/randaugment
|
keras-io
| null | 6 | 0 |
keras
| 0 | null | false | false | false |
apache-2.0
| null |
['cifar10']
| null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['keras', 'dataset:cifar10', 'arxiv:1909.13719', 'arxiv:1911.04252', 'arxiv:1904.12848', 'RandAugment', 'Image Classification', 'license:apache-2.0', 'has_space']
| false | true | true | 1,180 |
21,211 |
keras-io/recommender-transformers
|
keras-io
| null | 5 | 0 |
keras
| 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['keras']
| false | false | false | 0 |
21,212 |
keras-io/semantic-segmentation
|
keras-io
| null | 17 | 129 |
generic
| 15 |
image-segmentation
| false | true | false |
cc0-1.0
| null | null | null | 0 | 0 | 0 | 0 | 1 | 1 | 0 |
['tf', 'generic', 'image-segmentation', 'license:cc0-1.0', 'has_space']
| false | true | true | 1,847 |
21,213 |
keras-io/semi-supervised-classification-simclr
|
keras-io
| null | 6 | 28 |
keras
| 1 |
image-classification
| false | false | false |
apache-2.0
| null |
['STL-10']
| null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['keras', 'dataset:STL-10', 'image-classification', 'license:apache-2.0', 'has_space']
| false | true | true | 1,983 |
21,214 |
keras-io/sentiment-analysis
|
keras-io
|
distilbert
| 14 | 12 |
transformers
| 0 |
text-classification
| false | true | false |
apache-2.0
| null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['tf', 'tensorboard', 'distilbert', 'text-classification', 'transformers', 'generated_from_keras_callback', 'license:apache-2.0']
| true | true | true | 1,462 |
21,215 |
keras-io/simple-mnist-convnet
|
keras-io
| null | 7 | 0 |
keras
| 0 | null | false | false | false |
['cc0-1.0']
| null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['keras', 'lstm', 'license:cc0-1.0']
| false | true | true | 268 |
21,216 |
keras-io/super-resolution
|
keras-io
| null | 6 | 65 |
keras
| 12 |
image-to-image
| false | false | false |
mit
| null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['keras', 'image-to-image', 'license:mit', 'has_space']
| false | true | true | 483 |
21,217 |
keras-io/supervised-contrastive-learning-cifar10
|
keras-io
| null | 6 | 6 |
keras
| 3 |
image-classification
| false | false | false |
apache-2.0
| null |
['cifar10']
| null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['keras', 'dataset:cifar10', 'arxiv:2004.11362', 'image-classification', 'license:apache-2.0', 'has_space']
| false | true | true | 625 |
21,218 |
keras-io/swin-transformers
|
keras-io
| null | 6 | 0 |
transformers
| 1 |
image-classification
| false | false | false |
cc0-1.0
| null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['keras', 'arxiv:2103.14030', 'transformers', 'swin-transformers', 'Keras', 'image-classification', 'license:cc0-1.0', 'has_space']
| false | true | true | 1,479 |
21,219 |
keras-io/text-generation-miniature-gpt
|
keras-io
|
gpt2
| 7 | 3 |
keras
| 0 |
text-generation
| false | false | false |
gpl
|
['en']
| null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['keras', 'gpt2', 'en', 'gpt', 'text-generation', 'license:gpl', 'has_space']
| false | true | true | 641 |
21,220 |
keras-io/time-series-anomaly-detection-autoencoder
|
keras-io
| null | 6 | 4 |
keras
| 0 | null | false | false | false |
['cc0-1.0']
| null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['keras', 'autoencoder', 'time series', 'anomaly detection', 'license:cc0-1.0']
| false | true | true | 486 |
21,221 |
keras-io/timeseries_transformer_classification
|
keras-io
| null | 6 | 0 |
transformers
| 1 | null | false | false | false |
cc0-1.0
| null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['keras', 'transformers', 'time-series', 'license:cc0-1.0', 'has_space']
| false | true | true | 800 |
21,222 |
keras-io/transformers-qa
|
keras-io
|
distilbert
| 8 | 220 |
transformers
| 3 |
question-answering
| false | true | false |
apache-2.0
| null |
['squad']
| null | 1 | 1 | 0 | 0 | 0 | 0 | 0 |
['tf', 'distilbert', 'question-answering', 'dataset:squad', 'transformers', 'generated_from_keras_callback', 'license:apache-2.0', 'model-index', 'autotrain_compatible', 'has_space']
| true | true | true | 1,524 |
21,223 |
keras-io/video-classification-cnn-rnn
|
keras-io
| null | 11 | 2 |
keras
| 3 |
video-classification
| false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['tensorboard', 'keras', 'computer-vision', 'video-classification', 'has_space']
| false | true | true | 1,495 |
21,224 |
keras-io/video-vision-transformer
|
keras-io
| null | 6 | 3 |
keras
| 6 | null | false | false | false |
apache-2.0
| null | null | null | 0 | 0 | 0 | 0 | 1 | 1 | 0 |
['keras', 'arxiv:2103.15691', 'license:apache-2.0', 'has_space']
| false | true | true | 1,066 |
21,225 |
keras-io/vit-small-ds
|
keras-io
| null | 6 | 1 |
keras
| 0 |
image-classification
| false | false | false |
apache-2.0
| null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['keras', 'arxiv:2010.11929', 'arxiv:2112.13492', 'image-classification', 'license:apache-2.0']
| false | true | true | 2,915 |
21,226 |
keras-io/vit_small_ds_v2
|
keras-io
| null | 6 | 6 |
keras
| 1 |
image-classification
| false | false | false |
apache-2.0
| null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['keras', 'arxiv:2010.11929', 'arxiv:2112.13492', 'image-classification', 'license:apache-2.0', 'has_space']
| false | true | true | 2,160 |
21,227 |
keshan/SinhalaBERTo
|
keshan
|
roberta
| 11 | 127 |
transformers
| 0 |
fill-mask
| true | true | true | null |
['si']
|
['oscar']
| null | 1 | 0 | 1 | 0 | 0 | 0 | 0 |
['pytorch', 'tf', 'jax', 'safetensors', 'roberta', 'fill-mask', 'si', 'dataset:oscar', 'arxiv:1907.11692', 'transformers', 'SinhalaBERTo', 'Sinhala', 'autotrain_compatible']
| false | true | true | 966 |
21,228 |
keshan/sinhala-gpt2-newswire
|
keshan
|
gpt2
| 5 | 32 |
transformers
| 2 |
text-generation
| true | false | false | null |
['si']
| null | null | 1 | 1 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'gpt2', 'text-generation', 'si', 'transformers', 'sinhala', 'has_space']
| false | true | true | 175 |
21,229 |
keshan/sinhala-gpt2
|
keshan
|
gpt2
| 16 | 28 |
transformers
| 1 |
feature-extraction
| true | true | true | null |
['si']
|
['mc4']
| null | 1 | 1 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'tf', 'jax', 'tensorboard', 'gpt2', 'feature-extraction', 'si', 'dataset:mc4', 'transformers', 'Sinhala', 'text-generation']
| false | true | true | 757 |
21,230 |
keshan/sinhala-roberta-mc4
|
keshan
|
roberta
| 12 | 6 |
transformers
| 0 |
fill-mask
| true | false | true |
cc-by-4.0
|
['si']
| null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'jax', 'tensorboard', 'roberta', 'fill-mask', 'si', 'transformers', 'sinhala', 'license:cc-by-4.0', 'autotrain_compatible']
| false | true | true | 33 |
21,231 |
keshan/sinhala-roberta-oscar
|
keshan
|
roberta
| 12 | 18 |
transformers
| 0 |
fill-mask
| true | false | true | null |
['si']
|
['oscar']
| null | 1 | 1 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'jax', 'tensorboard', 'roberta', 'fill-mask', 'si', 'dataset:oscar', 'arxiv:1907.11692', 'transformers', 'oscar', 'Sinhala', 'autotrain_compatible']
| false | true | true | 985 |
21,232 |
keshan/sinhala-t5-small
|
keshan
|
t5
| 6 | 19 |
transformers
| 0 |
text2text-generation
| false | false | true | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['jax', 'tensorboard', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
| false | false | false | 0 |
21,233 |
keshan/wav2vec2-si-base
|
keshan
| null | 1 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,234 |
kevinbgray/test_model
|
kevinbgray
| null | 1 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,235 |
kevinfd/test
|
kevinfd
| null | 2 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,236 |
kevinlu1248/personificationgen
|
kevinlu1248
| null | 1 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,237 |
kevinrobinson/perturbations_table_quickstart
|
kevinrobinson
|
bert
| 7 | 5 |
transformers
| 0 | null | false | true | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['tf', 'bert', 'transformers']
| false | false | false | 0 |
21,238 |
kevinrobinson/perturbations_table_quickstart_sst
|
kevinrobinson
|
bert
| 8 | 5 |
transformers
| 0 | null | false | true | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['tf', 'bert', 'transformers']
| false | false | true | 238 |
21,239 |
kevinzyz/bert-base-chinese-finetuned-cola
|
kevinzyz
| null | 1 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,240 |
kevinzyz/chinese-bert-wwm-ext-finetuned-cola-e3
|
kevinzyz
|
bert
| 8 | 10 |
transformers
| 1 |
text-classification
| true | false | false | null | null | null | null | 1 | 1 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'bert', 'text-classification', 'transformers']
| false | false | false | 0 |
21,241 |
kevinzyz/chinese-bert-wwm-ext-finetuned-cola-hyper
|
kevinzyz
| null | 1 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,242 |
kevinzyz/chinese-bert-wwm-ext-finetuned-cola
|
kevinzyz
|
bert
| 16 | 11 |
transformers
| 0 |
text-classification
| true | false | false |
apache-2.0
| null | null | null | 1 | 1 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'tensorboard', 'bert', 'text-classification', 'transformers', 'generated_from_trainer', 'license:apache-2.0']
| true | true | true | 1,346 |
21,243 |
kevinzyz/chinese-roberta-wwm-ext-finetuned-MC-hyper
|
kevinzyz
|
bert
| 12 | 3 |
transformers
| 0 |
multiple-choice
| true | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'bert', 'multiple-choice', 'transformers']
| false | false | false | 0 |
21,244 |
kevinzyz/chinese-roberta-wwm-ext-test
|
kevinzyz
| null | 2 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,245 |
kevinzyz/chinese_roberta_L-12_H-768-finetuned-MC-hyper
|
kevinzyz
|
bert
| 8 | 3 |
transformers
| 0 |
multiple-choice
| true | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'bert', 'multiple-choice', 'transformers']
| false | false | false | 0 |
21,246 |
kevinzyz/chinese_roberta_L-2_H-128-finetuned-MC-hyper
|
kevinzyz
|
bert
| 12 | 3 |
transformers
| 0 |
multiple-choice
| true | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'bert', 'multiple-choice', 'transformers']
| false | false | false | 0 |
21,247 |
kevinzyz/distilbert-base-uncased-finetuned-cola
|
kevinzyz
| null | 1 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,248 |
kevinzyz/finetuned-cola-e3
|
kevinzyz
| null | 1 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,249 |
keyonvafa/compatible-gpt2
|
keyonvafa
|
gpt2
| 8 | 21 |
transformers
| 2 |
text-generation
| true | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'gpt2', 'text-generation', 'transformers']
| false | false | false | 0 |
21,250 |
kgerg/bert-base-cased_fine_tuned_glue_cola
|
kgerg
|
bert
| 7 | 4 |
transformers
| 0 |
text-classification
| false | false | true | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['jax', 'bert', 'text-classification', 'transformers']
| false | false | false | 0 |
21,251 |
kgerg/bert-glue-mrpc-test
|
kgerg
| null | 1 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,252 |
kh3060/bertTest
|
kh3060
| null | 1 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,253 |
kha-white/manga-ocr-base
|
kha-white
|
vision-encoder-decoder
| 8 | 25,451 |
transformers
| 38 |
image-to-text
| true | false | false |
apache-2.0
|
['ja']
|
['manga109s']
| null | 3 | 2 | 1 | 0 | 1 | 0 | 1 |
['pytorch', 'vision-encoder-decoder', 'ja', 'dataset:manga109s', 'transformers', 'image-to-text', 'license:apache-2.0', 'has_space']
| false | true | true | 620 |
21,254 |
khady/wolof-ASR
|
khady
|
wav2vec2
| 12 | 6 |
transformers
| 0 |
automatic-speech-recognition
| true | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'wav2vec2', 'automatic-speech-recognition', 'transformers']
| false | false | false | 0 |
21,255 |
khailai/roberta-offensive-classifier-beta
|
khailai
|
roberta
| 8 | 3 |
transformers
| 0 |
text-classification
| false | true | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['tf', 'roberta', 'text-classification', 'transformers']
| false | false | false | 0 |
21,256 |
khailai/roberta-offensive-classifier
|
khailai
|
roberta
| 10 | 6 |
transformers
| 0 |
text-classification
| false | true | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['tf', 'roberta', 'text-classification', 'transformers']
| false | false | false | 0 |
21,257 |
khailai/t5-wav2vec2-punctuator-2
|
khailai
|
t5
| 6 | 4 |
transformers
| 0 |
text2text-generation
| false | true | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['tf', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
| false | false | false | 0 |
21,258 |
khailai/t5-wav2vec2-punctuator
|
khailai
|
t5
| 6 | 4 |
transformers
| 0 |
text2text-generation
| false | true | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['tf', 't5', 'text2text-generation', 'transformers', 'autotrain_compatible']
| false | false | false | 0 |
21,259 |
khalidalt/DeBERTa-v3-large-mnli
|
khalidalt
|
deberta-v2
| 8 | 13,169 |
transformers
| 4 |
text-classification
| true | false | false | null |
['en']
| null | null | 1 | 1 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'deberta-v2', 'text-classification', 'en', 'arxiv:2006.03654', 'transformers', 'zero-shot-classification']
| false | true | true | 1,896 |
21,260 |
khalidsaifullaah/bengali-lyricist-gpt2
|
khalidsaifullaah
|
gpt2
| 11 | 19 |
transformers
| 0 |
text-generation
| true | false | false | null |
['bn']
| null | null | 1 | 1 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'tensorboard', 'gpt2', 'text-generation', 'bn', 'transformers', 'text generation', 'bengali', 'bangla', 'causal-lm', 'has_space']
| false | true | true | 2,936 |
21,261 |
khanghn/Emotion79
|
khanghn
| null | 1 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,262 |
khanglam7012/k2t-test3
|
khanglam7012
| null | 1 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,263 |
khanglam7012/t5-small
|
khanglam7012
|
t5
| 71 | 8 |
transformers
| 0 |
text2text-generation
| true | false | false |
mit
|
['en']
|
['WebNLG', 'Dart']
| null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 't5', 'text2text-generation', 'en', 'dataset:WebNLG', 'dataset:Dart', 'transformers', 'keytotext', 'k2t', 'Keywords to Sentences', 'license:mit', 'autotrain_compatible']
| false | true | true | 2,173 |
21,264 |
khanh98/model3
|
khanh98
|
roberta
| 10 | 9 |
transformers
| 0 |
text-classification
| true | true | true | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'tf', 'jax', 'roberta', 'text-classification', 'transformers']
| false | false | false | 0 |
21,265 |
khanhpd2/distilBERT-emotionv2
|
khanhpd2
|
distilbert
| 8 | 7 |
transformers
| 0 |
text-classification
| true | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'distilbert', 'text-classification', 'transformers']
| false | false | false | 0 |
21,266 |
khanhpd2/distilbert-emotion
|
khanhpd2
|
distilbert
| 8 | 7 |
transformers
| 0 |
text-classification
| true | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'distilbert', 'text-classification', 'transformers']
| false | false | false | 0 |
21,267 |
khavitidala/finetuned-indobartv2-id-su
|
khavitidala
|
mbart
| 7 | 9 |
transformers
| 0 |
text2text-generation
| true | false | false |
mit
|
['id']
|
['Indo4B+']
| null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'mbart', 'text2text-generation', 'id', 'dataset:Indo4B+', 'arxiv:2104.08200', 'transformers', 'indogpt', 'indobenchmark', 'indonlg', 'license:mit', 'autotrain_compatible']
| false | true | true | 1,458 |
21,268 |
khizon/bert-unreliable-news-eng-title
|
khizon
|
bert
| 3 | 7 |
transformers
| 0 |
text-classification
| true | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'bert', 'text-classification', 'transformers']
| false | false | false | 0 |
21,269 |
khizon/bert-unreliable-news-eng
|
khizon
|
bert
| 4 | 8 |
transformers
| 0 |
text-classification
| true | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'bert', 'text-classification', 'transformers']
| false | false | true | 415 |
21,270 |
khizon/distilbert-unreliable-news-eng-4L
|
khizon
|
distilbert
| 4 | 8 |
transformers
| 0 |
text-classification
| true | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'distilbert', 'text-classification', 'transformers', 'has_space']
| false | false | true | 549 |
21,271 |
khizon/distilbert-unreliable-news-eng-6L
|
khizon
|
distilbert
| 6 | 7 |
transformers
| 0 |
text-classification
| true | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'distilbert', 'text-classification', 'transformers']
| false | false | false | 0 |
21,272 |
khizon/greek-speech-emotion-classifier-demo
|
khizon
|
wav2vec2
| 3 | 6 |
transformers
| 0 | null | true | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'wav2vec2', 'transformers', 'has_space']
| false | false | false | 0 |
21,273 |
khleedkaaki/arcane
|
khleedkaaki
| null | 1 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,274 |
khosravani/transform_lang
|
khosravani
| null | 1 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,275 |
khursani8/distilgpt2-finetuned-wikitext2
|
khursani8
|
gpt2
| 22 | 10 |
transformers
| 0 |
text-generation
| true | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'tensorboard', 'gpt2', 'text-generation', 'transformers']
| false | false | false | 0 |
21,276 |
kiiim/wav2vec2-base-timit-demo-colab
|
kiiim
| null | 4 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,277 |
kika2000/wav2vec2-large-xls-r-300m-georgian
|
kika2000
| null | 1 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,278 |
kika2000/wav2vec2-large-xls-r-300m-kika-colab
|
kika2000
| null | 5 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,279 |
kika2000/wav2vec2-large-xls-r-300m-kika10
|
kika2000
|
wav2vec2
| 13 | 6 |
transformers
| 0 |
automatic-speech-recognition
| true | false | false |
apache-2.0
| null |
['common_voice']
| null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'tensorboard', 'wav2vec2', 'automatic-speech-recognition', 'dataset:common_voice', 'transformers', 'generated_from_trainer', 'license:apache-2.0']
| true | true | true | 1,673 |
21,280 |
kika2000/wav2vec2-large-xls-r-300m-kika2_my-colab
|
kika2000
| null | 5 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,281 |
kika2000/wav2vec2-large-xls-r-300m-kika3_my-colab
|
kika2000
|
wav2vec2
| 12 | 6 |
transformers
| 0 |
automatic-speech-recognition
| true | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'tensorboard', 'wav2vec2', 'automatic-speech-recognition', 'transformers']
| false | false | false | 0 |
21,282 |
kika2000/wav2vec2-large-xls-r-300m-kika4_my-colab
|
kika2000
|
wav2vec2
| 17 | 6 |
transformers
| 0 |
automatic-speech-recognition
| true | false | false |
apache-2.0
| null |
['common_voice']
| null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'tensorboard', 'wav2vec2', 'automatic-speech-recognition', 'dataset:common_voice', 'transformers', 'generated_from_trainer', 'license:apache-2.0']
| true | true | true | 1,082 |
21,283 |
kika2000/wav2vec2-large-xls-r-300m-kika5_my-colab
|
kika2000
|
wav2vec2
| 13 | 8 |
transformers
| 0 |
automatic-speech-recognition
| true | false | false |
apache-2.0
| null |
['common_voice']
| null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'tensorboard', 'wav2vec2', 'automatic-speech-recognition', 'dataset:common_voice', 'transformers', 'generated_from_trainer', 'license:apache-2.0']
| true | true | true | 2,513 |
21,284 |
kika2000/wav2vec2-large-xls-r-300m-kika6_my-colab
|
kika2000
| null | 1 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,285 |
kika2000/wav2vec2-large-xls-r-300m-kika_my-colab
|
kika2000
|
wav2vec2
| 15 | 6 |
transformers
| 0 |
automatic-speech-recognition
| true | false | false |
apache-2.0
| null |
['common_voice']
| null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'tensorboard', 'wav2vec2', 'automatic-speech-recognition', 'dataset:common_voice', 'transformers', 'generated_from_trainer', 'license:apache-2.0']
| true | true | true | 1,671 |
21,286 |
kika2000/wav2vec2-large-xls-r-300m-test80_my-colab
|
kika2000
|
wav2vec2
| 10 | 6 |
transformers
| 0 |
automatic-speech-recognition
| true | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'wav2vec2', 'automatic-speech-recognition', 'transformers']
| false | false | false | 0 |
21,287 |
kika2000/wav2vec2-large-xls-r-300m-test81_my-colab
|
kika2000
|
wav2vec2
| 10 | 8 |
transformers
| 0 |
automatic-speech-recognition
| true | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'wav2vec2', 'automatic-speech-recognition', 'transformers']
| false | false | false | 0 |
21,288 |
kika2000/wav2vec2-large-xls-r-300m-test_my-colab
|
kika2000
|
wav2vec2
| 10 | 7 |
transformers
| 0 |
automatic-speech-recognition
| true | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'wav2vec2', 'automatic-speech-recognition', 'transformers']
| false | false | false | 0 |
21,289 |
kikatikakity/DialoGPT-medium-Dio
|
kikatikakity
| null | 1 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,290 |
kikumaru818/easy_algebra
|
kikumaru818
|
gpt2
| 8 | 30 |
transformers
| 0 |
text-generation
| true | false | false | null | null | null | null | 1 | 1 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'gpt2', 'text-generation', 'transformers']
| false | false | false | 0 |
21,291 |
kimboyworkman/first_model
|
kimboyworkman
| null | 1 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,292 |
kinensake/gramformer-test
|
kinensake
| null | 1 | 0 | null | 0 | null | false | false | false | null | null | null | null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
[]
| false | false | false | 0 |
21,293 |
kingabzpro/DialoGPT-small-Rick-Bot
|
kingabzpro
|
gpt2
| 9 | 38 |
transformers
| 4 |
conversational
| true | false | false |
apache-2.0
|
['English']
|
['Andrada Olteanu Rickmorty-Scripts']
| null | 2 | 2 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'gpt2', 'text-generation', 'English', 'dataset:Andrada Olteanu Rickmorty-Scripts', 'transformers', 'conversational', 'Transformers', 'Chatbot', 'Rick&Morty', 'license:apache-2.0']
| false | true | true | 1,515 |
21,294 |
kingabzpro/Helsinki-NLP-opus-yor-mul-en
|
kingabzpro
|
marian
| 9 | 33 |
transformers
| 1 |
text2text-generation
| true | false | false |
apache-2.0
|
['Yorùbá']
|
['AI4D-Africa - Yorùbá Machine Translation Challenge']
| null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'marian', 'text2text-generation', 'Yorùbá', 'dataset:AI4D-Africa - Yorùbá Machine Translation Challenge', 'transformers', 'text', 'machine-translation', 'language-translation', 'seq2seq', 'helsinki-nlp', 'license:apache-2.0', 'autotrain_compatible']
| false | true | true | 881 |
21,295 |
kingabzpro/wav2vec2-60-Urdu-V8
|
kingabzpro
|
wav2vec2
| 19 | 7 |
transformers
| 1 |
automatic-speech-recognition
| true | false | false |
apache-2.0
|
['ur']
|
['mozilla-foundation/common_voice_8_0']
| null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'tensorboard', 'wav2vec2', 'automatic-speech-recognition', 'ur', 'dataset:mozilla-foundation/common_voice_8_0', 'transformers', 'robust-speech-event', 'hf-asr-leaderboard', 'license:apache-2.0']
| true | true | true | 1,585 |
21,296 |
kingabzpro/wav2vec2-60-urdu
|
kingabzpro
|
wav2vec2
| 43 | 12 |
transformers
| 1 |
automatic-speech-recognition
| true | false | false |
apache-2.0
|
['ur']
|
['mozilla-foundation/common_voice_7_0']
| null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'tensorboard', 'wav2vec2', 'automatic-speech-recognition', 'ur', 'dataset:mozilla-foundation/common_voice_7_0', 'transformers', 'hf-asr-leaderboard', 'robust-speech-event', 'license:apache-2.0']
| true | true | true | 1,713 |
21,297 |
kingabzpro/wav2vec2-large-xls-r-1b-Indonesian
|
kingabzpro
|
wav2vec2
| 11 | 12 |
transformers
| 1 |
automatic-speech-recognition
| true | false | false |
apache-2.0
|
['id']
|
['mozilla-foundation/common_voice_8_0']
| null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'wav2vec2', 'automatic-speech-recognition', 'id', 'dataset:mozilla-foundation/common_voice_8_0', 'transformers', 'hf-asr-leaderboard', 'robust-speech-event', 'license:apache-2.0', 'model-index']
| true | true | true | 1,576 |
21,298 |
kingabzpro/wav2vec2-large-xls-r-1b-Irish
|
kingabzpro
|
wav2vec2
| 18 | 6 |
transformers
| 0 |
automatic-speech-recognition
| true | false | false |
apache-2.0
|
['ga-IE']
|
['mozilla-foundation/common_voice_8_0']
| null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'wav2vec2', 'automatic-speech-recognition', 'ga-IE', 'dataset:mozilla-foundation/common_voice_8_0', 'transformers', 'robust-speech-event', 'hf-asr-leaderboard', 'license:apache-2.0', 'model-index']
| true | true | true | 2,745 |
21,299 |
kingabzpro/wav2vec2-large-xls-r-1b-Swedish
|
kingabzpro
|
wav2vec2
| 22 | 6 |
transformers
| 0 |
automatic-speech-recognition
| true | false | false |
apache-2.0
|
['sv-SE']
|
['mozilla-foundation/common_voice_8_0']
| null | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
['pytorch', 'wav2vec2', 'automatic-speech-recognition', 'sv-SE', 'dataset:mozilla-foundation/common_voice_8_0', 'transformers', 'robust-speech-event', 'hf-asr-leaderboard', 'license:apache-2.0']
| true | true | true | 2,811 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.