pipeline_tag
stringclasses
48 values
library_name
stringclasses
198 values
text
stringlengths
1
900k
metadata
stringlengths
2
438k
id
stringlengths
5
122
last_modified
null
tags
listlengths
1
1.84k
sha
null
created_at
stringlengths
25
25
arxiv
listlengths
0
201
languages
listlengths
0
1.83k
tags_str
stringlengths
17
9.34k
text_str
stringlengths
0
389k
text_lists
listlengths
0
722
processed_texts
listlengths
1
723
feature-extraction
transformers
# BioBERT-NLI This is the model [BioBERT](https://github.com/dmis-lab/biobert) [1] fine-tuned on the [SNLI](https://nlp.stanford.edu/projects/snli/) and the [MultiNLI](https://www.nyu.edu/projects/bowman/multinli/) datasets using the [`sentence-transformers` library](https://github.com/UKPLab/sentence-transformers/) to produce universal sentence embeddings [2]. The model uses the original BERT wordpiece vocabulary and was trained using the **average pooling strategy** and a **softmax loss**. **Base model**: `monologg/biobert_v1.1_pubmed` from HuggingFace's `AutoModel`. **Training time**: ~6 hours on the NVIDIA Tesla P100 GPU provided in Kaggle Notebooks. **Parameters**: | Parameter | Value | |------------------|-------| | Batch size | 64 | | Training steps | 30000 | | Warmup steps | 1450 | | Lowercasing | False | | Max. Seq. Length | 128 | **Performances**: The performance was evaluated on the test portion of the [STS dataset](http://ixa2.si.ehu.es/stswiki/index.php/STSbenchmark) using Spearman rank correlation and compared to the performances of a general BERT base model obtained with the same procedure to verify their similarity. | Model | Score | |-------------------------------|-------------| | `biobert-nli` (this) | 73.40 | | `gsarti/scibert-nli` | 74.50 | | `bert-base-nli-mean-tokens`[3]| 77.12 | An example usage for similarity-based scientific paper retrieval is provided in the [Covid Papers Browser](https://github.com/gsarti/covid-papers-browser) repository. **References:** [1] J. Lee et al, [BioBERT: a pre-trained biomedical language representation model for biomedical text mining](https://academic.oup.com/bioinformatics/article/36/4/1234/5566506) [2] A. Conneau et al., [Supervised Learning of Universal Sentence Representations from Natural Language Inference Data](https://www.aclweb.org/anthology/D17-1070/) [3] N. Reimers et I. Gurevych, [Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks](https://www.aclweb.org/anthology/D19-1410/)
{}
gsarti/biobert-nli
null
[ "transformers", "pytorch", "jax", "bert", "feature-extraction", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #bert #feature-extraction #endpoints_compatible #region-us
BioBERT-NLI =========== This is the model BioBERT [1] fine-tuned on the SNLI and the MultiNLI datasets using the 'sentence-transformers' library to produce universal sentence embeddings [2]. The model uses the original BERT wordpiece vocabulary and was trained using the average pooling strategy and a softmax loss. Base model: 'monologg/biobert\_v1.1\_pubmed' from HuggingFace's 'AutoModel'. Training time: ~6 hours on the NVIDIA Tesla P100 GPU provided in Kaggle Notebooks. Parameters: Performances: The performance was evaluated on the test portion of the STS dataset using Spearman rank correlation and compared to the performances of a general BERT base model obtained with the same procedure to verify their similarity. An example usage for similarity-based scientific paper retrieval is provided in the Covid Papers Browser repository. References: [1] J. Lee et al, BioBERT: a pre-trained biomedical language representation model for biomedical text mining [2] A. Conneau et al., Supervised Learning of Universal Sentence Representations from Natural Language Inference Data [3] N. Reimers et I. Gurevych, Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks
[]
[ "TAGS\n#transformers #pytorch #jax #bert #feature-extraction #endpoints_compatible #region-us \n" ]
feature-extraction
transformers
# CovidBERT-NLI This is the model **CovidBERT** trained by DeepSet on AllenAI's [CORD19 Dataset](https://pages.semanticscholar.org/coronavirus-research) of scientific articles about coronaviruses. The model uses the original BERT wordpiece vocabulary and was subsequently fine-tuned on the [SNLI](https://nlp.stanford.edu/projects/snli/) and the [MultiNLI](https://www.nyu.edu/projects/bowman/multinli/) datasets using the [`sentence-transformers` library](https://github.com/UKPLab/sentence-transformers/) to produce universal sentence embeddings [1] using the **average pooling strategy** and a **softmax loss**. Parameter details for the original training on CORD-19 are available on [DeepSet's MLFlow](https://public-mlflow.deepset.ai/#/experiments/2/runs/ba27d00c30044ef6a33b1d307b4a6cba) **Base model**: `deepset/covid_bert_base` from HuggingFace's `AutoModel`. **Training time**: ~6 hours on the NVIDIA Tesla P100 GPU provided in Kaggle Notebooks. **Parameters**: | Parameter | Value | |------------------|-------| | Batch size | 64 | | Training steps | 23000 | | Warmup steps | 1450 | | Lowercasing | True | | Max. Seq. Length | 128 | **Performances**: The performance was evaluated on the test portion of the [STS dataset](http://ixa2.si.ehu.es/stswiki/index.php/STSbenchmark) using Spearman rank correlation and compared to the performances of similar models obtained with the same procedure to verify its performances. | Model | Score | |-------------------------------|-------------| | `covidbert-nli` (this) | 67.52 | | `gsarti/biobert-nli` | 73.40 | | `gsarti/scibert-nli` | 74.50 | | `bert-base-nli-mean-tokens`[2]| 77.12 | An example usage for similarity-based scientific paper retrieval is provided in the [Covid-19 Semantic Browser](https://github.com/gsarti/covid-papers-browser) repository. **References:** [1] A. Conneau et al., [Supervised Learning of Universal Sentence Representations from Natural Language Inference Data](https://www.aclweb.org/anthology/D17-1070/) [2] N. Reimers et I. Gurevych, [Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks](https://www.aclweb.org/anthology/D19-1410/)
{}
gsarti/covidbert-nli
null
[ "transformers", "pytorch", "jax", "bert", "feature-extraction", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #bert #feature-extraction #endpoints_compatible #region-us
CovidBERT-NLI ============= This is the model CovidBERT trained by DeepSet on AllenAI's CORD19 Dataset of scientific articles about coronaviruses. The model uses the original BERT wordpiece vocabulary and was subsequently fine-tuned on the SNLI and the MultiNLI datasets using the 'sentence-transformers' library to produce universal sentence embeddings [1] using the average pooling strategy and a softmax loss. Parameter details for the original training on CORD-19 are available on DeepSet's MLFlow Base model: 'deepset/covid\_bert\_base' from HuggingFace's 'AutoModel'. Training time: ~6 hours on the NVIDIA Tesla P100 GPU provided in Kaggle Notebooks. Parameters: Performances: The performance was evaluated on the test portion of the STS dataset using Spearman rank correlation and compared to the performances of similar models obtained with the same procedure to verify its performances. An example usage for similarity-based scientific paper retrieval is provided in the Covid-19 Semantic Browser repository. References: [1] A. Conneau et al., Supervised Learning of Universal Sentence Representations from Natural Language Inference Data [2] N. Reimers et I. Gurevych, Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks
[]
[ "TAGS\n#transformers #pytorch #jax #bert #feature-extraction #endpoints_compatible #region-us \n" ]
text2text-generation
transformers
# Italian T5 Base (Oscar) 🇮🇹 *This repository contains the model formerly known as `gsarti/t5-base-it`* The [IT5](https://huggingface.co/models?search=it5) model family represents the first effort in pretraining large-scale sequence-to-sequence transformer models for the Italian language, following the approach adopted by the original [T5 model](https://github.com/google-research/text-to-text-transfer-transformer). This model is released as part of the project ["IT5: Large-Scale Text-to-Text Pretraining for Italian Language Understanding and Generation"](https://gsarti.com) (to be released), by [Gabriele Sarti](https://gsarti.com/) with the support of [Huggingface](https://discuss.huggingface.co/t/open-to-the-community-community-week-using-jax-flax-for-nlp-cv/7104) and with TPU usage sponsored by Google's [TPU Research Cloud](https://sites.research.google/trc/). All the training was conducted on a single TPU3v8-VM machine on Google Cloud. Refer to the Tensorboard tab of the repository for an overview of the training process. *The inference widget is deactivated because the model needs a task-specific seq2seq fine-tuning on a downstream task to be useful in practice. The model [`gsarti/it5-base-nli`](https://huggingface.co/gsarti/it5-base-nli) provides an example of this model fine-tuned on a downstream NLI task.* ## Model variants This repository contains the checkpoints for a `base` version of the model trained on the [OSCAR corpus](https://oscar-corpus.com/) using 🤗 Datasets. The original configuration for the model `t5-base` was adopted, with the exception of the parameter `dropout_rate` that was set at `0` instead of `0.1` during pre-training, following the implementation of [`t5-v1.1`](https://huggingface.co/google/t5-v1_1-base). The tokenizer is a `SentencePieceUnigramTokenizer` trained on the first 2M sentences of the Italian portion of the [`mC4`](https://huggingface.co/datasets/mc4) corpus. An improved version of the model trained on the [Thoroughly Cleaned Italian mC4 Corpus](https://huggingface.co/datasets/gsarti/clean_mc4_it) (~41B words, ~275GB) is also available under the name [`gsarti/it5-base`](https://huggingface.co/gsarti/it5-base). The training procedure is made available [on Github](https://github.com/gsarti/t5-flax-gcp). The following table summarizes the parameters for all available models | |`it5-small` |`it5-base` |`it5-large` |`it5-base-oscar` (this one) | |-----------------------|-----------------------|----------------------|-----------------------|----------------------------------| |`dataset` |`gsarti/clean_mc4_it` |`gsarti/clean_mc4_it` |`gsarti/clean_mc4_it` |`oscar/unshuffled_deduplicated_it`| |`architecture` |`google/t5-v1_1-small` |`google/t5-v1_1-base` |`google/t5-v1_1-large` |`t5-base` | |`learning rate` | 5e-3 | 5e-3 | 5e-3 | 1e-2 | |`steps` | 1'050'000 | 1'050'000 | 2'100'000 | 258'000 | |`training time` | 36 hours | 101 hours | 370 hours | 98 hours | |`ff projection` |`gated-gelu` |`gated-gelu` |`gated-gelu` |`relu` | |`tie embeds` |`false` |`false` |`false` |`true` | |`optimizer` | adafactor | adafactor | adafactor | adafactor | |`max seq. length` | 512 | 512 | 512 | 512 | |`per-device batch size`| 16 | 16 | 8 | 16 | |`tot. batch size` | 128 | 128 | 64 | 128 | |`weigth decay` | 1e-3 | 1e-3 | 1e-2 | 1e-3 | |`validation split size`| 15K examples | 15K examples | 15K examples | 15K examples | The high training time of `it5-base-oscar` was due to [a bug](https://github.com/huggingface/transformers/pull/13012) in the training script. For a list of individual model parameters, refer to the `config.json` file in the respective repositories. ## Using the models ```python from transformers import T5Tokenizer, T5ForConditionalGeneration tokenizer = T5Tokenizer.from_pretrained("gsarti/it5-base-oscar") model = T5ForConditionalGeneration.from_pretrained("gsarti/it5-base-oscar") ``` *Note: You will need to fine-tune the model on your downstream seq2seq task to use it. See an example [here](https://huggingface.co/gsarti/it5-base-nli).* Flax and Tensorflow versions of the model are also available: ```python from transformers import FlaxT5ForConditionalGeneration, TFT5ForConditionalGeneration model_flax = FlaxT5ForConditionalGeneration.from_pretrained("gsarti/it5-base-oscar") model_tf = TFT5ForConditionalGeneration.from_pretrained("gsarti/it5-base-oscar") ``` ## Limitations Due to the nature of the web-scraped corpus on which IT5 models were trained, it is likely that their usage could reproduce and amplify pre-existing biases in the data, resulting in potentially harmful content such as racial or gender stereotypes and conspiracist views. For this reason, the study of such biases is explicitly encouraged, and model usage should ideally be restricted to research-oriented and non-user-facing endeavors. ## Model curators For problems or updates on this model, please contact [[email protected]](mailto:[email protected]). ## Citation Information ```bibtex @article{sarti-nissim-2022-it5, title={IT5: Large-scale Text-to-text Pretraining for Italian Language Understanding and Generation}, author={Sarti, Gabriele and Nissim, Malvina}, journal={ArXiv preprint 2203.03759}, url={https://arxiv.org/abs/2203.03759}, year={2022}, month={mar} } ```
{"language": ["it"], "license": "apache-2.0", "tags": ["seq2seq", "lm-head"], "datasets": ["oscar"], "inference": false}
gsarti/it5-base-oscar
null
[ "transformers", "pytorch", "tf", "jax", "tensorboard", "t5", "text2text-generation", "seq2seq", "lm-head", "it", "dataset:oscar", "arxiv:2203.03759", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
[ "2203.03759" ]
[ "it" ]
TAGS #transformers #pytorch #tf #jax #tensorboard #t5 #text2text-generation #seq2seq #lm-head #it #dataset-oscar #arxiv-2203.03759 #license-apache-2.0 #autotrain_compatible #text-generation-inference #region-us
Italian T5 Base (Oscar) 🇮🇹 ========================== *This repository contains the model formerly known as 'gsarti/t5-base-it'* The IT5 model family represents the first effort in pretraining large-scale sequence-to-sequence transformer models for the Italian language, following the approach adopted by the original T5 model. This model is released as part of the project "IT5: Large-Scale Text-to-Text Pretraining for Italian Language Understanding and Generation" (to be released), by Gabriele Sarti with the support of Huggingface and with TPU usage sponsored by Google's TPU Research Cloud. All the training was conducted on a single TPU3v8-VM machine on Google Cloud. Refer to the Tensorboard tab of the repository for an overview of the training process. *The inference widget is deactivated because the model needs a task-specific seq2seq fine-tuning on a downstream task to be useful in practice. The model 'gsarti/it5-base-nli' provides an example of this model fine-tuned on a downstream NLI task.* Model variants -------------- This repository contains the checkpoints for a 'base' version of the model trained on the OSCAR corpus using Datasets. The original configuration for the model 't5-base' was adopted, with the exception of the parameter 'dropout\_rate' that was set at '0' instead of '0.1' during pre-training, following the implementation of 't5-v1.1'. The tokenizer is a 'SentencePieceUnigramTokenizer' trained on the first 2M sentences of the Italian portion of the 'mC4' corpus. An improved version of the model trained on the Thoroughly Cleaned Italian mC4 Corpus (~41B words, ~275GB) is also available under the name 'gsarti/it5-base'. The training procedure is made available on Github. The following table summarizes the parameters for all available models The high training time of 'it5-base-oscar' was due to a bug in the training script. For a list of individual model parameters, refer to the 'URL' file in the respective repositories. Using the models ---------------- *Note: You will need to fine-tune the model on your downstream seq2seq task to use it. See an example here.* Flax and Tensorflow versions of the model are also available: Limitations ----------- Due to the nature of the web-scraped corpus on which IT5 models were trained, it is likely that their usage could reproduce and amplify pre-existing biases in the data, resulting in potentially harmful content such as racial or gender stereotypes and conspiracist views. For this reason, the study of such biases is explicitly encouraged, and model usage should ideally be restricted to research-oriented and non-user-facing endeavors. Model curators -------------- For problems or updates on this model, please contact gabriele.sarti996@URL.
[]
[ "TAGS\n#transformers #pytorch #tf #jax #tensorboard #t5 #text2text-generation #seq2seq #lm-head #it #dataset-oscar #arxiv-2203.03759 #license-apache-2.0 #autotrain_compatible #text-generation-inference #region-us \n" ]
text2text-generation
transformers
# Italian T5 Base 🇮🇹 The [IT5](https://huggingface.co/models?search=it5) model family represents the first effort in pretraining large-scale sequence-to-sequence transformer models for the Italian language, following the approach adopted by the original [T5 model](https://github.com/google-research/text-to-text-transfer-transformer). This model is released as part of the project ["IT5: Large-Scale Text-to-Text Pretraining for Italian Language Understanding and Generation"](https://arxiv.org/abs/2203.03759), by [Gabriele Sarti](https://gsarti.com/) and [Malvina Nissim](https://malvinanissim.github.io/) with the support of [Huggingface](https://discuss.huggingface.co/t/open-to-the-community-community-week-using-jax-flax-for-nlp-cv/7104) and with TPU usage sponsored by Google's [TPU Research Cloud](https://sites.research.google/trc/). All the training was conducted on a single TPU3v8-VM machine on Google Cloud. Refer to the Tensorboard tab of the repository for an overview of the training process. *TThe inference widget is deactivated because the model needs a task-specific seq2seq fine-tuning on a downstream task to be useful in practice. The models in the [`it5`](https://huggingface.co/it5) organization provide some examples of this model fine-tuned on various downstream task.* ## Model variants This repository contains the checkpoints for the `base` version of the model. The model was trained for one epoch (1.05M steps) on the [Thoroughly Cleaned Italian mC4 Corpus](https://huggingface.co/datasets/gsarti/clean_mc4_it) (~41B words, ~275GB) using 🤗 Datasets and the `google/t5-v1_1-base` improved configuration. Another version of this model trained on the [OSCAR corpus](https://oscar-corpus.com/) is also available under the name [`gsarti/it5-base-oscar`](https://huggingface.co/gsartiit5-base-oscar). The training procedure is made available [on Github](https://github.com/gsarti/t5-flax-gcp). The following table summarizes the parameters for all available models | |`it5-small` |`it5-base` (this one) |`it5-large` |`it5-base-oscar` | |-----------------------|-----------------------|----------------------|-----------------------|----------------------------------| |`dataset` |`gsarti/clean_mc4_it` |`gsarti/clean_mc4_it` |`gsarti/clean_mc4_it` |`oscar/unshuffled_deduplicated_it`| |`architecture` |`google/t5-v1_1-small` |`google/t5-v1_1-base` |`google/t5-v1_1-large` |`t5-base` | |`learning rate` | 5e-3 | 5e-3 | 5e-3 | 1e-2 | |`steps` | 1'050'000 | 1'050'000 | 2'100'000 | 258'000 | |`training time` | 36 hours | 101 hours | 370 hours | 98 hours | |`ff projection` |`gated-gelu` |`gated-gelu` |`gated-gelu` |`relu` | |`tie embeds` |`false` |`false` |`false` |`true` | |`optimizer` | adafactor | adafactor | adafactor | adafactor | |`max seq. length` | 512 | 512 | 512 | 512 | |`per-device batch size`| 16 | 16 | 8 | 16 | |`tot. batch size` | 128 | 128 | 64 | 128 | |`weigth decay` | 1e-3 | 1e-3 | 1e-2 | 1e-3 | |`validation split size`| 15K examples | 15K examples | 15K examples | 15K examples | The high training time of `it5-base-oscar` was due to [a bug](https://github.com/huggingface/transformers/pull/13012) in the training script. For a list of individual model parameters, refer to the `config.json` file in the respective repositories. ## Using the models ```python from transformers import AutoTokenizer, AutoModelForSeq2SeqLM tokenizer = AutoTokenizer.from_pretrained("gsarti/it5-base") model = AutoModelForSeq2SeqLM.from_pretrained("gsarti/it5-base") ``` *Note: You will need to fine-tune the model on your downstream seq2seq task to use it. See an example [here](https://huggingface.co/it5/it5-base-news-summarization).* Flax and Tensorflow versions of the model are also available: ```python from transformers import FlaxT5ForConditionalGeneration, TFT5ForConditionalGeneration model_flax = FlaxT5ForConditionalGeneration.from_pretrained("gsarti/it5-base") model_tf = TFT5ForConditionalGeneration.from_pretrained("gsarti/it5-base") ``` ## Limitations Due to the nature of the web-scraped corpus on which IT5 models were trained, it is likely that their usage could reproduce and amplify pre-existing biases in the data, resulting in potentially harmful content such as racial or gender stereotypes and conspiracist views. For this reason, the study of such biases is explicitly encouraged, and model usage should ideally be restricted to research-oriented and non-user-facing endeavors. ## Model curators For problems or updates on this model, please contact [[email protected]](mailto:[email protected]). ## Citation Information ```bibtex @article{sarti-nissim-2022-it5, title={IT5: Large-scale Text-to-text Pretraining for Italian Language Understanding and Generation}, author={Sarti, Gabriele and Nissim, Malvina}, journal={ArXiv preprint 2203.03759}, url={https://arxiv.org/abs/2203.03759}, year={2022}, month={mar} } ```
{"language": ["it"], "license": "apache-2.0", "tags": ["seq2seq", "lm-head"], "datasets": ["gsarti/clean_mc4_it"], "inference": false, "thumbnail": "https://gsarti.com/publication/it5/featured.png"}
gsarti/it5-base
null
[ "transformers", "pytorch", "tf", "jax", "tensorboard", "t5", "text2text-generation", "seq2seq", "lm-head", "it", "dataset:gsarti/clean_mc4_it", "arxiv:2203.03759", "license:apache-2.0", "autotrain_compatible", "has_space", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
[ "2203.03759" ]
[ "it" ]
TAGS #transformers #pytorch #tf #jax #tensorboard #t5 #text2text-generation #seq2seq #lm-head #it #dataset-gsarti/clean_mc4_it #arxiv-2203.03759 #license-apache-2.0 #autotrain_compatible #has_space #text-generation-inference #region-us
Italian T5 Base 🇮🇹 ================== The IT5 model family represents the first effort in pretraining large-scale sequence-to-sequence transformer models for the Italian language, following the approach adopted by the original T5 model. This model is released as part of the project "IT5: Large-Scale Text-to-Text Pretraining for Italian Language Understanding and Generation", by Gabriele Sarti and Malvina Nissim with the support of Huggingface and with TPU usage sponsored by Google's TPU Research Cloud. All the training was conducted on a single TPU3v8-VM machine on Google Cloud. Refer to the Tensorboard tab of the repository for an overview of the training process. *TThe inference widget is deactivated because the model needs a task-specific seq2seq fine-tuning on a downstream task to be useful in practice. The models in the 'it5' organization provide some examples of this model fine-tuned on various downstream task.* Model variants -------------- This repository contains the checkpoints for the 'base' version of the model. The model was trained for one epoch (1.05M steps) on the Thoroughly Cleaned Italian mC4 Corpus (~41B words, ~275GB) using Datasets and the 'google/t5-v1\_1-base' improved configuration. Another version of this model trained on the OSCAR corpus is also available under the name 'gsarti/it5-base-oscar'. The training procedure is made available on Github. The following table summarizes the parameters for all available models The high training time of 'it5-base-oscar' was due to a bug in the training script. For a list of individual model parameters, refer to the 'URL' file in the respective repositories. Using the models ---------------- *Note: You will need to fine-tune the model on your downstream seq2seq task to use it. See an example here.* Flax and Tensorflow versions of the model are also available: Limitations ----------- Due to the nature of the web-scraped corpus on which IT5 models were trained, it is likely that their usage could reproduce and amplify pre-existing biases in the data, resulting in potentially harmful content such as racial or gender stereotypes and conspiracist views. For this reason, the study of such biases is explicitly encouraged, and model usage should ideally be restricted to research-oriented and non-user-facing endeavors. Model curators -------------- For problems or updates on this model, please contact gabriele.sarti996@URL.
[]
[ "TAGS\n#transformers #pytorch #tf #jax #tensorboard #t5 #text2text-generation #seq2seq #lm-head #it #dataset-gsarti/clean_mc4_it #arxiv-2203.03759 #license-apache-2.0 #autotrain_compatible #has_space #text-generation-inference #region-us \n" ]
text2text-generation
transformers
# Italian T5 Large 🇮🇹 The [IT5](https://huggingface.co/models?search=it5) model family represents the first effort in pretraining large-scale sequence-to-sequence transformer models for the Italian language, following the approach adopted by the original [T5 model](https://github.com/google-research/text-to-text-transfer-transformer). This model is released as part of the project ["IT5: Large-Scale Text-to-Text Pretraining for Italian Language Understanding and Generation"](https://arxiv.org/abs/2203.03759) (to be released), by [Gabriele Sarti](https://gsarti.com/) and [Malvina Nissim](https://malvinanissim.github.io/) with the support of [Huggingface](https://discuss.huggingface.co/t/open-to-the-community-community-week-using-jax-flax-for-nlp-cv/7104) and with TPU usage sponsored by Google's [TPU Research Cloud](https://sites.research.google/trc/). All the training was conducted on a single TPU3v8-VM machine on Google Cloud. Refer to the Tensorboard tab of the repository for an overview of the training process. *The inference widget is deactivated because the model needs a task-specific seq2seq fine-tuning on a downstream task to be useful in practice. The models in the [`it5`](https://huggingface.co/it5) organization provide some examples of this model fine-tuned on various downstream task.* ## Model variants This repository contains the checkpoints for the `base` version of the model. The model was trained for one epoch (1.05M steps) on the [Thoroughly Cleaned Italian mC4 Corpus](https://huggingface.co/datasets/gsarti/clean_mc4_it) (~41B words, ~275GB) using 🤗 Datasets and the `google/t5-v1_1-large` improved configuration. The training procedure is made available [on Github](https://github.com/gsarti/t5-flax-gcp). The following table summarizes the parameters for all available models | |`it5-small` |`it5-base` |`it5-large` (this one) |`it5-base-oscar` | |-----------------------|-----------------------|----------------------|-----------------------|----------------------------------| |`dataset` |`gsarti/clean_mc4_it` |`gsarti/clean_mc4_it` |`gsarti/clean_mc4_it` |`oscar/unshuffled_deduplicated_it`| |`architecture` |`google/t5-v1_1-small` |`google/t5-v1_1-base` |`google/t5-v1_1-large` |`t5-base` | |`learning rate` | 5e-3 | 5e-3 | 5e-3 | 1e-2 | |`steps` | 1'050'000 | 1'050'000 | 2'100'000 | 258'000 | |`training time` | 36 hours | 101 hours | 370 hours | 98 hours | |`ff projection` |`gated-gelu` |`gated-gelu` |`gated-gelu` |`relu` | |`tie embeds` |`false` |`false` |`false` |`true` | |`optimizer` | adafactor | adafactor | adafactor | adafactor | |`max seq. length` | 512 | 512 | 512 | 512 | |`per-device batch size`| 16 | 16 | 8 | 16 | |`tot. batch size` | 128 | 128 | 64 | 128 | |`weigth decay` | 1e-3 | 1e-3 | 1e-2 | 1e-3 | |`validation split size`| 15K examples | 15K examples | 15K examples | 15K examples | The high training time of `it5-base-oscar` was due to [a bug](https://github.com/huggingface/transformers/pull/13012) in the training script. For a list of individual model parameters, refer to the `config.json` file in the respective repositories. ## Using the models ```python from transformers import AutoTokenzier, AutoModelForSeq2SeqLM tokenizer = AutoTokenizer.from_pretrained("gsarti/it5-large") model = AutoModelForSeq2SeqLM.from_pretrained("gsarti/it5-large") ``` *Note: You will need to fine-tune the model on your downstream seq2seq task to use it. See an example [here](https://huggingface.co/gsarti/it5-base-nli).* Flax and Tensorflow versions of the model are also available: ```python from transformers import FlaxT5ForConditionalGeneration, TFT5ForConditionalGeneration model_flax = FlaxT5ForConditionalGeneration.from_pretrained("gsarti/it5-large") model_tf = TFT5ForConditionalGeneration.from_pretrained("gsarti/it5-large") ``` ## Limitations Due to the nature of the web-scraped corpus on which IT5 models were trained, it is likely that their usage could reproduce and amplify pre-existing biases in the data, resulting in potentially harmful content such as racial or gender stereotypes and conspiracist views. For this reason, the study of such biases is explicitly encouraged, and model usage should ideally be restricted to research-oriented and non-user-facing endeavors. ## Model curators For problems or updates on this model, please contact [[email protected]](mailto:[email protected]). ## Citation Information ```bibtex @article{sarti-nissim-2022-it5, title={IT5: Large-scale Text-to-text Pretraining for Italian Language Understanding and Generation}, author={Sarti, Gabriele and Nissim, Malvina}, journal={ArXiv preprint 2203.03759}, url={https://arxiv.org/abs/2203.03759}, year={2022}, month={mar} } ```
{"language": ["it"], "license": "apache-2.0", "tags": ["seq2seq", "lm-head"], "datasets": ["gsarti/clean_mc4_it"], "inference": false, "thumbnail": "https://gsarti.com/publication/it5/featured.png"}
gsarti/it5-large
null
[ "transformers", "pytorch", "tf", "jax", "tensorboard", "t5", "text2text-generation", "seq2seq", "lm-head", "it", "dataset:gsarti/clean_mc4_it", "arxiv:2203.03759", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
[ "2203.03759" ]
[ "it" ]
TAGS #transformers #pytorch #tf #jax #tensorboard #t5 #text2text-generation #seq2seq #lm-head #it #dataset-gsarti/clean_mc4_it #arxiv-2203.03759 #license-apache-2.0 #autotrain_compatible #text-generation-inference #region-us
Italian T5 Large 🇮🇹 =================== The IT5 model family represents the first effort in pretraining large-scale sequence-to-sequence transformer models for the Italian language, following the approach adopted by the original T5 model. This model is released as part of the project "IT5: Large-Scale Text-to-Text Pretraining for Italian Language Understanding and Generation" (to be released), by Gabriele Sarti and Malvina Nissim with the support of Huggingface and with TPU usage sponsored by Google's TPU Research Cloud. All the training was conducted on a single TPU3v8-VM machine on Google Cloud. Refer to the Tensorboard tab of the repository for an overview of the training process. *The inference widget is deactivated because the model needs a task-specific seq2seq fine-tuning on a downstream task to be useful in practice. The models in the 'it5' organization provide some examples of this model fine-tuned on various downstream task.* Model variants -------------- This repository contains the checkpoints for the 'base' version of the model. The model was trained for one epoch (1.05M steps) on the Thoroughly Cleaned Italian mC4 Corpus (~41B words, ~275GB) using Datasets and the 'google/t5-v1\_1-large' improved configuration. The training procedure is made available on Github. The following table summarizes the parameters for all available models The high training time of 'it5-base-oscar' was due to a bug in the training script. For a list of individual model parameters, refer to the 'URL' file in the respective repositories. Using the models ---------------- *Note: You will need to fine-tune the model on your downstream seq2seq task to use it. See an example here.* Flax and Tensorflow versions of the model are also available: Limitations ----------- Due to the nature of the web-scraped corpus on which IT5 models were trained, it is likely that their usage could reproduce and amplify pre-existing biases in the data, resulting in potentially harmful content such as racial or gender stereotypes and conspiracist views. For this reason, the study of such biases is explicitly encouraged, and model usage should ideally be restricted to research-oriented and non-user-facing endeavors. Model curators -------------- For problems or updates on this model, please contact gabriele.sarti996@URL.
[]
[ "TAGS\n#transformers #pytorch #tf #jax #tensorboard #t5 #text2text-generation #seq2seq #lm-head #it #dataset-gsarti/clean_mc4_it #arxiv-2203.03759 #license-apache-2.0 #autotrain_compatible #text-generation-inference #region-us \n" ]
text2text-generation
transformers
# Italian T5 Small 🇮🇹 The [IT5](https://huggingface.co/models?search=it5) model family represents the first effort in pretraining large-scale sequence-to-sequence transformer models for the Italian language, following the approach adopted by the original [T5 model](https://github.com/google-research/text-to-text-transfer-transformer). This model is released as part of the project ["IT5: Large-Scale Text-to-Text Pretraining for Italian Language Understanding and Generation"](https://arxiv.org/abs/2203.03759), by [Gabriele Sarti](https://gsarti.com/) and [Malvina Nissim](https://malvinanissim.github.io/) with the support of [Huggingface](https://discuss.huggingface.co/t/open-to-the-community-community-week-using-jax-flax-for-nlp-cv/7104) and with TPU usage sponsored by Google's [TPU Research Cloud](https://sites.research.google/trc/). All the training was conducted on a single TPU3v8-VM machine on Google Cloud. Refer to the Tensorboard tab of the repository for an overview of the training process. *The inference widget is deactivated because the model needs a task-specific seq2seq fine-tuning on a downstream task to be useful in practice. The models in the [`it5`](https://huggingface.co/it5) organization provide some examples of this model fine-tuned on various downstream task.* ## Model variants This repository contains the checkpoints for the `base` version of the model. The model was trained for one epoch (1.05M steps) on the [Thoroughly Cleaned Italian mC4 Corpus](https://huggingface.co/datasets/gsarti/clean_mc4_it) (~41B words, ~275GB) using 🤗 Datasets and the `google/t5-v1_1-small` improved configuration. The training procedure is made available [on Github](https://github.com/gsarti/t5-flax-gcp). The following table summarizes the parameters for all available models | |`it5-small` (this one) |`it5-base` |`it5-large` |`it5-base-oscar` | |-----------------------|-----------------------|----------------------|-----------------------|----------------------------------| |`dataset` |`gsarti/clean_mc4_it` |`gsarti/clean_mc4_it` |`gsarti/clean_mc4_it` |`oscar/unshuffled_deduplicated_it`| |`architecture` |`google/t5-v1_1-small` |`google/t5-v1_1-base` |`google/t5-v1_1-large` |`t5-base` | |`learning rate` | 5e-3 | 5e-3 | 5e-3 | 1e-2 | |`steps` | 1'050'000 | 1'050'000 | 2'100'000 | 258'000 | |`training time` | 36 hours | 101 hours | 370 hours | 98 hours | |`ff projection` |`gated-gelu` |`gated-gelu` |`gated-gelu` |`relu` | |`tie embeds` |`false` |`false` |`false` |`true` | |`optimizer` | adafactor | adafactor | adafactor | adafactor | |`max seq. length` | 512 | 512 | 512 | 512 | |`per-device batch size`| 16 | 16 | 8 | 16 | |`tot. batch size` | 128 | 128 | 64 | 128 | |`weigth decay` | 1e-3 | 1e-3 | 1e-2 | 1e-3 | |`validation split size`| 15K examples | 15K examples | 15K examples | 15K examples | The high training time of `it5-base-oscar` was due to [a bug](https://github.com/huggingface/transformers/pull/13012) in the training script. For a list of individual model parameters, refer to the `config.json` file in the respective repositories. ## Using the models ```python from transformers import AutoTokenzier, AutoModelForSeq2SeqLM tokenizer = AutoTokenizer.from_pretrained("gsarti/it5-small") model = AutoModelForSeq2SeqLM.from_pretrained("gsarti/it5-small") ``` *Note: You will need to fine-tune the model on your downstream seq2seq task to use it. See an example [here](https://huggingface.co/it5/it5-base-question-answering).* Flax and Tensorflow versions of the model are also available: ```python from transformers import FlaxT5ForConditionalGeneration, TFT5ForConditionalGeneration model_flax = FlaxT5ForConditionalGeneration.from_pretrained("gsarti/it5-small") model_tf = TFT5ForConditionalGeneration.from_pretrained("gsarti/it5-small") ``` ## Limitations Due to the nature of the web-scraped corpus on which IT5 models were trained, it is likely that their usage could reproduce and amplify pre-existing biases in the data, resulting in potentially harmful content such as racial or gender stereotypes and conspiracist views. For this reason, the study of such biases is explicitly encouraged, and model usage should ideally be restricted to research-oriented and non-user-facing endeavors. ## Model curators For problems or updates on this model, please contact [[email protected]](mailto:[email protected]). ## Citation Information ```bibtex @article{sarti-nissim-2022-it5, title={IT5: Large-scale Text-to-text Pretraining for Italian Language Understanding and Generation}, author={Sarti, Gabriele and Nissim, Malvina}, journal={ArXiv preprint 2203.03759}, url={https://arxiv.org/abs/2203.03759}, year={2022}, month={mar} } ```
{"language": ["it"], "license": "apache-2.0", "tags": ["seq2seq", "lm-head"], "datasets": ["gsarti/clean_mc4_it"], "inference": false, "thumbnail": "https://gsarti.com/publication/it5/featured.png"}
gsarti/it5-small
null
[ "transformers", "pytorch", "tf", "jax", "tensorboard", "t5", "text2text-generation", "seq2seq", "lm-head", "it", "dataset:gsarti/clean_mc4_it", "arxiv:2203.03759", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
[ "2203.03759" ]
[ "it" ]
TAGS #transformers #pytorch #tf #jax #tensorboard #t5 #text2text-generation #seq2seq #lm-head #it #dataset-gsarti/clean_mc4_it #arxiv-2203.03759 #license-apache-2.0 #autotrain_compatible #text-generation-inference #region-us
Italian T5 Small 🇮🇹 =================== The IT5 model family represents the first effort in pretraining large-scale sequence-to-sequence transformer models for the Italian language, following the approach adopted by the original T5 model. This model is released as part of the project "IT5: Large-Scale Text-to-Text Pretraining for Italian Language Understanding and Generation", by Gabriele Sarti and Malvina Nissim with the support of Huggingface and with TPU usage sponsored by Google's TPU Research Cloud. All the training was conducted on a single TPU3v8-VM machine on Google Cloud. Refer to the Tensorboard tab of the repository for an overview of the training process. *The inference widget is deactivated because the model needs a task-specific seq2seq fine-tuning on a downstream task to be useful in practice. The models in the 'it5' organization provide some examples of this model fine-tuned on various downstream task.* Model variants -------------- This repository contains the checkpoints for the 'base' version of the model. The model was trained for one epoch (1.05M steps) on the Thoroughly Cleaned Italian mC4 Corpus (~41B words, ~275GB) using Datasets and the 'google/t5-v1\_1-small' improved configuration. The training procedure is made available on Github. The following table summarizes the parameters for all available models The high training time of 'it5-base-oscar' was due to a bug in the training script. For a list of individual model parameters, refer to the 'URL' file in the respective repositories. Using the models ---------------- *Note: You will need to fine-tune the model on your downstream seq2seq task to use it. See an example here.* Flax and Tensorflow versions of the model are also available: Limitations ----------- Due to the nature of the web-scraped corpus on which IT5 models were trained, it is likely that their usage could reproduce and amplify pre-existing biases in the data, resulting in potentially harmful content such as racial or gender stereotypes and conspiracist views. For this reason, the study of such biases is explicitly encouraged, and model usage should ideally be restricted to research-oriented and non-user-facing endeavors. Model curators -------------- For problems or updates on this model, please contact gabriele.sarti996@URL.
[]
[ "TAGS\n#transformers #pytorch #tf #jax #tensorboard #t5 #text2text-generation #seq2seq #lm-head #it #dataset-gsarti/clean_mc4_it #arxiv-2203.03759 #license-apache-2.0 #autotrain_compatible #text-generation-inference #region-us \n" ]
feature-extraction
transformers
# SciBERT-NLI This is the model [SciBERT](https://github.com/allenai/scibert) [1] fine-tuned on the [SNLI](https://nlp.stanford.edu/projects/snli/) and the [MultiNLI](https://www.nyu.edu/projects/bowman/multinli/) datasets using the [`sentence-transformers` library](https://github.com/UKPLab/sentence-transformers/) to produce universal sentence embeddings [2]. The model uses the original `scivocab` wordpiece vocabulary and was trained using the **average pooling strategy** and a **softmax loss**. **Base model**: `allenai/scibert-scivocab-cased` from HuggingFace's `AutoModel`. **Training time**: ~4 hours on the NVIDIA Tesla P100 GPU provided in Kaggle Notebooks. **Parameters**: | Parameter | Value | |------------------|-------| | Batch size | 64 | | Training steps | 20000 | | Warmup steps | 1450 | | Lowercasing | True | | Max. Seq. Length | 128 | **Performances**: The performance was evaluated on the test portion of the [STS dataset](http://ixa2.si.ehu.es/stswiki/index.php/STSbenchmark) using Spearman rank correlation and compared to the performances of a general BERT base model obtained with the same procedure to verify their similarity. | Model | Score | |-------------------------------|-------------| | `scibert-nli` (this) | 74.50 | | `bert-base-nli-mean-tokens`[3]| 77.12 | An example usage for similarity-based scientific paper retrieval is provided in the [Covid Papers Browser](https://github.com/gsarti/covid-papers-browser) repository. **References:** [1] I. Beltagy et al, [SciBERT: A Pretrained Language Model for Scientific Text](https://www.aclweb.org/anthology/D19-1371/) [2] A. Conneau et al., [Supervised Learning of Universal Sentence Representations from Natural Language Inference Data](https://www.aclweb.org/anthology/D17-1070/) [3] N. Reimers et I. Gurevych, [Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks](https://www.aclweb.org/anthology/D19-1410/)
{}
gsarti/scibert-nli
null
[ "transformers", "pytorch", "jax", "bert", "feature-extraction", "doi:10.57967/hf/0038", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #bert #feature-extraction #doi-10.57967/hf/0038 #endpoints_compatible #region-us
SciBERT-NLI =========== This is the model SciBERT [1] fine-tuned on the SNLI and the MultiNLI datasets using the 'sentence-transformers' library to produce universal sentence embeddings [2]. The model uses the original 'scivocab' wordpiece vocabulary and was trained using the average pooling strategy and a softmax loss. Base model: 'allenai/scibert-scivocab-cased' from HuggingFace's 'AutoModel'. Training time: ~4 hours on the NVIDIA Tesla P100 GPU provided in Kaggle Notebooks. Parameters: Performances: The performance was evaluated on the test portion of the STS dataset using Spearman rank correlation and compared to the performances of a general BERT base model obtained with the same procedure to verify their similarity. An example usage for similarity-based scientific paper retrieval is provided in the Covid Papers Browser repository. References: [1] I. Beltagy et al, SciBERT: A Pretrained Language Model for Scientific Text [2] A. Conneau et al., Supervised Learning of Universal Sentence Representations from Natural Language Inference Data [3] N. Reimers et I. Gurevych, Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks
[]
[ "TAGS\n#transformers #pytorch #jax #bert #feature-extraction #doi-10.57967/hf/0038 #endpoints_compatible #region-us \n" ]
text-to-image
generic
ERROR: type should be string, got "\nhttps://github.com/borisdayma/dalle-mini"
{"language": ["en"], "library_name": "generic", "pipeline_tag": "text-to-image"}
gsurma/ai_dreamer
null
[ "generic", "jax", "bart", "text-to-image", "en", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #generic #jax #bart #text-to-image #en #region-us
URL
[]
[ "TAGS\n#generic #jax #bart #text-to-image #en #region-us \n" ]
fill-mask
transformers
# dummy model This is a dummy model
{}
gulabpatel/new-dummy-model
null
[ "transformers", "pytorch", "camembert", "fill-mask", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #camembert #fill-mask #autotrain_compatible #endpoints_compatible #region-us
# dummy model This is a dummy model
[ "# dummy model\n\nThis is a dummy model" ]
[ "TAGS\n#transformers #pytorch #camembert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n", "# dummy model\n\nThis is a dummy model" ]
automatic-speech-recognition
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-base-timit-demo-colab This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.4872 - Wer: 0.3417 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 32 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 1000 - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:------:| | 3.4857 | 4.0 | 500 | 1.4555 | 1.0040 | | 0.5994 | 8.0 | 1000 | 0.5011 | 0.4370 | | 0.2273 | 12.0 | 1500 | 0.4293 | 0.3903 | | 0.1235 | 16.0 | 2000 | 0.4602 | 0.3772 | | 0.084 | 20.0 | 2500 | 0.5055 | 0.3673 | | 0.0615 | 24.0 | 3000 | 0.4915 | 0.3486 | | 0.0468 | 28.0 | 3500 | 0.4872 | 0.3417 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.10.0+cu111 - Datasets 1.13.3 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "wav2vec2-base-timit-demo-colab", "results": []}]}
gullenasatish/wav2vec2-base-timit-demo-colab
null
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us
wav2vec2-base-timit-demo-colab ============================== This model is a fine-tuned version of facebook/wav2vec2-base on the None dataset. It achieves the following results on the evaluation set: * Loss: 0.4872 * Wer: 0.3417 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 0.0001 * train\_batch\_size: 32 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 1000 * num\_epochs: 30 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.11.3 * Pytorch 1.10.0+cu111 * Datasets 1.13.3 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 30\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.13.3\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 30\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.13.3\n* Tokenizers 0.10.3" ]
token-classification
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # gunghio/distilbert-base-multilingual-cased-finetuned-conll2003-ner This model was trained from scratch on an conll2003 dataset. It achieves the following results on the evaluation set: - Loss: 0.0388 - Precision: 0.9360 - Recall: 0.9458 - F1: 0.9409 - Accuracy: 0.9902 ## Model description It is based on distilbert-base-multilingual-cased ## Intended uses & limitations More information needed ## Training and evaluation data Training dataset: [conll2003](https://huggingface.co/datasets/conll2003) ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.1653 | 1.0 | 878 | 0.0465 | 0.9267 | 0.9300 | 0.9283 | 0.9883 | | 0.0322 | 2.0 | 1756 | 0.0404 | 0.9360 | 0.9431 | 0.9396 | 0.9897 | | 0.0185 | 3.0 | 2634 | 0.0388 | 0.9360 | 0.9458 | 0.9409 | 0.9902 | ### Framework versions - Transformers 4.6.1 - Pytorch 1.8.1+cu101 - Datasets 1.6.2 - Tokenizers 0.10.2
{"language": ["en", "de", "nl", "es", "multilingual"], "datasets": ["conll2003"], "metrics": [{"precision": 0.936}, {"recall": 0.9458}, {"f1": 0.9409}, {"accuracy": 0.9902}], "model-index": [{"name": "gunghio/distilbert-base-multilingual-cased-finetuned-conll2003-ner", "results": [{"task": {"type": "ner", "name": "Named Entity Recognition"}, "dataset": {"name": "ConLL 2003", "type": "conll2003"}, "metrics": [{"type": "f1-score", "value": 0.9409}]}]}]}
gunghio/distilbert-base-multilingual-cased-finetuned-conll2003-ner
null
[ "transformers", "pytorch", "distilbert", "token-classification", "en", "de", "nl", "es", "multilingual", "dataset:conll2003", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "en", "de", "nl", "es", "multilingual" ]
TAGS #transformers #pytorch #distilbert #token-classification #en #de #nl #es #multilingual #dataset-conll2003 #model-index #autotrain_compatible #endpoints_compatible #region-us
gunghio/distilbert-base-multilingual-cased-finetuned-conll2003-ner ================================================================== This model was trained from scratch on an conll2003 dataset. It achieves the following results on the evaluation set: * Loss: 0.0388 * Precision: 0.9360 * Recall: 0.9458 * F1: 0.9409 * Accuracy: 0.9902 Model description ----------------- It is based on distilbert-base-multilingual-cased Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- Training dataset: conll2003 Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3 ### Training results ### Framework versions * Transformers 4.6.1 * Pytorch 1.8.1+cu101 * Datasets 1.6.2 * Tokenizers 0.10.2
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3", "### Training results", "### Framework versions\n\n\n* Transformers 4.6.1\n* Pytorch 1.8.1+cu101\n* Datasets 1.6.2\n* Tokenizers 0.10.2" ]
[ "TAGS\n#transformers #pytorch #distilbert #token-classification #en #de #nl #es #multilingual #dataset-conll2003 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3", "### Training results", "### Framework versions\n\n\n* Transformers 4.6.1\n* Pytorch 1.8.1+cu101\n* Datasets 1.6.2\n* Tokenizers 0.10.2" ]
translation
transformers
This model is a fine-tuned version of [Helsinki-NLP/opus-tatoeba-es-zh](https://huggingface.co/Helsinki-NLP/opus-tatoeba-es-zh) on a dataset of legal domain constructed by the author himself. # Intended uses & limitations This model is the result of the master graduation thesis for the Tradumatics: Translation Technologies program at the Autonomous University of Barcelona. Please refer to the GitHub repo created for this thesis for the full-text and relative open-sourced materials: https://github.com/guocheng98/MUTTT2020_TFM_ZGC The thesis intends to explain various theories and certain algorithm details about neural machine translation, thus this fine-tuned model only serves as a hands-on practice example for that objective, without any intention of productive usage. # Training and evaluation data The dataset is constructed from the Chinese translation of Spanish Civil Code, Spanish Constitution, and many other laws & regulations found in the database China Law Info (北大法宝 Beida Fabao), along with their source text found on Boletín Oficial del Estado and EUR-Lex. There are 9972 sentence pairs constructed. 1000 are used for evaluation and the rest for training. # Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 10 - mixed_precision_training: Native AMP - weight_decay: 0.01 - early_stopping_patience: 8 # Training results Best validation loss achieved at step 5600. | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 2.9584 | 0.36 | 400 | 2.6800 | | 2.6402 | 0.71 | 800 | 2.5017 | | 2.5038 | 1.07 | 1200 | 2.3907 | | 2.3279 | 1.43 | 1600 | 2.2999 | | 2.2258 | 1.78 | 2000 | 2.2343 | | 2.1061 | 2.14 | 2400 | 2.1961 | | 1.9279 | 2.5 | 2800 | 2.1569 | | 1.9059 | 2.85 | 3200 | 2.1245 | | 1.7491 | 3.21 | 3600 | 2.1227 | | 1.6301 | 3.57 | 4000 | 2.1169 | | 1.6871 | 3.92 | 4400 | 2.0979 | | 1.5203 | 4.28 | 4800 | 2.1074 | | 1.4646 | 4.63 | 5200 | 2.1024 | | 1.4739 | 4.99 | 5600 | 2.0905 | | 1.338 | 5.35 | 6000 | 2.0946 | | 1.3152 | 5.7 | 6400 | 2.0974 | | 1.306 | 6.06 | 6800 | 2.0985 | | 1.1991 | 6.42 | 7200 | 2.0962 | | 1.2113 | 6.77 | 7600 | 2.1092 | | 1.1983 | 7.13 | 8000 | 2.1060 | | 1.1238 | 7.49 | 8400 | 2.1102 | | 1.1417 | 7.84 | 8800 | 2.1078 | # Framework versions - Transformers 4.7.0 - Pytorch 1.8.1+cu101 - Datasets 1.8.0 - Tokenizers 0.10.3
{"language": ["es", "zh"], "license": "apache-2.0", "tags": ["translation"]}
guocheng98/HelsinkiNLP-FineTuned-Legal-es-zh
null
[ "transformers", "pytorch", "marian", "text2text-generation", "translation", "es", "zh", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "es", "zh" ]
TAGS #transformers #pytorch #marian #text2text-generation #translation #es #zh #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
This model is a fine-tuned version of Helsinki-NLP/opus-tatoeba-es-zh on a dataset of legal domain constructed by the author himself. Intended uses & limitations =========================== This model is the result of the master graduation thesis for the Tradumatics: Translation Technologies program at the Autonomous University of Barcelona. Please refer to the GitHub repo created for this thesis for the full-text and relative open-sourced materials: URL The thesis intends to explain various theories and certain algorithm details about neural machine translation, thus this fine-tuned model only serves as a hands-on practice example for that objective, without any intention of productive usage. Training and evaluation data ============================ The dataset is constructed from the Chinese translation of Spanish Civil Code, Spanish Constitution, and many other laws & regulations found in the database China Law Info (北大法宝 Beida Fabao), along with their source text found on Boletín Oficial del Estado and EUR-Lex. There are 9972 sentence pairs constructed. 1000 are used for evaluation and the rest for training. Training hyperparameters ======================== The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 8 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 2000 * num\_epochs: 10 * mixed\_precision\_training: Native AMP * weight\_decay: 0.01 * early\_stopping\_patience: 8 Training results ================ Best validation loss achieved at step 5600. Framework versions ================== * Transformers 4.7.0 * Pytorch 1.8.1+cu101 * Datasets 1.8.0 * Tokenizers 0.10.3
[]
[ "TAGS\n#transformers #pytorch #marian #text2text-generation #translation #es #zh #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n" ]
null
null
# WudaoSailing WudaoSailing is a package for pretraining chinese Language Model and finetune tasks. Now it supports GLM, Bert, T5, Cogview and Roberta models. ## Get Started ### Docker Image We prepare two docker images based on CUDA 10.2 and CUDA 11.2. You can build images from the docker file [docs/docker/cuda102.dockerfile](docs/docker/cuda102.dcokerfile) or pull the pre-built images from Docker Hub and run with docker v19.03+ ```shell nvidia-docker run -id --hostname=V100 --network=host\ --ipc=host --shm-size=16gb --name=deepspeed-cuda \ -e NVIDIA_VISIBLE_DEVICES=0,1,2,3 \ -v /DATA/disk1/docker/containers/:/data deepspeed/cuda102:lastest ``` or replace `cuda102` with `cuda112`. ```shell docker build -f cuda102.dockerfile -t deepspeed/cuda102 . ``` ### Clone this repo ```shell git clone https://github.com/wangguojim/WudaoSailing.git cd WudaoSailing pip install -r requirements.txt ``` ## GLM We show some examples based on GLM model. ### finetuene We provide scripts for finetuning GLM on some downstream tasks. #### SuperGLUE - Download the [SuperGlue](https://super.gluebenchmark.com/tasks) data and check the experiment setup in [examples/glm/scripts/ds_finetune_superglue.sh](xamples/glm/scripts/ds_finetune_superglue.sh). Note that `DATA_ROOT, CHECKPOINT_PATH, SAVE_PATH` need to be changed to your local path. You may also change the `batch-size` and `nproc_per_node` according to your available hardware. - Run the following script for text similarity finetune task (use the afqmc dataset as an example) ``` cd examples/glm/ bash scripts/ds_finetune_superglue.sh\ config/model_blocklm_large_chinese.sh\ config_tasks/task_afqmc.sh ``` - Run the following script for text classification finetune task (use the thunews and thunews dataset as an example) ``` cd examples/glm/ bash scripts/ds_finetune_superglue.sh\ config/model_blocklm_large_chinese.sh\ config_tasks/task_tnews.sh ``` - Run the following script for causal inference finetune task (use the COPA dataset as an example) ``` cd examples/glm/ bash scripts/ds_finetune_superglue.sh\ config/model_blocklm_large_chinese.sh\ config_tasks/task_copa.sh ``` - To apply GLM to a new NLU dataset with cloze-filling finetuning, implement a `DataProcessor` in [examples/glm/tasks/superglue/dataset.py](examples/glm/tasks/superglue/dataset.py) for data loading and add a `PVP` in [examples/glm/tasks/superglue/pvp.py](examples/glm/tasks/superglue/pvp.py) for the cloze question. More details can be found [here](examples/glm/tasks/superglue/README.md). #### Blank Filling (Interactive) * Change `CHECKPOINT_PATH` to your local path. Run the following script ``` bash config/generate_block.sh\ config/model_blocklm_large_chinese.sh ``` ##### Example1 (Entity Prediction): Context: 凯旋门位于意大利米兰市古城堡旁。1807年为纪念[MASK]而建,门高25米,顶上矗立两武士青铜古兵车铸像。 GLM:拿破仑军队攻克米兰城 ##### Example2 (Sentence Prediction) Context: 工业互联网(Industrial Internet)是新一代信息通信技术与工业经济深度融合的新型基础设施、应用模式和工业生态,通过对人、机、物、系统等的全面连接,构建起覆盖全产业链、全价值链的全新制造和服务体系,为工业乃至产业数字化、网络化、智能化发展提供了实现途径,是第四次工业革命的重要基石。[sMASK]它以网络为基础、平台为中枢、数据为要素、安全为保障,既是工业数字化、网络化、智能化转型的基础设施,也是互联网、大数据、人工智能与实体经济深度融合的应用模式,同时也是一种新业态、新产业,将重塑企业形态、供应链和产业链。当前,工业互联网融合应用向国民经济重点行业广泛拓展,形成平台化设计、智能化制造、网络化协同、个性化定制、服务化延伸、数字化管理六大新模式,赋能、赋智、赋值作用不断显现,有力的促进了实体经济提质、增效、降本、绿色、安全发展。 GLM: 工业互联网是制造业技术、管理、模式的重大变革,是推动互联网、大数据、人工智能和实体经济深度融合的重要载体,是建设制造强国和网络强国的重要基础。 ##### Example3 (Long Text Generation) Context: 问题:高斯所在的国家有什么汽车品牌?答案:[gMASK] GLM:答案:[gMASK]<|startofpiece|>德国奔驰、德国大众、别克、沃尔沃、斯柯达、本田、雪铁龙. ### Ptuning Run the following script to integrate p-tuning with GLM: ```shell cd algutils/ptuning/ bash finetune_zy.sh ``` ### Pretrain Run the following script to pre-train the GLM-Large model ```shell cd examples/glm/ bash scripts/ds_pretrain_nvidia.sh config/ds_block_large.sh ``` The script [examples/glm/config/ds_pretrain_nvidia.sh](examples/glm/config/ds_pretrain_nvidia.sh) launches the training program with DeepSpeed. You should change `NUM_WORKERS` and `NUM_GPUS_PER_WORKER` to the number of workers and the number of gpus per worker. Also change `HOST_FILE_PATH` to the path to an OpenMPI-style hostfile. More details about DeepSpeed launcher can be found [here](https://www.deepspeed.ai/getting-started/#resource-configuration-multi-node). The file [examples/glm/config/ds_block_large.sh](examples/glm/config/ds_block_large.sh) defines the hyperparameters for pretraining. Most of the arguments are fairly self-explanatory. Specifically, `--train-data` can be multiple keywords defined in `NAMED_CORPORA` in [data_utils/corpora.py](data_utils/corpora.py). The hyperparameters of the optimizer are defined in the corresponding json file under `config`. The semantics of the json file can be found [here](https://www.deepspeed.ai/docs/config-json). ## Bert We show some examples based on GLM model. ### Pretrain Run the following script to pre-train the Bert model ```shell cd examples/bert/ python quick_start.py ``` ## CogView ### Pretrain Run the following script to pre-train the cogview model ```shell cd examples/cogview/ bash config/pretrain_multiple_nodes.sh ``` ### inference Run the following script to test the ability of text2image ```shell cd examples/cogview/ bash config/text2image_cogview.sh ```
{}
guoqiang/WuDaoSailing
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
# WudaoSailing WudaoSailing is a package for pretraining chinese Language Model and finetune tasks. Now it supports GLM, Bert, T5, Cogview and Roberta models. ## Get Started ### Docker Image We prepare two docker images based on CUDA 10.2 and CUDA 11.2. You can build images from the docker file docs/docker/cuda102.dockerfile or pull the pre-built images from Docker Hub and run with docker v19.03+ or replace 'cuda102' with 'cuda112'. ### Clone this repo ## GLM We show some examples based on GLM model. ### finetuene We provide scripts for finetuning GLM on some downstream tasks. #### SuperGLUE - Download the SuperGlue data and check the experiment setup in examples/glm/scripts/ds_finetune_superglue.sh. Note that 'DATA_ROOT, CHECKPOINT_PATH, SAVE_PATH' need to be changed to your local path. You may also change the 'batch-size' and 'nproc_per_node' according to your available hardware. - Run the following script for text similarity finetune task (use the afqmc dataset as an example) - Run the following script for text classification finetune task (use the thunews and thunews dataset as an example) - Run the following script for causal inference finetune task (use the COPA dataset as an example) - To apply GLM to a new NLU dataset with cloze-filling finetuning, implement a 'DataProcessor' in examples/glm/tasks/superglue/URL for data loading and add a 'PVP' in examples/glm/tasks/superglue/URL for the cloze question. More details can be found here. #### Blank Filling (Interactive) * Change 'CHECKPOINT_PATH' to your local path. Run the following script ##### Example1 (Entity Prediction): Context: 凯旋门位于意大利米兰市古城堡旁。1807年为纪念[MASK]而建,门高25米,顶上矗立两武士青铜古兵车铸像。 GLM:拿破仑军队攻克米兰城 ##### Example2 (Sentence Prediction) Context: 工业互联网(Industrial Internet)是新一代信息通信技术与工业经济深度融合的新型基础设施、应用模式和工业生态,通过对人、机、物、系统等的全面连接,构建起覆盖全产业链、全价值链的全新制造和服务体系,为工业乃至产业数字化、网络化、智能化发展提供了实现途径,是第四次工业革命的重要基石。[sMASK]它以网络为基础、平台为中枢、数据为要素、安全为保障,既是工业数字化、网络化、智能化转型的基础设施,也是互联网、大数据、人工智能与实体经济深度融合的应用模式,同时也是一种新业态、新产业,将重塑企业形态、供应链和产业链。当前,工业互联网融合应用向国民经济重点行业广泛拓展,形成平台化设计、智能化制造、网络化协同、个性化定制、服务化延伸、数字化管理六大新模式,赋能、赋智、赋值作用不断显现,有力的促进了实体经济提质、增效、降本、绿色、安全发展。 GLM: 工业互联网是制造业技术、管理、模式的重大变革,是推动互联网、大数据、人工智能和实体经济深度融合的重要载体,是建设制造强国和网络强国的重要基础。 ##### Example3 (Long Text Generation) Context: 问题:高斯所在的国家有什么汽车品牌?答案:[gMASK] GLM:答案:[gMASK]<|startofpiece|>德国奔驰、德国大众、别克、沃尔沃、斯柯达、本田、雪铁龙. ### Ptuning Run the following script to integrate p-tuning with GLM: ### Pretrain Run the following script to pre-train the GLM-Large model The script examples/glm/config/ds_pretrain_nvidia.sh launches the training program with DeepSpeed. You should change 'NUM_WORKERS' and 'NUM_GPUS_PER_WORKER' to the number of workers and the number of gpus per worker. Also change 'HOST_FILE_PATH' to the path to an OpenMPI-style hostfile. More details about DeepSpeed launcher can be found here. The file examples/glm/config/ds_block_large.sh defines the hyperparameters for pretraining. Most of the arguments are fairly self-explanatory. Specifically, '--train-data' can be multiple keywords defined in 'NAMED_CORPORA' in data_utils/URL. The hyperparameters of the optimizer are defined in the corresponding json file under 'config'. The semantics of the json file can be found here. ## Bert We show some examples based on GLM model. ### Pretrain Run the following script to pre-train the Bert model ## CogView ### Pretrain Run the following script to pre-train the cogview model ### inference Run the following script to test the ability of text2image
[ "# WudaoSailing\n\nWudaoSailing is a package for pretraining chinese Language Model and finetune tasks. Now it supports GLM, Bert, T5, Cogview and Roberta models.", "## Get Started", "### Docker Image\nWe prepare two docker images based on CUDA 10.2 and CUDA 11.2. You can build images from the docker file docs/docker/cuda102.dockerfile or pull the pre-built images from Docker Hub and run with docker v19.03+\n \n or replace 'cuda102' with 'cuda112'.", "### Clone this repo", "## GLM\n\nWe show some examples based on GLM model.", "### finetuene\nWe provide scripts for finetuning GLM on some downstream tasks.", "#### SuperGLUE\n\n- Download the SuperGlue data and check the experiment setup in \n examples/glm/scripts/ds_finetune_superglue.sh. Note that 'DATA_ROOT, CHECKPOINT_PATH, SAVE_PATH' \n need to be changed to your local path. You may also change the 'batch-size' and 'nproc_per_node' according to your \n available hardware.\n\n- Run the following script for text similarity finetune task (use the afqmc dataset as an example)\n\n\n\n\n- Run the following script for text classification finetune task (use the thunews and thunews dataset as an example)\n\n\n\n- Run the following script for causal inference finetune task (use the COPA dataset as an example)\n\n\n \n- To apply GLM to a new NLU dataset with cloze-filling finetuning, implement a 'DataProcessor' in\n examples/glm/tasks/superglue/URL for data loading and add a 'PVP' in \n examples/glm/tasks/superglue/URL for the cloze question. More details can be found \n here.", "#### Blank Filling (Interactive)\n* Change 'CHECKPOINT_PATH' to your local path. Run the following script", "##### Example1 (Entity Prediction):\n\nContext: 凯旋门位于意大利米兰市古城堡旁。1807年为纪念[MASK]而建,门高25米,顶上矗立两武士青铜古兵车铸像。\n\nGLM:拿破仑军队攻克米兰城", "##### Example2 (Sentence Prediction)\nContext: 工业互联网(Industrial Internet)是新一代信息通信技术与工业经济深度融合的新型基础设施、应用模式和工业生态,通过对人、机、物、系统等的全面连接,构建起覆盖全产业链、全价值链的全新制造和服务体系,为工业乃至产业数字化、网络化、智能化发展提供了实现途径,是第四次工业革命的重要基石。[sMASK]它以网络为基础、平台为中枢、数据为要素、安全为保障,既是工业数字化、网络化、智能化转型的基础设施,也是互联网、大数据、人工智能与实体经济深度融合的应用模式,同时也是一种新业态、新产业,将重塑企业形态、供应链和产业链。当前,工业互联网融合应用向国民经济重点行业广泛拓展,形成平台化设计、智能化制造、网络化协同、个性化定制、服务化延伸、数字化管理六大新模式,赋能、赋智、赋值作用不断显现,有力的促进了实体经济提质、增效、降本、绿色、安全发展。\n\nGLM: 工业互联网是制造业技术、管理、模式的重大变革,是推动互联网、大数据、人工智能和实体经济深度融合的重要载体,是建设制造强国和网络强国的重要基础。", "##### Example3 (Long Text Generation)\nContext: 问题:高斯所在的国家有什么汽车品牌?答案:[gMASK]\n\nGLM:答案:[gMASK]<|startofpiece|>德国奔驰、德国大众、别克、沃尔沃、斯柯达、本田、雪铁龙.", "### Ptuning\nRun the following script to integrate p-tuning with GLM:", "### Pretrain\nRun the following script to pre-train the GLM-Large model\n\n\nThe script examples/glm/config/ds_pretrain_nvidia.sh launches the training program with DeepSpeed. You should change 'NUM_WORKERS' and 'NUM_GPUS_PER_WORKER' to the number of workers and the number of gpus per worker. Also change 'HOST_FILE_PATH' to the path to an OpenMPI-style hostfile. More details about DeepSpeed launcher can be found here.\n\nThe file examples/glm/config/ds_block_large.sh defines the hyperparameters for pretraining. Most of the arguments are fairly self-explanatory. Specifically, '--train-data' can be multiple keywords defined in 'NAMED_CORPORA' in data_utils/URL. The hyperparameters of the optimizer are defined in the corresponding json file under 'config'. The semantics of the json file can be found here.", "## Bert\n\nWe show some examples based on GLM model.", "### Pretrain\nRun the following script to pre-train the Bert model", "## CogView", "### Pretrain\nRun the following script to pre-train the cogview model", "### inference\nRun the following script to test the ability of text2image" ]
[ "TAGS\n#region-us \n", "# WudaoSailing\n\nWudaoSailing is a package for pretraining chinese Language Model and finetune tasks. Now it supports GLM, Bert, T5, Cogview and Roberta models.", "## Get Started", "### Docker Image\nWe prepare two docker images based on CUDA 10.2 and CUDA 11.2. You can build images from the docker file docs/docker/cuda102.dockerfile or pull the pre-built images from Docker Hub and run with docker v19.03+\n \n or replace 'cuda102' with 'cuda112'.", "### Clone this repo", "## GLM\n\nWe show some examples based on GLM model.", "### finetuene\nWe provide scripts for finetuning GLM on some downstream tasks.", "#### SuperGLUE\n\n- Download the SuperGlue data and check the experiment setup in \n examples/glm/scripts/ds_finetune_superglue.sh. Note that 'DATA_ROOT, CHECKPOINT_PATH, SAVE_PATH' \n need to be changed to your local path. You may also change the 'batch-size' and 'nproc_per_node' according to your \n available hardware.\n\n- Run the following script for text similarity finetune task (use the afqmc dataset as an example)\n\n\n\n\n- Run the following script for text classification finetune task (use the thunews and thunews dataset as an example)\n\n\n\n- Run the following script for causal inference finetune task (use the COPA dataset as an example)\n\n\n \n- To apply GLM to a new NLU dataset with cloze-filling finetuning, implement a 'DataProcessor' in\n examples/glm/tasks/superglue/URL for data loading and add a 'PVP' in \n examples/glm/tasks/superglue/URL for the cloze question. More details can be found \n here.", "#### Blank Filling (Interactive)\n* Change 'CHECKPOINT_PATH' to your local path. Run the following script", "##### Example1 (Entity Prediction):\n\nContext: 凯旋门位于意大利米兰市古城堡旁。1807年为纪念[MASK]而建,门高25米,顶上矗立两武士青铜古兵车铸像。\n\nGLM:拿破仑军队攻克米兰城", "##### Example2 (Sentence Prediction)\nContext: 工业互联网(Industrial Internet)是新一代信息通信技术与工业经济深度融合的新型基础设施、应用模式和工业生态,通过对人、机、物、系统等的全面连接,构建起覆盖全产业链、全价值链的全新制造和服务体系,为工业乃至产业数字化、网络化、智能化发展提供了实现途径,是第四次工业革命的重要基石。[sMASK]它以网络为基础、平台为中枢、数据为要素、安全为保障,既是工业数字化、网络化、智能化转型的基础设施,也是互联网、大数据、人工智能与实体经济深度融合的应用模式,同时也是一种新业态、新产业,将重塑企业形态、供应链和产业链。当前,工业互联网融合应用向国民经济重点行业广泛拓展,形成平台化设计、智能化制造、网络化协同、个性化定制、服务化延伸、数字化管理六大新模式,赋能、赋智、赋值作用不断显现,有力的促进了实体经济提质、增效、降本、绿色、安全发展。\n\nGLM: 工业互联网是制造业技术、管理、模式的重大变革,是推动互联网、大数据、人工智能和实体经济深度融合的重要载体,是建设制造强国和网络强国的重要基础。", "##### Example3 (Long Text Generation)\nContext: 问题:高斯所在的国家有什么汽车品牌?答案:[gMASK]\n\nGLM:答案:[gMASK]<|startofpiece|>德国奔驰、德国大众、别克、沃尔沃、斯柯达、本田、雪铁龙.", "### Ptuning\nRun the following script to integrate p-tuning with GLM:", "### Pretrain\nRun the following script to pre-train the GLM-Large model\n\n\nThe script examples/glm/config/ds_pretrain_nvidia.sh launches the training program with DeepSpeed. You should change 'NUM_WORKERS' and 'NUM_GPUS_PER_WORKER' to the number of workers and the number of gpus per worker. Also change 'HOST_FILE_PATH' to the path to an OpenMPI-style hostfile. More details about DeepSpeed launcher can be found here.\n\nThe file examples/glm/config/ds_block_large.sh defines the hyperparameters for pretraining. Most of the arguments are fairly self-explanatory. Specifically, '--train-data' can be multiple keywords defined in 'NAMED_CORPORA' in data_utils/URL. The hyperparameters of the optimizer are defined in the corresponding json file under 'config'. The semantics of the json file can be found here.", "## Bert\n\nWe show some examples based on GLM model.", "### Pretrain\nRun the following script to pre-train the Bert model", "## CogView", "### Pretrain\nRun the following script to pre-train the cogview model", "### inference\nRun the following script to test the ability of text2image" ]
null
null
# WudaoSailing WudaoSailing is a package for pretraining chinese Language Model and finetune tasks. Now it supports GLM, Bert, T5, Cogview and Roberta models. ## Get Started ### Docker Image We prepare two docker images based on CUDA 10.2 and CUDA 11.2. You can build images from the docker file [docs/docker/cuda102.dockerfile](docs/docker/cuda102.dcokerfile) or pull the pre-built images from Docker Hub and run with docker v19.03+ ```shell nvidia-docker run -id --hostname=V100 --network=host\ --ipc=host --shm-size=16gb --name=deepspeed-cuda \ -e NVIDIA_VISIBLE_DEVICES=0,1,2,3 \ -v /DATA/disk1/docker/containers/:/data deepspeed/cuda102:lastest ``` or replace `cuda102` with `cuda112`. ```shell docker build -f cuda102.dockerfile -t deepspeed/cuda102 . ``` ### Clone this repo ```shell git clone https://github.com/wangguojim/WudaoSailing.git cd WudaoSailing pip install -r requirements.txt ``` ## GLM We show some examples based on GLM model. ### finetuene We provide scripts for finetuning GLM on some downstream tasks. #### SuperGLUE - Download the [SuperGlue](https://super.gluebenchmark.com/tasks) data and check the experiment setup in [examples/glm/scripts/ds_finetune_superglue.sh](xamples/glm/scripts/ds_finetune_superglue.sh). Note that `DATA_ROOT, CHECKPOINT_PATH, SAVE_PATH` need to be changed to your local path. You may also change the `batch-size` and `nproc_per_node` according to your available hardware. - Run the following script for text similarity finetune task (use the afqmc dataset as an example) ``` cd examples/glm/ bash scripts/ds_finetune_superglue.sh\ config/model_blocklm_large_chinese.sh\ config_tasks/task_afqmc.sh ``` - Run the following script for text classification finetune task (use the thunews and thunews dataset as an example) ``` cd examples/glm/ bash scripts/ds_finetune_superglue.sh\ config/model_blocklm_large_chinese.sh\ config_tasks/task_tnews.sh ``` - Run the following script for causal inference finetune task (use the COPA dataset as an example) ``` cd examples/glm/ bash scripts/ds_finetune_superglue.sh\ config/model_blocklm_large_chinese.sh\ config_tasks/task_copa.sh ``` - To apply GLM to a new NLU dataset with cloze-filling finetuning, implement a `DataProcessor` in [examples/glm/tasks/superglue/dataset.py](examples/glm/tasks/superglue/dataset.py) for data loading and add a `PVP` in [examples/glm/tasks/superglue/pvp.py](examples/glm/tasks/superglue/pvp.py) for the cloze question. More details can be found [here](examples/glm/tasks/superglue/README.md). #### Blank Filling (Interactive) * Change `CHECKPOINT_PATH` to your local path. Run the following script ``` bash config/generate_block.sh\ config/model_blocklm_large_chinese.sh ``` ##### Example1 (Entity Prediction): Context: 凯旋门位于意大利米兰市古城堡旁。1807年为纪念[MASK]而建,门高25米,顶上矗立两武士青铜古兵车铸像。 GLM:拿破仑军队攻克米兰城 ##### Example2 (Sentence Prediction) Context: 工业互联网(Industrial Internet)是新一代信息通信技术与工业经济深度融合的新型基础设施、应用模式和工业生态,通过对人、机、物、系统等的全面连接,构建起覆盖全产业链、全价值链的全新制造和服务体系,为工业乃至产业数字化、网络化、智能化发展提供了实现途径,是第四次工业革命的重要基石。[sMASK]它以网络为基础、平台为中枢、数据为要素、安全为保障,既是工业数字化、网络化、智能化转型的基础设施,也是互联网、大数据、人工智能与实体经济深度融合的应用模式,同时也是一种新业态、新产业,将重塑企业形态、供应链和产业链。当前,工业互联网融合应用向国民经济重点行业广泛拓展,形成平台化设计、智能化制造、网络化协同、个性化定制、服务化延伸、数字化管理六大新模式,赋能、赋智、赋值作用不断显现,有力的促进了实体经济提质、增效、降本、绿色、安全发展。 GLM: 工业互联网是制造业技术、管理、模式的重大变革,是推动互联网、大数据、人工智能和实体经济深度融合的重要载体,是建设制造强国和网络强国的重要基础。 ##### Example3 (Long Text Generation) Context: 问题:高斯所在的国家有什么汽车品牌?答案:[gMASK] GLM:答案:[gMASK]<|startofpiece|>德国奔驰、德国大众、别克、沃尔沃、斯柯达、本田、雪铁龙. ### Ptuning Run the following script to integrate p-tuning with GLM: ```shell cd algutils/ptuning/ bash finetune_zy.sh ``` ### Pretrain Run the following script to pre-train the GLM-Large model ```shell cd examples/glm/ bash scripts/ds_pretrain_nvidia.sh config/ds_block_large.sh ``` The script [examples/glm/config/ds_pretrain_nvidia.sh](examples/glm/config/ds_pretrain_nvidia.sh) launches the training program with DeepSpeed. You should change `NUM_WORKERS` and `NUM_GPUS_PER_WORKER` to the number of workers and the number of gpus per worker. Also change `HOST_FILE_PATH` to the path to an OpenMPI-style hostfile. More details about DeepSpeed launcher can be found [here](https://www.deepspeed.ai/getting-started/#resource-configuration-multi-node). The file [examples/glm/config/ds_block_large.sh](examples/glm/config/ds_block_large.sh) defines the hyperparameters for pretraining. Most of the arguments are fairly self-explanatory. Specifically, `--train-data` can be multiple keywords defined in `NAMED_CORPORA` in [data_utils/corpora.py](data_utils/corpora.py). The hyperparameters of the optimizer are defined in the corresponding json file under `config`. The semantics of the json file can be found [here](https://www.deepspeed.ai/docs/config-json). ## Bert We show some examples based on GLM model. ### Pretrain Run the following script to pre-train the Bert model ```shell cd examples/bert/ python quick_start.py ``` ## CogView ### Pretrain Run the following script to pre-train the cogview model ```shell cd examples/cogview/ bash config/pretrain_multiple_nodes.sh ``` ### inference Run the following script to test the ability of text2image ```shell cd examples/cogview/ bash config/text2image_cogview.sh ```
{}
guoqiang/glm
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
# WudaoSailing WudaoSailing is a package for pretraining chinese Language Model and finetune tasks. Now it supports GLM, Bert, T5, Cogview and Roberta models. ## Get Started ### Docker Image We prepare two docker images based on CUDA 10.2 and CUDA 11.2. You can build images from the docker file docs/docker/cuda102.dockerfile or pull the pre-built images from Docker Hub and run with docker v19.03+ or replace 'cuda102' with 'cuda112'. ### Clone this repo ## GLM We show some examples based on GLM model. ### finetuene We provide scripts for finetuning GLM on some downstream tasks. #### SuperGLUE - Download the SuperGlue data and check the experiment setup in examples/glm/scripts/ds_finetune_superglue.sh. Note that 'DATA_ROOT, CHECKPOINT_PATH, SAVE_PATH' need to be changed to your local path. You may also change the 'batch-size' and 'nproc_per_node' according to your available hardware. - Run the following script for text similarity finetune task (use the afqmc dataset as an example) - Run the following script for text classification finetune task (use the thunews and thunews dataset as an example) - Run the following script for causal inference finetune task (use the COPA dataset as an example) - To apply GLM to a new NLU dataset with cloze-filling finetuning, implement a 'DataProcessor' in examples/glm/tasks/superglue/URL for data loading and add a 'PVP' in examples/glm/tasks/superglue/URL for the cloze question. More details can be found here. #### Blank Filling (Interactive) * Change 'CHECKPOINT_PATH' to your local path. Run the following script ##### Example1 (Entity Prediction): Context: 凯旋门位于意大利米兰市古城堡旁。1807年为纪念[MASK]而建,门高25米,顶上矗立两武士青铜古兵车铸像。 GLM:拿破仑军队攻克米兰城 ##### Example2 (Sentence Prediction) Context: 工业互联网(Industrial Internet)是新一代信息通信技术与工业经济深度融合的新型基础设施、应用模式和工业生态,通过对人、机、物、系统等的全面连接,构建起覆盖全产业链、全价值链的全新制造和服务体系,为工业乃至产业数字化、网络化、智能化发展提供了实现途径,是第四次工业革命的重要基石。[sMASK]它以网络为基础、平台为中枢、数据为要素、安全为保障,既是工业数字化、网络化、智能化转型的基础设施,也是互联网、大数据、人工智能与实体经济深度融合的应用模式,同时也是一种新业态、新产业,将重塑企业形态、供应链和产业链。当前,工业互联网融合应用向国民经济重点行业广泛拓展,形成平台化设计、智能化制造、网络化协同、个性化定制、服务化延伸、数字化管理六大新模式,赋能、赋智、赋值作用不断显现,有力的促进了实体经济提质、增效、降本、绿色、安全发展。 GLM: 工业互联网是制造业技术、管理、模式的重大变革,是推动互联网、大数据、人工智能和实体经济深度融合的重要载体,是建设制造强国和网络强国的重要基础。 ##### Example3 (Long Text Generation) Context: 问题:高斯所在的国家有什么汽车品牌?答案:[gMASK] GLM:答案:[gMASK]<|startofpiece|>德国奔驰、德国大众、别克、沃尔沃、斯柯达、本田、雪铁龙. ### Ptuning Run the following script to integrate p-tuning with GLM: ### Pretrain Run the following script to pre-train the GLM-Large model The script examples/glm/config/ds_pretrain_nvidia.sh launches the training program with DeepSpeed. You should change 'NUM_WORKERS' and 'NUM_GPUS_PER_WORKER' to the number of workers and the number of gpus per worker. Also change 'HOST_FILE_PATH' to the path to an OpenMPI-style hostfile. More details about DeepSpeed launcher can be found here. The file examples/glm/config/ds_block_large.sh defines the hyperparameters for pretraining. Most of the arguments are fairly self-explanatory. Specifically, '--train-data' can be multiple keywords defined in 'NAMED_CORPORA' in data_utils/URL. The hyperparameters of the optimizer are defined in the corresponding json file under 'config'. The semantics of the json file can be found here. ## Bert We show some examples based on GLM model. ### Pretrain Run the following script to pre-train the Bert model ## CogView ### Pretrain Run the following script to pre-train the cogview model ### inference Run the following script to test the ability of text2image
[ "# WudaoSailing\n\nWudaoSailing is a package for pretraining chinese Language Model and finetune tasks. Now it supports GLM, Bert, T5, Cogview and Roberta models.", "## Get Started", "### Docker Image\nWe prepare two docker images based on CUDA 10.2 and CUDA 11.2. You can build images from the docker file docs/docker/cuda102.dockerfile or pull the pre-built images from Docker Hub and run with docker v19.03+\n \n or replace 'cuda102' with 'cuda112'.", "### Clone this repo", "## GLM\n\nWe show some examples based on GLM model.", "### finetuene\nWe provide scripts for finetuning GLM on some downstream tasks.", "#### SuperGLUE\n\n- Download the SuperGlue data and check the experiment setup in \n examples/glm/scripts/ds_finetune_superglue.sh. Note that 'DATA_ROOT, CHECKPOINT_PATH, SAVE_PATH' \n need to be changed to your local path. You may also change the 'batch-size' and 'nproc_per_node' according to your \n available hardware.\n\n- Run the following script for text similarity finetune task (use the afqmc dataset as an example)\n\n\n\n\n- Run the following script for text classification finetune task (use the thunews and thunews dataset as an example)\n\n\n\n- Run the following script for causal inference finetune task (use the COPA dataset as an example)\n\n\n \n- To apply GLM to a new NLU dataset with cloze-filling finetuning, implement a 'DataProcessor' in\n examples/glm/tasks/superglue/URL for data loading and add a 'PVP' in \n examples/glm/tasks/superglue/URL for the cloze question. More details can be found \n here.", "#### Blank Filling (Interactive)\n* Change 'CHECKPOINT_PATH' to your local path. Run the following script", "##### Example1 (Entity Prediction):\n\nContext: 凯旋门位于意大利米兰市古城堡旁。1807年为纪念[MASK]而建,门高25米,顶上矗立两武士青铜古兵车铸像。\n\nGLM:拿破仑军队攻克米兰城", "##### Example2 (Sentence Prediction)\nContext: 工业互联网(Industrial Internet)是新一代信息通信技术与工业经济深度融合的新型基础设施、应用模式和工业生态,通过对人、机、物、系统等的全面连接,构建起覆盖全产业链、全价值链的全新制造和服务体系,为工业乃至产业数字化、网络化、智能化发展提供了实现途径,是第四次工业革命的重要基石。[sMASK]它以网络为基础、平台为中枢、数据为要素、安全为保障,既是工业数字化、网络化、智能化转型的基础设施,也是互联网、大数据、人工智能与实体经济深度融合的应用模式,同时也是一种新业态、新产业,将重塑企业形态、供应链和产业链。当前,工业互联网融合应用向国民经济重点行业广泛拓展,形成平台化设计、智能化制造、网络化协同、个性化定制、服务化延伸、数字化管理六大新模式,赋能、赋智、赋值作用不断显现,有力的促进了实体经济提质、增效、降本、绿色、安全发展。\n\nGLM: 工业互联网是制造业技术、管理、模式的重大变革,是推动互联网、大数据、人工智能和实体经济深度融合的重要载体,是建设制造强国和网络强国的重要基础。", "##### Example3 (Long Text Generation)\nContext: 问题:高斯所在的国家有什么汽车品牌?答案:[gMASK]\n\nGLM:答案:[gMASK]<|startofpiece|>德国奔驰、德国大众、别克、沃尔沃、斯柯达、本田、雪铁龙.", "### Ptuning\nRun the following script to integrate p-tuning with GLM:", "### Pretrain\nRun the following script to pre-train the GLM-Large model\n\n\nThe script examples/glm/config/ds_pretrain_nvidia.sh launches the training program with DeepSpeed. You should change 'NUM_WORKERS' and 'NUM_GPUS_PER_WORKER' to the number of workers and the number of gpus per worker. Also change 'HOST_FILE_PATH' to the path to an OpenMPI-style hostfile. More details about DeepSpeed launcher can be found here.\n\nThe file examples/glm/config/ds_block_large.sh defines the hyperparameters for pretraining. Most of the arguments are fairly self-explanatory. Specifically, '--train-data' can be multiple keywords defined in 'NAMED_CORPORA' in data_utils/URL. The hyperparameters of the optimizer are defined in the corresponding json file under 'config'. The semantics of the json file can be found here.", "## Bert\n\nWe show some examples based on GLM model.", "### Pretrain\nRun the following script to pre-train the Bert model", "## CogView", "### Pretrain\nRun the following script to pre-train the cogview model", "### inference\nRun the following script to test the ability of text2image" ]
[ "TAGS\n#region-us \n", "# WudaoSailing\n\nWudaoSailing is a package for pretraining chinese Language Model and finetune tasks. Now it supports GLM, Bert, T5, Cogview and Roberta models.", "## Get Started", "### Docker Image\nWe prepare two docker images based on CUDA 10.2 and CUDA 11.2. You can build images from the docker file docs/docker/cuda102.dockerfile or pull the pre-built images from Docker Hub and run with docker v19.03+\n \n or replace 'cuda102' with 'cuda112'.", "### Clone this repo", "## GLM\n\nWe show some examples based on GLM model.", "### finetuene\nWe provide scripts for finetuning GLM on some downstream tasks.", "#### SuperGLUE\n\n- Download the SuperGlue data and check the experiment setup in \n examples/glm/scripts/ds_finetune_superglue.sh. Note that 'DATA_ROOT, CHECKPOINT_PATH, SAVE_PATH' \n need to be changed to your local path. You may also change the 'batch-size' and 'nproc_per_node' according to your \n available hardware.\n\n- Run the following script for text similarity finetune task (use the afqmc dataset as an example)\n\n\n\n\n- Run the following script for text classification finetune task (use the thunews and thunews dataset as an example)\n\n\n\n- Run the following script for causal inference finetune task (use the COPA dataset as an example)\n\n\n \n- To apply GLM to a new NLU dataset with cloze-filling finetuning, implement a 'DataProcessor' in\n examples/glm/tasks/superglue/URL for data loading and add a 'PVP' in \n examples/glm/tasks/superglue/URL for the cloze question. More details can be found \n here.", "#### Blank Filling (Interactive)\n* Change 'CHECKPOINT_PATH' to your local path. Run the following script", "##### Example1 (Entity Prediction):\n\nContext: 凯旋门位于意大利米兰市古城堡旁。1807年为纪念[MASK]而建,门高25米,顶上矗立两武士青铜古兵车铸像。\n\nGLM:拿破仑军队攻克米兰城", "##### Example2 (Sentence Prediction)\nContext: 工业互联网(Industrial Internet)是新一代信息通信技术与工业经济深度融合的新型基础设施、应用模式和工业生态,通过对人、机、物、系统等的全面连接,构建起覆盖全产业链、全价值链的全新制造和服务体系,为工业乃至产业数字化、网络化、智能化发展提供了实现途径,是第四次工业革命的重要基石。[sMASK]它以网络为基础、平台为中枢、数据为要素、安全为保障,既是工业数字化、网络化、智能化转型的基础设施,也是互联网、大数据、人工智能与实体经济深度融合的应用模式,同时也是一种新业态、新产业,将重塑企业形态、供应链和产业链。当前,工业互联网融合应用向国民经济重点行业广泛拓展,形成平台化设计、智能化制造、网络化协同、个性化定制、服务化延伸、数字化管理六大新模式,赋能、赋智、赋值作用不断显现,有力的促进了实体经济提质、增效、降本、绿色、安全发展。\n\nGLM: 工业互联网是制造业技术、管理、模式的重大变革,是推动互联网、大数据、人工智能和实体经济深度融合的重要载体,是建设制造强国和网络强国的重要基础。", "##### Example3 (Long Text Generation)\nContext: 问题:高斯所在的国家有什么汽车品牌?答案:[gMASK]\n\nGLM:答案:[gMASK]<|startofpiece|>德国奔驰、德国大众、别克、沃尔沃、斯柯达、本田、雪铁龙.", "### Ptuning\nRun the following script to integrate p-tuning with GLM:", "### Pretrain\nRun the following script to pre-train the GLM-Large model\n\n\nThe script examples/glm/config/ds_pretrain_nvidia.sh launches the training program with DeepSpeed. You should change 'NUM_WORKERS' and 'NUM_GPUS_PER_WORKER' to the number of workers and the number of gpus per worker. Also change 'HOST_FILE_PATH' to the path to an OpenMPI-style hostfile. More details about DeepSpeed launcher can be found here.\n\nThe file examples/glm/config/ds_block_large.sh defines the hyperparameters for pretraining. Most of the arguments are fairly self-explanatory. Specifically, '--train-data' can be multiple keywords defined in 'NAMED_CORPORA' in data_utils/URL. The hyperparameters of the optimizer are defined in the corresponding json file under 'config'. The semantics of the json file can be found here.", "## Bert\n\nWe show some examples based on GLM model.", "### Pretrain\nRun the following script to pre-train the Bert model", "## CogView", "### Pretrain\nRun the following script to pre-train the cogview model", "### inference\nRun the following script to test the ability of text2image" ]
text-classification
transformers
# Turkish News Text Classification Turkish text classification model obtained by fine-tuning the Turkish bert model (dbmdz/bert-base-turkish-cased) # Dataset Dataset consists of 11 classes were obtained from https://www.trthaber.com/. The model was created using the most distinctive 6 classes. Dataset can be accessed at https://github.com/gurkan08/datasets/tree/master/trt_11_category. label_dict = { 'LABEL_0': 'ekonomi', 'LABEL_1': 'spor', 'LABEL_2': 'saglik', 'LABEL_3': 'kultur_sanat', 'LABEL_4': 'bilim_teknoloji', 'LABEL_5': 'egitim' } 70% of the data were used for training and 30% for testing. train f1-weighted score = %97 test f1-weighted score = %94 # Usage from transformers import pipeline, AutoTokenizer, AutoModelForSequenceClassification tokenizer = AutoTokenizer.from_pretrained("gurkan08/bert-turkish-text-classification") model = AutoModelForSequenceClassification.from_pretrained("gurkan08/bert-turkish-text-classification") nlp = pipeline("sentiment-analysis", model=model, tokenizer=tokenizer) text = ["Süper Lig'in 6. haftasında Sivasspor ile Çaykur Rizespor karşı karşıya geldi...", "Son 24 saatte 69 kişi Kovid-19 nedeniyle yaşamını yitirdi, 1573 kişi iyileşti"] out = nlp(text) label_dict = { 'LABEL_0': 'ekonomi', 'LABEL_1': 'spor', 'LABEL_2': 'saglik', 'LABEL_3': 'kultur_sanat', 'LABEL_4': 'bilim_teknoloji', 'LABEL_5': 'egitim' } results = [] for result in out: result['label'] = label_dict[result['label']] results.append(result) print(results) # > [{'label': 'spor', 'score': 0.9992026090621948}, {'label': 'saglik', 'score': 0.9972177147865295}]
{"language": "tr"}
gurkan08/bert-turkish-text-classification
null
[ "transformers", "pytorch", "jax", "bert", "text-classification", "tr", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "tr" ]
TAGS #transformers #pytorch #jax #bert #text-classification #tr #autotrain_compatible #endpoints_compatible #region-us
# Turkish News Text Classification Turkish text classification model obtained by fine-tuning the Turkish bert model (dbmdz/bert-base-turkish-cased) # Dataset Dataset consists of 11 classes were obtained from URL The model was created using the most distinctive 6 classes. Dataset can be accessed at URL label_dict = { 'LABEL_0': 'ekonomi', 'LABEL_1': 'spor', 'LABEL_2': 'saglik', 'LABEL_3': 'kultur_sanat', 'LABEL_4': 'bilim_teknoloji', 'LABEL_5': 'egitim' } 70% of the data were used for training and 30% for testing. train f1-weighted score = %97 test f1-weighted score = %94 # Usage from transformers import pipeline, AutoTokenizer, AutoModelForSequenceClassification tokenizer = AutoTokenizer.from_pretrained("gurkan08/bert-turkish-text-classification") model = AutoModelForSequenceClassification.from_pretrained("gurkan08/bert-turkish-text-classification") nlp = pipeline("sentiment-analysis", model=model, tokenizer=tokenizer) text = ["Süper Lig'in 6. haftasında Sivasspor ile Çaykur Rizespor karşı karşıya geldi...", "Son 24 saatte 69 kişi Kovid-19 nedeniyle yaşamını yitirdi, 1573 kişi iyileşti"] out = nlp(text) label_dict = { 'LABEL_0': 'ekonomi', 'LABEL_1': 'spor', 'LABEL_2': 'saglik', 'LABEL_3': 'kultur_sanat', 'LABEL_4': 'bilim_teknoloji', 'LABEL_5': 'egitim' } results = [] for result in out: result['label'] = label_dict[result['label']] URL(result) print(results) # > [{'label': 'spor', 'score': 0.9992026090621948}, {'label': 'saglik', 'score': 0.9972177147865295}]
[ "# Turkish News Text Classification\n\n Turkish text classification model obtained by fine-tuning the Turkish bert model (dbmdz/bert-base-turkish-cased)", "# Dataset\n\nDataset consists of 11 classes were obtained from URL The model was created using the most distinctive 6 classes.\n\nDataset can be accessed at URL\n\n label_dict = {\n 'LABEL_0': 'ekonomi',\n 'LABEL_1': 'spor',\n 'LABEL_2': 'saglik',\n 'LABEL_3': 'kultur_sanat',\n 'LABEL_4': 'bilim_teknoloji',\n 'LABEL_5': 'egitim'\n }\n\n70% of the data were used for training and 30% for testing.\n\ntrain f1-weighted score = %97\n\ntest f1-weighted score = %94", "# Usage\n\n from transformers import pipeline, AutoTokenizer, AutoModelForSequenceClassification\n\n tokenizer = AutoTokenizer.from_pretrained(\"gurkan08/bert-turkish-text-classification\")\n model = AutoModelForSequenceClassification.from_pretrained(\"gurkan08/bert-turkish-text-classification\")\n\n nlp = pipeline(\"sentiment-analysis\", model=model, tokenizer=tokenizer)\n\n text = [\"Süper Lig'in 6. haftasında Sivasspor ile Çaykur Rizespor karşı karşıya geldi...\",\n \"Son 24 saatte 69 kişi Kovid-19 nedeniyle yaşamını yitirdi, 1573 kişi iyileşti\"]\n\n out = nlp(text)\n \n label_dict = {\n 'LABEL_0': 'ekonomi',\n 'LABEL_1': 'spor',\n 'LABEL_2': 'saglik',\n 'LABEL_3': 'kultur_sanat',\n 'LABEL_4': 'bilim_teknoloji',\n 'LABEL_5': 'egitim'\n }\n\n results = []\n for result in out:\n result['label'] = label_dict[result['label']]\n URL(result)\n print(results)\n\n # > [{'label': 'spor', 'score': 0.9992026090621948}, {'label': 'saglik', 'score': 0.9972177147865295}]" ]
[ "TAGS\n#transformers #pytorch #jax #bert #text-classification #tr #autotrain_compatible #endpoints_compatible #region-us \n", "# Turkish News Text Classification\n\n Turkish text classification model obtained by fine-tuning the Turkish bert model (dbmdz/bert-base-turkish-cased)", "# Dataset\n\nDataset consists of 11 classes were obtained from URL The model was created using the most distinctive 6 classes.\n\nDataset can be accessed at URL\n\n label_dict = {\n 'LABEL_0': 'ekonomi',\n 'LABEL_1': 'spor',\n 'LABEL_2': 'saglik',\n 'LABEL_3': 'kultur_sanat',\n 'LABEL_4': 'bilim_teknoloji',\n 'LABEL_5': 'egitim'\n }\n\n70% of the data were used for training and 30% for testing.\n\ntrain f1-weighted score = %97\n\ntest f1-weighted score = %94", "# Usage\n\n from transformers import pipeline, AutoTokenizer, AutoModelForSequenceClassification\n\n tokenizer = AutoTokenizer.from_pretrained(\"gurkan08/bert-turkish-text-classification\")\n model = AutoModelForSequenceClassification.from_pretrained(\"gurkan08/bert-turkish-text-classification\")\n\n nlp = pipeline(\"sentiment-analysis\", model=model, tokenizer=tokenizer)\n\n text = [\"Süper Lig'in 6. haftasında Sivasspor ile Çaykur Rizespor karşı karşıya geldi...\",\n \"Son 24 saatte 69 kişi Kovid-19 nedeniyle yaşamını yitirdi, 1573 kişi iyileşti\"]\n\n out = nlp(text)\n \n label_dict = {\n 'LABEL_0': 'ekonomi',\n 'LABEL_1': 'spor',\n 'LABEL_2': 'saglik',\n 'LABEL_3': 'kultur_sanat',\n 'LABEL_4': 'bilim_teknoloji',\n 'LABEL_5': 'egitim'\n }\n\n results = []\n for result in out:\n result['label'] = label_dict[result['label']]\n URL(result)\n print(results)\n\n # > [{'label': 'spor', 'score': 0.9992026090621948}, {'label': 'saglik', 'score': 0.9972177147865295}]" ]
text-generation
transformers
# Rick bot
{"tags": ["conversational"]}
gusintheshell/DialoGPT-small-rickbot
null
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Rick bot
[ "# Rick bot" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Rick bot" ]
text2text-generation
transformers
### Quantized BigScience's T0 3B with 8-bit weights This is a version of [BigScience's T0](https://huggingface.co/bigscience/T0_3B) with 3 billion parameters that is modified so you can generate **and fine-tune the model in colab or equivalent desktop gpu (e.g. single 1080Ti)**. Inspired by [GPT-J 8bit](https://huggingface.co/hivemind/gpt-j-6B-8bit). Here's how to run it: [![colab](https://camo.githubusercontent.com/84f0493939e0c4de4e6dbe113251b4bfb5353e57134ffd9fcab6b8714514d4d1/68747470733a2f2f636f6c61622e72657365617263682e676f6f676c652e636f6d2f6173736574732f636f6c61622d62616467652e737667)](https://colab.research.google.com/drive/1lMja-CPc0vm5_-gXNXAWU-9c0nom7vZ9) This model can be easily loaded using the `T5ForConditionalGeneration` functionality: ```python from transformers import T5ForConditionalGeneration model = T5ForConditionalGeneration.from_pretrained("gustavecortal/T0_3B-8bit") ``` Before loading, you have to Monkey-Patch T5: ```python class T5ForConditionalGeneration(transformers.models.t5.modeling_t5.T5ForConditionalGeneration): def __init__(self, config): super().__init__(config) convert_to_int8(self) transformers.models.t5.modeling_t5.T5ForConditionalGeneration = T5ForConditionalGeneration ``` ## Model Description T0* shows zero-shot task generalization on English natural language prompts, outperforming GPT-3 on many tasks, while being 16x smaller. It is a series of encoder-decoder models trained on a large set of different tasks specified in natural language prompts. We convert numerous English supervised datasets into prompts, each with multiple templates using varying formulations. These prompted datasets allow for benchmarking the ability of a model to perform completely unseen tasks specified in natural language. To obtain T0*, we fine-tune a pretrained language model on this multitask mixture covering many different NLP tasks. ## Links * [BigScience](https://bigscience.huggingface.co/) * [Hivemind](https://training-transformers-together.github.io/) * [Gustave Cortal](https://twitter.com/gustavecortal) ```bibtex @misc{sanh2021multitask, title={Multitask Prompted Training Enables Zero-Shot Task Generalization}, author={Victor Sanh and Albert Webson and Colin Raffel and Stephen H. Bach and Lintang Sutawika and Zaid Alyafeai and Antoine Chaffin and Arnaud Stiegler and Teven Le Scao and Arun Raja and Manan Dey and M Saiful Bari and Canwen Xu and Urmish Thakker and Shanya Sharma Sharma and Eliza Szczechla and Taewoon Kim and Gunjan Chhablani and Nihal Nayak and Debajyoti Datta and Jonathan Chang and Mike Tian-Jian Jiang and Han Wang and Matteo Manica and Sheng Shen and Zheng Xin Yong and Harshit Pandey and Rachel Bawden and Thomas Wang and Trishala Neeraj and Jos Rozen and Abheesht Sharma and Andrea Santilli and Thibault Fevry and Jason Alan Fries and Ryan Teehan and Stella Biderman and Leo Gao and Tali Bers and Thomas Wolf and Alexander M. Rush}, year={2021}, eprint={2110.08207}, archivePrefix={arXiv}, primaryClass={cs.LG} } ```
{"language": "fr", "license": "mit", "tags": ["en"], "datasets": ["bigscience/P3"]}
gustavecortal/T0_3B-8bit
null
[ "transformers", "pytorch", "t5", "text2text-generation", "en", "fr", "dataset:bigscience/P3", "arxiv:2110.08207", "license:mit", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
[ "2110.08207" ]
[ "fr" ]
TAGS #transformers #pytorch #t5 #text2text-generation #en #fr #dataset-bigscience/P3 #arxiv-2110.08207 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
### Quantized BigScience's T0 3B with 8-bit weights This is a version of BigScience's T0 with 3 billion parameters that is modified so you can generate and fine-tune the model in colab or equivalent desktop gpu (e.g. single 1080Ti). Inspired by GPT-J 8bit. Here's how to run it: ![colab](URL This model can be easily loaded using the 'T5ForConditionalGeneration' functionality: Before loading, you have to Monkey-Patch T5: ## Model Description T0* shows zero-shot task generalization on English natural language prompts, outperforming GPT-3 on many tasks, while being 16x smaller. It is a series of encoder-decoder models trained on a large set of different tasks specified in natural language prompts. We convert numerous English supervised datasets into prompts, each with multiple templates using varying formulations. These prompted datasets allow for benchmarking the ability of a model to perform completely unseen tasks specified in natural language. To obtain T0*, we fine-tune a pretrained language model on this multitask mixture covering many different NLP tasks. ## Links * BigScience * Hivemind * Gustave Cortal
[ "### Quantized BigScience's T0 3B with 8-bit weights\n\n\nThis is a version of BigScience's T0 with 3 billion parameters that is modified so you can generate and fine-tune the model in colab or equivalent desktop gpu (e.g. single 1080Ti). Inspired by GPT-J 8bit. \n\nHere's how to run it: ![colab](URL\n\nThis model can be easily loaded using the 'T5ForConditionalGeneration' functionality:\n\n\nBefore loading, you have to Monkey-Patch T5:", "## Model Description\n\nT0* shows zero-shot task generalization on English natural language prompts, outperforming GPT-3 on many tasks, while being 16x smaller. It is a series of encoder-decoder models trained on a large set of different tasks specified in natural language prompts. We convert numerous English supervised datasets into prompts, each with multiple templates using varying formulations. These prompted datasets allow for benchmarking the ability of a model to perform completely unseen tasks specified in natural language. To obtain T0*, we fine-tune a pretrained language model on this multitask mixture covering many different NLP tasks.", "## Links\n\n* BigScience\n* Hivemind\n* Gustave Cortal" ]
[ "TAGS\n#transformers #pytorch #t5 #text2text-generation #en #fr #dataset-bigscience/P3 #arxiv-2110.08207 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Quantized BigScience's T0 3B with 8-bit weights\n\n\nThis is a version of BigScience's T0 with 3 billion parameters that is modified so you can generate and fine-tune the model in colab or equivalent desktop gpu (e.g. single 1080Ti). Inspired by GPT-J 8bit. \n\nHere's how to run it: ![colab](URL\n\nThis model can be easily loaded using the 'T5ForConditionalGeneration' functionality:\n\n\nBefore loading, you have to Monkey-Patch T5:", "## Model Description\n\nT0* shows zero-shot task generalization on English natural language prompts, outperforming GPT-3 on many tasks, while being 16x smaller. It is a series of encoder-decoder models trained on a large set of different tasks specified in natural language prompts. We convert numerous English supervised datasets into prompts, each with multiple templates using varying formulations. These prompted datasets allow for benchmarking the ability of a model to perform completely unseen tasks specified in natural language. To obtain T0*, we fine-tune a pretrained language model on this multitask mixture covering many different NLP tasks.", "## Links\n\n* BigScience\n* Hivemind\n* Gustave Cortal" ]
text-generation
transformers
### Quantized Cedille/fr-boris with 8-bit weights This is a version of Cedille's GPT-J (fr-boris) with 6 billion parameters that is modified so you can generate **and fine-tune the model in colab or equivalent desktop gpu (e.g. single 1080Ti)**. Inspired by [GPT-J 8bit](https://huggingface.co/hivemind/gpt-j-6B-8bit). Here's how to run it: [![colab](https://camo.githubusercontent.com/84f0493939e0c4de4e6dbe113251b4bfb5353e57134ffd9fcab6b8714514d4d1/68747470733a2f2f636f6c61622e72657365617263682e676f6f676c652e636f6d2f6173736574732f636f6c61622d62616467652e737667)](https://colab.research.google.com/drive/1lMja-CPc0vm5_-gXNXAWU-9c0nom7vZ9) This model can be easily loaded using the `GPTJForCausalLM` functionality: ```python from transformers import GPTJForCausalLM model = GPTJForCausalLM.from_pretrained("gustavecortal/fr-boris-8bit") ``` ## fr-boris Boris is a 6B parameter autoregressive language model based on the GPT-J architecture and trained using the [mesh-transformer-jax](https://github.com/kingoflolz/mesh-transformer-jax) codebase. Boris was trained on around 78B tokens of French text from the [C4](https://huggingface.co/datasets/c4) dataset. ## Links * [Cedille](https://en.cedille.ai/) * [Hivemind](https://training-transformers-together.github.io/) * [Gustave Cortal](https://twitter.com/gustavecortal)
{"language": "fr", "license": "mit", "tags": ["causal-lm", "fr"], "datasets": ["c4", "The Pile"]}
gustavecortal/fr-boris-8bit
null
[ "transformers", "pytorch", "gptj", "text-generation", "causal-lm", "fr", "license:mit", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "fr" ]
TAGS #transformers #pytorch #gptj #text-generation #causal-lm #fr #license-mit #autotrain_compatible #endpoints_compatible #has_space #region-us
### Quantized Cedille/fr-boris with 8-bit weights This is a version of Cedille's GPT-J (fr-boris) with 6 billion parameters that is modified so you can generate and fine-tune the model in colab or equivalent desktop gpu (e.g. single 1080Ti). Inspired by GPT-J 8bit. Here's how to run it: ![colab](URL This model can be easily loaded using the 'GPTJForCausalLM' functionality: ## fr-boris Boris is a 6B parameter autoregressive language model based on the GPT-J architecture and trained using the mesh-transformer-jax codebase. Boris was trained on around 78B tokens of French text from the C4 dataset. ## Links * Cedille * Hivemind * Gustave Cortal
[ "### Quantized Cedille/fr-boris with 8-bit weights\n\n\nThis is a version of Cedille's GPT-J (fr-boris) with 6 billion parameters that is modified so you can generate and fine-tune the model in colab or equivalent desktop gpu (e.g. single 1080Ti). Inspired by GPT-J 8bit. \n\nHere's how to run it: ![colab](URL\n\nThis model can be easily loaded using the 'GPTJForCausalLM' functionality:", "## fr-boris\n\nBoris is a 6B parameter autoregressive language model based on the GPT-J architecture and trained using the mesh-transformer-jax codebase.\n\nBoris was trained on around 78B tokens of French text from the C4 dataset.", "## Links\n\n* Cedille\n* Hivemind\n* Gustave Cortal" ]
[ "TAGS\n#transformers #pytorch #gptj #text-generation #causal-lm #fr #license-mit #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "### Quantized Cedille/fr-boris with 8-bit weights\n\n\nThis is a version of Cedille's GPT-J (fr-boris) with 6 billion parameters that is modified so you can generate and fine-tune the model in colab or equivalent desktop gpu (e.g. single 1080Ti). Inspired by GPT-J 8bit. \n\nHere's how to run it: ![colab](URL\n\nThis model can be easily loaded using the 'GPTJForCausalLM' functionality:", "## fr-boris\n\nBoris is a 6B parameter autoregressive language model based on the GPT-J architecture and trained using the mesh-transformer-jax codebase.\n\nBoris was trained on around 78B tokens of French text from the C4 dataset.", "## Links\n\n* Cedille\n* Hivemind\n* Gustave Cortal" ]
text-generation
transformers
### Quantized EleutherAI/gpt-neo-2.7B with 8-bit weights This is a version of [EleutherAI's GPT-Neo](https://huggingface.co/EleutherAI/gpt-neo-2.7B) with 2.7 billion parameters that is modified so you can generate **and fine-tune the model in colab or equivalent desktop gpu (e.g. single 1080Ti)**. Inspired by [GPT-J 8bit](https://huggingface.co/hivemind/gpt-j-6B-8bit). Here's how to run it: [![colab](https://camo.githubusercontent.com/84f0493939e0c4de4e6dbe113251b4bfb5353e57134ffd9fcab6b8714514d4d1/68747470733a2f2f636f6c61622e72657365617263682e676f6f676c652e636f6d2f6173736574732f636f6c61622d62616467652e737667)](https://colab.research.google.com/drive/1lMja-CPc0vm5_-gXNXAWU-9c0nom7vZ9) ## Model Description GPT-Neo 2.7B is a transformer model designed using EleutherAI's replication of the GPT-3 architecture. GPT-Neo refers to the class of models, while 2.7B represents the number of parameters of this particular pre-trained model. ## Links * [EleutherAI](https://www.eleuther.ai) * [Hivemind](https://training-transformers-together.github.io/) * [Gustave Cortal](https://twitter.com/gustavecortal)
{"language": "en", "license": "mit", "tags": ["causal-lm"], "datasets": ["The_Pile"]}
gustavecortal/gpt-neo-2.7B-8bit
null
[ "transformers", "pytorch", "gpt_neo", "text-generation", "causal-lm", "en", "dataset:The_Pile", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt_neo #text-generation #causal-lm #en #dataset-The_Pile #license-mit #autotrain_compatible #endpoints_compatible #region-us
### Quantized EleutherAI/gpt-neo-2.7B with 8-bit weights This is a version of EleutherAI's GPT-Neo with 2.7 billion parameters that is modified so you can generate and fine-tune the model in colab or equivalent desktop gpu (e.g. single 1080Ti). Inspired by GPT-J 8bit. Here's how to run it: ![colab](URL ## Model Description GPT-Neo 2.7B is a transformer model designed using EleutherAI's replication of the GPT-3 architecture. GPT-Neo refers to the class of models, while 2.7B represents the number of parameters of this particular pre-trained model. ## Links * EleutherAI * Hivemind * Gustave Cortal
[ "### Quantized EleutherAI/gpt-neo-2.7B with 8-bit weights\n\n\nThis is a version of EleutherAI's GPT-Neo with 2.7 billion parameters that is modified so you can generate and fine-tune the model in colab or equivalent desktop gpu (e.g. single 1080Ti). Inspired by GPT-J 8bit. \n\nHere's how to run it: ![colab](URL", "## Model Description\n\nGPT-Neo 2.7B is a transformer model designed using EleutherAI's replication of the GPT-3 architecture. GPT-Neo refers to the class of models, while 2.7B represents the number of parameters of this particular pre-trained model.", "## Links\n\n* EleutherAI\n* Hivemind\n* Gustave Cortal" ]
[ "TAGS\n#transformers #pytorch #gpt_neo #text-generation #causal-lm #en #dataset-The_Pile #license-mit #autotrain_compatible #endpoints_compatible #region-us \n", "### Quantized EleutherAI/gpt-neo-2.7B with 8-bit weights\n\n\nThis is a version of EleutherAI's GPT-Neo with 2.7 billion parameters that is modified so you can generate and fine-tune the model in colab or equivalent desktop gpu (e.g. single 1080Ti). Inspired by GPT-J 8bit. \n\nHere's how to run it: ![colab](URL", "## Model Description\n\nGPT-Neo 2.7B is a transformer model designed using EleutherAI's replication of the GPT-3 architecture. GPT-Neo refers to the class of models, while 2.7B represents the number of parameters of this particular pre-trained model.", "## Links\n\n* EleutherAI\n* Hivemind\n* Gustave Cortal" ]
automatic-speech-recognition
transformers
# Wav2Vec2-Large-XLSR-53-ml Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on ml (Malayalam) using the [Indic TTS Malayalam Speech Corpus (via Kaggle)](https://www.kaggle.com/kavyamanohar/indic-tts-malayalam-speech-corpus), [Openslr Malayalam Speech Corpus](http://openslr.org/63/), [SMC Malayalam Speech Corpus](https://blog.smc.org.in/malayalam-speech-corpus/) and [IIIT-H Indic Speech Databases](http://speech.iiit.ac.in/index.php/research-svl/69.html). The notebooks used to train model are available [here](https://github.com/gauthamsuresh09/wav2vec2-large-xlsr-53-malayalam/). When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: ```python import torch import torchaudio from datasets import load_dataset from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor test_dataset = <load-test-split-of-combined-dataset> # Details on loading this dataset in the evaluation section processor = Wav2Vec2Processor.from_pretrained("gvs/wav2vec2-large-xlsr-malayalam") model = Wav2Vec2ForCTC.from_pretrained("gvs/wav2vec2-large-xlsr-malayalam") resampler = torchaudio.transforms.Resample(48_000, 16_000) # Preprocessing the datasets. # We need to read the audio files as arrays def speech_file_to_array_fn(batch): speech_array, sampling_rate = torchaudio.load(batch["path"]) batch["speech"] = resampler(speech_array).squeeze().numpy() return batch test_dataset = test_dataset.map(speech_file_to_array_fn) inputs = processor(test_dataset["speech"][:2], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): logits = model(inputs.input_values, attention_mask=inputs.attention_mask).logits predicted_ids = torch.argmax(logits, dim=-1) print("Prediction:", processor.batch_decode(predicted_ids)) print("Reference:", test_dataset["sentence"]) ``` ## Evaluation The model can be evaluated as follows on the test data of combined custom dataset. For more details on dataset preparation, check the notebooks mentioned at the end of this file. ```python import torch import torchaudio from datasets import load_dataset, load_metric from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor import re from datasets import load_dataset, load_metric from pathlib import Path # The custom dataset needs to be created using notebook mentioned at the end of this file data_dir = Path('<path-to-custom-dataset>') dataset_folders = { 'iiit': 'iiit_mal_abi', 'openslr': 'openslr', 'indic-tts': 'indic-tts-ml', 'msc-reviewed': 'msc-reviewed-speech-v1.0+20200825', } # Set directories for datasets openslr_male_dir = data_dir / dataset_folders['openslr'] / 'male' openslr_female_dir = data_dir / dataset_folders['openslr'] / 'female' iiit_dir = data_dir / dataset_folders['iiit'] indic_tts_male_dir = data_dir / dataset_folders['indic-tts'] / 'male' indic_tts_female_dir = data_dir / dataset_folders['indic-tts'] / 'female' msc_reviewed_dir = data_dir / dataset_folders['msc-reviewed'] # Load the datasets openslr_male = load_dataset("json", data_files=[f"{str(openslr_male_dir.absolute())}/sample_{i}.json" for i in range(2023)], split="train") openslr_female = load_dataset("json", data_files=[f"{str(openslr_female_dir.absolute())}/sample_{i}.json" for i in range(2103)], split="train") iiit = load_dataset("json", data_files=[f"{str(iiit_dir.absolute())}/sample_{i}.json" for i in range(1000)], split="train") indic_tts_male = load_dataset("json", data_files=[f"{str(indic_tts_male_dir.absolute())}/sample_{i}.json" for i in range(5649)], split="train") indic_tts_female = load_dataset("json", data_files=[f"{str(indic_tts_female_dir.absolute())}/sample_{i}.json" for i in range(2950)], split="train") msc_reviewed = load_dataset("json", data_files=[f"{str(msc_reviewed_dir.absolute())}/sample_{i}.json" for i in range(1541)], split="train") # Create test split as 20%, set random seed as well. test_size = 0.2 random_seed=1 openslr_male_splits = openslr_male.train_test_split(test_size=test_size, seed=random_seed) openslr_female_splits = openslr_female.train_test_split(test_size=test_size, seed=random_seed) iiit_splits = iiit.train_test_split(test_size=test_size, seed=random_seed) indic_tts_male_splits = indic_tts_male.train_test_split(test_size=test_size, seed=random_seed) indic_tts_female_splits = indic_tts_female.train_test_split(test_size=test_size, seed=random_seed) msc_reviewed_splits = msc_reviewed.train_test_split(test_size=test_size, seed=random_seed) # Get combined test dataset split_list = [openslr_male_splits, openslr_female_splits, indic_tts_male_splits, indic_tts_female_splits, msc_reviewed_splits, iiit_splits] test_dataset = datasets.concatenate_datasets([split['test'] for split in split_list) wer = load_metric("wer") processor = Wav2Vec2Processor.from_pretrained("gvs/wav2vec2-large-xlsr-malayalam") model = Wav2Vec2ForCTC.from_pretrained("gvs/wav2vec2-large-xlsr-malayalam") model.to("cuda") resamplers = { 48000: torchaudio.transforms.Resample(48_000, 16_000), } chars_to_ignore_regex = '[\\\\,\\\\?\\\\.\\\\!\\\\-\\\\;\\\\:\\\\"\\\\“\\\\%\\\\‘\\\\”\\\\�Utrnle\\\\_]' unicode_ignore_regex = r'[\\\\u200e]' # Preprocessing the datasets. # We need to read the audio files as arrays def speech_file_to_array_fn(batch): batch["sentence"] = re.sub(chars_to_ignore_regex, '', batch["sentence"]) batch["sentence"] = re.sub(unicode_ignore_regex, '', batch["sentence"]) speech_array, sampling_rate = torchaudio.load(batch["path"]) # Resample if its not in 16kHz if sampling_rate != 16000: batch["speech"] = resamplers[sampling_rate](speech_array).squeeze().numpy() else: batch["speech"] = speech_array.squeeze().numpy() # If more than one dimension is present, pick first one if batch["speech"].ndim > 1: batch["speech"] = batch["speech"][0] return batch test_dataset = test_dataset.map(speech_file_to_array_fn) # Preprocessing the datasets. # We need to read the audio files as arrays def evaluate(batch): inputs = processor(batch["speech"], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): logits = model(inputs.input_values.to("cuda"), attention_mask=inputs.attention_mask.to("cuda")).logits pred_ids = torch.argmax(logits, dim=-1) batch["pred_strings"] = processor.batch_decode(pred_ids) return batch result = test_dataset.map(evaluate, batched=True, batch_size=8) print("WER: {:2f}".format(100 * wer.compute(predictions=result["pred_strings"], references=result["sentence"]))) ``` **Test Result (WER)**: 28.43 % ## Training A combined dataset was created using [Indic TTS Malayalam Speech Corpus (via Kaggle)](https://www.kaggle.com/kavyamanohar/indic-tts-malayalam-speech-corpus), [Openslr Malayalam Speech Corpus](http://openslr.org/63/), [SMC Malayalam Speech Corpus](https://blog.smc.org.in/malayalam-speech-corpus/) and [IIIT-H Indic Speech Databases](http://speech.iiit.ac.in/index.php/research-svl/69.html). The datasets were downloaded and was converted to HF Dataset format using [this notebook](https://github.com/gauthamsuresh09/wav2vec2-large-xlsr-53-malayalam/blob/main/make_hf_dataset.ipynb) The notebook used for training and evaluation can be found [here](https://github.com/gauthamsuresh09/wav2vec2-large-xlsr-53-malayalam/blob/main/fine-tune-xlsr-wav2vec2-on-malayalam-asr-with-transformers_v2.ipynb)
{"language": "ml", "license": "apache-2.0", "tags": ["audio", "automatic-speech-recognition", "speech", "xlsr-fine-tuning-week"], "datasets": ["Indic TTS Malayalam Speech Corpus", "Openslr Malayalam Speech Corpus", "SMC Malayalam Speech Corpus", "IIIT-H Indic Speech Databases"], "metrics": ["wer"], "model-index": [{"name": "Malayalam XLSR Wav2Vec2 Large 53", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Speech Recognition"}, "dataset": {"name": "Test split of combined dataset using all datasets mentioned above", "type": "custom", "args": "ml"}, "metrics": [{"type": "wer", "value": 28.43, "name": "Test WER"}]}]}]}
gvs/wav2vec2-large-xlsr-malayalam
null
[ "transformers", "pytorch", "jax", "wav2vec2", "automatic-speech-recognition", "audio", "speech", "xlsr-fine-tuning-week", "ml", "license:apache-2.0", "model-index", "endpoints_compatible", "has_space", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "ml" ]
TAGS #transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #ml #license-apache-2.0 #model-index #endpoints_compatible #has_space #region-us
# Wav2Vec2-Large-XLSR-53-ml Fine-tuned facebook/wav2vec2-large-xlsr-53 on ml (Malayalam) using the Indic TTS Malayalam Speech Corpus (via Kaggle), Openslr Malayalam Speech Corpus, SMC Malayalam Speech Corpus and IIIT-H Indic Speech Databases. The notebooks used to train model are available here. When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: ## Evaluation The model can be evaluated as follows on the test data of combined custom dataset. For more details on dataset preparation, check the notebooks mentioned at the end of this file. Test Result (WER): 28.43 % ## Training A combined dataset was created using Indic TTS Malayalam Speech Corpus (via Kaggle), Openslr Malayalam Speech Corpus, SMC Malayalam Speech Corpus and IIIT-H Indic Speech Databases. The datasets were downloaded and was converted to HF Dataset format using this notebook The notebook used for training and evaluation can be found here
[ "# Wav2Vec2-Large-XLSR-53-ml\n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on ml (Malayalam) using the Indic TTS Malayalam Speech Corpus (via Kaggle), Openslr Malayalam Speech Corpus, SMC Malayalam Speech Corpus and IIIT-H Indic Speech Databases. The notebooks used to train model are available here. When using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\n\nThe model can be used directly (without a language model) as follows:", "## Evaluation\n\nThe model can be evaluated as follows on the test data of combined custom dataset. For more details on dataset preparation, check the notebooks mentioned at the end of this file.\n\n\n\n\nTest Result (WER): 28.43 %", "## Training\n\nA combined dataset was created using Indic TTS Malayalam Speech Corpus (via Kaggle), Openslr Malayalam Speech Corpus, SMC Malayalam Speech Corpus and IIIT-H Indic Speech Databases. The datasets were downloaded and was converted to HF Dataset format using this notebook\n\nThe notebook used for training and evaluation can be found here" ]
[ "TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #ml #license-apache-2.0 #model-index #endpoints_compatible #has_space #region-us \n", "# Wav2Vec2-Large-XLSR-53-ml\n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on ml (Malayalam) using the Indic TTS Malayalam Speech Corpus (via Kaggle), Openslr Malayalam Speech Corpus, SMC Malayalam Speech Corpus and IIIT-H Indic Speech Databases. The notebooks used to train model are available here. When using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\n\nThe model can be used directly (without a language model) as follows:", "## Evaluation\n\nThe model can be evaluated as follows on the test data of combined custom dataset. For more details on dataset preparation, check the notebooks mentioned at the end of this file.\n\n\n\n\nTest Result (WER): 28.43 %", "## Training\n\nA combined dataset was created using Indic TTS Malayalam Speech Corpus (via Kaggle), Openslr Malayalam Speech Corpus, SMC Malayalam Speech Corpus and IIIT-H Indic Speech Databases. The datasets were downloaded and was converted to HF Dataset format using this notebook\n\nThe notebook used for training and evaluation can be found here" ]
null
transformers
"5050_base_test"
{}
gwkim22/5050_b_disc
null
[ "transformers", "pytorch", "electra", "pretraining", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #electra #pretraining #endpoints_compatible #region-us
"5050_base_test"
[]
[ "TAGS\n#transformers #pytorch #electra #pretraining #endpoints_compatible #region-us \n" ]
null
transformers
"test_5050"
{}
gwkim22/5050_s_disc
null
[ "transformers", "pytorch", "electra", "pretraining", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #electra #pretraining #endpoints_compatible #region-us
"test_5050"
[]
[ "TAGS\n#transformers #pytorch #electra #pretraining #endpoints_compatible #region-us \n" ]
null
transformers
"domain_base_test"
{}
gwkim22/domain_b_disc
null
[ "transformers", "pytorch", "electra", "pretraining", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #electra #pretraining #endpoints_compatible #region-us
"domain_base_test"
[]
[ "TAGS\n#transformers #pytorch #electra #pretraining #endpoints_compatible #region-us \n" ]
null
transformers
"domain_base2_disc_0719"
{}
gwkim22/domain_base2_disc
null
[ "transformers", "pytorch", "electra", "pretraining", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #electra #pretraining #endpoints_compatible #region-us
"domain_base2_disc_0719"
[]
[ "TAGS\n#transformers #pytorch #electra #pretraining #endpoints_compatible #region-us \n" ]
null
transformers
"test_domain_only"
{}
gwkim22/domain_s_disc
null
[ "transformers", "pytorch", "electra", "pretraining", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #electra #pretraining #endpoints_compatible #region-us
"test_domain_only"
[]
[ "TAGS\n#transformers #pytorch #electra #pretraining #endpoints_compatible #region-us \n" ]
null
transformers
"general_base_test"
{}
gwkim22/general_b_disc
null
[ "transformers", "pytorch", "electra", "pretraining", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #electra #pretraining #endpoints_compatible #region-us
"general_base_test"
[]
[ "TAGS\n#transformers #pytorch #electra #pretraining #endpoints_compatible #region-us \n" ]
null
transformers
"general_test"
{}
gwkim22/general_s_disc
null
[ "transformers", "pytorch", "electra", "pretraining", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #electra #pretraining #endpoints_compatible #region-us
"general_test"
[]
[ "TAGS\n#transformers #pytorch #electra #pretraining #endpoints_compatible #region-us \n" ]
text2text-generation
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-small-finetuned-xsum This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:-------:|:------:|:-------:|:---------:|:-------:| | No log | 1.0 | 13 | 3.6429 | 15.3135 | 1.0725 | 12.0447 | 12.445 | 18.97 | ### Framework versions - Transformers 4.9.0 - Pytorch 1.9.0+cu102 - Datasets 1.10.2 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": [], "model_index": [{"name": "t5-small-finetuned-xsum", "results": [{"task": {"name": "Sequence-to-sequence Language Modeling", "type": "text2text-generation"}}]}]}
gwynethfae/t5-small-finetuned-xsum
null
[ "transformers", "pytorch", "tensorboard", "t5", "text2text-generation", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
t5-small-finetuned-xsum ======================= This model is a fine-tuned version of t5-small on the None dataset. Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 8 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 1 ### Training results ### Framework versions * Transformers 4.9.0 * Pytorch 1.9.0+cu102 * Datasets 1.10.2 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.9.0\n* Pytorch 1.9.0+cu102\n* Datasets 1.10.2\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.9.0\n* Pytorch 1.9.0+cu102\n* Datasets 1.10.2\n* Tokenizers 0.10.3" ]
null
null
# MultiLingual CLIP Multilingual CLIP is a pre-trained model which can be used for multilingual semantic search and zero-shot image classification in 100 languages. # Model Architecture Multilingual CLIP was built using [OpenAI CLIP](https://github.com/openai/CLIP) model. I have used the same Vision encoder (ResNet 50x4), but instead I replaced their text encoder (Transformer) with a Mulilingual Text Encoder ([XLM-Roberta](https://huggingface.co/xlm-roberta-large)) and a configurable number of projection heads, as seen below: ![Model Architecture](https://challengepost-s3-challengepost.netdna-ssl.com/photos/production/software_photos/001/858/046/datas/gallery.jpg) The model was trained in a distributed fashion on 16 Habana Gaudi Accelerators and with mixed Precision in two phases (using COCO Dataset for phase 1 and Google Conceptual Captions for phase 2). The training pipeline was built using PyTorch, PyTorch Lightning, and Distributed Data Parallel. # Datasets Three datasets have been used for building the model. COCO captions was used for training phase 1 and Google Conceptual Captions was used for training phase 2. Unsplash dataset was used for testing and inference. ## COCO Captions COCO (Common Objects in Context) is a large-scale object detection, segmentation, and captioning dataset. The COCO captions dataset has around ~85000 images and captions pairs. Run the following to download the dataset: ```bash ./download_coco.sh ``` This dataset was used for the first pre-training phase. ## Google Conceptual Captions Conceptual Captions is a dataset consisting of ~3.3 million images annotated with captions. In contrast with the curated style of other image caption annotations, Conceptual Caption images and their raw descriptions are harvested from the web, and therefore represent a wider variety of styles. Download the datasets urls/captions from [here](https://storage.cloud.google.com/gcc-data/Train/GCC-training.tsv?_ga=2.191230122.-1896153081.1529438250) as save it to `datasets/googlecc/googlecc.tsv`. The full dataset has over 3 million images, but you can select a subset by loading the `googlecc.tsv` file and saving only the number of rows you want (I have used 1 million images for training). Then run the following commands to download each image on the `googlecc.tsv` file: ```bash npm install node download_build_googlecc.js ``` This dataset was used for the second pre-training phase. ## Unplash This dataset was used as the test set during inference. Run `python3.8 download_unsplash.py` to download the dataset. # Training ![Training phase 1](https://challengepost-s3-challengepost.netdna-ssl.com/photos/production/software_photos/001/858/047/datas/gallery.jpg) ![Training phase 2](https://challengepost-s3-challengepost.netdna-ssl.com/photos/production/software_photos/001/858/048/datas/gallery.jpg) ## Setup Create two Habana instances ([AWS EC2 DL1](https://aws.amazon.com/ec2/instance-types/dl1/)) using [Habana® Deep Learning Base AMI (Ubuntu 20.04)](https://aws.amazon.com/marketplace/pp/prodview-fw46rwuxrtfse) Create the PyTorch docker container running: ```bash docker run --name pytorch -td --runtime=habana -e HABANA_VISIBLE_DEVICES=all -e OMPI_MCA_btl_vader_single_copy_mechanism=none --cap-add=sys_nice --net=host --ipc=host vault.habana.ai/gaudi-docker/1.2.0/ubuntu20.04/habanalabs/pytorch-installer-1.10.0:1.2.0-585 ``` Enter the docker image by running: ``` docker exec -it pytorch /bin/bash ``` #### Setup password-less ssh between all connected servers 1. Configure password-less ssh between all nodes: Do the following in all the nodes' docker sessions: ```bash mkdir ~/.ssh cd ~/.ssh ssh-keygen -t rsa -b 4096 ``` Copy id_rsa.pub contents from every node's docker to every other node's docker's ~/.ssh/authorized_keys (all public keys need to be in all hosts' authorized_keys): ```bash cat id_rsa.pub > authorized_keys vi authorized_keys ``` Copy the contents from inside to other systems. Paste all hosts' public keys in all hosts' “authorized_keys” file. 2. On each system: Add all hosts (including itself) to known_hosts. The IP addresses used below are just for illustration: ```bash ssh-keyscan -p 3022 -H $IP1 >> ~/.ssh/known_hosts ssh-keyscan -p 3022 -H $IP2 >> ~/.ssh/known_hosts ``` 3. Change Docker SSH port to 3022 ```bash sed -i 's/#Port 22/Port 3022/g' /etc/ssh/sshd_config sed -i 's/#PermitRootLogin prohibit-password/PermitRootLogin yes/' /etc/ssh/sshd_config service ssh restart ``` [Allow all TCP](https://docs.aws.amazon.com/vpc/latest/userguide/VPC_SecurityGroups.html) traffic between the nodes on AWS Clone the git repo: ```bash git clone https://github.com/gzomer/clip-multilingual ``` Create environment: ```bash python3.8 -m venv .env ``` Install requirements: ```bash python3.8 -r requirements.txt ``` Activate environment ```bash source .env/bin/activate ``` ## Training params Learning rate: 1e-3 Batch size: 64 Phase 1 - Epochs: 100 Phase 2 - Epochs: 15 ## Train script arguments ``` --dataset-num-workers Number of workers (default: 8) --dataset-type Dataset type (coco or googlecc) (default: coco) --dataset-dir Dataset dir (default: ./datasets/coco/) --dataset-subset-size Load only a subset of the dataset (useful for debugging) --dataset-train-split Dataset train split (default: 0.8) --train-device Type of device to use (default: hpu) --distributed-num-nodes Number of nodes (machines) (default: 2) --distributed-parallel-devices Number of parallel devices per node (default: 8) --distributed-master-address Master node IP address --distributed-master-port Master node port (default: 12345) --distributed-bucket-cap-mb DDP bucket cap MB (default: 200) --checkpoint-dir Model checkpoint dir (default: ./models) --checkpoint-save-every-n Save every n epochs (default: 1) --checkpoint-load-vision-path Load vision encoder checkpoint --checkpoint-load-text-path Load text encoder checkpoint --model-visual-name Which visual model to use (default: RN50x4) --model-textual-name Which textual model to use (default: xlm-roberta-base) --hyperparam-num-layers Number of layers (default: 3) --hyperparam-lr Model learning rate (default: 0.001) --hyperparam-epochs Max epochs (default: 100) --hyperparam-precision Precision (default: 16) --hyperparam-batch-size Batch size (default: 64) --wandb-project W&B project name (default: clip) --wandb-enabled W&B is enabled? (default: True) ``` ## Habana Gaudi - 8 accelerators ### Phase 1 training ```bash python3.8 train.py --train-device hpu --distributed-parallel-devices 8 --distributed-num-nodes 1 ``` ### Phase 2 training ```bash python3.8 train.py --train-device hpu --distributed-parallel-devices 8 --distributed-num-nodes 1 --hyperparam-epochs 15 --checkpoint-load-text-path /home/models/text-last.ckpt --checkpoint-load-vision-path /home/models/vision-last.ckpt --checkpoint-dir ./models_phase2 ``` ## Habana Gaudi - 16 accelerators (multi-server training) Change the master IP address based on your instances (use local IP, not public IP). ### Phase 1 training ```bash NODE_RANK=0 python3.8 train.py --distributed-master-address 172.31.86.231 --train-device hpu --distributed-parallel-devices 8 --distributed-num-nodes 2 ``` ```bash NODE_RANK=1 python3.8 train.py --distributed-master-address 172.31.86.231 --train-device hpu --distributed-parallel-devices 8 --distributed-num-nodes 2 ``` ### Phase 2 training ```bash NODE_RANK=0 python3.8 train.py --distributed-master-address 172.31.86.231 --train-device hpu --distributed-parallel-devices 8 --distributed-num-nodes 2 --hyperparam-epochs 10 --checkpoint-load-text-path /home/models/text-last.ckpt --checkpoint-load-vision-path /home/models/vision-last.ckpt --checkpoint-dir ./models_phase2 ``` ```bash NODE_RANK=1 python3.8 train.py --distributed-master-address 172.31.86.231 --train-device hpu --distributed-parallel-devices 8 --distributed-num-nodes 2 --hyperparam-epochs 15 --checkpoint-load-text-path /home/models/text-last.ckpt --checkpoint-load-vision-path /home/models/vision-last.ckpt --checkpoint-dir ./models_phase2 ``` ## Other devices If you don't have access to a Habana Gaudi accelerator yet, you can also train on CPU/GPU, although it will be way slower. To train on CPU, just pass `--train-device=cpu` and on GPU `--train-device=cuda` to the `train.py` script. # Inference ## Loading pre-trained model from Hugging Face HUB ```python from models import create_and_load_from_hub model = create_and_load_from_hub() ``` ## Loading model from local checkpoint ```python from models import MultiLingualCLIP, load_model text_checkpoint_path = '/path/to/text model checkpoint' vision_checkpoint_path = '/path/to/vision model checkpoint' model = MultiLingualCLIP(num_layers=3) load_model(model, vision_checkpoint_path, text_checkpoint_path) ``` ## Generate embeddings Run the following (after downloading Unplash dataset): `python3.8 ./generate_embeddings.py` ## Searching images ```python import numpy as np from search import MultiLingualSearch images_embeddings = np.load('/path/to/images_embeddings') images_data = [...] # List of image info for each row of the embeddings. For instance, it could be a list of urls, filepaths, ids. They will be returned when calling the search function semantic_search = MultiLingualSearch(model, images_embeddings, images_data) results = semantic_search.search('विद्यालय में') # Means at school print(results) ``` ```json [{"image": "https://images.unsplash.com/photo-1557804506-669a67965ba0?crop=entropy&cs=tinysrgb&fit=max&fm=jpg&ixid=MnwyNDg3OTV8MHwxfHNlYXJjaHwxM3x8bWVldGluZ3N8ZW58MHx8fHwxNjQ1NjA2MjQz&ixlib=rb-1.2.1&q=80&w=400", "prob": 0.2461608648300171}, {"image": "https://images.unsplash.com/photo-1558403194-611308249627?crop=entropy&cs=tinysrgb&fit=max&fm=jpg&ixid=MnwyNDg3OTV8MHwxfHNlYXJjaHwyMXx8cGVvcGxlJTIwd29ya2luZ3xlbnwwfHx8fDE2NDU2MDMyMjE&ixlib=rb-1.2.1&q=80&w=400", "prob": 0.16881239414215088}, {"image": "https://images.unsplash.com/photo-1531497865144-0464ef8fb9a9?crop=entropy&cs=tinysrgb&fit=max&fm=jpg&ixid=MnwyNDg3OTV8MHwxfHNlYXJjaHw4Nnx8cGVvcGxlJTIwd29ya2luZ3xlbnwwfHx8fDE2NDU2MDY5ODc&ixlib=rb-1.2.1&q=80&w=400", "prob": 0.14744874835014343}, {"image": "https://images.unsplash.com/photo-1561089489-f13d5e730d72?crop=entropy&cs=tinysrgb&fit=max&fm=jpg&ixid=MnwyNDg3OTV8MHwxfHNlYXJjaHw5MHx8ZWR1Y2F0aW9ufGVufDB8fHx8MTY0NTYwNjk1Nw&ixlib=rb-1.2.1&q=80&w=400", "prob": 0.095176100730896}, {"image": "https://images.unsplash.com/photo-1580582932707-520aed937b7b?crop=entropy&cs=tinysrgb&fit=max&fm=jpg&ixid=MnwyNDg3OTV8MHwxfHNlYXJjaHwxMnx8ZWR1Y2F0aW9ufGVufDB8fHx8MTY0NTYwMzIwMA&ixlib=rb-1.2.1&q=80&w=400", "prob": 0.05218643322587013}] ```
{"language": "multilingual", "license": "mit", "tags": ["clip", "vision", "text"]}
gzomer/clip-multilingual
null
[ "clip", "vision", "text", "multilingual", "license:mit", "has_space", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "multilingual" ]
TAGS #clip #vision #text #multilingual #license-mit #has_space #region-us
# MultiLingual CLIP Multilingual CLIP is a pre-trained model which can be used for multilingual semantic search and zero-shot image classification in 100 languages. # Model Architecture Multilingual CLIP was built using OpenAI CLIP model. I have used the same Vision encoder (ResNet 50x4), but instead I replaced their text encoder (Transformer) with a Mulilingual Text Encoder (XLM-Roberta) and a configurable number of projection heads, as seen below: !Model Architecture The model was trained in a distributed fashion on 16 Habana Gaudi Accelerators and with mixed Precision in two phases (using COCO Dataset for phase 1 and Google Conceptual Captions for phase 2). The training pipeline was built using PyTorch, PyTorch Lightning, and Distributed Data Parallel. # Datasets Three datasets have been used for building the model. COCO captions was used for training phase 1 and Google Conceptual Captions was used for training phase 2. Unsplash dataset was used for testing and inference. ## COCO Captions COCO (Common Objects in Context) is a large-scale object detection, segmentation, and captioning dataset. The COCO captions dataset has around ~85000 images and captions pairs. Run the following to download the dataset: This dataset was used for the first pre-training phase. ## Google Conceptual Captions Conceptual Captions is a dataset consisting of ~3.3 million images annotated with captions. In contrast with the curated style of other image caption annotations, Conceptual Caption images and their raw descriptions are harvested from the web, and therefore represent a wider variety of styles. Download the datasets urls/captions from here as save it to 'datasets/googlecc/URL'. The full dataset has over 3 million images, but you can select a subset by loading the 'URL' file and saving only the number of rows you want (I have used 1 million images for training). Then run the following commands to download each image on the 'URL' file: This dataset was used for the second pre-training phase. ## Unplash This dataset was used as the test set during inference. Run 'python3.8 download_unsplash.py' to download the dataset. # Training !Training phase 1 !Training phase 2 ## Setup Create two Habana instances (AWS EC2 DL1) using Habana® Deep Learning Base AMI (Ubuntu 20.04) Create the PyTorch docker container running: Enter the docker image by running: #### Setup password-less ssh between all connected servers 1. Configure password-less ssh between all nodes: Do the following in all the nodes' docker sessions: Copy id_rsa.pub contents from every node's docker to every other node's docker's ~/.ssh/authorized_keys (all public keys need to be in all hosts' authorized_keys): Copy the contents from inside to other systems. Paste all hosts' public keys in all hosts' “authorized_keys” file. 2. On each system: Add all hosts (including itself) to known_hosts. The IP addresses used below are just for illustration: 3. Change Docker SSH port to 3022 Allow all TCP traffic between the nodes on AWS Clone the git repo: Create environment: Install requirements: Activate environment ## Training params Learning rate: 1e-3 Batch size: 64 Phase 1 - Epochs: 100 Phase 2 - Epochs: 15 ## Train script arguments ## Habana Gaudi - 8 accelerators ### Phase 1 training ### Phase 2 training ## Habana Gaudi - 16 accelerators (multi-server training) Change the master IP address based on your instances (use local IP, not public IP). ### Phase 1 training ### Phase 2 training ## Other devices If you don't have access to a Habana Gaudi accelerator yet, you can also train on CPU/GPU, although it will be way slower. To train on CPU, just pass '--train-device=cpu' and on GPU '--train-device=cuda' to the 'URL' script. # Inference ## Loading pre-trained model from Hugging Face HUB ## Loading model from local checkpoint ## Generate embeddings Run the following (after downloading Unplash dataset): 'python3.8 ./generate_embeddings.py' ## Searching images
[ "# MultiLingual CLIP\n\nMultilingual CLIP is a pre-trained model which can be used for multilingual semantic search and zero-shot image classification in 100 languages.", "# Model Architecture\nMultilingual CLIP was built using OpenAI CLIP model. I have used the same Vision encoder (ResNet 50x4), but instead I replaced their text encoder (Transformer) with a Mulilingual Text Encoder (XLM-Roberta) and a configurable number of projection heads, as seen below:\n\n!Model Architecture\n\nThe model was trained in a distributed fashion on 16 Habana Gaudi Accelerators and with mixed Precision in two phases (using COCO Dataset for phase 1 and Google Conceptual Captions for phase 2). The training pipeline was built using PyTorch, PyTorch Lightning, and Distributed Data Parallel.", "# Datasets\n\nThree datasets have been used for building the model. COCO captions was used for training phase 1 and Google Conceptual Captions was used for training phase 2. Unsplash dataset was used for testing and inference.", "## COCO Captions\n\nCOCO (Common Objects in Context) is a large-scale object detection, segmentation, and captioning dataset. The COCO captions dataset has around ~85000 images and captions pairs.\n\nRun the following to download the dataset:\n\n\n\nThis dataset was used for the first pre-training phase.", "## Google Conceptual Captions\n\nConceptual Captions is a dataset consisting of ~3.3 million images annotated with captions. In contrast with the curated style of other image caption annotations, Conceptual Caption images and their raw descriptions are harvested from the web, and therefore represent a wider variety of styles.\n\nDownload the datasets urls/captions from here as save it to 'datasets/googlecc/URL'. The full dataset has over 3 million images, but you can select a subset by loading the 'URL' file and saving only the number of rows you want (I have used 1 million images for training).\n\nThen run the following commands to download each image on the 'URL' file:\n\n\n\nThis dataset was used for the second pre-training phase.", "## Unplash\n\nThis dataset was used as the test set during inference.\n\nRun 'python3.8 download_unsplash.py' to download the dataset.", "# Training\n\n!Training phase 1\n\n!Training phase 2", "## Setup\n\nCreate two Habana instances (AWS EC2 DL1) using Habana® Deep Learning Base AMI (Ubuntu 20.04)\n\n\nCreate the PyTorch docker container running:\n\n\n\nEnter the docker image by running:", "#### Setup password-less ssh between all connected servers\n\n1. Configure password-less ssh between all nodes:\n\n Do the following in all the nodes' docker sessions:\n \n Copy id_rsa.pub contents from every node's docker to every other node's docker's ~/.ssh/authorized_keys (all public keys need to be in all hosts' authorized_keys):\n \n Copy the contents from inside to other systems.\n Paste all hosts' public keys in all hosts' “authorized_keys” file.\n\n2. On each system:\n Add all hosts (including itself) to known_hosts. The IP addresses used below are just for illustration:\n \n\n3. Change Docker SSH port to 3022\n \n\nAllow all TCP traffic between the nodes on AWS\n\nClone the git repo:\n\n\n\nCreate environment:\n\n\n\nInstall requirements:\n\n\n\nActivate environment", "## Training params\n\nLearning rate: 1e-3\n\nBatch size: 64\n\nPhase 1 - Epochs: 100\n\nPhase 2 - Epochs: 15", "## Train script arguments", "## Habana Gaudi - 8 accelerators", "### Phase 1 training", "### Phase 2 training", "## Habana Gaudi - 16 accelerators (multi-server training)\n\nChange the master IP address based on your instances (use local IP, not public IP).", "### Phase 1 training", "### Phase 2 training", "## Other devices\nIf you don't have access to a Habana Gaudi accelerator yet, you can also train on CPU/GPU, although it will be way slower.\n\nTo train on CPU, just pass '--train-device=cpu' and on GPU '--train-device=cuda' to the 'URL' script.", "# Inference", "## Loading pre-trained model from Hugging Face HUB", "## Loading model from local checkpoint", "## Generate embeddings\n\nRun the following (after downloading Unplash dataset):\n\n'python3.8 ./generate_embeddings.py'", "## Searching images" ]
[ "TAGS\n#clip #vision #text #multilingual #license-mit #has_space #region-us \n", "# MultiLingual CLIP\n\nMultilingual CLIP is a pre-trained model which can be used for multilingual semantic search and zero-shot image classification in 100 languages.", "# Model Architecture\nMultilingual CLIP was built using OpenAI CLIP model. I have used the same Vision encoder (ResNet 50x4), but instead I replaced their text encoder (Transformer) with a Mulilingual Text Encoder (XLM-Roberta) and a configurable number of projection heads, as seen below:\n\n!Model Architecture\n\nThe model was trained in a distributed fashion on 16 Habana Gaudi Accelerators and with mixed Precision in two phases (using COCO Dataset for phase 1 and Google Conceptual Captions for phase 2). The training pipeline was built using PyTorch, PyTorch Lightning, and Distributed Data Parallel.", "# Datasets\n\nThree datasets have been used for building the model. COCO captions was used for training phase 1 and Google Conceptual Captions was used for training phase 2. Unsplash dataset was used for testing and inference.", "## COCO Captions\n\nCOCO (Common Objects in Context) is a large-scale object detection, segmentation, and captioning dataset. The COCO captions dataset has around ~85000 images and captions pairs.\n\nRun the following to download the dataset:\n\n\n\nThis dataset was used for the first pre-training phase.", "## Google Conceptual Captions\n\nConceptual Captions is a dataset consisting of ~3.3 million images annotated with captions. In contrast with the curated style of other image caption annotations, Conceptual Caption images and their raw descriptions are harvested from the web, and therefore represent a wider variety of styles.\n\nDownload the datasets urls/captions from here as save it to 'datasets/googlecc/URL'. The full dataset has over 3 million images, but you can select a subset by loading the 'URL' file and saving only the number of rows you want (I have used 1 million images for training).\n\nThen run the following commands to download each image on the 'URL' file:\n\n\n\nThis dataset was used for the second pre-training phase.", "## Unplash\n\nThis dataset was used as the test set during inference.\n\nRun 'python3.8 download_unsplash.py' to download the dataset.", "# Training\n\n!Training phase 1\n\n!Training phase 2", "## Setup\n\nCreate two Habana instances (AWS EC2 DL1) using Habana® Deep Learning Base AMI (Ubuntu 20.04)\n\n\nCreate the PyTorch docker container running:\n\n\n\nEnter the docker image by running:", "#### Setup password-less ssh between all connected servers\n\n1. Configure password-less ssh between all nodes:\n\n Do the following in all the nodes' docker sessions:\n \n Copy id_rsa.pub contents from every node's docker to every other node's docker's ~/.ssh/authorized_keys (all public keys need to be in all hosts' authorized_keys):\n \n Copy the contents from inside to other systems.\n Paste all hosts' public keys in all hosts' “authorized_keys” file.\n\n2. On each system:\n Add all hosts (including itself) to known_hosts. The IP addresses used below are just for illustration:\n \n\n3. Change Docker SSH port to 3022\n \n\nAllow all TCP traffic between the nodes on AWS\n\nClone the git repo:\n\n\n\nCreate environment:\n\n\n\nInstall requirements:\n\n\n\nActivate environment", "## Training params\n\nLearning rate: 1e-3\n\nBatch size: 64\n\nPhase 1 - Epochs: 100\n\nPhase 2 - Epochs: 15", "## Train script arguments", "## Habana Gaudi - 8 accelerators", "### Phase 1 training", "### Phase 2 training", "## Habana Gaudi - 16 accelerators (multi-server training)\n\nChange the master IP address based on your instances (use local IP, not public IP).", "### Phase 1 training", "### Phase 2 training", "## Other devices\nIf you don't have access to a Habana Gaudi accelerator yet, you can also train on CPU/GPU, although it will be way slower.\n\nTo train on CPU, just pass '--train-device=cpu' and on GPU '--train-device=cuda' to the 'URL' script.", "# Inference", "## Loading pre-trained model from Hugging Face HUB", "## Loading model from local checkpoint", "## Generate embeddings\n\nRun the following (after downloading Unplash dataset):\n\n'python3.8 ./generate_embeddings.py'", "## Searching images" ]
text-generation
transformers
hello
{}
ha-mulan/moby-dick
null
[ "transformers", "pytorch", "jax", "gpt2", "text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #jax #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
hello
[]
[ "TAGS\n#transformers #pytorch #jax #gpt2 #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
automatic-speech-recognition
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # egy-slang-model This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on the None dataset. It achieves the following results on the evaluation set: - Loss: 2.9273 - Wer: 1.0000 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.001 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 20 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:------:| | No log | 1.64 | 200 | 2.9735 | 1.0 | | 3.8098 | 3.28 | 400 | 2.9765 | 1.0 | | 3.8098 | 4.91 | 600 | 2.9662 | 1.0 | | 2.9531 | 6.56 | 800 | 2.9708 | 1.0 | | 2.9531 | 8.2 | 1000 | 2.9673 | 1.0 | | 2.9259 | 9.83 | 1200 | 2.9989 | 1.0 | | 2.9259 | 11.47 | 1400 | 2.9889 | 1.0 | | 2.9023 | 13.11 | 1600 | 2.9739 | 1.0 | | 2.9023 | 14.75 | 1800 | 3.0040 | 1.0000 | | 2.8832 | 16.39 | 2000 | 3.0170 | 1.0 | | 2.8832 | 18.03 | 2200 | 2.9963 | 0.9999 | | 2.8691 | 19.67 | 2400 | 2.9273 | 1.0000 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.10.1 - Datasets 1.13.3 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "egy-slang-model", "results": []}]}
habiba/egy-slang-model
null
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "license:apache-2.0", "endpoints_compatible", "has_space", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #has_space #region-us
egy-slang-model =============== This model is a fine-tuned version of facebook/wav2vec2-large-xlsr-53 on the None dataset. It achieves the following results on the evaluation set: * Loss: 2.9273 * Wer: 1.0000 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 0.001 * train\_batch\_size: 16 * eval\_batch\_size: 8 * seed: 42 * gradient\_accumulation\_steps: 2 * total\_train\_batch\_size: 32 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 500 * num\_epochs: 20 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.11.3 * Pytorch 1.10.1 * Datasets 1.13.3 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 20\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.1\n* Datasets 1.13.3\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #has_space #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 20\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.1\n* Datasets 1.13.3\n* Tokenizers 0.10.3" ]
fill-mask
transformers
This is a test!
{}
hackertec/dummy2
null
[ "transformers", "pytorch", "camembert", "fill-mask", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #camembert #fill-mask #autotrain_compatible #endpoints_compatible #region-us
This is a test!
[]
[ "TAGS\n#transformers #pytorch #camembert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n" ]
text-classification
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta-base-bne-finetuned-amazon_reviews_multi-taller This model is a fine-tuned version of [BSC-TeMU/roberta-base-bne](https://huggingface.co/BSC-TeMU/roberta-base-bne) on the amazon_reviews_multi dataset. It achieves the following results on the evaluation set: - Loss: 0.2463 - Accuracy: 0.9113 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.2474 | 1.0 | 125 | 0.2463 | 0.9113 | ### Framework versions - Transformers 4.9.2 - Pytorch 1.9.0+cu102 - Datasets 1.11.0 - Tokenizers 0.10.3
{"license": "cc-by-4.0", "tags": ["generated_from_trainer"], "datasets": ["amazon_reviews_multi"], "metrics": ["accuracy"], "model_index": [{"name": "roberta-base-bne-finetuned-amazon_reviews_multi-taller", "results": [{"task": {"name": "Text Classification", "type": "text-classification"}, "dataset": {"name": "amazon_reviews_multi", "type": "amazon_reviews_multi", "args": "es"}, "metric": {"name": "Accuracy", "type": "accuracy", "value": 0.91125}}]}]}
hackertec/roberta-base-bne-finetuned-amazon_reviews_multi-taller
null
[ "transformers", "pytorch", "tensorboard", "roberta", "text-classification", "generated_from_trainer", "dataset:amazon_reviews_multi", "license:cc-by-4.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #roberta #text-classification #generated_from_trainer #dataset-amazon_reviews_multi #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #region-us
roberta-base-bne-finetuned-amazon\_reviews\_multi-taller ======================================================== This model is a fine-tuned version of BSC-TeMU/roberta-base-bne on the amazon\_reviews\_multi dataset. It achieves the following results on the evaluation set: * Loss: 0.2463 * Accuracy: 0.9113 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 1 ### Training results ### Framework versions * Transformers 4.9.2 * Pytorch 1.9.0+cu102 * Datasets 1.11.0 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.9.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #roberta #text-classification #generated_from_trainer #dataset-amazon_reviews_multi #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.9.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3" ]
text-classification
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta-base-bne-finetuned-amazon_reviews_multi This model is a fine-tuned version of [BSC-TeMU/roberta-base-bne](https://huggingface.co/BSC-TeMU/roberta-base-bne) on the amazon_reviews_multi dataset. It achieves the following results on the evaluation set: - Loss: 0.2557 - Accuracy: 0.9085 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.2296 | 1.0 | 125 | 0.2557 | 0.9085 | ### Framework versions - Transformers 4.9.2 - Pytorch 1.9.0+cu102 - Datasets 1.11.0 - Tokenizers 0.10.3
{"license": "cc-by-4.0", "tags": ["generated_from_trainer"], "datasets": ["amazon_reviews_multi"], "metrics": ["accuracy"], "model_index": [{"name": "roberta-base-bne-finetuned-amazon_reviews_multi", "results": [{"task": {"name": "Text Classification", "type": "text-classification"}, "dataset": {"name": "amazon_reviews_multi", "type": "amazon_reviews_multi", "args": "es"}, "metric": {"name": "Accuracy", "type": "accuracy", "value": 0.9085}}]}]}
hackertec/roberta-base-bne-finetuned-amazon_reviews_multi
null
[ "transformers", "pytorch", "tensorboard", "roberta", "text-classification", "generated_from_trainer", "dataset:amazon_reviews_multi", "license:cc-by-4.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #roberta #text-classification #generated_from_trainer #dataset-amazon_reviews_multi #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #region-us
roberta-base-bne-finetuned-amazon\_reviews\_multi ================================================= This model is a fine-tuned version of BSC-TeMU/roberta-base-bne on the amazon\_reviews\_multi dataset. It achieves the following results on the evaluation set: * Loss: 0.2557 * Accuracy: 0.9085 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 1 ### Training results ### Framework versions * Transformers 4.9.2 * Pytorch 1.9.0+cu102 * Datasets 1.11.0 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.9.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #roberta #text-classification #generated_from_trainer #dataset-amazon_reviews_multi #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.9.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3" ]
text-classification
null
# Test
{"license": "afl-3.0", "tags": ["es", "bert"], "pipeline_tag": "text-classification", "widget": [{"text": "Mi nombre es Omar", "exdample_title": "Example 1"}, {"text": "Otra prueba", "example_title": "Test"}]}
hackertec9/test
null
[ "es", "bert", "text-classification", "license:afl-3.0", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #es #bert #text-classification #license-afl-3.0 #region-us
# Test
[ "# Test" ]
[ "TAGS\n#es #bert #text-classification #license-afl-3.0 #region-us \n", "# Test" ]
automatic-speech-recognition
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-base-timit-demo-colab This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 32 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 1000 - num_epochs: 1 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.11.3 - Pytorch 1.10.0+cu111 - Datasets 1.13.3 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "wav2vec2-base-timit-demo-colab", "results": []}]}
hady/wav2vec2-base-timit-demo-colab
null
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us
# wav2vec2-base-timit-demo-colab This model is a fine-tuned version of facebook/wav2vec2-base on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 32 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 1000 - num_epochs: 1 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.11.3 - Pytorch 1.10.0+cu111 - Datasets 1.13.3 - Tokenizers 0.10.3
[ "# wav2vec2-base-timit-demo-colab\n\nThis model is a fine-tuned version of facebook/wav2vec2-base on the None dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 32\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 1000\n- num_epochs: 1\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.10.0+cu111\n- Datasets 1.13.3\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us \n", "# wav2vec2-base-timit-demo-colab\n\nThis model is a fine-tuned version of facebook/wav2vec2-base on the None dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 32\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 1000\n- num_epochs: 1\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.10.0+cu111\n- Datasets 1.13.3\n- Tokenizers 0.10.3" ]
feature-extraction
transformers
Github: https://github.com/haisongzhang/roberta-tiny-cased
{}
haisongzhang/roberta-tiny-cased
null
[ "transformers", "pytorch", "tf", "jax", "bert", "feature-extraction", "endpoints_compatible", "has_space", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tf #jax #bert #feature-extraction #endpoints_compatible #has_space #region-us
Github: URL
[]
[ "TAGS\n#transformers #pytorch #tf #jax #bert #feature-extraction #endpoints_compatible #has_space #region-us \n" ]
fill-mask
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bertweet-base-SNS_BRANDS_100k This model is a fine-tuned version of [vinai/bertweet-base](https://huggingface.co/vinai/bertweet-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0483 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 0.0735 | 1.0 | 2928 | 0.0670 | | 0.0574 | 2.0 | 5856 | 0.0529 | | 0.0497 | 3.0 | 8784 | 0.0483 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "model-index": [{"name": "bertweet-base-SNS_BRANDS_100k", "results": []}]}
haji2438/bertweet-base-SNS_BRANDS_100k
null
[ "transformers", "pytorch", "tensorboard", "roberta", "fill-mask", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #roberta #fill-mask #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us
bertweet-base-SNS\_BRANDS\_100k =============================== This model is a fine-tuned version of vinai/bertweet-base on the None dataset. It achieves the following results on the evaluation set: * Loss: 0.0483 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 1e-05 * train\_batch\_size: 8 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3.0 ### Training results ### Framework versions * Transformers 4.15.0 * Pytorch 1.10.0+cu111 * Datasets 1.17.0 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.15.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.17.0\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #roberta #fill-mask #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.15.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.17.0\n* Tokenizers 0.10.3" ]
fill-mask
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bertweet-base-SNS_BRANDS_200k This model is a fine-tuned version of [vinai/bertweet-base](https://huggingface.co/vinai/bertweet-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0243 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:-----:|:---------------:| | 0.0428 | 1.0 | 5882 | 0.0336 | | 0.0276 | 2.0 | 11764 | 0.0241 | | 0.0251 | 3.0 | 17646 | 0.0243 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.0+cu111 - Datasets 1.18.0 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "model-index": [{"name": "bertweet-base-SNS_BRANDS_200k", "results": []}]}
haji2438/bertweet-base-SNS_BRANDS_200k
null
[ "transformers", "pytorch", "tensorboard", "roberta", "fill-mask", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #roberta #fill-mask #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us
bertweet-base-SNS\_BRANDS\_200k =============================== This model is a fine-tuned version of vinai/bertweet-base on the None dataset. It achieves the following results on the evaluation set: * Loss: 0.0243 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 8 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 500 * num\_epochs: 3 ### Training results ### Framework versions * Transformers 4.15.0 * Pytorch 1.10.0+cu111 * Datasets 1.18.0 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 3", "### Training results", "### Framework versions\n\n\n* Transformers 4.15.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.0\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #roberta #fill-mask #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 3", "### Training results", "### Framework versions\n\n\n* Transformers 4.15.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.0\n* Tokenizers 0.10.3" ]
fill-mask
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bertweet-base-SNS_BRANDS_50k This model is a fine-tuned version of [vinai/bertweet-base](https://huggingface.co/vinai/bertweet-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0490 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 4 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 0.0787 | 1.0 | 1465 | 0.0751 | | 0.0662 | 2.0 | 2930 | 0.0628 | | 0.053 | 3.0 | 4395 | 0.0531 | | 0.0452 | 4.0 | 5860 | 0.0490 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.0+cu111 - Datasets 1.18.0 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "model-index": [{"name": "bertweet-base-SNS_BRANDS_50k", "results": []}]}
haji2438/bertweet-base-SNS_BRANDS_50k
null
[ "transformers", "pytorch", "tensorboard", "roberta", "fill-mask", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #roberta #fill-mask #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us
bertweet-base-SNS\_BRANDS\_50k ============================== This model is a fine-tuned version of vinai/bertweet-base on the None dataset. It achieves the following results on the evaluation set: * Loss: 0.0490 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 8 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 500 * num\_epochs: 4 ### Training results ### Framework versions * Transformers 4.15.0 * Pytorch 1.10.0+cu111 * Datasets 1.18.0 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 4", "### Training results", "### Framework versions\n\n\n* Transformers 4.15.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.0\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #roberta #fill-mask #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 4", "### Training results", "### Framework versions\n\n\n* Transformers 4.15.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.0\n* Tokenizers 0.10.3" ]
fill-mask
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bertweet-base-finetuned-IGtext This model is a fine-tuned version of [vinai/bertweet-base](https://huggingface.co/vinai/bertweet-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 2.0334 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 4 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 2.6741 | 1.0 | 505 | 2.2096 | | 2.3183 | 2.0 | 1010 | 2.0934 | | 2.2089 | 3.0 | 1515 | 2.0595 | | 2.1473 | 4.0 | 2020 | 2.0246 | ### Framework versions - Transformers 4.12.3 - Pytorch 1.10.0+cu111 - Datasets 1.15.1 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "model-index": [{"name": "bertweet-base-finetuned-IGtext", "results": []}]}
haji2438/bertweet-base-finetuned-IGtext
null
[ "transformers", "pytorch", "tensorboard", "roberta", "fill-mask", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #roberta #fill-mask #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us
bertweet-base-finetuned-IGtext ============================== This model is a fine-tuned version of vinai/bertweet-base on the None dataset. It achieves the following results on the evaluation set: * Loss: 2.0334 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 32 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 4 ### Training results ### Framework versions * Transformers 4.12.3 * Pytorch 1.10.0+cu111 * Datasets 1.15.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 4", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #roberta #fill-mask #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 4", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
fill-mask
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bertweet-base-finetuned-SNS-brand-personality This model is a fine-tuned version of [vinai/bertweet-base](https://huggingface.co/vinai/bertweet-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0498 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 0.0757 | 1.0 | 1549 | 0.0723 | | 0.0605 | 2.0 | 3098 | 0.0573 | | 0.0498 | 3.0 | 4647 | 0.0498 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
{"tags": ["generated_from_trainer"], "model-index": [{"name": "bertweet-base-finetuned-SNS-brand-personality", "results": []}]}
haji2438/bertweet-base-finetuned-SNS-brand-personality
null
[ "transformers", "pytorch", "tensorboard", "roberta", "fill-mask", "generated_from_trainer", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #roberta #fill-mask #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us
bertweet-base-finetuned-SNS-brand-personality ============================================= This model is a fine-tuned version of vinai/bertweet-base on the None dataset. It achieves the following results on the evaluation set: * Loss: 0.0498 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 8 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3.0 ### Training results ### Framework versions * Transformers 4.15.0 * Pytorch 1.10.0+cu111 * Datasets 1.17.0 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.15.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.17.0\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #roberta #fill-mask #generated_from_trainer #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0", "### Training results", "### Framework versions\n\n\n* Transformers 4.15.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.17.0\n* Tokenizers 0.10.3" ]
text-generation
transformers
# XLNet-japanese ## Model description This model require Mecab and senetencepiece with XLNetTokenizer. See details https://qiita.com/mkt3/items/4d0ae36f3f212aee8002 This model uses NFKD as the normalization method for character encoding. Japanese muddle marks and semi-muddle marks will be lost. *日本語の濁点・半濁点がないモデルです* #### How to use ```python from fugashi import Tagger from transformers import ( pipeline, XLNetLMHeadModel, XLNetTokenizer ) class XLNet(): def __init__(self): self.m = Tagger('-Owakati') self.gen_model = XLNetLMHeadModel.from_pretrained("hajime9652/xlnet-japanese") self.gen_tokenizer = XLNetTokenizer.from_pretrained("hajime9652/xlnet-japanese") def generate(self, prompt="福岡のご飯は美味しい。コンパクトで暮らしやすい街。"): prompt = self.m.parse(prompt) inputs = self.gen_tokenizer.encode(prompt, add_special_tokens=False, return_tensors="pt") prompt_length = len(self.gen_tokenizer.decode(inputs[0], skip_special_tokens=True, clean_up_tokenization_spaces=True)) outputs = self.gen_model.generate(inputs, max_length=200, do_sample=True, top_p=0.95, top_k=60) generated = prompt + self.gen_tokenizer.decode(outputs[0])[prompt_length:] return generated ``` #### Limitations and bias This model's training use the Japanese Business News. # Important matter The company that created and published this model is called Stockmark. This repository is for use by HuggingFace and not for infringement. See this documents https://qiita.com/mkt3/items/4d0ae36f3f212aee8002 published by https://github.com/mkt3
{"language": ["ja"], "license": ["apache-2.0"], "tags": ["xlnet", "lm-head", "causal-lm"], "datasets": ["Japanese_Business_News"]}
hajime9652/xlnet-japanese
null
[ "transformers", "pytorch", "xlnet", "text-generation", "lm-head", "causal-lm", "ja", "dataset:Japanese_Business_News", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "ja" ]
TAGS #transformers #pytorch #xlnet #text-generation #lm-head #causal-lm #ja #dataset-Japanese_Business_News #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# XLNet-japanese ## Model description This model require Mecab and senetencepiece with XLNetTokenizer. See details URL This model uses NFKD as the normalization method for character encoding. Japanese muddle marks and semi-muddle marks will be lost. *日本語の濁点・半濁点がないモデルです* #### How to use #### Limitations and bias This model's training use the Japanese Business News. # Important matter The company that created and published this model is called Stockmark. This repository is for use by HuggingFace and not for infringement. See this documents URL published by URL
[ "# XLNet-japanese", "## Model description\nThis model require Mecab and senetencepiece with XLNetTokenizer.\nSee details URL\n\nThis model uses NFKD as the normalization method for character encoding.\nJapanese muddle marks and semi-muddle marks will be lost.\n\n*日本語の濁点・半濁点がないモデルです*", "#### How to use", "#### Limitations and bias\nThis model's training use the Japanese Business News.", "# Important matter\nThe company that created and published this model is called Stockmark.\nThis repository is for use by HuggingFace and not for infringement.\nSee this documents URL\npublished by URL" ]
[ "TAGS\n#transformers #pytorch #xlnet #text-generation #lm-head #causal-lm #ja #dataset-Japanese_Business_News #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# XLNet-japanese", "## Model description\nThis model require Mecab and senetencepiece with XLNetTokenizer.\nSee details URL\n\nThis model uses NFKD as the normalization method for character encoding.\nJapanese muddle marks and semi-muddle marks will be lost.\n\n*日本語の濁点・半濁点がないモデルです*", "#### How to use", "#### Limitations and bias\nThis model's training use the Japanese Business News.", "# Important matter\nThe company that created and published this model is called Stockmark.\nThis repository is for use by HuggingFace and not for infringement.\nSee this documents URL\npublished by URL" ]
text-generation
transformers
This model has been initialized with random values. It is supposed to be used for the purpose of debugging.
{}
hakurei/gpt-j-random-tinier
null
[ "transformers", "pytorch", "gptj", "text-generation", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gptj #text-generation #autotrain_compatible #endpoints_compatible #region-us
This model has been initialized with random values. It is supposed to be used for the purpose of debugging.
[]
[ "TAGS\n#transformers #pytorch #gptj #text-generation #autotrain_compatible #endpoints_compatible #region-us \n" ]
text-generation
transformers
# Lit-125M - A Small Fine-tuned Model For Fictional Storytelling Lit-125M is a GPT-Neo 125M model fine-tuned on 2GB of a diverse range of light novels, erotica, and annotated literature for the purpose of generating novel-like fictional text. ## Model Description The model used for fine-tuning is [GPT-Neo 125M](https://huggingface.co/EleutherAI/gpt-neo-125M), which is a 125 million parameter auto-regressive language model trained on [The Pile](https://pile.eleuther.ai/).. ## Training Data & Annotative Prompting The data used in fine-tuning has been gathered from various sources such as the [Gutenberg Project](https://www.gutenberg.org/). The annotated fiction dataset has prepended tags to assist in generating towards a particular style. Here is an example prompt that shows how to use the annotations. ``` [ Title: The Dunwich Horror; Author: H. P. Lovecraft; Genre: Horror; Tags: 3rdperson, scary; Style: Dark ] *** When a traveler in north central Massachusetts takes the wrong fork... ``` The annotations can be mixed and matched to help generate towards a specific style. ## Downstream Uses This model can be used for entertainment purposes and as a creative writing assistant for fiction writers. The small size of the model can also help for easy debugging or further development of other models with a similar purpose. ## Example Code ``` from transformers import AutoTokenizer, AutoModelForCausalLM model = AutoModelForCausalLM.from_pretrained('hakurei/lit-125M') tokenizer = AutoTokenizer.from_pretrained('hakurei/lit-125M') prompt = '''[ Title: The Dunwich Horror; Author: H. P. Lovecraft; Genre: Horror ] *** When a traveler''' input_ids = tokenizer.encode(prompt, return_tensors='pt') output = model.generate(input_ids, do_sample=True, temperature=1.0, top_p=0.9, repetition_penalty=1.2, max_length=len(input_ids[0])+100, pad_token_id=tokenizer.eos_token_id) generated_text = tokenizer.decode(output[0]) print(generated_text) ``` An example output from this code produces a result that will look similar to: ``` [ Title: The Dunwich Horror; Author: H. P. Lovecraft; Genre: Horror ] *** When a traveler takes a trip through the streets of the world, the traveler feels like a youkai with a whole world inside her mind. It can be very scary for a youkai. When someone goes in the opposite direction and knocks on your door, it is actually the first time you have ever come to investigate something like that. That's right: everyone has heard stories about youkai, right? If you have heard them, you know what I'm talking about. It's hard not to say you ``` ## Team members and Acknowledgements - [Anthony Mercurio](https://github.com/harubaru) - Imperishable_NEET
{"language": ["en"], "license": "mit", "tags": ["pytorch", "causal-lm"]}
hakurei/lit-125M
null
[ "transformers", "pytorch", "gpt_neo", "text-generation", "causal-lm", "en", "license:mit", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gpt_neo #text-generation #causal-lm #en #license-mit #autotrain_compatible #endpoints_compatible #has_space #region-us
# Lit-125M - A Small Fine-tuned Model For Fictional Storytelling Lit-125M is a GPT-Neo 125M model fine-tuned on 2GB of a diverse range of light novels, erotica, and annotated literature for the purpose of generating novel-like fictional text. ## Model Description The model used for fine-tuning is GPT-Neo 125M, which is a 125 million parameter auto-regressive language model trained on The Pile.. ## Training Data & Annotative Prompting The data used in fine-tuning has been gathered from various sources such as the Gutenberg Project. The annotated fiction dataset has prepended tags to assist in generating towards a particular style. Here is an example prompt that shows how to use the annotations. The annotations can be mixed and matched to help generate towards a specific style. ## Downstream Uses This model can be used for entertainment purposes and as a creative writing assistant for fiction writers. The small size of the model can also help for easy debugging or further development of other models with a similar purpose. ## Example Code An example output from this code produces a result that will look similar to: ## Team members and Acknowledgements - Anthony Mercurio - Imperishable_NEET
[ "# Lit-125M - A Small Fine-tuned Model For Fictional Storytelling\n\nLit-125M is a GPT-Neo 125M model fine-tuned on 2GB of a diverse range of light novels, erotica, and annotated literature for the purpose of generating novel-like fictional text.", "## Model Description\n\nThe model used for fine-tuning is GPT-Neo 125M, which is a 125 million parameter auto-regressive language model trained on The Pile..", "## Training Data & Annotative Prompting\n\nThe data used in fine-tuning has been gathered from various sources such as the Gutenberg Project. The annotated fiction dataset has prepended tags to assist in generating towards a particular style. Here is an example prompt that shows how to use the annotations.\n\n\n\nThe annotations can be mixed and matched to help generate towards a specific style.", "## Downstream Uses\n\nThis model can be used for entertainment purposes and as a creative writing assistant for fiction writers. The small size of the model can also help for easy debugging or further development of other models with a similar purpose.", "## Example Code\n\n\n\nAn example output from this code produces a result that will look similar to:", "## Team members and Acknowledgements\n\n- Anthony Mercurio\n- Imperishable_NEET" ]
[ "TAGS\n#transformers #pytorch #gpt_neo #text-generation #causal-lm #en #license-mit #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "# Lit-125M - A Small Fine-tuned Model For Fictional Storytelling\n\nLit-125M is a GPT-Neo 125M model fine-tuned on 2GB of a diverse range of light novels, erotica, and annotated literature for the purpose of generating novel-like fictional text.", "## Model Description\n\nThe model used for fine-tuning is GPT-Neo 125M, which is a 125 million parameter auto-regressive language model trained on The Pile..", "## Training Data & Annotative Prompting\n\nThe data used in fine-tuning has been gathered from various sources such as the Gutenberg Project. The annotated fiction dataset has prepended tags to assist in generating towards a particular style. Here is an example prompt that shows how to use the annotations.\n\n\n\nThe annotations can be mixed and matched to help generate towards a specific style.", "## Downstream Uses\n\nThis model can be used for entertainment purposes and as a creative writing assistant for fiction writers. The small size of the model can also help for easy debugging or further development of other models with a similar purpose.", "## Example Code\n\n\n\nAn example output from this code produces a result that will look similar to:", "## Team members and Acknowledgements\n\n- Anthony Mercurio\n- Imperishable_NEET" ]
null
transformers
# Lit-6B - A Large Fine-tuned Model For Fictional Storytelling Lit-6B is a GPT-J 6B model fine-tuned on 2GB of a diverse range of light novels, erotica, and annotated literature for the purpose of generating novel-like fictional text. ## Model Description The model used for fine-tuning is [GPT-J](https://github.com/kingoflolz/mesh-transformer-jax), which is a 6 billion parameter auto-regressive language model trained on [The Pile](https://pile.eleuther.ai/). ## Training Data & Annotative Prompting The data used in fine-tuning has been gathered from various sources such as the [Gutenberg Project](https://www.gutenberg.org/). The annotated fiction dataset has prepended tags to assist in generating towards a particular style. Here is an example prompt that shows how to use the annotations. ``` [ Title: The Dunwich Horror; Author: H. P. Lovecraft; Genre: Horror; Tags: 3rdperson, scary; Style: Dark ] *** When a traveler in north central Massachusetts takes the wrong fork... ``` The annotations can be mixed and matched to help generate towards a specific style. ## Downstream Uses This model can be used for entertainment purposes and as a creative writing assistant for fiction writers. ## Example Code ``` from transformers import AutoTokenizer, AutoModelForCausalLM model = AutoModelForCausalLM.from_pretrained('hakurei/lit-6B') tokenizer = AutoTokenizer.from_pretrained('hakurei/lit-6B') prompt = '''[ Title: The Dunwich Horror; Author: H. P. Lovecraft; Genre: Horror ] *** When a traveler''' input_ids = tokenizer.encode(prompt, return_tensors='pt') output = model.generate(input_ids, do_sample=True, temperature=1.0, top_p=0.9, repetition_penalty=1.2, max_length=len(input_ids[0])+100, pad_token_id=tokenizer.eos_token_id) generated_text = tokenizer.decode(output[0]) print(generated_text) ``` An example output from this code produces a result that will look similar to: ``` [ Title: The Dunwich Horror; Author: H. P. Lovecraft; Genre: Horror ] *** When a traveler comes to an unknown region, his thoughts turn inevitably towards the old gods and legends which cluster around its appearance. It is not that he believes in them or suspects their reality—but merely because they are present somewhere else in creation just as truly as himself, and so belong of necessity in any landscape whose features cannot be altogether strange to him. Moreover, man has been prone from ancient times to brood over those things most connected with the places where he dwells. Thus the Olympian deities who ruled Hyper ``` ## Team members and Acknowledgements This project would not have been possible without the computational resources graciously provided by the [TPU Research Cloud](https://sites.research.google/trc/) - [Anthony Mercurio](https://github.com/harubaru) - Imperishable_NEET
{"language": ["en"], "license": "mit", "tags": ["pytorch", "causal-lm"]}
hakurei/lit-6B-8bit
null
[ "transformers", "pytorch", "causal-lm", "en", "license:mit", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #causal-lm #en #license-mit #endpoints_compatible #region-us
# Lit-6B - A Large Fine-tuned Model For Fictional Storytelling Lit-6B is a GPT-J 6B model fine-tuned on 2GB of a diverse range of light novels, erotica, and annotated literature for the purpose of generating novel-like fictional text. ## Model Description The model used for fine-tuning is GPT-J, which is a 6 billion parameter auto-regressive language model trained on The Pile. ## Training Data & Annotative Prompting The data used in fine-tuning has been gathered from various sources such as the Gutenberg Project. The annotated fiction dataset has prepended tags to assist in generating towards a particular style. Here is an example prompt that shows how to use the annotations. The annotations can be mixed and matched to help generate towards a specific style. ## Downstream Uses This model can be used for entertainment purposes and as a creative writing assistant for fiction writers. ## Example Code An example output from this code produces a result that will look similar to: ## Team members and Acknowledgements This project would not have been possible without the computational resources graciously provided by the TPU Research Cloud - Anthony Mercurio - Imperishable_NEET
[ "# Lit-6B - A Large Fine-tuned Model For Fictional Storytelling\n\nLit-6B is a GPT-J 6B model fine-tuned on 2GB of a diverse range of light novels, erotica, and annotated literature for the purpose of generating novel-like fictional text.", "## Model Description\n\nThe model used for fine-tuning is GPT-J, which is a 6 billion parameter auto-regressive language model trained on The Pile.", "## Training Data & Annotative Prompting\n\nThe data used in fine-tuning has been gathered from various sources such as the Gutenberg Project. The annotated fiction dataset has prepended tags to assist in generating towards a particular style. Here is an example prompt that shows how to use the annotations.\n\n\n\nThe annotations can be mixed and matched to help generate towards a specific style.", "## Downstream Uses\n\nThis model can be used for entertainment purposes and as a creative writing assistant for fiction writers.", "## Example Code\n\n\n\nAn example output from this code produces a result that will look similar to:", "## Team members and Acknowledgements\n\nThis project would not have been possible without the computational resources graciously provided by the TPU Research Cloud\n\n- Anthony Mercurio\n- Imperishable_NEET" ]
[ "TAGS\n#transformers #pytorch #causal-lm #en #license-mit #endpoints_compatible #region-us \n", "# Lit-6B - A Large Fine-tuned Model For Fictional Storytelling\n\nLit-6B is a GPT-J 6B model fine-tuned on 2GB of a diverse range of light novels, erotica, and annotated literature for the purpose of generating novel-like fictional text.", "## Model Description\n\nThe model used for fine-tuning is GPT-J, which is a 6 billion parameter auto-regressive language model trained on The Pile.", "## Training Data & Annotative Prompting\n\nThe data used in fine-tuning has been gathered from various sources such as the Gutenberg Project. The annotated fiction dataset has prepended tags to assist in generating towards a particular style. Here is an example prompt that shows how to use the annotations.\n\n\n\nThe annotations can be mixed and matched to help generate towards a specific style.", "## Downstream Uses\n\nThis model can be used for entertainment purposes and as a creative writing assistant for fiction writers.", "## Example Code\n\n\n\nAn example output from this code produces a result that will look similar to:", "## Team members and Acknowledgements\n\nThis project would not have been possible without the computational resources graciously provided by the TPU Research Cloud\n\n- Anthony Mercurio\n- Imperishable_NEET" ]
text-generation
transformers
# Lit-6B - A Large Fine-tuned Model For Fictional Storytelling Lit-6B is a GPT-J 6B model fine-tuned on 2GB of a diverse range of light novels, erotica, and annotated literature for the purpose of generating novel-like fictional text. ## Model Description The model used for fine-tuning is [GPT-J](https://github.com/kingoflolz/mesh-transformer-jax), which is a 6 billion parameter auto-regressive language model trained on [The Pile](https://pile.eleuther.ai/). ## Training Data & Annotative Prompting The data used in fine-tuning has been gathered from various sources such as the [Gutenberg Project](https://www.gutenberg.org/). The annotated fiction dataset has prepended tags to assist in generating towards a particular style. Here is an example prompt that shows how to use the annotations. ``` [ Title: The Dunwich Horror; Author: H. P. Lovecraft; Genre: Horror; Tags: 3rdperson, scary; Style: Dark ] *** When a traveler in north central Massachusetts takes the wrong fork... ``` The annotations can be mixed and matched to help generate towards a specific style. ## Downstream Uses This model can be used for entertainment purposes and as a creative writing assistant for fiction writers. ## Example Code ``` from transformers import AutoTokenizer, AutoModelForCausalLM model = AutoModelForCausalLM.from_pretrained('hakurei/lit-6B') tokenizer = AutoTokenizer.from_pretrained('hakurei/lit-6B') prompt = '''[ Title: The Dunwich Horror; Author: H. P. Lovecraft; Genre: Horror ] *** When a traveler''' input_ids = tokenizer.encode(prompt, return_tensors='pt') output = model.generate(input_ids, do_sample=True, temperature=1.0, top_p=0.9, repetition_penalty=1.2, max_length=len(input_ids[0])+100, pad_token_id=tokenizer.eos_token_id) generated_text = tokenizer.decode(output[0]) print(generated_text) ``` An example output from this code produces a result that will look similar to: ``` [ Title: The Dunwich Horror; Author: H. P. Lovecraft; Genre: Horror ] *** When a traveler comes to an unknown region, his thoughts turn inevitably towards the old gods and legends which cluster around its appearance. It is not that he believes in them or suspects their reality—but merely because they are present somewhere else in creation just as truly as himself, and so belong of necessity in any landscape whose features cannot be altogether strange to him. Moreover, man has been prone from ancient times to brood over those things most connected with the places where he dwells. Thus the Olympian deities who ruled Hyper ``` ## Team members and Acknowledgements This project would not have been possible without the computational resources graciously provided by the [TPU Research Cloud](https://sites.research.google/trc/) - [Anthony Mercurio](https://github.com/harubaru) - Imperishable_NEET
{"language": ["en"], "license": "mit", "tags": ["pytorch", "causal-lm"]}
hakurei/lit-6B
null
[ "transformers", "pytorch", "gptj", "text-generation", "causal-lm", "en", "license:mit", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #gptj #text-generation #causal-lm #en #license-mit #autotrain_compatible #endpoints_compatible #has_space #region-us
# Lit-6B - A Large Fine-tuned Model For Fictional Storytelling Lit-6B is a GPT-J 6B model fine-tuned on 2GB of a diverse range of light novels, erotica, and annotated literature for the purpose of generating novel-like fictional text. ## Model Description The model used for fine-tuning is GPT-J, which is a 6 billion parameter auto-regressive language model trained on The Pile. ## Training Data & Annotative Prompting The data used in fine-tuning has been gathered from various sources such as the Gutenberg Project. The annotated fiction dataset has prepended tags to assist in generating towards a particular style. Here is an example prompt that shows how to use the annotations. The annotations can be mixed and matched to help generate towards a specific style. ## Downstream Uses This model can be used for entertainment purposes and as a creative writing assistant for fiction writers. ## Example Code An example output from this code produces a result that will look similar to: ## Team members and Acknowledgements This project would not have been possible without the computational resources graciously provided by the TPU Research Cloud - Anthony Mercurio - Imperishable_NEET
[ "# Lit-6B - A Large Fine-tuned Model For Fictional Storytelling\n\nLit-6B is a GPT-J 6B model fine-tuned on 2GB of a diverse range of light novels, erotica, and annotated literature for the purpose of generating novel-like fictional text.", "## Model Description\n\nThe model used for fine-tuning is GPT-J, which is a 6 billion parameter auto-regressive language model trained on The Pile.", "## Training Data & Annotative Prompting\n\nThe data used in fine-tuning has been gathered from various sources such as the Gutenberg Project. The annotated fiction dataset has prepended tags to assist in generating towards a particular style. Here is an example prompt that shows how to use the annotations.\n\n\n\nThe annotations can be mixed and matched to help generate towards a specific style.", "## Downstream Uses\n\nThis model can be used for entertainment purposes and as a creative writing assistant for fiction writers.", "## Example Code\n\n\n\nAn example output from this code produces a result that will look similar to:", "## Team members and Acknowledgements\n\nThis project would not have been possible without the computational resources graciously provided by the TPU Research Cloud\n\n- Anthony Mercurio\n- Imperishable_NEET" ]
[ "TAGS\n#transformers #pytorch #gptj #text-generation #causal-lm #en #license-mit #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "# Lit-6B - A Large Fine-tuned Model For Fictional Storytelling\n\nLit-6B is a GPT-J 6B model fine-tuned on 2GB of a diverse range of light novels, erotica, and annotated literature for the purpose of generating novel-like fictional text.", "## Model Description\n\nThe model used for fine-tuning is GPT-J, which is a 6 billion parameter auto-regressive language model trained on The Pile.", "## Training Data & Annotative Prompting\n\nThe data used in fine-tuning has been gathered from various sources such as the Gutenberg Project. The annotated fiction dataset has prepended tags to assist in generating towards a particular style. Here is an example prompt that shows how to use the annotations.\n\n\n\nThe annotations can be mixed and matched to help generate towards a specific style.", "## Downstream Uses\n\nThis model can be used for entertainment purposes and as a creative writing assistant for fiction writers.", "## Example Code\n\n\n\nAn example output from this code produces a result that will look similar to:", "## Team members and Acknowledgements\n\nThis project would not have been possible without the computational resources graciously provided by the TPU Research Cloud\n\n- Anthony Mercurio\n- Imperishable_NEET" ]
text-generation
transformers
# DOC DialoGPT Model
{"tags": ["conversational"]}
hama/Doctor_Bot
null
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# DOC DialoGPT Model
[ "# DOC DialoGPT Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# DOC DialoGPT Model" ]
text-generation
transformers
# Harry Potter DialoGPT Model
{"tags": ["conversational"]}
hama/Harry_Bot
null
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Harry Potter DialoGPT Model
[ "# Harry Potter DialoGPT Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Harry Potter DialoGPT Model" ]
text-generation
transformers
# BArney DialoGPT Model
{"tags": ["conversational"]}
hama/barney_bot
null
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# BArney DialoGPT Model
[ "# BArney DialoGPT Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# BArney DialoGPT Model" ]
text-generation
transformers
# me 101
{"tags": ["conversational"]}
hama/me0.01
null
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# me 101
[ "# me 101" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# me 101" ]
text-generation
transformers
# Rick and Morty DialoGPT Model
{"tags": ["conversational"]}
hama/rick_bot
null
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Rick and Morty DialoGPT Model
[ "# Rick and Morty DialoGPT Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Rick and Morty DialoGPT Model" ]
text2text-generation
transformers
# mBart50 for Zeroshot Azerbaijani-Turkish Translation The mBart50 model is finetuned on English-Azerbaijani-Turkish translation leaving Az<->Tr as zeroshot directions. The method of tied representations is used to enforce alignment between semantically equivalent sentences leading to superior zeroshot performance.
{}
hamishs/mBART50-en-az-tr1
null
[ "transformers", "pytorch", "mbart", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #mbart #text2text-generation #autotrain_compatible #endpoints_compatible #region-us
# mBart50 for Zeroshot Azerbaijani-Turkish Translation The mBart50 model is finetuned on English-Azerbaijani-Turkish translation leaving Az<->Tr as zeroshot directions. The method of tied representations is used to enforce alignment between semantically equivalent sentences leading to superior zeroshot performance.
[ "# mBart50 for Zeroshot Azerbaijani-Turkish Translation\nThe mBart50 model is finetuned on English-Azerbaijani-Turkish translation leaving Az<->Tr as zeroshot directions. The method of tied representations is used to enforce alignment between semantically equivalent sentences leading to superior zeroshot performance." ]
[ "TAGS\n#transformers #pytorch #mbart #text2text-generation #autotrain_compatible #endpoints_compatible #region-us \n", "# mBart50 for Zeroshot Azerbaijani-Turkish Translation\nThe mBart50 model is finetuned on English-Azerbaijani-Turkish translation leaving Az<->Tr as zeroshot directions. The method of tied representations is used to enforce alignment between semantically equivalent sentences leading to superior zeroshot performance." ]
null
null
hello
{}
hamxxxa/SBert
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
hello
[]
[ "TAGS\n#region-us \n" ]
question-answering
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # electra-small-discriminator-finetuned-squad This model is a fine-tuned version of [google/electra-small-discriminator](https://huggingface.co/google/electra-small-discriminator) on the squad dataset. It achieves the following results on the evaluation set: - Loss: 1.2174 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 1.5751 | 1.0 | 2767 | 1.3952 | | 1.2939 | 2.0 | 5534 | 1.2458 | | 1.1866 | 3.0 | 8301 | 1.2174 | ### Framework versions - Transformers 4.12.5 - Pytorch 1.10.0+cu111 - Datasets 1.16.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["squad"], "model-index": [{"name": "electra-small-discriminator-finetuned-squad", "results": []}]}
hankzhong/electra-small-discriminator-finetuned-squad
null
[ "transformers", "pytorch", "tensorboard", "electra", "question-answering", "generated_from_trainer", "dataset:squad", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #electra #question-answering #generated_from_trainer #dataset-squad #license-apache-2.0 #endpoints_compatible #region-us
electra-small-discriminator-finetuned-squad =========================================== This model is a fine-tuned version of google/electra-small-discriminator on the squad dataset. It achieves the following results on the evaluation set: * Loss: 1.2174 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 32 * eval\_batch\_size: 32 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3 ### Training results ### Framework versions * Transformers 4.12.5 * Pytorch 1.10.0+cu111 * Datasets 1.16.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #electra #question-answering #generated_from_trainer #dataset-squad #license-apache-2.0 #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3" ]
fill-mask
transformers
## Not yet
{}
hansgun/model_test
null
[ "transformers", "tf", "bert", "fill-mask", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #tf #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us
## Not yet
[ "## Not yet" ]
[ "TAGS\n#transformers #tf #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n", "## Not yet" ]
text2text-generation
transformers
# Helsinki-NLP/opus-mt-en-vi - This model is a fine-tune checkpoint of [Helsinki-NLP/opus-mt-en-vi](https://huggingface.co/Helsinki-NLP/opus-mt-en-vi). - This model reaches BLEU score = 33.086 on the test set of IWSLT'15 English-Vietnamese data. # Fine-tuning hyper-parameters - learning_rate = 1e-4 - batch_size = 4 - num_train_epochs = 3.0
{}
haotieu/en-vi-mt-model
null
[ "transformers", "pytorch", "marian", "text2text-generation", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #marian #text2text-generation #autotrain_compatible #endpoints_compatible #has_space #region-us
# Helsinki-NLP/opus-mt-en-vi - This model is a fine-tune checkpoint of Helsinki-NLP/opus-mt-en-vi. - This model reaches BLEU score = 33.086 on the test set of IWSLT'15 English-Vietnamese data. # Fine-tuning hyper-parameters - learning_rate = 1e-4 - batch_size = 4 - num_train_epochs = 3.0
[ "# Helsinki-NLP/opus-mt-en-vi\n- This model is a fine-tune checkpoint of Helsinki-NLP/opus-mt-en-vi.\n- This model reaches BLEU score = 33.086 on the test set of IWSLT'15 English-Vietnamese data.", "# Fine-tuning hyper-parameters\n- learning_rate = 1e-4\n- batch_size = 4\n- num_train_epochs = 3.0" ]
[ "TAGS\n#transformers #pytorch #marian #text2text-generation #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "# Helsinki-NLP/opus-mt-en-vi\n- This model is a fine-tune checkpoint of Helsinki-NLP/opus-mt-en-vi.\n- This model reaches BLEU score = 33.086 on the test set of IWSLT'15 English-Vietnamese data.", "# Fine-tuning hyper-parameters\n- learning_rate = 1e-4\n- batch_size = 4\n- num_train_epochs = 3.0" ]
feature-extraction
sentence-transformers
# multi-qa-MiniLM-L6-cos-v1 This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 384 dimensional dense vector space and was designed for **semantic search**. It has been trained on 215M (question, answer) pairs from diverse sources. For an introduction to semantic search, have a look at: [SBERT.net - Semantic Search](https://www.sbert.net/examples/applications/semantic-search/README.html) ## Usage (Sentence-Transformers) Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed: ``` pip install -U sentence-transformers ``` Then you can use the model like this: ```python from sentence_transformers import SentenceTransformer, util query = "How many people live in London?" docs = ["Around 9 Million people live in London", "London is known for its financial district"] #Load the model model = SentenceTransformer('sentence-transformers/multi-qa-MiniLM-L6-cos-v1') #Encode query and documents query_emb = model.encode(query) doc_emb = model.encode(docs) #Compute dot score between query and all document embeddings scores = util.dot_score(query_emb, doc_emb)[0].cpu().tolist() #Combine docs & scores doc_score_pairs = list(zip(docs, scores)) #Sort by decreasing score doc_score_pairs = sorted(doc_score_pairs, key=lambda x: x[1], reverse=True) #Output passages & scores for doc, score in doc_score_pairs: print(score, doc) ``` ## Usage (HuggingFace Transformers) Without [sentence-transformers](https://www.SBERT.net), you can use the model like this: First, you pass your input through the transformer model, then you have to apply the correct pooling-operation on-top of the contextualized word embeddings. ```python from transformers import AutoTokenizer, AutoModel import torch import torch.nn.functional as F #Mean Pooling - Take average of all tokens def mean_pooling(model_output, attention_mask): token_embeddings = model_output.last_hidden_state #First element of model_output contains all token embeddings input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float() return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9) #Encode text def encode(texts): # Tokenize sentences encoded_input = tokenizer(texts, padding=True, truncation=True, return_tensors='pt') # Compute token embeddings with torch.no_grad(): model_output = model(**encoded_input, return_dict=True) # Perform pooling embeddings = mean_pooling(model_output, encoded_input['attention_mask']) # Normalize embeddings embeddings = F.normalize(embeddings, p=2, dim=1) return embeddings # Sentences we want sentence embeddings for query = "How many people live in London?" docs = ["Around 9 Million people live in London", "London is known for its financial district"] # Load model from HuggingFace Hub tokenizer = AutoTokenizer.from_pretrained("sentence-transformers/multi-qa-MiniLM-L6-cos-v1") model = AutoModel.from_pretrained("sentence-transformers/multi-qa-MiniLM-L6-cos-v1") #Encode query and docs query_emb = encode(query) doc_emb = encode(docs) #Compute dot score between query and all document embeddings scores = torch.mm(query_emb, doc_emb.transpose(0, 1))[0].cpu().tolist() #Combine docs & scores doc_score_pairs = list(zip(docs, scores)) #Sort by decreasing score doc_score_pairs = sorted(doc_score_pairs, key=lambda x: x[1], reverse=True) #Output passages & scores for doc, score in doc_score_pairs: print(score, doc) ``` ## Technical Details In the following some technical details how this model must be used: | Setting | Value | | --- | :---: | | Dimensions | 384 | | Produces normalized embeddings | Yes | | Pooling-Method | Mean pooling | | Suitable score functions | dot-product (`util.dot_score`), cosine-similarity (`util.cos_sim`), or euclidean distance | Note: When loaded with `sentence-transformers`, this model produces normalized embeddings with length 1. In that case, dot-product and cosine-similarity are equivalent. dot-product is preferred as it is faster. Euclidean distance is proportional to dot-product and can also be used. ---- ## Background The project aims to train sentence embedding models on very large sentence level datasets using a self-supervised contrastive learning objective. We use a contrastive learning objective: given a sentence from the pair, the model should predict which out of a set of randomly sampled other sentences, was actually paired with it in our dataset. We developped this model during the [Community week using JAX/Flax for NLP & CV](https://discuss.huggingface.co/t/open-to-the-community-community-week-using-jax-flax-for-nlp-cv/7104), organized by Hugging Face. We developped this model as part of the project: [Train the Best Sentence Embedding Model Ever with 1B Training Pairs](https://discuss.huggingface.co/t/train-the-best-sentence-embedding-model-ever-with-1b-training-pairs/7354). We benefited from efficient hardware infrastructure to run the project: 7 TPUs v3-8, as well as intervention from Googles Flax, JAX, and Cloud team member about efficient deep learning frameworks. ## Intended uses Our model is intented to be used for semantic search: It encodes queries / questions and text paragraphs in a dense vector space. It finds relevant documents for the given passages. Note that there is a limit of 512 word pieces: Text longer than that will be truncated. Further note that the model was just trained on input text up to 250 word pieces. It might not work well for longer text. ## Training procedure The full training script is accessible in this current repository: `train_script.py`. ### Pre-training We use the pretrained [`nreimers/MiniLM-L6-H384-uncased`](https://huggingface.co/nreimers/MiniLM-L6-H384-uncased) model. Please refer to the model card for more detailed information about the pre-training procedure. #### Training We use the concatenation from multiple datasets to fine-tune our model. In total we have about 215M (question, answer) pairs. We sampled each dataset given a weighted probability which configuration is detailed in the `data_config.json` file. The model was trained with [MultipleNegativesRankingLoss](https://www.sbert.net/docs/package_reference/losses.html#multiplenegativesrankingloss) using Mean-pooling, cosine-similarity as similarity function, and a scale of 20. | Dataset | Number of training tuples | |--------------------------------------------------------|:--------------------------:| | [WikiAnswers](https://github.com/afader/oqa#wikianswers-corpus) Duplicate question pairs from WikiAnswers | 77,427,422 | | [PAQ](https://github.com/facebookresearch/PAQ) Automatically generated (Question, Paragraph) pairs for each paragraph in Wikipedia | 64,371,441 | | [Stack Exchange](https://huggingface.co/datasets/flax-sentence-embeddings/stackexchange_xml) (Title, Body) pairs from all StackExchanges | 25,316,456 | | [Stack Exchange](https://huggingface.co/datasets/flax-sentence-embeddings/stackexchange_xml) (Title, Answer) pairs from all StackExchanges | 21,396,559 | | [MS MARCO](https://microsoft.github.io/msmarco/) Triplets (query, answer, hard_negative) for 500k queries from Bing search engine | 17,579,773 | | [GOOAQ: Open Question Answering with Diverse Answer Types](https://github.com/allenai/gooaq) (query, answer) pairs for 3M Google queries and Google featured snippet | 3,012,496 | | [Amazon-QA](http://jmcauley.ucsd.edu/data/amazon/qa/) (Question, Answer) pairs from Amazon product pages | 2,448,839 | [Yahoo Answers](https://www.kaggle.com/soumikrakshit/yahoo-answers-dataset) (Title, Answer) pairs from Yahoo Answers | 1,198,260 | | [Yahoo Answers](https://www.kaggle.com/soumikrakshit/yahoo-answers-dataset) (Question, Answer) pairs from Yahoo Answers | 681,164 | | [Yahoo Answers](https://www.kaggle.com/soumikrakshit/yahoo-answers-dataset) (Title, Question) pairs from Yahoo Answers | 659,896 | | [SearchQA](https://huggingface.co/datasets/search_qa) (Question, Answer) pairs for 140k questions, each with Top5 Google snippets on that question | 582,261 | | [ELI5](https://huggingface.co/datasets/eli5) (Question, Answer) pairs from Reddit ELI5 (explainlikeimfive) | 325,475 | | [Stack Exchange](https://huggingface.co/datasets/flax-sentence-embeddings/stackexchange_xml) Duplicate questions pairs (titles) | 304,525 | | [Quora Question Triplets](https://quoradata.quora.com/First-Quora-Dataset-Release-Question-Pairs) (Question, Duplicate_Question, Hard_Negative) triplets for Quora Questions Pairs dataset | 103,663 | | [Natural Questions (NQ)](https://ai.google.com/research/NaturalQuestions) (Question, Paragraph) pairs for 100k real Google queries with relevant Wikipedia paragraph | 100,231 | | [SQuAD2.0](https://rajpurkar.github.io/SQuAD-explorer/) (Question, Paragraph) pairs from SQuAD2.0 dataset | 87,599 | | [TriviaQA](https://huggingface.co/datasets/trivia_qa) (Question, Evidence) pairs | 73,346 | | **Total** | **214,988,242** |
{"tags": ["sentence-transformers", "feature-extraction", "sentence-similarity"], "pipeline_tag": "feature-extraction"}
haqishen/test-mode-fe
null
[ "sentence-transformers", "pytorch", "bert", "feature-extraction", "sentence-similarity", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #sentence-transformers #pytorch #bert #feature-extraction #sentence-similarity #endpoints_compatible #region-us
multi-qa-MiniLM-L6-cos-v1 ========================= This is a sentence-transformers model: It maps sentences & paragraphs to a 384 dimensional dense vector space and was designed for semantic search. It has been trained on 215M (question, answer) pairs from diverse sources. For an introduction to semantic search, have a look at: URL - Semantic Search Usage (Sentence-Transformers) ----------------------------- Using this model becomes easy when you have sentence-transformers installed: Then you can use the model like this: Usage (HuggingFace Transformers) -------------------------------- Without sentence-transformers, you can use the model like this: First, you pass your input through the transformer model, then you have to apply the correct pooling-operation on-top of the contextualized word embeddings. Technical Details ----------------- In the following some technical details how this model must be used: Note: When loaded with 'sentence-transformers', this model produces normalized embeddings with length 1. In that case, dot-product and cosine-similarity are equivalent. dot-product is preferred as it is faster. Euclidean distance is proportional to dot-product and can also be used. --- Background ---------- The project aims to train sentence embedding models on very large sentence level datasets using a self-supervised contrastive learning objective. We use a contrastive learning objective: given a sentence from the pair, the model should predict which out of a set of randomly sampled other sentences, was actually paired with it in our dataset. We developped this model during the Community week using JAX/Flax for NLP & CV, organized by Hugging Face. We developped this model as part of the project: Train the Best Sentence Embedding Model Ever with 1B Training Pairs. We benefited from efficient hardware infrastructure to run the project: 7 TPUs v3-8, as well as intervention from Googles Flax, JAX, and Cloud team member about efficient deep learning frameworks. Intended uses ------------- Our model is intented to be used for semantic search: It encodes queries / questions and text paragraphs in a dense vector space. It finds relevant documents for the given passages. Note that there is a limit of 512 word pieces: Text longer than that will be truncated. Further note that the model was just trained on input text up to 250 word pieces. It might not work well for longer text. Training procedure ------------------ The full training script is accessible in this current repository: 'train\_script.py'. ### Pre-training We use the pretrained 'nreimers/MiniLM-L6-H384-uncased' model. Please refer to the model card for more detailed information about the pre-training procedure. #### Training We use the concatenation from multiple datasets to fine-tune our model. In total we have about 215M (question, answer) pairs. We sampled each dataset given a weighted probability which configuration is detailed in the 'data\_config.json' file. The model was trained with MultipleNegativesRankingLoss using Mean-pooling, cosine-similarity as similarity function, and a scale of 20.
[ "### Pre-training\n\n\nWe use the pretrained 'nreimers/MiniLM-L6-H384-uncased' model. Please refer to the model card for more detailed information about the pre-training procedure.", "#### Training\n\n\nWe use the concatenation from multiple datasets to fine-tune our model. In total we have about 215M (question, answer) pairs.\nWe sampled each dataset given a weighted probability which configuration is detailed in the 'data\\_config.json' file.\n\n\nThe model was trained with MultipleNegativesRankingLoss using Mean-pooling, cosine-similarity as similarity function, and a scale of 20." ]
[ "TAGS\n#sentence-transformers #pytorch #bert #feature-extraction #sentence-similarity #endpoints_compatible #region-us \n", "### Pre-training\n\n\nWe use the pretrained 'nreimers/MiniLM-L6-H384-uncased' model. Please refer to the model card for more detailed information about the pre-training procedure.", "#### Training\n\n\nWe use the concatenation from multiple datasets to fine-tune our model. In total we have about 215M (question, answer) pairs.\nWe sampled each dataset given a weighted probability which configuration is detailed in the 'data\\_config.json' file.\n\n\nThe model was trained with MultipleNegativesRankingLoss using Mean-pooling, cosine-similarity as similarity function, and a scale of 20." ]
question-answering
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-squad This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the squad dataset. It achieves the following results on the evaluation set: - Loss: 1.1642 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:-----:|:---------------:| | 1.2251 | 1.0 | 5533 | 1.1707 | | 0.9554 | 2.0 | 11066 | 1.1211 | | 0.7645 | 3.0 | 16599 | 1.1642 | ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.11.0
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["squad"], "model-index": [{"name": "distilbert-base-uncased-finetuned-squad", "results": []}]}
hark99/distilbert-base-uncased-finetuned-squad
null
[ "transformers", "pytorch", "tensorboard", "distilbert", "question-answering", "generated_from_trainer", "dataset:squad", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #distilbert #question-answering #generated_from_trainer #dataset-squad #license-apache-2.0 #endpoints_compatible #region-us
distilbert-base-uncased-finetuned-squad ======================================= This model is a fine-tuned version of distilbert-base-uncased on the squad dataset. It achieves the following results on the evaluation set: * Loss: 1.1642 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 3 ### Training results ### Framework versions * Transformers 4.16.2 * Pytorch 1.10.0+cu111 * Datasets 1.18.3 * Tokenizers 0.11.0
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
[ "TAGS\n#transformers #pytorch #tensorboard #distilbert #question-answering #generated_from_trainer #dataset-squad #license-apache-2.0 #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.0" ]
token-classification
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-ingredients This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the ingredients_yes_no dataset. It achieves the following results on the evaluation set: - Loss: 0.0105 - Precision: 0.9899 - Recall: 0.9932 - F1: 0.9915 - Accuracy: 0.9978 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | No log | 1.0 | 47 | 0.2783 | 0.4 | 0.5492 | 0.4629 | 0.8910 | | No log | 2.0 | 94 | 0.1089 | 0.8145 | 0.8780 | 0.8450 | 0.9718 | | No log | 3.0 | 141 | 0.0273 | 0.9865 | 0.9932 | 0.9899 | 0.9973 | | No log | 4.0 | 188 | 0.0168 | 0.9865 | 0.9932 | 0.9899 | 0.9973 | | No log | 5.0 | 235 | 0.0156 | 0.9865 | 0.9898 | 0.9882 | 0.9957 | | No log | 6.0 | 282 | 0.0129 | 0.9865 | 0.9932 | 0.9899 | 0.9973 | | No log | 7.0 | 329 | 0.0121 | 0.9899 | 0.9932 | 0.9915 | 0.9978 | | No log | 8.0 | 376 | 0.0115 | 0.9899 | 0.9932 | 0.9915 | 0.9978 | | No log | 9.0 | 423 | 0.0108 | 0.9899 | 0.9932 | 0.9915 | 0.9978 | | No log | 10.0 | 470 | 0.0105 | 0.9899 | 0.9932 | 0.9915 | 0.9978 | ### Framework versions - Transformers 4.10.2 - Pytorch 1.9.0+cu102 - Datasets 1.11.0 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["ingredients_yes_no"], "metrics": ["precision", "recall", "f1", "accuracy"], "model-index": [{"name": "distilbert-base-uncased-finetuned-ingredients", "results": [{"task": {"type": "token-classification", "name": "Token Classification"}, "dataset": {"name": "ingredients_yes_no", "type": "ingredients_yes_no", "args": "IngredientsYesNo"}, "metrics": [{"type": "precision", "value": 0.9898648648648649, "name": "Precision"}, {"type": "recall", "value": 0.9932203389830508, "name": "Recall"}, {"type": "f1", "value": 0.9915397631133671, "name": "F1"}, {"type": "accuracy", "value": 0.9978308026030369, "name": "Accuracy"}]}]}]}
harr/distilbert-base-uncased-finetuned-ingredients
null
[ "transformers", "pytorch", "tensorboard", "distilbert", "token-classification", "generated_from_trainer", "dataset:ingredients_yes_no", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #distilbert #token-classification #generated_from_trainer #dataset-ingredients_yes_no #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #has_space #region-us
distilbert-base-uncased-finetuned-ingredients ============================================= This model is a fine-tuned version of distilbert-base-uncased on the ingredients\_yes\_no dataset. It achieves the following results on the evaluation set: * Loss: 0.0105 * Precision: 0.9899 * Recall: 0.9932 * F1: 0.9915 * Accuracy: 0.9978 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 10 ### Training results ### Framework versions * Transformers 4.10.2 * Pytorch 1.9.0+cu102 * Datasets 1.11.0 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10", "### Training results", "### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #distilbert #token-classification #generated_from_trainer #dataset-ingredients_yes_no #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10", "### Training results", "### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3" ]
null
null
Simple Sentiment Ananlysis
{}
harsh2040/sentiment_ananlysis
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
Simple Sentiment Ananlysis
[]
[ "TAGS\n#region-us \n" ]
automatic-speech-recognition
transformers
# Wav2Vec2-Large-LV60-TIMIT Fine-tuned [facebook/wav2vec2-large-lv60](https://huggingface.co/facebook/wav2vec2-large-lv60) on the [timit_asr dataset](https://huggingface.co/datasets/timit_asr). When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: ```python import soundfile as sf import torch from datasets import load_dataset from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor model_name = "hktayal345/wav2vec2-large-lv60-timit-asr" processor = Wav2Vec2Processor.from_pretrained(model_name) model = Wav2Vec2ForCTC.from_pretrained(model_name) model.eval() dataset = load_dataset("timit_asr", split="test").shuffle().select(range(10)) char_translations = str.maketrans({"-": " ", ",": "", ".": "", "?": ""}) def prepare_example(example): example["speech"], _ = sf.read(example["file"]) example["text"] = example["text"].translate(char_translations) example["text"] = " ".join(example["text"].split()) # clean up whitespaces example["text"] = example["text"].lower() return example dataset = dataset.map(prepare_example, remove_columns=["file"]) inputs = processor(dataset["speech"], sampling_rate=16000, return_tensors="pt", padding="longest") with torch.no_grad(): predicted_ids = torch.argmax(model(inputs.input_values).logits, dim=-1) predicted_ids[predicted_ids == -100] = processor.tokenizer.pad_token_id # see fine-tuning script predicted_transcripts = processor.tokenizer.batch_decode(predicted_ids) for reference, predicted in zip(dataset["text"], predicted_transcripts): print("reference:", reference) print("predicted:", predicted) print("--") ``` Here's the output: ``` reference: the emblem depicts the acropolis all aglow predicted: the amblum depicts the acropolis all a glo -- reference: don't ask me to carry an oily rag like that predicted: don't ask me to carry an oily rag like that -- reference: they enjoy it when i audition predicted: they enjoy it when i addition -- reference: set aside to dry with lid on sugar bowl predicted: set aside to dry with a litt on shoogerbowl -- reference: a boring novel is a superb sleeping pill predicted: a bor and novel is a suberb sleeping peel -- reference: only the most accomplished artists obtain popularity predicted: only the most accomplished artists obtain popularity -- reference: he has never himself done anything for which to be hated which of us has predicted: he has never himself done anything for which to be hated which of us has -- reference: the fish began to leap frantically on the surface of the small lake predicted: the fish began to leap frantically on the surface of the small lake -- reference: or certain words or rituals that child and adult go through may do the trick predicted: or certain words or rituals that child an adult go through may do the trick -- reference: are your grades higher or lower than nancy's predicted: are your grades higher or lower than nancies -- ``` ## Fine-Tuning Script You can find the script used to produce this model [here](https://colab.research.google.com/drive/1gVaZhFuIXxBDN2pD0esW490azlbQtQ7C?usp=sharing). **Note:** This model can be fine-tuned further; [trainer_state.json](https://huggingface.co/harshit345/wav2vec2-large-lv60-timit/blob/main/trainer_state.json) shows useful details, namely the last state (this checkpoint): ```json { "epoch": 29.51, "eval_loss": 25.424150466918945, "eval_runtime": 182.9499, "eval_samples_per_second": 9.183, "eval_wer": 0.1351704233095107, "step": 8500 } ```
{"language": "en", "license": "apache-2.0", "tags": ["audio", "automatic-speech-recognition", "speech"], "datasets": ["timit_asr"]}
harshit345/wav2vec2-large-lv60-timit
null
[ "transformers", "pytorch", "jax", "wav2vec2", "automatic-speech-recognition", "audio", "speech", "en", "dataset:timit_asr", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #en #dataset-timit_asr #license-apache-2.0 #endpoints_compatible #region-us
# Wav2Vec2-Large-LV60-TIMIT Fine-tuned facebook/wav2vec2-large-lv60 on the timit_asr dataset. When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: Here's the output: ## Fine-Tuning Script You can find the script used to produce this model here. Note: This model can be fine-tuned further; trainer_state.json shows useful details, namely the last state (this checkpoint):
[ "# Wav2Vec2-Large-LV60-TIMIT\n\nFine-tuned facebook/wav2vec2-large-lv60\non the timit_asr dataset.\nWhen using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\n\nThe model can be used directly (without a language model) as follows:\n\n\n\nHere's the output:", "## Fine-Tuning Script\n\nYou can find the script used to produce this model\nhere.\n\nNote: This model can be fine-tuned further;\ntrainer_state.json\nshows useful details, namely the last state (this checkpoint):" ]
[ "TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #en #dataset-timit_asr #license-apache-2.0 #endpoints_compatible #region-us \n", "# Wav2Vec2-Large-LV60-TIMIT\n\nFine-tuned facebook/wav2vec2-large-lv60\non the timit_asr dataset.\nWhen using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\n\nThe model can be used directly (without a language model) as follows:\n\n\n\nHere's the output:", "## Fine-Tuning Script\n\nYou can find the script used to produce this model\nhere.\n\nNote: This model can be fine-tuned further;\ntrainer_state.json\nshows useful details, namely the last state (this checkpoint):" ]
automatic-speech-recognition
transformers
# Wav2Vec2-Large-XLSR-53-greek Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on greek using the [Common Voice](https://huggingface.co/datasets/common_voice) and [CSS10 Greek: Single Speaker Speech Dataset](https://www.kaggle.com/bryanpark/greek-single-speaker-speech-dataset). When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: ```python import torch import torchaudio from datasets import load_dataset from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor test_dataset = load_dataset("common_voice", "el", split="test[:2%]") #TODO: replace {lang_id} in your language code here. Make sure the code is one of the *ISO codes* of [this](https://huggingface.co/languages) site. processor = Wav2Vec2Processor.from_pretrained("vasilis/wav2vec2-large-xlsr-53-greek") #TODO: replace {model_id} with your model id. The model id consists of {your_username}/{your_modelname}, *e.g.* `elgeish/wav2vec2-large-xlsr-53-arabic` model = Wav2Vec2ForCTC.from_pretrained("vasilis/wav2vec2-large-xlsr-53-greek") #TODO: replace {model_id} with your model id. The model id consists of {your_username}/{your_modelname}, *e.g.* `elgeish/wav2vec2-large-xlsr-53-arabic` resampler = torchaudio.transforms.Resample(48_000, 16_000) # Preprocessing the datasets. # We need to read the aduio files as arrays def speech_file_to_array_fn(batch): speech_array, sampling_rate = torchaudio.load(batch["path"]) batch["speech"] = resampler(speech_array).squeeze().numpy() return batch test_dataset = test_dataset.map(speech_file_to_array_fn) inputs = processor(test_dataset["speech"][:2], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): logits = model(inputs.input_values, attention_mask=inputs.attention_mask).logits predicted_ids = torch.argmax(logits, dim=-1) print("Prediction:", processor.batch_decode(predicted_ids)) print("Reference:", test_dataset["sentence"][:2]) ``` | Reference | Prediction | | ------------- | ------------- | | ΤΟ ΒΑΣΙΛΌΠΟΥΛΟ, ΠΟΥ ΜΟΙΆΖΕΙ ΛΕΟΝΤΑΡΆΚΙ ΚΑΙ ΑΕΤΟΥΔΆΚΙ | ΤΟ ΒΑΣΙΛΌΠΟΥΛΟ ΠΟΥ ΜΙΑΣΕ ΛΙΟΝΤΑΡΑΚΉ ΚΑΙ ΑΪΤΟΥΔΆΚΙ | | ΣΥΝΆΜΑ ΞΕΠΡΌΒΑΛΑΝ ΑΠΌ ΜΈΣΑ ΑΠΌ ΤΑ ΔΈΝΤΡΑ, ΔΕΞΙΆ, ΑΡΜΑΤΩΜΈΝΟΙ ΚΑΒΑΛΑΡΈΟΙ. | ΣΥΝΆΜΑ ΚΑΙ ΤΡΌΒΑΛΑΝ ΑΠΌ ΜΈΣΑ ΑΠΌ ΤΑ ΔΈΝΤΡΑ ΔΕΞΙΆ ΑΡΜΑΤΩΜΈΝΟΙ ΚΑΒΑΛΑΡΈΟΙ | | ΤΑ ΣΥΣΚΕΥΑΣΜΈΝΑ ΒΙΟΛΟΓΙΚΆ ΛΑΧΑΝΙΚΆ ΔΕΝ ΠΕΡΙΈΧΟΥΝ ΣΥΝΤΗΡΗΤΙΚΆ ΚΑΙ ΟΡΜΌΝΕΣ | ΤΑ ΣΥΣΚΕΦΑΣΜΈΝΑ ΒΙΟΛΟΓΙΚΆ ΛΑΧΑΝΙΚΆ ΔΕΝ ΠΕΡΙΈΧΟΥΝ ΣΙΔΗΡΗΤΙΚΆ ΚΑΙ ΟΡΜΌΝΕΣ | | ΑΚΟΛΟΥΘΉΣΕΤΕ ΜΕ! | ΑΚΟΛΟΥΘΉΣΤΕ ΜΕ | | ΚΑΙ ΠΟΎ ΜΠΟΡΏ ΝΑ ΤΟΝ ΒΡΩ; | Ε ΠΟΎ ΜΠΟΡΏ ΝΑ ΤΙ ΕΒΡΩ | | ΝΑΙ! ΑΠΟΚΡΊΘΗΚΕ ΤΟ ΠΑΙΔΊ | ΝΑΙ ΑΠΟΚΡΊΘΗΚΕ ΤΟ ΠΑΙΔΊ | | ΤΟ ΠΑΛΆΤΙ ΜΟΥ ΤΟ ΠΡΟΜΉΘΕΥΕ. | ΤΟ ΠΑΛΆΤΙ ΜΟΥ ΤΟ ΠΡΟΜΉΘΕΥΕ | | ΉΛΘΕ ΜΉΝΥΜΑ ΑΠΌ ΤΟ ΘΕΊΟ ΒΑΣΙΛΙΆ; | ΉΛΘΑ ΜΕΊΝΕΙ ΜΕ ΑΠΌ ΤΟ ΘΕΊΟ ΒΑΣΊΛΙΑ | | ΠΑΡΑΚΆΤΩ, ΈΝΑ ΡΥΆΚΙ ΜΟΥΡΜΟΎΡΙΖΕ ΓΛΥΚΆ, ΚΥΛΏΝΤΑΣ ΤΑ ΚΡΥΣΤΑΛΛΈΝΙΑ ΝΕΡΆ ΤΟΥ ΑΝΆΜΕΣΑ ΣΤΑ ΠΥΚΝΆ ΧΑΜΌΔΕΝΤΡΑ. | ΠΑΡΑΚΆΤΩ ΈΝΑ ΡΥΆΚΙ ΜΟΥΡΜΟΎΡΙΖΕ ΓΛΥΚΆ ΚΥΛΏΝΤΑΣ ΤΑ ΚΡΥΣΤΑΛΛΈΝΙΑ ΝΕΡΆ ΤΟΥ ΑΝΆΜΕΣΑ ΣΤΑ ΠΥΚΡΆ ΧΑΜΌΔΕΝΤΡΑ | | ΠΡΆΓΜΑΤΙ, ΕΊΝΑΙ ΑΣΤΕΊΟ ΝΑ ΠΆΡΕΙ Ο ΔΙΆΒΟΛΟΣ | ΠΡΆΓΜΑΤΗ ΕΊΝΑΙ ΑΣΤΕΊΟ ΝΑ ΠΆΡΕΙ Ο ΔΙΆΒΟΛΟΣ | ## Evaluation The model can be evaluated as follows on the greek test data of Common Voice. ```python import torch import torchaudio from datasets import load_dataset, load_metric from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor import re test_dataset = load_dataset("common_voice", "el", split="test") #TODO: replace {lang_id} in your language code here. Make sure the code is one of the *ISO codes* of [this](https://huggingface.co/languages) site. wer = load_metric("wer") processor = Wav2Vec2Processor.from_pretrained("vasilis/wav2vec2-large-xlsr-53-greek") #TODO: replace {model_id} with your model id. The model id consists of {your_username}/{your_modelname}, *e.g.* `elgeish/wav2vec2-large-xlsr-53-arabic` model = Wav2Vec2ForCTC.from_pretrained("vasilis/wav2vec2-large-xlsr-53-greek") #TODO: replace {model_id} with your model id. The model id consists of {your_username}/{your_modelname}, *e.g.* `elgeish/wav2vec2-large-xlsr-53-arabic` model.to("cuda") chars_to_ignore_regex = '[\,\?\.\!\-\;\:\"\“]' # TODO: adapt this list to include all special characters you removed from the data normalize_greek_letters = {"ς": "σ"} # normalize_greek_letters = {"ά": "α", "έ": "ε", "ί": "ι", 'ϊ': "ι", "ύ": "υ", "ς": "σ", "ΐ": "ι", 'ϋ': "υ", "ή": "η", "ώ": "ω", 'ό': "ο"} remove_chars_greek = {"a": "", "h": "", "n": "", "g": "", "o": "", "v": "", "e": "", "r": "", "t": "", "«": "", "»": "", "m": "", '́': '', "·": "", "’": "", '´': ""} replacements = {**normalize_greek_letters, **remove_chars_greek} resampler = { 48_000: torchaudio.transforms.Resample(48_000, 16_000), 44100: torchaudio.transforms.Resample(44100, 16_000), 32000: torchaudio.transforms.Resample(32000, 16_000) } # Preprocessing the datasets. # We need to read the aduio files as arrays def speech_file_to_array_fn(batch): batch["sentence"] = re.sub(chars_to_ignore_regex, '', batch["sentence"]).lower() for key, value in replacements.items(): batch["sentence"] = batch["sentence"].replace(key, value) speech_array, sampling_rate = torchaudio.load(batch["path"]) batch["speech"] = resampler[sampling_rate](speech_array).squeeze().numpy() return batch test_dataset = test_dataset.map(speech_file_to_array_fn) # Preprocessing the datasets. # We need to read the aduio files as arrays def evaluate(batch): inputs = processor(batch["speech"], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): logits = model(inputs.input_values.to("cuda"), attention_mask=inputs.attention_mask.to("cuda")).logits pred_ids = torch.argmax(logits, dim=-1) batch["pred_strings"] = processor.batch_decode(pred_ids) return batch result = test_dataset.map(evaluate, batched=True, batch_size=8) print("WER: {:2f}".format(100 * wer.compute(predictions=result["pred_strings"], references=result["sentence"]))) print("CER: {:2f}".format(100 * wer.compute(predictions=[" ".join(list(entry)) for entry in result["pred_strings"]], references=[" ".join(list(entry)) for entry in result["sentence"]]))) ``` **Test Result**: 18.996669 % ## Training The Common Voice train dataset was used for training. Also all of `CSS10 Greek` was used using the normalized transcripts. During text preprocessing letter `ς` is normalized to `σ` the reason is that both letters sound the same with `ς` only used as the ending character of words. So, the change can be mapped up to proper dictation easily. I tried removing all accents from letters as well that improved `WER` significantly. The model was reaching `17%` WER easily without having converged. However, the text preprocessing needed to do after to fix transcrtiptions would be more complicated. A language model should fix things easily though. Another thing that could be tried out would be to change all of `ι`, `η` ... etc to a single character since all sound the same. similar for `o` and `ω` these should help the acoustic model part significantly since all these characters map to the same sound. But further text normlization would be needed.
{"language": "el", "license": "apache-2.0", "tags": ["audio", "automatic-speech-recognition", "speech", "xlsr-fine-tuning-week"], "datasets": ["common_voice"], "metrics": ["wer", "cer"], "model-index": [{"name": "V XLSR Wav2Vec2 Large 53 - greek", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Speech Recognition"}, "dataset": {"name": "Common Voice el", "type": "common_voice", "args": "el"}, "metrics": [{"type": "wer", "value": 18.996669, "name": "Test WER"}, {"type": "cer", "value": 5.781874, "name": "Test CER"}]}]}]}
harshit345/xlsr-53-wav2vec-greek
null
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "audio", "speech", "xlsr-fine-tuning-week", "el", "dataset:common_voice", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "el" ]
TAGS #transformers #pytorch #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #el #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us
Wav2Vec2-Large-XLSR-53-greek ============================ Fine-tuned facebook/wav2vec2-large-xlsr-53 on greek using the Common Voice and CSS10 Greek: Single Speaker Speech Dataset. When using this model, make sure that your speech input is sampled at 16kHz. Usage ----- The model can be used directly (without a language model) as follows: Evaluation ---------- The model can be evaluated as follows on the greek test data of Common Voice. Test Result: 18.996669 % Training -------- The Common Voice train dataset was used for training. Also all of 'CSS10 Greek' was used using the normalized transcripts. During text preprocessing letter 'ς' is normalized to 'σ' the reason is that both letters sound the same with 'ς' only used as the ending character of words. So, the change can be mapped up to proper dictation easily. I tried removing all accents from letters as well that improved 'WER' significantly. The model was reaching '17%' WER easily without having converged. However, the text preprocessing needed to do after to fix transcrtiptions would be more complicated. A language model should fix things easily though. Another thing that could be tried out would be to change all of 'ι', 'η' ... etc to a single character since all sound the same. similar for 'o' and 'ω' these should help the acoustic model part significantly since all these characters map to the same sound. But further text normlization would be needed.
[]
[ "TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #el #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us \n" ]
automatic-speech-recognition
transformers
# Wav2Vec2-Large-XLSR-53-hindi Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) hindi using the [Multilingual and code-switching ASR challenges for low resource Indian languages](https://navana-tech.github.io/IS21SS-indicASRchallenge/data.html). When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: ```python import torch import torchaudio from datasets import load_dataset from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor test_dataset = load_dataset("common_voice", "hi", split="test[:2%]") processor = Wav2Vec2Processor.from_pretrained("theainerd/Wav2Vec2-large-xlsr-hindi") model = Wav2Vec2ForCTC.from_pretrained("theainerd/Wav2Vec2-large-xlsr-hindi") resampler = torchaudio.transforms.Resample(48_000, 16_000) # Preprocessing the datasets. # We need to read the aduio files as arrays def speech_file_to_array_fn(batch): speech_array, sampling_rate = torchaudio.load(batch["path"]) batch["speech"] = resampler(speech_array).squeeze().numpy() return batch test_dataset = test_dataset.map(speech_file_to_array_fn) inputs = processor(test_dataset["speech"][:2], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): logits = model(inputs.input_values, attention_mask=inputs.attention_mask).logits predicted_ids = torch.argmax(logits, dim=-1) print("Prediction:", processor.batch_decode(predicted_ids)) print("Reference:", test_dataset["sentence"][:2]) ``` ## Evaluation The model can be evaluated as follows on the hindi test data of Common Voice. ```python import torch import torchaudio from datasets import load_dataset, load_metric from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor import re test_dataset = load_dataset("common_voice", "hi", split="test") wer = load_metric("wer") processor = Wav2Vec2Processor.from_pretrained("theainerd/Wav2Vec2-large-xlsr-hindi") model = Wav2Vec2ForCTC.from_pretrained("theainerd/Wav2Vec2-large-xlsr-hindi") model.to("cuda") resampler = torchaudio.transforms.Resample(48_000, 16_000) chars_to_ignore_regex = '[\,\?\.\!\-\;\:\"\“]' # Preprocessing the datasets. # We need to read the aduio files as arrays def speech_file_to_array_fn(batch): batch["sentence"] = re.sub(chars_to_ignore_regex, '', batch["sentence"]).lower() speech_array, sampling_rate = torchaudio.load(batch["path"]) batch["speech"] = resampler(speech_array).squeeze().numpy() return batch test_dataset = test_dataset.map(speech_file_to_array_fn) # Preprocessing the datasets. # We need to read the aduio files as arrays def evaluate(batch): inputs = processor(batch["speech"], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): logits = model(inputs.input_values.to("cuda"), attention_mask=inputs.attention_mask.to("cuda")).logits pred_ids = torch.argmax(logits, dim=-1) batch["pred_strings"] = processor.batch_decode(pred_ids) return batch result = test_dataset.map(evaluate, batched=True, batch_size=8) print("WER: {:2f}".format(100 * wer.compute(predictions=result["pred_strings"], references=result["sentence"]))) ``` **Test Result**:20.22 % ## Training The script used for training can be found [Hindi ASR Fine Tuning Wav2Vec2](https://colab.research.google.com/drive/1nY5WMj1oNlexD_qDeNYL7ZM427A021CV?usp=sharing)
{"language": "hi", "license": "apache-2.0", "tags": ["audio", "automatic-speech-recognition", "speech", "xlsr-fine-tuning-week"], "datasets": ["Interspeech 2021"], "metrics": ["wer"], "model-index": [{"name": "XLSR Wav2Vec2 Hindi by Shyam Sunder Kumar", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Speech Recognition"}, "dataset": {"name": "Common Voice hi", "type": "common_voice", "args": "hi"}, "metrics": [{"type": "wer", "value": 20.22, "name": "Test WER"}]}]}]}
harshit345/xlsr-53-wav2vec-hi
null
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "audio", "speech", "xlsr-fine-tuning-week", "hi", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "hi" ]
TAGS #transformers #pytorch #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #hi #license-apache-2.0 #model-index #endpoints_compatible #region-us
# Wav2Vec2-Large-XLSR-53-hindi Fine-tuned facebook/wav2vec2-large-xlsr-53 hindi using the Multilingual and code-switching ASR challenges for low resource Indian languages. When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows: ## Evaluation The model can be evaluated as follows on the hindi test data of Common Voice. Test Result:20.22 % ## Training The script used for training can be found Hindi ASR Fine Tuning Wav2Vec2
[ "# Wav2Vec2-Large-XLSR-53-hindi\n\nFine-tuned facebook/wav2vec2-large-xlsr-53 hindi using the Multilingual and code-switching ASR challenges for low resource Indian languages.\nWhen using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\n\nThe model can be used directly (without a language model) as follows:", "## Evaluation\n\nThe model can be evaluated as follows on the hindi test data of Common Voice. \n\n\n\n\nTest Result:20.22 %", "## Training\n\nThe script used for training can be found Hindi ASR Fine Tuning Wav2Vec2" ]
[ "TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #hi #license-apache-2.0 #model-index #endpoints_compatible #region-us \n", "# Wav2Vec2-Large-XLSR-53-hindi\n\nFine-tuned facebook/wav2vec2-large-xlsr-53 hindi using the Multilingual and code-switching ASR challenges for low resource Indian languages.\nWhen using this model, make sure that your speech input is sampled at 16kHz.", "## Usage\n\nThe model can be used directly (without a language model) as follows:", "## Evaluation\n\nThe model can be evaluated as follows on the hindi test data of Common Voice. \n\n\n\n\nTest Result:20.22 %", "## Training\n\nThe script used for training can be found Hindi ASR Fine Tuning Wav2Vec2" ]
audio-classification
transformers
~~~ # requirement packages !pip install git+https://github.com/huggingface/datasets.git !pip install git+https://github.com/huggingface/transformers.git !pip install torchaudio !pip install librosa ~~~ # prediction ~~~ import torch import torch.nn as nn import torch.nn.functional as F import torchaudio from transformers import AutoConfig, Wav2Vec2FeatureExtractor import librosa import IPython.display as ipd import numpy as np import pandas as pd ~~~ ~~~ device = torch.device("cuda" if torch.cuda.is_available() else "cpu") model_name_or_path = "harshit345/xlsr-wav2vec-speech-emotion-recognition" config = AutoConfig.from_pretrained(model_name_or_path) feature_extractor = Wav2Vec2FeatureExtractor.from_pretrained(model_name_or_path) sampling_rate = feature_extractor.sampling_rate model = Wav2Vec2ForSpeechClassification.from_pretrained(model_name_or_path).to(device) ~~~ ~~~ def speech_file_to_array_fn(path, sampling_rate): speech_array, _sampling_rate = torchaudio.load(path) resampler = torchaudio.transforms.Resample(_sampling_rate) speech = resampler(speech_array).squeeze().numpy() return speech def predict(path, sampling_rate): speech = speech_file_to_array_fn(path, sampling_rate) inputs = feature_extractor(speech, sampling_rate=sampling_rate, return_tensors="pt", padding=True) inputs = {key: inputs[key].to(device) for key in inputs} with torch.no_grad(): logits = model(**inputs).logits scores = F.softmax(logits, dim=1).detach().cpu().numpy()[0] outputs = [{"Emotion": config.id2label[i], "Score": f"{round(score * 100, 3):.1f}%"} for i, score in enumerate(scores)] return outputs ~~~ # prediction ~~~ # path for a sample path = '/data/jtes_v1.1/wav/f01/ang/f01_ang_01.wav' outputs = predict(path, sampling_rate) ~~~ ~~~ [{'Emotion': 'anger', 'Score': '78.3%'}, {'Emotion': 'disgust', 'Score': '11.7%'}, {'Emotion': 'fear', 'Score': '5.4%'}, {'Emotion': 'happiness', 'Score': '4.1%'}, {'Emotion': 'sadness', 'Score': '0.5%'}] ~~~ ## Evaluation The following tables summarize the scores obtained by model overall and per each class. | Emotions | precision | recall | f1-score | accuracy | |-----------|-----------|--------|----------|----------| | anger | 0.82 | 1.00 | 0.81 | | | disgust | 0.85 | 0.96 | 0.85 | | | fear | 0.78 | 0.88 | 0.80 | | | happiness | 0.84 | 0.71 | 0.78 | | | sadness | 0.86 | 1.00 | 0.79 | | | | | | Overall | 0.806 | ## Colab Notebook https://colab.research.google.com/drive/1aPPb_ZVS5dlFVZySly8Q80a44La1XjJu?usp=sharing
{"language": "en", "license": "apache-2.0", "tags": ["audio", "audio-classification", "speech"], "datasets": ["aesdd"]}
harshit345/xlsr-wav2vec-speech-emotion-recognition
null
[ "transformers", "pytorch", "wav2vec2", "audio", "audio-classification", "speech", "en", "dataset:aesdd", "license:apache-2.0", "endpoints_compatible", "has_space", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #wav2vec2 #audio #audio-classification #speech #en #dataset-aesdd #license-apache-2.0 #endpoints_compatible #has_space #region-us
``` # requirement packages !pip install git+URL !pip install git+URL !pip install torchaudio !pip install librosa ``` prediction ========== ``` import torch import URL as nn import URL.functional as F import torchaudio from transformers import AutoConfig, Wav2Vec2FeatureExtractor import librosa import IPython.display as ipd import numpy as np import pandas as pd ``` ``` device = URL("cuda" if URL.is_available() else "cpu") model_name_or_path = "harshit345/xlsr-wav2vec-speech-emotion-recognition" config = AutoConfig.from_pretrained(model_name_or_path) feature_extractor = Wav2Vec2FeatureExtractor.from_pretrained(model_name_or_path) sampling_rate = feature_extractor.sampling_rate model = Wav2Vec2ForSpeechClassification.from_pretrained(model_name_or_path).to(device) ``` ``` def speech_file_to_array_fn(path, sampling_rate): speech_array, _sampling_rate = URL(path) resampler = torchaudio.transforms.Resample(_sampling_rate) speech = resampler(speech_array).squeeze().numpy() return speech def predict(path, sampling_rate): speech = speech_file_to_array_fn(path, sampling_rate) inputs = feature_extractor(speech, sampling_rate=sampling_rate, return_tensors="pt", padding=True) inputs = {key: inputs[key].to(device) for key in inputs} with torch.no_grad(): logits = model(inputs).logits scores = F.softmax(logits, dim=1).detach().cpu().numpy()[0] outputs = [{"Emotion": config.id2label[i], "Score": f"{round(score * 100, 3):.1f}%"} for i, score in enumerate(scores)] return outputs ``` prediction ========== ``` # path for a sample path = '/data/jtes_v1.1/wav/f01/ang/f01_ang_01.wav' outputs = predict(path, sampling_rate) ``` ``` [{'Emotion': 'anger', 'Score': '78.3%'}, {'Emotion': 'disgust', 'Score': '11.7%'}, {'Emotion': 'fear', 'Score': '5.4%'}, {'Emotion': 'happiness', 'Score': '4.1%'}, {'Emotion': 'sadness', 'Score': '0.5%'}] ``` Evaluation ---------- The following tables summarize the scores obtained by model overall and per each class. Colab Notebook URL
[ "# requirement packages\n!pip install git+URL\n!pip install git+URL\n!pip install torchaudio\n!pip install librosa\n\n\n```\n\nprediction\n==========\n\n\n\n```\nimport torch\nimport URL as nn\nimport URL.functional as F\nimport torchaudio\nfrom transformers import AutoConfig, Wav2Vec2FeatureExtractor\nimport librosa\nimport IPython.display as ipd\nimport numpy as np\nimport pandas as pd\n\n```\n\n\n```\ndevice = URL(\"cuda\" if URL.is_available() else \"cpu\")\nmodel_name_or_path = \"harshit345/xlsr-wav2vec-speech-emotion-recognition\"\nconfig = AutoConfig.from_pretrained(model_name_or_path)\nfeature_extractor = Wav2Vec2FeatureExtractor.from_pretrained(model_name_or_path)\nsampling_rate = feature_extractor.sampling_rate\nmodel = Wav2Vec2ForSpeechClassification.from_pretrained(model_name_or_path).to(device)\n\n```\n\n\n```\ndef speech_file_to_array_fn(path, sampling_rate):\n speech_array, _sampling_rate = URL(path)\n resampler = torchaudio.transforms.Resample(_sampling_rate)\n speech = resampler(speech_array).squeeze().numpy()\n return speech\ndef predict(path, sampling_rate):\n speech = speech_file_to_array_fn(path, sampling_rate)\n inputs = feature_extractor(speech, sampling_rate=sampling_rate, return_tensors=\"pt\", padding=True)\n inputs = {key: inputs[key].to(device) for key in inputs}\n with torch.no_grad():\n logits = model(inputs).logits\n scores = F.softmax(logits, dim=1).detach().cpu().numpy()[0]\n outputs = [{\"Emotion\": config.id2label[i], \"Score\": f\"{round(score * 100, 3):.1f}%\"} for i, score in enumerate(scores)]\n return outputs\n\n```\n\nprediction\n==========\n\n\n\n```", "# path for a sample\npath = '/data/jtes_v1.1/wav/f01/ang/f01_ang_01.wav' \noutputs = predict(path, sampling_rate)\n\n```\n\n\n```\n[{'Emotion': 'anger', 'Score': '78.3%'},\n {'Emotion': 'disgust', 'Score': '11.7%'},\n {'Emotion': 'fear', 'Score': '5.4%'},\n {'Emotion': 'happiness', 'Score': '4.1%'},\n {'Emotion': 'sadness', 'Score': '0.5%'}]\n\n```\n\nEvaluation\n----------\n\n\nThe following tables summarize the scores obtained by model overall and per each class.\n\n\n\n\nColab Notebook\nURL" ]
[ "TAGS\n#transformers #pytorch #wav2vec2 #audio #audio-classification #speech #en #dataset-aesdd #license-apache-2.0 #endpoints_compatible #has_space #region-us \n", "# requirement packages\n!pip install git+URL\n!pip install git+URL\n!pip install torchaudio\n!pip install librosa\n\n\n```\n\nprediction\n==========\n\n\n\n```\nimport torch\nimport URL as nn\nimport URL.functional as F\nimport torchaudio\nfrom transformers import AutoConfig, Wav2Vec2FeatureExtractor\nimport librosa\nimport IPython.display as ipd\nimport numpy as np\nimport pandas as pd\n\n```\n\n\n```\ndevice = URL(\"cuda\" if URL.is_available() else \"cpu\")\nmodel_name_or_path = \"harshit345/xlsr-wav2vec-speech-emotion-recognition\"\nconfig = AutoConfig.from_pretrained(model_name_or_path)\nfeature_extractor = Wav2Vec2FeatureExtractor.from_pretrained(model_name_or_path)\nsampling_rate = feature_extractor.sampling_rate\nmodel = Wav2Vec2ForSpeechClassification.from_pretrained(model_name_or_path).to(device)\n\n```\n\n\n```\ndef speech_file_to_array_fn(path, sampling_rate):\n speech_array, _sampling_rate = URL(path)\n resampler = torchaudio.transforms.Resample(_sampling_rate)\n speech = resampler(speech_array).squeeze().numpy()\n return speech\ndef predict(path, sampling_rate):\n speech = speech_file_to_array_fn(path, sampling_rate)\n inputs = feature_extractor(speech, sampling_rate=sampling_rate, return_tensors=\"pt\", padding=True)\n inputs = {key: inputs[key].to(device) for key in inputs}\n with torch.no_grad():\n logits = model(inputs).logits\n scores = F.softmax(logits, dim=1).detach().cpu().numpy()[0]\n outputs = [{\"Emotion\": config.id2label[i], \"Score\": f\"{round(score * 100, 3):.1f}%\"} for i, score in enumerate(scores)]\n return outputs\n\n```\n\nprediction\n==========\n\n\n\n```", "# path for a sample\npath = '/data/jtes_v1.1/wav/f01/ang/f01_ang_01.wav' \noutputs = predict(path, sampling_rate)\n\n```\n\n\n```\n[{'Emotion': 'anger', 'Score': '78.3%'},\n {'Emotion': 'disgust', 'Score': '11.7%'},\n {'Emotion': 'fear', 'Score': '5.4%'},\n {'Emotion': 'happiness', 'Score': '4.1%'},\n {'Emotion': 'sadness', 'Score': '0.5%'}]\n\n```\n\nEvaluation\n----------\n\n\nThe following tables summarize the scores obtained by model overall and per each class.\n\n\n\n\nColab Notebook\nURL" ]
automatic-speech-recognition
transformers
# Wav2vec2-Large-English Fine-tuned [facebook/wav2vec2-large](https://huggingface.co/facebook/wav2vec2-large) on English using the [Common Voice](https://huggingface.co/datasets/common_voice). When using this model, make sure that your speech input is sampled at 16kHz. ## Usage The model can be used directly (without a language model) as follows... Using the [ASRecognition](https://github.com/jonatasgrosman/asrecognition) library: ```python from asrecognition import ASREngine asr = ASREngine("fr", model_path="jonatasgrosman/wav2vec2-large-english") audio_paths = ["/path/to/file.mp3", "/path/to/another_file.wav"] transcriptions = asr.transcribe(audio_paths) ``` Writing your own inference script: ```python import torch import librosa from datasets import load_dataset from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor LANG_ID = "en" MODEL_ID = "jonatasgrosman/wav2vec2-large-english" SAMPLES = 10 test_dataset = load_dataset("common_voice", LANG_ID, split=f"test[:{SAMPLES}]") processor = Wav2Vec2Processor.from_pretrained(MODEL_ID) model = Wav2Vec2ForCTC.from_pretrained(MODEL_ID) # Preprocessing the datasets. # We need to read the audio files as arrays def speech_file_to_array_fn(batch): speech_array, sampling_rate = librosa.load(batch["path"], sr=16_000) batch["speech"] = speech_array batch["sentence"] = batch["sentence"].upper() return batch test_dataset = test_dataset.map(speech_file_to_array_fn) inputs = processor(test_dataset["speech"], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): logits = model(inputs.input_values, attention_mask=inputs.attention_mask).logits predicted_ids = torch.argmax(logits, dim=-1) predicted_sentences = processor.batch_decode(predicted_ids) for i, predicted_sentence in enumerate(predicted_sentences): print("-" * 100) print("Reference:", test_dataset[i]["sentence"]) print("Prediction:", predicted_sentence) ``` | Reference | Prediction | | ------------- | ------------- | | "SHE'LL BE ALL RIGHT." | SHELL BE ALL RIGHT | | SIX | SIX | | "ALL'S WELL THAT ENDS WELL." | ALLAS WELL THAT ENDS WELL | | DO YOU MEAN IT? | W MEAN IT | | THE NEW PATCH IS LESS INVASIVE THAN THE OLD ONE, BUT STILL CAUSES REGRESSIONS. | THE NEW PATCH IS LESS INVASIVE THAN THE OLD ONE BUT STILL CAUSES REGRESTION | | HOW IS MOZILLA GOING TO HANDLE AMBIGUITIES LIKE QUEUE AND CUE? | HOW IS MOSILLA GOING TO BANDL AND BE WHIT IS LIKE QU AND QU | | "I GUESS YOU MUST THINK I'M KINDA BATTY." | RUSTION AS HAME AK AN THE POT | | NO ONE NEAR THE REMOTE MACHINE YOU COULD RING? | NO ONE NEAR THE REMOTE MACHINE YOU COULD RING | | SAUCE FOR THE GOOSE IS SAUCE FOR THE GANDER. | SAUCE FOR THE GUCE IS SAUCE FOR THE GONDER | | GROVES STARTED WRITING SONGS WHEN SHE WAS FOUR YEARS OLD. | GRAFS STARTED WRITING SONGS WHEN SHE WAS FOUR YEARS OLD | ## Evaluation The model can be evaluated as follows on the English (en) test data of Common Voice. ```python import torch import re import librosa from datasets import load_dataset, load_metric from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor LANG_ID = "en" MODEL_ID = "jonatasgrosman/wav2vec2-large-english" DEVICE = "cuda" CHARS_TO_IGNORE = [",", "?", "¿", ".", "!", "¡", ";", ";", ":", '""', "%", '"', "�", "ʿ", "·", "჻", "~", "՞", "؟", "،", "।", "॥", "«", "»", "„", "“", "”", "「", "」", "‘", "’", "《", "》", "(", ")", "[", "]", "{", "}", "=", "`", "_", "+", "<", ">", "…", "–", "°", "´", "ʾ", "‹", "›", "©", "®", "—", "→", "。", "、", "﹂", "﹁", "‧", "~", "﹏", ",", "{", "}", "(", ")", "[", "]", "【", "】", "‥", "〽", "『", "』", "〝", "〟", "⟨", "⟩", "〜", ":", "!", "?", "♪", "؛", "/", "\\", "º", "−", "^", "ʻ", "ˆ"] test_dataset = load_dataset("common_voice", LANG_ID, split="test") wer = load_metric("wer.py") # https://github.com/jonatasgrosman/wav2vec2-sprint/blob/main/wer.py cer = load_metric("cer.py") # https://github.com/jonatasgrosman/wav2vec2-sprint/blob/main/cer.py chars_to_ignore_regex = f"[{re.escape(''.join(CHARS_TO_IGNORE))}]" processor = Wav2Vec2Processor.from_pretrained(MODEL_ID) model = Wav2Vec2ForCTC.from_pretrained(MODEL_ID) model.to(DEVICE) # Preprocessing the datasets. # We need to read the audio files as arrays def speech_file_to_array_fn(batch): with warnings.catch_warnings(): warnings.simplefilter("ignore") speech_array, sampling_rate = librosa.load(batch["path"], sr=16_000) batch["speech"] = speech_array batch["sentence"] = re.sub(chars_to_ignore_regex, "", batch["sentence"]).upper() return batch test_dataset = test_dataset.map(speech_file_to_array_fn) # Preprocessing the datasets. # We need to read the audio files as arrays def evaluate(batch): inputs = processor(batch["speech"], sampling_rate=16_000, return_tensors="pt", padding=True) with torch.no_grad(): logits = model(inputs.input_values.to(DEVICE), attention_mask=inputs.attention_mask.to(DEVICE)).logits pred_ids = torch.argmax(logits, dim=-1) batch["pred_strings"] = processor.batch_decode(pred_ids) return batch result = test_dataset.map(evaluate, batched=True, batch_size=8) predictions = [x.upper() for x in result["pred_strings"]] references = [x.upper() for x in result["sentence"]] print(f"WER: {wer.compute(predictions=predictions, references=references, chunk_size=1000) * 100}") print(f"CER: {cer.compute(predictions=predictions, references=references, chunk_size=1000) * 100}") ``` **Test Result**: In the table below I report the Word Error Rate (WER) and the Character Error Rate (CER) of the model. I ran the evaluation script described above on other models as well. Note that the table below may show different results from those already reported, this may have been caused due to some specificity of the other evaluation scripts used. | Model | WER | CER | | ------------- | ------------- | ------------- | | wav2vec2-large-xlsr-53-english | **18.98%** | **8.29%** | | wav2vec2-large-xlsr-53-greek | 18.99% | 10.60% | | wav2vec2-large-xlsr-53-hindi | 20.01% | 9.66% | | wav2vec2-large-960h-lv60-english | 22.03% | 10.39% | | wav2vec2-base-100h-lv60-english | 24.97% | 11.14% | |
{"language": "en", "license": "apache-2.0", "tags": ["audio", "automatic-speech-recognition", "speech", "xlsr-fine-tuning-week"], "datasets": ["common_voice"], "metrics": ["wer", "cer"], "model-index": [{"name": "Wav2Vec2 English by Jonatas Grosman", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Speech Recognition"}, "dataset": {"name": "Common Voice en", "type": "common_voice", "args": "en"}, "metrics": [{"type": "wer", "value": 21.53, "name": "Test WER"}, {"type": "cer", "value": 9.66, "name": "Test CER"}]}]}]}
harshit345/xlsr_wav2vec_english
null
[ "transformers", "pytorch", "wav2vec2", "automatic-speech-recognition", "audio", "speech", "xlsr-fine-tuning-week", "en", "dataset:common_voice", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #en #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us
Wav2vec2-Large-English ====================== Fine-tuned facebook/wav2vec2-large on English using the Common Voice. When using this model, make sure that your speech input is sampled at 16kHz. Usage ----- The model can be used directly (without a language model) as follows... Using the ASRecognition library: Writing your own inference script: Evaluation ---------- The model can be evaluated as follows on the English (en) test data of Common Voice. Test Result: In the table below I report the Word Error Rate (WER) and the Character Error Rate (CER) of the model. I ran the evaluation script described above on other models as well. Note that the table below may show different results from those already reported, this may have been caused due to some specificity of the other evaluation scripts used. Model: wav2vec2-large-xlsr-53-english, WER: 18.98%, CER: 8.29% Model: wav2vec2-large-xlsr-53-greek, WER: 18.99%, CER: 10.60% Model: wav2vec2-large-xlsr-53-hindi, WER: 20.01%, CER: 9.66% Model: wav2vec2-large-960h-lv60-english, WER: 22.03%, CER: 10.39% Model: wav2vec2-base-100h-lv60-english, WER: 24.97%, CER: 11.14% Model: , WER: , CER:
[]
[ "TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #en #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #region-us \n" ]
fill-mask
transformers
## EsperBERTo: RoBERTa-like Language model trained on Esperanto
{"language": "eo", "thumbnail": "https://huggingface.co/blog/assets/01_how-to-train/EsperBERTo-thumbnail-v2.png", "widget": [{"text": "\u0108u vi paloras la <mask> Esperanto?"}]}
hashk1/EsperBERTo-malgranda
null
[ "transformers", "pytorch", "jax", "roberta", "fill-mask", "eo", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "eo" ]
TAGS #transformers #pytorch #jax #roberta #fill-mask #eo #autotrain_compatible #endpoints_compatible #region-us
## EsperBERTo: RoBERTa-like Language model trained on Esperanto
[ "## EsperBERTo: RoBERTa-like Language model trained on Esperanto" ]
[ "TAGS\n#transformers #pytorch #jax #roberta #fill-mask #eo #autotrain_compatible #endpoints_compatible #region-us \n", "## EsperBERTo: RoBERTa-like Language model trained on Esperanto" ]
token-classification
transformers
# Arabic Named Entity Recognition Model Pretrained BERT-based ([arabic-bert-base](https://huggingface.co/asafaya/bert-base-arabic)) Named Entity Recognition model for Arabic. The pre-trained model can recognize the following entities: 1. **PERSON** - و هذا ما نفاه المعاون السياسي للرئيس ***نبيه بري*** ، النائب ***علي حسن خليل*** - لكن أوساط ***الحريري*** تعتبر أنه ضحى كثيرا في سبيل البلد - و ستفقد الملكة ***إليزابيث الثانية*** بذلك سيادتها على واحدة من آخر ممالك الكومنولث 2. **ORGANIZATION** - حسب أرقام ***البنك الدولي*** - أعلن ***الجيش العراقي*** - و نقلت وكالة ***رويترز*** عن ثلاثة دبلوماسيين في ***الاتحاد الأوروبي*** ، أن ***بلجيكا*** و ***إيرلندا*** و ***لوكسمبورغ*** تريد أيضاً مناقشة - ***الحكومة الاتحادية*** و ***حكومة إقليم كردستان*** - و هو ما يثير الشكوك حول مشاركة النجم البرتغالي في المباراة المرتقبة أمام ***برشلونة*** الإسباني في 3. ***LOCATION*** - الجديد هو تمكين اللاجئين من “ مغادرة الجزيرة تدريجياً و بهدوء إلى ***أثينا*** ” - ***جزيرة ساكيز*** تبعد 1 كم عن ***إزمير*** 4. **DATE** - ***غدا الجمعة*** - ***06 أكتوبر 2020*** - ***العام السابق*** 5. **PRODUCT** - عبر حسابه ب ***تطبيق “ إنستغرام ”*** - الجيل الثاني من ***نظارة الواقع الافتراضي أوكولوس كويست*** تحت اسم " ***أوكولوس كويست 2*** " 6. **COMPETITION** - عدم المشاركة في ***بطولة فرنسا المفتوحة للتنس*** - في مباراة ***كأس السوبر الأوروبي*** 7. **PRIZE** - ***جائزة نوبل ل لآداب*** - الذي فاز ب ***جائزة “ إيمي ” لأفضل دور مساند*** 8. **EVENT** - تسجّل أغنية جديدة خاصة ب ***العيد الوطني السعودي*** - ***مهرجان المرأة يافوية*** في دورته الرابعة 9. **DISEASE** - في مكافحة فيروس ***كورونا*** و عدد من الأمراض - الأزمات المشابهة مثل “ ***انفلونزا الطيور*** ” و ” ***انفلونزا الخنازير*** ## Example [Find here a complete example to use this model](https://github.com/hatmimoha/arabic-ner) ## Training Corpus The training corpus is made of 378.000 tokens (14.000 sentences) collected from the Web and annotated manually. ## Results The results on a valid corpus made of 30.000 tokens shows an F-measure of ~87%.
{"language": "ar"}
hatmimoha/arabic-ner
null
[ "transformers", "pytorch", "tf", "jax", "safetensors", "bert", "token-classification", "ar", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "ar" ]
TAGS #transformers #pytorch #tf #jax #safetensors #bert #token-classification #ar #autotrain_compatible #endpoints_compatible #has_space #region-us
# Arabic Named Entity Recognition Model Pretrained BERT-based (arabic-bert-base) Named Entity Recognition model for Arabic. The pre-trained model can recognize the following entities: 1. PERSON - و هذا ما نفاه المعاون السياسي للرئيس *نبيه بري* ، النائب *علي حسن خليل* - لكن أوساط *الحريري* تعتبر أنه ضحى كثيرا في سبيل البلد - و ستفقد الملكة *إليزابيث الثانية* بذلك سيادتها على واحدة من آخر ممالك الكومنولث 2. ORGANIZATION - حسب أرقام *البنك الدولي* - أعلن *الجيش العراقي* - و نقلت وكالة *رويترز* عن ثلاثة دبلوماسيين في *الاتحاد الأوروبي* ، أن *بلجيكا* و *إيرلندا* و *لوكسمبورغ* تريد أيضاً مناقشة - *الحكومة الاتحادية* و *حكومة إقليم كردستان* - و هو ما يثير الشكوك حول مشاركة النجم البرتغالي في المباراة المرتقبة أمام *برشلونة* الإسباني في 3. *LOCATION* - الجديد هو تمكين اللاجئين من “ مغادرة الجزيرة تدريجياً و بهدوء إلى *أثينا* ” - *جزيرة ساكيز* تبعد 1 كم عن *إزمير* 4. DATE - *غدا الجمعة* - *06 أكتوبر 2020* - *العام السابق* 5. PRODUCT - عبر حسابه ب *تطبيق “ إنستغرام ”* - الجيل الثاني من *نظارة الواقع الافتراضي أوكولوس كويست* تحت اسم " *أوكولوس كويست 2* " 6. COMPETITION - عدم المشاركة في *بطولة فرنسا المفتوحة للتنس* - في مباراة *كأس السوبر الأوروبي* 7. PRIZE - *جائزة نوبل ل لآداب* - الذي فاز ب *جائزة “ إيمي ” لأفضل دور مساند* 8. EVENT - تسجّل أغنية جديدة خاصة ب *العيد الوطني السعودي* - *مهرجان المرأة يافوية* في دورته الرابعة 9. DISEASE - في مكافحة فيروس *كورونا* و عدد من الأمراض - الأزمات المشابهة مثل “ *انفلونزا الطيور* ” و ” *انفلونزا الخنازير* ## Example Find here a complete example to use this model ## Training Corpus The training corpus is made of 378.000 tokens (14.000 sentences) collected from the Web and annotated manually. ## Results The results on a valid corpus made of 30.000 tokens shows an F-measure of ~87%.
[ "# Arabic Named Entity Recognition Model\n\nPretrained BERT-based (arabic-bert-base) Named Entity Recognition model for Arabic.\n\nThe pre-trained model can recognize the following entities:\n1. PERSON\n\n- و هذا ما نفاه المعاون السياسي للرئيس *نبيه بري* ، النائب *علي حسن خليل* \n\n- لكن أوساط *الحريري* تعتبر أنه ضحى كثيرا في سبيل البلد \n\n- و ستفقد الملكة *إليزابيث الثانية* بذلك سيادتها على واحدة من آخر ممالك الكومنولث \n\n2. ORGANIZATION\n\n- حسب أرقام *البنك الدولي* \n\n- أعلن *الجيش العراقي* \n\n- و نقلت وكالة *رويترز* عن ثلاثة دبلوماسيين في *الاتحاد الأوروبي* ، أن *بلجيكا* و *إيرلندا* و *لوكسمبورغ* تريد أيضاً مناقشة \n\n- *الحكومة الاتحادية* و *حكومة إقليم كردستان* \n\n- و هو ما يثير الشكوك حول مشاركة النجم البرتغالي في المباراة المرتقبة أمام *برشلونة* الإسباني في \n\n\n3. *LOCATION*\n\n- الجديد هو تمكين اللاجئين من “ مغادرة الجزيرة تدريجياً و بهدوء إلى *أثينا* ” \n\n- *جزيرة ساكيز* تبعد 1 كم عن *إزمير* \n\n\n4. DATE\n\n- *غدا الجمعة* \n\n- *06 أكتوبر 2020* \n\n- *العام السابق* \n\n\n5. PRODUCT\n\n- عبر حسابه ب *تطبيق “ إنستغرام ”* \n\n- الجيل الثاني من *نظارة الواقع الافتراضي أوكولوس كويست* تحت اسم \" *أوكولوس كويست 2* \" \n\n\n6. COMPETITION\n\n- عدم المشاركة في *بطولة فرنسا المفتوحة للتنس* \n\n- في مباراة *كأس السوبر الأوروبي* \n\n7. PRIZE\n\n- *جائزة نوبل ل لآداب*\n\n- الذي فاز ب *جائزة “ إيمي ” لأفضل دور مساند*\n\n8. EVENT\n\n- تسجّل أغنية جديدة خاصة ب *العيد الوطني السعودي*\n\n- *مهرجان المرأة يافوية* في دورته الرابعة \n\n9. DISEASE\n\n- في مكافحة فيروس *كورونا* و عدد من الأمراض \n\n- الأزمات المشابهة مثل “ *انفلونزا الطيور* ” و ” *انفلونزا الخنازير*", "## Example\n\nFind here a complete example to use this model", "## Training Corpus\n\nThe training corpus is made of 378.000 tokens (14.000 sentences) collected from the Web and annotated manually.", "## Results\n\nThe results on a valid corpus made of 30.000 tokens shows an F-measure of ~87%." ]
[ "TAGS\n#transformers #pytorch #tf #jax #safetensors #bert #token-classification #ar #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "# Arabic Named Entity Recognition Model\n\nPretrained BERT-based (arabic-bert-base) Named Entity Recognition model for Arabic.\n\nThe pre-trained model can recognize the following entities:\n1. PERSON\n\n- و هذا ما نفاه المعاون السياسي للرئيس *نبيه بري* ، النائب *علي حسن خليل* \n\n- لكن أوساط *الحريري* تعتبر أنه ضحى كثيرا في سبيل البلد \n\n- و ستفقد الملكة *إليزابيث الثانية* بذلك سيادتها على واحدة من آخر ممالك الكومنولث \n\n2. ORGANIZATION\n\n- حسب أرقام *البنك الدولي* \n\n- أعلن *الجيش العراقي* \n\n- و نقلت وكالة *رويترز* عن ثلاثة دبلوماسيين في *الاتحاد الأوروبي* ، أن *بلجيكا* و *إيرلندا* و *لوكسمبورغ* تريد أيضاً مناقشة \n\n- *الحكومة الاتحادية* و *حكومة إقليم كردستان* \n\n- و هو ما يثير الشكوك حول مشاركة النجم البرتغالي في المباراة المرتقبة أمام *برشلونة* الإسباني في \n\n\n3. *LOCATION*\n\n- الجديد هو تمكين اللاجئين من “ مغادرة الجزيرة تدريجياً و بهدوء إلى *أثينا* ” \n\n- *جزيرة ساكيز* تبعد 1 كم عن *إزمير* \n\n\n4. DATE\n\n- *غدا الجمعة* \n\n- *06 أكتوبر 2020* \n\n- *العام السابق* \n\n\n5. PRODUCT\n\n- عبر حسابه ب *تطبيق “ إنستغرام ”* \n\n- الجيل الثاني من *نظارة الواقع الافتراضي أوكولوس كويست* تحت اسم \" *أوكولوس كويست 2* \" \n\n\n6. COMPETITION\n\n- عدم المشاركة في *بطولة فرنسا المفتوحة للتنس* \n\n- في مباراة *كأس السوبر الأوروبي* \n\n7. PRIZE\n\n- *جائزة نوبل ل لآداب*\n\n- الذي فاز ب *جائزة “ إيمي ” لأفضل دور مساند*\n\n8. EVENT\n\n- تسجّل أغنية جديدة خاصة ب *العيد الوطني السعودي*\n\n- *مهرجان المرأة يافوية* في دورته الرابعة \n\n9. DISEASE\n\n- في مكافحة فيروس *كورونا* و عدد من الأمراض \n\n- الأزمات المشابهة مثل “ *انفلونزا الطيور* ” و ” *انفلونزا الخنازير*", "## Example\n\nFind here a complete example to use this model", "## Training Corpus\n\nThe training corpus is made of 378.000 tokens (14.000 sentences) collected from the Web and annotated manually.", "## Results\n\nThe results on a valid corpus made of 30.000 tokens shows an F-measure of ~87%." ]
text-classification
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-cola This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the glue dataset. It achieves the following results on the evaluation set: - Loss: 0.8508 - Matthews Correlation: 0.5452 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Matthews Correlation | |:-------------:|:-----:|:----:|:---------------:|:--------------------:| | 0.5221 | 1.0 | 535 | 0.5370 | 0.4246 | | 0.3462 | 2.0 | 1070 | 0.5157 | 0.5183 | | 0.2332 | 3.0 | 1605 | 0.6324 | 0.5166 | | 0.1661 | 4.0 | 2140 | 0.7616 | 0.5370 | | 0.1263 | 5.0 | 2675 | 0.8508 | 0.5452 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.9.0+cu111 - Datasets 1.14.0 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["glue"], "metrics": ["matthews_correlation"], "model-index": [{"name": "distilbert-base-uncased-finetuned-cola", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "glue", "type": "glue", "args": "cola"}, "metrics": [{"type": "matthews_correlation", "value": 0.5451837431775948, "name": "Matthews Correlation"}]}]}]}
hchc/distilbert-base-uncased-finetuned-cola
null
[ "transformers", "pytorch", "tensorboard", "distilbert", "text-classification", "generated_from_trainer", "dataset:glue", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
distilbert-base-uncased-finetuned-cola ====================================== This model is a fine-tuned version of distilbert-base-uncased on the glue dataset. It achieves the following results on the evaluation set: * Loss: 0.8508 * Matthews Correlation: 0.5452 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 5 ### Training results ### Framework versions * Transformers 4.11.3 * Pytorch 1.9.0+cu111 * Datasets 1.14.0 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.14.0\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5", "### Training results", "### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.14.0\n* Tokenizers 0.10.3" ]
text-classification
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-cola This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the glue dataset. It achieves the following results on the evaluation set: - Loss: 0.8657 - Matthews Correlation: 0.5472 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Matthews Correlation | |:-------------:|:-----:|:----:|:---------------:|:--------------------:| | 0.527 | 1.0 | 535 | 0.5545 | 0.3893 | | 0.3518 | 2.0 | 1070 | 0.5170 | 0.4970 | | 0.2448 | 3.0 | 1605 | 0.6734 | 0.5142 | | 0.1779 | 4.0 | 2140 | 0.7728 | 0.5466 | | 0.1339 | 5.0 | 2675 | 0.8657 | 0.5472 | ### Framework versions - Transformers 4.12.3 - Pytorch 1.9.0+cu111 - Datasets 1.15.1 - Tokenizers 0.10.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["glue"], "metrics": ["matthews_correlation"], "model-index": [{"name": "distilbert-base-uncased-finetuned-cola", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "glue", "type": "glue", "args": "cola"}, "metrics": [{"type": "matthews_correlation", "value": 0.5471613867597194, "name": "Matthews Correlation"}]}]}]}
hcjang1987/distilbert-base-uncased-finetuned-cola
null
[ "transformers", "pytorch", "tensorboard", "distilbert", "text-classification", "generated_from_trainer", "dataset:glue", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
distilbert-base-uncased-finetuned-cola ====================================== This model is a fine-tuned version of distilbert-base-uncased on the glue dataset. It achieves the following results on the evaluation set: * Loss: 0.8657 * Matthews Correlation: 0.5472 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 5 ### Training results ### Framework versions * Transformers 4.12.3 * Pytorch 1.9.0+cu111 * Datasets 1.15.1 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5", "### Training results", "### Framework versions\n\n\n* Transformers 4.12.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.15.1\n* Tokenizers 0.10.3" ]
question-answering
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-squad This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the squad dataset. It achieves the following results on the evaluation set: - Loss: 1.2131 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 1.2672 | 1.0 | 5533 | 1.2131 | ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.11.6
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["squad"], "model-index": [{"name": "distilbert-base-uncased-finetuned-squad", "results": []}]}
hcy11/distilbert-base-uncased-finetuned-squad
null
[ "transformers", "pytorch", "tensorboard", "distilbert", "question-answering", "generated_from_trainer", "dataset:squad", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #distilbert #question-answering #generated_from_trainer #dataset-squad #license-apache-2.0 #endpoints_compatible #region-us
distilbert-base-uncased-finetuned-squad ======================================= This model is a fine-tuned version of distilbert-base-uncased on the squad dataset. It achieves the following results on the evaluation set: * Loss: 1.2131 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 1 ### Training results ### Framework versions * Transformers 4.16.2 * Pytorch 1.10.0+cu111 * Datasets 1.18.3 * Tokenizers 0.11.6
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.6" ]
[ "TAGS\n#transformers #pytorch #tensorboard #distilbert #question-answering #generated_from_trainer #dataset-squad #license-apache-2.0 #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.6" ]
null
null
# Fun with transformers
{"license": "mit"}
hcy11/transformer
null
[ "license:mit", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #license-mit #region-us
# Fun with transformers
[ "# Fun with transformers" ]
[ "TAGS\n#license-mit #region-us \n", "# Fun with transformers" ]
text-classification
transformers
Technique Classification for https://propaganda.qcri.org/ptc/index.html
{}
hd10/semeval2020_task11_tc
null
[ "transformers", "pytorch", "deberta", "text-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #deberta #text-classification #autotrain_compatible #endpoints_compatible #region-us
Technique Classification for URL
[]
[ "TAGS\n#transformers #pytorch #deberta #text-classification #autotrain_compatible #endpoints_compatible #region-us \n" ]
text-generation
transformers
# diablo GPT random
{"tags": ["conversational"]}
heabeoun/DiabloGPT-small-nuon-conv
null
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# diablo GPT random
[ "# diablo GPT random" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# diablo GPT random" ]
null
transformers
DPR context encoder for Biomedical slot filling see https://arxiv.org/abs/2109.08564 for details. Load with: ```python from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast ctx_encoder = DPRContextEncoder.from_pretrained('healx/biomedical-dpr-ctx-encoder') ctx_tokenizer = DPRContextEncoderTokenizerFast.from_pretrained('facebook/dpr-ctx_encoder-single-nq-base') ```
{}
healx/biomedical-dpr-ctx-encoder
null
[ "transformers", "pytorch", "dpr", "arxiv:2109.08564", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[ "2109.08564" ]
[]
TAGS #transformers #pytorch #dpr #arxiv-2109.08564 #endpoints_compatible #region-us
DPR context encoder for Biomedical slot filling see URL for details. Load with:
[]
[ "TAGS\n#transformers #pytorch #dpr #arxiv-2109.08564 #endpoints_compatible #region-us \n" ]
feature-extraction
transformers
DPR query encoder for Biomedical slot filling see https://arxiv.org/abs/2109.08564 for details. Load with: ```python from transformers import DPRQuestionEncoder, DPRQuestionEncoderTokenizerFast qry_encoder = DPRQuestionEncoder.from_pretrained('healx/biomedical-dpr-qry-encoder') qry_tokenizer = DPRQuestionEncoderTokenizer.from_pretrained('facebook/dpr-question_encoder-single-nq-base') ```
{}
healx/biomedical-dpr-qry-encoder
null
[ "transformers", "pytorch", "dpr", "feature-extraction", "arxiv:2109.08564", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[ "2109.08564" ]
[]
TAGS #transformers #pytorch #dpr #feature-extraction #arxiv-2109.08564 #endpoints_compatible #region-us
DPR query encoder for Biomedical slot filling see URL for details. Load with:
[]
[ "TAGS\n#transformers #pytorch #dpr #feature-extraction #arxiv-2109.08564 #endpoints_compatible #region-us \n" ]
question-answering
transformers
Reader model for Biomedical slot filling see https://arxiv.org/abs/2109.08564 for details. The model is initialized with [biobert-base](https://huggingface.co/dmis-lab/biobert-v1.1).
{}
healx/biomedical-slot-filling-reader-base
null
[ "transformers", "pytorch", "bert", "question-answering", "arxiv:2109.08564", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[ "2109.08564" ]
[]
TAGS #transformers #pytorch #bert #question-answering #arxiv-2109.08564 #endpoints_compatible #region-us
Reader model for Biomedical slot filling see URL for details. The model is initialized with biobert-base.
[]
[ "TAGS\n#transformers #pytorch #bert #question-answering #arxiv-2109.08564 #endpoints_compatible #region-us \n" ]
question-answering
transformers
Reader model for Biomedical slot filling see https://arxiv.org/abs/2109.08564 for details. The model is initialized with [biobert-large](https://huggingface.co/dmis-lab/biobert-large-cased-v1.1).
{}
healx/biomedical-slot-filling-reader-large
null
[ "transformers", "pytorch", "bert", "question-answering", "arxiv:2109.08564", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[ "2109.08564" ]
[]
TAGS #transformers #pytorch #bert #question-answering #arxiv-2109.08564 #endpoints_compatible #region-us
Reader model for Biomedical slot filling see URL for details. The model is initialized with biobert-large.
[]
[ "TAGS\n#transformers #pytorch #bert #question-answering #arxiv-2109.08564 #endpoints_compatible #region-us \n" ]
null
transformers
GPT-2 (774M model) finetuned on 0.5m PubMed abstracts. Used in the [writemeanabstract.com](writemeanabstract.com) and the following preprint: [Papanikolaou, Yannis, and Andrea Pierleoni. "DARE: Data Augmented Relation Extraction with GPT-2." arXiv preprint arXiv:2004.13845 (2020).](https://arxiv.org/abs/2004.13845)
{}
healx/gpt-2-pubmed-large
null
[ "transformers", "pytorch", "arxiv:2004.13845", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[ "2004.13845" ]
[]
TAGS #transformers #pytorch #arxiv-2004.13845 #endpoints_compatible #region-us
GPT-2 (774M model) finetuned on 0.5m PubMed abstracts. Used in the URL and the following preprint: Papanikolaou, Yannis, and Andrea Pierleoni. "DARE: Data Augmented Relation Extraction with GPT-2." arXiv preprint arXiv:2004.13845 (2020).
[]
[ "TAGS\n#transformers #pytorch #arxiv-2004.13845 #endpoints_compatible #region-us \n" ]
null
transformers
GPT-2 (355M model) finetuned on 0.5m PubMed abstracts. Used in the [writemeanabstract.com](writemeanabstract.com) and the following preprint: [Papanikolaou, Yannis, and Andrea Pierleoni. "DARE: Data Augmented Relation Extraction with GPT-2." arXiv preprint arXiv:2004.13845 (2020).](https://arxiv.org/abs/2004.13845)
{}
healx/gpt-2-pubmed-medium
null
[ "transformers", "pytorch", "arxiv:2004.13845", "endpoints_compatible", "has_space", "region:us" ]
null
2022-03-02T23:29:05+00:00
[ "2004.13845" ]
[]
TAGS #transformers #pytorch #arxiv-2004.13845 #endpoints_compatible #has_space #region-us
GPT-2 (355M model) finetuned on 0.5m PubMed abstracts. Used in the URL and the following preprint: Papanikolaou, Yannis, and Andrea Pierleoni. "DARE: Data Augmented Relation Extraction with GPT-2." arXiv preprint arXiv:2004.13845 (2020).
[]
[ "TAGS\n#transformers #pytorch #arxiv-2004.13845 #endpoints_compatible #has_space #region-us \n" ]
text-classification
transformers
# Model Trained Using AutoNLP - Problem type: Multi-class Classification - Model ID: 202661 ## Validation Metrics - Loss: 1.5369086265563965 - Accuracy: 0.30762817840766987 - Macro F1: 0.28034259092597485 - Micro F1: 0.30762817840766987 - Weighted F1: 0.28072818168048186 - Macro Precision: 0.3113843896292072 - Micro Precision: 0.30762817840766987 - Weighted Precision: 0.3128459166476807 - Macro Recall: 0.3071652685939504 - Micro Recall: 0.30762817840766987 - Weighted Recall: 0.30762817840766987 ## Usage You can use cURL to access this model: ``` $ curl -X POST -H "Authorization: Bearer YOUR_API_KEY" -H "Content-Type: application/json" -d '{"inputs": "I love AutoNLP"}' https://api-inference.huggingface.co/models/hectorcotelo/autonlp-spanish_songs-202661 ``` Or Python API: ``` from transformers import AutoModelForSequenceClassification, AutoTokenizer model = AutoModelForSequenceClassification.from_pretrained("hectorcotelo/autonlp-spanish_songs-202661", use_auth_token=True) tokenizer = AutoTokenizer.from_pretrained("hectorcotelo/autonlp-spanish_songs-202661", use_auth_token=True) inputs = tokenizer("I love AutoNLP", return_tensors="pt") outputs = model(**inputs) ```
{"language": "es", "tags": "autonlp", "datasets": ["hectorcotelo/autonlp-data-spanish_songs"], "widget": [{"text": "Y si me tomo una cerveza Vuelves a mi cabeza Y empiezo a recordarte Es que me gusta c\u00f3mo besas Con tu delicadeza Puede ser que T\u00fa y yo, somos el uno para el otro Que no dejo de pensarte Quise olvidarte y tom\u00e9 un poco Y result\u00f3 extra\u00f1arte, yeah"}]}
hectorcotelo/autonlp-spanish_songs-202661
null
[ "transformers", "pytorch", "bert", "text-classification", "autonlp", "es", "dataset:hectorcotelo/autonlp-data-spanish_songs", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "es" ]
TAGS #transformers #pytorch #bert #text-classification #autonlp #es #dataset-hectorcotelo/autonlp-data-spanish_songs #autotrain_compatible #endpoints_compatible #region-us
# Model Trained Using AutoNLP - Problem type: Multi-class Classification - Model ID: 202661 ## Validation Metrics - Loss: 1.5369086265563965 - Accuracy: 0.30762817840766987 - Macro F1: 0.28034259092597485 - Micro F1: 0.30762817840766987 - Weighted F1: 0.28072818168048186 - Macro Precision: 0.3113843896292072 - Micro Precision: 0.30762817840766987 - Weighted Precision: 0.3128459166476807 - Macro Recall: 0.3071652685939504 - Micro Recall: 0.30762817840766987 - Weighted Recall: 0.30762817840766987 ## Usage You can use cURL to access this model: Or Python API:
[ "# Model Trained Using AutoNLP\n\n- Problem type: Multi-class Classification\n- Model ID: 202661", "## Validation Metrics\n\n- Loss: 1.5369086265563965\n- Accuracy: 0.30762817840766987\n- Macro F1: 0.28034259092597485\n- Micro F1: 0.30762817840766987\n- Weighted F1: 0.28072818168048186\n- Macro Precision: 0.3113843896292072\n- Micro Precision: 0.30762817840766987\n- Weighted Precision: 0.3128459166476807\n- Macro Recall: 0.3071652685939504\n- Micro Recall: 0.30762817840766987\n- Weighted Recall: 0.30762817840766987", "## Usage\n\nYou can use cURL to access this model:\n\n\n\nOr Python API:" ]
[ "TAGS\n#transformers #pytorch #bert #text-classification #autonlp #es #dataset-hectorcotelo/autonlp-data-spanish_songs #autotrain_compatible #endpoints_compatible #region-us \n", "# Model Trained Using AutoNLP\n\n- Problem type: Multi-class Classification\n- Model ID: 202661", "## Validation Metrics\n\n- Loss: 1.5369086265563965\n- Accuracy: 0.30762817840766987\n- Macro F1: 0.28034259092597485\n- Micro F1: 0.30762817840766987\n- Weighted F1: 0.28072818168048186\n- Macro Precision: 0.3113843896292072\n- Micro Precision: 0.30762817840766987\n- Weighted Precision: 0.3128459166476807\n- Macro Recall: 0.3071652685939504\n- Micro Recall: 0.30762817840766987\n- Weighted Recall: 0.30762817840766987", "## Usage\n\nYou can use cURL to access this model:\n\n\n\nOr Python API:" ]
null
null
Trying out Hugging Face
{}
hegdeashwin/test-model
null
[ "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #region-us
Trying out Hugging Face
[]
[ "TAGS\n#region-us \n" ]
text-classification
transformers
## Offensive Language Detection Model in Turkish - uses Bert and pytorch - fine tuned with Twitter data. - UTF-8 configuration is done ### Training Data Number of training sentences: 31,277 **Example Tweets** - 19823 Daliaan yifng cok erken attin be... 1.38 ...| NOT| - 30525 @USER Bak biri kollarımda uyuyup gitmem diyor..|NOT| - 26468 Helal olsun be :) Norveçten sabaha karşı geldi aq... | OFF| - 14105 @USER Sunu cekecek ve güzel oldugunu söylecek aptal... |OFF| - 4958 Ya seni yerim ben şapşal şey 🤗 | NOT| - 12966 Herkesin akıllı geçindiği bir sosyal medyamız var ... |NOT| - 5788 Maçın özetlerini izleyenler futbolcular gidiyo... |NOT| |OFFENSIVE |RESULT | |--|--| |NOT | 25231| |OFF|6046| dtype: int64 ### Validation |epoch |Training Loss | Valid. Loss | Valid.Accuracy | Training Time | Validation Time | |--|--|--|--|--|--| |1 | 0.31| 0.28| 0.89| 0:07:14 | 0:00:13 |2 | 0.18| 0.29| 0.90| 0:07:18 | 0:00:13 |3 | 0.08| 0.40| 0.89| 0:07:16 | 0:00:13 |4 | 0.04| 0.59| 0.89| 0:07:13 | 0:00:13 **Matthews Corr. Coef. (-1 : +1):** Total MCC Score: 0.633
{"language": "tr", "widget": [{"text": "sevelim sevilelim bu dunya kimseye kalmaz"}]}
hemekci/off_detection_turkish
null
[ "transformers", "pytorch", "jax", "bert", "text-classification", "tr", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "tr" ]
TAGS #transformers #pytorch #jax #bert #text-classification #tr #autotrain_compatible #endpoints_compatible #region-us
Offensive Language Detection Model in Turkish --------------------------------------------- * uses Bert and pytorch * fine tuned with Twitter data. * UTF-8 configuration is done ### Training Data Number of training sentences: 31,277 Example Tweets * 19823 Daliaan yifng cok erken attin be... 1.38 ...| NOT| * 30525 @USER Bak biri kollarımda uyuyup gitmem diyor..|NOT| * 26468 Helal olsun be :) Norveçten sabaha karşı geldi aq... | OFF| * 14105 @USER Sunu cekecek ve güzel oldugunu söylecek aptal... |OFF| * 4958 Ya seni yerim ben şapşal şey | NOT| * 12966 Herkesin akıllı geçindiği bir sosyal medyamız var ... |NOT| * 5788 Maçın özetlerini izleyenler futbolcular gidiyo... |NOT| ### Validation Matthews Corr. Coef. (-1 : +1): Total MCC Score: 0.633
[ "### Training Data\n\n\nNumber of training sentences: 31,277\n\n\nExample Tweets\n\n\n* 19823 Daliaan yifng cok erken attin be... 1.38 ...| NOT|\n* 30525 @USER Bak biri kollarımda uyuyup gitmem diyor..|NOT|\n* 26468 Helal olsun be :) Norveçten sabaha karşı geldi aq... | OFF|\n* 14105 @USER Sunu cekecek ve güzel oldugunu söylecek aptal... |OFF|\n* 4958 Ya seni yerim ben şapşal şey | NOT|\n* 12966 Herkesin akıllı geçindiği bir sosyal medyamız var ... |NOT|\n* 5788 Maçın özetlerini izleyenler futbolcular gidiyo... |NOT|", "### Validation\n\n\n\nMatthews Corr. Coef. (-1 : +1):\nTotal MCC Score: 0.633" ]
[ "TAGS\n#transformers #pytorch #jax #bert #text-classification #tr #autotrain_compatible #endpoints_compatible #region-us \n", "### Training Data\n\n\nNumber of training sentences: 31,277\n\n\nExample Tweets\n\n\n* 19823 Daliaan yifng cok erken attin be... 1.38 ...| NOT|\n* 30525 @USER Bak biri kollarımda uyuyup gitmem diyor..|NOT|\n* 26468 Helal olsun be :) Norveçten sabaha karşı geldi aq... | OFF|\n* 14105 @USER Sunu cekecek ve güzel oldugunu söylecek aptal... |OFF|\n* 4958 Ya seni yerim ben şapşal şey | NOT|\n* 12966 Herkesin akıllı geçindiği bir sosyal medyamız var ... |NOT|\n* 5788 Maçın özetlerini izleyenler futbolcular gidiyo... |NOT|", "### Validation\n\n\n\nMatthews Corr. Coef. (-1 : +1):\nTotal MCC Score: 0.633" ]
question-answering
transformers
# Multilingual + Dutch SQuAD2.0 This model is the multilingual model provided by the Google research team with a fine-tuned dutch Q&A downstream task. ## Details of the language model Language model ([**bert-base-multilingual-cased**](https://github.com/google-research/bert/blob/master/multilingual.md)): 12-layer, 768-hidden, 12-heads, 110M parameters. Trained on cased text in the top 104 languages with the largest Wikipedias. ## Details of the downstream task Using the `mtranslate` Python module, [**SQuAD2.0**](https://rajpurkar.github.io/SQuAD-explorer/) was machine-translated. In order to find the start tokens, the direct translations of the answers were searched in the corresponding paragraphs. Due to the different translations depending on the context (missing context in the pure answer), the answer could not always be found in the text, and thus a loss of question-answer examples occurred. This is a potential problem where errors can occur in the data set. | Dataset | # Q&A | | ---------------------- | ----- | | SQuAD2.0 Train | 130 K | | Dutch SQuAD2.0 Train | 99 K | | SQuAD2.0 Dev | 12 K | | Dutch SQuAD2.0 Dev | 10 K | ## Model benchmark | Model | EM/F1 |HasAns (EM/F1) | NoAns | | ---------------------- | ----- | ----- | ----- | | [robBERT](https://huggingface.co/pdelobelle/robBERT-base) | 58.04/60.95 | 33.08/40.64 | 73.67 | | [dutchBERT](https://huggingface.co/wietsedv/bert-base-dutch-cased) | 64.25/68.45 | 45.59/56.49 | 75.94 | | [multiBERT](https://huggingface.co/bert-base-multilingual-cased) | **67.38**/**71.36** | 47.42/57.76 | 79.88 | ## Model training The model was trained on a **Tesla V100** GPU with the following command: ```python export SQUAD_DIR=path/to/nl_squad python run_squad.py --model_type bert \ --model_name_or_path bert-base-multilingual-cased \ --do_train \ --do_eval \ --train_file $SQUAD_DIR/nl_squadv2_train_clean.json \ --predict_file $SQUAD_DIR/nl_squadv2_dev_clean.json \ --num_train_epochs 2 \ --max_seq_length 384 \ --doc_stride 128 \ --save_steps=8000 \ --output_dir ../../output \ --overwrite_cache \ --overwrite_output_dir ``` **Results**: {'exact': 67.38028751680629, 'f1': 71.362297054268, 'total': 9669, 'HasAns_exact': 47.422126745435015, 'HasAns_f1': 57.761023151910734, 'HasAns_total': 3724, 'NoAns_exact': 79.88225399495374, 'NoAns_f1': 79.88225399495374, 'NoAns_total': 5945, 'best_exact': 67.53542248422795, 'best_exact_thresh': 0.0, 'best_f1': 71.36229705426837, 'best_f1_thresh': 0.0} ## Model in action Fast usage with **pipelines**: ```python from transformers import pipeline qa_pipeline = pipeline( "question-answering", model="henryk/bert-base-multilingual-cased-finetuned-dutch-squad2", tokenizer="henryk/bert-base-multilingual-cased-finetuned-dutch-squad2" ) qa_pipeline({ 'context': "Amsterdam is de hoofdstad en de dichtstbevolkte stad van Nederland.", 'question': "Wat is de hoofdstad van Nederland?"}) ``` # Output: ```json { "score": 0.83, "start": 0, "end": 9, "answer": "Amsterdam" } ``` ## Contact Please do not hesitate to contact me via [LinkedIn](https://www.linkedin.com/in/henryk-borzymowski-0755a2167/) if you want to discuss or get access to the Dutch version of SQuAD.
{"language": "nl"}
henryk/bert-base-multilingual-cased-finetuned-dutch-squad2
null
[ "transformers", "pytorch", "jax", "bert", "question-answering", "nl", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "nl" ]
TAGS #transformers #pytorch #jax #bert #question-answering #nl #endpoints_compatible #region-us
Multilingual + Dutch SQuAD2.0 ============================= This model is the multilingual model provided by the Google research team with a fine-tuned dutch Q&A downstream task. Details of the language model ----------------------------- Language model (bert-base-multilingual-cased): 12-layer, 768-hidden, 12-heads, 110M parameters. Trained on cased text in the top 104 languages with the largest Wikipedias. Details of the downstream task ------------------------------ Using the 'mtranslate' Python module, SQuAD2.0 was machine-translated. In order to find the start tokens, the direct translations of the answers were searched in the corresponding paragraphs. Due to the different translations depending on the context (missing context in the pure answer), the answer could not always be found in the text, and thus a loss of question-answer examples occurred. This is a potential problem where errors can occur in the data set. Model benchmark --------------- Model training -------------- The model was trained on a Tesla V100 GPU with the following command: Results: {'exact': 67.38028751680629, 'f1': 71.362297054268, 'total': 9669, 'HasAns\_exact': 47.422126745435015, 'HasAns\_f1': 57.761023151910734, 'HasAns\_total': 3724, 'NoAns\_exact': 79.88225399495374, 'NoAns\_f1': 79.88225399495374, 'NoAns\_total': 5945, 'best\_exact': 67.53542248422795, 'best\_exact\_thresh': 0.0, 'best\_f1': 71.36229705426837, 'best\_f1\_thresh': 0.0} Model in action --------------- Fast usage with pipelines: Output: ======= Contact ------- Please do not hesitate to contact me via LinkedIn if you want to discuss or get access to the Dutch version of SQuAD.
[]
[ "TAGS\n#transformers #pytorch #jax #bert #question-answering #nl #endpoints_compatible #region-us \n" ]
question-answering
transformers
# Multilingual + Polish SQuAD1.1 This model is the multilingual model provided by the Google research team with a fine-tuned polish Q&A downstream task. ## Details of the language model Language model ([**bert-base-multilingual-cased**](https://github.com/google-research/bert/blob/master/multilingual.md)): 12-layer, 768-hidden, 12-heads, 110M parameters. Trained on cased text in the top 104 languages with the largest Wikipedias. ## Details of the downstream task Using the `mtranslate` Python module, [**SQuAD1.1**](https://rajpurkar.github.io/SQuAD-explorer/) was machine-translated. In order to find the start tokens, the direct translations of the answers were searched in the corresponding paragraphs. Due to the different translations depending on the context (missing context in the pure answer), the answer could not always be found in the text, and thus a loss of question-answer examples occurred. This is a potential problem where errors can occur in the data set. | Dataset | # Q&A | | ---------------------- | ----- | | SQuAD1.1 Train | 87.7 K | | Polish SQuAD1.1 Train | 39.5 K | | SQuAD1.1 Dev | 10.6 K | | Polish SQuAD1.1 Dev | 2.6 K | ## Model benchmark | Model | EM | F1 | | ---------------------- | ----- | ----- | | [SlavicBERT](https://huggingface.co/DeepPavlov/bert-base-bg-cs-pl-ru-cased) | **60.89** | 71.68 | | [polBERT](https://huggingface.co/dkleczek/bert-base-polish-uncased-v1) | 57.46 | 68.87 | | [multiBERT](https://huggingface.co/bert-base-multilingual-cased) | 60.67 | **71.89** | | [xlm](https://huggingface.co/xlm-mlm-100-1280) | 47.98 | 59.42 | ## Model training The model was trained on a **Tesla V100** GPU with the following command: ```python export SQUAD_DIR=path/to/pl_squad python run_squad.py --model_type bert \ --model_name_or_path bert-base-multilingual-cased \ --do_train \ --do_eval \ --train_file $SQUAD_DIR/pl_squadv1_train_clean.json \ --predict_file $SQUAD_DIR/pl_squadv1_dev_clean.json \ --num_train_epochs 2 \ --max_seq_length 384 \ --doc_stride 128 \ --save_steps=8000 \ --output_dir ../../output \ --overwrite_cache \ --overwrite_output_dir ``` **Results**: {'exact': 60.670731707317074, 'f1': 71.8952193697293, 'total': 2624, 'HasAns_exact': 60.670731707317074, 'HasAns_f1': 71.8952193697293, 'HasAns_total': 2624, 'best_exact': 60.670731707317074, 'best_exact_thresh': 0.0, 'best_f1': 71.8952193697293, 'best_f1_thresh': 0.0} ## Model in action Fast usage with **pipelines**: ```python from transformers import pipeline qa_pipeline = pipeline( "question-answering", model="henryk/bert-base-multilingual-cased-finetuned-polish-squad1", tokenizer="henryk/bert-base-multilingual-cased-finetuned-polish-squad1" ) qa_pipeline({ 'context': "Warszawa jest największym miastem w Polsce pod względem liczby ludności i powierzchni", 'question': "Jakie jest największe miasto w Polsce?"}) ``` # Output: ```json { "score": 0.9988, "start": 0, "end": 8, "answer": "Warszawa" } ``` ## Contact Please do not hesitate to contact me via [LinkedIn](https://www.linkedin.com/in/henryk-borzymowski-0755a2167/) if you want to discuss or get access to the Polish version of SQuAD.
{"language": "pl"}
henryk/bert-base-multilingual-cased-finetuned-polish-squad1
null
[ "transformers", "pytorch", "jax", "bert", "question-answering", "pl", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "pl" ]
TAGS #transformers #pytorch #jax #bert #question-answering #pl #endpoints_compatible #region-us
Multilingual + Polish SQuAD1.1 ============================== This model is the multilingual model provided by the Google research team with a fine-tuned polish Q&A downstream task. Details of the language model ----------------------------- Language model (bert-base-multilingual-cased): 12-layer, 768-hidden, 12-heads, 110M parameters. Trained on cased text in the top 104 languages with the largest Wikipedias. Details of the downstream task ------------------------------ Using the 'mtranslate' Python module, SQuAD1.1 was machine-translated. In order to find the start tokens, the direct translations of the answers were searched in the corresponding paragraphs. Due to the different translations depending on the context (missing context in the pure answer), the answer could not always be found in the text, and thus a loss of question-answer examples occurred. This is a potential problem where errors can occur in the data set. Model benchmark --------------- Model: SlavicBERT, EM: 60.89, F1: 71.68 Model: polBERT, EM: 57.46, F1: 68.87 Model: multiBERT, EM: 60.67, F1: 71.89 Model: xlm, EM: 47.98, F1: 59.42 Model training -------------- The model was trained on a Tesla V100 GPU with the following command: Results: {'exact': 60.670731707317074, 'f1': 71.8952193697293, 'total': 2624, 'HasAns\_exact': 60.670731707317074, 'HasAns\_f1': 71.8952193697293, 'HasAns\_total': 2624, 'best\_exact': 60.670731707317074, 'best\_exact\_thresh': 0.0, 'best\_f1': 71.8952193697293, 'best\_f1\_thresh': 0.0} Model in action --------------- Fast usage with pipelines: Output: ======= Contact ------- Please do not hesitate to contact me via LinkedIn if you want to discuss or get access to the Polish version of SQuAD.
[]
[ "TAGS\n#transformers #pytorch #jax #bert #question-answering #pl #endpoints_compatible #region-us \n" ]
question-answering
transformers
# Multilingual + Polish SQuAD2.0 This model is the multilingual model provided by the Google research team with a fine-tuned polish Q&A downstream task. ## Details of the language model Language model ([**bert-base-multilingual-cased**](https://github.com/google-research/bert/blob/master/multilingual.md)): 12-layer, 768-hidden, 12-heads, 110M parameters. Trained on cased text in the top 104 languages with the largest Wikipedias. ## Details of the downstream task Using the `mtranslate` Python module, [**SQuAD2.0**](https://rajpurkar.github.io/SQuAD-explorer/) was machine-translated. In order to find the start tokens, the direct translations of the answers were searched in the corresponding paragraphs. Due to the different translations depending on the context (missing context in the pure answer), the answer could not always be found in the text, and thus a loss of question-answer examples occurred. This is a potential problem where errors can occur in the data set. | Dataset | # Q&A | | ---------------------- | ----- | | SQuAD2.0 Train | 130 K | | Polish SQuAD2.0 Train | 83.1 K | | SQuAD2.0 Dev | 12 K | | Polish SQuAD2.0 Dev | 8.5 K | ## Model benchmark | Model | EM/F1 |HasAns (EM/F1) | NoAns | | ---------------------- | ----- | ----- | ----- | | [SlavicBERT](https://huggingface.co/DeepPavlov/bert-base-bg-cs-pl-ru-cased) | 69.35/71.51 | 47.02/54.09 | 79.20 | | [polBERT](https://huggingface.co/dkleczek/bert-base-polish-uncased-v1) | 67.33/69.80| 45.73/53.80 | 76.87 | | [multiBERT](https://huggingface.co/bert-base-multilingual-cased) | **70.76**/**72.92** |45.00/52.04 | 82.13 | ## Model training The model was trained on a **Tesla V100** GPU with the following command: ```python export SQUAD_DIR=path/to/pl_squad python run_squad.py --model_type bert \ --model_name_or_path bert-base-multilingual-cased \ --do_train \ --do_eval \ --version_2_with_negative \ --train_file $SQUAD_DIR/pl_squadv2_train.json \ --predict_file $SQUAD_DIR/pl_squadv2_dev.json \ --num_train_epochs 2 \ --max_seq_length 384 \ --doc_stride 128 \ --save_steps=8000 \ --output_dir ../../output \ --overwrite_cache \ --overwrite_output_dir ``` **Results**: {'exact': 70.76671723655035, 'f1': 72.92156947155917, 'total': 8569, 'HasAns_exact': 45.00762195121951, 'HasAns_f1': 52.04456128116991, 'HasAns_total': 2624, 'NoAns_exact': 82.13624894869638, ' NoAns_f1': 82.13624894869638, 'NoAns_total': 5945, 'best_exact': 71.72365503559342, 'best_exact_thresh': 0.0, 'best_f1': 73.62662512059369, 'best_f1_thresh': 0.0} ## Model in action Fast usage with **pipelines**: ```python from transformers import pipeline qa_pipeline = pipeline( "question-answering", model="henryk/bert-base-multilingual-cased-finetuned-polish-squad2", tokenizer="henryk/bert-base-multilingual-cased-finetuned-polish-squad2" ) qa_pipeline({ 'context': "Warszawa jest największym miastem w Polsce pod względem liczby ludności i powierzchni", 'question': "Jakie jest największe miasto w Polsce?"}) ``` # Output: ```json { "score": 0.9986, "start": 0, "end": 8, "answer": "Warszawa" } ``` ## Contact Please do not hesitate to contact me via [LinkedIn](https://www.linkedin.com/in/henryk-borzymowski-0755a2167/) if you want to discuss or get access to the Polish version of SQuAD.
{"language": "pl"}
henryk/bert-base-multilingual-cased-finetuned-polish-squad2
null
[ "transformers", "pytorch", "jax", "bert", "question-answering", "pl", "endpoints_compatible", "has_space", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "pl" ]
TAGS #transformers #pytorch #jax #bert #question-answering #pl #endpoints_compatible #has_space #region-us
Multilingual + Polish SQuAD2.0 ============================== This model is the multilingual model provided by the Google research team with a fine-tuned polish Q&A downstream task. Details of the language model ----------------------------- Language model (bert-base-multilingual-cased): 12-layer, 768-hidden, 12-heads, 110M parameters. Trained on cased text in the top 104 languages with the largest Wikipedias. Details of the downstream task ------------------------------ Using the 'mtranslate' Python module, SQuAD2.0 was machine-translated. In order to find the start tokens, the direct translations of the answers were searched in the corresponding paragraphs. Due to the different translations depending on the context (missing context in the pure answer), the answer could not always be found in the text, and thus a loss of question-answer examples occurred. This is a potential problem where errors can occur in the data set. Model benchmark --------------- Model training -------------- The model was trained on a Tesla V100 GPU with the following command: Results: {'exact': 70.76671723655035, 'f1': 72.92156947155917, 'total': 8569, 'HasAns\_exact': 45.00762195121951, 'HasAns\_f1': 52.04456128116991, 'HasAns\_total': 2624, 'NoAns\_exact': 82.13624894869638, ' NoAns\_f1': 82.13624894869638, 'NoAns\_total': 5945, 'best\_exact': 71.72365503559342, 'best\_exact\_thresh': 0.0, 'best\_f1': 73.62662512059369, 'best\_f1\_thresh': 0.0} Model in action --------------- Fast usage with pipelines: Output: ======= Contact ------- Please do not hesitate to contact me via LinkedIn if you want to discuss or get access to the Polish version of SQuAD.
[]
[ "TAGS\n#transformers #pytorch #jax #bert #question-answering #pl #endpoints_compatible #has_space #region-us \n" ]
text-generation
transformers
# Rick and Morty DialoGPT Model
{"tags": ["conversational"]}
henryoce/DialoGPT-small-rick-and-morty
null
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Rick and Morty DialoGPT Model
[ "# Rick and Morty DialoGPT Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Rick and Morty DialoGPT Model" ]
summarization
transformers
## `t5-3b-samsum-deepspeed` This model was trained using Microsoft's `AzureML` and `DeepSpeed`'s ZeRO 2 optimization. It was fine-tuned on the `SAMSum` corpus from `t5-3b` checkpoint. More information on the fine-tuning process (includes samples and benchmarks): *(currently still WIP, updates coming soon: 7/6/21~7/9/21)* ## Resource Usage These results are retrieved from AzureML Studio's resource monitoring module. All experiments were ran on AzureML's low priority clusters. | key | value | | --- | ----- | | AzureML SKU | ND40rs_v2 (8 X V100 32GB) | | Region | US West 2 | | Run Duration | 43m 51.05s | | Compute Cost (LowPriority/Dedicated) | $3.22/$16.10 (USD) | | Average CPU Utilization | 46.0% | | Average GPU Utilization | 56.9% | | GPU Memory Usage (Avg/Peak) | 26.77/30.49 (GB) | | Total GPU Energy Usage | 2448.69 (kJ) | *Compute cost is calculated from run duration and SKU's price per hour. Updated SKU pricing could be found here: https://azure.microsoft.com/en-us/pricing/details/machine-learning/ *Peak memory usage is calculated from average peak across all utilized GPUs. ### Carbon Emissions These results are obtained using `codecarbon`. The carbon emission is estimated from training runtime only (excluding setup and evaluation runtime). CodeCarbon: https://github.com/mlco2/codecarbon | key | value | | --- | ----- | | timestamp | 2021-07-06T21:57:39 | | duration | 1841.4621863365173 | | emissions | 0.17802492531467784 | | energy_consumed | 0.5982020339874927 | | country_name | USA | | region | Washington | | cloud_provider | azure | | cloud_region | westus2 | ## Hyperparameters ```yaml fp16: True per device batch size: 2 effective batch size: 16 epoch: 3.0 learning rate: 3e-5 weight decay: 0.0 seed: 1 ``` *Same `per device batch size` for evaluations ### DeepSpeed Optimizer = `AdamW`, Scheduler = `WarmupDecayLR`, Offload = `none` ```json "zero_optimization": { "stage": 2, "allgather_partitions": true, "allgather_bucket_size": 1000000000, "overlap_comm": true, "reduce_scatter": true, "reduce_bucket_size": 1000000000, "contiguous_gradients": true } ``` ## Usage ```python from transformers import pipeline summarizer = pipeline("summarization", model="henryu-lin/t5-3b-samsum-deepspeed") conversation = '''Henry: Hey, is Nate coming over to watch the movie tonight? Kevin: Yea, he said he'll be arriving a bit later at around 7 since he gets off of work at 6. Have you taken out the garbage yet? It's starting to make the kitchen really smell. Henry: Oh I forgot. I'll do that once I'm finished with my assignment for my math class. Kevin: Yea, you should take it out as soon as possible. And also, Nate is bringing his girlfriend too. Henry: Nice, I'm really looking forward to seeing them again. ''' summarizer(conversation) ``` ## Results | ROUGE | Score | | ----- | ----- | | eval_rouge1 | 54.7875 | | eval_rouge2 | 30.565 | | eval_rougeL | 45.7625 | | eval_rougeLsum | 50.3915 | | predict_rouge1 | 53.6628 | | predict_rouge2 | 29.0196 | | predict_rougeL | 45.1257 | | predict_rougeLsum | 49.171 | | Metric | Value | | ------ | ----- | | eval_gen_len | 25.3399 | | predict_gen_len | 24.9133 | | train_loss | 1.1206104169494209 | | eval_loss | 1.0732421875 | | predict_loss | 1.087890625 | | train_runtime | 1841.3751 | | train_samples | 14732 | | train_samples_per_second | 24.002 | | train_steps_per_second | 1.501 | | eval_runtime | 163.8357 | | eval_samples | 818 | | eval_samples_per_second | 4.993 | | eval_steps_per_second | 0.317 | | predict_runtime | 168.8245 | | predict_samples | 819 | | predict_samples_per_second | 4.851 | | predict_steps_per_second | 0.308 | | total_steps | 2763 | | total_flos | 1.84452086400811e+17 |
{"language": "en", "license": "apache-2.0", "tags": ["azureml", "t5", "summarization", "deepspeed"], "datasets": ["samsum"], "widget": [{"text": "Henry: Hey, is Nate coming over to watch the movie tonight?\nKevin: Yea, he said he'll be arriving a bit later at around 7 since he gets off of work at 6. Have you taken out the garbage yet? It's starting to make the kitchen really smell.\nHenry: Oh I forgot. I'll do that once I'm finished with my assignment for my math class.\nKevin: Yea, you should take it out as soon as possible. And also, Nate is bringing his girlfriend too.\nHenry: Nice, I'm really looking forward to seeing them again."}]}
henryu-lin/t5-3b-samsum-deepspeed
null
[ "transformers", "pytorch", "t5", "text2text-generation", "azureml", "summarization", "deepspeed", "en", "dataset:samsum", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #t5 #text2text-generation #azureml #summarization #deepspeed #en #dataset-samsum #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
't5-3b-samsum-deepspeed' ------------------------ This model was trained using Microsoft's 'AzureML' and 'DeepSpeed''s ZeRO 2 optimization. It was fine-tuned on the 'SAMSum' corpus from 't5-3b' checkpoint. More information on the fine-tuning process (includes samples and benchmarks): *(currently still WIP, updates coming soon: 7/6/21~7/9/21)* Resource Usage -------------- These results are retrieved from AzureML Studio's resource monitoring module. All experiments were ran on AzureML's low priority clusters. \*Compute cost is calculated from run duration and SKU's price per hour. Updated SKU pricing could be found here: URL \*Peak memory usage is calculated from average peak across all utilized GPUs. ### Carbon Emissions These results are obtained using 'codecarbon'. The carbon emission is estimated from training runtime only (excluding setup and evaluation runtime). CodeCarbon: URL Hyperparameters --------------- \*Same 'per device batch size' for evaluations ### DeepSpeed Optimizer = 'AdamW', Scheduler = 'WarmupDecayLR', Offload = 'none' Usage ----- Results -------
[ "### Carbon Emissions\n\n\nThese results are obtained using 'codecarbon'. The carbon emission is estimated from training runtime only (excluding setup and evaluation runtime). \n\nCodeCarbon: URL\n\n\n\nHyperparameters\n---------------\n\n\n\\*Same 'per device batch size' for evaluations", "### DeepSpeed\n\n\nOptimizer = 'AdamW', Scheduler = 'WarmupDecayLR', Offload = 'none'\n\n\nUsage\n-----\n\n\nResults\n-------" ]
[ "TAGS\n#transformers #pytorch #t5 #text2text-generation #azureml #summarization #deepspeed #en #dataset-samsum #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Carbon Emissions\n\n\nThese results are obtained using 'codecarbon'. The carbon emission is estimated from training runtime only (excluding setup and evaluation runtime). \n\nCodeCarbon: URL\n\n\n\nHyperparameters\n---------------\n\n\n\\*Same 'per device batch size' for evaluations", "### DeepSpeed\n\n\nOptimizer = 'AdamW', Scheduler = 'WarmupDecayLR', Offload = 'none'\n\n\nUsage\n-----\n\n\nResults\n-------" ]
summarization
transformers
## `t5-large-samsum-deepspeed` This model was trained using Microsoft's `AzureML` and `DeepSpeed`'s ZeRO 2 optimization. It was fine-tuned on the `SAMSum` corpus from `t5-large` checkpoint. More information on the fine-tuning process (includes samples and benchmarks): *(currently still WIP, major updates coming soon: 7/6/21~7/9/21)* ## Resource Usage These results are retrieved from AzureML Studio's resource monitoring module. All experiments were ran on AzureML's low priority clusters. | key | value | | --- | ----- | | AzureML SKU | ND40rs_v2 (8 X V100 32GB) | | Region | US West 2 | | Run Duration | 12m 47.13s | | Compute Cost (LowPriority/Dedicated) | $0.94/$4.69 (USD) | | Average CPU Utilization | 51.2% | | Average GPU Utilization | 42.0% | | GPU Memory Usage (Avg/Peak) | 24.85/28.79 (GB) | | Total GPU Energy Usage | 670.38 (kJ) | *Compute cost is calculated from run duration and SKU's price per hour. Updated SKU pricing could be found here: https://azure.microsoft.com/en-us/pricing/details/machine-learning/ *Peak memory usage is calculated from average peak across all utilized GPUs. ### Carbon Emissions These results are obtained using `codecarbon`. The carbon emission is estimated from training runtime only (excluding setup and evaluation runtime). CodeCarbon: https://github.com/mlco2/codecarbon | key | value | | --- | ----- | | timestamp | 2021-07-08T06:29:27 | | duration | 515.5018835067749 | | emissions | 0.043562840982919106 | | energy_consumed | 0.14638051405550773 | | country_name | USA | | region | Washington | | cloud_provider | azure | | cloud_region | westus2 | ## Hyperparameters ```yaml fp16: True per device batch size: 8 effective batch size: 64 epoch: 3.0 learning rate: 1e-4 weight decay: 0.1 seed: 1 ``` *Same `per device batch size` for evaluations ### DeepSpeed Optimizer = `AdamW`, Scheduler = `WarmupDecayLR`, Offload = `none` ```json "zero_optimization": { "stage": 2, "allgather_partitions": true, "allgather_bucket_size": 1300000000, "overlap_comm": true, "reduce_scatter": true, "reduce_bucket_size": 1300000000, "contiguous_gradients": true } ``` ## Usage ```python from transformers import pipeline summarizer = pipeline("summarization", model="henryu-lin/t5-large-samsum-deepspeed") conversation = '''Kevin: Hey man, are you excited to watch Finding Nemo tonight? Henry: Yea, I can't wait to watch that same movie for the 89th time. Is Nate coming over to watch it with us tonight? Kevin: Yep, he said he'll be arriving a bit later at around 7 since he gets off of work at 6. Have you taken out the garbage yet? It's starting to make the kitchen really smell. Henry: Oh I forgot. I'll do that once I'm finished with my assignment for my math class. I didn't get to start on it until an hour ago, and it's due in 30 minutes. Kevin: Okay dude, you should take it out as soon as possible. By the way, Nate is bringing his girlfriend and their cat too. Henry: Nice, I'm really looking forward to seeing them again. ''' summarizer(conversation) ``` ## Results | ROUGE | Score | | ----- | ----- | | eval_rouge1 | 53.0823 | | eval_rouge2 | 28.7097 | | eval_rougeL | 43.939 | | eval_rougeLsum | 49.067 | | predict_rouge1 | 51.6716 | | predict_rouge2 | 26.5372 | | predict_rougeL | 42.9681 | | predict_rougeLsum | 47.4084 | | Metric | Value | | ------ | ----- | | eval_gen_len | 26.4071 | | predict_gen_len | 25.9451 | | train_loss | 1.3212629926497115 | | eval_loss | 1.23828125 | | predict_loss | 1.2333984375 | | train_runtime | 515.2198 | | train_samples | 14732 | | train_samples_per_second | 85.781 | | train_steps_per_second | 1.345 | | eval_runtime | 61.275 | | eval_samples | 818 | | eval_samples_per_second | 13.35 | | eval_steps_per_second | 0.212 | | predict_runtime | 63.3732 | | predict_samples | 819 | | predict_samples_per_second | 12.923 | | predict_steps_per_second | 0.205 | | total_steps | 693 | | total_flos | 7.20140924616704e+16 |
{"language": "en", "license": "apache-2.0", "tags": ["azureml", "t5", "summarization", "deepspeed"], "datasets": ["samsum"], "widget": [{"text": "Kevin: Hey man, are you excited to watch Finding Nemo tonight?\nHenry: Yea, I can't wait to watch that same movie for the 89th time. Is Nate coming over to watch it with us tonight?\nKevin: Yep, he said he'll be arriving a bit later at around 7 since he gets off of work at 6. Have you taken out the garbage yet? It's starting to make the kitchen really smell.\nHenry: Oh I forgot. I'll do that once I'm finished with my assignment for my math class. I didn't get to start on it until an hour ago, and it's due in 30 minutes.\nKevin: Okay dude, you should take it out as soon as possible. By the way, Nate is bringing his girlfriend and their cat too.\nHenry: Nice, I'm really looking forward to seeing them again."}]}
henryu-lin/t5-large-samsum-deepspeed
null
[ "transformers", "pytorch", "t5", "text2text-generation", "azureml", "summarization", "deepspeed", "en", "dataset:samsum", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #t5 #text2text-generation #azureml #summarization #deepspeed #en #dataset-samsum #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
't5-large-samsum-deepspeed' --------------------------- This model was trained using Microsoft's 'AzureML' and 'DeepSpeed''s ZeRO 2 optimization. It was fine-tuned on the 'SAMSum' corpus from 't5-large' checkpoint. More information on the fine-tuning process (includes samples and benchmarks): *(currently still WIP, major updates coming soon: 7/6/21~7/9/21)* Resource Usage -------------- These results are retrieved from AzureML Studio's resource monitoring module. All experiments were ran on AzureML's low priority clusters. \*Compute cost is calculated from run duration and SKU's price per hour. Updated SKU pricing could be found here: URL \*Peak memory usage is calculated from average peak across all utilized GPUs. ### Carbon Emissions These results are obtained using 'codecarbon'. The carbon emission is estimated from training runtime only (excluding setup and evaluation runtime). CodeCarbon: URL Hyperparameters --------------- \*Same 'per device batch size' for evaluations ### DeepSpeed Optimizer = 'AdamW', Scheduler = 'WarmupDecayLR', Offload = 'none' Usage ----- Results -------
[ "### Carbon Emissions\n\n\nThese results are obtained using 'codecarbon'. The carbon emission is estimated from training runtime only (excluding setup and evaluation runtime). \n\nCodeCarbon: URL\n\n\n\nHyperparameters\n---------------\n\n\n\\*Same 'per device batch size' for evaluations", "### DeepSpeed\n\n\nOptimizer = 'AdamW', Scheduler = 'WarmupDecayLR', Offload = 'none'\n\n\nUsage\n-----\n\n\nResults\n-------" ]
[ "TAGS\n#transformers #pytorch #t5 #text2text-generation #azureml #summarization #deepspeed #en #dataset-samsum #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Carbon Emissions\n\n\nThese results are obtained using 'codecarbon'. The carbon emission is estimated from training runtime only (excluding setup and evaluation runtime). \n\nCodeCarbon: URL\n\n\n\nHyperparameters\n---------------\n\n\n\\*Same 'per device batch size' for evaluations", "### DeepSpeed\n\n\nOptimizer = 'AdamW', Scheduler = 'WarmupDecayLR', Offload = 'none'\n\n\nUsage\n-----\n\n\nResults\n-------" ]
text-generation
transformers
# Harry Potter DialoGPT Model
{"tags": ["conversational"]}
hervetusse/DialogGPT-small-harrypotter
null
[ "transformers", "pytorch", "gpt2", "text-generation", "conversational", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Harry Potter DialoGPT Model
[ "# Harry Potter DialoGPT Model" ]
[ "TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Harry Potter DialoGPT Model" ]
text2text-generation
transformers
# T5-base for paraphrase generation Google's T5-base fine-tuned on [TaPaCo](https://huggingface.co/datasets/tapaco) dataset for paraphrasing. <!-- ## Model fine-tuning --> <!-- The training script is a slightly modified version of [this Colab Notebook](https://github.com/patil-suraj/exploring-T5/blob/master/t5_fine_tuning.ipynb) created by [Suraj Patil](https://github.com/patil-suraj), so all credits to him! --> ## Model in Action 🚀 ```python from transformers import T5ForConditionalGeneration, T5Tokenizer tokenizer = T5Tokenizer.from_pretrained("hetpandya/t5-base-tapaco") model = T5ForConditionalGeneration.from_pretrained("hetpandya/t5-base-tapaco") def get_paraphrases(sentence, prefix="paraphrase: ", n_predictions=5, top_k=120, max_length=256,device="cpu"): text = prefix + sentence + " </s>" encoding = tokenizer.encode_plus( text, pad_to_max_length=True, return_tensors="pt" ) input_ids, attention_masks = encoding["input_ids"].to(device), encoding[ "attention_mask" ].to(device) model_output = model.generate( input_ids=input_ids, attention_mask=attention_masks, do_sample=True, max_length=max_length, top_k=top_k, top_p=0.98, early_stopping=True, num_return_sequences=n_predictions, ) outputs = [] for output in model_output: generated_sent = tokenizer.decode( output, skip_special_tokens=True, clean_up_tokenization_spaces=True ) if ( generated_sent.lower() != sentence.lower() and generated_sent not in outputs ): outputs.append(generated_sent) return outputs paraphrases = get_paraphrases("The house will be cleaned by me every Saturday.") for sent in paraphrases: print(sent) ``` ## Output ``` The house will get cleaned for a whole week. The house is cleaning by me every weekend. What was going to do not get do with the house from me every Thursday. The house should be cleaned on Sunday--durse. It's time that I would be cleaning her house in tomorrow. ``` Created by [Het Pandya/@hetpandya](https://github.com/hetpandya) | [LinkedIn](https://www.linkedin.com/in/het-pandya) Made with <span style="color: red;">&hearts;</span> in India
{"language": "en", "datasets": ["tapaco"]}
hetpandya/t5-base-tapaco
null
[ "transformers", "pytorch", "safetensors", "t5", "text2text-generation", "en", "dataset:tapaco", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #safetensors #t5 #text2text-generation #en #dataset-tapaco #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# T5-base for paraphrase generation Google's T5-base fine-tuned on TaPaCo dataset for paraphrasing. ## Model in Action ## Output Created by Het Pandya/@hetpandya | LinkedIn Made with <span style="color: red;">&hearts;</span> in India
[ "# T5-base for paraphrase generation\n\nGoogle's T5-base fine-tuned on TaPaCo dataset for paraphrasing.", "## Model in Action", "## Output\n\n\nCreated by Het Pandya/@hetpandya | LinkedIn\n\nMade with <span style=\"color: red;\">&hearts;</span> in India" ]
[ "TAGS\n#transformers #pytorch #safetensors #t5 #text2text-generation #en #dataset-tapaco #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# T5-base for paraphrase generation\n\nGoogle's T5-base fine-tuned on TaPaCo dataset for paraphrasing.", "## Model in Action", "## Output\n\n\nCreated by Het Pandya/@hetpandya | LinkedIn\n\nMade with <span style=\"color: red;\">&hearts;</span> in India" ]
text2text-generation
transformers
# T5-small for paraphrase generation Google's T5-small fine-tuned on [Quora Question Pairs](https://huggingface.co/datasets/quora) dataset for paraphrasing. ## Model in Action 🚀 ```python from transformers import T5ForConditionalGeneration, T5Tokenizer tokenizer = T5Tokenizer.from_pretrained("hetpandya/t5-small-quora") model = T5ForConditionalGeneration.from_pretrained("hetpandya/t5-small-quora") def get_paraphrases(sentence, prefix="paraphrase: ", n_predictions=5, top_k=120, max_length=256,device="cpu"): text = prefix + sentence + " </s>" encoding = tokenizer.encode_plus( text, pad_to_max_length=True, return_tensors="pt" ) input_ids, attention_masks = encoding["input_ids"].to(device), encoding[ "attention_mask" ].to(device) model_output = model.generate( input_ids=input_ids, attention_mask=attention_masks, do_sample=True, max_length=max_length, top_k=top_k, top_p=0.98, early_stopping=True, num_return_sequences=n_predictions, ) outputs = [] for output in model_output: generated_sent = tokenizer.decode( output, skip_special_tokens=True, clean_up_tokenization_spaces=True ) if ( generated_sent.lower() != sentence.lower() and generated_sent not in outputs ): outputs.append(generated_sent) return outputs paraphrases = get_paraphrases("The house will be cleaned by me every Saturday.") for sent in paraphrases: print(sent) ``` ## Output ``` My house is up clean on Saturday morning. Thank you for this email. I'm introducing a new name and name. I'm running my house at home. I'm a taller myself. I'm gonna go with it on Monday. (the house will be up cleaned). Is there anything that will be cleaned every Saturday morning? The house is clean and will be cleaned each Saturday by my wife. I will clean the house for almost a week. I have to clean it all the weekend. I will be able to do it. My house is new. If I clean my house every Monday, I can call it clean. ``` Created by [Het Pandya/@hetpandya](https://github.com/hetpandya) | [LinkedIn](https://www.linkedin.com/in/het-pandya) Made with <span style="color: red;">&hearts;</span> in India
{"language": "en", "datasets": ["quora"]}
hetpandya/t5-small-quora
null
[ "transformers", "pytorch", "t5", "text2text-generation", "en", "dataset:quora", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #t5 #text2text-generation #en #dataset-quora #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# T5-small for paraphrase generation Google's T5-small fine-tuned on Quora Question Pairs dataset for paraphrasing. ## Model in Action ## Output Created by Het Pandya/@hetpandya | LinkedIn Made with <span style="color: red;">&hearts;</span> in India
[ "# T5-small for paraphrase generation\n\nGoogle's T5-small fine-tuned on Quora Question Pairs dataset for paraphrasing.", "## Model in Action", "## Output\n\n\nCreated by Het Pandya/@hetpandya | LinkedIn\n\nMade with <span style=\"color: red;\">&hearts;</span> in India" ]
[ "TAGS\n#transformers #pytorch #t5 #text2text-generation #en #dataset-quora #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# T5-small for paraphrase generation\n\nGoogle's T5-small fine-tuned on Quora Question Pairs dataset for paraphrasing.", "## Model in Action", "## Output\n\n\nCreated by Het Pandya/@hetpandya | LinkedIn\n\nMade with <span style=\"color: red;\">&hearts;</span> in India" ]
text2text-generation
transformers
# T5-small for paraphrase generation Google's T5 small fine-tuned on [TaPaCo](https://huggingface.co/datasets/tapaco) dataset for paraphrasing. ## Model in Action 🚀 ```python from transformers import T5ForConditionalGeneration, T5Tokenizer tokenizer = T5Tokenizer.from_pretrained("hetpandya/t5-small-tapaco") model = T5ForConditionalGeneration.from_pretrained("hetpandya/t5-small-tapaco") def get_paraphrases(sentence, prefix="paraphrase: ", n_predictions=5, top_k=120, max_length=256,device="cpu"): text = prefix + sentence + " </s>" encoding = tokenizer.encode_plus( text, pad_to_max_length=True, return_tensors="pt" ) input_ids, attention_masks = encoding["input_ids"].to(device), encoding[ "attention_mask" ].to(device) model_output = model.generate( input_ids=input_ids, attention_mask=attention_masks, do_sample=True, max_length=max_length, top_k=top_k, top_p=0.98, early_stopping=True, num_return_sequences=n_predictions, ) outputs = [] for output in model_output: generated_sent = tokenizer.decode( output, skip_special_tokens=True, clean_up_tokenization_spaces=True ) if ( generated_sent.lower() != sentence.lower() and generated_sent not in outputs ): outputs.append(generated_sent) return outputs paraphrases = get_paraphrases("The house will be cleaned by me every Saturday.") for sent in paraphrases: print(sent) ``` ## Output ``` The house is cleaned every Saturday by me. The house will be cleaned on Saturday. I will clean the house every Saturday. I get the house cleaned every Saturday. I will clean this house every Saturday. ``` ## Model fine-tuning Please find my guide on fine-tuning the model here: https://towardsdatascience.com/training-t5-for-paraphrase-generation-ab3b5be151a2 Created by [Het Pandya/@hetpandya](https://github.com/hetpandya) | [LinkedIn](https://www.linkedin.com/in/het-pandya) Made with <span style="color: red;">&hearts;</span> in India
{"language": "en", "datasets": ["tapaco"]}
hetpandya/t5-small-tapaco
null
[ "transformers", "pytorch", "t5", "text2text-generation", "en", "dataset:tapaco", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "en" ]
TAGS #transformers #pytorch #t5 #text2text-generation #en #dataset-tapaco #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# T5-small for paraphrase generation Google's T5 small fine-tuned on TaPaCo dataset for paraphrasing. ## Model in Action ## Output ## Model fine-tuning Please find my guide on fine-tuning the model here: URL Created by Het Pandya/@hetpandya | LinkedIn Made with <span style="color: red;">&hearts;</span> in India
[ "# T5-small for paraphrase generation\n\nGoogle's T5 small fine-tuned on TaPaCo dataset for paraphrasing.", "## Model in Action", "## Output", "## Model fine-tuning\nPlease find my guide on fine-tuning the model here:\n\nURL\n\n\nCreated by Het Pandya/@hetpandya | LinkedIn\n\nMade with <span style=\"color: red;\">&hearts;</span> in India" ]
[ "TAGS\n#transformers #pytorch #t5 #text2text-generation #en #dataset-tapaco #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# T5-small for paraphrase generation\n\nGoogle's T5 small fine-tuned on TaPaCo dataset for paraphrasing.", "## Model in Action", "## Output", "## Model fine-tuning\nPlease find my guide on fine-tuning the model here:\n\nURL\n\n\nCreated by Het Pandya/@hetpandya | LinkedIn\n\nMade with <span style=\"color: red;\">&hearts;</span> in India" ]
automatic-speech-recognition
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the MOZILLA-FOUNDATION/COMMON_VOICE_8_0 - SV-SE dataset. It achieves the following results on the evaluation set: **Without LM**: - Wer: 0.2465 - Cer: 0.0717 **With LM**: - Wer: 0.1710 - Cer: 0.0569 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 7.5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 50.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:-----:|:---------------:|:------:| | 3.3224 | 1.37 | 500 | 3.2676 | 1.0 | | 2.9319 | 2.74 | 1000 | 2.9287 | 1.0000 | | 2.1173 | 4.11 | 1500 | 1.1478 | 0.8788 | | 1.6973 | 5.48 | 2000 | 0.6749 | 0.6547 | | 1.5865 | 6.85 | 2500 | 0.5500 | 0.5634 | | 1.5094 | 8.22 | 3000 | 0.4840 | 0.5430 | | 1.4644 | 9.59 | 3500 | 0.4844 | 0.4142 | | 1.4061 | 10.96 | 4000 | 0.4356 | 0.3808 | | 1.3584 | 12.33 | 4500 | 0.4192 | 0.3698 | | 1.3438 | 13.7 | 5000 | 0.3980 | 0.3584 | | 1.3332 | 15.07 | 5500 | 0.3896 | 0.3572 | | 1.3025 | 16.44 | 6000 | 0.3835 | 0.3487 | | 1.2979 | 17.81 | 6500 | 0.3781 | 0.3417 | | 1.2736 | 19.18 | 7000 | 0.3734 | 0.3270 | | 1.2415 | 20.55 | 7500 | 0.3637 | 0.3316 | | 1.2255 | 21.92 | 8000 | 0.3546 | 0.3147 | | 1.2193 | 23.29 | 8500 | 0.3524 | 0.3196 | | 1.2104 | 24.66 | 9000 | 0.3403 | 0.3097 | | 1.1965 | 26.03 | 9500 | 0.3508 | 0.3093 | | 1.1976 | 27.4 | 10000 | 0.3419 | 0.3071 | | 1.182 | 28.77 | 10500 | 0.3364 | 0.2963 | | 1.158 | 30.14 | 11000 | 0.3338 | 0.2932 | | 1.1414 | 31.51 | 11500 | 0.3376 | 0.2940 | | 1.1402 | 32.88 | 12000 | 0.3370 | 0.2891 | | 1.1213 | 34.25 | 12500 | 0.3201 | 0.2874 | | 1.1207 | 35.62 | 13000 | 0.3261 | 0.2826 | | 1.1074 | 36.98 | 13500 | 0.3117 | 0.2786 | | 1.0818 | 38.36 | 14000 | 0.3194 | 0.2776 | | 1.0889 | 39.73 | 14500 | 0.3188 | 0.2738 | | 1.0672 | 41.1 | 15000 | 0.3196 | 0.2773 | | 1.0838 | 42.47 | 15500 | 0.3130 | 0.2739 | | 1.0553 | 43.83 | 16000 | 0.3165 | 0.2704 | | 1.0786 | 45.21 | 16500 | 0.3108 | 0.2706 | | 1.0546 | 46.57 | 17000 | 0.3102 | 0.2677 | | 1.0425 | 47.94 | 17500 | 0.3115 | 0.2679 | | 1.0398 | 49.31 | 18000 | 0.3131 | 0.2666 | ### Framework versions - Transformers 4.16.0.dev0 - Pytorch 1.10.1+cu113 - Datasets 1.18.1.dev0 - Tokenizers 0.10.3
{"language": ["sv-SE"], "license": "apache-2.0", "tags": ["automatic-speech-recognition", "mozilla-foundation/common_voice_8_0", "generated_from_trainer", "robust-speech-event", "hf-asr-leaderboard"], "datasets": ["mozilla-foundation/common_voice_8_0"], "model-index": [{"name": "XLS-R-300M - Swedish - CV8", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "Common Voice 8", "type": "mozilla-foundation/common_voice_8_0", "args": "sv-SE"}, "metrics": [{"type": "wer", "value": 17.1, "name": "Test WER"}, {"type": "cer", "value": 5.7, "name": "Test CER"}]}, {"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "Robust Speech Event - Dev Data", "type": "speech-recognition-community-v2/dev_data", "args": "sv"}, "metrics": [{"type": "wer", "value": 26.92, "name": "Test WER"}, {"type": "cer", "value": 12.53, "name": "Test CER"}]}]}]}
hf-test/xls-r-300m-sv-cv8
null
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "mozilla-foundation/common_voice_8_0", "generated_from_trainer", "robust-speech-event", "hf-asr-leaderboard", "dataset:mozilla-foundation/common_voice_8_0", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "sv-SE" ]
TAGS #transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #mozilla-foundation/common_voice_8_0 #generated_from_trainer #robust-speech-event #hf-asr-leaderboard #dataset-mozilla-foundation/common_voice_8_0 #license-apache-2.0 #model-index #endpoints_compatible #region-us
This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the MOZILLA-FOUNDATION/COMMON\_VOICE\_8\_0 - SV-SE dataset. It achieves the following results on the evaluation set: Without LM: * Wer: 0.2465 * Cer: 0.0717 With LM: * Wer: 0.1710 * Cer: 0.0569 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 7.5e-05 * train\_batch\_size: 8 * eval\_batch\_size: 8 * seed: 42 * gradient\_accumulation\_steps: 4 * total\_train\_batch\_size: 32 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 2000 * num\_epochs: 50.0 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.16.0.dev0 * Pytorch 1.10.1+cu113 * Datasets 1.18.1.dev0 * Tokenizers 0.10.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 7.5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 2000\n* num\\_epochs: 50.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.0.dev0\n* Pytorch 1.10.1+cu113\n* Datasets 1.18.1.dev0\n* Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #mozilla-foundation/common_voice_8_0 #generated_from_trainer #robust-speech-event #hf-asr-leaderboard #dataset-mozilla-foundation/common_voice_8_0 #license-apache-2.0 #model-index #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 7.5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 2000\n* num\\_epochs: 50.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.0.dev0\n* Pytorch 1.10.1+cu113\n* Datasets 1.18.1.dev0\n* Tokenizers 0.10.3" ]
automatic-speech-recognition
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # XLS-R-300m-SV This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the MOZILLA-FOUNDATION/COMMON_VOICE_7_0 - SV-SE dataset. It achieves the following results on the evaluation set: - Loss: 0.3171 - Wer: 0.2468 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 7.5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 2000 - num_epochs: 50.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:-----:|:---------------:|:------:| | 3.3349 | 1.45 | 500 | 3.2858 | 1.0 | | 2.9298 | 2.91 | 1000 | 2.9225 | 1.0000 | | 2.0839 | 4.36 | 1500 | 1.1546 | 0.8295 | | 1.7093 | 5.81 | 2000 | 0.6827 | 0.5701 | | 1.5855 | 7.27 | 2500 | 0.5597 | 0.4947 | | 1.4831 | 8.72 | 3000 | 0.4923 | 0.4527 | | 1.4416 | 10.17 | 3500 | 0.4670 | 0.4270 | | 1.3848 | 11.63 | 4000 | 0.4341 | 0.3980 | | 1.3749 | 13.08 | 4500 | 0.4203 | 0.4011 | | 1.3311 | 14.53 | 5000 | 0.4310 | 0.3961 | | 1.317 | 15.99 | 5500 | 0.3898 | 0.4322 | | 1.2799 | 17.44 | 6000 | 0.3806 | 0.3572 | | 1.2771 | 18.89 | 6500 | 0.3828 | 0.3427 | | 1.2451 | 20.35 | 7000 | 0.3702 | 0.3359 | | 1.2182 | 21.8 | 7500 | 0.3685 | 0.3270 | | 1.2152 | 23.26 | 8000 | 0.3650 | 0.3308 | | 1.1837 | 24.71 | 8500 | 0.3568 | 0.3187 | | 1.1721 | 26.16 | 9000 | 0.3659 | 0.3249 | | 1.1764 | 27.61 | 9500 | 0.3547 | 0.3145 | | 1.1606 | 29.07 | 10000 | 0.3514 | 0.3104 | | 1.1431 | 30.52 | 10500 | 0.3469 | 0.3062 | | 1.1047 | 31.97 | 11000 | 0.3313 | 0.2979 | | 1.1315 | 33.43 | 11500 | 0.3298 | 0.2992 | | 1.1022 | 34.88 | 12000 | 0.3296 | 0.2973 | | 1.0935 | 36.34 | 12500 | 0.3278 | 0.2926 | | 1.0676 | 37.79 | 13000 | 0.3208 | 0.2868 | | 1.0571 | 39.24 | 13500 | 0.3322 | 0.2885 | | 1.0536 | 40.7 | 14000 | 0.3245 | 0.2831 | | 1.0525 | 42.15 | 14500 | 0.3285 | 0.2826 | | 1.0464 | 43.6 | 15000 | 0.3223 | 0.2796 | | 1.0415 | 45.06 | 15500 | 0.3166 | 0.2774 | | 1.0356 | 46.51 | 16000 | 0.3177 | 0.2746 | | 1.04 | 47.96 | 16500 | 0.3150 | 0.2735 | | 1.0209 | 49.42 | 17000 | 0.3175 | 0.2731 | ### Framework versions - Transformers 4.16.0.dev0 - Pytorch 1.10.0+cu102 - Datasets 1.17.1.dev0 - Tokenizers 0.10.3 #### Evaluation Commands 1. To evaluate on `mozilla-foundation/common_voice_7_0` with split `test` ```bash python eval.py --model_id hf-test/xls-r-300m-sv --dataset mozilla-foundation/common_voice_7_0 --config sv-SE --split test ``` 2. To evaluate on `speech-recognition-community-v2/dev_data` ```bash python eval.py --model_id hf-test/xls-r-300m-sv --dataset speech-recognition-community-v2/dev_data --config sv --split validation --chunk_length_s 5.0 --stride_length_s 1.0 ``` ### Inference With LM ```python import torch from datasets import load_dataset from transformers import AutoModelForCTC, AutoProcessor import torchaudio.functional as F model_id = "hf-test/xls-r-300m-sv" sample_iter = iter(load_dataset("mozilla-foundation/common_voice_7_0", "sv-SE", split="test", streaming=True, use_auth_token=True)) sample = next(sample_iter) resampled_audio = F.resample(torch.tensor(sample["audio"]["array"]), 48_000, 16_000).numpy() model = AutoModelForCTC.from_pretrained(model_id) processor = AutoProcessor.from_pretrained(model_id) input_values = processor(resampled_audio, return_tensors="pt").input_values with torch.no_grad(): logits = model(input_values).logits transcription = processor.batch_decode(logits.numpy()).text # => "jag lämnade grovjobbet åt honom" ``` ### Eval results on Common Voice 7 "test" (WER): | Without LM | With LM (run `./eval.py`) | |---|---| | 24.68 | 16.98 |
{"language": ["sv-SE"], "license": "apache-2.0", "tags": ["automatic-speech-recognition", "generated_from_trainer", "hf-asr-leaderboard", "hello", "model_for_talk", "mozilla-foundation/common_voice_7_0", "robust-speech-event", "sv"], "datasets": ["mozilla-foundation/common_voice_7_0"], "model-index": [{"name": "XLS-R-300M - Swedish", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "Common Voice 7", "type": "mozilla-foundation/common_voice_7_0", "args": "sv-SE"}, "metrics": [{"type": "wer", "value": 16.98, "name": "Test WER"}, {"type": "cer", "value": 5.66, "name": "Test CER"}]}, {"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "Robust Speech Event - Dev Data", "type": "speech-recognition-community-v2/dev_data", "args": "sv"}, "metrics": [{"type": "wer", "value": 27.01, "name": "Test WER"}, {"type": "cer", "value": 13.14, "name": "Test CER"}]}]}]}
hf-test/xls-r-300m-sv
null
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "hf-asr-leaderboard", "hello", "model_for_talk", "mozilla-foundation/common_voice_7_0", "robust-speech-event", "sv", "dataset:mozilla-foundation/common_voice_7_0", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "sv-SE" ]
TAGS #transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #hf-asr-leaderboard #hello #model_for_talk #mozilla-foundation/common_voice_7_0 #robust-speech-event #sv #dataset-mozilla-foundation/common_voice_7_0 #license-apache-2.0 #model-index #endpoints_compatible #region-us
XLS-R-300m-SV ============= This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the MOZILLA-FOUNDATION/COMMON\_VOICE\_7\_0 - SV-SE dataset. It achieves the following results on the evaluation set: * Loss: 0.3171 * Wer: 0.2468 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 7.5e-05 * train\_batch\_size: 8 * eval\_batch\_size: 8 * seed: 42 * gradient\_accumulation\_steps: 4 * total\_train\_batch\_size: 32 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 2000 * num\_epochs: 50.0 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.16.0.dev0 * Pytorch 1.10.0+cu102 * Datasets 1.17.1.dev0 * Tokenizers 0.10.3 #### Evaluation Commands 1. To evaluate on 'mozilla-foundation/common\_voice\_7\_0' with split 'test' 2. To evaluate on 'speech-recognition-community-v2/dev\_data' ### Inference With LM ### Eval results on Common Voice 7 "test" (WER):
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 7.5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 2000\n* num\\_epochs: 50.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.0.dev0\n* Pytorch 1.10.0+cu102\n* Datasets 1.17.1.dev0\n* Tokenizers 0.10.3", "#### Evaluation Commands\n\n\n1. To evaluate on 'mozilla-foundation/common\\_voice\\_7\\_0' with split 'test'\n2. To evaluate on 'speech-recognition-community-v2/dev\\_data'", "### Inference With LM", "### Eval results on Common Voice 7 \"test\" (WER):" ]
[ "TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #hf-asr-leaderboard #hello #model_for_talk #mozilla-foundation/common_voice_7_0 #robust-speech-event #sv #dataset-mozilla-foundation/common_voice_7_0 #license-apache-2.0 #model-index #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 7.5e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 2000\n* num\\_epochs: 50.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.16.0.dev0\n* Pytorch 1.10.0+cu102\n* Datasets 1.17.1.dev0\n* Tokenizers 0.10.3", "#### Evaluation Commands\n\n\n1. To evaluate on 'mozilla-foundation/common\\_voice\\_7\\_0' with split 'test'\n2. To evaluate on 'speech-recognition-community-v2/dev\\_data'", "### Inference With LM", "### Eval results on Common Voice 7 \"test\" (WER):" ]
automatic-speech-recognition
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # This model is a fine-tuned version of [hf-test/xls-r-dummy](https://huggingface.co/hf-test/xls-r-dummy) on the MOZILLA-FOUNDATION/COMMON_VOICE_7_0 - AB dataset. It achieves the following results on the evaluation set: - Loss: 156.8787 - Wer: 1.3460 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 2 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - training_steps: 10 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.16.0.dev0 - Pytorch 1.10.0 - Datasets 1.16.1 - Tokenizers 0.10.3
{"language": ["ab"], "tags": ["automatic-speech-recognition", "mozilla-foundation/common_voice_7_0", "generated_from_trainer"], "datasets": ["common_voice"], "model-index": [{"name": "", "results": []}]}
hf-test/xls-r-ab-test
null
[ "transformers", "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "mozilla-foundation/common_voice_7_0", "generated_from_trainer", "ab", "dataset:common_voice", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[ "ab" ]
TAGS #transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #mozilla-foundation/common_voice_7_0 #generated_from_trainer #ab #dataset-common_voice #endpoints_compatible #region-us
# This model is a fine-tuned version of hf-test/xls-r-dummy on the MOZILLA-FOUNDATION/COMMON_VOICE_7_0 - AB dataset. It achieves the following results on the evaluation set: - Loss: 156.8787 - Wer: 1.3460 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0003 - train_batch_size: 2 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - training_steps: 10 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.16.0.dev0 - Pytorch 1.10.0 - Datasets 1.16.1 - Tokenizers 0.10.3
[ "# \n\nThis model is a fine-tuned version of hf-test/xls-r-dummy on the MOZILLA-FOUNDATION/COMMON_VOICE_7_0 - AB dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 156.8787\n- Wer: 1.3460", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0003\n- train_batch_size: 2\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- training_steps: 10\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.16.0.dev0\n- Pytorch 1.10.0\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #mozilla-foundation/common_voice_7_0 #generated_from_trainer #ab #dataset-common_voice #endpoints_compatible #region-us \n", "# \n\nThis model is a fine-tuned version of hf-test/xls-r-dummy on the MOZILLA-FOUNDATION/COMMON_VOICE_7_0 - AB dataset.\nIt achieves the following results on the evaluation set:\n- Loss: 156.8787\n- Wer: 1.3460", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0003\n- train_batch_size: 2\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- training_steps: 10\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- Transformers 4.16.0.dev0\n- Pytorch 1.10.0\n- Datasets 1.16.1\n- Tokenizers 0.10.3" ]
token-classification
transformers
# BERT base model (uncased) fine-tuned on CoNLL-2003 This model was trained following the PyTorch token-classification example from Hugging Face: https://github.com/huggingface/transformers/tree/master/examples/pytorch/token-classification. There were no tweaks to the model or dataset.
{}
hfeng/bert_base_uncased_conll2003
null
[ "transformers", "pytorch", "bert", "token-classification", "autotrain_compatible", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[]
[]
TAGS #transformers #pytorch #bert #token-classification #autotrain_compatible #endpoints_compatible #region-us
# BERT base model (uncased) fine-tuned on CoNLL-2003 This model was trained following the PyTorch token-classification example from Hugging Face: URL There were no tweaks to the model or dataset.
[ "# BERT base model (uncased) fine-tuned on CoNLL-2003\n\nThis model was trained following the PyTorch token-classification example from Hugging Face: URL\n\nThere were no tweaks to the model or dataset." ]
[ "TAGS\n#transformers #pytorch #bert #token-classification #autotrain_compatible #endpoints_compatible #region-us \n", "# BERT base model (uncased) fine-tuned on CoNLL-2003\n\nThis model was trained following the PyTorch token-classification example from Hugging Face: URL\n\nThere were no tweaks to the model or dataset." ]
fill-mask
transformers
## Chinese BERT with Whole Word Masking For further accelerating Chinese natural language processing, we provide **Chinese pre-trained BERT with Whole Word Masking**. **[Pre-Training with Whole Word Masking for Chinese BERT](https://arxiv.org/abs/1906.08101)** Yiming Cui, Wanxiang Che, Ting Liu, Bing Qin, Ziqing Yang, Shijin Wang, Guoping Hu This repository is developed based on:https://github.com/google-research/bert You may also interested in, - Chinese BERT series: https://github.com/ymcui/Chinese-BERT-wwm - Chinese MacBERT: https://github.com/ymcui/MacBERT - Chinese ELECTRA: https://github.com/ymcui/Chinese-ELECTRA - Chinese XLNet: https://github.com/ymcui/Chinese-XLNet - Knowledge Distillation Toolkit - TextBrewer: https://github.com/airaria/TextBrewer More resources by HFL: https://github.com/ymcui/HFL-Anthology ## Citation If you find the technical report or resource is useful, please cite the following technical report in your paper. - Primary: https://arxiv.org/abs/2004.13922 ``` @inproceedings{cui-etal-2020-revisiting, title = "Revisiting Pre-Trained Models for {C}hinese Natural Language Processing", author = "Cui, Yiming and Che, Wanxiang and Liu, Ting and Qin, Bing and Wang, Shijin and Hu, Guoping", booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: Findings", month = nov, year = "2020", address = "Online", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/2020.findings-emnlp.58", pages = "657--668", } ``` - Secondary: https://arxiv.org/abs/1906.08101 ``` @article{chinese-bert-wwm, title={Pre-Training with Whole Word Masking for Chinese BERT}, author={Cui, Yiming and Che, Wanxiang and Liu, Ting and Qin, Bing and Yang, Ziqing and Wang, Shijin and Hu, Guoping}, journal={arXiv preprint arXiv:1906.08101}, year={2019} } ```
{"language": ["zh"], "license": "apache-2.0"}
hfl/chinese-bert-wwm-ext
null
[ "transformers", "pytorch", "tf", "jax", "bert", "fill-mask", "zh", "arxiv:1906.08101", "arxiv:2004.13922", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
null
2022-03-02T23:29:05+00:00
[ "1906.08101", "2004.13922" ]
[ "zh" ]
TAGS #transformers #pytorch #tf #jax #bert #fill-mask #zh #arxiv-1906.08101 #arxiv-2004.13922 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
## Chinese BERT with Whole Word Masking For further accelerating Chinese natural language processing, we provide Chinese pre-trained BERT with Whole Word Masking. Pre-Training with Whole Word Masking for Chinese BERT Yiming Cui, Wanxiang Che, Ting Liu, Bing Qin, Ziqing Yang, Shijin Wang, Guoping Hu This repository is developed based on:URL You may also interested in, - Chinese BERT series: URL - Chinese MacBERT: URL - Chinese ELECTRA: URL - Chinese XLNet: URL - Knowledge Distillation Toolkit - TextBrewer: URL More resources by HFL: URL If you find the technical report or resource is useful, please cite the following technical report in your paper. - Primary: URL - Secondary: URL
[ "## Chinese BERT with Whole Word Masking\nFor further accelerating Chinese natural language processing, we provide Chinese pre-trained BERT with Whole Word Masking. \n\nPre-Training with Whole Word Masking for Chinese BERT \nYiming Cui, Wanxiang Che, Ting Liu, Bing Qin, Ziqing Yang, Shijin Wang, Guoping Hu\n\nThis repository is developed based on:URL\n\nYou may also interested in,\n- Chinese BERT series: URL\n- Chinese MacBERT: URL\n- Chinese ELECTRA: URL\n- Chinese XLNet: URL\n- Knowledge Distillation Toolkit - TextBrewer: URL\n\nMore resources by HFL: URL\n\nIf you find the technical report or resource is useful, please cite the following technical report in your paper.\n- Primary: URL\n\n- Secondary: URL" ]
[ "TAGS\n#transformers #pytorch #tf #jax #bert #fill-mask #zh #arxiv-1906.08101 #arxiv-2004.13922 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "## Chinese BERT with Whole Word Masking\nFor further accelerating Chinese natural language processing, we provide Chinese pre-trained BERT with Whole Word Masking. \n\nPre-Training with Whole Word Masking for Chinese BERT \nYiming Cui, Wanxiang Che, Ting Liu, Bing Qin, Ziqing Yang, Shijin Wang, Guoping Hu\n\nThis repository is developed based on:URL\n\nYou may also interested in,\n- Chinese BERT series: URL\n- Chinese MacBERT: URL\n- Chinese ELECTRA: URL\n- Chinese XLNet: URL\n- Knowledge Distillation Toolkit - TextBrewer: URL\n\nMore resources by HFL: URL\n\nIf you find the technical report or resource is useful, please cite the following technical report in your paper.\n- Primary: URL\n\n- Secondary: URL" ]
fill-mask
transformers
## Chinese BERT with Whole Word Masking For further accelerating Chinese natural language processing, we provide **Chinese pre-trained BERT with Whole Word Masking**. **[Pre-Training with Whole Word Masking for Chinese BERT](https://arxiv.org/abs/1906.08101)** Yiming Cui, Wanxiang Che, Ting Liu, Bing Qin, Ziqing Yang, Shijin Wang, Guoping Hu This repository is developed based on:https://github.com/google-research/bert You may also interested in, - Chinese BERT series: https://github.com/ymcui/Chinese-BERT-wwm - Chinese MacBERT: https://github.com/ymcui/MacBERT - Chinese ELECTRA: https://github.com/ymcui/Chinese-ELECTRA - Chinese XLNet: https://github.com/ymcui/Chinese-XLNet - Knowledge Distillation Toolkit - TextBrewer: https://github.com/airaria/TextBrewer More resources by HFL: https://github.com/ymcui/HFL-Anthology ## Citation If you find the technical report or resource is useful, please cite the following technical report in your paper. - Primary: https://arxiv.org/abs/2004.13922 ``` @inproceedings{cui-etal-2020-revisiting, title = "Revisiting Pre-Trained Models for {C}hinese Natural Language Processing", author = "Cui, Yiming and Che, Wanxiang and Liu, Ting and Qin, Bing and Wang, Shijin and Hu, Guoping", booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: Findings", month = nov, year = "2020", address = "Online", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/2020.findings-emnlp.58", pages = "657--668", } ``` - Secondary: https://arxiv.org/abs/1906.08101 ``` @article{chinese-bert-wwm, title={Pre-Training with Whole Word Masking for Chinese BERT}, author={Cui, Yiming and Che, Wanxiang and Liu, Ting and Qin, Bing and Yang, Ziqing and Wang, Shijin and Hu, Guoping}, journal={arXiv preprint arXiv:1906.08101}, year={2019} } ```
{"language": ["zh"], "license": "apache-2.0"}
hfl/chinese-bert-wwm
null
[ "transformers", "pytorch", "tf", "jax", "bert", "fill-mask", "zh", "arxiv:1906.08101", "arxiv:2004.13922", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
null
2022-03-02T23:29:05+00:00
[ "1906.08101", "2004.13922" ]
[ "zh" ]
TAGS #transformers #pytorch #tf #jax #bert #fill-mask #zh #arxiv-1906.08101 #arxiv-2004.13922 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
## Chinese BERT with Whole Word Masking For further accelerating Chinese natural language processing, we provide Chinese pre-trained BERT with Whole Word Masking. Pre-Training with Whole Word Masking for Chinese BERT Yiming Cui, Wanxiang Che, Ting Liu, Bing Qin, Ziqing Yang, Shijin Wang, Guoping Hu This repository is developed based on:URL You may also interested in, - Chinese BERT series: URL - Chinese MacBERT: URL - Chinese ELECTRA: URL - Chinese XLNet: URL - Knowledge Distillation Toolkit - TextBrewer: URL More resources by HFL: URL If you find the technical report or resource is useful, please cite the following technical report in your paper. - Primary: URL - Secondary: URL
[ "## Chinese BERT with Whole Word Masking\nFor further accelerating Chinese natural language processing, we provide Chinese pre-trained BERT with Whole Word Masking. \n\nPre-Training with Whole Word Masking for Chinese BERT \nYiming Cui, Wanxiang Che, Ting Liu, Bing Qin, Ziqing Yang, Shijin Wang, Guoping Hu\n\nThis repository is developed based on:URL\n\nYou may also interested in,\n- Chinese BERT series: URL\n- Chinese MacBERT: URL\n- Chinese ELECTRA: URL\n- Chinese XLNet: URL\n- Knowledge Distillation Toolkit - TextBrewer: URL\n\nMore resources by HFL: URL\n\nIf you find the technical report or resource is useful, please cite the following technical report in your paper.\n- Primary: URL\n\n- Secondary: URL" ]
[ "TAGS\n#transformers #pytorch #tf #jax #bert #fill-mask #zh #arxiv-1906.08101 #arxiv-2004.13922 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "## Chinese BERT with Whole Word Masking\nFor further accelerating Chinese natural language processing, we provide Chinese pre-trained BERT with Whole Word Masking. \n\nPre-Training with Whole Word Masking for Chinese BERT \nYiming Cui, Wanxiang Che, Ting Liu, Bing Qin, Ziqing Yang, Shijin Wang, Guoping Hu\n\nThis repository is developed based on:URL\n\nYou may also interested in,\n- Chinese BERT series: URL\n- Chinese MacBERT: URL\n- Chinese ELECTRA: URL\n- Chinese XLNet: URL\n- Knowledge Distillation Toolkit - TextBrewer: URL\n\nMore resources by HFL: URL\n\nIf you find the technical report or resource is useful, please cite the following technical report in your paper.\n- Primary: URL\n\n- Secondary: URL" ]
null
transformers
# This model is trained on 180G data, we recommend using this one than the original version. ## Chinese ELECTRA Google and Stanford University released a new pre-trained model called ELECTRA, which has a much compact model size and relatively competitive performance compared to BERT and its variants. For further accelerating the research of the Chinese pre-trained model, the Joint Laboratory of HIT and iFLYTEK Research (HFL) has released the Chinese ELECTRA models based on the official code of ELECTRA. ELECTRA-small could reach similar or even higher scores on several NLP tasks with only 1/10 parameters compared to BERT and its variants. This project is based on the official code of ELECTRA: [https://github.com/google-research/electra](https://github.com/google-research/electra) You may also interested in, - Chinese BERT series: https://github.com/ymcui/Chinese-BERT-wwm - Chinese ELECTRA: https://github.com/ymcui/Chinese-ELECTRA - Chinese XLNet: https://github.com/ymcui/Chinese-XLNet - Knowledge Distillation Toolkit - TextBrewer: https://github.com/airaria/TextBrewer More resources by HFL: https://github.com/ymcui/HFL-Anthology ## Citation If you find our resource or paper is useful, please consider including the following citation in your paper. - https://arxiv.org/abs/2004.13922 ``` @inproceedings{cui-etal-2020-revisiting, title = "Revisiting Pre-Trained Models for {C}hinese Natural Language Processing", author = "Cui, Yiming and Che, Wanxiang and Liu, Ting and Qin, Bing and Wang, Shijin and Hu, Guoping", booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: Findings", month = nov, year = "2020", address = "Online", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/2020.findings-emnlp.58", pages = "657--668", } ```
{"language": ["zh"], "license": "apache-2.0"}
hfl/chinese-electra-180g-base-discriminator
null
[ "transformers", "pytorch", "tf", "electra", "zh", "arxiv:2004.13922", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2022-03-02T23:29:05+00:00
[ "2004.13922" ]
[ "zh" ]
TAGS #transformers #pytorch #tf #electra #zh #arxiv-2004.13922 #license-apache-2.0 #endpoints_compatible #region-us
# This model is trained on 180G data, we recommend using this one than the original version. ## Chinese ELECTRA Google and Stanford University released a new pre-trained model called ELECTRA, which has a much compact model size and relatively competitive performance compared to BERT and its variants. For further accelerating the research of the Chinese pre-trained model, the Joint Laboratory of HIT and iFLYTEK Research (HFL) has released the Chinese ELECTRA models based on the official code of ELECTRA. ELECTRA-small could reach similar or even higher scores on several NLP tasks with only 1/10 parameters compared to BERT and its variants. This project is based on the official code of ELECTRA: URL You may also interested in, - Chinese BERT series: URL - Chinese ELECTRA: URL - Chinese XLNet: URL - Knowledge Distillation Toolkit - TextBrewer: URL More resources by HFL: URL If you find our resource or paper is useful, please consider including the following citation in your paper. - URL
[ "# This model is trained on 180G data, we recommend using this one than the original version.", "## Chinese ELECTRA\nGoogle and Stanford University released a new pre-trained model called ELECTRA, which has a much compact model size and relatively competitive performance compared to BERT and its variants.\nFor further accelerating the research of the Chinese pre-trained model, the Joint Laboratory of HIT and iFLYTEK Research (HFL) has released the Chinese ELECTRA models based on the official code of ELECTRA.\nELECTRA-small could reach similar or even higher scores on several NLP tasks with only 1/10 parameters compared to BERT and its variants.\n\nThis project is based on the official code of ELECTRA: URL\n\nYou may also interested in,\n- Chinese BERT series: URL\n- Chinese ELECTRA: URL\n- Chinese XLNet: URL\n- Knowledge Distillation Toolkit - TextBrewer: URL\n\nMore resources by HFL: URL\n\n\nIf you find our resource or paper is useful, please consider including the following citation in your paper.\n- URL" ]
[ "TAGS\n#transformers #pytorch #tf #electra #zh #arxiv-2004.13922 #license-apache-2.0 #endpoints_compatible #region-us \n", "# This model is trained on 180G data, we recommend using this one than the original version.", "## Chinese ELECTRA\nGoogle and Stanford University released a new pre-trained model called ELECTRA, which has a much compact model size and relatively competitive performance compared to BERT and its variants.\nFor further accelerating the research of the Chinese pre-trained model, the Joint Laboratory of HIT and iFLYTEK Research (HFL) has released the Chinese ELECTRA models based on the official code of ELECTRA.\nELECTRA-small could reach similar or even higher scores on several NLP tasks with only 1/10 parameters compared to BERT and its variants.\n\nThis project is based on the official code of ELECTRA: URL\n\nYou may also interested in,\n- Chinese BERT series: URL\n- Chinese ELECTRA: URL\n- Chinese XLNet: URL\n- Knowledge Distillation Toolkit - TextBrewer: URL\n\nMore resources by HFL: URL\n\n\nIf you find our resource or paper is useful, please consider including the following citation in your paper.\n- URL" ]