sha
stringlengths
40
40
text
stringlengths
0
13.4M
id
stringlengths
2
117
tags
list
created_at
stringlengths
25
25
metadata
stringlengths
2
31.7M
last_modified
stringlengths
25
25
6ebad500e4c26070bf0250887f2ea1add40535e9
# Dataset Card for "task-pages" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
osanseviero/task-pages
[ "region:us" ]
2022-11-18T06:58:02+00:00
{"dataset_info": {"features": [{"name": "CHANNEL_NAME", "dtype": "string"}, {"name": "URL", "dtype": "string"}, {"name": "TITLE", "dtype": "string"}, {"name": "DESCRIPTION", "dtype": "string"}, {"name": "TRANSCRIPTION", "dtype": "string"}, {"name": "SEGMENTS", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 27732, "num_examples": 2}], "download_size": 29958, "dataset_size": 27732}}
2022-11-18T07:01:05+00:00
1b7fe9e45386ba995a6e91128bcfd3b278fb7c42
# Dataset Card for "azure" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
osanseviero/azure
[ "task_categories:automatic-speech-recognition", "whisper", "whispering", "region:us" ]
2022-11-18T07:06:59+00:00
{"task_categories": ["automatic-speech-recognition"], "dataset_info": {"features": [{"name": "CHANNEL_NAME", "dtype": "string"}, {"name": "URL", "dtype": "string"}, {"name": "TITLE", "dtype": "string"}, {"name": "DESCRIPTION", "dtype": "string"}, {"name": "TRANSCRIPTION", "dtype": "string"}, {"name": "SEGMENTS", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 27732, "num_examples": 2}], "download_size": 29958, "dataset_size": 27732}, "tags": ["whisper", "whispering"]}
2022-11-18T07:07:02+00:00
744e1fd35ab07eb9d83860154b4298d453050009
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Binary Text Classification * Model: autoevaluate/binary-classification * Dataset: glue * Config: sst2 * Split: validation To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-2e778dac-2622-46c9-930e-6f9e705a27bf-2018
[ "autotrain", "evaluation", "region:us" ]
2022-11-18T10:00:57+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["glue"], "eval_info": {"task": "binary_classification", "model": "autoevaluate/binary-classification", "metrics": ["matthews_correlation"], "dataset_name": "glue", "dataset_config": "sst2", "dataset_split": "validation", "col_mapping": {"text": "sentence", "target": "label"}}}
2022-11-18T10:01:40+00:00
f1323cd10266ca8d8e135a7f567210d03a747139
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Binary Text Classification * Model: autoevaluate/binary-classification * Dataset: glue * Config: sst2 * Split: validation To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-61fd61be-9af8-4428-ac3c-2fe701ee60d1-2119
[ "autotrain", "evaluation", "region:us" ]
2022-11-18T10:10:19+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["glue"], "eval_info": {"task": "binary_classification", "model": "autoevaluate/binary-classification", "metrics": ["matthews_correlation"], "dataset_name": "glue", "dataset_config": "sst2", "dataset_split": "validation", "col_mapping": {"text": "sentence", "target": "label"}}}
2022-11-18T10:10:55+00:00
924ae6077edddf60f3ad2f2cbc54df3825a70930
# WikiCAT_es: Spanish Text Classification dataset ## Dataset Description - **Paper:** - **Point of Contact:** [email protected] **Repository** ### Dataset Summary WikiCAT_ca is a Spanish corpus for thematic Text Classification tasks. It is created automatically from Wikipedia and Wikidata sources, and contains 8401 articles from the Viquipedia classified under 12 different categories. This dataset was developed by BSC TeMU as part of the PlanTL project, and intended as an evaluation of LT capabilities to generate useful synthetic corpus. ### Supported Tasks and Leaderboards Text classification, Language Model ### Languages ES- Spanish ## Dataset Structure ### Data Instances Two json files, one for each split. ### Data Fields We used a simple model with the article text and associated labels, without further metadata. #### Example: <pre> {'sentence': 'La economía de Reunión se ha basado tradicionalmente en la agricultura. La caña de azúcar ha sido el cultivo principal durante más de un siglo, y en algunos años representa el 85% de las exportaciones. El gobierno ha estado impulsando el desarrollo de una industria turística para aliviar el alto desempleo, que representa más del 40% de la fuerza laboral.(...) El PIB total de la isla fue de 18.800 millones de dólares EE.UU. en 2007., 'label': 'Economía'} </pre> #### Labels 'Religión', 'Entretenimiento', 'Música', 'Ciencia_y_Tecnología', 'Política', 'Economía', 'Matemáticas', 'Humanidades', 'Deporte', 'Derecho', 'Historia', 'Filosofía' ### Data Splits * hfeval_esv5.json: 1681 label-document pairs * hftrain_esv5.json: 6716 label-document pairs ## Dataset Creation ### Methodology La páginas de "Categoría" representan los temas. para cada tema, extraemos las páginas asociadas a ese primer nivel de la jerarquía, y utilizamos el resúmen ("summary") como texto representativo. ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization The source data are thematic categories in the different Wikipedias #### Who are the source language producers? ### Annotations #### Annotation process Automatic annotation #### Who are the annotators? [N/A] ### Personal and Sensitive Information No personal or sensitive information included. ## Considerations for Using the Data ### Social Impact of Dataset We hope this corpus contributes to the development of language models in Spanish. ### Discussion of Biases We are aware that this data might contain biases. We have not applied any steps to reduce their impact. ### Other Known Limitations [N/A] ## Additional Information ### Dataset Curators Text Mining Unit (TeMU) at the Barcelona Supercomputing Center ([email protected]). For further information, send an email to ([email protected]). This work was funded by the [Spanish State Secretariat for Digitalization and Artificial Intelligence (SEDIA)](https://avancedigital.mineco.gob.es/en-us/Paginas/index.aspx) within the framework of the [Plan-TL](https://plantl.mineco.gob.es/Paginas/index.aspx). ### Licensing Information This work is licensed under [CC Attribution 4.0 International](https://creativecommons.org/licenses/by/4.0/) License. Copyright by the Spanish State Secretariat for Digitalization and Artificial Intelligence (SEDIA) (2022) ### Contributions [N/A]
PlanTL-GOB-ES/WikiCAT_esv2
[ "task_categories:text-classification", "task_ids:multi-class-classification", "annotations_creators:automatically-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:unknown", "language:es", "license:cc-by-sa-3.0", "region:us" ]
2022-11-18T10:18:53+00:00
{"annotations_creators": ["automatically-generated"], "language_creators": ["found"], "language": ["es"], "license": ["cc-by-sa-3.0"], "multilinguality": ["monolingual"], "size_categories": ["unknown"], "source_datasets": [], "task_categories": ["text-classification"], "task_ids": ["multi-class-classification"], "pretty_name": "wikicat_esv2"}
2023-07-27T08:13:16+00:00
facf3772e67d51b7d27508477f777565c6c720f5
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Binary Text Classification * Model: autoevaluate/binary-classification * Dataset: glue * Config: sst2 * Split: validation To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-ec388423-7e76-47a7-a778-e7cfff84a71c-2220
[ "autotrain", "evaluation", "region:us" ]
2022-11-18T10:28:05+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["glue"], "eval_info": {"task": "binary_classification", "model": "autoevaluate/binary-classification", "metrics": ["matthews_correlation"], "dataset_name": "glue", "dataset_config": "sst2", "dataset_split": "validation", "col_mapping": {"text": "sentence", "target": "label"}}}
2022-11-18T10:28:40+00:00
9c34d9a5a168152c846b6d3e07f34e97dff5f0bd
shiertier/conda
[ "license:other", "region:us" ]
2022-11-18T10:28:37+00:00
{"license": "other"}
2022-11-18T14:46:33+00:00
2278fe0ab44d4eaf561999862ac1a67ec0fbf4b7
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Binary Text Classification * Model: autoevaluate/binary-classification * Dataset: glue * Config: sst2 * Split: validation To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-fc4d51f7-9dde-4256-8b44-b5a68a081b2b-2321
[ "autotrain", "evaluation", "region:us" ]
2022-11-18T10:42:05+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["glue"], "eval_info": {"task": "binary_classification", "model": "autoevaluate/binary-classification", "metrics": ["matthews_correlation"], "dataset_name": "glue", "dataset_config": "sst2", "dataset_split": "validation", "col_mapping": {"text": "sentence", "target": "label"}}}
2022-11-18T10:42:41+00:00
2f1635ca28e9b62ec23b10311b737f02996d799b
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Binary Text Classification * Model: autoevaluate/binary-classification * Dataset: glue * Config: sst2 * Split: validation To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-ef91170b-c394-482c-8a00-6b7bc5ea5574-2422
[ "autotrain", "evaluation", "region:us" ]
2022-11-18T10:59:57+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["glue"], "eval_info": {"task": "binary_classification", "model": "autoevaluate/binary-classification", "metrics": ["matthews_correlation"], "dataset_name": "glue", "dataset_config": "sst2", "dataset_split": "validation", "col_mapping": {"text": "sentence", "target": "label"}}}
2022-11-18T11:00:32+00:00
febb64298be8a000cbc22029364c563b0b9c2105
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Binary Text Classification * Model: autoevaluate/binary-classification * Dataset: glue * Config: sst2 * Split: validation To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-622e0c30-b54d-415c-87b9-70c107d23cec-2523
[ "autotrain", "evaluation", "region:us" ]
2022-11-18T11:04:27+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["glue"], "eval_info": {"task": "binary_classification", "model": "autoevaluate/binary-classification", "metrics": ["matthews_correlation"], "dataset_name": "glue", "dataset_config": "sst2", "dataset_split": "validation", "col_mapping": {"text": "sentence", "target": "label"}}}
2022-11-18T11:05:03+00:00
b7cdd5ca43df8833b34d5ca2d4088051b6b82926
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Binary Text Classification * Model: autoevaluate/binary-classification * Dataset: glue * Config: sst2 * Split: validation To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-e6349348-5660-49a6-843b-4c305a6146f2-2624
[ "autotrain", "evaluation", "region:us" ]
2022-11-18T11:09:03+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["glue"], "eval_info": {"task": "binary_classification", "model": "autoevaluate/binary-classification", "metrics": ["matthews_correlation"], "dataset_name": "glue", "dataset_config": "sst2", "dataset_split": "validation", "col_mapping": {"text": "sentence", "target": "label"}}}
2022-11-18T11:09:39+00:00
bfc6a6071e5fc81c992e01784c8195aa1d23e910
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Binary Text Classification * Model: autoevaluate/binary-classification * Dataset: glue * Config: sst2 * Split: validation To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-6a0cd869-0e5a-4c97-8312-c7fea68b3609-2725
[ "autotrain", "evaluation", "region:us" ]
2022-11-18T11:19:34+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["glue"], "eval_info": {"task": "binary_classification", "model": "autoevaluate/binary-classification", "metrics": ["matthews_correlation"], "dataset_name": "glue", "dataset_config": "sst2", "dataset_split": "validation", "col_mapping": {"text": "sentence", "target": "label"}}}
2022-11-18T11:20:11+00:00
ceab8df9442a7649bc1f6656df30bda47e646039
prerana17/testing1
[ "license:afl-3.0", "region:us" ]
2022-11-18T11:25:01+00:00
{"license": "afl-3.0"}
2022-11-18T15:41:31+00:00
07b121a28410f42f09cf334d35a1042c48a60d67
# Citation ``` @article{DBLP:journals/corr/abs-1710-10196, author = {Tero Karras and Timo Aila and Samuli Laine and Jaakko Lehtinen}, title = {Progressive Growing of GANs for Improved Quality, Stability, and Variation}, journal = {CoRR}, volume = {abs/1710.10196}, year = {2017}, url = {http://arxiv.org/abs/1710.10196}, eprinttype = {arXiv}, eprint = {1710.10196}, timestamp = {Mon, 13 Aug 2018 16:46:42 +0200}, biburl = {https://dblp.org/rec/journals/corr/abs-1710-10196.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ```
Chris1/celebA-HQ
[ "arxiv:1710.10196", "region:us" ]
2022-11-18T14:01:17+00:00
{}
2022-11-18T14:30:59+00:00
23c0a29d28d0bbdb4c2bbcff56fda332379e69b0
# Dataset Card for "sidewalk-imagery" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
pattern123/sidewalk-imagery
[ "region:us" ]
2022-11-18T14:01:20+00:00
{"dataset_info": {"features": [{"name": "pixel_values", "dtype": "image"}, {"name": "label", "dtype": "image"}], "splits": [{"name": "train", "num_bytes": 3138394.0, "num_examples": 10}], "download_size": 3139599, "dataset_size": 3138394.0}}
2022-11-19T05:23:06+00:00
11be2afac6a84a3963bc34b66eb2c6fcfca4264e
arbml/arabic_pos_dialect
[ "license:apache-2.0", "region:us" ]
2022-11-18T14:17:01+00:00
{"license": "apache-2.0"}
2022-11-18T14:40:06+00:00
6dc072319bf629545aa7926c37d15c002d951c9e
awkwardneutrino/arkghn
[ "license:creativeml-openrail-m", "region:us" ]
2022-11-18T15:58:14+00:00
{"license": "creativeml-openrail-m"}
2022-11-18T16:00:46+00:00
21187d1891b1911eeb12022294a5681e28edb7eb
# Dataset Card for Dataset Name ## Dataset Description - **Homepage:** - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary This dataset card aims to be a base template for new datasets. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/datasetcard_template.md?plain=1). ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions [More Information Needed]
ccao/test
[ "region:us" ]
2022-11-18T16:17:38+00:00
{}
2023-01-19T05:02:01+00:00
2bc7902538956d23470bd31923ae3ff2d12757bb
# Dataset Card for "cstop_artificial" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
WillHeld/cstop_artificial
[ "region:us" ]
2022-11-18T20:13:29+00:00
{"dataset_info": {"features": [{"name": "utterance", "dtype": "string"}, {"name": " semantic_parse", "dtype": "string"}, {"name": "semantic_parse", "dtype": "string"}], "splits": [{"name": "eval", "num_bytes": 113084, "num_examples": 559}, {"name": "test", "num_bytes": 233020, "num_examples": 1167}, {"name": "train", "num_bytes": 819464, "num_examples": 4077}], "download_size": 371646, "dataset_size": 1165568}}
2022-12-10T17:54:16+00:00
9e0916d21f6fbedd8a1786e8be29b0df87b40bb1
<h4> Usage </h4> To use this embedding you have to download the file and put it into the "\stable-diffusion-webui\embeddings" folder To use it in a prompt add <em style="font-weight:600">art by rogue_style </em> add <b>[ ]</b> around it to reduce its weight. <h4> Included Files </h4> <ul> <li>500 steps <em>Usage: art by rogue_style-500</em></li> <li>3500 steps <em>Usage: art by rogue_style-3500</em></li> <li>6500 steps <em>Usage: art by rogue_style</em> </li> </ul> cheers<br> Wipeout <h4> Example Pictures </h4> <table> <tbody> <tr> <td><img height="100%/" width="100%" src="https://i.imgur.com/JefZ3cA.png"></td> <td><img height="100%/" width="100%" src="https://i.imgur.com/YBJzVIi.png"></td> <td><img height="100%/" width="100%" src="https://i.imgur.com/96iutfu.png"></td> <td><img height="100%/" width="100%" src="https://i.imgur.com/SBKfnc4.png"></td> </tr> </tbody> </table> <h4> prompt comparison </h4> <em> click the image to enlarge</em> <a href="https://i.imgur.com/a6te4zG.png" target="_blank"><img height="50%" width="50%" src="https://i.imgur.com/a6te4zG.png"></a>
zZWipeoutZz/rogue_style
[ "license:creativeml-openrail-m", "region:us" ]
2022-11-18T20:40:41+00:00
{"license": "creativeml-openrail-m"}
2022-11-19T15:03:02+00:00
74c07125220961838bfdecf0c0b205d32c471e1d
mamivac/db
[ "license:openrail", "region:us" ]
2022-11-18T21:17:16+00:00
{"license": "openrail"}
2022-11-18T21:17:16+00:00
45ae712ac42fa0209015db476c1d040e17442527
# Dataset Card for "highways-hacktum" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
ogimgio/highways-hacktum
[ "region:us" ]
2022-11-18T23:59:44+00:00
{"dataset_info": {"features": [{"name": "image", "dtype": "image"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "footway", "1": "primary"}}}}], "splits": [{"name": "train", "num_bytes": 1155915978.0, "num_examples": 500}, {"name": "validation", "num_bytes": 284161545.0, "num_examples": 125}], "download_size": 1431719317, "dataset_size": 1440077523.0}}
2022-11-19T00:04:45+00:00
74a44153625d0382b9d3c8af0a49a16e0c3cef0e
# Dataset Card for ravnursson_asr ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Ravnursson Faroese Speech and Transcripts](http://hdl.handle.net/20.500.12537/276) - **Repository:** [Clarin.is](http://hdl.handle.net/20.500.12537/276) - **Paper:** [ASR Language Resources for Faroese](https://aclanthology.org/2023.nodalida-1.4.pdf) - **Paper:** [Creating a basic language resource kit for faroese.](https://aclanthology.org/2022.lrec-1.495.pdf) - **Point of Contact:** [Annika Simonsen](mailto:[email protected]), [Carlos Mena](mailto:[email protected]) ### Dataset Summary The corpus "RAVNURSSON FAROESE SPEECH AND TRANSCRIPTS" (or RAVNURSSON Corpus for short) is a collection of speech recordings with transcriptions intended for Automatic Speech Recognition (ASR) applications in the language that is spoken at the Faroe Islands (Faroese). It was curated at the Reykjavík University (RU) in 2022. The RAVNURSSON Corpus is an extract of the "Basic Language Resource Kit 1.0" (BLARK 1.0) [1] developed by the Ravnur Project from the Faroe Islands [2]. As a matter of fact, the name RAVNURSSON comes from Ravnur (a tribute to the Ravnur Project) and the suffix "son" which in Icelandic means "son of". Therefore, the name "RAVNURSSON" means "The (Icelandic) son of Ravnur". The double "ss" is just for aesthetics. The audio was collected by recording speakers reading texts. The participants are aged 15-83, divided into 3 age groups: 15-35, 36-60 and 61+. The speech files are made of 249 female speakers and 184 male speakers; 433 speakers total. The recordings were made on TASCAM DR-40 Linear PCM audio recorders using the built-in stereo microphones in WAVE 16 bit with a sample rate of 48kHz, but then, downsampled to 16kHz@16bit mono for this corpus. [1] Simonsen, A., Debess, I. N., Lamhauge, S. S., & Henrichsen, P. J. Creating a basic language resource kit for Faroese. In LREC 2022. 13th International Conference on Language Resources and Evaluation. [2] Website. The Project Ravnur under the Talutøkni Foundation https://maltokni.fo/en/the-ravnur-project ### Example Usage The RAVNURSSON Corpus is divided in 3 splits: train, validation and test. To load a specific split pass its name as a config name: ```python from datasets import load_dataset ravnursson = load_dataset("carlosdanielhernandezmena/ravnursson_asr") ``` To load an specific split (for example, the validation split) do: ```python from datasets import load_dataset ravnursson = load_dataset("carlosdanielhernandezmena/ravnursson_asr",split="validation") ``` ### Supported Tasks automatic-speech-recognition: The dataset can be used to train a model for Automatic Speech Recognition (ASR). The model is presented with an audio file and asked to transcribe the audio file to written text. The most common evaluation metric is the word error rate (WER). ### Languages The audio is in Faroese. The reading prompts for the RAVNURSSON Corpus have been generated by expert linguists. The whole corpus was balanced for phonetic and dialectal coverage; Test and Dev subsets are gender-balanced. Tabular computer-searchable information is included as well as written documentation. ## Dataset Structure ### Data Instances ```python { 'audio_id': 'KAM06_151121_0101', 'audio': { 'path': '/home/carlos/.cache/HuggingFace/datasets/downloads/extracted/32b4a757027b72b8d2e25cd9c8be9c7c919cc8d4eb1a9a899e02c11fd6074536/dev/RDATA2/KAM06_151121/KAM06_151121_0101.flac', 'array': array([ 0.0010376 , -0.00521851, -0.00393677, ..., 0.00128174, 0.00076294, 0.00045776], dtype=float32), 'sampling_rate': 16000 }, 'speaker_id': 'KAM06_151121', 'gender': 'female', 'age': '36-60', 'duration': 4.863999843597412, 'normalized_text': 'endurskin eru týdningarmikil í myrkri', 'dialect': 'sandoy' } ``` ### Data Fields * `audio_id` (string) - id of audio segment * `audio` (datasets.Audio) - a dictionary containing the path to the audio, the decoded audio array, and the sampling rate. In non-streaming mode (default), the path points to the locally extracted audio. In streaming mode, the path is the relative path of an audio inside its archive (as files are not downloaded and extracted locally). * `speaker_id` (string) - id of speaker * `gender` (string) - gender of speaker (male or female) * `age` (string) - range of age of the speaker: Younger (15-35), Middle-aged (36-60) or Elderly (61+). * `duration` (float32) - duration of the audio file in seconds. * `normalized_text` (string) - normalized audio segment transcription * `dialect` (string) - dialect group, for example "Suðuroy" or "Sandoy". ### Data Splits The speech material has been subdivided into portions for training (train), development (evaluation) and testing (test). Lengths of each portion are: train = 100h08m, test = 4h30m, dev (evaluation)=4h30m. To load an specific portion please see the above section "Example Usage". The development and test portions have exactly 10 male and 10 female speakers each and both portions have exactly the same size in hours (4.5h each). ## Dataset Creation ### Curation Rationale The directory called "speech" contains all the speech files of the corpus. The files in the speech directory are divided in three directories: train, dev and test. The train portion is sub-divided in three types of recordings: RDATA1O, RDATA1OP and RDATA2; this is due to the organization of the recordings in the original BLARK 1.0. There, the recordings are divided in Rdata1 and Rdata2. One main difference between Rdata1 and Rdata2 is that the reading environment for Rdata2 was controlled by a software called "PushPrompt" which is included in the original BLARK 1.0. Another main difference is that in Rdata1 there are some available transcriptions labeled at the phoneme level. For this reason the audio files in the speech directory of the RAVNURSSON corpus are divided in the folders RDATA1O where "O" is for "Orthographic" and RDATA1OP where "O" is for Orthographic and "P" is for phonetic. In the case of the dev and test portions, the data come only from Rdata2 which does not have labels at the phonetic level. It is important to clarify that the RAVNURSSON Corpus only includes transcriptions at the orthographic level. ### Source Data #### Initial Data Collection and Normalization The dataset was released with normalized text only at an orthographic level in lower-case. The normalization process was performed by automatically removing punctuation marks and characters that are not present in the Faroese alphabet. #### Who are the source language producers? * The utterances were recorded using a TASCAM DR-40. * Participants self-reported their age group, gender, native language and dialect. * Participants are aged between 15 to 83 years. * The corpus contains 71949 speech files from 433 speakers, totalling 109 hours and 9 minutes. ### Annotations #### Annotation process Most of the reading prompts were selected by experts from a Faroese text corpus (news, blogs, Wikipedia etc.) and were edited to fit the format. Reading prompts that are within specific domains (such as Faroese place names, numbers, license plates, telling time etc.) were written by the Ravnur Project. Then, a software tool called PushPrompt were used for reading sessions (voice recordings). PushPromt presents the text items in the reading material to the reader, allowing him/her to manage the session interactively (adjusting the reading tempo, repeating speech productions at wish, inserting short breaks as needed, etc.). When the reading session is completed, a log file (with time stamps for each production) is written as a data table compliant with the TextGrid-format. #### Who are the annotators? The corpus was annotated by the [Ravnur Project](https://maltokni.fo/en/the-ravnur-project) ### Personal and Sensitive Information The dataset consists of people who have donated their voice. You agree to not attempt to determine the identity of speakers in this dataset. ## Considerations for Using the Data ### Social Impact of Dataset This is the first ASR corpus in Faroese. ### Discussion of Biases As the number of reading prompts was limited, the common denominator in the RAVNURSSON corpus is that one prompt is read by more than one speaker. This is relevant because is a common practice in ASR to create a language model using the prompts that are found in the train portion of the corpus. That is not recommended for the RAVNURSSON Corpus as it counts with many prompts shared by all the portions and that will produce an important bias in the language modeling task. In this section we present some statistics about the repeated prompts through all the portions of the corpus. - In the train portion: * Total number of prompts = 65616 * Number of unique prompts = 38646 There are 26970 repeated prompts in the train portion. In other words, 41.1% of the prompts are repeated. - In the test portion: * Total number of prompts = 3002 * Number of unique prompts = 2887 There are 115 repeated prompts in the test portion. In other words, 3.83% of the prompts are repeated. - In the dev portion: * Total number of prompts = 3331 * Number of unique prompts = 3302 There are 29 repeated prompts in the dev portion. In other words, 0.87% of the prompts are repeated. - Considering the corpus as a whole: * Total number of prompts = 71949 * Number of unique prompts = 39945 There are 32004 repeated prompts in the whole corpus. In other words, 44.48% of the prompts are repeated. NOTICE!: It is also important to clarify that none of the 3 portions of the corpus share speakers. ### Other Known Limitations "RAVNURSSON FAROESE SPEECH AND TRANSCRIPTS" by Carlos Daniel Hernández Mena and Annika Simonsen is licensed under a Creative Commons Attribution 4.0 International (CC BY 4.0) License with the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. ## Additional Information ### Dataset Curators The dataset was collected by Annika Simonsen and curated by Carlos Daniel Hernández Mena. ### Licensing Information [CC-BY-4.0](https://creativecommons.org/licenses/by/4.0/) ### Citation Information ``` @misc{carlosmenaravnursson2022, title={Ravnursson Faroese Speech and Transcripts}, author={Hernandez Mena, Carlos Daniel and Simonsen, Annika}, year={2022}, url={http://hdl.handle.net/20.500.12537/276}, } ``` ### Contributions This project was made possible under the umbrella of the Language Technology Programme for Icelandic 2019-2023. The programme, which is managed and coordinated by Almannarómur, is funded by the Icelandic Ministry of Education, Science and Culture. Special thanks to Dr. Jón Guðnason, professor at Reykjavík University and head of the Language and Voice Lab (LVL) for providing computational resources.
carlosdanielhernandezmena/ravnursson_asr
[ "task_categories:automatic-speech-recognition", "annotations_creators:expert-generated", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:fo", "license:cc-by-4.0", "faroe islands", "faroese", "ravnur project", "speech recognition in faroese", "region:us" ]
2022-11-19T00:02:04+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["expert-generated"], "language": ["fo"], "license": ["cc-by-4.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["automatic-speech-recognition"], "task_ids": [], "pretty_name": "RAVNURSSON FAROESE SPEECH AND TRANSCRIPTS", "tags": ["faroe islands", "faroese", "ravnur project", "speech recognition in faroese"]}
2023-07-10T20:20:03+00:00
45f2b87fa811b809c4482efc897ab9580fa3f910
munozariasjm/tab_pib_4_7
[ "license:mit", "region:us" ]
2022-11-19T02:31:12+00:00
{"license": "mit"}
2022-11-30T10:40:06+00:00
b908bad5ef0759d2c03baf09715a98aedda9ded1
# Dataset Card for kodak ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** <https://r0k.us/graphics/kodak/> - **Repository:** <https://github.com/MohamedBakrAli/Kodak-Lossless-True-Color-Image-Suite> - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary The pictures below link to lossless, true color (24 bits per pixel, aka "full color") images. It is my understanding they have been released by the Eastman Kodak Company for unrestricted usage. Many sites use them as a standard test suite for compression testing, etc. Prior to this site, they were only available in the Sun Raster format via ftp. This meant that the images could not be previewed before downloading. Since their release, however, the lossless PNG format has been incorporated into all the major browsers. Since PNG supports 24-bit lossless color (which GIF and JPEG do not), it is now possible to offer this browser-friendly access to the images. ### Supported Tasks and Leaderboards - Image compression ### Languages - en ## Dataset Structure ### Data Instances - [![kodak01](https://r0k.us/graphics/kodak/thumbs/kodim01t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim01.png) - [![kodak02](https://r0k.us/graphics/kodak/thumbs/kodim02t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim02.png) - [![kodak03](https://r0k.us/graphics/kodak/thumbs/kodim03t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim03.png) - [![kodak04](https://r0k.us/graphics/kodak/thumbs/kodim04t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim04.png) - [![kodak05](https://r0k.us/graphics/kodak/thumbs/kodim05t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim05.png) - [![kodak06](https://r0k.us/graphics/kodak/thumbs/kodim06t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim06.png) - [![kodak07](https://r0k.us/graphics/kodak/thumbs/kodim07t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim07.png) - [![kodak08](https://r0k.us/graphics/kodak/thumbs/kodim08t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim08.png) - [![kodak09](https://r0k.us/graphics/kodak/thumbs/kodim09t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim09.png) - [![kodak10](https://r0k.us/graphics/kodak/thumbs/kodim10t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim10.png) - [![kodak11](https://r0k.us/graphics/kodak/thumbs/kodim11t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim11.png) - [![kodak12](https://r0k.us/graphics/kodak/thumbs/kodim12t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim12.png) - [![kodak13](https://r0k.us/graphics/kodak/thumbs/kodim13t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim13.png) - [![kodak14](https://r0k.us/graphics/kodak/thumbs/kodim14t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim14.png) - [![kodak15](https://r0k.us/graphics/kodak/thumbs/kodim15t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim15.png) - [![kodak16](https://r0k.us/graphics/kodak/thumbs/kodim16t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim16.png) - [![kodak17](https://r0k.us/graphics/kodak/thumbs/kodim17t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim17.png) - [![kodak18](https://r0k.us/graphics/kodak/thumbs/kodim18t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim18.png) - [![kodak19](https://r0k.us/graphics/kodak/thumbs/kodim19t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim19.png) - [![kodak20](https://r0k.us/graphics/kodak/thumbs/kodim20t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim20.png) - [![kodak21](https://r0k.us/graphics/kodak/thumbs/kodim21t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim21.png) - [![kodak22](https://r0k.us/graphics/kodak/thumbs/kodim22t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim22.png) - [![kodak23](https://r0k.us/graphics/kodak/thumbs/kodim23t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim23.png) - [![kodak24](https://r0k.us/graphics/kodak/thumbs/kodim24t.jpg)](https://r0k.us/graphics/kodak/kodak/kodim24.png) ### Data Fields ### Data Splits ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization #### Who are the source language producers? <https://www.kodak.com> ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information ## Considerations for Using the Data ### Social Impact of Dataset ### Discussion of Biases ### Other Known Limitations ## Additional Information ### Dataset Curators ### Licensing Information [LICENSE](LICENSE) ### Citation Information ### Contributions Thanks to [@Freed-Wu](https://github.com/Freed-Wu) for adding this dataset.
Freed-Wu/kodak
[ "task_categories:other", "annotations_creators:no-annotation", "language_creators:found", "multilinguality:monolingual", "size_categories:n<1K", "source_datasets:original", "language:en", "license:gpl-3.0", "region:us" ]
2022-11-19T05:43:53+00:00
{"annotations_creators": ["no-annotation"], "language_creators": ["found"], "language": ["en"], "license": ["gpl-3.0"], "multilinguality": ["monolingual"], "size_categories": ["n<1K"], "source_datasets": ["original"], "task_categories": ["other"], "task_ids": [], "pretty_name": "kodak", "tags": [], "dataset_info": {"features": [{"name": "image", "dtype": "image"}], "splits": [{"name": "test", "num_bytes": 15072, "num_examples": 24}], "download_size": 15072, "dataset_size": 15072}}
2022-11-19T05:43:53+00:00
40df846508452e5b666a40f5def1e9dddd321975
uetchy/thesession
[ "license:odbl", "region:us" ]
2022-11-19T07:14:42+00:00
{"license": "odbl"}
2022-11-19T07:14:42+00:00
aadde6fe7f3a14364dcf4ed61b6173625beffead
# Dataset Card for "koikatsu-cards" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
chenrm/koikatsu-cards
[ "region:us" ]
2022-11-19T08:54:34+00:00
{"dataset_info": {"features": [{"name": "image", "dtype": "image"}], "splits": [{"name": "train", "num_bytes": 43368873054.078, "num_examples": 10178}, {"name": "test", "num_bytes": 20733059.0, "num_examples": 5}], "download_size": 56731523062, "dataset_size": 43389606113.078}}
2022-11-19T10:33:28+00:00
4eaa49ac06038400da1437d5cd98686ac3712ab0
# Dataset Card for "del" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
galman33/del
[ "region:us" ]
2022-11-19T12:26:10+00:00
{"dataset_info": {"features": [{"name": "lat", "dtype": "float64"}, {"name": "lon", "dtype": "float64"}, {"name": "country_code", "dtype": "string"}, {"name": "pixels", "dtype": {"array3_d": {"shape": [256, 256, 3], "dtype": "uint8"}}}], "splits": [{"name": "train", "num_bytes": 3816359256, "num_examples": 8300}], "download_size": 1455177025, "dataset_size": 3816359256}}
2022-11-19T12:45:41+00:00
4f15697f40bdb5be4c583942d825e48627d4bef5
# Dataset Card for "gal_yair_83000_256x256" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
galman33/gal_yair_83000_256x256
[ "region:us" ]
2022-11-19T12:29:05+00:00
{"dataset_info": {"features": [{"name": "lat", "dtype": "float64"}, {"name": "lon", "dtype": "float64"}, {"name": "country_code", "dtype": "string"}, {"name": "image", "dtype": "image"}], "splits": [{"name": "train", "num_bytes": 8075570913.0, "num_examples": 83000}], "download_size": 8075813262, "dataset_size": 8075570913.0}}
2022-11-19T12:33:25+00:00
24aea2a13efaddb810cdff6ae2b133ccb565d025
Hugging Face's logo Hugging Face Search models, datasets, users... Models Datasets Spaces Docs Solutions Pricing Hugging Face is way more fun with friends and colleagues! 🤗 Join an organization Datasets: Mostafa3zazi / Arabic_SQuAD Copied like 0 Dataset card Files and versions Community Arabic_SQuAD / README.md Mostafa3zazi's picture Mostafa3zazi Update README.md 17d5b9d 19 days ago raw history blame contribute delete Safe 2.18 kB --- dataset_info: features: - name: index dtype: string - name: question dtype: string - name: context dtype: string - name: text dtype: string - name: answer_start dtype: int64 - name: c_id dtype: int64 splits: - name: train num_bytes: 61868003 num_examples: 48344 download_size: 10512179 dataset_size: 61868003 --- # Dataset Card for "Arabic_SQuAD" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) --- # Citation ``` @inproceedings{mozannar-etal-2019-neural, title = "Neural {A}rabic Question Answering", author = "Mozannar, Hussein and Maamary, Elie and El Hajal, Karl and Hajj, Hazem", booktitle = "Proceedings of the Fourth Arabic Natural Language Processing Workshop", month = aug, year = "2019", address = "Florence, Italy", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W19-4612", doi = "10.18653/v1/W19-4612", pages = "108--118", abstract = "This paper tackles the problem of open domain factual Arabic question answering (QA) using Wikipedia as our knowledge source. This constrains the answer of any question to be a span of text in Wikipedia. Open domain QA for Arabic entails three challenges: annotated QA datasets in Arabic, large scale efficient information retrieval and machine reading comprehension. To deal with the lack of Arabic QA datasets we present the Arabic Reading Comprehension Dataset (ARCD) composed of 1,395 questions posed by crowdworkers on Wikipedia articles, and a machine translation of the Stanford Question Answering Dataset (Arabic-SQuAD). Our system for open domain question answering in Arabic (SOQAL) is based on two components: (1) a document retriever using a hierarchical TF-IDF approach and (2) a neural reading comprehension model using the pre-trained bi-directional transformer BERT. Our experiments on ARCD indicate the effectiveness of our approach with our BERT-based reader achieving a 61.3 F1 score, and our open domain system SOQAL achieving a 27.6 F1 score.", } ``` ---
ZIZOU/Arabic_Squad
[ "region:us" ]
2022-11-19T12:58:41+00:00
{}
2022-11-26T09:36:45+00:00
e9805e363e41f46225021bc9f3da6d0b0483b8e1
This dataset is for use in Automatic Speech Recognition (ASR) for a project at University of Zambia(UNZA)
unza/unza-nyanja
[ "region:us" ]
2022-11-19T14:18:39+00:00
{}
2022-11-19T17:42:38+00:00
d7c094f2a6ae22d41a3c143b42033e61c6ecfd72
# Dataset Card for "L1_poleval_korpus_pelny_train" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
Nikutka/L1_poleval_korpus_pelny_train
[ "region:us" ]
2022-11-19T14:43:15+00:00
{"dataset_info": {"features": [{"name": "content", "dtype": "string"}, {"name": "label", "dtype": "int64"}], "splits": [{"name": "train", "num_bytes": 764265, "num_examples": 9443}], "download_size": 509113, "dataset_size": 764265}}
2022-11-19T18:25:13+00:00
56bfdb5641038be046b772f3378e9b45b06a6bb2
# Dataset Card for "L1_poleval_korpus_pelny_test" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
Nikutka/L1_poleval_korpus_pelny_test
[ "region:us" ]
2022-11-19T14:43:36+00:00
{"dataset_info": {"features": [{"name": "content", "dtype": "string"}, {"name": "label", "dtype": "int64"}], "splits": [{"name": "train", "num_bytes": 71297, "num_examples": 891}], "download_size": 47500, "dataset_size": 71297}}
2022-11-19T18:25:17+00:00
c4d9f2faf033cc7f4f86b65814ae21df6e2ba768
# Dataset Card for "L1_poleval_korpus_wzorcowy_train" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
Nikutka/L1_poleval_korpus_wzorcowy_train
[ "region:us" ]
2022-11-19T14:55:23+00:00
{"dataset_info": {"features": [{"name": "content", "dtype": "string"}, {"name": "label", "dtype": "int64"}], "splits": [{"name": "train", "num_bytes": 20564, "num_examples": 253}], "download_size": 15381, "dataset_size": 20564}}
2022-11-19T18:25:28+00:00
85e49f1e1bc10c9cacc44a24d17112d42fb09ccb
# Dataset Card for "L1_poleval_korpus_wzorcowy_test" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
Nikutka/L1_poleval_korpus_wzorcowy_test
[ "region:us" ]
2022-11-19T14:55:27+00:00
{"dataset_info": {"features": [{"name": "content", "dtype": "string"}, {"name": "label", "dtype": "int64"}], "splits": [{"name": "train", "num_bytes": 1963, "num_examples": 25}], "download_size": 2784, "dataset_size": 1963}}
2022-11-19T18:25:32+00:00
3a031ae539b268125dad64f3ac9d559bd9ea9e22
# Dataset Card for "gal_yair_83000_100x100" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
galman33/gal_yair_83000_100x100
[ "region:us" ]
2022-11-19T14:56:55+00:00
{"dataset_info": {"features": [{"name": "lat", "dtype": "float64"}, {"name": "lon", "dtype": "float64"}, {"name": "country_code", "dtype": "string"}, {"name": "image", "dtype": "image"}], "splits": [{"name": "train", "num_bytes": 1423239502.0, "num_examples": 83000}], "download_size": 1423108777, "dataset_size": 1423239502.0}}
2022-11-19T14:57:47+00:00
88c830071713a2d643fe625eb6bb49fdc3b638a9
Juanchoxs/model1
[ "license:openrail", "region:us" ]
2022-11-19T15:20:07+00:00
{"license": "openrail"}
2022-11-19T15:21:08+00:00
e8aa28670485bbe09d7cc3ce8a47ea4186f01429
# Dataset Card for FIB ## Dataset Summary The FIB benchmark consists of 3579 examples for evaluating the factual inconsistency of large language models. Each example consists of a document and a pair of summaries: a factually consistent one and a factually inconsistent one. It is based on documents and summaries from XSum and CNN/DM. Since this dataset is intended to evaluate the factual inconsistency of large language models, there is only a test split. Accuracies should be reported separately for examples from XSum and for examples from CNN/DM. This is because the behavior of models on XSum and CNN/DM are expected to be very different. The factually inconsistent summaries are model-extracted from the document for CNN/DM but are model-generated for XSum. ### Citation Information ``` @article{tam2022fib, title={Evaluating the Factual Consistency of Large Language Models Through Summarization}, author={Tam, Derek and Mascarenhas, Anisha and Zhang, Shiyue and Kwan, Sarah and Bansal, Mohit and Raffel, Colin}, journal={arXiv preprint arXiv:2211.08412}, year={2022} } ``` ### Licensing Information license: cc-by-4.0
r-three/fib
[ "region:us" ]
2022-11-19T15:22:00+00:00
{}
2022-11-19T15:57:58+00:00
75e773476edee6fafc144ee1e0833fc06e106326
nlhappy/CLUE-NER
[ "license:mit", "region:us" ]
2022-11-19T15:36:25+00:00
{"license": "mit"}
2022-11-19T15:43:09+00:00
93d367c80cbd29aad9c9412a95b95ec782509b39
два датасета: один с оригинальными нуждиками и один поменьше с пупами на основе нуждиков. оба в аудио и текстовом формате.
4eJIoBek/nujdiki
[ "license:wtfpl", "region:us" ]
2022-11-19T16:23:30+00:00
{"license": "wtfpl"}
2023-02-13T15:35:25+00:00
51fbabf0496b056d0e018b99157a29261d6b9a93
# Dataset Card for "L1_scraped_korpus_pelny_train" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
Nikutka/L1_scraped_korpus_pelny_train
[ "region:us" ]
2022-11-19T16:38:58+00:00
{"dataset_info": {"features": [{"name": "content", "dtype": "string"}, {"name": "label", "dtype": "int64"}], "splits": [{"name": "train", "num_bytes": 118294409, "num_examples": 1249536}], "download_size": 86623523, "dataset_size": 118294409}}
2022-11-19T17:15:59+00:00
67a3ee8544898d09a5aefd43e59d2461a8231a65
# Dataset Card for "L1_scraped_korpus_pelny_test" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
Nikutka/L1_scraped_korpus_pelny_test
[ "region:us" ]
2022-11-19T16:39:05+00:00
{"dataset_info": {"features": [{"name": "content", "dtype": "string"}, {"name": "label", "dtype": "int64"}], "splits": [{"name": "train", "num_bytes": 29613125, "num_examples": 312385}], "download_size": 21671824, "dataset_size": 29613125}}
2022-11-19T17:16:03+00:00
208f15b308d39fe537796a914044a9d1f9656a23
# Dataset Card for "L1_scraped_korpus_wzorcowy_train" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
Nikutka/L1_scraped_korpus_wzorcowy_train
[ "region:us" ]
2022-11-19T16:39:36+00:00
{"dataset_info": {"features": [{"name": "content", "dtype": "string"}, {"name": "label", "dtype": "int64"}], "splits": [{"name": "train", "num_bytes": 4838134, "num_examples": 29488}], "download_size": 3466828, "dataset_size": 4838134}}
2022-11-19T17:18:32+00:00
7e3a8ff3857ecff1c74aaba90a379cb57dad6cae
# Dataset Card for "L1_scraped_korpus_wzorcowy_test" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
Nikutka/L1_scraped_korpus_wzorcowy_test
[ "region:us" ]
2022-11-19T16:39:40+00:00
{"dataset_info": {"features": [{"name": "content", "dtype": "string"}, {"name": "label", "dtype": "int64"}], "splits": [{"name": "train", "num_bytes": 1207567, "num_examples": 7372}], "download_size": 865883, "dataset_size": 1207567}}
2022-11-19T17:18:35+00:00
9dede1f6ae1d869828ed07cd1ef20c07ec4d6b2f
# Dataset Card for "L1_scraped_korpus_wzorcowy" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
Nikutka/L1_scraped_korpus_wzorcowy
[ "region:us" ]
2022-11-19T16:39:44+00:00
{"dataset_info": {"features": [{"name": "content", "dtype": "string"}, {"name": "label", "dtype": "int64"}], "splits": [{"name": "train", "num_bytes": 4838134, "num_examples": 29488}, {"name": "test", "num_bytes": 1207567, "num_examples": 7372}], "download_size": 4332711, "dataset_size": 6045701}}
2022-11-19T17:18:41+00:00
3aafab6b17f62690ee8aeba776319482cd1e9ce8
lIlBrother/KsponSpeech-Phonetic
[ "license:apache-2.0", "region:us" ]
2022-11-19T16:43:59+00:00
{"license": "apache-2.0"}
2022-11-19T16:43:59+00:00
5c4e18e9fa5fd20be2eb840b2f43d241bc073fc9
# Dataset Card for "L1_scraped_korpus_pelny" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
Nikutka/L1_scraped_korpus_pelny
[ "region:us" ]
2022-11-19T16:50:52+00:00
{"dataset_info": {"features": [{"name": "content", "dtype": "string"}, {"name": "label", "dtype": "int64"}], "splits": [{"name": "train", "num_bytes": 118294409, "num_examples": 1249536}, {"name": "test", "num_bytes": 29613125, "num_examples": 312385}], "download_size": 108295347, "dataset_size": 147907534}}
2022-11-19T17:15:54+00:00
a953228ef0aad99a67b9da3b5cc2a7e6dc3fff55
# Dataset Card for "L1_poleval_korpus_pelny" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
Nikutka/L1_poleval_korpus_pelny
[ "region:us" ]
2022-11-19T16:51:32+00:00
{"dataset_info": {"features": [{"name": "content", "dtype": "string"}, {"name": "label", "dtype": "int64"}], "splits": [{"name": "train", "num_bytes": 764265, "num_examples": 9443}, {"name": "test", "num_bytes": 71297, "num_examples": 891}], "download_size": 556613, "dataset_size": 835562}}
2022-11-19T16:51:39+00:00
12d7363c2f20943fa85687da7ac2a6947fb2923a
# Dataset Card for "L1_poleval_korpus_wzorcowy" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
Nikutka/L1_poleval_korpus_wzorcowy
[ "region:us" ]
2022-11-19T16:51:55+00:00
{"dataset_info": {"features": [{"name": "content", "dtype": "string"}, {"name": "label", "dtype": "int64"}], "splits": [{"name": "train", "num_bytes": 20564, "num_examples": 253}, {"name": "test", "num_bytes": 1963, "num_examples": 25}], "download_size": 18165, "dataset_size": 22527}}
2022-11-19T16:52:02+00:00
2b67f85fe6ae0762e5b6bcb5e2202477c204dc83
_The Dataset Teaser is now enabled instead! Isn't this better?_ ![preview of all texture sets](https://huggingface.co/datasets/texturedesign/td01_natural-ground-textures/resolve/main/docs/teaser.webp) # TD 01: Natural Ground Textures This dataset contains multi-photo texture captures in outdoor nature scenes — all focusing on the ground. Each set has different photos that showcase texture variety, making them ideal for training a domain-specific image generator! Overall information about this dataset: * **Format** — JPEG-XL, lossless RGB * **Resolution** — 4032 × 2268 * **Device** — mobile camera * **Technique** — hand-held * **Orientation** — portrait or landscape * **Author**: Alex J. Champandard * **Configurations**: 4K, 2K (default), 1K To load the medium- and high-resolution images of the dataset, you'll need to install `jxlpy` from [PyPI](https://pypi.org/project/jxlpy/) with `pip install jxlpy`: ```python # Recommended use, JXL at high-quality. from jxlpy import JXLImagePlugin from datasets import load_dataset d = load_dataset('texturedesign/td01_natural-ground-textures', 'JXL@4K', num_proc=4) print(len(d['train']), len(d['test'])) ``` The lowest-resolution images are available as PNG with a regular installation of `pillow`: ```python # Alternative use, PNG at low-quality. from datasets import load_dataset d = load_dataset('texturedesign/td01_natural-ground-textures', 'PNG@1K', num_proc=4) # EXAMPLE: Discard all other sets except Set #1. dataset = dataset.filter(lambda s: s['set'] == 1) # EXAMPLE: Only keep images with index 0 and 2. dataset = dataset.select([0, 2]) ``` Use built-in dataset `filter()` and `select()` to narrow down the loaded dataset for training, or to ease with development. ## Set #1: Rock and Gravel ![preview of the files in Set #1](https://huggingface.co/datasets/texturedesign/td01_natural-ground-textures/resolve/main/docs/set01.webp) * **Description**: - surface rocks with gravel and coarse sand - strong sunlight from the left, sharp shadows * **Number of Photos**: - 7 train - 2 test * **Edits**: - rotated photos to align sunlight - removed infrequent objects * **Size**: 77.8 Mb ## Set #2: Dry Grass with Pine Needles ![preview of the files in Set #2](https://huggingface.co/datasets/texturedesign/td01_natural-ground-textures/resolve/main/docs/set02.webp) * **Description**: - field of dry grass and pine needles - sunlight from the top right, some shadows * **Number of Photos**: - 6 train - 1 test * **Edits**: - removed dry leaves and large plants - removed sticks, rocks and sporadic daisies * **Size**: 95.2 Mb ## Set #3: Chipped Stones, Broken Leaves and Twiglets ![preview of the files in Set #3](https://huggingface.co/datasets/texturedesign/td01_natural-ground-textures/resolve/main/docs/set03.webp) * **Description**: - autumn path with chipped stones and dry broken leaves - diffuse light on a cloudy day, very soft shadows * **Number of Photos**: - 9 train - 3 test * **Edits**: - removed anything that looks green, fresh leaves - removed long sticks and large/odd stones * **Size**: 126.9 Mb ## Set #4: Grass Clumps and Cracked Dirt ![preview of the files in Set #4](https://huggingface.co/datasets/texturedesign/td01_natural-ground-textures/resolve/main/docs/set04.webp) * **Description**: - clumps of green grass, clover and patches of cracked dirt - diffuse light on cloudy day, shadows under large blades of grass * **Number of Photos**: - 9 train - 2 test * **Edits**: - removed dry leaves, sporadic dandelions, and large objects - histogram matching for two of the photos so the colors look similar * **Size**: 126.8 Mb ## Set #5: Dirt, Stones, Rock, Twigs... ![preview of the files in Set #5](https://huggingface.co/datasets/texturedesign/td01_natural-ground-textures/resolve/main/docs/set05.webp) * **Description**: - intricate micro-scene with grey dirt, surface rock, stones, twigs and organic debris - diffuse light on cloudy day, soft shadows around the larger objects * **Number of Photos**: - 9 train - 3 test * **Edits**: - removed odd objects that felt out-of-distribution * **Size**: 102.1 Mb ## Set #6: Plants with Flowers on Dry Leaves ![preview of the files in Set #6](https://huggingface.co/datasets/texturedesign/td01_natural-ground-textures/resolve/main/docs/set06.webp) * **Description**: - leafy plants with white flowers on a bed of dry brown leaves - soft diffuse light, shaded areas under the plants * **Number of Photos**: - 9 train - 2 test * **Edits**: - none yet, inpainting doesn't work well enough - would remove long sticks and pieces of wood * **Size**: 105.1 Mb ## Set #7: Frozen Footpath with Snow ![preview of the files in Set #7](https://huggingface.co/datasets/texturedesign/td01_natural-ground-textures/resolve/main/docs/set07.webp) * **Description**: - frozen ground on a path with footprints - areas with snow and dark brown ground beneath - diffuse lighting on a cloudy day * **Number of Photos**: - 11 train - 3 test * **Size**: 95.5 Mb ## Set #8: Pine Needles Forest Floor ![preview of the files in Set #8](https://huggingface.co/datasets/texturedesign/td01_natural-ground-textures/resolve/main/docs/set08.webp) * **Description**: - forest floor with a mix of brown soil and grass - variety of dry white leaves, sticks, pinecones, pine needles - diffuse lighting on a cloudy day * **Number of Photos**: - 15 train - 4 test * **Size**: 160.6 Mb ## Set #9: Snow on Grass and Dried Leaves ![preview of the files in Set #9](https://huggingface.co/datasets/texturedesign/td01_natural-ground-textures/resolve/main/docs/set09.webp) * **Description**: - field in a park with short green grass - large dried brown leaves and fallen snow on top - diffuse lighting on a cloudy day * **Number of Photos**: - 8 train - 3 test * **Size**: 99.8 Mb ## Set #10: Brown Leaves on Wet Ground ![preview of the files in Set #10](https://huggingface.co/datasets/texturedesign/td01_natural-ground-textures/resolve/main/docs/set10.webp) * **Description**: - fallew brown leaves on wet ground - occasional tree root and twiglets - diffuse lighting on a rainy day * **Number of Photos**: - 17 train - 4 test * **Size**: 186.2 Mb ## Set #11: Wet Sand Path with Debris ![preview of the files in Set #11](https://huggingface.co/datasets/texturedesign/td01_natural-ground-textures/resolve/main/docs/set11.webp) * **Description**: - hard sandy path in the rain - decomposing leaves and other organic debris - diffuse lighting on a rainy day * **Number of Photos**: - 17 train - 4 test * **Size**: 186.2 Mb ## Set #12: Wood Chips & Sawdust Sprinkled on Forest Path ![preview of the files in Set #12](https://huggingface.co/datasets/texturedesign/td01_natural-ground-textures/resolve/main/docs/set12.webp) * **Description**: - wood chips, sawdust, twigs and roots on forest path - intermittent sunlight with shadows of trees * **Number of Photos**: - 8 train - 2 test * **Size**: 110.4 Mb ## Set #13: Young Grass Growing in the Dog Park ![preview of the files in Set #13](https://huggingface.co/datasets/texturedesign/td01_natural-ground-textures/resolve/main/docs/set13.webp) * **Description**: - young grass growing in a dog park after overnight rain - occasional stones, sticks and twigs, pine needles - diffuse lighting on a cloudy day * **Number of Photos**: - 17 train - 4 test * **Size**: 193.4 Mb ## Set #14: Wavy Wet Beach Sand ![preview of the files in Set #14](https://huggingface.co/datasets/texturedesign/td01_natural-ground-textures/resolve/main/docs/set14.webp) * **Description**: - wavy wet sand on the beach after the tide retreated - some dirt and large pieces algae debris - diffuse lighting on a cloudy day * **Number of Photos**: - 11 train - 3 test * **Size**: 86.5 Mb ## Set #15: Dry Dirt Road and Debris from Trees ![preview of the files in Set #15](https://huggingface.co/datasets/texturedesign/td01_natural-ground-textures/resolve/main/docs/set15.webp) * **Description**: - dirt road of dry compacted sand with debris on top - old pine needles and dry brown leaves - diffuse lighting on a cloudy day * **Number of Photos**: - 8 train - 2 test * **Size**: 86.9 Mb ## Set #16: Sandy Beach Path with Grass Clumps ![preview of the files in Set #17](https://huggingface.co/datasets/texturedesign/td01_natural-ground-textures/resolve/main/docs/set17.webp) * **Description**: - path with sand and clumps grass heading towards the beach - occasional blueish stones, leafy weeds, and yellow flowers - diffuse lighting on a cloudy day * **Number of Photos**: - 10 train - 3 test * **Size**: 118.8 Mb ## Set #17: Pine Needles and Brown Leaves on Park Floor ![preview of the files in Set #16](https://huggingface.co/datasets/texturedesign/td01_natural-ground-textures/resolve/main/docs/set16.webp) * **Description**: - park floor with predominantly pine needles - brown leaves from nearby trees, green grass underneath - diffuse lighting on a cloudy day * **Number of Photos**: - 8 train - 2 test * **Size**: 99.9 Mb
texturedesign/td01_natural-ground-textures
[ "task_categories:unconditional-image-generation", "annotations_creators:expert-generated", "size_categories:n<1K", "source_datasets:original", "license:cc-by-nc-4.0", "texture-synthesis", "photography", "non-infringing", "region:us" ]
2022-11-19T17:43:30+00:00
{"annotations_creators": ["expert-generated"], "language_creators": [], "language": [], "license": ["cc-by-nc-4.0"], "multilinguality": [], "size_categories": ["n<1K"], "source_datasets": ["original"], "task_categories": ["unconditional-image-generation"], "task_ids": [], "pretty_name": "TD01: Natural Ground Texture Photos", "tags": ["texture-synthesis", "photography", "non-infringing"], "viewer": false}
2023-09-02T09:21:04+00:00
a7ae4cad6d4d34f32c1912718be9037f0015b779
starbotica/markkistler
[ "license:unknown", "region:us" ]
2022-11-19T19:23:57+00:00
{"license": "unknown"}
2022-11-19T19:37:19+00:00
bfc723a1831e441b95d5604a266ab939b48dd4f2
# AutoTrain Dataset for project: autotrain_goodreads_string ## Dataset Description This dataset has been automatically processed by AutoTrain for project autotrain_goodreads_string. ### Languages The BCP-47 code for the dataset's language is en. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```json [ { "target": 5, "text": "This book was absolutely ADORABLE!!!!!!!!!!! It was an awesome, light and FUN read. \n I loved the characters but I absolutely LOVED Cam!!!!!!!!!!!! Major Swoooon Worthy! J \n \"You've been checking me out, haven't you? In-between your flaming insults? I feel like man candy.\" \n Seriously, between being HOT, FUNNY and OH SO VERY ADORABLE, Cam was the perfect catch!! \n \" I'm not going out with you Cam.\" \n \" I didn't ask you at this moment, now did I\" One side of his lips curved up. \" But you will eventually.\" \n \"You're delusional\" \n \"I'm determined.\" \n \" More like annoying.\" \n \" Most would say amazing.\" \n Cam and Avery's relationship is tough due to the secrets she keeps but he is the perfect match for breaking her out of her shell and facing her fears. \n This book is definitely a MUST READ. \n Trust me when I say this YOU will not regret it! \n www.Jenreadit.com" }, { "target": 4, "text": "I FINISHED!!! This book took me FOREVER to read! But I am so glad I stuck with it, I really loved it. It took me a while to get into: this book has a TON of characters and storylines. But once I hit about the 100-page mark, I became very invested in the story and couldn't wait to see what would happen with Lizzie, Lane, Edward, Gin and the rest of the family. Oh, and Samuel T. There's a little bit of sex but mostly this is a sweeping romance novel, much like Dynasty and Dallas from the 1980's. If you loved those series, you will love this book. There's betrayal, unrequited love, family fortunes, and much scheming. \n There are many characters to love here and many to hate. Some are over-the-top but I loved the central storyline involving Lane and Lizzie. \n The author really gets the Southern mannerisms right, and the backdrop of the Kentucky Bourbon industry is fascinating. This book ends not so much on a cliffhanger but with many, many loose ends, and I will eagerly pick up the next book in this series." } ] ``` ### Dataset Fields The dataset has the following fields (also called "features"): ```json { "target": "ClassLabel(num_classes=6, names=['0_stars', '1_stars', '2_stars', '3_stars', '4_stars', '5_stars'], id=None)", "text": "Value(dtype='string', id=None)" } ``` ### Dataset Splits This dataset is split into a train and validation split. The split sizes are as follow: | Split name | Num samples | | ------------ | ------------------- | | train | 2357 | | valid | 592 |
fernanda-dionello/good-reads-string
[ "task_categories:text-classification", "language:en", "region:us" ]
2022-11-19T20:09:23+00:00
{"language": ["en"], "task_categories": ["text-classification"]}
2022-11-19T20:10:26+00:00
fd46dee0a684d1f651e353ce659e0f0ff11322e7
# Dataset Card for "stefano-finetune" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
albluc24/stefano-finetune
[ "region:us" ]
2022-11-19T20:30:42+00:00
{"dataset_info": {"features": [{"name": "text", "dtype": "string"}, {"name": "audio", "dtype": "audio"}], "splits": [{"name": "eval", "num_bytes": 3732782.0, "num_examples": 1}, {"name": "train", "num_bytes": 227326609.0, "num_examples": 55}], "download_size": 0, "dataset_size": 231059391.0}}
2022-11-19T20:46:15+00:00
ae1e4ca8b786994cd3192930ad28f1676b4b02ef
# Dataset Card for "two-minute-papers" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
Whispering-GPT/two-minute-papers
[ "task_categories:automatic-speech-recognition", "whisper", "whispering", "base", "region:us" ]
2022-11-19T20:52:17+00:00
{"task_categories": ["automatic-speech-recognition"], "dataset_info": {"features": [{"name": "CHANNEL_NAME", "dtype": "string"}, {"name": "URL", "dtype": "string"}, {"name": "TITLE", "dtype": "string"}, {"name": "DESCRIPTION", "dtype": "string"}, {"name": "TRANSCRIPTION", "dtype": "string"}, {"name": "SEGMENTS", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 10435074, "num_examples": 737}], "download_size": 4626170, "dataset_size": 10435074}, "tags": ["whisper", "whispering", "base"]}
2022-11-19T23:34:46+00:00
eb5f579468dc10ed0510d26e5f3de6b34f8700ca
# Dataset Card for "gal_yair_8300_100x100_fixed" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
galman33/gal_yair_8300_100x100_fixed
[ "region:us" ]
2022-11-19T22:43:01+00:00
{"dataset_info": {"features": [{"name": "lat", "dtype": "float64"}, {"name": "lon", "dtype": "float64"}, {"name": "country_code", "dtype": {"class_label": {"names": {"0": "ad", "1": "ae", "2": "al", "3": "aq", "4": "ar", "5": "au", "6": "bd", "7": "be", "8": "bg", "9": "bm", "10": "bo", "11": "br", "12": "bt", "13": "bw", "14": "ca", "15": "ch", "16": "cl", "17": "co", "18": "cz", "19": "de", "20": "dk", "21": "ec", "22": "ee", "23": "es", "24": "fi", "25": "fr", "26": "gb", "27": "gh", "28": "gl", "29": "gr", "30": "gt", "31": "hk", "32": "hr", "33": "hu", "34": "id", "35": "ie", "36": "il", "37": "is", "38": "it", "39": "ix", "40": "jp", "41": "kg", "42": "kh", "43": "kr", "44": "la", "45": "lk", "46": "ls", "47": "lt", "48": "lu", "49": "lv", "50": "me", "51": "mg", "52": "mk", "53": "mn", "54": "mo", "55": "mt", "56": "mx", "57": "my", "58": "nl", "59": "no", "60": "nz", "61": "pe", "62": "ph", "63": "pl", "64": "pt", "65": "ro", "66": "rs", "67": "ru", "68": "se", "69": "sg", "70": "si", "71": "sk", "72": "sn", "73": "sz", "74": "th", "75": "tn", "76": "tr", "77": "tw", "78": "ua", "79": "ug", "80": "us", "81": "uy", "82": "za"}}}}, {"name": "image", "dtype": "image"}], "splits": [{"name": "train", "num_bytes": 142019429.5, "num_examples": 8300}], "download_size": 141877783, "dataset_size": 142019429.5}}
2022-11-26T13:15:23+00:00
f7a01f58b105e3b089b9c492cab91d4bc5d9c0ca
hounst/whitn
[ "license:cc", "region:us" ]
2022-11-20T00:55:52+00:00
{"license": "cc"}
2022-11-20T01:18:07+00:00
6af6fbe04f185dabec6605e5c3858bac1dd23396
# Genshin Datasets for Diff-SVC ## 仓库地址 | 仓库 | 传送门 | | :--------------------------------------------------------: | :----------------------------------------------------------: | | Diff-SVC | [点此传送](https://github.com/prophesier/diff-svc) | | 44.1KHz声码器 | [点此传送](https://openvpi.github.io/vocoders) | | 原神语音数据集 | [点此传送](https://github.com/w4123/GenshinVoice) | | 训练用底模(如果想用这个数据集自己训练,非原神模型也可用) | [点此传送](https://huggingface.co/Erythrocyte/Diff-SVC_Pre-trained_Models) | | 已训练原神 Diff-SVC 模型(可用于二创整活) | [点此传送](https://huggingface.co/Erythrocyte/Diff-SVC_Genshin_Models) | ## 介绍 该数据集为训练 Diff-SVC 原神模型的数据集,上传的数据集均已进行 `长音频切割` 、`响度匹配`。可以 `直接用来预处理并训练`,由于每个角色数据集规模不够,训练时候建议配合 `预训练模型` 训练。预训练模型见上表,并且提供了 `详细的教程` 以及 `多种可选预训练模型`。 ## 使用教程(以下操作均需要在Diff-SVC目录下进行) ### Windows平台 1. 下载自己需要的数据集 2. 解压到 Diff-SVC 根目录,如果提示覆盖,请直接覆盖。 4. 依次输入如下指令进行预处理 ```bash set PYTHONPATH=. set CUDA_VISIBLE_DEVICES=0 python preprocessing/binarize.py --config training/config_nsf.yaml ``` 5. 按照教程进行加载预训练模型并训练: 教程地址:https://huggingface.co/Erythrocyte/Diff-SVC_Pre-trained_Models ### Linux平台 1. 输入如下命令下载需要的数据集(这里以纳西妲为例,其它角色可以从下表右击链接复制) ```bash wget https://huggingface.co/datasets/Erythrocyte/Diff-SVC_Genshin_Datasets/resolve/main/Sumeru/nahida.zip ``` 2. 输入如下命令解压到 Diff-SVC 根目录,如果提示覆盖请输入 y ```bash unzip nahida.zip ``` 4. 依次输入如下指令进行预处理 ```bash export PYTHONPATH=. CUDA_VISIBLE_DEVICES=0 python preprocessing/binarize.py --config training/config_nsf.yaml ``` 5. 按照教程进行加载预训练模型并训练: 教程地址:https://huggingface.co/Erythrocyte/Diff-SVC_Pre-trained_Models ## 下载地址 | 地区 | 角色名 | 下载地址 | | :--: | :----------: | :----------------------------------------------------------: | | 蒙德 | 优菈 | [eula.zip](https://huggingface.co/datasets/Erythrocyte/Diff-SVC_Genshin_Datasets/resolve/main/Mondstadt/eula.zip) | | 蒙德 | 阿贝多 | [albedo.zip](https://huggingface.co/datasets/Erythrocyte/Diff-SVC_Genshin_Datasets/resolve/main/Mondstadt/albedo.zip) | | 蒙德 | 温迪 | [venti.zip](https://huggingface.co/datasets/Erythrocyte/Diff-SVC_Genshin_Datasets/resolve/main/Mondstadt/venti.zip) | | 蒙德 | 莫娜 | [mona.zip](https://huggingface.co/datasets/Erythrocyte/Diff-SVC_Genshin_Datasets/resolve/main/Mondstadt/mona.zip) | | 蒙德 | 可莉 | [klee.zip](https://huggingface.co/datasets/Erythrocyte/Diff-SVC_Genshin_Datasets/resolve/main/Mondstadt/klee.zip) | | 蒙德 | 琴 | [jean.zip](https://huggingface.co/datasets/Erythrocyte/Diff-SVC_Genshin_Datasets/resolve/main/Mondstadt/jean.zip) | | 蒙德 | 迪卢克 | [diluc.zip](https://huggingface.co/datasets/Erythrocyte/Diff-SVC_Genshin_Datasets/resolve/main/Mondstadt/diluc.zip) | | 璃月 | 钟离 | [zhongli.zip](https://huggingface.co/datasets/Erythrocyte/Diff-SVC_Genshin_Datasets/resolve/main/Liyue/zhongli.zip) | | 稻妻 | 雷电将军 | [raiden.zip](https://huggingface.co/datasets/Erythrocyte/Diff-SVC_Genshin_Datasets/resolve/main/Inazuma/raiden.zip) | | 须弥 | 流浪者(散兵) | [wanderer.zip](https://huggingface.co/datasets/Erythrocyte/Diff-SVC_Genshin_Datasets/resolve/main/Sumeru/wanderer.zip) | | 须弥 | 纳西妲 | [nahida.zip](https://huggingface.co/datasets/Erythrocyte/Diff-SVC_Genshin_Datasets/resolve/main/Sumeru/nahida.zip) |
Erythrocyte/Diff-SVC_Genshin_Datasets
[ "Diff-SVC", "Genshin", "Genshin Impact", "Voice Data", "Voice Dataset", "region:us" ]
2022-11-20T01:25:40+00:00
{"tags": ["Diff-SVC", "Genshin", "Genshin Impact", "Voice Data", "Voice Dataset"]}
2022-12-19T01:33:22+00:00
d6115c5b9a52a758d6f0d08f0bd9b19df294edf9
# Dataset Card for EUWikipedias: A dataset of Wikipedias in the EU languages ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** [Joel Niklaus](mailto:[email protected]) ### Dataset Summary Wikipedia dataset containing cleaned articles of all languages. The datasets are built from the Wikipedia dump (https://dumps.wikimedia.org/) with one split per language. Each example contains the content of one full Wikipedia article with cleaning to strip markdown and unwanted sections (references, etc.). ### Supported Tasks and Leaderboards The dataset supports the tasks of fill-mask. ### Languages The following languages are supported: bg, cs, da, de, el, en, es, et, fi, fr, ga, hr, hu, it, lt, lv, mt, nl, pl, pt, ro, sk, sl, sv ## Dataset Structure It is structured in the following format: {date}/{language}_{shard}.jsonl.xz At the moment only the date '20221120' is supported. Use the dataset like this: ```python from datasets import load_dataset dataset = load_dataset('joelito/EU_Wikipedias', date="20221120", language="de", split='train', streaming=True) ``` ### Data Instances The file format is jsonl.xz and there is one split available (`train`). | Source | Size (MB) | Words | Documents | Words/Document | |:-------------|------------:|-----------:|------------:|-----------------:| | 20221120.all | 86034 | 9506846949 | 26481379 | 359 | | 20221120.bg | 1261 | 88138772 | 285876 | 308 | | 20221120.cs | 1904 | 189580185 | 513851 | 368 | | 20221120.da | 679 | 74546410 | 286864 | 259 | | 20221120.de | 11761 | 1191919523 | 2740891 | 434 | | 20221120.el | 1531 | 103504078 | 215046 | 481 | | 20221120.en | 26685 | 3192209334 | 6575634 | 485 | | 20221120.es | 6636 | 801322400 | 1583597 | 506 | | 20221120.et | 538 | 48618507 | 231609 | 209 | | 20221120.fi | 1391 | 115779646 | 542134 | 213 | | 20221120.fr | 9703 | 1140823165 | 2472002 | 461 | | 20221120.ga | 72 | 8025297 | 57808 | 138 | | 20221120.hr | 555 | 58853753 | 198746 | 296 | | 20221120.hu | 1855 | 167732810 | 515777 | 325 | | 20221120.it | 5999 | 687745355 | 1782242 | 385 | | 20221120.lt | 409 | 37572513 | 203233 | 184 | | 20221120.lv | 269 | 25091547 | 116740 | 214 | | 20221120.mt | 29 | 2867779 | 5030 | 570 | | 20221120.nl | 3208 | 355031186 | 2107071 | 168 | | 20221120.pl | 3608 | 349900622 | 1543442 | 226 | | 20221120.pt | 3315 | 389786026 | 1095808 | 355 | | 20221120.ro | 1017 | 111455336 | 434935 | 256 | | 20221120.sk | 506 | 49612232 | 238439 | 208 | | 20221120.sl | 543 | 58858041 | 178472 | 329 | | 20221120.sv | 2560 | 257872432 | 2556132 | 100 | ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation This dataset has been created by downloading the wikipedias using [olm/wikipedia](https://huggingface.co/datasets/olm/wikipedia) for the 24 EU languages. For more information about the creation of the dataset please refer to prepare_wikipedias.py ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information ``` TODO add citation ``` ### Contributions Thanks to [@JoelNiklaus](https://github.com/joelniklaus) for adding this dataset.
joelniklaus/EU_Wikipedias
[ "task_categories:fill-mask", "annotations_creators:other", "language_creators:found", "multilinguality:multilingual", "size_categories:10M<n<100M", "source_datasets:original", "language:bg", "language:cs", "language:da", "language:de", "language:el", "language:en", "language:es", "language:et", "language:fi", "language:fr", "language:ga", "language:hr", "language:hu", "language:it", "language:lt", "language:lv", "language:mt", "language:nl", "language:pl", "language:pt", "language:ro", "language:sk", "language:sl", "language:sv", "license:cc-by-4.0", "region:us" ]
2022-11-20T01:31:51+00:00
{"annotations_creators": ["other"], "language_creators": ["found"], "language": ["bg", "cs", "da", "de", "el", "en", "es", "et", "fi", "fr", "ga", "hr", "hu", "it", "lt", "lv", "mt", "nl", "pl", "pt", "ro", "sk", "sl", "sv"], "license": ["cc-by-4.0"], "multilinguality": ["multilingual"], "size_categories": ["10M<n<100M"], "source_datasets": ["original"], "task_categories": ["fill-mask"], "pretty_name": "EUWikipedias: A dataset of Wikipedias in the EU languages"}
2023-03-21T15:44:18+00:00
084fd1fb5aad0dc3f8c655807a21f07aaf27f5ff
# Dataset Card for "goog-tech-talks" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
juancopi81/goog-tech-talks
[ "task_categories:automatic-speech-recognition", "whisper", "whispering", "base", "region:us" ]
2022-11-20T02:08:42+00:00
{"task_categories": ["automatic-speech-recognition"], "dataset_info": {"features": [{"name": "CHANNEL_NAME", "dtype": "string"}, {"name": "URL", "dtype": "string"}, {"name": "TITLE", "dtype": "string"}, {"name": "DESCRIPTION", "dtype": "string"}, {"name": "TRANSCRIPTION", "dtype": "string"}, {"name": "SEGMENTS", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 101139, "num_examples": 1}], "download_size": 55503, "dataset_size": 101139}, "tags": ["whisper", "whispering", "base"]}
2022-11-20T02:08:46+00:00
cd69fbfac065aeebd820f8ed4a047fb266e2e170
# Dataset Card for "gpk-captions" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
gluten/gpk-captions
[ "doi:10.57967/hf/0123", "region:us" ]
2022-11-20T03:42:14+00:00
{"dataset_info": {"features": [{"name": "image", "dtype": "image"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 45126936.0, "num_examples": 83}], "download_size": 45128569, "dataset_size": 45126936.0}}
2022-11-20T04:03:10+00:00
89c13ef027bdb602bb91d8c16469fff1d4cf3862
This is the dataset used for making the model : https://huggingface.co/Guizmus/SD_PoW_Collection The images were made by the users of Stable Diffusion discord using CreativeML-OpenRail-M licenced models, in the intent to make this dataset. 60 pictures captioned with their content by hand, with the prefix "3D Style" The collection process was made public during a day, until enough variety was introduced to train through a Dreambooth method a style corresponding to the different members of this community The picture captioned are available in [this zip file](https://huggingface.co/datasets/Guizmus/3DChanStyle/resolve/main/3DChanStyle.zip)
Guizmus/3DChanStyle
[ "license:creativeml-openrail-m", "region:us" ]
2022-11-20T09:49:14+00:00
{"license": "creativeml-openrail-m"}
2022-11-23T09:18:31+00:00
77a3ee29ca65c47eaae250a3aa0164fd701d634a
# Dataset Card for "relbert/semeval2012_relational_similarity_V6" ## Dataset Description - **Repository:** [RelBERT](https://github.com/asahi417/relbert) - **Paper:** [https://aclanthology.org/S12-1047/](https://aclanthology.org/S12-1047/) - **Dataset:** SemEval2012: Relational Similarity ### Dataset Summary ***IMPORTANT***: This is the same dataset as [relbert/semeval2012_relational_similarity](https://huggingface.co/datasets/relbert/semeval2012_relational_similarity), but with a different dataset construction. Relational similarity dataset from [SemEval2012 task 2](https://aclanthology.org/S12-1047/), compiled to fine-tune [RelBERT](https://github.com/asahi417/relbert) model. The dataset contains a list of positive and negative word pair from 89 pre-defined relations. The relation types are constructed on top of following 10 parent relation types. ```shell { 1: "Class Inclusion", # Hypernym 2: "Part-Whole", # Meronym, Substance Meronym 3: "Similar", # Synonym, Co-hypornym 4: "Contrast", # Antonym 5: "Attribute", # Attribute, Event 6: "Non Attribute", 7: "Case Relation", 8: "Cause-Purpose", 9: "Space-Time", 10: "Representation" } ``` Each of the parent relation is further grouped into child relation types where the definition can be found [here](https://drive.google.com/file/d/0BzcZKTSeYL8VenY0QkVpZVpxYnc/view?resourcekey=0-ZP-UARfJj39PcLroibHPHw). ## Dataset Structure ### Data Instances An example of `train` looks as follows. ``` { 'relation_type': '8d', 'positives': [ [ "breathe", "live" ], [ "study", "learn" ], [ "speak", "communicate" ], ... ] 'negatives': [ [ "starving", "hungry" ], [ "clean", "bathe" ], [ "hungry", "starving" ], ... ] } ``` ### Data Splits | name |train|validation| |---------|----:|---------:| |semeval2012_relational_similarity| 89 | 89| ### Number of Positive/Negative Word-pairs in each Split | | positives | negatives | |:--------------------------------------------|------------:|------------:| | ('1', 'parent', 'train') | 88 | 544 | | ('1', 'parent', 'validation') | 22 | 136 | | ('10', 'parent', 'train') | 48 | 584 | | ('10', 'parent', 'validation') | 12 | 146 | | ('10a', 'child', 'train') | 8 | 1324 | | ('10a', 'child', 'validation') | 2 | 331 | | ('10a', 'child_prototypical', 'train') | 194 | 1917 | | ('10a', 'child_prototypical', 'validation') | 52 | 521 | | ('10b', 'child', 'train') | 8 | 1325 | | ('10b', 'child', 'validation') | 2 | 331 | | ('10b', 'child_prototypical', 'train') | 180 | 1558 | | ('10b', 'child_prototypical', 'validation') | 54 | 469 | | ('10c', 'child', 'train') | 8 | 1327 | | ('10c', 'child', 'validation') | 2 | 331 | | ('10c', 'child_prototypical', 'train') | 170 | 1640 | | ('10c', 'child_prototypical', 'validation') | 40 | 390 | | ('10d', 'child', 'train') | 8 | 1328 | | ('10d', 'child', 'validation') | 2 | 331 | | ('10d', 'child_prototypical', 'train') | 154 | 1390 | | ('10d', 'child_prototypical', 'validation') | 44 | 376 | | ('10e', 'child', 'train') | 8 | 1329 | | ('10e', 'child', 'validation') | 2 | 332 | | ('10e', 'child_prototypical', 'train') | 134 | 884 | | ('10e', 'child_prototypical', 'validation') | 40 | 234 | | ('10f', 'child', 'train') | 8 | 1328 | | ('10f', 'child', 'validation') | 2 | 331 | | ('10f', 'child_prototypical', 'train') | 160 | 1460 | | ('10f', 'child_prototypical', 'validation') | 38 | 306 | | ('1a', 'child', 'train') | 8 | 1324 | | ('1a', 'child', 'validation') | 2 | 331 | | ('1a', 'child_prototypical', 'train') | 212 | 1854 | | ('1a', 'child_prototypical', 'validation') | 34 | 338 | | ('1b', 'child', 'train') | 8 | 1324 | | ('1b', 'child', 'validation') | 2 | 331 | | ('1b', 'child_prototypical', 'train') | 190 | 1712 | | ('1b', 'child_prototypical', 'validation') | 56 | 480 | | ('1c', 'child', 'train') | 8 | 1327 | | ('1c', 'child', 'validation') | 2 | 331 | | ('1c', 'child_prototypical', 'train') | 160 | 1528 | | ('1c', 'child_prototypical', 'validation') | 50 | 502 | | ('1d', 'child', 'train') | 8 | 1323 | | ('1d', 'child', 'validation') | 2 | 330 | | ('1d', 'child_prototypical', 'train') | 224 | 2082 | | ('1d', 'child_prototypical', 'validation') | 46 | 458 | | ('1e', 'child', 'train') | 8 | 1329 | | ('1e', 'child', 'validation') | 2 | 332 | | ('1e', 'child_prototypical', 'train') | 126 | 775 | | ('1e', 'child_prototypical', 'validation') | 48 | 256 | | ('2', 'parent', 'train') | 80 | 552 | | ('2', 'parent', 'validation') | 20 | 138 | | ('2a', 'child', 'train') | 8 | 1324 | | ('2a', 'child', 'validation') | 2 | 330 | | ('2a', 'child_prototypical', 'train') | 186 | 1885 | | ('2a', 'child_prototypical', 'validation') | 72 | 736 | | ('2b', 'child', 'train') | 8 | 1327 | | ('2b', 'child', 'validation') | 2 | 331 | | ('2b', 'child_prototypical', 'train') | 172 | 1326 | | ('2b', 'child_prototypical', 'validation') | 38 | 284 | | ('2c', 'child', 'train') | 8 | 1325 | | ('2c', 'child', 'validation') | 2 | 331 | | ('2c', 'child_prototypical', 'train') | 192 | 1773 | | ('2c', 'child_prototypical', 'validation') | 42 | 371 | | ('2d', 'child', 'train') | 8 | 1328 | | ('2d', 'child', 'validation') | 2 | 331 | | ('2d', 'child_prototypical', 'train') | 158 | 1329 | | ('2d', 'child_prototypical', 'validation') | 40 | 338 | | ('2e', 'child', 'train') | 8 | 1327 | | ('2e', 'child', 'validation') | 2 | 331 | | ('2e', 'child_prototypical', 'train') | 164 | 1462 | | ('2e', 'child_prototypical', 'validation') | 46 | 463 | | ('2f', 'child', 'train') | 8 | 1327 | | ('2f', 'child', 'validation') | 2 | 331 | | ('2f', 'child_prototypical', 'train') | 176 | 1869 | | ('2f', 'child_prototypical', 'validation') | 34 | 371 | | ('2g', 'child', 'train') | 8 | 1323 | | ('2g', 'child', 'validation') | 2 | 330 | | ('2g', 'child_prototypical', 'train') | 216 | 1925 | | ('2g', 'child_prototypical', 'validation') | 54 | 480 | | ('2h', 'child', 'train') | 8 | 1327 | | ('2h', 'child', 'validation') | 2 | 331 | | ('2h', 'child_prototypical', 'train') | 168 | 1540 | | ('2h', 'child_prototypical', 'validation') | 42 | 385 | | ('2i', 'child', 'train') | 8 | 1328 | | ('2i', 'child', 'validation') | 2 | 332 | | ('2i', 'child_prototypical', 'train') | 144 | 1335 | | ('2i', 'child_prototypical', 'validation') | 42 | 371 | | ('2j', 'child', 'train') | 8 | 1328 | | ('2j', 'child', 'validation') | 2 | 331 | | ('2j', 'child_prototypical', 'train') | 160 | 1595 | | ('2j', 'child_prototypical', 'validation') | 38 | 369 | | ('3', 'parent', 'train') | 64 | 568 | | ('3', 'parent', 'validation') | 16 | 142 | | ('3a', 'child', 'train') | 8 | 1327 | | ('3a', 'child', 'validation') | 2 | 331 | | ('3a', 'child_prototypical', 'train') | 174 | 1597 | | ('3a', 'child_prototypical', 'validation') | 36 | 328 | | ('3b', 'child', 'train') | 8 | 1327 | | ('3b', 'child', 'validation') | 2 | 331 | | ('3b', 'child_prototypical', 'train') | 174 | 1833 | | ('3b', 'child_prototypical', 'validation') | 36 | 407 | | ('3c', 'child', 'train') | 8 | 1326 | | ('3c', 'child', 'validation') | 2 | 331 | | ('3c', 'child_prototypical', 'train') | 186 | 1664 | | ('3c', 'child_prototypical', 'validation') | 36 | 315 | | ('3d', 'child', 'train') | 8 | 1324 | | ('3d', 'child', 'validation') | 2 | 331 | | ('3d', 'child_prototypical', 'train') | 202 | 1943 | | ('3d', 'child_prototypical', 'validation') | 44 | 372 | | ('3e', 'child', 'train') | 8 | 1332 | | ('3e', 'child', 'validation') | 2 | 332 | | ('3e', 'child_prototypical', 'train') | 98 | 900 | | ('3e', 'child_prototypical', 'validation') | 40 | 368 | | ('3f', 'child', 'train') | 8 | 1327 | | ('3f', 'child', 'validation') | 2 | 331 | | ('3f', 'child_prototypical', 'train') | 180 | 1983 | | ('3f', 'child_prototypical', 'validation') | 30 | 362 | | ('3g', 'child', 'train') | 8 | 1331 | | ('3g', 'child', 'validation') | 2 | 332 | | ('3g', 'child_prototypical', 'train') | 122 | 1089 | | ('3g', 'child_prototypical', 'validation') | 28 | 251 | | ('3h', 'child', 'train') | 8 | 1328 | | ('3h', 'child', 'validation') | 2 | 331 | | ('3h', 'child_prototypical', 'train') | 142 | 1399 | | ('3h', 'child_prototypical', 'validation') | 56 | 565 | | ('4', 'parent', 'train') | 64 | 568 | | ('4', 'parent', 'validation') | 16 | 142 | | ('4a', 'child', 'train') | 8 | 1327 | | ('4a', 'child', 'validation') | 2 | 331 | | ('4a', 'child_prototypical', 'train') | 170 | 1766 | | ('4a', 'child_prototypical', 'validation') | 40 | 474 | | ('4b', 'child', 'train') | 8 | 1330 | | ('4b', 'child', 'validation') | 2 | 332 | | ('4b', 'child_prototypical', 'train') | 132 | 949 | | ('4b', 'child_prototypical', 'validation') | 30 | 214 | | ('4c', 'child', 'train') | 8 | 1326 | | ('4c', 'child', 'validation') | 2 | 331 | | ('4c', 'child_prototypical', 'train') | 172 | 1755 | | ('4c', 'child_prototypical', 'validation') | 50 | 446 | | ('4d', 'child', 'train') | 8 | 1332 | | ('4d', 'child', 'validation') | 2 | 333 | | ('4d', 'child_prototypical', 'train') | 92 | 531 | | ('4d', 'child_prototypical', 'validation') | 34 | 218 | | ('4e', 'child', 'train') | 8 | 1326 | | ('4e', 'child', 'validation') | 2 | 331 | | ('4e', 'child_prototypical', 'train') | 184 | 2021 | | ('4e', 'child_prototypical', 'validation') | 38 | 402 | | ('4f', 'child', 'train') | 8 | 1328 | | ('4f', 'child', 'validation') | 2 | 332 | | ('4f', 'child_prototypical', 'train') | 144 | 1464 | | ('4f', 'child_prototypical', 'validation') | 42 | 428 | | ('4g', 'child', 'train') | 8 | 1324 | | ('4g', 'child', 'validation') | 2 | 330 | | ('4g', 'child_prototypical', 'train') | 212 | 2057 | | ('4g', 'child_prototypical', 'validation') | 46 | 435 | | ('4h', 'child', 'train') | 8 | 1326 | | ('4h', 'child', 'validation') | 2 | 331 | | ('4h', 'child_prototypical', 'train') | 170 | 1787 | | ('4h', 'child_prototypical', 'validation') | 52 | 525 | | ('5', 'parent', 'train') | 72 | 560 | | ('5', 'parent', 'validation') | 18 | 140 | | ('5a', 'child', 'train') | 8 | 1324 | | ('5a', 'child', 'validation') | 2 | 331 | | ('5a', 'child_prototypical', 'train') | 202 | 1876 | | ('5a', 'child_prototypical', 'validation') | 44 | 439 | | ('5b', 'child', 'train') | 8 | 1329 | | ('5b', 'child', 'validation') | 2 | 332 | | ('5b', 'child_prototypical', 'train') | 140 | 1310 | | ('5b', 'child_prototypical', 'validation') | 34 | 330 | | ('5c', 'child', 'train') | 8 | 1327 | | ('5c', 'child', 'validation') | 2 | 331 | | ('5c', 'child_prototypical', 'train') | 170 | 1552 | | ('5c', 'child_prototypical', 'validation') | 40 | 373 | | ('5d', 'child', 'train') | 8 | 1324 | | ('5d', 'child', 'validation') | 2 | 330 | | ('5d', 'child_prototypical', 'train') | 204 | 1783 | | ('5d', 'child_prototypical', 'validation') | 54 | 580 | | ('5e', 'child', 'train') | 8 | 1329 | | ('5e', 'child', 'validation') | 2 | 332 | | ('5e', 'child_prototypical', 'train') | 136 | 1283 | | ('5e', 'child_prototypical', 'validation') | 38 | 357 | | ('5f', 'child', 'train') | 8 | 1327 | | ('5f', 'child', 'validation') | 2 | 331 | | ('5f', 'child_prototypical', 'train') | 154 | 1568 | | ('5f', 'child_prototypical', 'validation') | 56 | 567 | | ('5g', 'child', 'train') | 8 | 1328 | | ('5g', 'child', 'validation') | 2 | 332 | | ('5g', 'child_prototypical', 'train') | 158 | 1626 | | ('5g', 'child_prototypical', 'validation') | 28 | 266 | | ('5h', 'child', 'train') | 8 | 1324 | | ('5h', 'child', 'validation') | 2 | 330 | | ('5h', 'child_prototypical', 'train') | 218 | 2348 | | ('5h', 'child_prototypical', 'validation') | 40 | 402 | | ('5i', 'child', 'train') | 8 | 1324 | | ('5i', 'child', 'validation') | 2 | 331 | | ('5i', 'child_prototypical', 'train') | 192 | 2010 | | ('5i', 'child_prototypical', 'validation') | 54 | 551 | | ('6', 'parent', 'train') | 64 | 568 | | ('6', 'parent', 'validation') | 16 | 142 | | ('6a', 'child', 'train') | 8 | 1324 | | ('6a', 'child', 'validation') | 2 | 330 | | ('6a', 'child_prototypical', 'train') | 204 | 1962 | | ('6a', 'child_prototypical', 'validation') | 54 | 530 | | ('6b', 'child', 'train') | 8 | 1327 | | ('6b', 'child', 'validation') | 2 | 331 | | ('6b', 'child_prototypical', 'train') | 180 | 1840 | | ('6b', 'child_prototypical', 'validation') | 30 | 295 | | ('6c', 'child', 'train') | 8 | 1325 | | ('6c', 'child', 'validation') | 2 | 331 | | ('6c', 'child_prototypical', 'train') | 180 | 1968 | | ('6c', 'child_prototypical', 'validation') | 54 | 527 | | ('6d', 'child', 'train') | 8 | 1328 | | ('6d', 'child', 'validation') | 2 | 331 | | ('6d', 'child_prototypical', 'train') | 164 | 1903 | | ('6d', 'child_prototypical', 'validation') | 34 | 358 | | ('6e', 'child', 'train') | 8 | 1327 | | ('6e', 'child', 'validation') | 2 | 331 | | ('6e', 'child_prototypical', 'train') | 170 | 1737 | | ('6e', 'child_prototypical', 'validation') | 40 | 398 | | ('6f', 'child', 'train') | 8 | 1326 | | ('6f', 'child', 'validation') | 2 | 331 | | ('6f', 'child_prototypical', 'train') | 174 | 1652 | | ('6f', 'child_prototypical', 'validation') | 48 | 438 | | ('6g', 'child', 'train') | 8 | 1326 | | ('6g', 'child', 'validation') | 2 | 331 | | ('6g', 'child_prototypical', 'train') | 188 | 1740 | | ('6g', 'child_prototypical', 'validation') | 34 | 239 | | ('6h', 'child', 'train') | 8 | 1324 | | ('6h', 'child', 'validation') | 2 | 330 | | ('6h', 'child_prototypical', 'train') | 230 | 2337 | | ('6h', 'child_prototypical', 'validation') | 28 | 284 | | ('7', 'parent', 'train') | 64 | 568 | | ('7', 'parent', 'validation') | 16 | 142 | | ('7a', 'child', 'train') | 8 | 1324 | | ('7a', 'child', 'validation') | 2 | 331 | | ('7a', 'child_prototypical', 'train') | 198 | 2045 | | ('7a', 'child_prototypical', 'validation') | 48 | 516 | | ('7b', 'child', 'train') | 8 | 1330 | | ('7b', 'child', 'validation') | 2 | 332 | | ('7b', 'child_prototypical', 'train') | 138 | 905 | | ('7b', 'child_prototypical', 'validation') | 24 | 177 | | ('7c', 'child', 'train') | 8 | 1327 | | ('7c', 'child', 'validation') | 2 | 331 | | ('7c', 'child_prototypical', 'train') | 170 | 1402 | | ('7c', 'child_prototypical', 'validation') | 40 | 313 | | ('7d', 'child', 'train') | 8 | 1324 | | ('7d', 'child', 'validation') | 2 | 331 | | ('7d', 'child_prototypical', 'train') | 196 | 2064 | | ('7d', 'child_prototypical', 'validation') | 50 | 497 | | ('7e', 'child', 'train') | 8 | 1328 | | ('7e', 'child', 'validation') | 2 | 331 | | ('7e', 'child_prototypical', 'train') | 156 | 1270 | | ('7e', 'child_prototypical', 'validation') | 42 | 298 | | ('7f', 'child', 'train') | 8 | 1326 | | ('7f', 'child', 'validation') | 2 | 331 | | ('7f', 'child_prototypical', 'train') | 178 | 1377 | | ('7f', 'child_prototypical', 'validation') | 44 | 380 | | ('7g', 'child', 'train') | 8 | 1328 | | ('7g', 'child', 'validation') | 2 | 332 | | ('7g', 'child_prototypical', 'train') | 144 | 885 | | ('7g', 'child_prototypical', 'validation') | 42 | 263 | | ('7h', 'child', 'train') | 8 | 1324 | | ('7h', 'child', 'validation') | 2 | 331 | | ('7h', 'child_prototypical', 'train') | 188 | 1479 | | ('7h', 'child_prototypical', 'validation') | 58 | 467 | | ('8', 'parent', 'train') | 64 | 568 | | ('8', 'parent', 'validation') | 16 | 142 | | ('8a', 'child', 'train') | 8 | 1324 | | ('8a', 'child', 'validation') | 2 | 331 | | ('8a', 'child_prototypical', 'train') | 186 | 1640 | | ('8a', 'child_prototypical', 'validation') | 60 | 552 | | ('8b', 'child', 'train') | 8 | 1330 | | ('8b', 'child', 'validation') | 2 | 332 | | ('8b', 'child_prototypical', 'train') | 122 | 1126 | | ('8b', 'child_prototypical', 'validation') | 40 | 361 | | ('8c', 'child', 'train') | 8 | 1326 | | ('8c', 'child', 'validation') | 2 | 331 | | ('8c', 'child_prototypical', 'train') | 192 | 1547 | | ('8c', 'child_prototypical', 'validation') | 30 | 210 | | ('8d', 'child', 'train') | 8 | 1325 | | ('8d', 'child', 'validation') | 2 | 331 | | ('8d', 'child_prototypical', 'train') | 184 | 1472 | | ('8d', 'child_prototypical', 'validation') | 50 | 438 | | ('8e', 'child', 'train') | 8 | 1327 | | ('8e', 'child', 'validation') | 2 | 331 | | ('8e', 'child_prototypical', 'train') | 174 | 1340 | | ('8e', 'child_prototypical', 'validation') | 36 | 270 | | ('8f', 'child', 'train') | 8 | 1326 | | ('8f', 'child', 'validation') | 2 | 331 | | ('8f', 'child_prototypical', 'train') | 166 | 1416 | | ('8f', 'child_prototypical', 'validation') | 56 | 452 | | ('8g', 'child', 'train') | 8 | 1330 | | ('8g', 'child', 'validation') | 2 | 332 | | ('8g', 'child_prototypical', 'train') | 124 | 640 | | ('8g', 'child_prototypical', 'validation') | 38 | 199 | | ('8h', 'child', 'train') | 8 | 1324 | | ('8h', 'child', 'validation') | 2 | 331 | | ('8h', 'child_prototypical', 'train') | 200 | 1816 | | ('8h', 'child_prototypical', 'validation') | 46 | 499 | | ('9', 'parent', 'train') | 72 | 560 | | ('9', 'parent', 'validation') | 18 | 140 | | ('9a', 'child', 'train') | 8 | 1324 | | ('9a', 'child', 'validation') | 2 | 331 | | ('9a', 'child_prototypical', 'train') | 192 | 1520 | | ('9a', 'child_prototypical', 'validation') | 54 | 426 | | ('9b', 'child', 'train') | 8 | 1326 | | ('9b', 'child', 'validation') | 2 | 331 | | ('9b', 'child_prototypical', 'train') | 186 | 1783 | | ('9b', 'child_prototypical', 'validation') | 36 | 307 | | ('9c', 'child', 'train') | 8 | 1330 | | ('9c', 'child', 'validation') | 2 | 332 | | ('9c', 'child_prototypical', 'train') | 118 | 433 | | ('9c', 'child_prototypical', 'validation') | 44 | 163 | | ('9d', 'child', 'train') | 8 | 1328 | | ('9d', 'child', 'validation') | 2 | 332 | | ('9d', 'child_prototypical', 'train') | 156 | 1683 | | ('9d', 'child_prototypical', 'validation') | 30 | 302 | | ('9e', 'child', 'train') | 8 | 1329 | | ('9e', 'child', 'validation') | 2 | 332 | | ('9e', 'child_prototypical', 'train') | 132 | 1426 | | ('9e', 'child_prototypical', 'validation') | 42 | 475 | | ('9f', 'child', 'train') | 8 | 1328 | | ('9f', 'child', 'validation') | 2 | 331 | | ('9f', 'child_prototypical', 'train') | 158 | 1436 | | ('9f', 'child_prototypical', 'validation') | 40 | 330 | | ('9g', 'child', 'train') | 8 | 1324 | | ('9g', 'child', 'validation') | 2 | 331 | | ('9g', 'child_prototypical', 'train') | 200 | 1685 | | ('9g', 'child_prototypical', 'validation') | 46 | 384 | | ('9h', 'child', 'train') | 8 | 1325 | | ('9h', 'child', 'validation') | 2 | 331 | | ('9h', 'child_prototypical', 'train') | 190 | 1799 | | ('9h', 'child_prototypical', 'validation') | 44 | 462 | | ('9i', 'child', 'train') | 8 | 1328 | | ('9i', 'child', 'validation') | 2 | 332 | | ('9i', 'child_prototypical', 'train') | 158 | 1361 | | ('9i', 'child_prototypical', 'validation') | 28 | 252 | ### Citation Information ``` @inproceedings{jurgens-etal-2012-semeval, title = "{S}em{E}val-2012 Task 2: Measuring Degrees of Relational Similarity", author = "Jurgens, David and Mohammad, Saif and Turney, Peter and Holyoak, Keith", booktitle = "*{SEM} 2012: The First Joint Conference on Lexical and Computational Semantics {--} Volume 1: Proceedings of the main conference and the shared task, and Volume 2: Proceedings of the Sixth International Workshop on Semantic Evaluation ({S}em{E}val 2012)", month = "7-8 " # jun, year = "2012", address = "Montr{\'e}al, Canada", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/S12-1047", pages = "356--364", } ```
research-backup/semeval2012_relational_similarity_v6
[ "multilinguality:monolingual", "size_categories:1K<n<10K", "language:en", "license:other", "region:us" ]
2022-11-20T11:14:56+00:00
{"language": ["en"], "license": ["other"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "pretty_name": "SemEval2012 task 2 Relational Similarity"}
2022-11-20T11:30:10+00:00
3ca4be9c651bf7b10b8742b048bef587c01a7a5d
# Dataset Card for "relbert/semeval2012_relational_similarity_V7" ## Dataset Description - **Repository:** [RelBERT](https://github.com/asahi417/relbert) - **Paper:** [https://aclanthology.org/S12-1047/](https://aclanthology.org/S12-1047/) - **Dataset:** SemEval2012: Relational Similarity ### Dataset Summary ***IMPORTANT***: This is the same dataset as [relbert/semeval2012_relational_similarity](https://huggingface.co/datasets/relbert/semeval2012_relational_similarity), but with a different dataset construction. Relational similarity dataset from [SemEval2012 task 2](https://aclanthology.org/S12-1047/), compiled to fine-tune [RelBERT](https://github.com/asahi417/relbert) model. The dataset contains a list of positive and negative word pair from 89 pre-defined relations. The relation types are constructed on top of following 10 parent relation types. ```shell { 1: "Class Inclusion", # Hypernym 2: "Part-Whole", # Meronym, Substance Meronym 3: "Similar", # Synonym, Co-hypornym 4: "Contrast", # Antonym 5: "Attribute", # Attribute, Event 6: "Non Attribute", 7: "Case Relation", 8: "Cause-Purpose", 9: "Space-Time", 10: "Representation" } ``` Each of the parent relation is further grouped into child relation types where the definition can be found [here](https://drive.google.com/file/d/0BzcZKTSeYL8VenY0QkVpZVpxYnc/view?resourcekey=0-ZP-UARfJj39PcLroibHPHw). ## Dataset Structure ### Data Instances An example of `train` looks as follows. ``` { 'relation_type': '8d', 'positives': [ [ "breathe", "live" ], [ "study", "learn" ], [ "speak", "communicate" ], ... ] 'negatives': [ [ "starving", "hungry" ], [ "clean", "bathe" ], [ "hungry", "starving" ], ... ] } ``` ### Data Splits | name |train|validation| |---------|----:|---------:| |semeval2012_relational_similarity| 89 | 89| ### Number of Positive/Negative Word-pairs in each Split | | positives | negatives | |:------------------------------------------|------------:|------------:| | ('1', 'parent', 'train') | 110 | 680 | | ('10', 'parent', 'train') | 60 | 730 | | ('10a', 'child', 'train') | 10 | 1655 | | ('10a', 'child_prototypical', 'train') | 246 | 2438 | | ('10b', 'child', 'train') | 10 | 1656 | | ('10b', 'child_prototypical', 'train') | 234 | 2027 | | ('10c', 'child', 'train') | 10 | 1658 | | ('10c', 'child_prototypical', 'train') | 210 | 2030 | | ('10d', 'child', 'train') | 10 | 1659 | | ('10d', 'child_prototypical', 'train') | 198 | 1766 | | ('10e', 'child', 'train') | 10 | 1661 | | ('10e', 'child_prototypical', 'train') | 174 | 1118 | | ('10f', 'child', 'train') | 10 | 1659 | | ('10f', 'child_prototypical', 'train') | 198 | 1766 | | ('1a', 'child', 'train') | 10 | 1655 | | ('1a', 'child_prototypical', 'train') | 246 | 2192 | | ('1b', 'child', 'train') | 10 | 1655 | | ('1b', 'child_prototypical', 'train') | 246 | 2192 | | ('1c', 'child', 'train') | 10 | 1658 | | ('1c', 'child_prototypical', 'train') | 210 | 2030 | | ('1d', 'child', 'train') | 10 | 1653 | | ('1d', 'child_prototypical', 'train') | 270 | 2540 | | ('1e', 'child', 'train') | 10 | 1661 | | ('1e', 'child_prototypical', 'train') | 174 | 1031 | | ('2', 'parent', 'train') | 100 | 690 | | ('2a', 'child', 'train') | 10 | 1654 | | ('2a', 'child_prototypical', 'train') | 258 | 2621 | | ('2b', 'child', 'train') | 10 | 1658 | | ('2b', 'child_prototypical', 'train') | 210 | 1610 | | ('2c', 'child', 'train') | 10 | 1656 | | ('2c', 'child_prototypical', 'train') | 234 | 2144 | | ('2d', 'child', 'train') | 10 | 1659 | | ('2d', 'child_prototypical', 'train') | 198 | 1667 | | ('2e', 'child', 'train') | 10 | 1658 | | ('2e', 'child_prototypical', 'train') | 210 | 1925 | | ('2f', 'child', 'train') | 10 | 1658 | | ('2f', 'child_prototypical', 'train') | 210 | 2240 | | ('2g', 'child', 'train') | 10 | 1653 | | ('2g', 'child_prototypical', 'train') | 270 | 2405 | | ('2h', 'child', 'train') | 10 | 1658 | | ('2h', 'child_prototypical', 'train') | 210 | 1925 | | ('2i', 'child', 'train') | 10 | 1660 | | ('2i', 'child_prototypical', 'train') | 186 | 1706 | | ('2j', 'child', 'train') | 10 | 1659 | | ('2j', 'child_prototypical', 'train') | 198 | 1964 | | ('3', 'parent', 'train') | 80 | 710 | | ('3a', 'child', 'train') | 10 | 1658 | | ('3a', 'child_prototypical', 'train') | 210 | 1925 | | ('3b', 'child', 'train') | 10 | 1658 | | ('3b', 'child_prototypical', 'train') | 210 | 2240 | | ('3c', 'child', 'train') | 10 | 1657 | | ('3c', 'child_prototypical', 'train') | 222 | 1979 | | ('3d', 'child', 'train') | 10 | 1655 | | ('3d', 'child_prototypical', 'train') | 246 | 2315 | | ('3e', 'child', 'train') | 10 | 1664 | | ('3e', 'child_prototypical', 'train') | 138 | 1268 | | ('3f', 'child', 'train') | 10 | 1658 | | ('3f', 'child_prototypical', 'train') | 210 | 2345 | | ('3g', 'child', 'train') | 10 | 1663 | | ('3g', 'child_prototypical', 'train') | 150 | 1340 | | ('3h', 'child', 'train') | 10 | 1659 | | ('3h', 'child_prototypical', 'train') | 198 | 1964 | | ('4', 'parent', 'train') | 80 | 710 | | ('4a', 'child', 'train') | 10 | 1658 | | ('4a', 'child_prototypical', 'train') | 210 | 2240 | | ('4b', 'child', 'train') | 10 | 1662 | | ('4b', 'child_prototypical', 'train') | 162 | 1163 | | ('4c', 'child', 'train') | 10 | 1657 | | ('4c', 'child_prototypical', 'train') | 222 | 2201 | | ('4d', 'child', 'train') | 10 | 1665 | | ('4d', 'child_prototypical', 'train') | 126 | 749 | | ('4e', 'child', 'train') | 10 | 1657 | | ('4e', 'child_prototypical', 'train') | 222 | 2423 | | ('4f', 'child', 'train') | 10 | 1660 | | ('4f', 'child_prototypical', 'train') | 186 | 1892 | | ('4g', 'child', 'train') | 10 | 1654 | | ('4g', 'child_prototypical', 'train') | 258 | 2492 | | ('4h', 'child', 'train') | 10 | 1657 | | ('4h', 'child_prototypical', 'train') | 222 | 2312 | | ('5', 'parent', 'train') | 90 | 700 | | ('5a', 'child', 'train') | 10 | 1655 | | ('5a', 'child_prototypical', 'train') | 246 | 2315 | | ('5b', 'child', 'train') | 10 | 1661 | | ('5b', 'child_prototypical', 'train') | 174 | 1640 | | ('5c', 'child', 'train') | 10 | 1658 | | ('5c', 'child_prototypical', 'train') | 210 | 1925 | | ('5d', 'child', 'train') | 10 | 1654 | | ('5d', 'child_prototypical', 'train') | 258 | 2363 | | ('5e', 'child', 'train') | 10 | 1661 | | ('5e', 'child_prototypical', 'train') | 174 | 1640 | | ('5f', 'child', 'train') | 10 | 1658 | | ('5f', 'child_prototypical', 'train') | 210 | 2135 | | ('5g', 'child', 'train') | 10 | 1660 | | ('5g', 'child_prototypical', 'train') | 186 | 1892 | | ('5h', 'child', 'train') | 10 | 1654 | | ('5h', 'child_prototypical', 'train') | 258 | 2750 | | ('5i', 'child', 'train') | 10 | 1655 | | ('5i', 'child_prototypical', 'train') | 246 | 2561 | | ('6', 'parent', 'train') | 80 | 710 | | ('6a', 'child', 'train') | 10 | 1654 | | ('6a', 'child_prototypical', 'train') | 258 | 2492 | | ('6b', 'child', 'train') | 10 | 1658 | | ('6b', 'child_prototypical', 'train') | 210 | 2135 | | ('6c', 'child', 'train') | 10 | 1656 | | ('6c', 'child_prototypical', 'train') | 234 | 2495 | | ('6d', 'child', 'train') | 10 | 1659 | | ('6d', 'child_prototypical', 'train') | 198 | 2261 | | ('6e', 'child', 'train') | 10 | 1658 | | ('6e', 'child_prototypical', 'train') | 210 | 2135 | | ('6f', 'child', 'train') | 10 | 1657 | | ('6f', 'child_prototypical', 'train') | 222 | 2090 | | ('6g', 'child', 'train') | 10 | 1657 | | ('6g', 'child_prototypical', 'train') | 222 | 1979 | | ('6h', 'child', 'train') | 10 | 1654 | | ('6h', 'child_prototypical', 'train') | 258 | 2621 | | ('7', 'parent', 'train') | 80 | 710 | | ('7a', 'child', 'train') | 10 | 1655 | | ('7a', 'child_prototypical', 'train') | 246 | 2561 | | ('7b', 'child', 'train') | 10 | 1662 | | ('7b', 'child_prototypical', 'train') | 162 | 1082 | | ('7c', 'child', 'train') | 10 | 1658 | | ('7c', 'child_prototypical', 'train') | 210 | 1715 | | ('7d', 'child', 'train') | 10 | 1655 | | ('7d', 'child_prototypical', 'train') | 246 | 2561 | | ('7e', 'child', 'train') | 10 | 1659 | | ('7e', 'child_prototypical', 'train') | 198 | 1568 | | ('7f', 'child', 'train') | 10 | 1657 | | ('7f', 'child_prototypical', 'train') | 222 | 1757 | | ('7g', 'child', 'train') | 10 | 1660 | | ('7g', 'child_prototypical', 'train') | 186 | 1148 | | ('7h', 'child', 'train') | 10 | 1655 | | ('7h', 'child_prototypical', 'train') | 246 | 1946 | | ('8', 'parent', 'train') | 80 | 710 | | ('8a', 'child', 'train') | 10 | 1655 | | ('8a', 'child_prototypical', 'train') | 246 | 2192 | | ('8b', 'child', 'train') | 10 | 1662 | | ('8b', 'child_prototypical', 'train') | 162 | 1487 | | ('8c', 'child', 'train') | 10 | 1657 | | ('8c', 'child_prototypical', 'train') | 222 | 1757 | | ('8d', 'child', 'train') | 10 | 1656 | | ('8d', 'child_prototypical', 'train') | 234 | 1910 | | ('8e', 'child', 'train') | 10 | 1658 | | ('8e', 'child_prototypical', 'train') | 210 | 1610 | | ('8f', 'child', 'train') | 10 | 1657 | | ('8f', 'child_prototypical', 'train') | 222 | 1868 | | ('8g', 'child', 'train') | 10 | 1662 | | ('8g', 'child_prototypical', 'train') | 162 | 839 | | ('8h', 'child', 'train') | 10 | 1655 | | ('8h', 'child_prototypical', 'train') | 246 | 2315 | | ('9', 'parent', 'train') | 90 | 700 | | ('9a', 'child', 'train') | 10 | 1655 | | ('9a', 'child_prototypical', 'train') | 246 | 1946 | | ('9b', 'child', 'train') | 10 | 1657 | | ('9b', 'child_prototypical', 'train') | 222 | 2090 | | ('9c', 'child', 'train') | 10 | 1662 | | ('9c', 'child_prototypical', 'train') | 162 | 596 | | ('9d', 'child', 'train') | 10 | 1660 | | ('9d', 'child_prototypical', 'train') | 186 | 1985 | | ('9e', 'child', 'train') | 10 | 1661 | | ('9e', 'child_prototypical', 'train') | 174 | 1901 | | ('9f', 'child', 'train') | 10 | 1659 | | ('9f', 'child_prototypical', 'train') | 198 | 1766 | | ('9g', 'child', 'train') | 10 | 1655 | | ('9g', 'child_prototypical', 'train') | 246 | 2069 | | ('9h', 'child', 'train') | 10 | 1656 | | ('9h', 'child_prototypical', 'train') | 234 | 2261 | | ('9i', 'child', 'train') | 10 | 1660 | | ('9i', 'child_prototypical', 'train') | 186 | 1613 | | ('AtLocation', 'N/A', 'validation') | 960 | 4646 | | ('CapableOf', 'N/A', 'validation') | 536 | 4734 | | ('Causes', 'N/A', 'validation') | 194 | 4738 | | ('CausesDesire', 'N/A', 'validation') | 40 | 4730 | | ('CreatedBy', 'N/A', 'validation') | 4 | 3554 | | ('DefinedAs', 'N/A', 'validation') | 4 | 1182 | | ('Desires', 'N/A', 'validation') | 56 | 4732 | | ('HasA', 'N/A', 'validation') | 168 | 4772 | | ('HasFirstSubevent', 'N/A', 'validation') | 4 | 3554 | | ('HasLastSubevent', 'N/A', 'validation') | 10 | 4732 | | ('HasPrerequisite', 'N/A', 'validation') | 450 | 4744 | | ('HasProperty', 'N/A', 'validation') | 266 | 4766 | | ('HasSubevent', 'N/A', 'validation') | 330 | 4768 | | ('IsA', 'N/A', 'validation') | 816 | 4688 | | ('MadeOf', 'N/A', 'validation') | 48 | 4726 | | ('MotivatedByGoal', 'N/A', 'validation') | 50 | 4736 | | ('PartOf', 'N/A', 'validation') | 82 | 4742 | | ('ReceivesAction', 'N/A', 'validation') | 52 | 4726 | | ('SymbolOf', 'N/A', 'validation') | 4 | 1184 | | ('UsedFor', 'N/A', 'validation') | 660 | 4760 | ### Citation Information ``` @inproceedings{jurgens-etal-2012-semeval, title = "{S}em{E}val-2012 Task 2: Measuring Degrees of Relational Similarity", author = "Jurgens, David and Mohammad, Saif and Turney, Peter and Holyoak, Keith", booktitle = "*{SEM} 2012: The First Joint Conference on Lexical and Computational Semantics {--} Volume 1: Proceedings of the main conference and the shared task, and Volume 2: Proceedings of the Sixth International Workshop on Semantic Evaluation ({S}em{E}val 2012)", month = "7-8 " # jun, year = "2012", address = "Montr{\'e}al, Canada", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/S12-1047", pages = "356--364", } ```
research-backup/semeval2012_relational_similarity_v7
[ "multilinguality:monolingual", "size_categories:1K<n<10K", "language:en", "license:other", "region:us" ]
2022-11-20T11:42:11+00:00
{"language": ["en"], "license": ["other"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "pretty_name": "SemEval2012 task 2 Relational Similarity"}
2022-11-20T11:49:41+00:00
7287f4fdbaa8ecb13a8b4e6acdc299afe355e25a
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Binary Text Classification * Model: autoevaluate/binary-classification * Dataset: glue * Config: sst2 * Split: validation To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-37b497c4-c065-4454-9a21-53d55a38d3d3-2826
[ "autotrain", "evaluation", "region:us" ]
2022-11-20T13:02:16+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["glue"], "eval_info": {"task": "binary_classification", "model": "autoevaluate/binary-classification", "metrics": ["matthews_correlation"], "dataset_name": "glue", "dataset_config": "sst2", "dataset_split": "validation", "col_mapping": {"text": "sentence", "target": "label"}}}
2022-11-20T13:02:54+00:00
95c97f926a6acdebfa7392b75b0d9c80014851f3
# Dataset Card for "text_summarization_dataset5" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
shahidul034/text_summarization_dataset5
[ "region:us" ]
2022-11-20T13:07:21+00:00
{"dataset_info": {"features": [{"name": "title", "dtype": "string"}, {"name": "content", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 144216578, "num_examples": 129922}], "download_size": 49285071, "dataset_size": 144216578}}
2022-11-20T13:07:26+00:00
6c89dcbd2c8ba7b2a2851bfbed026a1986163fb1
# Dataset Card for "text_summarization_dataset6" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
shahidul034/text_summarization_dataset6
[ "region:us" ]
2022-11-20T13:08:28+00:00
{"dataset_info": {"features": [{"name": "title", "dtype": "string"}, {"name": "content", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 129713026, "num_examples": 113562}], "download_size": 44838616, "dataset_size": 129713026}}
2022-11-20T13:08:33+00:00
98dd635735326f77c2081ddcc667347d405ab126
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Binary Text Classification * Model: autoevaluate/binary-classification * Dataset: glue * Config: sst2 * Split: validation To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-304eb14a-d97c-4ab5-a495-bcda04ee4f5c-2927
[ "autotrain", "evaluation", "region:us" ]
2022-11-20T13:13:49+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["glue"], "eval_info": {"task": "binary_classification", "model": "autoevaluate/binary-classification", "metrics": ["matthews_correlation"], "dataset_name": "glue", "dataset_config": "sst2", "dataset_split": "validation", "col_mapping": {"text": "sentence", "target": "label"}}}
2022-11-20T13:14:26+00:00
a15f09cf78e46782ede2330135c7edcd030855b6
# Dataset Card for PokemonCards ### Languages All of the data is in English. ## Dataset Structure ### Data Instances ```json { "id": "pl1-1", "image_url": "https://images.pokemontcg.io/pl1/1_hires.png", "caption": "A Stage 2 Pokemon Card of type Lightning with the title ""Ampharos"" and 130 HP of rarity ""Rare Holo"" evolved from Flaaffy from the set Platinum and the flavor text: ""None"". It has the attack ""Gigavolt"" with the cost Lightning, Colorless, the energy cost 2 and the damage of 30+ with the description: ""Flip a coin. If heads, this attack does 30 damage plus 30 more damage. If tails, the Defending Pokemon is now Paralyzed."". It has the attack ""Reflect Energy"" with the cost Lightning, Colorless, Colorless, the energy cost 3 and the damage of 70 with the description: ""Move an Energy card attached to Ampharos to 1 of your Benched Pokemon."". It has the ability ""Damage Bind"" with the description: ""Each Pokemon that has any damage counters on it (both yours and your opponent's) can't use any Poke-Powers."". It has weakness against Fighting +30. It has resistance against Metal -20.", "name": "Ampharos", "hp": "130", "set_name": "Platinum" } ``` ### Data Fields - `id`: Unique ID of the pokemon card. - `image_url`: Static URL for downloading the image associated with the post. - `caption`: Caption generated for this card. - `name`: Name of the pokemon on that card. - `hp`: Health of the pokemon. - `set_name`: The name of the set the card is in. ### Data Splits All the data is contained in training set. The training set has nearly 13k instances. ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions
TheFusion21/PokemonCards
[ "task_categories:text-to-image", "task_categories:image-to-text", "task_ids:image-captioning", "annotations_creators:machine-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:cc-by-nc-4.0", "region:us" ]
2022-11-20T14:14:51+00:00
{"annotations_creators": ["machine-generated"], "language_creators": ["found"], "language": ["en"], "license": ["cc-by-nc-4.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["text-to-image", "image-to-text"], "task_ids": ["image-captioning"], "pretty_name": "Pokemoncards", "tags": []}
2022-11-21T18:28:25+00:00
7ef85b5e1c014167cf6af3a2e805f6b8e876089d
# Dataset Card for [Dataset Name] ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary [More Information Needed] ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@github-username](https://github.com/<github-username>) for adding this dataset.
DTU54DL/common-voice-test16k
[ "task_categories:token-classification", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:mit", "region:us" ]
2022-11-20T15:31:49+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["found"], "language": ["en"], "license": ["mit"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["token-classification"], "task_ids": ["token-classification-other-acronym-identification"], "paperswithcode_id": "acronym-identification", "pretty_name": "Acronym Identification Dataset", "dataset_info": {"features": [{"name": "audio", "dtype": {"audio": {"sampling_rate": 16000}}}, {"name": "sentence", "dtype": "string"}, {"name": "accent", "dtype": "string"}], "splits": [{"name": "test", "num_bytes": 710805761.334, "num_examples": 16354}], "download_size": 701898414, "dataset_size": 710805761.334}, "train-eval-index": [{"col_mapping": {"labels": "tags", "tokens": "tokens"}, "config": "default", "splits": {"eval_split": "test"}, "task": "token-classification", "task_id": "entity_extraction"}]}
2022-11-21T13:49:22+00:00
0c2e1e9f4f84bcbd79a604526f991e24f1e0e6a4
nlhappy/DuEE
[ "license:mit", "region:us" ]
2022-11-20T15:35:01+00:00
{"license": "mit"}
2022-11-20T15:36:40+00:00
a952b9be4c1e646e4f7d161901cf552ab42ea888
# Dataset Card for "QRsst2" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
gojiteji/QRsst2
[ "region:us" ]
2022-11-20T15:50:22+00:00
{"dataset_info": {"features": [{"name": "idx", "dtype": "int32"}, {"name": "text", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "negative", "1": "positive"}}}}, {"name": "image", "dtype": "image"}], "splits": [{"name": "train", "num_bytes": 150258864.979, "num_examples": 67349}], "download_size": 77510123, "dataset_size": 150258864.979}}
2022-11-20T15:50:42+00:00
289092973de457c245888e603961b204134cc3b9
# Dataset Card for "pokemon-split" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
yuvalkirstain/pokemon-split
[ "region:us" ]
2022-11-20T16:29:18+00:00
{"dataset_info": {"features": [{"name": "image", "dtype": "image"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 107374569.23529412, "num_examples": 749}, {"name": "test", "num_bytes": 12042007.764705881, "num_examples": 84}], "download_size": 99425904, "dataset_size": 119416577.0}}
2022-11-20T16:32:00+00:00
037011a224dc386aa3e0ffeeec57326f2700d635
# Dataset Card for "illusion-cards" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
chenrm/illusion-cards
[ "region:us" ]
2022-11-20T16:56:31+00:00
{"dataset_info": {"features": [{"name": "image", "dtype": "image"}], "splits": [{"name": "train", "num_bytes": 41920616810.06, "num_examples": 73190}], "download_size": 37899199783, "dataset_size": 41920616810.06}}
2022-11-20T19:14:34+00:00
d3940de7933a202d1224259db8ca95b72d54bbb5
# AutoTrain Dataset for project: goodreads_without_bookid ## Dataset Description This dataset has been automatically processed by AutoTrain for project goodreads_without_bookid. ### Languages The BCP-47 code for the dataset's language is en. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```json [ { "target": 5, "text": "This book was absolutely ADORABLE!!!!!!!!!!! It was an awesome light and FUN read. \n I loved the characters but I absolutely LOVED Cam!!!!!!!!!!!! Major Swoooon Worthy! J \n You've been checking me out haven't you? In-between your flaming insults? I feel like man candy. \n Seriously between being HOT FUNNY and OH SO VERY ADORABLE Cam was the perfect catch!! \n I'm not going out with you Cam. \n I didn't ask you at this moment now did I One side of his lips curved up. But you will eventually. \n You're delusional \n I'm determined. \n More like annoying. \n Most would say amazing. \n Cam and Avery's relationship is tough due to the secrets she keeps but he is the perfect match for breaking her out of her shell and facing her fears. \n This book is definitely a MUST READ. \n Trust me when I say this YOU will not regret it! \n www.Jenreadit.com" }, { "target": 4, "text": "3.5 stars! \n Abbi Glines' books are a guilty pleasure for me. I love the Southern charm the sexy boys and the beautiful sweet girls. When You're Back is the second book in Reese and Mase's story and other characters from my other favorite books all make appearances here. \n I loved River Captain Kipling! This guy is SEXY and broody. He is a bit mysterious and I am really looking forward to reading more about him! \n I can change your world too sweetheart. But I'll wait my turn. \n We also have Mase's cousin Aida here who gives the cold shoulder to sweet loving Reese. I really liked how Reese blossomed in this book and how loving and devoted Mase is to her. He is one of my favorite of Abbi Glines' characters and he is definitely book boyfriend material. Their scenes are touching sexy and sweet just what I expect from the Rosemary Beach series. I liked this book and recommend it if you are looking for a sexy quick summertime read!" } ] ``` ### Dataset Fields The dataset has the following fields (also called "features"): ```json { "target": "ClassLabel(num_classes=6, names=['0', '1', '2', '3', '4', '5'], id=None)", "text": "Value(dtype='string', id=None)" } ``` ### Dataset Splits This dataset is split into a train and validation split. The split sizes are as follow: | Split name | Num samples | | ------------ | ------------------- | | train | 2358 | | valid | 592 |
fernanda-dionello/autotrain-data-goodreads_without_bookid
[ "task_categories:text-classification", "language:en", "region:us" ]
2022-11-20T17:01:16+00:00
{"language": ["en"], "task_categories": ["text-classification"]}
2022-11-20T17:02:23+00:00
22983874cd7db0b3a00733fe7787ae26da4953b1
This dataset is the extension of the existing dataset Sent_anal_ukr_manual (which was made for binary classification). It includes negative (0), positive (1) and neutral (2) sentences, labeled manually. Part of sentences were taken from the book "Francesca. The Queen of Trajectories" by Dorje Batuu (Andriy Vasyliev), a modern Ukrainian author.
SergiiGurbych/Sent_anal_ukr_multi_manual
[ "region:us" ]
2022-11-20T19:04:37+00:00
{}
2022-11-20T19:15:23+00:00
f5fd4baefc53ccd725fbdea7d4dba5b8c3d482b4
# Dataset Card for "VMMRdb_make_model" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
venetis/VMMRdb_make_model
[ "region:us" ]
2022-11-20T19:09:20+00:00
{"dataset_info": {"features": [{"name": "image", "dtype": "image"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "acura_cl", "1": "acura_integra", "2": "acura_legend", "3": "acura_mdx", "4": "acura_rdx", "5": "acura_rl", "6": "acura_rsx", "7": "acura_tl", "8": "acura_tsx", "9": "audi_a3", "10": "audi_a4", "11": "audi_a6", "12": "audi_a8", "13": "audi_s4", "14": "audi_tt", "15": "bmw_323i", "16": "bmw_325i", "17": "bmw_328i", "18": "bmw_330ci", "19": "bmw_330i", "20": "bmw_335i", "21": "bmw_525i", "22": "bmw_528i", "23": "bmw_530i", "24": "bmw_535i", "25": "bmw_540i", "26": "bmw_545i", "27": "bmw_550i", "28": "bmw_740i", "29": "bmw_745i", "30": "bmw_750i", "31": "bmw_m3", "32": "bmw_m5", "33": "bmw_x3", "34": "bmw_x5", "35": "bmw_z3", "36": "bmw_z4", "37": "buick_century", "38": "buick_enclave", "39": "buick_lacrosse", "40": "buick_lesabre", "41": "buick_lucerne", "42": "buick_parkavenue", "43": "buick_regal", "44": "buick_rendezvous", "45": "buick_riviera", "46": "cadillac_catera", "47": "cadillac_cts", "48": "cadillac_deville", "49": "cadillac_eldorado", "50": "cadillac_escalade", "51": "cadillac_seville", "52": "cadillac_srx", "53": "cadillac_sts", "54": "chevrolet_astro", "55": "chevrolet_avalanche", "56": "chevrolet_aveo", "57": "chevrolet_bel air", "58": "chevrolet_blazer", "59": "chevrolet_c-k1500", "60": "chevrolet_c10", "61": "chevrolet_camaro", "62": "chevrolet_caprice", "63": "chevrolet_cavalier", "64": "chevrolet_chevelle", "65": "chevrolet_cobalt", "66": "chevrolet_colorado", "67": "chevrolet_corvette", "68": "chevrolet_cruze", "69": "chevrolet_el camino", "70": "chevrolet_equinox", "71": "chevrolet_express", "72": "chevrolet_hhr", "73": "chevrolet_impala", "74": "chevrolet_lumina", "75": "chevrolet_malibu", "76": "chevrolet_montecarlo", "77": "chevrolet_nova", "78": "chevrolet_prizm", "79": "chevrolet_s10", "80": "chevrolet_silverado", "81": "chevrolet_sonic", "82": "chevrolet_suburban", "83": "chevrolet_tahoe", "84": "chevrolet_tracker", "85": "chevrolet_trailblazer", "86": "chevrolet_traverse", "87": "chevrolet_uplander", "88": "chevrolet_venture", "89": "chrysler_200", "90": "chrysler_300", "91": "chrysler_concorde", "92": "chrysler_crossfire", "93": "chrysler_pacifica", "94": "chrysler_pt cruiser", "95": "chrysler_sebring", "96": "chrysler_town&country", "97": "chrysler_voyager", "98": "dodge_avenger", "99": "dodge_caliber", "100": "dodge_challenger", "101": "dodge_charger", "102": "dodge_dakota", "103": "dodge_dart", "104": "dodge_durango", "105": "dodge_grand caravan", "106": "dodge_intrepid", "107": "dodge_journey", "108": "dodge_magnum", "109": "dodge_neon", "110": "dodge_nitro", "111": "dodge_ram", "112": "dodge_stratus", "113": "fiat_five hundred", "114": "ford_bronco", "115": "ford_contour", "116": "ford_crown victoria", "117": "ford_e150", "118": "ford_e250", "119": "ford_e350", "120": "ford_edge", "121": "ford_escape", "122": "ford_escort", "123": "ford_excursion", "124": "ford_expedition", "125": "ford_explorer", "126": "ford_f100", "127": "ford_f150", "128": "ford_f250", "129": "ford_f350", "130": "ford_f450", "131": "ford_fiesta", "132": "ford_five hundred", "133": "ford_focus", "134": "ford_freestar", "135": "ford_fusion", "136": "ford_mustang", "137": "ford_ranger", "138": "ford_taurus", "139": "ford_thunderbird", "140": "ford_windstar", "141": "gmc_acadia", "142": "gmc_canyon", "143": "gmc_envoy", "144": "gmc_jimmy", "145": "gmc_sierra", "146": "gmc_sonoma", "147": "gmc_suburban", "148": "gmc_terrain", "149": "gmc_yukon", "150": "honda_accord", "151": "honda_civic", "152": "honda_cr-v", "153": "honda_delsol", "154": "honda_element", "155": "honda_fit", "156": "honda_odyssey", "157": "honda_passport", "158": "honda_pilot", "159": "honda_prelude", "160": "honda_ridgeline", "161": "honda_s2000", "162": "hummer_h2", "163": "hummer_h3", "164": "hyundai_accent", "165": "hyundai_azera", "166": "hyundai_elantra", "167": "hyundai_genesis", "168": "hyundai_santafe", "169": "hyundai_sonata", "170": "hyundai_tiburon", "171": "hyundai_tucson", "172": "infiniti_fx35", "173": "infiniti_g35", "174": "infiniti_g37", "175": "infiniti_i30", "176": "infiniti_i35", "177": "infiniti_m35", "178": "infiniti_q45", "179": "infiniti_qx4", "180": "infiniti_qx56", "181": "isuzu_rodeo", "182": "isuzu_trooper", "183": "jaguar_s-type", "184": "jaguar_x-type", "185": "jaguar_xj", "186": "jeep_cherokee", "187": "jeep_cj5", "188": "jeep_cj7", "189": "jeep_commander", "190": "jeep_compass", "191": "jeep_grand", "192": "jeep_liberty", "193": "jeep_patriot", "194": "jeep_wrangler", "195": "kia_amanti", "196": "kia_forte", "197": "kia_optima", "198": "kia_rio", "199": "kia_sedona", "200": "kia_sephia", "201": "kia_sorento", "202": "kia_soul", "203": "kia_spectra", "204": "kia_sportage", "205": "landrover_discovery", "206": "landrover_rangerover", "207": "lexus_es300", "208": "lexus_es330", "209": "lexus_es350", "210": "lexus_gs300", "211": "lexus_gx470", "212": "lexus_is250", "213": "lexus_is300", "214": "lexus_is350", "215": "lexus_ls400", "216": "lexus_ls430", "217": "lexus_rx300", "218": "lexus_rx330", "219": "lexus_sc430", "220": "lincoln_aviator", "221": "lincoln_continental", "222": "lincoln_ls", "223": "lincoln_mark", "224": "lincoln_mkx", "225": "lincoln_mkz", "226": "lincoln_navigator", "227": "lincoln_towncar", "228": "mazda_3", "229": "mazda_5", "230": "mazda_6", "231": "mazda_626", "232": "mazda_millenia", "233": "mazda_mpv", "234": "mazda_mx5", "235": "mazda_protege", "236": "mazda_rx7", "237": "mazda_rx8", "238": "mazda_tribute", "239": "mercedes benz_c230", "240": "mercedes benz_c240", "241": "mercedes benz_c280", "242": "mercedes benz_c300", "243": "mercedes benz_c320", "244": "mercedes benz_clk320", "245": "mercedes benz_e320", "246": "mercedes benz_e350", "247": "mercedes benz_e500", "248": "mercedes benz_ml320", "249": "mercedes benz_ml350", "250": "mercedes benz_ml500", "251": "mercedes benz_s430", "252": "mercedes benz_s500", "253": "mercedes benz_s550", "254": "mercedes benz_sl500", "255": "mercury_cougar", "256": "mercury_grandmarquis", "257": "mercury_mariner", "258": "mercury_milan", "259": "mercury_mountaineer", "260": "mercury_sable", "261": "mercury_villager", "262": "mini_cooper", "263": "mitsubishi_3000gt", "264": "mitsubishi_eclipse", "265": "mitsubishi_endeavor", "266": "mitsubishi_galant", "267": "mitsubishi_lancer", "268": "mitsubishi_mirage", "269": "mitsubishi_montero", "270": "mitsubishi_outlander", "271": "nissan_240sx", "272": "nissan_300zx", "273": "nissan_350z", "274": "nissan_altima", "275": "nissan_armada", "276": "nissan_frontier", "277": "nissan_maxima", "278": "nissan_murano", "279": "nissan_pathfinder", "280": "nissan_quest", "281": "nissan_rogue", "282": "nissan_sentra", "283": "nissan_titan", "284": "nissan_versa", "285": "nissan_xterra", "286": "oldsmobile_alero", "287": "oldsmobile_aurora", "288": "oldsmobile_bravada", "289": "oldsmobile_cutlass", "290": "oldsmobile_intrigue", "291": "oldsmobile_silhouette", "292": "plymouth_neon", "293": "plymouth_voyager", "294": "pontiac_bonneville", "295": "pontiac_firebird", "296": "pontiac_g5", "297": "pontiac_g6", "298": "pontiac_grandam", "299": "pontiac_grandprix", "300": "pontiac_gto", "301": "pontiac_montana", "302": "pontiac_sunfire", "303": "pontiac_torrent", "304": "pontiac_transam", "305": "pontiac_vibe", "306": "porsche_911", "307": "porsche_boxster", "308": "porsche_cayenne", "309": "ram_1500", "310": "saab_9-3", "311": "saab_9-5", "312": "saturn_aura", "313": "saturn_ion", "314": "saturn_l200", "315": "saturn_l300", "316": "saturn_sl1", "317": "saturn_sl2", "318": "saturn_vue", "319": "scion_tc", "320": "scion_xa", "321": "scion_xb", "322": "scion_xd", "323": "smart_fortwo", "324": "subaru_forester", "325": "subaru_impreza", "326": "subaru_legacy", "327": "subaru_outback", "328": "subaru_wrx", "329": "suzuki_forenza", "330": "suzuki_sx4", "331": "suzuki_xl7", "332": "toyota_4runner", "333": "toyota_avalon", "334": "toyota_camry", "335": "toyota_celica", "336": "toyota_corolla", "337": "toyota_echo", "338": "toyota_fjcruiser", "339": "toyota_highlander", "340": "toyota_landcruiser", "341": "toyota_matrix", "342": "toyota_mr2", "343": "toyota_pickup", "344": "toyota_prius", "345": "toyota_rav4", "346": "toyota_sequoia", "347": "toyota_sienna", "348": "toyota_solara", "349": "toyota_supra", "350": "toyota_t100", "351": "toyota_tacoma", "352": "toyota_tercel", "353": "toyota_tundra", "354": "toyota_yaris", "355": "volkswagen_beetle", "356": "volkswagen_bug", "357": "volkswagen_cc", "358": "volkswagen_eos", "359": "volkswagen_golf", "360": "volkswagen_gti", "361": "volkswagen_jetta", "362": "volkswagen_newbeetle", "363": "volkswagen_passat", "364": "volkswagen_rabbit", "365": "volkswagen_touareg", "366": "volvo_850", "367": "volvo_c70", "368": "volvo_s40", "369": "volvo_s60", "370": "volvo_s70", "371": "volvo_s80", "372": "volvo_v70", "373": "volvo_xc70", "374": "volvo_xc90"}}}}], "splits": [{"name": "train", "num_bytes": 4989542222.5, "num_examples": 268516}], "download_size": 4979329257, "dataset_size": 4989542222.5}}
2022-11-24T04:14:05+00:00
598b793d98b9ababc996a1854391cb04bb96dc43
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Binary Text Classification * Model: autoevaluate/binary-classification * Dataset: glue * Config: sst2 * Split: validation To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-7f2fc5f3-83ca-46d6-b659-8dd44f75eacd-3028
[ "autotrain", "evaluation", "region:us" ]
2022-11-20T19:35:53+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["glue"], "eval_info": {"task": "binary_classification", "model": "autoevaluate/binary-classification", "metrics": ["matthews_correlation"], "dataset_name": "glue", "dataset_config": "sst2", "dataset_split": "validation", "col_mapping": {"text": "sentence", "target": "label"}}}
2022-11-20T19:36:34+00:00
176a30c2fff5d186450f59e52a1b9635612a4912
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Binary Text Classification * Model: autoevaluate/binary-classification * Dataset: glue * Config: sst2 * Split: validation To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-3145a565-e770-4a8b-a969-f2fafcc9a1c0-3129
[ "autotrain", "evaluation", "region:us" ]
2022-11-20T19:50:23+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["glue"], "eval_info": {"task": "binary_classification", "model": "autoevaluate/binary-classification", "metrics": ["matthews_correlation"], "dataset_name": "glue", "dataset_config": "sst2", "dataset_split": "validation", "col_mapping": {"text": "sentence", "target": "label"}}}
2022-11-20T19:51:00+00:00
b8e5a3b4723ffe216f35850345aaa5784753631e
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Binary Text Classification * Model: autoevaluate/binary-classification * Dataset: glue * Config: sst2 * Split: validation To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
autoevaluate/autoeval-staging-eval-project-6598b244-9392-4c7f-a1a9-2f5ffa8b50f8-3230
[ "autotrain", "evaluation", "region:us" ]
2022-11-20T19:53:49+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["glue"], "eval_info": {"task": "binary_classification", "model": "autoevaluate/binary-classification", "metrics": ["matthews_correlation"], "dataset_name": "glue", "dataset_config": "sst2", "dataset_split": "validation", "col_mapping": {"text": "sentence", "target": "label"}}}
2022-11-20T19:54:25+00:00
b8078c22dec1be92a8bffb5b08b8bc5004cc003a
Horoscopes from website Rumbler.ru from 2004 to 2020. 73477 records. Train dataset size - 66481 Test dataset size - 6996 Split - 10%
dkagramanyan/horoscopes_ru
[ "task_categories:text-generation", "multilinguality:monolingual", "size_categories:10K<n<100K", "language:ru", "region:us" ]
2022-11-20T20:54:21+00:00
{"annotations_creators": [], "language_creators": [], "language": ["ru"], "license": [], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": [], "task_categories": ["text-generation"], "task_ids": [], "tags": [], "dataset_info": {"features": [{"name": "date", "dtype": "string"}, {"name": "sign", "dtype": "string"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "test", "num_bytes": 6532293, "num_examples": 6976}, {"name": "train", "num_bytes": 62194608, "num_examples": 66501}], "download_size": 31753326, "dataset_size": 68726901}}
2022-11-20T22:05:14+00:00
a3468539b1e0ea3aa17f9c8fc593cf307dd805d2
yapayzeka/meto
[ "license:openrail", "region:us" ]
2022-11-20T21:14:10+00:00
{"license": "openrail"}
2022-11-20T21:14:10+00:00
f0e67cd8ffe7a17c0f88f45d4c732780785aa638
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Token Classification * Model: classtest/berttest2 * Dataset: conll2003 * Config: conll2003 * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@classtest](https://huggingface.co/classtest) for evaluating this model.
autoevaluate/autoeval-eval-conll2003-conll2003-962530-2172769890
[ "autotrain", "evaluation", "region:us" ]
2022-11-20T22:35:45+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["conll2003"], "eval_info": {"task": "entity_extraction", "model": "classtest/berttest2", "metrics": [], "dataset_name": "conll2003", "dataset_config": "conll2003", "dataset_split": "test", "col_mapping": {"tokens": "tokens", "tags": "ner_tags"}}}
2022-11-20T22:37:01+00:00
baf5e74e4d6c1e0d37dbd58daf51f7e0309cc9bb
# Dataset Card for "VMMRdb_make_model_splits" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
venetis/VMMRdb_make_model_splits
[ "region:us" ]
2022-11-21T02:27:37+00:00
{"dataset_info": {"features": [{"name": "image", "dtype": "image"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "acura_cl", "1": "acura_integra", "2": "acura_legend", "3": "acura_mdx", "4": "acura_rdx", "5": "acura_rl", "6": "acura_rsx", "7": "acura_tl", "8": "acura_tsx", "9": "audi_a3", "10": "audi_a4", "11": "audi_a6", "12": "audi_a8", "13": "audi_s4", "14": "audi_tt", "15": "bmw_323i", "16": "bmw_325i", "17": "bmw_328i", "18": "bmw_330ci", "19": "bmw_330i", "20": "bmw_335i", "21": "bmw_525i", "22": "bmw_528i", "23": "bmw_530i", "24": "bmw_535i", "25": "bmw_540i", "26": "bmw_545i", "27": "bmw_550i", "28": "bmw_740i", "29": "bmw_745i", "30": "bmw_750i", "31": "bmw_m3", "32": "bmw_m5", "33": "bmw_x3", "34": "bmw_x5", "35": "bmw_z3", "36": "bmw_z4", "37": "buick_century", "38": "buick_enclave", "39": "buick_lacrosse", "40": "buick_lesabre", "41": "buick_lucerne", "42": "buick_parkavenue", "43": "buick_regal", "44": "buick_rendezvous", "45": "buick_riviera", "46": "cadillac_catera", "47": "cadillac_cts", "48": "cadillac_deville", "49": "cadillac_eldorado", "50": "cadillac_escalade", "51": "cadillac_seville", "52": "cadillac_srx", "53": "cadillac_sts", "54": "chevrolet_astro", "55": "chevrolet_avalanche", "56": "chevrolet_aveo", "57": "chevrolet_bel air", "58": "chevrolet_blazer", "59": "chevrolet_c-k1500", "60": "chevrolet_c10", "61": "chevrolet_camaro", "62": "chevrolet_caprice", "63": "chevrolet_cavalier", "64": "chevrolet_chevelle", "65": "chevrolet_cobalt", "66": "chevrolet_colorado", "67": "chevrolet_corvette", "68": "chevrolet_cruze", "69": "chevrolet_el camino", "70": "chevrolet_equinox", "71": "chevrolet_express", "72": "chevrolet_hhr", "73": "chevrolet_impala", "74": "chevrolet_lumina", "75": "chevrolet_malibu", "76": "chevrolet_montecarlo", "77": "chevrolet_nova", "78": "chevrolet_prizm", "79": "chevrolet_s10", "80": "chevrolet_silverado", "81": "chevrolet_sonic", "82": "chevrolet_suburban", "83": "chevrolet_tahoe", "84": "chevrolet_tracker", "85": "chevrolet_trailblazer", "86": "chevrolet_traverse", "87": "chevrolet_uplander", "88": "chevrolet_venture", "89": "chrysler_200", "90": "chrysler_300", "91": "chrysler_concorde", "92": "chrysler_crossfire", "93": "chrysler_pacifica", "94": "chrysler_pt cruiser", "95": "chrysler_sebring", "96": "chrysler_town&country", "97": "chrysler_voyager", "98": "dodge_avenger", "99": "dodge_caliber", "100": "dodge_challenger", "101": "dodge_charger", "102": "dodge_dakota", "103": "dodge_dart", "104": "dodge_durango", "105": "dodge_grand caravan", "106": "dodge_intrepid", "107": "dodge_journey", "108": "dodge_magnum", "109": "dodge_neon", "110": "dodge_nitro", "111": "dodge_ram", "112": "dodge_stratus", "113": "fiat_five hundred", "114": "ford_bronco", "115": "ford_contour", "116": "ford_crown victoria", "117": "ford_e150", "118": "ford_e250", "119": "ford_e350", "120": "ford_edge", "121": "ford_escape", "122": "ford_escort", "123": "ford_excursion", "124": "ford_expedition", "125": "ford_explorer", "126": "ford_f100", "127": "ford_f150", "128": "ford_f250", "129": "ford_f350", "130": "ford_f450", "131": "ford_fiesta", "132": "ford_five hundred", "133": "ford_focus", "134": "ford_freestar", "135": "ford_fusion", "136": "ford_mustang", "137": "ford_ranger", "138": "ford_taurus", "139": "ford_thunderbird", "140": "ford_windstar", "141": "gmc_acadia", "142": "gmc_canyon", "143": "gmc_envoy", "144": "gmc_jimmy", "145": "gmc_sierra", "146": "gmc_sonoma", "147": "gmc_suburban", "148": "gmc_terrain", "149": "gmc_yukon", "150": "honda_accord", "151": "honda_civic", "152": "honda_cr-v", "153": "honda_delsol", "154": "honda_element", "155": "honda_fit", "156": "honda_odyssey", "157": "honda_passport", "158": "honda_pilot", "159": "honda_prelude", "160": "honda_ridgeline", "161": "honda_s2000", "162": "hummer_h2", "163": "hummer_h3", "164": "hyundai_accent", "165": "hyundai_azera", "166": "hyundai_elantra", "167": "hyundai_genesis", "168": "hyundai_santafe", "169": "hyundai_sonata", "170": "hyundai_tiburon", "171": "hyundai_tucson", "172": "infiniti_fx35", "173": "infiniti_g35", "174": "infiniti_g37", "175": "infiniti_i30", "176": "infiniti_i35", "177": "infiniti_m35", "178": "infiniti_q45", "179": "infiniti_qx4", "180": "infiniti_qx56", "181": "isuzu_rodeo", "182": "isuzu_trooper", "183": "jaguar_s-type", "184": "jaguar_x-type", "185": "jaguar_xj", "186": "jeep_cherokee", "187": "jeep_cj5", "188": "jeep_cj7", "189": "jeep_commander", "190": "jeep_compass", "191": "jeep_grand", "192": "jeep_liberty", "193": "jeep_patriot", "194": "jeep_wrangler", "195": "kia_amanti", "196": "kia_forte", "197": "kia_optima", "198": "kia_rio", "199": "kia_sedona", "200": "kia_sephia", "201": "kia_sorento", "202": "kia_soul", "203": "kia_spectra", "204": "kia_sportage", "205": "landrover_discovery", "206": "landrover_rangerover", "207": "lexus_es300", "208": "lexus_es330", "209": "lexus_es350", "210": "lexus_gs300", "211": "lexus_gx470", "212": "lexus_is250", "213": "lexus_is300", "214": "lexus_is350", "215": "lexus_ls400", "216": "lexus_ls430", "217": "lexus_rx300", "218": "lexus_rx330", "219": "lexus_sc430", "220": "lincoln_aviator", "221": "lincoln_continental", "222": "lincoln_ls", "223": "lincoln_mark", "224": "lincoln_mkx", "225": "lincoln_mkz", "226": "lincoln_navigator", "227": "lincoln_towncar", "228": "mazda_3", "229": "mazda_5", "230": "mazda_6", "231": "mazda_626", "232": "mazda_millenia", "233": "mazda_mpv", "234": "mazda_mx5", "235": "mazda_protege", "236": "mazda_rx7", "237": "mazda_rx8", "238": "mazda_tribute", "239": "mercedes benz_c230", "240": "mercedes benz_c240", "241": "mercedes benz_c280", "242": "mercedes benz_c300", "243": "mercedes benz_c320", "244": "mercedes benz_clk320", "245": "mercedes benz_e320", "246": "mercedes benz_e350", "247": "mercedes benz_e500", "248": "mercedes benz_ml320", "249": "mercedes benz_ml350", "250": "mercedes benz_ml500", "251": "mercedes benz_s430", "252": "mercedes benz_s500", "253": "mercedes benz_s550", "254": "mercedes benz_sl500", "255": "mercury_cougar", "256": "mercury_grandmarquis", "257": "mercury_mariner", "258": "mercury_milan", "259": "mercury_mountaineer", "260": "mercury_sable", "261": "mercury_villager", "262": "mini_cooper", "263": "mitsubishi_3000gt", "264": "mitsubishi_eclipse", "265": "mitsubishi_endeavor", "266": "mitsubishi_galant", "267": "mitsubishi_lancer", "268": "mitsubishi_mirage", "269": "mitsubishi_montero", "270": "mitsubishi_outlander", "271": "nissan_240sx", "272": "nissan_300zx", "273": "nissan_350z", "274": "nissan_altima", "275": "nissan_armada", "276": "nissan_frontier", "277": "nissan_maxima", "278": "nissan_murano", "279": "nissan_pathfinder", "280": "nissan_quest", "281": "nissan_rogue", "282": "nissan_sentra", "283": "nissan_titan", "284": "nissan_versa", "285": "nissan_xterra", "286": "oldsmobile_alero", "287": "oldsmobile_aurora", "288": "oldsmobile_bravada", "289": "oldsmobile_cutlass", "290": "oldsmobile_intrigue", "291": "oldsmobile_silhouette", "292": "plymouth_neon", "293": "plymouth_voyager", "294": "pontiac_bonneville", "295": "pontiac_firebird", "296": "pontiac_g5", "297": "pontiac_g6", "298": "pontiac_grandam", "299": "pontiac_grandprix", "300": "pontiac_gto", "301": "pontiac_montana", "302": "pontiac_sunfire", "303": "pontiac_torrent", "304": "pontiac_transam", "305": "pontiac_vibe", "306": "porsche_911", "307": "porsche_boxster", "308": "porsche_cayenne", "309": "ram_1500", "310": "saab_9-3", "311": "saab_9-5", "312": "saturn_aura", "313": "saturn_ion", "314": "saturn_l200", "315": "saturn_l300", "316": "saturn_sl1", "317": "saturn_sl2", "318": "saturn_vue", "319": "scion_tc", "320": "scion_xa", "321": "scion_xb", "322": "scion_xd", "323": "smart_fortwo", "324": "subaru_forester", "325": "subaru_impreza", "326": "subaru_legacy", "327": "subaru_outback", "328": "subaru_wrx", "329": "suzuki_forenza", "330": "suzuki_sx4", "331": "suzuki_xl7", "332": "toyota_4runner", "333": "toyota_avalon", "334": "toyota_camry", "335": "toyota_celica", "336": "toyota_corolla", "337": "toyota_echo", "338": "toyota_fjcruiser", "339": "toyota_highlander", "340": "toyota_landcruiser", "341": "toyota_matrix", "342": "toyota_mr2", "343": "toyota_pickup", "344": "toyota_prius", "345": "toyota_rav4", "346": "toyota_sequoia", "347": "toyota_sienna", "348": "toyota_solara", "349": "toyota_supra", "350": "toyota_t100", "351": "toyota_tacoma", "352": "toyota_tercel", "353": "toyota_tundra", "354": "toyota_yaris", "355": "volkswagen_beetle", "356": "volkswagen_bug", "357": "volkswagen_cc", "358": "volkswagen_eos", "359": "volkswagen_golf", "360": "volkswagen_gti", "361": "volkswagen_jetta", "362": "volkswagen_newbeetle", "363": "volkswagen_passat", "364": "volkswagen_rabbit", "365": "volkswagen_touareg", "366": "volvo_850", "367": "volvo_c70", "368": "volvo_s40", "369": "volvo_s60", "370": "volvo_s70", "371": "volvo_s80", "372": "volvo_v70", "373": "volvo_xc70", "374": "volvo_xc90"}}}}], "splits": [{"name": "train", "num_bytes": 10727024689.275133, "num_examples": 241664}, {"name": "test", "num_bytes": 1178142648.144868, "num_examples": 26852}, {"name": "valid", "num_bytes": 1178142648.144868, "num_examples": 26852}], "download_size": 13055655663, "dataset_size": 13083309985.564869}}
2022-11-24T08:55:41+00:00
5ef0e53b97165af68e3c9d8897b7c28ebc4dc14a
179 images of Ataturk's face, resized to 512x512, extracted from a collection of 196 of the best quality Ataturk photos I could find online.
cansa/ataturk-photos
[ "license:cc-by-4.0", "region:us" ]
2022-11-21T02:49:59+00:00
{"license": "cc-by-4.0"}
2022-11-21T03:03:31+00:00
d521086f5182a98003a6f45b28260cf0b25db3de
# Dataset Card for "yannic-w-large" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
osanseviero/yannic-w-large
[ "task_categories:automatic-speech-recognition", "whisper", "whispering", "large", "region:us" ]
2022-11-21T02:55:16+00:00
{"task_categories": ["automatic-speech-recognition"], "dataset_info": {"features": [{"name": "CHANNEL_NAME", "dtype": "null"}, {"name": "URL", "dtype": "null"}, {"name": "TITLE", "dtype": "null"}, {"name": "DESCRIPTION", "dtype": "null"}, {"name": "TRANSCRIPTION", "dtype": "null"}, {"name": "SEGMENTS", "dtype": "null"}, {"name": "__index_level_0__", "dtype": "null"}], "splits": [{"name": "train"}], "download_size": 812, "dataset_size": 0}, "tags": ["whisper", "whispering", "large"]}
2022-11-21T02:55:20+00:00
d1756ac12f4dae524444d9e48e1426f75d7a0a92
yuniv/KoSST
[ "region:us" ]
2022-11-21T03:18:13+00:00
{}
2022-11-30T14:59:26+00:00
e8a1e6ad8f9cfa6af283de61cc3e3beaa38c9b63
# Dataset Card for "yannic_test2" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
osanseviero/yannic_test2
[ "task_categories:automatic-speech-recognition", "whisper", "whispering", "large", "region:us" ]
2022-11-21T03:23:00+00:00
{"task_categories": ["automatic-speech-recognition"], "dataset_info": {"features": [{"name": "CHANNEL_NAME", "dtype": "string"}, {"name": "URL", "dtype": "string"}, {"name": "TITLE", "dtype": "string"}, {"name": "DESCRIPTION", "dtype": "string"}, {"name": "TRANSCRIPTION", "dtype": "string"}, {"name": "SEGMENTS", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 22634324, "num_examples": 260}], "download_size": 10407175, "dataset_size": 22634324}, "tags": ["whisper", "whispering", "large"]}
2022-11-22T06:54:05+00:00
0a80f88ecb8dbae73ef3c8c7893c4bb3244c142b
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Zero-Shot Text Classification * Model: bigscience/bloom-1b7 * Dataset: futin/feed * Config: top_vi * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@futin](https://huggingface.co/futin) for evaluating this model.
autoevaluate/autoeval-eval-futin__feed-top_vi-b5257d-2174969943
[ "autotrain", "evaluation", "region:us" ]
2022-11-21T04:36:12+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["futin/feed"], "eval_info": {"task": "text_zero_shot_classification", "model": "bigscience/bloom-1b7", "metrics": [], "dataset_name": "futin/feed", "dataset_config": "top_vi", "dataset_split": "test", "col_mapping": {"text": "text", "classes": "classes", "target": "target"}}}
2022-11-21T05:28:44+00:00
a19f83d5ac97ccf20c3af988b6636677b7b4efe3
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Zero-Shot Text Classification * Model: bigscience/bloom-1b7 * Dataset: futin/feed * Config: top_en * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@futin](https://huggingface.co/futin) for evaluating this model.
autoevaluate/autoeval-eval-futin__feed-top_en-246167-2175069948
[ "autotrain", "evaluation", "region:us" ]
2022-11-21T04:36:13+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["futin/feed"], "eval_info": {"task": "text_zero_shot_classification", "model": "bigscience/bloom-1b7", "metrics": [], "dataset_name": "futin/feed", "dataset_config": "top_en", "dataset_split": "test", "col_mapping": {"text": "text", "classes": "classes", "target": "target"}}}
2022-11-21T05:28:44+00:00
dbb75621cd252fd50effb09730a14f8cd6a2a265
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Zero-Shot Text Classification * Model: bigscience/bloom-3b * Dataset: futin/feed * Config: top_vi * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@futin](https://huggingface.co/futin) for evaluating this model.
autoevaluate/autoeval-eval-futin__feed-top_vi-b5257d-2174969941
[ "autotrain", "evaluation", "region:us" ]
2022-11-21T04:36:13+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["futin/feed"], "eval_info": {"task": "text_zero_shot_classification", "model": "bigscience/bloom-3b", "metrics": [], "dataset_name": "futin/feed", "dataset_config": "top_vi", "dataset_split": "test", "col_mapping": {"text": "text", "classes": "classes", "target": "target"}}}
2022-11-21T05:51:38+00:00
b6719a97ce763a47bc4fd6ac508cc895390f1d5d
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Zero-Shot Text Classification * Model: bigscience/bloom-7b1 * Dataset: futin/feed * Config: sen_vi * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@futin](https://huggingface.co/futin) for evaluating this model.
autoevaluate/autoeval-eval-futin__feed-sen_vi-b48d12-2175169952
[ "autotrain", "evaluation", "region:us" ]
2022-11-21T04:36:13+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["futin/feed"], "eval_info": {"task": "text_zero_shot_classification", "model": "bigscience/bloom-7b1", "metrics": [], "dataset_name": "futin/feed", "dataset_config": "sen_vi", "dataset_split": "test", "col_mapping": {"text": "text", "classes": "classes", "target": "target"}}}
2022-11-21T07:20:12+00:00
2ad8c45f06c0578deb1feb9a9a019173befa34cc
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Zero-Shot Text Classification * Model: bigscience/bloom-3b * Dataset: futin/feed * Config: sen_vi * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@futin](https://huggingface.co/futin) for evaluating this model.
autoevaluate/autoeval-eval-futin__feed-sen_vi-b48d12-2175169951
[ "autotrain", "evaluation", "region:us" ]
2022-11-21T04:36:13+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["futin/feed"], "eval_info": {"task": "text_zero_shot_classification", "model": "bigscience/bloom-3b", "metrics": [], "dataset_name": "futin/feed", "dataset_config": "sen_vi", "dataset_split": "test", "col_mapping": {"text": "text", "classes": "classes", "target": "target"}}}
2022-11-21T05:33:19+00:00
ae22cb923e6d3288b5f059d7bfd0ca10db40b023
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Zero-Shot Text Classification * Model: bigscience/bloom-560m * Dataset: futin/feed * Config: top_vi * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@futin](https://huggingface.co/futin) for evaluating this model.
autoevaluate/autoeval-eval-futin__feed-top_vi-b5257d-2174969945
[ "autotrain", "evaluation", "region:us" ]
2022-11-21T04:36:14+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["futin/feed"], "eval_info": {"task": "text_zero_shot_classification", "model": "bigscience/bloom-560m", "metrics": [], "dataset_name": "futin/feed", "dataset_config": "top_vi", "dataset_split": "test", "col_mapping": {"text": "text", "classes": "classes", "target": "target"}}}
2022-11-21T05:01:06+00:00
aab4840d78c43bf642400aa88809dc09ff86f52a
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Zero-Shot Text Classification * Model: bigscience/bloom-3b * Dataset: futin/feed * Config: top_en * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@futin](https://huggingface.co/futin) for evaluating this model.
autoevaluate/autoeval-eval-futin__feed-top_en-246167-2175069946
[ "autotrain", "evaluation", "region:us" ]
2022-11-21T04:36:14+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["futin/feed"], "eval_info": {"task": "text_zero_shot_classification", "model": "bigscience/bloom-3b", "metrics": [], "dataset_name": "futin/feed", "dataset_config": "top_en", "dataset_split": "test", "col_mapping": {"text": "text", "classes": "classes", "target": "target"}}}
2022-11-21T05:55:13+00:00
eb7c5c108505231073ef09d9661066b99a848aee
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Zero-Shot Text Classification * Model: bigscience/bloom-7b1 * Dataset: futin/feed * Config: top_en * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@futin](https://huggingface.co/futin) for evaluating this model.
autoevaluate/autoeval-eval-futin__feed-top_en-246167-2175069947
[ "autotrain", "evaluation", "region:us" ]
2022-11-21T04:36:14+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["futin/feed"], "eval_info": {"task": "text_zero_shot_classification", "model": "bigscience/bloom-7b1", "metrics": [], "dataset_name": "futin/feed", "dataset_config": "top_en", "dataset_split": "test", "col_mapping": {"text": "text", "classes": "classes", "target": "target"}}}
2022-11-21T08:24:24+00:00
00d8d32cd2e78d915cb93452571265a01dca8c31
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Zero-Shot Text Classification * Model: bigscience/bloom-1b1 * Dataset: futin/feed * Config: top_vi * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@futin](https://huggingface.co/futin) for evaluating this model.
autoevaluate/autoeval-eval-futin__feed-top_vi-b5257d-2174969944
[ "autotrain", "evaluation", "region:us" ]
2022-11-21T04:36:14+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["futin/feed"], "eval_info": {"task": "text_zero_shot_classification", "model": "bigscience/bloom-1b1", "metrics": [], "dataset_name": "futin/feed", "dataset_config": "top_vi", "dataset_split": "test", "col_mapping": {"text": "text", "classes": "classes", "target": "target"}}}
2022-11-21T05:09:21+00:00
4f012f14d9a2c0f6930e2fa9b63267cf14264f1a
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Zero-Shot Text Classification * Model: bigscience/bloom-1b1 * Dataset: futin/feed * Config: top_en * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@futin](https://huggingface.co/futin) for evaluating this model.
autoevaluate/autoeval-eval-futin__feed-top_en-246167-2175069949
[ "autotrain", "evaluation", "region:us" ]
2022-11-21T04:36:18+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["futin/feed"], "eval_info": {"task": "text_zero_shot_classification", "model": "bigscience/bloom-1b1", "metrics": [], "dataset_name": "futin/feed", "dataset_config": "top_en", "dataset_split": "test", "col_mapping": {"text": "text", "classes": "classes", "target": "target"}}}
2022-11-21T05:08:39+00:00
023ac3c3e252d345e639ea0e05da4fb9cc82ed03
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Zero-Shot Text Classification * Model: bigscience/bloom-7b1 * Dataset: futin/feed * Config: top_vi * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@futin](https://huggingface.co/futin) for evaluating this model.
autoevaluate/autoeval-eval-futin__feed-top_vi-b5257d-2174969942
[ "autotrain", "evaluation", "region:us" ]
2022-11-21T04:36:18+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["futin/feed"], "eval_info": {"task": "text_zero_shot_classification", "model": "bigscience/bloom-7b1", "metrics": [], "dataset_name": "futin/feed", "dataset_config": "top_vi", "dataset_split": "test", "col_mapping": {"text": "text", "classes": "classes", "target": "target"}}}
2022-11-21T08:14:17+00:00
b3769ebe31b993f074422307636d25c7203e7c35
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Zero-Shot Text Classification * Model: bigscience/bloom-1b1 * Dataset: futin/feed * Config: sen_vi * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@futin](https://huggingface.co/futin) for evaluating this model.
autoevaluate/autoeval-eval-futin__feed-sen_vi-b48d12-2175169954
[ "autotrain", "evaluation", "region:us" ]
2022-11-21T04:36:20+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["futin/feed"], "eval_info": {"task": "text_zero_shot_classification", "model": "bigscience/bloom-1b1", "metrics": [], "dataset_name": "futin/feed", "dataset_config": "sen_vi", "dataset_split": "test", "col_mapping": {"text": "text", "classes": "classes", "target": "target"}}}
2022-11-21T05:01:10+00:00
ea127bb85aafc69e7799f77e3c5993fe28800528
# Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Zero-Shot Text Classification * Model: bigscience/bloom-560m * Dataset: futin/feed * Config: top_en * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@futin](https://huggingface.co/futin) for evaluating this model.
autoevaluate/autoeval-eval-futin__feed-top_en-246167-2175069950
[ "autotrain", "evaluation", "region:us" ]
2022-11-21T04:36:23+00:00
{"type": "predictions", "tags": ["autotrain", "evaluation"], "datasets": ["futin/feed"], "eval_info": {"task": "text_zero_shot_classification", "model": "bigscience/bloom-560m", "metrics": [], "dataset_name": "futin/feed", "dataset_config": "top_en", "dataset_split": "test", "col_mapping": {"text": "text", "classes": "classes", "target": "target"}}}
2022-11-21T05:06:31+00:00