sha
stringlengths 40
40
| text
stringlengths 1
13.4M
| id
stringlengths 2
117
| tags
listlengths 1
7.91k
| created_at
stringlengths 25
25
| metadata
stringlengths 2
875k
| last_modified
stringlengths 25
25
| arxiv
listlengths 0
25
| languages
listlengths 0
7.91k
| tags_str
stringlengths 17
159k
| text_str
stringlengths 1
447k
| text_lists
listlengths 0
352
| processed_texts
listlengths 1
353
| tokens_length
listlengths 1
353
| input_texts
listlengths 1
40
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
5d9b85ae1f28e60a8f092a3164595be89b8de5da | # Dataset Card for "k8s-kubectl-cot-20k"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | ComponentSoft/k8s-kubectl-cot-20k | [
"region:us"
]
| 2023-10-26T19:30:51+00:00 | {"dataset_info": {"features": [{"name": "objective", "dtype": "string"}, {"name": "command_name", "dtype": "string"}, {"name": "command", "dtype": "string"}, {"name": "description", "dtype": "string"}, {"name": "syntax", "dtype": "string"}, {"name": "flags", "list": [{"name": "default", "dtype": "string"}, {"name": "description", "dtype": "string"}, {"name": "option", "dtype": "string"}, {"name": "short", "dtype": "string"}]}, {"name": "question", "dtype": "string"}, {"name": "chain_of_thought", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 51338358, "num_examples": 19661}], "download_size": 0, "dataset_size": 51338358}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-27T02:54:10+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "k8s-kubectl-cot-20k"
More Information needed | [
"# Dataset Card for \"k8s-kubectl-cot-20k\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"k8s-kubectl-cot-20k\"\n\nMore Information needed"
]
| [
6,
21
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"k8s-kubectl-cot-20k\"\n\nMore Information needed"
]
|
772224dd9662654a989ddd899c0bb82331abe7f1 | # Dataset Card for "billsum"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | CJWeiss/billsum | [
"region:us"
]
| 2023-10-26T19:40:03+00:00 | {"dataset_info": {"features": [{"name": "text", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "title", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 193223866, "num_examples": 16664}, {"name": "test", "num_bytes": 38326645, "num_examples": 3332}, {"name": "valid", "num_bytes": 25911836, "num_examples": 2222}], "download_size": 107645045, "dataset_size": 257462347}} | 2023-10-26T19:40:16+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "billsum"
More Information needed | [
"# Dataset Card for \"billsum\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"billsum\"\n\nMore Information needed"
]
| [
6,
13
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"billsum\"\n\nMore Information needed"
]
|
9105b53345f48fed5d7001a73389701ac6e26b28 | this is where i put mark rober voice clips | DuckyPolice/Mark-Rober-Voice | [
"language:en",
"license:wtfpl",
"region:us"
]
| 2023-10-26T19:40:52+00:00 | {"language": ["en"], "license": "wtfpl", "pretty_name": "Mark Rober Voice Dataset"} | 2023-10-27T19:14:24+00:00 | []
| [
"en"
]
| TAGS
#language-English #license-wtfpl #region-us
| this is where i put mark rober voice clips | []
| [
"TAGS\n#language-English #license-wtfpl #region-us \n"
]
| [
18
]
| [
"passage: TAGS\n#language-English #license-wtfpl #region-us \n"
]
|
bdfce8f6b446d87fecc0d202c398bbaae836b791 | # Dataset Card for "inabs"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | CJWeiss/inabs | [
"region:us"
]
| 2023-10-26T19:42:23+00:00 | {"dataset_info": {"features": [{"name": "text", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "file", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 159441006, "num_examples": 5346}, {"name": "test", "num_bytes": 32277886, "num_examples": 1069}, {"name": "valid", "num_bytes": 21628228, "num_examples": 713}], "download_size": 103927432, "dataset_size": 213347120}} | 2023-10-26T19:42:33+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "inabs"
More Information needed | [
"# Dataset Card for \"inabs\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"inabs\"\n\nMore Information needed"
]
| [
6,
12
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"inabs\"\n\nMore Information needed"
]
|
02591fe276e813f3993f50926f43d9d00bdc9946 | # Dataset Card for "ukabs"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | CJWeiss/ukabs | [
"region:us"
]
| 2023-10-26T19:44:33+00:00 | {"dataset_info": {"features": [{"name": "judgement", "dtype": "string"}, {"name": "summary", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 54887059, "num_examples": 595}, {"name": "test", "num_bytes": 9859833, "num_examples": 119}, {"name": "valid", "num_bytes": 6659871, "num_examples": 79}], "download_size": 33858783, "dataset_size": 71406763}} | 2023-10-26T19:44:41+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "ukabs"
More Information needed | [
"# Dataset Card for \"ukabs\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"ukabs\"\n\nMore Information needed"
]
| [
6,
12
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"ukabs\"\n\nMore Information needed"
]
|
4bc95b5d3e136f00aef52a666d6a944a76787e9a | # Dataset Card for "eurlexsum"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | CJWeiss/eurlexsum | [
"region:us"
]
| 2023-10-26T19:46:45+00:00 | {"dataset_info": {"features": [{"name": "celex_id", "dtype": "string"}, {"name": "reference", "dtype": "string"}, {"name": "summary", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 109972638, "num_examples": 1128}, {"name": "test", "num_bytes": 18741974, "num_examples": 225}, {"name": "valid", "num_bytes": 12084163, "num_examples": 151}], "download_size": 56318842, "dataset_size": 140798775}} | 2023-10-26T19:46:54+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "eurlexsum"
More Information needed | [
"# Dataset Card for \"eurlexsum\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"eurlexsum\"\n\nMore Information needed"
]
| [
6,
13
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"eurlexsum\"\n\nMore Information needed"
]
|
9bcacb0446b0f895dd352164bd39938710df4a1e |
# Dataset Card for MimicGen Datasets
## Dataset Summary
This repository contains the official release of datasets for the [CoRL 2023](https://www.corl2023.org/) paper "MimicGen: A Data Generation System for Scalable Robot Learning using Human Demonstrations".
The datasets contain over 48,000 task demonstrations across 12 tasks, grouped into the following categories:
- **source**: 120 human demonstrations across 12 tasks used to automatically generate the other datasets
- **core**: 26,000 task demonstrations across 12 tasks (26 task variants)
- **object**: 2000 task demonstrations on the Mug Cleanup task with different mugs
- **robot**: 16,000 task demonstrations across 4 different robot arms on 2 tasks (4 task variants)
- **large_interpolation**: 6000 task demonstrations across 6 tasks that pose significant challenges for modern imitation learning methods
For more information please see the [website](https://mimicgen.github.io), the [paper](https://arxiv.org/abs/2310.17596), and the [code](https://github.com/NVlabs/mimicgen_environments).
## Dataset Structure
Each dataset is an hdf5 file that is readily compatible with [robomimic](https://robomimic.github.io/) --- the structure is explained [here](https://robomimic.github.io/docs/datasets/overview.html#dataset-structure).
As described in the paper, each task has a default reset distribution (D_0). Source human demonstrations (usually 10 demos) were collected on this distribution and MimicGen was subsequently used to generate large datasets (usually 1000 demos) across different task reset distributions (e.g. D_0, D_1, D_2), objects, and robots.
The datasets are split into different types:
- **source**: source human datasets used to generate all data -- this generally consists of 10 human demonstrations collected on the D_0 variant for each task.
- **core**: datasets generated with MimicGen for different task reset distributions. These correspond to the core set of results in Figure 4 of the paper.
- **object**: datasets generated with MimicGen for different objects. These correspond to the results in Appendix G of the paper.
- **robot**: datasets generated with MimicGen for different robots. These correspond to the results in Appendix F of the paper.
- **large_interpolation**: datasets generated with MimicGen using much larger interpolation segments. These correspond to the results in Appendix H in the paper.
**Note**: We found that the large_interpolation datasets pose a significant challenge for imitation learning, and have substantial room for improvement.
## Citation
Please cite the [MimicGen paper](https://arxiv.org/abs/2310.17596) if you use these datasets in your work:
```bibtex
@inproceedings{mandlekar2023mimicgen,
title={MimicGen: A Data Generation System for Scalable Robot Learning using Human Demonstrations},
author={Mandlekar, Ajay and Nasiriany, Soroush and Wen, Bowen and Akinola, Iretiayo and Narang, Yashraj and Fan, Linxi and Zhu, Yuke and Fox, Dieter},
booktitle={7th Annual Conference on Robot Learning},
year={2023}
}
``` | amandlek/mimicgen_datasets | [
"license:cc-by-nc-sa-4.0",
"arxiv:2310.17596",
"region:us"
]
| 2023-10-26T19:47:14+00:00 | {"license": "cc-by-nc-sa-4.0"} | 2023-10-27T00:21:47+00:00 | [
"2310.17596"
]
| []
| TAGS
#license-cc-by-nc-sa-4.0 #arxiv-2310.17596 #region-us
|
# Dataset Card for MimicGen Datasets
## Dataset Summary
This repository contains the official release of datasets for the CoRL 2023 paper "MimicGen: A Data Generation System for Scalable Robot Learning using Human Demonstrations".
The datasets contain over 48,000 task demonstrations across 12 tasks, grouped into the following categories:
- source: 120 human demonstrations across 12 tasks used to automatically generate the other datasets
- core: 26,000 task demonstrations across 12 tasks (26 task variants)
- object: 2000 task demonstrations on the Mug Cleanup task with different mugs
- robot: 16,000 task demonstrations across 4 different robot arms on 2 tasks (4 task variants)
- large_interpolation: 6000 task demonstrations across 6 tasks that pose significant challenges for modern imitation learning methods
For more information please see the website, the paper, and the code.
## Dataset Structure
Each dataset is an hdf5 file that is readily compatible with robomimic --- the structure is explained here.
As described in the paper, each task has a default reset distribution (D_0). Source human demonstrations (usually 10 demos) were collected on this distribution and MimicGen was subsequently used to generate large datasets (usually 1000 demos) across different task reset distributions (e.g. D_0, D_1, D_2), objects, and robots.
The datasets are split into different types:
- source: source human datasets used to generate all data -- this generally consists of 10 human demonstrations collected on the D_0 variant for each task.
- core: datasets generated with MimicGen for different task reset distributions. These correspond to the core set of results in Figure 4 of the paper.
- object: datasets generated with MimicGen for different objects. These correspond to the results in Appendix G of the paper.
- robot: datasets generated with MimicGen for different robots. These correspond to the results in Appendix F of the paper.
- large_interpolation: datasets generated with MimicGen using much larger interpolation segments. These correspond to the results in Appendix H in the paper.
Note: We found that the large_interpolation datasets pose a significant challenge for imitation learning, and have substantial room for improvement.
Please cite the MimicGen paper if you use these datasets in your work:
| [
"# Dataset Card for MimicGen Datasets",
"## Dataset Summary\n\nThis repository contains the official release of datasets for the CoRL 2023 paper \"MimicGen: A Data Generation System for Scalable Robot Learning using Human Demonstrations\".\n\nThe datasets contain over 48,000 task demonstrations across 12 tasks, grouped into the following categories:\n- source: 120 human demonstrations across 12 tasks used to automatically generate the other datasets\n- core: 26,000 task demonstrations across 12 tasks (26 task variants)\n- object: 2000 task demonstrations on the Mug Cleanup task with different mugs\n- robot: 16,000 task demonstrations across 4 different robot arms on 2 tasks (4 task variants)\n- large_interpolation: 6000 task demonstrations across 6 tasks that pose significant challenges for modern imitation learning methods\n\nFor more information please see the website, the paper, and the code.",
"## Dataset Structure\n\nEach dataset is an hdf5 file that is readily compatible with robomimic --- the structure is explained here. \n\nAs described in the paper, each task has a default reset distribution (D_0). Source human demonstrations (usually 10 demos) were collected on this distribution and MimicGen was subsequently used to generate large datasets (usually 1000 demos) across different task reset distributions (e.g. D_0, D_1, D_2), objects, and robots.\n\nThe datasets are split into different types:\n\n- source: source human datasets used to generate all data -- this generally consists of 10 human demonstrations collected on the D_0 variant for each task.\n- core: datasets generated with MimicGen for different task reset distributions. These correspond to the core set of results in Figure 4 of the paper.\n- object: datasets generated with MimicGen for different objects. These correspond to the results in Appendix G of the paper.\n- robot: datasets generated with MimicGen for different robots. These correspond to the results in Appendix F of the paper.\n- large_interpolation: datasets generated with MimicGen using much larger interpolation segments. These correspond to the results in Appendix H in the paper.\n\nNote: We found that the large_interpolation datasets pose a significant challenge for imitation learning, and have substantial room for improvement.\n\nPlease cite the MimicGen paper if you use these datasets in your work:"
]
| [
"TAGS\n#license-cc-by-nc-sa-4.0 #arxiv-2310.17596 #region-us \n",
"# Dataset Card for MimicGen Datasets",
"## Dataset Summary\n\nThis repository contains the official release of datasets for the CoRL 2023 paper \"MimicGen: A Data Generation System for Scalable Robot Learning using Human Demonstrations\".\n\nThe datasets contain over 48,000 task demonstrations across 12 tasks, grouped into the following categories:\n- source: 120 human demonstrations across 12 tasks used to automatically generate the other datasets\n- core: 26,000 task demonstrations across 12 tasks (26 task variants)\n- object: 2000 task demonstrations on the Mug Cleanup task with different mugs\n- robot: 16,000 task demonstrations across 4 different robot arms on 2 tasks (4 task variants)\n- large_interpolation: 6000 task demonstrations across 6 tasks that pose significant challenges for modern imitation learning methods\n\nFor more information please see the website, the paper, and the code.",
"## Dataset Structure\n\nEach dataset is an hdf5 file that is readily compatible with robomimic --- the structure is explained here. \n\nAs described in the paper, each task has a default reset distribution (D_0). Source human demonstrations (usually 10 demos) were collected on this distribution and MimicGen was subsequently used to generate large datasets (usually 1000 demos) across different task reset distributions (e.g. D_0, D_1, D_2), objects, and robots.\n\nThe datasets are split into different types:\n\n- source: source human datasets used to generate all data -- this generally consists of 10 human demonstrations collected on the D_0 variant for each task.\n- core: datasets generated with MimicGen for different task reset distributions. These correspond to the core set of results in Figure 4 of the paper.\n- object: datasets generated with MimicGen for different objects. These correspond to the results in Appendix G of the paper.\n- robot: datasets generated with MimicGen for different robots. These correspond to the results in Appendix F of the paper.\n- large_interpolation: datasets generated with MimicGen using much larger interpolation segments. These correspond to the results in Appendix H in the paper.\n\nNote: We found that the large_interpolation datasets pose a significant challenge for imitation learning, and have substantial room for improvement.\n\nPlease cite the MimicGen paper if you use these datasets in your work:"
]
| [
28,
11,
183,
345
]
| [
"passage: TAGS\n#license-cc-by-nc-sa-4.0 #arxiv-2310.17596 #region-us \n# Dataset Card for MimicGen Datasets## Dataset Summary\n\nThis repository contains the official release of datasets for the CoRL 2023 paper \"MimicGen: A Data Generation System for Scalable Robot Learning using Human Demonstrations\".\n\nThe datasets contain over 48,000 task demonstrations across 12 tasks, grouped into the following categories:\n- source: 120 human demonstrations across 12 tasks used to automatically generate the other datasets\n- core: 26,000 task demonstrations across 12 tasks (26 task variants)\n- object: 2000 task demonstrations on the Mug Cleanup task with different mugs\n- robot: 16,000 task demonstrations across 4 different robot arms on 2 tasks (4 task variants)\n- large_interpolation: 6000 task demonstrations across 6 tasks that pose significant challenges for modern imitation learning methods\n\nFor more information please see the website, the paper, and the code."
]
|
f3975ef5cbcc43f83f2110863e86d9e4e6bdf3e9 | # Dataset Card for "multilexsum"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | CJWeiss/multilexsum | [
"region:us"
]
| 2023-10-26T19:49:26+00:00 | {"dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "sources", "sequence": "string"}, {"name": "summary/long", "dtype": "string"}, {"name": "summary/short", "dtype": "string"}, {"name": "summary/tiny", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 1381375968, "num_examples": 3404}, {"name": "test", "num_bytes": 265556706, "num_examples": 681}, {"name": "valid", "num_bytes": 199444854, "num_examples": 454}], "download_size": 833868199, "dataset_size": 1846377528}} | 2023-10-26T19:50:07+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "multilexsum"
More Information needed | [
"# Dataset Card for \"multilexsum\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"multilexsum\"\n\nMore Information needed"
]
| [
6,
13
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"multilexsum\"\n\nMore Information needed"
]
|
30118844b5ff8f0447595c7faadb4ec9569cafcb | # Dataset Card for "govreport"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | CJWeiss/govreport | [
"region:us"
]
| 2023-10-26T19:52:05+00:00 | {"dataset_info": {"features": [{"name": "report", "dtype": "string"}, {"name": "summary", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 799538925, "num_examples": 14598}, {"name": "test", "num_bytes": 157374869, "num_examples": 2919}, {"name": "valid", "num_bytes": 103818773, "num_examples": 1946}], "download_size": 506671700, "dataset_size": 1060732567}} | 2023-10-26T19:52:30+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "govreport"
More Information needed | [
"# Dataset Card for \"govreport\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"govreport\"\n\nMore Information needed"
]
| [
6,
12
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"govreport\"\n\nMore Information needed"
]
|
e6ad81ff7d3ead31b9220ebbf2f63709fdc9387b | # Dataset Card for "es-2610-no-demoji-m"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | gg-ai/es-2610-no-demoji-m | [
"region:us"
]
| 2023-10-26T19:54:05+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}, {"split": "test", "path": "data/test-*"}, {"split": "val", "path": "data/val-*"}]}], "dataset_info": {"features": [{"name": "text", "dtype": "string"}, {"name": "clean_text", "dtype": "string"}, {"name": "sent", "dtype": "int64"}], "splits": [{"name": "train", "num_bytes": 14582372, "num_examples": 37614}, {"name": "test", "num_bytes": 2804158, "num_examples": 7523}, {"name": "val", "num_bytes": 728021, "num_examples": 1881}], "download_size": 12052915, "dataset_size": 18114551}} | 2023-10-26T19:54:11+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "es-2610-no-demoji-m"
More Information needed | [
"# Dataset Card for \"es-2610-no-demoji-m\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"es-2610-no-demoji-m\"\n\nMore Information needed"
]
| [
6,
20
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"es-2610-no-demoji-m\"\n\nMore Information needed"
]
|
b6c90df07e7f5b15bf473990dc57b41c3a6ffb96 | # Dataset Card for "lcr_final"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | CJWeiss/lcr_final | [
"region:us"
]
| 2023-10-26T19:54:27+00:00 | {"dataset_info": {"features": [{"name": "Long Text", "dtype": "string"}, {"name": "Summary", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 87287943, "num_examples": 2918}, {"name": "test", "num_bytes": 16210230, "num_examples": 584}, {"name": "valid", "num_bytes": 10483063, "num_examples": 389}], "download_size": 55981252, "dataset_size": 113981236}} | 2023-10-26T19:54:35+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "lcr_final"
More Information needed | [
"# Dataset Card for \"lcr_final\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"lcr_final\"\n\nMore Information needed"
]
| [
6,
14
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"lcr_final\"\n\nMore Information needed"
]
|
53de1a31e8a6af6b81a2951f2d3aa4833fd43640 | # Dataset Card for "TestUpload"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | MaxReynolds/TestUpload | [
"region:us"
]
| 2023-10-26T20:10:22+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}], "dataset_info": {"features": [{"name": "image", "dtype": "image"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 1216137.0, "num_examples": 10}], "download_size": 1217696, "dataset_size": 1216137.0}} | 2023-10-26T20:10:24+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "TestUpload"
More Information needed | [
"# Dataset Card for \"TestUpload\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"TestUpload\"\n\nMore Information needed"
]
| [
6,
13
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"TestUpload\"\n\nMore Information needed"
]
|
558053a10034958afe9d0348e7665f0614ce3c6e | # Dataset Card for "chemnlp-text-mofdscribe"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | kjappelbaum/chemnlp-text-mofdscribe | [
"region:us"
]
| 2023-10-26T20:23:35+00:00 | {"dataset_info": {"features": [{"name": "cif", "dtype": "string"}, {"name": "description", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 6488768, "num_examples": 1267}], "download_size": 1948433, "dataset_size": 6488768}, "configs": [{"config_name": "core", "data_files": [{"split": "train", "path": "core/train-*"}]}, {"config_name": "qmof", "data_files": [{"split": "train", "path": "qmof/train-*"}]}]} | 2023-11-01T17:27:58+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "chemnlp-text-mofdscribe"
More Information needed | [
"# Dataset Card for \"chemnlp-text-mofdscribe\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"chemnlp-text-mofdscribe\"\n\nMore Information needed"
]
| [
6,
20
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"chemnlp-text-mofdscribe\"\n\nMore Information needed"
]
|
9b13dd236f772a96d31fed88286401c7e90c5c55 | # Dataset Card for "arxiv-papers"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | zelalt/arxiv-papers | [
"region:us"
]
| 2023-10-26T20:27:44+00:00 | {"dataset_info": {"features": [{"name": "chunk", "dtype": "string"}, {"name": "id", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "summary", "dtype": "string"}, {"name": "source", "dtype": "string"}, {"name": "authors", "sequence": "string"}, {"name": "text_length", "dtype": "int64"}], "splits": [{"name": "train", "num_bytes": 681551, "num_examples": 423}], "download_size": 405180, "dataset_size": 681551}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-26T20:27:45+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "arxiv-papers"
More Information needed | [
"# Dataset Card for \"arxiv-papers\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"arxiv-papers\"\n\nMore Information needed"
]
| [
6,
16
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"arxiv-papers\"\n\nMore Information needed"
]
|
386dac514ecda8d4737c71b2ba00b1dcf8253f61 | # Dataset Card for "multitiny"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | CJWeiss/multitiny | [
"region:us"
]
| 2023-10-26T20:32:01+00:00 | {"dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "sources", "sequence": "string"}, {"name": "summary/long", "dtype": "string"}, {"name": "summary/short", "dtype": "string"}, {"name": "summary/tiny", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 489812218.2614571, "num_examples": 1207}, {"name": "test", "num_bytes": 97877726.43171807, "num_examples": 251}, {"name": "valid", "num_bytes": 63699346.36563877, "num_examples": 145}], "download_size": 465403499, "dataset_size": 651389291.0588139}} | 2023-10-26T20:32:27+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "multitiny"
More Information needed | [
"# Dataset Card for \"multitiny\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"multitiny\"\n\nMore Information needed"
]
| [
6,
12
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"multitiny\"\n\nMore Information needed"
]
|
6bfb779da07467cb2fe3a07c8dc35981ba6b3ce7 | # M-SYNTH
<!-- Provide a quick summary of the dataset. -->
M-SYNTH is a synthetic digital mammography (DM) dataset with four breast fibroglandular density distributions imaged using Monte Carlo x-ray simulations with the publicly available [Virtual Imaging Clinical Trial for Regulatory Evaluation (VICTRE)](https://github.com/DIDSR/VICTRE) toolkit.
## Dataset Details
The dataset has the following characteristics:
* Breast density: dense, heterogeneously dense, scattered, fatty
* Mass radius (mm): 5.00, 7.00, 9.00
* Mass density: 1.0, 1.06, 1.1 (ratio of radiodensity of the mass to that of fibroglandular tissue)
* Relative dose: 20%, 40%, 60%, 80%, 100% of the clinically recommended dose for each density
<p align="center">
<img src='https://raw.githubusercontent.com/DIDSR/msynth-release/main/images/examples.png' width='700'>
</p>
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [Elena Sizikova](https://esizikova.github.io/), [Niloufar Saharkhiz](https://www.linkedin.com/in/niloufar-saharkhiz/), [Diksha Sharma](https://www.linkedin.com/in/diksha-sharma-6059977/), [Miguel Lago](https://www.linkedin.com/in/milaan/), [Berkman Sahiner](https://www.linkedin.com/in/berkman-sahiner-6aa9a919/), [Jana Gut Delfino](https://www.linkedin.com/in/janadelfino/), [Aldo Badano](https://www.linkedin.com/in/aldobadano/)
- **License:** Creative Commons 1.0 Universal License (CC0)
### Dataset Sources
<!-- Provide the basic links for the dataset. -->
- **Code:** [https://github.com/DIDSR/msynth-release](https://github.com/DIDSR/msynth-release)
- **Paper:** [https://arxiv.org/pdf/2310.18494.pdf](https://arxiv.org/pdf/2310.18494.pdf)
- **Demo:** [https://github.com/DIDSR/msynth-release/tree/master/examples](https://github.com/DIDSR/msynth-release/tree/master/examples)
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
M-SYNTH is intended to facilitate testing of AI with pre-computed synthetic mammography data.
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
M-SYNTH can be used to evaluate the effect of mass size and density, breast density, and dose on AI performance in lesion detection.
M-SYNTH can be used to either train or test pre-trained AI models.
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
M-SYNTH cannot be used in lieu of real patient examples to make performance determinations.
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
M-SYNTH is organized into a directory structure that indicates the parameters. The folder
```
device_data_VICTREPhantoms_spic_[LESION_DENSITY]/[DOSE]/[BREAST_DENSITY]/2/[LESION_SIZE]/SIM/P2_[LESION_SIZE]_[BREAST_DENSITY].8337609.[PHANTOM_FILE_ID]/[PHANTOM_FILEID]/
```
contains image files imaged with the specified parameters. Note that only examples with odd PHANTOM_FILEID contain lesions, others do not.
```
$ tree data/device_data_VICTREPhantoms_spic_1.0/1.02e10/hetero/2/5.0/SIM/P2_5.0_hetero.8337609.1/1/
data/device_data_VICTREPhantoms_spic_1.0/1.02e10/hetero/2/5.0/SIM/P2_5.0_hetero.8337609.1/1/
├── DICOM_dm
│ └── 000.dcm
├── projection_DM1.loc
├── projection_DM1.mhd
└── projection_DM1.raw
```
Each folder contains mammogram data that can be read from .raw format (.mhd contains supporting data), or DICOM (.dcm) format.
Coordinates of lesions can be found in .loc files. Segmentations are stored in .raw format and can be found in data/segmentation_masks/* .
See [Github](https://github.com/DIDSR/msynth-release/tree/main/code) for examples of how to access the files, and [examples](https://github.com/DIDSR/msynth-release/tree/main/examples) for code to load each type of file.
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
Simulation-based testing is constrained to the parameter variability represented in the object model and the acquisition system.
There is a risk of misjudging model performance if the simulated examples do not capture the variability in real patients. Please
see the paper for a full discussion of biases, risks, and limitations.
## How to use it
The msynth dataset is a very large dataset so for most use cases it is recommended to make use of the streaming API of `datasets`.
The msynth dataset has three configurations: 1) device_data, 2) segmentation_mask, and 3) metadata
You can load and iterate through the dataset using the configurations with the following lines of code:
```python
from datasets import load_dataset
ds = load_dataset("didsr/msynth", 'device_data') # For device data for all breast density, mass redius, mass density, and relative dose, change configuration to 'segmentation_mask' and 'metadata' to load the segmentation masks and bound information
print(ds_data["device_data"])
# A sample data instance
{'Raw': '~\\.cache\\huggingface\\datasets\\downloads\\extracted\\59384cf05fc44e8c0cb23bb19e1fcd8f0c39720b282109d204a85561fe66bdb1\\SIM\\P2_5.0_fatty.8336179.1\\1\\projection_DM1.raw',
'mhd': '~/.cache/huggingface/datasets/downloads/extracted/59384cf05fc44e8c0cb23bb19e1fcd8f0c39720b282109d204a85561fe66bdb1/SIM/P2_5.0_fatty.8336179.1/1\\projection_DM1.mhd',
'loc': '~/.cache/huggingface/datasets/downloads/extracted/59384cf05fc44e8c0cb23bb19e1fcd8f0c39720b282109d204a85561fe66bdb1/SIM/P2_5.0_fatty.8336179.1/1\\projection_DM1.loc',
'dcm': '~/.cache/huggingface/datasets/downloads/extracted/59384cf05fc44e8c0cb23bb19e1fcd8f0c39720b282109d204a85561fe66bdb1/SIM/P2_5.0_fatty.8336179.1/1\\DICOM_dm\\000.dcm',
'density': 'fatty',
'mass_radius': 5.0}
```
Msynth dataset can also be loaded using custom breast density, mass redius, mass density, and relative dose information
```python
from datasets import load_dataset
# Dataset properties. change to 'all' to include all the values of breast density, mass redius, mass density, and relative dose information
config_kwargs = {
"lesion_density": ["1.0"],
"dose": ["20%"],
"density": ["fatty"],
"size": ["5.0"]
}
# Loading device data
ds_data = load_dataset("didsr/msynth", 'device_data', **config_kwargs)
# Loading segmentation-mask
ds_seg = load_dataset("didsr/msynth", 'segmentation_mask', **config_kwargs)
```
The meta data can also be loaded using the datasets API. An example of using metadata is given in **Demo:** [https://github.com/DIDSR/msynth-release/tree/master/examples](https://github.com/DIDSR/msynth-release/tree/master/examples)
```python
from datasets import load_dataset
# Loading metadata
ds_meta = load_dataset("didsr/msynth", 'metadata')
# A sample data instance
ds_meta['metadata'][0]
# Output
{'fatty': '~\\.cache\\huggingface\\datasets\\downloads\\extracted\\3ea85fc6b3fcc253ac8550b5d1b21db406ca9a59ea125ff8fc63d9b754c88348\\bounds\\bounds_fatty.npy',
'dense': '~\\.cache\\huggingface\\datasets\\downloads\\extracted\\3ea85fc6b3fcc253ac8550b5d1b21db406ca9a59ea125ff8fc63d9b754c88348\\bounds\\bounds_dense.npy',
'hetero': '~\\.cache\\huggingface\\datasets\\downloads\\extracted\\3ea85fc6b3fcc253ac8550b5d1b21db406ca9a59ea125ff8fc63d9b754c88348\\bounds\\bounds_hetero.npy',
'scattered': '~\\.cache\\huggingface\\datasets\\downloads\\extracted\\3ea85fc6b3fcc253ac8550b5d1b21db406ca9a59ea125ff8fc63d9b754c88348\\bounds\\bounds_scattered.npy'}
```
## Citation
```
@article{sizikova2023knowledge,
title={Knowledge-based in silico models and dataset for the comparative evaluation of mammography AI for a range of breast characteristics, lesion conspicuities and doses},
author={Sizikova, Elena and Saharkhiz, Niloufar and Sharma, Diksha and Lago, Miguel and Sahiner, Berkman and Delfino, Jana G. and Badano, Aldo},
journal={Advances in Neural Information Processing Systems},
volume={},
pages={},
year={2023}
}
```
## Related Links
1. [Virtual Imaging Clinical Trial for Regulatory Evaluation (VICTRE)](https://www.fda.gov/medical-devices/science-and-research-medical-devices/victre-silico-breast-imaging-pipeline).
2. [FDA Catalog of Regulatory Science Tools to Help Assess New Medical Devices](https://www.fda.gov/medical-devices/science-and-research-medical-devices/catalog-regulatory-science-tools-help-assess-new-medical-devices).
3. A. Badano, C. G. Graff, A. Badal, D. Sharma, R. Zeng, F. W. Samuelson, S. Glick, K. J. Myers. [Evaluation of Digital Breast Tomosynthesis as Replacement of Full-Field Digital Mammography Using an In Silico Imaging Trial](http://dx.doi.org/10.1001/jamanetworkopen.2018.5474). JAMA Network Open 2018.
4. A. Badano, M. Lago, E. Sizikova, J. G. Delfino, S. Guan, M. A. Anastasio, B. Sahiner. [The stochastic digital human is now enrolling for in silico imaging trials—methods and tools for generating digital cohorts.](http://dx.doi.org/10.1088/2516-1091/ad04c0) Progress in Biomedical Engineering 2023.
5. E. Sizikova, N. Saharkhiz, D. Sharma, M. Lago, B. Sahiner, J. G. Delfino, A. Badano. [Knowledge-based in silico models and dataset for the comparative evaluation of mammography AI](https://github.com/DIDSR/msynth-release). NeurIPS 2023 Workshop on Synthetic Data Generation with Generative AI. | didsr/msynth | [
"task_categories:image-classification",
"task_categories:image-segmentation",
"size_categories:10K<n<100K",
"license:cc0-1.0",
"medical",
"arxiv:2310.18494",
"region:us"
]
| 2023-10-26T20:32:23+00:00 | {"license": "cc0-1.0", "size_categories": ["10K<n<100K"], "task_categories": ["image-classification", "image-segmentation"], "pretty_name": "M-SYNTH", "tags": ["medical"]} | 2023-12-07T18:58:27+00:00 | [
"2310.18494"
]
| []
| TAGS
#task_categories-image-classification #task_categories-image-segmentation #size_categories-10K<n<100K #license-cc0-1.0 #medical #arxiv-2310.18494 #region-us
| # M-SYNTH
M-SYNTH is a synthetic digital mammography (DM) dataset with four breast fibroglandular density distributions imaged using Monte Carlo x-ray simulations with the publicly available Virtual Imaging Clinical Trial for Regulatory Evaluation (VICTRE) toolkit.
## Dataset Details
The dataset has the following characteristics:
* Breast density: dense, heterogeneously dense, scattered, fatty
* Mass radius (mm): 5.00, 7.00, 9.00
* Mass density: 1.0, 1.06, 1.1 (ratio of radiodensity of the mass to that of fibroglandular tissue)
* Relative dose: 20%, 40%, 60%, 80%, 100% of the clinically recommended dose for each density
<p align="center">
<img src='URL width='700'>
</p>
### Dataset Description
- Curated by: Elena Sizikova, Niloufar Saharkhiz, Diksha Sharma, Miguel Lago, Berkman Sahiner, Jana Gut Delfino, Aldo Badano
- License: Creative Commons 1.0 Universal License (CC0)
### Dataset Sources
- Code: URL
- Paper: URL
- Demo: URL
## Uses
M-SYNTH is intended to facilitate testing of AI with pre-computed synthetic mammography data.
### Direct Use
M-SYNTH can be used to evaluate the effect of mass size and density, breast density, and dose on AI performance in lesion detection.
M-SYNTH can be used to either train or test pre-trained AI models.
### Out-of-Scope Use
M-SYNTH cannot be used in lieu of real patient examples to make performance determinations.
## Dataset Structure
M-SYNTH is organized into a directory structure that indicates the parameters. The folder
contains image files imaged with the specified parameters. Note that only examples with odd PHANTOM_FILEID contain lesions, others do not.
Each folder contains mammogram data that can be read from .raw format (.mhd contains supporting data), or DICOM (.dcm) format.
Coordinates of lesions can be found in .loc files. Segmentations are stored in .raw format and can be found in data/segmentation_masks/* .
See Github for examples of how to access the files, and examples for code to load each type of file.
## Bias, Risks, and Limitations
Simulation-based testing is constrained to the parameter variability represented in the object model and the acquisition system.
There is a risk of misjudging model performance if the simulated examples do not capture the variability in real patients. Please
see the paper for a full discussion of biases, risks, and limitations.
## How to use it
The msynth dataset is a very large dataset so for most use cases it is recommended to make use of the streaming API of 'datasets'.
The msynth dataset has three configurations: 1) device_data, 2) segmentation_mask, and 3) metadata
You can load and iterate through the dataset using the configurations with the following lines of code:
Msynth dataset can also be loaded using custom breast density, mass redius, mass density, and relative dose information
The meta data can also be loaded using the datasets API. An example of using metadata is given in Demo: URL
## Related Links
1. Virtual Imaging Clinical Trial for Regulatory Evaluation (VICTRE).
2. FDA Catalog of Regulatory Science Tools to Help Assess New Medical Devices.
3. A. Badano, C. G. Graff, A. Badal, D. Sharma, R. Zeng, F. W. Samuelson, S. Glick, K. J. Myers. Evaluation of Digital Breast Tomosynthesis as Replacement of Full-Field Digital Mammography Using an In Silico Imaging Trial. JAMA Network Open 2018.
4. A. Badano, M. Lago, E. Sizikova, J. G. Delfino, S. Guan, M. A. Anastasio, B. Sahiner. The stochastic digital human is now enrolling for in silico imaging trials—methods and tools for generating digital cohorts. Progress in Biomedical Engineering 2023.
5. E. Sizikova, N. Saharkhiz, D. Sharma, M. Lago, B. Sahiner, J. G. Delfino, A. Badano. Knowledge-based in silico models and dataset for the comparative evaluation of mammography AI. NeurIPS 2023 Workshop on Synthetic Data Generation with Generative AI. | [
"# M-SYNTH\n\n\n\nM-SYNTH is a synthetic digital mammography (DM) dataset with four breast fibroglandular density distributions imaged using Monte Carlo x-ray simulations with the publicly available Virtual Imaging Clinical Trial for Regulatory Evaluation (VICTRE) toolkit.",
"## Dataset Details\n\nThe dataset has the following characteristics:\n\n* Breast density: dense, heterogeneously dense, scattered, fatty\n* Mass radius (mm): 5.00, 7.00, 9.00\n* Mass density: 1.0, 1.06, 1.1 (ratio of radiodensity of the mass to that of fibroglandular tissue)\n* Relative dose: 20%, 40%, 60%, 80%, 100% of the clinically recommended dose for each density\n\n<p align=\"center\">\n<img src='URL width='700'>\n</p>",
"### Dataset Description\n\n\n\n- Curated by: Elena Sizikova, Niloufar Saharkhiz, Diksha Sharma, Miguel Lago, Berkman Sahiner, Jana Gut Delfino, Aldo Badano\n- License: Creative Commons 1.0 Universal License (CC0)",
"### Dataset Sources\n\n\n\n- Code: URL\n- Paper: URL \n- Demo: URL",
"## Uses\n\n\n\nM-SYNTH is intended to facilitate testing of AI with pre-computed synthetic mammography data.",
"### Direct Use\n\n\n\nM-SYNTH can be used to evaluate the effect of mass size and density, breast density, and dose on AI performance in lesion detection. \nM-SYNTH can be used to either train or test pre-trained AI models.",
"### Out-of-Scope Use\n\n\n\nM-SYNTH cannot be used in lieu of real patient examples to make performance determinations.",
"## Dataset Structure\n\n\n\nM-SYNTH is organized into a directory structure that indicates the parameters. The folder\n\ncontains image files imaged with the specified parameters. Note that only examples with odd PHANTOM_FILEID contain lesions, others do not.\n\n\n\nEach folder contains mammogram data that can be read from .raw format (.mhd contains supporting data), or DICOM (.dcm) format. \nCoordinates of lesions can be found in .loc files. Segmentations are stored in .raw format and can be found in data/segmentation_masks/* .\n\nSee Github for examples of how to access the files, and examples for code to load each type of file.",
"## Bias, Risks, and Limitations\n\n\n\nSimulation-based testing is constrained to the parameter variability represented in the object model and the acquisition system. \nThere is a risk of misjudging model performance if the simulated examples do not capture the variability in real patients. Please\nsee the paper for a full discussion of biases, risks, and limitations.",
"## How to use it\nThe msynth dataset is a very large dataset so for most use cases it is recommended to make use of the streaming API of 'datasets'. \nThe msynth dataset has three configurations: 1) device_data, 2) segmentation_mask, and 3) metadata \nYou can load and iterate through the dataset using the configurations with the following lines of code:\n\n\nMsynth dataset can also be loaded using custom breast density, mass redius, mass density, and relative dose information\n\n\nThe meta data can also be loaded using the datasets API. An example of using metadata is given in Demo: URL",
"## Related Links\n1. Virtual Imaging Clinical Trial for Regulatory Evaluation (VICTRE).\n2. FDA Catalog of Regulatory Science Tools to Help Assess New Medical Devices.\n3. A. Badano, C. G. Graff, A. Badal, D. Sharma, R. Zeng, F. W. Samuelson, S. Glick, K. J. Myers. Evaluation of Digital Breast Tomosynthesis as Replacement of Full-Field Digital Mammography Using an In Silico Imaging Trial. JAMA Network Open 2018.\n4. A. Badano, M. Lago, E. Sizikova, J. G. Delfino, S. Guan, M. A. Anastasio, B. Sahiner. The stochastic digital human is now enrolling for in silico imaging trials—methods and tools for generating digital cohorts. Progress in Biomedical Engineering 2023. \n5. E. Sizikova, N. Saharkhiz, D. Sharma, M. Lago, B. Sahiner, J. G. Delfino, A. Badano. Knowledge-based in silico models and dataset for the comparative evaluation of mammography AI. NeurIPS 2023 Workshop on Synthetic Data Generation with Generative AI."
]
| [
"TAGS\n#task_categories-image-classification #task_categories-image-segmentation #size_categories-10K<n<100K #license-cc0-1.0 #medical #arxiv-2310.18494 #region-us \n",
"# M-SYNTH\n\n\n\nM-SYNTH is a synthetic digital mammography (DM) dataset with four breast fibroglandular density distributions imaged using Monte Carlo x-ray simulations with the publicly available Virtual Imaging Clinical Trial for Regulatory Evaluation (VICTRE) toolkit.",
"## Dataset Details\n\nThe dataset has the following characteristics:\n\n* Breast density: dense, heterogeneously dense, scattered, fatty\n* Mass radius (mm): 5.00, 7.00, 9.00\n* Mass density: 1.0, 1.06, 1.1 (ratio of radiodensity of the mass to that of fibroglandular tissue)\n* Relative dose: 20%, 40%, 60%, 80%, 100% of the clinically recommended dose for each density\n\n<p align=\"center\">\n<img src='URL width='700'>\n</p>",
"### Dataset Description\n\n\n\n- Curated by: Elena Sizikova, Niloufar Saharkhiz, Diksha Sharma, Miguel Lago, Berkman Sahiner, Jana Gut Delfino, Aldo Badano\n- License: Creative Commons 1.0 Universal License (CC0)",
"### Dataset Sources\n\n\n\n- Code: URL\n- Paper: URL \n- Demo: URL",
"## Uses\n\n\n\nM-SYNTH is intended to facilitate testing of AI with pre-computed synthetic mammography data.",
"### Direct Use\n\n\n\nM-SYNTH can be used to evaluate the effect of mass size and density, breast density, and dose on AI performance in lesion detection. \nM-SYNTH can be used to either train or test pre-trained AI models.",
"### Out-of-Scope Use\n\n\n\nM-SYNTH cannot be used in lieu of real patient examples to make performance determinations.",
"## Dataset Structure\n\n\n\nM-SYNTH is organized into a directory structure that indicates the parameters. The folder\n\ncontains image files imaged with the specified parameters. Note that only examples with odd PHANTOM_FILEID contain lesions, others do not.\n\n\n\nEach folder contains mammogram data that can be read from .raw format (.mhd contains supporting data), or DICOM (.dcm) format. \nCoordinates of lesions can be found in .loc files. Segmentations are stored in .raw format and can be found in data/segmentation_masks/* .\n\nSee Github for examples of how to access the files, and examples for code to load each type of file.",
"## Bias, Risks, and Limitations\n\n\n\nSimulation-based testing is constrained to the parameter variability represented in the object model and the acquisition system. \nThere is a risk of misjudging model performance if the simulated examples do not capture the variability in real patients. Please\nsee the paper for a full discussion of biases, risks, and limitations.",
"## How to use it\nThe msynth dataset is a very large dataset so for most use cases it is recommended to make use of the streaming API of 'datasets'. \nThe msynth dataset has three configurations: 1) device_data, 2) segmentation_mask, and 3) metadata \nYou can load and iterate through the dataset using the configurations with the following lines of code:\n\n\nMsynth dataset can also be loaded using custom breast density, mass redius, mass density, and relative dose information\n\n\nThe meta data can also be loaded using the datasets API. An example of using metadata is given in Demo: URL",
"## Related Links\n1. Virtual Imaging Clinical Trial for Regulatory Evaluation (VICTRE).\n2. FDA Catalog of Regulatory Science Tools to Help Assess New Medical Devices.\n3. A. Badano, C. G. Graff, A. Badal, D. Sharma, R. Zeng, F. W. Samuelson, S. Glick, K. J. Myers. Evaluation of Digital Breast Tomosynthesis as Replacement of Full-Field Digital Mammography Using an In Silico Imaging Trial. JAMA Network Open 2018.\n4. A. Badano, M. Lago, E. Sizikova, J. G. Delfino, S. Guan, M. A. Anastasio, B. Sahiner. The stochastic digital human is now enrolling for in silico imaging trials—methods and tools for generating digital cohorts. Progress in Biomedical Engineering 2023. \n5. E. Sizikova, N. Saharkhiz, D. Sharma, M. Lago, B. Sahiner, J. G. Delfino, A. Badano. Knowledge-based in silico models and dataset for the comparative evaluation of mammography AI. NeurIPS 2023 Workshop on Synthetic Data Generation with Generative AI."
]
| [
61,
72,
133,
55,
18,
30,
61,
30,
169,
83,
145,
282
]
| [
"passage: TAGS\n#task_categories-image-classification #task_categories-image-segmentation #size_categories-10K<n<100K #license-cc0-1.0 #medical #arxiv-2310.18494 #region-us \n# M-SYNTH\n\n\n\nM-SYNTH is a synthetic digital mammography (DM) dataset with four breast fibroglandular density distributions imaged using Monte Carlo x-ray simulations with the publicly available Virtual Imaging Clinical Trial for Regulatory Evaluation (VICTRE) toolkit.## Dataset Details\n\nThe dataset has the following characteristics:\n\n* Breast density: dense, heterogeneously dense, scattered, fatty\n* Mass radius (mm): 5.00, 7.00, 9.00\n* Mass density: 1.0, 1.06, 1.1 (ratio of radiodensity of the mass to that of fibroglandular tissue)\n* Relative dose: 20%, 40%, 60%, 80%, 100% of the clinically recommended dose for each density\n\n<p align=\"center\">\n<img src='URL width='700'>\n</p>### Dataset Description\n\n\n\n- Curated by: Elena Sizikova, Niloufar Saharkhiz, Diksha Sharma, Miguel Lago, Berkman Sahiner, Jana Gut Delfino, Aldo Badano\n- License: Creative Commons 1.0 Universal License (CC0)### Dataset Sources\n\n\n\n- Code: URL\n- Paper: URL \n- Demo: URL## Uses\n\n\n\nM-SYNTH is intended to facilitate testing of AI with pre-computed synthetic mammography data.### Direct Use\n\n\n\nM-SYNTH can be used to evaluate the effect of mass size and density, breast density, and dose on AI performance in lesion detection. \nM-SYNTH can be used to either train or test pre-trained AI models.### Out-of-Scope Use\n\n\n\nM-SYNTH cannot be used in lieu of real patient examples to make performance determinations.",
"passage: ## Dataset Structure\n\n\n\nM-SYNTH is organized into a directory structure that indicates the parameters. The folder\n\ncontains image files imaged with the specified parameters. Note that only examples with odd PHANTOM_FILEID contain lesions, others do not.\n\n\n\nEach folder contains mammogram data that can be read from .raw format (.mhd contains supporting data), or DICOM (.dcm) format. \nCoordinates of lesions can be found in .loc files. Segmentations are stored in .raw format and can be found in data/segmentation_masks/* .\n\nSee Github for examples of how to access the files, and examples for code to load each type of file.## Bias, Risks, and Limitations\n\n\n\nSimulation-based testing is constrained to the parameter variability represented in the object model and the acquisition system. \nThere is a risk of misjudging model performance if the simulated examples do not capture the variability in real patients. Please\nsee the paper for a full discussion of biases, risks, and limitations.## How to use it\nThe msynth dataset is a very large dataset so for most use cases it is recommended to make use of the streaming API of 'datasets'. \nThe msynth dataset has three configurations: 1) device_data, 2) segmentation_mask, and 3) metadata \nYou can load and iterate through the dataset using the configurations with the following lines of code:\n\n\nMsynth dataset can also be loaded using custom breast density, mass redius, mass density, and relative dose information\n\n\nThe meta data can also be loaded using the datasets API. An example of using metadata is given in Demo: URL"
]
|
5b12191a2eebf084d24f40aa8472b1a27661cbfc | # Dataset Card for "multishort"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | CJWeiss/multishort | [
"region:us"
]
| 2023-10-26T20:34:18+00:00 | {"dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "sources", "sequence": "string"}, {"name": "summary/long", "dtype": "string"}, {"name": "summary/short", "dtype": "string"}, {"name": "summary/tiny", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 949594524.2185664, "num_examples": 2340}, {"name": "test", "num_bytes": 189516235.24229074, "num_examples": 486}, {"name": "valid", "num_bytes": 137063421.14537445, "num_examples": 312}], "download_size": 762638149, "dataset_size": 1276174180.6062317}} | 2023-10-26T20:34:51+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "multishort"
More Information needed | [
"# Dataset Card for \"multishort\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"multishort\"\n\nMore Information needed"
]
| [
6,
12
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"multishort\"\n\nMore Information needed"
]
|
efa1b341fc7a55d578042d552bd0a97968526d9d | # Dataset Card for "multilong"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | CJWeiss/multilong | [
"region:us"
]
| 2023-10-26T20:38:00+00:00 | {"dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "sources", "sequence": "string"}, {"name": "summary/long", "dtype": "string"}, {"name": "summary/short", "dtype": "string"}, {"name": "summary/tiny", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 1381375966.0, "num_examples": 3404}, {"name": "test", "num_bytes": 265556700.0, "num_examples": 681}, {"name": "valid", "num_bytes": 199444850.0, "num_examples": 454}], "download_size": 835227494, "dataset_size": 1846377516.0}} | 2023-10-26T20:38:41+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "multilong"
More Information needed | [
"# Dataset Card for \"multilong\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"multilong\"\n\nMore Information needed"
]
| [
6,
12
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"multilong\"\n\nMore Information needed"
]
|
6aa1ae65597fd6d17112a90c5e6851451cefe67c |
## Overview
This data was used to evaluate the two models below to decide whether convergence was reached.
https://huggingface.co/PDBEurope/BiomedNLP-PubMedBERT-ProteinStructure-NER-v2.1
https://huggingface.co/PDBEurope/BiomedNLP-PubMedBERT-ProteinStructure-NER-v3.1
There are 20 different entity types in this dataset:
"bond_interaction", "chemical", "complex_assembly", "evidence", "experimental_method", "gene",
"mutant", "oligomeric_state", "protein", "protein_state", "protein_type", "ptm", "residue_name",
"residue_name_number","residue_number", "residue_range", "site", "species", "structure_element",
"taxonomy_domain"
Annotation was carried out with the free annotation tool TeamTat (https://www.teamtat.org/) and
documents were downloaded as BioC XML before converting them to IOB, annotation only JSON and CSV format.
The number of annotations and sentences in each file is given below:
| document ID | number of annotations in BioC XML | number of annotations in IOB/JSON/CSV | number of sentences |
| --- | --- | --- | --- |
| PMC5173035 | 885 | 885 | 195 |
| PMC4993997 | 1052 | 1051 | 217 |
| PMC5014086 | 676 | 676 | 136 |
| PMC5063996 | 1048 | 1046 | 243 |
| PMC4980666 | 669 | 669 | 164 |
| PMC4817029 | 897 | 897 | 180 |
| PMC5012862 | 2203 | 2202 | 438 |
| PMC4981400 | 570 | 570 | 121 |
| PMC4806292 | 760 | 760 | 167 |
| PMC5603727 | 1353 | 1353 | 240 |
| total | 10113 | 10109 | 2101 |
Documents and annotations are easiest viewed by using the BioC XML files and opening
them in free annotation tool TeamTat (https://www.teamtat.org/). More about the BioC
format can be found here: https://bioc.sourceforge.net/
## Raw BioC XML files
These are the raw, un-annotated XML files for the publications in the dataset in BioC format.
The files are found in the directory: "raw_BioC_XML"
There is one file for each document and they follow standard naming
"unique PubMedCentral ID"_raw.xml
## Annotations in IOB format
The IOB formated files can be found in the directory: "annotation_IOB". There is one file for each
document in the dataset and they all follow the naming "unique PubMedCentral ID".tsv.
## Annotations in BioC JSON
The BioC formated JSON files of the publications have been downloaded from the annotation
tool TeamTat. The files are found in the directory: "annotated_BioC_JSON"
There is one file for each document and they follow standard naming
"unique PubMedCentral ID"_ann.json
Each document JSON contains the following relevant keys:
* "sourceid" --> giving the numerical part of the unique PubMedCentral ID
* "text" --> containing the complete raw text of the publication as a string
* "denotations" --> containing a list of all the annotations for the text
Each annotation is a dictionary with the following keys:
* "span" --> gives the start and end of the annotatiom span defined by sub keys:
* "begin" --> character start position of annotation
* "end" --> character end position of annotation
* "obj" --> a string containing a number of terms that can be separated by ","; the order
of the terms gives the following: entity type, reference to ontology, annotator,
time stamp
* "id" --> unique annotation ID
Here an example:
```json
[{"sourceid":"4784909",
"sourcedb":"",
"project":"",
"target":"",
"text":"",
"denotations":[{"span":{"begin":24,
"end":34},
"obj":"chemical,CHEBI:,[email protected],2023-03-21T15:19:42Z",
"id":"4500"},
{"span":{"begin":50,
"end":59},
"obj":"taxonomy_domain,DUMMY:,[email protected],2023-03-21T15:15:03Z",
"id":"1281"}]
}
]
```
## Annotations in BioC XML
The BioC formated XML files of the publications have been downloaded from the annotation
tool TeamTat. The files are found in the directory: "annotated_BioC_XML"
There is one file for each document and they follow standard naming
"unique PubMedCentral ID_ann.xml
The key XML tags to be able to visualise the annotations in TeamTat as well as extracting
them to create the training data are "passage" and "offset". The "passage" tag encloses a
text passage or paragraph to which the annotations are linked. "Offset" gives the passage/
paragraph offset and allows to determine the character starting and ending postions of the
annotations. The tag "text" encloses the raw text of the passage.
Each annotation in the XML file is tagged as below:
* "annotation id=" --> giving the unique ID of the annotation
* "infon key="type"" --> giving the entity type of the annotation
* "infon key="identifier"" --> giving a reference to an ontology for the annotation
* "infon key="annotator"" --> giving the annotator
* "infon key="updated_at"" --> providing a time stamp for annotation creation/update
* "location" --> start and end character positions for the annotated text span
* "offset" --> start character position as defined by offset value
* "length" --> length of the annotation span; sum of "offset" and "length" creates
the end character position
Here is a basic example of what the BioC XML looks like. Additional tags for document
management are not given. Please refer to the documenttation to find out more.
```xml
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE collection SYSTEM "BioC.dtd">
<collection>
<source>PMC</source>
<date>20140719</date>
<key>pmc.key</key>
<document>
<id>4784909</id>
<passage>
<offset>0</offset>
<text>The Structural Basis of Coenzyme A Recycling in a Bacterial Organelle</text>
<annotation id="4500">
<infon key="type">chemical</infon>
<infon key="identifier">CHEBI:</infon>
<infon key="annotator">[email protected]</infon>
<infon key="updated_at">2023-03-21T15:19:42Z</infon>
<location offset="24" length="10"/>
<text>Coenzyme A</text>
</annotation>
</passage>
</document>
</collection>
```
## Annotations in CSV
The annotations and the relevant sentences they have been found in have also been made
available as tab-separated CSV files, one for each publication in the dataset. The files can
be found in directory "annotation_CSV". Each file is named as "unique PubMedCentral ID".csv.
The column labels in the CSV files are as follows:
* "anno_start" --> character start position of the annotation
* "anno_end" --> character end position of the annotation
* "anno_text" --> text covered by the annotation
* "entity_type" --> entity type of the annotation
* "sentence" --> sentence text in which the annotation was found
* "section" --> publication section in which the annotation was found
## Annotations in JSON
A combined JSON file was created only containing the relevant sentences and associated
annotations for each publication in the dataset. The file can be found in directory
"annotation_JSON" under the name "annotations.json".
The following keys are used:
* "PMC4850273" --> unique PubMedCentral of the publication
* "annotations" --> list of dictionaries for the relevant, annotated sentences of the
document; each dictionary has the following sub keys
* "sid" --> unique sentence ID
* "sent" --> sentence text as string
* "section" --> publication section the sentence is in
* "ner" --> nested list of annotations; each sublist contains the following items:
start character position, end character position, annotation text,
entity type
Here is an example of a sentence and its annotations:
```json
{"PMC4850273": {"annotations":
[{"sid": 0,
"sent": "Molecular Dissection of Xyloglucan Recognition in a Prominent Human Gut Symbiont",
"section": "TITLE",
"ner": [
[24,34,"Xyloglucan","chemical"],
[62,67,"Human","species"],]
},]
}}
```
| PDBEurope/protein_structure_NER_independent_val_set | [
"language:en",
"license:mit",
"biology",
"protein structure",
"token classification",
"region:us"
]
| 2023-10-26T20:41:13+00:00 | {"language": ["en"], "license": "mit", "tags": ["biology", "protein structure", "token classification"]} | 2023-11-01T09:47:21+00:00 | []
| [
"en"
]
| TAGS
#language-English #license-mit #biology #protein structure #token classification #region-us
| Overview
--------
This data was used to evaluate the two models below to decide whether convergence was reached.
URL
URL
There are 20 different entity types in this dataset:
"bond\_interaction", "chemical", "complex\_assembly", "evidence", "experimental\_method", "gene",
"mutant", "oligomeric\_state", "protein", "protein\_state", "protein\_type", "ptm", "residue\_name",
"residue\_name\_number","residue\_number", "residue\_range", "site", "species", "structure\_element",
"taxonomy\_domain"
Annotation was carried out with the free annotation tool TeamTat (URL and
documents were downloaded as BioC XML before converting them to IOB, annotation only JSON and CSV format.
The number of annotations and sentences in each file is given below:
Documents and annotations are easiest viewed by using the BioC XML files and opening
them in free annotation tool TeamTat (URL More about the BioC
format can be found here: URL
Raw BioC XML files
------------------
These are the raw, un-annotated XML files for the publications in the dataset in BioC format.
The files are found in the directory: "raw\_BioC\_XML"
There is one file for each document and they follow standard naming
"unique PubMedCentral ID"\_raw.xml
Annotations in IOB format
-------------------------
The IOB formated files can be found in the directory: "annotation\_IOB". There is one file for each
document in the dataset and they all follow the naming "unique PubMedCentral ID".tsv.
Annotations in BioC JSON
------------------------
The BioC formated JSON files of the publications have been downloaded from the annotation
tool TeamTat. The files are found in the directory: "annotated\_BioC\_JSON"
There is one file for each document and they follow standard naming
"unique PubMedCentral ID"\_ann.json
Each document JSON contains the following relevant keys:
* "sourceid" --> giving the numerical part of the unique PubMedCentral ID
* "text" --> containing the complete raw text of the publication as a string
* "denotations" --> containing a list of all the annotations for the text
Each annotation is a dictionary with the following keys:
* "span" --> gives the start and end of the annotatiom span defined by sub keys:
+ "begin" --> character start position of annotation
+ "end" --> character end position of annotation
* "obj" --> a string containing a number of terms that can be separated by ","; the order
of the terms gives the following: entity type, reference to ontology, annotator,
time stamp
* "id" --> unique annotation ID
Here an example:
Annotations in BioC XML
-----------------------
The BioC formated XML files of the publications have been downloaded from the annotation
tool TeamTat. The files are found in the directory: "annotated\_BioC\_XML"
There is one file for each document and they follow standard naming
"unique PubMedCentral ID\_ann.xml
The key XML tags to be able to visualise the annotations in TeamTat as well as extracting
them to create the training data are "passage" and "offset". The "passage" tag encloses a
text passage or paragraph to which the annotations are linked. "Offset" gives the passage/
paragraph offset and allows to determine the character starting and ending postions of the
annotations. The tag "text" encloses the raw text of the passage.
Each annotation in the XML file is tagged as below:
* "annotation id=" --> giving the unique ID of the annotation
* "infon key="type"" --> giving the entity type of the annotation
* "infon key="identifier"" --> giving a reference to an ontology for the annotation
* "infon key="annotator"" --> giving the annotator
* "infon key="updated\_at"" --> providing a time stamp for annotation creation/update
* "location" --> start and end character positions for the annotated text span
+ "offset" --> start character position as defined by offset value
+ "length" --> length of the annotation span; sum of "offset" and "length" creates
the end character position
Here is a basic example of what the BioC XML looks like. Additional tags for document
management are not given. Please refer to the documenttation to find out more.
Annotations in CSV
------------------
The annotations and the relevant sentences they have been found in have also been made
available as tab-separated CSV files, one for each publication in the dataset. The files can
be found in directory "annotation\_CSV". Each file is named as "unique PubMedCentral ID".csv.
The column labels in the CSV files are as follows:
* "anno\_start" --> character start position of the annotation
* "anno\_end" --> character end position of the annotation
* "anno\_text" --> text covered by the annotation
* "entity\_type" --> entity type of the annotation
* "sentence" --> sentence text in which the annotation was found
* "section" --> publication section in which the annotation was found
Annotations in JSON
-------------------
A combined JSON file was created only containing the relevant sentences and associated
annotations for each publication in the dataset. The file can be found in directory
"annotation\_JSON" under the name "URL".
The following keys are used:
* "PMC4850273" --> unique PubMedCentral of the publication
* "annotations" --> list of dictionaries for the relevant, annotated sentences of the
document; each dictionary has the following sub keys
+ "sid" --> unique sentence ID
+ "sent" --> sentence text as string
+ "section" --> publication section the sentence is in
+ "ner" --> nested list of annotations; each sublist contains the following items:
start character position, end character position, annotation text,
entity type
Here is an example of a sentence and its annotations:
| []
| [
"TAGS\n#language-English #license-mit #biology #protein structure #token classification #region-us \n"
]
| [
27
]
| [
"passage: TAGS\n#language-English #license-mit #biology #protein structure #token classification #region-us \n"
]
|
9c98b22f5aaa4005d0b27f06939fe4c267f1eb68 |
## Overview
This data was used to evaluate the two models below to decide whether convergence was reached.
https://huggingface.co/PDBEurope/BiomedNLP-PubMedBERT-ProteinStructure-NER-v2.1
https://huggingface.co/PDBEurope/BiomedNLP-PubMedBERT-ProteinStructure-NER-v3.1
There are 20 different entity types in this dataset:
"bond_interaction", "chemical", "complex_assembly", "evidence", "experimental_method", "gene",
"mutant", "oligomeric_state", "protein", "protein_state", "protein_type", "ptm", "residue_name",
"residue_name_number","residue_number", "residue_range", "site", "species", "structure_element",
"taxonomy_domain"
Annotation was carried out with the free annotation tool TeamTat (https://www.teamtat.org/) and
documents were downloaded as BioC XML before converting them to IOB, annotation only JSON and CSV format.
The number of annotations and sentences in each file is given below:
| document ID | number of annotations in BioC XML | number of annotations in IOB/JSON/CSV | number of sentences |
| --- | --- | --- | --- |
| PMC5173035 | 885 | 885 | 195 |
| PMC4993997 | 1052 | 1051 | 217 |
| PMC5014086 | 676 | 676 | 136 |
| PMC5063996 | 1048 | 1046 | 243 |
| PMC4980666 | 669 | 669 | 164 |
| PMC4817029 | 897 | 897 | 180 |
| PMC5012862 | 2203 | 2202 | 438 |
| PMC4981400 | 570 | 570 | 121 |
| PMC4806292 | 760 | 760 | 167 |
| PMC5603727 | 1353 | 1353 | 240 |
| total | 10113 | 10109 | 2101 |
Documents and annotations are easiest viewed by using the BioC XML files and opening
them in free annotation tool TeamTat (https://www.teamtat.org/). More about the BioC
format can be found here: https://bioc.sourceforge.net/
## Raw BioC XML files
These are the raw, un-annotated XML files for the publications in the dataset in BioC format.
The files are found in the directory: "raw_BioC_XML"
There is one file for each document and they follow standard naming
"unique PubMedCentral ID"_raw.xml
## Annotations in IOB format
The IOB formated files can be found in the directory: "annotation_IOB". There is one file for each
document in the dataset and they all follow the naming "unique PubMedCentral ID".tsv.
## Annotations in BioC JSON
The BioC formated JSON files of the publications have been downloaded from the annotation
tool TeamTat. The files are found in the directory: "annotated_BioC_JSON"
There is one file for each document and they follow standard naming
"unique PubMedCentral ID"_ann.json
Each document JSON contains the following relevant keys:
* "sourceid" --> giving the numerical part of the unique PubMedCentral ID
* "text" --> containing the complete raw text of the publication as a string
* "denotations" --> containing a list of all the annotations for the text
Each annotation is a dictionary with the following keys:
* "span" --> gives the start and end of the annotatiom span defined by sub keys:
* "begin" --> character start position of annotation
* "end" --> character end position of annotation
* "obj" --> a string containing a number of terms that can be separated by ","; the order
of the terms gives the following: entity type, reference to ontology, annotator,
time stamp
* "id" --> unique annotation ID
Here an example:
```json
[{"sourceid":"4784909",
"sourcedb":"",
"project":"",
"target":"",
"text":"",
"denotations":[{"span":{"begin":24,
"end":34},
"obj":"chemical,CHEBI:,[email protected],2023-03-21T15:19:42Z",
"id":"4500"},
{"span":{"begin":50,
"end":59},
"obj":"taxonomy_domain,DUMMY:,[email protected],2023-03-21T15:15:03Z",
"id":"1281"}]
}
]
```
## Annotations in BioC XML
The BioC formated XML files of the publications have been downloaded from the annotation
tool TeamTat. The files are found in the directory: "annotated_BioC_XML"
There is one file for each document and they follow standard naming
"unique PubMedCentral ID_ann.xml
The key XML tags to be able to visualise the annotations in TeamTat as well as extracting
them to create the training data are "passage" and "offset". The "passage" tag encloses a
text passage or paragraph to which the annotations are linked. "Offset" gives the passage/
paragraph offset and allows to determine the character starting and ending postions of the
annotations. The tag "text" encloses the raw text of the passage.
Each annotation in the XML file is tagged as below:
* "annotation id=" --> giving the unique ID of the annotation
* "infon key="type"" --> giving the entity type of the annotation
* "infon key="identifier"" --> giving a reference to an ontology for the annotation
* "infon key="annotator"" --> giving the annotator
* "infon key="updated_at"" --> providing a time stamp for annotation creation/update
* "location" --> start and end character positions for the annotated text span
* "offset" --> start character position as defined by offset value
* "length" --> length of the annotation span; sum of "offset" and "length" creates
the end character position
Here is a basic example of what the BioC XML looks like. Additional tags for document
management are not given. Please refer to the documenttation to find out more.
```xml
<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE collection SYSTEM "BioC.dtd">
<collection>
<source>PMC</source>
<date>20140719</date>
<key>pmc.key</key>
<document>
<id>4784909</id>
<passage>
<offset>0</offset>
<text>The Structural Basis of Coenzyme A Recycling in a Bacterial Organelle</text>
<annotation id="4500">
<infon key="type">chemical</infon>
<infon key="identifier">CHEBI:</infon>
<infon key="annotator">[email protected]</infon>
<infon key="updated_at">2023-03-21T15:19:42Z</infon>
<location offset="24" length="10"/>
<text>Coenzyme A</text>
</annotation>
</passage>
</document>
</collection>
```
## Annotations in CSV
The annotations and the relevant sentences they have been found in have also been made
available as tab-separated CSV files, one for each publication in the dataset. The files can
be found in directory "annotation_CSV". Each file is named as "unique PubMedCentral ID".csv.
The column labels in the CSV files are as follows:
* "anno_start" --> character start position of the annotation
* "anno_end" --> character end position of the annotation
* "anno_text" --> text covered by the annotation
* "entity_type" --> entity type of the annotation
* "sentence" --> sentence text in which the annotation was found
* "section" --> publication section in which the annotation was found
## Annotations in JSON
A combined JSON file was created only containing the relevant sentences and associated
annotations for each publication in the dataset. The file can be found in directory
"annotation_JSON" under the name "annotations.json".
The following keys are used:
* "PMC4850273" --> unique PubMedCentral of the publication
* "annotations" --> list of dictionaries for the relevant, annotated sentences of the
document; each dictionary has the following sub keys
* "sid" --> unique sentence ID
* "sent" --> sentence text as string
* "section" --> publication section the sentence is in
* "ner" --> nested list of annotations; each sublist contains the following items:
start character position, end character position, annotation text,
entity type
Here is an example of a sentence and its annotations:
```json
{"PMC4850273": {"annotations":
[{"sid": 0,
"sent": "Molecular Dissection of Xyloglucan Recognition in a Prominent Human Gut Symbiont",
"section": "TITLE",
"ner": [
[24,34,"Xyloglucan","chemical"],
[62,67,"Human","species"],]
},]
}}
```
| mevol/protein_structure_NER_independent_val_set | [
"language:en",
"license:mit",
"biology",
"protein structure",
"token classification",
"region:us"
]
| 2023-10-26T20:46:25+00:00 | {"language": ["en"], "license": "mit", "tags": ["biology", "protein structure", "token classification"]} | 2023-11-01T10:28:29+00:00 | []
| [
"en"
]
| TAGS
#language-English #license-mit #biology #protein structure #token classification #region-us
| Overview
--------
This data was used to evaluate the two models below to decide whether convergence was reached.
URL
URL
There are 20 different entity types in this dataset:
"bond\_interaction", "chemical", "complex\_assembly", "evidence", "experimental\_method", "gene",
"mutant", "oligomeric\_state", "protein", "protein\_state", "protein\_type", "ptm", "residue\_name",
"residue\_name\_number","residue\_number", "residue\_range", "site", "species", "structure\_element",
"taxonomy\_domain"
Annotation was carried out with the free annotation tool TeamTat (URL and
documents were downloaded as BioC XML before converting them to IOB, annotation only JSON and CSV format.
The number of annotations and sentences in each file is given below:
Documents and annotations are easiest viewed by using the BioC XML files and opening
them in free annotation tool TeamTat (URL More about the BioC
format can be found here: URL
Raw BioC XML files
------------------
These are the raw, un-annotated XML files for the publications in the dataset in BioC format.
The files are found in the directory: "raw\_BioC\_XML"
There is one file for each document and they follow standard naming
"unique PubMedCentral ID"\_raw.xml
Annotations in IOB format
-------------------------
The IOB formated files can be found in the directory: "annotation\_IOB". There is one file for each
document in the dataset and they all follow the naming "unique PubMedCentral ID".tsv.
Annotations in BioC JSON
------------------------
The BioC formated JSON files of the publications have been downloaded from the annotation
tool TeamTat. The files are found in the directory: "annotated\_BioC\_JSON"
There is one file for each document and they follow standard naming
"unique PubMedCentral ID"\_ann.json
Each document JSON contains the following relevant keys:
* "sourceid" --> giving the numerical part of the unique PubMedCentral ID
* "text" --> containing the complete raw text of the publication as a string
* "denotations" --> containing a list of all the annotations for the text
Each annotation is a dictionary with the following keys:
* "span" --> gives the start and end of the annotatiom span defined by sub keys:
+ "begin" --> character start position of annotation
+ "end" --> character end position of annotation
* "obj" --> a string containing a number of terms that can be separated by ","; the order
of the terms gives the following: entity type, reference to ontology, annotator,
time stamp
* "id" --> unique annotation ID
Here an example:
Annotations in BioC XML
-----------------------
The BioC formated XML files of the publications have been downloaded from the annotation
tool TeamTat. The files are found in the directory: "annotated\_BioC\_XML"
There is one file for each document and they follow standard naming
"unique PubMedCentral ID\_ann.xml
The key XML tags to be able to visualise the annotations in TeamTat as well as extracting
them to create the training data are "passage" and "offset". The "passage" tag encloses a
text passage or paragraph to which the annotations are linked. "Offset" gives the passage/
paragraph offset and allows to determine the character starting and ending postions of the
annotations. The tag "text" encloses the raw text of the passage.
Each annotation in the XML file is tagged as below:
* "annotation id=" --> giving the unique ID of the annotation
* "infon key="type"" --> giving the entity type of the annotation
* "infon key="identifier"" --> giving a reference to an ontology for the annotation
* "infon key="annotator"" --> giving the annotator
* "infon key="updated\_at"" --> providing a time stamp for annotation creation/update
* "location" --> start and end character positions for the annotated text span
+ "offset" --> start character position as defined by offset value
+ "length" --> length of the annotation span; sum of "offset" and "length" creates
the end character position
Here is a basic example of what the BioC XML looks like. Additional tags for document
management are not given. Please refer to the documenttation to find out more.
Annotations in CSV
------------------
The annotations and the relevant sentences they have been found in have also been made
available as tab-separated CSV files, one for each publication in the dataset. The files can
be found in directory "annotation\_CSV". Each file is named as "unique PubMedCentral ID".csv.
The column labels in the CSV files are as follows:
* "anno\_start" --> character start position of the annotation
* "anno\_end" --> character end position of the annotation
* "anno\_text" --> text covered by the annotation
* "entity\_type" --> entity type of the annotation
* "sentence" --> sentence text in which the annotation was found
* "section" --> publication section in which the annotation was found
Annotations in JSON
-------------------
A combined JSON file was created only containing the relevant sentences and associated
annotations for each publication in the dataset. The file can be found in directory
"annotation\_JSON" under the name "URL".
The following keys are used:
* "PMC4850273" --> unique PubMedCentral of the publication
* "annotations" --> list of dictionaries for the relevant, annotated sentences of the
document; each dictionary has the following sub keys
+ "sid" --> unique sentence ID
+ "sent" --> sentence text as string
+ "section" --> publication section the sentence is in
+ "ner" --> nested list of annotations; each sublist contains the following items:
start character position, end character position, annotation text,
entity type
Here is an example of a sentence and its annotations:
| []
| [
"TAGS\n#language-English #license-mit #biology #protein structure #token classification #region-us \n"
]
| [
27
]
| [
"passage: TAGS\n#language-English #license-mit #biology #protein structure #token classification #region-us \n"
]
|
efb9445c38933548af394c17e206505343e10348 | # Dataset Card for "scifi-book"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | faizalnf1800/scifi-book | [
"region:us"
]
| 2023-10-26T20:49:13+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}], "dataset_info": {"features": [{"name": "filename", "dtype": "string"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 3557355, "num_examples": 7}], "download_size": 2057340, "dataset_size": 3557355}} | 2023-10-27T10:43:33+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "scifi-book"
More Information needed | [
"# Dataset Card for \"scifi-book\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"scifi-book\"\n\nMore Information needed"
]
| [
6,
14
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"scifi-book\"\n\nMore Information needed"
]
|
87dffe8bd9679f87631ee696c47f49580ab7d383 |
## OncQA: The Impact of Using an AI Chatbot to Respond to Patient Messages
### Importance
Documentation burden is a major factor contributing to clinician burnout, which is increasing across the country and threatens our capacity to provide patient care in the U.S. While AI chatbots show potential in reducing this burden by aiding in documentation and are being incorporated into electronic health record systems, their influence on clinical decision-making remains understudied for this purpose.
### Objective
Investigate the acceptability, safety, and potential human factors issues when utilizing an AI-powered chatbot to draft responses to patients' inquiries.
### Design
- A 2-stage cross-sectional study was designed around 100 synthetic cancer patient scenarios couples with patient messages.
- These questions emulate realistic oncology scenarios.
- **Stage 1: Manual Reponse**: Six oncologists were randomly allocated 26 questions for response.
- **Stage 2: AI-Assisted Response**: The same oncologists received 26 new questions, alongside GPT-4 generated responses for editing.
- Informed consent was obtained.
- Participants were blinded to the source of the drafts.
- Surveys were undertaken for every scenario/response.
### About this repo
The dataset shows here is the complete stage2 parsed data with all physician edits.
If you wish to see the full data for stage1 and others please visit https://github.com/AIM-Harvard/OncQA/
### Settings
This research was conducted at the Brigham and Women’s Hospital, Boston, MA in 2023.
Q1: 'How challenging was it to respond to this message?'
Q2: 'Do you believe this patient is experiencing a severe medical event?'
Q3: 'How would you rate the acceptability of the draft response?'
Q4: 'How likely is it that the unedited draft response could cause harm?'
Q5: 'If the unedited draft does cause harm, what would be the extent, or clinical impact on the patient?'
Q6: 'Do you believe the provided unedited draft response improved your documentation efficiency?'
Q7: 'Do you believe the provided draft response was written by an AI or by a human?'
### Participants
Six board-certified oncologists participated.
### Intervention
Employment of GPT-4, an AI chatbot, for drafting responses to patient inquiries.
### Main Outcomes & Measures
- Evaluate the impact and utility of an AI chatbot in assisting responses to patient messages.
- Impact was determined by comparing response length and readability, using the Flesch reading ease score, and content.
- Utility was ascertained through physician feedback on surveys regarding acceptability, potential harm, and efficiency of chatbot-crafted drafts.

### Results
- On average, manual responses were more concise than those by GPT-4 or AI-assisted (34 vs. 169 vs. 160 words, p<0.001).
- Manual responses were more readable than GPT-4 or AI-assisted messages (Flesch score 67 vs. 45 vs. 46, p<0.001).
- About 58% of GPT-4 drafts were immediately acceptable, with 82% posing a low risk of harm.
- Utilizing the GPT-4 draft enhanced documentation efficiency in 77% of replies.
- Surprisingly, 31% of GPT-4 responses were perceived to be human-written, despite being AI-generated.
- 7.7% of survey responses felt unedited GPT-4 drafts could lead to severe harm or death.
- Among 56 dual-annotated responses, annotation agreement was low for manual responses (Cohen's kappa 0.10), but improved for AI-assisted responses (Cohen's kappa 0.52).
- AI-assistance led to differences in clinical content in the responses (p=0.001).
- Manual replies were more likely to advise direct clinical actions, while GPT-4 drafts often provided educational and self-management suggestions.
- AI-aided replies closely mirrored GPT-4 drafts but introduced some direct clinical actions.
### Conclusions & Relevance
AI-generated chatbot responses, while lengthier and less accessible, were overall safe and improved efficiency. AI-assistance altered the nature of physician feedback and reduced variability. AI chatbots are a promising avenue to address physician burnout and could improve patient care, however interactions between humans and AI might affect clinical decisions in unexpected ways. Addressing these interactions is vital for the safe incorporation of such technologies.
**Note**: It's imperative to delve deeper into human-AI interactions and their potential impact on outcomes.
# Citation:
```
@misc{chen2023impact,
title={The impact of using an AI chatbot to respond to patient messages},
author={Shan Chen and Marco Guevara and Shalini Moningi and Frank Hoebers and Hesham Elhalawani and Benjamin H. Kann and Fallon E. Chipidza and Jonathan Leeman and Hugo J. W. L. Aerts and Timothy Miller and Guergana K. Savova and Raymond H. Mak and Maryam Lustberg and Majid Afshar and Danielle S. Bitterman},
year={2023},
eprint={2310.17703},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
``` | shanchen/OncQA | [
"task_categories:conversational",
"task_categories:text2text-generation",
"language:en",
"license:cc-by-sa-4.0",
"medical",
"arxiv:2310.17703",
"region:us"
]
| 2023-10-26T20:56:33+00:00 | {"language": ["en"], "license": "cc-by-sa-4.0", "task_categories": ["conversational", "text2text-generation"], "tags": ["medical"]} | 2023-12-26T19:33:28+00:00 | [
"2310.17703"
]
| [
"en"
]
| TAGS
#task_categories-conversational #task_categories-text2text-generation #language-English #license-cc-by-sa-4.0 #medical #arxiv-2310.17703 #region-us
|
## OncQA: The Impact of Using an AI Chatbot to Respond to Patient Messages
### Importance
Documentation burden is a major factor contributing to clinician burnout, which is increasing across the country and threatens our capacity to provide patient care in the U.S. While AI chatbots show potential in reducing this burden by aiding in documentation and are being incorporated into electronic health record systems, their influence on clinical decision-making remains understudied for this purpose.
### Objective
Investigate the acceptability, safety, and potential human factors issues when utilizing an AI-powered chatbot to draft responses to patients' inquiries.
### Design
- A 2-stage cross-sectional study was designed around 100 synthetic cancer patient scenarios couples with patient messages.
- These questions emulate realistic oncology scenarios.
- Stage 1: Manual Reponse: Six oncologists were randomly allocated 26 questions for response.
- Stage 2: AI-Assisted Response: The same oncologists received 26 new questions, alongside GPT-4 generated responses for editing.
- Informed consent was obtained.
- Participants were blinded to the source of the drafts.
- Surveys were undertaken for every scenario/response.
### About this repo
The dataset shows here is the complete stage2 parsed data with all physician edits.
If you wish to see the full data for stage1 and others please visit URL
### Settings
This research was conducted at the Brigham and Women’s Hospital, Boston, MA in 2023.
Q1: 'How challenging was it to respond to this message?'
Q2: 'Do you believe this patient is experiencing a severe medical event?'
Q3: 'How would you rate the acceptability of the draft response?'
Q4: 'How likely is it that the unedited draft response could cause harm?'
Q5: 'If the unedited draft does cause harm, what would be the extent, or clinical impact on the patient?'
Q6: 'Do you believe the provided unedited draft response improved your documentation efficiency?'
Q7: 'Do you believe the provided draft response was written by an AI or by a human?'
### Participants
Six board-certified oncologists participated.
### Intervention
Employment of GPT-4, an AI chatbot, for drafting responses to patient inquiries.
### Main Outcomes & Measures
- Evaluate the impact and utility of an AI chatbot in assisting responses to patient messages.
- Impact was determined by comparing response length and readability, using the Flesch reading ease score, and content.
- Utility was ascertained through physician feedback on surveys regarding acceptability, potential harm, and efficiency of chatbot-crafted drafts.
!Workflow Diagram
### Results
- On average, manual responses were more concise than those by GPT-4 or AI-assisted (34 vs. 169 vs. 160 words, p<0.001).
- Manual responses were more readable than GPT-4 or AI-assisted messages (Flesch score 67 vs. 45 vs. 46, p<0.001).
- About 58% of GPT-4 drafts were immediately acceptable, with 82% posing a low risk of harm.
- Utilizing the GPT-4 draft enhanced documentation efficiency in 77% of replies.
- Surprisingly, 31% of GPT-4 responses were perceived to be human-written, despite being AI-generated.
- 7.7% of survey responses felt unedited GPT-4 drafts could lead to severe harm or death.
- Among 56 dual-annotated responses, annotation agreement was low for manual responses (Cohen's kappa 0.10), but improved for AI-assisted responses (Cohen's kappa 0.52).
- AI-assistance led to differences in clinical content in the responses (p=0.001).
- Manual replies were more likely to advise direct clinical actions, while GPT-4 drafts often provided educational and self-management suggestions.
- AI-aided replies closely mirrored GPT-4 drafts but introduced some direct clinical actions.
### Conclusions & Relevance
AI-generated chatbot responses, while lengthier and less accessible, were overall safe and improved efficiency. AI-assistance altered the nature of physician feedback and reduced variability. AI chatbots are a promising avenue to address physician burnout and could improve patient care, however interactions between humans and AI might affect clinical decisions in unexpected ways. Addressing these interactions is vital for the safe incorporation of such technologies.
Note: It's imperative to delve deeper into human-AI interactions and their potential impact on outcomes.
:
| [
"## OncQA: The Impact of Using an AI Chatbot to Respond to Patient Messages",
"### Importance\n\nDocumentation burden is a major factor contributing to clinician burnout, which is increasing across the country and threatens our capacity to provide patient care in the U.S. While AI chatbots show potential in reducing this burden by aiding in documentation and are being incorporated into electronic health record systems, their influence on clinical decision-making remains understudied for this purpose.",
"### Objective\n\nInvestigate the acceptability, safety, and potential human factors issues when utilizing an AI-powered chatbot to draft responses to patients' inquiries.",
"### Design\n\n- A 2-stage cross-sectional study was designed around 100 synthetic cancer patient scenarios couples with patient messages.\n- These questions emulate realistic oncology scenarios.\n- Stage 1: Manual Reponse: Six oncologists were randomly allocated 26 questions for response.\n- Stage 2: AI-Assisted Response: The same oncologists received 26 new questions, alongside GPT-4 generated responses for editing.\n- Informed consent was obtained.\n- Participants were blinded to the source of the drafts.\n- Surveys were undertaken for every scenario/response.",
"### About this repo\n\nThe dataset shows here is the complete stage2 parsed data with all physician edits.\nIf you wish to see the full data for stage1 and others please visit URL",
"### Settings\n\nThis research was conducted at the Brigham and Women’s Hospital, Boston, MA in 2023.\n\nQ1: 'How challenging was it to respond to this message?'\n\nQ2: 'Do you believe this patient is experiencing a severe medical event?'\n\nQ3: 'How would you rate the acceptability of the draft response?'\n\nQ4: 'How likely is it that the unedited draft response could cause harm?'\n\nQ5: 'If the unedited draft does cause harm, what would be the extent, or clinical impact on the patient?'\n\nQ6: 'Do you believe the provided unedited draft response improved your documentation efficiency?'\n\nQ7: 'Do you believe the provided draft response was written by an AI or by a human?'",
"### Participants\n\nSix board-certified oncologists participated.",
"### Intervention\n\nEmployment of GPT-4, an AI chatbot, for drafting responses to patient inquiries.",
"### Main Outcomes & Measures\n\n- Evaluate the impact and utility of an AI chatbot in assisting responses to patient messages.\n- Impact was determined by comparing response length and readability, using the Flesch reading ease score, and content.\n- Utility was ascertained through physician feedback on surveys regarding acceptability, potential harm, and efficiency of chatbot-crafted drafts.\n !Workflow Diagram",
"### Results\n\n- On average, manual responses were more concise than those by GPT-4 or AI-assisted (34 vs. 169 vs. 160 words, p<0.001).\n- Manual responses were more readable than GPT-4 or AI-assisted messages (Flesch score 67 vs. 45 vs. 46, p<0.001).\n- About 58% of GPT-4 drafts were immediately acceptable, with 82% posing a low risk of harm.\n- Utilizing the GPT-4 draft enhanced documentation efficiency in 77% of replies.\n- Surprisingly, 31% of GPT-4 responses were perceived to be human-written, despite being AI-generated.\n- 7.7% of survey responses felt unedited GPT-4 drafts could lead to severe harm or death.\n- Among 56 dual-annotated responses, annotation agreement was low for manual responses (Cohen's kappa 0.10), but improved for AI-assisted responses (Cohen's kappa 0.52).\n- AI-assistance led to differences in clinical content in the responses (p=0.001).\n- Manual replies were more likely to advise direct clinical actions, while GPT-4 drafts often provided educational and self-management suggestions.\n- AI-aided replies closely mirrored GPT-4 drafts but introduced some direct clinical actions.",
"### Conclusions & Relevance\n\nAI-generated chatbot responses, while lengthier and less accessible, were overall safe and improved efficiency. AI-assistance altered the nature of physician feedback and reduced variability. AI chatbots are a promising avenue to address physician burnout and could improve patient care, however interactions between humans and AI might affect clinical decisions in unexpected ways. Addressing these interactions is vital for the safe incorporation of such technologies.\n\nNote: It's imperative to delve deeper into human-AI interactions and their potential impact on outcomes.\n\n\n:"
]
| [
"TAGS\n#task_categories-conversational #task_categories-text2text-generation #language-English #license-cc-by-sa-4.0 #medical #arxiv-2310.17703 #region-us \n",
"## OncQA: The Impact of Using an AI Chatbot to Respond to Patient Messages",
"### Importance\n\nDocumentation burden is a major factor contributing to clinician burnout, which is increasing across the country and threatens our capacity to provide patient care in the U.S. While AI chatbots show potential in reducing this burden by aiding in documentation and are being incorporated into electronic health record systems, their influence on clinical decision-making remains understudied for this purpose.",
"### Objective\n\nInvestigate the acceptability, safety, and potential human factors issues when utilizing an AI-powered chatbot to draft responses to patients' inquiries.",
"### Design\n\n- A 2-stage cross-sectional study was designed around 100 synthetic cancer patient scenarios couples with patient messages.\n- These questions emulate realistic oncology scenarios.\n- Stage 1: Manual Reponse: Six oncologists were randomly allocated 26 questions for response.\n- Stage 2: AI-Assisted Response: The same oncologists received 26 new questions, alongside GPT-4 generated responses for editing.\n- Informed consent was obtained.\n- Participants were blinded to the source of the drafts.\n- Surveys were undertaken for every scenario/response.",
"### About this repo\n\nThe dataset shows here is the complete stage2 parsed data with all physician edits.\nIf you wish to see the full data for stage1 and others please visit URL",
"### Settings\n\nThis research was conducted at the Brigham and Women’s Hospital, Boston, MA in 2023.\n\nQ1: 'How challenging was it to respond to this message?'\n\nQ2: 'Do you believe this patient is experiencing a severe medical event?'\n\nQ3: 'How would you rate the acceptability of the draft response?'\n\nQ4: 'How likely is it that the unedited draft response could cause harm?'\n\nQ5: 'If the unedited draft does cause harm, what would be the extent, or clinical impact on the patient?'\n\nQ6: 'Do you believe the provided unedited draft response improved your documentation efficiency?'\n\nQ7: 'Do you believe the provided draft response was written by an AI or by a human?'",
"### Participants\n\nSix board-certified oncologists participated.",
"### Intervention\n\nEmployment of GPT-4, an AI chatbot, for drafting responses to patient inquiries.",
"### Main Outcomes & Measures\n\n- Evaluate the impact and utility of an AI chatbot in assisting responses to patient messages.\n- Impact was determined by comparing response length and readability, using the Flesch reading ease score, and content.\n- Utility was ascertained through physician feedback on surveys regarding acceptability, potential harm, and efficiency of chatbot-crafted drafts.\n !Workflow Diagram",
"### Results\n\n- On average, manual responses were more concise than those by GPT-4 or AI-assisted (34 vs. 169 vs. 160 words, p<0.001).\n- Manual responses were more readable than GPT-4 or AI-assisted messages (Flesch score 67 vs. 45 vs. 46, p<0.001).\n- About 58% of GPT-4 drafts were immediately acceptable, with 82% posing a low risk of harm.\n- Utilizing the GPT-4 draft enhanced documentation efficiency in 77% of replies.\n- Surprisingly, 31% of GPT-4 responses were perceived to be human-written, despite being AI-generated.\n- 7.7% of survey responses felt unedited GPT-4 drafts could lead to severe harm or death.\n- Among 56 dual-annotated responses, annotation agreement was low for manual responses (Cohen's kappa 0.10), but improved for AI-assisted responses (Cohen's kappa 0.52).\n- AI-assistance led to differences in clinical content in the responses (p=0.001).\n- Manual replies were more likely to advise direct clinical actions, while GPT-4 drafts often provided educational and self-management suggestions.\n- AI-aided replies closely mirrored GPT-4 drafts but introduced some direct clinical actions.",
"### Conclusions & Relevance\n\nAI-generated chatbot responses, while lengthier and less accessible, were overall safe and improved efficiency. AI-assistance altered the nature of physician feedback and reduced variability. AI chatbots are a promising avenue to address physician burnout and could improve patient care, however interactions between humans and AI might affect clinical decisions in unexpected ways. Addressing these interactions is vital for the safe incorporation of such technologies.\n\nNote: It's imperative to delve deeper into human-AI interactions and their potential impact on outcomes.\n\n\n:"
]
| [
56,
20,
87,
38,
138,
40,
169,
16,
29,
95,
309,
129
]
| [
"passage: TAGS\n#task_categories-conversational #task_categories-text2text-generation #language-English #license-cc-by-sa-4.0 #medical #arxiv-2310.17703 #region-us \n## OncQA: The Impact of Using an AI Chatbot to Respond to Patient Messages### Importance\n\nDocumentation burden is a major factor contributing to clinician burnout, which is increasing across the country and threatens our capacity to provide patient care in the U.S. While AI chatbots show potential in reducing this burden by aiding in documentation and are being incorporated into electronic health record systems, their influence on clinical decision-making remains understudied for this purpose.### Objective\n\nInvestigate the acceptability, safety, and potential human factors issues when utilizing an AI-powered chatbot to draft responses to patients' inquiries.### Design\n\n- A 2-stage cross-sectional study was designed around 100 synthetic cancer patient scenarios couples with patient messages.\n- These questions emulate realistic oncology scenarios.\n- Stage 1: Manual Reponse: Six oncologists were randomly allocated 26 questions for response.\n- Stage 2: AI-Assisted Response: The same oncologists received 26 new questions, alongside GPT-4 generated responses for editing.\n- Informed consent was obtained.\n- Participants were blinded to the source of the drafts.\n- Surveys were undertaken for every scenario/response.### About this repo\n\nThe dataset shows here is the complete stage2 parsed data with all physician edits.\nIf you wish to see the full data for stage1 and others please visit URL",
"passage: ### Settings\n\nThis research was conducted at the Brigham and Women’s Hospital, Boston, MA in 2023.\n\nQ1: 'How challenging was it to respond to this message?'\n\nQ2: 'Do you believe this patient is experiencing a severe medical event?'\n\nQ3: 'How would you rate the acceptability of the draft response?'\n\nQ4: 'How likely is it that the unedited draft response could cause harm?'\n\nQ5: 'If the unedited draft does cause harm, what would be the extent, or clinical impact on the patient?'\n\nQ6: 'Do you believe the provided unedited draft response improved your documentation efficiency?'\n\nQ7: 'Do you believe the provided draft response was written by an AI or by a human?'### Participants\n\nSix board-certified oncologists participated.### Intervention\n\nEmployment of GPT-4, an AI chatbot, for drafting responses to patient inquiries.### Main Outcomes & Measures\n\n- Evaluate the impact and utility of an AI chatbot in assisting responses to patient messages.\n- Impact was determined by comparing response length and readability, using the Flesch reading ease score, and content.\n- Utility was ascertained through physician feedback on surveys regarding acceptability, potential harm, and efficiency of chatbot-crafted drafts.\n !Workflow Diagram### Results\n\n- On average, manual responses were more concise than those by GPT-4 or AI-assisted (34 vs. 169 vs. 160 words, p<0.001).\n- Manual responses were more readable than GPT-4 or AI-assisted messages (Flesch score 67 vs. 45 vs. 46, p<0.001).\n- About 58% of GPT-4 drafts were immediately acceptable, with 82% posing a low risk of harm.\n- Utilizing the GPT-4 draft enhanced documentation efficiency in 77% of replies.\n- Surprisingly, 31% of GPT-4 responses were perceived to be human-written, despite being AI-generated.\n- 7.7% of survey responses felt unedited GPT-4 drafts could lead to severe harm or death.\n- Among 56 dual-annotated responses, annotation agreement was low for manual responses (Cohen's kappa 0.10), but improved for AI-assisted responses (Cohen's kappa 0.52).\n- AI-assistance led to differences in clinical content in the responses (p=0.001).\n- Manual replies were more likely to advise direct clinical actions, while GPT-4 drafts often provided educational and self-management suggestions.\n- AI-aided replies closely mirrored GPT-4 drafts but introduced some direct clinical actions."
]
|
0dc179c2caa0d474664b8a98470a15a4c3c9a37d | # Dataset Card for "combined_train_dataset"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | ajdesh2000/combined_train_dataset | [
"region:us"
]
| 2023-10-26T21:12:50+00:00 | {"dataset_info": {"features": [{"name": "mmlu_id", "dtype": "string"}, {"name": "group_id", "dtype": "string"}, {"name": "category", "dtype": "string"}, {"name": "perturb_type", "dtype": "string"}, {"name": "split_used", "dtype": "string"}, {"name": "instruction", "dtype": "string"}, {"name": "output", "dtype": "string"}, {"name": "combined_id", "dtype": "string"}, {"name": "bbq_id", "dtype": "string"}, {"name": "is_ambiguous", "dtype": "string"}, {"name": "is_negative", "dtype": "string"}, {"name": "bb_id", "dtype": "string"}, {"name": "section", "dtype": "string"}, {"name": "task", "dtype": "string"}, {"name": "subtask", "dtype": "string"}, {"name": "org_task", "dtype": "string"}, {"name": "bb_stem_id", "dtype": "string"}, {"name": "math_id", "dtype": "string"}, {"name": "tqa_id", "dtype": "string"}, {"name": "gsm_id", "dtype": "string"}, {"name": "verbose", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 3617631, "num_examples": 6548}], "download_size": 1503383, "dataset_size": 3617631}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-27T03:51:13+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "combined_train_dataset"
More Information needed | [
"# Dataset Card for \"combined_train_dataset\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"combined_train_dataset\"\n\nMore Information needed"
]
| [
6,
19
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"combined_train_dataset\"\n\nMore Information needed"
]
|
75d957f4afdd340ea9ffc1e0ffcdfe0b82cf2f9e | # Dataset Card for "new_sentiment"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | marcus2000/new_sentiment | [
"region:us"
]
| 2023-10-26T21:21:23+00:00 | {"dataset_info": {"features": [{"name": "text", "dtype": "string"}, {"name": "sentiment", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 9012929, "num_examples": 6195}], "download_size": 4355943, "dataset_size": 9012929}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-27T10:45:52+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "new_sentiment"
More Information needed | [
"# Dataset Card for \"new_sentiment\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"new_sentiment\"\n\nMore Information needed"
]
| [
6,
14
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"new_sentiment\"\n\nMore Information needed"
]
|
7ee82297d06bf894c5ae1044b535c8a3e38041aa | # Dataset Card for "96ca277a"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | result-kand2-sdxl-wuerst-karlo/96ca277a | [
"region:us"
]
| 2023-10-26T21:44:39+00:00 | {"dataset_info": {"features": [{"name": "result", "dtype": "string"}, {"name": "id", "dtype": "int64"}], "splits": [{"name": "train", "num_bytes": 173, "num_examples": 10}], "download_size": 1332, "dataset_size": 173}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-26T21:44:39+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "96ca277a"
More Information needed | [
"# Dataset Card for \"96ca277a\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"96ca277a\"\n\nMore Information needed"
]
| [
6,
15
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"96ca277a\"\n\nMore Information needed"
]
|
8923386c9b0b2ad47abbd7925c3153a31c589e47 | # Dataset Card for "tencentdata_encodec"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | zion84006/tencentdata_encodec | [
"region:us"
]
| 2023-10-26T21:49:15+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}, {"split": "valid", "path": "data/valid-*"}, {"split": "test", "path": "data/test-*"}]}], "dataset_info": {"features": [{"name": "file_id", "dtype": "int64"}, {"name": "wav_id", "dtype": "int64"}, {"name": "instruction", "dtype": "string"}, {"name": "transcription", "dtype": "string"}, {"name": "src_encodec_0", "sequence": "int64"}, {"name": "src_encodec_1", "sequence": "int64"}, {"name": "src_encodec_2", "sequence": "int64"}, {"name": "src_encodec_3", "sequence": "int64"}, {"name": "src_encodec_4", "sequence": "int64"}, {"name": "src_encodec_5", "sequence": "int64"}, {"name": "src_encodec_6", "sequence": "int64"}, {"name": "src_encodec_7", "sequence": "int64"}, {"name": "tgt_encodec_0", "sequence": "int64"}, {"name": "tgt_encodec_1", "sequence": "int64"}, {"name": "tgt_encodec_2", "sequence": "int64"}, {"name": "tgt_encodec_3", "sequence": "int64"}, {"name": "tgt_encodec_4", "sequence": "int64"}, {"name": "tgt_encodec_5", "sequence": "int64"}, {"name": "tgt_encodec_6", "sequence": "int64"}, {"name": "tgt_encodec_7", "sequence": "int64"}], "splits": [{"name": "train", "num_bytes": 18584711340, "num_examples": 266780}, {"name": "valid", "num_bytes": 527848804, "num_examples": 7620}, {"name": "test", "num_bytes": 508405068, "num_examples": 7620}], "download_size": 471012646, "dataset_size": 19620965212}} | 2023-11-10T09:44:35+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "tencentdata_encodec"
More Information needed | [
"# Dataset Card for \"tencentdata_encodec\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"tencentdata_encodec\"\n\nMore Information needed"
]
| [
6,
16
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"tencentdata_encodec\"\n\nMore Information needed"
]
|
8989b08fac00ec85836f7ab99a0c3563352709a5 | # Dataset Card for "ddb_baseprompts"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | Sree1994/ddb_baseprompts | [
"region:us"
]
| 2023-10-26T21:52:33+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}, {"split": "test", "path": "data/test-*"}, {"split": "valid", "path": "data/valid-*"}]}], "dataset_info": {"features": [{"name": "Base_prompt", "dtype": "string"}, {"name": "Prompt", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 14886028, "num_examples": 51602}, {"name": "test", "num_bytes": 2096918, "num_examples": 7299}, {"name": "valid", "num_bytes": 4301342, "num_examples": 14817}], "download_size": 10829614, "dataset_size": 21284288}} | 2023-10-26T21:52:37+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "ddb_baseprompts"
More Information needed | [
"# Dataset Card for \"ddb_baseprompts\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"ddb_baseprompts\"\n\nMore Information needed"
]
| [
6,
17
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"ddb_baseprompts\"\n\nMore Information needed"
]
|
62c58b430ded8e1e3d64b6c4de0e3d0854d2a5b4 | # Dataset Card for "SDv2-CLIP-aligned-15000"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | Doub7e/SDv2-CLIP-aligned-15000 | [
"region:us"
]
| 2023-10-26T22:24:55+00:00 | {"dataset_info": {"features": [{"name": "image", "dtype": "image"}, {"name": "prompt", "dtype": "string"}, {"name": "type", "dtype": "string"}, {"name": "aligned", "dtype": "bool"}], "splits": [{"name": "train", "num_bytes": 14641004231.25, "num_examples": 15238}], "download_size": 14641382612, "dataset_size": 14641004231.25}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-26T22:37:50+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "SDv2-CLIP-aligned-15000"
More Information needed | [
"# Dataset Card for \"SDv2-CLIP-aligned-15000\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"SDv2-CLIP-aligned-15000\"\n\nMore Information needed"
]
| [
6,
20
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"SDv2-CLIP-aligned-15000\"\n\nMore Information needed"
]
|
ff69d3f972a78a38743dd6401bdd84b1721b513f | # Dataset Card for "content-papers-withprompt"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | zelalt/content-papers-withprompt | [
"region:us"
]
| 2023-10-26T23:27:53+00:00 | {"dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "authors", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "abstract", "dtype": "string"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 1283997, "num_examples": 992}], "download_size": 797519, "dataset_size": 1283997}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-26T23:27:54+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "content-papers-withprompt"
More Information needed | [
"# Dataset Card for \"content-papers-withprompt\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"content-papers-withprompt\"\n\nMore Information needed"
]
| [
6,
19
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"content-papers-withprompt\"\n\nMore Information needed"
]
|
74ec21f891366875d4433f30b741d0bbfee38953 | # AutoTrain Dataset for project: coffee-beans
## Dataset Description
This dataset has been automatically processed by AutoTrain for project coffee-beans.
### Languages
The BCP-47 code for the dataset's language is unk.
## Dataset Structure
### Data Instances
A sample from this dataset looks as follows:
```json
[
{
"image": "<224x224 RGB PIL image>",
"feat_width": 224,
"feat_height": 224,
"target": 1,
"feat_xmin": 22,
"feat_ymin": 61,
"feat_xmax": 140,
"feat_ymax": 160
},
{
"image": "<224x224 RGB PIL image>",
"feat_width": 224,
"feat_height": 224,
"target": 1,
"feat_xmin": 34,
"feat_ymin": 13,
"feat_xmax": 205,
"feat_ymax": 164
}
]
```
### Dataset Fields
The dataset has the following fields (also called "features"):
```json
{
"image": "Image(decode=True, id=None)",
"feat_width": "Value(dtype='int64', id=None)",
"feat_height": "Value(dtype='int64', id=None)",
"target": "ClassLabel(names=['defect', 'good'], id=None)",
"feat_xmin": "Value(dtype='int64', id=None)",
"feat_ymin": "Value(dtype='int64', id=None)",
"feat_xmax": "Value(dtype='int64', id=None)",
"feat_ymax": "Value(dtype='int64', id=None)"
}
```
### Dataset Splits
This dataset is split into a train and validation split. The split sizes are as follow:
| Split name | Num samples |
| ------------ | ------------------- |
| train | 3348 |
| valid | 1237 |
| everycoffee/autotrain-data-coffee-beans | [
"task_categories:image-classification",
"region:us"
]
| 2023-10-26T23:36:03+00:00 | {"task_categories": ["image-classification"]} | 2023-10-26T23:51:34+00:00 | []
| []
| TAGS
#task_categories-image-classification #region-us
| AutoTrain Dataset for project: coffee-beans
===========================================
Dataset Description
-------------------
This dataset has been automatically processed by AutoTrain for project coffee-beans.
### Languages
The BCP-47 code for the dataset's language is unk.
Dataset Structure
-----------------
### Data Instances
A sample from this dataset looks as follows:
### Dataset Fields
The dataset has the following fields (also called "features"):
### Dataset Splits
This dataset is split into a train and validation split. The split sizes are as follow:
| [
"### Languages\n\n\nThe BCP-47 code for the dataset's language is unk.\n\n\nDataset Structure\n-----------------",
"### Data Instances\n\n\nA sample from this dataset looks as follows:",
"### Dataset Fields\n\n\nThe dataset has the following fields (also called \"features\"):",
"### Dataset Splits\n\n\nThis dataset is split into a train and validation split. The split sizes are as follow:"
]
| [
"TAGS\n#task_categories-image-classification #region-us \n",
"### Languages\n\n\nThe BCP-47 code for the dataset's language is unk.\n\n\nDataset Structure\n-----------------",
"### Data Instances\n\n\nA sample from this dataset looks as follows:",
"### Dataset Fields\n\n\nThe dataset has the following fields (also called \"features\"):",
"### Dataset Splits\n\n\nThis dataset is split into a train and validation split. The split sizes are as follow:"
]
| [
17,
27,
17,
23,
27
]
| [
"passage: TAGS\n#task_categories-image-classification #region-us \n### Languages\n\n\nThe BCP-47 code for the dataset's language is unk.\n\n\nDataset Structure\n-----------------### Data Instances\n\n\nA sample from this dataset looks as follows:### Dataset Fields\n\n\nThe dataset has the following fields (also called \"features\"):### Dataset Splits\n\n\nThis dataset is split into a train and validation split. The split sizes are as follow:"
]
|
fe0041b71ec1820c98476099f932a94012bc3b8c | # Dataset Card for "soict_train_dataset_filter"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | thanhduycao/soict_train_dataset_filter | [
"region:us"
]
| 2023-10-27T00:01:46+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}, {"split": "test", "path": "data/test-*"}]}], "dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "sentence", "dtype": "string"}, {"name": "intent", "dtype": "string"}, {"name": "sentence_annotation", "dtype": "string"}, {"name": "entities", "list": [{"name": "type", "dtype": "string"}, {"name": "filler", "dtype": "string"}]}, {"name": "file", "dtype": "string"}, {"name": "audio", "struct": [{"name": "array", "sequence": "float64"}, {"name": "path", "dtype": "string"}, {"name": "sampling_rate", "dtype": "int64"}]}, {"name": "origin_transcription", "dtype": "string"}, {"name": "sentence_norm", "dtype": "string"}, {"name": "sentence_norm_v2", "dtype": "string"}, {"name": "w2v2_large_transcription", "dtype": "string"}, {"name": "wer", "dtype": "float64"}], "splits": [{"name": "train", "num_bytes": 3205296038.433596, "num_examples": 6184}, {"name": "test", "num_bytes": 566006350.9006286, "num_examples": 1092}], "download_size": 902006355, "dataset_size": 3771302389.3342247}} | 2023-10-27T00:02:51+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "soict_train_dataset_filter"
More Information needed | [
"# Dataset Card for \"soict_train_dataset_filter\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"soict_train_dataset_filter\"\n\nMore Information needed"
]
| [
6,
21
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"soict_train_dataset_filter\"\n\nMore Information needed"
]
|
c3340e05e6c866ae6e98dfd0473f14b8b3e97071 |
# Dataset Card for Evaluation run of ajibawa-2023/carl-7b
## Dataset Description
- **Homepage:**
- **Repository:** https://huggingface.co/ajibawa-2023/carl-7b
- **Paper:**
- **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
- **Point of Contact:** [email protected]
### Dataset Summary
Dataset automatically created during the evaluation run of model [ajibawa-2023/carl-7b](https://huggingface.co/ajibawa-2023/carl-7b) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 3 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the agregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_ajibawa-2023__carl-7b",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2023-10-27T01:32:56.729607](https://huggingface.co/datasets/open-llm-leaderboard/details_ajibawa-2023__carl-7b/blob/main/results_2023-10-27T01-32-56.729607.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"em": 0.0350251677852349,
"em_stderr": 0.001882728759888054,
"f1": 0.09129823825503375,
"f1_stderr": 0.002205658806843774,
"acc": 0.35468739621810547,
"acc_stderr": 0.008609150192858253
},
"harness|drop|3": {
"em": 0.0350251677852349,
"em_stderr": 0.001882728759888054,
"f1": 0.09129823825503375,
"f1_stderr": 0.002205658806843774
},
"harness|gsm8k|5": {
"acc": 0.02350265352539803,
"acc_stderr": 0.004172883669643945
},
"harness|winogrande|5": {
"acc": 0.6858721389108129,
"acc_stderr": 0.01304541671607256
}
}
```
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] | open-llm-leaderboard/details_ajibawa-2023__carl-7b | [
"region:us"
]
| 2023-10-27T00:33:01+00:00 | {"pretty_name": "Evaluation run of ajibawa-2023/carl-7b", "dataset_summary": "Dataset automatically created during the evaluation run of model [ajibawa-2023/carl-7b](https://huggingface.co/ajibawa-2023/carl-7b) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\nThe dataset is composed of 3 configuration, each one coresponding to one of the evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The \"train\" split is always pointing to the latest results.\n\nAn additional configuration \"results\" store all the aggregated results of the run (and is used to compute and display the agregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\nTo load the details from a run, you can for instance do the following:\n```python\nfrom datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_ajibawa-2023__carl-7b\",\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\nThese are the [latest results from run 2023-10-27T01:32:56.729607](https://huggingface.co/datasets/open-llm-leaderboard/details_ajibawa-2023__carl-7b/blob/main/results_2023-10-27T01-32-56.729607.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the \"latest\" split for each eval):\n\n```python\n{\n \"all\": {\n \"em\": 0.0350251677852349,\n \"em_stderr\": 0.001882728759888054,\n \"f1\": 0.09129823825503375,\n \"f1_stderr\": 0.002205658806843774,\n \"acc\": 0.35468739621810547,\n \"acc_stderr\": 0.008609150192858253\n },\n \"harness|drop|3\": {\n \"em\": 0.0350251677852349,\n \"em_stderr\": 0.001882728759888054,\n \"f1\": 0.09129823825503375,\n \"f1_stderr\": 0.002205658806843774\n },\n \"harness|gsm8k|5\": {\n \"acc\": 0.02350265352539803,\n \"acc_stderr\": 0.004172883669643945\n },\n \"harness|winogrande|5\": {\n \"acc\": 0.6858721389108129,\n \"acc_stderr\": 0.01304541671607256\n }\n}\n```", "repo_url": "https://huggingface.co/ajibawa-2023/carl-7b", "leaderboard_url": "https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard", "point_of_contact": "[email protected]", "configs": [{"config_name": "harness_drop_3", "data_files": [{"split": "2023_10_27T01_32_56.729607", "path": ["**/details_harness|drop|3_2023-10-27T01-32-56.729607.parquet"]}, {"split": "latest", "path": ["**/details_harness|drop|3_2023-10-27T01-32-56.729607.parquet"]}]}, {"config_name": "harness_gsm8k_5", "data_files": [{"split": "2023_10_27T01_32_56.729607", "path": ["**/details_harness|gsm8k|5_2023-10-27T01-32-56.729607.parquet"]}, {"split": "latest", "path": ["**/details_harness|gsm8k|5_2023-10-27T01-32-56.729607.parquet"]}]}, {"config_name": "harness_winogrande_5", "data_files": [{"split": "2023_10_27T01_32_56.729607", "path": ["**/details_harness|winogrande|5_2023-10-27T01-32-56.729607.parquet"]}, {"split": "latest", "path": ["**/details_harness|winogrande|5_2023-10-27T01-32-56.729607.parquet"]}]}, {"config_name": "results", "data_files": [{"split": "2023_10_27T01_32_56.729607", "path": ["results_2023-10-27T01-32-56.729607.parquet"]}, {"split": "latest", "path": ["results_2023-10-27T01-32-56.729607.parquet"]}]}]} | 2023-10-27T00:33:08+00:00 | []
| []
| TAGS
#region-us
|
# Dataset Card for Evaluation run of ajibawa-2023/carl-7b
## Dataset Description
- Homepage:
- Repository: URL
- Paper:
- Leaderboard: URL
- Point of Contact: clementine@URL
### Dataset Summary
Dataset automatically created during the evaluation run of model ajibawa-2023/carl-7b on the Open LLM Leaderboard.
The dataset is composed of 3 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the agregated metrics on the Open LLM Leaderboard).
To load the details from a run, you can for instance do the following:
## Latest results
These are the latest results from run 2023-10-27T01:32:56.729607(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
### Supported Tasks and Leaderboards
### Languages
## Dataset Structure
### Data Instances
### Data Fields
### Data Splits
## Dataset Creation
### Curation Rationale
### Source Data
#### Initial Data Collection and Normalization
#### Who are the source language producers?
### Annotations
#### Annotation process
#### Who are the annotators?
### Personal and Sensitive Information
## Considerations for Using the Data
### Social Impact of Dataset
### Discussion of Biases
### Other Known Limitations
## Additional Information
### Dataset Curators
### Licensing Information
### Contributions
| [
"# Dataset Card for Evaluation run of ajibawa-2023/carl-7b",
"## Dataset Description\n\n- Homepage: \n- Repository: URL\n- Paper: \n- Leaderboard: URL\n- Point of Contact: clementine@URL",
"### Dataset Summary\n\nDataset automatically created during the evaluation run of model ajibawa-2023/carl-7b on the Open LLM Leaderboard.\n\nThe dataset is composed of 3 configuration, each one coresponding to one of the evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The \"train\" split is always pointing to the latest results.\n\nAn additional configuration \"results\" store all the aggregated results of the run (and is used to compute and display the agregated metrics on the Open LLM Leaderboard).\n\nTo load the details from a run, you can for instance do the following:",
"## Latest results\n\nThese are the latest results from run 2023-10-27T01:32:56.729607(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the \"latest\" split for each eval):",
"### Supported Tasks and Leaderboards",
"### Languages",
"## Dataset Structure",
"### Data Instances",
"### Data Fields",
"### Data Splits",
"## Dataset Creation",
"### Curation Rationale",
"### Source Data",
"#### Initial Data Collection and Normalization",
"#### Who are the source language producers?",
"### Annotations",
"#### Annotation process",
"#### Who are the annotators?",
"### Personal and Sensitive Information",
"## Considerations for Using the Data",
"### Social Impact of Dataset",
"### Discussion of Biases",
"### Other Known Limitations",
"## Additional Information",
"### Dataset Curators",
"### Licensing Information",
"### Contributions"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for Evaluation run of ajibawa-2023/carl-7b",
"## Dataset Description\n\n- Homepage: \n- Repository: URL\n- Paper: \n- Leaderboard: URL\n- Point of Contact: clementine@URL",
"### Dataset Summary\n\nDataset automatically created during the evaluation run of model ajibawa-2023/carl-7b on the Open LLM Leaderboard.\n\nThe dataset is composed of 3 configuration, each one coresponding to one of the evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The \"train\" split is always pointing to the latest results.\n\nAn additional configuration \"results\" store all the aggregated results of the run (and is used to compute and display the agregated metrics on the Open LLM Leaderboard).\n\nTo load the details from a run, you can for instance do the following:",
"## Latest results\n\nThese are the latest results from run 2023-10-27T01:32:56.729607(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the \"latest\" split for each eval):",
"### Supported Tasks and Leaderboards",
"### Languages",
"## Dataset Structure",
"### Data Instances",
"### Data Fields",
"### Data Splits",
"## Dataset Creation",
"### Curation Rationale",
"### Source Data",
"#### Initial Data Collection and Normalization",
"#### Who are the source language producers?",
"### Annotations",
"#### Annotation process",
"#### Who are the annotators?",
"### Personal and Sensitive Information",
"## Considerations for Using the Data",
"### Social Impact of Dataset",
"### Discussion of Biases",
"### Other Known Limitations",
"## Additional Information",
"### Dataset Curators",
"### Licensing Information",
"### Contributions"
]
| [
6,
19,
31,
167,
67,
10,
4,
6,
6,
5,
5,
5,
7,
4,
10,
10,
5,
5,
9,
8,
8,
7,
8,
7,
5,
6,
6,
5
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for Evaluation run of ajibawa-2023/carl-7b## Dataset Description\n\n- Homepage: \n- Repository: URL\n- Paper: \n- Leaderboard: URL\n- Point of Contact: clementine@URL### Dataset Summary\n\nDataset automatically created during the evaluation run of model ajibawa-2023/carl-7b on the Open LLM Leaderboard.\n\nThe dataset is composed of 3 configuration, each one coresponding to one of the evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The \"train\" split is always pointing to the latest results.\n\nAn additional configuration \"results\" store all the aggregated results of the run (and is used to compute and display the agregated metrics on the Open LLM Leaderboard).\n\nTo load the details from a run, you can for instance do the following:## Latest results\n\nThese are the latest results from run 2023-10-27T01:32:56.729607(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the \"latest\" split for each eval):### Supported Tasks and Leaderboards### Languages## Dataset Structure### Data Instances### Data Fields### Data Splits## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information### Dataset Curators### Licensing Information### Contributions"
]
|
5db6065ab6b0ba03a1aad26d86aeb2561df6e2e9 | # Dataset Card for Wikipedia summary-only dataset
<!-- Provide a quick summary of the dataset. -->
This dataset contains only the summary of English wikipedia, generated from [jordiclive/wikipedia-summary-dataset](https://huggingface.co/datasets/jordiclive/wikipedia-summary-dataset).
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Language(s) (NLP):** English | yanbozhang/wikipedia-summary-only | [
"task_categories:text-generation",
"size_categories:1M<n<10M",
"language:en",
"region:us"
]
| 2023-10-27T00:45:47+00:00 | {"language": ["en"], "size_categories": ["1M<n<10M"], "task_categories": ["text-generation"]} | 2023-10-27T02:34:16+00:00 | []
| [
"en"
]
| TAGS
#task_categories-text-generation #size_categories-1M<n<10M #language-English #region-us
| # Dataset Card for Wikipedia summary-only dataset
This dataset contains only the summary of English wikipedia, generated from jordiclive/wikipedia-summary-dataset.
## Dataset Details
### Dataset Description
- Language(s) (NLP): English | [
"# Dataset Card for Wikipedia summary-only dataset\n\n\n\nThis dataset contains only the summary of English wikipedia, generated from jordiclive/wikipedia-summary-dataset.",
"## Dataset Details",
"### Dataset Description\n\n\n\n\n\n- Language(s) (NLP): English"
]
| [
"TAGS\n#task_categories-text-generation #size_categories-1M<n<10M #language-English #region-us \n",
"# Dataset Card for Wikipedia summary-only dataset\n\n\n\nThis dataset contains only the summary of English wikipedia, generated from jordiclive/wikipedia-summary-dataset.",
"## Dataset Details",
"### Dataset Description\n\n\n\n\n\n- Language(s) (NLP): English"
]
| [
33,
40,
4,
15
]
| [
"passage: TAGS\n#task_categories-text-generation #size_categories-1M<n<10M #language-English #region-us \n# Dataset Card for Wikipedia summary-only dataset\n\n\n\nThis dataset contains only the summary of English wikipedia, generated from jordiclive/wikipedia-summary-dataset.## Dataset Details### Dataset Description\n\n\n\n\n\n- Language(s) (NLP): English"
]
|
184b0fcae54e865852d36da5a55e2c78c0c953f3 | # Dataset Card for "UECFOODPIXCOMPLETE"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | justinsiow/UECFOODPIXCOMPLETE | [
"region:us"
]
| 2023-10-27T00:58:49+00:00 | {"dataset_info": {"features": [{"name": "image", "dtype": "image"}, {"name": "label", "dtype": "image"}], "splits": [{"name": "train", "num_bytes": 572997006.0, "num_examples": 9000}, {"name": "validation", "num_bytes": 77988933.0, "num_examples": 1000}], "download_size": 721976398, "dataset_size": 650985939.0}} | 2023-10-27T01:15:36+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "UECFOODPIXCOMPLETE"
More Information needed | [
"# Dataset Card for \"UECFOODPIXCOMPLETE\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"UECFOODPIXCOMPLETE\"\n\nMore Information needed"
]
| [
6,
18
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"UECFOODPIXCOMPLETE\"\n\nMore Information needed"
]
|
b7e6ba0185d322a725083f9e59d4e8747f908ac4 | # Dataset Card for "prompt_learning_paper"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | hudssntao/prompt_learning_paper | [
"region:us"
]
| 2023-10-27T01:07:27+00:00 | {"dataset_info": {"features": [{"name": "newColumn", "dtype": "string"}, {"name": "new_colmmn", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 80, "num_examples": 6}], "download_size": 0, "dataset_size": 80}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-27T01:12:10+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "prompt_learning_paper"
More Information needed | [
"# Dataset Card for \"prompt_learning_paper\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"prompt_learning_paper\"\n\nMore Information needed"
]
| [
6,
17
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"prompt_learning_paper\"\n\nMore Information needed"
]
|
49b0d9e0ea15bc37baec4f8588199869d5cecb11 | # Dataset Card for "prueba-arg"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | finiteautomata/prueba-arg | [
"region:us"
]
| 2023-10-27T01:35:34+00:00 | {"dataset_info": {"features": [{"name": "tweet_id", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "user", "dtype": "string"}, {"name": "body", "dtype": "string"}, {"name": "created_at", "dtype": "string"}, {"name": "comments", "list": [{"name": "created_at", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "user_id", "dtype": "string"}]}], "splits": [{"name": "train", "num_bytes": 909617906, "num_examples": 73423}], "download_size": 0, "dataset_size": 909617906}} | 2023-10-27T03:34:43+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "prueba-arg"
More Information needed | [
"# Dataset Card for \"prueba-arg\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"prueba-arg\"\n\nMore Information needed"
]
| [
6,
15
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"prueba-arg\"\n\nMore Information needed"
]
|
03c56862951c04ca0c0d64db0f54d9d6d7d0ebf6 | # Dataset Card for "test_dataset"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | hudssntao/test_dataset | [
"region:us"
]
| 2023-10-27T01:43:19+00:00 | {"dataset_info": {"features": [{"name": "column1", "dtype": "string"}, {"name": "column2", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 40, "num_examples": 2}], "download_size": 1227, "dataset_size": 40}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-27T02:27:15+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "test_dataset"
More Information needed | [
"# Dataset Card for \"test_dataset\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"test_dataset\"\n\nMore Information needed"
]
| [
6,
14
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"test_dataset\"\n\nMore Information needed"
]
|
44e2fedccc8bed7aee39fd9144f94b0df673d360 | # Disclaimer:
this dataset is curated for NeurIPS 2023 LLM efficiency challange, and currently work in progress. Please use at your own risk.
# Data composition:
All data were derived from the training set portion of the open source dataset.
**gsm2k_dolly15k_cnnadd6k_mmlulog1.7w_bbqabc8k.json**:
-gsm8k_2000: https://huggingface.co/datasets/gsm8k
-dolly_15000: https://huggingface.co/datasets/databricks/databricks-dolly-15k
-cnn_dailymail_6000: https://huggingface.co/datasets/cnn_dailymail
-mmlu_17000: https://huggingface.co/datasets/cais/mmlu
-bbq_8000: https://huggingface.co/datasets/tasksource/bigbench
**lima_4kall.json**
-lima_1000: https://huggingface.co/datasets/GAIR/lima
-3000 of gsm8k_dolly15k_cnnadd8k_mmlulog1.7w_bbqabc8k.json: https://huggingface.co/datasets/zhongshupeng/dataset_4090_1 | zhongshupeng/dataset_4090_2 | [
"region:us"
]
| 2023-10-27T01:47:49+00:00 | {} | 2023-10-27T02:01:33+00:00 | []
| []
| TAGS
#region-us
| # Disclaimer:
this dataset is curated for NeurIPS 2023 LLM efficiency challange, and currently work in progress. Please use at your own risk.
# Data composition:
All data were derived from the training set portion of the open source dataset.
gsm2k_dolly15k_cnnadd6k_mmlulog1.7w_bbqabc8k.json:
-gsm8k_2000: URL
-dolly_15000: URL
-cnn_dailymail_6000: URL
-mmlu_17000: URL
-bbq_8000: URL
lima_4kall.json
-lima_1000: URL
-3000 of gsm8k_dolly15k_cnnadd8k_mmlulog1.7w_bbqabc8k.json: URL | [
"# Disclaimer: \nthis dataset is curated for NeurIPS 2023 LLM efficiency challange, and currently work in progress. Please use at your own risk.",
"# Data composition: \nAll data were derived from the training set portion of the open source dataset.\n\ngsm2k_dolly15k_cnnadd6k_mmlulog1.7w_bbqabc8k.json:\n\n-gsm8k_2000: URL\n\n-dolly_15000: URL\n\n-cnn_dailymail_6000: URL\n\n-mmlu_17000: URL\n\n-bbq_8000: URL\n\n\nlima_4kall.json\n\n-lima_1000: URL\n\n-3000 of gsm8k_dolly15k_cnnadd8k_mmlulog1.7w_bbqabc8k.json: URL"
]
| [
"TAGS\n#region-us \n",
"# Disclaimer: \nthis dataset is curated for NeurIPS 2023 LLM efficiency challange, and currently work in progress. Please use at your own risk.",
"# Data composition: \nAll data were derived from the training set portion of the open source dataset.\n\ngsm2k_dolly15k_cnnadd6k_mmlulog1.7w_bbqabc8k.json:\n\n-gsm8k_2000: URL\n\n-dolly_15000: URL\n\n-cnn_dailymail_6000: URL\n\n-mmlu_17000: URL\n\n-bbq_8000: URL\n\n\nlima_4kall.json\n\n-lima_1000: URL\n\n-3000 of gsm8k_dolly15k_cnnadd8k_mmlulog1.7w_bbqabc8k.json: URL"
]
| [
6,
34,
143
]
| [
"passage: TAGS\n#region-us \n# Disclaimer: \nthis dataset is curated for NeurIPS 2023 LLM efficiency challange, and currently work in progress. Please use at your own risk.# Data composition: \nAll data were derived from the training set portion of the open source dataset.\n\ngsm2k_dolly15k_cnnadd6k_mmlulog1.7w_bbqabc8k.json:\n\n-gsm8k_2000: URL\n\n-dolly_15000: URL\n\n-cnn_dailymail_6000: URL\n\n-mmlu_17000: URL\n\n-bbq_8000: URL\n\n\nlima_4kall.json\n\n-lima_1000: URL\n\n-3000 of gsm8k_dolly15k_cnnadd8k_mmlulog1.7w_bbqabc8k.json: URL"
]
|
38a5323b10abbc4baa08e4c6455339a830c87f10 | # Disclaimer:
this dataset is curated for NeurIPS 2023 LLM efficiency challange, and currently work in progress. Please use at your own risk.
# Data composition:
All data were derived from the training set portion of the open source dataset.
**gsm2k_dolly12k_cnnadd4k_mmlulog1.7w_bbqabc8k.json**:
-gsm8k_2000: https://huggingface.co/datasets/gsm8k
-dolly_12000: https://huggingface.co/datasets/databricks/databricks-dolly-15k
-cnn_dailymail_4000: https://huggingface.co/datasets/cnn_dailymail
-mmlu_17000: https://huggingface.co/datasets/cais/mmlu
-bbq_8000: https://huggingface.co/datasets/tasksource/bigbench | zhongshupeng/dataset_4090_3 | [
"region:us"
]
| 2023-10-27T02:02:27+00:00 | {} | 2023-10-27T02:04:31+00:00 | []
| []
| TAGS
#region-us
| # Disclaimer:
this dataset is curated for NeurIPS 2023 LLM efficiency challange, and currently work in progress. Please use at your own risk.
# Data composition:
All data were derived from the training set portion of the open source dataset.
gsm2k_dolly12k_cnnadd4k_mmlulog1.7w_bbqabc8k.json:
-gsm8k_2000: URL
-dolly_12000: URL
-cnn_dailymail_4000: URL
-mmlu_17000: URL
-bbq_8000: URL | [
"# Disclaimer: \nthis dataset is curated for NeurIPS 2023 LLM efficiency challange, and currently work in progress. Please use at your own risk.",
"# Data composition: \nAll data were derived from the training set portion of the open source dataset.\n\ngsm2k_dolly12k_cnnadd4k_mmlulog1.7w_bbqabc8k.json:\n\n-gsm8k_2000: URL\n\n-dolly_12000: URL\n\n-cnn_dailymail_4000: URL\n\n-mmlu_17000: URL\n\n-bbq_8000: URL"
]
| [
"TAGS\n#region-us \n",
"# Disclaimer: \nthis dataset is curated for NeurIPS 2023 LLM efficiency challange, and currently work in progress. Please use at your own risk.",
"# Data composition: \nAll data were derived from the training set portion of the open source dataset.\n\ngsm2k_dolly12k_cnnadd4k_mmlulog1.7w_bbqabc8k.json:\n\n-gsm8k_2000: URL\n\n-dolly_12000: URL\n\n-cnn_dailymail_4000: URL\n\n-mmlu_17000: URL\n\n-bbq_8000: URL"
]
| [
6,
34,
93
]
| [
"passage: TAGS\n#region-us \n# Disclaimer: \nthis dataset is curated for NeurIPS 2023 LLM efficiency challange, and currently work in progress. Please use at your own risk.# Data composition: \nAll data were derived from the training set portion of the open source dataset.\n\ngsm2k_dolly12k_cnnadd4k_mmlulog1.7w_bbqabc8k.json:\n\n-gsm8k_2000: URL\n\n-dolly_12000: URL\n\n-cnn_dailymail_4000: URL\n\n-mmlu_17000: URL\n\n-bbq_8000: URL"
]
|
49efe6772d3b077422307c8e54a0e93909d66ee9 | # Dataset Card for "7f9071c2"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | result-kand2-sdxl-wuerst-karlo/7f9071c2 | [
"region:us"
]
| 2023-10-27T02:05:39+00:00 | {"dataset_info": {"features": [{"name": "result", "dtype": "string"}, {"name": "id", "dtype": "int64"}], "splits": [{"name": "train", "num_bytes": 171, "num_examples": 10}], "download_size": 1324, "dataset_size": 171}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-27T02:05:40+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "7f9071c2"
More Information needed | [
"# Dataset Card for \"7f9071c2\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"7f9071c2\"\n\nMore Information needed"
]
| [
6,
16
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"7f9071c2\"\n\nMore Information needed"
]
|
8bc6ffe4f6c712bdf8313d9ccfd3626e8b06874d | # Dataset Card for Dataset Name: Indian Temple Destruction Dataset
## Dataset Details
### Dataset Description
The Indian Temple Destruction Dataset provides information about historical temples that were destroyed in the past in India, including details on the locations of these temples, the entities responsible for their destruction, and contact information for inquiries.
- **Curated by:** Gaurav Sinha
- **Funded by [optional]:** [Information Not Available]
- **Shared by [optional]:** [Information Not Available]
- **Language(s) (NLP):** English
- **License:** [Information Not Available]
### Dataset Sources [optional]
- **Repository:** [Link to the dataset repository]
- **Paper [optional]:** [Link to any associated research paper]
- **Demo [optional]:** [Link to a demo or usage example]
## Uses
### Direct Use
This dataset can be used for historical research, cultural preservation efforts, and to understand the history of temple destruction in India.
### Out-of-Scope Use
Misuse of this dataset for promoting hatred, violence, or discrimination is strictly out of scope.
## Dataset Structure
[Information Not Available]
## Dataset Creation
### Curation Rationale
The dataset was created to document the historical information about the destruction of temples in India for research, education, and preservation purposes. It includes data from books authored by Sir Sita Ram Goel and contributions by Gaurav Sinha.
### Source Data
#### Data Collection and Processing
The data for this dataset was collected from historical records, scholarly research, and reputable sources, including books authored by Sir Sita Ram Goel. It also includes contributions by Gaurav Sinha. The dataset was carefully compiled to provide accurate and valuable information.
#### Who are the source data producers?
The source data was produced by historians, researchers, and scholars, including Sir Sita Ram Goel, who documented the destruction of temples in India. Contributions by Gaurav Sinha are also part of the dataset.
### Annotations [optional]
[Information Not Available]
## Bias, Risks, and Limitations
This dataset may contain historical events that could be sensitive to some communities. It is essential to use this data responsibly and with cultural sensitivity.
### Recommendations
Users should exercise caution when using this dataset to ensure that it is used for educational and research purposes and not for promoting hatred or discrimination.
## Citation [optional]
**BibTeX:**
[Information Not Available]
**APA:**
[Information Not Available]
## Glossary [optional]
[Information Not Available]
## More Information [optional]
[Information Not Available]
## Dataset Card Authors [optional]
Gaurav Sinha
## Dataset Card Contact
For inquiries related to this dataset, please contact [Your Email Address]. | gaurav16/temples_dataset | [
"task_categories:question-answering",
"size_categories:1M<n<10M",
"language:en",
"license:apache-2.0",
"art",
"region:us"
]
| 2023-10-27T02:15:21+00:00 | {"language": ["en"], "license": "apache-2.0", "size_categories": ["1M<n<10M"], "task_categories": ["question-answering"], "tags": ["art"]} | 2023-10-27T02:22:25+00:00 | []
| [
"en"
]
| TAGS
#task_categories-question-answering #size_categories-1M<n<10M #language-English #license-apache-2.0 #art #region-us
| # Dataset Card for Dataset Name: Indian Temple Destruction Dataset
## Dataset Details
### Dataset Description
The Indian Temple Destruction Dataset provides information about historical temples that were destroyed in the past in India, including details on the locations of these temples, the entities responsible for their destruction, and contact information for inquiries.
- Curated by: Gaurav Sinha
- Funded by [optional]: [Information Not Available]
- Shared by [optional]: [Information Not Available]
- Language(s) (NLP): English
- License: [Information Not Available]
### Dataset Sources [optional]
- Repository: [Link to the dataset repository]
- Paper [optional]: [Link to any associated research paper]
- Demo [optional]: [Link to a demo or usage example]
## Uses
### Direct Use
This dataset can be used for historical research, cultural preservation efforts, and to understand the history of temple destruction in India.
### Out-of-Scope Use
Misuse of this dataset for promoting hatred, violence, or discrimination is strictly out of scope.
## Dataset Structure
[Information Not Available]
## Dataset Creation
### Curation Rationale
The dataset was created to document the historical information about the destruction of temples in India for research, education, and preservation purposes. It includes data from books authored by Sir Sita Ram Goel and contributions by Gaurav Sinha.
### Source Data
#### Data Collection and Processing
The data for this dataset was collected from historical records, scholarly research, and reputable sources, including books authored by Sir Sita Ram Goel. It also includes contributions by Gaurav Sinha. The dataset was carefully compiled to provide accurate and valuable information.
#### Who are the source data producers?
The source data was produced by historians, researchers, and scholars, including Sir Sita Ram Goel, who documented the destruction of temples in India. Contributions by Gaurav Sinha are also part of the dataset.
### Annotations [optional]
[Information Not Available]
## Bias, Risks, and Limitations
This dataset may contain historical events that could be sensitive to some communities. It is essential to use this data responsibly and with cultural sensitivity.
### Recommendations
Users should exercise caution when using this dataset to ensure that it is used for educational and research purposes and not for promoting hatred or discrimination.
[optional]
BibTeX:
[Information Not Available]
APA:
[Information Not Available]
## Glossary [optional]
[Information Not Available]
## More Information [optional]
[Information Not Available]
## Dataset Card Authors [optional]
Gaurav Sinha
## Dataset Card Contact
For inquiries related to this dataset, please contact [Your Email Address]. | [
"# Dataset Card for Dataset Name: Indian Temple Destruction Dataset",
"## Dataset Details",
"### Dataset Description\n\nThe Indian Temple Destruction Dataset provides information about historical temples that were destroyed in the past in India, including details on the locations of these temples, the entities responsible for their destruction, and contact information for inquiries.\n\n- Curated by: Gaurav Sinha\n- Funded by [optional]: [Information Not Available]\n- Shared by [optional]: [Information Not Available]\n- Language(s) (NLP): English\n- License: [Information Not Available]",
"### Dataset Sources [optional]\n\n- Repository: [Link to the dataset repository]\n- Paper [optional]: [Link to any associated research paper]\n- Demo [optional]: [Link to a demo or usage example]",
"## Uses",
"### Direct Use\n\nThis dataset can be used for historical research, cultural preservation efforts, and to understand the history of temple destruction in India.",
"### Out-of-Scope Use\n\nMisuse of this dataset for promoting hatred, violence, or discrimination is strictly out of scope.",
"## Dataset Structure\n\n[Information Not Available]",
"## Dataset Creation",
"### Curation Rationale\n\nThe dataset was created to document the historical information about the destruction of temples in India for research, education, and preservation purposes. It includes data from books authored by Sir Sita Ram Goel and contributions by Gaurav Sinha.",
"### Source Data",
"#### Data Collection and Processing\n\nThe data for this dataset was collected from historical records, scholarly research, and reputable sources, including books authored by Sir Sita Ram Goel. It also includes contributions by Gaurav Sinha. The dataset was carefully compiled to provide accurate and valuable information.",
"#### Who are the source data producers?\n\nThe source data was produced by historians, researchers, and scholars, including Sir Sita Ram Goel, who documented the destruction of temples in India. Contributions by Gaurav Sinha are also part of the dataset.",
"### Annotations [optional]\n\n[Information Not Available]",
"## Bias, Risks, and Limitations\n\nThis dataset may contain historical events that could be sensitive to some communities. It is essential to use this data responsibly and with cultural sensitivity.",
"### Recommendations\n\nUsers should exercise caution when using this dataset to ensure that it is used for educational and research purposes and not for promoting hatred or discrimination.\n\n[optional]\n\nBibTeX:\n\n[Information Not Available]\n\nAPA:\n\n[Information Not Available]",
"## Glossary [optional]\n\n[Information Not Available]",
"## More Information [optional]\n\n[Information Not Available]",
"## Dataset Card Authors [optional]\n\nGaurav Sinha",
"## Dataset Card Contact\n\nFor inquiries related to this dataset, please contact [Your Email Address]."
]
| [
"TAGS\n#task_categories-question-answering #size_categories-1M<n<10M #language-English #license-apache-2.0 #art #region-us \n",
"# Dataset Card for Dataset Name: Indian Temple Destruction Dataset",
"## Dataset Details",
"### Dataset Description\n\nThe Indian Temple Destruction Dataset provides information about historical temples that were destroyed in the past in India, including details on the locations of these temples, the entities responsible for their destruction, and contact information for inquiries.\n\n- Curated by: Gaurav Sinha\n- Funded by [optional]: [Information Not Available]\n- Shared by [optional]: [Information Not Available]\n- Language(s) (NLP): English\n- License: [Information Not Available]",
"### Dataset Sources [optional]\n\n- Repository: [Link to the dataset repository]\n- Paper [optional]: [Link to any associated research paper]\n- Demo [optional]: [Link to a demo or usage example]",
"## Uses",
"### Direct Use\n\nThis dataset can be used for historical research, cultural preservation efforts, and to understand the history of temple destruction in India.",
"### Out-of-Scope Use\n\nMisuse of this dataset for promoting hatred, violence, or discrimination is strictly out of scope.",
"## Dataset Structure\n\n[Information Not Available]",
"## Dataset Creation",
"### Curation Rationale\n\nThe dataset was created to document the historical information about the destruction of temples in India for research, education, and preservation purposes. It includes data from books authored by Sir Sita Ram Goel and contributions by Gaurav Sinha.",
"### Source Data",
"#### Data Collection and Processing\n\nThe data for this dataset was collected from historical records, scholarly research, and reputable sources, including books authored by Sir Sita Ram Goel. It also includes contributions by Gaurav Sinha. The dataset was carefully compiled to provide accurate and valuable information.",
"#### Who are the source data producers?\n\nThe source data was produced by historians, researchers, and scholars, including Sir Sita Ram Goel, who documented the destruction of temples in India. Contributions by Gaurav Sinha are also part of the dataset.",
"### Annotations [optional]\n\n[Information Not Available]",
"## Bias, Risks, and Limitations\n\nThis dataset may contain historical events that could be sensitive to some communities. It is essential to use this data responsibly and with cultural sensitivity.",
"### Recommendations\n\nUsers should exercise caution when using this dataset to ensure that it is used for educational and research purposes and not for promoting hatred or discrimination.\n\n[optional]\n\nBibTeX:\n\n[Information Not Available]\n\nAPA:\n\n[Information Not Available]",
"## Glossary [optional]\n\n[Information Not Available]",
"## More Information [optional]\n\n[Information Not Available]",
"## Dataset Card Authors [optional]\n\nGaurav Sinha",
"## Dataset Card Contact\n\nFor inquiries related to this dataset, please contact [Your Email Address]."
]
| [
44,
16,
4,
111,
56,
3,
30,
34,
11,
5,
59,
4,
67,
61,
14,
42,
61,
13,
12,
14,
24
]
| [
"passage: TAGS\n#task_categories-question-answering #size_categories-1M<n<10M #language-English #license-apache-2.0 #art #region-us \n# Dataset Card for Dataset Name: Indian Temple Destruction Dataset## Dataset Details### Dataset Description\n\nThe Indian Temple Destruction Dataset provides information about historical temples that were destroyed in the past in India, including details on the locations of these temples, the entities responsible for their destruction, and contact information for inquiries.\n\n- Curated by: Gaurav Sinha\n- Funded by [optional]: [Information Not Available]\n- Shared by [optional]: [Information Not Available]\n- Language(s) (NLP): English\n- License: [Information Not Available]### Dataset Sources [optional]\n\n- Repository: [Link to the dataset repository]\n- Paper [optional]: [Link to any associated research paper]\n- Demo [optional]: [Link to a demo or usage example]## Uses### Direct Use\n\nThis dataset can be used for historical research, cultural preservation efforts, and to understand the history of temple destruction in India.### Out-of-Scope Use\n\nMisuse of this dataset for promoting hatred, violence, or discrimination is strictly out of scope.## Dataset Structure\n\n[Information Not Available]## Dataset Creation### Curation Rationale\n\nThe dataset was created to document the historical information about the destruction of temples in India for research, education, and preservation purposes. It includes data from books authored by Sir Sita Ram Goel and contributions by Gaurav Sinha.### Source Data#### Data Collection and Processing\n\nThe data for this dataset was collected from historical records, scholarly research, and reputable sources, including books authored by Sir Sita Ram Goel. It also includes contributions by Gaurav Sinha. The dataset was carefully compiled to provide accurate and valuable information.#### Who are the source data producers?\n\nThe source data was produced by historians, researchers, and scholars, including Sir Sita Ram Goel, who documented the destruction of temples in India. Contributions by Gaurav Sinha are also part of the dataset."
]
|
eac8109e67622d396288001a0c1ccca16ba10d43 | # Dataset Card for "test2"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | hudssntao/test2 | [
"region:us"
]
| 2023-10-27T02:32:52+00:00 | {"dataset_info": {"features": [{"name": "column1", "dtype": "string"}, {"name": "column2", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 100, "num_examples": 5}], "download_size": 1255, "dataset_size": 100}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-30T05:11:50+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "test2"
More Information needed | [
"# Dataset Card for \"test2\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"test2\"\n\nMore Information needed"
]
| [
6,
12
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"test2\"\n\nMore Information needed"
]
|
d7d43d0c5b3b7d7ecba2600269fbea222f4747ff |
# Dataset Card for Dataset Name
<!-- Provide a quick summary of the dataset. -->
This dataset card aims to be a base template for new datasets. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/datasetcard_template.md?plain=1).
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] | kakafei/444 | [
"region:us"
]
| 2023-10-27T02:36:18+00:00 | {} | 2023-10-27T02:37:53+00:00 | []
| []
| TAGS
#region-us
|
# Dataset Card for Dataset Name
This dataset card aims to be a base template for new datasets. It has been generated using this raw template.
## Dataset Details
### Dataset Description
- Curated by:
- Funded by [optional]:
- Shared by [optional]:
- Language(s) (NLP):
- License:
### Dataset Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Out-of-Scope Use
## Dataset Structure
## Dataset Creation
### Curation Rationale
### Source Data
#### Data Collection and Processing
#### Who are the source data producers?
### Annotations [optional]
#### Annotation process
#### Who are the annotators?
#### Personal and Sensitive Information
## Bias, Risks, and Limitations
### Recommendations
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Dataset Card Authors [optional]
## Dataset Card Contact
| [
"# Dataset Card for Dataset Name\n\n\n\nThis dataset card aims to be a base template for new datasets. It has been generated using this raw template.",
"## Dataset Details",
"### Dataset Description\n\n\n\n\n\n- Curated by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Language(s) (NLP): \n- License:",
"### Dataset Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Out-of-Scope Use",
"## Dataset Structure",
"## Dataset Creation",
"### Curation Rationale",
"### Source Data",
"#### Data Collection and Processing",
"#### Who are the source data producers?",
"### Annotations [optional]",
"#### Annotation process",
"#### Who are the annotators?",
"#### Personal and Sensitive Information",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Dataset Card Authors [optional]",
"## Dataset Card Contact"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for Dataset Name\n\n\n\nThis dataset card aims to be a base template for new datasets. It has been generated using this raw template.",
"## Dataset Details",
"### Dataset Description\n\n\n\n\n\n- Curated by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Language(s) (NLP): \n- License:",
"### Dataset Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Out-of-Scope Use",
"## Dataset Structure",
"## Dataset Creation",
"### Curation Rationale",
"### Source Data",
"#### Data Collection and Processing",
"#### Who are the source data producers?",
"### Annotations [optional]",
"#### Annotation process",
"#### Who are the annotators?",
"#### Personal and Sensitive Information",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Dataset Card Authors [optional]",
"## Dataset Card Contact"
]
| [
6,
34,
4,
40,
29,
3,
4,
9,
6,
5,
7,
4,
7,
10,
9,
5,
9,
8,
10,
46,
8,
7,
10,
5
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for Dataset Name\n\n\n\nThis dataset card aims to be a base template for new datasets. It has been generated using this raw template.## Dataset Details### Dataset Description\n\n\n\n\n\n- Curated by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Language(s) (NLP): \n- License:### Dataset Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Out-of-Scope Use## Dataset Structure## Dataset Creation### Curation Rationale### Source Data#### Data Collection and Processing#### Who are the source data producers?### Annotations [optional]#### Annotation process#### Who are the annotators?#### Personal and Sensitive Information## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Dataset Card Authors [optional]## Dataset Card Contact"
]
|
67d32349bbf1d3e57ec4e8a1269546e63b1e07f1 | # Dataset Card for "GRE_words_gregmat-marriam_webster"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | chirunder/GRE_words_gregmat-marriam_webster | [
"region:us"
]
| 2023-10-27T02:59:56+00:00 | {"dataset_info": {"features": [{"name": "word", "dtype": "string"}, {"name": "html", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 599332498, "num_examples": 1112}], "download_size": 167890595, "dataset_size": 599332498}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-27T03:00:11+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "GRE_words_gregmat-marriam_webster"
More Information needed | [
"# Dataset Card for \"GRE_words_gregmat-marriam_webster\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"GRE_words_gregmat-marriam_webster\"\n\nMore Information needed"
]
| [
6,
23
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"GRE_words_gregmat-marriam_webster\"\n\nMore Information needed"
]
|
20ce4801b1b13c8bc4dc601663b2281e0b084c6a | # Dataset Card for "capstone_hal_without_gold"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | Deojoandco/capstone_hal_without_gold | [
"region:us"
]
| 2023-10-27T03:02:56+00:00 | {"dataset_info": {"features": [{"name": "dialog_id", "dtype": "int32"}, {"name": "source", "sequence": "string"}, {"name": "tags", "sequence": {"class_label": {"names": {"0": "C", "1": "M", "2": "N", "3": "O", "4": "OB", "5": "W"}}}}], "splits": [{"name": "train", "num_bytes": 239933, "num_examples": 76}, {"name": "validation", "num_bytes": 47958, "num_examples": 12}, {"name": "test", "num_bytes": 27286, "num_examples": 12}], "download_size": 35488, "dataset_size": 315177}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}, {"split": "validation", "path": "data/validation-*"}, {"split": "test", "path": "data/test-*"}]}]} | 2023-10-27T03:03:05+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "capstone_hal_without_gold"
More Information needed | [
"# Dataset Card for \"capstone_hal_without_gold\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"capstone_hal_without_gold\"\n\nMore Information needed"
]
| [
6,
19
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"capstone_hal_without_gold\"\n\nMore Information needed"
]
|
f4c240b3ce773cdca0431a98a793f500b731f37c | # Dataset Card for "capstone_hal_with_gold"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | Deojoandco/capstone_hal_with_gold | [
"region:us"
]
| 2023-10-27T03:03:31+00:00 | {"dataset_info": {"features": [{"name": "dialog_id", "dtype": "int32"}, {"name": "source", "sequence": "string"}, {"name": "tags", "sequence": {"class_label": {"names": {"0": "C", "1": "M", "2": "N", "3": "O", "4": "OB", "5": "W"}}}}], "splits": [{"name": "train", "num_bytes": 268989, "num_examples": 76}, {"name": "validation", "num_bytes": 53862, "num_examples": 12}, {"name": "test", "num_bytes": 31570, "num_examples": 12}], "download_size": 39058, "dataset_size": 354421}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}, {"split": "validation", "path": "data/validation-*"}, {"split": "test", "path": "data/test-*"}]}]} | 2023-10-27T03:03:41+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "capstone_hal_with_gold"
More Information needed | [
"# Dataset Card for \"capstone_hal_with_gold\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"capstone_hal_with_gold\"\n\nMore Information needed"
]
| [
6,
18
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"capstone_hal_with_gold\"\n\nMore Information needed"
]
|
2b7c30e79edffd2aa3b743627f3411f419567802 | # Dataset Card for "oasst_lima_arc"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | ycchen/oasst_lima_arc | [
"region:us"
]
| 2023-10-27T03:12:45+00:00 | {"dataset_info": {"features": [{"name": "conversations", "sequence": "string"}, {"name": "source", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 8102880, "num_examples": 4970}], "download_size": 4569911, "dataset_size": 8102880}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-27T03:18:07+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "oasst_lima_arc"
More Information needed | [
"# Dataset Card for \"oasst_lima_arc\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"oasst_lima_arc\"\n\nMore Information needed"
]
| [
6,
18
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"oasst_lima_arc\"\n\nMore Information needed"
]
|
b9f68b227e69065cf97786ff8db785f09d668195 | # Dataset Card for "quirky_math_bob_grader_last_1.0e_0.0p_finetuning"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | atmallen/quirky_math_bob_grader_last_1.0e_0.0p_finetuning | [
"region:us"
]
| 2023-10-27T03:33:46+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}, {"split": "validation", "path": "data/validation-*"}, {"split": "test", "path": "data/test-*"}]}], "dataset_info": {"features": [{"name": "statement", "dtype": "string"}, {"name": "choices", "sequence": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "False", "1": "True"}}}}, {"name": "true_label", "dtype": "bool"}], "splits": [{"name": "train", "num_bytes": 11540623, "num_examples": 200000}, {"name": "validation", "num_bytes": 1159427, "num_examples": 20000}, {"name": "test", "num_bytes": 1159757, "num_examples": 20000}], "download_size": 3315827, "dataset_size": 13859807}} | 2023-10-27T03:33:48+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "quirky_math_bob_grader_last_1.0e_0.0p_finetuning"
More Information needed | [
"# Dataset Card for \"quirky_math_bob_grader_last_1.0e_0.0p_finetuning\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"quirky_math_bob_grader_last_1.0e_0.0p_finetuning\"\n\nMore Information needed"
]
| [
6,
31
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"quirky_math_bob_grader_last_1.0e_0.0p_finetuning\"\n\nMore Information needed"
]
|
1141ad6606b089b1491a3e2fd37bce54f142d8ff | # Disclaimer:
this dataset is curated for NeurIPS 2023 LLM efficiency challange, and currently work in progress. Please use at your own risk.
# Data composition:
All data were derived from the training set portion of the open source dataset.
**Data Sources**:
-dolly: https://huggingface.co/datasets/databricks/databricks-dolly-15k
-cnn_dailymail: https://huggingface.co/datasets/cnn_dailymail
-mmlu: https://huggingface.co/datasets/cais/mmlu
-bbq: https://huggingface.co/datasets/tasksource/bigbench
-ScienceQA: https://huggingface.co/datasets/tasksource/ScienceQA_text_only | zhongshupeng/dataset_A100 | [
"region:us"
]
| 2023-10-27T03:41:46+00:00 | {} | 2023-10-27T04:41:49+00:00 | []
| []
| TAGS
#region-us
| # Disclaimer:
this dataset is curated for NeurIPS 2023 LLM efficiency challange, and currently work in progress. Please use at your own risk.
# Data composition:
All data were derived from the training set portion of the open source dataset.
Data Sources:
-dolly: URL
-cnn_dailymail: URL
-mmlu: URL
-bbq: URL
-ScienceQA: URL | [
"# Disclaimer: \nthis dataset is curated for NeurIPS 2023 LLM efficiency challange, and currently work in progress. Please use at your own risk.",
"# Data composition: \nAll data were derived from the training set portion of the open source dataset.\n\nData Sources:\n\n-dolly: URL\n\n-cnn_dailymail: URL\n\n-mmlu: URL\n\n-bbq: URL\n\n-ScienceQA: URL"
]
| [
"TAGS\n#region-us \n",
"# Disclaimer: \nthis dataset is curated for NeurIPS 2023 LLM efficiency challange, and currently work in progress. Please use at your own risk.",
"# Data composition: \nAll data were derived from the training set portion of the open source dataset.\n\nData Sources:\n\n-dolly: URL\n\n-cnn_dailymail: URL\n\n-mmlu: URL\n\n-bbq: URL\n\n-ScienceQA: URL"
]
| [
6,
34,
53
]
| [
"passage: TAGS\n#region-us \n# Disclaimer: \nthis dataset is curated for NeurIPS 2023 LLM efficiency challange, and currently work in progress. Please use at your own risk.# Data composition: \nAll data were derived from the training set portion of the open source dataset.\n\nData Sources:\n\n-dolly: URL\n\n-cnn_dailymail: URL\n\n-mmlu: URL\n\n-bbq: URL\n\n-ScienceQA: URL"
]
|
8bc8e93de29b36770b1b309517f574452a472747 | # Dataset Card for "kan-ds-mini"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | anonymouse03052002/kan-ds-mini | [
"region:us"
]
| 2023-10-27T03:54:12+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}, {"split": "validation", "path": "data/validation-*"}]}], "dataset_info": {"features": [{"name": "text", "dtype": "string"}, {"name": "text_length", "dtype": "int64"}], "splits": [{"name": "train", "num_bytes": 49353.04, "num_examples": 88}, {"name": "validation", "num_bytes": 5608.3, "num_examples": 10}], "download_size": 0, "dataset_size": 54961.340000000004}} | 2023-10-27T05:34:59+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "kan-ds-mini"
More Information needed | [
"# Dataset Card for \"kan-ds-mini\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"kan-ds-mini\"\n\nMore Information needed"
]
| [
6,
15
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"kan-ds-mini\"\n\nMore Information needed"
]
|
2a371e1f5ad60ee482798added87285f15fc094c | # Dataset Card for "repo_dedup_sep2023"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | tianyang/repo_dedup_sep2023 | [
"region:us"
]
| 2023-10-27T04:21:30+00:00 | {"dataset_info": {"features": [{"name": "repo_name", "dtype": "string"}, {"name": "language", "dtype": "string"}, {"name": "created_at", "dtype": "timestamp[ns]"}, {"name": "license", "dtype": "string"}, {"name": "description", "dtype": "string"}, {"name": "stars", "dtype": "int64"}, {"name": "forks", "dtype": "int64"}, {"name": "url", "dtype": "string"}, {"name": "repo_code", "list": [{"name": "code", "dtype": "string"}, {"name": "path", "dtype": "string"}, {"name": "repo_name", "dtype": "string"}, {"name": "size", "dtype": "int64"}]}], "splits": [{"name": "train", "num_bytes": 219555370, "num_examples": 1474}], "download_size": 71458940, "dataset_size": 219555370}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-27T04:26:00+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "repo_dedup_sep2023"
More Information needed | [
"# Dataset Card for \"repo_dedup_sep2023\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"repo_dedup_sep2023\"\n\nMore Information needed"
]
| [
6,
19
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"repo_dedup_sep2023\"\n\nMore Information needed"
]
|
a0aa544a9c6c861f49f0c75148ea32f02877d031 | AIDA/testc introduced in the paper [SPEL: Structured Prediction for Entity Linking (EMNLP 2023)](https://arxiv.org/abs/2310.14684), contains 131 Reuters news articles published between December 5th and 7th, 2020.
We have meticulously linked the named entity mentions in the newly annotated NER test set of (Liu and Ritter, 2023) to their corresponding Wikipedia pages, using the same linking procedure employed in the original AIDA dataset.
Our new entity linking test set, AIDA/testc, has 1,160 unique Wikipedia identifiers, spanning over 3,777 mentions and encompassing a total of 46,456 words.
This dataset is in NIF format and can be easily integrated into [GERBIL](https://github.com/dice-group/gerbil).
### How can I integrate AIDA/testc into GERBIL?
Here is the simple modifications you need to do:
1. If you are running GERBIL, stop the process.
2. Put [`aida_testc.ttl`](aida_testc.ttl) in `gerbil/gerbil_data/datasets/aida`
3. Open `gerbil/src/main/properties/datasets.properties` (this properties file contains the dataset configurations for GERBIL).
4. Copy the following lines underneath the last line defining AIDA/CoNLL-Test B:
```
org.aksw.gerbil.datasets.AIDATestC.file=${org.aksw.gerbil.DataPath}/datasets/aida/aida_testc.ttl
org.aksw.gerbil.datasets.definition.AIDATestC.name=AIDA/CoNLL-Test C
org.aksw.gerbil.datasets.definition.AIDATestC.class=org.aksw.gerbil.dataset.impl.nif.FileBasedNIFDataset
org.aksw.gerbil.datasets.definition.AIDATestC.cacheable=true
org.aksw.gerbil.datasets.definition.AIDATestC.experimentType=A2KB
org.aksw.gerbil.datasets.definition.AIDATestC.constructorArgs=${org.aksw.gerbil.datasets.AIDATestC.file},${org.aksw.gerbil.datasets.definition.AIDATestC.name}
```
5. Run GERBIL, the new dataset should show up.
| sshavara/AIDA_testc | [
"license:cc-by-4.0",
"arxiv:2310.14684",
"region:us"
]
| 2023-10-27T04:42:13+00:00 | {"license": "cc-by-4.0", "pretty_name": "AIDA/testc", "viewer": false} | 2023-10-27T04:55:54+00:00 | [
"2310.14684"
]
| []
| TAGS
#license-cc-by-4.0 #arxiv-2310.14684 #region-us
| AIDA/testc introduced in the paper SPEL: Structured Prediction for Entity Linking (EMNLP 2023), contains 131 Reuters news articles published between December 5th and 7th, 2020.
We have meticulously linked the named entity mentions in the newly annotated NER test set of (Liu and Ritter, 2023) to their corresponding Wikipedia pages, using the same linking procedure employed in the original AIDA dataset.
Our new entity linking test set, AIDA/testc, has 1,160 unique Wikipedia identifiers, spanning over 3,777 mentions and encompassing a total of 46,456 words.
This dataset is in NIF format and can be easily integrated into GERBIL.
### How can I integrate AIDA/testc into GERBIL?
Here is the simple modifications you need to do:
1. If you are running GERBIL, stop the process.
2. Put 'aida_testc.ttl' in 'gerbil/gerbil_data/datasets/aida'
3. Open 'gerbil/src/main/properties/datasets.properties' (this properties file contains the dataset configurations for GERBIL).
4. Copy the following lines underneath the last line defining AIDA/CoNLL-Test B:
5. Run GERBIL, the new dataset should show up.
| [
"### How can I integrate AIDA/testc into GERBIL?\nHere is the simple modifications you need to do:\n\n1. If you are running GERBIL, stop the process.\n2. Put 'aida_testc.ttl' in 'gerbil/gerbil_data/datasets/aida'\n3. Open 'gerbil/src/main/properties/datasets.properties' (this properties file contains the dataset configurations for GERBIL).\n4. Copy the following lines underneath the last line defining AIDA/CoNLL-Test B:\n \n5. Run GERBIL, the new dataset should show up."
]
| [
"TAGS\n#license-cc-by-4.0 #arxiv-2310.14684 #region-us \n",
"### How can I integrate AIDA/testc into GERBIL?\nHere is the simple modifications you need to do:\n\n1. If you are running GERBIL, stop the process.\n2. Put 'aida_testc.ttl' in 'gerbil/gerbil_data/datasets/aida'\n3. Open 'gerbil/src/main/properties/datasets.properties' (this properties file contains the dataset configurations for GERBIL).\n4. Copy the following lines underneath the last line defining AIDA/CoNLL-Test B:\n \n5. Run GERBIL, the new dataset should show up."
]
| [
24,
142
]
| [
"passage: TAGS\n#license-cc-by-4.0 #arxiv-2310.14684 #region-us \n### How can I integrate AIDA/testc into GERBIL?\nHere is the simple modifications you need to do:\n\n1. If you are running GERBIL, stop the process.\n2. Put 'aida_testc.ttl' in 'gerbil/gerbil_data/datasets/aida'\n3. Open 'gerbil/src/main/properties/datasets.properties' (this properties file contains the dataset configurations for GERBIL).\n4. Copy the following lines underneath the last line defining AIDA/CoNLL-Test B:\n \n5. Run GERBIL, the new dataset should show up."
]
|
42fd82dc8f4eeab29c67cda460b804c1b2d3a409 | # Dataset Card for "anno1_w_elimination"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | anlp/anno1_w_elimination | [
"region:us"
]
| 2023-10-27T04:53:09+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}], "dataset_info": {"features": [{"name": "sentences", "sequence": "string"}, {"name": "ner_tags", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 1239484, "num_examples": 917}], "download_size": 249472, "dataset_size": 1239484}} | 2023-10-27T04:53:10+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "anno1_w_elimination"
More Information needed | [
"# Dataset Card for \"anno1_w_elimination\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"anno1_w_elimination\"\n\nMore Information needed"
]
| [
6,
17
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"anno1_w_elimination\"\n\nMore Information needed"
]
|
6bc723bc3dc7cc588ec67f9d4ea2f2a5d830d377 | # Dataset Card for "filtered_lemma41kV0.0.2"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | fia24/filtered_lemma41kV0.0.2 | [
"region:us"
]
| 2023-10-27T05:08:14+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}, {"split": "test", "path": "data/test-*"}, {"split": "val", "path": "data/val-*"}]}], "dataset_info": {"features": [{"name": "Unnamed: 0", "dtype": "int64"}, {"name": "Inflected_Word", "dtype": "string"}, {"name": "Lemma", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 1794357.4941723635, "num_examples": 28553}, {"name": "test", "num_bytes": 224349.67443684858, "num_examples": 3570}, {"name": "val", "num_bytes": 224286.83139078785, "num_examples": 3569}], "download_size": 1201505, "dataset_size": 2242994.0}} | 2023-10-27T05:08:22+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "filtered_lemma41kV0.0.2"
More Information needed | [
"# Dataset Card for \"filtered_lemma41kV0.0.2\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"filtered_lemma41kV0.0.2\"\n\nMore Information needed"
]
| [
6,
21
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"filtered_lemma41kV0.0.2\"\n\nMore Information needed"
]
|
8471a2411e3294a628ac958122891948dcf6b879 | OpenOrcaデータセットの日本語翻訳版です
https://huggingface.co/datasets/Open-Orca/OpenOrca
現在翻訳作業が続行中で、OpenOrca全体の1/5程度の翻訳が終わった状態でひとまず公開します。商用利用可能です。
| shi3z/OpenOrcaJapanese | [
"task_categories:table-question-answering",
"size_categories:100M<n<1B",
"language:ja",
"license:mit",
"region:us"
]
| 2023-10-27T05:15:27+00:00 | {"language": ["ja"], "license": "mit", "size_categories": ["100M<n<1B"], "task_categories": ["table-question-answering"]} | 2023-10-28T01:50:27+00:00 | []
| [
"ja"
]
| TAGS
#task_categories-table-question-answering #size_categories-100M<n<1B #language-Japanese #license-mit #region-us
| OpenOrcaデータセットの日本語翻訳版です
URL
現在翻訳作業が続行中で、OpenOrca全体の1/5程度の翻訳が終わった状態でひとまず公開します。商用利用可能です。
| []
| [
"TAGS\n#task_categories-table-question-answering #size_categories-100M<n<1B #language-Japanese #license-mit #region-us \n"
]
| [
43
]
| [
"passage: TAGS\n#task_categories-table-question-answering #size_categories-100M<n<1B #language-Japanese #license-mit #region-us \n"
]
|
792342ba43adfa7bd25d48a72a21103da39b28c4 | # Dataset Card for "LORA_ONE_DATA"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | Adminhuggingface/LORA_ONE_DATA | [
"region:us"
]
| 2023-10-27T05:18:32+00:00 | {"dataset_info": {"features": [{"name": "image", "dtype": "image"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 2493084.0, "num_examples": 6}], "download_size": 2495157, "dataset_size": 2493084.0}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-27T05:18:33+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "LORA_ONE_DATA"
More Information needed | [
"# Dataset Card for \"LORA_ONE_DATA\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"LORA_ONE_DATA\"\n\nMore Information needed"
]
| [
6,
17
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"LORA_ONE_DATA\"\n\nMore Information needed"
]
|
e3d49f61fa580686ad58ce0527bc34cb3cedcfcd | # 中文论文问答数据集
* 来自知网的论文数据,版权受限,不能直接公开。下载后请勿上传到公开场合。
* 包括 为论文写摘要、基于论文内容的问答 两个任务。论文摘要任务已经迁移到[论文摘要数据集](https://huggingface.co/datasets/yuyijiong/Chinese_Paper_Abstract/settings)中。
## 改进版
* 此数据集中筛选出较长的论文,并为每篇论文设计多个任务,形成新数据集:[中文论文多任务数据集](https://huggingface.co/datasets/yuyijiong/Paper_mutli_QA_Chinese) | yuyijiong/Chinese_Paper_QA | [
"size_categories:1K<n<10K",
"language:zh",
"license:cc-by-nc-4.0",
"region:us"
]
| 2023-10-27T05:26:05+00:00 | {"language": ["zh"], "license": "cc-by-nc-4.0", "size_categories": ["1K<n<10K"]} | 2023-11-21T05:56:27+00:00 | []
| [
"zh"
]
| TAGS
#size_categories-1K<n<10K #language-Chinese #license-cc-by-nc-4.0 #region-us
| # 中文论文问答数据集
* 来自知网的论文数据,版权受限,不能直接公开。下载后请勿上传到公开场合。
* 包括 为论文写摘要、基于论文内容的问答 两个任务。论文摘要任务已经迁移到论文摘要数据集中。
## 改进版
* 此数据集中筛选出较长的论文,并为每篇论文设计多个任务,形成新数据集:中文论文多任务数据集 | [
"# 中文论文问答数据集\n* 来自知网的论文数据,版权受限,不能直接公开。下载后请勿上传到公开场合。\n* 包括 为论文写摘要、基于论文内容的问答 两个任务。论文摘要任务已经迁移到论文摘要数据集中。",
"## 改进版\n* 此数据集中筛选出较长的论文,并为每篇论文设计多个任务,形成新数据集:中文论文多任务数据集"
]
| [
"TAGS\n#size_categories-1K<n<10K #language-Chinese #license-cc-by-nc-4.0 #region-us \n",
"# 中文论文问答数据集\n* 来自知网的论文数据,版权受限,不能直接公开。下载后请勿上传到公开场合。\n* 包括 为论文写摘要、基于论文内容的问答 两个任务。论文摘要任务已经迁移到论文摘要数据集中。",
"## 改进版\n* 此数据集中筛选出较长的论文,并为每篇论文设计多个任务,形成新数据集:中文论文多任务数据集"
]
| [
34,
66,
35
]
| [
"passage: TAGS\n#size_categories-1K<n<10K #language-Chinese #license-cc-by-nc-4.0 #region-us \n# 中文论文问答数据集\n* 来自知网的论文数据,版权受限,不能直接公开。下载后请勿上传到公开场合。\n* 包括 为论文写摘要、基于论文内容的问答 两个任务。论文摘要任务已经迁移到论文摘要数据集中。## 改进版\n* 此数据集中筛选出较长的论文,并为每篇论文设计多个任务,形成新数据集:中文论文多任务数据集"
]
|
0af454a518c4575fd7f43677612b9bad7dcdb2a5 | # Dataset Card for "one_piece_dataset"
<h1> This dataset contains 922 images taken from the one piece anime, with each row containing the coloured image and the sketch one. </h1>
<img alt='example images' src="./combined.png"/ >
<h2> Example Setup (the images are not normalized) </h2>
<code>
transform = Compose([
transforms.Resize((128, 128)),
transforms.ToTensor(),
transforms.Lambda(lambda t: (t * 2) - 1)
])
def train():
def transforms(examples):
examples["sketch_pixel_values"] = [transform(image.convert("RGB")) for image in examples["sketch"]]
examples["full_colour_pixel_values"] = [ transform(image.convert("RGB")) for image in examples["full_colour"]]
del examples["sketch"]
del examples["full_colour"]
return examples
dataset = load_dataset("pawlo2013/one_piece_dataset", split="train")
transformed_full_colour_dataset = dataset.with_transform(transforms)
dataloader = DataLoader(transformed_full_colour_dataset, batch_size=16, shuffle=True, num_workers=0)
</code> | pawlo2013/one_piece_dataset | [
"region:us"
]
| 2023-10-27T05:34:37+00:00 | {"dataset_info": {"features": [{"name": "full_colour", "dtype": "image"}, {"name": "sketch", "dtype": "image"}], "splits": [{"name": "train", "num_bytes": 170204480.0, "num_examples": 922}], "download_size": 170225532, "dataset_size": 170204480.0}} | 2023-10-28T12:45:08+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "one_piece_dataset"
<h1> This dataset contains 922 images taken from the one piece anime, with each row containing the coloured image and the sketch one. </h1>
<img alt='example images' src="./URL"/ >
<h2> Example Setup (the images are not normalized) </h2>
<code>
transform = Compose([
transforms.Resize((128, 128)),
transforms.ToTensor(),
transforms.Lambda(lambda t: (t * 2) - 1)
])
def train():
def transforms(examples):
examples["sketch_pixel_values"] = [transform(image.convert("RGB")) for image in examples["sketch"]]
examples["full_colour_pixel_values"] = [ transform(image.convert("RGB")) for image in examples["full_colour"]]
del examples["sketch"]
del examples["full_colour"]
return examples
dataset = load_dataset("pawlo2013/one_piece_dataset", split="train")
transformed_full_colour_dataset = dataset.with_transform(transforms)
dataloader = DataLoader(transformed_full_colour_dataset, batch_size=16, shuffle=True, num_workers=0)
</code> | [
"# Dataset Card for \"one_piece_dataset\"\n\n<h1> This dataset contains 922 images taken from the one piece anime, with each row containing the coloured image and the sketch one. </h1>\n<img alt='example images' src=\"./URL\"/ > \n\n<h2> Example Setup (the images are not normalized) </h2>\n<code>\n\n \n transform = Compose([\n transforms.Resize((128, 128)),\n transforms.ToTensor(),\n transforms.Lambda(lambda t: (t * 2) - 1)\n ])\n \n def train():\n\n def transforms(examples):\n examples[\"sketch_pixel_values\"] = [transform(image.convert(\"RGB\")) for image in examples[\"sketch\"]]\n examples[\"full_colour_pixel_values\"] = [ transform(image.convert(\"RGB\")) for image in examples[\"full_colour\"]]\n del examples[\"sketch\"]\n del examples[\"full_colour\"]\n\n return examples\n\n\n dataset = load_dataset(\"pawlo2013/one_piece_dataset\", split=\"train\")\n\n transformed_full_colour_dataset = dataset.with_transform(transforms)\n\n dataloader = DataLoader(transformed_full_colour_dataset, batch_size=16, shuffle=True, num_workers=0)\n</code>"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"one_piece_dataset\"\n\n<h1> This dataset contains 922 images taken from the one piece anime, with each row containing the coloured image and the sketch one. </h1>\n<img alt='example images' src=\"./URL\"/ > \n\n<h2> Example Setup (the images are not normalized) </h2>\n<code>\n\n \n transform = Compose([\n transforms.Resize((128, 128)),\n transforms.ToTensor(),\n transforms.Lambda(lambda t: (t * 2) - 1)\n ])\n \n def train():\n\n def transforms(examples):\n examples[\"sketch_pixel_values\"] = [transform(image.convert(\"RGB\")) for image in examples[\"sketch\"]]\n examples[\"full_colour_pixel_values\"] = [ transform(image.convert(\"RGB\")) for image in examples[\"full_colour\"]]\n del examples[\"sketch\"]\n del examples[\"full_colour\"]\n\n return examples\n\n\n dataset = load_dataset(\"pawlo2013/one_piece_dataset\", split=\"train\")\n\n transformed_full_colour_dataset = dataset.with_transform(transforms)\n\n dataloader = DataLoader(transformed_full_colour_dataset, batch_size=16, shuffle=True, num_workers=0)\n</code>"
]
| [
6,
354
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"one_piece_dataset\"\n\n<h1> This dataset contains 922 images taken from the one piece anime, with each row containing the coloured image and the sketch one. </h1>\n<img alt='example images' src=\"./URL\"/ > \n\n<h2> Example Setup (the images are not normalized) </h2>\n<code>\n\n \n transform = Compose([\n transforms.Resize((128, 128)),\n transforms.ToTensor(),\n transforms.Lambda(lambda t: (t * 2) - 1)\n ])\n \n def train():\n\n def transforms(examples):\n examples[\"sketch_pixel_values\"] = [transform(image.convert(\"RGB\")) for image in examples[\"sketch\"]]\n examples[\"full_colour_pixel_values\"] = [ transform(image.convert(\"RGB\")) for image in examples[\"full_colour\"]]\n del examples[\"sketch\"]\n del examples[\"full_colour\"]\n\n return examples\n\n\n dataset = load_dataset(\"pawlo2013/one_piece_dataset\", split=\"train\")\n\n transformed_full_colour_dataset = dataset.with_transform(transforms)\n\n dataloader = DataLoader(transformed_full_colour_dataset, batch_size=16, shuffle=True, num_workers=0)\n</code>"
]
|
14b9e4ad3055b6da95fab184aa623bac80db4d11 |
# Dataset Card for Evaluation run of Weyaxi/llama-2-alpacagpt4-1000step
## Dataset Description
- **Homepage:**
- **Repository:** https://huggingface.co/Weyaxi/llama-2-alpacagpt4-1000step
- **Paper:**
- **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
- **Point of Contact:** [email protected]
### Dataset Summary
Dataset automatically created during the evaluation run of model [Weyaxi/llama-2-alpacagpt4-1000step](https://huggingface.co/Weyaxi/llama-2-alpacagpt4-1000step) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
The dataset is composed of 3 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the agregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).
To load the details from a run, you can for instance do the following:
```python
from datasets import load_dataset
data = load_dataset("open-llm-leaderboard/details_Weyaxi__llama-2-alpacagpt4-1000step",
"harness_winogrande_5",
split="train")
```
## Latest results
These are the [latest results from run 2023-10-27T06:34:41.482780](https://huggingface.co/datasets/open-llm-leaderboard/details_Weyaxi__llama-2-alpacagpt4-1000step/blob/main/results_2023-10-27T06-34-41.482780.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
```python
{
"all": {
"em": 0.007864932885906041,
"em_stderr": 0.0009046332824008166,
"f1": 0.07264681208053722,
"f1_stderr": 0.0016288280664899088,
"acc": 0.4836165974438314,
"acc_stderr": 0.010680678903995254
},
"harness|drop|3": {
"em": 0.007864932885906041,
"em_stderr": 0.0009046332824008166,
"f1": 0.07264681208053722,
"f1_stderr": 0.0016288280664899088
},
"harness|gsm8k|5": {
"acc": 0.16376042456406367,
"acc_stderr": 0.010193237214420942
},
"harness|winogrande|5": {
"acc": 0.8034727703235991,
"acc_stderr": 0.011168120593569567
}
}
```
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] | open-llm-leaderboard/details_Weyaxi__llama-2-alpacagpt4-1000step | [
"region:us"
]
| 2023-10-27T05:34:45+00:00 | {"pretty_name": "Evaluation run of Weyaxi/llama-2-alpacagpt4-1000step", "dataset_summary": "Dataset automatically created during the evaluation run of model [Weyaxi/llama-2-alpacagpt4-1000step](https://huggingface.co/Weyaxi/llama-2-alpacagpt4-1000step) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\nThe dataset is composed of 3 configuration, each one coresponding to one of the evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The \"train\" split is always pointing to the latest results.\n\nAn additional configuration \"results\" store all the aggregated results of the run (and is used to compute and display the agregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\nTo load the details from a run, you can for instance do the following:\n```python\nfrom datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_Weyaxi__llama-2-alpacagpt4-1000step\",\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\nThese are the [latest results from run 2023-10-27T06:34:41.482780](https://huggingface.co/datasets/open-llm-leaderboard/details_Weyaxi__llama-2-alpacagpt4-1000step/blob/main/results_2023-10-27T06-34-41.482780.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the \"latest\" split for each eval):\n\n```python\n{\n \"all\": {\n \"em\": 0.007864932885906041,\n \"em_stderr\": 0.0009046332824008166,\n \"f1\": 0.07264681208053722,\n \"f1_stderr\": 0.0016288280664899088,\n \"acc\": 0.4836165974438314,\n \"acc_stderr\": 0.010680678903995254\n },\n \"harness|drop|3\": {\n \"em\": 0.007864932885906041,\n \"em_stderr\": 0.0009046332824008166,\n \"f1\": 0.07264681208053722,\n \"f1_stderr\": 0.0016288280664899088\n },\n \"harness|gsm8k|5\": {\n \"acc\": 0.16376042456406367,\n \"acc_stderr\": 0.010193237214420942\n },\n \"harness|winogrande|5\": {\n \"acc\": 0.8034727703235991,\n \"acc_stderr\": 0.011168120593569567\n }\n}\n```", "repo_url": "https://huggingface.co/Weyaxi/llama-2-alpacagpt4-1000step", "leaderboard_url": "https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard", "point_of_contact": "[email protected]", "configs": [{"config_name": "harness_drop_3", "data_files": [{"split": "2023_10_27T06_34_41.482780", "path": ["**/details_harness|drop|3_2023-10-27T06-34-41.482780.parquet"]}, {"split": "latest", "path": ["**/details_harness|drop|3_2023-10-27T06-34-41.482780.parquet"]}]}, {"config_name": "harness_gsm8k_5", "data_files": [{"split": "2023_10_27T06_34_41.482780", "path": ["**/details_harness|gsm8k|5_2023-10-27T06-34-41.482780.parquet"]}, {"split": "latest", "path": ["**/details_harness|gsm8k|5_2023-10-27T06-34-41.482780.parquet"]}]}, {"config_name": "harness_winogrande_5", "data_files": [{"split": "2023_10_27T06_34_41.482780", "path": ["**/details_harness|winogrande|5_2023-10-27T06-34-41.482780.parquet"]}, {"split": "latest", "path": ["**/details_harness|winogrande|5_2023-10-27T06-34-41.482780.parquet"]}]}, {"config_name": "results", "data_files": [{"split": "2023_10_27T06_34_41.482780", "path": ["results_2023-10-27T06-34-41.482780.parquet"]}, {"split": "latest", "path": ["results_2023-10-27T06-34-41.482780.parquet"]}]}]} | 2023-10-27T05:34:53+00:00 | []
| []
| TAGS
#region-us
|
# Dataset Card for Evaluation run of Weyaxi/llama-2-alpacagpt4-1000step
## Dataset Description
- Homepage:
- Repository: URL
- Paper:
- Leaderboard: URL
- Point of Contact: clementine@URL
### Dataset Summary
Dataset automatically created during the evaluation run of model Weyaxi/llama-2-alpacagpt4-1000step on the Open LLM Leaderboard.
The dataset is composed of 3 configuration, each one coresponding to one of the evaluated task.
The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results.
An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the agregated metrics on the Open LLM Leaderboard).
To load the details from a run, you can for instance do the following:
## Latest results
These are the latest results from run 2023-10-27T06:34:41.482780(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval):
### Supported Tasks and Leaderboards
### Languages
## Dataset Structure
### Data Instances
### Data Fields
### Data Splits
## Dataset Creation
### Curation Rationale
### Source Data
#### Initial Data Collection and Normalization
#### Who are the source language producers?
### Annotations
#### Annotation process
#### Who are the annotators?
### Personal and Sensitive Information
## Considerations for Using the Data
### Social Impact of Dataset
### Discussion of Biases
### Other Known Limitations
## Additional Information
### Dataset Curators
### Licensing Information
### Contributions
| [
"# Dataset Card for Evaluation run of Weyaxi/llama-2-alpacagpt4-1000step",
"## Dataset Description\n\n- Homepage: \n- Repository: URL\n- Paper: \n- Leaderboard: URL\n- Point of Contact: clementine@URL",
"### Dataset Summary\n\nDataset automatically created during the evaluation run of model Weyaxi/llama-2-alpacagpt4-1000step on the Open LLM Leaderboard.\n\nThe dataset is composed of 3 configuration, each one coresponding to one of the evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The \"train\" split is always pointing to the latest results.\n\nAn additional configuration \"results\" store all the aggregated results of the run (and is used to compute and display the agregated metrics on the Open LLM Leaderboard).\n\nTo load the details from a run, you can for instance do the following:",
"## Latest results\n\nThese are the latest results from run 2023-10-27T06:34:41.482780(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the \"latest\" split for each eval):",
"### Supported Tasks and Leaderboards",
"### Languages",
"## Dataset Structure",
"### Data Instances",
"### Data Fields",
"### Data Splits",
"## Dataset Creation",
"### Curation Rationale",
"### Source Data",
"#### Initial Data Collection and Normalization",
"#### Who are the source language producers?",
"### Annotations",
"#### Annotation process",
"#### Who are the annotators?",
"### Personal and Sensitive Information",
"## Considerations for Using the Data",
"### Social Impact of Dataset",
"### Discussion of Biases",
"### Other Known Limitations",
"## Additional Information",
"### Dataset Curators",
"### Licensing Information",
"### Contributions"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for Evaluation run of Weyaxi/llama-2-alpacagpt4-1000step",
"## Dataset Description\n\n- Homepage: \n- Repository: URL\n- Paper: \n- Leaderboard: URL\n- Point of Contact: clementine@URL",
"### Dataset Summary\n\nDataset automatically created during the evaluation run of model Weyaxi/llama-2-alpacagpt4-1000step on the Open LLM Leaderboard.\n\nThe dataset is composed of 3 configuration, each one coresponding to one of the evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The \"train\" split is always pointing to the latest results.\n\nAn additional configuration \"results\" store all the aggregated results of the run (and is used to compute and display the agregated metrics on the Open LLM Leaderboard).\n\nTo load the details from a run, you can for instance do the following:",
"## Latest results\n\nThese are the latest results from run 2023-10-27T06:34:41.482780(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the \"latest\" split for each eval):",
"### Supported Tasks and Leaderboards",
"### Languages",
"## Dataset Structure",
"### Data Instances",
"### Data Fields",
"### Data Splits",
"## Dataset Creation",
"### Curation Rationale",
"### Source Data",
"#### Initial Data Collection and Normalization",
"#### Who are the source language producers?",
"### Annotations",
"#### Annotation process",
"#### Who are the annotators?",
"### Personal and Sensitive Information",
"## Considerations for Using the Data",
"### Social Impact of Dataset",
"### Discussion of Biases",
"### Other Known Limitations",
"## Additional Information",
"### Dataset Curators",
"### Licensing Information",
"### Contributions"
]
| [
6,
24,
31,
172,
67,
10,
4,
6,
6,
5,
5,
5,
7,
4,
10,
10,
5,
5,
9,
8,
8,
7,
8,
7,
5,
6,
6,
5
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for Evaluation run of Weyaxi/llama-2-alpacagpt4-1000step## Dataset Description\n\n- Homepage: \n- Repository: URL\n- Paper: \n- Leaderboard: URL\n- Point of Contact: clementine@URL### Dataset Summary\n\nDataset automatically created during the evaluation run of model Weyaxi/llama-2-alpacagpt4-1000step on the Open LLM Leaderboard.\n\nThe dataset is composed of 3 configuration, each one coresponding to one of the evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The \"train\" split is always pointing to the latest results.\n\nAn additional configuration \"results\" store all the aggregated results of the run (and is used to compute and display the agregated metrics on the Open LLM Leaderboard).\n\nTo load the details from a run, you can for instance do the following:## Latest results\n\nThese are the latest results from run 2023-10-27T06:34:41.482780(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the \"latest\" split for each eval):### Supported Tasks and Leaderboards### Languages## Dataset Structure### Data Instances### Data Fields### Data Splits## Dataset Creation### Curation Rationale### Source Data#### Initial Data Collection and Normalization#### Who are the source language producers?### Annotations#### Annotation process#### Who are the annotators?### Personal and Sensitive Information## Considerations for Using the Data### Social Impact of Dataset### Discussion of Biases### Other Known Limitations## Additional Information### Dataset Curators### Licensing Information### Contributions"
]
|
6097a1038c5724ff0e74654f43e75045fab7832b | # Dataset Card for "multi-choices-text"
Bộ dữ liệu trắc nghiệm gồm 58,290 dòng từ vungoi. Bộ này có một số đặc điểm sau:
```
- Các câu hỏi đều là câu hỏi hoàn chỉnh với "?" cuối câu
- Các câu hỏi tiếng Anh đều đã bị bỏ qua
- Các phần "Đáp án.*[ABCD]" của field "solution" bị thay bằng ""
- Đã bỏ "." ở từng "answer" của "options" và cả "solution". Chủ yếu để dễ làm prompt.
```
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | nlplabtdtu/multi-choices-text | [
"region:us"
]
| 2023-10-27T05:35:45+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}], "dataset_info": {"features": [{"name": "question", "dtype": "string"}, {"name": "options", "list": [{"name": "answer", "dtype": "string"}, {"name": "key", "dtype": "string"}]}, {"name": "answer", "struct": [{"name": "answer", "dtype": "string"}, {"name": "key", "dtype": "string"}]}, {"name": "solution", "dtype": "string"}, {"name": "type", "dtype": "string"}, {"name": "alnum_start", "dtype": "bool"}, {"name": "prompt", "dtype": "string"}, {"name": "response", "dtype": "string"}, {"name": "grade", "dtype": "string"}, {"name": "subject", "dtype": "string"}, {"name": "prompt_type", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 93596608, "num_examples": 58286}], "download_size": 48223987, "dataset_size": 93596608}} | 2023-11-09T10:07:38+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "multi-choices-text"
Bộ dữ liệu trắc nghiệm gồm 58,290 dòng từ vungoi. Bộ này có một số đặc điểm sau:
More Information needed | [
"# Dataset Card for \"multi-choices-text\"\nBộ dữ liệu trắc nghiệm gồm 58,290 dòng từ vungoi. Bộ này có một số đặc điểm sau:\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"multi-choices-text\"\nBộ dữ liệu trắc nghiệm gồm 58,290 dòng từ vungoi. Bộ này có một số đặc điểm sau:\n\nMore Information needed"
]
| [
6,
43
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"multi-choices-text\"\nBộ dữ liệu trắc nghiệm gồm 58,290 dòng từ vungoi. Bộ này có một số đặc điểm sau:\n\nMore Information needed"
]
|
070c9aec228b6e5231444bd52c85a261cc269c2d | # Lamini Mistral
Lamini Docs dataset formatted for fine-tuning with Mistral-7B Instruct model | mwitiderrick/lamini_mistral | [
"language:en",
"license:apache-2.0",
"region:us"
]
| 2023-10-27T06:03:20+00:00 | {"language": ["en"], "license": "apache-2.0"} | 2023-10-27T06:05:54+00:00 | []
| [
"en"
]
| TAGS
#language-English #license-apache-2.0 #region-us
| # Lamini Mistral
Lamini Docs dataset formatted for fine-tuning with Mistral-7B Instruct model | [
"# Lamini Mistral \n\nLamini Docs dataset formatted for fine-tuning with Mistral-7B Instruct model"
]
| [
"TAGS\n#language-English #license-apache-2.0 #region-us \n",
"# Lamini Mistral \n\nLamini Docs dataset formatted for fine-tuning with Mistral-7B Instruct model"
]
| [
18,
26
]
| [
"passage: TAGS\n#language-English #license-apache-2.0 #region-us \n# Lamini Mistral \n\nLamini Docs dataset formatted for fine-tuning with Mistral-7B Instruct model"
]
|
a2d74f7f6e2f97292ad55af2b44a24ab1e299dbf | # Dataset Card for "t2i_reward"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | toilaluan/t2i_reward | [
"region:us"
]
| 2023-10-27T06:35:19+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}], "dataset_info": {"features": [{"name": "model_type", "dtype": "string"}, {"name": "request_id", "dtype": "int64"}, {"name": "topic", "dtype": "string"}, {"name": "reward", "dtype": "float64"}, {"name": "individual_rewards", "struct": [{"name": "image_rewarder", "dtype": "float64"}, {"name": "hps_v2_rewarder", "dtype": "float64"}]}], "splits": [{"name": "train", "num_bytes": 154200, "num_examples": 1800}], "download_size": 36440, "dataset_size": 154200}} | 2023-10-28T06:27:46+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "t2i_reward"
More Information needed | [
"# Dataset Card for \"t2i_reward\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"t2i_reward\"\n\nMore Information needed"
]
| [
6,
16
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"t2i_reward\"\n\nMore Information needed"
]
|
91c65a9bdd74522444e73fb8fb2e631ef73a099c |
# Dataset Card for Dataset Name
<!-- Provide a quick summary of the dataset. -->
This dataset card aims to be a base template for new datasets. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/datasetcard_template.md?plain=1).
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] | KendrickPham/fine-tuning-csv | [
"region:us"
]
| 2023-10-27T06:45:07+00:00 | {} | 2023-10-27T06:47:00+00:00 | []
| []
| TAGS
#region-us
|
# Dataset Card for Dataset Name
This dataset card aims to be a base template for new datasets. It has been generated using this raw template.
## Dataset Details
### Dataset Description
- Curated by:
- Funded by [optional]:
- Shared by [optional]:
- Language(s) (NLP):
- License:
### Dataset Sources [optional]
- Repository:
- Paper [optional]:
- Demo [optional]:
## Uses
### Direct Use
### Out-of-Scope Use
## Dataset Structure
## Dataset Creation
### Curation Rationale
### Source Data
#### Data Collection and Processing
#### Who are the source data producers?
### Annotations [optional]
#### Annotation process
#### Who are the annotators?
#### Personal and Sensitive Information
## Bias, Risks, and Limitations
### Recommendations
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Dataset Card Authors [optional]
## Dataset Card Contact
| [
"# Dataset Card for Dataset Name\n\n\n\nThis dataset card aims to be a base template for new datasets. It has been generated using this raw template.",
"## Dataset Details",
"### Dataset Description\n\n\n\n\n\n- Curated by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Language(s) (NLP): \n- License:",
"### Dataset Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Out-of-Scope Use",
"## Dataset Structure",
"## Dataset Creation",
"### Curation Rationale",
"### Source Data",
"#### Data Collection and Processing",
"#### Who are the source data producers?",
"### Annotations [optional]",
"#### Annotation process",
"#### Who are the annotators?",
"#### Personal and Sensitive Information",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Dataset Card Authors [optional]",
"## Dataset Card Contact"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for Dataset Name\n\n\n\nThis dataset card aims to be a base template for new datasets. It has been generated using this raw template.",
"## Dataset Details",
"### Dataset Description\n\n\n\n\n\n- Curated by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Language(s) (NLP): \n- License:",
"### Dataset Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
"## Uses",
"### Direct Use",
"### Out-of-Scope Use",
"## Dataset Structure",
"## Dataset Creation",
"### Curation Rationale",
"### Source Data",
"#### Data Collection and Processing",
"#### Who are the source data producers?",
"### Annotations [optional]",
"#### Annotation process",
"#### Who are the annotators?",
"#### Personal and Sensitive Information",
"## Bias, Risks, and Limitations",
"### Recommendations\n\n\n\nUsers should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
"## Glossary [optional]",
"## More Information [optional]",
"## Dataset Card Authors [optional]",
"## Dataset Card Contact"
]
| [
6,
34,
4,
40,
29,
3,
4,
9,
6,
5,
7,
4,
7,
10,
9,
5,
9,
8,
10,
46,
8,
7,
10,
5
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for Dataset Name\n\n\n\nThis dataset card aims to be a base template for new datasets. It has been generated using this raw template.## Dataset Details### Dataset Description\n\n\n\n\n\n- Curated by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Language(s) (NLP): \n- License:### Dataset Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Out-of-Scope Use## Dataset Structure## Dataset Creation### Curation Rationale### Source Data#### Data Collection and Processing#### Who are the source data producers?### Annotations [optional]#### Annotation process#### Who are the annotators?#### Personal and Sensitive Information## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Dataset Card Authors [optional]## Dataset Card Contact"
]
|
8e4ea7e31846b7bd34d099794b413aa11ff0d291 |
This dataset has an accompanying paper "Introducing a novel dataset for product matching: A new challenge for matching systems" that is accepted at The 3rd International Conference on Computers and Automation (CompAuto 2023) and will be published in IEEE Xplore.
The structure of the dataset is as follows: Each data point consists of a pair products and a binary label that indicates if these two product refer to the same real-world entity.
It consists of four subsets that differ in size and class distribution:
| Dataset |Data points | Negative | Positive | Imbalance Ratio |
|---|---:|---:|---:|---:|
| Full | 960,532| 665,831 | 294,701 | 2.3 |
| L | 243,954| 199,749 | 44,205 | 4.5 |
| M |66,556 | 59,925 | 6,631 | 9.0 |
| S | 18,973 |17,978 | 995 | 18.1 |
The test set consists of 5,000 manually checked data points and is shared across all four subsets.
| trettenmeier/markt-pilot | [
"task_categories:text-classification",
"size_categories:100K<n<1M",
"language:de",
"language:en",
"license:cc-by-sa-4.0",
"entity resolution",
"product matching",
"region:us"
]
| 2023-10-27T06:53:40+00:00 | {"language": ["de", "en"], "license": "cc-by-sa-4.0", "size_categories": ["100K<n<1M"], "task_categories": ["text-classification"], "pretty_name": "Markt-Pilot Dataset for Product Matching", "tags": ["entity resolution", "product matching"]} | 2023-10-27T07:30:11+00:00 | []
| [
"de",
"en"
]
| TAGS
#task_categories-text-classification #size_categories-100K<n<1M #language-German #language-English #license-cc-by-sa-4.0 #entity resolution #product matching #region-us
| This dataset has an accompanying paper "Introducing a novel dataset for product matching: A new challenge for matching systems" that is accepted at The 3rd International Conference on Computers and Automation (CompAuto 2023) and will be published in IEEE Xplore.
The structure of the dataset is as follows: Each data point consists of a pair products and a binary label that indicates if these two product refer to the same real-world entity.
It consists of four subsets that differ in size and class distribution:
The test set consists of 5,000 manually checked data points and is shared across all four subsets.
| []
| [
"TAGS\n#task_categories-text-classification #size_categories-100K<n<1M #language-German #language-English #license-cc-by-sa-4.0 #entity resolution #product matching #region-us \n"
]
| [
56
]
| [
"passage: TAGS\n#task_categories-text-classification #size_categories-100K<n<1M #language-German #language-English #license-cc-by-sa-4.0 #entity resolution #product matching #region-us \n"
]
|
60770d00c01cac62ddd0c984eabab9c1b839a29f | Japanese translated file from Vicuna MT bench question.jsonl
https://github.com/lm-sys/FastChat/blob/main/fastchat/llm_judge/data/mt_bench/question.jsonl
fixed by npaka
https://note.com/npaka/n/na28f31e96599 | shi3z/MTbenchJapanese | [
"license:mit",
"region:us"
]
| 2023-10-27T06:58:07+00:00 | {"license": "mit"} | 2023-10-28T04:43:54+00:00 | []
| []
| TAGS
#license-mit #region-us
| Japanese translated file from Vicuna MT bench URL
URL
fixed by npaka
URL | []
| [
"TAGS\n#license-mit #region-us \n"
]
| [
11
]
| [
"passage: TAGS\n#license-mit #region-us \n"
]
|
6830bd2b5ba437cc1187a306d93885056efcf5cd | # Dataset Card for "Bitcoin Tweets ("
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Dataset Distribution](#dataset-distribution)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
### Dataset Summary
This dataset contains a collection of 16 million tweets related to Bitcoin, collected from Twitter. Each tweet is tagged with sentiment (positive, negative, neutral). The dataset was originally created and uploaded to Kaggle by user gauravduttakiit. It is a valuable resource for training and evaluating models for sentiment analysis within the context of cryptocurrency discussions.
### Supported Tasks and Leaderboards
- `text-classification`: This dataset can be used to train a model for sentiment analysis. The performance of the model can be evaluated using standard metrics like accuracy, F1 score, precision, and recall.
### Languages
The text data is primarily in English.
## Dataset Structure
### Data Instances
Each instance in the dataset contains the following fields:
- `tweet`: the text of the tweet.
- `sentiment`: the sentiment of the tweet, labeled as either "positive", "negative", or "neutral".
### Data Fields
- `tweet`: a string containing the text of the tweet.
- `sentiment`: a string indicating the sentiment of the tweet.
### Data Splits
The dataset is not explicitly split into training, validation, and test sets. Users will need to create these splits as per their requirements.
## Dataset Creation
### Curation Rationale
The dataset was curated to analyze the sentiment within the cryptocurrency community, specifically focusing on Bitcoin.
### Source Data
#### Initial Data Collection and Normalization
The data was collected from Twitter using specific keywords related to Bitcoin. For more details regarding data collection, one can refer to the [original Kaggle dataset](https://www.kaggle.com/datasets/gauravduttakiit/bitcoin-tweets-16m-tweets-with-sentiment-tagged).
#### Who are the source data providers?
The data was provided by Kaggle user gauravduttakiit.
### Annotations
The sentiment labels were generated using automated sentiment analysis tools. For more details, refer to the [original Kaggle dataset](https://www.kaggle.com/datasets/gauravduttakiit/bitcoin-tweets-16m-tweets-with-sentiment-tagged).
## Dataset Distribution
### Dataset Curators
The dataset was curated by gauravduttakiit and uploaded to Kaggle.
### Licensing Information
Refer to the [original Kaggle dataset](https://www.kaggle.com/datasets/gauravduttakiit/bitcoin-tweets-16m-tweets-with-sentiment-tagged) for licensing information. | ckandemir/bitcoin_tweets_sentiment_kaggle | [
"task_categories:text-classification",
"task_ids:sentiment-classification",
"multilinguality:multilingual",
"size_categories:100K<n<1M",
"source_datasets:kaggle",
"datasets",
"bitcoin",
"text-classification",
"sentiment-analysis",
"region:us"
]
| 2023-10-27T07:01:02+00:00 | {"multilinguality": ["multilingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["kaggle"], "task_categories": ["text-classification"], "task_ids": ["sentiment-classification"], "tags": ["datasets", "bitcoin", "text-classification", "sentiment-analysis"], "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}, {"split": "test", "path": "data/test-*"}, {"split": "eval", "path": "data/eval-*"}]}], "dataset_info": {"features": [{"name": "Date", "dtype": "string"}, {"name": "text", "dtype": "string"}, {"name": "Sentiment", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 12842246, "num_examples": 77791}, {"name": "test", "num_bytes": 1609120, "num_examples": 9724}, {"name": "eval", "num_bytes": 1598297, "num_examples": 9724}], "download_size": 9868625, "dataset_size": 16049663}} | 2023-11-06T07:25:54+00:00 | []
| []
| TAGS
#task_categories-text-classification #task_ids-sentiment-classification #multilinguality-multilingual #size_categories-100K<n<1M #source_datasets-kaggle #datasets #bitcoin #text-classification #sentiment-analysis #region-us
| # Dataset Card for "Bitcoin Tweets ("
## Table of Contents
- Dataset Description
- Dataset Summary
- Supported Tasks and Leaderboards
- Languages
- Dataset Structure
- Data Instances
- Data Fields
- Data Splits
- Dataset Creation
- Curation Rationale
- Source Data
- Annotations
- Dataset Distribution
- Dataset Curators
- Licensing Information
### Dataset Summary
This dataset contains a collection of 16 million tweets related to Bitcoin, collected from Twitter. Each tweet is tagged with sentiment (positive, negative, neutral). The dataset was originally created and uploaded to Kaggle by user gauravduttakiit. It is a valuable resource for training and evaluating models for sentiment analysis within the context of cryptocurrency discussions.
### Supported Tasks and Leaderboards
- 'text-classification': This dataset can be used to train a model for sentiment analysis. The performance of the model can be evaluated using standard metrics like accuracy, F1 score, precision, and recall.
### Languages
The text data is primarily in English.
## Dataset Structure
### Data Instances
Each instance in the dataset contains the following fields:
- 'tweet': the text of the tweet.
- 'sentiment': the sentiment of the tweet, labeled as either "positive", "negative", or "neutral".
### Data Fields
- 'tweet': a string containing the text of the tweet.
- 'sentiment': a string indicating the sentiment of the tweet.
### Data Splits
The dataset is not explicitly split into training, validation, and test sets. Users will need to create these splits as per their requirements.
## Dataset Creation
### Curation Rationale
The dataset was curated to analyze the sentiment within the cryptocurrency community, specifically focusing on Bitcoin.
### Source Data
#### Initial Data Collection and Normalization
The data was collected from Twitter using specific keywords related to Bitcoin. For more details regarding data collection, one can refer to the original Kaggle dataset.
#### Who are the source data providers?
The data was provided by Kaggle user gauravduttakiit.
### Annotations
The sentiment labels were generated using automated sentiment analysis tools. For more details, refer to the original Kaggle dataset.
## Dataset Distribution
### Dataset Curators
The dataset was curated by gauravduttakiit and uploaded to Kaggle.
### Licensing Information
Refer to the original Kaggle dataset for licensing information. | [
"# Dataset Card for \"Bitcoin Tweets (\"",
"## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n- Dataset Distribution\n - Dataset Curators\n - Licensing Information",
"### Dataset Summary\n\nThis dataset contains a collection of 16 million tweets related to Bitcoin, collected from Twitter. Each tweet is tagged with sentiment (positive, negative, neutral). The dataset was originally created and uploaded to Kaggle by user gauravduttakiit. It is a valuable resource for training and evaluating models for sentiment analysis within the context of cryptocurrency discussions.",
"### Supported Tasks and Leaderboards\n\n- 'text-classification': This dataset can be used to train a model for sentiment analysis. The performance of the model can be evaluated using standard metrics like accuracy, F1 score, precision, and recall.",
"### Languages\n\nThe text data is primarily in English.",
"## Dataset Structure",
"### Data Instances\n\nEach instance in the dataset contains the following fields:\n- 'tweet': the text of the tweet.\n- 'sentiment': the sentiment of the tweet, labeled as either \"positive\", \"negative\", or \"neutral\".",
"### Data Fields\n\n- 'tweet': a string containing the text of the tweet.\n- 'sentiment': a string indicating the sentiment of the tweet.",
"### Data Splits\n\nThe dataset is not explicitly split into training, validation, and test sets. Users will need to create these splits as per their requirements.",
"## Dataset Creation",
"### Curation Rationale\n\nThe dataset was curated to analyze the sentiment within the cryptocurrency community, specifically focusing on Bitcoin.",
"### Source Data",
"#### Initial Data Collection and Normalization\n\nThe data was collected from Twitter using specific keywords related to Bitcoin. For more details regarding data collection, one can refer to the original Kaggle dataset.",
"#### Who are the source data providers?\n\nThe data was provided by Kaggle user gauravduttakiit.",
"### Annotations\n\nThe sentiment labels were generated using automated sentiment analysis tools. For more details, refer to the original Kaggle dataset.",
"## Dataset Distribution",
"### Dataset Curators\n\nThe dataset was curated by gauravduttakiit and uploaded to Kaggle.",
"### Licensing Information\n\nRefer to the original Kaggle dataset for licensing information."
]
| [
"TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #multilinguality-multilingual #size_categories-100K<n<1M #source_datasets-kaggle #datasets #bitcoin #text-classification #sentiment-analysis #region-us \n",
"# Dataset Card for \"Bitcoin Tweets (\"",
"## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n- Dataset Distribution\n - Dataset Curators\n - Licensing Information",
"### Dataset Summary\n\nThis dataset contains a collection of 16 million tweets related to Bitcoin, collected from Twitter. Each tweet is tagged with sentiment (positive, negative, neutral). The dataset was originally created and uploaded to Kaggle by user gauravduttakiit. It is a valuable resource for training and evaluating models for sentiment analysis within the context of cryptocurrency discussions.",
"### Supported Tasks and Leaderboards\n\n- 'text-classification': This dataset can be used to train a model for sentiment analysis. The performance of the model can be evaluated using standard metrics like accuracy, F1 score, precision, and recall.",
"### Languages\n\nThe text data is primarily in English.",
"## Dataset Structure",
"### Data Instances\n\nEach instance in the dataset contains the following fields:\n- 'tweet': the text of the tweet.\n- 'sentiment': the sentiment of the tweet, labeled as either \"positive\", \"negative\", or \"neutral\".",
"### Data Fields\n\n- 'tweet': a string containing the text of the tweet.\n- 'sentiment': a string indicating the sentiment of the tweet.",
"### Data Splits\n\nThe dataset is not explicitly split into training, validation, and test sets. Users will need to create these splits as per their requirements.",
"## Dataset Creation",
"### Curation Rationale\n\nThe dataset was curated to analyze the sentiment within the cryptocurrency community, specifically focusing on Bitcoin.",
"### Source Data",
"#### Initial Data Collection and Normalization\n\nThe data was collected from Twitter using specific keywords related to Bitcoin. For more details regarding data collection, one can refer to the original Kaggle dataset.",
"#### Who are the source data providers?\n\nThe data was provided by Kaggle user gauravduttakiit.",
"### Annotations\n\nThe sentiment labels were generated using automated sentiment analysis tools. For more details, refer to the original Kaggle dataset.",
"## Dataset Distribution",
"### Dataset Curators\n\nThe dataset was curated by gauravduttakiit and uploaded to Kaggle.",
"### Licensing Information\n\nRefer to the original Kaggle dataset for licensing information."
]
| [
76,
13,
78,
84,
60,
13,
6,
58,
36,
38,
5,
30,
4,
43,
24,
32,
5,
25,
20
]
| [
"passage: TAGS\n#task_categories-text-classification #task_ids-sentiment-classification #multilinguality-multilingual #size_categories-100K<n<1M #source_datasets-kaggle #datasets #bitcoin #text-classification #sentiment-analysis #region-us \n# Dataset Card for \"Bitcoin Tweets (\"## Table of Contents\n- Dataset Description\n - Dataset Summary\n - Supported Tasks and Leaderboards\n - Languages\n- Dataset Structure\n - Data Instances\n - Data Fields\n - Data Splits\n- Dataset Creation\n - Curation Rationale\n - Source Data\n - Annotations\n- Dataset Distribution\n - Dataset Curators\n - Licensing Information### Dataset Summary\n\nThis dataset contains a collection of 16 million tweets related to Bitcoin, collected from Twitter. Each tweet is tagged with sentiment (positive, negative, neutral). The dataset was originally created and uploaded to Kaggle by user gauravduttakiit. It is a valuable resource for training and evaluating models for sentiment analysis within the context of cryptocurrency discussions.### Supported Tasks and Leaderboards\n\n- 'text-classification': This dataset can be used to train a model for sentiment analysis. The performance of the model can be evaluated using standard metrics like accuracy, F1 score, precision, and recall.### Languages\n\nThe text data is primarily in English.## Dataset Structure### Data Instances\n\nEach instance in the dataset contains the following fields:\n- 'tweet': the text of the tweet.\n- 'sentiment': the sentiment of the tweet, labeled as either \"positive\", \"negative\", or \"neutral\".### Data Fields\n\n- 'tweet': a string containing the text of the tweet.\n- 'sentiment': a string indicating the sentiment of the tweet.### Data Splits\n\nThe dataset is not explicitly split into training, validation, and test sets. Users will need to create these splits as per their requirements.## Dataset Creation### Curation Rationale\n\nThe dataset was curated to analyze the sentiment within the cryptocurrency community, specifically focusing on Bitcoin.### Source Data"
]
|
2c5a7de688dbe3cf29fb0fcdb376be2242fadb02 | # Dataset Card for "relabel_SciERC"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | anlp/relabel_SciERC | [
"region:us"
]
| 2023-10-27T07:13:15+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}], "dataset_info": {"features": [{"name": "sentences", "sequence": "string"}, {"name": "ner_tags", "sequence": "string"}, {"name": "predict", "sequence": "string"}, {"name": "new_gt", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 2267323, "num_examples": 3238}], "download_size": 312123, "dataset_size": 2267323}} | 2023-10-27T17:37:16+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "relabel_SciERC"
More Information needed | [
"# Dataset Card for \"relabel_SciERC\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"relabel_SciERC\"\n\nMore Information needed"
]
| [
6,
16
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"relabel_SciERC\"\n\nMore Information needed"
]
|
5787baaae24a33baee4cf57a4cc743e8a02cd548 | # Bamboogle
This repo contains the data for ["Measuring and Narrowing the Compositionality Gap in Language Models" paper](https://arxiv.org/abs/2210.03350).
The original data link is here: https://docs.google.com/spreadsheets/d/1jwcsA5kE4TObr9YHn9Gc-wQHYjTbLhDGx6tmIzMhl_U/edit?usp=sharing
This dataset is distributed with the MIT license. | chiayewken/bamboogle | [
"arxiv:2210.03350",
"region:us"
]
| 2023-10-27T07:15:25+00:00 | {"dataset_info": {"features": [{"name": "Question", "dtype": "string"}, {"name": "Answer", "dtype": "string"}], "splits": [{"name": "test", "num_bytes": 10747, "num_examples": 125}], "download_size": 8383, "dataset_size": 10747}, "configs": [{"config_name": "default", "data_files": [{"split": "test", "path": "data/test-*"}]}]} | 2023-10-27T08:22:40+00:00 | [
"2210.03350"
]
| []
| TAGS
#arxiv-2210.03350 #region-us
| # Bamboogle
This repo contains the data for "Measuring and Narrowing the Compositionality Gap in Language Models" paper.
The original data link is here: URL
This dataset is distributed with the MIT license. | [
"# Bamboogle\n\nThis repo contains the data for \"Measuring and Narrowing the Compositionality Gap in Language Models\" paper.\n\nThe original data link is here: URL\n\nThis dataset is distributed with the MIT license."
]
| [
"TAGS\n#arxiv-2210.03350 #region-us \n",
"# Bamboogle\n\nThis repo contains the data for \"Measuring and Narrowing the Compositionality Gap in Language Models\" paper.\n\nThe original data link is here: URL\n\nThis dataset is distributed with the MIT license."
]
| [
15,
51
]
| [
"passage: TAGS\n#arxiv-2210.03350 #region-us \n# Bamboogle\n\nThis repo contains the data for \"Measuring and Narrowing the Compositionality Gap in Language Models\" paper.\n\nThe original data link is here: URL\n\nThis dataset is distributed with the MIT license."
]
|
bb30327f7f48c9372092599ac99ee25cc270e2e4 | # Dataset Card for "sentence_w_elimination"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | anlp/sentence_w_elimination | [
"region:us"
]
| 2023-10-27T07:22:29+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}], "dataset_info": {"features": [{"name": "sentences", "sequence": "string"}, {"name": "new_gt", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 1201528, "num_examples": 990}], "download_size": 244599, "dataset_size": 1201528}} | 2023-10-27T07:43:53+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "sentence_w_elimination"
More Information needed | [
"# Dataset Card for \"sentence_w_elimination\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"sentence_w_elimination\"\n\nMore Information needed"
]
| [
6,
17
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"sentence_w_elimination\"\n\nMore Information needed"
]
|
a6cb4b7f04ab331af96a71654c54fe6e49d2313b |
# This datasets is contains reviews on movies on IMDB
## Columns include :
- review
- sentiment
# what can we do with this datasets
- perform fine tuning using your preferred models
- text -generation
# More rows and column might be added | damerajee/IMDB-sentiment-reviews | [
"task_categories:text-classification",
"size_categories:10K<n<100K",
"language:en",
"license:mit",
"art",
"region:us"
]
| 2023-10-27T07:23:26+00:00 | {"language": ["en"], "license": "mit", "size_categories": ["10K<n<100K"], "task_categories": ["text-classification"], "pretty_name": "IMDB -reviews-Sentiment", "tags": ["art"]} | 2023-10-27T07:28:26+00:00 | []
| [
"en"
]
| TAGS
#task_categories-text-classification #size_categories-10K<n<100K #language-English #license-mit #art #region-us
|
# This datasets is contains reviews on movies on IMDB
## Columns include :
- review
- sentiment
# what can we do with this datasets
- perform fine tuning using your preferred models
- text -generation
# More rows and column might be added | [
"# This datasets is contains reviews on movies on IMDB",
"## Columns include :\n\n- review\n- sentiment",
"# what can we do with this datasets\n\n- perform fine tuning using your preferred models\n- text -generation",
"# More rows and column might be added"
]
| [
"TAGS\n#task_categories-text-classification #size_categories-10K<n<100K #language-English #license-mit #art #region-us \n",
"# This datasets is contains reviews on movies on IMDB",
"## Columns include :\n\n- review\n- sentiment",
"# what can we do with this datasets\n\n- perform fine tuning using your preferred models\n- text -generation",
"# More rows and column might be added"
]
| [
40,
14,
9,
25,
12
]
| [
"passage: TAGS\n#task_categories-text-classification #size_categories-10K<n<100K #language-English #license-mit #art #region-us \n# This datasets is contains reviews on movies on IMDB## Columns include :\n\n- review\n- sentiment# what can we do with this datasets\n\n- perform fine tuning using your preferred models\n- text -generation# More rows and column might be added"
]
|
b6577c661b17e500e322b88f6aec6a83a32b1874 | # Dataset Card for "superb_ks"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | Codec-SUPERB/superb_ks | [
"region:us"
]
| 2023-10-27T07:42:59+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "original", "path": "data/original-*"}, {"split": "descript_audio_codec", "path": "data/descript_audio_codec-*"}, {"split": "encodec_hf", "path": "data/encodec_hf-*"}, {"split": "speech_tokenizer", "path": "data/speech_tokenizer-*"}]}], "dataset_info": {"features": [{"name": "audio", "dtype": {"audio": {"sampling_rate": 16000}}}, {"name": "id", "dtype": "string"}], "splits": [{"name": "original", "num_bytes": 98824867.676, "num_examples": 3081}, {"name": "descript_audio_codec", "num_bytes": 272081821.676, "num_examples": 3081}, {"name": "encodec_hf", "num_bytes": 148225621.676, "num_examples": 3081}, {"name": "speech_tokenizer", "num_bytes": 98929621.676, "num_examples": 3081}], "download_size": 544447448, "dataset_size": 618061932.704}} | 2023-10-27T07:43:49+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "superb_ks"
More Information needed | [
"# Dataset Card for \"superb_ks\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"superb_ks\"\n\nMore Information needed"
]
| [
6,
14
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"superb_ks\"\n\nMore Information needed"
]
|
555eeb8ef70ef5fed19adae053c622eb5dc9bf2c | # Dataset Card for "sentence_anno"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | anlp/sentence_anno | [
"region:us"
]
| 2023-10-27T07:49:00+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}], "dataset_info": {"features": [{"name": "sentences", "sequence": "string"}, {"name": "new_gt", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 1201528, "num_examples": 990}], "download_size": 244599, "dataset_size": 1201528}} | 2023-10-27T07:49:02+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "sentence_anno"
More Information needed | [
"# Dataset Card for \"sentence_anno\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"sentence_anno\"\n\nMore Information needed"
]
| [
6,
14
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"sentence_anno\"\n\nMore Information needed"
]
|
16f7d72e24cad8b9a64e48d424d410bdd357168c |
# Dataset Card for JudgeLM
## Dataset Description
- **Repository:** https://github.com/baaivision/JudgeLM
- **Paper:** https://arxiv.org/abs/2310.17631
### Dataset Summary
JudgeLM-100K dataset contains 100,000 judge samples for training and 5,000 judge samples for validation. All the judge samples have the GPT-4-generated high-quality judgements.
This instruction data can be used to conduct instruction-tuning for language models and make the language model has ability to judge open-ended answer pairs.
See more details in the "Dataset" section and the appendix sections of this [paper](https://arxiv.org/abs/2310.17631).
This produced a judge dataset for instruction-following with 100K examples, obtained at a cost of about $4,000.
### Supported Tasks and Leaderboards
The JudgeLM-100K dataset designed for fine-tuning large language models as judges.
### Languages
The data in JudgeLM-100K are in English.
## Dataset Structure
### Data Instances
An example of "train" looks as follows:
```json
{
"review_id": "HZKJaPyrQSdPyxiz2hUZx9",
"question_id": 0,
"answer1_id": "",
"answer2_id": "",
"reviewer_id": 1,
"metadata": {},
"text": "2 8\nAssistant 1's response is not very helpful or detailed. It simply suggests a website for professional tattoo advice, but does not provide any suggestions or guidance on what kind of design the user might consider. On the other hand, Assistant 2's response is much more detailed and helpful. It provides several suggestions for tattoo designs, such as something meaningful, something based on personal interests, or something based on the user's own body. It also advises the user to get a tattoo they like and to seek help from a reputable tattoo artist. The response is relevant, accurate, and provides a good level of detail.",
"score": [2.0, 8.0],
"question_body": "\nI want to get a tattoo but I'm not sure what kind of design to get. What do you suggest?",
"answer1_body": "myDesigner.com\n\nA:\n\n I suggest to ask for professional tattoo advice on Inkchat.com",
"answer2_body": "1. Something meaningful. There are so many ways to go on this.\n2. Something based on personal interests. Something you love, something funny, something you do for fun.\n3. Something based on your own body. The idea of having a tattoo that covers part of your skin is, well, a little unappealing. If you're open to more creative solutions, and you don't mind having a tattoo all over your body, you can go for a design that is meant to be covered up.\nI wouldn't suggest getting a tattoo just to cover up a bad tattoo or because it would be a good story. Get a tattoo you like, that you would love to look at a hundred years from now as a reminder of a time and place in your life.I\u2019m sure you know this, but if you don\u2019t, you don\u2019t have to go at it alone. You have a whole army of people who can help you.\nThe best way to avoid a bad tattoo is to get a lot of help. The most important thing is to find a reputable tattoo artist. Tattoo artists are like any other skilled professional. You want someone who has a lot of experience and training.\nThe best way to",
"answer1_model_id": "dolly-v2-12b",
"answer2_model_id": "mpt-7b",
"answer1_metadata": {"decoding_method": "top_p_sampling"},
"answer2_metadata": {"decoding_method": "top_p_sampling"},
"reference": {
"decoding_method": "None",
"model": "unified_chip2/83622",
"text": "As an AI language model, I cannot recommend any specific tattoo designs. However, I recommend researching various designs and styles to find one that resonates with you and has a personal meaning. It's also important to find a reputable tattoo artist that you trust and feel comfortable with. Best of luck in your tattoo journey!",
"scores": {}
},
"score_w_reference": [3.0, 7.0],
"text_w_reference": "3 7\n\nAssistant 1's response was not very helpful or detailed. It simply suggested a website for professional tattoo advice without providing any context or explanation. The assistant did not address the user's uncertainty about the kind of design to get, which was the main point of the question. \n\nOn the other hand, Assistant 2's response was more detailed and relevant. It provided several suggestions for the user to consider when choosing a tattoo design, such as choosing something meaningful or based on personal interests. It also emphasized the importance of finding a reputable tattoo artist, which aligns with the reference answer. However, the response was a bit rambling and could have been more concise and organized."}
```
### Data Fields
The main data fields are as follows:
* `question_id`: describes the index of seed questions.
* `question_body`: the text of seed questions.
* `answer1_body` / `answer2_body`: the text of answer that generated by LLMs.
* `answer1_model_id` / `answer2_model_id`: the name of LLM that produced this answer.
* `answer1_metadata` / `answer1_metadata`: some metadata informations of answers, including `decoding_method`.
* `text`: the judgements produced by the GPT-4 teacher.
* `score`: the structured score of GPT-4-generated Judgements.
* `reference`: the reference answer that could be used in judging, which could provide extra knowledge or a specific preference.
* `text_w_reference`: the GPT-4-generated Judgements when given `reference` as the reference answer.
- `score_w_reference`: the structured score of `text_w_reference`.
### Data Splits
| | train | val (w/o reference) | val (w/ reference) |
|---------------|------:|----:|----:|
| JudgeLM-100K | 99647 | 4849 | 4942 |
## Additional Information
### Licensing Information
The dataset is available under the [Creative Commons NonCommercial (CC BY-NC 4.0)](https://creativecommons.org/licenses/by-nc/4.0/legalcode).
### Citation Information
```
@article{zhu2023judgelm,
title={JudgeLM: Fine-tuned Large Language Models are Scalable Judges},
author={Lianghui Zhu and Xinggang Wang and Xinlong Wang},
year={2023},
eprint={2310.17631},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
| BAAI/JudgeLM-100K | [
"task_categories:text-generation",
"language:en",
"license:cc-by-nc-4.0",
"instruction-finetuning",
"arxiv:2310.17631",
"region:us"
]
| 2023-10-27T07:54:35+00:00 | {"language": ["en"], "license": "cc-by-nc-4.0", "task_categories": ["text-generation"], "pretty_name": "JudgeLM-100K", "tags": ["instruction-finetuning"]} | 2023-10-27T08:42:09+00:00 | [
"2310.17631"
]
| [
"en"
]
| TAGS
#task_categories-text-generation #language-English #license-cc-by-nc-4.0 #instruction-finetuning #arxiv-2310.17631 #region-us
| Dataset Card for JudgeLM
========================
Dataset Description
-------------------
* Repository: URL
* Paper: URL
### Dataset Summary
JudgeLM-100K dataset contains 100,000 judge samples for training and 5,000 judge samples for validation. All the judge samples have the GPT-4-generated high-quality judgements.
This instruction data can be used to conduct instruction-tuning for language models and make the language model has ability to judge open-ended answer pairs.
See more details in the "Dataset" section and the appendix sections of this paper.
This produced a judge dataset for instruction-following with 100K examples, obtained at a cost of about $4,000.
### Supported Tasks and Leaderboards
The JudgeLM-100K dataset designed for fine-tuning large language models as judges.
### Languages
The data in JudgeLM-100K are in English.
Dataset Structure
-----------------
### Data Instances
An example of "train" looks as follows:
### Data Fields
The main data fields are as follows:
* 'question\_id': describes the index of seed questions.
* 'question\_body': the text of seed questions.
* 'answer1\_body' / 'answer2\_body': the text of answer that generated by LLMs.
* 'answer1\_model\_id' / 'answer2\_model\_id': the name of LLM that produced this answer.
* 'answer1\_metadata' / 'answer1\_metadata': some metadata informations of answers, including 'decoding\_method'.
* 'text': the judgements produced by the GPT-4 teacher.
* 'score': the structured score of GPT-4-generated Judgements.
* 'reference': the reference answer that could be used in judging, which could provide extra knowledge or a specific preference.
* 'text\_w\_reference': the GPT-4-generated Judgements when given 'reference' as the reference answer.
* 'score\_w\_reference': the structured score of 'text\_w\_reference'.
### Data Splits
Additional Information
----------------------
### Licensing Information
The dataset is available under the Creative Commons NonCommercial (CC BY-NC 4.0).
| [
"### Dataset Summary\n\n\nJudgeLM-100K dataset contains 100,000 judge samples for training and 5,000 judge samples for validation. All the judge samples have the GPT-4-generated high-quality judgements.\nThis instruction data can be used to conduct instruction-tuning for language models and make the language model has ability to judge open-ended answer pairs.\n\n\nSee more details in the \"Dataset\" section and the appendix sections of this paper.\n\n\nThis produced a judge dataset for instruction-following with 100K examples, obtained at a cost of about $4,000.",
"### Supported Tasks and Leaderboards\n\n\nThe JudgeLM-100K dataset designed for fine-tuning large language models as judges.",
"### Languages\n\n\nThe data in JudgeLM-100K are in English.\n\n\nDataset Structure\n-----------------",
"### Data Instances\n\n\nAn example of \"train\" looks as follows:",
"### Data Fields\n\n\nThe main data fields are as follows:\n\n\n* 'question\\_id': describes the index of seed questions.\n* 'question\\_body': the text of seed questions.\n* 'answer1\\_body' / 'answer2\\_body': the text of answer that generated by LLMs.\n* 'answer1\\_model\\_id' / 'answer2\\_model\\_id': the name of LLM that produced this answer.\n* 'answer1\\_metadata' / 'answer1\\_metadata': some metadata informations of answers, including 'decoding\\_method'.\n* 'text': the judgements produced by the GPT-4 teacher.\n* 'score': the structured score of GPT-4-generated Judgements.\n* 'reference': the reference answer that could be used in judging, which could provide extra knowledge or a specific preference.\n* 'text\\_w\\_reference': the GPT-4-generated Judgements when given 'reference' as the reference answer.\n\n\n* 'score\\_w\\_reference': the structured score of 'text\\_w\\_reference'.",
"### Data Splits\n\n\n\nAdditional Information\n----------------------",
"### Licensing Information\n\n\nThe dataset is available under the Creative Commons NonCommercial (CC BY-NC 4.0)."
]
| [
"TAGS\n#task_categories-text-generation #language-English #license-cc-by-nc-4.0 #instruction-finetuning #arxiv-2310.17631 #region-us \n",
"### Dataset Summary\n\n\nJudgeLM-100K dataset contains 100,000 judge samples for training and 5,000 judge samples for validation. All the judge samples have the GPT-4-generated high-quality judgements.\nThis instruction data can be used to conduct instruction-tuning for language models and make the language model has ability to judge open-ended answer pairs.\n\n\nSee more details in the \"Dataset\" section and the appendix sections of this paper.\n\n\nThis produced a judge dataset for instruction-following with 100K examples, obtained at a cost of about $4,000.",
"### Supported Tasks and Leaderboards\n\n\nThe JudgeLM-100K dataset designed for fine-tuning large language models as judges.",
"### Languages\n\n\nThe data in JudgeLM-100K are in English.\n\n\nDataset Structure\n-----------------",
"### Data Instances\n\n\nAn example of \"train\" looks as follows:",
"### Data Fields\n\n\nThe main data fields are as follows:\n\n\n* 'question\\_id': describes the index of seed questions.\n* 'question\\_body': the text of seed questions.\n* 'answer1\\_body' / 'answer2\\_body': the text of answer that generated by LLMs.\n* 'answer1\\_model\\_id' / 'answer2\\_model\\_id': the name of LLM that produced this answer.\n* 'answer1\\_metadata' / 'answer1\\_metadata': some metadata informations of answers, including 'decoding\\_method'.\n* 'text': the judgements produced by the GPT-4 teacher.\n* 'score': the structured score of GPT-4-generated Judgements.\n* 'reference': the reference answer that could be used in judging, which could provide extra knowledge or a specific preference.\n* 'text\\_w\\_reference': the GPT-4-generated Judgements when given 'reference' as the reference answer.\n\n\n* 'score\\_w\\_reference': the structured score of 'text\\_w\\_reference'.",
"### Data Splits\n\n\n\nAdditional Information\n----------------------",
"### Licensing Information\n\n\nThe dataset is available under the Creative Commons NonCommercial (CC BY-NC 4.0)."
]
| [
47,
135,
31,
23,
18,
276,
12,
26
]
| [
"passage: TAGS\n#task_categories-text-generation #language-English #license-cc-by-nc-4.0 #instruction-finetuning #arxiv-2310.17631 #region-us \n### Dataset Summary\n\n\nJudgeLM-100K dataset contains 100,000 judge samples for training and 5,000 judge samples for validation. All the judge samples have the GPT-4-generated high-quality judgements.\nThis instruction data can be used to conduct instruction-tuning for language models and make the language model has ability to judge open-ended answer pairs.\n\n\nSee more details in the \"Dataset\" section and the appendix sections of this paper.\n\n\nThis produced a judge dataset for instruction-following with 100K examples, obtained at a cost of about $4,000.### Supported Tasks and Leaderboards\n\n\nThe JudgeLM-100K dataset designed for fine-tuning large language models as judges.### Languages\n\n\nThe data in JudgeLM-100K are in English.\n\n\nDataset Structure\n-----------------### Data Instances\n\n\nAn example of \"train\" looks as follows:"
]
|
0838e5201ba527c60f4eaeaf79cca2e51bd2c71f | # Dataset Card for "english_learn"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | VuongQuoc/english_learn | [
"region:us"
]
| 2023-10-27T08:05:55+00:00 | {"dataset_info": {"features": [{"name": "image", "dtype": "image"}, {"name": "caption", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 4602747761.0, "num_examples": 77456}], "download_size": 4600511540, "dataset_size": 4602747761.0}} | 2023-10-27T08:09:46+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "english_learn"
More Information needed | [
"# Dataset Card for \"english_learn\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"english_learn\"\n\nMore Information needed"
]
| [
6,
16
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"english_learn\"\n\nMore Information needed"
]
|
7af922a51e188fb05ad076b2b46e0f35ce2dd6a8 | # Dataset Card for "rsna_5k_512_a"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | Phaedrus/rsna_5k_512_a | [
"region:us"
]
| 2023-10-27T08:10:55+00:00 | {"dataset_info": {"features": [{"name": "image", "dtype": "image"}, {"name": "label1", "dtype": "image"}, {"name": "label2", "dtype": "image"}, {"name": "label3", "dtype": "image"}, {"name": "label4", "dtype": "image"}], "splits": [{"name": "train", "num_bytes": 8605017463.0, "num_examples": 2000}], "download_size": 574221474, "dataset_size": 8605017463.0}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-27T08:16:28+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "rsna_5k_512_a"
More Information needed | [
"# Dataset Card for \"rsna_5k_512_a\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"rsna_5k_512_a\"\n\nMore Information needed"
]
| [
6,
20
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"rsna_5k_512_a\"\n\nMore Information needed"
]
|
f8dbba0ba0c4efa185a33545f22f30a6760ca82e | {
"intents": [
{
"tag": "busy",
"pattern": [
"बाहर हु",
"व्यस्त हु",
"बिजी हु",
"बात नई कर सकता",
"व्यस्त",
"बिजी हूँ",
"व्यस्त हूँ",
"मीटिंग में हूँ",
"अभी समय नहीं है",
"बात नहीं कर सकते",
"वर्तमान में कोई फ्री समय नहीं",
"डेडलाइन पर ध्यान केंद्रित",
"पूरी तरह व्यस्त, कोई विश्राम नहीं",
"एक के बाद एक मीटिंगों में व्यस्त",
"रात-दिन काम कर रहा हूँ",
"काम में उलझ रहा हूँ",
"बहुत काम बचा है",
"समय की कमी है",
"वापस कॉल करो",
"रविवार को कॉल करो",
"बात नई कर सकता",
"कॉल करो",
"बदमे कॉल करो",
"थोड़ी देर बाद कॉल करो",
"अभी बात नई कर सकता",
"नहीं"
],
"response": [
"यह कॉल आपके लिए महत्वपूर्ण है। इसे नजरंदाज करने से आपके सीबील स्कोर पर नकारात्मक प्रभाव पड़ेगा"
]
},
{
"tag": "hi",
"pattern": [
"हैलो",
"हई",
"हेलो",
"हेल",
"नमस्ते",
"नमस्कार",
"हैलो",
"हई",
"नमस्ते",
"हेलो",
"हाइ"
],
"response": [
"नमस्कार , सोनाटा फाइनैन्स से मे स्वरा बोल रही हु। हमने <client_name> के लोन की राशी लौटाने के संधरभ मे कॉल किया है। यह कॉल गुणवत्ता एवं ट्रैनिंग केलिए रेकॉर्ड हो राहा है"
]
},
{
"tag": "agent visit date",
"pattern": [
"कल भेज दो",
"परसों भेज दो",
"दो दिन बाद भेजो",
"एक",
"दो",
"तीन",
"चार",
"पांच",
"छे",
"सात",
"आठ",
"जानेवारी",
"फेब्रुएरी",
"मार्च",
"एप्रिल",
"मे",
"जून",
"जुलाई",
"अउगेस्ट",
"सप्टेंबर",
"ऑक्टोबर",
"नवंबर",
"दिसम्बर"
],
"response": [
"धन्यवाद, हमारा एजेंट आपसे <callback_time> पर संपर्क करेगा"
]
},
{
"tag": "Refuse to Pay",
"pattern": [
"नई दे सकता",
"पैसे नई है",
"नई दे पाऊँगा",
"पैसे नई है",
"नई",
"मजबूरी है",
"पैसे काम है",
"प्रयास कर रहा हु",
"नई दे सकता",
"ना",
"नहीं",
"नही",
"नहीं पाऊंगा",
"नहीं कर सकता",
"मत भेजो",
"भेजना मत"
],
"response": [
"हम समजते है की आपको कठिनाई है । लेकिन भूकतान जमा करने मे किसी भी तरीके की देरी आपका सीबील स्कोर खराब कर सकती है । जिससे आपको भविष्य मे लोन मिलने मे कठिनाई होंगी"
]
},
{
"tag": "miscellaneous question",
"pattern": [
"कोनसा लॉन",
"क्या लोन",
"मैंने कब लोन लिया",
"किस लोन की बात कर रहे हो",
"मैंने कोई लोन नई लिया",
"किसका मेम",
"किस चीज का पैसा मेम"
],
"response": [
"ये कॉल आपने जो सोनाटा के साथ लोन लिया था उसके संबंध मे किया गया है । आपकी पिछली <due_installments> किश्त और <emi_amount> रुपए बकाया है।"
]
},
{
"tag": "Agree to Pay",
"pattern": [
"ठीक है",
"दे दूंगा",
"आजाओ",
"दे सकता हु",
"दे दूंगा",
"हा कर दूंगा",
"कर दूंगा",
"कर देता हु",
"हाँ",
"हा कर दूंगा",
"कर दूंगा",
"आज अजाऊँगा ब्रांच",
"ब्रांच मे अजाऊँगा",
"का आ जाऊंगा",
"जी",
"जी बिल्कुल कर देंगे",
"जी कर देंगे आज",
"हाँ कर देंगे",
"ठीक है",
"हा कर सकते हैं",
"जी कर सकते हैं",
"आज नई हो पाएगा",
"आज नई हो सकता",
"आज बिजी हु",
"कल करता हु",
"आज मुश्किल है",
"आज नहीं हो पाएगा कल कर देंगे",
"नहीं हो पाएगा जी आज",
"कल देंगे कल"
],
"response": [
"कृपया आप <emi_amount> तैयार रखें हमारा एजेंट पैसे लेने के लिए आपके पास आएगा,धन्यवाद."
]
},
{
"tag": "insist to pay tomorrow",
"pattern": [
"बाद मे देता हु",
"नहीं कर सकता",
"नई कर",
"नहीं",
"नहीं दे सकता",
"आज नहीं हो पाएगा",
"नहीं कर पाऊंगा",
"नई कर सकता",
"नई हो पाएगा",
"नई हो पाएगा",
"नहीं"
],
"response": [
"कृपया आप <emi_amount> तैयार रखें हमारा एजेंट पैसे लेने के लिए आपके पास आएगा,धन्यवाद."
]
}
]
} | MrDontKnowWhatToDo/Intent_Recognition | [
"region:us"
]
| 2023-10-27T08:19:22+00:00 | {} | 2023-10-27T10:01:55+00:00 | []
| []
| TAGS
#region-us
| {
"intents": [
{
"tag": "busy",
"pattern": [
"बाहर हु",
"व्यस्त हु",
"बिजी हु",
"बात नई कर सकता",
"व्यस्त",
"बिजी हूँ",
"व्यस्त हूँ",
"मीटिंग में हूँ",
"अभी समय नहीं है",
"बात नहीं कर सकते",
"वर्तमान में कोई फ्री समय नहीं",
"डेडलाइन पर ध्यान केंद्रित",
"पूरी तरह व्यस्त, कोई विश्राम नहीं",
"एक के बाद एक मीटिंगों में व्यस्त",
"रात-दिन काम कर रहा हूँ",
"काम में उलझ रहा हूँ",
"बहुत काम बचा है",
"समय की कमी है",
"वापस कॉल करो",
"रविवार को कॉल करो",
"बात नई कर सकता",
"कॉल करो",
"बदमे कॉल करो",
"थोड़ी देर बाद कॉल करो",
"अभी बात नई कर सकता",
"नहीं"
],
"response": [
"यह कॉल आपके लिए महत्वपूर्ण है। इसे नजरंदाज करने से आपके सीबील स्कोर पर नकारात्मक प्रभाव पड़ेगा"
]
},
{
"tag": "hi",
"pattern": [
"हैलो",
"हई",
"हेलो",
"हेल",
"नमस्ते",
"नमस्कार",
"हैलो",
"हई",
"नमस्ते",
"हेलो",
"हाइ"
],
"response": [
"नमस्कार , सोनाटा फाइनैन्स से मे स्वरा बोल रही हु। हमने <client_name> के लोन की राशी लौटाने के संधरभ मे कॉल किया है। यह कॉल गुणवत्ता एवं ट्रैनिंग केलिए रेकॉर्ड हो राहा है"
]
},
{
"tag": "agent visit date",
"pattern": [
"कल भेज दो",
"परसों भेज दो",
"दो दिन बाद भेजो",
"एक",
"दो",
"तीन",
"चार",
"पांच",
"छे",
"सात",
"आठ",
"जानेवारी",
"फेब्रुएरी",
"मार्च",
"एप्रिल",
"मे",
"जून",
"जुलाई",
"अउगेस्ट",
"सप्टेंबर",
"ऑक्टोबर",
"नवंबर",
"दिसम्बर"
],
"response": [
"धन्यवाद, हमारा एजेंट आपसे <callback_time> पर संपर्क करेगा"
]
},
{
"tag": "Refuse to Pay",
"pattern": [
"नई दे सकता",
"पैसे नई है",
"नई दे पाऊँगा",
"पैसे नई है",
"नई",
"मजबूरी है",
"पैसे काम है",
"प्रयास कर रहा हु",
"नई दे सकता",
"ना",
"नहीं",
"नही",
"नहीं पाऊंगा",
"नहीं कर सकता",
"मत भेजो",
"भेजना मत"
],
"response": [
"हम समजते है की आपको कठिनाई है । लेकिन भूकतान जमा करने मे किसी भी तरीके की देरी आपका सीबील स्कोर खराब कर सकती है । जिससे आपको भविष्य मे लोन मिलने मे कठिनाई होंगी"
]
},
{
"tag": "miscellaneous question",
"pattern": [
"कोनसा लॉन",
"क्या लोन",
"मैंने कब लोन लिया",
"किस लोन की बात कर रहे हो",
"मैंने कोई लोन नई लिया",
"किसका मेम",
"किस चीज का पैसा मेम"
],
"response": [
"ये कॉल आपने जो सोनाटा के साथ लोन लिया था उसके संबंध मे किया गया है । आपकी पिछली <due_installments> किश्त और <emi_amount> रुपए बकाया है।"
]
},
{
"tag": "Agree to Pay",
"pattern": [
"ठीक है",
"दे दूंगा",
"आजाओ",
"दे सकता हु",
"दे दूंगा",
"हा कर दूंगा",
"कर दूंगा",
"कर देता हु",
"हाँ",
"हा कर दूंगा",
"कर दूंगा",
"आज अजाऊँगा ब्रांच",
"ब्रांच मे अजाऊँगा",
"का आ जाऊंगा",
"जी",
"जी बिल्कुल कर देंगे",
"जी कर देंगे आज",
"हाँ कर देंगे",
"ठीक है",
"हा कर सकते हैं",
"जी कर सकते हैं",
"आज नई हो पाएगा",
"आज नई हो सकता",
"आज बिजी हु",
"कल करता हु",
"आज मुश्किल है",
"आज नहीं हो पाएगा कल कर देंगे",
"नहीं हो पाएगा जी आज",
"कल देंगे कल"
],
"response": [
"कृपया आप <emi_amount> तैयार रखें हमारा एजेंट पैसे लेने के लिए आपके पास आएगा,धन्यवाद."
]
},
{
"tag": "insist to pay tomorrow",
"pattern": [
"बाद मे देता हु",
"नहीं कर सकता",
"नई कर",
"नहीं",
"नहीं दे सकता",
"आज नहीं हो पाएगा",
"नहीं कर पाऊंगा",
"नई कर सकता",
"नई हो पाएगा",
"नई हो पाएगा",
"नहीं"
],
"response": [
"कृपया आप <emi_amount> तैयार रखें हमारा एजेंट पैसे लेने के लिए आपके पास आएगा,धन्यवाद."
]
}
]
} | []
| [
"TAGS\n#region-us \n"
]
| [
6
]
| [
"passage: TAGS\n#region-us \n"
]
|
77a71d54b75a1442ef5d071b6f073c22f0184b71 | # Simsamu dataset
This repository contains recordings of simulated medical dispatch dialogs in the
french language, annotated for diarization and transcription. It is published
under the MIT license.
These dialogs were recorded as part of the training of emergency medicine
interns, which consisted in simulating a medical dispatch call where the interns
took turns playing the caller and the regulating doctor.
Each situation was decided randomly in advance, blind to who was playing the
medical dispatcher (e.g., road accident, chest pain, burns, etc.). The
affiliations between the caller and the patient (family, friend, colleague...)
and the caller's communication mode is then randomly selected. The caller had to
adapt his or her performance to the communication mode associated with the
situation. Seven communication modes were defined: shy, procedural, angry,
cooperative, frightened, impassive, incomprehensible.
Regarding sound quality, the voice of the regulating doctor is directly picked
up by a microphone, whereas the voice of the caller is transmitted through the
phone network and re-emitted by a phone speaker before being picked up by the
microphone. This leads to different acoustic characteristics between the
caller's voice and the regulator's, the later one often being much clearer. This
phenomena is also present in actual dispatch services recordings, where the
regulator's voice is directly recorded in a quiet room whereas the caller is
often calling from noisier environments and its voice is altered by the phone
network compression.
The dataset is composed of 61 audio recordings with a total duration of 3h 15
and an average duration per recording of 3 minutes 11 seconds. Each recording is
available as a `.m4a` audio file with 8KHz sample rate and a 128 Kbps bitrate.
The diarization data is available in a corresponding `.rttm` file and the
transcription in an `.srt` file.
An additional `metadata.csv` contains speaker ids for callers and regulators in
each recording.
See also: [Simsamu diarization
pipeline](https://huggingface.co/medkit/simsamu-diarization)
See also: [Simsamu transcription
model](https://huggingface.co/medkit/simsamu-transcription)
| medkit/simsamu | [
"task_categories:automatic-speech-recognition",
"task_categories:voice-activity-detection",
"multilinguality:monolingual",
"language:fr",
"license:mit",
"region:us"
]
| 2023-10-27T08:57:09+00:00 | {"language": "fr", "license": "mit", "multilinguality": "monolingual", "task_categories": ["automatic-speech-recognition", "voice-activity-detection"]} | 2023-12-15T13:16:11+00:00 | []
| [
"fr"
]
| TAGS
#task_categories-automatic-speech-recognition #task_categories-voice-activity-detection #multilinguality-monolingual #language-French #license-mit #region-us
| # Simsamu dataset
This repository contains recordings of simulated medical dispatch dialogs in the
french language, annotated for diarization and transcription. It is published
under the MIT license.
These dialogs were recorded as part of the training of emergency medicine
interns, which consisted in simulating a medical dispatch call where the interns
took turns playing the caller and the regulating doctor.
Each situation was decided randomly in advance, blind to who was playing the
medical dispatcher (e.g., road accident, chest pain, burns, etc.). The
affiliations between the caller and the patient (family, friend, colleague...)
and the caller's communication mode is then randomly selected. The caller had to
adapt his or her performance to the communication mode associated with the
situation. Seven communication modes were defined: shy, procedural, angry,
cooperative, frightened, impassive, incomprehensible.
Regarding sound quality, the voice of the regulating doctor is directly picked
up by a microphone, whereas the voice of the caller is transmitted through the
phone network and re-emitted by a phone speaker before being picked up by the
microphone. This leads to different acoustic characteristics between the
caller's voice and the regulator's, the later one often being much clearer. This
phenomena is also present in actual dispatch services recordings, where the
regulator's voice is directly recorded in a quiet room whereas the caller is
often calling from noisier environments and its voice is altered by the phone
network compression.
The dataset is composed of 61 audio recordings with a total duration of 3h 15
and an average duration per recording of 3 minutes 11 seconds. Each recording is
available as a '.m4a' audio file with 8KHz sample rate and a 128 Kbps bitrate.
The diarization data is available in a corresponding '.rttm' file and the
transcription in an '.srt' file.
An additional 'URL' contains speaker ids for callers and regulators in
each recording.
See also: Simsamu diarization
pipeline
See also: Simsamu transcription
model
| [
"# Simsamu dataset\n\nThis repository contains recordings of simulated medical dispatch dialogs in the\nfrench language, annotated for diarization and transcription. It is published\nunder the MIT license.\n\nThese dialogs were recorded as part of the training of emergency medicine\ninterns, which consisted in simulating a medical dispatch call where the interns\ntook turns playing the caller and the regulating doctor. \n\nEach situation was decided randomly in advance, blind to who was playing the\nmedical dispatcher (e.g., road accident, chest pain, burns, etc.). The\naffiliations between the caller and the patient (family, friend, colleague...)\nand the caller's communication mode is then randomly selected. The caller had to\nadapt his or her performance to the communication mode associated with the\nsituation. Seven communication modes were defined: shy, procedural, angry,\ncooperative, frightened, impassive, incomprehensible. \n\nRegarding sound quality, the voice of the regulating doctor is directly picked\nup by a microphone, whereas the voice of the caller is transmitted through the\nphone network and re-emitted by a phone speaker before being picked up by the\nmicrophone. This leads to different acoustic characteristics between the\ncaller's voice and the regulator's, the later one often being much clearer. This\nphenomena is also present in actual dispatch services recordings, where the\nregulator's voice is directly recorded in a quiet room whereas the caller is\noften calling from noisier environments and its voice is altered by the phone\nnetwork compression.\n\nThe dataset is composed of 61 audio recordings with a total duration of 3h 15\nand an average duration per recording of 3 minutes 11 seconds. Each recording is\navailable as a '.m4a' audio file with 8KHz sample rate and a 128 Kbps bitrate.\nThe diarization data is available in a corresponding '.rttm' file and the\ntranscription in an '.srt' file.\n\nAn additional 'URL' contains speaker ids for callers and regulators in\neach recording.\n\nSee also: Simsamu diarization\npipeline\n\nSee also: Simsamu transcription\nmodel"
]
| [
"TAGS\n#task_categories-automatic-speech-recognition #task_categories-voice-activity-detection #multilinguality-monolingual #language-French #license-mit #region-us \n",
"# Simsamu dataset\n\nThis repository contains recordings of simulated medical dispatch dialogs in the\nfrench language, annotated for diarization and transcription. It is published\nunder the MIT license.\n\nThese dialogs were recorded as part of the training of emergency medicine\ninterns, which consisted in simulating a medical dispatch call where the interns\ntook turns playing the caller and the regulating doctor. \n\nEach situation was decided randomly in advance, blind to who was playing the\nmedical dispatcher (e.g., road accident, chest pain, burns, etc.). The\naffiliations between the caller and the patient (family, friend, colleague...)\nand the caller's communication mode is then randomly selected. The caller had to\nadapt his or her performance to the communication mode associated with the\nsituation. Seven communication modes were defined: shy, procedural, angry,\ncooperative, frightened, impassive, incomprehensible. \n\nRegarding sound quality, the voice of the regulating doctor is directly picked\nup by a microphone, whereas the voice of the caller is transmitted through the\nphone network and re-emitted by a phone speaker before being picked up by the\nmicrophone. This leads to different acoustic characteristics between the\ncaller's voice and the regulator's, the later one often being much clearer. This\nphenomena is also present in actual dispatch services recordings, where the\nregulator's voice is directly recorded in a quiet room whereas the caller is\noften calling from noisier environments and its voice is altered by the phone\nnetwork compression.\n\nThe dataset is composed of 61 audio recordings with a total duration of 3h 15\nand an average duration per recording of 3 minutes 11 seconds. Each recording is\navailable as a '.m4a' audio file with 8KHz sample rate and a 128 Kbps bitrate.\nThe diarization data is available in a corresponding '.rttm' file and the\ntranscription in an '.srt' file.\n\nAn additional 'URL' contains speaker ids for callers and regulators in\neach recording.\n\nSee also: Simsamu diarization\npipeline\n\nSee also: Simsamu transcription\nmodel"
]
| [
56,
494
]
| [
"passage: TAGS\n#task_categories-automatic-speech-recognition #task_categories-voice-activity-detection #multilinguality-monolingual #language-French #license-mit #region-us \n"
]
|
da080600d603671167ab1cbe472d1971dafd54c9 | # Dataset Card for "filtered_lemma41kV0.0.3"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | fia24/filtered_lemma41kV0.0.3 | [
"region:us"
]
| 2023-10-27T09:04:26+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}, {"split": "test", "path": "data/test-*"}, {"split": "val", "path": "data/val-*"}]}], "dataset_info": {"features": [{"name": "Unnamed: 0", "dtype": "int64"}, {"name": "Inflected_Word", "dtype": "string"}, {"name": "Lemma", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 1789172.8, "num_examples": 28460}, {"name": "test", "num_bytes": 223678.03311314125, "num_examples": 3558}, {"name": "val", "num_bytes": 223615.16688685876, "num_examples": 3557}], "download_size": 1196489, "dataset_size": 2236466.0}} | 2023-10-27T09:04:32+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "filtered_lemma41kV0.0.3"
More Information needed | [
"# Dataset Card for \"filtered_lemma41kV0.0.3\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"filtered_lemma41kV0.0.3\"\n\nMore Information needed"
]
| [
6,
21
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"filtered_lemma41kV0.0.3\"\n\nMore Information needed"
]
|
ccfb26f6de1bee9163b59f4e628af291526caca9 | # Dataset Card for "accepted_pairs_50"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | makram93/accepted_pairs_50 | [
"region:us"
]
| 2023-10-27T09:15:10+00:00 | {"dataset_info": {"features": [{"name": "url", "dtype": "string"}, {"name": "doc_id", "dtype": "string"}, {"name": "original_title", "sequence": "string"}, {"name": "right", "dtype": "string"}, {"name": "left", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 88447.0623234648, "num_examples": 100}], "download_size": 78941, "dataset_size": 88447.0623234648}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-27T11:18:50+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "accepted_pairs_50"
More Information needed | [
"# Dataset Card for \"accepted_pairs_50\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"accepted_pairs_50\"\n\nMore Information needed"
]
| [
6,
17
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"accepted_pairs_50\"\n\nMore Information needed"
]
|
0b7abfa3a398c9ad4eef07d3bf239ef5307c7d15 | # Dataset Card for "rejected_pairs_50"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | makram93/rejected_pairs_50 | [
"region:us"
]
| 2023-10-27T09:15:13+00:00 | {"dataset_info": {"features": [{"name": "url", "dtype": "string"}, {"name": "doc_id", "dtype": "string"}, {"name": "original_title", "sequence": "string"}, {"name": "right", "dtype": "string"}, {"name": "left", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 88447.0623234648, "num_examples": 100}], "download_size": 85583, "dataset_size": 88447.0623234648}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-27T11:18:53+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "rejected_pairs_50"
More Information needed | [
"# Dataset Card for \"rejected_pairs_50\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"rejected_pairs_50\"\n\nMore Information needed"
]
| [
6,
17
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"rejected_pairs_50\"\n\nMore Information needed"
]
|
c2939e68454800714ed965e8c1d5f789f86b51ba | # Dataset Card for "autotrain-data-1w6s-u4vt-i7yo"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | Kabatubare/autotrain-data-1w6s-u4vt-i7yo | [
"region:us"
]
| 2023-10-27T09:42:08+00:00 | {"dataset_info": {"features": [{"name": "autotrain_text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 19109937, "num_examples": 23437}, {"name": "validation", "num_bytes": 19109937, "num_examples": 23437}], "download_size": 20605004, "dataset_size": 38219874}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}, {"split": "validation", "path": "data/validation-*"}]}]} | 2023-10-27T09:42:10+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "autotrain-data-1w6s-u4vt-i7yo"
More Information needed | [
"# Dataset Card for \"autotrain-data-1w6s-u4vt-i7yo\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"autotrain-data-1w6s-u4vt-i7yo\"\n\nMore Information needed"
]
| [
6,
27
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"autotrain-data-1w6s-u4vt-i7yo\"\n\nMore Information needed"
]
|
117154eb8c86fb0c18f763f859e4c5df2029777f | # Dataset Card for "rsna_5k_512_b"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | Phaedrus/rsna_5k_512_b | [
"region:us"
]
| 2023-10-27T09:42:12+00:00 | {"dataset_info": {"features": [{"name": "image", "dtype": "image"}, {"name": "label1", "dtype": "image"}, {"name": "label2", "dtype": "image"}, {"name": "label3", "dtype": "image"}, {"name": "label4", "dtype": "image"}], "splits": [{"name": "train", "num_bytes": 8605017463.0, "num_examples": 2000}], "download_size": 549148202, "dataset_size": 8605017463.0}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-27T09:47:31+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "rsna_5k_512_b"
More Information needed | [
"# Dataset Card for \"rsna_5k_512_b\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"rsna_5k_512_b\"\n\nMore Information needed"
]
| [
6,
20
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"rsna_5k_512_b\"\n\nMore Information needed"
]
|
07831747dfcc7394b0193c02dde1df65ef832309 |
# Dataset Card for "Cross-Lingual Data Augmentation For Thai QA"
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Structure](#dataset-structure)
- [Acknowledgements](#acknowledgements)
- [Authors](#authors)
- [Additional Information](#additional-information)
## Dataset Description
### Abstract
This dataset accompanies the paper titled "Cross-Lingual Data Augmentation For Thai Question Answering" by Parinthapat Pengpun, Can Udomcharoenchaikit, Weerayut Buaphet, and Peerat Limkonchotiwat, to be presented at GenBench in EMNLP 2023. The paper introduces an innovative framework for data augmentation with quality control measures, aimed at enhancing the robustness of Thai QA models. This dataset is designed to improve model performance in low-resource language settings like Thai, by increasing linguistic diversity through monolingual and cross-lingual data augmentation techniques.
### Links
- ACL Link: [PDF](https://aclanthology.org/2023.genbench-1.16/)
- ResearchGate Link: [PDF](https://www.researchgate.net/publication/374977605_Cross-Lingual_Data_Augmentation_For_Thai_Question-Answering#fullTextFileContent)
## Dataset Structure
### Dataset Info
The dataset, available at [Hugging Face Datasets](https://huggingface.co/datasets/parinzee/claq-qa-thai-dataset), is structured with the following features:
- `id`: string
- `question`: string
- `context`: string
- `answers`: string
- `source`: string
- Augmentation columns for Thai (e.g., `th_aug`, `th_fasttext_aug`, `th_llm_gec_aug`, etc.)
- Augmentation columns for English (e.g., `en_aug`, `en_llm_gec_aug`, `en_llm_paraphrase_aug`, etc.)
- Semantic distance columns for various augmentations (e.g., `dis_aug`, `dis_fasttext_aug`, `dis_llm_gec_aug`, etc.)
### Splits (No Designated Train/Test Splits)
- Train:
- Number of rows: **16,980**
- Number of augmentation sets: **10**
- Total Number of Examples = 16,980 * 11 = **186,780**
- Size: 117,313,078 bytes
### Download Size
- 35,147,642 bytes
### Total Dataset Size
- 117,313,078 bytes
## Acknowledgements

## Authors
- Parinthapat Pengpun
- Can Udomcharoenchaikit
- Weerayut Buaphet
- Peerat Limkonchotiwat
## Additional Information
- The dataset is intended for research purposes, especially in the field of machine learning and natural language processing.
- This work is a significant contribution to enhancing the capabilities of QA models in Thai, a low-resource language, by addressing the challenges of limited and varied quality training data. | parinzee/claq-qa-thai-dataset | [
"region:us"
]
| 2023-10-27T10:07:58+00:00 | {"dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "question", "dtype": "string"}, {"name": "context", "dtype": "string"}, {"name": "answers", "dtype": "string"}, {"name": "source", "dtype": "string"}, {"name": "th_aug", "dtype": "string"}, {"name": "th_fasttext_aug", "dtype": "string"}, {"name": "th_llm_gec_aug", "dtype": "string"}, {"name": "th_llm_paraphrase_aug", "dtype": "string"}, {"name": "th_ltw2v_aug", "dtype": "string"}, {"name": "th_qcpg_0.2_aug", "dtype": "string"}, {"name": "th_qcpg_0.2_llm_gec_aug", "dtype": "string"}, {"name": "th_qcpg_0.5_aug", "dtype": "string"}, {"name": "th_qcpg_0.5_llm_gec_aug", "dtype": "string"}, {"name": "th_qcpg_0.8_aug", "dtype": "string"}, {"name": "th_qcpg_0.8_llm_gec_aug", "dtype": "string"}, {"name": "th_thai2fit_aug", "dtype": "string"}, {"name": "th_thai2trans_aug", "dtype": "string"}, {"name": "th_wordnet_aug", "dtype": "string"}, {"name": "en_aug", "dtype": "string"}, {"name": "en_llm_gec_aug", "dtype": "string"}, {"name": "en_llm_paraphrase_aug", "dtype": "string"}, {"name": "en_qcpg_0.2_aug", "dtype": "string"}, {"name": "en_qcpg_0.2_llm_gec_aug", "dtype": "string"}, {"name": "en_qcpg_0.5_aug", "dtype": "string"}, {"name": "en_qcpg_0.5_llm_gec_aug", "dtype": "string"}, {"name": "en_qcpg_0.8_aug", "dtype": "string"}, {"name": "en_qcpg_0.8_llm_gec_aug", "dtype": "string"}, {"name": "dis_aug", "dtype": "float64"}, {"name": "dis_fasttext_aug", "dtype": "float64"}, {"name": "dis_llm_gec_aug", "dtype": "float64"}, {"name": "dis_llm_paraphrase_aug", "dtype": "float64"}, {"name": "dis_ltw2v_aug", "dtype": "float64"}, {"name": "dis_qcpg_0.2_aug", "dtype": "float64"}, {"name": "dis_qcpg_0.2_llm_gec_aug", "dtype": "float64"}, {"name": "dis_qcpg_0.5_aug", "dtype": "float64"}, {"name": "dis_qcpg_0.5_llm_gec_aug", "dtype": "float64"}, {"name": "dis_qcpg_0.8_aug", "dtype": "float64"}, {"name": "dis_qcpg_0.8_llm_gec_aug", "dtype": "float64"}, {"name": "dis_thai2fit_aug", "dtype": "float64"}, {"name": "dis_thai2trans_aug", "dtype": "float64"}, {"name": "dis_wordnet_aug", "dtype": "float64"}], "splits": [{"name": "train", "num_bytes": 117313078, "num_examples": 16980}], "download_size": 35147642, "dataset_size": 117313078}} | 2024-01-06T03:31:33+00:00 | []
| []
| TAGS
#region-us
|
# Dataset Card for "Cross-Lingual Data Augmentation For Thai QA"
## Table of Contents
- Dataset Description
- Dataset Structure
- Acknowledgements
- Authors
- Additional Information
## Dataset Description
### Abstract
This dataset accompanies the paper titled "Cross-Lingual Data Augmentation For Thai Question Answering" by Parinthapat Pengpun, Can Udomcharoenchaikit, Weerayut Buaphet, and Peerat Limkonchotiwat, to be presented at GenBench in EMNLP 2023. The paper introduces an innovative framework for data augmentation with quality control measures, aimed at enhancing the robustness of Thai QA models. This dataset is designed to improve model performance in low-resource language settings like Thai, by increasing linguistic diversity through monolingual and cross-lingual data augmentation techniques.
### Links
- ACL Link: PDF
- ResearchGate Link: PDF
## Dataset Structure
### Dataset Info
The dataset, available at Hugging Face Datasets, is structured with the following features:
- 'id': string
- 'question': string
- 'context': string
- 'answers': string
- 'source': string
- Augmentation columns for Thai (e.g., 'th_aug', 'th_fasttext_aug', 'th_llm_gec_aug', etc.)
- Augmentation columns for English (e.g., 'en_aug', 'en_llm_gec_aug', 'en_llm_paraphrase_aug', etc.)
- Semantic distance columns for various augmentations (e.g., 'dis_aug', 'dis_fasttext_aug', 'dis_llm_gec_aug', etc.)
### Splits (No Designated Train/Test Splits)
- Train:
- Number of rows: 16,980
- Number of augmentation sets: 10
- Total Number of Examples = 16,980 * 11 = 186,780
- Size: 117,313,078 bytes
### Download Size
- 35,147,642 bytes
### Total Dataset Size
- 117,313,078 bytes
## Acknowledgements
\n- Augmentation columns for English (e.g., 'en_aug', 'en_llm_gec_aug', 'en_llm_paraphrase_aug', etc.)\n- Semantic distance columns for various augmentations (e.g., 'dis_aug', 'dis_fasttext_aug', 'dis_llm_gec_aug', etc.)",
"### Splits (No Designated Train/Test Splits)\n- Train: \n - Number of rows: 16,980\n - Number of augmentation sets: 10\n - Total Number of Examples = 16,980 * 11 = 186,780\n - Size: 117,313,078 bytes",
"### Download Size\n- 35,147,642 bytes",
"### Total Dataset Size\n- 117,313,078 bytes",
"## Acknowledgements\n\n- Augmentation columns for English (e.g., 'en_aug', 'en_llm_gec_aug', 'en_llm_paraphrase_aug', etc.)\n- Semantic distance columns for various augmentations (e.g., 'dis_aug', 'dis_fasttext_aug', 'dis_llm_gec_aug', etc.)",
"### Splits (No Designated Train/Test Splits)\n- Train: \n - Number of rows: 16,980\n - Number of augmentation sets: 10\n - Total Number of Examples = 16,980 * 11 = 186,780\n - Size: 117,313,078 bytes",
"### Download Size\n- 35,147,642 bytes",
"### Total Dataset Size\n- 117,313,078 bytes",
"## Acknowledgements\n\n- Augmentation columns for English (e.g., 'en_aug', 'en_llm_gec_aug', 'en_llm_paraphrase_aug', etc.)\n- Semantic distance columns for various augmentations (e.g., 'dis_aug', 'dis_fasttext_aug', 'dis_llm_gec_aug', etc.)### Splits (No Designated Train/Test Splits)\n- Train: \n - Number of rows: 16,980\n - Number of augmentation sets: 10\n - Total Number of Examples = 16,980 * 11 = 186,780\n - Size: 117,313,078 bytes"
]
|
fb433acf03c600b3de361188acc42a76f4c87a87 |
# Dataset Card for Zip Code to Timezone Offset Mapping
<!-- Provide a quick summary of the dataset. -->
This dataset maps Zip Codes and Postal Codes for the USA and Canada to the relevant timezone offset.
## Dataset Details
### Dataset Description
In addition to providing a mapping from a Zip Code or Postal Code to timezone offset, it also contains the timezone offset for DST (if observed).
- **Curated by:** Bobby Gill, BlueLabel
### Acknowledgements
<!-- Provide the basic links for the dataset. -->
- **Based off the Work Here:** [https://www.kaggle.com/datasets/joeleichter/us-zip-codes-with-lat-and-long]
| omgbobbyg/Zip-Code-to-Timezone | [
"region:us"
]
| 2023-10-27T10:14:09+00:00 | {} | 2023-10-27T10:29:34+00:00 | []
| []
| TAGS
#region-us
|
# Dataset Card for Zip Code to Timezone Offset Mapping
This dataset maps Zip Codes and Postal Codes for the USA and Canada to the relevant timezone offset.
## Dataset Details
### Dataset Description
In addition to providing a mapping from a Zip Code or Postal Code to timezone offset, it also contains the timezone offset for DST (if observed).
- Curated by: Bobby Gill, BlueLabel
### Acknowledgements
- Based off the Work Here: [URL
| [
"# Dataset Card for Zip Code to Timezone Offset Mapping\n\n\n\nThis dataset maps Zip Codes and Postal Codes for the USA and Canada to the relevant timezone offset.",
"## Dataset Details",
"### Dataset Description\n\nIn addition to providing a mapping from a Zip Code or Postal Code to timezone offset, it also contains the timezone offset for DST (if observed).\n\n- Curated by: Bobby Gill, BlueLabel",
"### Acknowledgements\n\n\n\n- Based off the Work Here: [URL"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for Zip Code to Timezone Offset Mapping\n\n\n\nThis dataset maps Zip Codes and Postal Codes for the USA and Canada to the relevant timezone offset.",
"## Dataset Details",
"### Dataset Description\n\nIn addition to providing a mapping from a Zip Code or Postal Code to timezone offset, it also contains the timezone offset for DST (if observed).\n\n- Curated by: Bobby Gill, BlueLabel",
"### Acknowledgements\n\n\n\n- Based off the Work Here: [URL"
]
| [
6,
40,
4,
54,
16
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for Zip Code to Timezone Offset Mapping\n\n\n\nThis dataset maps Zip Codes and Postal Codes for the USA and Canada to the relevant timezone offset.## Dataset Details### Dataset Description\n\nIn addition to providing a mapping from a Zip Code or Postal Code to timezone offset, it also contains the timezone offset for DST (if observed).\n\n- Curated by: Bobby Gill, BlueLabel### Acknowledgements\n\n\n\n- Based off the Work Here: [URL"
]
|
b8a52b61fb0d57227dfe95e9dff41dad560b609e | # Dataset Card for "filtered_lemma41kV0.0.04"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | fia24/filtered_lemma41kV0.0.04 | [
"region:us"
]
| 2023-10-27T10:20:15+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}, {"split": "test", "path": "data/test-*"}, {"split": "val", "path": "data/val-*"}]}], "dataset_info": {"features": [{"name": "Unnamed: 0", "dtype": "int64"}, {"name": "Inflected_Word", "dtype": "string"}, {"name": "Lemma", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 1665775.7149052797, "num_examples": 26562}, {"name": "test", "num_bytes": 208268.99891576063, "num_examples": 3321}, {"name": "val", "num_bytes": 208206.28617895974, "num_examples": 3320}], "download_size": 1113260, "dataset_size": 2082251.0}} | 2023-10-27T10:20:22+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "filtered_lemma41kV0.0.04"
More Information needed | [
"# Dataset Card for \"filtered_lemma41kV0.0.04\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"filtered_lemma41kV0.0.04\"\n\nMore Information needed"
]
| [
6,
21
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"filtered_lemma41kV0.0.04\"\n\nMore Information needed"
]
|
871ed6ed565efb2725e415a1ca0ce85b8e48d3fa | # Dataset Card for "74441c7b"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | result-kand2-sdxl-wuerst-karlo/74441c7b | [
"region:us"
]
| 2023-10-27T10:35:55+00:00 | {"dataset_info": {"features": [{"name": "result", "dtype": "string"}, {"name": "id", "dtype": "int64"}], "splits": [{"name": "train", "num_bytes": 178, "num_examples": 10}], "download_size": 1358, "dataset_size": 178}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-27T10:35:57+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "74441c7b"
More Information needed | [
"# Dataset Card for \"74441c7b\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"74441c7b\"\n\nMore Information needed"
]
| [
6,
15
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"74441c7b\"\n\nMore Information needed"
]
|
8a993e13496d926089c9efad6e0841e881ee3738 | ## Writing a Novel
写小说:
(1)下载小说。
(2)利用大模型将小说重新整理成一句一行。
(3)利用大模型对小说的各段落写一个摘要。
(4)摘要做 query,小说内容做 response,训练模型。
| qgyd2021/writing_a_novel | [
"task_categories:question-answering",
"task_categories:text-generation",
"task_categories:text2text-generation",
"size_categories:100M<n<1B",
"language:zh",
"license:apache-2.0",
"region:us"
]
| 2023-10-27T10:47:10+00:00 | {"language": ["zh"], "license": "apache-2.0", "size_categories": ["100M<n<1B"], "task_categories": ["question-answering", "text-generation", "text2text-generation"]} | 2023-10-27T11:00:25+00:00 | []
| [
"zh"
]
| TAGS
#task_categories-question-answering #task_categories-text-generation #task_categories-text2text-generation #size_categories-100M<n<1B #language-Chinese #license-apache-2.0 #region-us
| ## Writing a Novel
写小说:
(1)下载小说。
(2)利用大模型将小说重新整理成一句一行。
(3)利用大模型对小说的各段落写一个摘要。
(4)摘要做 query,小说内容做 response,训练模型。
| [
"## Writing a Novel\n\n写小说:\n\n(1)下载小说。\n\n(2)利用大模型将小说重新整理成一句一行。\n\n(3)利用大模型对小说的各段落写一个摘要。\n\n(4)摘要做 query,小说内容做 response,训练模型。"
]
| [
"TAGS\n#task_categories-question-answering #task_categories-text-generation #task_categories-text2text-generation #size_categories-100M<n<1B #language-Chinese #license-apache-2.0 #region-us \n",
"## Writing a Novel\n\n写小说:\n\n(1)下载小说。\n\n(2)利用大模型将小说重新整理成一句一行。\n\n(3)利用大模型对小说的各段落写一个摘要。\n\n(4)摘要做 query,小说内容做 response,训练模型。"
]
| [
67,
52
]
| [
"passage: TAGS\n#task_categories-question-answering #task_categories-text-generation #task_categories-text2text-generation #size_categories-100M<n<1B #language-Chinese #license-apache-2.0 #region-us \n## Writing a Novel\n\n写小说:\n\n(1)下载小说。\n\n(2)利用大模型将小说重新整理成一句一行。\n\n(3)利用大模型对小说的各段落写一个摘要。\n\n(4)摘要做 query,小说内容做 response,训练模型。"
]
|
8f66612f2148d7232a877343606949d323d64a1c | # Dataset Card for "Synthetic_English_MMS_EL"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | mekaneeky/Synthetic_English_MMS | [
"region:us"
]
| 2023-10-27T11:36:58+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}, {"split": "dev", "path": "data/dev-*"}, {"split": "test", "path": "data/test-*"}]}], "dataset_info": {"features": [{"name": "eng", "dtype": "string"}, {"name": "lug", "dtype": "string"}, {"name": "ach", "dtype": "string"}, {"name": "teo", "dtype": "string"}, {"name": "lgg", "dtype": "string"}, {"name": "nyn", "dtype": "string"}, {"name": "ID", "dtype": "string"}, {"name": "eng_tts", "sequence": {"sequence": "float32"}}], "splits": [{"name": "train", "num_bytes": 12857414976, "num_examples": 23947}, {"name": "dev", "num_bytes": 267728460, "num_examples": 500}, {"name": "test", "num_bytes": 266636552, "num_examples": 500}], "download_size": 13400072749, "dataset_size": 13391779988}} | 2023-10-27T11:46:42+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "Synthetic_English_MMS_EL"
More Information needed | [
"# Dataset Card for \"Synthetic_English_MMS_EL\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"Synthetic_English_MMS_EL\"\n\nMore Information needed"
]
| [
6,
21
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"Synthetic_English_MMS_EL\"\n\nMore Information needed"
]
|
f0813b28d36379962f7831ec8209959e421e1c13 | # Dataset Card for "accepted_pairs_base"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | makram93/accepted_pairs_base | [
"region:us"
]
| 2023-10-27T12:04:20+00:00 | {"dataset_info": {"features": [{"name": "url", "dtype": "string"}, {"name": "doc_id", "dtype": "string"}, {"name": "original_title", "sequence": "string"}, {"name": "right", "dtype": "string"}, {"name": "left", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 88447.0623234648, "num_examples": 100}], "download_size": 0, "dataset_size": 88447.0623234648}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-27T14:03:13+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "accepted_pairs_base"
More Information needed | [
"# Dataset Card for \"accepted_pairs_base\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"accepted_pairs_base\"\n\nMore Information needed"
]
| [
6,
17
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"accepted_pairs_base\"\n\nMore Information needed"
]
|
769b89f1df9c265376efb93faa997261a6f21188 | # Dataset Card for "rejected_pairs_base"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | makram93/rejected_pairs_base | [
"region:us"
]
| 2023-10-27T12:04:22+00:00 | {"dataset_info": {"features": [{"name": "url", "dtype": "string"}, {"name": "doc_id", "dtype": "string"}, {"name": "original_title", "sequence": "string"}, {"name": "right", "dtype": "string"}, {"name": "left", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 88447.0623234648, "num_examples": 100}], "download_size": 0, "dataset_size": 88447.0623234648}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-27T14:03:14+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "rejected_pairs_base"
More Information needed | [
"# Dataset Card for \"rejected_pairs_base\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"rejected_pairs_base\"\n\nMore Information needed"
]
| [
6,
17
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"rejected_pairs_base\"\n\nMore Information needed"
]
|
1082f8589abacdc9ed7ba5cb82ef7846df1a7b29 | # Dataset Card for "web_crawl_docs"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | mponty/web_crawl_docs | [
"region:us"
]
| 2023-10-27T12:18:10+00:00 | {"dataset_info": {"features": [{"name": "text", "dtype": "string"}, {"name": "lang", "dtype": "string"}, {"name": "source", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "id", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 1157430740, "num_examples": 87370}], "download_size": 492924255, "dataset_size": 1157430740}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-27T12:18:47+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "web_crawl_docs"
More Information needed | [
"# Dataset Card for \"web_crawl_docs\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"web_crawl_docs\"\n\nMore Information needed"
]
| [
6,
17
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"web_crawl_docs\"\n\nMore Information needed"
]
|
882aaea1a07cac86691b9dcd0893c3c6622e1c92 | # Dataset Card for "GRE_synonyms_gregmat"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | chirunder/GRE_synonyms_gregmat | [
"region:us"
]
| 2023-10-27T12:20:03+00:00 | {"dataset_info": {"features": [{"name": "html", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 1025161, "num_examples": 310}], "download_size": 212008, "dataset_size": 1025161}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-27T12:20:07+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "GRE_synonyms_gregmat"
More Information needed | [
"# Dataset Card for \"GRE_synonyms_gregmat\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"GRE_synonyms_gregmat\"\n\nMore Information needed"
]
| [
6,
18
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"GRE_synonyms_gregmat\"\n\nMore Information needed"
]
|
5e81bdfc0a2764a61438f7fd79516f2f473304cd | # Dataset Card for "sft-train-samples"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | facat/sft-train-samples | [
"region:us"
]
| 2023-10-27T12:49:23+00:00 | {"dataset_info": {"features": [{"name": "prompt", "dtype": "string"}, {"name": "output", "dtype": "string"}, {"name": "task", "dtype": "string"}, {"name": "name", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 7997635, "num_examples": 2420}], "download_size": 4481416, "dataset_size": 7997635}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-28T03:27:28+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "sft-train-samples"
More Information needed | [
"# Dataset Card for \"sft-train-samples\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"sft-train-samples\"\n\nMore Information needed"
]
| [
6,
18
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"sft-train-samples\"\n\nMore Information needed"
]
|
d226bb3bea1dbdf562d42878589720a47aef59ad | # Dataset Card for "soict_sentence_filter"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | thanhduycao/soict_sentence_filter | [
"region:us"
]
| 2023-10-27T12:49:40+00:00 | {"dataset_info": {"features": [{"name": "sentence", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 12169, "num_examples": 197}], "download_size": 6583, "dataset_size": 12169}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}]} | 2023-10-27T12:49:41+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "soict_sentence_filter"
More Information needed | [
"# Dataset Card for \"soict_sentence_filter\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"soict_sentence_filter\"\n\nMore Information needed"
]
| [
6,
18
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"soict_sentence_filter\"\n\nMore Information needed"
]
|
a3ef28f5a3592b5897c79a2611c24fda3cde21dd |
# Dataset Card for "Arab States Analogy Dataset (ASAD)"
This dataset is created using 20 Arab States<sup>1</sup> with their corresponding capital cities, nationalities, currencies, and on which continents they are located, consisting of four sets: country-capital set, country-currency set, country-nationality set, and country-continent set. Each set has 380 word analogies, and the total number of word analogies in the ASAD dataset is 1520. This dataset is used to evaluate Arabic Word Embedding Models (WEMs).
For more details about the dataset, please **read** and **cite** our paper:
```bash
@inproceedings{alshahrani-etal-2023-performance,
title = "{Performance Implications of Using Unrepresentative Corpora in {A}rabic Natural Language Processing}",
author = "Alshahrani, Saied and Alshahrani, Norah and Dey, Soumyabrata and Matthews, Jeanna",
booktitle = "Proceedings of the The First Arabic Natural Language Processing Conference (ArabicNLP 2023)",
month = December,
year = "2023",
address = "Singapore (Hybrid)",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.arabicnlp-1.19",
doi = "10.18653/v1/2023.arabicnlp-1.19",
pages = "218--231",
abstract = "Wikipedia articles are a widely used source of training data for Natural Language Processing (NLP) research, particularly as corpora for low-resource languages like Arabic. However, it is essential to understand the extent to which these corpora reflect the representative contributions of native speakers, especially when many entries in a given language are directly translated from other languages or automatically generated through automated mechanisms. In this paper, we study the performance implications of using inorganic corpora that are not representative of native speakers and are generated through automated techniques such as bot generation or automated template-based translation. The case of the Arabic Wikipedia editions gives a unique case study of this since the Moroccan Arabic Wikipedia edition (ARY) is small but representative, the Egyptian Arabic Wikipedia edition (ARZ) is large but unrepresentative, and the Modern Standard Arabic Wikipedia edition (AR) is both large and more representative. We intrinsically evaluate the performance of two main NLP upstream tasks, namely word representation and language modeling, using word analogy evaluations and fill-mask evaluations using our two newly created datasets: Arab States Analogy Dataset (ASAD) and Masked Arab States Dataset (MASD). We demonstrate that for good NLP performance, we need both large and organic corpora; neither alone is sufficient. We show that producing large corpora through automated means can be a counter-productive, producing models that both perform worse and lack cultural richness and meaningful representation of the Arabic language and its native speakers.",
}
```
<sub>1. We only drop two Arab states: the United Arab Emirates (الإمارات العربية المتحدة) and Comoros (جزر القمر), because they or their capital cities are written as open compound words (two words), which cannot be directly handled by the word embedding models, like Abu Dhabi (أبو ظبي).</sub> | SaiedAlshahrani/ASAD | [
"size_categories:1K<n<10K",
"language:ar",
"license:mit",
"region:us"
]
| 2023-10-27T12:55:52+00:00 | {"language": ["ar"], "license": "mit", "size_categories": ["1K<n<10K"], "pretty_name": "ASAD"} | 2024-01-05T15:13:16+00:00 | []
| [
"ar"
]
| TAGS
#size_categories-1K<n<10K #language-Arabic #license-mit #region-us
|
# Dataset Card for "Arab States Analogy Dataset (ASAD)"
This dataset is created using 20 Arab States<sup>1</sup> with their corresponding capital cities, nationalities, currencies, and on which continents they are located, consisting of four sets: country-capital set, country-currency set, country-nationality set, and country-continent set. Each set has 380 word analogies, and the total number of word analogies in the ASAD dataset is 1520. This dataset is used to evaluate Arabic Word Embedding Models (WEMs).
For more details about the dataset, please read and cite our paper:
<sub>1. We only drop two Arab states: the United Arab Emirates (الإمارات العربية المتحدة) and Comoros (جزر القمر), because they or their capital cities are written as open compound words (two words), which cannot be directly handled by the word embedding models, like Abu Dhabi (أبو ظبي).</sub> | [
"# Dataset Card for \"Arab States Analogy Dataset (ASAD)\"\n\nThis dataset is created using 20 Arab States<sup>1</sup> with their corresponding capital cities, nationalities, currencies, and on which continents they are located, consisting of four sets: country-capital set, country-currency set, country-nationality set, and country-continent set. Each set has 380 word analogies, and the total number of word analogies in the ASAD dataset is 1520. This dataset is used to evaluate Arabic Word Embedding Models (WEMs). \n\nFor more details about the dataset, please read and cite our paper:\n\n\n\n<sub>1. We only drop two Arab states: the United Arab Emirates (الإمارات العربية المتحدة) and Comoros (جزر القمر), because they or their capital cities are written as open compound words (two words), which cannot be directly handled by the word embedding models, like Abu Dhabi (أبو ظبي).</sub>"
]
| [
"TAGS\n#size_categories-1K<n<10K #language-Arabic #license-mit #region-us \n",
"# Dataset Card for \"Arab States Analogy Dataset (ASAD)\"\n\nThis dataset is created using 20 Arab States<sup>1</sup> with their corresponding capital cities, nationalities, currencies, and on which continents they are located, consisting of four sets: country-capital set, country-currency set, country-nationality set, and country-continent set. Each set has 380 word analogies, and the total number of word analogies in the ASAD dataset is 1520. This dataset is used to evaluate Arabic Word Embedding Models (WEMs). \n\nFor more details about the dataset, please read and cite our paper:\n\n\n\n<sub>1. We only drop two Arab states: the United Arab Emirates (الإمارات العربية المتحدة) and Comoros (جزر القمر), because they or their capital cities are written as open compound words (two words), which cannot be directly handled by the word embedding models, like Abu Dhabi (أبو ظبي).</sub>"
]
| [
28,
224
]
| [
"passage: TAGS\n#size_categories-1K<n<10K #language-Arabic #license-mit #region-us \n# Dataset Card for \"Arab States Analogy Dataset (ASAD)\"\n\nThis dataset is created using 20 Arab States<sup>1</sup> with their corresponding capital cities, nationalities, currencies, and on which continents they are located, consisting of four sets: country-capital set, country-currency set, country-nationality set, and country-continent set. Each set has 380 word analogies, and the total number of word analogies in the ASAD dataset is 1520. This dataset is used to evaluate Arabic Word Embedding Models (WEMs). \n\nFor more details about the dataset, please read and cite our paper:\n\n\n\n<sub>1. We only drop two Arab states: the United Arab Emirates (الإمارات العربية المتحدة) and Comoros (جزر القمر), because they or their capital cities are written as open compound words (two words), which cannot be directly handled by the word embedding models, like Abu Dhabi (أبو ظبي).</sub>"
]
|
0bf5cbf8dc419eb2962a1b42b3c33eb0529a6012 |
# Dataset Card for "Masked Arab States Dataset (MASD)"
This dataset is created using 20 Arab States<sup>1</sup> with their corresponding capital cities, nationalities, currencies, and on which continents they are located, consisting of four categories: country-capital
prompts, country-currency prompts, country-nationality prompts, and country-continent prompts. Each prompts category has 40 masked prompts, and the total number of masked prompts in the MASD dataset is 160. This dataset is used to evaluate these Arabic Masked Language Models (MLMs):
1. [SaiedAlshahrani/arwiki_20230101_roberta_mlm_bots](https://huggingface.co/SaiedAlshahrani/arwiki_20230101_roberta_mlm_bots).
2. [SaiedAlshahrani/arwiki_20230101_roberta_mlm_nobots](https://huggingface.co/SaiedAlshahrani/arwiki_20230101_roberta_mlm_nobots).
3. [SaiedAlshahrani/arzwiki_20230101_roberta_mlm](https://huggingface.co/SaiedAlshahrani/arzwiki_20230101_roberta_mlm).
4. [SaiedAlshahrani/arywiki_20230101_roberta_mlm_bots](https://huggingface.co/SaiedAlshahrani/arywiki_20230101_roberta_mlm_bots).
5. [SaiedAlshahrani/arywiki_20230101_roberta_mlm_nobots](https://huggingface.co/SaiedAlshahrani/arywiki_20230101_roberta_mlm_nobots).
For more details about the dataset, please **read** and **cite** our paper:
```bash
@inproceedings{alshahrani-etal-2023-performance,
title = "{Performance Implications of Using Unrepresentative Corpora in {A}rabic Natural Language Processing}",
author = "Alshahrani, Saied and Alshahrani, Norah and Dey, Soumyabrata and Matthews, Jeanna",
booktitle = "Proceedings of the The First Arabic Natural Language Processing Conference (ArabicNLP 2023)",
month = December,
year = "2023",
address = "Singapore (Hybrid)",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.arabicnlp-1.19",
doi = "10.18653/v1/2023.arabicnlp-1.19",
pages = "218--231",
abstract = "Wikipedia articles are a widely used source of training data for Natural Language Processing (NLP) research, particularly as corpora for low-resource languages like Arabic. However, it is essential to understand the extent to which these corpora reflect the representative contributions of native speakers, especially when many entries in a given language are directly translated from other languages or automatically generated through automated mechanisms. In this paper, we study the performance implications of using inorganic corpora that are not representative of native speakers and are generated through automated techniques such as bot generation or automated template-based translation. The case of the Arabic Wikipedia editions gives a unique case study of this since the Moroccan Arabic Wikipedia edition (ARY) is small but representative, the Egyptian Arabic Wikipedia edition (ARZ) is large but unrepresentative, and the Modern Standard Arabic Wikipedia edition (AR) is both large and more representative. We intrinsically evaluate the performance of two main NLP upstream tasks, namely word representation and language modeling, using word analogy evaluations and fill-mask evaluations using our two newly created datasets: Arab States Analogy Dataset (ASAD) and Masked Arab States Dataset (MASD). We demonstrate that for good NLP performance, we need both large and organic corpora; neither alone is sufficient. We show that producing large corpora through automated means can be a counter-productive, producing models that both perform worse and lack cultural richness and meaningful representation of the Arabic language and its native speakers.",
}
```
<sub>1. We only drop two Arab states: the United Arab Emirates (الإمارات العربية المتحدة) and Comoros (جزر القمر), because they or their capital cities are written as open compound words (two words), which cannot be directly handled by the word embedding models, like Abu Dhabi (أبو ظبي).</sub> | SaiedAlshahrani/MASD | [
"size_categories:n<1K",
"language:ar",
"license:mit",
"region:us"
]
| 2023-10-27T13:07:44+00:00 | {"language": ["ar"], "license": "mit", "size_categories": ["n<1K"], "pretty_name": "MASD"} | 2024-01-05T15:13:59+00:00 | []
| [
"ar"
]
| TAGS
#size_categories-n<1K #language-Arabic #license-mit #region-us
|
# Dataset Card for "Masked Arab States Dataset (MASD)"
This dataset is created using 20 Arab States<sup>1</sup> with their corresponding capital cities, nationalities, currencies, and on which continents they are located, consisting of four categories: country-capital
prompts, country-currency prompts, country-nationality prompts, and country-continent prompts. Each prompts category has 40 masked prompts, and the total number of masked prompts in the MASD dataset is 160. This dataset is used to evaluate these Arabic Masked Language Models (MLMs):
1. SaiedAlshahrani/arwiki_20230101_roberta_mlm_bots.
2. SaiedAlshahrani/arwiki_20230101_roberta_mlm_nobots.
3. SaiedAlshahrani/arzwiki_20230101_roberta_mlm.
4. SaiedAlshahrani/arywiki_20230101_roberta_mlm_bots.
5. SaiedAlshahrani/arywiki_20230101_roberta_mlm_nobots.
For more details about the dataset, please read and cite our paper:
<sub>1. We only drop two Arab states: the United Arab Emirates (الإمارات العربية المتحدة) and Comoros (جزر القمر), because they or their capital cities are written as open compound words (two words), which cannot be directly handled by the word embedding models, like Abu Dhabi (أبو ظبي).</sub> | [
"# Dataset Card for \"Masked Arab States Dataset (MASD)\"\nThis dataset is created using 20 Arab States<sup>1</sup> with their corresponding capital cities, nationalities, currencies, and on which continents they are located, consisting of four categories: country-capital\nprompts, country-currency prompts, country-nationality prompts, and country-continent prompts. Each prompts category has 40 masked prompts, and the total number of masked prompts in the MASD dataset is 160. This dataset is used to evaluate these Arabic Masked Language Models (MLMs):\n1. SaiedAlshahrani/arwiki_20230101_roberta_mlm_bots.\n2. SaiedAlshahrani/arwiki_20230101_roberta_mlm_nobots.\n3. SaiedAlshahrani/arzwiki_20230101_roberta_mlm.\n4. SaiedAlshahrani/arywiki_20230101_roberta_mlm_bots.\n5. SaiedAlshahrani/arywiki_20230101_roberta_mlm_nobots.\n\nFor more details about the dataset, please read and cite our paper:\n\n\n\n<sub>1. We only drop two Arab states: the United Arab Emirates (الإمارات العربية المتحدة) and Comoros (جزر القمر), because they or their capital cities are written as open compound words (two words), which cannot be directly handled by the word embedding models, like Abu Dhabi (أبو ظبي).</sub>"
]
| [
"TAGS\n#size_categories-n<1K #language-Arabic #license-mit #region-us \n",
"# Dataset Card for \"Masked Arab States Dataset (MASD)\"\nThis dataset is created using 20 Arab States<sup>1</sup> with their corresponding capital cities, nationalities, currencies, and on which continents they are located, consisting of four categories: country-capital\nprompts, country-currency prompts, country-nationality prompts, and country-continent prompts. Each prompts category has 40 masked prompts, and the total number of masked prompts in the MASD dataset is 160. This dataset is used to evaluate these Arabic Masked Language Models (MLMs):\n1. SaiedAlshahrani/arwiki_20230101_roberta_mlm_bots.\n2. SaiedAlshahrani/arwiki_20230101_roberta_mlm_nobots.\n3. SaiedAlshahrani/arzwiki_20230101_roberta_mlm.\n4. SaiedAlshahrani/arywiki_20230101_roberta_mlm_bots.\n5. SaiedAlshahrani/arywiki_20230101_roberta_mlm_nobots.\n\nFor more details about the dataset, please read and cite our paper:\n\n\n\n<sub>1. We only drop two Arab states: the United Arab Emirates (الإمارات العربية المتحدة) and Comoros (جزر القمر), because they or their capital cities are written as open compound words (two words), which cannot be directly handled by the word embedding models, like Abu Dhabi (أبو ظبي).</sub>"
]
| [
26,
344
]
| [
"passage: TAGS\n#size_categories-n<1K #language-Arabic #license-mit #region-us \n# Dataset Card for \"Masked Arab States Dataset (MASD)\"\nThis dataset is created using 20 Arab States<sup>1</sup> with their corresponding capital cities, nationalities, currencies, and on which continents they are located, consisting of four categories: country-capital\nprompts, country-currency prompts, country-nationality prompts, and country-continent prompts. Each prompts category has 40 masked prompts, and the total number of masked prompts in the MASD dataset is 160. This dataset is used to evaluate these Arabic Masked Language Models (MLMs):\n1. SaiedAlshahrani/arwiki_20230101_roberta_mlm_bots.\n2. SaiedAlshahrani/arwiki_20230101_roberta_mlm_nobots.\n3. SaiedAlshahrani/arzwiki_20230101_roberta_mlm.\n4. SaiedAlshahrani/arywiki_20230101_roberta_mlm_bots.\n5. SaiedAlshahrani/arywiki_20230101_roberta_mlm_nobots.\n\nFor more details about the dataset, please read and cite our paper:\n\n\n\n<sub>1. We only drop two Arab states: the United Arab Emirates (الإمارات العربية المتحدة) and Comoros (جزر القمر), because they or their capital cities are written as open compound words (two words), which cannot be directly handled by the word embedding models, like Abu Dhabi (أبو ظبي).</sub>"
]
|
c7240c40751b0df45d86f01a33b3692ec84973ac | # Dataset Card for "LLaVa-instruction-trasaleted"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | Frorozcol/LLaVa-instruction-trasaleted | [
"region:us"
]
| 2023-10-27T13:18:29+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}], "dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "conversations", "list": [{"name": "from", "dtype": "string"}, {"name": "value", "dtype": "string"}]}, {"name": "image", "dtype": "string"}, {"name": "conversations_translated", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 397720156, "num_examples": 157500}], "download_size": 197927858, "dataset_size": 397720156}} | 2023-10-27T13:18:56+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "LLaVa-instruction-trasaleted"
More Information needed | [
"# Dataset Card for \"LLaVa-instruction-trasaleted\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"LLaVa-instruction-trasaleted\"\n\nMore Information needed"
]
| [
6,
19
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"LLaVa-instruction-trasaleted\"\n\nMore Information needed"
]
|
bd00482f06128c4c91c8417e76e2b9db8b78ff84 | # Dataset Card for "ner_assignment"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | FunkyQ/NER_Assignment | [
"region:us"
]
| 2023-10-27T13:23:09+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}, {"split": "validation", "path": "data/validation-*"}, {"split": "test", "path": "data/test-*"}]}], "dataset_info": {"features": [{"name": "sentence", "sequence": "string"}, {"name": "labels", "sequence": "string"}, {"name": "word_idx", "sequence": "int64"}, {"name": "label_idx", "sequence": "int64"}], "splits": [{"name": "train", "num_bytes": 6345988, "num_examples": 14041}, {"name": "validation", "num_bytes": 1595927, "num_examples": 3250}, {"name": "test", "num_bytes": 1449601, "num_examples": 3453}], "download_size": 2208622, "dataset_size": 9391516}} | 2023-10-27T17:07:56+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "ner_assignment"
More Information needed | [
"# Dataset Card for \"ner_assignment\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"ner_assignment\"\n\nMore Information needed"
]
| [
6,
15
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"ner_assignment\"\n\nMore Information needed"
]
|
bc2deecb599eb65e9c82eaf980b578485128c4e1 | Japanese Spitz, resolution above 256x256, no 1 to 1 copies (but there are processed versions of one photo), serial names.
mistaken breed at 632, 639, 640, 644, 1003, 1810, 2000 numbers | Morevorot/Dog_breed_Japanese_Spitz | [
"size_categories:1K<n<10K",
"license:cc-by-sa-4.0",
"region:us"
]
| 2023-10-27T13:23:53+00:00 | {"license": "cc-by-sa-4.0", "size_categories": ["1K<n<10K"], "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}]}], "dataset_info": {"features": [{"name": "image", "dtype": "image"}], "splits": [{"name": "train", "num_bytes": 4336374796.1, "num_examples": 3497}], "download_size": 3682840051, "dataset_size": 4336374796.1}} | 2023-10-30T15:49:24+00:00 | []
| []
| TAGS
#size_categories-1K<n<10K #license-cc-by-sa-4.0 #region-us
| Japanese Spitz, resolution above 256x256, no 1 to 1 copies (but there are processed versions of one photo), serial names.
mistaken breed at 632, 639, 640, 644, 1003, 1810, 2000 numbers | []
| [
"TAGS\n#size_categories-1K<n<10K #license-cc-by-sa-4.0 #region-us \n"
]
| [
29
]
| [
"passage: TAGS\n#size_categories-1K<n<10K #license-cc-by-sa-4.0 #region-us \n"
]
|
6afb5cdf2dce7c1f916452680ef38e1ad56c6373 | # Dataset Card for "Synthetic_English_VITS_22.5k"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | mekaneeky/Synthetic_English_VITS_22.5k | [
"region:us"
]
| 2023-10-27T13:28:34+00:00 | {"configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}, {"split": "dev", "path": "data/dev-*"}, {"split": "test", "path": "data/test-*"}]}], "dataset_info": {"features": [{"name": "eng", "dtype": "string"}, {"name": "lug", "dtype": "string"}, {"name": "ach", "dtype": "string"}, {"name": "teo", "dtype": "string"}, {"name": "lgg", "dtype": "string"}, {"name": "nyn", "dtype": "string"}, {"name": "ID", "dtype": "string"}, {"name": "eng_tts", "sequence": {"sequence": "float32"}}], "splits": [{"name": "train", "num_bytes": 14292643136, "num_examples": 23947}, {"name": "dev", "num_bytes": 286070348, "num_examples": 500}, {"name": "test", "num_bytes": 300760328, "num_examples": 500}], "download_size": 14889677577, "dataset_size": 14879473812}} | 2023-10-27T13:42:36+00:00 | []
| []
| TAGS
#region-us
| # Dataset Card for "Synthetic_English_VITS_22.5k"
More Information needed | [
"# Dataset Card for \"Synthetic_English_VITS_22.5k\"\n\nMore Information needed"
]
| [
"TAGS\n#region-us \n",
"# Dataset Card for \"Synthetic_English_VITS_22.5k\"\n\nMore Information needed"
]
| [
6,
23
]
| [
"passage: TAGS\n#region-us \n# Dataset Card for \"Synthetic_English_VITS_22.5k\"\n\nMore Information needed"
]
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.