sha
stringlengths
40
40
text
stringlengths
0
13.4M
id
stringlengths
2
117
tags
sequence
created_at
stringlengths
25
25
metadata
stringlengths
2
31.7M
last_modified
stringlengths
25
25
6c1a1284eb3557055d7c57b91cd7e68e3252b32c
## Dataset Description - **Homepage:** [Human Action Recognition (HAR) Dataset](https://www.kaggle.com/datasets/meetnagadia/human-action-recognition-har-dataset) - **Repository:** N/A - **Paper:** N/A - **Leaderboard:** N/A - **Point of Contact:** N/A ## Dataset Summary A dataset from [kaggle](https://www.kaggle.com/datasets/meetnagadia/human-action-recognition-har-dataset). origin: https://dphi.tech/challenges/data-sprint-76-human-activity-recognition/233/data ### Introduction - The dataset features 15 different classes of Human Activities. - The dataset contains about 12k+ labelled images including the validation images. - Each image has only one human activity category and are saved in separate folders of the labelled classes ### PROBLEM STATEMENT - Human Action Recognition (HAR) aims to understand human behavior and assign a label to each action. It has a wide range of applications, and therefore has been attracting increasing attention in the field of computer vision. Human actions can be represented using various data modalities, such as RGB, skeleton, depth, infrared, point cloud, event stream, audio, acceleration, radar, and WiFi signal, which encode different sources of useful yet distinct information and have various advantages depending on the application scenarios. - Consequently, lots of existing works have attempted to investigate different types of approaches for HAR using various modalities. - Your Task is to build an Image Classification Model using CNN that classifies to which class of activity a human is performing. ### About Files - Train - contains all the images that are to be used for training your model. In this folder you will find 15 folders namely - 'calling', ’clapping’, ’cycling’, ’dancing’, ‘drinking’, ‘eating’, ‘fighting’, ‘hugging’, ‘laughing’, ‘listeningtomusic’, ‘running’, ‘sitting’, ‘sleeping’, texting’, ‘using_laptop’ which contain the images of the respective human activities. - Test - contains 5400 images of Human Activities. For these images you are required to make predictions as the respective class names -'calling', ’clapping’, ’cycling’, ’dancing’, ‘drinking’, ‘eating’, ‘fighting’, ‘hugging’, ‘laughing’, ‘listeningtomusic’, ‘running’, ‘sitting’, ‘sleeping’, texting’, ‘using_laptop’. - Testing_set.csv - this is the order of the predictions for each image that is to be submitted on the platform. Make sure the predictions you download are with their image’s filename in the same order as given in this file. - sample_submission: This is a csv file that contains the sample submission for the data sprint. ### Data Fields The data instances have the following fields: - `image`: A `PIL.Image.Image` object containing the image. Note that when accessing the image column: `dataset[0]["image"]` the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the `"image"` column, *i.e.* `dataset[0]["image"]` should **always** be preferred over `dataset["image"][0]`. - `labels`: an `int` classification label. All `test` data is labeled 0. ### Class Label Mappings: ``` { 'calling': 0, 'clapping': 1, 'cycling': 2, 'dancing': 3, 'drinking': 4, 'eating': 5, 'fighting': 6, 'hugging': 7, 'laughing': 8, 'listening_to_music': 9, 'running': 10, 'sitting': 11, 'sleeping': 12, 'texting': 13, 'using_laptop': 14 } ``` ### Data Splits | | train | test | |---------------|--------|-----:| | # of examples | 12600 | 5400 | ### Data Size - download: 311.96 MiB - generated: 312.59 MiB - total: 624.55 MiB ```pycon >>> from datasets import load_dataset >>> ds = load_dataset("Bingsu/Human_Action_Recognition") >>> ds DatasetDict({ test: Dataset({ features: ['image', 'labels'], num_rows: 5400 }) train: Dataset({ features: ['image', 'labels'], num_rows: 12600 }) }) >>> ds["train"].features {'image': Image(decode=True, id=None), 'labels': ClassLabel(num_classes=15, names=['calling', 'clapping', 'cycling', 'dancing', 'drinking', 'eating', 'fighting', 'hugging', 'laughing', 'listening_to_music', 'running', 'sitting', 'sleeping', 'texting', 'using_laptop'], id=None)} >>> ds["train"][0] {'image': <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=240x160>, 'labels': 11} ```
Bingsu/Human_Action_Recognition
[ "task_categories:image-classification", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:odbl", "region:us" ]
2022-06-09T01:00:52+00:00
{"language": ["en"], "license": ["odbl"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["image-classification"], "pretty_name": "Human Action Recognition"}
2022-07-05T01:48:56+00:00
e86a47f2f12179b719e73ae853eb4547c40c5cf1
AkaGuda/fate
[ "license:other", "region:us" ]
2022-06-09T05:51:23+00:00
{"license": "other"}
2022-06-09T05:51:23+00:00
f46f986fff162cdbfe9f35874a08d9cec2446b6e
# AutoTrain Dataset for project: quality-customer-reviews ## Dataset Descritpion This dataset has been automatically processed by AutoTrain for project quality-customer-reviews. ### Languages The BCP-47 code for the dataset's language is en. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```json [ { "text": " Love this truck, I think it is light years better than the competition. I have driven or owned all [...]", "target": 1 }, { "text": " I purchased this to haul our 4 horse trailer since the standard iterations of the domestic vehicles[...]", "target": 0 } ] ``` ### Dataset Fields The dataset has the following fields (also called "features"): ```json { "text": "Value(dtype='string', id=None)", "target": "ClassLabel(num_classes=5, names=['good', 'great', 'ok', 'poor', 'terrible'], id=None)" } ``` ### Dataset Splits This dataset is split into a train and validation split. The split sizes are as follow: | Split name | Num samples | | ------------ | ------------------- | | train | 9166 | | valid | 2295 |
florentgbelidji/autotrain-data-quality-customer-reviews
[ "task_categories:text-classification", "language:en", "region:us" ]
2022-06-09T08:35:36+00:00
{"language": ["en"], "task_categories": ["text-classification"]}
2022-10-25T09:29:24+00:00
f666ec81588b1b9df9f93bcbc0ee19a5ca264ad9
# Dataset Card for Quick, Draw! ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Quick, Draw! homepage](https://quickdraw.withgoogle.com/data) - **Repository:** [Quick, Draw! repository](https://github.com/googlecreativelab/quickdraw-dataset) - **Paper:** [A Neural Representation of Sketch Drawings](https://arxiv.org/abs/1704.03477v4) - **Leaderboard:** [Quick, Draw! Doodle Recognition Challenge](https://www.kaggle.com/competitions/quickdraw-doodle-recognition/leaderboard) - **Point of Contact:** [Quick, Draw! support](mailto:[email protected]) ### Dataset Summary The Quick Draw Dataset is a collection of 50 million drawings across 345 categories, contributed by players of the game Quick, Draw!. The drawings were captured as timestamped vectors, tagged with metadata including what the player was asked to draw and in which country the player was located. ### Supported Tasks and Leaderboards - `image-classification`: The goal of this task is to classify a given sketch into one of 345 classes. The (closed) leaderboard for this task is available [here](https://www.kaggle.com/competitions/quickdraw-doodle-recognition/leaderboard). ### Languages English. ## Dataset Structure ### Data Instances #### `raw` A data point comprises a drawing and its metadata. ``` { 'key_id': '5475678961008640', 'word': 0, 'recognized': True, 'timestamp': datetime.datetime(2017, 3, 28, 13, 28, 0, 851730), 'countrycode': 'MY', 'drawing': { 'x': [[379.0, 380.0, 381.0, 381.0, 381.0, 381.0, 382.0], [362.0, 368.0, 375.0, 380.0, 388.0, 393.0, 399.0, 404.0, 409.0, 410.0, 410.0, 405.0, 397.0, 392.0, 384.0, 377.0, 370.0, 363.0, 356.0, 348.0, 342.0, 336.0, 333.0], ..., [477.0, 473.0, 471.0, 469.0, 468.0, 466.0, 464.0, 462.0, 461.0, 469.0, 475.0, 483.0, 491.0, 499.0, 510.0, 521.0, 531.0, 540.0, 548.0, 558.0, 566.0, 576.0, 583.0, 590.0, 595.0, 598.0, 597.0, 596.0, 594.0, 592.0, 590.0, 589.0, 588.0, 586.0]], 'y': [[1.0, 7.0, 15.0, 21.0, 27.0, 32.0, 32.0], [17.0, 17.0, 17.0, 17.0, 16.0, 16.0, 16.0, 16.0, 18.0, 23.0, 29.0, 32.0, 32.0, 32.0, 29.0, 27.0, 25.0, 23.0, 21.0, 19.0, 17.0, 16.0, 14.0], ..., [151.0, 146.0, 139.0, 131.0, 125.0, 119.0, 113.0, 107.0, 102.0, 99.0, 98.0, 98.0, 98.0, 98.0, 98.0, 98.0, 98.0, 98.0, 98.0, 98.0, 98.0, 100.0, 102.0, 104.0, 105.0, 110.0, 115.0, 121.0, 126.0, 131.0, 137.0, 142.0, 148.0, 150.0]], 't': [[0, 84, 100, 116, 132, 148, 260], [573, 636, 652, 660, 676, 684, 701, 724, 796, 838, 860, 956, 973, 979, 989, 995, 1005, 1012, 1020, 1028, 1036, 1053, 1118], ..., [8349, 8446, 8468, 8484, 8500, 8516, 8541, 8557, 8573, 8685, 8693, 8702, 8710, 8718, 8724, 8732, 8741, 8748, 8757, 8764, 8773, 8780, 8788, 8797, 8804, 8965, 8996, 9029, 9045, 9061, 9076, 9092, 9109, 9167]] } } ``` #### `preprocessed_simplified_drawings` The simplified version of the dataset generated from the `raw` data with the simplified vectors, removed timing information, and the data positioned and scaled into a 256x256 region. The simplification process was: 1.Align the drawing to the top-left corner, to have minimum values of 0. 2.Uniformly scale the drawing, to have a maximum value of 255. 3.Resample all strokes with a 1 pixel spacing. 4.Simplify all strokes using the [Ramer-Douglas-Peucker algorithm](https://en.wikipedia.org/wiki/Ramer%E2%80%93Douglas%E2%80%93Peucker_algorithm) with an epsilon value of 2.0. ``` { 'key_id': '5475678961008640', 'word': 0, 'recognized': True, 'timestamp': datetime.datetime(2017, 3, 28, 15, 28), 'countrycode': 'MY', 'drawing': { 'x': [[31, 32], [27, 37, 38, 35, 21], [25, 28, 38, 39], [33, 34, 32], [5, 188, 254, 251, 241, 185, 45, 9, 0], [35, 35, 43, 125, 126], [35, 76, 80, 77], [53, 50, 54, 80, 78]], 'y': [[0, 7], [4, 4, 6, 7, 3], [5, 10, 10, 7], [4, 33, 44], [50, 50, 54, 83, 86, 90, 86, 77, 52], [85, 91, 92, 96, 90], [35, 37, 41, 47], [34, 23, 22, 23, 34]] } } ``` #### `preprocessed_bitmaps` (default configuration) This configuration contains the 28x28 grayscale bitmap images that were generated from the simplified data, but are aligned to the center of the drawing's bounding box rather than the top-left corner. The code that was used for generation is available [here](https://github.com/googlecreativelab/quickdraw-dataset/issues/19#issuecomment-402247262). ``` { 'image': <PIL.PngImagePlugin.PngImageFile image mode=L size=28x28 at 0x10B5B102828>, 'label': 0 } ``` #### `sketch_rnn` and `sketch_rnn_full` The `sketch_rnn_full` configuration stores the data in the format suitable for inputs into a recurrent neural network and was used for for training the [Sketch-RNN](https://arxiv.org/abs/1704.03477) model. Unlike `sketch_rnn` where the samples have been randomly selected from each category, the `sketch_rnn_full` configuration contains the full data for each category. ``` { 'word': 0, 'drawing': [[132, 0, 0], [23, 4, 0], [61, 1, 0], [76, 0, 0], [22, -4, 0], [152, 0, 0], [50, -5, 0], [36, -10, 0], [8, 26, 0], [0, 69, 0], [-2, 11, 0], [-8, 10, 0], [-56, 24, 0], [-23, 14, 0], [-99, 40, 0], [-45, 6, 0], [-21, 6, 0], [-170, 2, 0], [-81, 0, 0], [-29, -9, 0], [-94, -19, 0], [-48, -24, 0], [-6, -16, 0], [2, -36, 0], [7, -29, 0], [23, -45, 0], [13, -6, 0], [41, -8, 0], [42, -2, 1], [392, 38, 0], [2, 19, 0], [11, 33, 0], [13, 0, 0], [24, -9, 0], [26, -27, 0], [0, -14, 0], [-8, -10, 0], [-18, -5, 0], [-14, 1, 0], [-23, 4, 0], [-21, 12, 1], [-152, 18, 0], [10, 46, 0], [26, 6, 0], [38, 0, 0], [31, -2, 0], [7, -2, 0], [4, -6, 0], [-10, -21, 0], [-2, -33, 0], [-6, -11, 0], [-46, 1, 0], [-39, 18, 0], [-19, 4, 1], [-122, 0, 0], [-2, 38, 0], [4, 16, 0], [6, 4, 0], [78, 0, 0], [4, -8, 0], [-8, -36, 0], [0, -22, 0], [-6, -2, 0], [-32, 14, 0], [-58, 13, 1], [-96, -12, 0], [-10, 27, 0], [2, 32, 0], [102, 0, 0], [1, -7, 0], [-27, -17, 0], [-4, -6, 0], [-1, -34, 0], [-64, 8, 1], [129, -138, 0], [-108, 0, 0], [-8, 12, 0], [-1, 15, 0], [12, 15, 0], [20, 5, 0], [61, -3, 0], [24, 6, 0], [19, 0, 0], [5, -4, 0], [2, 14, 1]] } ``` ### Data Fields #### `raw` - `key_id`: A unique identifier across all drawings. - `word`: Category the player was prompted to draw. - `recognized`: Whether the word was recognized by the game. - `timestamp`: When the drawing was created. - `countrycode`: A two letter country code ([ISO 3166-1 alpha-2](https://en.wikipedia.org/wiki/ISO_3166-1_alpha-2)) of where the player was located. - `drawing`: A dictionary where `x` and `y` are the pixel coordinates, and `t` is the time in milliseconds since the first point. `x` and `y` are real-valued while `t` is an integer. `x`, `y` and `t` match in lenght and are represented as lists of lists where each sublist corresponds to a single stroke. The raw drawings can have vastly different bounding boxes and number of points due to the different devices used for display and input. #### `preprocessed_simplified_drawings` - `key_id`: A unique identifier across all drawings. - `word`: Category the player was prompted to draw. - `recognized`: Whether the word was recognized by the game. - `timestamp`: When the drawing was created. - `countrycode`: A two letter country code ([ISO 3166-1 alpha-2](https://en.wikipedia.org/wiki/ISO_3166-1_alpha-2)) of where the player was located. - `drawing`: A simplified drawing represented as a dictionary where `x` and `y` are the pixel coordinates. The simplification processed is described in the `Data Instances` section. #### `preprocessed_bitmaps` (default configuration) - `image`: A `PIL.Image.Image` object containing the 28x28 grayscale bitmap. Note that when accessing the image column: `dataset[0]["image"]` the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the `"image"` column, *i.e.* `dataset[0]["image"]` should **always** be preferred over `dataset["image"][0]`. - `label`: Category the player was prompted to draw. <details> <summary> Click here to see the full class labels mapping: </summary> |id|class| |---|---| |0|aircraft carrier| |1|airplane| |2|alarm clock| |3|ambulance| |4|angel| |5|animal migration| |6|ant| |7|anvil| |8|apple| |9|arm| |10|asparagus| |11|axe| |12|backpack| |13|banana| |14|bandage| |15|barn| |16|baseball bat| |17|baseball| |18|basket| |19|basketball| |20|bat| |21|bathtub| |22|beach| |23|bear| |24|beard| |25|bed| |26|bee| |27|belt| |28|bench| |29|bicycle| |30|binoculars| |31|bird| |32|birthday cake| |33|blackberry| |34|blueberry| |35|book| |36|boomerang| |37|bottlecap| |38|bowtie| |39|bracelet| |40|brain| |41|bread| |42|bridge| |43|broccoli| |44|broom| |45|bucket| |46|bulldozer| |47|bus| |48|bush| |49|butterfly| |50|cactus| |51|cake| |52|calculator| |53|calendar| |54|camel| |55|camera| |56|camouflage| |57|campfire| |58|candle| |59|cannon| |60|canoe| |61|car| |62|carrot| |63|castle| |64|cat| |65|ceiling fan| |66|cell phone| |67|cello| |68|chair| |69|chandelier| |70|church| |71|circle| |72|clarinet| |73|clock| |74|cloud| |75|coffee cup| |76|compass| |77|computer| |78|cookie| |79|cooler| |80|couch| |81|cow| |82|crab| |83|crayon| |84|crocodile| |85|crown| |86|cruise ship| |87|cup| |88|diamond| |89|dishwasher| |90|diving board| |91|dog| |92|dolphin| |93|donut| |94|door| |95|dragon| |96|dresser| |97|drill| |98|drums| |99|duck| |100|dumbbell| |101|ear| |102|elbow| |103|elephant| |104|envelope| |105|eraser| |106|eye| |107|eyeglasses| |108|face| |109|fan| |110|feather| |111|fence| |112|finger| |113|fire hydrant| |114|fireplace| |115|firetruck| |116|fish| |117|flamingo| |118|flashlight| |119|flip flops| |120|floor lamp| |121|flower| |122|flying saucer| |123|foot| |124|fork| |125|frog| |126|frying pan| |127|garden hose| |128|garden| |129|giraffe| |130|goatee| |131|golf club| |132|grapes| |133|grass| |134|guitar| |135|hamburger| |136|hammer| |137|hand| |138|harp| |139|hat| |140|headphones| |141|hedgehog| |142|helicopter| |143|helmet| |144|hexagon| |145|hockey puck| |146|hockey stick| |147|horse| |148|hospital| |149|hot air balloon| |150|hot dog| |151|hot tub| |152|hourglass| |153|house plant| |154|house| |155|hurricane| |156|ice cream| |157|jacket| |158|jail| |159|kangaroo| |160|key| |161|keyboard| |162|knee| |163|knife| |164|ladder| |165|lantern| |166|laptop| |167|leaf| |168|leg| |169|light bulb| |170|lighter| |171|lighthouse| |172|lightning| |173|line| |174|lion| |175|lipstick| |176|lobster| |177|lollipop| |178|mailbox| |179|map| |180|marker| |181|matches| |182|megaphone| |183|mermaid| |184|microphone| |185|microwave| |186|monkey| |187|moon| |188|mosquito| |189|motorbike| |190|mountain| |191|mouse| |192|moustache| |193|mouth| |194|mug| |195|mushroom| |196|nail| |197|necklace| |198|nose| |199|ocean| |200|octagon| |201|octopus| |202|onion| |203|oven| |204|owl| |205|paint can| |206|paintbrush| |207|palm tree| |208|panda| |209|pants| |210|paper clip| |211|parachute| |212|parrot| |213|passport| |214|peanut| |215|pear| |216|peas| |217|pencil| |218|penguin| |219|piano| |220|pickup truck| |221|picture frame| |222|pig| |223|pillow| |224|pineapple| |225|pizza| |226|pliers| |227|police car| |228|pond| |229|pool| |230|popsicle| |231|postcard| |232|potato| |233|power outlet| |234|purse| |235|rabbit| |236|raccoon| |237|radio| |238|rain| |239|rainbow| |240|rake| |241|remote control| |242|rhinoceros| |243|rifle| |244|river| |245|roller coaster| |246|rollerskates| |247|sailboat| |248|sandwich| |249|saw| |250|saxophone| |251|school bus| |252|scissors| |253|scorpion| |254|screwdriver| |255|sea turtle| |256|see saw| |257|shark| |258|sheep| |259|shoe| |260|shorts| |261|shovel| |262|sink| |263|skateboard| |264|skull| |265|skyscraper| |266|sleeping bag| |267|smiley face| |268|snail| |269|snake| |270|snorkel| |271|snowflake| |272|snowman| |273|soccer ball| |274|sock| |275|speedboat| |276|spider| |277|spoon| |278|spreadsheet| |279|square| |280|squiggle| |281|squirrel| |282|stairs| |283|star| |284|steak| |285|stereo| |286|stethoscope| |287|stitches| |288|stop sign| |289|stove| |290|strawberry| |291|streetlight| |292|string bean| |293|submarine| |294|suitcase| |295|sun| |296|swan| |297|sweater| |298|swing set| |299|sword| |300|syringe| |301|t-shirt| |302|table| |303|teapot| |304|teddy-bear| |305|telephone| |306|television| |307|tennis racquet| |308|tent| |309|The Eiffel Tower| |310|The Great Wall of China| |311|The Mona Lisa| |312|tiger| |313|toaster| |314|toe| |315|toilet| |316|tooth| |317|toothbrush| |318|toothpaste| |319|tornado| |320|tractor| |321|traffic light| |322|train| |323|tree| |324|triangle| |325|trombone| |326|truck| |327|trumpet| |328|umbrella| |329|underwear| |330|van| |331|vase| |332|violin| |333|washing machine| |334|watermelon| |335|waterslide| |336|whale| |337|wheel| |338|windmill| |339|wine bottle| |340|wine glass| |341|wristwatch| |342|yoga| |343|zebra| |344|zigzag| </details> #### `sketch_rnn` and `sketch_rnn_full` - `word`: Category the player was prompted to draw. - `drawing`: An array of strokes. Strokes are represented as 3-tuples consisting of x-offset, y-offset, and a binary variable which is 1 if the pen is lifted between this position and the next, and 0 otherwise. <details> <summary> Click here to see the code for visualizing drawings in Jupyter Notebook or Google Colab: </summary> ```python import numpy as np import svgwrite # pip install svgwrite from IPython.display import SVG, display def draw_strokes(drawing, factor=0.045): """Displays vector drawing as SVG. Args: drawing: a list of strokes represented as 3-tuples factor: scaling factor. The smaller the scaling factor, the bigger the SVG picture and vice versa. """ def get_bounds(data, factor): """Return bounds of data.""" min_x = 0 max_x = 0 min_y = 0 max_y = 0 abs_x = 0 abs_y = 0 for i in range(len(data)): x = float(data[i, 0]) / factor y = float(data[i, 1]) / factor abs_x += x abs_y += y min_x = min(min_x, abs_x) min_y = min(min_y, abs_y) max_x = max(max_x, abs_x) max_y = max(max_y, abs_y) return (min_x, max_x, min_y, max_y) data = np.array(drawing) min_x, max_x, min_y, max_y = get_bounds(data, factor) dims = (50 + max_x - min_x, 50 + max_y - min_y) dwg = svgwrite.Drawing(size=dims) dwg.add(dwg.rect(insert=(0, 0), size=dims,fill='white')) lift_pen = 1 abs_x = 25 - min_x abs_y = 25 - min_y p = "M%s,%s " % (abs_x, abs_y) command = "m" for i in range(len(data)): if (lift_pen == 1): command = "m" elif (command != "l"): command = "l" else: command = "" x = float(data[i,0])/factor y = float(data[i,1])/factor lift_pen = data[i, 2] p += command+str(x)+","+str(y)+" " the_color = "black" stroke_width = 1 dwg.add(dwg.path(p).stroke(the_color,stroke_width).fill("none")) display(SVG(dwg.tostring())) ``` </details> > **Note**: Sketch-RNN takes for input strokes represented as 5-tuples with drawings padded to a common maximum length and prefixed by the special start token `[0, 0, 1, 0, 0]`. The 5-tuple representation consists of x-offset, y-offset, and p_1, p_2, p_3, a binary one-hot vector of 3 possible pen states: pen down, pen up, end of sketch. More precisely, the first two elements are the offset distance in the x and y directions of the pen from the previous point. The last 3 elements represents a binary one-hot vector of 3 possible states. The first pen state, p1, indicates that the pen is currently touching the paper, and that a line will be drawn connecting the next point with the current point. The second pen state, p2, indicates that the pen will be lifted from the paper after the current point, and that no line will be drawn next. The final pen state, p3, indicates that the drawing has ended, and subsequent points, including the current point, will not be rendered. ><details> > <summary> > Click here to see the code for converting drawings to Sketch-RNN input format: > </summary> > > ```python > def to_sketch_rnn_format(drawing, max_len): > """Converts a drawing to Sketch-RNN input format. > > Args: > drawing: a list of strokes represented as 3-tuples > max_len: maximum common length of all drawings > > Returns: > NumPy array > """ > drawing = np.array(drawing) > result = np.zeros((max_len, 5), dtype=float) > l = len(drawing) > assert l <= max_len > result[0:l, 0:2] = drawing[:, 0:2] > result[0:l, 3] = drawing[:, 2] > result[0:l, 2] = 1 - result[0:l, 3] > result[l:, 4] = 1 > # Prepend special start token > result = np.vstack([[0, 0, 1, 0, 0], result]) > return result > ``` > ></details> ### Data Splits In the configurations `raw`, `preprocessed_simplified_drawings` and `preprocessed_bitamps` (default configuration), all the data is contained in the training set, which has 50426266 examples. `sketch_rnn` and `sketch_rnn_full` have the data split into training, validation and test split. In the `sketch_rnn` configuration, 75K samples (70K Training, 2.5K Validation, 2.5K Test) have been randomly selected from each category. Therefore, the training set contains 24150000 examples, the validation set 862500 examples and the test set 862500 examples. The `sketch_rnn_full` configuration has the full (training) data for each category, which leads to the training set having 43988874 examples, the validation set 862500 and the test set 862500 examples. ## Dataset Creation ### Curation Rationale From the GitHub repository: > The Quick Draw Dataset is a collection of 50 million drawings across [345 categories](categories.txt), contributed by players of the game [Quick, Draw!](https://quickdraw.withgoogle.com). The drawings were captured as timestamped vectors, tagged with metadata including what the player was asked to draw and in which country the player was located. You can browse the recognized drawings on [quickdraw.withgoogle.com/data](https://quickdraw.withgoogle.com/data). > > We're sharing them here for developers, researchers, and artists to explore, study, and learn from ### Source Data #### Initial Data Collection and Normalization This dataset contains vector drawings obtained from [Quick, Draw!](https://quickdraw.withgoogle.com/), an online game where the players are asked to draw objects belonging to a particular object class in less than 20 seconds. #### Who are the source language producers? The participants in the [Quick, Draw!](https://quickdraw.withgoogle.com/) game. ### Annotations #### Annotation process The annotations are machine-generated and match the category the player was prompted to draw. #### Who are the annotators? The annotations are machine-generated. ### Personal and Sensitive Information Some sketches are known to be problematic (see https://github.com/googlecreativelab/quickdraw-dataset/issues/74 and https://github.com/googlecreativelab/quickdraw-dataset/issues/18). ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations ## Additional Information ### Dataset Curators Jonas Jongejan, Henry Rowley, Takashi Kawashima, Jongmin Kim and Nick Fox-Gieg. ### Licensing Information The data is made available by Google, Inc. under the [Creative Commons Attribution 4.0 International](https://creativecommons.org/licenses/by/4.0/) license. ### Citation Information ```bibtex @article{DBLP:journals/corr/HaE17, author = {David Ha and Douglas Eck}, title = {A Neural Representation of Sketch Drawings}, journal = {CoRR}, volume = {abs/1704.03477}, year = {2017}, url = {http://arxiv.org/abs/1704.03477}, archivePrefix = {arXiv}, eprint = {1704.03477}, timestamp = {Mon, 13 Aug 2018 16:48:30 +0200}, biburl = {https://dblp.org/rec/bib/journals/corr/HaE17}, bibsource = {dblp computer science bibliography, https://dblp.org} } ``` ### Contributions Thanks to [@mariosasko](https://github.com/mariosasko) for adding this dataset.
quickdraw
[ "task_categories:image-classification", "task_ids:multi-class-image-classification", "annotations_creators:machine-generated", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:10M<n<100M", "source_datasets:original", "language:en", "license:cc-by-4.0", "arxiv:1704.03477", "region:us" ]
2022-06-09T08:56:43+00:00
{"annotations_creators": ["machine-generated"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["cc-by-4.0"], "multilinguality": ["monolingual"], "size_categories": ["10M<n<100M"], "source_datasets": ["original"], "task_categories": ["image-classification"], "task_ids": ["multi-class-image-classification"], "paperswithcode_id": "quick-draw-dataset", "pretty_name": "Quick, Draw!", "dataset_info": [{"config_name": "raw", "features": [{"name": "key_id", "dtype": "string"}, {"name": "word", "dtype": {"class_label": {"names": {"0": "aircraft carrier", "1": "airplane", "2": "alarm clock", "3": "ambulance", "4": "angel", "5": "animal migration", "6": "ant", "7": "anvil", "8": "apple", "9": "arm", "10": "asparagus", "11": "axe", "12": "backpack", "13": "banana", "14": "bandage", "15": "barn", "16": "baseball bat", "17": "baseball", "18": "basket", "19": "basketball", "20": "bat", "21": "bathtub", "22": "beach", "23": "bear", "24": "beard", "25": "bed", "26": "bee", "27": "belt", "28": "bench", "29": "bicycle", "30": "binoculars", "31": "bird", "32": "birthday cake", "33": "blackberry", "34": "blueberry", "35": "book", "36": "boomerang", "37": "bottlecap", "38": "bowtie", "39": "bracelet", "40": "brain", "41": "bread", "42": "bridge", "43": "broccoli", "44": "broom", "45": "bucket", "46": "bulldozer", "47": "bus", "48": "bush", "49": "butterfly", "50": "cactus", "51": "cake", "52": "calculator", "53": "calendar", "54": "camel", "55": "camera", "56": "camouflage", "57": "campfire", "58": "candle", "59": "cannon", "60": "canoe", "61": "car", "62": "carrot", "63": "castle", "64": "cat", "65": "ceiling fan", "66": "cell phone", "67": "cello", "68": "chair", "69": "chandelier", "70": "church", "71": "circle", "72": "clarinet", "73": "clock", "74": "cloud", "75": "coffee cup", "76": "compass", "77": "computer", "78": "cookie", "79": "cooler", "80": "couch", "81": "cow", "82": "crab", "83": "crayon", "84": "crocodile", "85": "crown", "86": "cruise ship", "87": "cup", "88": "diamond", "89": "dishwasher", "90": "diving board", "91": "dog", "92": "dolphin", "93": "donut", "94": "door", "95": "dragon", "96": "dresser", "97": "drill", "98": "drums", "99": "duck", "100": "dumbbell", "101": "ear", "102": "elbow", "103": "elephant", "104": "envelope", "105": "eraser", "106": "eye", "107": "eyeglasses", "108": "face", "109": "fan", "110": "feather", "111": "fence", "112": "finger", "113": "fire hydrant", "114": "fireplace", "115": "firetruck", "116": "fish", "117": "flamingo", "118": "flashlight", "119": "flip flops", "120": "floor lamp", "121": "flower", "122": "flying saucer", "123": "foot", "124": "fork", "125": "frog", "126": "frying pan", "127": "garden hose", "128": "garden", "129": "giraffe", "130": "goatee", "131": "golf club", "132": "grapes", "133": "grass", "134": "guitar", "135": "hamburger", "136": "hammer", "137": "hand", "138": "harp", "139": "hat", "140": "headphones", "141": "hedgehog", "142": "helicopter", "143": "helmet", "144": "hexagon", "145": "hockey puck", "146": "hockey stick", "147": "horse", "148": "hospital", "149": "hot air balloon", "150": "hot dog", "151": "hot tub", "152": "hourglass", "153": "house plant", "154": "house", "155": "hurricane", "156": "ice cream", "157": "jacket", "158": "jail", "159": "kangaroo", "160": "key", "161": "keyboard", "162": "knee", "163": "knife", "164": "ladder", "165": "lantern", "166": "laptop", "167": "leaf", "168": "leg", "169": "light bulb", "170": "lighter", "171": "lighthouse", "172": "lightning", "173": "line", "174": "lion", "175": "lipstick", "176": "lobster", "177": "lollipop", "178": "mailbox", "179": "map", "180": "marker", "181": "matches", "182": "megaphone", "183": "mermaid", "184": "microphone", "185": "microwave", "186": "monkey", "187": "moon", "188": "mosquito", "189": "motorbike", "190": "mountain", "191": "mouse", "192": "moustache", "193": "mouth", "194": "mug", "195": "mushroom", "196": "nail", "197": "necklace", "198": "nose", "199": "ocean", "200": "octagon", "201": "octopus", "202": "onion", "203": "oven", "204": "owl", "205": "paint can", "206": "paintbrush", "207": "palm tree", "208": "panda", "209": "pants", "210": "paper clip", "211": "parachute", "212": "parrot", "213": "passport", "214": "peanut", "215": "pear", "216": "peas", "217": "pencil", "218": "penguin", "219": "piano", "220": "pickup truck", "221": "picture frame", "222": "pig", "223": "pillow", "224": "pineapple", "225": "pizza", "226": "pliers", "227": "police car", "228": "pond", "229": "pool", "230": "popsicle", "231": "postcard", "232": "potato", "233": "power outlet", "234": "purse", "235": "rabbit", "236": "raccoon", "237": "radio", "238": "rain", "239": "rainbow", "240": "rake", "241": "remote control", "242": "rhinoceros", "243": "rifle", "244": "river", "245": "roller coaster", "246": "rollerskates", "247": "sailboat", "248": "sandwich", "249": "saw", "250": "saxophone", "251": "school bus", "252": "scissors", "253": "scorpion", "254": "screwdriver", "255": "sea turtle", "256": "see saw", "257": "shark", "258": "sheep", "259": "shoe", "260": "shorts", "261": "shovel", "262": "sink", "263": "skateboard", "264": "skull", "265": "skyscraper", "266": "sleeping bag", "267": "smiley face", "268": "snail", "269": "snake", "270": "snorkel", "271": "snowflake", "272": "snowman", "273": "soccer ball", "274": "sock", "275": "speedboat", "276": "spider", "277": "spoon", "278": "spreadsheet", "279": "square", "280": "squiggle", "281": "squirrel", "282": "stairs", "283": "star", "284": "steak", "285": "stereo", "286": "stethoscope", "287": "stitches", "288": "stop sign", "289": "stove", "290": "strawberry", "291": "streetlight", "292": "string bean", "293": "submarine", "294": "suitcase", "295": "sun", "296": "swan", "297": "sweater", "298": "swing set", "299": "sword", "300": "syringe", "301": "t-shirt", "302": "table", "303": "teapot", "304": "teddy-bear", "305": "telephone", "306": "television", "307": "tennis racquet", "308": "tent", "309": "The Eiffel Tower", "310": "The Great Wall of China", "311": "The Mona Lisa", "312": "tiger", "313": "toaster", "314": "toe", "315": "toilet", "316": "tooth", "317": "toothbrush", "318": "toothpaste", "319": "tornado", "320": "tractor", "321": "traffic light", "322": "train", "323": "tree", "324": "triangle", "325": "trombone", "326": "truck", "327": "trumpet", "328": "umbrella", "329": "underwear", "330": "van", "331": "vase", "332": "violin", "333": "washing machine", "334": "watermelon", "335": "waterslide", "336": "whale", "337": "wheel", "338": "windmill", "339": "wine bottle", "340": "wine glass", "341": "wristwatch", "342": "yoga", "343": "zebra", "344": "zigzag"}}}}, {"name": "recognized", "dtype": "bool"}, {"name": "timestamp", "dtype": "timestamp[us, tz=UTC]"}, {"name": "countrycode", "dtype": "string"}, {"name": "drawing", "sequence": [{"name": "x", "sequence": "float32"}, {"name": "y", "sequence": "float32"}, {"name": "t", "sequence": "int32"}]}], "splits": [{"name": "train", "num_bytes": 134763164880, "num_examples": 50426266}], "download_size": 194810597157, "dataset_size": 134763164880}, {"config_name": "preprocessed_simplified_drawings", "features": [{"name": "key_id", "dtype": "string"}, {"name": "word", "dtype": {"class_label": {"names": {"0": "aircraft carrier", "1": "airplane", "2": "alarm clock", "3": "ambulance", "4": "angel", "5": "animal migration", "6": "ant", "7": "anvil", "8": "apple", "9": "arm", "10": "asparagus", "11": "axe", "12": "backpack", "13": "banana", "14": "bandage", "15": "barn", "16": "baseball bat", "17": "baseball", "18": "basket", "19": "basketball", "20": "bat", "21": "bathtub", "22": "beach", "23": "bear", "24": "beard", "25": "bed", "26": "bee", "27": "belt", "28": "bench", "29": "bicycle", "30": "binoculars", "31": "bird", "32": "birthday cake", "33": "blackberry", "34": "blueberry", "35": "book", "36": "boomerang", "37": "bottlecap", "38": "bowtie", "39": "bracelet", "40": "brain", "41": "bread", "42": "bridge", "43": "broccoli", "44": "broom", "45": "bucket", "46": "bulldozer", "47": "bus", "48": "bush", "49": "butterfly", "50": "cactus", "51": "cake", "52": "calculator", "53": "calendar", "54": "camel", "55": "camera", "56": "camouflage", "57": "campfire", "58": "candle", "59": "cannon", "60": "canoe", "61": "car", "62": "carrot", "63": "castle", "64": "cat", "65": "ceiling fan", "66": "cell phone", "67": "cello", "68": "chair", "69": "chandelier", "70": "church", "71": "circle", "72": "clarinet", "73": "clock", "74": "cloud", "75": "coffee cup", "76": "compass", "77": "computer", "78": "cookie", "79": "cooler", "80": "couch", "81": "cow", "82": "crab", "83": "crayon", "84": "crocodile", "85": "crown", "86": "cruise ship", "87": "cup", "88": "diamond", "89": "dishwasher", "90": "diving board", "91": "dog", "92": "dolphin", "93": "donut", "94": "door", "95": "dragon", "96": "dresser", "97": "drill", "98": "drums", "99": "duck", "100": "dumbbell", "101": "ear", "102": "elbow", "103": "elephant", "104": "envelope", "105": "eraser", "106": "eye", "107": "eyeglasses", "108": "face", "109": "fan", "110": "feather", "111": "fence", "112": "finger", "113": "fire hydrant", "114": "fireplace", "115": "firetruck", "116": "fish", "117": "flamingo", "118": "flashlight", "119": "flip flops", "120": "floor lamp", "121": "flower", "122": "flying saucer", "123": "foot", "124": "fork", "125": "frog", "126": "frying pan", "127": "garden hose", "128": "garden", "129": "giraffe", "130": "goatee", "131": "golf club", "132": "grapes", "133": "grass", "134": "guitar", "135": "hamburger", "136": "hammer", "137": "hand", "138": "harp", "139": "hat", "140": "headphones", "141": "hedgehog", "142": "helicopter", "143": "helmet", "144": "hexagon", "145": "hockey puck", "146": "hockey stick", "147": "horse", "148": "hospital", "149": "hot air balloon", "150": "hot dog", "151": "hot tub", "152": "hourglass", "153": "house plant", "154": "house", "155": "hurricane", "156": "ice cream", "157": "jacket", "158": "jail", "159": "kangaroo", "160": "key", "161": "keyboard", "162": "knee", "163": "knife", "164": "ladder", "165": "lantern", "166": "laptop", "167": "leaf", "168": "leg", "169": "light bulb", "170": "lighter", "171": "lighthouse", "172": "lightning", "173": "line", "174": "lion", "175": "lipstick", "176": "lobster", "177": "lollipop", "178": "mailbox", "179": "map", "180": "marker", "181": "matches", "182": "megaphone", "183": "mermaid", "184": "microphone", "185": "microwave", "186": "monkey", "187": "moon", "188": "mosquito", "189": "motorbike", "190": "mountain", "191": "mouse", "192": "moustache", "193": "mouth", "194": "mug", "195": "mushroom", "196": "nail", "197": "necklace", "198": "nose", "199": "ocean", "200": "octagon", "201": "octopus", "202": "onion", "203": "oven", "204": "owl", "205": "paint can", "206": "paintbrush", "207": "palm tree", "208": "panda", "209": "pants", "210": "paper clip", "211": "parachute", "212": "parrot", "213": "passport", "214": "peanut", "215": "pear", "216": "peas", "217": "pencil", "218": "penguin", "219": "piano", "220": "pickup truck", "221": "picture frame", "222": "pig", "223": "pillow", "224": "pineapple", "225": "pizza", "226": "pliers", "227": "police car", "228": "pond", "229": "pool", "230": "popsicle", "231": "postcard", "232": "potato", "233": "power outlet", "234": "purse", "235": "rabbit", "236": "raccoon", "237": "radio", "238": "rain", "239": "rainbow", "240": "rake", "241": "remote control", "242": "rhinoceros", "243": "rifle", "244": "river", "245": "roller coaster", "246": "rollerskates", "247": "sailboat", "248": "sandwich", "249": "saw", "250": "saxophone", "251": "school bus", "252": "scissors", "253": "scorpion", "254": "screwdriver", "255": "sea turtle", "256": "see saw", "257": "shark", "258": "sheep", "259": "shoe", "260": "shorts", "261": "shovel", "262": "sink", "263": "skateboard", "264": "skull", "265": "skyscraper", "266": "sleeping bag", "267": "smiley face", "268": "snail", "269": "snake", "270": "snorkel", "271": "snowflake", "272": "snowman", "273": "soccer ball", "274": "sock", "275": "speedboat", "276": "spider", "277": "spoon", "278": "spreadsheet", "279": "square", "280": "squiggle", "281": "squirrel", "282": "stairs", "283": "star", "284": "steak", "285": "stereo", "286": "stethoscope", "287": "stitches", "288": "stop sign", "289": "stove", "290": "strawberry", "291": "streetlight", "292": "string bean", "293": "submarine", "294": "suitcase", "295": "sun", "296": "swan", "297": "sweater", "298": "swing set", "299": "sword", "300": "syringe", "301": "t-shirt", "302": "table", "303": "teapot", "304": "teddy-bear", "305": "telephone", "306": "television", "307": "tennis racquet", "308": "tent", "309": "The Eiffel Tower", "310": "The Great Wall of China", "311": "The Mona Lisa", "312": "tiger", "313": "toaster", "314": "toe", "315": "toilet", "316": "tooth", "317": "toothbrush", "318": "toothpaste", "319": "tornado", "320": "tractor", "321": "traffic light", "322": "train", "323": "tree", "324": "triangle", "325": "trombone", "326": "truck", "327": "trumpet", "328": "umbrella", "329": "underwear", "330": "van", "331": "vase", "332": "violin", "333": "washing machine", "334": "watermelon", "335": "waterslide", "336": "whale", "337": "wheel", "338": "windmill", "339": "wine bottle", "340": "wine glass", "341": "wristwatch", "342": "yoga", "343": "zebra", "344": "zigzag"}}}}, {"name": "recognized", "dtype": "bool"}, {"name": "timestamp", "dtype": "timestamp[us, tz=UTC]"}, {"name": "countrycode", "dtype": "string"}, {"name": "drawing", "sequence": [{"name": "x", "sequence": "uint8"}, {"name": "y", "sequence": "uint8"}]}], "splits": [{"name": "train", "num_bytes": 9741454188, "num_examples": 50426266}], "download_size": 5889968422, "dataset_size": 9741454188}, {"config_name": "preprocessed_bitmaps", "features": [{"name": "image", "dtype": "image"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "aircraft carrier", "1": "airplane", "2": "alarm clock", "3": "ambulance", "4": "angel", "5": "animal migration", "6": "ant", "7": "anvil", "8": "apple", "9": "arm", "10": "asparagus", "11": "axe", "12": "backpack", "13": "banana", "14": "bandage", "15": "barn", "16": "baseball bat", "17": "baseball", "18": "basket", "19": "basketball", "20": "bat", "21": "bathtub", "22": "beach", "23": "bear", "24": "beard", "25": "bed", "26": "bee", "27": "belt", "28": "bench", "29": "bicycle", "30": "binoculars", "31": "bird", "32": "birthday cake", "33": "blackberry", "34": "blueberry", "35": "book", "36": "boomerang", "37": "bottlecap", "38": "bowtie", "39": "bracelet", "40": "brain", "41": "bread", "42": "bridge", "43": "broccoli", "44": "broom", "45": "bucket", "46": "bulldozer", "47": "bus", "48": "bush", "49": "butterfly", "50": "cactus", "51": "cake", "52": "calculator", "53": "calendar", "54": "camel", "55": "camera", "56": "camouflage", "57": "campfire", "58": "candle", "59": "cannon", "60": "canoe", "61": "car", "62": "carrot", "63": "castle", "64": "cat", "65": "ceiling fan", "66": "cell phone", "67": "cello", "68": "chair", "69": "chandelier", "70": "church", "71": "circle", "72": "clarinet", "73": "clock", "74": "cloud", "75": "coffee cup", "76": "compass", "77": "computer", "78": "cookie", "79": "cooler", "80": "couch", "81": "cow", "82": "crab", "83": "crayon", "84": "crocodile", "85": "crown", "86": "cruise ship", "87": "cup", "88": "diamond", "89": "dishwasher", "90": "diving board", "91": "dog", "92": "dolphin", "93": "donut", "94": "door", "95": "dragon", "96": "dresser", "97": "drill", "98": "drums", "99": "duck", "100": "dumbbell", "101": "ear", "102": "elbow", "103": "elephant", "104": "envelope", "105": "eraser", "106": "eye", "107": "eyeglasses", "108": "face", "109": "fan", "110": "feather", "111": "fence", "112": "finger", "113": "fire hydrant", "114": "fireplace", "115": "firetruck", "116": "fish", "117": "flamingo", "118": "flashlight", "119": "flip flops", "120": "floor lamp", "121": "flower", "122": "flying saucer", "123": "foot", "124": "fork", "125": "frog", "126": "frying pan", "127": "garden hose", "128": "garden", "129": "giraffe", "130": "goatee", "131": "golf club", "132": "grapes", "133": "grass", "134": "guitar", "135": "hamburger", "136": "hammer", "137": "hand", "138": "harp", "139": "hat", "140": "headphones", "141": "hedgehog", "142": "helicopter", "143": "helmet", "144": "hexagon", "145": "hockey puck", "146": "hockey stick", "147": "horse", "148": "hospital", "149": "hot air balloon", "150": "hot dog", "151": "hot tub", "152": "hourglass", "153": "house plant", "154": "house", "155": "hurricane", "156": "ice cream", "157": "jacket", "158": "jail", "159": "kangaroo", "160": "key", "161": "keyboard", "162": "knee", "163": "knife", "164": "ladder", "165": "lantern", "166": "laptop", "167": "leaf", "168": "leg", "169": "light bulb", "170": "lighter", "171": "lighthouse", "172": "lightning", "173": "line", "174": "lion", "175": "lipstick", "176": "lobster", "177": "lollipop", "178": "mailbox", "179": "map", "180": "marker", "181": "matches", "182": "megaphone", "183": "mermaid", "184": "microphone", "185": "microwave", "186": "monkey", "187": "moon", "188": "mosquito", "189": "motorbike", "190": "mountain", "191": "mouse", "192": "moustache", "193": "mouth", "194": "mug", "195": "mushroom", "196": "nail", "197": "necklace", "198": "nose", "199": "ocean", "200": "octagon", "201": "octopus", "202": "onion", "203": "oven", "204": "owl", "205": "paint can", "206": "paintbrush", "207": "palm tree", "208": "panda", "209": "pants", "210": "paper clip", "211": "parachute", "212": "parrot", "213": "passport", "214": "peanut", "215": "pear", "216": "peas", "217": "pencil", "218": "penguin", "219": "piano", "220": "pickup truck", "221": "picture frame", "222": "pig", "223": "pillow", "224": "pineapple", "225": "pizza", "226": "pliers", "227": "police car", "228": "pond", "229": "pool", "230": "popsicle", "231": "postcard", "232": "potato", "233": "power outlet", "234": "purse", "235": "rabbit", "236": "raccoon", "237": "radio", "238": "rain", "239": "rainbow", "240": "rake", "241": "remote control", "242": "rhinoceros", "243": "rifle", "244": "river", "245": "roller coaster", "246": "rollerskates", "247": "sailboat", "248": "sandwich", "249": "saw", "250": "saxophone", "251": "school bus", "252": "scissors", "253": "scorpion", "254": "screwdriver", "255": "sea turtle", "256": "see saw", "257": "shark", "258": "sheep", "259": "shoe", "260": "shorts", "261": "shovel", "262": "sink", "263": "skateboard", "264": "skull", "265": "skyscraper", "266": "sleeping bag", "267": "smiley face", "268": "snail", "269": "snake", "270": "snorkel", "271": "snowflake", "272": "snowman", "273": "soccer ball", "274": "sock", "275": "speedboat", "276": "spider", "277": "spoon", "278": "spreadsheet", "279": "square", "280": "squiggle", "281": "squirrel", "282": "stairs", "283": "star", "284": "steak", "285": "stereo", "286": "stethoscope", "287": "stitches", "288": "stop sign", "289": "stove", "290": "strawberry", "291": "streetlight", "292": "string bean", "293": "submarine", "294": "suitcase", "295": "sun", "296": "swan", "297": "sweater", "298": "swing set", "299": "sword", "300": "syringe", "301": "t-shirt", "302": "table", "303": "teapot", "304": "teddy-bear", "305": "telephone", "306": "television", "307": "tennis racquet", "308": "tent", "309": "The Eiffel Tower", "310": "The Great Wall of China", "311": "The Mona Lisa", "312": "tiger", "313": "toaster", "314": "toe", "315": "toilet", "316": "tooth", "317": "toothbrush", "318": "toothpaste", "319": "tornado", "320": "tractor", "321": "traffic light", "322": "train", "323": "tree", "324": "triangle", "325": "trombone", "326": "truck", "327": "trumpet", "328": "umbrella", "329": "underwear", "330": "van", "331": "vase", "332": "violin", "333": "washing machine", "334": "watermelon", "335": "waterslide", "336": "whale", "337": "wheel", "338": "windmill", "339": "wine bottle", "340": "wine glass", "341": "wristwatch", "342": "yoga", "343": "zebra", "344": "zigzag"}}}}], "splits": [{"name": "train", "num_bytes": 20372624628, "num_examples": 50426266}], "download_size": 39534220144, "dataset_size": 20372624628}, {"config_name": "sketch_rnn", "features": [{"name": "word", "dtype": {"class_label": {"names": {"0": "aircraft carrier", "1": "airplane", "2": "alarm clock", "3": "ambulance", "4": "angel", "5": "animal migration", "6": "ant", "7": "anvil", "8": "apple", "9": "arm", "10": "asparagus", "11": "axe", "12": "backpack", "13": "banana", "14": "bandage", "15": "barn", "16": "baseball bat", "17": "baseball", "18": "basket", "19": "basketball", "20": "bat", "21": "bathtub", "22": "beach", "23": "bear", "24": "beard", "25": "bed", "26": "bee", "27": "belt", "28": "bench", "29": "bicycle", "30": "binoculars", "31": "bird", "32": "birthday cake", "33": "blackberry", "34": "blueberry", "35": "book", "36": "boomerang", "37": "bottlecap", "38": "bowtie", "39": "bracelet", "40": "brain", "41": "bread", "42": "bridge", "43": "broccoli", "44": "broom", "45": "bucket", "46": "bulldozer", "47": "bus", "48": "bush", "49": "butterfly", "50": "cactus", "51": "cake", "52": "calculator", "53": "calendar", "54": "camel", "55": "camera", "56": "camouflage", "57": "campfire", "58": "candle", "59": "cannon", "60": "canoe", "61": "car", "62": "carrot", "63": "castle", "64": "cat", "65": "ceiling fan", "66": "cell phone", "67": "cello", "68": "chair", "69": "chandelier", "70": "church", "71": "circle", "72": "clarinet", "73": "clock", "74": "cloud", "75": "coffee cup", "76": "compass", "77": "computer", "78": "cookie", "79": "cooler", "80": "couch", "81": "cow", "82": "crab", "83": "crayon", "84": "crocodile", "85": "crown", "86": "cruise ship", "87": "cup", "88": "diamond", "89": "dishwasher", "90": "diving board", "91": "dog", "92": "dolphin", "93": "donut", "94": "door", "95": "dragon", "96": "dresser", "97": "drill", "98": "drums", "99": "duck", "100": "dumbbell", "101": "ear", "102": "elbow", "103": "elephant", "104": "envelope", "105": "eraser", "106": "eye", "107": "eyeglasses", "108": "face", "109": "fan", "110": "feather", "111": "fence", "112": "finger", "113": "fire hydrant", "114": "fireplace", "115": "firetruck", "116": "fish", "117": "flamingo", "118": "flashlight", "119": "flip flops", "120": "floor lamp", "121": "flower", "122": "flying saucer", "123": "foot", "124": "fork", "125": "frog", "126": "frying pan", "127": "garden hose", "128": "garden", "129": "giraffe", "130": "goatee", "131": "golf club", "132": "grapes", "133": "grass", "134": "guitar", "135": "hamburger", "136": "hammer", "137": "hand", "138": "harp", "139": "hat", "140": "headphones", "141": "hedgehog", "142": "helicopter", "143": "helmet", "144": "hexagon", "145": "hockey puck", "146": "hockey stick", "147": "horse", "148": "hospital", "149": "hot air balloon", "150": "hot dog", "151": "hot tub", "152": "hourglass", "153": "house plant", "154": "house", "155": "hurricane", "156": "ice cream", "157": "jacket", "158": "jail", "159": "kangaroo", "160": "key", "161": "keyboard", "162": "knee", "163": "knife", "164": "ladder", "165": "lantern", "166": "laptop", "167": "leaf", "168": "leg", "169": "light bulb", "170": "lighter", "171": "lighthouse", "172": "lightning", "173": "line", "174": "lion", "175": "lipstick", "176": "lobster", "177": "lollipop", "178": "mailbox", "179": "map", "180": "marker", "181": "matches", "182": "megaphone", "183": "mermaid", "184": "microphone", "185": "microwave", "186": "monkey", "187": "moon", "188": "mosquito", "189": "motorbike", "190": "mountain", "191": "mouse", "192": "moustache", "193": "mouth", "194": "mug", "195": "mushroom", "196": "nail", "197": "necklace", "198": "nose", "199": "ocean", "200": "octagon", "201": "octopus", "202": "onion", "203": "oven", "204": "owl", "205": "paint can", "206": "paintbrush", "207": "palm tree", "208": "panda", "209": "pants", "210": "paper clip", "211": "parachute", "212": "parrot", "213": "passport", "214": "peanut", "215": "pear", "216": "peas", "217": "pencil", "218": "penguin", "219": "piano", "220": "pickup truck", "221": "picture frame", "222": "pig", "223": "pillow", "224": "pineapple", "225": "pizza", "226": "pliers", "227": "police car", "228": "pond", "229": "pool", "230": "popsicle", "231": "postcard", "232": "potato", "233": "power outlet", "234": "purse", "235": "rabbit", "236": "raccoon", "237": "radio", "238": "rain", "239": "rainbow", "240": "rake", "241": "remote control", "242": "rhinoceros", "243": "rifle", "244": "river", "245": "roller coaster", "246": "rollerskates", "247": "sailboat", "248": "sandwich", "249": "saw", "250": "saxophone", "251": "school bus", "252": "scissors", "253": "scorpion", "254": "screwdriver", "255": "sea turtle", "256": "see saw", "257": "shark", "258": "sheep", "259": "shoe", "260": "shorts", "261": "shovel", "262": "sink", "263": "skateboard", "264": "skull", "265": "skyscraper", "266": "sleeping bag", "267": "smiley face", "268": "snail", "269": "snake", "270": "snorkel", "271": "snowflake", "272": "snowman", "273": "soccer ball", "274": "sock", "275": "speedboat", "276": "spider", "277": "spoon", "278": "spreadsheet", "279": "square", "280": "squiggle", "281": "squirrel", "282": "stairs", "283": "star", "284": "steak", "285": "stereo", "286": "stethoscope", "287": "stitches", "288": "stop sign", "289": "stove", "290": "strawberry", "291": "streetlight", "292": "string bean", "293": "submarine", "294": "suitcase", "295": "sun", "296": "swan", "297": "sweater", "298": "swing set", "299": "sword", "300": "syringe", "301": "t-shirt", "302": "table", "303": "teapot", "304": "teddy-bear", "305": "telephone", "306": "television", "307": "tennis racquet", "308": "tent", "309": "The Eiffel Tower", "310": "The Great Wall of China", "311": "The Mona Lisa", "312": "tiger", "313": "toaster", "314": "toe", "315": "toilet", "316": "tooth", "317": "toothbrush", "318": "toothpaste", "319": "tornado", "320": "tractor", "321": "traffic light", "322": "train", "323": "tree", "324": "triangle", "325": "trombone", "326": "truck", "327": "trumpet", "328": "umbrella", "329": "underwear", "330": "van", "331": "vase", "332": "violin", "333": "washing machine", "334": "watermelon", "335": "waterslide", "336": "whale", "337": "wheel", "338": "windmill", "339": "wine bottle", "340": "wine glass", "341": "wristwatch", "342": "yoga", "343": "zebra", "344": "zigzag"}}}}, {"name": "drawing", "dtype": {"array2_d": {"shape": [3], "dtype": "int16"}}}], "splits": [{"name": "train", "num_bytes": 13056229420, "num_examples": 24150000}, {"name": "validation", "num_bytes": 466485546, "num_examples": 862500}, {"name": "test", "num_bytes": 466191706, "num_examples": 862500}], "download_size": 3928904911, "dataset_size": 13988906672}, {"config_name": "sketch_rnn_full", "features": [{"name": "word", "dtype": {"class_label": {"names": {"0": "aircraft carrier", "1": "airplane", "2": "alarm clock", "3": "ambulance", "4": "angel", "5": "animal migration", "6": "ant", "7": "anvil", "8": "apple", "9": "arm", "10": "asparagus", "11": "axe", "12": "backpack", "13": "banana", "14": "bandage", "15": "barn", "16": "baseball bat", "17": "baseball", "18": "basket", "19": "basketball", "20": "bat", "21": "bathtub", "22": "beach", "23": "bear", "24": "beard", "25": "bed", "26": "bee", "27": "belt", "28": "bench", "29": "bicycle", "30": "binoculars", "31": "bird", "32": "birthday cake", "33": "blackberry", "34": "blueberry", "35": "book", "36": "boomerang", "37": "bottlecap", "38": "bowtie", "39": "bracelet", "40": "brain", "41": "bread", "42": "bridge", "43": "broccoli", "44": "broom", "45": "bucket", "46": "bulldozer", "47": "bus", "48": "bush", "49": "butterfly", "50": "cactus", "51": "cake", "52": "calculator", "53": "calendar", "54": "camel", "55": "camera", "56": "camouflage", "57": "campfire", "58": "candle", "59": "cannon", "60": "canoe", "61": "car", "62": "carrot", "63": "castle", "64": "cat", "65": "ceiling fan", "66": "cell phone", "67": "cello", "68": "chair", "69": "chandelier", "70": "church", "71": "circle", "72": "clarinet", "73": "clock", "74": "cloud", "75": "coffee cup", "76": "compass", "77": "computer", "78": "cookie", "79": "cooler", "80": "couch", "81": "cow", "82": "crab", "83": "crayon", "84": "crocodile", "85": "crown", "86": "cruise ship", "87": "cup", "88": "diamond", "89": "dishwasher", "90": "diving board", "91": "dog", "92": "dolphin", "93": "donut", "94": "door", "95": "dragon", "96": "dresser", "97": "drill", "98": "drums", "99": "duck", "100": "dumbbell", "101": "ear", "102": "elbow", "103": "elephant", "104": "envelope", "105": "eraser", "106": "eye", "107": "eyeglasses", "108": "face", "109": "fan", "110": "feather", "111": "fence", "112": "finger", "113": "fire hydrant", "114": "fireplace", "115": "firetruck", "116": "fish", "117": "flamingo", "118": "flashlight", "119": "flip flops", "120": "floor lamp", "121": "flower", "122": "flying saucer", "123": "foot", "124": "fork", "125": "frog", "126": "frying pan", "127": "garden hose", "128": "garden", "129": "giraffe", "130": "goatee", "131": "golf club", "132": "grapes", "133": "grass", "134": "guitar", "135": "hamburger", "136": "hammer", "137": "hand", "138": "harp", "139": "hat", "140": "headphones", "141": "hedgehog", "142": "helicopter", "143": "helmet", "144": "hexagon", "145": "hockey puck", "146": "hockey stick", "147": "horse", "148": "hospital", "149": "hot air balloon", "150": "hot dog", "151": "hot tub", "152": "hourglass", "153": "house plant", "154": "house", "155": "hurricane", "156": "ice cream", "157": "jacket", "158": "jail", "159": "kangaroo", "160": "key", "161": "keyboard", "162": "knee", "163": "knife", "164": "ladder", "165": "lantern", "166": "laptop", "167": "leaf", "168": "leg", "169": "light bulb", "170": "lighter", "171": "lighthouse", "172": "lightning", "173": "line", "174": "lion", "175": "lipstick", "176": "lobster", "177": "lollipop", "178": "mailbox", "179": "map", "180": "marker", "181": "matches", "182": "megaphone", "183": "mermaid", "184": "microphone", "185": "microwave", "186": "monkey", "187": "moon", "188": "mosquito", "189": "motorbike", "190": "mountain", "191": "mouse", "192": "moustache", "193": "mouth", "194": "mug", "195": "mushroom", "196": "nail", "197": "necklace", "198": "nose", "199": "ocean", "200": "octagon", "201": "octopus", "202": "onion", "203": "oven", "204": "owl", "205": "paint can", "206": "paintbrush", "207": "palm tree", "208": "panda", "209": "pants", "210": "paper clip", "211": "parachute", "212": "parrot", "213": "passport", "214": "peanut", "215": "pear", "216": "peas", "217": "pencil", "218": "penguin", "219": "piano", "220": "pickup truck", "221": "picture frame", "222": "pig", "223": "pillow", "224": "pineapple", "225": "pizza", "226": "pliers", "227": "police car", "228": "pond", "229": "pool", "230": "popsicle", "231": "postcard", "232": "potato", "233": "power outlet", "234": "purse", "235": "rabbit", "236": "raccoon", "237": "radio", "238": "rain", "239": "rainbow", "240": "rake", "241": "remote control", "242": "rhinoceros", "243": "rifle", "244": "river", "245": "roller coaster", "246": "rollerskates", "247": "sailboat", "248": "sandwich", "249": "saw", "250": "saxophone", "251": "school bus", "252": "scissors", "253": "scorpion", "254": "screwdriver", "255": "sea turtle", "256": "see saw", "257": "shark", "258": "sheep", "259": "shoe", "260": "shorts", "261": "shovel", "262": "sink", "263": "skateboard", "264": "skull", "265": "skyscraper", "266": "sleeping bag", "267": "smiley face", "268": "snail", "269": "snake", "270": "snorkel", "271": "snowflake", "272": "snowman", "273": "soccer ball", "274": "sock", "275": "speedboat", "276": "spider", "277": "spoon", "278": "spreadsheet", "279": "square", "280": "squiggle", "281": "squirrel", "282": "stairs", "283": "star", "284": "steak", "285": "stereo", "286": "stethoscope", "287": "stitches", "288": "stop sign", "289": "stove", "290": "strawberry", "291": "streetlight", "292": "string bean", "293": "submarine", "294": "suitcase", "295": "sun", "296": "swan", "297": "sweater", "298": "swing set", "299": "sword", "300": "syringe", "301": "t-shirt", "302": "table", "303": "teapot", "304": "teddy-bear", "305": "telephone", "306": "television", "307": "tennis racquet", "308": "tent", "309": "The Eiffel Tower", "310": "The Great Wall of China", "311": "The Mona Lisa", "312": "tiger", "313": "toaster", "314": "toe", "315": "toilet", "316": "tooth", "317": "toothbrush", "318": "toothpaste", "319": "tornado", "320": "tractor", "321": "traffic light", "322": "train", "323": "tree", "324": "triangle", "325": "trombone", "326": "truck", "327": "trumpet", "328": "umbrella", "329": "underwear", "330": "van", "331": "vase", "332": "violin", "333": "washing machine", "334": "watermelon", "335": "waterslide", "336": "whale", "337": "wheel", "338": "windmill", "339": "wine bottle", "340": "wine glass", "341": "wristwatch", "342": "yoga", "343": "zebra", "344": "zigzag"}}}}, {"name": "drawing", "dtype": {"array2_d": {"shape": [3], "dtype": "int16"}}}], "splits": [{"name": "train", "num_bytes": 23725242280, "num_examples": 43988874}, {"name": "validation", "num_bytes": 466485546, "num_examples": 862500}, {"name": "test", "num_bytes": 466191706, "num_examples": 862500}], "download_size": 6928245966, "dataset_size": 24657919532}]}
2024-01-18T11:19:15+00:00
5df91be3ec941e3ce0e9e214d0be2d208bcb6b05
## Auto Miles per Gallon (MPG) Dataset Following description was taken from [UCI machine learning repository](https://archive.ics.uci.edu/ml/datasets/auto+mpg). Source: This dataset was taken from the StatLib library which is maintained at Carnegie Mellon University. The dataset was used in the 1983 American Statistical Association Exposition. ## Data Set Information: This dataset is a slightly modified version of the dataset provided in the StatLib library. In line with the use by Ross Quinlan (1993) in predicting the attribute "mpg", 8 of the original instances were removed because they had unknown values for the "mpg" attribute. The original dataset is available in the file "auto-mpg.data-original". "The data concerns city-cycle fuel consumption in miles per gallon, to be predicted in terms of 3 multivalued discrete and 5 continuous attributes." (Quinlan, 1993) ## Attribute Information: - mpg: continuous - cylinders: multi-valued discrete - displacement: continuous - horsepower: continuous - weight: continuous - acceleration: continuous - model year: multi-valued discrete - origin: multi-valued discrete - car name: string (unique for each instance)
scikit-learn/auto-mpg
[ "task_categories:tabular-classification", "task_categories:tabular-regression", "language:en", "license:apache-2.0", "scikit-learn", "region:us" ]
2022-06-09T09:05:01+00:00
{"language": ["en"], "license": "apache-2.0", "task_categories": ["tabular-classification", "tabular-regression"], "pretty_name": "auto-mpg", "tags": ["scikit-learn"]}
2023-12-05T12:45:05+00:00
a0c9094f4ff5841ef6daef20cf62fa9c3e6c3927
anton-l/images
[ "license:apache-2.0", "region:us" ]
2022-06-09T09:11:50+00:00
{"license": "apache-2.0"}
2022-09-05T15:08:48+00:00
00ccd79fc0c83d159cf835b236132f5903ecf6c7
chradden/face_age
[ "region:us" ]
2022-06-09T09:31:42+00:00
{}
2022-06-16T14:21:17+00:00
7d0c06fa172853f1eb41358c1c9ec081c478d24a
# AutoTrain Dataset for project: qa-team-car-review-project ## Dataset Descritpion This dataset has been automatically processed by AutoTrain for project qa-team-car-review-project. ### Languages The BCP-47 code for the dataset's language is en. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```json [ { "text": " ", "target": 1 }, { "text": " Mazda truck costs less than the sister look-a-like Ford; Mazda is a \"A\" series of the Ford Ranger, [...]", "target": 2 } ] ``` ### Dataset Fields The dataset has the following fields (also called "features"): ```json { "text": "Value(dtype='string', id=None)", "target": "ClassLabel(num_classes=3, names=['great', 'ok', 'poor'], id=None)" } ``` ### Dataset Splits This dataset is split into a train and validation split. The split sizes are as follow: | Split name | Num samples | | ------------ | ------------------- | | train | 19731 | | valid | 4935 |
florentgbelidji/autotrain-data-qa-team-car-review-project
[ "task_categories:text-classification", "language:en", "region:us" ]
2022-06-09T09:47:22+00:00
{"language": ["en"], "task_categories": ["text-classification"]}
2022-10-25T09:29:30+00:00
660956f28b0c98cf634d693dfb25156fceeef638
## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) <!-- - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) --> ## Dataset Description - **Homepage:** [SIL AI](https://ai.sil.org/) - **Point of Contact:** [SIL AI email](mailto:[email protected]) - **Source Data:** [Bloom Library](https://bloomlibrary.org/) ![logo for Bloom Library](https://bloom-vist.s3.amazonaws.com/bloom_logo.png) ![sil-ai logo](https://s3.amazonaws.com/moonup/production/uploads/1661440873726-6108057a823007eaf0c7bd10.png) ## Dataset Summary **Bloom** is free, open-source software and an associated website [Bloom Library](https://bloomlibrary.org/), app, and services developed by [SIL International](https://www.sil.org/). Bloom’s primary goal is to equip non-dominant language communities and their members to create the literature they want for their community and children. Bloom also serves organizations that help such communities develop literature and education or other aspects of community development. This version of the Bloom Library data is developed specifically for the automatic speech recognition and speech-to-text tasks. It includes data from 56 languages across 18 language families. There is a mean of 458 and median of 138 audio records per language. **Note**: If you speak one of these languages and can help provide feedback or corrections, please let us know! **Note**: Although data from [bloom-lm](https://huggingface.co/datasets/sil-ai/bloom-lm) was used in the training of the [BLOOM model](https://huggingface.co/bigscience/bloom), the dataset only represents a small portion of the data used to train that model. Data from "Bloom Library" was combined with a large number of other datasets to train that model. "Bloom Library" is a project that existed prior to the BLOOM model, and is something separate. All that to say... We were using the "Bloom" name before it was cool. 😉 ## Languages Of the 500+ languages listed at BloomLibrary.org, there are 56 languages available in this dataset. Here are the corresponding ISO 639-3 codes: ajz, bam, bis, bjn, boz, bze, bzi, cak, ceb, chd, chp, clo, csw, eng, fli, fra, guj, hbb, hin, ind, jmx, jra, kan, kbq, kek, kjb, kmu, kqr, kwu, loh, mai, mal, mam, mar, mle, mya, myk, nas, nsk, nsn, oji, omw, por, quc, sdk, snk, spa, stk, taj, tam, tbj, tdc, tgl, tpi, tuz, tzj ## Dataset Statistics Some of the languages included in the dataset include few audio cuts. These are not split between training, validation, and test. For those with higher numbers of available stories we include the following numbers of stories in each split: | ISO 639-3 | Name | Train Cuts | Validation Cuts | Test Cuts | |:------------|:------------------------------|----------------:|---------------------:|---------------:| | ajz | Amri Karbi | 135 | 34 | 50 | | bam | Bamanankan | 203 | 50 | 50 | | bis | Bislama | 0 | 0 | 46 | | bjn | Banjar | 80 | 20 | 50 | | boz | Bozo, Tieyaxo | 427 | 50 | 52 | | bze | Bozo, Jenaama | 101 | 26 | 50 | | bzi | Bisu | 1363 | 50 | 157 | | cak | Kaqchikel | 989 | 50 | 115 | | ceb | Cebuano | 553 | 50 | 67 | | chd | Chontal, Highland Oaxaca | 205 | 50 | 50 | | chp | Dene | 0 | 0 | 14 | | clo | Chontal, Lowland Oaxaca | 120 | 30 | 50 | | csw | Cree, Swampy | 0 | 0 | 45 | | eng | English | 4143 | 48 | 455 | | fli | Fali Muchella | 59 | 15 | 50 | | fra | French | 261 | 49 | 50 | | guj | Gujarati | 27 | 0 | 48 | | hbb | Nya Huba | 558 | 50 | 67 | | hin | Hindi | 62 | 15 | 49 | | ind | Indonesian | 0 | 0 | 14 | | jmx | Mixtec, Western Juxtlahuaca | 39 | 0 | 50 | | jra | Jarai | 203 | 50 | 50 | | kan | Kannada | 281 | 43 | 50 | | kbq | Kamano | 0 | 0 | 27 | | kek | Q’eqchi’ | 1676 | 49 | 190 | | kjb | Q’anjob’al | 770 | 50 | 91 | | kmu | Kanite | 0 | 0 | 28 | | kqr | Kimaragang | 0 | 0 | 18 | | kwu | Kwakum | 58 | 15 | 50 | | loh | Narim | 0 | 0 | 15 | | mai | Maithili | 0 | 0 | 11 | | mal | Malayalam | 125 | 31 | 44 | | mam | Mam | 1313 | 50 | 151 | | mar | Marathi | 25 | 0 | 49 | | mle | Manambu | 0 | 0 | 8 | | mya | Burmese | 321 | 50 | 50 | | myk | Sénoufo, Mamara | 669 | 50 | 80 | | nas | Naasioi | 13 | 0 | 50 | | nsk | Naskapi | 0 | 0 | 15 | | nsn | Nehan | 0 | 0 | 31 | | oji | Ojibwa | 0 | 0 | 25 | | omw | Tairora, South | 0 | 0 | 34 | | por | Portuguese | 0 | 0 | 34 | | quc | K’iche’ | 1460 | 50 | 167 | | sdk | Sos Kundi | 312 | 50 | 50 | | snk | Soninke | 546 | 50 | 66 | | spa | Spanish | 1816 | 50 | 207 | | stk | Aramba | 180 | 45 | 50 | | taj | Tamang, Eastern | 0 | 0 | 24 | | tam | Tamil | 159 | 39 | 46 | | tbj | Tiang | 0 | 0 | 24 | | tdc | Ẽpẽra Pedea | 0 | 0 | 19 | | tgl | Tagalog | 352 | 48 | 50 | | tpi | Tok Pisin | 1061 | 50 | 123 | | tuz | Turka | 48 | 13 | 50 | | tzj | Tz’utujil | 0 | 0 | 41 | ## Dataset Structure ### Data Instances The examples look like this for Hindi: ``` from datasets import load_dataset # Specify the language code. dataset = load_dataset('sil-ai/bloom-speech', 'hin', use_auth_token=True) #note you must login to HuggingFace via the huggingface hub or huggingface cli # A data point consists of transcribed audio in the specified language code. # To see a transcription: print(dataset['train']['text'][0]) ``` This would produce an output: ``` चित्र: बो और शैम्पू की बोतल ``` Whereas if you wish to gather all the text for a language you may use this: ``` dataset['train']['text'] ``` ### Data Fields The metadata fields are below. In terms of licenses, all stories included in the current release are released under a Creative Commons license (even if the individual story metadata fields are missing). - **file**: the local path to the audio file - **audio**: a dictionary with a path, array, and sampling_rate as is standard for Hugging Face audio - **text**: the transcribed text - **book**: title of the book, e.g. "बो मेस्सी और शैम्पू". - **instance**: unique ID for each book/translation assigned by Bloom Library. For example the Hindi version of 'बो मेस्सी और शैम्पू' is 'eba60f56-eade-4d78-a66f-f52870f6bfdd' - **license**: specific license used, e.g. "cc-by-sa" for "Creative Commons, by attribution, share-alike". - **credits**: attribution of contributors as described in the book metadata, including authors, editors, etc. if available - **original_lang_tag**: the language tag originally assigned in Bloom Library. This may include information on script type, etc. ### Data Splits All languages include a train, validation, and test split. However, for language having a small number of stories, certain of these splits maybe empty. In such cases, we recommend using any data for testing only or for zero-shot experiments. ## Changelog - **26 September 2022** Page initiated
sil-ai/bloom-speech
[ "task_categories:automatic-speech-recognition", "task_categories:text-to-speech", "annotations_creators:expert-generated", "language_creators:expert-generated", "multilinguality:multilingual", "size_categories:10K<n<100K", "source_datasets:original", "language:ajz", "language:bam", "language:bi", "language:bis", "language:bjn", "language:bm", "language:boz", "language:bze", "language:bzi", "language:cak", "language:ceb", "language:chd", "language:chp", "language:clo", "language:csw", "language:en", "language:eng", "language:es", "language:fli", "language:fr", "language:fra", "language:gu", "language:guj", "language:hbb", "language:hi", "language:hin", "language:id", "language:ind", "language:jmx", "language:jra", "language:kan", "language:kbq", "language:kek", "language:kjb", "language:kmu", "language:kn", "language:kqr", "language:kwu", "language:loh", "language:mai", "language:mal", "language:mam", "language:mar", "language:ml", "language:mle", "language:mr", "language:my", "language:mya", "language:myk", "language:nas", "language:nsk", "language:nsn", "language:oj", "language:oji", "language:omw", "language:por", "language:pt", "language:quc", "language:sdk", "language:snk", "language:spa", "language:stk", "language:ta", "language:taj", "language:tam", "language:tbj", "language:tdc", "language:tgl", "language:tl", "language:tpi", "language:tuz", "language:tzj", "license:cc-by-nc-4.0", "license:cc-by-sa-4.0", "license:cc-by-nc-nd-4.0", "license:cc-by-nc-sa-4.0", "region:us" ]
2022-06-09T11:08:44+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["expert-generated"], "language": ["ajz", "bam", "bi", "bis", "bjn", "bm", "boz", "bze", "bzi", "cak", "ceb", "chd", "chp", "clo", "csw", "en", "eng", "es", "fli", "fr", "fra", "gu", "guj", "hbb", "hi", "hin", "id", "ind", "jmx", "jra", "kan", "kbq", "kek", "kjb", "kmu", "kn", "kqr", "kwu", "loh", "mai", "mal", "mam", "mar", "ml", "mle", "mr", "my", "mya", "myk", "nas", "nsk", "nsn", "oj", "oji", "omw", "por", "pt", "quc", "sdk", "snk", "spa", "stk", "ta", "taj", "tam", "tbj", "tdc", "tgl", "tl", "tpi", "tuz", "tzj"], "license": ["cc-by-nc-4.0", "cc-by-sa-4.0", "cc-by-nc-nd-4.0", "cc-by-nc-sa-4.0"], "multilinguality": ["multilingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["automatic-speech-recognition", "text-to-speech"], "pretty_name": "BloomSpeech", "extra_gated_prompt": "One more step before getting this dataset. This dataset is open access and available only for non-commercial use (except for portions of the dataset labeled with a `cc-by-sa` license). A \"license\" field paired with each of the dataset entries/samples specifies the Creative Commons license for that entry/sample.\n\nThese [Creative Commons licenses](https://creativecommons.org/about/cclicenses/) specify that: \n1. You cannot use the dataset for or directed toward commercial advantage or monetary compensation (except for those portions of the dataset labeled specifically with a `cc-by-sa` license. If you would like to ask about commercial uses of this dataset, please [email us](mailto:[email protected]).\n2. Any public, non-commercial use of the data must give appropriate credit, provide a link to the license, and indicate if changes were made. You may do so in any reasonable manner, but not in any way that suggests the licensor endorses you or your use. \n3. For those portions of the dataset marked with an ND license, you cannot remix, transform, or build upon the material, and you may not distribute modified material. \n\nIn addition to the above implied by Creative Commons and when clicking \"Access Repository\" below, you agree: \n\n1. Not to use the dataset for any use intended to or which has the effect of harming or enabling discrimination against individuals or groups based on legally protected characteristics or categories, including but not limited to discrimination against Indigenous People as outlined in Articles 2; 13-16; and 31 of the United Nations Declaration on the Rights of Indigenous People, 13 September 2007 and as subsequently amended and revised.\n2. That your *contact information* (email address and username) can be shared with the model authors as well.\n ", "extra_gated_fields": {"I have read the License and agree with its terms": "checkbox"}}
2023-02-15T13:28:59+00:00
3a2f92dc83d67d89f1eb1885d1c75961b32722ec
# Title 1 hahahoho
otheng03/test1
[ "license:apache-2.0", "region:us" ]
2022-06-09T11:16:55+00:00
{"license": "apache-2.0"}
2022-06-09T11:20:57+00:00
9699ef019676b4ae1504e9c156bdb4cfda059bb5
# AutoTrain Dataset for project: car-review-project ## Dataset Descritpion This dataset has been automatically processed by AutoTrain for project car-review-project. It contains consumer car ratings and reviews from [Edmunds website](https://www.kaggle.com/datasets/ankkur13/edmundsconsumer-car-ratings-and-reviews) ### Languages The BCP-47 code for the dataset's language is en. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```json [ { "text": " ", "target": 1 }, { "text": " Mazda truck costs less than the sister look-a-like Ford; Mazda is a \"A\" series of the Ford Ranger, [...]", "target": 2 } ] ``` ### Dataset Fields The dataset has the following fields (also called "features"): ```json { "text": "Value(dtype='string', id=None)", "target": "ClassLabel(num_classes=3, names=['great', 'ok', 'poor'], id=None)" } ``` ### Dataset Splits This dataset is split into a train and validation split. The split sizes are as follow: | Split name | Num samples | | ------------ | ------------------- | | train | 19731 | | valid | 4935 |
qualitydatalab/autotrain-data-car-review-project
[ "task_categories:text-classification", "language:en", "region:us" ]
2022-06-09T11:27:44+00:00
{"language": ["en"], "task_categories": ["text-classification"]}
2022-10-25T09:29:37+00:00
13eadc735ff81c0e0537276f729f2f391e594bb8
# Dataset Card for Gigaspeech ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) - [Terms of Access](#terms-of-access) ## Dataset Description - **Homepage:** https://github.com/SpeechColab/GigaSpeech - **Repository:** https://github.com/SpeechColab/GigaSpeech - **Paper:** https://arxiv.org/abs/2106.06909 - **Leaderboard:** https://github.com/SpeechColab/GigaSpeech#leaderboard - **Point of Contact:** [[email protected]](mailto:[email protected]) ## Dataset Description GigaSpeech is an evolving, multi-domain English speech recognition corpus with 10,000 hours of high quality labeled audio suitable for supervised training. The transcribed audio data is collected from audiobooks, podcasts and YouTube, covering both read and spontaneous speaking styles, and a variety of topics, such as arts, science, sports, etc. ### Example Usage The training split has several configurations of various size: XS, S, M, L, XL. See the Section on "Data Splits" for more information. To download the XS configuration: ```python from datasets import load_dataset gs = load_dataset("speechcolab/gigaspeech", "xs", use_auth_token=True) # see structure print(gs) # load audio sample on the fly audio_input = gs["train"][0]["audio"] # first decoded audio sample transcription = gs["train"][0]["text"] # first transcription ``` It is possible to download only the development or test data: ```python gs_dev = load_dataset("speechcolab/gigaspeech", "dev", use_auth_token=True) gs_test = load_dataset("speechcolab/gigaspeech", "test", use_auth_token=True) ``` ### Supported Tasks and Leaderboards - `automatic-speech-recognition`: The dataset can be used to train a model for Automatic Speech Recognition (ASR). The model is presented with an audio file and asked to transcribe the audio file to written text. The most common evaluation metric is the word error rate (WER). The task has an active leaderboard which can be found at https://github.com/SpeechColab/GigaSpeech#leaderboard and ranks models based on their WER. - `text-to-speech`, `text-to-audio`: The dataset can also be used to train a model for Text-To-Speech (TTS). ### Languages Gigaspeech contains audio and transcription data in English. ## Dataset Structure ### Data Instances ```python { 'segment_id': 'YOU0000000315_S0000660', 'speaker': 'N/A', 'text': "AS THEY'RE LEAVING <COMMA> CAN KASH PULL ZAHRA ASIDE REALLY QUICKLY <QUESTIONMARK>", 'audio': { # in streaming mode 'path' will be 'xs_chunks_0000/YOU0000000315_S0000660.wav' 'path': '/home/user/.cache/huggingface/datasets/downloads/extracted/9d48cf31/xs_chunks_0000/YOU0000000315_S0000660.wav', 'array': array([0.0005188 , 0.00085449, 0.00012207, ..., 0.00125122, 0.00076294, 0.00036621], dtype=float32), 'sampling_rate': 16000 }, 'begin_time': 2941.889892578125, 'end_time': 2945.070068359375, 'audio_id': 'YOU0000000315', 'title': 'Return to Vasselheim | Critical Role: VOX MACHINA | Episode 43', 'url': 'https://www.youtube.com/watch?v=zr2n1fLVasU', 'source': 2, 'category': 24, 'original_full_path': 'audio/youtube/P0004/YOU0000000315.opus' } ``` ### Data Fields * segment_id (string) - string id of the segment. * speaker (string) - string id of the speaker (can be "N/A"). * text (string) - transcription of the segment. * begin_time (float) - start time of the segment in an original full audio. * end_time (float32) - end time of the segment in an original full audio. * audio (Audio feature) - a dictionary containing the path to the audio, the decoded audio array, and the sampling rate. In non-streaming mode (default), the path point to the locally extracted audio. In streaming mode, the path is the relative path of an audio. segment inside its archive (as files are not downloaded and extracted locally). * audio_id (string) - string idea of the original full audio. * title (string) - title of the original full audio. * url (string) - url of the original full audio. * source (ClassLabel) - id of the audio source. Sources are audiobook (0), podcast (1), and YouYube (2). * category (ClassLabel) - id of the audio category, categories are listed below. * original_full_path (string) - the relative path to the original full audio sample in the original data directory. Categories are assigned from the following labels: "People and Blogs", "Business", "Nonprofits and Activism", "Crime", "History", "Pets and Animals", "News and Politics", "Travel and Events", "Kids and Family", "Leisure", "N/A", "Comedy", "News and Politics", "Sports", "Arts", "Science and Technology", "Autos and Vehicles", "Science and Technology", "People and Blogs", "Music", "Society and Culture", "Education", "Howto and Style", "Film and Animation", "Gaming", "Entertainment", "Travel and Events", "Health and Fitness", "audiobook". ### Data Splits The dataset has three splits: train, evaluation (dev) and test. The train split has five configurations of various sizes: XS, S, M, L, XL. Larger subsets are supersets of smaller subsets, e.g., the L subset contains all the data from the M subset. #### Transcribed Training Subsets Size | Subset | Hours | Remarks | |:---------------:|:-------------:|:-------------| | XS | 10 | System building and debugging | | S | 250 | Quick research experiments | | M | 1,000 | Large-scale research experiments | | L | 2,500 | Medium-scale industrial experiments | | XL | 10,000 | Large-scale industrial experiments | #### Transcribed Evaluation Subsets | Subset | Hours | Remarks | |:------:|:-----:|:--------| | Dev | 12 | Randomly selected from the crawled Podcast and YouTube Data | | Test | 40 | Part of the subset was randomly selected from the crawled Podcast and YouTube data; part of it was manually collected through other channels to have better coverage. | ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data | Audio Source | Transcribed Hours | Acoustic Condition | |:-------------|:----------------------:|:-------------------| | Audiobook | 2,655 | <li>Reading</li><li>Various ages and accents</li> | | Podcast | 3,498 | <li>Clean or background music</li><li>Indoor</li><li>Near-field</li><li>Spontaneous</li><li>Various ages and accents</li>| | YouTube | 3,845 | <li>Clean and noisy</li><li>Indoor and outdoor</li><li>Near- and far-field</li><li>Reading and spontaneous</li><li>Various ages and accents</li> | | ***Total*** | ***10,000*** || #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? Development and test subsets are annotated by professional human annotators. ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information SpeechColab does not own the copyright of the audio files. For researchers and educators who wish to use the audio files for non-commercial research and/or educational purposes, we can provide access through our site under certain conditions and terms. In general, when training a machine learning model on a given dataset, the license of the model is **independent** to that of the dataset. That is to say, speech recognition models trained on the GigaSpeech dataset may be eligible for commercial license, provided they abide to the 'Fair Use' terms of the underlying data and do not violate any explicit copyright restrictions. This is likely to be true in most use-cases. However, it is your responsiblity to verify the appropriate model license for your specific use-case by confirming that the dataset usage abides by the Fair Use terms. SpeechColab is not responsible for the license of any machine learning model trained on the GigaSpeech dataset. ### Citation Information Please cite this paper if you find this work useful: ```bibtext @inproceedings{GigaSpeech2021, title={GigaSpeech: An Evolving, Multi-domain ASR Corpus with 10,000 Hours of Transcribed Audio}, booktitle={Proc. Interspeech 2021}, year=2021, author={Guoguo Chen, Shuzhou Chai, Guanbo Wang, Jiayu Du, Wei-Qiang Zhang, Chao Weng, Dan Su, Daniel Povey, Jan Trmal, Junbo Zhang, Mingjie Jin, Sanjeev Khudanpur, Shinji Watanabe, Shuaijiang Zhao, Wei Zou, Xiangang Li, Xuchen Yao, Yongqing Wang, Yujun Wang, Zhao You, Zhiyong Yan} } ``` ### Contributions Thanks to [@polinaeterna](https://github.com/polinaeterna) and [@sanchit-gandhi](https://github.com/sanchit-gandhi) for adding this dataset. ## Terms of Access The "Researcher" has requested permission to use the GigaSpeech database (the "Database") at Tsinghua University. In exchange for such permission, Researcher hereby agrees to the following terms and conditions: 1. Researcher shall use the Database only for non-commercial research and educational purposes. 2. The SpeechColab team and Tsinghua University make no representations or warranties regarding the Database, including but not limited to warranties of non-infringement or fitness for a particular purpose. 3. Researcher accepts full responsibility for his or her use of the Database and shall defend and indemnify the SpeechColab team and Tsinghua University, including their employees, Trustees, officers and agents, against any and all claims arising from Researcher's use of the Database, including but not limited to Researcher's use of any copies of copyrighted audio files that he or she may create from the Database. 4. Researcher may provide research associates and colleagues with access to the Database provided that they first agree to be bound by these terms and conditions. 5. The SpeechColab team and Tsinghua University reserve the right to terminate Researcher's access to the Database at any time. 6. If Researcher is employed by a for-profit, commercial entity, Researcher's employer shall also be bound by these terms and conditions, and Researcher hereby represents that he or she is fully authorized to enter into this agreement on behalf of such employer.
speechcolab/gigaspeech
[ "task_categories:automatic-speech-recognition", "task_categories:text-to-speech", "task_categories:text-to-audio", "multilinguality:monolingual", "language:en", "license:apache-2.0", "arxiv:2106.06909", "region:us" ]
2022-06-09T13:51:58+00:00
{"annotations_creators": [], "language_creators": [], "language": ["en"], "license": ["apache-2.0"], "multilinguality": ["monolingual"], "source_datasets": [], "task_categories": ["automatic-speech-recognition", "text-to-speech", "text-to-audio"], "pretty_name": "Gigaspeech", "extra_gated_prompt": "SpeechColab does not own the copyright of the audio files. For researchers and educators who wish to use the audio files for non-commercial research and/or educational purposes, we can provide access through the Hub under certain conditions and terms. \nTerms of Access:\nThe \"Researcher\" has requested permission to use the GigaSpeech database (the \"Database\") at Tsinghua University. In exchange for such permission, Researcher hereby agrees to the following terms and conditions:\n1. Researcher shall use the Database only for non-commercial research and educational purposes.\n2. The SpeechColab team and Tsinghua University make no representations or warranties regarding the Database, including but not limited to warranties of non-infringement or fitness for a particular purpose.\n3. Researcher accepts full responsibility for his or her use of the Database and shall defend and indemnify the SpeechColab team and Tsinghua University, including their employees, Trustees, officers and agents, against any and all claims arising from Researcher's use of the Database, including but not limited to Researcher's use of any copies of copyrighted audio files that he or she may create from the Database.\n4. Researcher may provide research associates and colleagues with access to the Database provided that they first agree to be bound by these terms and conditions.\n5. The SpeechColab team and Tsinghua University reserve the right to terminate Researcher's access to the Database at any time.\n6. If Researcher is employed by a for-profit, commercial entity, Researcher's employer shall also be bound by these terms and conditions, and Researcher hereby represents that he or she is fully authorized to enter into this agreement on behalf of such employer.\n\n!!! Please also fill out the Google Form https://forms.gle/UuGQAPyscGRrUMLq6 to request access to the Gigaspeech dataset.", "extra_gated_fields": {"Name": "text", "Email": "text", "Organization": "text", "Address": "text", "I hereby confirm that I have requested access via the Google Form provided above": "checkbox", "I accept the terms of access": "checkbox"}}
2023-11-23T14:08:34+00:00
468e1cc664d11602655e3180e8648a9d5703a761
# Dataset Card for answersumm ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** https://github.com/Alex-Fabbri/AnswerSumm - **Paper:** [AnswerSumm: A Manually-Curated Dataset and Pipeline for Answer Summarization](https://arxiv.org/abs/2111.06474) - **Point of Contact:** [Alex Fabbri](mailto:[email protected]) ### Dataset Summary The AnswerSumm dataset is an English-language dataset of questions and answers collected from a [StackExchange data dump](https://archive.org/details/stackexchange). The dataset was created to support the task of query-focused answer summarization with an emphasis on multi-perspective answers. The dataset consists of over 4200 such question-answer threads annotated by professional linguists and includes over 8700 summaries. We decompose the task into several annotation stages, including sentence selection, sentence clustering, cluster summarization, and overall summarization. For each thread, the annotator writes two summaries, one in which the annotator is asked to mark sentences that are included in the final summary and instructed to more closely use the words in these sentences rather than abstract. We have multiple annotators for a subset of the examples in the test set. ### Languages The text in the dataset is in English. ## Dataset Structure ### Data Instances A data point comprises a question with a `title` field containing the overview of the question and a `question` that elaborates on the title. The answers are sentence tokenized and contain relevance labels, labels for inclusion in the final summary, and cluster labels. We include cluster summaries, overall summaries, and additional metadata. An example from the AnswerSumm test set looks as follows: ```json { "example_id": 9_24, "annotator_id": [1], "question": { "author": "gaming.stackexchange.com/users/11/Jeffrey", "forum": "gaming.stackexchange.com", "link": "gaming.stackexchange.com/questions/1", "question": "Now that the Engineer update has come, there will be lots of Engineers building up everywhere. How should this best be handled?", "question_tags": "\<team-fortress-2\>", "title": "What is a good strategy to deal with lots of engineers turtling on the other team?" }, "answers": [ { "answer_details": { "author": "gaming.stackexchange.com/users/44/Corv1nus", "score": 49 } "sents": [ "text": "Lots of medics with lots of ubers on high-damage-dealing classes." "label": [0], "label_summ": [0], "cluster_id": [[-1]], ] ... }, ... ] "summaries": [ [ "Demomen usually work best against a sentry farm. Heavies or pyros can also be effective. Medics should be in the frontline to absorb the shock. Build a teleporter to help your team through.", "Demomen are best against a sentry farm. Heavies or pyros can also be effective. The medic should lead the uber combo. ..." ] ] "cluster_summaries":[ "Demomen are best against a sentry farm.", "Heavies or pyros can also be effective.", ... ] } ``` ### Data Fields - question: contains metadata about the question and forum - question: the body of the question post - title: the title of the question post - question_tags: user-provided question tags - link: link to the original question - author: link to the author's user page (as requested by StackExchange's attribution policy) - answers: list of sentence-tokenized answers - answer_details: dictionary consisting of link to answer author's user page (author) and community-assigned score (score) - sents: sentences that compose the answer - text: the sentence text - label: a list (to generalize to multi-annotator scenarios) of whether the sentence is labeled as relevant or not for answering the question. - label_summ: a list of whether the sentence was used to write the first annotator-created summary (that is the first summary in `summaries`) - cluster_id: a list of lists (potentially multiple annotators and a sentence can be in potentially multiple clusters) of the clusters a sentence belongs to. -1 implies no cluster. This label can be used to aggregate sentences into clusters across answers. - summaries: list of list of summaries. Each annotator wrote two summaries. The first in the list is the summary in which the instructor was told to mark sentences relevant for inclusion in the summary and then closely use the words of these sentences, while for the second summary the annotator was asked to paraphrase and condense the cluster summaries but was not asked to reduce abstraction. - annotator_id: a list of the ids of the annotator(s) who completed all tasks related to that thread. - mismatch_info: a dict of any issues in processing the excel files on which annotations were completed. - rel_sent_not_in_cluster: list of booleans indicating whether there are sentences that are labeled as relevant but were not included in a cluster. - cluster_sents_not_matched: list of sentences that were found in a cluster but which our processing script didn't automatically match to sentences in the source answers. If cluster summarization is of interest to the user you may want to process these examples separately using clusters_orig. ### Data Splits The data is split into training, validation, and test sets using stratified sampling on the source forums. There are 2783, 500, and 1000 train/validation/test threads, respectively. ## Dataset Creation ### Curation Rationale AnswerSumm was built to provide a testbed for query-focused summarization of multi-perspective answers. The data collection was designed to tackle multiple subtasks including sentence selection, clustering, cluster summarization, and overall summarization. ### Source Data #### Initial Data Collection and Normalization The data was obtained by filtering examples based on a whitelist of forums from StackExchange which we believed would be able to be summarized by a lay person. We describe. We asked annotators to remove examples which required technical knowledge or additional context beyond what was present in the answers. #### Who are the source language producers? The language producers are the users of the StackExchange forums sampled. ### Annotations #### Annotation process Please see our [paper](https://arxiv.org/pdf/2111.06474.pdf) for additional annotation details. We began with a pre-pilot of 50 examples, followed by a pilot of 500 and a final annotation of 5000 examples. This release contains the results of the final data collection. We will release the instructions used in data collection. #### Who are the annotators? The annotators are professional linguists who were obtained through an internal contractor. ### Personal and Sensitive Information We did not anonymize the data. We followed the specifications from StackExchange [here](https://archive.org/details/stackexchange) to include author information. ## Considerations for Using the Data ### Social Impact of Dataset The purpose of this dataset is to help develop systems that automatically summarize multi-perspective answers. A system that succeeds at this task would be able to summarize many perspectives present in an answer and not limit itself to a single perspective. ### Discussion of Biases While StackExchange allows for the exchange of information and ideas, hate and harassment may exist on this site. While our annotators did not flag examples in this process, we encourage users of the dataset to reach out with concerns. We also note that this dataset is limited in its monolingual coverage. ## Additional Information ### Dataset Curators The dataset was collected by Alex Fabbri, Xiaojian Wu, Srini Iyer, Haoran Li, and Mona Diab during work done at Facebook. ### Licensing Information The data is released under cc-by-sa 4.0 following the original StackExchange [release](https://archive.org/details/stackexchange). ### Citation Information ```bibtex @misc{fabbri-etal-2022-answersumm, title={AnswerSumm: A Manually-Curated Dataset and Pipeline for Answer Summarization}, author={Alexander R. Fabbri and Xiaojian Wu and Srini Iyer and Haoran Li and Mona Diab }, year={2022}, eprint={2111.06474}, archivePrefix={arXiv}, primaryClass={cs.CL}, url={https://arxiv.org/abs/2111.06474} } ```
alexfabbri/answersumm
[ "task_categories:summarization", "annotations_creators:found", "language_creators:found", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "license:cc-by-sa-4.0", "query-based-summarization", "arxiv:2111.06474", "region:us" ]
2022-06-09T13:58:23+00:00
{"annotations_creators": ["found"], "language_creators": ["found"], "language": ["en"], "license": ["cc-by-sa-4.0"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["summarization"], "task_ids": [], "tags": ["query-based-summarization"]}
2022-12-14T20:18:28+00:00
4afddd9cc59089a6a59650cd847e1650be1e5399
MrClean/Dalleproject
[ "license:apache-2.0", "region:us" ]
2022-06-09T17:29:33+00:00
{"license": "apache-2.0", "title": "DALL\u00b7E mini", "emoji": "\ud83e\udd51", "colorFrom": "yellow", "colorTo": "green", "sdk": "static", "pinned": true}
2022-06-09T17:33:12+00:00
07eeed48418a6392700eda3bba5d3eb077036864
https://github.com/eladrich/pixel2style2pixel.git
Impostor/Pixel
[ "license:cc-by-4.0", "region:us" ]
2022-06-09T20:15:07+00:00
{"license": "cc-by-4.0"}
2022-06-09T20:15:33+00:00
1968c2e5f786501e647c46386dac435e5babd32d
# MuP - Multi Perspective Scientific Document Summarization Generating summaries of scientific documents is known to be a challenging task. Majority of existing work in summarization assumes only one single best gold summary for each given document. Having only one gold summary negatively impacts our ability to evaluate the quality of summarization systems as writing summaries is a subjective activity. At the same time, annotating multiple gold summaries for scientific documents can be extremely expensive as it requires domain experts to read and understand long scientific documents. This shared task will enable exploring methods for generating multi-perspective summaries. We introduce a novel summarization corpus, leveraging data from scientific peer reviews to capture diverse perspectives from the reader's point of view. For more information about the dataset please refer to: https://github.com/allenai/mup
allenai/mup-full
[ "license:odc-by", "region:us" ]
2022-06-09T23:07:46+00:00
{"license": ["odc-by"]}
2022-10-25T09:29:44+00:00
35ed298434fb9458d27546dc64ce88b1eb93a2d1
## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Other Known Limitations](#other-known-limitations) ## Dataset Description - **Point of Contact:** [Nart Tlisha](mailto:[email protected]) - **Size of the generated dataset:** 33.5 MB ### Dataset Summary The Abkhaz Russian parallel corpus dataset is a collection of 205,665 sentences/words extracted from different sources; e-books, web scrapping. ## Dataset Creation ### Source Data Here is a link to the source on [github](https://github.com/danielinux7/Multilingual-Parallel-Corpus/blob/master/references.md) ## Considerations for Using the Data ### Other Known Limitations The accuracy of the dataset is around 95% (gramatical, arthographical errors)
Nart/parallel-ab-ru
[ "task_categories:text-generation", "task_categories:translation", "language_creators:expert-generated", "multilinguality:translation", "multilinguality:multilingual", "size_categories:100K<n<1M", "source_datasets:original", "language:ab", "language:ru", "license:cc0-1.0", "region:us" ]
2022-06-10T12:08:42+00:00
{"language_creators": ["expert-generated"], "language": ["ab", "ru"], "license": ["cc0-1.0"], "multilinguality": ["translation", "multilingual"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["text-generation", "translation"], "task_ids": [], "pretty_name": "Abkhazian Russian parallel corpus"}
2023-04-08T06:52:41+00:00
4ac984b4226120f96ea25525647a997cf73f76a6
sagantime/NBAData
[ "region:us" ]
2022-06-10T12:51:56+00:00
{}
2022-06-10T12:55:59+00:00
33313bc359ab206b3dedc32c3017ba5fc2b26a78
# Glue WSC Fixed This dataset is a port of the official [`wsc.fixed` dataset](https://huggingface.co/datasets/super_glue/viewer/wsc.fixed/train) on the Hub. Also, the test split is not labeled; the label column values are always -1.
SetFit/wsc_fixed
[ "region:us" ]
2022-06-10T12:53:16+00:00
{}
2022-06-10T12:55:19+00:00
4c95d91e7feaf48078405d0ba520f3da9de3d23d
sudo-s/datasetexample
[ "region:us" ]
2022-06-10T12:57:00+00:00
{}
2022-06-10T12:57:31+00:00
8694ce7ea420cbcce8a7e4316bfebce9ee4a0665
# Glue WSC This dataset is a port of the official [`wsc` dataset](https://huggingface.co/datasets/super_glue) on the Hub. Also, the test split is not labeled; the label column values are always -1.
SetFit/wsc
[ "region:us" ]
2022-06-10T12:57:36+00:00
{}
2022-06-10T12:59:09+00:00
01e427e689e9d3a9097f85eab7a91ce937cf5f98
# Customer Reviews This dataset is a port of the official [`CR` dataset](https://github.com/hiyouga/Dual-Contrastive-Learning/tree/main/data) from [this paper](https://www.cs.uic.edu/~liub/FBS/opinion-mining-final-WSDM.pdf). There is no validation split.
SetFit/CR
[ "region:us" ]
2022-06-10T13:30:21+00:00
{}
2022-06-21T08:04:33+00:00
ddd2bbc0e2119770e28033421296e74818981e33
# Italian Tweets Test Dataset This is a dataset with 10M italian tweets. It still contains errors. Please do not use. ## How to Use ```python from datasets import load_dataset data = load_dataset("pere/italian_tweets_10M") ```
pere/italian_tweets_10M
[ "region:us" ]
2022-06-10T15:12:45+00:00
{}
2022-06-12T17:26:39+00:00
05f68a2dbe784d24da08c4f35fda61a86a21d2e6
Theivaprakasham/wildreceipt
[ "license:apache-2.0", "region:us" ]
2022-06-10T17:01:28+00:00
{"license": "apache-2.0"}
2022-06-10T20:46:37+00:00
37a6f25b900d852495080624a3432b5cc231bc9e
ggnm/ewww
[ "license:afl-3.0", "region:us" ]
2022-06-10T19:08:10+00:00
{"license": "afl-3.0"}
2022-06-10T19:08:10+00:00
933704c272f00bd518ade032a0b1cac1c80a0938
ggnm/grdgytygh
[ "license:afl-3.0", "region:us" ]
2022-06-10T19:10:23+00:00
{"license": "afl-3.0"}
2022-06-10T19:10:23+00:00
f7c998cc409db7ac860b5b84ef6a3ea53ae56954
deusprofano/images
[ "license:other", "region:us" ]
2022-06-10T22:23:23+00:00
{"license": "other"}
2022-06-10T22:23:23+00:00
cd2b1b4ea8112bdae1f80ae74d290715833d4169
ReverseThings/lol
[ "license:afl-3.0", "region:us" ]
2022-06-11T02:24:01+00:00
{"license": "afl-3.0"}
2022-06-11T02:24:01+00:00
439a75851cb739c75b7179a3d1d9a17e7224b8f1
Wendigofucker/GeneratedHorror
[ "license:other", "region:us" ]
2022-06-11T02:55:07+00:00
{"license": "other"}
2022-06-11T02:55:07+00:00
49db1aafbad19ee8a494342f74c1a640b5a70e75
# Dataset Card for [Dataset Name] ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary The Ultimate Arabic News Dataset is a collection of single-label modern Arabic texts that are used in news websites and press articles. Arabic news data was collected by web scraping techniques from many famous news sites such as Al-Arabiya, Al-Youm Al-Sabea (Youm7), the news published on the Google search engine and other various sources. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information license: cc-by-4.0 ### Citation Information ``` @book{url, author = {Al-Dulaimi, Ahmed Hashim}, year = {2022}, month = {05}, website = {Mendeley Data, V1}, title = {Ultimate Arabic News Dataset}, doi = {10.17632/jz56k5wxz7.1} } ``` ### Contributions [More Information Needed]
khalidalt/ultimate_arabic_news
[ "region:us" ]
2022-06-11T05:06:25+00:00
{}
2022-06-15T13:46:10+00:00
9a6a5918ca709882044f23470120423e0297d986
noob123/small_augemented_nlp_dataset
[ "license:other", "region:us" ]
2022-06-11T05:17:55+00:00
{"license": "other"}
2022-06-11T05:23:35+00:00
674d842241096b770b86bf5c69ac85d7a68a5fc3
# Dataset Card for "XKCD" ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Dataset Creation](#dataset-creation) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Additional Information](#additional-information) - [Licensing Information](#licensing-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://xkcd.com/](https://xkcd.com/), [https://www.explainxkcd.com](https://www.explainxkcd.com) - **Repository:** [Hugging Face repository](https://huggingface.co/datasets/olivierdehaene/xkcd/tree/main) ### Dataset Summary XKCD is an export of all XKCD comics with their transcript and explanation scrapped from [https://explainxkcd.com](https://explainxkcd.com). ## Dataset Structure ### Data Instances - `id`: `1` - `title`: `Barrel - Part 1` - `image_title`: `Barrel - Part 1` - `url`: `https://www.xkcd.com/1` - `image_url`: `https://imgs.xkcd.com/comics/barrel_cropped_(1).jpg` - `explained_url`: `https://www.explainxkcd.com/wiki/index.php/1:_Barrel_-_Part_1` - `transcript`: `[A boy sits in a barrel which is floating in an ocean.] Boy: i wonder where i'll float next? [A smaller frame with a zoom out of the boy in the barrel seen from afar. The barrel drifts into the distance. Nothing else can be seen.]` - `explanation`: `The comic shows a young boy floating in a barrel in an ocean that doesn't have a visible end. It comments on the unlikely optimism and perhaps naïveté people sometimes display. The boy is completely lost and seems hopelessly alone, without any plan or control of the situation. Yet, rather than afraid or worried, he is instead quietly curious: "I wonder where I'll float next?" Although not necessarily the situation in this comic, this is a behavior people often exhibit when there is nothing they can do about a problematic situation for a long time; they may have given up hope or developed a cavalier attitude as a coping mechanism. The title text expands on the philosophical content, with the boy representing the average human being: wandering through life with no real plan, quietly optimistic, always opportunistic and clueless as to what the future may hold. The isolation of the boy may also represent the way in which we often feel lost through life, never knowing quite where we are, believing that there is no one to whom to turn. This comic could also reflect on Randall's feelings towards creating xkcd in the first place; unsure of what direction the web comic would turn towards, but hopeful that it would eventually become the popular web comic that we know today. This is the first in a six-part series of comics whose parts were randomly published during the first several dozen strips. The series features a character that is not consistent with what would quickly become the xkcd stick figure style. The character is in a barrel. In 1110: Click and Drag there is a reference to this comic at 1 North, 48 East . After Randall released the full The Boy and his Barrel story on xkcd, it has been clear that the original Ferret story should also be included as part of the barrel series. The full series can be found here . They are listed below in the order Randall chose for the short story above: ` ### Data Fields - `id` - `title` - `url`: xkcd.com URL - `image_url` - `explained_url`: explainxkcd.com URL - `transcript`: english text transcript of the comic - `explanation`: english explanation of the comic ## Dataset Creation The dataset was scrapped from both explainxkcd.com and xkcd.com. The dataset is therefore licensed under the Creative Commons Attribution-ShareAlike 3.0 license for the `transcript` and `explanation` fields, while the image itself is licensed under the Creative Commons Attribution-NonCommercial 2.5 license. See the [Copyrights](https://www.explainxkcd.com/wiki/index.php/explain_xkcd:Copyrights) page from explainxkcd.com for more explanations. ### Update You can update the dataset by using the `scrapper.py` script. First install the dependencies: ```bash pip install aiolimiter aiohttp beautifulsoup4 pandas ``` Then run the script: ```bash python scrapper.py ``` ## Considerations for Using the Data As the data was scrapped, it is entirely possible that some fields are missing part of the original data. ## Additional Information ### Licensing Information The dataset is licensed under the Creative Commons Attribution-ShareAlike 3.0 license for the `transcript` and `explanation` fields, while the images are licensed under the Creative Commons Attribution-NonCommercial 2.5 license. ### Contributions Thanks to [@OlivierDehaene](https://github.com/OlivierDehaene) for adding this dataset.
olivierdehaene/xkcd
[ "task_categories:image-to-text", "task_categories:feature-extraction", "language_creators:other", "multilinguality:monolingual", "size_categories:1K<n<10K", "language:en", "license:cc-by-sa-3.0", "license:other", "region:us" ]
2022-06-11T19:32:01+00:00
{"annotations_creators": [], "language_creators": ["other"], "language": ["en"], "license": ["cc-by-sa-3.0", "other"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": [], "task_categories": ["image-to-text", "feature-extraction"], "task_ids": [], "pretty_name": "XKCD"}
2022-10-25T09:31:55+00:00
9961dbe05d583ef6a90e0556be01d4d883c4e259
TencentMedicalNet/MRBrains18
[ "license:mit", "region:us" ]
2022-06-12T00:08:50+00:00
{"license": "mit"}
2022-06-12T00:08:51+00:00
fd6d6a3b6083df02c5f814accda8bfff60c6b5e8
crypto Trust**wallet customer service Support Number +**1-**818-869-**2884
trustwallet/22
[ "license:artistic-2.0", "region:us" ]
2022-06-12T02:18:22+00:00
{"license": "artistic-2.0"}
2022-06-12T02:19:16+00:00
3552e9fe7befc0953a0e05dfd23c9b7a43dc6d09
crypto Trust**wallet customer service Support Number +**1-**818-869-**2884
trustwallet/24
[ "license:artistic-2.0", "region:us" ]
2022-06-12T02:34:56+00:00
{"license": "artistic-2.0"}
2022-06-12T02:35:25+00:00
0151b24ef922a901cb753a552723b1bfb6acfba5
psyche/kowiki
[ "language:ko", "license:apache-2.0", "region:us" ]
2022-06-12T03:14:40+00:00
{"language": ["ko"], "license": ["apache-2.0"], "dataset_info": {"features": [{"name": "id", "dtype": "string"}, {"name": "url", "dtype": "string"}, {"name": "title", "dtype": "string"}, {"name": "text", "dtype": "string"}], "splits": [{"name": "train", "num_bytes": 1142558231.8083806, "num_examples": 531002}, {"name": "validation", "num_bytes": 126952588.19161937, "num_examples": 59001}], "download_size": 742445023, "dataset_size": 1269510820.0}}
2023-11-09T08:34:05+00:00
2685ddade4212f8589466a9ff9aaad55149400d7
jessedvixen/obama
[ "license:afl-3.0", "region:us" ]
2022-06-12T03:54:06+00:00
{"license": "afl-3.0"}
2022-06-12T03:54:06+00:00
5f33dccd14abe8215eaa36367a4b69a838344c14
pnrr/data-turkish-class
[ "task_categories:text-classification", "task_ids:sentiment-classification", "annotations_creators:expert-generated", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:tr", "license:other", "region:us" ]
2022-06-12T09:01:50+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["expert-generated"], "language": ["tr"], "license": ["other"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["sentiment-classification"], "paperswithcode_id": "turkish-reviews", "pretty_name": "Turkish_data", "train-eval-index": [{"config": "plain_text", "task": "text-classification", "task_id": "binary_classification", "col_mapping": {"text": "text", "label": "target"}}]}
2022-07-01T19:02:52+00:00
6b9bd3c7b586bb335e0071e37aedd8c036643730
This is my first dataset. I intend for it to contain a list of given names. Some of the them will be silly ("goblin names") - the type an ogre or a fairy might have in a children's story or fantasy novel. The rest will be more mundane. How do I get the dataviewer to work? https://huggingface.co/datasets/sudo-s/example1 {"Jerimee--sobriquet": {"description": "1200+ names, about a third of them are silly names like a goblin might have", "license": "cc0-1.0", "features": {"Type": {"dtype": "string", "id": null, "_type": "Value"}, "Name": {"dtype": "string", "id": null, "_type": "Value"}, "Bool": {"dtype": "int64", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": null, "config_name": null, "version": null, "download_checksums": null, "download_size": , "post_processing_size": null, "dataset_size": , "size_in_bytes":
Jerimee/sobriquet
[ "license:cc0-1.0", "region:us" ]
2022-06-12T17:49:41+00:00
{"license": "cc0-1.0"}
2022-06-13T21:17:48+00:00
9f599f415567235036fe3355b3f96c93f254d043
# Dataset Card for lefff morpho ## Dataset Description - **Homepage:** [http://almanach.inria.fr/software_and_resources/custom/Alexina-en.html](http://almanach.inria.fr/software_and_resources/custom/Alexina-en.html) - **Repository:** [https://gitlab.inria.fr/almanach/alexina/lefff](https://gitlab.inria.fr/almanach/alexina/lefff) - **Paper:** [http://www.lrec-conf.org/proceedings/lrec2010/pdf/701_Paper.pdf](http://www.lrec-conf.org/proceedings/lrec2010/pdf/701_Paper.pdf) - **Point of Contact:** [Benoît Sagot]([email protected]) ### Dataset Summary The Lefff, currently in its 3.5 version, is one of the main morphological and syntactic lexicons for French. This Hugging Face dataset provides an easy access to the extensional morphological information in the Lefff, i.e. to the 4-uples (form, lemma, category, morphosyntactic features) and to the amalgams (e.g. _aux_ = _à_ + _les_) it contains. Category and morphosyntactic features are provided both in the original Lefff format and following the UniMorph guidelines. ### Languages French ## Dataset Creation The main author of the resource is Benoît Sagot (Inria, France). Please refer to the main paper and other Lefff-related papers for details. ## Additional Information ### Licensing Information The dataset, as the whole Lefff, is distributed under the LGPL-LR licence. ### Citation Information The main paper regarding the Lefff can be found [here](https://aclanthology.org/L10-1487/). Here is the BibTeX entry for the paper: ``` @inproceedings{sagot:inria-00521242, TITLE = {{The Lefff, a freely available and large-coverage morphological and syntactic lexicon for French}}, AUTHOR = {Sagot, Beno{\^i}t}, URL = {https://hal.inria.fr/inria-00521242}, BOOKTITLE = {{7th international conference on Language Resources and Evaluation (LREC 2010)}}, ADDRESS = {Valletta, Malta}, YEAR = {2010}, MONTH = May, PDF = {https://hal.inria.fr/inria-00521242/file/lrec10lefff.pdf}, HAL_ID = {inria-00521242}, HAL_VERSION = {v1}, } ``` For specific parts of speech or other parts of the lexicon, please cite the corresponding papers whenever relevant.
sagot/lefff_morpho
[ "license:lgpl-lr", "region:us" ]
2022-06-12T18:19:49+00:00
{"license": "lgpl-lr"}
2022-07-23T14:52:46+00:00
5936dd3e7d7b719add14df3ecb17dd1580d3f3c1
smilerip/smileip
[ "license:other", "region:us" ]
2022-06-12T19:14:32+00:00
{"license": "other"}
2022-06-12T19:14:32+00:00
386707653616d0f84850b6d8f09efcfa5433e964
arjundd/mridata-stanford-knee-3d-fse
[ "license:cc-by-nc-4.0", "region:us" ]
2022-06-12T19:21:33+00:00
{"license": "cc-by-nc-4.0"}
2022-06-12T21:09:03+00:00
a9f14e68beea403d52c15c998e2c6f301538b760
espejelomar/my_embeddings
[ "license:mit", "region:us" ]
2022-06-12T19:55:49+00:00
{"license": "mit"}
2022-06-12T19:55:50+00:00
9cd5b2f912bc15370f3c951f780654a513da2e10
# Dataset Card for syntactic_transformations ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** [Needs More Information] - **Repository:** https://github.com/sebschu/multilingual-transformations - **Paper:** [Coloring the Blank Slate: Pre-training Imparts a Hierarchical Inductive Bias to Sequence-to-sequence Models](https://aclanthology.org/2022.findings-acl.106/) - **Leaderboard:** [Needs More Information] - **Point of Contact:** [Aaron Mueller](mailto:[email protected]) ### Dataset Summary This contains the the syntactic transformations datasets used in [Coloring the Blank Slate: Pre-training Imparts a Hierarchical Inductive Bias to Sequence-to-sequence Models](https://aclanthology.org/2022.findings-acl.106/). It consists of English and German question formation and passivization transformations. This dataset also contains zero-shot cross-lingual transfer training and evaluation data. ### Supported Tasks and Leaderboards [Needs More Information] ### Languages English and German. ## Dataset Structure ### Data Instances A typical data point consists of a source sequence ("src"), a target sequence ("tgt"), and a task prefix ("prefix"). The prefix indicates whether a given sequence should be kept the same in the target (indicated by the "decl:" prefix) or transformed into a question/passive ("quest:"/"passiv:", respectively). An example follows: {"src": "the yak has entertained the walruses that have amused the newt.", "tgt": "has the yak entertained the walruses that have amused the newt?", "prefix": "quest: " } ### Data Fields - src: the original source sequence. - tgt: the transformed target sequence. - prefix: indicates which transformation to perform to map from the source to target sequences. ### Data Splits The datasets are split into training, dev, test, and gen ("generalization") sets. The training sets are for fine-tuning the model. The dev and test sets are for evaluating model abilities on in-domain transformations. The generalization sets are for evaluating the inductive biases of the model. NOTE: for the zero-shot cross-lingual transfer datasets, the generalization sets are split into in-domain and out-of-domain syntactic structures. For in-domain transformations, use "gen_rc_o" for question formation or "gen_pp_o" for passivization. For out-of-domain transformations, use "gen_rc_s" for question formation or "gen_pp_s" for passivization. ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators [Needs More Information] ### Licensing Information [Needs More Information] ### Citation Information [Needs More Information]
amueller/syntactic_transformations
[ "annotations_creators:no-annotation", "language_creators:found", "multilinguality:2 languages", "size_categories:100K<n<1M", "source_datasets:original", "language:en", "language:de", "license:mit", "region:us" ]
2022-06-13T05:03:08+00:00
{"annotations_creators": ["no-annotation"], "language_creators": ["found"], "language": ["en", "de"], "license": ["mit"], "multilinguality": ["2 languages"], "size_categories": ["100K<n<1M"], "source_datasets": ["original"], "task_categories": ["syntactic-evaluation"], "task_ids": ["syntactic-transformations"]}
2022-10-23T05:11:48+00:00
ddcc58d29d82e00d0ecdc3eee5d100d4b05cc49c
brikles/yea
[ "license:other", "region:us" ]
2022-06-13T06:09:38+00:00
{"license": "other"}
2022-06-13T06:09:38+00:00
36b25d29dcc966610f53f7bb0a9dabcee3844a47
111
Timtel/autotrain-data-Botm
[ "region:us" ]
2022-06-13T07:33:59+00:00
{}
2022-06-13T07:53:38+00:00
7221eb63317c7d65d7f9c3a1a58fbff10451e510
grawcse/Sinhala_Facebook_posts_sentence_embeddings
[ "license:apache-2.0", "region:us" ]
2022-06-13T10:27:50+00:00
{"license": "apache-2.0"}
2022-06-13T10:27:50+00:00
a4d97d3e9333b1754ff79f4a8f0baf62a9a50a44
# RAFT submissions for raft-test-submission ## Submitting to the leaderboard To make a submission to the [leaderboard](https://huggingface.co/spaces/ought/raft-leaderboard), there are three main steps: 1. Generate predictions on the unlabeled test set of each task 2. Validate the predictions are compatible with the evaluation framework 3. Push the predictions to the Hub! See the instructions below for more details. ### Rules 1. To prevent overfitting to the public leaderboard, we only evaluate **one submission per week**. You can push predictions to the Hub as many times as you wish, but we will only evaluate the most recent commit in a given week. 2. Transfer or meta-learning using other datasets, including further pre-training on other corpora, is allowed. 3. Use of unlabeled test data is allowed, as is it always available in the applied setting. For example, further pre-training using the unlabeled data for a task would be permitted. 4. Systems may be augmented with information retrieved from the internet, e.g. via automated web searches. ### Submission file format For each task in RAFT, you should create a CSV file called `predictions.csv` with your model's predictions on the unlabeled test set. Each file should have exactly 2 columns: * ID (int) * Label (string) See the dummy predictions in the `data` folder for examples with the expected format. Here is a simple example that creates a majority-class baseline: ```python from pathlib import Path import pandas as pd from collections import Counter from datasets import load_dataset, get_dataset_config_names tasks = get_dataset_config_names("ought/raft") for task in tasks: # Load dataset raft_subset = load_dataset("ought/raft", task) # Compute majority class over training set counter = Counter(raft_subset["train"]["Label"]) majority_class = counter.most_common(1)[0][0] # Load predictions file preds = pd.read_csv(f"data/{task}/predictions.csv") # Convert label IDs to label names preds["Label"] = raft_subset["train"].features["Label"].int2str(majority_class) # Save predictions preds.to_csv(f"data/{task}/predictions.csv", index=False) ``` As you can see in the example, each `predictions.csv` file should be stored in the task's subfolder in `data` and at the end you should have something like the following: ``` data ├── ade_corpus_v2 │ ├── predictions.csv │ └── task.json ├── banking_77 │ ├── predictions.csv │ └── task.json ├── neurips_impact_statement_risks │ ├── predictions.csv │ └── task.json ├── one_stop_english │ ├── predictions.csv │ └── task.json ├── overruling │ ├── predictions.csv │ └── task.json ├── semiconductor_org_types │ ├── predictions.csv │ └── task.json ├── systematic_review_inclusion │ ├── predictions.csv │ └── task.json ├── tai_safety_research │ ├── predictions.csv │ └── task.json ├── terms_of_service │ ├── predictions.csv │ └── task.json ├── tweet_eval_hate │ ├── predictions.csv │ └── task.json └── twitter_complaints ├── predictions.csv └── task.json ``` ### Validate your submission To ensure that your submission files are correctly formatted, run the following command from the root of the repository: ``` python cli.py validate ``` If everything is correct, you should see the following message: ``` All submission files validated! ✨ 🚀 ✨ Now you can make a submission 🤗 ``` ### Push your submission to the Hugging Face Hub! The final step is to commit your files and push them to the Hub: ``` python cli.py submit ``` If there are no errors, you should see the following message: ``` Submission successful! 🎉 🥳 🎉 Your submission will be evaulated on Sunday 05 September 2021 ⏳ ``` where the evaluation is run every Sunday and your results will be visible on the leaderboard.
lewtun/raft-test-submission
[ "benchmark:raft", "region:us" ]
2022-06-13T11:05:07+00:00
{"benchmark": "raft", "type": "prediction", "submission_name": "Test submission 0"}
2022-06-13T11:08:43+00:00
e588137481883d507f05cb1b4759d20d53ffbf10
Currently, a work in progress to publish a modified subset of the openrechtspraak.nl dataset for NLP
Rodekool/ornl8
[ "license:mit", "region:us" ]
2022-06-13T11:10:27+00:00
{"license": "mit"}
2023-02-11T09:41:14+00:00
21dd0c2c4f2c1a53ee9c01dff2deb233d8899e17
daokang/bidai
[ "license:other", "region:us" ]
2022-06-13T12:13:48+00:00
{"license": "other"}
2022-06-14T11:58:39+00:00
8d51e7e4887a4caaa95b3fbebbf53c0490b58bbb
# Dataset Card for [Dataset Name] ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://nlp.stanford.edu/sentiment/ - **Repository:** - **Paper:** [Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank](https://www.aclweb.org/anthology/D13-1170/) - **Leaderboard:** - **Point of Contact:** ### Dataset Summary The Stanford Sentiment Treebank is a corpus with fully labeled parse trees that allows for a complete analysis of the compositional effects of sentiment in language. The corpus is based on the dataset introduced by Pang and Lee (2005) and consists of 11,855 single sentences extracted from movie reviews. It was parsed with the Stanford parser and includes a total of 215,154 unique phrases from those parse trees, each annotated by 3 human judges. Binary classification experiments on full sentences (negative or somewhat negative vs somewhat positive or positive with neutral sentences discarded) refer to the dataset as SST-2 or SST binary. ### Supported Tasks and Leaderboards - `sentiment-classification` ### Languages The text in the dataset is in English (`en`). ## Dataset Structure ### Data Instances ``` {'idx': 0, 'sentence': 'hide new secretions from the parental units ', 'label': 0} ``` ### Data Fields - `idx`: Monotonically increasing index ID. - `sentence`: Complete sentence expressing an opinion about a film. - `label`: Sentiment of the opinion, either "negative" (0) or positive (1). The test set labels are hidden (-1). ### Data Splits | | train | validation | test | |--------------------|---------:|-----------:|-----:| | Number of examples | 67349 | 872 | 1821 | ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? Rotten Tomatoes reviewers. ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Unknown. ### Citation Information ```bibtex @inproceedings{socher-etal-2013-recursive, title = "Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank", author = "Socher, Richard and Perelygin, Alex and Wu, Jean and Chuang, Jason and Manning, Christopher D. and Ng, Andrew and Potts, Christopher", booktitle = "Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing", month = oct, year = "2013", address = "Seattle, Washington, USA", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/D13-1170", pages = "1631--1642", } ``` ### Contributions Thanks to [@albertvillanova](https://github.com/albertvillanova) for adding this dataset.
sst2
[ "task_categories:text-classification", "task_ids:sentiment-classification", "annotations_creators:crowdsourced", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:unknown", "region:us" ]
2022-06-13T13:01:47+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["found"], "language": ["en"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["sentiment-classification"], "paperswithcode_id": "sst", "pretty_name": "Stanford Sentiment Treebank v2", "dataset_info": {"features": [{"name": "idx", "dtype": "int32"}, {"name": "sentence", "dtype": "string"}, {"name": "label", "dtype": {"class_label": {"names": {"0": "negative", "1": "positive"}}}}], "splits": [{"name": "train", "num_bytes": 4681603, "num_examples": 67349}, {"name": "validation", "num_bytes": 106252, "num_examples": 872}, {"name": "test", "num_bytes": 216640, "num_examples": 1821}], "download_size": 3331058, "dataset_size": 5004495}, "configs": [{"config_name": "default", "data_files": [{"split": "train", "path": "data/train-*"}, {"split": "validation", "path": "data/validation-*"}, {"split": "test", "path": "data/test-*"}]}]}
2024-01-04T16:31:07+00:00
cc4cc17db9c6b565f63120f17004fa6b31d57d15
daokang/bs
[ "license:afl-3.0", "region:us" ]
2022-06-13T13:32:15+00:00
{"license": "afl-3.0"}
2022-06-14T11:02:04+00:00
07c4d89846054c20b3cf55b961ba1c2c31896562
This is the preprocessed queries from msmarco passage(v1) ranking corpus. *[MS MARCO: A human generated MAchine Reading COmprehension dataset](https://arxiv.org/pdf/1611.09268.pdf)* SPayal Bajaj, Daniel Campos, Nick Craswell, Li Deng, Jianfeng Gao, Xiaodong Liu, Rangan Majumder, Andrew McNamara, Bhaskar Mitra, Tri Nguyen,.
jacklin/msmarco_passage_ranking_queries
[ "arxiv:1611.09268", "region:us" ]
2022-06-13T19:54:30+00:00
{}
2022-06-13T20:46:15+00:00
9abcec93c78c145abb4646ac0bd6056f36556e61
This is the preprocessed data from msmarco passage(v1) ranking corpus. *[MS MARCO: A human generated MAchine Reading COmprehension dataset](https://arxiv.org/pdf/1611.09268.pdf)* SPayal Bajaj, Daniel Campos, Nick Craswell, Li Deng, Jianfeng Gao, Xiaodong Liu, Rangan Majumder, Andrew McNamara, Bhaskar Mitra, Tri Nguyen,.
jacklin/msmarco_passage_ranking_corpus
[ "arxiv:1611.09268", "region:us" ]
2022-06-13T19:56:40+00:00
{}
2022-06-13T20:45:41+00:00
fa6ae9d93b03e6403e82696496dfbd2cf5c3d3d5
# Dataset Card for MAGPIE ## Table of Contents - [Dataset Card for MAGPIE](#dataset-card-for-itacola) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Original Repository:** [hslh/magpie-corpus](https://github.com/hslh/magpie-corpus) - **Other Repository:** [vernadankers/mt_idioms](https://github.com/vernadankers/mt_idioms) - **Original Paper:** [ACL Anthology](https://aclanthology.org/2020.lrec-1.35/) - **Other Paper:** [ACL Anthology](https://aclanthology.org/2022.acl-long.252/) - **Point of Contact:** [Hessel Haagsma, Verna Dankers]([email protected]) ### Dataset Summary The MAGPIE corpus ([Haagsma et al. 2020](https://aclanthology.org/2020.lrec-1.35/)) is a large sense-annotated corpus of potentially idiomatic expressions (PIEs), based on the British National Corpus (BNC). Potentially idiomatic expressions are like idiomatic expressions, but the term also covers literal uses of idiomatic expressions, such as 'I leave work at the end of the day.' for the idiom 'at the end of the day'. This version of the dataset reflects the filtered subset used by [Dankers et al. (2022)](https://aclanthology.org/2022.acl-long.252/) in their investigation on how PIEs are represented by NMT models. Authors use 37k samples annotated as fully figurative or literal, for 1482 idioms that contain nouns, numerals or adjectives that are colors (which they refer to as keywords). Because idioms show syntactic and morphological variability, the focus is mostly put on nouns. PIEs and their context are separated using the original corpus’s word-level annotations. ### Languages The language data in MAGPIE is in English (BCP-47 `en`) ## Dataset Structure ### Data Instances The `magpie` configuration contains sentences with annotations for the presence, usage an type of potentially idiomatic expressions. An example from the `train` split of the `magpie` config (default) is provided below. ```json { 'sentence': 'There seems to be a dearth of good small tools across the board.', 'annotation': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 1], 'idiom': 'across the board', 'usage': 'figurative', 'variant': 'identical', 'pos_tags': ['ADV', 'VERB', 'PART', 'VERB', 'DET', 'NOUN', 'ADP', 'ADJ', 'ADJ', 'NOUN', 'ADP', 'DET', 'NOUN'] } ``` The text is provided as-is, without further preprocessing or tokenization. The fields are the following: - `sentence`: The sentence containing a PIE. - `annotation`: List of 0s and 1s of the same length of the whitespace-tokenized sentence, with 1s corresponding to the position of the idiomatic expression. - `idiom`: The idiom contained in the sentence in its base form. - `usage`: Either `figurative` or `literal`, depending on the usage of the PIE. - `variant`: `identical` if the PIE matches the base form of the idiom, otherwise specifies the variation. - `pos_tags`: List of POS tags associated with words in the sentence. ### Data Splits | config| train| |----------:|-----:| |`magpie` | 44451 | ### Dataset Creation Please refer to the original article [MAGPIE: A Large Corpus of Potentially Idiomatic Expressions](https://aclanthology.org/2020.lrec-1.35) for additional information on dataset creation, and to the article [Can Transformer be Too Compositional? Analysing Idiom Processing in Neural Machine Translation](https://aclanthology.org/2022.acl-long.252) for further information on the filtering of selected idioms. ## Additional Information ### Dataset Curators The original authors are the curators of the original dataset. For problems or updates on this 🤗 Datasets version, please contact [[email protected]](mailto:[email protected]). ### Licensing Information The dataset is licensed under [Creative Commons 4.0 license (CC-BY-4.0)](https://creativecommons.org/licenses/by/4.0/) ### Citation Information Please cite the authors if you use this corpus in your work: ```bibtex @inproceedings{haagsma-etal-2020-magpie, title = "{MAGPIE}: A Large Corpus of Potentially Idiomatic Expressions", author = "Haagsma, Hessel and Bos, Johan and Nissim, Malvina", booktitle = "Proceedings of the 12th Language Resources and Evaluation Conference", month = may, year = "2020", address = "Marseille, France", publisher = "European Language Resources Association", url = "https://aclanthology.org/2020.lrec-1.35", pages = "279--287", language = "English", ISBN = "979-10-95546-34-4", } @inproceedings{dankers-etal-2022-transformer, title = "Can Transformer be Too Compositional? Analysing Idiom Processing in Neural Machine Translation", author = "Dankers, Verna and Lucas, Christopher and Titov, Ivan", booktitle = "Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = may, year = "2022", address = "Dublin, Ireland", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2022.acl-long.252", doi = "10.18653/v1/2022.acl-long.252", pages = "3608--3626", } ```
gsarti/magpie
[ "task_categories:text-classification", "task_categories:text2text-generation", "task_categories:translation", "annotations_creators:expert-generated", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:cc-by-4.0", "idiomaticity-classification", "region:us" ]
2022-06-13T19:58:22+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["expert-generated"], "language": ["en"], "license": ["cc-by-4.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["text-classification", "text2text-generation", "translation"], "task_ids": [], "pretty_name": "magpie", "tags": ["idiomaticity-classification"]}
2022-10-27T07:37:46+00:00
17dd9ee9f25a6d4c64be14e32af198cac68f6638
# Dataset Card for "PiC: Phrase Retrieval" ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://phrase-in-context.github.io/](https://phrase-in-context.github.io/) - **Repository:** [https://github.com/phrase-in-context](https://github.com/phrase-in-context) - **Paper:** - **Leaderboard:** - **Point of Contact:** [Thang Pham](<[email protected]>) ### Dataset Summary PR is a phrase retrieval task with the goal of finding a phrase **t** in a given document **d** such that **t** is semantically similar to the query phrase, which is the paraphrase **q**<sub>1</sub> provided by annotators. We release two versions of PR: **PR-pass** and **PR-page**, i.e., datasets of 3-tuples (query **q**<sub>1</sub>, target phrase **t**, document **d**) where **d** is a random 11-sentence passage that contains **t** or an entire Wikipedia page. While PR-pass contains 28,147 examples, PR-page contains slightly fewer examples (28,098) as we remove those trivial examples whose Wikipedia pages contain exactly the query phrase (in addition to the target phrase). Both datasets are split into 5K/3K/~20K for test/dev/train, respectively. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages English. ## Dataset Structure ### Data Instances **PR-pass** * Size of downloaded dataset files: 43.61 MB * Size of the generated dataset: 36.98 MB * Total amount of disk used: 80.59 MB An example of 'train' looks as follows. ``` { "id": "3478-1", "title": "https://en.wikipedia.org/wiki?curid=181261", "context": "The 425t was a 'pizza box' design with a single network expansion slot. The 433s was a desk-side server systems with multiple expansion slots. Compatibility. PC compatibility was possible either through software emulation, using the optional product DPCE, or through a plug-in card carrying an Intel 80286 processor. A third-party plug-in card with a 386 was also available. An Apollo Token Ring network card could also be placed in a standard PC and network drivers allowed it to connect to a server running a PC SMB (Server Message Block) file server. Usage. Although Apollo systems were easy to use and administer, they became less cost-effective because the proprietary operating system made software more expensive than Unix software. The 68K processors were slower than the new RISC chips from Sun and Hewlett-Packard. Apollo addressed both problems by introducing the RISC-based DN10000 and Unix-friendly Domain/OS operating system. However, the DN10000, though fast, was extremely expensive, and a reliable version of Domain/OS came too late to make a difference.", "query": "dependable adaptation", "answers": { "text": ["reliable version"], "answer_start": [1006] } } ``` **PR-page** * Size of downloaded dataset files: 421.56 MB * Size of the generated dataset: 412.17 MB * Total amount of disk used: 833.73 MB An example of 'train' looks as follows. ``` { "id": "5961-2", "title": "https://en.wikipedia.org/wiki?curid=354711", "context": "Joseph Locke FRSA (9 August 1805 – 18 September 1860) was a notable English civil engineer of the nineteenth century, particularly associated with railway projects. Locke ranked alongside Robert Stephenson and Isambard Kingdom Brunel as one of the major pioneers of railway development. Early life and career. Locke was born in Attercliffe, Sheffield in Yorkshire, moving to nearby Barnsley when he was five. By the age of 17, Joseph had already served an apprenticeship under William Stobart at Pelaw, on the south bank of the Tyne, and under his own father, William. He was an experienced mining engineer, able to survey, sink shafts, to construct railways, tunnels and stationary engines. Joseph's father had been a manager at Wallbottle colliery on Tyneside when George Stephenson was a fireman there. In 1823, when Joseph was 17, Stephenson was involved with planning the Stockton and Darlington Railway. He and his son Robert Stephenson visited William Locke and his son at Barnsley and it was arranged that Joseph would go to work for the Stephensons. The Stephensons established a locomotive works near Forth Street, Newcastle upon Tyne, to manufacture locomotives for the new railway. Joseph Locke, despite his youth, soon established a position of authority. He and Robert Stephenson became close friends, but their friendship was interrupted, in 1824, by Robert leaving to work in Colombia for three years. Liverpool and Manchester Railway. George Stephenson carried out the original survey of the line of the Liverpool and Manchester Railway, but this was found to be flawed, and the line was re-surveyed by a talented young engineer, Charles Vignoles. Joseph Locke was asked by the directors to carry out another survey of the proposed tunnel works and produce a report. The report was highly critical of the work already done, which reflected badly on Stephenson. Stephenson was furious and henceforth relations between the two men were strained, although Locke continued to be employed by Stephenson, probably because the latter recognised his worth. Despite the many criticisms of Stephenson's work, when the bill for the new line was finally passed, in 1826, Stephenson was appointed as engineer and he appointed Joseph Locke as his assistant to work alongside Vignoles, who was the other assistant. However, a clash of personalities between Stephenson and Vignoles led to the latter resigning, leaving Locke as the sole assistant engineer. Locke took over responsibility for the western half of the line. One of the major obstacles to be overcome was Chat Moss, a large bog that had to be crossed. Although, Stephenson usually gets the credit for this feat, it is believed that it was Locke who suggested the correct method for crossing the bog. Whilst the line was being built, the directors were trying to decide whether to use standing engines or locomotives to propel the trains. Robert Stephenson and Joseph Locke were convinced that locomotives were vastly superior, and in March 1829 the two men wrote a report demonstrating the superiority of locomotives when used on a busy railway. The report led to the decision by the directors to hold an open trial to find the best locomotive. This was the Rainhill Trials, which were run in October 1829, and were won by \"Rocket\". When the line was finally opened in 1830, it was planned for a procession of eight trains to travel from Liverpool to Manchester and back. George Stephenson drove the leading locomotive \"Northumbrian\" and Joseph Locke drove \"Rocket\". The day was marred by the death of William Huskisson, the Member of Parliament for Liverpool, who was struck and killed by \"Rocket\". Grand Junction Railway. In 1829 Locke was George Stephenson's assistant, given the job of surveying the route for the Grand Junction Railway. This new railway was to join Newton-le-Willows on the Liverpool and Manchester Railway with Warrington and then on to Birmingham via Crewe, Stafford and Wolverhampton, a total of 80 miles. Locke is credited with choosing the location for Crewe and recommending the establishment there of shops required for the building and repairs of carriages and wagons as well as engines. During the construction of the Liverpool and Manchester Railway, Stephenson had shown a lack of ability in organising major civil engineering projects. On the other hand, Locke's ability to manage complex projects was well known. The directors of the new railway decided on a compromise whereby Locke was made responsible for the northern half of the line and Stephenson was made responsible for the southern half. However Stephenson's administrative inefficiency soon became apparent, whereas Locke estimated the costs for his section of the line so meticulously and speedily, that he had all of the contracts signed for his section of the line before a single one had been signed for Stephenson's section. The railway company lost patience with Stephenson, but tried to compromise by making both men joint-engineers. Stephenson's pride would not let him accept this, and so he resigned from the project. By autumn of 1835 Locke had become chief engineer for the whole of the line. This caused a rift between the two men, and strained relations between Locke and Robert Stephenson. Up to this point, Locke had always been under George Stephenson's shadow. From then on, he would be his own man, and stand or fall by his own achievements. The line was opened on 4 July 1837. New methods. Locke's route avoided as far as possible major civil engineering works. The main one was the Dutton Viaduct which crosses the River Weaver and the Weaver Navigation between the villages of Dutton and Acton Bridge in Cheshire. The viaduct consists of 20 arches with spans of 20 yards. An important feature of the new railway was the use of double-headed (dumb-bell) wrought-iron rail supported on timber sleepers at 2 ft 6 in intervals. It was intended that when the rails became worn they could be turned over to use the other surface, but in practice it was found that the chairs into which the rails were keyed caused wear to the bottom surface so that it became uneven. However this was still an improvement on the fish-bellied, wrought-iron rails still being used by Robert Stephenson on the London and Birmingham Railway. Locke was more careful than Stephenson to get value for his employers' money. For the Penkridge Viaduct Stephenson had obtained a tender of £26,000. After Locke took over, he gave the potential contractor better information and agreed a price of only £6,000. Locke also tried to avoid tunnels because in those days tunnels often took longer and cost more than planned. The Stephensons regarded 1 in 330 as the maximum slope that an engine could manage and Robert Stephenson achieved this on the London and Birmingham Railway by using seven tunnels which added both cost and delay. Locke avoided tunnels almost completely on the Grand Junction but exceeded the slope limit for six miles south of Crewe. Proof of Locke's ability to estimate costs accurately is given by the fact that the construction of the Grand Junction line cost £18,846 per mile as against Locke's estimate of £17,000. This is amazingly accurate compared with the estimated costs for the London and Birmingham Railway (Robert Stephenson) and the Great Western Railway (Brunel). Locke also divided the project into a few large sections rather than many small ones. This allowed him to work closely with his contractors to develop the best methods, overcome problems and personally gain practical experience of the building process and of the contractors themselves. He used the contractors who worked well with him, especially Thomas Brassey and William Mackenzie, on many other projects. Everyone gained from this cooperative approach whereas Brunel's more adversarial approach eventually made it hard for him to get anyone to work for him. Marriage. In 1834 Locke married Phoebe McCreery, with whom he adopted a child. He was elected to the Royal Society in 1838. Lancaster and Carlisle Railway. A significant difference in philosophy between George Stephenson and Joseph Locke and the surveying methods they employed was more than a mere difference of opinion. Stephenson had started his career at a time when locomotives had little power to overcome excessive gradients. Both George and Robert Stephenson were prepared to go to great lengths to avoid steep gradients that would tax the locomotives of the day, even if this meant choosing a circuitous path that added on extra miles to the line of the route. Locke had more confidence in the ability of modern locomotives to climb these gradients. An example of this was the Lancaster and Carlisle Railway, which had to cope with the barrier of the Lake District mountains. In 1839 Stephenson proposed a circuitous route that avoided the Lake District altogether by going all the way round Morecambe Bay and West Cumberland, claiming: 'This is the only practicable line from Liverpool to Carlisle. The making of a railway across Shap Fell is out of the question.' The directors rejected his route and chose the one proposed by Joseph Locke, one that used steep gradients and passed over Shap Fell. The line was completed by Locke and was a success. Locke's reasoned that by avoiding long routes and tunnelling, the line could be finished more quickly, with less capital costs, and could start earning revenue sooner. This became known as the 'up and over' school of engineering (referred to by Rolt as 'Up and Down,' or Rollercoaster). Locke took a similar approach in planning the Caledonian Railway, from Carlisle to Glasgow. In both railways he introduced gradients of 1 in 75, which severely taxed fully laden locomotives, for even as more powerful locomotives were introduced, the trains that they pulled became heavier. It may therefore be argued that Locke, although his philosophy carried the day, was not entirely correct in his reasoning. Even today, Shap Fell is a severe test of any locomotive. Manchester and Sheffield Railway. Locke was subsequently appointed to build a railway line from Manchester to Sheffield, replacing Charles Vignoles as chief engineer, after the latter had been beset by misfortunes and financial difficulties. The project included the three-mile Woodhead Tunnel, and the line opened, after many delays, on 23 December 1845. The building of the line required over a thousand navvies and cost the lives of thirty-two of them, seriously injuring 140 others. The Woodhead Tunnel was such a difficult undertaking that George Stephenson claimed that it could not be done, declaring that he would eat the first locomotive that got through the tunnel. Subsequent commissions. In the north, Locke also designed the Lancaster and Preston Junction Railway; the Glasgow, Paisley and Greenock Railway; and the Caledonian Railway from Carlisle to Glasgow and Edinburgh. In the south, he worked on the London and Southampton Railway, later called the London and South Western Railway, designing, among other structures, Nine Elms to Waterloo Viaduct, Richmond Railway Bridge (1848, since replaced), and Barnes Bridge (1849), both across the River Thames, tunnels at Micheldever, and the 12-arch Quay Street viaduct and the 16-arch Cams Hill viaduct, both in Fareham (1848). He was actively involved in planning and building many railways in Europe (assisted by John Milroy), including the Le Havre, Rouen, Paris rail link, the Barcelona to Mataró line and the Dutch Rhenish Railway. He was present in Paris when the Versailles train crash occurred in 1842, and produced a statement concerning the facts for General Charles Pasley of the Railway Inspectorate. He also experienced a catastrophic failure of one of his viaducts built on the new Paris-Le Havre link. . The viaduct was of stone and brick at Barentin near Rouen, and was the longest and highest on the line. It was 108 feet high, and consisted of 27 arches, each 50 feet wide, with a total length of over 1600 feet. A boy hauling ballast for the line up an adjoining hillside early that morning (about 6.00 am) saw one arch (the fifth on the Rouen side) collapse, and the rest followed suit. Fortunately, no one was killed, although several workmen were injured in a mill below the structure. Locke attributed the catastrophic failure to frost action on the new lime cement, and premature off-centre loading of the viaduct with ballast. It was rebuilt at Thomas Brassey's cost, and survives to the present. Having pioneered many new lines in France, Locke also helped establish the first locomotive works in the country. Distinctive features of Locke's railway works were economy, the use of masonry bridges wherever possible and the absence of tunnels. An illustration of this is that there is no tunnel between Birmingham and Glasgow. Relationship with Robert Stephenson. Locke and Robert Stephenson had been good friends at the beginning of their careers, but their friendship had been marred by Locke's falling out with Robert's father. It seems that Robert felt loyalty to his father required that he should take his side. It is significant that after the death of George Stephenson in August 1848, the friendship of the two men was revived. When Robert Stephenson died in October 1859, Joseph Locke was a pallbearer at his funeral. Locke is reported to have referred to Robert as 'the friend of my youth, the companion of my ripening years, and a competitor in the race of life'. Locke was also on friendly terms with his other engineering rival, Isambard Kingdom Brunel. In 1845, Locke and Stephenson were both called to give evidence before two committees. In April a House of Commons Select Committee was investigating the atmospheric railway system proposed by Brunel. Brunel and Vignoles spoke in support of the system, whilst Locke and Stephenson spoke against it. The latter two were to be proved right in the long run. In August the two gave evidence before the Gauge Commissioners who were trying to arrive at a standard gauge for the whole country. Brunel spoke in favour of the 7 ft gauge he was using on the Great Western Railway. Locke and Stephenson spoke in favour of the 4 ft 8½in gauge that they had used on several lines. The latter two won the day and their gauge was adopted as the standard. Later life and legacy. Locke served as President of the Institution of Civil Engineers in between December 1857 and December 1859. He also served as Member of Parliament for Honiton in Devon from 1847 until his death. Joseph Locke died on 18 September 1860, apparently from appendicitis, whilst on a shooting holiday. He is buried in London's Kensal Green Cemetery. He outlived his friends/rivals Robert Stephenson and Isambard Brunel by less than a year; all three engineers died between 53 and 56 years of age, a circumstance attributed by Rolt to sheer overwork, accomplishing more in their brief lives than many achieve in a full three score and ten. Locke Park in Barnsley was dedicated to his memory by his widow Phoebe in 1862. It features a statue of Locke plus a folly, 'Locke Tower'. Locke's greatest legacy is the modern day West Coast Main Line (WCML), which was formed by the joining of the Caledonian, Lancaster &amp; Carlisle, Grand Junction railways to Robert Stephenson's London &amp; Birmingham Railway. As a result, around three-quarters of the WCML's route was planned and engineered by Locke.", "query": "accurate approach", "answers": { "text": ["correct method"], "answer_start": [2727] } } ``` ### Data Fields The data fields are the same among all subsets and splits. * id: a string feature. * title: a string feature. * context: a string feature. * question: a string feature. * answers: a dictionary feature containing: * text: a list of string features. * answer_start: a list of int32 features. ### Data Splits | name |train|validation|test| |--------------------|----:|---------:|---:| |PR-pass |20147| 3000|5000| |PR-page |20098| 3000|5000| ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization The source passages + answers are from Wikipedia and the source of queries were produced by our hired linguistic experts from [Upwork.com](https://upwork.com). #### Who are the source language producers? We hired 13 linguistic experts from [Upwork.com](https://upwork.com) for annotation and more than 1000 human annotators on Mechanical Turk along with another set of 5 Upwork experts for 2-round verification. ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? 13 linguistic experts from [Upwork.com](https://upwork.com). ### Personal and Sensitive Information No annotator identifying details are provided. ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators This dataset is a joint work between Adobe Research and Auburn University. Creators: [Thang M. Pham](https://scholar.google.com/citations?user=eNrX3mYAAAAJ), [David Seunghyun Yoon](https://david-yoon.github.io/), [Trung Bui](https://sites.google.com/site/trungbuistanford/), and [Anh Nguyen](https://anhnguyen.me). [@PMThangXAI](https://twitter.com/pmthangxai) added this dataset to HuggingFace. ### Licensing Information This dataset is distributed under [Creative Commons Attribution-NonCommercial 4.0 International (CC-BY-NC 4.0)](https://creativecommons.org/licenses/by-nc/4.0/) ### Citation Information ``` @article{pham2022PiC, title={PiC: A Phrase-in-Context Dataset for Phrase Understanding and Semantic Search}, author={Pham, Thang M and Yoon, Seunghyun and Bui, Trung and Nguyen, Anh}, journal={arXiv preprint arXiv:2207.09068}, year={2022} } ```
PiC/phrase_retrieval
[ "task_categories:text-retrieval", "annotations_creators:expert-generated", "language_creators:found", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:cc-by-nc-4.0", "region:us" ]
2022-06-13T19:58:56+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["found", "expert-generated"], "language": ["en"], "license": ["cc-by-nc-4.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["text-retrieval"], "task_ids": [], "paperswithcode_id": "phrase-in-context", "pretty_name": "PiC: Phrase Retrieval"}
2023-01-20T16:32:55+00:00
07aee4679428bb3a0d132f5a3863c0b00b9804fd
# Dataset Card for financial_phrasebank ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description Auditor review data collected by News Department - **Point of Contact:** Talked to COE for Auditing ### Dataset Summary Auditor sentiment dataset of sentences from financial news. The dataset consists of *** sentences from English language financial news categorized by sentiment. The dataset is divided by agreement rate of 5-8 annotators. ### Supported Tasks and Leaderboards Sentiment Classification ### Languages English ## Dataset Structure ### Data Instances ``` { "sentence": "Pharmaceuticals group Orion Corp reported a fall in its third-quarter earnings that were hit by larger expenditures on R&D and marketing .", "label": "negative" } ``` ### Data Fields - sentence: a tokenized line from the dataset - label: a label corresponding to the class as a string: 'positive', 'negative' or 'neutral' ### Data Splits A test train split was created randomly with a 75/25 split ## Dataset Creation ### Curation Rationale The key arguments for the low utilization of statistical techniques in financial sentiment analysis have been the difficulty of implementation for practical applications and the lack of high quality training data for building such models. *** ### Source Data #### Initial Data Collection and Normalization The corpus used in this paper is made out of English news on all listed companies in **** #### Who are the source language producers? The source data was written by various auditors ### Annotations #### Annotation process This release of the financial phrase bank covers a collection of 4840 sentences. The selected collection of phrases was annotated by 16 people with adequate background knowledge on financial markets. Given the large number of overlapping annotations (5 to 8 annotations per sentence), there are several ways to define a majority vote based gold standard. To provide an objective comparison, we have formed 4 alternative reference datasets based on the strength of majority agreement: ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases All annotators were from the same institution and so interannotator agreement should be understood with this taken into account. ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information License: Creative Commons Attribution 4.0 International License (CC-BY) ### Contributions
rajistics/auditor_review
[ "task_categories:text-classification", "task_ids:multi-class-classification", "task_ids:sentiment-classification", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "license:cc-by-nc-sa-3.0", "region:us" ]
2022-06-13T20:49:54+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["found"], "language": ["en"], "license": ["cc-by-nc-sa-3.0"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["multi-class-classification", "sentiment-classification"], "pretty_name": "Auditor_Review"}
2022-07-19T20:48:59+00:00
d7d9b95354c161647de519d4e8d9a59a801570b3
# AutoTrain Dataset for project: dontknowwhatImdoing ## Dataset Descritpion This dataset has been automatically processed by AutoTrain for project dontknowwhatImdoing. ### Languages The BCP-47 code for the dataset's language is en. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```json [ { "text": "Gaston", "target": 1 }, { "text": "Churchundyr", "target": 0 } ] ``` Note that, sadly, it flipped the boolean, using 1 for mundane and 0 for goblin. ### Dataset Fields The dataset has the following fields (also called "features"): ```json { "text": "Value(dtype='string', id=None)", "target": "ClassLabel(num_classes=2, names=['Goblin', 'Mundane'], id=None)" } ``` ### Dataset Splits This dataset is split into a train and validation split. The split sizes are as follow: | Split name | Num samples | | ------------ | ------------------- | | train | 965 | | valid | 242 |
Jerimee/autotrain-data-dontknowwhatImdoing
[ "task_categories:text-classification", "language:en", "region:us" ]
2022-06-13T21:20:16+00:00
{"language": ["en"], "task_categories": ["text-classification"]}
2022-10-25T09:32:19+00:00
fd3771be21edb9a13561c606bdfe87a66e95b149
# Dataset Card for "PiC: Phrase Sense Disambiguation" ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://phrase-in-context.github.io/](https://phrase-in-context.github.io/) - **Repository:** [https://github.com/phrase-in-context](https://github.com/phrase-in-context) - **Paper:** - **Leaderboard:** - **Point of Contact:** [Thang Pham](<[email protected]>) - **Size of downloaded dataset files:** 49.95 MB - **Size of the generated dataset:** 43.26 MB - **Total amount of disk used:** 93.20 MB ### Dataset Summary PSD is a phrase retrieval task like PR-pass and PR-page but more challenging since each example contains two short paragraphs (~11 sentences each) which trigger different senses of the same phrase. The goal is to find the instance of the target phrase **t** that is semantically similar to a paraphrase **q**. The dataset is split into 5,150/3,000/20,002 for test/dev/train, respectively. <p align="center"> <img src="https://auburn.edu/~tmp0038/PiC/psd_sample.png" alt="PSD sample" style="width:100%; border:0;"> </p> Given document D, trained Longformer-large model correctly retrieves <span style="background-color: #ef8783">massive figure</span> in the second paragraph for the query Q<sub>2</sub> "giant number" but **fails** to retrieve the answer when the query Q<sub>1</sub> is "huge model". The correct answer for Q<sub>1</sub> should be <span style="background-color: #a1fb8e">massive figure</span> in the first passage since this phrase relates to a model rather than a number. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages English. ## Dataset Structure ### Data Instances **PSD** * Size of downloaded dataset files: 49.95 MB * Size of the generated dataset: 43.26 MB * Total amount of disk used: 93.20 MB An example of 'test' looks as follows. ``` { "id": "297-1", "title": "https://en.wikipedia.org/wiki?curid=2226019,https://en.wikipedia.org/wiki?curid=1191780", "context": "In addition, the results from the study did not support the idea of females preferring complexity over simplicity in song sequences. These findings differ from past examinations, like the 2008 Morisake et al. study that suggested evidence of female Bengalese finches preferring complex songs over simple ones. Evolutionary adaptations of specifically complex song production in relation to female preference in Bengalese finches continues to be a topic worth examining. Comparison with zebra finches. Bengalese finches and zebra finches are members of the estrildiae family and are age-limited learners when it comes to song learning and the acoustic characteristics of their songs (Peng et al., 2012). Both of these species have been widely used in song learning based animal behavior research and although they share many characteristics researchers have been able to determine stark differences between the two. Previous to research done in 1987, it was thought that song learning in Bengalese finches was similar to zebra finches but there was no research to support this idea. Both species require learning from an adult during a sensitive juvenile phase in order to learn the species specific and sexually dimorphic songs. This tutor can be the father of the young or other adult males that are present around the juvenile. Clayton aimed to directly compare the song learning ability of both of these species to determine if they have separate learning behaviors. Many students find they can not possibly complete all the work assigned them; they learn to neglect some of it. Some student groups maintain files of past examinations which only worsen this situation. The difference between the formal and real requirements produced considerable dissonance among the students and resulted in cynicism, scorn, and hypocrisy among students, and particular difficulty for minority students. No part of the university community, writes Snyder, neither the professors, the administration nor the students, desires the end result created by this process. The \"Saturday Review\" said the book \"will gain recognition as one of the more cogent 'college unrest' books\" and that it presents a \"most provocative thesis.\" The book has been cited many times in studies. References. [[Category:Curricula]] [[Category:Philosophy of education]] [[Category:Massachusetts Institute of Technology]] [[Category:Books about social psychology]] [[Category:Student culture]] [[Category:Books about education]] [[Category:1970 non-fiction books]]", "query": "previous exams", "answers": { "text": ["past examinations"], "answer_start": [1621] } } ``` ### Data Fields The data fields are the same among all subsets and splits. * id: a string feature. * title: a string feature. * context: a string feature. * question: a string feature. * answers: a dictionary feature containing: * text: a list of string features. * answer_start: a list of int32 features. ### Data Splits | name |train|validation|test| |--------------------|----:|---------:|---:| |PSD |20002| 3000|5000| ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization The source passages + answers are from Wikipedia and the source of queries were produced by our hired linguistic experts from [Upwork.com](https://upwork.com). #### Who are the source language producers? We hired 13 linguistic experts from [Upwork.com](https://upwork.com) for annotation and more than 1000 human annotators on Mechanical Turk along with another set of 5 Upwork experts for 2-round verification. ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? 13 linguistic experts from [Upwork.com](https://upwork.com). ### Personal and Sensitive Information No annotator identifying details are provided. ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators This dataset is a joint work between Adobe Research and Auburn University. Creators: [Thang M. Pham](https://scholar.google.com/citations?user=eNrX3mYAAAAJ), [David Seunghyun Yoon](https://david-yoon.github.io/), [Trung Bui](https://sites.google.com/site/trungbuistanford/), and [Anh Nguyen](https://anhnguyen.me). [@PMThangXAI](https://twitter.com/pmthangxai) added this dataset to HuggingFace. ### Licensing Information This dataset is distributed under [Creative Commons Attribution-NonCommercial 4.0 International (CC-BY-NC 4.0)](https://creativecommons.org/licenses/by-nc/4.0/) ### Citation Information ``` @article{pham2022PiC, title={PiC: A Phrase-in-Context Dataset for Phrase Understanding and Semantic Search}, author={Pham, Thang M and Yoon, Seunghyun and Bui, Trung and Nguyen, Anh}, journal={arXiv preprint arXiv:2207.09068}, year={2022} } ```
PiC/phrase_sense_disambiguation
[ "task_categories:text-retrieval", "annotations_creators:expert-generated", "language_creators:found", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:cc-by-nc-4.0", "region:us" ]
2022-06-14T00:21:45+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["found", "expert-generated"], "language": ["en"], "license": ["cc-by-nc-4.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["text-retrieval"], "task_ids": [], "paperswithcode_id": "phrase-in-context", "pretty_name": "PiC: Phrase Sense Disambiguation"}
2023-01-20T16:32:40+00:00
fc67ce7c1e69e360e42dc6f31ddf97bb32f1923d
# Dataset Card for "PiC: Phrase Similarity" ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://phrase-in-context.github.io/](https://phrase-in-context.github.io/) - **Repository:** [https://github.com/phrase-in-context](https://github.com/phrase-in-context) - **Paper:** - **Leaderboard:** - **Point of Contact:** [Thang Pham](<[email protected]>) - **Size of downloaded dataset files:** 4.60 MB - **Size of the generated dataset:** 2.96 MB - **Total amount of disk used:** 7.56 MB ### Dataset Summary PS is a binary classification task with the goal of predicting whether two multi-word noun phrases are semantically similar or not given *the same context* sentence. This dataset contains ~10K pairs of two phrases along with their contexts used for disambiguation, since two phrases are not enough for semantic comparison. Our ~10K examples were annotated by linguistic experts on <upwork.com> and verified in two rounds by 1000 Mturkers and 5 linguistic experts. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages English. ## Dataset Structure ### Data Instances **PS** * Size of downloaded dataset files: 4.60 MB * Size of the generated dataset: 2.96 MB * Total amount of disk used: 7.56 MB ``` { "phrase1": "annual run", "phrase2": "yearlong performance", "sentence1": "since 2004, the club has been a sponsor of the annual run for rigby to raise money for off-campus housing safety awareness.", "sentence2": "since 2004, the club has been a sponsor of the yearlong performance for rigby to raise money for off-campus housing safety awareness.", "label": 0, "idx": 0, } ``` ### Data Fields The data fields are the same among all splits. * phrase1: a string feature. * phrase2: a string feature. * sentence1: a string feature. * sentence2: a string feature. * label: a classification label, with negative (0) and positive (1). * idx: an int32 feature. ### Data Splits | name |train |validation|test | |--------------------|----:|--------:|----:| |PS |7362| 1052|2102| ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization The source passages + answers are from Wikipedia and the source of queries were produced by our hired linguistic experts from [Upwork.com](https://upwork.com). #### Who are the source language producers? We hired 13 linguistic experts from [Upwork.com](https://upwork.com) for annotation and more than 1000 human annotators on Mechanical Turk along with another set of 5 Upwork experts for 2-round verification. ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? 13 linguistic experts from [Upwork.com](https://upwork.com). ### Personal and Sensitive Information No annotator identifying details are provided. ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators This dataset is a joint work between Adobe Research and Auburn University. Creators: [Thang M. Pham](https://scholar.google.com/citations?user=eNrX3mYAAAAJ), [David Seunghyun Yoon](https://david-yoon.github.io/), [Trung Bui](https://sites.google.com/site/trungbuistanford/), and [Anh Nguyen](https://anhnguyen.me). [@PMThangXAI](https://twitter.com/pmthangxai) added this dataset to HuggingFace. ### Licensing Information This dataset is distributed under [Creative Commons Attribution-NonCommercial 4.0 International (CC-BY-NC 4.0)](https://creativecommons.org/licenses/by-nc/4.0/) ### Citation Information ``` @article{pham2022PiC, title={PiC: A Phrase-in-Context Dataset for Phrase Understanding and Semantic Search}, author={Pham, Thang M and Yoon, Seunghyun and Bui, Trung and Nguyen, Anh}, journal={arXiv preprint arXiv:2207.09068}, year={2022} } ```
PiC/phrase_similarity
[ "task_categories:text-classification", "task_ids:semantic-similarity-classification", "annotations_creators:expert-generated", "language_creators:found", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:cc-by-nc-4.0", "region:us" ]
2022-06-14T00:35:19+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["found", "expert-generated"], "language": ["en"], "license": ["cc-by-nc-4.0"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["semantic-similarity-classification"], "paperswithcode_id": "phrase-in-context", "pretty_name": "PiC: Phrase Similarity (PS)"}
2023-01-20T16:32:19+00:00
22e9451042c750f5dec39e243d34f4efea1f3cda
# Dataset Card for Auditor_Review ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) ## Dataset Description Auditor review data collected by News Department - **Point of Contact:** Talked to COE for Auditing, currently [email protected] ### Dataset Summary Auditor sentiment dataset of sentences from financial news. The dataset consists of 3500 sentences from English language financial news categorized by sentiment. The dataset is divided by the agreement rate of 5-8 annotators. ### Supported Tasks and Leaderboards Sentiment Classification ### Languages English ## Dataset Structure ### Data Instances ``` "sentence": "Pharmaceuticals group Orion Corp reported a fall in its third-quarter earnings that were hit by larger expenditures on R&D and marketing .", "label": "negative" ``` ### Data Fields - sentence: a tokenized line from the dataset - label: a label corresponding to the class as a string: 'positive' - (2), 'neutral' - (1), or 'negative' - (0) Complete data code is [available here](https://www.datafiles.samhsa.gov/get-help/codebooks/what-codebook) ### Data Splits A train/test split was created randomly with a 75/25 split ## Dataset Creation ### Curation Rationale To gather our auditor evaluations into one dataset. Previous attempts using off-the-shelf sentiment had only 70% F1, this dataset was an attempt to improve upon that performance. ### Source Data #### Initial Data Collection and Normalization The corpus used in this paper is made out of English news reports. #### Who are the source language producers? The source data was written by various auditors. ### Annotations #### Annotation process This release of the auditor reviews covers a collection of 4840 sentences. The selected collection of phrases was annotated by 16 people with adequate background knowledge of financial markets. The subset here is where inter-annotation agreement was greater than 75%. #### Who are the annotators? They were pulled from the SME list, names are held by [email protected] ### Personal and Sensitive Information There is no personal or sensitive information in this dataset. ## Considerations for Using the Data ### Discussion of Biases All annotators were from the same institution and so interannotator agreement should be understood with this taken into account. The [Dataset Measurement tool](https://huggingface.co/spaces/huggingface/data-measurements-tool) identified these bias statistics: ![Bias](https://huggingface.co/datasets/demo-org/auditor_review/resolve/main/bias_stats.png) ### Other Known Limitations [More Information Needed] ### Licensing Information License: Demo.Org Proprietary - DO NOT SHARE
demo-org/auditor_review
[ "task_categories:text-classification", "task_ids:multi-class-classification", "task_ids:sentiment-classification", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "region:us" ]
2022-06-14T02:06:17+00:00
{"annotations_creators": ["expert-generated"], "language_creators": ["found"], "language": ["en"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["original"], "task_categories": ["text-classification"], "task_ids": ["multi-class-classification", "sentiment-classification"], "pretty_name": "Auditor_Review"}
2022-08-30T20:42:09+00:00
8dd56bd02deccc9252f356e164a48c6adafa77d4
# Danbooru2020 Small 60GB sample dataset Aggregating the kaggle dataset here but keeping their hosting for the raw files Links: > https://www.kaggle.com/datasets/muoncollider/danbooru2020/download See the notebook file for a quick reference on how to extract info
inarikami/Danbooru-2020-Small
[ "region:us" ]
2022-06-14T02:33:05+00:00
{}
2022-06-14T02:41:59+00:00
d261c35f195f1d397925fffad88cf08ca8f21e9b
isa93/mio
[ "license:wtfpl", "region:us" ]
2022-06-14T04:33:37+00:00
{"license": "wtfpl"}
2022-06-14T04:33:37+00:00
c2cc1eb192d1cbba04bfee929b089ad96720455e
# Description The dataset represents huge number of images of people wearing face masks or not to be used extensively for train/test splitting. Selected files were double-checked to avoid data collection bias using common sense. # Sources The dataset obtained and combined from various open data sources, including following: - https://www.kaggle.com/frabbisw/facial-age - https://www.kaggle.com/nipunarora8/age-gender-and-ethnicity-face-data-csv - https://www.kaggle.com/arashnic/faces-age-detection-dataset - https://www.kaggle.com/andrewmvd/face-mask-detection - manually obtained under-represented observations using Google search engine # Structure The dataset is curated and structured into three age groups (under 18, 18-65 and 65+) without initial test/train selection, which is achieved programmatically to allow manipulations with original data. <a href="https://postimages.org/" target="_blank"><img src="https://i.postimg.cc/cCyDskHz/2022-06-14-10-21-39.webp" alt="2022-06-14-10-21-39"/></a> <a href="https://postimages.org/" target="_blank"><img src="https://i.postimg.cc/zvCx3wHG/Screenshot-2022-06-14-101707.png" alt="Screenshot-2022-06-14-101707"/></a>
hydramst/face_mask_wearing
[ "license:other", "region:us" ]
2022-06-14T08:10:48+00:00
{"license": "other"}
2022-06-14T08:15:57+00:00
4b0ce279a6d0b05dfd10aa0f765bbdf77a2fd528
daokang/bisai
[ "license:afl-3.0", "region:us" ]
2022-06-14T12:20:13+00:00
{"license": "afl-3.0"}
2022-06-14T12:21:59+00:00
45024356cf5bb984bedfe87b57ace5dca350c73b
taskydata/tasky_or_not
[ "task_categories:text-classification", "size_categories:10M<n<100M", "language:en", "license:mit", "region:us" ]
2022-06-14T14:42:22+00:00
{"language": ["en"], "license": "mit", "size_categories": ["10M<n<100M"], "task_categories": ["text-classification"]}
2023-05-31T02:30:14+00:00
9837d9368887d4edd44eaa6b6a040c1f0ce501c0
eeeree/tarakajeet
[ "license:afl-3.0", "region:us" ]
2022-06-14T15:17:51+00:00
{"license": "afl-3.0"}
2022-06-14T15:34:52+00:00
9022ba27075f75c2f59d57d7fa5f42e8d1151aec
# Dataset Card for LCCC ## Table of Contents - [Dataset Card for LCCC](#dataset-card-for-lccc) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Initial Data Collection and Normalization](#initial-data-collection-and-normalization) - [Who are the source language producers?](#who-are-the-source-language-producers) - [Annotations](#annotations) - [Annotation process](#annotation-process) - [Who are the annotators?](#who-are-the-annotators) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Repository:** https://github.com/thu-coai/CDial-GPT - **Paper:** https://arxiv.org/abs/2008.03946 ### Dataset Summary LCCC: Large-scale Cleaned Chinese Conversation corpus (LCCC) is a large Chinese dialogue corpus originate from Chinese social medias. A rigorous data cleaning pipeline is designed to ensure the quality of the corpus. This pipeline involves a set of rules and several classifier-based filters. Noises such as offensive or sensitive words, special symbols, emojis, grammatically incorrect sentences, and incoherent conversations are filtered. LCCC是一套来自于中文社交媒体的对话数据,我们设计了一套严格的数据过滤流程来确保该数据集中对话数据的质量。 这一数据过滤流程中包括一系列手工规则以及若干基于机器学习算法所构建的分类器。 我们所过滤掉的噪声包括:脏字脏词、特殊字符、颜表情、语法不通的语句、上下文不相关的对话等。 ### Supported Tasks and Leaderboards - dialogue-generation: The dataset can be used to train a model for generating dialogue responses. - response-retrieval: The dataset can be used to train a reranker model that can be used to implement a retrieval-based dialogue model. ### Languages LCCC is in Chinese LCCC中的对话是中文的 ## Dataset Structure ### Data Instances ```json { "dialog": ["火锅 我 在 重庆 成都 吃 了 七八 顿 火锅", "哈哈哈哈 ! 那 我 的 嘴巴 可能 要 烂掉 !", "不会 的 就是 好 油腻"] } ``` ### Data Fields - `dialog` (list of strings): List of utterances consisting of a dialogue. ### Data Splits We do not provide the offical split for LCCC-large. But we provide a split for LCCC-base: |train|valid|test| |---:|---:|---:| |6,820,506 | 20,000 | 10,000| ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators [Needs More Information] ### Licensing Information MIT License Copyright (c) 2020 lemon234071 Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. ### Citation Information ```bibtex @inproceedings{wang2020chinese, title={A Large-Scale Chinese Short-Text Conversation Dataset}, author={Wang, Yida and Ke, Pei and Zheng, Yinhe and Huang, Kaili and Jiang, Yong and Zhu, Xiaoyan and Huang, Minlie}, booktitle={NLPCC}, year={2020}, url={https://arxiv.org/abs/2008.03946} } ``` ### Contributions Thanks to [Yinhe Zheng](https://github.com/silverriver) for adding this dataset.
lccc
[ "task_categories:conversational", "task_ids:dialogue-generation", "annotations_creators:other", "language_creators:other", "multilinguality:monolingual", "size_categories:10M<n<100M", "source_datasets:original", "language:zh", "license:mit", "arxiv:2008.03946", "region:us" ]
2022-06-14T17:05:32+00:00
{"annotations_creators": ["other"], "language_creators": ["other"], "language": ["zh"], "license": ["mit"], "multilinguality": ["monolingual"], "size_categories": ["10M<n<100M"], "source_datasets": ["original"], "task_categories": ["conversational"], "task_ids": ["dialogue-generation"], "paperswithcode_id": "lccc", "pretty_name": "LCCC: Large-scale Cleaned Chinese Conversation corpus", "dataset_info": [{"config_name": "large", "features": [{"name": "dialog", "list": "string"}], "splits": [{"name": "train", "num_bytes": 1530827965, "num_examples": 12007759}], "download_size": 607605643, "dataset_size": 1530827965}, {"config_name": "base", "features": [{"name": "dialog", "list": "string"}], "splits": [{"name": "train", "num_bytes": 932634902, "num_examples": 6820506}, {"name": "test", "num_bytes": 1498216, "num_examples": 10000}, {"name": "validation", "num_bytes": 2922731, "num_examples": 20000}], "download_size": 371475095, "dataset_size": 937055849}]}
2024-01-18T11:19:16+00:00
f379fd9af10f2178159a6c7fdf6d3a8f10dccc74
# Dataset Card for "dane-mini" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
ScandEval/dane-mini
[ "task_categories:token-classification", "size_categories:1K<n<10K", "language:da", "license:cc-by-sa-4.0", "region:us" ]
2022-06-14T17:20:34+00:00
{"language": ["da"], "license": "cc-by-sa-4.0", "size_categories": ["1K<n<10K"], "task_categories": ["token-classification"], "dataset_info": {"features": [{"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "labels", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 355712, "num_examples": 1024}, {"name": "test", "num_bytes": 747809, "num_examples": 2048}, {"name": "val", "num_bytes": 92001, "num_examples": 256}], "download_size": 532720, "dataset_size": 1195522}}
2023-07-05T08:40:02+00:00
67710dfa772469bf5d3653bf0ca9f431e782d85c
# Dataset Card for "norne-nb-mini" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
ScandEval/norne-nb-mini
[ "task_categories:token-classification", "size_categories:1K<n<10K", "language:nb", "license:other", "region:us" ]
2022-06-14T17:21:00+00:00
{"language": ["nb"], "license": "other", "size_categories": ["1K<n<10K"], "task_categories": ["token-classification"], "dataset_info": {"features": [{"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "labels", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 317673, "num_examples": 1024}, {"name": "test", "num_bytes": 626004, "num_examples": 2048}, {"name": "val", "num_bytes": 87124, "num_examples": 256}], "download_size": 455512, "dataset_size": 1030801}}
2023-07-05T08:42:22+00:00
b34982a3ffc391bf3c5d6c999a2c1804ae170780
# Dataset Card for "norne-nn-mini" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
ScandEval/norne-nn-mini
[ "task_categories:token-classification", "size_categories:1K<n<10K", "language:nn", "license:other", "region:us" ]
2022-06-14T17:21:22+00:00
{"language": ["nn"], "license": "other", "size_categories": ["1K<n<10K"], "task_categories": ["token-classification"], "dataset_info": {"features": [{"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "labels", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 341534, "num_examples": 1024}, {"name": "test", "num_bytes": 721476, "num_examples": 2048}, {"name": "val", "num_bytes": 90956, "num_examples": 256}], "download_size": 502871, "dataset_size": 1153966}}
2023-07-05T08:41:26+00:00
d8243105818293a2f5b4ba7ab6406e73098c9b8b
# Dataset Card for "suc3-mini" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
ScandEval/suc3-mini
[ "task_categories:token-classification", "size_categories:1K<n<10K", "language:sv", "license:cc-by-4.0", "region:us" ]
2022-06-14T17:21:45+00:00
{"language": ["sv"], "license": "cc-by-4.0", "size_categories": ["1K<n<10K"], "task_categories": ["token-classification"], "dataset_info": {"features": [{"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "labels", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 344855, "num_examples": 1024}, {"name": "test", "num_bytes": 681936, "num_examples": 2048}, {"name": "val", "num_bytes": 81547, "num_examples": 256}], "download_size": 509020, "dataset_size": 1108338}}
2023-07-05T08:42:05+00:00
457785e82655186a01a89026cb35611fb2d71663
ScandEval/angry-tweets-mini
[ "task_categories:text-classification", "size_categories:1K<n<10K", "language:da", "license:cc-by-4.0", "region:us" ]
2022-06-14T17:30:11+00:00
{"language": ["da"], "license": "cc-by-4.0", "size_categories": ["1K<n<10K"], "task_categories": ["text-classification"]}
2023-07-05T08:52:07+00:00
07b99ab3363c2e7f8f87015b01c21f4d9b917ce3
ScandEval/norec-mini
[ "task_categories:text-classification", "size_categories:1K<n<10K", "language:no", "language:nb", "language:nn", "license:cc-by-nc-4.0", "region:us" ]
2022-06-14T17:30:30+00:00
{"language": ["no", "nb", "nn"], "license": "cc-by-nc-4.0", "size_categories": ["1K<n<10K"], "task_categories": ["text-classification"]}
2023-07-05T08:51:39+00:00
1273f672d0ac2b5e42a188a170de0513e4e5b5c1
ScandEval/absabank-imm-mini
[ "task_categories:text-classification", "size_categories:1K<n<10K", "language:sv", "license:cc-by-4.0", "region:us" ]
2022-06-14T17:30:48+00:00
{"language": ["sv"], "license": "cc-by-4.0", "size_categories": ["1K<n<10K"], "task_categories": ["text-classification"]}
2023-07-05T08:50:27+00:00
10d04e5710d83c882ae62acb2dd928958b49516a
# Dataset Card for "mim-gold-ner-mini" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
ScandEval/mim-gold-ner-mini
[ "task_categories:token-classification", "size_categories:1K<n<10K", "language:is", "license:other", "region:us" ]
2022-06-14T17:35:40+00:00
{"language": ["is"], "license": "other", "size_categories": ["1K<n<10K"], "task_categories": ["token-classification"], "dataset_info": {"features": [{"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "labels", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 377525, "num_examples": 1024}, {"name": "test", "num_bytes": 746049, "num_examples": 2048}, {"name": "val", "num_bytes": 93607, "num_examples": 256}], "download_size": 562012, "dataset_size": 1217181}}
2023-07-05T08:43:10+00:00
564cc836587aa94894fffba0ed28857eb5b6f939
# Dataset Card for "wikiann-fo-mini" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
ScandEval/wikiann-fo-mini
[ "language:fo", "region:us" ]
2022-06-14T17:39:30+00:00
{"language": ["fo"], "dataset_info": {"features": [{"name": "text", "dtype": "string"}, {"name": "tokens", "sequence": "string"}, {"name": "labels", "sequence": "string"}], "splits": [{"name": "train", "num_bytes": 233626, "num_examples": 1024}, {"name": "test", "num_bytes": 467705, "num_examples": 2048}, {"name": "val", "num_bytes": 60873, "num_examples": 256}], "download_size": 338188, "dataset_size": 762204}}
2023-07-05T07:08:24+00:00
54319f7cf5d56172bcae5acaa90b046ec6fe4ae6
rr/dd
[ "license:afl-3.0", "region:us" ]
2022-06-14T18:33:29+00:00
{"license": "afl-3.0"}
2022-06-23T03:09:14+00:00
7bdf563492accd06815580ffdd685adad8b8674b
# Dataset Card for ADE 20K Tiny This is a tiny subset of the ADE 20K dataset, which you can find [here](https://huggingface.co/datasets/scene_parse_150).
nateraw/ade20k-tiny
[ "task_categories:image-segmentation", "task_ids:semantic-segmentation", "annotations_creators:crowdsourced", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:n<1K", "source_datasets:extended|ade20k", "language:en", "license:bsd-3-clause", "region:us" ]
2022-06-15T03:32:58+00:00
{"annotations_creators": ["crowdsourced", "expert-generated"], "language_creators": ["found"], "language": ["en"], "license": ["bsd-3-clause"], "multilinguality": ["monolingual"], "size_categories": ["n<1K"], "source_datasets": ["extended|ade20k"], "task_categories": ["image-segmentation"], "task_ids": ["semantic-segmentation"], "pretty_name": "ADE 20K Tiny"}
2022-07-08T05:58:09+00:00
22b3b59656bf17b64ef0294318274afc7b5cf6a2
# Dataset Card for Country211 The [Country 211 Dataset](https://github.com/openai/CLIP/blob/main/data/country211.md) from OpenAI. This dataset was built by filtering the images from the YFCC100m dataset that have GPS coordinate corresponding to a ISO-3166 country code. The dataset is balanced by sampling 150 train images, 50 validation images, and 100 test images images for each country.
nateraw/country211
[ "task_categories:image-classification", "task_ids:multi-class-image-classification", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:extended|yfcc100m", "language:en", "license:unknown", "region:us" ]
2022-06-15T04:11:59+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["10K<n<100K"], "source_datasets": ["extended|yfcc100m"], "task_categories": ["image-classification"], "task_ids": ["multi-class-image-classification"], "pretty_name": "Country 211"}
2022-07-25T19:27:00+00:00
813d20cfb22b7ac76cb6a272cc8510bd85e8a66e
# Rendered SST-2 The [Rendered SST-2 Dataset](https://github.com/openai/CLIP/blob/main/data/rendered-sst2.md) from Open AI. Rendered SST2 is an image classification dataset used to evaluate the models capability on optical character recognition. This dataset was generated by rendering sentences in the Standford Sentiment Treebank v2 dataset. This dataset contains two classes (positive and negative) and is divided in three splits: a train split containing 6920 images (3610 positive and 3310 negative), a validation split containing 872 images (444 positive and 428 negative), and a test split containing 1821 images (909 positive and 912 negative).
nateraw/rendered-sst2
[ "task_categories:image-classification", "task_ids:multi-class-image-classification", "annotations_creators:machine-generated", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:extended|sst2", "language:en", "license:unknown", "region:us" ]
2022-06-15T04:32:09+00:00
{"annotations_creators": ["machine-generated"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "source_datasets": ["extended|sst2"], "task_categories": ["image-classification"], "task_ids": ["multi-class-image-classification"], "pretty_name": "Rendered SST-2"}
2022-10-25T09:32:21+00:00
5d1705be26da650adea619ee9bc5bf45571bb653
# Dataset Card for Kitti The [Kitti](http://www.cvlibs.net/datasets/kitti/eval_object.php) dataset. The Kitti object detection and object orientation estimation benchmark consists of 7481 training images and 7518 test images, comprising a total of 80.256 labeled objects
nateraw/kitti
[ "task_categories:object-detection", "annotations_creators:found", "language_creators:crowdsourced", "multilinguality:monolingual", "size_categories:1K<n<10K", "language:en", "license:unknown", "region:us" ]
2022-06-15T04:58:44+00:00
{"annotations_creators": ["found"], "language_creators": ["crowdsourced"], "language": ["en"], "license": ["unknown"], "multilinguality": ["monolingual"], "size_categories": ["1K<n<10K"], "task_categories": ["object-detection"], "task_ids": ["object-detection"], "pretty_name": "Kitti"}
2022-07-15T17:17:21+00:00
ccdfad1bd9c9f5558f073cd7d40e59901a527d85
EulerianKnight/SeaSpongeDetection
[ "license:apache-2.0", "region:us" ]
2022-06-15T05:57:16+00:00
{"license": "apache-2.0"}
2022-06-15T06:00:17+00:00
023111f79120d8f955d3691c5ed612b8f0a92d00
rr/DDR
[ "license:pddl", "region:us" ]
2022-06-15T06:57:47+00:00
{"license": "pddl"}
2022-06-15T06:59:20+00:00
603ca7858c8c00d7b762ff96d3aa29f1507c6954
# Dataset Card for Sketch Data Model Dataset ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** https://github.com/sketchai - **Repository:** https://github.com/sketchai/preprocessing - **Paper:** [Needs More Information] - **Leaderboard:** [Needs More Information] - **Point of Contact:** [Needs More Information] ### Dataset Summary This dataset contains over 6M CAD 2D sketches extracted from Onshape. Sketches are stored as python objects in the custom SAM format. SAM leverages the [Sketchgraphs](https://github.com/PrincetonLIPS/SketchGraphs) dataset for industrial needs and allows for easier transfer learning on other CAD softwares. ### Supported Tasks and Leaderboards Tasks: Automatic Sketch Generation, Auto Constraint ## Dataset Structure ### Data Instances The presented npy files contain python pickled objects and require the [flat_array](https://github.com/PrincetonLIPS/SketchGraphs/blob/master/sketchgraphs/data/flat_array.py) module of Sketchgraphs to be loaded. The normalization_output_merged.npy file contains sketch sequences represented as a list of SAM Primitives and Constraints. The sg_merged_final_*.npy files contain encoded constraint graphs of the sketches represented as a dictionnary of arrays. ### Data Fields [Needs More Information] ### Data Splits |Train |Val |Test | |------|------|------| |6M |50k | 50k | ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators [Needs More Information] ### Licensing Information [Needs More Information] ### Citation Information [Needs More Information]
sketchai/sam-dataset
[ "annotations_creators:no-annotation", "language_creators:other", "size_categories:1M<n<10M", "license:lgpl-3.0", "region:us" ]
2022-06-15T08:18:23+00:00
{"annotations_creators": ["no-annotation"], "language_creators": ["other"], "language": [], "license": ["lgpl-3.0"], "multilinguality": [], "size_categories": ["1M<n<10M"], "task_categories": [], "task_ids": [], "pretty_name": "Sketch Data Model Dataset"}
2022-07-13T12:03:40+00:00
b7f5ca3b82fd40f1b5eaae91c720817eb477a2cd
# Dataset Card for frwiki_good_pages_el ## Dataset Description - Repository: [frwiki_el](https://github.com/GaaH/frwiki_el) - Point of Contact: [Gaëtan Caillaut](mailto://[email protected]) ### Dataset Summary This dataset contains articles from the French Wikipédia. It is intended to be used to train Entity Linking (EL) systems. Links in articles are used to detect named entities. The dataset `frwiki` contains sentences of each Wikipedia pages. The dataset `entities` contains description for each Wikipedia pages. ### Languages - French ## Dataset Structure ### frwiki ``` { "name": "Title of the page", "wikidata_id": "Identifier of the related Wikidata entity. Can be null.", "wikipedia_id": "Identifier of the Wikipedia page", "wikipedia_url": "URL to the Wikipedia page", "wikidata_url": "URL to the Wikidata page. Can be null.", "sentences" : [ { "text": "text of the current sentence", "ner": ["list", "of", "ner", "labels"], "mention_mappings": [ (start_of_first_mention, end_of_first_mention), (start_of_second_mention, end_of_second_mention) ], "el_wikidata_id": ["wikidata id of first mention", "wikidata id of second mention"], "el_wikipedia_id": [wikipedia id of first mention, wikipedia id of second mention], "el_wikipedia_title": ["wikipedia title of first mention", "wikipedia title of second mention"] } ] "words": ["words", "in", "the", "sentence"], "ner": ["ner", "labels", "of", "each", "words"], "el": ["el", "labels", "of", "each", "words"] } ``` ### entities ``` { "name": "Title of the page", "wikidata_id": "Identifier of the related Wikidata entity. Can be null.", "wikipedia_id": "Identifier of the Wikipedia page", "wikipedia_url": "URL to the Wikipedia page", "wikidata_url": "URL to the Wikidata page. Can be null.", "description": "Description of the entity" } ```
gcaillaut/frwiki_el
[ "task_categories:token-classification", "annotations_creators:crowdsourced", "language_creators:machine-generated", "multilinguality:monolingual", "size_categories:1M<n<10M", "source_datasets:original", "language:fr", "license:wtfpl", "region:us" ]
2022-06-15T08:37:40+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["machine-generated"], "language": ["fr"], "license": ["wtfpl"], "multilinguality": ["monolingual"], "size_categories": ["1M<n<10M"], "source_datasets": ["original"], "task_categories": ["token-classification"], "task_ids": [], "pretty_name": "French Wikipedia dataset for Entity Linking"}
2022-09-28T07:52:12+00:00
ae6988dd116193043f09ac8afb7e768e23cb53fd
ksbk/gg
[ "license:afl-3.0", "region:us" ]
2022-06-15T08:37:41+00:00
{"license": "afl-3.0"}
2022-06-15T08:39:15+00:00
5d39a097127b8a6c8342cfc602967ce396478678
# multiFC - a dataset for the task of **automatic claim verification** - License is currently unknown, please refer to the original paper/[dataset site](http://www.copenlu.com/publication/2019_emnlp_augenstein/): - https://arxiv.org/abs/1909.03242 ## Dataset contents - **IMPORTANT:** the `label` column in the `test` set has dummy values as these were not provided (see original readme section for explanation) ``` DatasetDict({ train: Dataset({ features: ['claimID', 'claim', 'label', 'claimURL', 'reason', 'categories', 'speaker', 'checker', 'tags', 'article title', 'publish date', 'climate', 'entities'], num_rows: 27871 }) test: Dataset({ features: ['claimID', 'claim', 'label', 'claimURL', 'reason', 'categories', 'speaker', 'checker', 'tags', 'article title', 'publish date', 'climate', 'entities'], num_rows: 3487 }) validation: Dataset({ features: ['claimID', 'claim', 'label', 'claimURL', 'reason', 'categories', 'speaker', 'checker', 'tags', 'article title', 'publish date', 'climate', 'entities'], num_rows: 3484 }) }) ``` ## Paper Abstract / Citation > We contribute the largest publicly available dataset of naturally occurring factual claims for the purpose of automatic claim verification. It is collected from 26 fact checking websites in English, paired with textual sources and rich metadata, and labelled for veracity by human expert journalists. We present an in-depth analysis of the dataset, highlighting characteristics and challenges. Further, we present results for automatic veracity prediction, both with established baselines and with a novel method for joint ranking of evidence pages and predicting veracity that outperforms all baselines. Significant performance increases are achieved by encoding evidence, and by modelling metadata. Our best-performing model achieves a Macro F1 of 49.2%, showing that this is a challenging testbed for claim veracity prediction. ``` @inproceedings{conf/emnlp2019/Augenstein, added-at = {2019-10-27T00:00:00.000+0200}, author = {Augenstein, Isabelle and Lioma, Christina and Wang, Dongsheng and Chaves Lima, Lucas and Hansen, Casper and Hansen, Christian and Grue Simonsen, Jakob}, booktitle = {EMNLP}, crossref = {conf/emnlp/2019}, publisher = {Association for Computational Linguistics}, title = {MultiFC: A Real-World Multi-Domain Dataset for Evidence-Based Fact Checking of Claims}, year = 2019 } ``` ## Original README Real-World Multi-Domain Dataset for Evidence-Based Fact Checking of Claims The MultiFC is the largest publicly available dataset of naturally occurring factual claims for automatic claim verification. It is collected from 26 English fact-checking websites paired with textual sources and rich metadata and labeled for veracity by human expert journalists. ###### TRAIN and DEV ####### The train and dev files are (tab-separated) and contain the following metadata: claimID, claim, label, claimURL, reason, categories, speaker, checker, tags, article title, publish date, climate, entities Fields that could not be crawled were set as "None." Please refer to Table 11 of our paper to see the summary statistics. ###### TEST ####### The test file follows the same structure. However, we have removed the label. Thus, it only presents 12 metadata. claimID, claim, claim, reason, categories, speaker, checker, tags, article title, publish date, climate, entities Fields that could not be crawled were set as "None." Please refer to Table 11 of our paper to see the summary statistics. ###### Snippets ###### The text of each claim is submitted verbatim as a query to the Google Search API (without quotes). In the folder snippet, we provide the top 10 snippets retrieved. In some cases, fewer snippets are provided since we have excluded the claimURL from the snippets. Each file in the snippets folder is named after the claimID of the claim submitted as a query. Snippets file is (tab-separated) and contains the following metadata: rank_position, title, snippet, snippet_url For more information, please refer to our paper: References: Isabelle Augenstein, Christina Lioma, Dongsheng Wang, Lucas Chaves Lima, Casper Hansen, Christian Hansen, and Jakob Grue Simonsen. 2019. MultiFC: A Real-World Multi-Domain Dataset for Evidence-Based Fact Checking of Claims. In EMNLP. Association for Computational Linguistics. https://copenlu.github.io/publication/2019_emnlp_augenstein/
pszemraj/multi_fc
[ "license:other", "automatic claim verification", "claims", "arxiv:1909.03242", "region:us" ]
2022-06-15T10:27:47+00:00
{"license": "other", "tags": ["automatic claim verification", "claims"]}
2022-06-16T10:57:52+00:00
7b79121ac9e503a2a396851c1e670387ec6ed10a
themindorchestra/SoundHealing
[ "license:cc-by-nc-sa-4.0", "region:us" ]
2022-06-15T12:01:21+00:00
{"license": "cc-by-nc-sa-4.0"}
2022-06-15T12:01:21+00:00
21e74ddf8de1a21436da12e3e653065c5213e9d1
# APPS Dataset ## Dataset Description [APPS](https://arxiv.org/abs/2105.09938) is a benchmark for code generation with 10000 problems. It can be used to evaluate the ability of language models to generate code from natural language specifications. You can also find **APPS metric** in the hub here [codeparrot/apps_metric](https://huggingface.co/spaces/codeparrot/apps_metric). ## Languages The dataset contains questions in English and code solutions in Python. ## Dataset Structure ```python from datasets import load_dataset load_dataset("codeparrot/apps") DatasetDict({ train: Dataset({ features: ['problem_id', 'question', 'solutions', 'input_output', 'difficulty', 'url', 'starter_code'], num_rows: 5000 }) test: Dataset({ features: ['problem_id', 'question', 'solutions', 'input_output', 'difficulty', 'url', 'starter_code'], num_rows: 5000 }) }) ``` ### How to use it You can load and iterate through the dataset with the following two lines of code for the train split: ```python from datasets import load_dataset import json ds = load_dataset("codeparrot/apps", split="train") sample = next(iter(ds)) # non-empty solutions and input_output features can be parsed from text format this way: sample["solutions"] = json.loads(sample["solutions"]) sample["input_output"] = json.loads(sample["input_output"]) print(sample) #OUTPUT: { 'problem_id': 0, 'question': 'Polycarp has $n$ different binary words. A word called binary if it contains only characters \'0\' and \'1\'. For example...', 'solutions': ["for _ in range(int(input())):\n n = int(input())\n mass = []\n zo = 0\n oz = 0\n zz = 0\n oo = 0\n...",...], 'input_output': {'inputs': ['4\n4\n0001\n1000\n0011\n0111\n3\n010\n101\n0\n2\n00000\n00001\n4\n01\n001\n0001\n00001\n'], 'outputs': ['1\n3 \n-1\n0\n\n2\n1 2 \n']}, 'difficulty': 'interview', 'url': 'https://codeforces.com/problemset/problem/1259/D', 'starter_code': ''} } ``` Each sample consists of a programming problem formulation in English, some ground truth Python solutions, test cases that are defined by their inputs and outputs and function name if provided, as well as some metadata regarding the difficulty level of the problem and its source. If a sample has non empty `input_output` feature, you can read it as a dictionary with keys `inputs` and `outputs` and `fn_name` if it exists, and similarily you can parse the solutions into a list of solutions as shown in the code above. You can also filter the dataset for the difficulty level: Introductory, Interview and Competition. Just pass the list of difficulties as a list. E.g. if you want the most challenging problems, you need to select the competition level: ```python ds = load_dataset("codeparrot/apps", split="train", difficulties=["competition"]) print(next(iter(ds))["question"]) #OUTPUT: """\ Codefortia is a small island country located somewhere in the West Pacific. It consists of $n$ settlements connected by ... For each settlement $p = 1, 2, \dots, n$, can you tell what is the minimum time required to travel between the king's residence and the parliament house (located in settlement $p$) after some roads are abandoned? -----Input----- The first line of the input contains four integers $n$, $m$, $a$ and $b$ ... -----Output----- Output a single line containing $n$ integers ... -----Examples----- Input 5 5 20 25 1 2 25 ... Output 0 25 60 40 20 ... ``` ### Data Fields |Field|Type|Description| |---|---|---| |problem_id|int|problem id| |question|string|problem description| |solutions|string|some python solutions| |input_output|string|Json string with "inputs" and "outputs" of the test cases, might also include "fn_name" the name of the function| |difficulty|string|difficulty level of the problem| |url|string|url of the source of the problem| |starter_code|string|starter code to include in prompts| we mention that only few samples have `fn_name` and `starter_code` specified ### Data Splits The dataset contains a train and test splits with 5000 samples each. ### Dataset Statistics * 10000 coding problems * 131777 test cases * all problems have a least one test case except 195 samples in the train split * for tests split, the average number of test cases is 21.2 * average length of a problem is 293.2 words * all files have ground-truth solutions except 1235 samples in the test split ## Dataset Creation To create the APPS dataset, the authors manually curated problems from open-access sites where programmers share problems with each other, including Codewars, AtCoder, Kattis, and Codeforces. For more details please refer to the original [paper](https://arxiv.org/pdf/2105.09938.pdf). ## Considerations for Using the Data In [AlphaCode](https://arxiv.org/pdf/2203.07814v1.pdf) the authors found that this dataset can generate many false positives during evaluation, where incorrect submissions are marked as correct due to lack of test coverage. ## Citation Information ``` @article{hendrycksapps2021, title={Measuring Coding Challenge Competence With APPS}, author={Dan Hendrycks and Steven Basart and Saurav Kadavath and Mantas Mazeika and Akul Arora and Ethan Guo and Collin Burns and Samir Puranik and Horace He and Dawn Song and Jacob Steinhardt}, journal={NeurIPS}, year={2021} } ```
codeparrot/apps
[ "task_categories:text-generation", "task_ids:language-modeling", "language_creators:crowdsourced", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:unknown", "language:code", "license:mit", "arxiv:2105.09938", "arxiv:2203.07814", "region:us" ]
2022-06-15T12:20:26+00:00
{"annotations_creators": [], "language_creators": ["crowdsourced", "expert-generated"], "language": ["code"], "license": ["mit"], "multilinguality": ["monolingual"], "size_categories": ["unknown"], "source_datasets": [], "task_categories": ["text-generation"], "task_ids": ["language-modeling"], "pretty_name": "APPS"}
2022-10-20T14:00:15+00:00
83e07480c44954d638b087ecd1f6af7934ba9d68
Trust**wallet customer service Support Number +1-818*751*8351
trustwallet/33
[ "license:apache-2.0", "region:us" ]
2022-06-15T16:31:08+00:00
{"license": "apache-2.0"}
2022-06-15T16:31:41+00:00
1413ed90879c1ebb1b7016388a6ef43e7765d295
# AutoTrain Dataset for project: Psychiatry_Article_Identifier ## Dataset Descritpion This dataset has been automatically processed by AutoTrain for project Psychiatry_Article_Identifier. ### Languages The BCP-47 code for the dataset's language is unk. ## Dataset Structure ### Data Instances A sample from this dataset looks as follows: ```json [ { "text": "diffuse actinic keratinocyte dysplasia", "target": 15 }, { "text": "cholesterol atheroembolism", "target": 8 } ] ``` ### Dataset Fields The dataset has the following fields (also called "features"): ```json { "text": "Value(dtype='string', id=None)", "target": "ClassLabel(num_classes=20, names=['Certain infectious or parasitic diseases', 'Developmental anaomalies', 'Diseases of the blood or blood forming organs', 'Diseases of the genitourinary system', 'Mental behavioural or neurodevelopmental disorders', 'Neoplasms', 'certain conditions originating in the perinatal period', 'conditions related to sexual health', 'diseases of the circulatroy system', 'diseases of the digestive system', 'diseases of the ear or mastoid process', 'diseases of the immune system', 'diseases of the musculoskeletal system or connective tissue', 'diseases of the nervous system', 'diseases of the respiratory system', 'diseases of the skin', 'diseases of the visual system', 'endocrine nutritional or metabolic diseases', 'pregnanacy childbirth or the puerperium', 'sleep-wake disorders'], id=None)" } ``` ### Dataset Splits This dataset is split into a train and validation split. The split sizes are as follow: | Split name | Num samples | | ------------ | ------------------- | | train | 9828 | | valid | 2468 |
justpyschitry/autotrain-data-Psychiatry_Article_Identifier
[ "task_categories:text-classification", "region:us" ]
2022-06-15T18:02:36+00:00
{"task_categories": ["text-classification"]}
2022-06-15T20:34:39+00:00
ca5836dcea910a720fe456e3d3c9b68206507eeb
# Dataset Card for MAWPS_ar ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description ### Dataset Summary MAWPS: A Math Word Problem Repository ### Supported Tasks Math Word Problem Solving ### Languages Supports Arabic and English ## Dataset Structure ### Data Fields - `text_en`: a `string` feature. - `text_ar`: a `string` feature. - `eqn`: a `string` feature. ### Data Splits |train|validation|test| |----:|---------:|---:| | 3636| 1040| 520| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [Rik Koncel-Kedziorski**, Subhro Roy**, Aida Amini, Nate Kushman and Hannaneh Hajishirzi.](https://aclanthology.org/N16-1136.pdf) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Contributions Special thanks to Associate Professor Marwan Torki and all my colleagues in CC491N (NLP) class for helping me translate this dataset.
omarxadel/MaWPS-ar
[ "task_categories:text2text-generation", "task_ids:explanation-generation", "annotations_creators:crowdsourced", "language_creators:found", "multilinguality:multilingual", "size_categories:1K<n<10K", "language:en", "language:ar", "license:mit", "region:us" ]
2022-06-15T19:39:14+00:00
{"annotations_creators": ["crowdsourced"], "language_creators": ["found"], "language": ["en", "ar"], "license": ["mit"], "multilinguality": ["multilingual"], "size_categories": ["1K<n<10K"], "source_datasets": [], "task_categories": ["text2text-generation"], "task_ids": ["explanation-generation"], "pretty_name": "MAWPS_ar"}
2022-07-12T14:31:07+00:00
6bfc581b1b87cbd2ff08a7369f34a10fbed01c4b
sounakray1997/CoNLLU_WikiNEuRal
[ "license:apache-2.0", "region:us" ]
2022-06-15T22:16:23+00:00
{"license": "apache-2.0"}
2022-06-15T22:16:23+00:00
2f2f43e7e54c6dcebe550767c8a284c26186e030
nateraw/kinetics
[ "license:cc-by-4.0", "region:us" ]
2022-06-16T01:29:09+00:00
{"license": "cc-by-4.0"}
2022-06-16T01:30:12+00:00
3906031fe2e1a2d7766f0847014d3eb6229c6351
HALLA/Rhubarbify
[ "license:other", "region:us" ]
2022-06-16T02:19:36+00:00
{"license": "other"}
2022-06-16T02:25:29+00:00
6adb717a2503f1d49af178c0497c2529f1a8e68f
# CORD: A Consolidated Receipt Dataset for Post-OCR Parsing CORD dataset is cloned from [clovaai](https://github.com/clovaai/cord) GitHub repo - Box coordinates are normalized against image width/height - Labels with very few occurrences are replaced with O: ``` replacing_labels = ['menu.etc', 'menu.itemsubtotal', 'menu.sub_etc', 'menu.sub_unitprice', 'menu.vatyn', 'void_menu.nm', 'void_menu.price', 'sub_total.othersvc_price'] ``` Check for more info [Sparrow](https://github.com/katanaml/sparrow) ## Citation ### CORD: A Consolidated Receipt Dataset for Post-OCR Parsing ``` @article{park2019cord, title={CORD: A Consolidated Receipt Dataset for Post-OCR Parsing}, author={Park, Seunghyun and Shin, Seung and Lee, Bado and Lee, Junyeop and Surh, Jaeheung and Seo, Minjoon and Lee, Hwalsuk} booktitle={Document Intelligence Workshop at Neural Information Processing Systems} year={2019} } ``` ### Post-OCR parsing: building simple and robust parser via BIO tagging ``` @article{hwang2019post, title={Post-OCR parsing: building simple and robust parser via BIO tagging}, author={Hwang, Wonseok and Kim, Seonghyeon and Yim, Jinyeong and Seo, Minjoon and Park, Seunghyun and Park, Sungrae and Lee, Junyeop and Lee, Bado and Lee, Hwalsuk} booktitle={Document Intelligence Workshop at Neural Information Processing Systems} year={2019} } ```
nehruperumalla/forms
[ "region:us" ]
2022-06-16T05:34:48+00:00
{}
2022-06-16T05:38:45+00:00
f27efa2241b715868b4e2c6a2ead19ce067b3b48
This is the sentiment analysis dataset based on IMDB reviews initially released by Stanford University. ``` This is a dataset for binary sentiment classification containing substantially more data than previous benchmark datasets. We provide a set of 25,000 highly polar movie reviews for training, and 25,000 for testing. There is additional unlabeled data for use as well. Raw text and already processed bag of words formats are provided. See the README file contained in the release for more details. ``` [Here](http://ai.stanford.edu/~amaas/data/sentiment/) is the redirection. ``` @InProceedings{maas-EtAl:2011:ACL-HLT2011, author = {Maas, Andrew L. and Daly, Raymond E. and Pham, Peter T. and Huang, Dan and Ng, Andrew Y. and Potts, Christopher}, title = {Learning Word Vectors for Sentiment Analysis}, booktitle = {Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies}, month = {June}, year = {2011}, address = {Portland, Oregon, USA}, publisher = {Association for Computational Linguistics}, pages = {142--150}, url = {http://www.aclweb.org/anthology/P11-1015} } ```
scikit-learn/imdb
[ "license:other", "region:us" ]
2022-06-16T08:07:41+00:00
{"license": "other"}
2022-06-16T08:11:24+00:00
428766bc07a9e8699b7782e8557b7e07d32923f3
# CodeParrot 🦜 Dataset after near deduplication (validation) ## Dataset Description A dataset of Python files from Github. We performed near deduplication of this dataset split [codeparrot-clean-train](https://huggingface.co/datasets/codeparrot/codeparrot-clean-valid) from [codeparrot-clean](https://huggingface.co/datasets/codeparrot/codeparrot-clean#codeparrot-%F0%9F%A6%9C-dataset-cleaned). Exact deduplication can miss a fair amount of nearly identical files. We used MinHash with a Jaccard threshold (default=0.85) to create duplicate clusters. Then these clusters are reduced to unique files based on the exact Jaccard similarity. Fore more details, please refer to this [repo](https://github.com/huggingface/transformers/tree/main/examples/research_projects/codeparrot).
codeparrot/codeparrot-valid-near-deduplication
[ "region:us" ]
2022-06-16T13:29:26+00:00
{}
2022-06-21T18:06:58+00:00
e77668a8fb0cedb8ff0fb97ca9e4699b1f095841
# Dataset Card for OffendES ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Source Data](#source-data) - [Annotations](#annotations) - [Additional Information](#additional-information) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Paper: OffendES:** [A New Corpus in Spanish for Offensive Language Research](https://aclanthology.org/2021.ranlp-1.123.pdf) - **Leaderboard:** [Leaderboard for OffendES / Spanish](http://journal.sepln.org/sepln/ojs/ojs/index.php/pln/article/view/6388) - **Point of Contact: [email protected]** ### Dataset Summary Focusing on young influencers from the well-known social platforms of Twitter, Instagram, and YouTube, we have collected a corpus composed of Spanish comments manually labeled on offensive pre-defined categories. From the total corpus, we selected 30,416 posts to be publicly published, they correspond to the ones used in the MeOffendES competition at IberLEF 2021. The posts are labeled with the following categories: - Offensive, the target is a person (OFP). Offensive text targeting a specific individual. - Offensive, the target is a group of people or collective (OFG). Offensive text targeting a group of people belonging to the same ethnic group, gender or sexual orientation, political ideology, religious belief, or other common characteristics. - Non-offensive, but with expletive language (NOE). A text that contains rude words, blasphemes, or swearwords but without the aim of offending, and usually with a positive connotation. - Non-offensive (NO). Text that is neither offensive nor contains expletive language ### Supported Tasks and Leaderboards This dataset is intended for multi-class offensive classification and binary offensive classification. Competition [MeOffendES task on offensive detection for Spanish at IberLEF 2021](http://journal.sepln.org/sepln/ojs/ojs/index.php/pln/article/view/6388) ### Languages - Spanish ## Dataset Structure ### Data Instances For each instance, there is a string for the id of the tweet, a string for the emotion class, a string for the offensive class, and a string for the event. See the []() to explore more examples. ``` {'comment_id': '8003', 'influencer': 'dalas', 'comment': 'Estupido aburrido', 'label': 'NO', 'influencer_gender': 'man', 'media': youtube } ``` ### Data Fields - `comment_id`: a string to identify the comment - `influencer`: a string containing the influencer associated with the comment - `comment`: a string containing the text of the comment - `label`: a string containing the offensive gold label - `influencer_gender`: a string containing the genre of the influencer - `media`: a string containing the social media platform where the comment has been retrieved ### Data Splits The OffendES dataset contains 3 splits: _train_, _validation_, and _test_. Below are the statistics for each class. | OffendES | Number of Instances in Split per class| | | | ------------- | ---------------------------------|---------------------------------|------------------------------------------| | `Class` | `Train` | `Validation` | `Test` | | NO | 13,212 | 64 | 9,651 | | NOE | 1,235 | 22 | 2,340 | | OFP | 2,051 | 10 | 1,404 | | OFG | 212 | 4 | 211 | | Total | 16,710 | 100 | 13,606 | ## Dataset Creation ### Source Data Twitter, Youtube, Instagram #### Who are the annotators? Amazon Mechanical Turkers ## Additional Information ### Licensing Information The OffendES dataset is released under the [Apache-2.0 License](http://www.apache.org/licenses/LICENSE-2.0). ### Citation Information ``` @inproceedings{plaza-del-arco-etal-2021-offendes, title = "{O}ffend{ES}: A New Corpus in {S}panish for Offensive Language Research", author = "{Plaza-del-Arco}, Flor Miriam and Montejo-R{\'a}ez, Arturo and Ure{\~n}a-L{\'o}pez, L. Alfonso and Mart{\'\i}n-Valdivia, Mar{\'\i}a-Teresa", booktitle = "Proceedings of the 12th Language Resources and Evaluation Conference", month = sep, year = "2021", address = "Held Online", url = "https://aclanthology.org/2021.ranlp-1.123.pdf", language = "English", pages = "1096--1108" } ``` ``` @article{meoffendes2021, title="{{Overview of MeOffendEs at IberLEF 2021: Offensive Language Detection in Spanish Variants}}", author="{Flor Miriam Plaza-del-Arco and Casavantes, Marco and Jair Escalante, Hugo and Martín-Valdivia, M. Teresa and Montejo-Ráez, Arturo and {Montes-y-Gómez}, Manuel and Jarquín-Vásquez, Horacio and Villaseñor-Pineda, Luis}", journal="Procesamiento del Lenguaje Natural", url = "https://bit.ly/3QpRDfy", volume="67", pages="183--194", year="2021" } ```
fmplaza/offendes
[ "language:es", "license:apache-2.0", "region:us" ]
2022-06-16T13:32:03+00:00
{"language": ["es"], "license": "apache-2.0"}
2024-02-06T14:26:55+00:00