id
stringlengths
2
115
author
stringlengths
2
42
last_modified
timestamp[us, tz=UTC]
downloads
int64
0
8.87M
likes
int64
0
3.84k
paperswithcode_id
stringlengths
2
45
tags
list
lastModified
timestamp[us, tz=UTC]
createdAt
stringlengths
24
24
key
stringclasses
1 value
created
timestamp[us]
card
stringlengths
1
1.01M
embedding
list
library_name
stringclasses
21 values
pipeline_tag
stringclasses
27 values
mask_token
null
card_data
null
widget_data
null
model_index
null
config
null
transformers_info
null
spaces
null
safetensors
null
transformersInfo
null
modelId
stringlengths
5
111
embeddings
list
charsiu/libriphrase_meta
charsiu
2023-10-27T18:33:50Z
40
0
null
[ "region:us" ]
2023-10-27T18:33:50Z
2023-10-27T18:30:29.000Z
2023-10-27T18:30:29
--- configs: - config_name: default data_files: - split: train path: data/train-* dataset_info: features: - name: anchor dtype: string - name: anchor_spk dtype: int64 - name: anchor_text dtype: string - name: anchor_dur dtype: float64 - name: comparison dtype: string - name: comparison_spk dtype: int64 - name: comparison_text dtype: string - name: comparison_dur dtype: float64 - name: type dtype: string - name: target dtype: int64 - name: class dtype: int64 - name: anchor_phone dtype: string - name: comparison_phone dtype: string splits: - name: train num_bytes: 53970720 num_examples: 203013 download_size: 8382220 dataset_size: 53970720 --- # Dataset Card for "libriphrase_meta" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.35992351174354553, -0.2675364017486572, 0.4148982763290405, 0.031053833663463593, -0.25862935185432434, -0.03867526352405548, 0.2413831502199173, -0.019139137119054794, 0.8892353773117065, 0.5554897785186768, -0.763922929763794, -0.834837794303894, -0.5488250851631165, -0.04956544563174...
null
null
null
null
null
null
null
null
null
null
null
null
null
Ka4on/mri
Ka4on
2023-10-28T01:03:33Z
40
0
null
[ "region:us" ]
2023-10-28T01:03:33Z
2023-10-28T01:02:50.000Z
2023-10-28T01:02:50
Entry not found
[ -0.32276472449302673, -0.22568407654762268, 0.8622258901596069, 0.4346148371696472, -0.5282984972000122, 0.7012965679168701, 0.7915717363357544, 0.07618629932403564, 0.7746022939682007, 0.2563222646713257, -0.785281777381897, -0.22573848068714142, -0.9104482531547546, 0.5715669393539429, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
automated-research-group/gpt2-winogrande
automated-research-group
2023-10-28T06:48:10Z
40
0
null
[ "region:us" ]
2023-10-28T06:48:10Z
2023-10-28T06:48:09.000Z
2023-10-28T06:48:09
--- dataset_info: features: - name: answer dtype: string - name: id dtype: string - name: question dtype: string - name: input_perplexity dtype: float64 - name: input_likelihood dtype: float64 - name: output_perplexity dtype: float64 - name: output_likelihood dtype: float64 splits: - name: validation num_bytes: 357232 num_examples: 1267 download_size: 162550 dataset_size: 357232 configs: - config_name: default data_files: - split: validation path: data/validation-* --- # Dataset Card for "gpt2-winogrande" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.41015559434890747, -0.17709650099277496, 0.29955539107322693, 0.138153538107872, -0.3233467638492584, -0.05001383274793625, 0.15053588151931763, -0.2849363684654236, 0.513012707233429, 0.20205198228359222, -0.6006694436073303, -0.585280179977417, -0.817493736743927, -0.5743923783302307,...
null
null
null
null
null
null
null
null
null
null
null
null
null
ayoub999/dataset_for_orange_factures
ayoub999
2023-11-09T13:59:48Z
40
0
null
[ "region:us" ]
2023-11-09T13:59:48Z
2023-10-30T15:15:47.000Z
2023-10-30T15:15:47
--- configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* dataset_info: features: - name: id dtype: string - name: image dtype: image - name: bboxes sequence: sequence: int64 - name: ner_tags sequence: class_label: names: '0': O '1': Ref '2': NumFa '3': Fourniss '4': DateFa '5': DateLim '6': TotalHT '7': TVA '8': TotalTTc '9': unitP '10': Qt '11': TVAP '12': Désignation '13': Adresse - name: tokens sequence: string splits: - name: train num_bytes: 2942860.8 num_examples: 12 - name: test num_bytes: 735715.2 num_examples: 3 download_size: 2799104 dataset_size: 3678576.0 --- # Dataset Card for "dataset_for_orange_factures" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.5826767683029175, -0.1584644615650177, 0.2013721466064453, 0.4803481996059418, -0.1977083683013916, -0.2482631802558899, 0.25698891282081604, -0.2779044806957245, 0.938926100730896, 0.630354642868042, -0.562171220779419, -0.8030178546905518, -0.45039695501327515, -0.09437257796525955, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
yuvalkirstain/task_prediction_train2
yuvalkirstain
2023-10-31T18:48:49Z
40
0
null
[ "region:us" ]
2023-10-31T18:48:49Z
2023-10-31T18:48:28.000Z
2023-10-31T18:48:28
--- configs: - config_name: default data_files: - split: train path: data/train-* - split: validation path: data/validation-* dataset_info: features: - name: path dtype: string - name: text dtype: string - name: task_name dtype: string splits: - name: train num_bytes: 659890949 num_examples: 5663600 - name: validation num_bytes: 7823929 num_examples: 60002 download_size: 148156628 dataset_size: 667714878 --- # Dataset Card for "task_prediction_train2" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.22719773650169373, -0.04342102259397507, 0.19181403517723083, 0.3092663586139679, -0.05620702728629112, -0.20392447710037231, 0.17041121423244476, -0.3135908842086792, 0.496336430311203, 0.35913708806037903, -0.8980210423469543, -0.45277833938598633, -0.8460240364074707, -0.518759369850...
null
null
null
null
null
null
null
null
null
null
null
null
null
btt-mining-coalation/red_pajama_random_5000
btt-mining-coalation
2023-11-09T14:45:49Z
40
0
null
[ "region:us" ]
2023-11-09T14:45:49Z
2023-11-05T01:42:27.000Z
2023-11-05T01:42:27
--- dataset_info: features: - name: text dtype: string - name: summary dtype: string - name: reward_dpo dtype: float64 splits: - name: train num_bytes: 304835574 num_examples: 5000 download_size: 137390511 dataset_size: 304835574 configs: - config_name: default data_files: - split: train path: data/train-* --- # Dataset Card for "red_pajama_random_5000" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.6282620429992676, -0.21327221393585205, 0.14692799746990204, 0.6847658753395081, -0.36777690052986145, -0.33240270614624023, 0.1714601218700409, -0.24441415071487427, 1.0220221281051636, 0.6109746098518372, -0.9150915145874023, -0.6857317090034485, -0.6623834371566772, -0.01580510474741...
null
null
null
null
null
null
null
null
null
null
null
null
null
ostapeno/qa-openai_batched_long_icl0_clen512_maxD-1_maxC2000_08000_cleaned_train
ostapeno
2023-11-06T14:50:24Z
40
0
null
[ "region:us" ]
2023-11-06T14:50:24Z
2023-11-06T14:45:47.000Z
2023-11-06T14:45:47
--- configs: - config_name: default data_files: - split: abstract_algebra path: data/abstract_algebra-* - split: college_biology path: data/college_biology-* - split: formal_logic path: data/formal_logic-* - split: global_facts path: data/global_facts-* - split: high_school_government_and_politics path: data/high_school_government_and_politics-* - split: high_school_physics path: data/high_school_physics-* - split: machine_learning path: data/machine_learning-* - split: prehistory path: data/prehistory-* - split: security_studies path: data/security_studies-* - split: sociology path: data/sociology-* dataset_info: features: - name: id dtype: string - name: context dtype: string - name: docno dtype: string - name: subject dtype: string - name: icl_examples dtype: 'null' - name: author_instr dtype: string - name: instruction dtype: string - name: response dtype: string splits: - name: abstract_algebra num_bytes: 23879722.12837838 num_examples: 8000 - name: college_biology num_bytes: 25235647.10957722 num_examples: 8000 - name: formal_logic num_bytes: 24553218.679549113 num_examples: 8000 - name: global_facts num_bytes: 24477439.56043956 num_examples: 8000 - name: high_school_government_and_politics num_bytes: 25633829.363360763 num_examples: 8000 - name: high_school_physics num_bytes: 26191171.15244726 num_examples: 8000 - name: machine_learning num_bytes: 26083439.57671958 num_examples: 8000 - name: prehistory num_bytes: 25412395.782298435 num_examples: 8000 - name: security_studies num_bytes: 25639530.37483843 num_examples: 8000 - name: sociology num_bytes: 24402985.15227051 num_examples: 8000 download_size: 140394435 dataset_size: 251509378.8798793 --- # Dataset Card for "qa-openai_batched_long_icl0_clen512_maxD-1_maxC2000_08000_cleaned_train" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.586467444896698, 0.029071280732750893, 0.15121541917324066, 0.3275112211704254, -0.4351518452167511, -0.344635009765625, -0.021098937839269638, -0.0950879231095314, 0.5700687766075134, 0.6171809434890747, -0.8575785160064697, -0.5152838826179504, -0.33041974902153015, -0.055025629699230...
null
null
null
null
null
null
null
null
null
null
null
null
null
evanfrick/lichess
evanfrick
2023-11-08T08:36:24Z
40
0
null
[ "license:mit", "region:us" ]
2023-11-08T08:36:24Z
2023-11-08T05:48:31.000Z
2023-11-08T05:48:31
--- license: mit ---
[ -0.12853392958641052, -0.18616779148578644, 0.6529127955436707, 0.49436280131340027, -0.19319361448287964, 0.23607419431209564, 0.36072003841400146, 0.050563063472509384, 0.579365611076355, 0.7400140762329102, -0.6508104205131531, -0.23783954977989197, -0.7102249264717102, -0.0478260256350...
null
null
null
null
null
null
null
null
null
null
null
null
null
rjaiswal/tubogas-dataset
rjaiswal
2023-11-20T09:42:33Z
40
0
null
[ "region:us" ]
2023-11-20T09:42:33Z
2023-11-16T15:57:04.000Z
2023-11-16T15:57:04
--- dataset_info: features: - name: image dtype: image - name: text dtype: string splits: - name: train num_bytes: 4389802.0 num_examples: 42 download_size: 2186125 dataset_size: 4389802.0 configs: - config_name: default data_files: - split: train path: data/train-* --- # Dataset Card for "tubogas-dataset" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.7161089777946472, -0.19296574592590332, 0.11274369806051254, 0.2646186351776123, -0.2646429240703583, -0.009472347795963287, 0.3281655013561249, 0.1251978576183319, 0.8259877562522888, 0.5093401670455933, -0.6648954749107361, -0.7734971046447754, -0.6830518245697021, -0.1351934671401977...
null
null
null
null
null
null
null
null
null
null
null
null
null
hodgesz/llama2-sql-create-context4
hodgesz
2023-11-18T01:13:21Z
40
0
null
[ "license:apache-2.0", "region:us" ]
2023-11-18T01:13:21Z
2023-11-18T01:12:52.000Z
2023-11-18T01:12:52
--- license: apache-2.0 ---
[ -0.12853392958641052, -0.18616779148578644, 0.6529127955436707, 0.49436280131340027, -0.19319361448287964, 0.23607419431209564, 0.36072003841400146, 0.050563063472509384, 0.579365611076355, 0.7400140762329102, -0.6508104205131531, -0.23783954977989197, -0.7102249264717102, -0.0478260256350...
null
null
null
null
null
null
null
null
null
null
null
null
null
ngoxuanphong/zalo
ngoxuanphong
2023-11-21T10:00:50Z
40
0
null
[ "size_categories:1K<n<10K", "language:en", "license:apache-2.0", "region:us" ]
2023-11-21T10:00:50Z
2023-11-20T04:46:01.000Z
2023-11-20T04:46:01
--- license: apache-2.0 language: - en size_categories: - 1K<n<10K ---
[ -0.12853392958641052, -0.18616779148578644, 0.6529127955436707, 0.49436280131340027, -0.19319361448287964, 0.23607419431209564, 0.36072003841400146, 0.050563063472509384, 0.579365611076355, 0.7400140762329102, -0.6508104205131531, -0.23783954977989197, -0.7102249264717102, -0.0478260256350...
null
null
null
null
null
null
null
null
null
null
null
null
null
SUSTech/mt_bench_judge
SUSTech
2023-11-23T10:18:55Z
40
0
null
[ "region:us" ]
2023-11-23T10:18:55Z
2023-11-23T10:03:09.000Z
2023-11-23T10:03:09
--- dataset_info: features: - name: question_id dtype: int64 - name: model dtype: string - name: conversation list: - name: content dtype: string - name: role dtype: string - name: turn dtype: int64 - name: judge sequence: string - name: user_prompt dtype: string - name: judgment dtype: string - name: score dtype: float64 - name: tstamp dtype: float64 - name: category dtype: string - name: reference sequence: string splits: - name: train num_bytes: 4409406 num_examples: 800 download_size: 949262 dataset_size: 4409406 configs: - config_name: default data_files: - split: train path: data/train-* --- # Dataset Card for "mt_bench_judge" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.5070008635520935, -0.13040265440940857, 0.5846325159072876, 0.24061515927314758, -0.3629312813282013, -0.3571460545063019, 0.254354327917099, 0.10948209464550018, 0.6192311644554138, 0.6882535815238953, -0.6617558002471924, -0.786194384098053, -0.786971390247345, -0.4348928928375244, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
jashmehta3300/social-injection-bold
jashmehta3300
2023-11-26T04:32:23Z
40
0
null
[ "region:us" ]
2023-11-26T04:32:23Z
2023-11-25T18:23:38.000Z
2023-11-25T18:23:38
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
lavis-nlp/german_legal_sentences
lavis-nlp
2022-10-20T18:34:19Z
39
3
null
[ "task_categories:text-retrieval", "task_ids:semantic-similarity-scoring", "annotations_creators:machine-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:n>1M", "source_datasets:original", "language:de", "license:unknown", "arxiv:2005.13342", "arxiv:2010.1025...
2022-10-20T18:34:19Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
--- annotations_creators: - machine-generated language_creators: - found language: - de license: - unknown multilinguality: - monolingual size_categories: - n>1M source_datasets: - original task_categories: - text-retrieval - text-scoring task_ids: - semantic-similarity-scoring - text-retrieval-other-example-based-retrieval --- # Dataset Card for German Legal Sentences ## Table of Contents - [Dataset Card for [Dataset Name]](#dataset-card-for-dataset-name) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Initial Data Collection and Normalization](#initial-data-collection-and-normalization) - [Who are the source language producers?](#who-are-the-source-language-producers) - [Annotations](#annotations) - [Annotation process](#annotation-process) - [Who are the annotators?](#who-are-the-annotators) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://lavis-nlp.github.io/german_legal_sentences/ - **Repository:** https://github.com/lavis-nlp/german_legal_sentences - **Paper:** coming soon - **Leaderboard:** - **Point of Contact:** [Marco Wrzalik](mailto:marco.wrzalik@hs-rm.de) ### Dataset Summary German Legal Sentences (GLS) is an automatically generated training dataset for semantic sentence matching and citation recommendation in the domain in german legal documents. It follows the concept of weak supervision, where imperfect labels are generated using multiple heuristics. For this purpose we use a combination of legal citation matching and BM25 similarity. The contained sentences and their citations are parsed from real judicial decisions provided by [Open Legal Data](http://openlegaldata.io/) (https://arxiv.org/abs/2005.13342). ### Supported Tasks and Leaderboards The main associated task is *Semantic Similarity Ranking*. We propose to use the *Mean Reciprocal Rank* (MRR) cut at the tenth position as well as MAP and Recall on Rankings of size 200. As baselines we provide the follows: | Method | MRR@10 | MAP@200 | Recall@200 | |-----------------------------------|---------:|-----------:|------------:| | BM25 - default `(k1=1.2; b=0.75)` | 25.7 | 17.6 | 42.9 | | BM25 - tuned `(k1=0.47; b=0.97)` | 26.2 | 18.1 | 43.3 | | [CoRT](https://arxiv.org/abs/2010.10252) | 31.2 | 21.4 | 56.2 | | [CoRT + BM25](https://arxiv.org/abs/2010.10252) | 32.1 | 22.1 | 67.1 | In addition, we want to support a *Citation Recommendation* task in the future. If you wish to contribute evaluation measures or give any suggestion or critique, please write an [e-mail](mailto:marco.wrzalik@hs-rm.de). ### Languages This dataset contains texts from the specific domain of German court decisions. ## Dataset Structure ### Data Instances ``` {'query.doc_id': 28860, 'query.ref_ids': [6215, 248, 248], 'query.sent_id': 304863, 'query.text': 'Zudem ist zu berücksichtigen , dass die Vollverzinsung nach ' '[REF] i. V. m. [REF] gleichermaßen zugunsten wie zulasten des ' 'Steuerpflichtigen wirkt , sodass bei einer Überzahlung durch ' 'den Steuerpflichtigen der Staat dem Steuerpflichtigen neben ' 'der Erstattung ebenfalls den entstandenen potentiellen Zins- ' 'und Liquiditätsnachteil in der pauschalierten Höhe des [REF] ' 'zu ersetzen hat , unabhängig davon , in welcher Höhe dem ' 'Berechtigten tatsächlich Zinsen entgangen sind .', 'related.doc_id': 56348, 'related.ref_ids': [248, 6215, 62375], 'related.sent_id': 558646, 'related.text': 'Ferner ist zu berücksichtigen , dass der Zinssatz des [REF] ' 'im Rahmen des [REF] sowohl für Steuernachforderung wie auch ' 'für Steuererstattungen und damit gleichermaßen zugunsten wie ' 'zulasten des Steuerpflichtigen wirkt , Vgl. BVerfG , ' 'Nichtannahmebeschluss vom [DATE] [REF] , juris , mit der ' 'Folge , dass auch Erstattungsansprüche unabhängig davon , ob ' 'und in welcher Höhe dem Berechtigten tatsächlich Zinsen ' 'entgangen sind , mit monatlich 0,0 % verzinst werden .'} ``` ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization The documents we take from [Open Legal Data](http://openlegaldata.io/) (https://arxiv.org/abs/2005.13342) are first preprocessed by removing line breaks, enumeration characters and headings. Afterwards we parse legal citations using hand-crafted regular expressions. Each citation is split into it components and normalized, thus different variants of the same citation are matched together. For instance, "§211 Absatz 1 des Strafgesetzbuches" is normalized to "§ 211 Abs. 1 StGB". Every time we discover an unknown citation, we assign an unique id to it. We use these ids to replace parsed citations in the document text with a simple reference tag containing this id (e.g `[REF321]`). At the same time we parse dates and replace them with the date tag `[DATE]`. Both remove dots which can may be confused with the end of a sentence, which makes the next stage easier. We use [SoMaJo](https://github.com/tsproisl/SoMaJo) to perform sentence tokenizing on the pre-processed documents. Each sentence that does not contain at least one legal citation is discarded. For the rest we assign sentence ids, remove all reference ids from them as well as any contents in braces (braces often contain large enumerations of citations and their sources). At the same time we keep track of the corresponding document from which a sentence originates and which references occur in it. #### Who are the source language producers? The source language originates in the context of German court proceedings. ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? The annotations are machine-generated. ### Personal and Sensitive Information The source documents are already public and anonymized. ## Considerations for Using the Data ### Social Impact of Dataset With this dataset, we strive towards better accessibility of court decisions to the general public by accelerating research on semantic search technologies. We hope that emerging search technologies will enable the layperson to find relevant information without knowing the specific terms used by lawyers. ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information Coming soon! ### Contributions Thanks to [@mwrzalik](https://github.com/mwrzalik) for adding this dataset.
[ -0.3053070306777954, -0.6488118171691895, 0.5966860055923462, 0.05480438098311424, -0.34128203988075256, -0.3778536915779114, -0.33543047308921814, -0.22077585756778717, 0.36270809173583984, 0.4365076720714569, -0.3717495799064636, -0.9940224289894104, -0.5817584991455078, 0.25738072395324...
null
null
null
null
null
null
null
null
null
null
null
null
null
m-newhauser/senator-tweets
m-newhauser
2022-03-07T16:37:44Z
39
1
null
[ "region:us" ]
2022-03-07T16:37:44Z
2022-03-07T16:37:35.000Z
2022-03-07T16:37:35
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
taln-ls2n/kp20k
taln-ls2n
2023-09-13T13:15:04Z
39
1
null
[ "task_categories:text-generation", "annotations_creators:unknown", "language_creators:unknown", "multilinguality:monolingual", "size_categories:100K<n<1M", "language:en", "license:unknown", "keyphrase-generation", "keyphrase-extraction", "text-mining", "region:us" ]
2023-09-13T13:15:04Z
2022-04-14T09:00:02.000Z
2022-04-14T09:00:02
--- annotations_creators: - unknown language_creators: - unknown language: - en license: - unknown multilinguality: - monolingual size_categories: - 100K<n<1M task_categories: - text-generation task_ids: [] pretty_name: KP20k tags: - keyphrase-generation - keyphrase-extraction - text-mining --- # KP20k Benchmark Dataset for Keyphrase Generation ## About KP20k is a dataset for benchmarking keyphrase extraction and generation models. The data is composed of 570 809 abstracts and their associated titles from scientific articles. Details about the dataset can be found in the original paper: - Meng et al 2017. [Deep keyphrase Generation](https://aclanthology.org/P17-1054.pdf) Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, pages 582–592 Reference (indexer-assigned) keyphrases are also categorized under the PRMU (<u>P</u>resent-<u>R</u>eordered-<u>M</u>ixed-<u>U</u>nseen) scheme as proposed in the following paper: - Florian Boudin and Ygor Gallina. 2021. [Redefining Absent Keyphrases and their Effect on Retrieval Effectiveness](https://aclanthology.org/2021.naacl-main.330/). In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 4185–4193, Online. Association for Computational Linguistics. Text pre-processing (tokenization) is carried out using spacy (en_core_web_sm model) with a special rule to avoid splitting words with hyphens (e.g. graph-based is kept as one token). Stemming (Porter's stemmer implementation provided in nltk) is applied before reference keyphrases are matched against the source text. ## Content The dataset is divided into the following three splits: | Split | # documents | # keyphrases by document (average) | % Present | % Reordered | % Mixed | % Unseen | | :--------- | ----------: | -----------: | --------: | ----------: | ------: | -------: | | Train | 530 809 | 5.29 | 58.19 | 10.93 | 17.36 | 13.52 | | Test | 20 000 | 5.28 | 58.40 | 10.84 | 17.20 | 13.56 | | Validation | 20 000 | 5.27 | 58.20 | 10.94 | 17.26 | 13.61 | The following data fields are available: - **id**: unique identifier of the document. **NB** There were no ids in the original dataset. The ids were generated using the python module shortuuid (https://pypi.org/project/shortuuid/) - **title**: title of the document. - **abstract**: abstract of the document. - **keyphrases**: list of the author assigned keyphrases. - **prmu**: list of <u>P</u>resent-<u>R</u>eordered-<u>M</u>ixed-<u>U</u>nseen categories for reference keyphrases. **NB**: The present keyphrases (represented by the "P" label in the PRMU column) are sorted by their apparition order in the text (title + abstract).
[ -0.2397076040506363, -0.27083244919776917, 0.4346559941768646, 0.24682028591632843, -0.39116397500038147, 0.18899905681610107, -0.06481043249368668, -0.18186625838279724, 0.10181933641433716, 0.4054800570011139, -0.5717522501945496, -0.8443481922149658, -0.5878648161888123, 0.5735462903976...
null
null
null
null
null
null
null
null
null
null
null
null
null
KevinZ/oLMpics
KevinZ
2022-04-19T18:08:06Z
39
0
null
[ "region:us" ]
2022-04-19T18:08:06Z
2022-04-18T02:14:53.000Z
2022-04-18T02:14:53
oLMpics README
[ -0.5414970517158508, -0.13784530758857727, 0.29386138916015625, -0.061467669904232025, -0.6916552186012268, -0.16386118531227112, 0.08057012408971786, -0.6702588200569153, 0.11930129677057266, 0.9155616164207458, -0.8323282599449158, -0.5169150233268738, -0.31646761298179626, -0.3293415009...
null
null
null
null
null
null
null
null
null
null
null
null
null
GroNLP/divemt
GroNLP
2023-02-10T11:04:33Z
39
2
null
[ "task_categories:translation", "annotations_creators:machine-generated", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:translation", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "language:it", "language:vi", "language:nl", "langu...
2023-02-10T11:04:33Z
2022-05-23T19:56:55.000Z
2022-05-23T19:56:55
--- annotations_creators: - machine-generated - expert-generated language_creators: - found language: - en - it - vi - nl - uk - tr - ar license: - gpl-3.0 multilinguality: - translation pretty_name: divemt size_categories: - 1K<n<10K source_datasets: - original task_categories: - translation --- # Dataset Card for DivEMT *For more details on DivEMT, see our [EMNLP 2022 Paper](https://arxiv.org/abs/2205.12215) and our [Github repository](https://github.com/gsarti/divemt)* ## Dataset Description - **Source:** [Github](https://github.com/gsarti/divemt) - **Paper:** [Arxiv](https://arxiv.org/abs/2205.12215) - **Point of Contact:** [Gabriele Sarti](mailto:g.sarti@rug.nl) [Gabriele Sarti](https://gsarti.com) • [Arianna Bisazza](https://www.cs.rug.nl/~bisazza/) • [Ana Guerberof Arenas](https://scholar.google.com/citations?user=i6bqaTsAAAAJ) • [Antonio Toral](https://antoniotor.al/) <img src="https://huggingface.co/datasets/GroNLP/divemt/resolve/main/divemt.png" alt="DivEMT annotation pipeline" width="600"/> >We introduce DivEMT, the first publicly available post-editing study of Neural Machine Translation (NMT) over a typologically diverse set of target languages. Using a strictly controlled setup, 18 professional translators were instructed to translate or post-edit the same set of English documents into Arabic, Dutch, Italian, Turkish, Ukrainian, and Vietnamese. During the process, their edits, keystrokes, editing times and pauses were recorded, enabling an in-depth, cross-lingual evaluation of NMT quality and post-editing effectiveness. Using this new dataset, we assess the impact of two state-of-the-art NMT systems, Google Translate and the multilingual mBART-50 model, on translation productivity. We find that post-editing is consistently faster than translation from scratch. However, the magnitude of productivity gains varies widely across systems and languages, highlighting major disparities in post-editing effectiveness for languages at different degrees of typological relatedness to English, even when controlling for system architecture and training data size. We publicly release the complete dataset including all collected behavioral data, to foster new research on the translation capabilities of NMT systems for typologically diverse languages. ### Dataset Summary This dataset contains the processed `warmup` and `main` splits of the DivEMT dataset. A sample of documents extracted from the Flores-101 corpus were either translated from scratch or post-edited from an existing automatic translation by a total of 18 professional translators across six typologically diverse languages (Arabic, Dutch, Italian, Turkish, Ukrainian, Vietnamese). During the translation, behavioral data (keystrokes, pauses, editing times) were collected using the [PET](https://github.com/wilkeraziz/PET) platform. We publicly release the processed dataset including all collected behavioural data, to foster new research on the ability of state-of-the-art NMT systems to generate text in typologically diverse languages. ### News 🎉 **February, 2023**: The DivEMT dataset now contains linguistic annotations (`*_annotations` fields) computed with Stanza and word-level quality estimation tags (`src_wmt22_qe`, `mt_wmt22_qe`) obtained using the same scripts adopted for the WMT22 QE Task 2. ### Languages The language data of DivEMT is in English (BCP-47 `en`), Italian (BCP-47 `it`), Dutch (BCP-47 `nl`), Arabic (BCP-47 `ar`), Turkish (BCP-47 `tr`), Ukrainian (BCP-47 `uk`) and Vietnamese (BCP-47 `vi`) ## Dataset Structure ### Data Instances The dataset contains two configurations: `main` and `warmup`. `main` contains the full data collected during the main task and analyzed during our experiments. `warmup` contains the data collected in the verification phase, before the main task begins. ### Data Fields The following fields are contained in the training set: |Field|Description| |-----|-----------| |`unit_id` | The full entry identifier. Format: `flores101-{config}-{lang}-{doc_id}-{modality}-{sent_in_doc_num}` | |`flores_id` | Index of the sentence in the original [Flores-101](https://huggingface.co/datasets/gsarti/flores_101) dataset | |`item_id` | The sentence identifier. The first digits of the number represent the document containing the sentence, while the last digit of the number represents the sentence position inside the document. Documents can contain from 3 to 5 contiguous sentences each. | |`subject_id` | The identifier for the translator performing the translation from scratch or post-editing task. Values: `t1`, `t2` or `t3`. | |`lang_id` | Language identifier for the sentence, using Flores-101 three-letter format (e.g. `ara`, `nld`)| |`doc_id` | Document identifier for the sentence | |`task_type` | The modality of the translation task. Values: `ht` (translation from scratch), `pe1` (post-editing Google Translate translations), `pe2` (post-editing [mBART 1-to-50](https://huggingface.co/facebook/mbart-large-50-one-to-many-mmt) translations). | |`translation_type` | Either `ht` for from scratch or `pe` for post-editing | |`src_len_chr` | Length of the English source text in number of characters | |`mt_len_chr` | Length of the machine translation in number of characters (NaN for ht) | |`tgt_len_chr` | Length of the target text in number of characters | |`src_len_wrd` | Length of the English source text in number of words | |`mt_len_wrd` | Length of the machine translation in number of words (NaN for ht) | |`tgt_len_wrd` | Length of the target text in number of words | |`edit_time` | Total editing time for the translation in seconds. | |`k_total` | Total number of keystrokes for the translation. | |`k_letter` | Total number of letter keystrokes for the translation. | |`k_digit` | Total number of digit keystrokes for the translation. | |`k_white` | Total number of whitespace keystrokes for the translation. | |`k_symbol` | Total number of symbol (punctuation, etc.) keystrokes for the translation. | |`k_nav` | Total number of navigation keystrokes (left-right arrows, mouse clicks) for the translation. | |`k_erase` | Total number of erase keystrokes (backspace, cancel) for the translation. | |`k_copy` | Total number of copy (Ctrl + C) actions during the translation. | |`k_cut` | Total number of cut (Ctrl + X) actions during the translation. | |`k_paste` | Total number of paste (Ctrl + V) actions during the translation. | |`k_do` | Total number of Enter actions during the translation. | |`n_pause_geq_300` | Number of pauses of 300ms or more during the translation. | |`len_pause_geq_300` | Total duration of pauses of 300ms or more, in milliseconds. | |`n_pause_geq_1000` | Number of pauses of 1s or more during the translation. | |`len_pause_geq_1000` | Total duration of pauses of 1000ms or more, in milliseconds. | |`event_time` | Total time summed across all translation events, should be comparable to `edit_time` in most cases. | |`num_annotations` | Number of times the translator focused the textbox for performing the translation of the sentence during the translation session. E.g. 1 means the translation was performed once and never revised. | |`n_insert` | Number of post-editing insertions (empty for modality `ht`) computed using the [tercom](https://github.com/jhclark/tercom) library. | |`n_delete` | Number of post-editing deletions (empty for modality `ht`) computed using the [tercom](https://github.com/jhclark/tercom) library. | |`n_substitute` | Number of post-editing substitutions (empty for modality `ht`) computed using the [tercom](https://github.com/jhclark/tercom) library. | |`n_shift` | Number of post-editing shifts (empty for modality `ht`) computed using the [tercom](https://github.com/jhclark/tercom) library. | |`tot_shifted_words` | Total amount of shifted words from all shifts present in the sentence. | |`tot_edits` | Total of all edit types for the sentence. | |`hter` | Human-mediated Translation Edit Rate score computed between MT and post-edited TGT (empty for modality `ht`) using the [tercom](https://github.com/jhclark/tercom) library. | |`cer` | Character-level HTER score computed between MT and post-edited TGT (empty for modality `ht`) using [CharacTER](https://github.com/rwth-i6/CharacTER). |`bleu` | Sentence-level BLEU score between MT and post-edited TGT (empty for modality `ht`) computed using the [SacreBLEU](https://github.com/mjpost/sacrebleu) library with default parameters. | |`chrf` | Sentence-level chrF score between MT and post-edited TGT (empty for modality `ht`) computed using the [SacreBLEU](https://github.com/mjpost/sacrebleu) library with default parameters. | |`time_s` | Edit time expressed in seconds. | |`time_m` | Edit time expressed in minutes. | |`time_h` | Edit time expressed in hours. | |`time_per_char` | Edit time per source character, expressed in seconds. | |`time_per_word` | Edit time per source word, expressed in seconds. | |`key_per_char` | Proportion of keys per character needed to perform the translation. | |`words_per_hour` | Amount of source words translated or post-edited per hour. | |`words_per_minute` | Amount of source words translated or post-edited per minute. | |`per_subject_visit_order` | Id denoting the order in which the translator accessed documents. 1 correspond to the first accessed document. | |`src_text` | The original source sentence extracted from Wikinews, wikibooks or wikivoyage. | |`mt_text` | Missing if tasktype is `ht`. Otherwise, contains the automatically-translated sentence before post-editing. | |`tgt_text` | Final sentence produced by the translator (either via translation from scratch of `sl_text` or post-editing `mt_text`) | |`aligned_edit` | Aligned visual representation of REF (`mt_text`), HYP (`tl_text`) and edit operations (I = Insertion, D = Deletion, S = Substitution) performed on the field. Replace `\\n` with `\n` to show the three aligned rows.| |`src_tokens` | List of tokens obtained tokenizing `src_text` with Stanza using default params. | |`src_annotations` | List of lists (one per `src_tokens` token) containing dictionaries (one per word, >1 for mwt) with pos, ner and other info parsed by Stanza | |`mt_tokens` | List of tokens obtained tokenizing `mt_text` with Stanza using default params. | |`mt_annotations` | List of lists (one per `mt_tokens` token) containing dictionaries (one per word, >1 for mwt) with pos, ner and other info parsed by Stanza | |`tgt_tokens` | List of tokens obtained tokenizing `tgt_text` with Stanza using default params. | |`tgt_annotations` | List of lists (one per `tgt_tokens` token) containing dictionaries (one per word, >1 for mwt) with pos, ner and other info parsed by Stanza | ### Data Splits | config | train| |-------:|-----:| |`main` | 7740 (107 docs i.e. 430 sents x 18 translators) | |`warmup`| 360 (5 docs i.e. 20 sents x 18 translators) | #### Train Split The `train` split contains the totality of triplets (or pairs, when translation from scratch is performed) annotated with behavioral data produced during the translation. The following is an example of the subject `t1` post-editing a machine translation produced by Google Translate (task_type `pe1`) taken from the `train` split for Turkish. The field `aligned_edit` is showed over three lines to provide a visual understanding of its contents. ```json { 'unit_id': 'flores101-main-tur-46-pe1-3', 'flores_id': 871, 'item_id': 'flores101-main-463', 'subject_id': 'tur_t1', 'task_type': 'pe1', 'translation_type': 'pe', 'src_len_chr': 109, 'mt_len_chr': 129.0, 'tgt_len_chr': 120, 'src_len_wrd': 17, 'mt_len_wrd': 15.0, 'tgt_len_wrd': 13, 'edit_time': 11.762999534606934, 'k_total': 31, 'k_letter': 9, 'k_digit': 0, 'k_white': 0, 'k_symbol': 0, 'k_nav': 20, 'k_erase': 2, 'k_copy': 0, 'k_cut': 0, 'k_paste': 0, 'k_do': 0, 'n_pause_geq_300': 2, 'len_pause_geq_300': 4986, 'n_pause_geq_1000': 1, 'len_pause_geq_1000': 4490, 'event_time': 11763, 'num_annotations': 2, 'last_modification_time': 1643569484, 'n_insert': 0.0, 'n_delete': 2.0, 'n_substitute': 1.0, 'n_shift': 0.0, 'tot_shifted_words': 0.0, 'tot_edits': 3.0, 'hter': 20.0, 'cer': 0.10, 'bleu': 0.0, 'chrf': 2.569999933242798, 'lang_id': 'tur', 'doc_id': 46, 'time_s': 11.762999534606934, 'time_m': 0.1960500031709671, 'time_h': 0.0032675000838935375, 'time_per_char': 0.1079174280166626, 'time_per_word': 0.6919412016868591, 'key_per_char': 0.2844036817550659, 'words_per_hour': 5202.75439453125, 'words_per_minute': 86.71257019042969, 'per_subject_visit_order': 201, 'src_text': 'As one example, American citizens in the Middle East might face different situations from Europeans or Arabs.', 'mt_text': "Bir örnek olarak, Orta Doğu'daki Amerikan vatandaşları, Avrupalılardan veya Araplardan farklı durumlarla karşı karşıya kalabilir.", 'tgt_text': "Örneğin, Orta Doğu'daki Amerikan vatandaşları, Avrupalılardan veya Araplardan farklı durumlarla karşı karşıya kalabilir.", 'aligned_edit': "REF: bir örnek olarak, orta doğu'daki amerikan vatandaşları, avrupalılardan veya araplardan farklı durumlarla karşı karşıya kalabilir.\\n HYP: *** ***** örneğin, orta doğu'daki amerikan vatandaşları, avrupalılardan veya araplardan farklı durumlarla karşı karşıya kalabilir.\\n EVAL: D D S" } ``` The text is provided as-is, without further preprocessing or tokenization. ### Dataset Creation The dataset was parsed from PET XML files into CSV format using the scripts available in the [DivEMT Github repository](https://github.com/gsarti/divemt). Those are adapted from the ones by [Antonio Toral](https://research.rug.nl/en/persons/antonio-toral-ruiz) found at the following link: [https://github.com/antot/postediting_novel_frontiers](https://github.com/antot/postediting_novel_frontiers). ## Additional Information ### Dataset Curators For problems related to this 🤗 Datasets version, please contact me at [g.sarti@rug.nl](mailto:g.sarti@rug.nl). ### Citation Information ```bibtex @inproceedings{sarti-etal-2022-divemt, title = "{D}iv{EMT}: Neural Machine Translation Post-Editing Effort Across Typologically Diverse Languages", author = "Sarti, Gabriele and Bisazza, Arianna and Guerberof-Arenas, Ana and Toral, Antonio", booktitle = "Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing", month = dec, year = "2022", address = "Abu Dhabi, United Arab Emirates", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2022.emnlp-main.532", pages = "7795--7816", } ```
[ -0.3049851953983307, -0.6313626766204834, 0.44244933128356934, 0.30382850766181946, -0.4271915853023529, -0.10384097695350647, -0.4326137602329254, -0.24820920825004578, 0.37696701288223267, 0.39765429496765137, -0.6339412927627563, -0.881414532661438, -0.6099082231521606, 0.48315957188606...
null
null
null
null
null
null
null
null
null
null
null
null
null
taln-ls2n/pubmed
taln-ls2n
2022-10-26T19:14:46Z
39
1
null
[ "task_categories:text-generation", "annotations_creators:unknown", "language_creators:unknown", "multilinguality:monolingual", "size_categories:1k<n<10k", "language:en", "license:unknown", "keyphrase-generation", "keyphrase-extraction", "text-mining", "region:us" ]
2022-10-26T19:14:46Z
2022-05-24T08:34:08.000Z
2022-05-24T08:34:08
--- annotations_creators: - unknown language_creators: - unknown language: - en license: - unknown multilinguality: - monolingual size_categories: - 1k<n<10k task_categories: - text-generation task_ids: [] pretty_name: PubMed tags: - keyphrase-generation - keyphrase-extraction - text-mining --- # Schutz 2008 PubMed dataset for keyphrase extraction ## About This dataset is made of 1320 articles with full text and author assigned keyphrases. Details about the dataset can be found in the original paper: Keyphrase extraction from single documents in the open domain exploiting linguistic and statistical methods. Alexander Thorsten Schutz. Master's thesis, National University of Ireland (2008). Reference (indexer-assigned) keyphrases are also categorized under the PRMU (<u>P</u>resent-<u>R</u>eordered-<u>M</u>ixed-<u>U</u>nseen) scheme as proposed in the following paper: - Florian Boudin and Ygor Gallina. 2021. [Redefining Absent Keyphrases and their Effect on Retrieval Effectiveness](https://aclanthology.org/2021.naacl-main.330/). In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 4185–4193, Online. Association for Computational Linguistics. Text pre-processing (tokenization) is carried out using spacy (en_core_web_sm model) with a special rule to avoid splitting words with hyphens (e.g. graph-based is kept as one token). Stemming (Porter's stemmer implementation provided in nltk) is applied before reference keyphrases are matched against the source text. ## Content The details of the dataset are in the table below: | Split | # documents | # keyphrases by document (average) | % Present | % Reordered | % Mixed | % Unseen | | :--------- | ----------: | -----------: | --------: | ----------: | ------: | -------: | | Test | 1320 | 5.40 | 84.54 | 9.14 | 3.84 | 2.47 | The following data fields are available: - **id**: unique identifier of the document. - **title**: title of the document. - **text**: full article minus the title. - **keyphrases**: list of reference keyphrases. - **prmu**: list of <u>P</u>resent-<u>R</u>eordered-<u>M</u>ixed-<u>U</u>nseen categories for reference keyphrases. **NB**: The present keyphrases (represented by the "P" label in the PRMU column) are sorted by their apparition order in the text (title + text).
[ -0.005555886309593916, -0.2898070514202118, 0.48875126242637634, 0.24048380553722382, -0.46108391880989075, -0.09348160028457642, -0.1663980484008789, -0.06701760739088058, 0.2617139220237732, 0.5958781838417053, -0.3619893193244934, -0.7547374963760376, -0.6398683786392212, 0.675595998764...
null
null
null
null
null
null
null
null
null
null
null
null
null
AhmedSSabir/Textual-Image-Caption-Dataset
AhmedSSabir
2023-10-14T12:32:07Z
39
5
null
[ "task_categories:image-to-text", "task_categories:image-classification", "task_categories:visual-question-answering", "task_categories:sentence-similarity", "language:en", "image captioning", "language grounding", "visual semantic", "semantic similarity", "arxiv:2301.08784", "arxiv:1408.5882", ...
2023-10-14T12:32:07Z
2022-06-08T10:36:12.000Z
2022-06-08T10:36:12
--- task_categories: - image-to-text - image-classification - visual-question-answering - sentence-similarity language: - en tags: - image captioning - language grounding - visual semantic - semantic similarity pretty_name: ' image captioning language grounding visual semantic ' --- #### Update: OCT-2023 ### Add v2 with recent SoTA model **swinV2 classifier** for both soft/*hard-label* visual_caption_cosine_score_v2 with _person_ label (0.2, 0.3 and 0.4) # Introduction Modern image captaining relies heavily on extracting knowledge, from images such as objects, to capture the concept of static story in the image. In this paper, we propose a textual visual context dataset for captioning, where the publicly available dataset COCO caption (Lin et al., 2014) has been extended with information about the scene (such as objects in the image). Since this information has textual form, it can be used to leverage any NLP task, such as text similarity or semantic relation methods, into captioning systems, either as an end-to-end training strategy or a post-processing based approach. Please refer to [project page](https://sabirdvd.github.io/project_page/Dataset_2022/index.html) and [Github](https://github.com/ahmedssabir/Visual-Semantic-Relatedness-Dataset-for-Image-Captioning) for more information. [![arXiv](https://img.shields.io/badge/arXiv-2301.08784-b31b1b.svg)](https://arxiv.org/abs/2301.08784) [![Website shields.io](https://img.shields.io/website-up-down-green-red/http/shields.io.svg)](https://ahmed.jp/project_page/Dataset_2022/index.html) For quick start please have a look this [demo](https://github.com/ahmedssabir/Textual-Visual-Semantic-Dataset/blob/main/BERT_CNN_Visual_re_ranker_demo.ipynb) and [pre-trained model with th 0.2, 0.3, 0.4](https://huggingface.co/AhmedSSabir/BERT-CNN-Visual-Semantic) # Overview We enrich COCO-Caption with textual Visual Context information. We use ResNet152, CLIP, and Faster R-CNN to extract object information for each image. We use three filter approaches to ensure the quality of the dataset (1) Threshold: to filter out predictions where the object classifier is not confident enough, and (2) semantic alignment with semantic similarity to remove duplicated objects. (3) semantic relatedness score as soft-label: to guarantee the visual context and caption have a strong relation. In particular, we use Sentence-RoBERTa-sts via cosine similarity to give a soft score, and then we use a threshold to annotate the final label (if th ≥ 0.2, 0.3, 0.4 then 1,0). Finally, to take advantage of the visual overlap between caption and visual context, and to extract global information, we use BERT followed by a shallow 1D-CNN (Kim, 2014) to estimate the visual relatedness score. <!-- ## Dataset (<a href="https://arxiv.org/abs/1408.5882">Kim, 2014</a>) ### Sample ``` |---------------+--------------+---------+---------------------------------------------------| | VC1 | VC2 | VC3 | human annoated caption | | ------------- | ----------- | --------| ------------------------------------------------- | | cheeseburger | plate | hotdog | a plate with a hamburger fries and tomatoes | | bakery | dining table | website | a table having tea and a cake on it | | gown | groom | apron | its time to cut the cake at this couples wedding | |---------------+--------------+---------+---------------------------------------------------| ``` --> ### Download 0. [Dowload Raw data with ID and Visual context](https://www.dropbox.com/s/xuov24on8477zg8/All_Caption_ID.csv?dl=0) -> original dataset with related ID caption [train2014](https://cocodataset.org/#download) 1. [Downlod Data with cosine score](https://www.dropbox.com/s/55sit8ow9tems4u/visual_caption_cosine_score.zip?dl=0)-> soft cosine lable with **th** 0.2, 0.3, 0.4 and 0.5 and hardlabel [0,1] 2. [Dowload Overlaping visual with caption](https://www.dropbox.com/s/br8nhnlf4k2czo8/COCO_overlaping_dataset.txt?dl=0)-> Overlap visual context and the human annotated caption 3. [Download Dataset (tsv file)](https://www.dropbox.com/s/dh38xibtjpohbeg/train_all.zip?dl=0) 0.0-> raw data with hard lable without cosine similairty and with **th**reshold cosine sim degree of the relation beteween the visual and caption = 0.2, 0.3, 0.4 4. [Download Dataset GenderBias](https://www.dropbox.com/s/1wki0b0d21078mj/gender%20natural.zip?dl=0)-> man/woman replaced with person class label For future work, we plan to extract the visual context from the caption (without using a visual classifier) and estimate the visual relatedness score by employing unsupervised learning (i.e. contrastive learning). (work in progress) 1. [Download CC](https://www.dropbox.com/s/pc1uv2rf6nqdp57/CC_caption_40.txt.zip) -> Caption dataset from Conceptinal Caption (CC) 2M (2255927 captions) 2. [Download CC+wiki](https://www.dropbox.com/s/xuov24on8477zg8/All_Caption_ID.csv?dl=0) -> CC+1M-wiki 3M (3255928) 3. [Download CC+wiki+COCO](https://www.dropbox.com/s/k7oqwr9a1a0h8x1/CC_caption_40%2Bwiki%2BCOCO.txt.zip) -> CC+wiki+COCO-Caption 3.5M (366984) 4. [Download COCO-caption+wiki](https://www.dropbox.com/s/wc4k677wp24kzhh/COCO%2Bwiki.txt.zip) -> COCO-caption +wiki 1.4M (1413915) 5. [Download COCO-caption+wiki+CC+8Mwiki](https://www.dropbox.com/s/xhfx32sjy2z5bpa/11M_wiki_7M%2BCC%2BCOCO.txt.zip) -> COCO-caption+wiki+CC+8Mwiki 11M (11541667) ## Citation The details of this repo are described in the following paper. If you find this repo useful, please kindly cite it: ```bibtex @article{sabir2023visual, title={Visual Semantic Relatedness Dataset for Image Captioning}, author={Sabir, Ahmed and Moreno-Noguer, Francesc and Padr{\'o}, Llu{\'\i}s}, journal={arXiv preprint arXiv:2301.08784}, year={2023} } ```
[ -0.7025397419929504, -0.6237320303916931, 0.056404128670692444, 0.3135712444782257, -0.5657690763473511, -0.09819695353507996, -0.18751709163188934, -0.6892073154449463, 0.44647255539894104, 0.5506954789161682, -0.7087982892990112, -0.7152636647224426, -0.4892575740814209, 0.27318418025970...
null
null
null
null
null
null
null
null
null
null
null
null
null
BeIR/trec-news-generated-queries
BeIR
2022-10-23T06:13:54Z
39
1
beir
[ "task_categories:text-retrieval", "task_ids:entity-linking-retrieval", "task_ids:fact-checking-retrieval", "multilinguality:monolingual", "language:en", "license:cc-by-sa-4.0", "region:us" ]
2022-10-23T06:13:54Z
2022-06-17T13:04:13.000Z
2022-06-17T13:04:13
--- annotations_creators: [] language_creators: [] language: - en license: - cc-by-sa-4.0 multilinguality: - monolingual paperswithcode_id: beir pretty_name: BEIR Benchmark size_categories: msmarco: - 1M<n<10M trec-covid: - 100k<n<1M nfcorpus: - 1K<n<10K nq: - 1M<n<10M hotpotqa: - 1M<n<10M fiqa: - 10K<n<100K arguana: - 1K<n<10K touche-2020: - 100K<n<1M cqadupstack: - 100K<n<1M quora: - 100K<n<1M dbpedia: - 1M<n<10M scidocs: - 10K<n<100K fever: - 1M<n<10M climate-fever: - 1M<n<10M scifact: - 1K<n<10K source_datasets: [] task_categories: - text-retrieval - zero-shot-retrieval - information-retrieval - zero-shot-information-retrieval task_ids: - passage-retrieval - entity-linking-retrieval - fact-checking-retrieval - tweet-retrieval - citation-prediction-retrieval - duplication-question-retrieval - argument-retrieval - news-retrieval - biomedical-information-retrieval - question-answering-retrieval --- # Dataset Card for BEIR Benchmark ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://github.com/UKPLab/beir - **Repository:** https://github.com/UKPLab/beir - **Paper:** https://openreview.net/forum?id=wCu6T5xFjeJ - **Leaderboard:** https://docs.google.com/spreadsheets/d/1L8aACyPaXrL8iEelJLGqlMqXKPX2oSP_R10pZoy77Ns - **Point of Contact:** nandan.thakur@uwaterloo.ca ### Dataset Summary BEIR is a heterogeneous benchmark that has been built from 18 diverse datasets representing 9 information retrieval tasks: - Fact-checking: [FEVER](http://fever.ai), [Climate-FEVER](http://climatefever.ai), [SciFact](https://github.com/allenai/scifact) - Question-Answering: [NQ](https://ai.google.com/research/NaturalQuestions), [HotpotQA](https://hotpotqa.github.io), [FiQA-2018](https://sites.google.com/view/fiqa/) - Bio-Medical IR: [TREC-COVID](https://ir.nist.gov/covidSubmit/index.html), [BioASQ](http://bioasq.org), [NFCorpus](https://www.cl.uni-heidelberg.de/statnlpgroup/nfcorpus/) - News Retrieval: [TREC-NEWS](https://trec.nist.gov/data/news2019.html), [Robust04](https://trec.nist.gov/data/robust/04.guidelines.html) - Argument Retrieval: [Touche-2020](https://webis.de/events/touche-20/shared-task-1.html), [ArguAna](tp://argumentation.bplaced.net/arguana/data) - Duplicate Question Retrieval: [Quora](https://www.quora.com/q/quoradata/First-Quora-Dataset-Release-Question-Pairs), [CqaDupstack](http://nlp.cis.unimelb.edu.au/resources/cqadupstack/) - Citation-Prediction: [SCIDOCS](https://allenai.org/data/scidocs) - Tweet Retrieval: [Signal-1M](https://research.signal-ai.com/datasets/signal1m-tweetir.html) - Entity Retrieval: [DBPedia](https://github.com/iai-group/DBpedia-Entity/) All these datasets have been preprocessed and can be used for your experiments. ```python ``` ### Supported Tasks and Leaderboards The dataset supports a leaderboard that evaluates models against task-specific metrics such as F1 or EM, as well as their ability to retrieve supporting information from Wikipedia. The current best performing models can be found [here](https://eval.ai/web/challenges/challenge-page/689/leaderboard/). ### Languages All tasks are in English (`en`). ## Dataset Structure All BEIR datasets must contain a corpus, queries and qrels (relevance judgments file). They must be in the following format: - `corpus` file: a `.jsonl` file (jsonlines) that contains a list of dictionaries, each with three fields `_id` with unique document identifier, `title` with document title (optional) and `text` with document paragraph or passage. For example: `{"_id": "doc1", "title": "Albert Einstein", "text": "Albert Einstein was a German-born...."}` - `queries` file: a `.jsonl` file (jsonlines) that contains a list of dictionaries, each with two fields `_id` with unique query identifier and `text` with query text. For example: `{"_id": "q1", "text": "Who developed the mass-energy equivalence formula?"}` - `qrels` file: a `.tsv` file (tab-seperated) that contains three columns, i.e. the `query-id`, `corpus-id` and `score` in this order. Keep 1st row as header. For example: `q1 doc1 1` ### Data Instances A high level example of any beir dataset: ```python corpus = { "doc1" : { "title": "Albert Einstein", "text": "Albert Einstein was a German-born theoretical physicist. who developed the theory of relativity, \ one of the two pillars of modern physics (alongside quantum mechanics). His work is also known for \ its influence on the philosophy of science. He is best known to the general public for his mass–energy \ equivalence formula E = mc2, which has been dubbed 'the world's most famous equation'. He received the 1921 \ Nobel Prize in Physics 'for his services to theoretical physics, and especially for his discovery of the law \ of the photoelectric effect', a pivotal step in the development of quantum theory." }, "doc2" : { "title": "", # Keep title an empty string if not present "text": "Wheat beer is a top-fermented beer which is brewed with a large proportion of wheat relative to the amount of \ malted barley. The two main varieties are German Weißbier and Belgian witbier; other types include Lambic (made\ with wild yeast), Berliner Weisse (a cloudy, sour beer), and Gose (a sour, salty beer)." }, } queries = { "q1" : "Who developed the mass-energy equivalence formula?", "q2" : "Which beer is brewed with a large proportion of wheat?" } qrels = { "q1" : {"doc1": 1}, "q2" : {"doc2": 1}, } ``` ### Data Fields Examples from all configurations have the following features: ### Corpus - `corpus`: a `dict` feature representing the document title and passage text, made up of: - `_id`: a `string` feature representing the unique document id - `title`: a `string` feature, denoting the title of the document. - `text`: a `string` feature, denoting the text of the document. ### Queries - `queries`: a `dict` feature representing the query, made up of: - `_id`: a `string` feature representing the unique query id - `text`: a `string` feature, denoting the text of the query. ### Qrels - `qrels`: a `dict` feature representing the query document relevance judgements, made up of: - `_id`: a `string` feature representing the query id - `_id`: a `string` feature, denoting the document id. - `score`: a `int32` feature, denoting the relevance judgement between query and document. ### Data Splits | Dataset | Website| BEIR-Name | Type | Queries | Corpus | Rel D/Q | Down-load | md5 | | -------- | -----| ---------| --------- | ----------- | ---------| ---------| :----------: | :------:| | MSMARCO | [Homepage](https://microsoft.github.io/msmarco/)| ``msmarco`` | ``train``<br>``dev``<br>``test``| 6,980 | 8.84M | 1.1 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/msmarco.zip) | ``444067daf65d982533ea17ebd59501e4`` | | TREC-COVID | [Homepage](https://ir.nist.gov/covidSubmit/index.html)| ``trec-covid``| ``test``| 50| 171K| 493.5 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/trec-covid.zip) | ``ce62140cb23feb9becf6270d0d1fe6d1`` | | NFCorpus | [Homepage](https://www.cl.uni-heidelberg.de/statnlpgroup/nfcorpus/) | ``nfcorpus`` | ``train``<br>``dev``<br>``test``| 323 | 3.6K | 38.2 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/nfcorpus.zip) | ``a89dba18a62ef92f7d323ec890a0d38d`` | | BioASQ | [Homepage](http://bioasq.org) | ``bioasq``| ``train``<br>``test`` | 500 | 14.91M | 8.05 | No | [How to Reproduce?](https://github.com/UKPLab/beir/blob/main/examples/dataset#2-bioasq) | | NQ | [Homepage](https://ai.google.com/research/NaturalQuestions) | ``nq``| ``train``<br>``test``| 3,452 | 2.68M | 1.2 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/nq.zip) | ``d4d3d2e48787a744b6f6e691ff534307`` | | HotpotQA | [Homepage](https://hotpotqa.github.io) | ``hotpotqa``| ``train``<br>``dev``<br>``test``| 7,405 | 5.23M | 2.0 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/hotpotqa.zip) | ``f412724f78b0d91183a0e86805e16114`` | | FiQA-2018 | [Homepage](https://sites.google.com/view/fiqa/) | ``fiqa`` | ``train``<br>``dev``<br>``test``| 648 | 57K | 2.6 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/fiqa.zip) | ``17918ed23cd04fb15047f73e6c3bd9d9`` | | Signal-1M(RT) | [Homepage](https://research.signal-ai.com/datasets/signal1m-tweetir.html)| ``signal1m`` | ``test``| 97 | 2.86M | 19.6 | No | [How to Reproduce?](https://github.com/UKPLab/beir/blob/main/examples/dataset#4-signal-1m) | | TREC-NEWS | [Homepage](https://trec.nist.gov/data/news2019.html) | ``trec-news`` | ``test``| 57 | 595K | 19.6 | No | [How to Reproduce?](https://github.com/UKPLab/beir/blob/main/examples/dataset#1-trec-news) | | ArguAna | [Homepage](http://argumentation.bplaced.net/arguana/data) | ``arguana``| ``test`` | 1,406 | 8.67K | 1.0 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/arguana.zip) | ``8ad3e3c2a5867cdced806d6503f29b99`` | | Touche-2020| [Homepage](https://webis.de/events/touche-20/shared-task-1.html) | ``webis-touche2020``| ``test``| 49 | 382K | 19.0 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/webis-touche2020.zip) | ``46f650ba5a527fc69e0a6521c5a23563`` | | CQADupstack| [Homepage](http://nlp.cis.unimelb.edu.au/resources/cqadupstack/) | ``cqadupstack``| ``test``| 13,145 | 457K | 1.4 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/cqadupstack.zip) | ``4e41456d7df8ee7760a7f866133bda78`` | | Quora| [Homepage](https://www.quora.com/q/quoradata/First-Quora-Dataset-Release-Question-Pairs) | ``quora``| ``dev``<br>``test``| 10,000 | 523K | 1.6 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/quora.zip) | ``18fb154900ba42a600f84b839c173167`` | | DBPedia | [Homepage](https://github.com/iai-group/DBpedia-Entity/) | ``dbpedia-entity``| ``dev``<br>``test``| 400 | 4.63M | 38.2 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/dbpedia-entity.zip) | ``c2a39eb420a3164af735795df012ac2c`` | | SCIDOCS| [Homepage](https://allenai.org/data/scidocs) | ``scidocs``| ``test``| 1,000 | 25K | 4.9 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/scidocs.zip) | ``38121350fc3a4d2f48850f6aff52e4a9`` | | FEVER | [Homepage](http://fever.ai) | ``fever``| ``train``<br>``dev``<br>``test``| 6,666 | 5.42M | 1.2| [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/fever.zip) | ``5a818580227bfb4b35bb6fa46d9b6c03`` | | Climate-FEVER| [Homepage](http://climatefever.ai) | ``climate-fever``|``test``| 1,535 | 5.42M | 3.0 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/climate-fever.zip) | ``8b66f0a9126c521bae2bde127b4dc99d`` | | SciFact| [Homepage](https://github.com/allenai/scifact) | ``scifact``| ``train``<br>``test``| 300 | 5K | 1.1 | [Link](https://public.ukp.informatik.tu-darmstadt.de/thakur/BEIR/datasets/scifact.zip) | ``5f7d1de60b170fc8027bb7898e2efca1`` | | Robust04 | [Homepage](https://trec.nist.gov/data/robust/04.guidelines.html) | ``robust04``| ``test``| 249 | 528K | 69.9 | No | [How to Reproduce?](https://github.com/UKPLab/beir/blob/main/examples/dataset#3-robust04) | ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators [Needs More Information] ### Licensing Information [Needs More Information] ### Citation Information Cite as: ``` @inproceedings{ thakur2021beir, title={{BEIR}: A Heterogeneous Benchmark for Zero-shot Evaluation of Information Retrieval Models}, author={Nandan Thakur and Nils Reimers and Andreas R{\"u}ckl{\'e} and Abhishek Srivastava and Iryna Gurevych}, booktitle={Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2)}, year={2021}, url={https://openreview.net/forum?id=wCu6T5xFjeJ} } ``` ### Contributions Thanks to [@Nthakur20](https://github.com/Nthakur20) for adding this dataset.
[ -0.5227212905883789, -0.5249219536781311, 0.14435674250125885, 0.04820423573255539, 0.055916160345077515, 0.0011022627586498857, -0.1081070527434349, -0.24874727427959442, 0.28598034381866455, 0.07840226590633392, -0.45233607292175293, -0.7186435461044312, -0.347678542137146, 0.20300328731...
null
null
null
null
null
null
null
null
null
null
null
null
null
codeparrot/codecomplex
codeparrot
2022-10-25T09:30:16Z
39
11
null
[ "task_categories:text-generation", "task_ids:language-modeling", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:unknown", "language:code", "license:apache-2.0", "region:us" ]
2022-10-25T09:30:16Z
2022-06-24T20:18:43.000Z
2022-06-24T20:18:43
--- annotations_creators: [] language_creators: - expert-generated language: - code license: - apache-2.0 multilinguality: - monolingual size_categories: - unknown source_datasets: [] task_categories: - text-generation task_ids: - language-modeling pretty_name: CodeComplex --- # CodeComplex Dataset ## Dataset Description [CodeComplex](https://github.com/yonsei-toc/CodeComple) consists of 4,200 Java codes submitted to programming competitions by human programmers and their complexity labels annotated by a group of algorithm experts. ### How to use it You can load and iterate through the dataset with the following two lines of code: ```python from datasets import load_dataset ds = load_dataset("codeparrot/codecomplex", split="train") print(next(iter(ds))) ``` ## Data Structure ``` DatasetDict({ train: Dataset({ features: ['src', 'complexity', 'problem', 'from'], num_rows: 4517 }) }) ``` ### Data Instances ```python {'src': 'import java.io.*;\nimport java.math.BigInteger;\nimport java.util.InputMismatchException;...', 'complexity': 'quadratic', 'problem': '1179_B. Tolik and His Uncle', 'from': 'CODEFORCES'} ``` ### Data Fields * src: a string feature, representing the source code in Java. * complexity: a string feature, giving program complexity. * problem: a string of the feature, representing the problem name. * from: a string feature, representing the source of the problem. complexity filed has 7 classes, where each class has around 500 codes each. The seven classes are constant, linear, quadratic, cubic, log(n), nlog(n) and NP-hard. ### Data Splits The dataset only contains a train split. ## Dataset Creation The authors first collected problem and solution codes in Java from CodeForces and they were inspected by experienced human annotators to label each code by their time complexity. After the labelling, they used different programming experts to verify the class of each data that the human annotators assigned. ## Citation Information ``` @article{JeonBHHK22, author = {Mingi Jeon and Seung-Yeop Baik and Joonghyuk Hahn and Yo-Sub Han and Sang-Ki Ko}, title = {{Deep Learning-based Code Complexity Prediction}}, year = {2022}, } ```
[ -0.5097919702529907, -0.3065548837184906, 0.17924386262893677, 0.25846609473228455, 0.07703936100006104, 0.27944818139076233, -0.3755101263523102, -0.3563574552536011, -0.17258967459201813, 0.34529778361320496, -0.338498055934906, -0.5452361106872559, -0.6716673970222473, 0.241822689771652...
null
null
null
null
null
null
null
null
null
null
null
null
null
imvladikon/nemo_corpus
imvladikon
2023-11-24T10:36:57Z
39
0
null
[ "task_categories:token-classification", "task_ids:named-entity-recognition", "annotations_creators:crowdsourced", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:extended|other-reuters-corpus", "language:he", "region:us" ]
2023-11-24T10:36:57Z
2022-06-28T16:51:45.000Z
2022-06-28T16:51:45
--- annotations_creators: - crowdsourced language_creators: - found language: - he multilinguality: - monolingual size_categories: - 10K<n<100K source_datasets: - extended|other-reuters-corpus task_categories: - token-classification task_ids: - named-entity-recognition train-eval-index: - config: nemo_corpus task: token-classification task_id: entity_extraction splits: train_split: train eval_split: validation test_split: test col_mapping: tokens: tokens ner_tags: tags metrics: - type: seqeval name: seqeval --- # NEMO-Corpus - The Hebrew Named Entities and Morphology Corpus ## Config and Usage Config: * flat_token - flatten tags * nested_token - nested tags * flat_morph - flatten tags with morphologically presegmentized tokens * nested_morph - nested tags with morphologically presegmentized tokens Note: It seems that a couple of samples for the flat_token and nested_token are mistakenly presegmented, and as a result, these samples have white space in the token. ```python from datasets import load_dataset # the main corpus ds = load_dataset('imvladikon/nemo_corpus', "flat_token") for sample in ds["train"]: print(sample) # the nested corpus ds = load_dataset('imvladikon/nemo_corpus', "nested_morph") ``` Getting classes and encoding/decoding could be done through these functions: ``` idx2label = dataset["train"].features["ner_tags"].feature.int2str label2idx = dataset["train"].features["ner_tags"].feature.str2int ``` or just use raw_tags field. ## Fields available fields (flat): * "id" * "sentence" * "tokens" * "raw_tags" * "ner_tags" Example of the one record for `flat`: ```json {'id': '0', 'tokens': ['"', 'תהיה', 'נקמה', 'ו', 'בגדול', '.'], 'sentence': '" תהיה נקמה ו בגדול .', 'raw_tags': ['O', 'O', 'O', 'O', 'O', 'O'], 'ner_tags': [24, 24, 24, 24, 24, 24]} ``` Example of the one record for `nested`: ```json {'id': '0', 'tokens': ['"', 'תהיה', 'נקמה', 'ו', 'בגדול', '.'], 'ner_tags': [24, 24, 24, 24, 24, 24], 'ner_tags_2': [24, 24, 24, 24, 24, 24], 'ner_tags_3': [24, 24, 24, 24, 24, 24], 'ner_tags_4': [24, 24, 24, 24, 24, 24]} ``` ## Dataset Description it's README.md of the [original repository](https://github.com/OnlpLab/NEMO-Corpus) Named Entity (NER) annotations of the Hebrew Treebank (Haaretz newspaper) corpus, including: morpheme and token level NER labels, nested mentions, and more. We publish the NEMO corpus in the TACL paper [*"Neural Modeling for Named Entities and Morphology (NEMO<sup>2</sup>)"*](https://direct.mit.edu/tacl/article/doi/10.1162/tacl_a_00404/107206/Neural-Modeling-for-Named-Entities-and-Morphology) [1], where we use it in extensive experiments and analyses, showing the importance of morphological boundaries for neural modeling of NER in morphologically rich languages. Code for these models and experiments can be found in the [NEMO code repo](https://github.com/OnlpLab/NEMO). ## Main features: 1. Morpheme, token-single and token-multi sequence labels. Morpheme labels provide exact boundaries, token-multi provide partial sub-word morphological but no exact boundaries, token-single provides only token-level information. 1. All annotations are in `BIOSE` format (`B`=Begin, `I`=Inside, `O`=Outside, `S`=Singleton, `E`=End). 1. Widely-used OntoNotes entity category set: `GPE` (geo-political entity), `PER` (person), `LOC` (location), `ORG` (organization), `FAC` (facility), `EVE` (event), `WOA` (work-of-art), `ANG` (language), `DUC` (product). 1. NEMO includes NER annotations for the two major versions of the Hebrew Treebank, UD (Universal Dependency) and SPMRL. These can be aligned to the other morphosyntactic information layers of the treebank using [bclm](https://github.com/OnlpLab/bclm) 1. We provide nested mentions. Only the first, widest, layer is used in the NEMO<sup>2</sup> paper. We invite you to take on this challenge! 1. Guidelines used for annotation are provided [here](./guidelines/). 1. Corpus was annotated by two native Hebrew speakers of academic education, and curated by the project manager. We provide the original annotations made by the annotators as well to promote work on [learning with disagreements](https://sites.google.com/view/semeval2021-task12/home). 1. Annotation was performed using [WebAnno](https://webanno.github.io/webanno/) (version 3.4.5) ## Legend for Files and Folder Structure 1. The two main [data](./data/) folders are [ud](./data/ud/) and [spmrl](./data/spmrl/), corresponding to the relevant Hebrew Treebank corpus version. 1. Both contain a `gold` folder ([spmrl/gold](./data/spmrl/gold/), [ud/gold](./data/ud/gold/)) of gold curated annotations. 1. Each `gold` folder contains files of the three input-output variants (morph, token-multi, token-single), for each of the treebank splits (train,dev,test). 1. Each `gold` folder also contains a `nested` subfolder ([spmrl/nested](./data/spmrl/gold/nested/), [ud/nested](./data/ud/gold/nested/)), which contains all layers of nested mentions (the first layer is the layer used in the non-nested files, and in the NEMO<sup>2</sup> paper [1]) 1. The `ud` folder also contains an [ab_annotators](./data/ud/ab_annotators/) folder. This folder contains the original annotations made by each annotator (named `a`, `b`), including first-layer and nested annotatations. 1. *\*UPDATE 2021-09-06\** `ud` folder now contains a [pilot_annotations](./data/ud/pilot_annotations/) folder. This folder contains the original annotations made by each annotator in our two phase pilot (phase I - sentences 1-200 of dev; phase II - sentences 201-400 of dev). ## Basic Corpus Statistics | | train | dev | test | |------------------------------| --:| --:| --:| | Sentences | 4,937 | 500 | 706 | | Tokens | 93,504 | 8,531 | 12,619 | | Morphemes | 127,031 | 11,301 | 16,828 | | All mentions | 6,282 | 499 | 932 | | Type: Person (PER) | 2,128 | 193 | 267 | | Type: Organization (ORG) | 2,043 | 119 | 408 | | Type: Geo-Political (GPE) | 1,377 | 121 | 195 | | Type: Location (LOC) | 331 | 28 | 41 | | Type: Facility (FAC) | 163 | 12 | 11 | | Type: Work-of-Art (WOA) | 114 | 9 | 6 | | Type: Event (EVE) | 57 | 12 | 0 | | Type: Product (DUC) | 36 | 2 | 3 | | Type: Language (ANG) | 33 | 3 | 1 | ## Aligned Treenbank Versions The NEMO corpus matches the treebank version of [bclm v.1.0.0](https://github.com/OnlpLab/bclm/releases/tag/v1.0.0-alpha). This version is based on the [HTB UD v2.2](https://github.com/UniversalDependencies/UD_Hebrew-HTB/releases/tag/r2.2) and the [latest SPMRL HTB version](https://github.com/OnlpLab/HebrewResources/tree/102674bb030f5836e1ab827feb63954ad7a6f8fe/HebrewTreebank/hebtb). The changes contain (but might not be limited to the following): 1. Flagged and dropped duplicate and leaking sentences (between train and test). In addition to the sentences already removed in the bclm v1.0.0 HTB version, the following duplicate sentences were dropped as well (SPMRL sentence IDs): 5438, 5444, 5445, 5446, 5448, 5449, 5450, 5451, 5453, 5459 (in the bclm dataframes, these are marked in the `duplicate_sent_id` column). To read the treebank (UD/SPMRL) in a way that matches the NEMO corpus, you can use the following: ```python import bclm dropped = [5438, 5444, 5445, 5446, 5448, 5449, 5450, 5451, 5453, 5459] spdf = bclm.read_dataframe('spmrl') # load SPMRL treebank dataframe global_dropped = [spdf[spdf.sent_id==d].global_sent_id.iat[0] for d in dropped] uddf = bclm.read_dataframe('ud') # load UD treebank dataframe uddf = uddf[(~uddf.global_sent_id.isin(global_dropped))] # remove extra duplicates spdf = spdf[(~spdf.sent_id.isin(dropped))] # remove extra duplicates # The resulting dataframes contain gold morph NER labels in the `biose_layer0`, `biose_layer1`... columns. ``` 2. The UD treebank contains many more duplicates. In this version: all sentences exist in both UD and SPMRL versions, and all sentences and tokens are aligned between UD and SPMRL. 2. Fixed numbers that were originally reversed. 2. Fixed mismatches between tokens and morphemes. 2. Added Binyan feature. 2. No individual morphemes or tokens were added or removed, only complete sentences. ## Evaluation An evaluation script is provided in the [NEMO code repo](https://github.com/OnlpLab/NEMO#evaluation) along with evaluation instructions. ## Citations ##### [1] If you use the NEMO corpus in your research, please cite the NEMO<sup>2</sup> paper: ```bibtex @article{10.1162/tacl_a_00404, author = {Bareket, Dan and Tsarfaty, Reut}, title = "{Neural Modeling for Named Entities and Morphology (NEMO2)}", journal = {Transactions of the Association for Computational Linguistics}, volume = {9}, pages = {909-928}, year = {2021}, month = {09}, abstract = "{Named Entity Recognition (NER) is a fundamental NLP task, commonly formulated as classification over a sequence of tokens. Morphologically rich languages (MRLs) pose a challenge to this basic formulation, as the boundaries of named entities do not necessarily coincide with token boundaries, rather, they respect morphological boundaries. To address NER in MRLs we then need to answer two fundamental questions, namely, what are the basic units to be labeled, and how can these units be detected and classified in realistic settings (i.e., where no gold morphology is available). We empirically investigate these questions on a novel NER benchmark, with parallel token- level and morpheme-level NER annotations, which we develop for Modern Hebrew, a morphologically rich-and-ambiguous language. Our results show that explicitly modeling morphological boundaries leads to improved NER performance, and that a novel hybrid architecture, in which NER precedes and prunes morphological decomposition, greatly outperforms the standard pipeline, where morphological decomposition strictly precedes NER, setting a new performance bar for both Hebrew NER and Hebrew morphological decomposition tasks.}", issn = {2307-387X}, doi = {10.1162/tacl_a_00404}, url = {https://doi.org/10.1162/tacl\_a\_00404}, eprint = {https://direct.mit.edu/tacl/article-pdf/doi/10.1162/tacl\_a\_00404/1962472/tacl\_a\_00404.pdf}, } ``` ##### [2] Please cite the Hebrew Treebank as well, described the following paper: ```bibtex @article{sima2001building, title={Building a tree-bank of modern Hebrew text}, author={Sima’an, Khalil and Itai, Alon and Winter, Yoad and Altman, Alon and Nativ, Noa}, journal={Traitement Automatique des Langues}, volume={42}, number={2}, pages={247--380}, year={2001}, publisher={Citeseer} } ``` ##### [3] The UD version of the Hebrew Treebank is described in: ```bibtex @inproceedings{sade-etal-2018-hebrew, title = "The {H}ebrew {U}niversal {D}ependency Treebank: Past Present and Future", author = "Sade, Shoval and Seker, Amit and Tsarfaty, Reut", booktitle = "Proceedings of the Second Workshop on Universal Dependencies ({UDW} 2018)", month = nov, year = "2018", address = "Brussels, Belgium", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W18-6016", doi = "10.18653/v1/W18-6016", pages = "133--143", abstract = "The Hebrew treebank (HTB), consisting of 6221 morpho-syntactically annotated newspaper sentences, has been the only resource for training and validating statistical parsers and taggers for Hebrew, for almost two decades now. During these decades, the HTB has gone through a trajectory of automatic and semi-automatic conversions, until arriving at its UDv2 form. In this work we manually validate the UDv2 version of the HTB, and, according to our findings, we apply scheme changes that bring the UD HTB to the same theoretical grounds as the rest of UD. Our experimental parsing results with UDv2New confirm that improving the coherence and internal consistency of the UD HTB indeed leads to improved parsing performance. At the same time, our analysis demonstrates that there is more to be done at the point of intersection of UD with other linguistic processing layers, in particular, at the points where UD interfaces external morphological and lexical resources.", } ```
[ -0.40148067474365234, -0.6540820598602295, 0.029762238264083862, 0.11701666563749313, -0.28764280676841736, -0.1565522998571396, -0.23660533130168915, -0.5411955118179321, 0.4286149740219116, 0.4966883659362793, -0.5310245752334595, -0.7641593813896179, -0.6775803565979004, 0.3211990296840...
null
null
null
null
null
null
null
null
null
null
null
null
null
ShapeNet/ShapeNetCore
ShapeNet
2023-09-20T15:05:48Z
39
19
null
[ "language:en", "license:other", "3D shapes", "region:us" ]
2023-09-20T15:05:48Z
2022-08-26T09:34:57.000Z
2022-08-26T09:34:57
--- language: - en pretty_name: ShapeNetCore tags: - 3D shapes license: other extra_gated_heading: Acknowledge license to accept the repository extra_gated_prompt: >- To request access to this ShapeNet repo, you will need to provide your **full name** (please provide both your first and last name), the name of your **advisor or the principal investigator (PI)** of your lab (in the PI/Advisor) fields, and the **school or company** that you are affiliated with (the **Affiliation** field). After requesting access to this ShapeNet repo, you will be considered for access approval. After access approval, you (the "Researcher") receive permission to use the ShapeNet database (the "Database") at Princeton University and Stanford University. In exchange for being able to join the ShapeNet community and receive such permission, Researcher hereby agrees to the following terms and conditions: Researcher shall use the Database only for non-commercial research and educational purposes. Princeton University and Stanford University make no representations or warranties regarding the Database, including but not limited to warranties of non-infringement or fitness for a particular purpose. Researcher accepts full responsibility for his or her use of the Database and shall defend and indemnify Princeton University and Stanford University, including their employees, Trustees, officers and agents, against any and all claims arising from Researcher's use of the Database, including but not limited to Researcher's use of any copies of copyrighted 3D models that he or she may create from the Database. Researcher may provide research associates and colleagues with access to the Database provided that they first agree to be bound by these terms and conditions. Princeton University and Stanford University reserve the right to terminate Researcher's access to the Database at any time. If Researcher is employed by a for-profit, commercial entity, Researcher's employer shall also be bound by these terms and conditions, and Researcher hereby represents that he or she is fully authorized to enter into this agreement on behalf of such employer. The law of the State of New Jersey shall apply to all disputes under this agreement. For access to the data, please fill in your **full name** (both first and last name), the name of your **advisor or principal investigator (PI)**, and the name of the **school or company** you are affliated with. Please actually fill out the fields (DO NOT put the word "Advisor" for PI/Advisor and the word "School" for "Affiliation", please specify the name of your advisor and the name of your school). extra_gated_fields: Name: text PI/Advisor: text Affiliation: text Purpose: text Country: text I agree to use this dataset for non-commercial use ONLY: checkbox --- This repository contains ShapeNetCore (v2), a subset of [ShapeNet](https://shapenet.org). ShapeNetCore is a densely annotated subset of ShapeNet covering 55 common object categories with ~51,300 unique 3D models. Each model in ShapeNetCore are linked to an appropriate synset in [WordNet 3.0](https://wordnet.princeton.edu/). Please see [DATA.md](DATA.md) for details about the data. If you use ShapeNet data, you agree to abide by the [ShapeNet terms of use](https://shapenet.org/terms). You are only allowed to redistribute the data to your research associates and colleagues provided that they first agree to be bound by these terms and conditions. If you use this data, please cite the main ShapeNet technical report. ``` @techreport{shapenet2015, title = {{ShapeNet: An Information-Rich 3D Model Repository}}, author = {Chang, Angel X. and Funkhouser, Thomas and Guibas, Leonidas and Hanrahan, Pat and Huang, Qixing and Li, Zimo and Savarese, Silvio and Savva, Manolis and Song, Shuran and Su, Hao and Xiao, Jianxiong and Yi, Li and Yu, Fisher}, number = {arXiv:1512.03012 [cs.GR]}, institution = {Stanford University --- Princeton University --- Toyota Technological Institute at Chicago}, year = {2015} } ``` For more information, please contact us at shapenetwebmaster@gmail.com and indicate ShapeNetCore v2 in the title of your email.
[ -0.09440137445926666, -0.21979136765003204, 0.4112263023853302, 0.07662332057952881, -0.18683123588562012, -0.4498729407787323, 0.1571209579706192, -0.6546458005905151, 0.35690563917160034, 0.6133618354797363, -0.48073774576187134, -0.6689183115959167, -0.46643468737602234, -0.016502317041...
null
null
null
null
null
null
null
null
null
null
null
null
null
truongpdd/Covid19-NER-Vietnamese
truongpdd
2022-09-09T03:03:32Z
39
0
null
[ "region:us" ]
2022-09-09T03:03:32Z
2022-09-09T03:03:24.000Z
2022-09-09T03:03:24
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
TheGreatRambler/mm2_level
TheGreatRambler
2022-11-11T08:07:34Z
39
5
null
[ "task_categories:other", "task_categories:object-detection", "task_categories:text-retrieval", "task_categories:token-classification", "task_categories:text-generation", "multilinguality:multilingual", "size_categories:10M<n<100M", "source_datasets:original", "language:multilingual", "license:cc-b...
2022-11-11T08:07:34Z
2022-09-18T20:15:00.000Z
2022-09-18T20:15:00
--- language: - multilingual license: - cc-by-nc-sa-4.0 multilinguality: - multilingual size_categories: - 10M<n<100M source_datasets: - original task_categories: - other - object-detection - text-retrieval - token-classification - text-generation task_ids: [] pretty_name: Mario Maker 2 levels tags: - text-mining --- # Mario Maker 2 levels Part of the [Mario Maker 2 Dataset Collection](https://tgrcode.com/posts/mario_maker_2_datasets) ## Dataset Description The Mario Maker 2 levels dataset consists of 26.6 million levels from Nintendo's online service totaling around 100GB of data. The dataset was created using the self-hosted [Mario Maker 2 api](https://tgrcode.com/posts/mario_maker_2_api) over the course of 1 month in February 2022. ### How to use it The Mario Maker 2 levels dataset is a very large dataset so for most use cases it is recommended to make use of the streaming API of `datasets`. You can load and iterate through the dataset with the following code: ```python from datasets import load_dataset ds = load_dataset("TheGreatRambler/mm2_level", streaming=True, split="train") print(next(iter(ds))) #OUTPUT: { 'data_id': 3000004, 'name': 'カベキック', 'description': 'カベキックをとにかくするコースです。', 'uploaded': 1561644329, 'created': 1561674240, 'gamestyle': 4, 'theme': 0, 'difficulty': 0, 'tag1': 7, 'tag2': 10, 'game_version': 1, 'world_record': 8049, 'upload_time': 193540, 'upload_attempts': 1, 'num_comments': 60, 'clear_condition': 0, 'clear_condition_magnitude': 0, 'timer': 300, 'autoscroll_speed': 0, 'clears': 1646, 'attempts': 3168, 'clear_rate': 51.957070707070706, 'plays': 1704, 'versus_matches': 80, 'coop_matches': 27, 'likes': 152, 'boos': 118, 'unique_players_and_versus': 1391, 'weekly_likes': 0, 'weekly_plays': 1, 'uploader_pid': '5218390885570355093', 'first_completer_pid': '16824392528839047213', 'record_holder_pid': '5411258160547085075', 'level_data': [some binary data], 'unk2': 0, 'unk3': [some binary data], 'unk9': 3, 'unk10': 4, 'unk11': 1, 'unk12': 1 } ``` Level data is a binary blob describing the actual level and is equivalent to the level format Nintendo uses in-game. It is gzip compressed and needs to be decompressed to be read. To read it you only need to use the provided `level.ksy` kaitai struct file and install the kaitai struct runtime to parse it into an object: ```python from datasets import load_dataset from kaitaistruct import KaitaiStream from io import BytesIO from level import Level import zlib ds = load_dataset("TheGreatRambler/mm2_level", streaming=True, split="train") level_data = next(iter(ds))["level_data"] level = Level(KaitaiStream(BytesIO(zlib.decompress(level_data)))) # NOTE level.overworld.objects is a fixed size (limitation of Kaitai struct) # must iterate by object_count or null objects will be included for i in range(level.overworld.object_count): obj = level.overworld.objects[i] print("X: %d Y: %d ID: %s" % (obj.x, obj.y, obj.id)) #OUTPUT: X: 1200 Y: 400 ID: ObjId.block X: 1360 Y: 400 ID: ObjId.block X: 1360 Y: 240 ID: ObjId.block X: 1520 Y: 240 ID: ObjId.block X: 1680 Y: 240 ID: ObjId.block X: 1680 Y: 400 ID: ObjId.block X: 1840 Y: 400 ID: ObjId.block X: 2000 Y: 400 ID: ObjId.block X: 2160 Y: 400 ID: ObjId.block X: 2320 Y: 400 ID: ObjId.block X: 2480 Y: 560 ID: ObjId.block X: 2480 Y: 720 ID: ObjId.block X: 2480 Y: 880 ID: ObjId.block X: 2160 Y: 880 ID: ObjId.block ``` Rendering the level data into an image can be done using [Toost](https://github.com/TheGreatRambler/toost) if desired. You can also download the full dataset. Note that this will download ~100GB: ```python ds = load_dataset("TheGreatRambler/mm2_level", split="train") ``` ## Data Structure ### Data Instances ```python { 'data_id': 3000004, 'name': 'カベキック', 'description': 'カベキックをとにかくするコースです。', 'uploaded': 1561644329, 'created': 1561674240, 'gamestyle': 4, 'theme': 0, 'difficulty': 0, 'tag1': 7, 'tag2': 10, 'game_version': 1, 'world_record': 8049, 'upload_time': 193540, 'upload_attempts': 1, 'num_comments': 60, 'clear_condition': 0, 'clear_condition_magnitude': 0, 'timer': 300, 'autoscroll_speed': 0, 'clears': 1646, 'attempts': 3168, 'clear_rate': 51.957070707070706, 'plays': 1704, 'versus_matches': 80, 'coop_matches': 27, 'likes': 152, 'boos': 118, 'unique_players_and_versus': 1391, 'weekly_likes': 0, 'weekly_plays': 1, 'uploader_pid': '5218390885570355093', 'first_completer_pid': '16824392528839047213', 'record_holder_pid': '5411258160547085075', 'level_data': [some binary data], 'unk2': 0, 'unk3': [some binary data], 'unk9': 3, 'unk10': 4, 'unk11': 1, 'unk12': 1 } ``` ### Data Fields |Field|Type|Description| |---|---|---| |data_id|int|Data IDs are unique identifiers, gaps in the table are due to levels deleted by users or Nintendo| |name|string|Course name| |description|string|Course description| |uploaded|int|UTC timestamp for when the level was uploaded| |created|int|Local timestamp for when the level was created| |gamestyle|int|Gamestyle, enum below| |theme|int|Theme, enum below| |difficulty|int|Difficulty, enum below| |tag1|int|The first tag, if it exists, enum below| |tag2|int|The second tag, if it exists, enum below| |game_version|int|The version of the game this level was made on| |world_record|int|The world record in milliseconds| |upload_time|int|The upload time in milliseconds| |upload_attempts|int|The number of attempts it took the uploader to upload| |num_comments|int|Number of comments, may not reflect the archived comments if there were more than 1000 comments| |clear_condition|int|Clear condition, enum below| |clear_condition_magnitude|int|If applicable, the magnitude of the clear condition| |timer|int|The timer of the level| |autoscroll_speed|int|A unit of how fast the configured autoscroll speed is for the level| |clears|int|Course clears| |attempts|int|Course attempts| |clear_rate|float|Course clear rate as a float between 0 and 1| |plays|int|Course plays, or "footprints"| |versus_matches|int|Course versus matches| |coop_matches|int|Course coop matches| |likes|int|Course likes| |boos|int|Course boos| |unique_players_and_versus|int|All unique players that have ever played this level, including the number of versus matches| |weekly_likes|int|The weekly likes on this course| |weekly_plays|int|The weekly plays on this course| |uploader_pid|string|The player ID of the uploader| |first_completer_pid|string|The player ID of the user who first cleared this course| |record_holder_pid|string|The player ID of the user who held the world record at time of archival | |level_data|bytes|The GZIP compressed decrypted level data, kaitai struct file is provided for reading| |unk2|int|Unknown| |unk3|bytes|Unknown| |unk9|int|Unknown| |unk10|int|Unknown| |unk11|int|Unknown| |unk12|int|Unknown| ### Data Splits The dataset only contains a train split. ## Enums The dataset contains some enum integer fields. This can be used to convert back to their string equivalents: ```python GameStyles = { 0: "SMB1", 1: "SMB3", 2: "SMW", 3: "NSMBU", 4: "SM3DW" } Difficulties = { 0: "Easy", 1: "Normal", 2: "Expert", 3: "Super expert" } CourseThemes = { 0: "Overworld", 1: "Underground", 2: "Castle", 3: "Airship", 4: "Underwater", 5: "Ghost house", 6: "Snow", 7: "Desert", 8: "Sky", 9: "Forest" } TagNames = { 0: "None", 1: "Standard", 2: "Puzzle solving", 3: "Speedrun", 4: "Autoscroll", 5: "Auto mario", 6: "Short and sweet", 7: "Multiplayer versus", 8: "Themed", 9: "Music", 10: "Art", 11: "Technical", 12: "Shooter", 13: "Boss battle", 14: "Single player", 15: "Link" } ClearConditions = { 137525990: "Reach the goal without landing after leaving the ground.", 199585683: "Reach the goal after defeating at least/all (n) Mechakoopa(s).", 272349836: "Reach the goal after defeating at least/all (n) Cheep Cheep(s).", 375673178: "Reach the goal without taking damage.", 426197923: "Reach the goal as Boomerang Mario.", 436833616: "Reach the goal while wearing a Shoe.", 713979835: "Reach the goal as Fire Mario.", 744927294: "Reach the goal as Frog Mario.", 751004331: "Reach the goal after defeating at least/all (n) Larry(s).", 900050759: "Reach the goal as Raccoon Mario.", 947659466: "Reach the goal after defeating at least/all (n) Blooper(s).", 976173462: "Reach the goal as Propeller Mario.", 994686866: "Reach the goal while wearing a Propeller Box.", 998904081: "Reach the goal after defeating at least/all (n) Spike(s).", 1008094897: "Reach the goal after defeating at least/all (n) Boom Boom(s).", 1051433633: "Reach the goal while holding a Koopa Shell.", 1061233896: "Reach the goal after defeating at least/all (n) Porcupuffer(s).", 1062253843: "Reach the goal after defeating at least/all (n) Charvaargh(s).", 1079889509: "Reach the goal after defeating at least/all (n) Bullet Bill(s).", 1080535886: "Reach the goal after defeating at least/all (n) Bully/Bullies.", 1151250770: "Reach the goal while wearing a Goomba Mask.", 1182464856: "Reach the goal after defeating at least/all (n) Hop-Chops.", 1219761531: "Reach the goal while holding a Red POW Block. OR Reach the goal after activating at least/all (n) Red POW Block(s).", 1221661152: "Reach the goal after defeating at least/all (n) Bob-omb(s).", 1259427138: "Reach the goal after defeating at least/all (n) Spiny/Spinies.", 1268255615: "Reach the goal after defeating at least/all (n) Bowser(s)/Meowser(s).", 1279580818: "Reach the goal after defeating at least/all (n) Ant Trooper(s).", 1283945123: "Reach the goal on a Lakitu's Cloud.", 1344044032: "Reach the goal after defeating at least/all (n) Boo(s).", 1425973877: "Reach the goal after defeating at least/all (n) Roy(s).", 1429902736: "Reach the goal while holding a Trampoline.", 1431944825: "Reach the goal after defeating at least/all (n) Morton(s).", 1446467058: "Reach the goal after defeating at least/all (n) Fish Bone(s).", 1510495760: "Reach the goal after defeating at least/all (n) Monty Mole(s).", 1656179347: "Reach the goal after picking up at least/all (n) 1-Up Mushroom(s).", 1665820273: "Reach the goal after defeating at least/all (n) Hammer Bro(s.).", 1676924210: "Reach the goal after hitting at least/all (n) P Switch(es). OR Reach the goal while holding a P Switch.", 1715960804: "Reach the goal after activating at least/all (n) POW Block(s). OR Reach the goal while holding a POW Block.", 1724036958: "Reach the goal after defeating at least/all (n) Angry Sun(s).", 1730095541: "Reach the goal after defeating at least/all (n) Pokey(s).", 1780278293: "Reach the goal as Superball Mario.", 1839897151: "Reach the goal after defeating at least/all (n) Pom Pom(s).", 1969299694: "Reach the goal after defeating at least/all (n) Peepa(s).", 2035052211: "Reach the goal after defeating at least/all (n) Lakitu(s).", 2038503215: "Reach the goal after defeating at least/all (n) Lemmy(s).", 2048033177: "Reach the goal after defeating at least/all (n) Lava Bubble(s).", 2076496776: "Reach the goal while wearing a Bullet Bill Mask.", 2089161429: "Reach the goal as Big Mario.", 2111528319: "Reach the goal as Cat Mario.", 2131209407: "Reach the goal after defeating at least/all (n) Goomba(s)/Galoomba(s).", 2139645066: "Reach the goal after defeating at least/all (n) Thwomp(s).", 2259346429: "Reach the goal after defeating at least/all (n) Iggy(s).", 2549654281: "Reach the goal while wearing a Dry Bones Shell.", 2694559007: "Reach the goal after defeating at least/all (n) Sledge Bro(s.).", 2746139466: "Reach the goal after defeating at least/all (n) Rocky Wrench(es).", 2749601092: "Reach the goal after grabbing at least/all (n) 50-Coin(s).", 2855236681: "Reach the goal as Flying Squirrel Mario.", 3036298571: "Reach the goal as Buzzy Mario.", 3074433106: "Reach the goal as Builder Mario.", 3146932243: "Reach the goal as Cape Mario.", 3174413484: "Reach the goal after defeating at least/all (n) Wendy(s).", 3206222275: "Reach the goal while wearing a Cannon Box.", 3314955857: "Reach the goal as Link.", 3342591980: "Reach the goal while you have Super Star invincibility.", 3346433512: "Reach the goal after defeating at least/all (n) Goombrat(s)/Goombud(s).", 3348058176: "Reach the goal after grabbing at least/all (n) 10-Coin(s).", 3353006607: "Reach the goal after defeating at least/all (n) Buzzy Beetle(s).", 3392229961: "Reach the goal after defeating at least/all (n) Bowser Jr.(s).", 3437308486: "Reach the goal after defeating at least/all (n) Koopa Troopa(s).", 3459144213: "Reach the goal after defeating at least/all (n) Chain Chomp(s).", 3466227835: "Reach the goal after defeating at least/all (n) Muncher(s).", 3481362698: "Reach the goal after defeating at least/all (n) Wiggler(s).", 3513732174: "Reach the goal as SMB2 Mario.", 3649647177: "Reach the goal in a Koopa Clown Car/Junior Clown Car.", 3725246406: "Reach the goal as Spiny Mario.", 3730243509: "Reach the goal in a Koopa Troopa Car.", 3748075486: "Reach the goal after defeating at least/all (n) Piranha Plant(s)/Jumping Piranha Plant(s).", 3797704544: "Reach the goal after defeating at least/all (n) Dry Bones.", 3824561269: "Reach the goal after defeating at least/all (n) Stingby/Stingbies.", 3833342952: "Reach the goal after defeating at least/all (n) Piranha Creeper(s).", 3842179831: "Reach the goal after defeating at least/all (n) Fire Piranha Plant(s).", 3874680510: "Reach the goal after breaking at least/all (n) Crates(s).", 3974581191: "Reach the goal after defeating at least/all (n) Ludwig(s).", 3977257962: "Reach the goal as Super Mario.", 4042480826: "Reach the goal after defeating at least/all (n) Skipsqueak(s).", 4116396131: "Reach the goal after grabbing at least/all (n) Coin(s).", 4117878280: "Reach the goal after defeating at least/all (n) Magikoopa(s).", 4122555074: "Reach the goal after grabbing at least/all (n) 30-Coin(s).", 4153835197: "Reach the goal as Balloon Mario.", 4172105156: "Reach the goal while wearing a Red POW Box.", 4209535561: "Reach the Goal while riding Yoshi.", 4269094462: "Reach the goal after defeating at least/all (n) Spike Top(s).", 4293354249: "Reach the goal after defeating at least/all (n) Banzai Bill(s)." } ``` <!-- TODO create detailed statistics --> ## Dataset Creation The dataset was created over a little more than a month in Febuary 2022 using the self hosted [Mario Maker 2 api](https://tgrcode.com/posts/mario_maker_2_api). As requests made to Nintendo's servers require authentication the process had to be done with upmost care and limiting download speed as to not overload the API and risk a ban. There are no intentions to create an updated release of this dataset. ## Considerations for Using the Data The dataset consists of levels from many different Mario Maker 2 players globally and as such their titles and descriptions could contain harmful language. Harmful depictions could also be present in the level data, should you choose to render it.
[ -0.5189028978347778, -0.5137307047843933, 0.24674804508686066, 0.1613769382238388, -0.02466094307601452, 0.1609397977590561, -0.048928435891866684, -0.5459641218185425, 0.44795769453048706, 0.3772777020931244, -0.729365885257721, -0.7693312168121338, -0.6606497168540955, 0.1669801771640777...
null
null
null
null
null
null
null
null
null
null
null
null
null
dgrnd4/animals-10
dgrnd4
2022-10-04T16:45:42Z
39
3
null
[ "license:other", "region:us" ]
2022-10-04T16:45:42Z
2022-10-04T16:39:10.000Z
2022-10-04T16:39:10
--- license: other ---
[ -0.1285335123538971, -0.1861683875322342, 0.6529128551483154, 0.49436232447624207, -0.19319400191307068, 0.23607441782951355, 0.36072009801864624, 0.05056373029947281, 0.5793656706809998, 0.7400146722793579, -0.650810182094574, -0.23784008622169495, -0.7102247476577759, -0.0478255338966846...
null
null
null
null
null
null
null
null
null
null
null
null
null
tomekkorbak/detoxify-pile-chunk3-4200000-4250000
tomekkorbak
2022-10-06T04:32:21Z
39
0
null
[ "region:us" ]
2022-10-06T04:32:21Z
2022-10-06T04:32:13.000Z
2022-10-06T04:32:13
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
maderix/flickr_bw_rgb
maderix
2022-10-12T15:34:25Z
39
5
null
[ "task_categories:text-to-image", "annotations_creators:machine-generated", "language_creators:other", "multilinguality:monolingual", "size_categories:n<1K", "source_datasets:N/A", "language:en", "license:cc-by-nc-sa-4.0", "region:us" ]
2022-10-12T15:34:25Z
2022-10-12T15:09:17.000Z
2022-10-12T15:09:17
--- license: cc-by-nc-sa-4.0 annotations_creators: - machine-generated language: - en language_creators: - other multilinguality: - monolingual pretty_name: 'flickr_bw_rgb' size_categories: - n<1K source_datasets: - N/A tags: [] task_categories: - text-to-image task_ids: [] --- # Dataset Card for Flickr_bw_rgb _Dataset A image-caption dataset which stores group of black and white and color images with corresponding captions mentioning the content of the image with a 'colorized photograph of' or 'Black and white photograph of' suffix. This dataset can then be used for fine-tuning image to text models.. Only a train split is provided. ## Examples "train/<filename>.jpg" : containing the images in JPEG format "train/metadata.jsonl" : Contains the metadata and the fields. Dataset columns: "file_name" "caption" ## Citation If you use this dataset, please cite it as: ``` @misc{maderix2022flickrbwrgb, author = {maderix: maderix@gmail.com}, title = {flickr_bw_rgb}, year={2022}, howpublished= {\url{https://huggingface.co/datasets/maderix/flickr_bw_rgb/}} } ```
[ -0.5219917297363281, -0.14638066291809082, -0.09657303243875504, 0.39638638496398926, -0.7074393033981323, 0.0282095056027174, 0.17267389595508575, -0.08784617483615875, 0.23621518909931183, 0.4198598563671112, -0.7453266382217407, -0.37532925605773926, -0.48996180295944214, 0.014057286083...
null
null
null
null
null
null
null
null
null
null
null
null
null
jamescalam/channel-metadata
jamescalam
2022-10-26T01:05:55Z
39
1
null
[ "task_categories:other", "annotations_creators:no-annotation", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:en", "license:afl-3.0", "youtube", "video", "video metadata", "tech", "science and tech", "region:us"...
2022-10-26T01:05:55Z
2022-10-14T05:29:45.000Z
2022-10-14T05:29:45
--- annotations_creators: - no-annotation language: - en language_creators: - found license: - afl-3.0 multilinguality: - monolingual pretty_name: Tech Channels Metadata size_categories: - 10K<n<100K source_datasets: - original tags: - youtube - video - video metadata - tech - science and tech task_categories: - other task_ids: [] --- Dataset containing video metadata from a few tech channels, i.e. * [James Briggs](https://youtube.com/c/JamesBriggs) * [Yannic Kilcher](https://www.youtube.com/c/YannicKilcher) * [sentdex](https://www.youtube.com/c/sentdex) * [Daniel Bourke](https://www.youtube.com/channel/UCr8O8l5cCX85Oem1d18EezQ) * [AI Coffee Break with Letitia](https://www.youtube.com/c/AICoffeeBreak) * [Alex Ziskind](https://youtube.com/channel/UCajiMK_CY9icRhLepS8_3ug)
[ -0.6173519492149353, -0.4831438958644867, 0.2907837927341461, 0.07249698042869568, -0.023873712867498398, 0.09017015248537064, -0.11383123695850372, 0.5779020190238953, 0.6869503855705261, 0.6531174182891846, -1.237919569015503, -0.8261027336120605, -0.9402349591255188, 0.02633868716657161...
null
null
null
null
null
null
null
null
null
null
null
null
null
helena-balabin/pereira_fMRI_passages
helena-balabin
2023-07-27T13:34:57Z
39
0
null
[ "region:us" ]
2023-07-27T13:34:57Z
2022-10-19T14:45:37.000Z
2022-10-19T14:45:37
--- dataset_info: features: - name: language_lh sequence: sequence: float64 - name: language_rh sequence: sequence: float64 - name: vision_body sequence: sequence: float64 - name: vision_face sequence: sequence: float64 - name: vision_object sequence: sequence: float64 - name: vision_scene sequence: sequence: float64 - name: vision sequence: sequence: float64 - name: dmn sequence: sequence: float64 - name: task sequence: sequence: float64 - name: all sequence: sequence: float64 - name: paragraphs sequence: string - name: topic_indices sequence: uint8 - name: permuted_paragraphs sequence: string splits: - name: train num_bytes: 1649652912 num_examples: 8 download_size: 1658872446 dataset_size: 1649652912 --- # Dataset Card for "pereira_fMRI_passages" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.4318171441555023, -0.2581547796726227, 0.5305353999137878, 0.4338859021663666, -0.2196269929409027, -0.24627931416034698, 0.18162603676319122, -0.295053094625473, 0.7362645864486694, 0.3161599636077881, -0.8109650611877441, -0.6059537529945374, -0.7699500918388367, -0.2177482545375824, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
Rosenberg/genia
Rosenberg
2022-10-23T12:08:03Z
39
2
null
[ "license:mit", "region:us" ]
2022-10-23T12:08:03Z
2022-10-23T12:07:06.000Z
2022-10-23T12:07:06
--- license: mit ---
[ -0.1285339742898941, -0.18616800010204315, 0.6529127359390259, 0.4943626821041107, -0.1931934952735901, 0.2360742688179016, 0.360720157623291, 0.05056300014257431, 0.5793654322624207, 0.7400140166282654, -0.6508105993270874, -0.23783984780311584, -0.7102248668670654, -0.047826044261455536,...
null
null
null
null
null
null
null
null
null
null
null
null
null
bond005/sberdevices_golos_100h_farfield
bond005
2022-10-27T04:23:04Z
39
0
golos
[ "task_categories:automatic-speech-recognition", "task_categories:audio-classification", "annotations_creators:expert-generated", "language_creators:crowdsourced", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:10K<n<100k", "source_datasets:extended", "language:...
2022-10-27T04:23:04Z
2022-10-26T05:04:50.000Z
2022-10-26T05:04:50
--- pretty_name: Golos annotations_creators: - expert-generated language_creators: - crowdsourced - expert-generated language: - ru license: - other multilinguality: - monolingual paperswithcode_id: golos size_categories: - 10K<n<100k source_datasets: - extended task_categories: - automatic-speech-recognition - audio-classification --- # Dataset Card for sberdevices_golos_100h_farfield ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Golos ASR corpus](https://www.openslr.org/114) - **Repository:** [Golos dataset](https://github.com/sberdevices/golos) - **Paper:** [Golos: Russian Dataset for Speech Research](https://arxiv.org/pdf/2106.10161.pdf) - **Leaderboard:** [The 🤗 Speech Bench](https://huggingface.co/spaces/huggingface/hf-speech-bench) - **Point of Contact:** [Nikolay Karpov](mailto:karpnv@gmail.com) ### Dataset Summary Sberdevices Golos is a corpus of approximately 1200 hours of 16kHz Russian speech from crowd (reading speech) and farfield (communication with smart devices) domains, prepared by SberDevices Team (Alexander Denisenko, Angelina Kovalenko, Fedor Minkin, and Nikolay Karpov). The data is derived from the crowd-sourcing platform, and has been manually annotated. Authors divide all dataset into train and test subsets. The training subset includes approximately 1000 hours. For experiments with a limited number of records, authors identified training subsets of shorter length: 100 hours, 10 hours, 1 hour, 10 minutes. This dataset is a simpler version of the above mentioned Golos: - it includes the farfield domain only (without any sound from the crowd domain); - validation split is built on the 10-hour training subset; - training split corresponds to the 100-hour training subset without sounds from the 10-hour training subset; - test split is a full original test split. ### Supported Tasks and Leaderboards - `automatic-speech-recognition`: The dataset can be used to train a model for Automatic Speech Recognition (ASR). The model is presented with an audio file and asked to transcribe the audio file to written text. The most common evaluation metric is the word error rate (WER). The task has an active Hugging Face leaderboard which can be found at https://huggingface.co/spaces/huggingface/hf-speech-bench. The leaderboard ranks models uploaded to the Hub based on their WER. ### Languages The audio is in Russian. ## Dataset Structure ### Data Instances A typical data point comprises the audio data, usually called `audio` and its transcription, called `transcription`. Any additional information about the speaker and the passage which contains the transcription is not provided. ``` {'audio': {'path': None, 'array': array([ 1.22070312e-04, 1.22070312e-04, 9.15527344e-05, ..., 6.10351562e-05, 6.10351562e-05, 3.05175781e-05]), dtype=float64), 'sampling_rate': 16000}, 'transcription': 'джой источники истории турции'} ``` ### Data Fields - audio: A dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate. Note that when accessing the audio column: `dataset[0]["audio"]` the audio file is automatically decoded and resampled to `dataset.features["audio"].sampling_rate`. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus it is important to first query the sample index before the `"audio"` column, *i.e.* `dataset[0]["audio"]` should **always** be preferred over `dataset["audio"][0]`. - transcription: the transcription of the audio file. ### Data Splits This dataset is a simpler version of the original Golos: - it includes the farfield domain only (without any sound from the crowd domain); - validation split is built on the 10-hour training subset; - training split corresponds to the 100-hour training subset without sounds from the 10-hour training subset; - test split is a full original test split. | | Train | Validation | Test | | ----- | ------ | ---------- | ----- | | examples | 9570 | 933 | 1916 | | hours | 10.3h | 1.0h | 1.4h | ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process All recorded audio files were manually annotated on the crowd-sourcing platform. #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information The dataset consists of people who have donated their voice. You agree to not attempt to determine the identity of speakers in this dataset. ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators The dataset was initially created by Alexander Denisenko, Angelina Kovalenko, Fedor Minkin, and Nikolay Karpov. ### Licensing Information [Public license with attribution and conditions reserved](https://github.com/sberdevices/golos/blob/master/license/en_us.pdf) ### Citation Information ``` @misc{karpov2021golos, author = {Karpov, Nikolay and Denisenko, Alexander and Minkin, Fedor}, title = {Golos: Russian Dataset for Speech Research}, publisher = {arXiv}, year = {2021}, url = {https://arxiv.org/abs/2106.10161} } ``` ### Contributions Thanks to [@bond005](https://github.com/bond005) for adding this dataset.
[ -0.27049878239631653, -0.40137481689453125, -0.06635230779647827, 0.2545919418334961, -0.3084430694580078, -0.06672288477420807, -0.43783310055732727, -0.39769086241722107, 0.4271133244037628, 0.3422606289386749, -0.7570516467094421, -0.6925211548805237, -0.5663592219352722, 0.066475190222...
null
null
null
null
null
null
null
null
null
null
null
null
null
SALT-NLP/spider_VALUE
SALT-NLP
2022-10-27T21:40:03Z
39
0
null
[ "region:us" ]
2022-10-27T21:40:03Z
2022-10-27T21:21:27.000Z
2022-10-27T21:21:27
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
allenai/objaverse
allenai
2023-03-31T11:05:57Z
39
251
null
[ "language:en", "license:odc-by", "arxiv:2212.08051", "region:us" ]
2023-03-31T11:05:57Z
2022-12-12T19:06:33.000Z
2022-12-12T19:06:33
--- license: odc-by language: - en viewer: false --- # Objaverse Objaverse is a Massive Dataset with 800K+ Annotated 3D Objects. More documentation is coming soon. In the meantime, please see our [paper](https://arxiv.org/abs/2212.08051) and [website](https://objaverse.allenai.org/) for additional details. # License The use of the dataset as a whole is licensed under the [ODC-By v1.0](https://opendatacommons.org/licenses/by/1-0/) license. Individual objects in Objaverse are all licensed as creative commons distributable objects, and may be under the following licenses: - [CC-BY 4.0](https://creativecommons.org/licenses/by/4.0/) - 721K objects - [CC-BY-NC 4.0](https://creativecommons.org/licenses/by-nc/4.0/) - 25K objects - [CC-BY-NC-SA 4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/) - 52K objects - [CC-BY-SA 4.0](https://creativecommons.org/licenses/by-sa/4.0/) - 16K objects - [CC0 1.0](https://creativecommons.org/publicdomain/zero/1.0/) - 3.5K objects The metadata will provide the license for each object. # Citation To cite Objaverse, please use the following BibTeX entry: ```bibtex @article{objaverse, title={Objaverse: A Universe of Annotated 3D Objects}, author={Matt Deitke and Dustin Schwenk and Jordi Salvador and Luca Weihs and Oscar Michel and Eli VanderBilt and Ludwig Schmidt and Kiana Ehsani and Aniruddha Kembhavi and Ali Farhadi}, journal={arXiv preprint arXiv:2212.08051}, year={2022} } ```
[ -0.8047175407409668, -0.9002016186714172, 0.6280531287193298, 0.04201669618487358, -0.11375489085912704, -0.3924335837364197, 0.12576957046985626, -0.8114016056060791, 0.1468675136566162, 1.0898005962371826, -0.4304323196411133, -0.5058952569961548, -0.5298776030540466, 0.40819844603538513...
null
null
null
null
null
null
null
null
null
null
null
null
null
DFKI-SLT/fabner
DFKI-SLT
2023-04-05T23:20:21Z
39
0
null
[ "task_categories:token-classification", "task_ids:named-entity-recognition", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "language:en", "license:other", "manufacturing", "2000-2020", "region:us" ]
2023-04-05T23:20:21Z
2023-01-13T13:01:38.000Z
2023-01-13T13:01:38
--- annotations_creators: - expert-generated language: - en language_creators: - found license: - other multilinguality: - monolingual pretty_name: FabNER is a manufacturing text dataset for Named Entity Recognition. size_categories: - 10K<n<100K source_datasets: [] tags: - manufacturing - 2000-2020 task_categories: - token-classification task_ids: - named-entity-recognition dataset_info: - config_name: fabner features: - name: id dtype: string - name: tokens sequence: string - name: ner_tags sequence: class_label: names: '0': O '1': B-MATE '2': I-MATE '3': O-MATE '4': E-MATE '5': S-MATE '6': B-MANP '7': I-MANP '8': O-MANP '9': E-MANP '10': S-MANP '11': B-MACEQ '12': I-MACEQ '13': O-MACEQ '14': E-MACEQ '15': S-MACEQ '16': B-APPL '17': I-APPL '18': O-APPL '19': E-APPL '20': S-APPL '21': B-FEAT '22': I-FEAT '23': O-FEAT '24': E-FEAT '25': S-FEAT '26': B-PRO '27': I-PRO '28': O-PRO '29': E-PRO '30': S-PRO '31': B-CHAR '32': I-CHAR '33': O-CHAR '34': E-CHAR '35': S-CHAR '36': B-PARA '37': I-PARA '38': O-PARA '39': E-PARA '40': S-PARA '41': B-ENAT '42': I-ENAT '43': O-ENAT '44': E-ENAT '45': S-ENAT '46': B-CONPRI '47': I-CONPRI '48': O-CONPRI '49': E-CONPRI '50': S-CONPRI '51': B-MANS '52': I-MANS '53': O-MANS '54': E-MANS '55': S-MANS '56': B-BIOP '57': I-BIOP '58': O-BIOP '59': E-BIOP '60': S-BIOP splits: - name: train num_bytes: 4394010 num_examples: 9435 - name: validation num_bytes: 934347 num_examples: 2183 - name: test num_bytes: 940136 num_examples: 2064 download_size: 3793613 dataset_size: 6268493 - config_name: fabner_bio features: - name: id dtype: string - name: tokens sequence: string - name: ner_tags sequence: class_label: names: '0': O '1': B-MATE '2': I-MATE '3': B-MANP '4': I-MANP '5': B-MACEQ '6': I-MACEQ '7': B-APPL '8': I-APPL '9': B-FEAT '10': I-FEAT '11': B-PRO '12': I-PRO '13': B-CHAR '14': I-CHAR '15': B-PARA '16': I-PARA '17': B-ENAT '18': I-ENAT '19': B-CONPRI '20': I-CONPRI '21': B-MANS '22': I-MANS '23': B-BIOP '24': I-BIOP splits: - name: train num_bytes: 4394010 num_examples: 9435 - name: validation num_bytes: 934347 num_examples: 2183 - name: test num_bytes: 940136 num_examples: 2064 download_size: 3793613 dataset_size: 6268493 - config_name: fabner_simple features: - name: id dtype: string - name: tokens sequence: string - name: ner_tags sequence: class_label: names: '0': O '1': MATE '2': MANP '3': MACEQ '4': APPL '5': FEAT '6': PRO '7': CHAR '8': PARA '9': ENAT '10': CONPRI '11': MANS '12': BIOP splits: - name: train num_bytes: 4394010 num_examples: 9435 - name: validation num_bytes: 934347 num_examples: 2183 - name: test num_bytes: 940136 num_examples: 2064 download_size: 3793613 dataset_size: 6268493 - config_name: text2tech features: - name: id dtype: string - name: tokens sequence: string - name: ner_tags sequence: class_label: names: '0': O '1': Technological System '2': Method '3': Material '4': Technical Field splits: - name: train num_bytes: 4394010 num_examples: 9435 - name: validation num_bytes: 934347 num_examples: 2183 - name: test num_bytes: 940136 num_examples: 2064 download_size: 3793613 dataset_size: 6268493 --- # Dataset Card for [Dataset Name] ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://figshare.com/articles/dataset/Dataset_NER_Manufacturing_-_FabNER_Information_Extraction_from_Manufacturing_Process_Science_Domain_Literature_Using_Named_Entity_Recognition/14782407](https://figshare.com/articles/dataset/Dataset_NER_Manufacturing_-_FabNER_Information_Extraction_from_Manufacturing_Process_Science_Domain_Literature_Using_Named_Entity_Recognition/14782407) - **Paper:** ["FabNER": information extraction from manufacturing process science domain literature using named entity recognition](https://par.nsf.gov/servlets/purl/10290810) - **Size of downloaded dataset files:** 3.79 MB - **Size of the generated dataset:** 6.27 MB ### Dataset Summary FabNER is a manufacturing text corpus of 350,000+ words for Named Entity Recognition. It is a collection of abstracts obtained from Web of Science through known journals available in manufacturing process science research. For every word, there were categories/entity labels defined namely Material (MATE), Manufacturing Process (MANP), Machine/Equipment (MACEQ), Application (APPL), Features (FEAT), Mechanical Properties (PRO), Characterization (CHAR), Parameters (PARA), Enabling Technology (ENAT), Concept/Principles (CONPRI), Manufacturing Standards (MANS) and BioMedical (BIOP). Annotation was performed in all categories along with the output tag in 'BIOES' format: B=Beginning, I-Intermediate, O=Outside, E=End, S=Single. For details about the dataset, please refer to the paper: ["FabNER": information extraction from manufacturing process science domain literature using named entity recognition](https://par.nsf.gov/servlets/purl/10290810) ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages The language in the dataset is English. ## Dataset Structure ### Data Instances - **Size of downloaded dataset files:** 3.79 MB - **Size of the generated dataset:** 6.27 MB An example of 'train' looks as follows: ```json { "id": "0", "tokens": ["Revealed", "the", "location-specific", "flow", "patterns", "and", "quantified", "the", "speeds", "of", "various", "types", "of", "flow", "."], "ner_tags": [0, 0, 0, 46, 49, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] } ``` ### Data Fields #### fabner - `id`: the instance id of this sentence, a `string` feature. - `tokens`: the list of tokens of this sentence, a `list` of `string` features. - `ner_tags`: the list of entity tags, a `list` of classification labels. ```json {"O": 0, "B-MATE": 1, "I-MATE": 2, "O-MATE": 3, "E-MATE": 4, "S-MATE": 5, "B-MANP": 6, "I-MANP": 7, "O-MANP": 8, "E-MANP": 9, "S-MANP": 10, "B-MACEQ": 11, "I-MACEQ": 12, "O-MACEQ": 13, "E-MACEQ": 14, "S-MACEQ": 15, "B-APPL": 16, "I-APPL": 17, "O-APPL": 18, "E-APPL": 19, "S-APPL": 20, "B-FEAT": 21, "I-FEAT": 22, "O-FEAT": 23, "E-FEAT": 24, "S-FEAT": 25, "B-PRO": 26, "I-PRO": 27, "O-PRO": 28, "E-PRO": 29, "S-PRO": 30, "B-CHAR": 31, "I-CHAR": 32, "O-CHAR": 33, "E-CHAR": 34, "S-CHAR": 35, "B-PARA": 36, "I-PARA": 37, "O-PARA": 38, "E-PARA": 39, "S-PARA": 40, "B-ENAT": 41, "I-ENAT": 42, "O-ENAT": 43, "E-ENAT": 44, "S-ENAT": 45, "B-CONPRI": 46, "I-CONPRI": 47, "O-CONPRI": 48, "E-CONPRI": 49, "S-CONPRI": 50, "B-MANS": 51, "I-MANS": 52, "O-MANS": 53, "E-MANS": 54, "S-MANS": 55, "B-BIOP": 56, "I-BIOP": 57, "O-BIOP": 58, "E-BIOP": 59, "S-BIOP": 60} ``` #### fabner_bio - `id`: the instance id of this sentence, a `string` feature. - `tokens`: the list of tokens of this sentence, a `list` of `string` features. - `ner_tags`: the list of entity tags, a `list` of classification labels. ```json {"O": 0, "B-MATE": 1, "I-MATE": 2, "B-MANP": 3, "I-MANP": 4, "B-MACEQ": 5, "I-MACEQ": 6, "B-APPL": 7, "I-APPL": 8, "B-FEAT": 9, "I-FEAT": 10, "B-PRO": 11, "I-PRO": 12, "B-CHAR": 13, "I-CHAR": 14, "B-PARA": 15, "I-PARA": 16, "B-ENAT": 17, "I-ENAT": 18, "B-CONPRI": 19, "I-CONPRI": 20, "B-MANS": 21, "I-MANS": 22, "B-BIOP": 23, "I-BIOP": 24} ``` #### fabner_simple - `id`: the instance id of this sentence, a `string` feature. - `tokens`: the list of tokens of this sentence, a `list` of `string` features. - `ner_tags`: the list of entity tags, a `list` of classification labels. ```json {"O": 0, "MATE": 1, "MANP": 2, "MACEQ": 3, "APPL": 4, "FEAT": 5, "PRO": 6, "CHAR": 7, "PARA": 8, "ENAT": 9, "CONPRI": 10, "MANS": 11, "BIOP": 12} ``` #### text2tech - `id`: the instance id of this sentence, a `string` feature. - `tokens`: the list of tokens of this sentence, a `list` of `string` features. - `ner_tags`: the list of entity tags, a `list` of classification labels. ```json {"O": 0, "Technological System": 1, "Method": 2, "Material": 3, "Technical Field": 4} ``` ### Data Splits | | Train | Dev | Test | |--------|-------|------|------| | fabner | 9435 | 2183 | 2064 | ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @article{DBLP:journals/jim/KumarS22, author = {Aman Kumar and Binil Starly}, title = {"FabNER": information extraction from manufacturing process science domain literature using named entity recognition}, journal = {J. Intell. Manuf.}, volume = {33}, number = {8}, pages = {2393--2407}, year = {2022}, url = {https://doi.org/10.1007/s10845-021-01807-x}, doi = {10.1007/s10845-021-01807-x}, timestamp = {Sun, 13 Nov 2022 17:52:57 +0100}, biburl = {https://dblp.org/rec/journals/jim/KumarS22.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ``` ### Contributions Thanks to [@phucdev](https://github.com/phucdev) for adding this dataset.
[ -0.5756768584251404, -0.8826786875724792, 0.2648547291755676, 0.01075439341366291, -0.15777701139450073, 0.08048886805772781, -0.23495151102542877, -0.27424541115760803, 0.610088586807251, 0.4400820732116699, -0.7336682677268982, -0.9710372090339661, -0.6461793184280396, 0.2602635025978088...
null
null
null
null
null
null
null
null
null
null
null
null
null
Aniemore/resd_annotated
Aniemore
2023-07-14T07:59:51Z
39
3
null
[ "task_categories:audio-classification", "size_categories:1K<n<10K", "language:ru", "license:mit", "voice", "emotions", "annotated", "classification", "doi:10.57967/hf/1272", "region:us" ]
2023-07-14T07:59:51Z
2023-02-15T20:00:40.000Z
2023-02-15T20:00:40
--- language: ru dataset_info: features: - name: name dtype: string - name: path dtype: string - name: speech dtype: audio - name: text dtype: string - name: emotion dtype: string splits: - name: train num_bytes: 398878916.336 num_examples: 1116 - name: test num_bytes: 96643276 num_examples: 280 download_size: 485513605 dataset_size: 495522192.336 license: mit task_categories: - audio-classification tags: - voice - emotions - annotated - classification pretty_name: RESD size_categories: - 1K<n<10K --- # Dataset Card for "resd_annotated" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.7211395502090454, -0.29460155963897705, 0.09668513387441635, 0.07449304312467575, -0.3462425768375397, -0.07418710738420486, 0.14536386728286743, -0.2936820685863495, 0.9887826442718506, 0.6588473916053772, -0.8836620450019836, -0.6517181992530823, -0.4705813527107239, 0.023981187492609...
null
null
null
null
null
null
null
null
null
null
null
null
null
mstz/student_performance
mstz
2023-04-07T14:54:45Z
39
0
null
[ "task_categories:tabular-classification", "size_categories:n<1K", "language:en", "license:cc", "student performance", "tabular_classification", "binary_classification", "region:us" ]
2023-04-07T14:54:45Z
2023-03-24T13:53:31.000Z
2023-03-24T13:53:31
--- language: - en tags: - student performance - tabular_classification - binary_classification pretty_name: Student Performance size_categories: - n<1K task_categories: - tabular-classification configs: - encoding - math - writing - reading license: cc --- # Student performance The [Student performance dataset](https://www.kaggle.com/datasets/ulrikthygepedersen/student_performances) from Kaggle. | **Configuration** | **Task** | **Description** | |-------------------|---------------------------|-----------------------------------------------------------------| | encoding | | Encoding dictionary showing original values of encoded features.| | math | Binary classification | Has the student passed the math exam? | | writing | Binary classification | Has the student passed the writing exam? | | reading | Binary classification | Has the student passed the reading exam? | # Usage ```python from datasets import load_dataset dataset = load_dataset("mstz/student_performance", "math")["train"] ``` # Features |**Feature** |**Type** | |-----------------------------------|-----------| |`is_male` |`bool` | |`ethnicity` |`string` | |`parental_level_of_education` |`int8` | |`has_standard_lunch` |`bool` | |`has_completed_preparation_test` |`bool` | |`reading_score` |`int64` | |`writing_score` |`int64` | |`math_score` |`int64` |
[ -0.21350066363811493, -0.6017742156982422, 0.12320129573345184, 0.24792751669883728, 0.23635320365428925, 0.1676507145166397, -0.23396801948547363, 0.10366969555616379, 0.21187058091163635, 0.39957311749458313, -0.5634909272193909, -0.7222284078598022, -0.617937445640564, 0.052162747830152...
null
null
null
null
null
null
null
null
null
null
null
null
null
sandhyasachidanandan/workshop-path-dataset
sandhyasachidanandan
2023-04-26T19:29:25Z
39
0
null
[ "region:us" ]
2023-04-26T19:29:25Z
2023-04-26T19:29:21.000Z
2023-04-26T19:29:21
--- dataset_info: features: - name: product_name dtype: string - name: product_description dtype: string - name: category_path dtype: string splits: - name: train num_bytes: 26820 num_examples: 100 download_size: 6735 dataset_size: 26820 --- # Dataset Card for "workshop-path-dataset" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.44064465165138245, -0.2534318268299103, 0.4245816171169281, 0.32829755544662476, -0.20217686891555786, -0.028285805135965347, 0.33760908246040344, -0.2488863170146942, 0.7732841968536377, 0.46486157178878784, -1.277385950088501, -0.8457223773002625, -0.5358489751815796, -0.4812611043453...
null
null
null
null
null
null
null
null
null
null
null
null
null
leemeng/jcommonsenseqa-v1.1
leemeng
2023-04-28T08:13:50Z
39
2
null
[ "license:cc-by-4.0", "region:us" ]
2023-04-28T08:13:50Z
2023-04-28T07:50:46.000Z
2023-04-28T07:50:46
--- license: cc-by-4.0 dataset_info: features: - name: q_id dtype: int64 - name: question dtype: string - name: choice0 dtype: string - name: choice1 dtype: string - name: choice2 dtype: string - name: choice3 dtype: string - name: choice4 dtype: string - name: label dtype: int64 splits: - name: train num_bytes: 1183829 num_examples: 8939 - name: validation num_bytes: 148293 num_examples: 1119 download_size: 887894 dataset_size: 1332122 ---
[ -0.1285339742898941, -0.18616800010204315, 0.6529127359390259, 0.4943626821041107, -0.1931934952735901, 0.2360742688179016, 0.360720157623291, 0.05056300014257431, 0.5793654322624207, 0.7400140166282654, -0.6508105993270874, -0.23783984780311584, -0.7102248668670654, -0.047826044261455536,...
null
null
null
null
null
null
null
null
null
null
null
null
null
EleutherAI/fever
EleutherAI
2023-04-30T00:09:28Z
39
1
fever
[ "task_categories:text-classification", "annotations_creators:crowdsourced", "language_creators:found", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:extended|wikipedia", "language:en", "license:cc-by-sa-3.0", "license:gpl-3.0", "knowledge-verification", "region:us"...
2023-04-30T00:09:28Z
2023-04-30T00:07:16.000Z
2023-04-30T00:07:16
--- language: - en paperswithcode_id: fever annotations_creators: - crowdsourced language_creators: - found license: - cc-by-sa-3.0 - gpl-3.0 multilinguality: - monolingual pretty_name: FEVER size_categories: - 100K<n<1M source_datasets: - extended|wikipedia task_categories: - text-classification task_ids: [] tags: - knowledge-verification dataset_info: - config_name: v1.0 features: - name: id dtype: int32 - name: label dtype: string - name: claim dtype: string - name: evidence_annotation_id dtype: int32 - name: evidence_id dtype: int32 - name: evidence_wiki_url dtype: string - name: evidence_sentence_id dtype: int32 splits: - name: train num_bytes: 24147163 num_examples: 263822 - name: dev num_bytes: 2696375 num_examples: 28625 - name: paper_dev num_bytes: 1348943 num_examples: 14475 - name: paper_test num_bytes: 1347432 num_examples: 14150 download_size: 44853972 dataset_size: 40043693 - config_name: v2.0 features: - name: id dtype: int32 - name: label dtype: string - name: claim dtype: string - name: evidence_annotation_id dtype: int32 - name: evidence_id dtype: int32 - name: evidence_wiki_url dtype: string - name: evidence_sentence_id dtype: int32 splits: - name: validation num_bytes: 306243 num_examples: 2384 download_size: 392466 dataset_size: 306243 - config_name: wiki_pages features: - name: id dtype: string - name: text dtype: string - name: lines dtype: string splits: - name: wikipedia_pages num_bytes: 7254115038 num_examples: 5416537 download_size: 1713485474 dataset_size: 7254115038 --- # Dataset Card for "fever" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://fever.ai/](https://fever.ai/) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Dataset Summary With billions of individual pages on the web providing information on almost every conceivable topic, we should have the ability to collect facts that answer almost every conceivable question. However, only a small fraction of this information is contained in structured sources (Wikidata, Freebase, etc.) – we are therefore limited by our ability to transform free-form text to structured knowledge. There is, however, another problem that has become the focus of a lot of recent research and media coverage: false information coming from unreliable sources. The FEVER workshops are a venue for work in verifiable knowledge extraction and to stimulate progress in this direction. - FEVER Dataset: FEVER (Fact Extraction and VERification) consists of 185,445 claims generated by altering sentences extracted from Wikipedia and subsequently verified without knowledge of the sentence they were derived from. The claims are classified as Supported, Refuted or NotEnoughInfo. For the first two classes, the annotators also recorded the sentence(s) forming the necessary evidence for their judgment. - FEVER 2.0 Adversarial Attacks Dataset: The FEVER 2.0 Dataset consists of 1174 claims created by the submissions of participants in the Breaker phase of the 2019 shared task. Participants (Breakers) were tasked with generating adversarial examples that induce classification errors for the existing systems. Breakers submitted a dataset of up to 1000 instances with equal number of instances for each of the three classes (Supported, Refuted NotEnoughInfo). Only novel claims (i.e. not contained in the original FEVER dataset) were considered as valid entries to the shared task. The submissions were then manually evaluated for Correctness (grammatical, appropriately labeled and meet the FEVER annotation guidelines requirements). ### Supported Tasks and Leaderboards The task is verification of textual claims against textual sources. When compared to textual entailment (TE)/natural language inference, the key difference is that in these tasks the passage to verify each claim is given, and in recent years it typically consists a single sentence, while in verification systems it is retrieved from a large set of documents in order to form the evidence. ### Languages The dataset is in English. ## Dataset Structure ### Data Instances #### v1.0 - **Size of downloaded dataset files:** 44.86 MB - **Size of the generated dataset:** 40.05 MB - **Total amount of disk used:** 84.89 MB An example of 'train' looks as follows. ``` 'claim': 'Nikolaj Coster-Waldau worked with the Fox Broadcasting Company.', 'evidence_wiki_url': 'Nikolaj_Coster-Waldau', 'label': 'SUPPORTS', 'id': 75397, 'evidence_id': 104971, 'evidence_sentence_id': 7, 'evidence_annotation_id': 92206} ``` #### v2.0 - **Size of downloaded dataset files:** 0.39 MB - **Size of the generated dataset:** 0.30 MB - **Total amount of disk used:** 0.70 MB #### wiki_pages - **Size of downloaded dataset files:** 1.71 GB - **Size of the generated dataset:** 7.25 GB - **Total amount of disk used:** 8.97 GB An example of 'wikipedia_pages' looks as follows. ``` {'text': 'The following are the football -LRB- soccer -RRB- events of the year 1928 throughout the world . ', 'lines': '0\tThe following are the football -LRB- soccer -RRB- events of the year 1928 throughout the world .\n1\t', 'id': '1928_in_association_football'} ``` ### Data Fields The data fields are the same among all splits. #### v1.0 - `id`: a `int32` feature. - `label`: a `string` feature. - `claim`: a `string` feature. - `evidence_annotation_id`: a `int32` feature. - `evidence_id`: a `int32` feature. - `evidence_wiki_url`: a `string` feature. - `evidence_sentence_id`: a `int32` feature. #### v2.0 - `id`: a `int32` feature. - `label`: a `string` feature. - `claim`: a `string` feature. - `evidence_annotation_id`: a `int32` feature. - `evidence_id`: a `int32` feature. - `evidence_wiki_url`: a `string` feature. - `evidence_sentence_id`: a `int32` feature. #### wiki_pages - `id`: a `string` feature. - `text`: a `string` feature. - `lines`: a `string` feature. ### Data Splits #### v1.0 | | train | dev | paper_dev | paper_test | |------|-------:|------:|----------:|-----------:| | v1.0 | 311431 | 37566 | 18999 | 18567 | #### v2.0 | | validation | |------|-----------:| | v2.0 | 2384 | #### wiki_pages | | wikipedia_pages | |------------|----------------:| | wiki_pages | 5416537 | ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information FEVER license: ``` These data annotations incorporate material from Wikipedia, which is licensed pursuant to the Wikipedia Copyright Policy. These annotations are made available under the license terms described on the applicable Wikipedia article pages, or, where Wikipedia license terms are unavailable, under the Creative Commons Attribution-ShareAlike License (version 3.0), available at http://creativecommons.org/licenses/by-sa/3.0/ (collectively, the “License Terms”). You may not use these files except in compliance with the applicable License Terms. ``` ### Citation Information If you use "FEVER Dataset", please cite: ```bibtex @inproceedings{Thorne18Fever, author = {Thorne, James and Vlachos, Andreas and Christodoulopoulos, Christos and Mittal, Arpit}, title = {{FEVER}: a Large-scale Dataset for Fact Extraction and {VERification}}, booktitle = {NAACL-HLT}, year = {2018} } ``` If you use "FEVER 2.0 Adversarial Attacks Dataset", please cite: ```bibtex @inproceedings{Thorne19FEVER2, author = {Thorne, James and Vlachos, Andreas and Cocarascu, Oana and Christodoulopoulos, Christos and Mittal, Arpit}, title = {The {FEVER2.0} Shared Task}, booktitle = {Proceedings of the Second Workshop on {Fact Extraction and VERification (FEVER)}}, year = {2018} } ``` ### Contributions Thanks to [@thomwolf](https://github.com/thomwolf), [@lhoestq](https://github.com/lhoestq), [@mariamabarham](https://github.com/mariamabarham), [@lewtun](https://github.com/lewtun), [@albertvillanova](https://github.com/albertvillanova) for adding this dataset.
[ -0.5471979975700378, -0.6866545081138611, 0.15270830690860748, 0.14032886922359467, -0.2059536874294281, -0.09807195514440536, -0.2811519503593445, -0.5318810343742371, 0.6165002584457397, 0.4107823073863983, -0.5871056914329529, -0.8343188762664795, -0.5745019912719727, 0.2787407338619232...
null
null
null
null
null
null
null
null
null
null
null
null
null
yuchenlin/G-PlanET
yuchenlin
2023-07-15T07:33:33Z
39
3
null
[ "task_categories:text-generation", "task_categories:table-to-text", "task_categories:table-question-answering", "language:en", "license:apache-2.0", "arxiv:2209.00465", "region:us" ]
2023-07-15T07:33:33Z
2023-05-11T00:54:50.000Z
2023-05-11T00:54:50
--- task_categories: - text-generation - table-to-text - table-question-answering language: - en license: apache-2.0 --- # Dataset Card for Dataset Name ## Dataset Description - **Homepage:** - **Repository:** - **Paper:** https://arxiv.org/abs/2209.00465 - **Leaderboard:** - **Point of Contact:** yuchenlin1995@gmail.com ### Dataset Summary This **G-PlanET** dataset is built on AI2 [ALFRED](https://leaderboard.allenai.org/alfred/submissions/get-started). ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions [More Information Needed]
[ -0.46052879095077515, -0.40509793162345886, 0.20610801875591278, 0.03483505919575691, -0.34303203225135803, 0.13329318165779114, -0.2464527040719986, -0.41952842473983765, 0.42752352356910706, 0.6362455487251282, -0.8097149133682251, -0.9344141483306885, -0.8518665432929993, -0.02077401801...
null
null
null
null
null
null
null
null
null
null
null
null
null
lighteval/pile_helm
lighteval
2023-05-16T13:02:31Z
39
0
null
[ "region:us" ]
2023-05-16T13:02:31Z
2023-05-12T10:03:48.000Z
2023-05-12T10:03:48
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
a6kme/minds14-mirror
a6kme
2023-05-13T11:42:15Z
39
0
null
[ "task_categories:automatic-speech-recognition", "task_ids:keyword-spotting", "annotations_creators:expert-generated", "annotations_creators:crowdsourced", "annotations_creators:machine-generated", "language_creators:crowdsourced", "language_creators:expert-generated", "multilinguality:multilingual", ...
2023-05-13T11:42:15Z
2023-05-13T07:56:01.000Z
2023-05-13T07:56:01
--- annotations_creators: - expert-generated - crowdsourced - machine-generated language_creators: - crowdsourced - expert-generated language: - en - fr - it - es - pt - de - nl - ru - pl - cs - ko - zh language_bcp47: - en - en-GB - en-US - en-AU - fr - it - es - pt - de - nl - ru - pl - cs - ko - zh license: - cc-by-4.0 multilinguality: - multilingual pretty_name: 'MInDS-14' size_categories: - 10K<n<100K task_categories: - automatic-speech-recognition - speech-processing task_ids: - speech-recognition - keyword-spotting --- # MInDS-14 ## Dataset Description - **Fine-Tuning script:** [pytorch/audio-classification](https://github.com/huggingface/transformers/tree/main/examples/pytorch/audio-classification) - **Paper:** [Multilingual and Cross-Lingual Intent Detection from Spoken Data](https://arxiv.org/abs/2104.08524) - **Total amount of disk used:** ca. 500 MB MINDS-14 is training and evaluation resource for intent detection task with spoken data. It covers 14 intents extracted from a commercial system in the e-banking domain, associated with spoken examples in 14 diverse language varieties. ## Example MInDS-14 can be downloaded and used as follows: ```py from datasets import load_dataset minds_14 = load_dataset("PolyAI/minds14", "fr-FR") # for French # to download all data for multi-lingual fine-tuning uncomment following line # minds_14 = load_dataset("PolyAI/all", "all") # see structure print(minds_14) # load audio sample on the fly audio_input = minds_14["train"][0]["audio"] # first decoded audio sample intent_class = minds_14["train"][0]["intent_class"] # first transcription intent = minds_14["train"].features["intent_class"].names[intent_class] # use audio_input and language_class to fine-tune your model for audio classification ``` ## Dataset Structure We show detailed information the example configurations `fr-FR` of the dataset. All other configurations have the same structure. ### Data Instances **fr-FR** - Size of downloaded dataset files: 471 MB - Size of the generated dataset: 300 KB - Total amount of disk used: 471 MB An example of a datainstance of the config `fr-FR` looks as follows: ``` { "path": "/home/patrick/.cache/huggingface/datasets/downloads/extracted/3ebe2265b2f102203be5e64fa8e533e0c6742e72268772c8ac1834c5a1a921e3/fr-FR~ADDRESS/response_4.wav", "audio": { "path": "/home/patrick/.cache/huggingface/datasets/downloads/extracted/3ebe2265b2f102203be5e64fa8e533e0c6742e72268772c8ac1834c5a1a921e3/fr-FR~ADDRESS/response_4.wav", "array": array( [0.0, 0.0, 0.0, ..., 0.0, 0.00048828, -0.00024414], dtype=float32 ), "sampling_rate": 8000, }, "transcription": "je souhaite changer mon adresse", "english_transcription": "I want to change my address", "intent_class": 1, "lang_id": 6, } ``` ### Data Fields The data fields are the same among all splits. - **path** (str): Path to the audio file - **audio** (dict): Audio object including loaded audio array, sampling rate and path ot audio - **transcription** (str): Transcription of the audio file - **english_transcription** (str): English transcription of the audio file - **intent_class** (int): Class id of intent - **lang_id** (int): Id of language ### Data Splits Every config only has the `"train"` split containing of *ca.* 600 examples. ## Dataset Creation [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information All datasets are licensed under the [Creative Commons license (CC-BY)](https://creativecommons.org/licenses/). ### Citation Information ``` @article{DBLP:journals/corr/abs-2104-08524, author = {Daniela Gerz and Pei{-}Hao Su and Razvan Kusztos and Avishek Mondal and Michal Lis and Eshan Singhal and Nikola Mrksic and Tsung{-}Hsien Wen and Ivan Vulic}, title = {Multilingual and Cross-Lingual Intent Detection from Spoken Data}, journal = {CoRR}, volume = {abs/2104.08524}, year = {2021}, url = {https://arxiv.org/abs/2104.08524}, eprinttype = {arXiv}, eprint = {2104.08524}, timestamp = {Mon, 26 Apr 2021 17:25:10 +0200}, biburl = {https://dblp.org/rec/journals/corr/abs-2104-08524.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ``` ### Contributions Thanks to [@patrickvonplaten](https://github.com/patrickvonplaten) for adding this dataset
[ -0.46738189458847046, -0.5980509519577026, 0.36973637342453003, 0.2669966518878937, -0.12211816012859344, -0.34158211946487427, -0.5432057976722717, -0.3907387852668762, 0.32248517870903015, 0.4533486068248749, -0.7834721803665161, -0.9560457468032837, -0.5506324172019958, 0.09728632122278...
null
null
null
null
null
null
null
null
null
null
null
null
null
FanChen0116/syn_few7_7100_chat_all_data_pvi
FanChen0116
2023-06-01T02:38:40Z
39
0
null
[ "region:us" ]
2023-06-01T02:38:40Z
2023-05-31T08:34:59.000Z
2023-05-31T08:34:59
--- dataset_info: features: - name: id dtype: int64 - name: tokens sequence: string - name: labels sequence: class_label: names: '0': O '1': I-time '2': B-date '3': B-last_name '4': B-people '5': I-date '6': I-people '7': I-last_name '8': I-first_name '9': B-first_name '10': B-time - name: request_slot sequence: string splits: - name: train num_bytes: 558759 num_examples: 3335 - name: validation num_bytes: 646729 num_examples: 3731 - name: test num_bytes: 646729 num_examples: 3731 download_size: 92716 dataset_size: 1852217 --- # Dataset Card for "syn_few7_7100_chat_all_data_pvi" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.31743016839027405, -0.2044840157032013, 0.18833808600902557, 0.26547691226005554, -0.2494935393333435, -0.08976920694112778, 0.22371573746204376, -0.16362687945365906, 1.0481188297271729, 0.3653820753097534, -0.8517699837684631, -0.6325012445449829, -0.35506224632263184, -0.175031125545...
null
null
null
null
null
null
null
null
null
null
null
null
null
CIRAL/ciral-corpus
CIRAL
2023-06-27T19:01:03Z
39
0
null
[ "language:ha", "language:so", "language:sw", "language:yo", "license:apache-2.0", "region:us" ]
2023-06-27T19:01:03Z
2023-06-01T20:05:01.000Z
2023-06-01T20:05:01
--- language: - ha - so - sw - yo mutilinguality: - multilingual task-categories: - text-retrieval license: apache-2.0 viewer: true --- # Dataset Summary CIRAL is a collection for cross-lingual information retrieval research across four (4) African languages. The collection comprises English queries and query-passage relevance judgements manually annotated by native speakers. This dataset stores passages which have been culled from news websites for CIRAL. ## Dataset Structure This dataset is configured by language. An example of a passage data entry is ```json { 'docid': 'DOCID#0#0', 'title': 'This is the title of a sample passage', 'text': 'This is the content of a sample passage', 'url': 'https:/\/\this-is-a-sample-url.com' } ``` ## Load Dataset An example to load the dataset ```python language = "hausa" dataset = load_dataset("ciral/ciral-corpus", language) ``` ## Citation ...
[ -0.4407525956630707, -0.4228622317314148, 0.19248168170452118, 0.28470179438591003, -0.2174438238143921, 0.08636093884706497, -0.35011234879493713, -0.1746327430009842, 0.576088011264801, 0.7195374369621277, -0.4049180746078491, -0.9527519941329956, -0.37629103660583496, 0.3297294378280639...
null
null
null
null
null
null
null
null
null
null
null
null
null
dmayhem93/agieval-sat-en-without-passage
dmayhem93
2023-06-18T17:31:43Z
39
0
null
[ "license:mit", "arxiv:2304.06364", "region:us" ]
2023-06-18T17:31:43Z
2023-06-18T12:51:12.000Z
2023-06-18T12:51:12
--- dataset_info: features: - name: query dtype: string - name: choices sequence: string - name: gold sequence: int64 splits: - name: test num_bytes: 154762 num_examples: 206 download_size: 85136 dataset_size: 154762 license: mit --- # Dataset Card for "agieval-sat-en-without-passage" Dataset taken from https://github.com/microsoft/AGIEval and processed as in that repo. MIT License Copyright (c) Microsoft Corporation. Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE @misc{zhong2023agieval, title={AGIEval: A Human-Centric Benchmark for Evaluating Foundation Models}, author={Wanjun Zhong and Ruixiang Cui and Yiduo Guo and Yaobo Liang and Shuai Lu and Yanlin Wang and Amin Saied and Weizhu Chen and Nan Duan}, year={2023}, eprint={2304.06364}, archivePrefix={arXiv}, primaryClass={cs.CL} }
[ -0.2787812352180481, -0.5386655330657959, 0.3505707085132599, 0.36788153648376465, -0.38620051741600037, -0.21414190530776978, 0.11536765843629837, -0.4942888021469116, -0.0017134598456323147, 0.6574452519416809, -0.7383328080177307, -0.5668697953224182, -0.4717787504196167, -0.06120911613...
null
null
null
null
null
null
null
null
null
null
null
null
null
dmayhem93/agieval-sat-math
dmayhem93
2023-06-18T17:32:05Z
39
6
null
[ "license:mit", "arxiv:2304.06364", "region:us" ]
2023-06-18T17:32:05Z
2023-06-18T12:51:24.000Z
2023-06-18T12:51:24
--- dataset_info: features: - name: query dtype: string - name: choices sequence: string - name: gold sequence: int64 splits: - name: test num_bytes: 110388 num_examples: 220 download_size: 57002 dataset_size: 110388 license: mit --- # Dataset Card for "agieval-sat-math" Dataset taken from https://github.com/microsoft/AGIEval and processed as in that repo. MIT License Copyright (c) Microsoft Corporation. Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE @misc{zhong2023agieval, title={AGIEval: A Human-Centric Benchmark for Evaluating Foundation Models}, author={Wanjun Zhong and Ruixiang Cui and Yiduo Guo and Yaobo Liang and Shuai Lu and Yanlin Wang and Amin Saied and Weizhu Chen and Nan Duan}, year={2023}, eprint={2304.06364}, archivePrefix={arXiv}, primaryClass={cs.CL} }
[ -0.30836835503578186, -0.6597541570663452, 0.08315117657184601, 0.35629308223724365, -0.22811061143875122, -0.06573422253131866, 0.03990735858678818, -0.24673113226890564, -0.051092468202114105, 0.5163847804069519, -0.7176838517189026, -0.484518826007843, -0.43332910537719727, -0.272025018...
null
null
null
null
null
null
null
null
null
null
null
null
null
Patt/MultiRC_TH_drop
Patt
2023-07-20T15:26:22Z
39
0
null
[ "task_categories:text-classification", "language:en", "language:th", "arxiv:1907.04307", "region:us" ]
2023-07-20T15:26:22Z
2023-06-22T13:20:37.000Z
2023-06-22T13:20:37
--- task_categories: - text-classification language: - en - th dataset_info: features: - name: paragraph dtype: string - name: paragraph_TH dtype: string - name: question dtype: string - name: question_TH dtype: string - name: answer dtype: string - name: answer_TH dtype: string - name: idx struct: - name: answer dtype: int64 - name: paragraph dtype: int64 - name: question dtype: int64 - name: label dtype: int64 - name: score_paragraph dtype: float64 - name: score_question dtype: float64 - name: score_answer dtype: float64 splits: - name: train num_bytes: 133061823 num_examples: 23520 - name: validation num_bytes: 22534453 num_examples: 4212 - name: test num_bytes: 42757726 num_examples: 8272 download_size: 5756232 dataset_size: 198354002 --- # Dataset Card for MultiRC_TH_drop ### Dataset Description This dataset is Thai translated version of [multirc](https://huggingface.co/datasets/super_glue/viewer/multirc) using google translate with [Multilingual Universal Sentence Encoder](https://arxiv.org/abs/1907.04307) to calculate score for Thai translation. The score was penalized by the length of original text compare to translated text. The row that any score < 0.66 was dropped.
[ -0.5792210102081299, -0.5747436881065369, 0.017340945079922676, 0.21569469571113586, -0.5390503406524658, 0.13692818582057953, -0.4918600618839264, -0.16126692295074463, 0.4688900113105774, 0.47556254267692566, -0.9017436504364014, -0.7712934017181396, -0.5849436521530151, 0.31541979312896...
null
null
null
null
null
null
null
null
null
null
null
null
null
yulongmannlp/dev_orig
yulongmannlp
2023-06-26T00:14:57Z
39
0
null
[ "region:us" ]
2023-06-26T00:14:57Z
2023-06-26T00:04:11.000Z
2023-06-26T00:04:11
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622264862060547, 0.43461528420448303, -0.52829909324646, 0.7012971639633179, 0.7915720343589783, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104477167129517, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
yulongmannlp/adv_para
yulongmannlp
2023-06-26T00:37:52Z
39
0
null
[ "region:us" ]
2023-06-26T00:37:52Z
2023-06-26T00:36:04.000Z
2023-06-26T00:36:04
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622264862060547, 0.43461528420448303, -0.52829909324646, 0.7012971639633179, 0.7915720343589783, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104477167129517, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
DataHammer/scimrc
DataHammer
2023-06-28T12:00:41Z
39
6
null
[ "task_categories:question-answering", "task_categories:text-generation", "size_categories:10K<n<100K", "language:en", "license:apache-2.0", "region:us" ]
2023-06-28T12:00:41Z
2023-06-28T06:15:50.000Z
2023-06-28T06:15:50
--- license: apache-2.0 task_categories: - question-answering - text-generation language: - en size_categories: - 10K<n<100K --- # Scientific Emotional Dialogue ## Dataset Description - **Homepage:** - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary This is a dataset for question answering on scientific research papers. It consists of 21.297 questions-answer-evidence pairs. ### Supported Tasks and Leaderboards - question-answering: The dataset can be used to train a model for Scientific Question Answering. Success on this task is typically measured by achieving a high F1 score. ### Languages English ## Dataset Structure ### Data Instances A typical instance in the dataset: ``` { "question": "What aim do the authors have by improving Wiki(GOLD) results?", "answer": "The aim is not to tune their model specifically on this class hierarchy. They instead aim to present a framework which can be modified easily to any domain hierarchy and has acceptable out-of-the-box performances to any fine-grained dataset.", "evidence": "The results for each class type are shown in Table TABREF19 , with some specific examples shown in Figure FIGREF18 . For the Wiki(gold) we quote the micro-averaged F-1 scores for the entire top level entity category. The total F-1 score on the OntoNotes dataset is 88%, and the total F-1 cross-validation score on the 112 class Wiki(gold) dataset is 53%. It is worth noting that one could improve Wiki(gold) results by training directly using this dataset. However, the aim is not to tune our model specifically on this class hierarchy. We instead aim to present a framework which can be modified easily to any domain hierarchy and has acceptable out-of-the-box performances to any fine-grained dataset. The results in Table TABREF19 (OntoNotes) only show the main 7 categories in OntoNotes which map to Wiki(gold) for clarity. The other categories (date, time, norp, language, ordinal, cardinal, quantity, percent, money, law) have F-1 scores between 80-90%, with the exception of time (65%)\nIt is worth noting that one could improve Wiki(GOLD) results by training directly using this dataset. However, the aim is not to tune our model specifically on this class hierarchy. We instead aim to present a framework which can be modified easily to any domain hierarchy and has acceptable out-of-the-box performances to any fine-grained dataset.", "yes_no": false } ```
[ -0.4366612136363983, -0.6903831362724304, 0.06848426163196564, 0.07479256391525269, -0.041460469365119934, 0.042942702770233154, -0.16534225642681122, -0.48465242981910706, 0.0808868333697319, 0.2540450394153595, -0.7345383763313293, -0.7663187384605408, -0.4187333881855011, 0.383084475994...
null
null
null
null
null
null
null
null
null
null
null
null
null
wu981526092/MGSD
wu981526092
2023-08-26T06:22:56Z
39
1
null
[ "task_categories:text-classification", "task_categories:token-classification", "size_categories:10K<n<100K", "language:en", "license:mit", "region:us" ]
2023-08-26T06:22:56Z
2023-06-29T18:05:33.000Z
2023-06-29T18:05:33
--- license: mit task_categories: - text-classification - token-classification language: - en size_categories: - 10K<n<100K --- # MULTI-GRAIN STEREOTYPE DATASET (MGSD) The MULTI-GRAIN STEREOTYPE DATASET (MGSD) is a comprehensive dataset designed for the research and analysis of stereotypes in natural language processing. It provides granular annotations at both the sentence and token levels, enabling various studies and applications in stereotype detection. ## Dataset Structure The dataset contains the following columns: - **text_with_marker**: Contains the original text with markers (`===`) highlighting potential stereotype tokens. - **text_no_marker**: The text without any markers, suitable for models that operate at the sentence level. - **label**: Indicates if the sentence is a stereotype, anti-stereotype, or unrelated. - **stereotype_type**: Describes the type of stereotype e.g., race, gender, profession. - **binary_class**: A binary classification of the stereotype e.g., stereotype_race, unrelated. - **multi_class**: A multi-class classification label e.g., stereotype_race, stereotype_gender. - **original_dataset**: Source of the data. ## Usage This dataset can be used to train models for various tasks: 1. **Sentence-level Stereotype Detection**: Using the `text_no_marker` column as input and `binary_label` or `multi_label` as target. 2. **Token-level Stereotype Detection**: Using the `text_with_marker` to identify the position of the token in the sentence.
[ -0.7164341807365417, -0.588527500629425, 0.4777514636516571, 0.32386961579322815, -0.2815364599227905, 0.03608746826648712, 0.10990317165851593, -0.26180145144462585, 0.4935207664966583, 0.34791335463523865, -0.6475872993469238, -1.1421780586242676, -0.7329581379890442, 0.2063559740781784,...
null
null
null
null
null
null
null
null
null
null
null
null
null
beyond/rlhf-reward-single-round-trans_chinese
beyond
2023-07-05T13:03:15Z
39
29
null
[ "region:us" ]
2023-07-05T13:03:15Z
2023-07-05T13:02:55.000Z
2023-07-05T13:02:55
--- dataset_info: features: - name: prompt dtype: string - name: chosen dtype: string - name: rejected dtype: string splits: - name: train num_bytes: 12139022 num_examples: 19862 - name: test num_bytes: 3117841 num_examples: 4996 download_size: 10699367 dataset_size: 15256863 --- # Dataset Card for "rlhf-reward-single-round-trans_chinese" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.25583842396736145, -0.3446066677570343, -0.012887611985206604, 0.37746837735176086, -0.4797392785549164, -0.06881333887577057, 0.15251226723194122, -0.24965617060661316, 0.8944696187973022, 0.5960627794265747, -1.1349353790283203, -0.8066648840904236, -0.47788572311401367, -0.1700821220...
null
null
null
null
null
null
null
null
null
null
null
null
null
pytorch-survival/gbsg_pycox
pytorch-survival
2023-07-12T01:53:49Z
39
0
null
[ "region:us" ]
2023-07-12T01:53:49Z
2023-07-12T00:32:24.000Z
2023-07-12T00:32:24
--- dataset_info: features: - name: x0 dtype: float32 - name: x1 dtype: float32 - name: x2 dtype: float32 - name: x3 dtype: float32 - name: x4 dtype: float32 - name: x5 dtype: float32 - name: x6 dtype: float32 - name: event_time dtype: float32 - name: event_indicator dtype: int32 splits: - name: train num_bytes: 80352 num_examples: 2232 download_size: 34711 dataset_size: 80352 --- # Dataset Card for "gbsg_pycox" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.5810496807098389, -0.22967544198036194, 0.21131274104118347, 0.2978135645389557, -0.12982800602912903, -0.11002959311008453, 0.23675566911697388, -0.10191719979047775, 0.7569771409034729, 0.40560218691825867, -0.6488245725631714, -0.7448176145553589, -0.5963921546936035, -0.304426372051...
null
null
null
null
null
null
null
null
null
null
null
null
null
fhirfly/medicalquestions
fhirfly
2023-10-28T17:54:21Z
39
4
null
[ "task_categories:text-classification", "size_categories:10K<n<100K", "language:en", "license:mit", "medical", "region:us" ]
2023-10-28T17:54:21Z
2023-07-13T16:46:49.000Z
2023-07-13T16:46:49
--- license: mit task_categories: - text-classification language: - en tags: - medical pretty_name: FhirFly Medical Questions size_categories: - 10K<n<100K --- # 🤗 Dataset Card: fhirfly/medicalquestions ## Dataset Overview - Dataset name: fhirfly/medicalquestions - Dataset size: 25,102 questions - Labels: 1 (medical), 0 (non-medical) - Distribution: Evenly distributed between medical and non-medical questions ## Dataset Description The fhirfly/medicalquestions dataset is a collection of 25,102 questions labeled as either medical or non-medical. The dataset aims to provide a diverse range of questions covering various medical and non-medical domains. The questions in the dataset have been manually labeled by domain experts based on the context and content of each question. Each question is assigned a label of 1 if it is determined to be a medical question and a label of 0 if it is classified as a non-medical question. ## Dataset Structure The dataset consists of a single file containing the following columns: - **Text**: The text of the question. - **Label**: The label assigned to each question, either 1 (medical) or 0 (non-medical). The questions are evenly distributed between medical and non-medical categories, ensuring a balanced dataset for training and evaluation. ## Potential Biases Efforts have been made to ensure that the dataset is representative of various medical and non-medical topics. However, it is important to acknowledge that biases may exist in the dataset due to the subjective nature of labeling questions. Biases could be present in terms of the types of questions included, the representation of certain medical conditions or non-medical topics, or the labeling process itself. It is recommended to perform thorough evaluation and analysis of the dataset to identify and mitigate potential biases during model training and deployment. Care should be taken to address any biases to ensure fair and unbiased predictions. ## Dataset Quality The fhirfly/medicalquestions dataset has undergone manual labeling by domain experts, which helps maintain a high level of quality and accuracy. However, human labeling is not entirely immune to errors or subjectivity. To ensure the quality of the dataset, a thorough review process has been conducted to minimize errors and maintain consistency in labeling. Nonetheless, it is advisable to validate and verify the data as part of your specific use case to ensure it meets your requirements. ## Data License The fhirfly/medicalquestions dataset is released under the MIT license. Please refer to the license file accompanying the dataset for more information on its usage and any restrictions that may apply. ## Dataset Citation If you use the fhirfly/medicalquestions dataset in your work, please cite it as: ``` @dataset{fhirfly/medicalquestions, title = {fhirfly/medicalquestions}, author = {fhirfly}, year = {2023}, publisher = {Hugging Face}, version = {1.0.0}, url = {https://huggingface.co/datasets/fhirfly/medicalquestions} } ```
[ -0.2944018542766571, -0.6048973798751831, 0.10124421864748001, 0.24898818135261536, -0.12482338398694992, -0.09073282033205032, 0.1320602297782898, -0.6793378591537476, 0.5940116047859192, 0.48330405354499817, -0.7879933714866638, -0.663722038269043, -0.6272023320198059, 0.1957480162382125...
null
null
null
null
null
null
null
null
null
null
null
null
null
nlpkevinl/whatsthatbook
nlpkevinl
2023-08-15T07:29:24Z
39
0
null
[ "task_categories:text-retrieval", "language:en", "license:odc-by", "arxiv:2305.15053", "region:us" ]
2023-08-15T07:29:24Z
2023-07-26T15:29:14.000Z
2023-07-26T15:29:14
--- license: odc-by task_categories: - text-retrieval language: - en pretty_name: whatsthatbook extra_gated_prompt: "To access this dataset, you agree to the terms and conditions from the GoodReads website stated here: https://www.goodreads.com/about/terms" extra_gated_fields: I agree to use to the terms and conditions: checkbox --- # Dataset Card for WhatsThatBook ## Dataset Description - **Paper: https://arxiv.org/abs/2305.15053** - **Point of Contact: k-lin@berkeley.edu** ### Dataset Summary A collection of tip-of-the-tongue queries for book searches. The dataset was curated from GoodReads community forum user queries. It seves as a training and evaluation resource for tip-of-the-tongue book queries. The user queries contain the interactions on the community forum and the documents are books with associated metadata. ### Supported Tasks and Leaderboards WhatsThatBook is intended for information retrieval tasks including but not limited to standard retrieval, using just the original query posted by the user and interactive settings, where the system asks clarification queries to narrow down the user's information needs. ### Languages The dataset is primary in English, some book descriptions may contain other languages. ## Dataset Structure ### Data Fields Data fields for WhatsThatBook queries: - `question`: Inital query posted to the community forum - `question_posted_date`: The date that the query was posted in YYYY-MM-DD format - `book_id`: ID of the gold book used for evaluation - `answers`: List of the gold book descriptions The fields for the books: - `title`: The title of the book - `author`: The author of the book - `author_url`: Link to the author page - `description` The blurb of the book that contains description of the plot or - `isbn_13`: The ISBN 13 number - `date`: String representation of the date from the book webpage - `parsed_dates`:A list of the publication date parsed out in YYYY-MM-DD format - `image_link`: original link to image - `ratings`: Total number of ratings - `reviews`: Total number of reviews - `genres`: Dictionary of genre tags to number of times tagged with that genre - `id`: ID of the book, corresponding to the query file ### Data Splits The dataset is comprised of two parts, WTB (WhatsThatBook), as well as TOMT (tip-of-my-tongue). WhatsThatBook contains standard train, dev, and test splits, and TOMT serves as an evaluation set. ## Dataset Creation ### Source Data ## Additional Information ### Dataset Curators 1. Kevin Lin, UC Berkeley, k-lin@berkeley.edu 2. Kyle Lo, Allen Institue For Artificial Intelligence, kylel@allenai.org ### Citation Information ``` @article{lin2023decomposing, title={Decomposing Complex Queries for Tip-of-the-tongue Retrieval}, author={Lin, Kevin and Lo, Kyle and Gonzalez, Joseph E and Klein, Dan}, journal={arXiv preprint arXiv:2305.15053}, year={2023} } ```
[ -0.4200352132320404, -0.5765782594680786, 0.14817215502262115, -0.04216628521680832, -0.261917382478714, -0.23704031109809875, -0.0584222637116909, -0.4352584481239319, 0.4287104904651642, 0.6613031029701233, -0.8145191073417664, -0.8793155550956726, -0.38024088740348816, 0.077982209622859...
null
null
null
null
null
null
null
null
null
null
null
null
null
morpheuslord/cve-llm-training
morpheuslord
2023-08-30T06:26:48Z
39
7
null
[ "task_categories:text-generation", "language:en", "license:apache-2.0", "code", "region:us" ]
2023-08-30T06:26:48Z
2023-08-15T06:15:00.000Z
2023-08-15T06:15:00
--- license: apache-2.0 task_categories: - text-generation language: - en tags: - code pretty_name: CVE-LLm-Dataset data_source: Custom data collected from the CVE database data_formats: JSONL --- # CVE-llm_dataset This dataset is intended to train an LLM model for an utterly CVE-focused input and output. ## Data extraction: For the data extraction, I first downloaded the CVE database from NVD lists and then loaded them using the `cve_dataset_2.py` and `cve_dataset.py` both have produce different datasets one is for llama and the other is for openai GPT. The CVE json files are mapped in this format: ``` cves: | ├─1999 | ├─0xxx | | ├─CVE-1999-0001.json | | ├─.... | | └─CVE-1999-0999.json | └─1xxx | ├─CVE-1999-1000.json | ├─.... | └─CVE-1999-1598.json └─2023 ``` The programs traverse through these folders extract the data in the files and arrange them into usable formats for the fine-tuning process. ## llama2 Model dataset: The llama2 fine-tuned dataset follows this format: ``` { "instruction": "Explain CVE-1999-0001", "input": "Explain the vulnerability: CVE-1999-0001", "output": "ip_input.c in BSD-derived TCP/IP implementations allows remote attackers to cause a denial of service (crash or hang) via crafted packets.\nAffected Products: n/a\nReferences: [{'tags': ['x_refsource_CONFIRM'], 'url': 'http://www.openbsd.org/errata23.html#tcpfix'}, {'name': '5707', 'tags': ['vdb-entry', 'x_refsource_OSVDB'], 'url': 'http://www.osvdb.org/5707'}]\nCVE State: PUBLISHED" } ``` The instruction is what we instruct the AI to do with the data provided For example we can command the AI `To take in user input analyze it and then based on what he asks returns an answer` This is also where we can add a `role` or a `personal` to the AI. The input is the user Input of the main query or data that must be processed by the AI. This is a crucial piece of information that the AI will process in order to provide an output. The output is the format that we define and tell the AI to generate answers in that format or provide that answer to the question asked.
[ -0.32709071040153503, -0.7903976440429688, 0.2792648673057556, 0.050472162663936615, -0.3236471712589264, -0.21515271067619324, 0.12558431923389435, -0.18171393871307373, 0.16840407252311707, 0.9651035666465759, -0.7433644533157349, -0.6795248985290527, -0.49047425389289856, -0.09481573104...
null
null
null
null
null
null
null
null
null
null
null
null
null
Sofoklis/vrna_dataset
Sofoklis
2023-10-10T15:58:27Z
39
0
null
[ "region:us" ]
2023-10-10T15:58:27Z
2023-09-06T13:22:48.000Z
2023-09-06T13:22:48
--- configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* - split: valid path: data/valid-* dataset_info: features: - name: name dtype: string - name: caption dtype: string - name: split_caption dtype: string - name: matrix dtype: image splits: - name: train num_bytes: 328952.0 num_examples: 80 - name: test num_bytes: 82238.0 num_examples: 20 - name: valid num_bytes: 263161.6 num_examples: 64 download_size: 549331 dataset_size: 674351.6 --- # Dataset Card for "vrna_dataset" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.5897886753082275, -0.2661450207233429, 0.13073422014713287, 0.22046460211277008, -0.35173746943473816, 0.03430962935090065, 0.4675815999507904, -0.04801258444786072, 1.0472861528396606, 0.4685933589935303, -0.6928264498710632, -0.6814365386962891, -0.4413866698741913, -0.307684928178787...
null
null
null
null
null
null
null
null
null
null
null
null
null
ahmed000000000/cybersec
ahmed000000000
2023-09-17T21:16:25Z
39
0
null
[ "region:us" ]
2023-09-17T21:16:25Z
2023-09-17T21:15:07.000Z
2023-09-17T21:15:07
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622264862060547, 0.43461528420448303, -0.52829909324646, 0.7012971639633179, 0.7915720343589783, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104477167129517, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
shossain/govreport-summarization-tokenized
shossain
2023-09-20T07:04:40Z
39
0
null
[ "region:us" ]
2023-09-20T07:04:40Z
2023-09-20T06:19:21.000Z
2023-09-20T06:19:21
--- dataset_info: features: - name: input_ids sequence: int32 - name: attention_mask sequence: int8 - name: labels sequence: int64 splits: - name: train num_bytes: 69604 num_examples: 973 download_size: 22673 dataset_size: 69604 configs: - config_name: default data_files: - split: train path: data/train-* --- # Dataset Card for "govreport-summarization-tokenized" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.4529753625392914, -0.22076362371444702, 0.08056457340717316, 0.20722338557243347, -0.40371373295783997, 0.058917880058288574, 0.11116214096546173, -0.022441258653998375, 1.12617826461792, 0.554128885269165, -0.5307256579399109, -0.8273975849151611, -0.7575028538703918, -0.22969198226928...
null
null
null
null
null
null
null
null
null
null
null
null
null
FreedomIntelligence/Arabic-preference-data-RLHF
FreedomIntelligence
2023-09-21T09:13:49Z
39
1
null
[ "region:us" ]
2023-09-21T09:13:49Z
2023-09-21T09:11:51.000Z
2023-09-21T09:11:51
Entry not found
[ -0.32276472449302673, -0.22568407654762268, 0.8622258901596069, 0.4346148371696472, -0.5282984972000122, 0.7012965679168701, 0.7915717363357544, 0.07618629932403564, 0.7746022939682007, 0.2563222646713257, -0.785281777381897, -0.22573848068714142, -0.9104482531547546, 0.5715669393539429, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
Aples/FineTune_Dataset_Aples_1K
Aples
2023-09-27T19:26:31Z
39
0
null
[ "license:mit", "region:us" ]
2023-09-27T19:26:31Z
2023-09-27T19:23:20.000Z
2023-09-27T19:23:20
--- license: mit ---
[ -0.12853392958641052, -0.18616779148578644, 0.6529127955436707, 0.49436280131340027, -0.19319361448287964, 0.23607419431209564, 0.36072003841400146, 0.050563063472509384, 0.579365611076355, 0.7400140762329102, -0.6508104205131531, -0.23783954977989197, -0.7102249264717102, -0.0478260256350...
null
null
null
null
null
null
null
null
null
null
null
null
null
yashnbx/l27b-E02-large-b10-1314-3
yashnbx
2023-09-30T16:29:18Z
39
0
null
[ "region:us" ]
2023-09-30T16:29:18Z
2023-09-30T16:28:57.000Z
2023-09-30T16:28:57
--- dataset_info: features: - name: id dtype: int64 - name: conversations list: - name: from dtype: string - name: value dtype: string splits: - name: test num_bytes: 1013014 num_examples: 146 - name: train num_bytes: 9077266 num_examples: 1314 download_size: 1662927 dataset_size: 10090280 --- # Dataset Card for "l27b-E02-large-b10-1314-3" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.6715719699859619, -0.20642130076885223, 0.3868418037891388, 0.40806037187576294, -0.2040187120437622, -0.03809454292058945, 0.11966609954833984, -0.19930621981620789, 1.109570026397705, 0.4201990067958832, -0.8161139488220215, -0.6142500042915344, -0.46714338660240173, -0.08092524856328...
null
null
null
null
null
null
null
null
null
null
null
null
null
renumics/emodb
renumics
2023-11-09T09:18:16Z
39
0
null
[ "region:us" ]
2023-11-09T09:18:16Z
2023-10-04T04:49:02.000Z
2023-10-04T04:49:02
--- dataset_info: features: - name: age dtype: float32 - name: gender dtype: class_label: names: '0': female '1': male - name: emotion dtype: class_label: names: '0': anger '1': boredom '2': disgust '3': fear '4': happiness '5': neutral '6': sadness - name: audio dtype: audio splits: - name: train num_bytes: 47623397.0 num_examples: 535 download_size: 46870260 dataset_size: 47623397.0 configs: - config_name: default data_files: - split: train path: data/train-* --- # Dataset Card for "emodb" This is a mirror for the emodb dataset. You can find the original version here: http://emodb.bilderbar.info/docu/ ## Explore this dataset You can interactively explore this dataset with Spotlight: ```python import datasets from renumics import spotlight ds = datasets.load_dataset('renumics/emodb', split='all') spotlight.show(ds) ``` ![image/gif](https://cdn-uploads.huggingface.co/production/uploads/63dd29ffaf221a78fa4ec8d1/X2L07kbwQCY9uup98YyhT.gif)
[ -0.5913172364234924, -0.7897250056266785, 0.23155692219734192, 0.040708672255277634, -0.5085639953613281, -0.02530750073492527, 0.05150545388460159, -0.0004105107218492776, 1.0236210823059082, 0.6258023381233215, -0.8178410530090332, -0.6434116959571838, -0.3034009039402008, 0.264182329177...
null
null
null
null
null
null
null
null
null
null
null
null
null
Trelis/big_patent_sample
Trelis
2023-10-09T13:32:05Z
39
1
bigpatent
[ "task_categories:summarization", "annotations_creators:no-annotation", "language_creators:found", "multilinguality:monolingual", "size_categories:n<1k", "source_datasets:big_patent", "language:en", "license:cc-by-4.0", "patent-summarization", "arxiv:1906.03741", "region:us" ]
2023-10-09T13:32:05Z
2023-10-06T12:07:45.000Z
2023-10-06T12:07:45
--- annotations_creators: - no-annotation language_creators: - found language: - en license: - cc-by-4.0 multilinguality: - monolingual size_categories: - n<1k source_datasets: - big_patent task_categories: - summarization task_ids: [] paperswithcode_id: bigpatent pretty_name: Big Patent Sample tags: - patent-summarization --- # Sampled big_patent Dataset This is a sampled big_patent dataset - sampled down for shorter fine-tunings. The data is sampled with the aim of providing an even distribution across data lengths. The distribution is quite flat up until 1 million characters in length, making the dataset good for training on lengths up to 250,000 tokens. # Dataset Card for Big Patent ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [Big Patent](https://evasharma.github.io/bigpatent/) - **Repository:** - **Paper:** [BIGPATENT: A Large-Scale Dataset for Abstractive and Coherent Summarization](https://arxiv.org/abs/1906.03741) - **Leaderboard:** - **Point of Contact:** [Lu Wang](mailto:wangluxy@umich.edu) ### Dataset Summary BIGPATENT, consisting of 1.3 million records of U.S. patent documents along with human written abstractive summaries. Each US patent application is filed under a Cooperative Patent Classification (CPC) code. There are nine such classification categories: - a: Human Necessities - b: Performing Operations; Transporting - c: Chemistry; Metallurgy - d: Textiles; Paper - e: Fixed Constructions - f: Mechanical Engineering; Lightning; Heating; Weapons; Blasting - g: Physics - h: Electricity - y: General tagging of new or cross-sectional technology Current defaults are 2.1.2 version (fix update to cased raw strings) and 'all' CPC codes: ```python from datasets import load_dataset ds = load_dataset("big_patent") # default is 'all' CPC codes ds = load_dataset("big_patent", "all") # the same as above ds = load_dataset("big_patent", "a") # only 'a' CPC codes ds = load_dataset("big_patent", codes=["a", "b"]) ``` To use 1.0.0 version (lower cased tokenized words), pass both parameters `codes` and `version`: ```python ds = load_dataset("big_patent", codes="all", version="1.0.0") ds = load_dataset("big_patent", codes="a", version="1.0.0") ds = load_dataset("big_patent", codes=["a", "b"], version="1.0.0") ``` ### Supported Tasks and Leaderboards [More Information Needed] ### Languages English ## Dataset Structure ### Data Instances Each instance contains a pair of `description` and `abstract`. `description` is extracted from the Description section of the Patent while `abstract` is extracted from the Abstract section. ``` { 'description': 'FIELD OF THE INVENTION \n [0001] This invention relates to novel calcium phosphate-coated implantable medical devices and processes of making same. The unique calcium-phosphate coated implantable medical devices minimize...', 'abstract': 'This invention relates to novel calcium phosphate-coated implantable medical devices...' } ``` ### Data Fields - `description`: detailed description of patent. - `abstract`: Patent abastract. ### Data Splits | | train | validation | test | |:----|------------------:|-------------:|-------:| | all | 1207222 | 67068 | 67072 | | a | 174134 | 9674 | 9675 | | b | 161520 | 8973 | 8974 | | c | 101042 | 5613 | 5614 | | d | 10164 | 565 | 565 | | e | 34443 | 1914 | 1914 | | f | 85568 | 4754 | 4754 | | g | 258935 | 14385 | 14386 | | h | 257019 | 14279 | 14279 | | y | 124397 | 6911 | 6911 | ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information ```bibtex @article{DBLP:journals/corr/abs-1906-03741, author = {Eva Sharma and Chen Li and Lu Wang}, title = {{BIGPATENT:} {A} Large-Scale Dataset for Abstractive and Coherent Summarization}, journal = {CoRR}, volume = {abs/1906.03741}, year = {2019}, url = {http://arxiv.org/abs/1906.03741}, eprinttype = {arXiv}, eprint = {1906.03741}, timestamp = {Wed, 26 Jun 2019 07:14:58 +0200}, biburl = {https://dblp.org/rec/journals/corr/abs-1906-03741.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ``` ### Contributions Thanks to [@mattbui](https://github.com/mattbui) for adding this dataset.
[ -0.3162391185760498, -0.37655138969421387, 0.1370326578617096, 0.7797120809555054, -0.22105616331100464, -0.009585268795490265, -0.19997930526733398, -0.38152649998664856, 0.5463310480117798, 0.3582291901111603, -0.11289192736148834, -0.7792425155639648, -0.5546914339065552, 0.055502187460...
null
null
null
null
null
null
null
null
null
null
null
null
null
dhkim123/jy_finetune_sd
dhkim123
2023-10-11T21:56:16Z
39
0
null
[ "region:us" ]
2023-10-11T21:56:16Z
2023-10-11T05:38:44.000Z
2023-10-11T05:38:44
--- configs: - config_name: default data_files: - split: train path: data/train-* dataset_info: features: - name: image dtype: image splits: - name: train num_bytes: 37668449.2 num_examples: 1300 download_size: 35715363 dataset_size: 37668449.2 --- # Dataset Card for "jy_finetune_sd" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.777073085308075, -0.3366057872772217, 0.26705724000930786, 0.13095347583293915, -0.20835894346237183, -0.23654191195964813, 0.06268839538097382, 0.01697261445224285, 0.9778012633323669, 0.5303617119789124, -0.8772494196891785, -0.6979880928993225, -0.5478960275650024, -0.287829399108886...
null
null
null
null
null
null
null
null
null
null
null
null
null
satyanshu404/trec-cast-2019
satyanshu404
2023-11-02T14:16:22Z
39
1
null
[ "arxiv:2003.13624", "region:us" ]
2023-11-02T14:16:22Z
2023-10-12T10:07:14.000Z
2023-10-12T10:07:14
# TREC Conversational Assistance Track (CAsT) There are currently few datasets appropriate for training and evaluating models for Conversational Information Seeking (CIS). The main aim of TREC CAsT is to advance research on conversational search systems. The goal of the track is to create a reusable benchmark for open-domain information centric conversational dialogues. # Year 1 (TREC 2019) * Read the [TREC 2019 Overview](https://arxiv.org/abs/2003.13624) paper. ## 2019 Data ### Topics * [Training topics] - 30 example training topics * [Training judgments] - The judgments are graded on a three point scale (2 very relevant, 1 relevant, and 0 not relevant). * [Evaluation topics]- 50 evaluation topics ### Sample of Dataset * Title: US Judicial history * Description: Judicial history in the US including key court cases and what they established. * Prompts: 1. What are the most important US Supreme Court cases? 2. What did plessy v. ferguson establish? 3. How about marbury vs madison? 4. Was it unanimous? 5. What was the implication of roe vs wade? 6. What were the main arguments? 7. What was the point of the brown v board of education? 8. What were the main arguments? 9. Why is it important today? ### Collection * The corpus is a combination of three standard TREC collections: MARCO Ranking passages, Wikipedia (TREC CAR), and News (Washington Post) * The [MS MARCO Passage Ranking collection](https://msmarco.blob.core.windows.net/msmarcoranking/collection.tar.gz) - This file only includes the passage id and passage text. For convenience, we also provide a passage id -> URL mapping file in TSV format [pid to URL file](http://boston.lti.cs.cmu.edu/vaibhav2/cast/marco_pas_url.tsv). * The [TREC CAR paragraph collection v2.0](http://trec-car.cs.unh.edu/datareleases/v2.0/paragraphCorpus.v2.0.tar.xz) * The [TREC Washington Post Corpus version 2](https://ir.nist.gov/wapo/WashingtonPost.v2.tar.gz): Note this is behind a password and requires an organizational agreement, to obtain it see: https://ir.nist.gov/wapo/ ### Document ID format * The document id format is `[collection_id_paragraph_id]` with collection id and paragraph id separated by an underscore. * The collection ids are in the set: `{MARCO, CAR, WAPO}`. * The paragraph ids are: standard provided by MARCO and CAR. For WAPO the paragraph ID is `[article_id-paragraph_index]` where the paragraph_index is the *starting from 1-based* index of the paragraph using the provided paragraph markup separated by a single dash. * Example WaPo combined document id: `[WAPO_903cc1eab726b829294d1abdd755d5ab-1]`, or CAR: `[CAR_6869dee46ab12f0f7060874f7fc7b1c57d53144a]` ## Code and tools * [TREC-CAsT Tools](https://github.com/gla-ial/trec-cast-tools) repository with code and scripts for processing data. * The tools contain scripts for parsing the collection into standard indexing formats. It also provides APIs for working with the topics (in text, json, and protocol buffer formats).
[ -0.46340087056159973, -0.7443260550498962, 0.529461681842804, 0.07836559414863586, -0.3774261176586151, 0.25244298577308655, -0.18427389860153198, -0.1561838835477829, 0.23337818682193756, 0.43045487999916077, -0.47247424721717834, -0.7557024359703064, -0.4349628686904907, -0.0713202953338...
null
null
null
null
null
null
null
null
null
null
null
null
null
erico25/aminer_title_abstract_v10
erico25
2023-10-23T20:43:49Z
39
0
null
[ "size_categories:1M<n<10M", "language:en", "region:us" ]
2023-10-23T20:43:49Z
2023-10-23T19:07:34.000Z
2023-10-23T19:07:34
--- dataset_info: features: - name: title dtype: string - name: abstract dtype: string splits: - name: train num_bytes: 2628760201 num_examples: 2548532 download_size: 0 dataset_size: 2628760201 language: - en size_categories: - 1M<n<10M --- # Dataset Card for "aminer_title_abstract_v10" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.6322628259658813, 0.04629809781908989, 0.4828195571899414, 0.17103658616542816, -0.15485644340515137, -0.12102637439966202, 0.3893585801124573, -0.224775493144989, 0.7910004258155823, 0.3142724335193634, -0.6722745299339294, -0.789857804775238, -0.7411952018737793, 0.10314728319644928, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
krasaee/nietzsche
krasaee
2023-10-26T07:47:27Z
39
0
null
[ "region:us" ]
2023-10-26T07:47:27Z
2023-10-24T16:54:51.000Z
2023-10-24T16:54:51
--- configs: - config_name: default data_files: - split: train path: data/train-* dataset_info: features: - name: text dtype: string splits: - name: train num_bytes: 9929433 num_examples: 60480 download_size: 6288420 dataset_size: 9929433 --- # Dataset Card for "nietzsche" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.79643315076828, -0.14036957919597626, 0.4908256530761719, 0.16667641699314117, -0.18795287609100342, -0.14197322726249695, 0.2122792750597, -0.19943545758724213, 0.9703095555305481, 0.3507809042930603, -0.9706792235374451, -0.7587977647781372, -0.7481629848480225, -0.38234788179397583, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
baohuynhbk14/vietnamese-guanaco-llama2-1k
baohuynhbk14
2023-10-27T08:01:25Z
39
0
null
[ "region:us" ]
2023-10-27T08:01:25Z
2023-10-27T07:39:39.000Z
2023-10-27T07:39:39
Entry not found
[ -0.32276472449302673, -0.22568407654762268, 0.8622258901596069, 0.4346148371696472, -0.5282984972000122, 0.7012965679168701, 0.7915717363357544, 0.07618629932403564, 0.7746022939682007, 0.2563222646713257, -0.785281777381897, -0.22573848068714142, -0.9104482531547546, 0.5715669393539429, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
anforsm/movie_posters-100k-torchvision
anforsm
2023-10-30T15:06:04Z
39
1
null
[ "region:us" ]
2023-10-30T15:06:04Z
2023-10-30T06:44:24.000Z
2023-10-30T06:44:24
--- configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* dataset_info: features: - name: id dtype: int64 - name: image sequence: sequence: sequence: float32 - name: title dtype: string - name: genres list: - name: id dtype: int64 - name: name dtype: string - name: overview dtype: string - name: popularity dtype: float64 - name: release_date dtype: string - name: budget dtype: int64 - name: revenue dtype: int64 - name: tagline dtype: string - name: original_language dtype: string - name: runtime dtype: int64 splits: - name: train num_bytes: 25531277848.2 num_examples: 85770 - name: test num_bytes: 2836808649.8 num_examples: 9530 download_size: 20999210873 dataset_size: 28368086498.0 --- # Dataset Card for "movie_posters-100k-torchvision" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.5597794055938721, 0.008476204238831997, 0.17086321115493774, 0.2648391127586365, -0.36732152104377747, 0.12690255045890808, 0.2852441072463989, 0.1357639729976654, 0.8266597986221313, 0.7473640441894531, -0.7474534511566162, -0.771815836429596, -0.776897668838501, -0.2516210377216339, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
singh-aditya/MACCROBAT_biomedical_ner
singh-aditya
2023-11-05T02:19:17Z
39
2
null
[ "task_categories:token-classification", "size_categories:1M<n<10M", "language:en", "license:mit", "biology", "medical", "region:us" ]
2023-11-05T02:19:17Z
2023-11-04T19:57:50.000Z
2023-11-04T19:57:50
--- license: mit task_categories: - token-classification language: - en tags: - biology - medical size_categories: - 1M<n<10M field: - data --- # MACCROBAT-biomedical-ner This data is the same data from [here](https://figshare.com/articles/dataset/MACCROBAT2018/9764942), the only difference is that it has been converted into the Huggingface dataset format. So it can be easily loaded and can be used wherever need. To convert from the orginal format to huggingface dataset format, followed the following steps (**To know in more detail look at the `create_dataset.py` file**): * Read corresponding `*.txt` and `*.ann` file. * Used `pandas` to convert the `*.ann` file into dataframe. * After converting into dataframe, did some processing and converted NER label information into: ```JSON { "text": "ner-text", "label": "ner-label", "start": 10, "end": 20 } ``` * Standard labels are converted into `B-Tag` and `I-tag`, where `B`- stands for begning of the tag and `I` - stands for inside the tag. * Finally the JSON is created and uploaded here. ## Source Data This ZIP-compressed file contains 200 source documents (in plain text, on sentence per line) and 200 annotation documents (in brat standoff format). Documents are named using PubMed document IDs, e.g. "15939911.txt" contains text from the document "A young man with palpitations and Ebstein's anomaly of the tricuspid valve" by Marcu and Donohue. Text is from PubMed Central full-text documents but has been edited to include only clinical case report details. All annotations were created manually. "MACCROBAT2020" is the second release of this dataset, following "MACCROBAT2018". The consistency and format of annotations has been improved in the newest version. ## Uses Use below snippet to load the data properly and it can be used to finetune medical based NER model with some additional processing. ```Python from datasets import load_dataset # load the data medical_ner_data = load_dataset("singh-aditya/MACCROBAT_biomedical_ner") print(medical_ner_data) ``` ``` DatasetDict({ train: Dataset({ features: ['ner_labels', 'tokens', 'full_text', 'ner_info'], num_rows: 200 }) }) ``` <!-- Address questions around how the dataset is intended to be used. --> ## Dataset Structure ``` { 'full_text': "CASE: A 28-year-old previously healthy man presented with a 6-week history of palpitations.\nThe symptoms occurred during rest, 2–3 times per week, lasted up to 30 minutes at a time and were associated with dyspnea.\nExcept for a grade 2/6 holosystolic tricuspid regurgitation murmur (best heard at the left sternal border with inspiratory accentuation), physical examination yielded unremarkable findings.\nAn electrocardiogram (ECG) revealed normal sinus rhythm and a Wolff– Parkinson– White pre-excitation pattern (Fig.1: Top), produced by a right-sided accessory pathway.\nTransthoracic echocardiography demonstrated the presence of Ebstein's anomaly of the tricuspid valve, with apical displacement of the valve and formation of an “atrialized” right ventricle (a functional unit between the right atrium and the inlet [inflow] portion of the right ventricle) (Fig.2).\nThe anterior tricuspid valve leaflet was elongated (Fig.2C, arrow), whereas the septal leaflet was rudimentary (Fig.2C, arrowhead).\nContrast echocardiography using saline revealed a patent foramen ovale with right-to-left shunting and bubbles in the left atrium (Fig.2D).\nThe patient underwent an electrophysiologic study with mapping of the accessory pathway, followed by radiofrequency ablation (interruption of the pathway using the heat generated by electromagnetic waves at the tip of an ablation catheter).\nHis post-ablation ECG showed a prolonged PR interval and an odd “second” QRS complex in leads III, aVF and V2–V4 (Fig.1Bottom), a consequence of abnormal impulse conduction in the “atrialized” right ventricle.\nThe patient reported no recurrence of palpitations at follow-up 6 months after the ablation.\n", 'ner_info': [ { 'text': '28-year-old', 'label': 'AGE', 'start': 8, 'end': 19 }, {'text': 'previously healthy', 'label': 'HISTORY', 'start': 20, 'end': 38}, {'text': 'man', 'label': 'SEX', 'start': 39, 'end': 42}, {'text': 'presented', 'label': 'CLINICAL_EVENT', 'start': 43, 'end': 52}, {'text': '6-week', 'label': 'DURATION', 'start': 60, 'end': 66}, {'text': 'palpitations', 'label': 'SIGN_SYMPTOM', 'start': 78, 'end': 90}, {'text': 'symptoms', 'label': 'COREFERENCE', 'start': 96, 'end': 104}, {'text': 'rest', 'label': 'CLINICAL_EVENT', 'start': 121, 'end': 125}, {'text': '2–3 times per week', 'label': 'FREQUENCY', 'start': 127, 'end': 145}, {'text': 'up to 30 minutes at a time', 'label': 'DETAILED_DESCRIPTION', 'start': 154, 'end': 180}, {'text': 'dyspnea', 'label': 'SIGN_SYMPTOM', 'start': 206, 'end': 213}, {'text': 'grade 2/6', 'label': 'LAB_VALUE', 'start': 228, 'end': 237}, {'text': 'holosystolic', 'label': 'DETAILED_DESCRIPTION', 'start': 238, 'end': 250}, {'text': 'tricuspid', 'label': 'BIOLOGICAL_STRUCTURE', 'start': 251, 'end': 260}, {'text': 'regurgitation murmur', 'label': 'SIGN_SYMPTOM', 'start': 261, 'end': 281}, {'text': 'left sternal border', 'label': 'BIOLOGICAL_STRUCTURE', 'start': 301, 'end': 320}, {'text': 'inspiratory accentuation', 'label': 'DETAILED_DESCRIPTION', 'start': 326, 'end': 350}, {'text': 'physical examination', 'label': 'DIAGNOSTIC_PROCEDURE', 'start': 353, 'end': 373}, {'text': 'unremarkable', 'label': 'LAB_VALUE', 'start': 382, 'end': 394}, {'text': 'electrocardiogram', 'label': 'DIAGNOSTIC_PROCEDURE', 'start': 408, 'end': 425}, {'text': 'ECG', 'label': 'DIAGNOSTIC_PROCEDURE', 'start': 427, 'end': 430}, {'text': 'normal', 'label': 'LAB_VALUE', 'start': 441, 'end': 447}, {'text': 'sinus rhythm', 'label': 'DIAGNOSTIC_PROCEDURE', 'start': 448, 'end': 460}, {'text': 'Wolff– Parkinson– White pre-excitation pattern', 'label': 'SIGN_SYMPTOM', 'start': 467, 'end': 513}, {'text': 'right-sided', 'label': 'DETAILED_DESCRIPTION', 'start': 542, 'end': 553}, {'text': 'accessory pathway', 'label': 'DISEASE_DISORDER', 'start': 554, 'end': 571}, {'text': 'Transthoracic', 'label': 'BIOLOGICAL_STRUCTURE', 'start': 573, 'end': 586}, {'text': 'echocardiography', 'label': 'DIAGNOSTIC_PROCEDURE', 'start': 587, 'end': 603}, {'text': "Ebstein's anomaly", 'label': 'DISEASE_DISORDER', 'start': 633, 'end': 650}, {'text': 'tricuspid valve', 'label': 'BIOLOGICAL_STRUCTURE', 'start': 658, 'end': 673}, {'text': 'apical displacement', 'label': 'SIGN_SYMPTOM', 'start': 680, 'end': 699}, {'text': 'valve', 'label': 'COREFERENCE', 'start': 707, 'end': 712}, {'text': 'atrialized', 'label': 'DISEASE_DISORDER', 'start': 734, 'end': 744}, {'text': 'right ventricle', 'label': 'BIOLOGICAL_STRUCTURE', 'start': 746, 'end': 761}, {'text': 'right atrium', 'label': 'BIOLOGICAL_STRUCTURE', 'start': 793, 'end': 805}, {'text': 'inlet', 'label': 'BIOLOGICAL_STRUCTURE', 'start': 814, 'end': 819}, {'text': 'right ventricle', 'label': 'BIOLOGICAL_STRUCTURE', 'start': 844, 'end': 859}, {'text': 'anterior tricuspid valve leaflet', 'label': 'BIOLOGICAL_STRUCTURE', 'start': 874, 'end': 906}, {'text': 'elongated', 'label': 'SIGN_SYMPTOM', 'start': 911, 'end': 920}, {'text': 'septal leaflet', 'label': 'BIOLOGICAL_STRUCTURE', 'start': 950, 'end': 964}, {'text': 'rudimentary', 'label': 'SIGN_SYMPTOM', 'start': 969, 'end': 980}, {'text': 'Contrast', 'label': 'DETAILED_DESCRIPTION', 'start': 1002, 'end': 1010}, {'text': 'echocardiography', 'label': 'DIAGNOSTIC_PROCEDURE', 'start': 1011, 'end': 1027}, {'text': 'using saline', 'label': 'DETAILED_DESCRIPTION', 'start': 1028, 'end': 1040}, {'text': 'patent foramen ovale', 'label': 'DISEASE_DISORDER', 'start': 1052, 'end': 1072}, {'text': 'right-to-left shunting', 'label': 'SIGN_SYMPTOM', 'start': 1078, 'end': 1100}, {'text': 'bubbles', 'label': 'SIGN_SYMPTOM', 'start': 1105, 'end': 1112}, {'text': 'left atrium', 'label': 'BIOLOGICAL_STRUCTURE', 'start': 1120, 'end': 1131}, {'text': 'electrophysiologic study', 'label': 'DIAGNOSTIC_PROCEDURE', 'start': 1167, 'end': 1191}, {'text': 'mapping', 'label': 'DIAGNOSTIC_PROCEDURE', 'start': 1197, 'end': 1204}, {'text': 'accessory pathway', 'label': 'BIOLOGICAL_STRUCTURE', 'start': 1212, 'end': 1229}, {'text': 'radiofrequency', 'label': 'DETAILED_DESCRIPTION', 'start': 1243, 'end': 1257}, {'text': 'ablation', 'label': 'THERAPEUTIC_PROCEDURE', 'start': 1258, 'end': 1266}, {'text': 'ablation catheter', 'label': 'THERAPEUTIC_PROCEDURE', 'start': 1363, 'end': 1380}, {'text': 'ECG', 'label': 'DIAGNOSTIC_PROCEDURE', 'start': 1401, 'end': 1404}, {'text': 'prolonged', 'label': 'LAB_VALUE', 'start': 1414, 'end': 1423}, {'text': 'PR interval', 'label': 'DIAGNOSTIC_PROCEDURE', 'start': 1424, 'end': 1435}, {'text': 'odd', 'label': 'LAB_VALUE', 'start': 1443, 'end': 1446}, {'text': '“second”', 'label': 'LAB_VALUE', 'start': 1447, 'end': 1455}, {'text': 'QRS complex', 'label': 'DIAGNOSTIC_PROCEDURE', 'start': 1456, 'end': 1467}, {'text': 'leads III, aVF and V2–V4', 'label': 'DETAILED_DESCRIPTION', 'start': 1471, 'end': 1495}, {'text': 'abnormal impulse conduction', 'label': 'DISEASE_DISORDER', 'start': 1528, 'end': 1555}, {'text': 'atrialized', 'label': 'DISEASE_DISORDER', 'start': 1564, 'end': 1574}, {'text': 'right ventricle', 'label': 'BIOLOGICAL_STRUCTURE', 'start': 1576, 'end': 1591}, {'text': 'palpitations', 'label': 'SIGN_SYMPTOM', 'start': 1631, 'end': 1643}, {'text': 'follow-up', 'label': 'CLINICAL_EVENT', 'start': 1647, 'end': 1656}, {'text': '6 months after', 'label': 'DATE', 'start': 1657, 'end': 1671}], 'tokens': ['CASE: A ', '28-year-old', ' ', 'previously healthy', ' ', 'man', ' ', 'presented', ' with a ', '6-week', ' history of ', 'palpitations', '.\nThe ', 'symptoms', ' occurred during ', 'rest', ', ', '2–3 times per week', ', lasted ', 'up to 30 minutes at a time', ' and were associated with ', 'dyspnea', '.\nExcept for a ', 'grade 2/6', ' ', 'holosystolic', ' ', 'tricuspid', ' ', 'regurgitation murmur', ' (best heard at the ', 'left sternal border', ' with ', 'inspiratory accentuation', '), ', 'physical examination', ' yielded ', 'unremarkable', ' findings.\nAn ', 'electrocardiogram', ' (', 'ECG', ') revealed ', 'normal', ' ', 'sinus rhythm', ' and a ', 'Wolff– Parkinson– White pre-excitation pattern', ' (Fig.1: Top), produced by a ', 'right-sided', ' ', 'accessory pathway', '.\n', 'Transthoracic', ' ', 'echocardiography', ' demonstrated the presence of ', "Ebstein's anomaly", ' of the ', 'tricuspid valve', ', with ', 'apical displacement', ' of the ', 'valve', ' and formation of an “', 'atrialized', '” ', 'right ventricle', ' (a functional unit between the ', 'right atrium', ' and the ', 'inlet', ' [inflow] portion of the ', 'right ventricle', ') (Fig.2).\nThe ', 'anterior tricuspid valve leaflet', ' was ', 'elongated', ' (Fig.2C, arrow), whereas the ', 'septal leaflet', ' was ', 'rudimentary', ' (Fig.2C, arrowhead).\n', 'Contrast', ' ', 'echocardiography', ' ', 'using saline', ' revealed a ', 'patent foramen ovale', ' with ', 'right-to-left shunting', ' and ', 'bubbles', ' in the ', 'left atrium', ' (Fig.2D).\nThe patient underwent an ', 'electrophysiologic study', ' with ', 'mapping', ' of the ', 'accessory pathway', ', followed by ', 'radiofrequency', ' ', 'ablation', ' (interruption of the pathway using the heat generated by electromagnetic waves at the tip of an ', 'ablation catheter', ').\nHis post-ablation ', 'ECG', ' showed a ', 'prolonged', ' ', 'PR interval', ' and an ', 'odd', ' ', '“second”', ' ', 'QRS complex', ' in ', 'leads III, aVF and V2–V4', ' (Fig.1Bottom), a consequence of ', 'abnormal impulse conduction', ' in the “', 'atrialized', '” ', 'right ventricle', '.\nThe patient reported no recurrence of ', 'palpitations', ' at ', 'follow-up', ' ', '6 months after', ' the ablation.\n'], 'ner_labels': [0, 5, 0, 39, 0, 65, 0, 13, 0, 32, 0, 69, 0, 18, 0, 13, 0, 35, 0, 22, 0, 69, 0, 42, 0, 22, 0, 12, 0, 69, 0, 12, 0, 22, 0, 24, 0, 42, 0, 24, 0, 24, 0, 42, 0, 24, 0, 69, 0, 22, 0, 26, 0, 12, 0, 24, 0, 26, 0, 12, 0, 69, 0, 18, 0, 26, 0, 12, 0, 12, 0, 12, 0, 12, 0, 12, 0, 69, 0, 12, 0, 69, 0, 22, 0, 24, 0, 22, 0, 26, 0, 69, 0, 69, 0, 12, 0, 24, 0, 24, 0, 12, 0, 22, 0, 75, 0, 75, 0, 24, 0, 42, 0, 24, 0, 42, 0, 42, 0, 24, 0, 22, 0, 26, 0, 26, 0, 12, 0, 69, 0, 13, 0, 19, 0]} ``` ## NER-Lables ```Python NER_lables = [ "O", "B-ACTIVITY", "I-ACTIVITY", "I-ADMINISTRATION", "B-ADMINISTRATION", "B-AGE", "I-AGE", "I-AREA", "B-AREA", "B-BIOLOGICAL_ATTRIBUTE", "I-BIOLOGICAL_ATTRIBUTE", "I-BIOLOGICAL_STRUCTURE", "B-BIOLOGICAL_STRUCTURE", "B-CLINICAL_EVENT", "I-CLINICAL_EVENT", "B-COLOR", "I-COLOR", "I-COREFERENCE", "B-COREFERENCE", "B-DATE", "I-DATE", "I-DETAILED_DESCRIPTION", "B-DETAILED_DESCRIPTION", "I-DIAGNOSTIC_PROCEDURE", "B-DIAGNOSTIC_PROCEDURE", "I-DISEASE_DISORDER", "B-DISEASE_DISORDER", "B-DISTANCE", "I-DISTANCE", "B-DOSAGE", "I-DOSAGE", "I-DURATION", "B-DURATION", "I-FAMILY_HISTORY", "B-FAMILY_HISTORY", "B-FREQUENCY", "I-FREQUENCY", "I-HEIGHT", "B-HEIGHT", "B-HISTORY", "I-HISTORY", "I-LAB_VALUE", "B-LAB_VALUE", "I-MASS", "B-MASS", "I-MEDICATION", "B-MEDICATION", "I-NONBIOLOGICAL_LOCATION", "B-NONBIOLOGICAL_LOCATION", "I-OCCUPATION", "B-OCCUPATION", "B-OTHER_ENTITY", "I-OTHER_ENTITY", "B-OTHER_EVENT", "I-OTHER_EVENT", "I-OUTCOME", "B-OUTCOME", "I-PERSONAL_BACKGROUND", "B-PERSONAL_BACKGROUND", "B-QUALITATIVE_CONCEPT", "I-QUALITATIVE_CONCEPT", "I-QUANTITATIVE_CONCEPT", "B-QUANTITATIVE_CONCEPT", "B-SEVERITY", "I-SEVERITY", "B-SEX", "I-SEX", "B-SHAPE", "I-SHAPE", "B-SIGN_SYMPTOM", "I-SIGN_SYMPTOM", "B-SUBJECT", "I-SUBJECT", "B-TEXTURE", "I-TEXTURE", "B-THERAPEUTIC_PROCEDURE", "I-THERAPEUTIC_PROCEDURE", "I-TIME", "B-TIME", "B-VOLUME", "I-VOLUME", "I-WEIGHT", "B-WEIGHT", ] ``` **BibTeX:** ```JSON { article= Caufield2020, author = "J. Harry Caufield", title = "{MACCROBAT}", year = "2020", month = "1", url = "https://figshare.com/articles/dataset/MACCROBAT2018/9764942", doi = "10.6084/m9.figshare.9764942.v2" } ```
[ -0.7240431308746338, -0.34267812967300415, 0.24641256034374237, 0.24626493453979492, -0.5149473547935486, -0.07266843318939209, 0.18726283311843872, -0.08830682188272476, 0.7281467914581299, 0.49851158261299133, -0.6854816675186157, -1.0104631185531616, -0.32193148136138916, 0.308141082525...
null
null
null
null
null
null
null
null
null
null
null
null
null
ktrinh38/eva-pix2pix
ktrinh38
2023-11-04T20:33:49Z
39
0
null
[ "region:us" ]
2023-11-04T20:33:49Z
2023-11-04T20:33:10.000Z
2023-11-04T20:33:10
--- dataset_info: features: - name: input_image dtype: image - name: edit_prompt dtype: string - name: edited_image dtype: image splits: - name: train num_bytes: 712910099.55 num_examples: 4291 download_size: 337563830 dataset_size: 712910099.55 configs: - config_name: default data_files: - split: train path: data/train-* --- # Dataset Card for "eva-pix2pix" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.603121280670166, -0.25258252024650574, 0.34406232833862305, 0.1643005907535553, -0.10537345707416534, 0.10179026424884796, 0.25309184193611145, -0.14049184322357178, 0.8131775856018066, 0.6203403472900391, -0.743351936340332, -0.5905857086181641, -0.578197181224823, -0.36665433645248413...
null
null
null
null
null
null
null
null
null
null
null
null
null
ShrinivasSK/hi_te_1
ShrinivasSK
2023-11-06T19:22:07Z
39
0
null
[ "region:us" ]
2023-11-06T19:22:07Z
2023-11-06T19:21:57.000Z
2023-11-06T19:21:57
--- dataset_info: features: - name: source dtype: string - name: target dtype: string splits: - name: train num_bytes: 5287422.6 num_examples: 18000 - name: test num_bytes: 587491.4 num_examples: 2000 download_size: 2682481 dataset_size: 5874914.0 --- # Dataset Card for "hi_te_1" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.6980561017990112, -0.4967464208602905, 0.07140454649925232, 0.21622030436992645, -0.18666799366474152, -0.2524581551551819, 0.33228743076324463, -0.18986739218235016, 1.09060800075531, 0.5998806953430176, -0.7905035614967346, -0.8799585103988647, -0.6054392457008362, -0.2538867294788360...
null
null
null
null
null
null
null
null
null
null
null
null
null
epptt/erukaLabels
epptt
2023-11-16T05:51:08Z
39
0
null
[ "region:us" ]
2023-11-16T05:51:08Z
2023-11-06T23:07:46.000Z
2023-11-06T23:07:46
--- configs: - config_name: default data_files: - split: train path: "train.json" --- # Dataset Card for Dataset Name <!-- Provide a quick summary of the dataset. --> This dataset card aims to be a base template for new datasets. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/datasetcard_template.md?plain=1). ## Dataset Details ### Dataset Description <!-- Provide a longer summary of what this dataset is. --> - **Curated by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] ### Dataset Sources [optional] <!-- Provide the basic links for the dataset. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the dataset is intended to be used. --> ### Direct Use <!-- This section describes suitable use cases for the dataset. --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. --> [More Information Needed] ## Dataset Structure <!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. --> [More Information Needed] ## Dataset Creation ### Curation Rationale <!-- Motivation for the creation of this dataset. --> [More Information Needed] ### Source Data <!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). --> #### Data Collection and Processing <!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. --> [More Information Needed] #### Who are the source data producers? <!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. --> [More Information Needed] ### Annotations [optional] <!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. --> #### Annotation process <!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. --> [More Information Needed] #### Who are the annotators? <!-- This section describes the people or systems who created the annotations. --> [More Information Needed] #### Personal and Sensitive Information <!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations. ## Citation [optional] <!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Dataset Card Authors [optional] [More Information Needed] ## Dataset Card Contact [More Information Needed]
[ -0.5322356224060059, -0.5534716844558716, 0.1290130317211151, 0.23470574617385864, -0.39626216888427734, -0.11762470752000809, -0.03545304760336876, -0.6389272212982178, 0.5699821710586548, 0.7838326096534729, -0.7834625840187073, -0.917327344417572, -0.55633145570755, 0.13078095018863678,...
null
null
null
null
null
null
null
null
null
null
null
null
null
stas/openwebtext-synthetic-testing
stas
2023-11-14T07:31:20Z
39
3
null
[ "license:apache-2.0", "region:us" ]
2023-11-14T07:31:20Z
2023-11-14T07:30:30.000Z
2023-11-14T07:30:30
--- license: apache-2.0 --- Using 10 records from [openwebtext-10k](https://huggingface.co/datasets/stas/openwebtext-10k) this dataset is written for very fast testing and can produce a repeat of these 10 records in a form of 1, 2, 3, 4, 5, 10, 100, 300 or 1k records splits, e.g.: ``` $ python -c 'from datasets import load_dataset; \ ds=load_dataset("stas/openwebtext-synthetic-testing", split="10.repeat"); print(len(ds))' 10 $ python -c 'from datasets import load_dataset; \ ds=load_dataset("stas/openwebtext-synthetic-testing", split="1k.repeat"); print(len(ds))' 1000 ``` Each record is just a single `text` record of several paragraphs long - web articles. As this is used for very fast functional testing on CI there is no `train` or `validation` splits, you can just repeat the same records.
[ -0.6987168788909912, -0.5949839949607849, 0.05590592697262764, 0.6328563690185547, -0.251590371131897, -0.2100653052330017, -0.26669326424598694, -0.2642074525356293, 0.28601399064064026, 0.21166633069515228, -0.5344089269638062, -0.31489768624305725, 0.04629320278763771, 0.475589871406555...
null
null
null
null
null
null
null
null
null
null
null
null
null
marvy/book-covers
marvy
2023-11-19T16:24:11Z
39
0
null
[ "region:us" ]
2023-11-19T16:24:11Z
2023-11-19T16:22:43.000Z
2023-11-19T16:22:43
--- dataset_info: features: - name: image dtype: image - name: title dtype: string splits: - name: train num_bytes: 286874817.68 num_examples: 32581 download_size: 283302050 dataset_size: 286874817.68 configs: - config_name: default data_files: - split: train path: data/train-* ---
[ -0.1285335123538971, -0.1861683875322342, 0.6529128551483154, 0.49436232447624207, -0.19319400191307068, 0.23607441782951355, 0.36072009801864624, 0.05056373029947281, 0.5793656706809998, 0.7400146722793579, -0.650810182094574, -0.23784008622169495, -0.7102247476577759, -0.0478255338966846...
null
null
null
null
null
null
null
null
null
null
null
null
null
doubledsbv/german-prefs_v4_prepared
doubledsbv
2023-11-20T09:23:48Z
39
0
null
[ "region:us" ]
2023-11-20T09:23:48Z
2023-11-20T09:23:39.000Z
2023-11-20T09:23:39
--- dataset_info: features: - name: prompt dtype: string - name: chosen dtype: string - name: rejected dtype: string - name: __index_level_0__ dtype: int64 splits: - name: train num_bytes: 207351857 num_examples: 57380 - name: test num_bytes: 6452941 num_examples: 1775 download_size: 120828622 dataset_size: 213804798 configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* ---
[ -0.1285335123538971, -0.1861683875322342, 0.6529128551483154, 0.49436232447624207, -0.19319400191307068, 0.23607441782951355, 0.36072009801864624, 0.05056373029947281, 0.5793656706809998, 0.7400146722793579, -0.650810182094574, -0.23784008622169495, -0.7102247476577759, -0.0478255338966846...
null
null
null
null
null
null
null
null
null
null
null
null
null
mrbmaryam/zephyr_train_2500
mrbmaryam
2023-11-20T21:26:26Z
39
0
null
[ "region:us" ]
2023-11-20T21:26:26Z
2023-11-20T21:26:09.000Z
2023-11-20T21:26:09
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622264862060547, 0.43461528420448303, -0.52829909324646, 0.7012971639633179, 0.7915720343589783, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104477167129517, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
barbaroo/Sprotin_parallel
barbaroo
2023-11-21T13:30:57Z
39
0
null
[ "region:us" ]
2023-11-21T13:30:57Z
2023-11-21T13:30:24.000Z
2023-11-21T13:30:24
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
result-kand2-sdxl-wuerst-karlo/b50562e5
result-kand2-sdxl-wuerst-karlo
2023-11-23T03:38:23Z
39
0
null
[ "region:us" ]
2023-11-23T03:38:23Z
2023-11-23T03:38:22.000Z
2023-11-23T03:38:22
--- dataset_info: features: - name: result dtype: string - name: id dtype: int64 splits: - name: train num_bytes: 170 num_examples: 10 download_size: 1334 dataset_size: 170 configs: - config_name: default data_files: - split: train path: data/train-* --- # Dataset Card for "b50562e5" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.650661289691925, -0.0033841398544609547, 0.1924002468585968, 0.3684273958206177, -0.08945410698652267, -0.16792859137058258, 0.34630462527275085, -0.33025801181793213, 0.7411776185035706, 0.429939329624176, -0.8680625557899475, -0.6793990135192871, -0.5009608268737793, -0.07270675152540...
null
null
null
null
null
null
null
null
null
null
null
null
null
euisuh15/python-piss-my-name
euisuh15
2023-11-26T19:11:01Z
39
0
null
[ "region:us" ]
2023-11-26T19:11:01Z
2023-11-26T19:10:16.000Z
2023-11-26T19:10:16
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
rbawden/DiaBLa
rbawden
2022-10-25T14:21:10Z
38
1
null
[ "task_categories:translation", "annotations_creators:expert-generated", "language_creators:crowdsourced", "multilinguality:translation", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "language:fr", "license:cc-by-sa-4.0", "region:us" ]
2022-10-25T14:21:10Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
--- annotations_creators: - expert-generated language_creators: - crowdsourced language: - en - fr license: - cc-by-sa-4.0 multilinguality: - translation size_categories: - 1K<n<10K source_datasets: - original task_categories: - translation task_ids: [] pretty_name: DiaBLa language_bcp47: - en-UK - fr-FR --- # Dataset Card for DiaBLa: Bilingual dialogue parallel evaluation set ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [almanach.inria.fr/software_and_resources/custom/DiaBLa-en.html](http://almanach.inria.fr/software_and_resources/custom/DiaBLa-en.html) - **Repository:** [github.com/rbawden/DiaBLa-dataset](https://github.com/rbawden/DiaBLa-dataset) - **Paper:** [Bawden et al. (2021). DiaBLa: A Corpus of Bilingual Spontaneous Written Dialogues for Machine Translation. Language Resources and Evaluation(55). Pages 635–660. Springer Verlag. 10.1007/s10579-020-09514-4.](https://hal.inria.fr/hal-03021633) - **Point of contact:** rachel.bawden[at]inria.fr ### Dataset Summary The dataset is an English-French dataset for the evaluation of Machine Translation (MT) for informal, written bilingual dialogue. The dataset contains 144 spontaneous dialogues (5,700+ sentences) between native English and French speakers, mediated by one of two neural MT systems in a range of role-play settings. See below for some basic statistics. The dialogues are accompanied by fine-grained sentence-level judgments of MT quality, produced by the dialogue participants themselves, as well as by manually normalised versions and reference translations produced a posteriori. See here for information about evaluation. The motivation for the corpus is two-fold: to provide: - a unique resource for evaluating MT models for dialogue (i.e. in context) - a corpus for the analysis of MT-mediated communication ### Supported Tasks and Leaderboards [More Information Needed] ### Languages English (mainly UK) and French ## Dataset Structure ### Data Instances - **Size of downloaded dataset files:** 37 MB - **Number of parallel utterances:** 5748 Each example is highly annotated and is associated with dialogue context. An example from the test set looks as follows (only the first and last utterances of the dialogue history are shown for readability purposes): ``` { "id": "dialogue-2018-04-25T16-20-36.087170_french_english_1_2_25", "mt": "Tu m'en veux pour \u00e7a ?", "norm": "", "orig": "Are you blaming me for this?", "ref": "C'est moi que vous critiquez pour \u00e7a\u00a0?", "utterance_meta": { "eval_judgment": "medium", "eval_verbatim": "", "eval_problems": [ "coherence" ], "lang": "english" }, "dialogue_meta": { "start_time": "2018-04-25T16:20:36.087170", "end_time": "", "translation_model": "baseline", "final_evaluation_user1": { "style": "average", "coherence": "average", "grammaticality": "good", "meaning": "average", "word_choice": "average" }, "final_evaluation_user2": { "style": "", "coherence": "", "grammaticality": "", "meaning": "", "word_choice": "" }, "scenario": [ [ "You are both stuck in a lift at work.", "Vous \u00eates tous les deux bloqu\u00e9(e)s dans un ascenseur au travail." ], [ "You are an employee and you are with your boss.", "Vous \u00eates un(e) employ\u00e9(e) et vous \u00eates avez votre patron(ne)" ], [ "You are the boss and are with an employee.", "Vous \u00eates le ou la patron(ne) et vous \u00eates avec un(e) employ\u00e9(e)" ] ], "user1": { "role_num": 1, "role": [ "You are an employee and you are with your boss.", "Vous \u00eates un(e) employ\u00e9(e) et vous \u00eates avez votre patron(ne)" ], "initiated_dialogue": true, "turn_number": 2, "lang": "french" }, "user2": { "role_num": 2, "role": [ "You are the boss and are with an employee.", "Vous \u00eates le ou la patron(ne) et vous \u00eates avec un(e) employ\u00e9(e)" ], "initiated_dialogue": false, "turn_number": 1, "lang": "english" } }, "dialogue_history": [ { "id": "dialogue-2018-04-25T16-20-36.087170_french_english_1_2_0", "orig": "We appear to have stopped moving.", "norm": "", "mt": "On semble avoir arr\u00eat\u00e9 de bouger.", "ref": "J'ai l'impression qu'on s'est arr\u00eat\u00e9s.", "utterance_meta": { "eval_judgment": "medium", "eval_verbatim": "", "eval_problems": [ "style" ], "lang": "english" } }, [...] { "id": "dialogue-2018-04-25T16-20-36.087170_french_english_1_2_24", "orig": "La sonnerie s'est arr\u00eat\u00e9, je pense que personne ne va nous r\u00e9pondre.", "norm": "", "mt": "The ringing stopped, and I don't think anyone's gonna answer us.", "ref": "It stopped ringing. I don't think anybody's going to reply.", "utterance_meta": { "eval_judgment": "perfect", "eval_verbatim": "", "eval_problems": [], "lang": "french" } } ] } ``` ### Data Fields #### plain_text - `id`: a `string` feature. - `orig`: a `string` feature. - `norm`: a `string` feature. - `mt`: a `string` feature. - `ref`: a `string` feature. - `utterance_meta`: a dictionary feature containing: - `eval_judgment`: a `string` feature. - `eval_verbatim`: a `string` feature. - `eval_problems`: a list feature containing: - up to 5 `string` features. - `lang`: a `string` feature. - `dialogue_meta`: a dictionary feature containing: - `start_time` : a `string` feature. - `end_time`: a `string` feature. - `translation_model`: a `string` feature. - `final_evaluation_user1`: a dictionary feature containing: - `style`: a `string` feature. - `coherence`: a `string` feature. - `grammaticality`: a `string` feature. - `meaning`: a `string` feature. - `word_choice`: a `string` feature. - `final_evaluation_user2`: a dictionary feature containing: - `style`: a `string` feature. - `coherence`: a `string` feature. - `grammaticality`: a `string` feature. - `meaning`: a `string` feature. - `word_choice`: a `string` feature. - `scenario`: a list feature containing - 3 lists each containing 2 `string` features. - `user1`: a dictionary feature containing: - `role_num`: an `int` feature. - `role`: a list feature containing: - 2 `string` features. - `initiated_dialogue`: a `bool` feature. - `turn_number`: an `int` value. - `lang`: a `string` value. - `user2`: a dictionary feature containing: - `role_num`: an `int` feature. - `role`: a list feature containing: - 2 `string` features. - `initiated_dialogue`: a `bool` feature. - `turn_number`: an `int` value. - `lang`: a `string` value. - `dialogue_history`: a list feature containing: - dictionary features containing: - `id`: a `string` feature. - `orig`: a `string` feature. - `norm`: a `string` feature. - `mt`: a `string` feature. - `ref`: a `string` feature. - `utterance_meta`: a dictionary feature containing: - `eval_judgment`: a `string` feature. - `eval_verbatim`: a `string` feature. - `eval_problems`: a list feature containing: - up to 5 `string` features. - `lang`: a `string` feature. ### Data Splits DiaBLa is a test set only. | name |test | |----------|------:| |plain_text| 5748| ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization Original data was collected through a [dedicated online chat platform](https://github.com/rbawden/diabla-chat-interface) and involved native speakers of English and of French. As well as producing the original text, participants also annotated the quality of the machine-translated outputs of their partners' utterances (which they saw instead of their partners' original text) based on their monolingual intuitions and the dialogue context. Each dialogue is assigned one of 12 role-play scenarios and where appropriate each participant is assigned a role to play in the dialogue. #### Who are the source language producers? The source text producers were native French and native English volunteers (mainly British English). See the paper for very basic information concerning their backgrounds (age categories and experience in NLP). ### Annotations #### Annotation process On top of the original dialogue text (a mixture of utterances in English and in French), the following "annotations" are provided: - machine translated version of the original text (produced in real time and presented during the dialogue), produced by one of two MT systems, both trained using [Marian](https://github.com/marian-nmt/marian). - judgments of MT quality by participants (overall quality, particular problems, verbatim comments) - manually produced normalised version of the original text (for spelling mistakes, grammatical errors, missing punctuation, etc.) - manually produced reference translations #### Who are the annotators? The judgments of MT quality were produced by the dialogue participants themselves in real time. The normalised version of the text and the reference translations were manually produced by the authors of the paper. Translations were always done into the translator's native language and all translations were verified and post-edited by a bilingual English-French speaker. ### Personal and Sensitive Information A priori the dataset does not contain personal and sensitive information. Participants were instructed not to give any personal information and to assume the roles assigned in the role play scenario. Usernames were anonymised prior to distribution and any mention of either usernames or real names in the dialogues were replaced by generic names of the same gender as the participant. Only basic user information was collected to get an idea of the distribution of participants and to potentially see how multilingual ability influences quality judgments (rough age categories, experience in NLP or research, native languages, familiarity with the other language (either English or French), other languages spoken and gender). Gender was included because it is an important factor in translation (particularly for the direction English-to-French), and this was explained in advance to the participants in the FAQs. ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators The dataset was collected by Rachel Bawden, Eric Bilinski, Thomas Lavergne and Sophie Rosset (see citation below). ### Licensing Information The dataset is available under a CC BY-SA 4.0 licence. ### Citation Information If you use or are inspired by this dataset, please cite: ``` @article{bawden_DiaBLa:-A-Corpus-of_2021, author = {Bawden, Rachel and Bilinski, Eric and Lavergne, Thomas and Rosset, Sophie}, doi = {10.1007/s10579-020-09514-4}, title = {DiaBLa: A Corpus of Bilingual Spontaneous Written Dialogues for Machine Translation}, year = {2021}, journal = {Language Resources and Evaluation}, publisher = {Springer Verlag}, volume = {55}, pages = {635--660}, url = {https://hal.inria.fr/hal-03021633}, pdf = {https://hal.inria.fr/hal-03021633/file/diabla-lre-personal-formatting.pdf}, } ``` ### Contributions This dataset was added by Rachel Bawden [@rbawden](https://github.com/rbawden).
[ -0.5425506234169006, -0.7322291731834412, 0.3697434365749359, 0.5639804601669312, -0.1805303692817688, 0.09924111515283585, -0.12267187982797623, -0.12711305916309357, 0.41842740774154663, 0.652106523513794, -0.8804033994674683, -0.8662673234939575, -0.49101147055625916, 0.5491994619369507...
null
null
null
null
null
null
null
null
null
null
null
null
null
jquiros/suicide
jquiros
2022-03-08T11:23:20Z
38
4
null
[ "region:us" ]
2022-03-08T11:23:20Z
2022-03-08T11:20:09.000Z
2022-03-08T11:20:09
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
crystina-z/no-nonself-mrtydi
crystina-z
2022-04-10T02:02:35Z
38
0
null
[ "region:us" ]
2022-04-10T02:02:35Z
2022-03-08T20:04:07.000Z
2022-03-08T20:04:07
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
stjokerli/TextToText_wic_seqio
stjokerli
2022-03-18T04:56:49Z
38
0
null
[ "region:us" ]
2022-03-18T04:56:49Z
2022-03-13T09:30:58.000Z
2022-03-13T09:30:58
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
Chris1/GTA5
Chris1
2022-04-06T14:44:22Z
38
0
null
[ "region:us" ]
2022-04-06T14:44:22Z
2022-04-06T13:02:07.000Z
2022-04-06T13:02:07
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
scikit-learn/credit-card-clients
scikit-learn
2022-06-20T15:42:14Z
38
0
null
[ "license:cc0-1.0", "region:us" ]
2022-06-20T15:42:14Z
2022-06-20T14:57:10.000Z
2022-06-20T14:57:10
--- license: cc0-1.0 --- ## Default of Credit Card Clients Dataset The following was retrieved from [UCI machine learning repository](https://archive.ics.uci.edu/ml/datasets/default+of+credit+card+clients). **Dataset Information** This dataset contains information on default payments, demographic factors, credit data, history of payment, and bill statements of credit card clients in Taiwan from April 2005 to September 2005. **Content** There are 25 variables: - ID: ID of each client - LIMIT_BAL: Amount of given credit in NT dollars (includes individual and family/supplementary credit - SEX: Gender (1=male, 2=female) - EDUCATION: (1=graduate school, 2=university, 3=high school, 4=others, 5=unknown, 6=unknown) - MARRIAGE: Marital status (1=married, 2=single, 3=others) - AGE: Age in years - PAY_0: Repayment status in September, 2005 (-1=pay duly, 1=payment delay for one month, 2=payment delay for two months, … 8=payment delay for eight months, 9=payment delay for nine months and above) - PAY_2: Repayment status in August, 2005 (scale same as above) - PAY_3: Repayment status in July, 2005 (scale same as above) - PAY_4: Repayment status in June, 2005 (scale same as above) - PAY_5: Repayment status in May, 2005 (scale same as above) - PAY_6: Repayment status in April, 2005 (scale same as above) - BILL_AMT1: Amount of bill statement in September, 2005 (NT dollar) - BILL_AMT2: Amount of bill statement in August, 2005 (NT dollar) - BILL_AMT3: Amount of bill statement in July, 2005 (NT dollar) - BILL_AMT4: Amount of bill statement in June, 2005 (NT dollar) - BILL_AMT5: Amount of bill statement in May, 2005 (NT dollar) - BILL_AMT6: Amount of bill statement in April, 2005 (NT dollar) - PAY_AMT1: Amount of previous payment in September, 2005 (NT dollar) - PAY_AMT2: Amount of previous payment in August, 2005 (NT dollar) - PAY_AMT3: Amount of previous payment in July, 2005 (NT dollar) - PAY_AMT4: Amount of previous payment in June, 2005 (NT dollar) - PAY_AMT5: Amount of previous payment in May, 2005 (NT dollar) - PAY_AMT6: Amount of previous payment in April, 2005 (NT dollar) - default.payment.next.month: Default payment (1=yes, 0=no) **Inspiration** Some ideas for exploration: How does the probability of default payment vary by categories of different demographic variables? Which variables are the strongest predictors of default payment? **Acknowledgements** Any publications based on this dataset should acknowledge the following: Lichman, M. (2013). UCI Machine Learning Repository [http://archive.ics.uci.edu/ml]. Irvine, CA: University of California, School of Information and Computer Science.
[ -0.20623581111431122, -0.4662413001060486, 0.21763896942138672, -0.06409598141908646, -0.13395977020263672, -0.2815612852573395, 0.223565936088562, -0.1311955451965332, 0.07311364263296127, 0.9917935132980347, -0.382113516330719, -0.7576791644096375, -0.36529484391212463, -0.09609708935022...
null
null
null
null
null
null
null
null
null
null
null
null
null
jakartaresearch/indoqa
jakartaresearch
2022-12-17T06:07:27Z
38
1
null
[ "task_categories:question-answering", "task_ids:extractive-qa", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:id", "license:cc-by-nd-4.0", "indoqa", "qa", "question-answering"...
2022-12-17T06:07:27Z
2022-08-13T10:54:08.000Z
2022-08-13T10:54:08
--- annotations_creators: - expert-generated language: - id language_creators: - found license: - cc-by-nd-4.0 multilinguality: - monolingual pretty_name: Indonesian Question Answering Dataset size_categories: - 1K<n<10K source_datasets: - original tags: - indoqa - qa - question-answering - indonesian task_categories: - question-answering task_ids: - extractive-qa --- # Dataset Card for Indonesian Question Answering Dataset ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary [More Information Needed] ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@fhrzn](https://github.com/fhrzn)[@Kalzaik](https://github.com/Kalzaik) [@ibamibrahim](https://github.com/ibamibrahim) [@andreaschandra](https://github.com/andreaschandra) for adding this dataset.
[ -0.5596947073936462, -0.779777467250824, 0.0233513955026865, 0.31879952549934387, -0.326363205909729, -0.018054915592074394, -0.16507212817668915, -0.32631006836891174, 0.5443535447120667, 0.8801376819610596, -0.7869605422019958, -0.8764442801475525, -0.6577150821685791, 0.4416554272174835...
null
null
null
null
null
null
null
null
null
null
null
null
null
teticio/audio-diffusion-256
teticio
2022-11-09T10:49:48Z
38
3
null
[ "task_categories:image-to-image", "size_categories:10K<n<100K", "audio", "spectrograms", "region:us" ]
2022-11-09T10:49:48Z
2022-08-25T17:32:42.000Z
2022-08-25T17:32:42
--- annotations_creators: [] language: [] language_creators: [] license: [] multilinguality: [] pretty_name: Mel spectrograms of music size_categories: - 10K<n<100K source_datasets: [] tags: - audio - spectrograms task_categories: - image-to-image task_ids: [] --- Over 20,000 256x256 mel spectrograms of 5 second samples of music from my Spotify liked playlist. The code to convert from audio to spectrogram and vice versa can be found in https://github.com/teticio/audio-diffusion along with scripts to train and run inference using De-noising Diffusion Probabilistic Models. ``` x_res = 256 y_res = 256 sample_rate = 22050 n_fft = 2048 hop_length = 512 ```
[ -0.7777500152587891, -0.2157420665025711, 0.7321594953536987, 0.4212033450603485, -0.07960247993469238, -0.18307003378868103, -0.28970441222190857, -0.3536110818386078, 0.5059683918952942, 0.34585556387901306, -0.42328378558158875, -0.9210429787635803, -0.313353568315506, -0.26719149947166...
null
null
null
null
null
null
null
null
null
null
null
null
null
din0s/asqa
din0s
2022-09-20T16:14:54Z
38
0
null
[ "task_categories:question-answering", "task_ids:open-domain-qa", "annotations_creators:crowdsourced", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:extended|ambig_qa", "language:en", "license:apache-2.0", "factoid questions", "l...
2022-09-20T16:14:54Z
2022-09-19T22:25:51.000Z
2022-09-19T22:25:51
--- annotations_creators: - crowdsourced language: - en language_creators: - expert-generated license: - apache-2.0 multilinguality: - monolingual pretty_name: ASQA size_categories: - 1K<n<10K source_datasets: - extended|ambig_qa tags: - factoid questions - long-form answers task_categories: - question-answering task_ids: - open-domain-qa --- # Dataset Card for ASQA ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Additional Information](#additional-information) - [Contributions](#contributions) ## Dataset Description - **Repository:** https://github.com/google-research/language/tree/master/language/asqa - **Paper:** https://arxiv.org/abs/2204.06092 - **Leaderboard:** https://ambigqa.github.io/asqa_leaderboard.html ### Dataset Summary ASQA is the first long-form question answering dataset that focuses on ambiguous factoid questions. Different from previous long-form answers datasets, each question is annotated with both long-form answers and extractive question-answer pairs, which should be answerable by the generated passage. A generated long-form answer will be evaluated using both ROUGE and QA accuracy. In the paper, we show that these evaluation metrics are well-correlated with human judgments. ### Supported Tasks and Leaderboards Long-form Question Answering. [Leaderboard](https://ambigqa.github.io/asqa_leaderboard.html) ### Languages - English ## Dataset Structure ### Data Instances ```py { "ambiguous_question": "Where does the civil liberties act place the blame for the internment of u.s. citizens?", "qa_pairs": [ { "context": "No context provided", "question": "Where does the civil liberties act place the blame for the internment of u.s. citizens by apologizing on behalf of them?", "short_answers": [ "the people of the United States" ], "wikipage": None }, { "context": "No context provided", "question": "Where does the civil liberties act place the blame for the internment of u.s. citizens by making them pay reparations?", "short_answers": [ "United States government" ], "wikipage": None } ], "wikipages": [ { "title": "Civil Liberties Act of 1988", "url": "https://en.wikipedia.org/wiki/Civil%20Liberties%20Act%20of%201988" } ], "annotations": [ { "knowledge": [ { "content": "The Civil Liberties Act of 1988 (Pub.L. 100–383, title I, August 10, 1988, 102 Stat. 904, 50a U.S.C. § 1989b et seq.) is a United States federal law that granted reparations to Japanese Americans who had been interned by the United States government during World War II.", "wikipage": "Civil Liberties Act of 1988" } ], "long_answer": "The Civil Liberties Act of 1988 is a United States federal law that granted reparations to Japanese Americans who had been interned by the United States government during World War II. In the act, the blame for the internment of U.S. citizens was placed on the people of the United States, by apologizing on behalf of them. Furthermore, the blame for the internment was placed on the United States government, by making them pay reparations." } ], "sample_id": -4557617869928758000 } ``` ### Data Fields - `ambiguous_question`: ambiguous question from AmbigQA. - `annotations`: long-form answers to the ambiguous question constructed by ASQA annotators. - `annotations/knowledge`: list of additional knowledge pieces. - `annotations/knowledge/content`: a passage from Wikipedia. - `annotations/knowledge/wikipage`: title of the Wikipedia page the passage was taken from. - `annotations/long_answer`: annotation. - `qa_pairs`: Q&A pairs from AmbigQA which are used for disambiguation. - `qa_pairs/context`: additional context provided. - `qa_pairs/question`: disambiguated question from AmbigQA. - `qa_pairs/short_answers`: list of short answers from AmbigQA. - `qa_pairs/wikipage`: title of the Wikipedia page the additional context was taken from. - `sample_id`: the unique id of the sample - `wikipages`: list of Wikipedia pages visited by AmbigQA annotators. - `wikipages/title`: title of the Wikipedia page. - `wikipages/url`: link to the Wikipedia page. ### Data Splits | **Split** | **Instances** | |-----------|---------------| | Train | 4353 | | Dev | 948 | ## Additional Information ### Contributions Thanks to [@din0s](https://github.com/din0s) for adding this dataset.
[ -0.5463582873344421, -0.7262665629386902, 0.49687543511390686, -0.09552009403705597, -0.14116840064525604, -0.037118230015039444, -0.08747050911188126, -0.2835697531700134, 0.4076807498931885, 0.4947029948234558, -0.596950113773346, -0.5265423655509949, -0.30089280009269714, 0.179322287440...
null
null
null
null
null
null
null
null
null
null
null
null
null
IIC/SQUAC
IIC
2022-10-11T11:52:45Z
38
1
null
[ "region:us" ]
2022-10-11T11:52:45Z
2022-10-11T11:52:34.000Z
2022-10-11T11:52:34
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
Dizex/InstaFoodSet
Dizex
2022-12-11T20:07:40Z
38
0
null
[ "region:us" ]
2022-12-11T20:07:40Z
2022-11-06T19:39:47.000Z
2022-11-06T19:39:47
--- dataset_info: features: - name: tokens sequence: string - name: iob_tags sequence: string - name: iob_tags_ids sequence: int64 splits: - name: train num_bytes: 346804 num_examples: 320 - name: val num_bytes: 37219 num_examples: 40 - name: test num_bytes: 39352 num_examples: 40 download_size: 84698 dataset_size: 423375 --- # Dataset Card for "InstaFoodSet" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.41006234288215637, -0.16151085495948792, 0.060299165546894073, 0.16038522124290466, -0.12415237724781036, 0.22218400239944458, 0.24877263605594635, -0.328553169965744, 1.0188878774642944, 0.525925874710083, -0.7715709805488586, -0.7286593317985535, -0.49534305930137634, -0.0631333142518...
null
null
null
null
null
null
null
null
null
null
null
null
null
101arrowz/vox_celeb
101arrowz
2023-08-20T03:04:07Z
38
1
null
[ "task_categories:automatic-speech-recognition", "task_categories:audio-classification", "task_categories:image-classification", "task_ids:speaker-identification", "annotations_creators:crowdsourced", "language_creators:crowdsourced", "multilinguality:multilingual", "size_categories:1K<n<10K", "size_...
2023-08-20T03:04:07Z
2022-11-13T01:43:46.000Z
2022-11-13T01:43:46
--- annotations_creators: - crowdsourced language: [] language_creators: - crowdsourced license: - cc-by-4.0 multilinguality: - multilingual pretty_name: VoxCeleb size_categories: - 1K<n<10K - 10K<n<100K - 100K<n<1M source_datasets: [] tags: [] task_categories: - automatic-speech-recognition - audio-classification - image-classification task_ids: - speaker-identification --- # Dataset Card for VoxCeleb ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description ### Dataset Summary VoxCeleb is an audio-visual dataset consisting of short clips of human speech, extracted from interview videos uploaded to YouTube. NOTE: Although this dataset can be automatically downloaded, you must manually request credentials to access it from the creators' website. ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances Each datapoint has a path to the audio/video clip along with metadata about the speaker. ``` { 'file': '/datasets/downloads/extracted/[hash]/wav/id10271/_YimahVgI1A/00003.wav', 'file_format': 'wav', 'dataset_id': 'vox1', 'speaker_id': 'id10271', 'speaker_gender': 'm', 'speaker_name': 'Ed_Westwick', 'speaker_nationality': 'UK', 'video_id': '_YimahVgI1A', 'clip_id': '00003', 'audio': { 'path': '/datasets/downloads/extracted/[hash]/wav/id10271/_YimahVgI1A/00003.wav', 'array': array([...], dtype=float32), 'sampling_rate': 16000 } } ``` ### Data Fields Each row includes the following fields: - `file`: The path to the audio/video clip - `file_format`: The file format in which the clip is stored (e.g. `wav`, `aac`, `mp4`) - `dataset_id`: The ID of the dataset this clip is from (`vox1`, `vox2`) - `speaker_id`: The ID of the speaker in this clip - `speaker_gender`: The gender of the speaker (`m`/`f`) - `speaker_name` (VoxCeleb1 only): The full name of the speaker in the clip - `speaker_nationality` (VoxCeleb1 only): The speaker's country of origin - `video_id`: The ID of the video from which this clip was taken - `clip_index`: The index of the clip for this specific video - `audio` (Audio dataset only): The audio signal data ### Data Splits The dataset has a predefined dev set and test set. The dev set has been renamed to a "train" split. ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information The dataset includes recordings of clips (mostly of celebrities and public figures) from public YouTube videos. The names of speakers in VoxCeleb1 are provided. ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information The VoxCeleb authors request that anyone who uses VoxCeleb1 or VoxCeleb2 includes the following three citations: ``` @Article{Nagrani19, author = "Arsha Nagrani and Joon~Son Chung and Weidi Xie and Andrew Zisserman", title = "Voxceleb: Large-scale speaker verification in the wild", journal = "Computer Science and Language", year = "2019", publisher = "Elsevier", } @InProceedings{Chung18b, author = "Chung, J.~S. and Nagrani, A. and Zisserman, A.", title = "VoxCeleb2: Deep Speaker Recognition", booktitle = "INTERSPEECH", year = "2018", } @InProceedings{Nagrani17, author = "Nagrani, A. and Chung, J.~S. and Zisserman, A.", title = "VoxCeleb: a large-scale speaker identification dataset", booktitle = "INTERSPEECH", year = "2017", } ``` ### Contributions Thanks to [@101arrowz](https://github.com/101arrowz) for adding this dataset.
[ -0.7007195949554443, -0.7795397639274597, 0.12930825352668762, 0.24614810943603516, -0.19460181891918182, 0.06788404285907745, -0.5131685137748718, -0.03965647146105766, 0.37010809779167175, 0.3776523470878601, -0.77531498670578, -1.0008189678192139, -0.5688537359237671, -0.034280583262443...
null
null
null
null
null
null
null
null
null
null
null
null
null
shjwudp/chinese-c4
shjwudp
2023-06-20T11:40:06Z
38
12
null
[ "language:zh", "license:cc-by-4.0", "region:us" ]
2023-06-20T11:40:06Z
2022-11-15T01:27:26.000Z
2022-11-15T01:27:26
--- license: cc-by-4.0 language: - zh --- ## Introduction Chinese-C4 is a clean Chinese internet dataset based on Common Crawl. The dataset is 46.29GB and has undergone multiple cleaning strategies, including Chinese filtering, heuristic cleaning based on punctuation, line-based hashing for deduplication, and repetition removal. The dataset is open source and free for commercial use, and you are welcome to use the data and the cleaning strategies provided and contribute your cleaning strategies. You can find the cleaning script for the dataset on GitHub [c4-dataset-script](https://github.com/shjwudp/c4-dataset-script).
[ -0.3483254611492157, -0.33034029603004456, 0.2448779046535492, 0.18547800183296204, -0.4120955169200897, -0.0557364858686924, -0.20413517951965332, -0.508743941783905, 0.07251019030809402, 0.6238188743591309, -0.5017893314361572, -0.8337077498435974, 0.10429196804761887, 0.2934835851192474...
null
null
null
null
null
null
null
null
null
null
null
null
null
HuggingFaceM4/TextCaps
HuggingFaceM4
2022-12-09T01:38:32Z
38
2
null
[ "license:cc-by-4.0", "region:us" ]
2022-12-09T01:38:32Z
2022-12-06T20:56:12.000Z
2022-12-06T20:56:12
--- license: cc-by-4.0 ---
[ -0.1285339742898941, -0.18616800010204315, 0.6529127359390259, 0.4943626821041107, -0.1931934952735901, 0.2360742688179016, 0.360720157623291, 0.05056300014257431, 0.5793654322624207, 0.7400140166282654, -0.6508105993270874, -0.23783984780311584, -0.7102248668670654, -0.047826044261455536,...
null
null
null
null
null
null
null
null
null
null
null
null
null
jonathan-roberts1/Optimal-31
jonathan-roberts1
2023-03-31T17:06:29Z
38
0
null
[ "task_categories:image-classification", "task_categories:zero-shot-image-classification", "license:other", "region:us" ]
2023-03-31T17:06:29Z
2023-02-17T15:53:58.000Z
2023-02-17T15:53:58
--- dataset_info: features: - name: image dtype: image - name: label dtype: class_label: names: '0': airplane '1': airport '2': baseball diamond '3': basketball court '4': beach '5': bridge '6': chaparral '7': church '8': circular farmland '9': commercial area '10': dense residential '11': desert '12': forest '13': freeway '14': golf course '15': ground track field '16': harbor '17': industrial area '18': intersection '19': island '20': lake '21': meadow '22': medium residential '23': mobile home park '24': mountain '25': overpass '26': parking lot '27': railway '28': rectangular farmland '29': roundabout '30': runway splits: - name: train num_bytes: 25100636.72 num_examples: 1860 download_size: 25105452 dataset_size: 25100636.72 license: other task_categories: - image-classification - zero-shot-image-classification --- # Dataset Card for "Optimal-31" ## Dataset Description - **Paper** [Scene classification with recurrent attention of VHR remote sensing images](https://ieeexplore.ieee.org/iel7/5/8045830/07891544.pdf) ### Licensing Information [No license for now, cite the paper below.] ## Citation Information [Scene classification with recurrent attention of VHR remote sensing images](https://ieeexplore.ieee.org/iel7/5/8045830/07891544.pdf) ``` @article{wang2018scene, title = {Scene classification with recurrent attention of VHR remote sensing images}, author = {Wang, Qi and Liu, Shaoteng and Chanussot, Jocelyn and Li, Xuelong}, year = 2018, journal = {IEEE Transactions on Geoscience and Remote Sensing}, publisher = {IEEE}, volume = 57, number = 2, pages = {1155--1167} } ```
[ -0.6191813945770264, -0.16142036020755768, 0.26190951466560364, 0.1113378256559372, -0.5422631502151489, -0.24539227783679962, 0.10389702767133713, -0.4494291841983795, -0.38256919384002686, 0.30216631293296814, -0.6494527459144592, -0.6190582513809204, -0.24737858772277832, 0.130532786250...
null
null
null
null
null
null
null
null
null
null
null
null
null
sid6i7/patient-doctor
sid6i7
2023-03-30T20:02:27Z
38
4
null
[ "region:us" ]
2023-03-30T20:02:27Z
2023-03-30T20:01:09.000Z
2023-03-30T20:01:09
Entry not found
[ -0.32276487350463867, -0.22568444907665253, 0.8622263073921204, 0.43461570143699646, -0.5282988548278809, 0.7012969255447388, 0.7915717363357544, 0.07618642598390579, 0.7746027112007141, 0.25632190704345703, -0.7852815389633179, -0.22573848068714142, -0.910447895526886, 0.5715675354003906,...
null
null
null
null
null
null
null
null
null
null
null
null
null
SkyHuReal/DrugBank-Alpaca
SkyHuReal
2023-04-03T17:37:30Z
38
0
null
[ "license:afl-3.0", "region:us" ]
2023-04-03T17:37:30Z
2023-04-03T15:39:50.000Z
2023-04-03T15:39:50
--- license: afl-3.0 ---
[ -0.1285339742898941, -0.18616800010204315, 0.6529127359390259, 0.4943626821041107, -0.1931934952735901, 0.2360742688179016, 0.360720157623291, 0.05056300014257431, 0.5793654322624207, 0.7400140166282654, -0.6508105993270874, -0.23783984780311584, -0.7102248668670654, -0.047826044261455536,...
null
null
null
null
null
null
null
null
null
null
null
null
null