id
stringlengths
2
115
author
stringlengths
2
42
last_modified
timestamp[us, tz=UTC]
downloads
int64
0
8.87M
likes
int64
0
3.84k
paperswithcode_id
stringlengths
2
45
tags
list
lastModified
timestamp[us, tz=UTC]
createdAt
stringlengths
24
24
key
stringclasses
1 value
created
timestamp[us]
card
stringlengths
1
1.01M
embedding
list
library_name
stringclasses
21 values
pipeline_tag
stringclasses
27 values
mask_token
null
card_data
null
widget_data
null
model_index
null
config
null
transformers_info
null
spaces
null
safetensors
null
transformersInfo
null
modelId
stringlengths
5
111
embeddings
list
abdoutony207/en_ar_dataset
abdoutony207
2022-05-28T18:52:55Z
13
0
null
[ "region:us" ]
2022-05-28T18:52:55Z
2022-05-28T18:28:41.000Z
2022-05-28T18:28:41
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
vcv/sentiment-banking
vcv
2022-06-06T09:42:57Z
13
0
null
[ "region:us" ]
2022-06-06T09:42:57Z
2022-05-28T18:36:22.000Z
2022-05-28T18:36:22
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
Rexhaif/cedr-full
Rexhaif
2022-05-29T22:31:49Z
13
0
null
[ "region:us" ]
2022-05-29T22:31:49Z
2022-05-29T22:29:09.000Z
2022-05-29T22:29:09
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
meetyildiz/toqad-aug
meetyildiz
2022-05-30T06:26:00Z
13
0
null
[ "region:us" ]
2022-05-30T06:26:00Z
2022-05-30T06:24:43.000Z
2022-05-30T06:24:43
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
NLPC-UOM/Sentiment-tagger
NLPC-UOM
2022-10-25T10:29:09Z
13
0
null
[ "language:si", "license:mit", "region:us" ]
2022-10-25T10:29:09Z
2022-06-03T15:51:41.000Z
2022-06-03T15:51:41
--- language: - si license: - mit --- *Sentiment Analysis of Sinhala News Comments* Dataset used in this project is collected by crawling Sinhala online news sites, mainly www.lankadeepa.lk. contact Please contact us if you need more information. Surangika Ranathunga-surangika@cse.mrt.ac.lk Isuru Liyanage-theisuru@gmail.com https://github.com/theisuru/sentiment-tagger cite If you use this data please cite this work Ranathunga, S., & Liyanage, I. U. (2021). Sentiment Analysis of Sinhala News Comments. Transactions on Asian and Low-Resource Language Information Processing, 20(4), 1-23.
[ -0.14582060277462006, -0.498007595539093, -0.07850293815135956, 0.58656245470047, -0.733677327632904, 0.07148933410644531, -0.12579485774040222, -0.24300114810466766, 0.7380655407905579, 0.38943737745285034, -0.486223042011261, -0.17131152749061584, -0.2566794157028198, 0.426265150308609, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
PLN-Activdad-2/sentiment-banking
PLN-Activdad-2
2022-06-03T17:08:37Z
13
0
null
[ "region:us" ]
2022-06-03T17:08:37Z
2022-06-03T17:08:05.000Z
2022-06-03T17:08:05
Entry not found
[ -0.32276472449302673, -0.22568407654762268, 0.8622258901596069, 0.4346148371696472, -0.5282984972000122, 0.7012965679168701, 0.7915717363357544, 0.07618629932403564, 0.7746022939682007, 0.2563222646713257, -0.785281777381897, -0.22573848068714142, -0.9104482531547546, 0.5715669393539429, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
yanekyuk/wikikey-fr
yanekyuk
2022-09-17T02:21:13Z
13
0
null
[ "language:fr", "region:us" ]
2022-09-17T02:21:13Z
2022-06-03T18:05:45.000Z
2022-06-03T18:05:45
--- language: fr ---
[ -0.12853392958641052, -0.18616779148578644, 0.6529127955436707, 0.49436280131340027, -0.19319361448287964, 0.23607419431209564, 0.36072003841400146, 0.050563063472509384, 0.579365611076355, 0.7400140762329102, -0.6508104205131531, -0.23783954977989197, -0.7102249264717102, -0.0478260256350...
null
null
null
null
null
null
null
null
null
null
null
null
null
limsc/mlm-tapt-requirements
limsc
2022-06-04T04:30:23Z
13
0
null
[ "region:us" ]
2022-06-04T04:30:23Z
2022-06-04T04:25:57.000Z
2022-06-04T04:25:57
Entry not found
[ -0.32276472449302673, -0.22568407654762268, 0.8622258901596069, 0.4346148371696472, -0.5282984972000122, 0.7012965679168701, 0.7915717363357544, 0.07618629932403564, 0.7746022939682007, 0.2563222646713257, -0.785281777381897, -0.22573848068714142, -0.9104482531547546, 0.5715669393539429, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
limsc/fr-nfr-classification
limsc
2022-06-05T00:16:48Z
13
0
null
[ "region:us" ]
2022-06-05T00:16:48Z
2022-06-04T04:33:19.000Z
2022-06-04T04:33:19
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
limsc/req-subclass-classification
limsc
2022-06-14T22:19:20Z
13
1
null
[ "region:us" ]
2022-06-14T22:19:20Z
2022-06-04T04:33:38.000Z
2022-06-04T04:33:38
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
awghuku/infore25
awghuku
2022-06-04T05:55:22Z
13
0
null
[ "license:cc-by-4.0", "region:us" ]
2022-06-04T05:55:22Z
2022-06-04T04:53:29.000Z
2022-06-04T04:53:29
--- license: cc-by-4.0 ---
[ -0.12853367626667023, -0.18616794049739838, 0.6529126763343811, 0.4943627417087555, -0.19319313764572144, 0.23607443273067474, 0.36071979999542236, 0.05056338757276535, 0.5793654322624207, 0.7400138974189758, -0.6508103013038635, -0.23783987760543823, -0.710224986076355, -0.047825977206230...
null
null
null
null
null
null
null
null
null
null
null
null
null
nateraw/imagefolder-metadata-test
nateraw
2022-06-07T03:06:07Z
13
0
null
[ "region:us" ]
2022-06-07T03:06:07Z
2022-06-07T02:48:56.000Z
2022-06-07T02:48:56
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
limsc/concept-recognition-not-iob
limsc
2022-06-08T04:41:48Z
13
0
null
[ "region:us" ]
2022-06-08T04:41:48Z
2022-06-08T04:41:44.000Z
2022-06-08T04:41:44
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
demo-org/auditor_review
demo-org
2022-08-30T21:42:09Z
13
0
null
[ "task_categories:text-classification", "task_ids:multi-class-classification", "task_ids:sentiment-classification", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "region:us" ]
2022-08-30T21:42:09Z
2022-06-14T03:06:17.000Z
2022-06-14T03:06:17
--- annotations_creators: - expert-generated language_creators: - found language: - en multilinguality: - monolingual size_categories: - 1K<n<10K source_datasets: - original task_categories: - text-classification task_ids: - multi-class-classification - sentiment-classification paperswithcode_id: null pretty_name: Auditor_Review --- # Dataset Card for Auditor_Review ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) ## Dataset Description Auditor review data collected by News Department - **Point of Contact:** Talked to COE for Auditing, currently sue@demo.org ### Dataset Summary Auditor sentiment dataset of sentences from financial news. The dataset consists of 3500 sentences from English language financial news categorized by sentiment. The dataset is divided by the agreement rate of 5-8 annotators. ### Supported Tasks and Leaderboards Sentiment Classification ### Languages English ## Dataset Structure ### Data Instances ``` "sentence": "Pharmaceuticals group Orion Corp reported a fall in its third-quarter earnings that were hit by larger expenditures on R&D and marketing .", "label": "negative" ``` ### Data Fields - sentence: a tokenized line from the dataset - label: a label corresponding to the class as a string: 'positive' - (2), 'neutral' - (1), or 'negative' - (0) Complete data code is [available here](https://www.datafiles.samhsa.gov/get-help/codebooks/what-codebook) ### Data Splits A train/test split was created randomly with a 75/25 split ## Dataset Creation ### Curation Rationale To gather our auditor evaluations into one dataset. Previous attempts using off-the-shelf sentiment had only 70% F1, this dataset was an attempt to improve upon that performance. ### Source Data #### Initial Data Collection and Normalization The corpus used in this paper is made out of English news reports. #### Who are the source language producers? The source data was written by various auditors. ### Annotations #### Annotation process This release of the auditor reviews covers a collection of 4840 sentences. The selected collection of phrases was annotated by 16 people with adequate background knowledge of financial markets. The subset here is where inter-annotation agreement was greater than 75%. #### Who are the annotators? They were pulled from the SME list, names are held by sue@demo.org ### Personal and Sensitive Information There is no personal or sensitive information in this dataset. ## Considerations for Using the Data ### Discussion of Biases All annotators were from the same institution and so interannotator agreement should be understood with this taken into account. The [Dataset Measurement tool](https://huggingface.co/spaces/huggingface/data-measurements-tool) identified these bias statistics: ![Bias](https://huggingface.co/datasets/demo-org/auditor_review/resolve/main/bias_stats.png) ### Other Known Limitations [More Information Needed] ### Licensing Information License: Demo.Org Proprietary - DO NOT SHARE
[ -0.5004733800888062, -0.27193865180015564, 0.014816772192716599, 0.3621878921985626, -0.48215603828430176, -0.1183815598487854, -0.120775505900383, -0.47154954075813293, 0.4688456058502197, 0.5331881046295166, -0.5158435702323914, -0.7979121208190918, -0.6103346943855286, 0.292323529720306...
null
null
null
null
null
null
null
null
null
null
null
null
null
pszemraj/multi_fc
pszemraj
2022-06-16T11:57:52Z
13
0
null
[ "license:other", "automatic claim verification", "claims", "arxiv:1909.03242", "region:us" ]
2022-06-16T11:57:52Z
2022-06-15T11:27:47.000Z
2022-06-15T11:27:47
--- license: other tags: - automatic claim verification - claims --- # multiFC - a dataset for the task of **automatic claim verification** - License is currently unknown, please refer to the original paper/[dataset site](http://www.copenlu.com/publication/2019_emnlp_augenstein/): - https://arxiv.org/abs/1909.03242 ## Dataset contents - **IMPORTANT:** the `label` column in the `test` set has dummy values as these were not provided (see original readme section for explanation) ``` DatasetDict({ train: Dataset({ features: ['claimID', 'claim', 'label', 'claimURL', 'reason', 'categories', 'speaker', 'checker', 'tags', 'article title', 'publish date', 'climate', 'entities'], num_rows: 27871 }) test: Dataset({ features: ['claimID', 'claim', 'label', 'claimURL', 'reason', 'categories', 'speaker', 'checker', 'tags', 'article title', 'publish date', 'climate', 'entities'], num_rows: 3487 }) validation: Dataset({ features: ['claimID', 'claim', 'label', 'claimURL', 'reason', 'categories', 'speaker', 'checker', 'tags', 'article title', 'publish date', 'climate', 'entities'], num_rows: 3484 }) }) ``` ## Paper Abstract / Citation > We contribute the largest publicly available dataset of naturally occurring factual claims for the purpose of automatic claim verification. It is collected from 26 fact checking websites in English, paired with textual sources and rich metadata, and labelled for veracity by human expert journalists. We present an in-depth analysis of the dataset, highlighting characteristics and challenges. Further, we present results for automatic veracity prediction, both with established baselines and with a novel method for joint ranking of evidence pages and predicting veracity that outperforms all baselines. Significant performance increases are achieved by encoding evidence, and by modelling metadata. Our best-performing model achieves a Macro F1 of 49.2%, showing that this is a challenging testbed for claim veracity prediction. ``` @inproceedings{conf/emnlp2019/Augenstein, added-at = {2019-10-27T00:00:00.000+0200}, author = {Augenstein, Isabelle and Lioma, Christina and Wang, Dongsheng and Chaves Lima, Lucas and Hansen, Casper and Hansen, Christian and Grue Simonsen, Jakob}, booktitle = {EMNLP}, crossref = {conf/emnlp/2019}, publisher = {Association for Computational Linguistics}, title = {MultiFC: A Real-World Multi-Domain Dataset for Evidence-Based Fact Checking of Claims}, year = 2019 } ``` ## Original README Real-World Multi-Domain Dataset for Evidence-Based Fact Checking of Claims The MultiFC is the largest publicly available dataset of naturally occurring factual claims for automatic claim verification. It is collected from 26 English fact-checking websites paired with textual sources and rich metadata and labeled for veracity by human expert journalists. ###### TRAIN and DEV ####### The train and dev files are (tab-separated) and contain the following metadata: claimID, claim, label, claimURL, reason, categories, speaker, checker, tags, article title, publish date, climate, entities Fields that could not be crawled were set as "None." Please refer to Table 11 of our paper to see the summary statistics. ###### TEST ####### The test file follows the same structure. However, we have removed the label. Thus, it only presents 12 metadata. claimID, claim, claim, reason, categories, speaker, checker, tags, article title, publish date, climate, entities Fields that could not be crawled were set as "None." Please refer to Table 11 of our paper to see the summary statistics. ###### Snippets ###### The text of each claim is submitted verbatim as a query to the Google Search API (without quotes). In the folder snippet, we provide the top 10 snippets retrieved. In some cases, fewer snippets are provided since we have excluded the claimURL from the snippets. Each file in the snippets folder is named after the claimID of the claim submitted as a query. Snippets file is (tab-separated) and contains the following metadata: rank_position, title, snippet, snippet_url For more information, please refer to our paper: References: Isabelle Augenstein, Christina Lioma, Dongsheng Wang, Lucas Chaves Lima, Casper Hansen, Christian Hansen, and Jakob Grue Simonsen. 2019. MultiFC: A Real-World Multi-Domain Dataset for Evidence-Based Fact Checking of Claims. In EMNLP. Association for Computational Linguistics. https://copenlu.github.io/publication/2019_emnlp_augenstein/
[ -0.389263391494751, -0.645091712474823, 0.3027518391609192, 0.19997328519821167, 0.05859309807419777, 0.1025141179561615, -0.2937428057193756, -0.6046632528305054, 0.19315548241138458, 0.30019092559814453, -0.4197578728199005, -0.7343537211418152, -0.5887863636016846, 0.3787493109703064, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
EddieChen372/js2jest
EddieChen372
2022-06-17T05:51:23Z
13
0
null
[ "region:us" ]
2022-06-17T05:51:23Z
2022-06-17T05:51:00.000Z
2022-06-17T05:51:00
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
EddieChen372/js2jestFiles
EddieChen372
2022-06-17T05:56:57Z
13
0
null
[ "region:us" ]
2022-06-17T05:56:57Z
2022-06-17T05:56:48.000Z
2022-06-17T05:56:48
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
marcomameli01/gear
marcomameli01
2022-06-17T07:37:26Z
13
0
null
[ "region:us" ]
2022-06-17T07:37:26Z
2022-06-17T07:06:41.000Z
2022-06-17T07:06:41
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
iohadrubin/qampari
iohadrubin
2022-06-28T11:43:03Z
13
0
null
[ "region:us" ]
2022-06-28T11:43:03Z
2022-06-19T11:44:36.000Z
2022-06-19T11:44:36
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
athar/cc
athar
2022-06-20T19:24:00Z
13
0
null
[ "region:us" ]
2022-06-20T19:24:00Z
2022-06-20T19:23:25.000Z
2022-06-20T19:23:25
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
Nexdata/Mandarin_Heavy_Accent_Speech_Data
Nexdata
2023-08-30T10:38:08Z
13
0
null
[ "region:us" ]
2023-08-30T10:38:08Z
2022-06-22T06:26:29.000Z
2022-06-22T06:26:29
--- YAML tags: - copy-paste the tags obtained with the tagging app: https://github.com/huggingface/datasets-tagging --- # Dataset Card for Nexdata/Mandarin_Heavy_Accent_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/44?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary It collects 2,034 local Chinese from 26 provinces like Henan, Shanxi, Sichuan, Hunan, Fujian, etc. It is mandarin speech data with heavy accent. The recoring contents are finance and economics, entertainment, policy, news, TV, and movies. For more details, please refer to the link: https://www.nexdata.ai/datasets/44?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Mandarin Chinese ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
[ -0.34986183047294617, -0.525627851486206, -0.2101520299911499, 0.4491666555404663, -0.15685033798217773, 0.01937953196465969, -0.48308202624320984, -0.39874619245529175, 0.5859537720680237, 0.7277745604515076, -0.6200095415115356, -0.8941694498062134, -0.3983449339866638, 0.158361062407493...
null
null
null
null
null
null
null
null
null
null
null
null
null
Nexdata/Hong_Kong_Cantonese_Average_Tone_Speech_Synthesis_Corpus
Nexdata
2023-11-10T07:28:26Z
13
0
null
[ "task_categories:text-to-speech", "region:us" ]
2023-11-10T07:28:26Z
2022-06-22T06:29:49.000Z
2022-06-22T06:29:49
--- task_categories: - text-to-speech --- # Dataset Card for Nexdata/Hong_Kong_Cantonese_Average_Tone_Speech_Synthesis_Corpus ## Description 38 People - Hong Kong Cantonese Average Tone Speech Synthesis Corpus, It is recorded by Hong Kong native speakers. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: https://www.nexdata.ai/datasets/1201?source=Huggingface # Specifications ## Format 44,100Hz, 16bit, uncompressed wav, mono channel; ## Recording environment quiet indoor environment, low background noise, without echo; ## Recording content news and colloquial sentences; ## Speaker 9 males, 29 females; ## Device microphone; ## Language Cantonese, English; ## Annotation word and phoneme transcription, prosodic boundary annotation; ## Application scenarios speech synthesis. # Licensing Information Commercial License
[ -0.5782986879348755, -0.5346205830574036, -0.05397948622703552, 0.7543428540229797, -0.20985743403434753, 0.029333626851439476, 0.019770175218582153, -0.34381210803985596, 0.7241925597190857, 0.48058798909187317, -0.4898296296596527, -0.9191222190856934, -0.10315611958503723, 0.24369911849...
null
null
null
null
null
null
null
null
null
null
null
null
null
Nexdata/Chinese_Mandarin_Speech_Synthesis_Corpus-Female_Imitating_Children
Nexdata
2023-08-30T10:37:14Z
13
0
null
[ "region:us" ]
2023-08-30T10:37:14Z
2022-06-22T06:44:34.000Z
2022-06-22T06:44:34
--- YAML tags: - copy-paste the tags obtained with the tagging app: https://github.com/huggingface/datasets-tagging --- # Dataset Card for Nexdata/Chinese_Mandarin_Speech_Synthesis_Corpus-Female_Imitating_Children ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1091?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary Female audio data of adults imitating children, 6599 sentences in total and 6.78 hours. It is recorded by Chinese native speakers, with authentic accent and sweet sound. The phoneme coverage is balanced. Professional phonetician participates in the annotation. It precisely matches with the research and development needs of the speech synthesis. For more details, please refer to the link: https://www.nexdata.ai/datasets/1091?source=Huggingface ### Supported Tasks and Leaderboards tts: The dataset can be used to train a model for Text to Speech (TTS). ### Languages Mandarin Chinese ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
[ -0.32478588819503784, -0.671369194984436, -0.15203078091144562, 0.4658643305301666, -0.07453054189682007, 0.03561443090438843, -0.284728467464447, -0.40323591232299805, 0.47192803025245667, 0.5587317943572998, -0.8008309006690979, -0.8488197326660156, -0.4153824746608734, 0.144513517618179...
null
null
null
null
null
null
null
null
null
null
null
null
null
Nexdata/Sichuan_Dialect_Speech_Data_by_Mobile_Phone
Nexdata
2023-08-30T10:41:35Z
13
0
null
[ "region:us" ]
2023-08-30T10:41:35Z
2022-06-22T06:47:11.000Z
2022-06-22T06:47:11
--- YAML tags: - copy-paste the tags obtained with the tagging app: https://github.com/huggingface/datasets-tagging --- # Dataset Card for Nexdata/Sichuan_Dialect_Speech_Data_by_Mobile_Phone ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/52?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary It collects 2,507 speakers from Sichuan Basin and is recorded in quiet indoor environment. The recorded content covers customer consultation and text messages in many fields. The average number of repetitions is 1.3 and the average sentence length is 12.5 words. Sichuan natives participate in quality inspection and proofreading to ensure the accuracy of the text transcription. For more details, please refer to the link: https://www.nexdata.ai/datasets/52?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Sichuan Dialect ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
[ -0.4030243158340454, -0.5431419610977173, -0.15489724278450012, 0.5324883460998535, -0.17368686199188232, -0.18357619643211365, -0.49006858468055725, -0.3750428855419159, 0.4709964394569397, 0.563991367816925, -0.5898035764694214, -0.885397732257843, -0.5027638077735901, 0.2107145637273788...
null
null
null
null
null
null
null
null
null
null
null
null
null
Nexdata/Korean_Speech_Data_by_Mobile_Phone_Guiding
Nexdata
2023-08-30T10:38:49Z
13
0
null
[ "region:us" ]
2023-08-30T10:38:49Z
2022-06-22T06:48:33.000Z
2022-06-22T06:48:33
--- YAML tags: - copy-paste the tags obtained with the tagging app: https://github.com/huggingface/datasets-tagging --- # Dataset Card for Nexdata/Korean_Speech_Data_by_Mobile_Phone_Guiding ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/61?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary It collects 211 Korean locals and is recorded in quiet indoor environment. 99 females, 112 males. Recording devices are mainstream Android phones and iPhones. For more details, please refer to the link: https://www.nexdata.ai/datasets/61?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Korean ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
[ -0.4510541260242462, -0.5149637460708618, -0.01344594731926918, 0.4139609932899475, -0.10474998503923416, 0.07964155822992325, -0.31559517979621887, -0.3846301734447479, 0.6008352041244507, 0.6768893003463745, -0.8202303647994995, -1.0467610359191895, -0.5717546343803406, 0.122414641082286...
null
null
null
null
null
null
null
null
null
null
null
null
null
Nexdata/Taiwanese_Mandarin_Speech_Data_by_Mobile_Phone_Guiding
Nexdata
2023-08-30T10:39:25Z
13
0
null
[ "region:us" ]
2023-08-30T10:39:25Z
2022-06-22T06:56:08.000Z
2022-06-22T06:56:08
--- YAML tags: - copy-paste the tags obtained with the tagging app: https://github.com/huggingface/datasets-tagging --- # Dataset Card for Nexdata/Taiwanese_Mandarin_Speech_Data_by_Mobile_Phone_Guiding ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/64?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary The data collected 203 Taiwan people, covering Taipei, Kaohsiung, Taichung, Tainan, etc. 137 females, 66 males. It is recorded in quiet indoor environment. It can be used in speech recognition, machine translation, voiceprint recognition model training and algorithm research. For more details, please refer to the link: https://www.nexdata.ai/datasets/64?source=Huggingface ### Supported Tasks and Leaderboards automatic-speech-recognition, audio-speaker-identification: The dataset can be used to train a model for Automatic Speech Recognition (ASR). ### Languages Taiwanese Mandarin ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
[ -0.3281753957271576, -0.6601064801216125, -0.12466676533222198, 0.36825689673423767, -0.269060879945755, 0.10615433752536774, -0.27883440256118774, -0.47372087836265564, 0.5875499248504639, 0.6507775187492371, -0.6788925528526306, -0.85504549741745, -0.5194734334945679, 0.15844836831092834...
null
null
null
null
null
null
null
null
null
null
null
null
null
GEM-submissions/lewtun__this-is-a-test-name__1655928558
GEM-submissions
2022-06-22T20:09:24Z
13
0
null
[ "benchmark:gem", "evaluation", "benchmark", "region:us" ]
2022-06-22T20:09:24Z
2022-06-22T20:09:21.000Z
2022-06-22T20:09:21
--- benchmark: gem type: prediction submission_name: This is a test name tags: - evaluation - benchmark --- # GEM Submission Submission name: This is a test name
[ 0.05652708560228348, -0.9112688899040222, 0.6085131764411926, 0.11336709558963776, -0.19541293382644653, 0.5379477739334106, 0.13777051866054535, 0.34700918197631836, 0.4342630207538605, 0.346922904253006, -1.0974667072296143, -0.14333292841911316, -0.5057262778282166, -0.01758344098925590...
null
null
null
null
null
null
null
null
null
null
null
null
null
martosinc/morrowtext
martosinc
2022-06-22T23:17:49Z
13
0
null
[ "license:mit", "region:us" ]
2022-06-22T23:17:49Z
2022-06-22T23:10:16.000Z
2022-06-22T23:10:16
--- license: mit --- Contains all TES3:Morrowind dialogues and journal queries. There are in total 4 labels: Journal, Greeting, Persuasion, Topic (Last one being the usual dialogues). The text is already formatted and does not contain duplicates or NaNs.
[ -0.2828662097454071, -0.5672226548194885, 0.8096381425857544, 0.44372493028640747, -0.1460341364145279, -0.22332894802093506, 0.08111917972564697, -0.18562045693397522, 0.6481438875198364, 0.9023251533508301, -0.9733201265335083, -0.7005314230918884, -0.3662962019443512, 0.4803466200828552...
null
null
null
null
null
null
null
null
null
null
null
null
null
vimal-quilt/tweet-eval-emotion
vimal-quilt
2022-06-23T03:39:59Z
13
0
null
[ "region:us" ]
2022-06-23T03:39:59Z
2022-06-23T03:37:10.000Z
2022-06-23T03:37:10
Entry not found
[ -0.32276472449302673, -0.22568407654762268, 0.8622258901596069, 0.4346148371696472, -0.5282984972000122, 0.7012965679168701, 0.7915717363357544, 0.07618629932403564, 0.7746022939682007, 0.2563222646713257, -0.785281777381897, -0.22573848068714142, -0.9104482531547546, 0.5715669393539429, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-e1907042-7494831
autoevaluate
2022-06-26T11:26:20Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-06-26T11:26:20Z
2022-06-26T11:25:44.000Z
2022-06-26T11:25:44
--- type: predictions tags: - autotrain - evaluation datasets: - clinc_oos eval_info: task: multi_class_classification model: Omar95farag/distilbert-base-uncased-distilled-clinc metrics: [] dataset_name: clinc_oos dataset_config: small dataset_split: test col_mapping: text: text target: intent --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Multi-class Text Classification * Model: Omar95farag/distilbert-base-uncased-distilled-clinc * Dataset: clinc_oos To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
[ -0.31949207186698914, -0.26833996176719666, 0.3607659935951233, 0.14645831286907196, 0.0040358672849833965, -0.04914836585521698, -0.10591689497232437, -0.43623045086860657, -0.0530460961163044, 0.40053316950798035, -0.7586050033569336, -0.2800822854042053, -0.8490405082702637, 0.070189878...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-e1907042-7494832
autoevaluate
2022-06-26T11:26:25Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-06-26T11:26:25Z
2022-06-26T11:25:49.000Z
2022-06-26T11:25:49
--- type: predictions tags: - autotrain - evaluation datasets: - clinc_oos eval_info: task: multi_class_classification model: abdelkader/distilbert-base-uncased-distilled-clinc metrics: [] dataset_name: clinc_oos dataset_config: small dataset_split: test col_mapping: text: text target: intent --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Multi-class Text Classification * Model: abdelkader/distilbert-base-uncased-distilled-clinc * Dataset: clinc_oos To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
[ -0.32194066047668457, -0.2816174030303955, 0.3827234208583832, 0.10699622333049774, 0.003293168032541871, -0.10404712706804276, -0.11786581575870514, -0.42947590351104736, -0.07240438461303711, 0.3833231031894684, -0.7860928177833557, -0.27853214740753174, -0.8555137515068054, 0.0856940746...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-e1907042-7494835
autoevaluate
2022-06-26T11:26:45Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-06-26T11:26:45Z
2022-06-26T11:26:07.000Z
2022-06-26T11:26:07
--- type: predictions tags: - autotrain - evaluation datasets: - clinc_oos eval_info: task: multi_class_classification model: jackmleitch/distilbert-base-uncased-distilled-clinc metrics: [] dataset_name: clinc_oos dataset_config: small dataset_split: test col_mapping: text: text target: intent --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Multi-class Text Classification * Model: jackmleitch/distilbert-base-uncased-distilled-clinc * Dataset: clinc_oos To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
[ -0.3433530628681183, -0.30816471576690674, 0.35031887888908386, 0.12738046050071716, -0.0031165170948952436, -0.10865698009729385, -0.09273131191730499, -0.441851407289505, -0.03247803822159767, 0.3959439992904663, -0.7833848595619202, -0.26483628153800964, -0.8351922631263733, 0.096942335...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-e1907042-7494833
autoevaluate
2022-06-26T11:29:12Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-06-26T11:29:12Z
2022-06-26T11:26:09.000Z
2022-06-26T11:26:09
--- type: predictions tags: - autotrain - evaluation datasets: - clinc_oos eval_info: task: multi_class_classification model: aytugkaya/distilbert-base-uncased-finetuned-clinc metrics: [] dataset_name: clinc_oos dataset_config: small dataset_split: test col_mapping: text: text target: intent --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Multi-class Text Classification * Model: aytugkaya/distilbert-base-uncased-finetuned-clinc * Dataset: clinc_oos To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
[ -0.31203827261924744, -0.33138152956962585, 0.3543614447116852, 0.11892668902873993, 0.011499466374516487, -0.16472606360912323, -0.14739196002483368, -0.48598596453666687, -0.050841912627220154, 0.3747353255748749, -0.8550172448158264, -0.30836862325668335, -0.7470162510871887, 0.01838394...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-e1907042-7494836
autoevaluate
2022-06-26T11:26:51Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-06-26T11:26:51Z
2022-06-26T11:26:13.000Z
2022-06-26T11:26:13
--- type: predictions tags: - autotrain - evaluation datasets: - clinc_oos eval_info: task: multi_class_classification model: moshew/distilbert-base-uncased-finetuned-clinc metrics: [] dataset_name: clinc_oos dataset_config: small dataset_split: test col_mapping: text: text target: intent --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Multi-class Text Classification * Model: moshew/distilbert-base-uncased-finetuned-clinc * Dataset: clinc_oos To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
[ -0.38052496314048767, -0.3064555525779724, 0.3406575620174408, 0.1091904416680336, 0.01989217847585678, -0.18188518285751343, -0.11674565821886063, -0.4794706106185913, -0.05930149182677269, 0.3945445120334625, -0.8394550681114197, -0.31698304414749146, -0.7823950052261353, 0.0173779763281...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-e1907042-7494834
autoevaluate
2022-06-26T11:29:24Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-06-26T11:29:24Z
2022-06-26T11:26:18.000Z
2022-06-26T11:26:18
--- type: predictions tags: - autotrain - evaluation datasets: - clinc_oos eval_info: task: multi_class_classification model: calcworks/distilbert-base-uncased-distilled-clinc metrics: [] dataset_name: clinc_oos dataset_config: small dataset_split: test col_mapping: text: text target: intent --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Multi-class Text Classification * Model: calcworks/distilbert-base-uncased-distilled-clinc * Dataset: clinc_oos To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
[ -0.28226587176322937, -0.2623085379600525, 0.40725159645080566, 0.14612950384616852, 0.05031273514032364, -0.06574318557977676, -0.08689369261264801, -0.3947750926017761, -0.06106806546449661, 0.414557546377182, -0.7575061321258545, -0.28824758529663086, -0.8363151550292969, 0.065889842808...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-019e0f0d-7644945
autoevaluate
2022-06-26T23:46:29Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-06-26T23:46:29Z
2022-06-26T20:21:02.000Z
2022-06-26T20:21:02
--- type: predictions tags: - autotrain - evaluation datasets: - scientific_papers eval_info: task: summarization model: google/bigbird-pegasus-large-pubmed metrics: [] dataset_name: scientific_papers dataset_config: pubmed dataset_split: test col_mapping: text: article target: abstract --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: google/bigbird-pegasus-large-pubmed * Dataset: scientific_papers To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
[ -0.3518012464046478, -0.19682638347148895, 0.3941453993320465, 0.21348708868026733, -0.08354036509990692, -0.28260329365730286, 0.028309855610132217, -0.45023077726364136, 0.3896665573120117, 0.3712954521179199, -0.8963701725006104, -0.23054854571819305, -0.687137246131897, 0.0808169320225...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-d47ba8c2-7654948
autoevaluate
2022-06-26T23:44:04Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-06-26T23:44:04Z
2022-06-26T20:22:09.000Z
2022-06-26T20:22:09
--- type: predictions tags: - autotrain - evaluation datasets: - scientific_papers eval_info: task: summarization model: google/bigbird-pegasus-large-arxiv metrics: [] dataset_name: scientific_papers dataset_config: arxiv dataset_split: test col_mapping: text: article target: abstract --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: google/bigbird-pegasus-large-arxiv * Dataset: scientific_papers To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
[ -0.42531710863113403, -0.2194414734840393, 0.319227010011673, 0.2434995472431183, -0.1096610352396965, -0.29965341091156006, 0.03896467015147209, -0.480110764503479, 0.35373005270957947, 0.3629288673400879, -0.9244767427444458, -0.19262343645095825, -0.6723511815071106, 0.03721149638295173...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-d47ba8c2-7654949
autoevaluate
2022-06-26T23:45:21Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-06-26T23:45:21Z
2022-06-26T20:22:15.000Z
2022-06-26T20:22:15
--- type: predictions tags: - autotrain - evaluation datasets: - scientific_papers eval_info: task: summarization model: google/bigbird-pegasus-large-pubmed metrics: [] dataset_name: scientific_papers dataset_config: arxiv dataset_split: test col_mapping: text: article target: abstract --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: google/bigbird-pegasus-large-pubmed * Dataset: scientific_papers To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
[ -0.35180070996284485, -0.19682584702968597, 0.394145131111145, 0.21348708868026733, -0.08354011178016663, -0.28260308504104614, 0.02831018902361393, -0.4502306878566742, 0.38966622948646545, 0.3712952733039856, -0.896369993686676, -0.2305484563112259, -0.6871371865272522, 0.080816946923732...
null
null
null
null
null
null
null
null
null
null
null
null
null
carlosejimenez/flickr30k_captions_simCSE
carlosejimenez
2022-06-26T23:04:54Z
13
0
null
[ "region:us" ]
2022-06-26T23:04:54Z
2022-06-26T22:27:46.000Z
2022-06-26T22:27:46
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622264862060547, 0.43461528420448303, -0.52829909324646, 0.7012971639633179, 0.7915720343589783, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104477167129517, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
Nexdata/Human_Facial_Skin_Defects_Data
Nexdata
2023-08-31T02:40:21Z
13
1
null
[ "region:us" ]
2023-08-31T02:40:21Z
2022-06-27T08:53:34.000Z
2022-06-27T08:53:34
--- YAML tags: - copy-paste the tags obtained with the tagging app: https://github.com/huggingface/datasets-tagging --- # Dataset Card for Nexdata/Human_Facial_Skin_Defects_Data ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1052?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 4,788 Chinese people 5,105 images Human Facial Skin Defects Data. The data includes the following five types of facial skin defects: acne, acne marks, stains, wrinkles and dark circles. This data can be used for tasks such as skin defects detection. For more details, please refer to the link: https://www.nexdata.ai/datasets/1052?source=Huggingface ### Supported Tasks and Leaderboards face-detection, computer-vision: The dataset can be used to train a model for face detection. ### Languages English ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
[ -0.40658465027809143, -0.6985246539115906, 0.10469955205917358, 0.23981395363807678, -0.09863337874412537, 0.02823297493159771, -0.008773049339652061, -0.6064411401748657, 0.7048903703689575, 0.8691247701644897, -0.8061517477035522, -1.147760033607483, -0.37564077973365784, -0.002353720832...
null
null
null
null
null
null
null
null
null
null
null
null
null
Nexdata/Re-ID_Data_in_Surveillance_Scenes
Nexdata
2023-08-31T02:20:46Z
13
0
null
[ "region:us" ]
2023-08-31T02:20:46Z
2022-06-27T09:01:22.000Z
2022-06-27T09:01:22
--- YAML tags: - copy-paste the tags obtained with the tagging app: https://github.com/huggingface/datasets-tagging --- # Dataset Card for Nexdata/Re-ID_Data_in_Surveillance_Scenes ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://www.nexdata.ai/datasets/1129?source=Huggingface - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary 10,000 People - Re-ID Data in Surveillance Scenes. The data includes indoor scenes and outdoor scenes. The data includes males and females, and the age distribution is from children to the elderly. The data diversity includes different age groups, different time periods, different shooting angles, different human body orientations and postures, clothing for different seasons. For annotation, the rectangular bounding boxes and 15 attributes of human body were annotated. The data can be used for re-id and other tasks. For more details, please refer to the link: https://www.nexdata.ai/datasets/1129?source=Huggingface ### Supported Tasks and Leaderboards face-detection, computer-vision: The dataset can be used to train a model for face detection. ### Languages ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information Commerical License: https://drive.google.com/file/d/1saDCPm74D4UWfBL17VbkTsZLGfpOQj1J/view?usp=sharing ### Citation Information [More Information Needed] ### Contributions
[ -0.6654296517372131, -0.6003051996231079, 0.17197982966899872, 0.1583690345287323, -0.19097965955734253, 0.02731768600642681, 0.13537205755710602, -0.5031320452690125, 0.7713476419448853, 0.6443144083023071, -0.9605520367622375, -1.0390599966049194, -0.6385934948921204, 0.20034746825695038...
null
null
null
null
null
null
null
null
null
null
null
null
null
knibu/best_sellers_june2022
knibu
2022-06-27T15:18:12Z
13
0
null
[ "region:us" ]
2022-06-27T15:18:12Z
2022-06-27T15:18:05.000Z
2022-06-27T15:18:05
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622264862060547, 0.43461528420448303, -0.52829909324646, 0.7012971639633179, 0.7915720343589783, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104477167129517, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
davanstrien/testhmd
davanstrien
2022-06-28T11:14:43Z
13
0
null
[ "region:us" ]
2022-06-28T11:14:43Z
2022-06-27T16:50:49.000Z
2022-06-27T16:50:49
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622264862060547, 0.43461528420448303, -0.52829909324646, 0.7012971639633179, 0.7915720343589783, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104477167129517, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-d42d3c12-7815006
autoevaluate
2022-06-27T20:36:00Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-06-27T20:36:00Z
2022-06-27T20:33:13.000Z
2022-06-27T20:33:13
--- type: predictions tags: - autotrain - evaluation datasets: - xtreme eval_info: task: entity_extraction model: jg/xlm-roberta-base-finetuned-panx-de metrics: [] dataset_name: xtreme dataset_config: PAN-X.de dataset_split: test col_mapping: tokens: tokens tags: ner_tags --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Token Classification * Model: jg/xlm-roberta-base-finetuned-panx-de * Dataset: xtreme To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
[ -0.4211260676383972, -0.1634199470281601, 0.21589486300945282, 0.06297922134399414, -0.01105608232319355, -0.07975266873836517, 0.03087488003075123, -0.4349895119667053, 0.17342139780521393, 0.5043210983276367, -0.9389719367027283, -0.3818478286266327, -0.7219592332839966, -0.1004316732287...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-c967fc98-8385125
autoevaluate
2022-06-29T01:09:37Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-06-29T01:09:37Z
2022-06-28T21:43:42.000Z
2022-06-28T21:43:42
--- type: predictions tags: - autotrain - evaluation datasets: - scientific_papers eval_info: task: summarization model: google/bigbird-pegasus-large-arxiv metrics: ['bertscore', 'meteor'] dataset_name: scientific_papers dataset_config: pubmed dataset_split: test col_mapping: text: article target: abstract --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: google/bigbird-pegasus-large-arxiv * Dataset: scientific_papers To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@Blaise_g](https://huggingface.co/Blaise_g) for evaluating this model.
[ -0.42039141058921814, -0.22555091977119446, 0.3265030086040497, 0.2382073849439621, -0.08488292247056961, -0.30111590027809143, 0.015598186291754246, -0.48111170530319214, 0.3385500907897949, 0.3945448100566864, -0.9255398511886597, -0.2058098316192627, -0.6951705813407898, 0.0431844778358...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-72edae24-8665151
autoevaluate
2022-06-30T05:04:02Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-06-30T05:04:02Z
2022-06-29T02:15:21.000Z
2022-06-29T02:15:21
--- type: predictions tags: - autotrain - evaluation datasets: - cnn_dailymail eval_info: task: summarization model: eslamxm/mbart-finetune-en-cnn metrics: [] dataset_name: cnn_dailymail dataset_config: 3.0.0 dataset_split: train col_mapping: text: article target: highlights --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: eslamxm/mbart-finetune-en-cnn * Dataset: cnn_dailymail To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@](https://huggingface.co/) for evaluating this model.
[ -0.5430601239204407, -0.3788489103317261, 0.10842783004045486, 0.16845019161701202, -0.14706912636756897, -0.16288210451602936, 0.00938444398343563, -0.33508649468421936, 0.22479328513145447, 0.42847633361816406, -0.9740955233573914, -0.28103163838386536, -0.6808013916015625, -0.0654140636...
null
null
null
null
null
null
null
null
null
null
null
null
null
joelniklaus/german_argument_mining
joelniklaus
2022-09-22T13:44:35Z
13
3
null
[ "task_categories:text-classification", "task_ids:multi-class-classification", "annotations_creators:expert-generated", "annotations_creators:found", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:original", "language:de", "license:cc-by-4.0...
2022-09-22T13:44:35Z
2022-07-01T11:30:58.000Z
2022-07-01T11:30:58
--- annotations_creators: - expert-generated - found language_creators: - found language: - de license: - cc-by-4.0 multilinguality: - monolingual pretty_name: Annotated German Legal Decision Corpus size_categories: - 10K<n<100K source_datasets: - original task_categories: - text-classification task_ids: - multi-class-classification --- # Dataset Card for Annotated German Legal Decision Corpus ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** - **Repository:** https://zenodo.org/record/3936490#.X1ed7ovgomK - **Paper:** Urchs., S., Mitrović., J., & Granitzer., M. (2021). Design and Implementation of German Legal Decision Corpora. Proceedings of the 13th International Conference on Agents and Artificial Intelligence - Volume 2: ICAART, 515–521. https://doi.org/10.5220/0010187305150521 - **Leaderboard:** - **Point of Contact:** [Joel Niklaus](mailto:joel.niklaus.2@bfh.ch) ### Dataset Summary This dataset consists of 200 randomly chosen judgments. In these judgments a legal expert annotated the components conclusion, definition and subsumption of the German legal writing style Urteilsstil. *"Overall 25,075 sentences are annotated. 5% (1,202) of these sentences are marked as conclusion, 21% (5,328) as definition, 53% (13,322) are marked as subsumption and the remaining 21% (6,481) as other. The length of judgments in sentences ranges from 38 to 862 sentences. The median of judgments have 97 sentences, the length of most judgments is on the shorter side."* (Urchs. et al., 2021) *"Judgments from 22 of the 131 courts are selected for the corpus. Most judgments originate from the VG Augsburg (59 / 30%) followed by the VG Ansbach (39 / 20%) and LSG Munich (33 / 17%)."* (Urchs. et al., 2021) *"29% (58) of all selected judgments are issued in the year 2016, followed by 22% (44) from the year 2017 and 21% (41) issued in the year 2015. [...] The percentages of selected judgments and decisions issued in 2018 and 2019 are roughly the same. No judgments from 2020 are selected."* (Urchs. et al., 2021) ### Supported Tasks and Leaderboards The dataset can be used for multi-class text classification tasks, more specifically, for argument mining. ### Languages The language in the dataset is German as it is used in Bavarian courts in Germany. ## Dataset Structure ### Data Instances Each sentence is saved as a json object on a line in one of the three files `train.jsonl`, `validation.jsonl` or `test.jsonl`. The file `meta.jsonl` contains meta information for each court. The `file_number` is present in all files for identification. Each sentence of the court decision was categorized according to its function. ### Data Fields The file `meta.jsonl` contains for each row the following fields: - `meta_title`: Title provided by the website, it is used for saving the decision - `court`: Issuing court - `decision_style`: Style of the decision; the corpus contains either *Urteil* (='judgment') or *Endurteil* ( ='end-judgment') - `date`: Date when the decision was issued by the court - `file_number`: Identification number used for this decision by the court - `title`: Title provided by the court - `norm_chains`: Norms related to the decision - `decision_guidelines`: Short summary of the decision - `keywords`: Keywords associated with the decision - `lower_court`: Court that decided on the decision before - `additional_information`: Additional Information - `decision_reference`: References to the location of the decision in beck-online - `tenor`: Designation of the legal consequence ordered by the court (list of paragraphs) - `legal_facts`: Facts that form the base for the decision (list of paragraphs) The files `train.jsonl`, `validation.jsonl` and `test.jsonl` contain the following fields: - `file_number`: Identification number for linkage with the file `meta.jsonl` - `input_sentence`: The sentence to be classified - `label`: In depth explanation of the court decision. Each sentence is assigned to one of the major components of German *Urteilsstil* (Urchs. et al., 2021) (list of paragraphs, each paragraph containing list of sentences, each sentence annotated with one of the following four labels): - `conclusion`: Overall result - `definition`: Abstract legal facts and consequences - `subsumption`: Determination sentence / Concrete facts - `other`: Anything else - `context_before`: Context in the same paragraph before the input_sentence - `context_after`: Context in the same paragraph after the input_sentence ### Data Splits No split provided in the original release. Splits created by Joel Niklaus. We randomly split the dataset into 80% (160 decisions, 19271 sentences) train, 10% validation (20 decisions, 2726 sentences) and 10% test (20 decisions, 3078 sentences). We made sure, that a decision only occurs in one split and is not dispersed over multiple splits. Label Distribution | label | train | validation | test | |:---------------|-----------:|-------------:|----------:| | conclusion | 975 | 115 | 112 | | definition | 4105 | 614 | 609 | | subsumption | 10034 | 1486 | 1802 | | other | 4157 | 511 | 555 | | total | **19271** | **2726** | **3078** | ## Dataset Creation ### Curation Rationale Creating a publicly available German legal text corpus consisting of judgments that have been annotated by a legal expert. The annotated components consist of *conclusion*, *definition* and *subsumption* of the German legal writing style *Urteilsstil*. ### Source Data #### Initial Data Collection and Normalization *“The decision corpus is a collection of the decisions published on the website www.gesetze-bayern.de. At the time of the crawling the website offered 32,748 decisions of 131 Bavarian courts, dating back to 2015. The decisions are provided from the Bavarian state after the courts agreed to a publication. All decisions are processed by the publisher C.H.BECK, commissioned by the Bavarian state. This processing includes anonymisation, key-wording, and adding of editorial guidelines to the decisions.”* (Urchs. et al., 2021) #### Who are the source language producers? German courts from Bavaria ### Annotations #### Annotation process *“As stated above, the judgment corpus consist of 200 randomly chosen judgments that are annotated by a legal expert, who holds a first legal state exam. Due to financial, staff and time reasons the presented iteration of the corpus was only annotated by a single expert. In a future version several other experts will annotate the corpus and the inter-annotator agreement will be calculated.”* (Urchs. et al., 2021) #### Who are the annotators? A legal expert, who holds a first legal state exam. ### Personal and Sensitive Information *"All decisions are processed by the publisher C.H.BECK, commissioned by the Bavarian state. This processing includes ** anonymisation**, key-wording, and adding of editorial guidelines to the decisions.”* (Urchs. et al., 2021) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations The SoMaJo Sentence Splitter has been used. Upon manual inspection of the dataset, we could see that the sentence splitter had poor accuracy in some cases (see ```analyze_dataset()``` in ```convert_to_hf_dataset.py```). When creating the splits, we thought about merging small sentences with their neighbors or removing them all together. However, since we could not find an straightforward way to do this, we decided to leave the dataset content untouched. Note that the information given in this dataset card refer to the dataset version as provided by Joel Niklaus and Veton Matoshi. The dataset at hand is intended to be part of a bigger benchmark dataset. Creating a benchmark dataset consisting of several other datasets from different sources requires postprocessing. Therefore, the structure of the dataset at hand, including the folder structure, may differ considerably from the original dataset. In addition to that, differences with regard to dataset statistics as give in the respective papers can be expected. The reader is advised to have a look at the conversion script ```convert_to_hf_dataset.py``` in order to retrace the steps for converting the original dataset into the present jsonl-format. For further information on the original dataset structure, we refer to the bibliographical references and the original Github repositories and/or web pages provided in this dataset card. ## Additional Information ### Dataset Curators The names of the original dataset curators and creators can be found in references given below, in the section *Citation Information*. Additional changes were made by Joel Niklaus ([Email](mailto:joel.niklaus.2@bfh.ch) ; [Github](https://github.com/joelniklaus)) and Veton Matoshi ([Email](mailto:veton.matoshi@bfh.ch) ; [Github](https://github.com/kapllan)). ### Licensing Information [Creative Commons Attribution 4.0 International](https://creativecommons.org/licenses/by/4.0/legalcode) ### Citation Information ``` @dataset{urchs_stefanie_2020_3936490, author = {Urchs, Stefanie and Mitrović, Jelena}, title = {{German legal jugements annotated with judement style components}}, month = jul, year = 2020, publisher = {Zenodo}, doi = {10.5281/zenodo.3936490}, url = {https://doi.org/10.5281/zenodo.3936490} } ``` ``` @conference{icaart21, author = {Urchs., Stefanie and Mitrovi{\'{c}}., Jelena and Granitzer., Michael}, booktitle = {Proceedings of the 13th International Conference on Agents and Artificial Intelligence - Volume 2: ICAART,}, doi = {10.5220/0010187305150521}, isbn = {978-989-758-484-8}, issn = {2184-433X}, organization = {INSTICC}, pages = {515--521}, publisher = {SciTePress}, title = {{Design and Implementation of German Legal Decision Corpora}}, year = {2021} } ``` ### Contributions Thanks to [@kapllan](https://github.com/kapllan) and [@joelniklaus](https://github.com/joelniklaus) for adding this dataset.
[ -0.5094854831695557, -0.7207943201065063, 0.7167198657989502, 0.11467380076646805, -0.43842339515686035, -0.4282744526863098, -0.36066651344299316, -0.25122082233428955, 0.14490598440170288, 0.5950449109077454, -0.34376445412635803, -1.001030445098877, -0.791209876537323, 0.051024936139583...
null
null
null
null
null
null
null
null
null
null
null
null
null
joelniklaus/online_terms_of_service
joelniklaus
2022-09-22T13:45:42Z
13
5
null
[ "task_categories:text-classification", "task_ids:multi-class-classification", "task_ids:multi-label-classification", "annotations_creators:found", "annotations_creators:other", "language_creators:found", "multilinguality:multilingual", "size_categories:1K<n<10K", "source_datasets:original", "langu...
2022-09-22T13:45:42Z
2022-07-01T11:42:49.000Z
2022-07-01T11:42:49
--- annotations_creators: - found - other language_creators: - found language: - de - en - it - pl license: - other multilinguality: - multilingual pretty_name: A Corpus for Multilingual Analysis of Online Terms of Service size_categories: - 1K<n<10K source_datasets: - original task_categories: - text-classification task_ids: - multi-class-classification - multi-label-classification --- # Dataset Card for A Corpus for Multilingual Analysis of Online Terms of Service ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** - **Repository:** http://claudette.eui.eu/corpus_multilingual_NLLP2021.zip - **Paper:** Drawzeski, K., Galassi, A., Jablonowska, A., Lagioia, F., Lippi, M., Micklitz, H. W., Sartor, G., Tagiuri, G., & Torroni, P. (2021). A Corpus for Multilingual Analysis of Online Terms of Service. Proceedings of the Natural Legal Language Processing Workshop 2021, 1–8. https://doi.org/10.18653/v1/2021.nllp-1.1 - **Leaderboard:** - **Point of Contact:** [Joel Niklaus](mailto:joel.niklaus.2@bfh.ch) ### Dataset Summary *"We present the first annotated corpus for multilingual analysis of potentially unfair clauses in online Terms of Service [=ToS]. The data set comprises a total of 100 contracts, obtained from 25 documents annotated in four different languages: English, German, Italian, and Polish. For each contract, potentially unfair clauses for the consumer are annotated, for nine different unfairness categories."* (Drawzeski et al., 2021) ### Supported Tasks and Leaderboards The dataset can be used for multi-class multi-label text classification tasks, more specifically, for classifying unfair clauses in ToS. ### Languages English, German, Italian, and Polish. ## Dataset Structure ### Data Instances The file format is jsonl and three data splits are present (train, validation and test). ### Data Fields The dataset contains the following fields: - `language`: The language of the sentence/document. - `company`: The company of the document. - `line_number`: The line number of the sentence in the document. - `sentence`: The sentence to be classified. - `unfairness_level`: The unfairness level assigned to the sentence (if two clauses apply, the higher unfairness level is assigned here). The documents have been annotated using nine tags that represent different categories of clause unfairness. These boolean tags are: - `a` = Arbitration: *”This clause requires or allows the parties to resolve their disputes through an arbitration process, before the case could go to court. It is therefore considered a kind of forum selection clause. However, such a clause may or may not specify that arbitration should occur within a specific jurisdiction. Clauses stipulating that the arbitration should (1) take place in a state other than the state of consumer’s residence and/or (2) be based not on law but on arbiter’s discretion were marked as clearly unfair.”* (Lippi et al., 2019) - `ch` = Unilateral change: *"This clause specifies the conditions under which the service provider could amend and modify the terms of service and/or the service itself. Such clauses were always considered as potentially unfair. This is because the ECJ has not yet issued a judgment in this regard, though the Annex to the Direc- tive contains several examples supporting such a qualification."* (Lippi et al., 2019) - `cr` = Content removal : *"This gives the provider a right to modify/delete user’s content, including in-app purchases, and sometimes specifies the conditions under which the service provider may do so. As in the case of unilateral termination, clauses that indicate conditions for content removal were marked as potentially unfair, whereas clauses stipulating that the service provider may remove content in his full discretion, and/or at any time for any or no reasons and/or without notice nor possibility to retrieve the content were marked as clearly unfair."* (Lippi et al., 2019) - `j` = Jurisdiction : *"This type of clause stipulates what courts will have the competence to adjudicate disputes under the contract. Jurisdiction clauses giving consumers a right to bring disputes in their place of residence were marked as clearly fair, whereas clauses stating that any judicial proceeding takes a residence away (i.e. in a different city, different country) were marked as clearly unfair. This assessment is grounded in ECJ’s case law, see for example Oceano case number C-240/98."* (Lippi et al., 2019) - `law` = Choice of law: *"This clause specifies what law will govern the contract, meaning also what law will be applied in potential adjudication of a dispute arising under the contract. Clauses defining the applicable law as the law of the consumer’s country of residence were marked as clearly fair [...]"* (Lippi et al., 2019) - `ltd` = Limitation of liability: *"This clause stipulates that the duty to pay damages is limited or excluded, for certain kinds of losses and under certain conditions. Clauses that explicitly affirm non-excludable providers’ liabilities were marked as clearly fair."* (Lippi et al., 2019) - `ter` = Unilateral termination: *"This clause gives provider the right to suspend and/or terminate the service and/or the contract, and sometimes details the circumstances under which the provider claims to have a right to do so. Unilateral termination clauses that specify reasons for termination were marked as potentially unfair. Whereas clauses stipulating that the service provider may suspend or terminate the service at any time for any or no reasons and/or without notice were marked as clearly unfair."* (Lippi et al., 2019) - `use` = Contract by using: *"This clause stipulates that the consumer is bound by the terms of use of a specific service, simply by using the service, without even being required to mark that he or she has read and accepted them. We always marked such clauses as potentially unfair. The reason for this choice is that a good argument can be offered for these clauses to be unfair, because they originate an imbalance in rights and duties of the parties, but this argument has no decisive authoritative backing yet, since the ECJ has never assessed a clause of this type."* (Lippi et al., 2019) - `pinc` = Privacy included: This tag identifies *"clauses stating that consumers consent to the privacy policy simply by using the service. Such clauses have been always considered potentially unfair"* (Drawzeski et al., 2021) - `all_topics` = an aggregate column containing all applicable topics combined *”We assumed that each type of clause could be classified as either clearly fair, or potentially unfair, or clearly unfair. In order to mark the different degrees of (un)fairness we appended a numeric value to each XML tag, with 1 meaning clearly fair, 2 potentially unfair, and 3 clearly unfair. Nested tags were used to annotate text segments relevant to more than one type of clause. With clauses covering multiple paragraphs, we chose to tag each paragraph separately, possibly with different degrees of (un)fairness.”* (Lippi et al., 2019) ### Data Splits No splits provided in the original paper. Joel Niklaus created the splits manually. The train split contains the 20 (80%) first companies in alphabetic order (*Booking, Dropbox, Electronic_Arts, Evernote, Facebook, Garmin, Google, Grindr, Linkedin, Mozilla, Pinterest, Quora, Ryanair, Skype, Skyscanner, Snap, Spotify, Terravision, Tinder, Tripadvisor*). The validation split contains the 2 (8%) companies *Tumblr* and *Uber*. The test split contains the 3 (12%) companies *Weebly*, *Yelp*, *Zynga*. There are two tasks possible for this dataset. #### Clause Topics By only considering the clause topic, we separated the clause topic from the fairness level classification. Thus, the label set could be reduced to just 9 classes. This dataset poses a multi-label multi-class sentence classification problem. The following label distribution shows the number of occurrences per label per split. `total occurrences` sums up the previous rows (number of clause topics per split). `split size` is the number of sentences per split. | clause topic | train | validation | test | |:----------------------|------------:|-----------------:|-----------:| | a | 117 | 6 | 21 | | ch | 308 | 45 | 53 | | cr | 155 | 4 | 44 | | j | 206 | 8 | 36 | | law | 178 | 8 | 26 | | ltd | 714 | 84 | 161 | | ter | 361 | 39 | 83 | | use | 185 | 14 | 32 | | pinc | 71 | 0 | 8 | | **total occurrences** | **2295** | **208** | **464** | | **split size** | **19942** | **1690** | **4297** | #### Unfairness Levels When predicting unfairness levels, all untagged sentences can be removed. This reduces the dataset size considerably. This dataset poses a single-label multi-class sentence classification problem. | unfairness_level | train | validation | test | |:---------------------------|------------:|-----------:|----------:| | untagged | 17868 | 1499 | 3880 | | potentially_unfair | 1560 | 142 | 291 | | clearly_unfair | 259 | 31 | 65 | | clearly_fair | 156 | 5 | 32 | | **total without untagged** | **1975** | **178** | **388** | | **total** | **19942** | **1690** | **4297** | ## Dataset Creation ### Curation Rationale The EU legislation is published in all official languages. This multilingualism comes with costs and challenges, such as limited cross-linguistical interpretability. The EU has refrained from regulating languages in which standard terms in consumer contracts should be drafted, allowing for differing approaches to emerge in various jurisdictions. Consumer protection authorities and non-governmental organizations in Europe tend to operate only in their respective languages. Therefore, consumer protection technologies are needed that are capable of dealing with multiple languages. The dataset at hand can be used for the automated detection of unfair clauses in ToS which, in most cases, are available in multiple languages. (Drawzeski et al., 2021) ### Source Data #### Initial Data Collection and Normalization *"The analysed ToS were retrieved from the [Claudette pre-existing corpus](http://claudette.eui.eu/ToS.zip), covering 100 English ToS (Lippi et al., 2019; Ruggeri et al., 2021). Such terms mainly concern popular digital services provided to consumers, including leading online platforms (such as search engines and social media). The predominant language of drafting of these ToS is English, with differing availability of corresponding ToS in other languages. To carry out the present study, the ultimate 25 ToS were selected on the basis of three main criteria: a) their availability in the four selected languages; b) the possibility of identifying a correspondence between the different versions, given their publication date; and c) the similarity of their structure (e.g. number of clauses, sections, etc.). To illustrate, while ToS in both German and Italian were identified for 63 out of the 100 ToS contained in the pre-existing Claudette training corpus, Polish versions were found for only 42 of these 63 ToS. Out of the 42 ToS available in the four languages, we selected those with the more closely corresponding versions based on criteria b) and c) above. Perfect correspondence across the 4 languages, however, could not be achieved for all 25 ToS."* (Drawzeski et al., 2021) #### Who are the source language producers? The source language producers are likely to be lawyers. ### Annotations #### Annotation process The dataset at hand is described by Drawzeski et al. (2021). The ToS of the dataset were retrieved from the pre-existing and mono-lingual (English) Claudette corpus which is described in (Lippi et al., 2019). Drawzeski et al. (2021) *“investigate methods for automatically transferring the annotations made on ToS in the context of the Claudette project onto the corresponding versions of the same documents in a target language, where such resources and expertise may be lacking.”* Therefore, in the following, we will present the annotation process for the Claudette corpus as described in (Lippi et al., 2019). *”The corpus consists of 50 relevant on-line consumer contracts, i.e., ToS of on-line platforms. Such contracts were selected among those offered by some of the major players in terms of number of users, global relevance, and time the service was established. Such contracts are usually quite detailed in content, are frequently updated to reflect changes both in the service and in the applicable law, and are often available in different versions for different jurisdictions. Given multiple versions of the same contract, we selected the most recent version available on-line to European customers. The mark-up was done in XML by three annotators, which jointly worked for the formulation of the annotation guidelines. The whole annotation process included several revisions, where some corrections were also suggested by an analysis of the false positives and false negatives retrieved by the initial machine learning prototypes. Due to the large interaction among the annotators during this process, in order to assess inter-annotation agreement, a further test set consisting of 10 additional contracts was tagged, following the final version of the guidelines. […] We produced an additional test set consisting of 10 more annotated contracts. Such documents were independently tagged by two distinct annotators who had carefully studied the guidelines. In order to quantitatively measure the inter-annotation agreement, for this test set we computed the standard Cohen’s 𝜅 metric […] which resulted to be 0.871 […].”* #### Who are the annotators? Not specified. ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases It is very likely that some ToS in German, Italian and Polish are direct translations from English. Drawzeski et al. (2021) write: *“Although we could not assess this comprehensively in the present study, we infer from the wording of the ToS that at least in 9 out of 25 cases, German, Italian and Polish documents were indeed translations of the English originals.”* ### Other Known Limitations Note that the information given in this dataset card refer to the dataset version as provided by Joel Niklaus and Veton Matoshi. The dataset at hand is intended to be part of a bigger benchmark dataset. Creating a benchmark dataset consisting of several other datasets from different sources requires postprocessing. Therefore, the structure of the dataset at hand, including the folder structure, may differ considerably from the original dataset. In addition to that, differences with regard to dataset statistics as give in the respective papers can be expected. The reader is advised to have a look at the conversion script ```convert_to_hf_dataset.py``` in order to retrace the steps for converting the original dataset into the present jsonl-format. For further information on the original dataset structure, we refer to the bibliographical references and the original Github repositories and/or web pages provided in this dataset card. ## Additional Information ### Dataset Curators The names of the original dataset curators and creators can be found in references given below, in the section *Citation Information*. Additional changes were made by Joel Niklaus ([Email](mailto:joel.niklaus.2@bfh.ch); [Github](https://github.com/joelniklaus)) and Veton Matoshi ([Email](mailto:veton.matoshi@bfh.ch); [Github](https://github.com/kapllan)). ### Licensing Information cc-by-nc-2.5 ### Citation Information ``` @inproceedings{drawzeski-etal-2021-corpus, address = {Punta Cana, Dominican Republic}, author = {Drawzeski, Kasper and Galassi, Andrea and Jablonowska, Agnieszka and Lagioia, Francesca and Lippi, Marco and Micklitz, Hans Wolfgang and Sartor, Giovanni and Tagiuri, Giacomo and Torroni, Paolo}, booktitle = {Proceedings of the Natural Legal Language Processing Workshop 2021}, doi = {10.18653/v1/2021.nllp-1.1}, month = {nov}, pages = {1--8}, publisher = {Association for Computational Linguistics}, title = {{A Corpus for Multilingual Analysis of Online Terms of Service}}, url = {https://aclanthology.org/2021.nllp-1.1}, year = {2021} } ``` ### Contributions Thanks to [@JoelNiklaus](https://github.com/joelniklaus) and [@kapllan](https://github.com/kapllan) for adding this dataset.
[ -0.5070019364356995, -0.42616909742355347, 0.48139894008636475, 0.3539533317089081, -0.3731659948825836, -0.19150391221046448, -0.06511253118515015, -0.74333655834198, 0.23090477287769318, 0.6771756410598755, -0.2964329719543457, -0.430957555770874, -0.4660895764827728, 0.2860812842845917,...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-854c0218-9415245
autoevaluate
2022-07-02T22:28:44Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-07-02T22:28:44Z
2022-07-02T22:28:07.000Z
2022-07-02T22:28:07
--- type: predictions tags: - autotrain - evaluation datasets: - emotion eval_info: task: multi_class_classification model: lewtun/sagemaker-distilbert-emotion metrics: [] dataset_name: emotion dataset_config: default dataset_split: test col_mapping: text: text target: label --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Multi-class Text Classification * Model: lewtun/sagemaker-distilbert-emotion * Dataset: emotion To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@gabrielaltay](https://huggingface.co/gabrielaltay) for evaluating this model.
[ -0.3877298831939697, -0.2390030175447464, 0.37889599800109863, 0.19549250602722168, 0.11279939860105515, -0.13436368107795715, -0.06497868150472641, -0.44824928045272827, 0.03676901012659073, 0.27198272943496704, -0.9652850031852722, -0.25893670320510864, -0.7740851640701294, 0.00685207638...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-562e1223-9425246
autoevaluate
2022-07-02T23:01:39Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-07-02T23:01:39Z
2022-07-02T23:00:47.000Z
2022-07-02T23:00:47
--- type: predictions tags: - autotrain - evaluation datasets: - bigscience-biomedical/tmp-scitail eval_info: task: binary_classification model: gabrielaltay/autotrain-at-test-bb-tmp-scitail-1078438446 metrics: [] dataset_name: bigscience-biomedical/tmp-scitail dataset_config: scitail_bigbio_te dataset_split: test col_mapping: text: premise target: label --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Binary Text Classification * Model: gabrielaltay/autotrain-at-test-bb-tmp-scitail-1078438446 * Dataset: bigscience-biomedical/tmp-scitail To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@gabrielaltay](https://huggingface.co/gabrielaltay) for evaluating this model.
[ -0.14881522953510284, -0.2819705903530121, 0.4090959429740906, 0.252414733171463, -0.027160221710801125, -0.01185489259660244, 0.013425585813820362, -0.5224810242652893, 0.1039905995130539, 0.33764299750328064, -0.8238725066184998, -0.45052656531333923, -0.8119018077850342, 0.0105761801823...
null
null
null
null
null
null
null
null
null
null
null
null
null
CloverSearch/distribution
CloverSearch
2022-07-06T02:38:34Z
13
0
null
[ "region:us" ]
2022-07-06T02:38:34Z
2022-07-05T07:13:18.000Z
2022-07-05T07:13:18
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622264862060547, 0.43461528420448303, -0.52829909324646, 0.7012971639633179, 0.7915720343589783, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104477167129517, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
Paul/hatecheck-polish
Paul
2022-07-05T10:26:41Z
13
1
null
[ "task_categories:text-classification", "task_ids:hate-speech-detection", "annotations_creators:crowdsourced", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:pl", "license:cc-by-4.0", "arxiv:2206.09917", "regi...
2022-07-05T10:26:41Z
2022-07-05T10:24:24.000Z
2022-07-05T10:24:24
--- annotations_creators: - crowdsourced language_creators: - expert-generated language: - pl license: - cc-by-4.0 multilinguality: - monolingual pretty_name: Polish HateCheck size_categories: - 1K<n<10K source_datasets: - original task_categories: - text-classification task_ids: - hate-speech-detection --- # Dataset Card for Multilingual HateCheck ## Dataset Description Multilingual HateCheck (MHC) is a suite of functional tests for hate speech detection models in 10 different languages: Arabic, Dutch, French, German, Hindi, Italian, Mandarin, Polish, Portuguese and Spanish. For each language, there are 25+ functional tests that correspond to distinct types of hate and challenging non-hate. This allows for targeted diagnostic insights into model performance. For more details, please refer to our paper about MHC, published at the 2022 Workshop on Online Abuse and Harms (WOAH) at NAACL 2022. If you are using MHC, please cite our work! - **Paper:** Röttger et al. (2022) - Multilingual HateCheck: Functional Tests for Multilingual Hate Speech Detection Models. https://arxiv.org/abs/2206.09917 - **Repository:** https://github.com/rewire-online/multilingual-hatecheck - **Point of Contact:** paul@rewire.online ## Dataset Structure The csv format mostly matches the original HateCheck data, with some adjustments for specific languages. **mhc_case_id** The test case ID that is unique to each test case across languages (e.g., "mandarin-1305") **functionality** The shorthand for the functionality tested by the test case (e.g, "target_obj_nh"). The same functionalities are tested in all languages, except for Mandarin and Arabic, where non-Latin script required adapting the tests for spelling variations. **test_case** The test case text. **label_gold** The gold standard label ("hateful" or "non-hateful") of the test case. All test cases within a given functionality have the same gold standard label. **target_ident** Where applicable, the protected group that is targeted or referenced in the test case. All HateChecks cover seven target groups, but their composition varies across languages. **ref_case_id** For hateful cases, where applicable, the ID of the hateful case which was perturbed to generate this test case. For non-hateful cases, where applicable, the ID of the hateful case which is contrasted by this test case. **ref_templ_id** The equivalent to ref_case_id, but for template IDs. **templ_id** The ID of the template from which the test case was generated. **case_templ** The template from which the test case was generated (where applicable). **gender_male** and **gender_female** For gender-inflected languages (French, Spanish, Portuguese, Hindi, Arabic, Italian, Polish, German), only for cases where gender inflection is relevant, separate entries for gender_male and gender_female replace case_templ. **label_annotated** A list of labels given by the three annotators who reviewed the test case (e.g., "['hateful', 'hateful', 'hateful']"). **label_annotated_maj** The majority vote of the three annotators (e.g., "hateful"). In some cases this differs from the gold label given by our language experts. **disagreement_in_case** True if label_annotated_maj does not match label_gold for the entry. **disagreement_in_template** True if the test case is generated from an IDENT template and there is at least one case with disagreement_in_case generated from the same template. This can be used to exclude entire templates from MHC.
[ -0.6419409513473511, -0.7158888578414917, -0.05510092154145241, 0.09203927218914032, -0.11549574881792068, 0.10751984268426895, -0.030292538926005363, -0.5101843476295471, 0.39948996901512146, 0.3274587094783783, -0.7589273452758789, -0.7721040844917297, -0.5623311400413513, 0.460262417793...
null
null
null
null
null
null
null
null
null
null
null
null
null
Paul/hatecheck-hindi
Paul
2022-07-05T10:36:37Z
13
0
null
[ "task_categories:text-classification", "task_ids:hate-speech-detection", "annotations_creators:crowdsourced", "language_creators:expert-generated", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:hi", "license:cc-by-4.0", "arxiv:2206.09917", "regi...
2022-07-05T10:36:37Z
2022-07-05T10:35:40.000Z
2022-07-05T10:35:40
--- annotations_creators: - crowdsourced language_creators: - expert-generated language: - hi license: - cc-by-4.0 multilinguality: - monolingual pretty_name: Hindi HateCheck size_categories: - 1K<n<10K source_datasets: - original task_categories: - text-classification task_ids: - hate-speech-detection --- # Dataset Card for Multilingual HateCheck ## Dataset Description Multilingual HateCheck (MHC) is a suite of functional tests for hate speech detection models in 10 different languages: Arabic, Dutch, French, German, Hindi, Italian, Mandarin, Polish, Portuguese and Spanish. For each language, there are 25+ functional tests that correspond to distinct types of hate and challenging non-hate. This allows for targeted diagnostic insights into model performance. For more details, please refer to our paper about MHC, published at the 2022 Workshop on Online Abuse and Harms (WOAH) at NAACL 2022. If you are using MHC, please cite our work! - **Paper:** Röttger et al. (2022) - Multilingual HateCheck: Functional Tests for Multilingual Hate Speech Detection Models. https://arxiv.org/abs/2206.09917 - **Repository:** https://github.com/rewire-online/multilingual-hatecheck - **Point of Contact:** paul@rewire.online ## Dataset Structure The csv format mostly matches the original HateCheck data, with some adjustments for specific languages. **mhc_case_id** The test case ID that is unique to each test case across languages (e.g., "mandarin-1305") **functionality** The shorthand for the functionality tested by the test case (e.g, "target_obj_nh"). The same functionalities are tested in all languages, except for Mandarin and Arabic, where non-Latin script required adapting the tests for spelling variations. **test_case** The test case text. **label_gold** The gold standard label ("hateful" or "non-hateful") of the test case. All test cases within a given functionality have the same gold standard label. **target_ident** Where applicable, the protected group that is targeted or referenced in the test case. All HateChecks cover seven target groups, but their composition varies across languages. **ref_case_id** For hateful cases, where applicable, the ID of the hateful case which was perturbed to generate this test case. For non-hateful cases, where applicable, the ID of the hateful case which is contrasted by this test case. **ref_templ_id** The equivalent to ref_case_id, but for template IDs. **templ_id** The ID of the template from which the test case was generated. **case_templ** The template from which the test case was generated (where applicable). **gender_male** and **gender_female** For gender-inflected languages (French, Spanish, Portuguese, Hindi, Arabic, Italian, Polish, German), only for cases where gender inflection is relevant, separate entries for gender_male and gender_female replace case_templ. **label_annotated** A list of labels given by the three annotators who reviewed the test case (e.g., "['hateful', 'hateful', 'hateful']"). **label_annotated_maj** The majority vote of the three annotators (e.g., "hateful"). In some cases this differs from the gold label given by our language experts. **disagreement_in_case** True if label_annotated_maj does not match label_gold for the entry. **disagreement_in_template** True if the test case is generated from an IDENT template and there is at least one case with disagreement_in_case generated from the same template. This can be used to exclude entire templates from MHC.
[ -0.6419410109519958, -0.7158888578414917, -0.05510091781616211, 0.09203927218914032, -0.11549574136734009, 0.10751984268426895, -0.030292538926005363, -0.5101842880249023, 0.39948996901512146, 0.3274586796760559, -0.7589271664619446, -0.7721040844917297, -0.5623311400413513, 0.460262417793...
null
null
null
null
null
null
null
null
null
null
null
null
null
OGB/ogbg-code2
OGB
2023-02-07T16:40:02Z
13
1
null
[ "task_categories:graph-ml", "license:mit", "region:us" ]
2023-02-07T16:40:02Z
2022-07-07T13:51:15.000Z
2022-07-07T13:51:15
--- license: mit task_categories: - graph-ml --- # Dataset Card for ogbg-code2 ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [External Use](#external-use) - [PyGeometric](#pygeometric) - [Dataset Structure](#dataset-structure) - [Data Properties](#data-properties) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Additional Information](#additional-information) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **[Homepage](https://ogb.stanford.edu/docs/graphprop/#ogbg-code2)** - **[Repository](https://github.com/snap-stanford/ogb):**: - **Paper:**: Open Graph Benchmark: Datasets for Machine Learning on Graphs (see citation) - **Leaderboard:**: [OGB leaderboard](https://ogb.stanford.edu/docs/leader_graphprop/#ogbg-code2) and [Papers with code leaderboard](https://paperswithcode.com/sota/graph-property-prediction-on-ogbg-code2) ### Dataset Summary The `ogbg-code2` dataset contains Abstract Syntax Trees (ASTs) obtained from 450 thousands Python method definitions, from GitHub CodeSearchNet. "Methods are extracted from a total of 13,587 different repositories across the most popular projects on GitHub.", by teams at Stanford, to be a part of the Open Graph Benchmark. See their website or paper for dataset postprocessing. ### Supported Tasks and Leaderboards "The task is to predict the sub-tokens forming the method name, given the Python method body represented by AST and its node features. This task is often referred to as “code summarization”, because the model is trained to find succinct and precise description for a complete logical unit." The score is the F1 score of sub-token prediction. ## External Use ### PyGeometric To load in PyGeometric, do the following: ```python from datasets import load_dataset from torch_geometric.data import Data from torch_geometric.loader import DataLoader graphs_dataset = load_dataset("graphs-datasets/ogbg-code2) # For the train set (replace by valid or test as needed) graphs_list = [Data(graph) for graph in graphs_dataset["train"]] graphs_pygeometric = DataLoader(graph_list) ``` ## Dataset Structure ### Data Properties | property | value | |---|---| | scale | medium | | #graphs | 452,741 | | average #nodes | 125.2 | | average #edges | 124.2 | | average node degree | 2.0 | | average cluster coefficient | 0.0 | | MaxSCC ratio | 1.000 | | graph diameter | 13.5 | ### Data Fields Each row of a given file is a graph, with: - `edge_index` (list: 2 x #edges): pairs of nodes constituting edges - `edge_feat` (list: #edges x #edge-features): features of edges - `node_feat` (list: #nodes x #node-features): the nodes features, embedded - `node_feat_expanded` (list: #nodes x #node-features): the nodes features, as code - `node_is_attributed` (list: 1 x #nodes): ? - `node_dfs_order` (list: #nodes x #1): the nodes order in the abstract tree, if parsed using a depth first search - `node_depth` (list: #nodes x #1): the nodes depth in the abstract tree - `y` (list: 1 x #tokens): contains the tokens to predict as method name - `num_nodes` (int): number of nodes of the graph - `ptr` (list: 2): index of first and last node of the graph - `batch` (list: 1 x #nodes): ? ### Data Splits This data comes from the PyGeometric version of the dataset provided by OGB, and follows the provided data splits. This information can be found back using ```python from ogb.graphproppred import PygGraphPropPredDataset dataset = PygGraphPropPredDataset(name = 'ogbg-code2') split_idx = dataset.get_idx_split() train = dataset[split_idx['train']] # valid, test ``` More information (`node_feat_expanded`) has been added through the typeidx2type and attridx2attr csv files of the repo. ## Additional Information ### Licensing Information The dataset has been released under MIT license license. ### Citation Information ``` @inproceedings{hu-etal-2020-open, author = {Weihua Hu and Matthias Fey and Marinka Zitnik and Yuxiao Dong and Hongyu Ren and Bowen Liu and Michele Catasta and Jure Leskovec}, editor = {Hugo Larochelle and Marc Aurelio Ranzato and Raia Hadsell and Maria{-}Florina Balcan and Hsuan{-}Tien Lin}, title = {Open Graph Benchmark: Datasets for Machine Learning on Graphs}, booktitle = {Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual}, year = {2020}, url = {https://proceedings.neurips.cc/paper/2020/hash/fb60d411a5c5b72b2e7d3527cfc84fd0-Abstract.html}, } ``` ### Contributions Thanks to [@clefourrier](https://github.com/clefourrier) for adding this dataset.
[ -0.4622497856616974, -0.6599794626235962, 0.15342338383197784, -0.07968278229236603, -0.16179662942886353, -0.09635152667760849, -0.29984724521636963, -0.4812818467617035, 0.1274871677160263, 0.1874391883611679, -0.3462800085544586, -0.7914268970489502, -0.606979489326477, -0.0780058130621...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-xsum-7db0303b-10095338
autoevaluate
2022-07-07T15:18:12Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-07-07T15:18:12Z
2022-07-07T14:41:22.000Z
2022-07-07T14:41:22
--- type: predictions tags: - autotrain - evaluation datasets: - xsum eval_info: task: summarization model: philschmid/distilbart-cnn-12-6-samsum metrics: [] dataset_name: xsum dataset_config: default dataset_split: test col_mapping: text: document target: summary --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: philschmid/distilbart-cnn-12-6-samsum * Dataset: xsum To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@ArjunPrSarkhel](https://huggingface.co/ArjunPrSarkhel) for evaluating this model.
[ -0.5184828042984009, -0.05735490843653679, 0.06428296118974686, 0.12116266041994095, -0.20340724289417267, -0.1127840206027031, -0.0181773342192173, -0.3126896917819977, 0.27948665618896484, 0.32453224062919617, -1.0642526149749756, -0.12290552258491516, -0.7990909814834595, -0.06574476510...
null
null
null
null
null
null
null
null
null
null
null
null
null
erickdp/mndt
erickdp
2022-07-07T20:00:33Z
13
0
null
[ "region:us" ]
2022-07-07T20:00:33Z
2022-07-07T19:58:07.000Z
2022-07-07T19:58:07
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
erickdp/dt22
erickdp
2022-07-07T20:15:04Z
13
0
null
[ "region:us" ]
2022-07-07T20:15:04Z
2022-07-07T20:14:30.000Z
2022-07-07T20:14:30
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622264862060547, 0.43461528420448303, -0.52829909324646, 0.7012971639633179, 0.7915720343589783, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104477167129517, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
erickdp/ds23
erickdp
2022-07-07T20:36:48Z
13
0
null
[ "region:us" ]
2022-07-07T20:36:48Z
2022-07-07T20:35:43.000Z
2022-07-07T20:35:43
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622264862060547, 0.43461528420448303, -0.52829909324646, 0.7012971639633179, 0.7915720343589783, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104477167129517, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
juancopi81/mutopia_guitar_dataset
juancopi81
2022-07-22T00:09:34Z
13
1
null
[ "task_categories:text-generation", "task_ids:language-modeling", "multilinguality:other-music", "license:cc", "arxiv:2008.06048", "region:us" ]
2022-07-22T00:09:34Z
2022-07-08T00:06:39.000Z
2022-07-08T00:06:39
--- license: - cc multilinguality: - other-music pretty_name: Mutopia Guitar Dataset task_categories: - text-generation task_ids: - language-modeling --- # Mutopia Guitar Dataset ## Table of Contents - [Dataset Card Creation Guide](#mutopia-guitar-dataset) - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) ## Dataset Description - **Homepage:** [Mutopia Project](https://www.mutopiaproject.org/) - **Repository implementation of the paper:** [MMM: Exploring Conditional Multi-Track Music Generation with the Transformer and the Johann Sebastian Bach Chorales Dataset](https://github.com/AI-Guru/MMM-JSB) - **Based on Paper:** [MMM: Exploring Conditional Multi-Track Music Generation with the Transformer](https://arxiv.org/abs/2008.06048) - **Point of Contact:** [Juan Carlos Piñeros](https://www.linkedin.com/in/juancarlospinerosp/) ### Dataset Summary Mutopia guitar dataset consists of the soloist guitar pieces of the [Mutopia Project](https://www.mutopiaproject.org/). I encoded the MIDI files into text tokens using the excellent [implementation](https://github.com/AI-Guru/MMM-JSB) of Dr. Tristan Beheren of the paper: [MMM: Exploring Conditional Multi-Track Music Generation with the Transformer](https://arxiv.org/abs/2008.06048). The dataset mainly contains guitar music from western classical composers, such as Sor, Aguado, Carcassi, and Giuliani. ### Supported Tasks and Leaderboards Anyone interested can use the dataset to train a model for symbolic music generation, which consists in treating symbols for music sounds (notes) as text tokens. Then, one can implement a generative model using NLP techniques, such as Transformers. ## Dataset Structure ### Data Instances Each guitar piece is represented as a line of text that contains a series of tokens, for instance: PIECE_START: Where the piece begins PIECE_ENDS: Where the piece ends TIME_SIGNATURE: Time signature for the piece BPM: Tempo of the piece BAR_START: Begining of a new bar NOTE_ON: Start of a new musical note specifying its MIDI note number TIME_DELTA: Duration until the next event NOTE_OFF: End of musical note specifying its MIDI note number ``` { 'text': PIECE_START TIME_SIGNATURE=2_4 BPM=74 TRACK_START INST=0 DENSITY=4 BAR_START NOTE_ON=52 TIME_DELTA=2.0 NOTE_OFF=52 NOTE_ON=45 NOTE_ON=49 TIME_DELTA=2.0 NOTE_OFF=49 NOTE_ON=52 TIME_DELTA=2.0 NOTE_OFF=45 NOTE_ON=47 NOTE_OFF=52 NOTE_ON=44 TIME_DELTA=2.0, ... } ``` ### Data Fields - `text`: Sequence of tokens that represent the guitar piece as explained in the paper [MMM: Exploring Conditional Multi-Track Music Generation with the Transformer](https://arxiv.org/abs/2008.06048). ### Data Splits There are, at this moment, 395 MIDI guitar files in the Mutopia Project. I removed files of pieces that were not music for soloist guitar. After this removal, there were 372 MIDI files. I used an 80/20 split and augmented the training dataset by transposing the piece 1 octave above and below (24 semitones). The final result is then: **Train dataset:** 7325 pieces **Test dataset:** 74 pieces
[ -0.7006639838218689, -0.4498589336872101, 0.3559260070323944, 0.3368558883666992, -0.19320198893547058, 0.16617299616336823, -0.47104883193969727, -0.13974492251873016, 0.4598298668861389, 0.5003478527069092, -1.0665791034698486, -0.61394864320755, -0.3424125611782074, 0.06500672549009323,...
null
null
null
null
null
null
null
null
null
null
null
null
null
codecrafters/github-avatars
codecrafters
2022-07-08T00:53:09Z
13
0
null
[ "license:mit", "region:us" ]
2022-07-08T00:53:09Z
2022-07-08T00:28:48.000Z
2022-07-08T00:28:48
--- license: mit ---
[ -0.1285335123538971, -0.1861683875322342, 0.6529128551483154, 0.49436232447624207, -0.19319400191307068, 0.23607441782951355, 0.36072009801864624, 0.05056373029947281, 0.5793656706809998, 0.7400146722793579, -0.650810182094574, -0.23784008622169495, -0.7102247476577759, -0.0478255338966846...
null
null
null
null
null
null
null
null
null
null
null
null
null
gagan3012/alexuw
gagan3012
2022-07-08T05:38:42Z
13
0
null
[ "region:us" ]
2022-07-08T05:38:42Z
2022-07-08T00:36:47.000Z
2022-07-08T00:36:47
--- viewer: true ---
[ -0.12853392958641052, -0.18616779148578644, 0.6529127955436707, 0.49436280131340027, -0.19319361448287964, 0.23607419431209564, 0.36072003841400146, 0.050563063472509384, 0.579365611076355, 0.7400140762329102, -0.6508104205131531, -0.23783954977989197, -0.7102249264717102, -0.0478260256350...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-xsum-e02a2fb8-10255357
autoevaluate
2022-07-08T03:47:13Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-07-08T03:47:13Z
2022-07-08T02:56:23.000Z
2022-07-08T02:56:23
--- type: predictions tags: - autotrain - evaluation datasets: - xsum eval_info: task: summarization model: Ameer05/bart-large-cnn-samsum-rescom-finetuned-resume-summarizer-10-epoch metrics: [] dataset_name: xsum dataset_config: default dataset_split: test col_mapping: text: document target: summary --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: Ameer05/bart-large-cnn-samsum-rescom-finetuned-resume-summarizer-10-epoch * Dataset: xsum To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@Abhijeet3922](https://huggingface.co/Abhijeet3922) for evaluating this model.
[ -0.508518397808075, -0.12419038265943527, 0.15681514143943787, 0.12257508933544159, -0.10815177112817764, -0.14608369767665863, -0.16629958152770996, -0.3018364906311035, 0.3609079122543335, 0.429513156414032, -0.9998050332069397, -0.20805518329143524, -0.6670230031013489, -0.0557976216077...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-samsum-0c672345-10275361
autoevaluate
2022-07-08T08:32:13Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-07-08T08:32:13Z
2022-07-08T04:33:50.000Z
2022-07-08T04:33:50
--- type: predictions tags: - autotrain - evaluation datasets: - samsum eval_info: task: summarization model: google/pegasus-large metrics: [] dataset_name: samsum dataset_config: samsum dataset_split: train col_mapping: text: dialogue target: summary --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: google/pegasus-large * Dataset: samsum To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@ikadebi](https://huggingface.co/ikadebi) for evaluating this model.
[ -0.5094285607337952, -0.09877500683069229, 0.28210532665252686, 0.10647346079349518, -0.19695419073104858, -0.2568368911743164, 0.11076974868774414, -0.41445741057395935, 0.47118669748306274, 0.47185635566711426, -1.149802803993225, -0.21385899186134338, -0.6967419981956482, -0.14423564076...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-samsum-0c672345-10275362
autoevaluate
2022-07-08T05:16:37Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-07-08T05:16:37Z
2022-07-08T04:33:54.000Z
2022-07-08T04:33:54
--- type: predictions tags: - autotrain - evaluation datasets: - samsum eval_info: task: summarization model: google/pegasus-xsum metrics: [] dataset_name: samsum dataset_config: samsum dataset_split: train col_mapping: text: dialogue target: summary --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: google/pegasus-xsum * Dataset: samsum To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@ikadebi](https://huggingface.co/ikadebi) for evaluating this model.
[ -0.48340484499931335, -0.0001488219277234748, 0.26752209663391113, 0.036167874932289124, -0.19628281891345978, -0.19830793142318726, 0.1354266107082367, -0.39664554595947266, 0.517988920211792, 0.47248584032058716, -1.1817065477371216, -0.19058825075626373, -0.6849720478057861, -0.19922797...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-samsum-0c672345-10275364
autoevaluate
2022-07-08T05:59:46Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-07-08T05:59:46Z
2022-07-08T04:34:06.000Z
2022-07-08T04:34:06
--- type: predictions tags: - autotrain - evaluation datasets: - samsum eval_info: task: summarization model: google/pegasus-reddit_tifu metrics: [] dataset_name: samsum dataset_config: samsum dataset_split: train col_mapping: text: dialogue target: summary --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: google/pegasus-reddit_tifu * Dataset: samsum To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@ikadebi](https://huggingface.co/ikadebi) for evaluating this model.
[ -0.5055578947067261, -0.2340993583202362, 0.21989840269088745, 0.17605307698249817, -0.19649481773376465, -0.2007790207862854, 0.11981233954429626, -0.378601610660553, 0.4571053683757782, 0.4010181427001953, -1.12259840965271, -0.18488039076328278, -0.6678062677383423, -0.06094668433070183...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-samsum-0c672345-10275365
autoevaluate
2022-07-08T05:41:09Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-07-08T05:41:09Z
2022-07-08T05:16:57.000Z
2022-07-08T05:16:57
--- type: predictions tags: - autotrain - evaluation datasets: - samsum eval_info: task: summarization model: facebook/bart-large-xsum metrics: [] dataset_name: samsum dataset_config: samsum dataset_split: train col_mapping: text: dialogue target: summary --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: facebook/bart-large-xsum * Dataset: samsum To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@ikadebi](https://huggingface.co/ikadebi) for evaluating this model.
[ -0.43643149733543396, -0.19937092065811157, 0.3007427155971527, 0.14195643365383148, -0.06724821776151657, -0.05425050109624863, 0.02553047239780426, -0.38449227809906006, 0.5072949528694153, 0.4954293966293335, -1.1453310251235962, -0.1954723298549652, -0.5937836170196533, -0.127551525831...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-samsum-0c672345-10275366
autoevaluate
2022-07-08T06:11:02Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-07-08T06:11:02Z
2022-07-08T05:41:25.000Z
2022-07-08T05:41:25
--- type: predictions tags: - autotrain - evaluation datasets: - samsum eval_info: task: summarization model: knkarthick/bart-large-xsum-samsum metrics: [] dataset_name: samsum dataset_config: samsum dataset_split: train col_mapping: text: dialogue target: summary --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: knkarthick/bart-large-xsum-samsum * Dataset: samsum To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@ikadebi](https://huggingface.co/ikadebi) for evaluating this model.
[ -0.4612419009208679, -0.05640745535492897, 0.35649797320365906, 0.06294597685337067, -0.18138378858566284, -0.05707351490855217, -0.016651805490255356, -0.40236881375312805, 0.4544757306575775, 0.5290325284004211, -1.0893206596374512, -0.21077926456928253, -0.6362278461456299, -0.068324208...
null
null
null
null
null
null
null
null
null
null
null
null
null
GEM-submissions/lewtun__this-is-a-test-submission-3__1657282248
GEM-submissions
2022-07-08T12:10:53Z
13
0
null
[ "benchmark:gem", "evaluation", "benchmark", "region:us" ]
2022-07-08T12:10:53Z
2022-07-08T12:10:50.000Z
2022-07-08T12:10:50
--- benchmark: gem type: prediction submission_name: This is a test submission 3 tags: - evaluation - benchmark --- # GEM Submission Submission name: This is a test submission 3
[ 0.14071477949619293, -0.831126868724823, 0.7007408738136292, 0.2553391754627228, -0.16222964227199554, 0.26310965418815613, 0.4091661870479584, 0.17561329901218414, 0.2834315299987793, 0.4843377470970154, -0.994183361530304, -0.1469837874174118, -0.31571251153945923, 0.0889996811747551, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
MicPie/unpredictable_cluster13
MicPie
2022-08-04T19:52:42Z
13
0
null
[ "task_categories:multiple-choice", "task_categories:question-answering", "task_categories:zero-shot-classification", "task_categories:text2text-generation", "task_categories:table-question-answering", "task_categories:text-generation", "task_categories:text-classification", "task_categories:tabular-cl...
2022-08-04T19:52:42Z
2022-07-08T17:23:44.000Z
2022-07-08T17:23:44
--- annotations_creators: - no-annotation language_creators: - found language: - en license: - apache-2.0 multilinguality: - monolingual pretty_name: UnpredicTable-cluster13 size_categories: - 100K<n<1M source_datasets: [] task_categories: - multiple-choice - question-answering - zero-shot-classification - text2text-generation - table-question-answering - text-generation - text-classification - tabular-classification task_ids: - multiple-choice-qa - extractive-qa - open-domain-qa - closed-domain-qa - closed-book-qa - open-book-qa - language-modeling - multi-class-classification - natural-language-inference - topic-classification - multi-label-classification - tabular-multi-class-classification - tabular-multi-label-classification --- # Dataset Card for "UnpredicTable-cluster13" - Dataset of Few-shot Tasks from Tables ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** https://ethanperez.net/unpredictable - **Repository:** https://github.com/JunShern/few-shot-adaptation - **Paper:** Few-shot Adaptation Works with UnpredicTable Data - **Point of Contact:** junshern@nyu.edu, perez@nyu.edu ### Dataset Summary The UnpredicTable dataset consists of web tables formatted as few-shot tasks for fine-tuning language models to improve their few-shot performance. There are several dataset versions available: * [UnpredicTable-full](https://huggingface.co/datasets/MicPie/unpredictable_full): Starting from the initial WTC corpus of 50M tables, we apply our tables-to-tasks procedure to produce our resulting dataset, [UnpredicTable-full](https://huggingface.co/datasets/MicPie/unpredictable_full), which comprises 413,299 tasks from 23,744 unique websites. * [UnpredicTable-unique](https://huggingface.co/datasets/MicPie/unpredictable_unique): This is the same as [UnpredicTable-full](https://huggingface.co/datasets/MicPie/unpredictable_full) but filtered to have a maximum of one task per website. [UnpredicTable-unique](https://huggingface.co/datasets/MicPie/unpredictable_unique) contains exactly 23,744 tasks from 23,744 websites. * [UnpredicTable-5k](https://huggingface.co/datasets/MicPie/unpredictable_5k): This dataset contains 5k random tables from the full dataset. * UnpredicTable data subsets based on a manual human quality rating (please see our publication for details of the ratings): * [UnpredicTable-rated-low](https://huggingface.co/datasets/MicPie/unpredictable_rated-low) * [UnpredicTable-rated-medium](https://huggingface.co/datasets/MicPie/unpredictable_rated-medium) * [UnpredicTable-rated-high](https://huggingface.co/datasets/MicPie/unpredictable_rated-high) * UnpredicTable data subsets based on the website of origin: * [UnpredicTable-baseball-fantasysports-yahoo-com](https://huggingface.co/datasets/MicPie/unpredictable_baseball-fantasysports-yahoo-com) * [UnpredicTable-bulbapedia-bulbagarden-net](https://huggingface.co/datasets/MicPie/unpredictable_bulbapedia-bulbagarden-net) * [UnpredicTable-cappex-com](https://huggingface.co/datasets/MicPie/unpredictable_cappex-com) * [UnpredicTable-cram-com](https://huggingface.co/datasets/MicPie/unpredictable_cram-com) * [UnpredicTable-dividend-com](https://huggingface.co/datasets/MicPie/unpredictable_dividend-com) * [UnpredicTable-dummies-com](https://huggingface.co/datasets/MicPie/unpredictable_dummies-com) * [UnpredicTable-en-wikipedia-org](https://huggingface.co/datasets/MicPie/unpredictable_en-wikipedia-org) * [UnpredicTable-ensembl-org](https://huggingface.co/datasets/MicPie/unpredictable_ensembl-org) * [UnpredicTable-gamefaqs-com](https://huggingface.co/datasets/MicPie/unpredictable_gamefaqs-com) * [UnpredicTable-mgoblog-com](https://huggingface.co/datasets/MicPie/unpredictable_mgoblog-com) * [UnpredicTable-mmo-champion-com](https://huggingface.co/datasets/MicPie/unpredictable_mmo-champion-com) * [UnpredicTable-msdn-microsoft-com](https://huggingface.co/datasets/MicPie/unpredictable_msdn-microsoft-com) * [UnpredicTable-phonearena-com](https://huggingface.co/datasets/MicPie/unpredictable_phonearena-com) * [UnpredicTable-sittercity-com](https://huggingface.co/datasets/MicPie/unpredictable_sittercity-com) * [UnpredicTable-sporcle-com](https://huggingface.co/datasets/MicPie/unpredictable_sporcle-com) * [UnpredicTable-studystack-com](https://huggingface.co/datasets/MicPie/unpredictable_studystack-com) * [UnpredicTable-support-google-com](https://huggingface.co/datasets/MicPie/unpredictable_support-google-com) * [UnpredicTable-w3-org](https://huggingface.co/datasets/MicPie/unpredictable_w3-org) * [UnpredicTable-wiki-openmoko-org](https://huggingface.co/datasets/MicPie/unpredictable_wiki-openmoko-org) * [UnpredicTable-wkdu-org](https://huggingface.co/datasets/MicPie/unpredictable_wkdu-org) * UnpredicTable data subsets based on clustering (for the clustering details please see our publication): * [UnpredicTable-cluster00](https://huggingface.co/datasets/MicPie/unpredictable_cluster00) * [UnpredicTable-cluster01](https://huggingface.co/datasets/MicPie/unpredictable_cluster01) * [UnpredicTable-cluster02](https://huggingface.co/datasets/MicPie/unpredictable_cluster02) * [UnpredicTable-cluster03](https://huggingface.co/datasets/MicPie/unpredictable_cluster03) * [UnpredicTable-cluster04](https://huggingface.co/datasets/MicPie/unpredictable_cluster04) * [UnpredicTable-cluster05](https://huggingface.co/datasets/MicPie/unpredictable_cluster05) * [UnpredicTable-cluster06](https://huggingface.co/datasets/MicPie/unpredictable_cluster06) * [UnpredicTable-cluster07](https://huggingface.co/datasets/MicPie/unpredictable_cluster07) * [UnpredicTable-cluster08](https://huggingface.co/datasets/MicPie/unpredictable_cluster08) * [UnpredicTable-cluster09](https://huggingface.co/datasets/MicPie/unpredictable_cluster09) * [UnpredicTable-cluster10](https://huggingface.co/datasets/MicPie/unpredictable_cluster10) * [UnpredicTable-cluster11](https://huggingface.co/datasets/MicPie/unpredictable_cluster11) * [UnpredicTable-cluster12](https://huggingface.co/datasets/MicPie/unpredictable_cluster12) * [UnpredicTable-cluster13](https://huggingface.co/datasets/MicPie/unpredictable_cluster13) * [UnpredicTable-cluster14](https://huggingface.co/datasets/MicPie/unpredictable_cluster14) * [UnpredicTable-cluster15](https://huggingface.co/datasets/MicPie/unpredictable_cluster15) * [UnpredicTable-cluster16](https://huggingface.co/datasets/MicPie/unpredictable_cluster16) * [UnpredicTable-cluster17](https://huggingface.co/datasets/MicPie/unpredictable_cluster17) * [UnpredicTable-cluster18](https://huggingface.co/datasets/MicPie/unpredictable_cluster18) * [UnpredicTable-cluster19](https://huggingface.co/datasets/MicPie/unpredictable_cluster19) * [UnpredicTable-cluster20](https://huggingface.co/datasets/MicPie/unpredictable_cluster20) * [UnpredicTable-cluster21](https://huggingface.co/datasets/MicPie/unpredictable_cluster21) * [UnpredicTable-cluster22](https://huggingface.co/datasets/MicPie/unpredictable_cluster22) * [UnpredicTable-cluster23](https://huggingface.co/datasets/MicPie/unpredictable_cluster23) * [UnpredicTable-cluster24](https://huggingface.co/datasets/MicPie/unpredictable_cluster24) * [UnpredicTable-cluster25](https://huggingface.co/datasets/MicPie/unpredictable_cluster25) * [UnpredicTable-cluster26](https://huggingface.co/datasets/MicPie/unpredictable_cluster26) * [UnpredicTable-cluster27](https://huggingface.co/datasets/MicPie/unpredictable_cluster27) * [UnpredicTable-cluster28](https://huggingface.co/datasets/MicPie/unpredictable_cluster28) * [UnpredicTable-cluster29](https://huggingface.co/datasets/MicPie/unpredictable_cluster29) * [UnpredicTable-cluster-noise](https://huggingface.co/datasets/MicPie/unpredictable_cluster-noise) ### Supported Tasks and Leaderboards Since the tables come from the web, the distribution of tasks and topics is very broad. The shape of our dataset is very wide, i.e., we have 1000's of tasks, while each task has only a few examples, compared to most current NLP datasets which are very deep, i.e., 10s of tasks with many examples. This implies that our dataset covers a broad range of potential tasks, e.g., multiple-choice, question-answering, table-question-answering, text-classification, etc. The intended use of this dataset is to improve few-shot performance by fine-tuning/pre-training on our dataset. ### Languages English ## Dataset Structure ### Data Instances Each task is represented as a jsonline file and consists of several few-shot examples. Each example is a dictionary containing a field 'task', which identifies the task, followed by an 'input', 'options', and 'output' field. The 'input' field contains several column elements of the same row in the table, while the 'output' field is a target which represents an individual column of the same row. Each task contains several such examples which can be concatenated as a few-shot task. In the case of multiple choice classification, the 'options' field contains the possible classes that a model needs to choose from. There are also additional meta-data fields such as 'pageTitle', 'title', 'outputColName', 'url', 'wdcFile'. ### Data Fields 'task': task identifier 'input': column elements of a specific row in the table. 'options': for multiple choice classification, it provides the options to choose from. 'output': target column element of the same row as input. 'pageTitle': the title of the page containing the table. 'outputColName': output column name 'url': url to the website containing the table 'wdcFile': WDC Web Table Corpus file ### Data Splits The UnpredicTable datasets do not come with additional data splits. ## Dataset Creation ### Curation Rationale Few-shot training on multi-task datasets has been demonstrated to improve language models' few-shot learning (FSL) performance on new tasks, but it is unclear which training tasks lead to effective downstream task adaptation. Few-shot learning datasets are typically produced with expensive human curation, limiting the scale and diversity of the training tasks available to study. As an alternative source of few-shot data, we automatically extract 413,299 tasks from diverse internet tables. We provide this as a research resource to investigate the relationship between training data and few-shot learning. ### Source Data #### Initial Data Collection and Normalization We use internet tables from the English-language Relational Subset of the WDC Web Table Corpus 2015 (WTC). The WTC dataset tables were extracted from the July 2015 Common Crawl web corpus (http://webdatacommons.org/webtables/2015/EnglishStatistics.html). The dataset contains 50,820,165 tables from 323,160 web domains. We then convert the tables into few-shot learning tasks. Please see our publication for more details on the data collection and conversion pipeline. #### Who are the source language producers? The dataset is extracted from [WDC Web Table Corpora](http://webdatacommons.org/webtables/). ### Annotations #### Annotation process Manual annotation was only carried out for the [UnpredicTable-rated-low](https://huggingface.co/datasets/MicPie/unpredictable_rated-low), [UnpredicTable-rated-medium](https://huggingface.co/datasets/MicPie/unpredictable_rated-medium), and [UnpredicTable-rated-high](https://huggingface.co/datasets/MicPie/unpredictable_rated-high) data subsets to rate task quality. Detailed instructions of the annotation instructions can be found in our publication. #### Who are the annotators? Annotations were carried out by a lab assistant. ### Personal and Sensitive Information The data was extracted from [WDC Web Table Corpora](http://webdatacommons.org/webtables/), which in turn extracted tables from the [Common Crawl](https://commoncrawl.org/). We did not filter the data in any way. Thus any user identities or otherwise sensitive information (e.g., data that reveals racial or ethnic origins, sexual orientations, religious beliefs, political opinions or union memberships, or locations; financial or health data; biometric or genetic data; forms of government identification, such as social security numbers; criminal history, etc.) might be contained in our dataset. ## Considerations for Using the Data ### Social Impact of Dataset This dataset is intended for use as a research resource to investigate the relationship between training data and few-shot learning. As such, it contains high- and low-quality data, as well as diverse content that may be untruthful or inappropriate. Without careful investigation, it should not be used for training models that will be deployed for use in decision-critical or user-facing situations. ### Discussion of Biases Since our dataset contains tables that are scraped from the web, it will also contain many toxic, racist, sexist, and otherwise harmful biases and texts. We have not run any analysis on the biases prevalent in our datasets. Neither have we explicitly filtered the content. This implies that a model trained on our dataset may potentially reflect harmful biases and toxic text that exist in our dataset. ### Other Known Limitations No additional known limitations. ## Additional Information ### Dataset Curators Jun Shern Chan, Michael Pieler, Jonathan Jao, Jérémy Scheurer, Ethan Perez ### Licensing Information Apache 2.0 ### Citation Information ``` @misc{chan2022few, author = {Chan, Jun Shern and Pieler, Michael and Jao, Jonathan and Scheurer, Jérémy and Perez, Ethan}, title = {Few-shot Adaptation Works with UnpredicTable Data}, publisher={arXiv}, year = {2022}, url = {https://arxiv.org/abs/2208.01009} } ```
[ -0.5828394889831543, -0.5666294097900391, 0.4744274318218231, 0.33533042669296265, 0.08311328291893005, 0.1527356654405594, -0.14912305772304535, -0.60282963514328, 0.5245921015739441, 0.272038072347641, -1.0352694988250732, -0.682465672492981, -0.6594418287277222, 0.20154514908790588, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
MicPie/unpredictable_cluster25
MicPie
2022-08-04T20:00:11Z
13
0
null
[ "task_categories:multiple-choice", "task_categories:question-answering", "task_categories:zero-shot-classification", "task_categories:text2text-generation", "task_categories:table-question-answering", "task_categories:text-generation", "task_categories:text-classification", "task_categories:tabular-cl...
2022-08-04T20:00:11Z
2022-07-08T18:35:02.000Z
2022-07-08T18:35:02
--- annotations_creators: - no-annotation language_creators: - found language: - en license: - apache-2.0 multilinguality: - monolingual pretty_name: UnpredicTable-cluster25 size_categories: - 100K<n<1M source_datasets: [] task_categories: - multiple-choice - question-answering - zero-shot-classification - text2text-generation - table-question-answering - text-generation - text-classification - tabular-classification task_ids: - multiple-choice-qa - extractive-qa - open-domain-qa - closed-domain-qa - closed-book-qa - open-book-qa - language-modeling - multi-class-classification - natural-language-inference - topic-classification - multi-label-classification - tabular-multi-class-classification - tabular-multi-label-classification --- # Dataset Card for "UnpredicTable-cluster25" - Dataset of Few-shot Tasks from Tables ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** https://ethanperez.net/unpredictable - **Repository:** https://github.com/JunShern/few-shot-adaptation - **Paper:** Few-shot Adaptation Works with UnpredicTable Data - **Point of Contact:** junshern@nyu.edu, perez@nyu.edu ### Dataset Summary The UnpredicTable dataset consists of web tables formatted as few-shot tasks for fine-tuning language models to improve their few-shot performance. There are several dataset versions available: * [UnpredicTable-full](https://huggingface.co/datasets/MicPie/unpredictable_full): Starting from the initial WTC corpus of 50M tables, we apply our tables-to-tasks procedure to produce our resulting dataset, [UnpredicTable-full](https://huggingface.co/datasets/MicPie/unpredictable_full), which comprises 413,299 tasks from 23,744 unique websites. * [UnpredicTable-unique](https://huggingface.co/datasets/MicPie/unpredictable_unique): This is the same as [UnpredicTable-full](https://huggingface.co/datasets/MicPie/unpredictable_full) but filtered to have a maximum of one task per website. [UnpredicTable-unique](https://huggingface.co/datasets/MicPie/unpredictable_unique) contains exactly 23,744 tasks from 23,744 websites. * [UnpredicTable-5k](https://huggingface.co/datasets/MicPie/unpredictable_5k): This dataset contains 5k random tables from the full dataset. * UnpredicTable data subsets based on a manual human quality rating (please see our publication for details of the ratings): * [UnpredicTable-rated-low](https://huggingface.co/datasets/MicPie/unpredictable_rated-low) * [UnpredicTable-rated-medium](https://huggingface.co/datasets/MicPie/unpredictable_rated-medium) * [UnpredicTable-rated-high](https://huggingface.co/datasets/MicPie/unpredictable_rated-high) * UnpredicTable data subsets based on the website of origin: * [UnpredicTable-baseball-fantasysports-yahoo-com](https://huggingface.co/datasets/MicPie/unpredictable_baseball-fantasysports-yahoo-com) * [UnpredicTable-bulbapedia-bulbagarden-net](https://huggingface.co/datasets/MicPie/unpredictable_bulbapedia-bulbagarden-net) * [UnpredicTable-cappex-com](https://huggingface.co/datasets/MicPie/unpredictable_cappex-com) * [UnpredicTable-cram-com](https://huggingface.co/datasets/MicPie/unpredictable_cram-com) * [UnpredicTable-dividend-com](https://huggingface.co/datasets/MicPie/unpredictable_dividend-com) * [UnpredicTable-dummies-com](https://huggingface.co/datasets/MicPie/unpredictable_dummies-com) * [UnpredicTable-en-wikipedia-org](https://huggingface.co/datasets/MicPie/unpredictable_en-wikipedia-org) * [UnpredicTable-ensembl-org](https://huggingface.co/datasets/MicPie/unpredictable_ensembl-org) * [UnpredicTable-gamefaqs-com](https://huggingface.co/datasets/MicPie/unpredictable_gamefaqs-com) * [UnpredicTable-mgoblog-com](https://huggingface.co/datasets/MicPie/unpredictable_mgoblog-com) * [UnpredicTable-mmo-champion-com](https://huggingface.co/datasets/MicPie/unpredictable_mmo-champion-com) * [UnpredicTable-msdn-microsoft-com](https://huggingface.co/datasets/MicPie/unpredictable_msdn-microsoft-com) * [UnpredicTable-phonearena-com](https://huggingface.co/datasets/MicPie/unpredictable_phonearena-com) * [UnpredicTable-sittercity-com](https://huggingface.co/datasets/MicPie/unpredictable_sittercity-com) * [UnpredicTable-sporcle-com](https://huggingface.co/datasets/MicPie/unpredictable_sporcle-com) * [UnpredicTable-studystack-com](https://huggingface.co/datasets/MicPie/unpredictable_studystack-com) * [UnpredicTable-support-google-com](https://huggingface.co/datasets/MicPie/unpredictable_support-google-com) * [UnpredicTable-w3-org](https://huggingface.co/datasets/MicPie/unpredictable_w3-org) * [UnpredicTable-wiki-openmoko-org](https://huggingface.co/datasets/MicPie/unpredictable_wiki-openmoko-org) * [UnpredicTable-wkdu-org](https://huggingface.co/datasets/MicPie/unpredictable_wkdu-org) * UnpredicTable data subsets based on clustering (for the clustering details please see our publication): * [UnpredicTable-cluster00](https://huggingface.co/datasets/MicPie/unpredictable_cluster00) * [UnpredicTable-cluster01](https://huggingface.co/datasets/MicPie/unpredictable_cluster01) * [UnpredicTable-cluster02](https://huggingface.co/datasets/MicPie/unpredictable_cluster02) * [UnpredicTable-cluster03](https://huggingface.co/datasets/MicPie/unpredictable_cluster03) * [UnpredicTable-cluster04](https://huggingface.co/datasets/MicPie/unpredictable_cluster04) * [UnpredicTable-cluster05](https://huggingface.co/datasets/MicPie/unpredictable_cluster05) * [UnpredicTable-cluster06](https://huggingface.co/datasets/MicPie/unpredictable_cluster06) * [UnpredicTable-cluster07](https://huggingface.co/datasets/MicPie/unpredictable_cluster07) * [UnpredicTable-cluster08](https://huggingface.co/datasets/MicPie/unpredictable_cluster08) * [UnpredicTable-cluster09](https://huggingface.co/datasets/MicPie/unpredictable_cluster09) * [UnpredicTable-cluster10](https://huggingface.co/datasets/MicPie/unpredictable_cluster10) * [UnpredicTable-cluster11](https://huggingface.co/datasets/MicPie/unpredictable_cluster11) * [UnpredicTable-cluster12](https://huggingface.co/datasets/MicPie/unpredictable_cluster12) * [UnpredicTable-cluster13](https://huggingface.co/datasets/MicPie/unpredictable_cluster13) * [UnpredicTable-cluster14](https://huggingface.co/datasets/MicPie/unpredictable_cluster14) * [UnpredicTable-cluster15](https://huggingface.co/datasets/MicPie/unpredictable_cluster15) * [UnpredicTable-cluster16](https://huggingface.co/datasets/MicPie/unpredictable_cluster16) * [UnpredicTable-cluster17](https://huggingface.co/datasets/MicPie/unpredictable_cluster17) * [UnpredicTable-cluster18](https://huggingface.co/datasets/MicPie/unpredictable_cluster18) * [UnpredicTable-cluster19](https://huggingface.co/datasets/MicPie/unpredictable_cluster19) * [UnpredicTable-cluster20](https://huggingface.co/datasets/MicPie/unpredictable_cluster20) * [UnpredicTable-cluster21](https://huggingface.co/datasets/MicPie/unpredictable_cluster21) * [UnpredicTable-cluster22](https://huggingface.co/datasets/MicPie/unpredictable_cluster22) * [UnpredicTable-cluster23](https://huggingface.co/datasets/MicPie/unpredictable_cluster23) * [UnpredicTable-cluster24](https://huggingface.co/datasets/MicPie/unpredictable_cluster24) * [UnpredicTable-cluster25](https://huggingface.co/datasets/MicPie/unpredictable_cluster25) * [UnpredicTable-cluster26](https://huggingface.co/datasets/MicPie/unpredictable_cluster26) * [UnpredicTable-cluster27](https://huggingface.co/datasets/MicPie/unpredictable_cluster27) * [UnpredicTable-cluster28](https://huggingface.co/datasets/MicPie/unpredictable_cluster28) * [UnpredicTable-cluster29](https://huggingface.co/datasets/MicPie/unpredictable_cluster29) * [UnpredicTable-cluster-noise](https://huggingface.co/datasets/MicPie/unpredictable_cluster-noise) ### Supported Tasks and Leaderboards Since the tables come from the web, the distribution of tasks and topics is very broad. The shape of our dataset is very wide, i.e., we have 1000's of tasks, while each task has only a few examples, compared to most current NLP datasets which are very deep, i.e., 10s of tasks with many examples. This implies that our dataset covers a broad range of potential tasks, e.g., multiple-choice, question-answering, table-question-answering, text-classification, etc. The intended use of this dataset is to improve few-shot performance by fine-tuning/pre-training on our dataset. ### Languages English ## Dataset Structure ### Data Instances Each task is represented as a jsonline file and consists of several few-shot examples. Each example is a dictionary containing a field 'task', which identifies the task, followed by an 'input', 'options', and 'output' field. The 'input' field contains several column elements of the same row in the table, while the 'output' field is a target which represents an individual column of the same row. Each task contains several such examples which can be concatenated as a few-shot task. In the case of multiple choice classification, the 'options' field contains the possible classes that a model needs to choose from. There are also additional meta-data fields such as 'pageTitle', 'title', 'outputColName', 'url', 'wdcFile'. ### Data Fields 'task': task identifier 'input': column elements of a specific row in the table. 'options': for multiple choice classification, it provides the options to choose from. 'output': target column element of the same row as input. 'pageTitle': the title of the page containing the table. 'outputColName': output column name 'url': url to the website containing the table 'wdcFile': WDC Web Table Corpus file ### Data Splits The UnpredicTable datasets do not come with additional data splits. ## Dataset Creation ### Curation Rationale Few-shot training on multi-task datasets has been demonstrated to improve language models' few-shot learning (FSL) performance on new tasks, but it is unclear which training tasks lead to effective downstream task adaptation. Few-shot learning datasets are typically produced with expensive human curation, limiting the scale and diversity of the training tasks available to study. As an alternative source of few-shot data, we automatically extract 413,299 tasks from diverse internet tables. We provide this as a research resource to investigate the relationship between training data and few-shot learning. ### Source Data #### Initial Data Collection and Normalization We use internet tables from the English-language Relational Subset of the WDC Web Table Corpus 2015 (WTC). The WTC dataset tables were extracted from the July 2015 Common Crawl web corpus (http://webdatacommons.org/webtables/2015/EnglishStatistics.html). The dataset contains 50,820,165 tables from 323,160 web domains. We then convert the tables into few-shot learning tasks. Please see our publication for more details on the data collection and conversion pipeline. #### Who are the source language producers? The dataset is extracted from [WDC Web Table Corpora](http://webdatacommons.org/webtables/). ### Annotations #### Annotation process Manual annotation was only carried out for the [UnpredicTable-rated-low](https://huggingface.co/datasets/MicPie/unpredictable_rated-low), [UnpredicTable-rated-medium](https://huggingface.co/datasets/MicPie/unpredictable_rated-medium), and [UnpredicTable-rated-high](https://huggingface.co/datasets/MicPie/unpredictable_rated-high) data subsets to rate task quality. Detailed instructions of the annotation instructions can be found in our publication. #### Who are the annotators? Annotations were carried out by a lab assistant. ### Personal and Sensitive Information The data was extracted from [WDC Web Table Corpora](http://webdatacommons.org/webtables/), which in turn extracted tables from the [Common Crawl](https://commoncrawl.org/). We did not filter the data in any way. Thus any user identities or otherwise sensitive information (e.g., data that reveals racial or ethnic origins, sexual orientations, religious beliefs, political opinions or union memberships, or locations; financial or health data; biometric or genetic data; forms of government identification, such as social security numbers; criminal history, etc.) might be contained in our dataset. ## Considerations for Using the Data ### Social Impact of Dataset This dataset is intended for use as a research resource to investigate the relationship between training data and few-shot learning. As such, it contains high- and low-quality data, as well as diverse content that may be untruthful or inappropriate. Without careful investigation, it should not be used for training models that will be deployed for use in decision-critical or user-facing situations. ### Discussion of Biases Since our dataset contains tables that are scraped from the web, it will also contain many toxic, racist, sexist, and otherwise harmful biases and texts. We have not run any analysis on the biases prevalent in our datasets. Neither have we explicitly filtered the content. This implies that a model trained on our dataset may potentially reflect harmful biases and toxic text that exist in our dataset. ### Other Known Limitations No additional known limitations. ## Additional Information ### Dataset Curators Jun Shern Chan, Michael Pieler, Jonathan Jao, Jérémy Scheurer, Ethan Perez ### Licensing Information Apache 2.0 ### Citation Information ``` @misc{chan2022few, author = {Chan, Jun Shern and Pieler, Michael and Jao, Jonathan and Scheurer, Jérémy and Perez, Ethan}, title = {Few-shot Adaptation Works with UnpredicTable Data}, publisher={arXiv}, year = {2022}, url = {https://arxiv.org/abs/2208.01009} } ```
[ -0.5835341215133667, -0.5662290453910828, 0.46426764130592346, 0.3361322283744812, 0.09068774431943893, 0.1591319888830185, -0.15546613931655884, -0.5924043655395508, 0.5226195454597473, 0.2890837788581848, -1.031678318977356, -0.6929382681846619, -0.6448044776916504, 0.18168815970420837, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
kasumi222/busy
kasumi222
2022-08-15T01:55:40Z
13
0
null
[ "region:us" ]
2022-08-15T01:55:40Z
2022-07-09T18:21:26.000Z
2022-07-09T18:21:26
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
kasumi222/busy2
kasumi222
2022-07-09T18:23:19Z
13
0
null
[ "region:us" ]
2022-07-09T18:23:19Z
2022-07-09T18:22:42.000Z
2022-07-09T18:22:42
Dataset1
[ -0.25494322180747986, 0.0017639786237850785, -0.297717809677124, 0.3378487527370453, -0.028897887095808983, -0.2552754282951355, 0.3406207263469696, 0.3764059543609619, 0.4681522846221924, 1.1269928216934204, -1.2244112491607666, -0.2884278893470764, -0.7127476930618286, -0.030882194638252...
null
null
null
null
null
null
null
null
null
null
null
null
null
jorge-henao/historias_conflicto_colombia
jorge-henao
2022-07-10T15:26:41Z
13
0
null
[ "license:apache-2.0", "region:us" ]
2022-07-10T15:26:41Z
2022-07-09T18:26:16.000Z
2022-07-09T18:26:16
--- license: apache-2.0 ---
[ -0.12853367626667023, -0.18616794049739838, 0.6529126763343811, 0.4943627417087555, -0.19319313764572144, 0.23607443273067474, 0.36071979999542236, 0.05056338757276535, 0.5793654322624207, 0.7400138974189758, -0.6508103013038635, -0.23783987760543823, -0.710224986076355, -0.047825977206230...
null
null
null
null
null
null
null
null
null
null
null
null
null
sonicpierre/Rick-bot-flags
sonicpierre
2022-07-10T14:52:40Z
13
0
null
[ "region:us" ]
2022-07-10T14:52:40Z
2022-07-09T20:09:41.000Z
2022-07-09T20:09:41
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
etan18/SampleMCDataset
etan18
2022-07-09T20:13:27Z
13
0
null
[ "license:unknown", "region:us" ]
2022-07-09T20:13:27Z
2022-07-09T20:12:18.000Z
2022-07-09T20:12:18
--- license: unknown ---
[ -0.12853367626667023, -0.18616794049739838, 0.6529126763343811, 0.4943627417087555, -0.19319313764572144, 0.23607443273067474, 0.36071979999542236, 0.05056338757276535, 0.5793654322624207, 0.7400138974189758, -0.6508103013038635, -0.23783987760543823, -0.710224986076355, -0.047825977206230...
null
null
null
null
null
null
null
null
null
null
null
null
null
azuur/en_corpora_parliament_processed
azuur
2022-07-10T01:36:27Z
13
0
null
[ "region:us" ]
2022-07-10T01:36:27Z
2022-07-09T22:03:35.000Z
2022-07-09T22:03:35
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-pn_summary-5464695d-10495406
autoevaluate
2022-07-11T14:22:50Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-07-11T14:22:50Z
2022-07-10T11:47:54.000Z
2022-07-10T11:47:54
--- type: predictions tags: - autotrain - evaluation datasets: - pn_summary eval_info: task: summarization model: google/pegasus-large metrics: [] dataset_name: pn_summary dataset_config: 1.0.0 dataset_split: train col_mapping: text: article target: summary --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: google/pegasus-large * Dataset: pn_summary * Config: 1.0.0 * Split: train To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@marsraker09](https://huggingface.co/marsraker09) for evaluating this model.
[ -0.5578094124794006, -0.1442231833934784, 0.20517006516456604, 0.18276523053646088, -0.22872860729694366, -0.23483586311340332, 0.0574411079287529, -0.387175589799881, 0.4514274001121521, 0.4650925397872925, -1.09524667263031, -0.14462733268737793, -0.7255716919898987, -0.1420077532529831,...
null
null
null
null
null
null
null
null
null
null
null
null
null
zzySJ/sv_corpora_parliament_processed
zzySJ
2022-07-10T12:36:49Z
13
0
null
[ "region:us" ]
2022-07-10T12:36:49Z
2022-07-10T12:36:25.000Z
2022-07-10T12:36:25
Entry not found
[ -0.32276472449302673, -0.22568407654762268, 0.8622258901596069, 0.4346148371696472, -0.5282984972000122, 0.7012965679168701, 0.7915717363357544, 0.07618629932403564, 0.7746022939682007, 0.2563222646713257, -0.785281777381897, -0.22573848068714142, -0.9104482531547546, 0.5715669393539429, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
BirdL/SimulacraUnsupervised
BirdL
2022-12-19T20:31:16Z
13
1
null
[ "task_categories:unconditional-image-generation", "size_categories:100K<n<1M", "region:us" ]
2022-12-19T20:31:16Z
2022-07-10T19:17:34.000Z
2022-07-10T19:17:34
--- annotations_creators: [] language: [] language_creators: [] multilinguality: [] pretty_name: Simulacra Aes Captions Unsupervised size_categories: - 100K<n<1M source_datasets: [] tags: [] task_categories: - unconditional-image-generation task_ids: [] --- SimulacraUnsupervised is a download of Simulacra Aesthetic Captions from JDP converted to a JPEG compressed parquet. Under the BirdL-AirL License
[ -0.5351921319961548, -0.40463966131210327, 0.022966142743825912, 0.7717231512069702, -0.9486656188964844, -0.016440965235233307, -0.0355517603456974, -0.3299877941608429, 0.5984933376312256, 0.7790433764457703, -0.8340150713920593, -0.32155904173851013, -0.38772591948509216, 0.208355173468...
null
null
null
null
null
null
null
null
null
null
null
null
null
Li-Tang/cn_text
Li-Tang
2022-07-11T09:50:11Z
13
0
null
[ "license:apache-2.0", "region:us" ]
2022-07-11T09:50:11Z
2022-07-11T09:48:59.000Z
2022-07-11T09:48:59
--- license: apache-2.0 ---
[ -0.12853392958641052, -0.18616779148578644, 0.6529127955436707, 0.49436280131340027, -0.19319361448287964, 0.23607419431209564, 0.36072003841400146, 0.050563063472509384, 0.579365611076355, 0.7400140762329102, -0.6508104205131531, -0.23783954977989197, -0.7102249264717102, -0.0478260256350...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-dane-2d14d683-10645434
autoevaluate
2022-07-11T13:14:03Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-07-11T13:14:03Z
2022-07-11T13:13:24.000Z
2022-07-11T13:13:24
--- type: predictions tags: - autotrain - evaluation datasets: - dane eval_info: task: entity_extraction model: saattrupdan/nbailab-base-ner-scandi metrics: [] dataset_name: dane dataset_config: default dataset_split: test col_mapping: tokens: tokens tags: ner_tags --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Token Classification * Model: saattrupdan/nbailab-base-ner-scandi * Dataset: dane * Config: default * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@KennethEnevoldsen](https://huggingface.co/KennethEnevoldsen) for evaluating this model.
[ -0.4328569173812866, -0.15250727534294128, 0.11283785849809647, 0.17715291678905487, -0.013649769127368927, -0.0124126011505723, 0.1538112312555313, -0.4186851978302002, 0.22423289716243744, 0.31457364559173584, -0.9272310137748718, -0.24246057868003845, -0.6661699414253235, 0.001454861601...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-cnn_dailymail-da2ad07e-10655435
autoevaluate
2022-07-12T05:57:39Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-07-12T05:57:39Z
2022-07-11T13:15:22.000Z
2022-07-11T13:15:22
--- type: predictions tags: - autotrain - evaluation datasets: - cnn_dailymail eval_info: task: summarization model: patrickvonplaten/bert2bert_cnn_daily_mail metrics: [] dataset_name: cnn_dailymail dataset_config: 3.0.0 dataset_split: train col_mapping: text: article target: highlights --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: patrickvonplaten/bert2bert_cnn_daily_mail * Dataset: cnn_dailymail * Config: 3.0.0 * Split: train To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@mumumumu](https://huggingface.co/mumumumu) for evaluating this model.
[ -0.42580702900886536, -0.2904776930809021, 0.14407290518283844, 0.275132954120636, -0.20691242814064026, -0.12589506804943085, -0.04095559939742088, -0.39063364267349243, 0.27347418665885925, 0.3640539050102234, -0.9069266319274902, -0.1843891590833664, -0.7665406465530396, -0.113391757011...
null
null
null
null
null
null
null
null
null
null
null
null
null
postbot/aeslc_kw
postbot
2022-08-07T12:14:34Z
13
0
null
[ "multilinguality:monolingual", "source_datasets:aeslc", "language:en", "license:mit", "text2text generation", "email", "email generation", "enron", "region:us" ]
2022-08-07T12:14:34Z
2022-07-11T13:23:36.000Z
2022-07-11T13:23:36
--- language: - en license: - mit multilinguality: - monolingual pretty_name: AESLC - Cleaned & Keyword Extracted source_datasets: - aeslc tags: - text2text generation - email - email generation - enron --- ## about - aeslc dataset but cleaned and keywords extracted to a new column - an EDA website generated via pandas profiling [is on netlify here](https://aeslc-kw-train-eda.netlify.app/) ``` DatasetDict({ train: Dataset({ features: ['email_body', 'subject_line', 'clean_email', 'clean_email_keywords'], num_rows: 14436 }) test: Dataset({ features: ['email_body', 'subject_line', 'clean_email', 'clean_email_keywords'], num_rows: 1906 }) validation: Dataset({ features: ['email_body', 'subject_line', 'clean_email', 'clean_email_keywords'], num_rows: 1960 }) }) ``` ## Python usage Basic example notebook [here](https://colab.research.google.com/gist/pszemraj/18742da8db4a99e57e95824eaead285a/scratchpad.ipynb). ```python from datasets import load_dataset dataset = load_dataset("postbot/aeslc_kw") ``` ## Citation ``` @InProceedings{zhang2019slg, author = "Rui Zhang and Joel Tetreault", title = "This Email Could Save Your Life: Introducing the Task of Email Subject Line Generation", booktitle = "Proceedings of The 57th Annual Meeting of the Association for Computational Linguistics", year = "2019", address = "Florence, Italy" } ```
[ -0.3074687421321869, -0.6695820689201355, -0.06160715967416763, -0.09610442072153091, -0.4871649742126465, -0.04782997444272041, -0.1384427398443222, -0.10793311148881912, 0.5014170408248901, 0.7499548196792603, -0.37542062997817993, -0.9313392043113708, -0.5731022953987122, 0.509926795959...
null
null
null
null
null
null
null
null
null
null
null
null
null
bhadresh-savani/image-to-style
bhadresh-savani
2022-07-20T08:58:29Z
13
0
null
[ "region:us" ]
2022-07-20T08:58:29Z
2022-07-11T14:22:03.000Z
2022-07-11T14:22:03
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622264862060547, 0.43461528420448303, -0.52829909324646, 0.7012971639633179, 0.7915720343589783, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104477167129517, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-conll2003-e2bfcc2b-10665436
autoevaluate
2022-07-11T14:24:36Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-07-11T14:24:36Z
2022-07-11T14:23:17.000Z
2022-07-11T14:23:17
--- type: predictions tags: - autotrain - evaluation datasets: - conll2003 eval_info: task: entity_extraction model: huggingface-course/bert-finetuned-ner metrics: ['jordyvl/ece'] dataset_name: conll2003 dataset_config: conll2003 dataset_split: test col_mapping: tokens: tokens tags: ner_tags --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Token Classification * Model: huggingface-course/bert-finetuned-ner * Dataset: conll2003 * Config: conll2003 * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@jordyvl](https://huggingface.co/jordyvl) for evaluating this model.
[ -0.49724650382995605, -0.3537701964378357, 0.06800234317779541, 0.06561827659606934, -0.005386353470385075, -0.11347762495279312, 0.0136004239320755, -0.5443346500396729, 0.18229134380817413, 0.34105199575424194, -0.9800329208374023, -0.10084112733602524, -0.6565409302711487, -0.0423544161...
null
null
null
null
null
null
null
null
null
null
null
null
null
carlosejimenez/flickr30k_clip-SimCLRv2-caption_pairs
carlosejimenez
2022-07-14T16:20:37Z
13
0
null
[ "region:us" ]
2022-07-14T16:20:37Z
2022-07-14T14:34:22.000Z
2022-07-14T14:34:22
Entry not found
[ -0.32276472449302673, -0.22568407654762268, 0.8622258901596069, 0.4346148371696472, -0.5282984972000122, 0.7012965679168701, 0.7915717363357544, 0.07618629932403564, 0.7746022939682007, 0.2563222646713257, -0.785281777381897, -0.22573848068714142, -0.9104482531547546, 0.5715669393539429, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-kmfoda__booksum-79c1c0d8-10905465
autoevaluate
2022-07-15T20:08:50Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-07-15T20:08:50Z
2022-07-14T18:31:38.000Z
2022-07-14T18:31:38
--- type: predictions tags: - autotrain - evaluation datasets: - kmfoda/booksum eval_info: task: summarization model: pszemraj/pegasus-large-summary-explain metrics: ['bleu', 'perplexity'] dataset_name: kmfoda/booksum dataset_config: kmfoda--booksum dataset_split: test col_mapping: text: chapter target: summary_text --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: pszemraj/pegasus-large-summary-explain * Dataset: kmfoda/booksum * Config: kmfoda--booksum * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@pszemraj](https://huggingface.co/pszemraj) for evaluating this model.
[ -0.5990710854530334, -0.061411503702402115, 0.2137870490550995, 0.1611979901790619, -0.26684078574180603, -0.2543359398841858, 0.07703611254692078, -0.35060882568359375, 0.26774120330810547, 0.5225921869277954, -1.0886160135269165, -0.2571146786212921, -0.696104109287262, -0.13508635759353...
null
null
null
null
null
null
null
null
null
null
null
null
null
tarteel-ai/EA-DI
tarteel-ai
2022-07-15T00:03:00Z
13
2
null
[ "region:us" ]
2022-07-15T00:03:00Z
2022-07-14T23:21:53.000Z
2022-07-14T23:21:53
Entry not found
[ -0.32276472449302673, -0.22568407654762268, 0.8622258901596069, 0.4346148371696472, -0.5282984972000122, 0.7012965679168701, 0.7915717363357544, 0.07618629932403564, 0.7746022939682007, 0.2563222646713257, -0.785281777381897, -0.22573848068714142, -0.9104482531547546, 0.5715669393539429, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
tarteel-ai/EA-UD
tarteel-ai
2022-07-15T03:04:41Z
13
0
null
[ "region:us" ]
2022-07-15T03:04:41Z
2022-07-15T02:28:41.000Z
2022-07-15T02:28:41
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622264862060547, 0.43461528420448303, -0.52829909324646, 0.7012971639633179, 0.7915720343589783, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104477167129517, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
DaliaSmirnov/HIMYM
DaliaSmirnov
2022-07-15T14:17:28Z
13
0
null
[ "region:us" ]
2022-07-15T14:17:28Z
2022-07-15T12:02:08.000Z
2022-07-15T12:02:08
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622264862060547, 0.43461528420448303, -0.52829909324646, 0.7012971639633179, 0.7915720343589783, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104477167129517, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
nbroad/mediasum
nbroad
2022-10-25T10:40:11Z
13
1
null
[ "task_categories:summarization", "multilinguality:monolingual", "size_categories:100K<n<1M", "language:en", "license:cc-by-nc-sa-4.0", "arxiv:2103.06410", "region:us" ]
2022-10-25T10:40:11Z
2022-07-15T21:42:51.000Z
2022-07-15T21:42:51
--- language: - en license: - cc-by-nc-sa-4.0 multilinguality: - monolingual size_categories: - 100K<n<1M task_categories: - summarization --- # MediaSum ## Description This large-scale media interview dataset contains 463.6K transcripts with abstractive summaries, collected from interview transcripts and overview / topic descriptions from NPR and CNN. ### **NOTE: The authors have requested that this dataset be used for research purposes only** ## Homepage https://github.com/zcgzcgzcg1/MediaSum ## Paper https://arxiv.org/abs/2103.06410 ## Authors ### Chenguang Zhu*, Yang Liu*, Jie Mei, Michael Zeng #### Microsoft Cognitive Services Research Group {chezhu,yaliu10,jimei,nzeng}@microsoft.com ## Citation @article{zhu2021mediasum, title={MediaSum: A Large-scale Media Interview Dataset for Dialogue Summarization}, author={Zhu, Chenguang and Liu, Yang and Mei, Jie and Zeng, Michael}, journal={arXiv preprint arXiv:2103.06410}, year={2021} } ## Dataset size Train: 443,596 Validation: 10,000 Test: 10,000 The splits were made by using the file located here: https://github.com/zcgzcgzcg1/MediaSum/tree/main/data ## Data details - id (string): unique identifier - program (string): the program this transcript came from - date (string): date of program - url (string): link to where audio and transcript are located - title (string): title of the program. some datapoints do not have a title - summary (string): summary of the program - utt (list of string): list of utterances by the speakers in the program. corresponds with `speaker` - speaker (list of string): list of speakers, corresponds with `utt` Example: ``` { "id": "NPR-11", "program": "Day to Day", "date": "2008-06-10", "url": "https://www.npr.org/templates/story/story.php?storyId=91356794", "title": "Researchers Find Discriminating Plants", "summary": "The \"sea rocket\" shows preferential treatment to plants that are its kin. Evolutionary plant ecologist Susan Dudley of McMaster University in Ontario discusses her discovery.", "utt": [ "This is Day to Day. I'm Madeleine Brand.", "And I'm Alex Cohen.", "Coming up, the question of who wrote a famous religious poem turns into a very unchristian battle.", "First, remember the 1970s? People talked to their houseplants, played them classical music. They were convinced plants were sensuous beings and there was that 1979 movie, \"The Secret Life of Plants.\"", "Only a few daring individuals, from the scientific establishment, have come forward with offers to replicate his experiments, or test his results. The great majority are content simply to condemn his efforts without taking the trouble to investigate their validity.", ... "OK. Thank you.", "That's Susan Dudley. She's an associate professor of biology at McMaster University in Hamilt on Ontario. She discovered that there is a social life of plants." ], "speaker": [ "MADELEINE BRAND, host", "ALEX COHEN, host", "ALEX COHEN, host", "MADELEINE BRAND, host", "Unidentified Male", ..." Professor SUSAN DUDLEY (Biology, McMaster University)", "MADELEINE BRAND, host" ] } ``` ## Using the dataset ```python from datasets import load_dataset ds = load_dataset("nbroad/mediasum") ``` ## Data location https://drive.google.com/file/d/1ZAKZM1cGhEw2A4_n4bGGMYyF8iPjLZni/view?usp=sharing ## License No license specified, but the authors have requested that this dataset be used for research purposes only.
[ -0.3479301929473877, -0.6266974806785583, 0.34550604224205017, 0.12358355522155762, -0.010281036607921124, 0.02670997567474842, -0.2733040153980255, -0.28375548124313354, 0.5061780214309692, 0.24566981196403503, -0.6919527053833008, -0.4490374028682709, -0.600079357624054, 0.15847733616828...
null
null
null
null
null
null
null
null
null
null
null
null
null
QuoQA-NLP/KoCC3M
QuoQA-NLP
2022-07-22T08:37:34Z
13
3
null
[ "region:us" ]
2022-07-22T08:37:34Z
2022-07-22T07:24:26.000Z
2022-07-22T07:24:26
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622264862060547, 0.43461528420448303, -0.52829909324646, 0.7012971639633179, 0.7915720343589783, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104477167129517, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
MrSemyon12/data_frame
MrSemyon12
2022-07-25T04:12:33Z
13
0
null
[ "region:us" ]
2022-07-25T04:12:33Z
2022-07-25T01:12:04.000Z
2022-07-25T01:12:04
Entry not found
[ -0.32276472449302673, -0.22568407654762268, 0.8622258901596069, 0.4346148371696472, -0.5282984972000122, 0.7012965679168701, 0.7915717363357544, 0.07618629932403564, 0.7746022939682007, 0.2563222646713257, -0.785281777381897, -0.22573848068714142, -0.9104482531547546, 0.5715669393539429, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
emilylearning/cond_ft_none_on_reddit__prcnt_100__test_run_False__bert-base-uncased
emilylearning
2022-07-25T12:55:20Z
13
0
null
[ "region:us" ]
2022-07-25T12:55:20Z
2022-07-25T06:22:17.000Z
2022-07-25T06:22:17
Entry not found
[ -0.32276472449302673, -0.22568407654762268, 0.8622258901596069, 0.4346148371696472, -0.5282984972000122, 0.7012965679168701, 0.7915717363357544, 0.07618629932403564, 0.7746022939682007, 0.2563222646713257, -0.785281777381897, -0.22573848068714142, -0.9104482531547546, 0.5715669393539429, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-squad_v2-fdec2e9c-11705559
autoevaluate
2022-07-25T07:29:26Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-07-25T07:29:26Z
2022-07-25T07:24:04.000Z
2022-07-25T07:24:04
--- type: predictions tags: - autotrain - evaluation datasets: - squad_v2 eval_info: task: extractive_question_answering model: deepset/xlm-roberta-large-squad2 metrics: [] dataset_name: squad_v2 dataset_config: squad_v2 dataset_split: validation col_mapping: context: context question: question answers-text: answers.text answers-answer_start: answers.answer_start --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Question Answering * Model: deepset/xlm-roberta-large-squad2 * Dataset: squad_v2 * Config: squad_v2 * Split: validation To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@sjrlee](https://huggingface.co/sjrlee) for evaluating this model.
[ -0.478790283203125, -0.4222498834133148, 0.4279681146144867, 0.11446921527385712, 0.0824565514922142, 0.1316302865743637, 0.02959442138671875, -0.4153931736946106, -0.04280597344040871, 0.5222002863883972, -1.3187062740325928, -0.12530925869941711, -0.5563628077507019, 0.03532204031944275,...
null
null
null
null
null
null
null
null
null
null
null
null
null
autoevaluate/autoeval-staging-eval-project-squad-810261fd-11725561
autoevaluate
2022-07-25T09:36:36Z
13
0
null
[ "autotrain", "evaluation", "region:us" ]
2022-07-25T09:36:36Z
2022-07-25T09:33:54.000Z
2022-07-25T09:33:54
--- type: predictions tags: - autotrain - evaluation datasets: - squad eval_info: task: extractive_question_answering model: Shanny/bert-finetuned-squad metrics: ['accuracy'] dataset_name: squad dataset_config: plain_text dataset_split: validation col_mapping: context: context question: question answers-text: answers.text answers-answer_start: answers.answer_start --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Question Answering * Model: Shanny/bert-finetuned-squad * Dataset: squad * Config: plain_text * Split: validation To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@ola13](https://huggingface.co/ola13) for evaluating this model.
[ -0.5294521450996399, -0.5205848813056946, 0.3441304564476013, 0.18083371222019196, 0.062188711017370224, -0.016785424202680588, 0.10639319568872452, -0.5022252798080444, 0.10807598382234573, 0.4439399242401123, -1.2792896032333374, -0.05655934289097786, -0.418372243642807, 0.07572413235902...
null
null
null
null
null
null
null
null
null
null
null
null
null
ttxy/weibo_4_moods
ttxy
2022-07-25T09:55:43Z
13
0
null
[ "region:us" ]
2022-07-25T09:55:43Z
2022-07-25T09:55:07.000Z
2022-07-25T09:55:07
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
naem1023/augmented-namuwiki
naem1023
2022-07-25T12:45:56Z
13
0
null
[ "license:apache-2.0", "region:us" ]
2022-07-25T12:45:56Z
2022-07-25T11:02:14.000Z
2022-07-25T11:02:14
--- license: apache-2.0 ---
[ -0.12853367626667023, -0.18616794049739838, 0.6529126763343811, 0.4943627417087555, -0.19319313764572144, 0.23607443273067474, 0.36071979999542236, 0.05056338757276535, 0.5793654322624207, 0.7400138974189758, -0.6508103013038635, -0.23783987760543823, -0.710224986076355, -0.047825977206230...
null
null
null
null
null
null
null
null
null
null
null
null
null
davanstrien/hf_model_metadata
davanstrien
2022-07-25T12:13:05Z
13
0
null
[ "region:us" ]
2022-07-25T12:13:05Z
2022-07-25T12:12:59.000Z
2022-07-25T12:12:59
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
naem1023/augmented-concat-100000
naem1023
2022-07-25T14:30:47Z
13
0
null
[ "license:apache-2.0", "region:us" ]
2022-07-25T14:30:47Z
2022-07-25T14:04:21.000Z
2022-07-25T14:04:21
--- license: apache-2.0 ---
[ -0.12853367626667023, -0.18616794049739838, 0.6529126763343811, 0.4943627417087555, -0.19319313764572144, 0.23607443273067474, 0.36071979999542236, 0.05056338757276535, 0.5793654322624207, 0.7400138974189758, -0.6508103013038635, -0.23783987760543823, -0.710224986076355, -0.047825977206230...
null
null
null
null
null
null
null
null
null
null
null
null
null