id
stringlengths
2
115
author
stringlengths
2
42
last_modified
timestamp[us, tz=UTC]
downloads
int64
0
8.87M
likes
int64
0
3.84k
paperswithcode_id
stringlengths
2
45
tags
list
lastModified
timestamp[us, tz=UTC]
createdAt
stringlengths
24
24
key
stringclasses
1 value
created
timestamp[us]
card
stringlengths
1
1.01M
embedding
list
library_name
stringclasses
21 values
pipeline_tag
stringclasses
27 values
mask_token
null
card_data
null
widget_data
null
model_index
null
config
null
transformers_info
null
spaces
null
safetensors
null
transformersInfo
null
modelId
stringlengths
5
111
embeddings
list
cahya/persona_empathetic
cahya
2022-02-19T22:49:35Z
153
0
null
[ "license:mit", "region:us" ]
2022-02-19T22:49:35Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
--- license: mit ---
[ -0.12853367626667023, -0.18616794049739838, 0.6529126763343811, 0.4943627417087555, -0.19319313764572144, 0.23607443273067474, 0.36071979999542236, 0.05056338757276535, 0.5793654322624207, 0.7400138974189758, -0.6508103013038635, -0.23783987760543823, -0.710224986076355, -0.047825977206230...
null
null
null
null
null
null
null
null
null
null
null
null
null
cakiki/arxiv-metadata
cakiki
2022-02-03T20:57:23Z
153
0
null
[ "license:cc0-1.0", "region:us" ]
2022-02-03T20:57:23Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
--- license: cc0-1.0 ---
[ -0.12853367626667023, -0.18616794049739838, 0.6529126763343811, 0.4943627417087555, -0.19319313764572144, 0.23607443273067474, 0.36071979999542236, 0.05056338757276535, 0.5793654322624207, 0.7400138974189758, -0.6508103013038635, -0.23783987760543823, -0.710224986076355, -0.047825977206230...
null
null
null
null
null
null
null
null
null
null
null
null
null
caythuoc/caoduoclieu
caythuoc
2023-06-15T10:41:13Z
153
0
null
[ "region:us" ]
2023-06-15T10:41:13Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
ccccccc/hdjw_94ejrjr
ccccccc
2021-02-18T07:41:38Z
153
0
null
[ "region:us" ]
2021-02-18T07:41:38Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
cdleong/temp_africaNLP_keyword_spotting_for_african_languages
cdleong
2022-10-25T09:07:32Z
153
0
null
[ "language:wo", "language:fuc", "language:srr", "language:mnk", "language:snk", "region:us" ]
2022-10-25T09:07:32Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
--- language: - wo - fuc - srr - mnk - snk --- ## Dataset Description - **Homepage:** https://zenodo.org/record/4661645 TEMPORARY TEST DATASET Not for actual use! Attempting to test out a dataset script for loading https://zenodo.org/record/4661645
[ -0.2611089050769806, -0.16739805042743683, 0.1438436210155487, 0.5390400886535645, -0.7581366896629333, 0.13277389109134674, -0.10329926759004593, 0.1648896485567093, 0.1767561435699463, 0.660321831703186, -1.231587529182434, -0.6473388671875, 0.040851250290870667, 0.34319373965263367, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
cem/film
cem
2021-12-23T22:02:57Z
153
0
null
[ "region:us" ]
2021-12-23T22:02:57Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.32276472449302673, -0.22568407654762268, 0.8622258901596069, 0.4346148371696472, -0.5282984972000122, 0.7012965679168701, 0.7915717363357544, 0.07618629932403564, 0.7746022939682007, 0.2563222646713257, -0.785281777381897, -0.22573848068714142, -0.9104482531547546, 0.5715669393539429, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
cemigo/taylor_vs_shakes
cemigo
2021-03-14T23:45:59Z
153
0
null
[ "region:us" ]
2021-03-14T23:45:59Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
This dataset has 336 pieces of quotes from William Shakespeare and Taylor Swift (labeled) for supervised classification. Source: https://www.kaggle.com/kellylougheed/tswift-vs-shakespeare
[ -0.3528725206851959, -0.49498608708381653, 0.17421595752239227, 0.3363476097583771, 0.0019012205302715302, 0.06886040419340134, -0.025744786486029625, -0.41607120633125305, 0.08843306452035904, 0.862523078918457, -0.9132828712463379, -0.16851173341274261, -0.5095446109771729, 0.04360641166...
null
null
null
null
null
null
null
null
null
null
null
null
null
cemigo/test-data
cemigo
2021-02-07T23:49:41Z
153
0
null
[ "region:us" ]
2021-02-07T23:49:41Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.32276472449302673, -0.22568407654762268, 0.8622258901596069, 0.4346148371696472, -0.5282984972000122, 0.7012965679168701, 0.7915717363357544, 0.07618629932403564, 0.7746022939682007, 0.2563222646713257, -0.785281777381897, -0.22573848068714142, -0.9104482531547546, 0.5715669393539429, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
csarron/25m-img-caps
csarron
2022-03-28T18:51:26Z
153
1
null
[ "region:us" ]
2022-03-28T18:51:26Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
see https://huggingface.co/datasets/csarron/4m-img-caps for example usage
[ -0.6945481896400452, -0.27406325936317444, 0.12098154425621033, 0.5052483081817627, -0.13724862039089203, 0.032137878239154816, 0.19517825543880463, -0.45081761479377747, 0.7188369631767273, 0.5162168145179749, -1.0182642936706543, -0.7857872247695923, -0.539254903793335, 0.264833718538284...
null
null
null
null
null
null
null
null
null
null
null
null
null
csarron/4m-img-caps
csarron
2022-03-28T18:50:53Z
153
1
null
[ "region:us" ]
2022-03-28T18:50:53Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
see [read_pyarrow.py](https://gist.github.com/csarron/df712e53c9e0dcaad4eb6843e7a3d51c#file-read_pyarrow-py) for how to read one pyarrow file. example PyTorch dataset: ```python from torch.utils.data import Dataset class ImageCaptionArrowDataset(Dataset): def __init__( self, dataset_file, tokenizer, ): import pyarrow as pa data = [pa.ipc.open_file(pa.memory_map(f, "rb")).read_all() for f in glob.glob(dataset_file)] self.data = pa.concat_tables(data) # do other initialization, like init image preprocessing fn, def __getitem__(self, index): # item_id = self.data["id"][index].as_py() text = self.data["text"][index].as_py() # get text if isinstance(text, list): text = random.choice(text) img_bytes = self.data["image"][index].as_py() # get image bytes # do some processing with image and text, return the features # img_feat = self.image_bytes_to_tensor(img_bytes) # inputs = self.tokenizer( # text, # padding="max_length", # max_length=self.max_text_len, # truncation=True, # return_token_type_ids=True, # return_attention_mask=True, # add_special_tokens=True, # return_tensors="pt", # ) # input_ids = inputs.input_ids.squeeze(0) # attention_mask = inputs.attention_mask.squeeze(0) # return { # # "item_ids": item_id, # "text_ids": input_ids, # "input_ids": input_ids, # "text_masks": attention_mask, # "pixel_values": img_feat, # } def __len__(self): return len(self.data) ```
[ -0.2299148440361023, -0.544771671295166, 0.10518138855695724, 0.14956609904766083, -0.3305014967918396, -0.08597679436206818, -0.01706288382411003, -0.33755263686180115, 0.17606347799301147, 0.2263861745595932, -0.3430895507335663, -0.5375257730484009, -0.6523412466049194, 0.02987181767821...
null
null
null
null
null
null
null
null
null
null
null
null
null
image-search-2/unsplash_lite_image_dataset
image-search-2
2021-11-19T12:44:46Z
153
1
null
[ "region:us" ]
2021-11-19T12:44:46Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
# The Unsplash Dataset ![](https://unsplash.com/blog/content/images/2020/08/dataheader.jpg) The Unsplash Dataset is made up of over 250,000+ contributing global photographers and data sourced from hundreds of millions of searches across a nearly unlimited number of uses and contexts. Due to the breadth of intent and semantics contained within the Unsplash dataset, it enables new opportunities for research and learning. The Unsplash Dataset is offered in two datasets: - the Lite dataset: available for commercial and noncommercial usage, containing 25k nature-themed Unsplash photos, 25k keywords, and 1M searches - the Full dataset: available for noncommercial usage, containing 3M+ high-quality Unsplash photos, 5M keywords, and over 250M searches As the Unsplash library continues to grow, we’ll release updates to the dataset with new fields and new images, with each subsequent release being [semantically versioned](https://semver.org/). We welcome any feedback regarding the content of the datasets or their format. With your input, we hope to close the gap between the data we provide and the data that you would like to leverage. You can [open an issue](https://github.com/unsplash/datasets/issues/new/choose) to report a problem or to let us know what you would like to see in the next release of the datasets. For more on the Unsplash Dataset, see [our announcement](https://unsplash.com/blog/the-unsplash-dataset/) and [site](https://unsplash.com/data). ## Download ### Lite Dataset The Lite dataset contains all of the same fields as the Full dataset, but is limited to ~25,000 photos. It can be used for both commercial and non-commercial usage, provided you abide by [the terms](https://github.com/unsplash/datasets/blob/master/TERMS.md). [⬇️ Download the Lite dataset](https://unsplash.com/data/lite/latest) [~650MB compressed, ~1.4GB raw] ### Full Dataset The Full dataset is available for non-commercial usage and all uses must abide by [the terms](https://github.com/unsplash/datasets/blob/master/TERMS.md). To access, please go to [unsplash.com/data](https://unsplash.com/data) and request access. The dataset weighs ~20 GB compressed (~43GB raw)). ## Documentation See the [documentation for a complete list of tables and fields](https://github.com/unsplash/datasets/blob/master/DOCS.md). ## Usage You can follow these examples to load the dataset in these common formats: - [Load the dataset in a PostgreSQL database](https://github.com/unsplash/datasets/tree/master/how-to/psql) - [Load the dataset in a Python environment](https://github.com/unsplash/datasets/tree/master/how-to/python) - [Submit an example doc](https://github.com/unsplash/datasets/blob/master/how-to/README.md#submit-an-example) ## Share your work We're making this data open and available with the hopes of enabling researchers and developers to discover interesting and useful connections in the data. We'd love to see what you create, whether that's a research paper, a machine learning model, a blog post, or just an interesting discovery in the data. Send us an email at [data@unsplash.com](mailto:data@unsplash.com). If you're using the dataset in a research paper, you can attribute the dataset as `Unsplash Lite Dataset 1.2.0` or `Unsplash Full Dataset 1.2.0` and link to the permalink [`unsplash.com/data`](https://unsplash.com/data). ---- The Unsplash Dataset is made available for research purposes. [It cannot be used to redistribute the images contained within](https://github.com/unsplash/datasets/blob/master/TERMS.md). To use the Unsplash library in a product, see [the Unsplash API](https://unsplash.com/developers). ![](https://unsplash.com/blog/content/images/2020/08/footer-alt.jpg)
[ -0.0999462902545929, -0.18598991632461548, 0.14069034159183502, 0.09842837601900101, -0.5665063858032227, 0.19609595835208893, -0.3083202540874481, -0.3755965828895569, 0.3953779339790344, 0.5846114158630371, -0.5519187450408936, -0.7459720373153687, -0.13306768238544464, 0.127736344933509...
null
null
null
null
null
null
null
null
null
null
null
null
null
jimregan/foinse
jimregan
2021-10-06T20:42:52Z
153
0
null
[ "region:us" ]
2021-10-06T20:42:52Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
jakartaresearch/google-play-review
jakartaresearch
2022-08-06T16:24:49Z
153
4
null
[ "task_categories:text-classification", "task_ids:sentiment-classification", "annotations_creators:found", "language_creators:found", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:id", "license:cc-by-4.0", "sentiment", "google-play", "indonesia...
2022-08-06T16:24:49Z
2022-08-06T05:00:32.000Z
2022-08-06T05:00:32
--- annotations_creators: - found language: - id language_creators: - found license: - cc-by-4.0 multilinguality: - monolingual pretty_name: Indonesian Google Play Review size_categories: - 1K<n<10K source_datasets: - original tags: - sentiment - google-play - indonesian task_categories: - text-classification task_ids: - sentiment-classification --- # Dataset Card for Indonesian Google Play Review ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary Scrapped from e-commerce app on Google Play. ### Supported Tasks and Leaderboards Sentiment Analysis ### Languages Indonesian ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@andreaschandra](https://github.com/andreaschandra) for adding this dataset.
[ -0.45517972111701965, -0.49983277916908264, -0.051154736429452896, 0.49620047211647034, -0.42576295137405396, 0.08203400671482086, -0.2973781228065491, -0.37442344427108765, 0.6587487459182739, 0.8926987051963806, -0.9955716133117676, -0.9578273892402649, -0.6353380680084229, 0.21590693295...
null
null
null
null
null
null
null
null
null
null
null
null
null
cardiffnlp/tweet_topic_single
cardiffnlp
2022-11-27T11:25:34Z
153
3
null
[ "task_categories:text-classification", "task_ids:sentiment-classification", "multilinguality:monolingual", "size_categories:1k<10K", "language:en", "license:other", "arxiv:2209.09824", "region:us" ]
2022-11-27T11:25:34Z
2022-09-02T00:20:17.000Z
2022-09-02T00:20:17
--- language: - en license: - other multilinguality: - monolingual size_categories: - 1k<10K task_categories: - text-classification task_ids: - sentiment-classification pretty_name: TweetTopicSingle --- # Dataset Card for "cardiffnlp/tweet_topic_single" ## Dataset Description - **Paper:** [https://arxiv.org/abs/2209.09824](https://arxiv.org/abs/2209.09824) - **Dataset:** Tweet Topic Dataset - **Domain:** Twitter - **Number of Class:** 6 ### Dataset Summary This is the official repository of TweetTopic (["Twitter Topic Classification , COLING main conference 2022"](https://arxiv.org/abs/2209.09824)), a topic classification dataset on Twitter with 6 labels. Each instance of TweetTopic comes with a timestamp which distributes from September 2019 to August 2021. See [cardiffnlp/tweet_topic_multi](https://huggingface.co/datasets/cardiffnlp/tweet_topic_multi) for multi label version of TweetTopic. The tweet collection used in TweetTopic is same as what used in [TweetNER7](https://huggingface.co/datasets/tner/tweetner7). The dataset is integrated in [TweetNLP](https://tweetnlp.org/) too. ### Preprocessing We pre-process tweets before the annotation to normalize some artifacts, converting URLs into a special token `{{URL}}` and non-verified usernames into `{{USERNAME}}`. For verified usernames, we replace its display name (or account name) with symbols `{@}`. For example, a tweet ``` Get the all-analog Classic Vinyl Edition of "Takin' Off" Album from @herbiehancock via @bluenoterecords link below: http://bluenote.lnk.to/AlbumOfTheWeek ``` is transformed into the following text. ``` Get the all-analog Classic Vinyl Edition of "Takin' Off" Album from {@herbiehancock@} via {@bluenoterecords@} link below: {{URL}} ``` A simple function to format tweet follows below. ```python import re from urlextract import URLExtract extractor = URLExtract() def format_tweet(tweet): # mask web urls urls = extractor.find_urls(tweet) for url in urls: tweet = tweet.replace(url, "{{URL}}") # format twitter account tweet = re.sub(r"\b(\s*)(@[\S]+)\b", r'\1{\2@}', tweet) return tweet target = """Get the all-analog Classic Vinyl Edition of "Takin' Off" Album from @herbiehancock via @bluenoterecords link below: http://bluenote.lnk.to/AlbumOfTheWeek""" target_format = format_tweet(target) print(target_format) 'Get the all-analog Classic Vinyl Edition of "Takin\' Off" Album from {@herbiehancock@} via {@bluenoterecords@} link below: {{URL}}' ``` ### Data Splits | split | number of texts | description | |:------------------------|-----:|------:| | test_2020 | 376 | test dataset from September 2019 to August 2020 | | test_2021 | 1693 | test dataset from September 2020 to August 2021 | | train_2020 | 2858 | training dataset from September 2019 to August 2020 | | train_2021 | 1516 | training dataset from September 2020 to August 2021 | | train_all | 4374 | combined training dataset of `train_2020` and `train_2021` | | validation_2020 | 352 | validation dataset from September 2019 to August 2020 | | validation_2021 | 189 | validation dataset from September 2020 to August 2021 | | train_random | 2830 | randomly sampled training dataset with the same size as `train_2020` from `train_all` | | validation_random | 354 | randomly sampled training dataset with the same size as `validation_2020` from `validation_all` | | test_coling2022_random | 3399 | random split used in the COLING 2022 paper | | train_coling2022_random | 3598 | random split used in the COLING 2022 paper | | test_coling2022 | 3399 | temporal split used in the COLING 2022 paper | | train_coling2022 | 3598 | temporal split used in the COLING 2022 paper | For the temporal-shift setting, model should be trained on `train_2020` with `validation_2020` and evaluate on `test_2021`. In general, model would be trained on `train_all`, the most representative training set with `validation_2021` and evaluate on `test_2021`. **IMPORTANT NOTE:** To get a result that is comparable with the results of the COLING 2022 Tweet Topic paper, please use `train_coling2022` and `test_coling2022` for temporal-shift, and `train_coling2022_random` and `test_coling2022_random` fir random split (the coling2022 split does not have validation set). ### Models | model | training data | F1 | F1 (macro) | Accuracy | |:------------------------------------------------------------------------------------------------------------------------------------------------------------|:------------------|---------:|-------------:|-----------:| | [cardiffnlp/roberta-large-tweet-topic-single-all](https://huggingface.co/cardiffnlp/roberta-large-tweet-topic-single-all) | all (2020 + 2021) | 0.896043 | 0.800061 | 0.896043 | | [cardiffnlp/roberta-base-tweet-topic-single-all](https://huggingface.co/cardiffnlp/roberta-base-tweet-topic-single-all) | all (2020 + 2021) | 0.887773 | 0.79793 | 0.887773 | | [cardiffnlp/twitter-roberta-base-2019-90m-tweet-topic-single-all](https://huggingface.co/cardiffnlp/twitter-roberta-base-2019-90m-tweet-topic-single-all) | all (2020 + 2021) | 0.892499 | 0.774494 | 0.892499 | | [cardiffnlp/twitter-roberta-base-dec2020-tweet-topic-single-all](https://huggingface.co/cardiffnlp/twitter-roberta-base-dec2020-tweet-topic-single-all) | all (2020 + 2021) | 0.890136 | 0.776025 | 0.890136 | | [cardiffnlp/twitter-roberta-base-dec2021-tweet-topic-single-all](https://huggingface.co/cardiffnlp/twitter-roberta-base-dec2021-tweet-topic-single-all) | all (2020 + 2021) | 0.894861 | 0.800952 | 0.894861 | | [cardiffnlp/roberta-large-tweet-topic-single-2020](https://huggingface.co/cardiffnlp/roberta-large-tweet-topic-single-2020) | 2020 only | 0.878913 | 0.70565 | 0.878913 | | [cardiffnlp/roberta-base-tweet-topic-single-2020](https://huggingface.co/cardiffnlp/roberta-base-tweet-topic-single-2020) | 2020 only | 0.868281 | 0.729667 | 0.868281 | | [cardiffnlp/twitter-roberta-base-2019-90m-tweet-topic-single-2020](https://huggingface.co/cardiffnlp/twitter-roberta-base-2019-90m-tweet-topic-single-2020) | 2020 only | 0.882457 | 0.740187 | 0.882457 | | [cardiffnlp/twitter-roberta-base-dec2020-tweet-topic-single-2020](https://huggingface.co/cardiffnlp/twitter-roberta-base-dec2020-tweet-topic-single-2020) | 2020 only | 0.87596 | 0.746275 | 0.87596 | | [cardiffnlp/twitter-roberta-base-dec2021-tweet-topic-single-2020](https://huggingface.co/cardiffnlp/twitter-roberta-base-dec2021-tweet-topic-single-2020) | 2020 only | 0.877732 | 0.746119 | 0.877732 | Model fine-tuning script can be found [here](https://huggingface.co/datasets/cardiffnlp/tweet_topic_single/blob/main/lm_finetuning.py). ## Dataset Structure ### Data Instances An example of `train` looks as follows. ```python { "text": "Game day for {{USERNAME}} U18\u2019s against {{USERNAME}} U18\u2019s. Even though it\u2019s a \u2018home\u2019 game for the people that have settled in Mid Wales it\u2019s still a 4 hour round trip for us up to Colwyn Bay. Still enjoy it though!", "date": "2019-09-08", "label": 4, "id": "1170606779568463874", "label_name": "sports_&_gaming" } ``` ### Label ID The label2id dictionary can be found at [here](https://huggingface.co/datasets/tner/tweet_topic_single/raw/main/dataset/label.single.json). ```python { "arts_&_culture": 0, "business_&_entrepreneurs": 1, "pop_culture": 2, "daily_life": 3, "sports_&_gaming": 4, "science_&_technology": 5 } ``` ### Citation Information ``` @inproceedings{dimosthenis-etal-2022-twitter, title = "{T}witter {T}opic {C}lassification", author = "Antypas, Dimosthenis and Ushio, Asahi and Camacho-Collados, Jose and Neves, Leonardo and Silva, Vitor and Barbieri, Francesco", booktitle = "Proceedings of the 29th International Conference on Computational Linguistics", month = oct, year = "2022", address = "Gyeongju, Republic of Korea", publisher = "International Committee on Computational Linguistics" } ```
[ -0.41150397062301636, -0.5397923588752747, 0.19379831850528717, 0.31822553277015686, -0.39666521549224854, 0.013945408165454865, -0.26671135425567627, -0.5954558253288269, 0.6959244608879089, 0.07942402362823486, -0.8865302801132202, -0.7433744072914124, -0.5327351093292236, 0.099146649241...
null
null
null
null
null
null
null
null
null
null
null
null
null
LinhDuong/chatdoctor-5k
LinhDuong
2023-03-28T07:32:21Z
153
0
null
[ "license:apache-2.0", "arxiv:2303.14070", "region:us" ]
2023-03-28T07:32:21Z
2023-03-28T07:23:57.000Z
2023-03-28T07:23:57
--- license: apache-2.0 --- This ChatDoctor-5K dataset is collected from this paper https://arxiv.org/pdf/2303.14070.pdf Alternatively, you can download the original dataset from this link https://drive.google.com/file/d/1nDTKZ3wZbZWTkFMBkxlamrzbNz0frugg/view?usp=sharing
[ -0.515437662601471, -0.3415053188800812, 0.05897929519414902, -0.17993639409542084, -0.11729735136032104, 0.03404417634010315, 0.1897708922624588, -0.27391883730888367, 0.2561528980731964, 0.8184594511985779, -0.6972082853317261, -0.43813708424568176, -0.5306776165962219, 0.039448883384466...
null
null
null
null
null
null
null
null
null
null
null
null
null
bigheiniuJ/EvalMetaICLAll
bigheiniuJ
2023-07-24T06:39:16Z
153
0
null
[ "region:us" ]
2023-07-24T06:39:16Z
2023-07-23T20:34:43.000Z
2023-07-23T20:34:43
--- dataset_info: features: - name: task dtype: string - name: input dtype: string - name: output dtype: string - name: options sequence: string - name: seed dtype: string - name: split dtype: string splits: - name: meta_train num_bytes: 648803062 num_examples: 1111614 - name: meta_eval_100shot num_bytes: 1798838431 num_examples: 2725939 download_size: 1076308849 dataset_size: 2447641493 --- # Dataset Card for "EvalMetaICLAll" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.5287930965423584, -0.5164644718170166, 0.127943754196167, 0.29996296763420105, -0.1837684065103531, 0.3474554121494293, 0.2346542626619339, -0.2214013636112213, 0.9856589436531067, 0.6218628883361816, -0.7527875900268555, -0.8235347867012024, -0.5683114528656006, -0.19690248370170593, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
RealTimeData/bbc_latest
RealTimeData
2023-11-27T00:35:05Z
153
0
null
[ "region:us" ]
2023-11-27T00:35:05Z
2023-08-19T23:03:43.000Z
2023-08-19T23:03:43
--- {} --- # Latest BBC News You could always access the latest BBC News articles via this dataset. We update the dataset weekly, on every Sunday. So the dataset always provides the latest BBC News article from the last week. The current dataset on main branch contains the latest BBC News articles submitted from 2023-11-20 to 2023-11-27. The data collection is conducted on 2023-11-27. Use the dataset via: ``` ds = datasets.load_dataset('RealTimeData/bbc_latest') ``` # Previsou versions You could access previous versions by requesting different branches. For example, you could find the 2023-08-20 version via: ``` ds = datasets.load_dataset('RealTimeData/bbc_latest', revision = '2023-08-20') ``` Check all available versions by clicking the "Files and versions" button on the top bar.
[ -0.4674071669578552, -0.5754712224006653, 0.4863301217556, 0.25712665915489197, -0.6121233701705933, 0.21553201973438263, -0.2681798040866852, -0.6393352150917053, 0.5421441793441772, 0.7454524636268616, -0.9205769300460815, -0.4685693681240082, -0.5543708801269531, 0.15941977500915527, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
jamescalam/agent-conversations-retrieval-tool
jamescalam
2023-08-27T12:57:37Z
153
7
null
[ "region:us" ]
2023-08-27T12:57:37Z
2023-08-27T12:56:16.000Z
2023-08-27T12:56:16
Entry not found
[ -0.32276472449302673, -0.22568407654762268, 0.8622258901596069, 0.4346148371696472, -0.5282984972000122, 0.7012965679168701, 0.7915717363357544, 0.07618629932403564, 0.7746022939682007, 0.2563222646713257, -0.785281777381897, -0.22573848068714142, -0.9104482531547546, 0.5715669393539429, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
EleutherAI/sycophancy
EleutherAI
2023-09-05T15:14:40Z
153
1
null
[ "region:us" ]
2023-09-05T15:14:40Z
2023-08-29T07:58:29.000Z
2023-08-29T07:58:29
Entry not found
[ -0.32276472449302673, -0.22568407654762268, 0.8622258901596069, 0.4346148371696472, -0.5282984972000122, 0.7012965679168701, 0.7915717363357544, 0.07618629932403564, 0.7746022939682007, 0.2563222646713257, -0.785281777381897, -0.22573848068714142, -0.9104482531547546, 0.5715669393539429, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
benjis/bigvul
benjis
2023-08-31T03:02:50Z
153
0
null
[ "region:us" ]
2023-08-31T03:02:50Z
2023-08-31T02:55:32.000Z
2023-08-31T02:55:32
--- configs: - config_name: default data_files: - split: train path: data/train-* - split: validation path: data/validation-* - split: test path: data/test-* dataset_info: features: - name: CVE ID dtype: string - name: CVE Page dtype: string - name: CWE ID dtype: string - name: codeLink dtype: string - name: commit_id dtype: string - name: commit_message dtype: string - name: func_after dtype: string - name: func_before dtype: string - name: lang dtype: string - name: project dtype: string - name: vul dtype: int8 splits: - name: train num_bytes: 404950685.2579571 num_examples: 150908 - name: validation num_bytes: 88684597.21877055 num_examples: 33049 - name: test num_bytes: 88687280.64632414 num_examples: 33050 download_size: 252969708 dataset_size: 582322563.1230518 --- # Dataset Card for "bigvul" Unofficial, not affiliated with the authors. - **Paper:** https://doi.org/10.1145/3379597.3387501 - **Repository:** https://github.com/ZeoVan/MSR_20_Code_vulnerability_CSV_Dataset
[ 0.12298933416604996, -0.43564391136169434, 0.101936936378479, 0.19315655529499054, -0.4602789878845215, -0.020963475108146667, -0.2328115552663803, -0.24503077566623688, 0.26893147826194763, 0.30113109946250916, -0.35672104358673096, -1.0613905191421509, -0.6851311922073364, -0.07864366471...
null
null
null
null
null
null
null
null
null
null
null
null
null
legacy107/spamming-email-classification
legacy107
2023-10-02T09:39:55Z
153
0
null
[ "region:us" ]
2023-10-02T09:39:55Z
2023-09-25T14:22:14.000Z
2023-09-25T14:22:14
--- configs: - config_name: default data_files: - split: train path: data/train-* - split: val path: data/val-* - split: test path: data/test-* dataset_info: features: - name: Text dtype: string - name: Spam dtype: int64 splits: - name: train num_bytes: 7196065 num_examples: 4556 - name: val num_bytes: 819608 num_examples: 569 - name: test num_bytes: 925859 num_examples: 570 download_size: 4959617 dataset_size: 8941532 --- # Dataset Card for "spamming-email-classification" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.5479223132133484, -0.4092443585395813, -0.08031372725963593, 0.35796186327934265, 0.10786420106887817, -0.017265083268284798, 0.3324497640132904, -0.17351588606834412, 0.4617313742637634, 0.7126885056495667, -0.6400539875030518, -0.6551244258880615, -0.9767588973045349, -0.2762328684329...
null
null
null
null
null
null
null
null
null
null
null
null
null
LDJnr/Pure-Dove
LDJnr
2023-11-21T17:55:19Z
153
21
null
[ "task_categories:conversational", "task_categories:question-answering", "task_categories:text-generation", "size_categories:1K<n<10K", "language:en", "license:apache-2.0", "Physics", "Biology", "Math", "Chemistry", "Culture", "Logic", "Roleplay", "region:us" ]
2023-11-21T17:55:19Z
2023-09-26T02:06:24.000Z
2023-09-26T02:06:24
--- license: apache-2.0 task_categories: - conversational - question-answering - text-generation language: - en tags: - Physics - Biology - Math - Chemistry - Culture - Logic - Roleplay pretty_name: Pure-Dove size_categories: - 1K<n<10K --- ## This is the Official Pure-Dove dataset. Over 3K multi-turn examples, and many more coming soon! This dataset aims to be the largest highest quality cluster of real human back and forth conversations with GPT-4. Steps have even been done to ensure that only the best GPT-4 conversations in comparisons are kept, there are many instances where two GPT-4 responses are rated as equal to eachother or as both bad. We exclude all such responses from Pure Dove and make sure to only include ChatBot Arena responses that are voted as being better even against another instance of GPT-4. - Comprised of over 3000 highly filtered multi-turn conversations between GPT-4 and real humans. - Average context length per conversation is over 800 tokens. ## Purpose? - This dataset is not particularly intended to be trained on by itself, however, the size and quality of this dataset can work wonderfully as a supplemmentary addition to virtually any multi-turn compatible dataset. I encourage this use, all I ask is proper credits given for such! ## Quality filtering and cleaning. - The conversations were sourced from openly datasets such as ShareGPT and ChatBotArena by Lmsys, however, a large portion of these chats were riddled with hallucinations and abnormal distributions of different languages. - Extensive cleaning was done to filter out instances of overt AI moralizing or related behaviour, such as "As an AI language model" and "September 2021", not just in english, but other languages too! ## Credits During the curation process, there can be some relatively arduos steps when it comes to actually executing on the best experimentation or concepts for how to filter examples out. Luckily there is folks over at NousResearch that helped expedite this process with little to no sacrifices in quality, big credit to J-Supha within NousResearch specifically for making these types of significant contributions. ## Future Plans & How you can help! This is a relatively early build amongst the grand plans for the future of what I plan to work on! In the near future we plan on leveraging the help of domain specific expert volunteers to eliminate any mathematically/verifiably incorrect answers from training curations of different types of datasets. If you have at-least a bachelors in mathematics, physics, biology or chemistry and would like to volunteer even just 30 minutes of your expertise time, please contact LDJ on discord! Citation: ``` @article{daniele2023amplify-instruct, title={Amplify-Instruct: Synthetically Generated Diverse Multi-turn Conversations for Effecient LLM Training.}, author={Daniele, Luigi and Suphavadeeprasit}, journal={arXiv preprint arXiv:(comming soon)}, year={2023} } ```
[ -0.4659762978553772, -0.82603520154953, 0.4255504310131073, 0.17199984192848206, -0.15100425481796265, 0.035484857857227325, -0.17246367037296295, -0.44513750076293945, 0.16209010779857635, 0.5178656578063965, -0.7029075026512146, -0.4168851673603058, -0.4708016812801361, 0.071703262627124...
null
null
null
null
null
null
null
null
null
null
null
null
null
maxolotl/must-c-en-es-wait3-02
maxolotl
2023-10-22T07:48:24Z
153
0
null
[ "region:us" ]
2023-10-22T07:48:24Z
2023-10-22T07:48:05.000Z
2023-10-22T07:48:05
--- dataset_info: features: - name: current_source dtype: string - name: current_target dtype: string - name: target_token dtype: string splits: - name: train num_bytes: 995120593 num_examples: 5240243 - name: test num_bytes: 9960448 num_examples: 57187 - name: validation num_bytes: 5429701 num_examples: 27549 download_size: 184348036 dataset_size: 1010510742 --- # Dataset Card for "must-c-en-es-wait3-02" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.6564971208572388, -0.11159498989582062, 0.4751039445400238, 0.7593249082565308, -0.10837173461914062, -0.11256479471921921, 0.3184361457824707, -0.4941241443157196, 0.7914745807647705, 0.6035370826721191, -1.1325933933258057, -0.5855128765106201, -0.6314719319343567, 0.19302424788475037...
null
null
null
null
null
null
null
null
null
null
null
null
null
Karavet/pioNER-Armenian-Named-Entity
Karavet
2022-10-21T16:07:06Z
152
1
null
[ "task_ids:named-entity-recognition", "multilinguality:monolingual", "language:hy", "license:apache-2.0", "region:us" ]
2022-10-21T16:07:06Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
--- language: [hy] task_categories: [named-entity-recognition] multilinguality: [monolingual] task_ids: [named-entity-recognition] license: [apache-2.0] --- ## Table of Contents - [Table of Contents](#table-of-contents) - [pioNER - named entity annotated datasets](#pioNER---named-entity-annotated-datasets) - [Silver-standard dataset](#silver-standard-dataset) - [Gold-standard dataset](#gold-standard-dataset) # pioNER - named entity annotated datasets pioNER corpus provides gold-standard and automatically generated named-entity datasets for the Armenian language. Alongside the datasets, we release 50-, 100-, 200-, and 300-dimensional GloVe word embeddings trained on a collection of Armenian texts from Wikipedia, news, blogs, and encyclopedia. ## Silver-standard dataset The generated corpus is automatically extracted and annotated using Armenian Wikipedia. We used a modification of [Nothman et al](https://www.researchgate.net/publication/256660013_Learning_multilingual_named_entity_recognition_from_Wikipedia) and [Sysoev and Andrianov](http://www.dialog-21.ru/media/3433/sysoevaaandrianovia.pdf) approaches to create this corpus. This approach uses links between Wikipedia articles to extract fragments of named-entity annotated texts. The corpus is split into train and development sets. *Table 1. Statistics for pioNER train, development and test sets* | dataset | #tokens | #sents | annotation | texts' source | |-------------|:--------:|:-----:|:--------:|:-----:| | train | 130719 | 5964 | automatic | Wikipedia | | dev | 32528 | 1491 | automatic | Wikipedia | | test | 53606 | 2529 | manual | iLur.am | ## Gold-standard dataset This dataset is a collection of over 250 news articles from iLur.am with manual named-entity annotation. It includes sentences from political, sports, local and world news, and is comparable in size with the test sets of other languages (Table 2). We aim it to serve as a benchmark for future named entity recognition systems designed for the Armenian language. The dataset contains annotations for 3 popular named entity classes: people (PER), organizations (ORG), and locations (LOC), and is released in CoNLL03 format with IOB tagging scheme. During annotation, we generally relied on categories and [guidelines assembled by BBN](https://catalog.ldc.upenn.edu/docs/LDC2005T33/BBN-Types-Subtypes.html) Technologies for TREC 2002 question answering track Tokens and sentences were segmented according to the UD standards for the Armenian language from [ArmTreebank project](http://armtreebank.yerevann.com/tokenization/process/). *Table 2. Comparison of pioNER gold-standard test set with test sets for English, Russian, Spanish and German* | test dataset | #tokens | #LOC | #ORG | #PER | |-------------|:--------:|:-----:|:--------:|:-----:| | Armenian pioNER | 53606 | 1312 | 1338 | 1274 | | Russian factRuEval-2016 | 59382 | 1239 | 1595 | 1353 | | German CoNLL03 | 51943 | 1035 | 773 | 1195 | | Spanish CoNLL02 | 51533 | 1084 | 1400 | 735 | | English CoNLL03 | 46453 | 1668 | 1661 | 1671 |
[ -0.8337982296943665, -0.6106128096580505, 0.0016679394757375121, -0.024099867790937424, -0.41270124912261963, -0.18098530173301697, -0.5245099067687988, -0.5758296251296997, 0.3895459771156311, 0.6980950832366943, -0.25429248809814453, -0.6129705309867859, -0.6595730781555176, 0.5282713174...
null
null
null
null
null
null
null
null
null
null
null
null
null
Niciu/test-cre-dataset-issues
Niciu
2022-03-01T14:06:43Z
152
0
null
[ "region:us" ]
2022-03-01T14:06:43Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
WyrdCurt/AO4W
WyrdCurt
2021-07-26T12:03:27Z
152
0
null
[ "region:us" ]
2021-07-26T12:03:27Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
# Archive Of Our Own Original Works (AO4W) **Warning! Many/most of these files may be NSFW!** Approximately 2GB of text files from Archive of Our Own; specifically, files labeled "original work" or some variation. For training fiction models. I recommend that you clean the text as needed for your purposes.
[ -0.27497172355651855, -0.8118712902069092, 0.7034948468208313, 0.07102394849061966, -0.32298487424850464, -0.4371403753757477, 0.533312201499939, -0.7302860021591187, 0.12044892460107803, 1.4352080821990967, -0.5271919369697571, -0.22597596049308777, -0.7737706899642944, 0.4774087369441986...
null
null
null
null
null
null
null
null
null
null
null
null
null
abidlabs/crowdsourced-speech3
abidlabs
2022-01-21T16:12:06Z
152
0
null
[ "region:us" ]
2022-01-21T16:12:06Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
abwicke/koplo
abwicke
2021-03-18T15:43:39Z
152
0
null
[ "region:us" ]
2021-03-18T15:43:39Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
https://jobs.shrm.org/profile/free-watch-full-raya-and-the-last-dragon-2021/1589725/ https://jobs.shrm.org/profile/full-watch-raya-and-the-last-dragon-2021-hd-online-full-free-123movies/1589732/ https://jobs.shrm.org/profile/123movies-watch-raya-and-the-last-dragon-2021-hd-online-full-free-streaming/1589735/ https://jobs.shrm.org/profile/full-watch-zack-snyders-justice-league-2021-full-free/1591000/ https://jobs.shrm.org/profile/online-watch-zack-snyders-justice-league-2021-123movies-full-version-/1591028/ https://jobs.shrm.org/profile/watch-zack-snyders-justice-league-2021-online-movie-full-version-hd/1591260/ https://jobs.shrm.org/profile/full-watch-zack-snyders-justice-league-2021-hd-online-full-free-123movies/1591268/ https://jobs.shrm.org/profile/watch-zack-snyders-justice-league-2021-full-free/1591274/ https://jobs.shrm.org/profile/watch-zack-snyders-justice-league-2021-full-123movies/1591294/ https://jobs.shrm.org/profile/123movies-watch-zack-snyders-justice-league-online-2021-full-free-hd/1591301/ https://jobs.aapor.org/profile/full-watch-365-days-2020-hd-online-full-free-123movies/1592853/ https://jobs.aapor.org/profile/full-watch-army-of-the-dead-2021-hd-online-full-free-123movies/1592863/ https://jobs.aapor.org/profile/full-watch-barb-and-star-go-to-vista-del-mar-2021-hd-online-full-free-123movies/1592894/ https://jobs.aapor.org/profile/full-watch-billie-eilish-the-worlds-a-little-blurry-2021-hd-online-full-free-123movies/1592902/ https://jobs.aapor.org/profile/full-watch-black-widow-2021-hd-online-full-free-123movies/1592920/ https://jobs.aapor.org/profile/full-watch-bliss-2021-hd-online-full-free-123movies/1592926/ https://jobs.aapor.org/profile/full-watch-borat-subsequent-moviefilm-2020-hd-online-full-free-123movies/1592939/ https://jobs.aapor.org/profile/full-watch-boss-level-2021-hd-online-full-free-123movies/1592952/ https://jobs.aapor.org/profile/full-watch-chaos-walking-2021-hd-online-full-free-123movies/1592958/ https://jobs.aapor.org/profile/full-watch-cherry-2021-hd-online-full-free-123movies/1592984/ https://jobs.aapor.org/profile/full-watch-coming-2-america-2021-hd-online-full-free-123movies/1592999/ https://jobs.aapor.org/profile/full-watch-cosmic-sin-2021-hd-online-full-free-123movies/1593012/ https://jobs.aapor.org/profile/full-watch-crisis-2021-hd-online-full-free-123movies/1593017/ https://jobs.aapor.org/profile/full-watch-demon-slayer-kimetsu-no-yaiba-mugen-train-2020-hd-online-full-free-123movies/1593028/ https://jobs.aapor.org/profile/full-watch-eternals-2021-hd-online-full-free-123movies/1593038/ https://jobs.aapor.org/profile/full-watch-flora-ulysses-2021-hd-online-full-free-123movies/1593050/ https://jobs.aapor.org/profile/full-watch-girl-in-the-basement-2021-hd-online-full-free-123movies/1593056/ https://jobs.aapor.org/profile/full-watch-godzilla-vs-kong-2021-hd-online-full-free-123movies/1593071/ https://jobs.aapor.org/profile/full-watch-greenland-2020-hd-online-full-free-123movies/1593089/ https://jobs.aapor.org/profile/full-watch-i-care-a-lot-2021-hd-online-full-free-123movies/1593100/ https://jobs.aapor.org/profile/full-watch-judas-and-the-black-messiah-2021-hd-online-full-free-123movies/1593112/ https://jobs.aapor.org/profile/full-watch-land-2021-hd-online-full-free-123movies/1593128/ https://jobs.aapor.org/profile/full-watch-to-all-the-boys-always-and-forever-2021-hd-online-full-free-123movies/1592831/ https://jobs.shrm.org/profile/full-watch-zack-snyders-justice-league-2021-hd-online-movie-free-123movies/1592215/ https://jobs.shrm.org/profile/full-watch-coming-2-america-2021-hd-movie-online-free-123movies/1592252/ https://jobs.shrm.org/profile/full-watch-raya-and-the-last-dragon-2021-hd-movie-online-free-123movies/1592256/ https://jobs.shrm.org/profile/full-watch-demon-slayer-kimetsu-no-yaiba-the-movie-mugen-train-2020-hd-movie-online-free-123movies/1592270/ https://jobs.shrm.org/profile/hbo-max-watch-zack-snyders-justice-league-2021-hd-online-version/1592272/ https://jobs.aapor.org/profile/full-episodes-the-falcon-and-the-winter-soldier-season-1-episode-1-e1-online/1592374/ https://jobs.aapor.org/profile/watch-tom-jerry-2021-online-full-movie-free-123movies/1592338/ https://jobs.aapor.org/profile/hbo-max-watch-zack-snyders-justice-league-2021-hd-online-version/1592290/ https://jobs.aapor.org/profile/watch-zack-snyders-justice-league-2021-online-full-movie-hd-free/1592288/ https://jobs.aapor.org/profile/downloadhdzack-snyders-justice-league-2021-1080p-full-online-version/1592297/ https://jobs.aapor.org/profile/watch-coming-2-america-2021-online-full-movie-free-123movies/1592307/ https://jobs.aapor.org/profile/watch-zack-snyders-justice-league-2021-full-free-online-movie-here-hbomax/1592312/ https://jobs.aapor.org/profile/watch-raya-and-the-last-dragon-2021-online-full-movie-free-123movies/1592326/ https://jobs.aapor.org/profile/full-watch-tom-and-jerry-2021-full-movie-online-free-hd-stream/1592328/ https://jobs.aapor.org/profile/watch-billie-eilish-the-worlds-a-little-blurry-2021-full/1592344/ https://jobs.aapor.org/profile/watch-billie-eilish-the-worlds-a-little-blurry-2021-online-full-movie-free-123movies/1592353/ https://jobs.aapor.org/profile/watch-demon-slayer-kimetsu-no-yaiba-the-movie-mugen-train-2020-online-full-movie-free-123movies/1592363/ https://jobs.aapor.org/profile/full-watch-demon-slayer-kimetsu-no-yaiba-mugen-train-2020-full-free/1592355/ https://jobs.aapor.org/profile/watch-the-falcon-and-the-winter-soldier-season-1-e1-2021-online-free/1592389/
[ -0.4497397840023041, -0.047093573957681656, 0.301297664642334, 0.046867504715919495, -0.5837980508804321, 0.09756352007389069, 0.6407689452171326, -0.6415932178497314, 0.44099298119544983, 0.15395814180374146, -0.5891977548599243, -0.055207084864377975, -0.30397966504096985, 0.031818091869...
null
null
null
null
null
null
null
null
null
null
null
null
null
albertvillanova/tmp-tests
albertvillanova
2021-12-02T14:12:12Z
152
0
null
[ "region:us" ]
2021-12-02T14:12:12Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.32276496291160583, -0.22568435966968536, 0.8622260093688965, 0.43461480736732483, -0.5282987952232361, 0.7012965083122253, 0.7915714979171753, 0.07618625462055206, 0.7746025323867798, 0.25632181763648987, -0.7852815389633179, -0.22573819756507874, -0.9104480743408203, 0.5715669393539429...
null
null
null
null
null
null
null
null
null
null
null
null
null
aliabd/crowdsourced-speech4
aliabd
2022-01-21T17:36:51Z
152
0
null
[ "region:us" ]
2022-01-21T17:36:51Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.32276496291160583, -0.22568435966968536, 0.8622260093688965, 0.43461480736732483, -0.5282987952232361, 0.7012965083122253, 0.7915714979171753, 0.07618625462055206, 0.7746025323867798, 0.25632181763648987, -0.7852815389633179, -0.22573819756507874, -0.9104480743408203, 0.5715669393539429...
null
null
null
null
null
null
null
null
null
null
null
null
null
alireza655/alireza655
alireza655
2021-02-08T23:24:50Z
152
0
null
[ "region:us" ]
2021-02-08T23:24:50Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.32276496291160583, -0.22568435966968536, 0.8622260093688965, 0.43461480736732483, -0.5282987952232361, 0.7012965083122253, 0.7915714979171753, 0.07618625462055206, 0.7746025323867798, 0.25632181763648987, -0.7852815389633179, -0.22573819756507874, -0.9104480743408203, 0.5715669393539429...
null
null
null
null
null
null
null
null
null
null
null
null
null
clem/autonlp-data-french_word_detection
clem
2021-09-14T09:45:38Z
152
1
null
[ "region:us" ]
2021-09-14T09:45:38Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
This is a very good dataset!
[ -0.2994808256626129, -0.058056823909282684, 0.21844859421253204, 0.5398688912391663, -0.13290078938007355, -0.1372750997543335, -0.031124968081712723, -0.07468447834253311, 0.49724268913269043, 0.7139886617660522, -0.2566799223423004, -0.7480513453483582, -0.5223762392997742, -0.3120886683...
null
null
null
null
null
null
null
null
null
null
null
null
null
cnrcastroli/aaaa
cnrcastroli
2021-03-04T21:51:21Z
152
0
null
[ "region:us" ]
2021-03-04T21:51:21Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
fwefwefewf
[ -0.25071993470191956, -0.5934706330299377, 0.34227949380874634, 1.0664750337600708, -0.4548211097717285, -0.0020777417812496424, 0.21651989221572876, -0.7492340207099915, 0.5542600154876709, 0.812045693397522, -0.859272301197052, 0.08489752560853958, -1.113316535949707, 0.29075518250465393...
null
null
null
null
null
null
null
null
null
null
null
null
null
coala/kkk
coala
2021-09-14T07:56:22Z
152
0
null
[ "region:us" ]
2021-09-14T07:56:22Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.32276496291160583, -0.22568435966968536, 0.8622260093688965, 0.43461480736732483, -0.5282987952232361, 0.7012965083122253, 0.7915714979171753, 0.07618625462055206, 0.7746025323867798, 0.25632181763648987, -0.7852815389633179, -0.22573819756507874, -0.9104480743408203, 0.5715669393539429...
null
null
null
null
null
null
null
null
null
null
null
null
null
congpt/dstc23_asr
congpt
2021-04-06T18:04:04Z
152
0
null
[ "region:us" ]
2021-04-06T18:04:04Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.32276496291160583, -0.22568435966968536, 0.8622260093688965, 0.43461480736732483, -0.5282987952232361, 0.7012965083122253, 0.7915714979171753, 0.07618625462055206, 0.7746025323867798, 0.25632181763648987, -0.7852815389633179, -0.22573819756507874, -0.9104480743408203, 0.5715669393539429...
null
null
null
null
null
null
null
null
null
null
null
null
null
crich/cider
crich
2021-11-15T18:26:56Z
152
0
null
[ "region:us" ]
2021-11-15T18:26:56Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.32276496291160583, -0.22568435966968536, 0.8622260093688965, 0.43461480736732483, -0.5282987952232361, 0.7012965083122253, 0.7915714979171753, 0.07618625462055206, 0.7746025323867798, 0.25632181763648987, -0.7852815389633179, -0.22573819756507874, -0.9104480743408203, 0.5715669393539429...
null
null
null
null
null
null
null
null
null
null
null
null
null
csarron/image-captions
csarron
2021-11-29T04:31:34Z
152
0
null
[ "region:us" ]
2021-11-29T04:31:34Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
ctu-aic/multinli_cs
ctu-aic
2021-11-21T19:58:09Z
152
0
null
[ "region:us" ]
2021-11-21T19:58:09Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
dasago78/dasago78dataset
dasago78
2021-04-02T17:57:24Z
152
0
null
[ "region:us" ]
2021-04-02T17:57:24Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
davanstrien/hipe2020
davanstrien
2022-02-15T11:40:24Z
152
0
null
[ "region:us" ]
2022-02-15T11:40:24Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
davanstrien/iiif_labeled
davanstrien
2022-02-28T11:06:07Z
152
0
null
[ "region:us" ]
2022-02-28T11:06:07Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622264862060547, 0.43461528420448303, -0.52829909324646, 0.7012971639633179, 0.7915720343589783, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104477167129517, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
dongpil/test
dongpil
2021-07-29T10:34:34Z
152
0
null
[ "region:us" ]
2021-07-29T10:34:34Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622264862060547, 0.43461528420448303, -0.52829909324646, 0.7012971639633179, 0.7915720343589783, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104477167129517, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
eason929/test
eason929
2021-03-15T04:02:59Z
152
0
null
[ "region:us" ]
2021-03-15T04:02:59Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622264862060547, 0.43461528420448303, -0.52829909324646, 0.7012971639633179, 0.7915720343589783, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104477167129517, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
ehcalabres/ravdess_speech
ehcalabres
2022-10-24T15:51:41Z
152
3
null
[ "task_categories:audio-classification", "annotations_creators:no-annotation", "language_creators:found", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "license:cc-by-nc-sa-4.0", "region:us" ]
2022-10-24T15:51:41Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
--- annotations_creators: - no-annotation language_creators: - found language: - en license: - cc-by-nc-sa-4.0 multilinguality: - monolingual size_categories: - 1K<n<10K source_datasets: - original task_categories: - audio-classification task_ids: - speech-emotion-recognition --- # Dataset Card for ravdess_speech ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-instances) - [Data Splits](#data-instances) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** [Needs More Information] - **Repository:** https://zenodo.org/record/1188976#.YUS4MrozZdS - **Paper:** https://doi.org/10.1371/journal.pone.0196391 - **Leaderboard:** [Needs More Information] - **Point of Contact:** ravdess@gmail.com ### Dataset Summary The Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS) contains 24 professional actors (12 female, 12 male), vocalizing two lexically-matched statements in a neutral North American accent. Speech includes calm, happy, sad, angry, fearful, surprise, and disgust expressions. Each expression is produced at two levels of emotional intensity (normal, strong), with an additional neutral expression. The conditions of the audio files are: 16bit, 48kHz .wav. ### Supported Tasks and Leaderboards - audio-classification: The dataset can be used to train a model for Audio Classification tasks, which consists in predict the latent emotion presented on the audios. ### Languages The audios available in the dataset are in English spoken by actors in a neutral North American accent. ## Dataset Structure ### Data Instances [Needs More Information] ### Data Fields [Needs More Information] ### Data Splits [Needs More Information] ## Dataset Creation ### Curation Rationale [Needs More Information] ### Source Data #### Initial Data Collection and Normalization [Needs More Information] #### Who are the source language producers? [Needs More Information] ### Annotations #### Annotation process [Needs More Information] #### Who are the annotators? [Needs More Information] ### Personal and Sensitive Information [Needs More Information] ## Considerations for Using the Data ### Social Impact of Dataset [Needs More Information] ### Discussion of Biases [Needs More Information] ### Other Known Limitations [Needs More Information] ## Additional Information ### Dataset Curators [Needs More Information] ### Licensing Information The RAVDESS is released under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License, CC BY-NC-SA 4.0 Commercial licenses for the RAVDESS can also be purchased. For more information, please visit our license fee page, or contact us at ravdess@gmail.com. ### Citation Information Livingstone SR, Russo FA (2018) The Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS): A dynamic, multimodal set of facial and vocal expressions in North American English. PLoS ONE 13(5): e0196391. https://doi.org/10.1371/journal.pone.0196391.
[ -0.43191179633140564, -0.7637857794761658, -0.1252378225326538, 0.4789009094238281, -0.06024398282170296, 0.16646696627140045, -0.4492100775241852, -0.5712066292762756, 0.5647458434104919, 0.5642333030700684, -0.9924535155296326, -1.0414401292800903, -0.6020432710647583, -0.045269902795553...
null
null
null
null
null
null
null
null
null
null
null
null
null
Norod78/simpsons-blip-captions
Norod78
2022-11-09T16:27:19Z
152
3
null
[ "task_categories:text-to-image", "annotations_creators:machine-generated", "language_creators:other", "multilinguality:monolingual", "size_categories:n<1K", "language:en", "license:cc-by-nc-sa-4.0", "region:us" ]
2022-11-09T16:27:19Z
2022-11-06T11:11:36.000Z
2022-11-06T11:11:36
--- dataset_info: features: - name: image dtype: image - name: text dtype: string splits: - name: train num_bytes: 51605730.0 num_examples: 755 download_size: 50553165 dataset_size: 51605730.0 pretty_name: 'Simpsons BLIP captions' size_categories: - n<1K tags: [] task_categories: - text-to-image license: cc-by-nc-sa-4.0 annotations_creators: - machine-generated language: - en language_creators: - other multilinguality: - monolingual --- # Dataset Card for "simpsons-blip-captions"
[ -0.2915683090686798, -0.0044026849791407585, -0.2134803980588913, 0.5665171146392822, -0.6229936480522156, 0.39689597487449646, -0.2998969852924347, 0.36532121896743774, 0.4522024393081665, 0.6106967329978943, -0.5644146203994751, -0.5934778451919556, -0.5761433243751526, 0.114859238266944...
null
null
null
null
null
null
null
null
null
null
null
null
null
Jzuluaga/uwb_atcc
Jzuluaga
2022-12-05T11:15:20Z
152
0
null
[ "task_categories:automatic-speech-recognition", "multilinguality:monolingual", "language:en", "license:cc-by-nc-sa-4.0", "audio", "automatic-speech-recognition", "en-atc", "en", "noisy-speech-recognition", "speech-recognition", "arxiv:2203.16822", "region:us" ]
2022-12-05T11:15:20Z
2022-11-28T07:12:02.000Z
2022-11-28T07:12:02
--- dataset_info: features: - name: id dtype: string - name: audio dtype: audio: sampling_rate: 16000 - name: text dtype: string - name: segment_start_time dtype: float32 - name: segment_end_time dtype: float32 - name: duration dtype: float32 splits: - name: test num_bytes: 140620332.25 num_examples: 2822 - name: train num_bytes: 608597323.625 num_examples: 11291 download_size: 711464914 dataset_size: 749217655.875 tags: - audio - automatic-speech-recognition - en-atc - en - noisy-speech-recognition - speech-recognition task_categories: - automatic-speech-recognition language: - en multilinguality: - monolingual license: - cc-by-nc-sa-4.0 --- # Dataset Card for UWB-ATCC corpus ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages and Other Details](#languages-and-other-details) - [Dataset Structure](#dataset-structure) - [Data Fields](#data-fields) - [Additional Information](#additional-information) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** [UWB-ATCC corpus homepage](https://lindat.mff.cuni.cz/repository/xmlui/handle/11858/00-097C-0000-0001-CCA1-0) - **Repository:** [GitHub repository (used in research)](https://github.com/idiap/w2v2-air-traffic) - **Paper:** [Air traffic control communication (ATCC) speech corpora and their use for ASR and TTS development](https://link.springer.com/article/10.1007/s10579-019-09449-5) - **Paper of this research:** [How Does Pre-trained Wav2Vec 2.0 Perform on Domain Shifted ASR? An Extensive Benchmark on Air Traffic Control Communications](https://arxiv.org/abs/2203.16822) ### Dataset Summary The UWB-ATCC Corpus is provided provided by University of West Bohemia, Department of Cybernetics. The corpus contains recordings of communication between air traffic controllers and pilots. The speech is manually transcribed and labeled with the information about the speaker (pilot/controller, not the full identity of the person). The corpus is currently small (20 hours) but we plan to search for additional data next year. The audio data format is: 8kHz, 16bit PCM, mono. Important, from the `<id (string)>` field, you can obtain the speaker roles. For instance: - `_PI`: segment with only pilot speech - `_AT`: segment with only ATCO speech - `PIAT`: segment with both, ATCO and pilot speech ### Supported Tasks and Leaderboards - `automatic-speech-recognition`. Already adapted/fine-tuned models are available here --> [XLS-R-300m](https://huggingface.co/Jzuluaga/wav2vec2-large-960h-lv60-self-en-atc-atcosim). ### Languages and other details The text and the recordings are in English. The authors took advantage of the fact that one of their industrial partners develops complex IT solutions for several ATC authorities and airports and, as such, has access to the ATC communication recordings collected in the Czech airspace. This partner was able to secure the following data: - Ground control—communication before takeoff and after landing—19.2 h of data. - Tower control—communication during takeoff, landing and landing standby—22.5 h. - Approach control—communication during landing approach—25.5 h. - Area control—communication during overflights and cruises—71.3 h. (Not all data is released. Check their website [here](https://lindat.mff.cuni.cz/repository/xmlui/handle/11858/00-097C-0000-0001-CCA1-0)) ## Dataset Structure ### Data Fields - `id (string)`: a string of recording identifier for each example, corresponding to its. - `audio (audio)`: audio data for the given ID - `text (string)`: transcript of the file already normalized. Follow these repositories for more details [w2v2-air-traffic](https://github.com/idiap/w2v2-air-traffic) and [bert-text-diarization-atc](https://github.com/idiap/bert-text-diarization-atc) - `segment_start_time (float32)`: segment start time (normally 0) - `segment_end_time (float32): segment end time - `duration (float32)`: duration of the recording, compute as segment_end_time - segment_start_time ## Additional Information ### Licensing Information The licensing status of the dataset hinges on the legal status of the [UWB-ATCC corpus](https://lindat.mff.cuni.cz/repository/xmlui/handle/11858/00-097C-0000-0001-CCA1-0) creators. They used [Creative Commons - Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0)](https://creativecommons.org/licenses/by-nc-sa/4.0/) licensing. ### Citation Information Contributors who prepared, processed, normalized and uploaded the dataset in HuggingFace: ``` @article{zuluaga2022how, title={How Does Pre-trained Wav2Vec2. 0 Perform on Domain Shifted ASR? An Extensive Benchmark on Air Traffic Control Communications}, author={Zuluaga-Gomez, Juan and Prasad, Amrutha and Nigmatulina, Iuliia and Sarfjoo, Saeed and others}, journal={IEEE Spoken Language Technology Workshop (SLT), Doha, Qatar}, year={2022} } @article{zuluaga2022bertraffic, title={BERTraffic: BERT-based Joint Speaker Role and Speaker Change Detection for Air Traffic Control Communications}, author={Zuluaga-Gomez, Juan and Sarfjoo, Seyyed Saeed and Prasad, Amrutha and others}, journal={IEEE Spoken Language Technology Workshop (SLT), Doha, Qatar}, year={2022} } @article{zuluaga2022atco2, title={ATCO2 corpus: A Large-Scale Dataset for Research on Automatic Speech Recognition and Natural Language Understanding of Air Traffic Control Communications}, author={Zuluaga-Gomez, Juan and Vesel{\`y}, Karel and Sz{\"o}ke, Igor and Motlicek, Petr and others}, journal={arXiv preprint arXiv:2211.04054}, year={2022} } ``` Authors of the dataset: ``` @article{vsmidl2019air, title={Air traffic control communication (ATCC) speech corpora and their use for ASR and TTS development}, author={{\v{S}}m{\'\i}dl, Lubo{\v{s}} and {\v{S}}vec, Jan and Tihelka, Daniel and Matou{\v{s}}ek, Jind{\v{r}}ich and Romportl, Jan and Ircing, Pavel}, journal={Language Resources and Evaluation}, volume={53}, number={3}, pages={449--464}, year={2019}, publisher={Springer} } ```
[ -0.39164409041404724, -0.5750758051872253, 0.05795922875404358, 0.18492422997951508, -0.40153899788856506, 0.200161412358284, -0.5382319688796997, -0.6016563773155212, 0.12122086435556412, 0.39749327301979065, -0.45469599962234497, -0.7072417140007019, -0.5663379430770874, -0.1574173867702...
null
null
null
null
null
null
null
null
null
null
null
null
null
12ml/e-CARE
12ml
2023-01-06T18:50:03Z
152
1
null
[ "task_categories:multiple-choice", "region:us" ]
2023-01-06T18:50:03Z
2022-12-21T11:38:01.000Z
2022-12-21T11:38:01
--- task_categories: - multiple-choice --- # Dataset of (Du et al., 2022) ## Abstract >Understanding causality has vital importance for various Natural Language Processing (NLP) applications. Beyond the labeled instances, conceptual explanations of the causality can provide deep understanding of the causal fact to facilitate the causal reasoning process. However, such explanation information still remains absent in existing causal reasoning resources. In this paper, we fill this gap by presenting a human-annotated explainable CAusal REasoning dataset (e-CARE), which contains over 20K causal reasoning questions, together with natural language formed explanations of the causal questions. Experimental results show that generating valid explanations for causal facts still remains especially challenging for the state-of-the-art models, and the explanation information can be helpful for promoting the accuracy and stability of causal reasoning models. ## Notes Please note that the original dataset has been modified so that the variable names match with those in the COPA dataset (Roemmele et al., 2011). In addition, only the training and the development sets are [publicly available](https://github.com/waste-wood/e-care). ## References Du, L., Ding, X., Xiong, K., Liu, T., & Qin, B. (2022). e-CARE: a New Dataset for Exploring Explainable Causal Reasoning. arXiv preprint arXiv:2205.05849. Roemmele, M., Bejan, C., and Gordon, A. (2011) Choice of Plausible Alternatives: An Evaluation of Commonsense Causal Reasoning. AAAI Spring Symposium on Logical Formalizations of Commonsense Reasoning, Stanford University, March 21-23, 2011.
[ -0.1367652863264084, -0.7371608018875122, 0.6847586631774902, 0.03632257506251335, -0.2050774097442627, -0.7123168706893921, -0.006128290202468634, -0.5719534754753113, 0.041625797748565674, 0.2689487934112549, -0.9064939618110657, -0.4440196752548218, -0.4125695526599884, 0.39446032047271...
null
null
null
null
null
null
null
null
null
null
null
null
null
heegyu/news-category-balanced-top10
heegyu
2023-02-13T02:56:31Z
152
1
null
[ "license:cc-by-4.0", "region:us" ]
2023-02-13T02:56:31Z
2023-02-13T02:45:28.000Z
2023-02-13T02:45:28
--- license: cc-by-4.0 --- ### Top10 sampled news category dataset randomly sampled news data original dataset: https://www.kaggle.com/datasets/rmisra/news-category-dataset ### Value Counts per Category ``` ENTERTAINMENT 10000 POLITICS 10000 WELLNESS 10000 TRAVEL 9900 STYLE & BEAUTY 9814 PARENTING 8791 HEALTHY LIVING 6694 QUEER VOICES 6347 FOOD & DRINK 6340 BUSINESS 5992 ```
[ -0.20592691004276276, -0.4814252257347107, 0.023958196863532066, 0.34358859062194824, -0.4472077488899231, 0.2468661218881607, 0.14254577457904816, 0.1494312286376953, 0.9594330191612244, 0.4604412019252777, -0.6774774789810181, -0.7773218750953674, -0.8453078269958496, 0.01241811644285917...
null
null
null
null
null
null
null
null
null
null
null
null
null
FastJobs/Visual_Emotional_Analysis
FastJobs
2023-03-13T06:31:17Z
152
8
null
[ "task_categories:image-classification", "size_categories:10K<n<100K", "language:en", "region:us" ]
2023-03-13T06:31:17Z
2023-03-03T06:23:19.000Z
2023-03-03T06:23:19
--- task_categories: - image-classification language: - en size_categories: - 10K<n<100K ---
[ -0.1285335123538971, -0.1861683875322342, 0.6529128551483154, 0.49436232447624207, -0.19319400191307068, 0.23607441782951355, 0.36072009801864624, 0.05056373029947281, 0.5793656706809998, 0.7400146722793579, -0.650810182094574, -0.23784008622169495, -0.7102247476577759, -0.0478255338966846...
null
null
null
null
null
null
null
null
null
null
null
null
null
llm-book/aio-retriever
llm-book
2023-10-25T15:31:08Z
152
0
null
[ "size_categories:10K<n<100K", "language:ja", "region:us" ]
2023-10-25T15:31:08Z
2023-07-04T04:53:47.000Z
2023-07-04T04:53:47
--- language: - ja size_categories: - 10K<n<100K dataset_info: features: - name: qid dtype: string - name: competition dtype: string - name: timestamp dtype: string - name: section dtype: string - name: number dtype: string - name: original_question dtype: string - name: original_answer dtype: string - name: original_additional_info dtype: string - name: question dtype: string - name: answers list: string - name: passages list: - name: passage_id dtype: int32 - name: title dtype: string - name: text dtype: string - name: positive_passage_indices list: int32 - name: negative_passage_indices list: int32 splits: - name: train num_bytes: 1742881639 num_examples: 22335 - name: validation num_bytes: 78671502 num_examples: 1000 download_size: 665253451 dataset_size: 1821553141 --- # Dataset Card for llm-book/aio-retriever 書籍『大規模言語モデル入門』で使用する、「AI王」コンペティションのQAデータセット(文書検索モデル訓練用)です。 GitHub リポジトリ [cl-tohoku/quiz-datasets](https://github.com/cl-tohoku/quiz-datasets) で公開されているデータセットを利用しています。 ## Licence 本データセットに含まれる一部のクイズ問題の著作権は [abc/EQIDEN 実行委員会](https://abc-dive.com/portal/)に帰属するものであり、これらのクイズ問題は本書における使用許諾を得ているものです。 本データセットに含まれる一部のクイズ問題は[株式会社キュービック](http://www.qbik.co.jp/)および[株式会社カプリティオ](https://capriccio.tokyo/)に依頼し作成したものであり、これらのクイズ問題は[クリエイティブ・コモンズ表示・継承ライセンス 4.0 (CC BY-SA 4.0)](https://creativecommons.org/licenses/by-sa/4.0/deed.ja) ライセンスの下に提供されています。 本データセットにパッセージとして付与されている Wikipedia のコンテンツは、[クリエイティブ・コモンズ表示・継承ライセンス 3.0 (CC BY-SA 3.0)](https://creativecommons.org/licenses/by-sa/3.0/deed.ja) および [GNU 自由文書ライセンス (GFDL)](https://www.gnu.org/licenses/fdl.html) の下に配布されているものです。 クイズ問題のライセンスについて、詳細は [cl-tohoku/quiz-datasets](https://github.com/cl-tohoku/quiz-datasets) を参照してください。
[ -0.4279440641403198, -0.6333563327789307, 0.3278101980686188, -0.08201396465301514, -0.6405560374259949, -0.10540910065174103, -0.12112569063901901, -0.25877928733825684, 0.28232577443122864, 0.5071145296096802, -0.7687205076217651, -1.06296706199646, -0.3952067792415619, 0.136961743235588...
null
null
null
null
null
null
null
null
null
null
null
null
null
TrainingDataPro/email-spam-classification
TrainingDataPro
2023-09-14T16:37:38Z
152
1
null
[ "task_categories:text-classification", "language:en", "license:cc-by-nc-nd-4.0", "finance", "code", "region:us" ]
2023-09-14T16:37:38Z
2023-07-25T12:09:29.000Z
2023-07-25T12:09:29
--- license: cc-by-nc-nd-4.0 task_categories: - text-classification language: - en tags: - finance - code --- # Email Spam Classification The dataset consists of a collection of emails categorized into two major classes: **spam** and **not spam**. It is designed to facilitate the development and evaluation of spam detection or email filtering systems. **The spam emails** in the dataset are typically unsolicited and unwanted messages that aim to promote products or services, spread malware, or deceive recipients for various malicious purposes. These emails often contain misleading subject lines, excessive use of advertisements, unauthorized links, or attempts to collect personal information. The **non-spam emails** in the dataset are genuine and legitimate messages sent by individuals or organizations. They may include personal or professional communication, newsletters, transaction receipts, or any other non-malicious content. The dataset encompasses emails of varying *lengths, languages, and writing styles*, reflecting the inherent heterogeneity of email communication. This diversity aids in training algorithms that can generalize well to different types of emails, making them robust against different spammer tactics and variations in non-spam email content. ### The dataset's possible applications: - spam detection - fraud detection - email filtering systems - customer support automation - natural language processing ![](https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F618942%2F4d1fdedb2827152696dd0c0af05fd8da%2Ff.png?generation=1690286497115141&alt=media) # Get the dataset ### This is just an example of the data Leave a request on [**https://trainingdata.pro/data-market**](https://trainingdata.pro/data-market?utm_source=huggingface&utm_medium=cpc&utm_campaign=email-spam-classification) to discuss your requirements, learn about the price and buy the dataset. # File with the extension .csv includes the following information: - **title**: title of the email, - **text**: text of the email, - **type**: type of the email # Email spam might be collected in accordance with your requirements. ## [**TrainingData**](https://trainingdata.pro/data-market?utm_source=huggingface&utm_medium=cpc&utm_campaign=email-spam-classification) provides high-quality data annotation tailored to your needs More datasets in TrainingData's Kaggle account: **https://www.kaggle.com/trainingdatapro/datasets** TrainingData's GitHub: **https://github.com/Trainingdata-datamarket/TrainingData_All_datasets**
[ -0.28808820247650146, -0.7617974877357483, -0.26661285758018494, 0.26028385758399963, 0.027178220450878143, 0.17718401551246643, -0.18194563686847687, -0.34256088733673096, 0.146999791264534, 0.9328370690345764, -0.5162990689277649, -0.8438812494277954, -0.9306033253669739, 0.0051321787759...
null
null
null
null
null
null
null
null
null
null
null
null
null
jitx/Methods2Test_java_unit_test_code
jitx
2023-08-30T19:31:25Z
152
3
null
[ "task_categories:text-generation", "language:en", "license:mit", "unit test", "java", "code", "arxiv:2203.12776", "region:us" ]
2023-08-30T19:31:25Z
2023-08-30T18:59:03.000Z
2023-08-30T18:59:03
--- license: mit configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* dataset_info: features: - name: target dtype: string - name: src_fm dtype: string - name: src_fm_fc dtype: string - name: src_fm_fc_co dtype: string - name: src_fm_fc_ms dtype: string - name: src_fm_fc_ms_ff dtype: string splits: - name: train num_bytes: 3399525755 num_examples: 624022 - name: test num_bytes: 907751466 num_examples: 156922 download_size: 558984469 dataset_size: 4307277221 task_categories: - text-generation language: - en tags: - unit test - java - code --- ## Dataset Description Microsoft created this large dataset of Java Junit test cases with its corresponding focal methods. It contains 780k pairs of JUnit test cases and focal methods which were extracted from a total of 91K Java open source project hosted on GitHub. The mapping between test case and focal methods are based heuristics rules and Java developer's best practice. More information could be found here: - [methods2test Github repo](https://github.com/microsoft/methods2test) - [Methods2Test: A dataset of focal methods mapped to test cases](https://arxiv.org/pdf/2203.12776.pdf) ## Dataset Schema ``` target: <TEST_CASE> src_fm: <FOCAL_METHOD> src_fm_fc: <FOCAL_CLASS_NAME> <FOCAL_METHOD> src_fm_fc_co: <FOCAL_CLASS_NAME> <FOCAL_METHOD> <CONTRSUCTORS> src_fm_fc_ms: <FOCAL_CLASS_NAME> <FOCAL_METHOD> <CONTRSUCTORS> <METHOD_SIGNATURES> src_fm_fc_ms_ff: <FOCAL_CLASS_NAME> <FOCAL_METHOD> <CONTRSUCTORS> <METHOD_SIGNATURES> <FIELDS> ``` ## Focal Context - fm: this representation incorporates exclusively the source code of the focal method. Intuitively, this contains the most important information for generating accurate test cases for the given method. - fm+fc: this representations adds the focal class name, which can provide meaningful semantic information to the model. - fm+fc+c: this representation adds the signatures of the constructor methods of the focal class. The idea behind this augmentation is that the test case may require instantiating an object of the focal class in order to properly test the focal method. - fm+fc+c+m: this representation adds the signatures of the other public methods in the focal class. The rationale which motivated this inclusion is that the test case may need to invoke other auxiliary methods within the class (e.g., getters, setters) to set up or tear down the testing environment. - fm+fc+c+m+f : this representation adds the public fields of the focal class. The motivation is that test cases may need to inspect the status of the public fields to properly test a focal method. ![image/png](https://cdn-uploads.huggingface.co/production/uploads/642382bb6e61cda1b3a20983/PYpNYXweRZyFOp6TbNkyt.png) The different levels of focal contexts are the following: ``` FM: focal method FM_FC: focal method + focal class name FM_FC_CO: focal method + focal class name + constructor signatures FM_FC_MS: focal method + focal class name + constructor signatures + public method signatures FM_FC_MS_FF: focal method + focal class name + constructor signatures + public method signatures + public fields ``` ## Lmitations The original authors validate the heuristics by inspecting a statistically significant sample (confidence level of 95% within 10% margin of error) of 97 samples from the training set. Two authors independently evaluated the sample, then met to discuss the disagreements. We found that 90.72% of the samples have a correct link between the test case and the corresponding focal method ## Contribution All the thanks to the original authors.
[ -0.7303378582000732, -0.4665660858154297, 0.3914388418197632, 0.4036341607570648, -0.07431723177433014, -0.23491407930850983, -0.13066262006759644, -0.6311100721359253, -0.3108639419078827, 0.14804449677467346, -0.48852038383483887, -0.41052672266960144, -0.4552026391029358, -0.02452009171...
null
null
null
null
null
null
null
null
null
null
null
null
null
taishi-i/awesome-japanese-nlp-classification-dataset
taishi-i
2023-09-09T11:09:04Z
152
3
null
[ "task_categories:text-classification", "size_categories:1K<n<10K", "language:en", "language:ja", "license:other", "code", "region:us" ]
2023-09-09T11:09:04Z
2023-09-09T06:37:36.000Z
2023-09-09T06:37:36
--- license: other task_categories: - text-classification language: - en - ja tags: - code size_categories: - 1K<n<10K --- # Dataset overview This dataset identifies whether a GitHub repository description pertains to Japanese natural language processing (NLP). The labels are categorized as **"Relevant (1)" and "Not Relevant (0)"**. Problem Setting: - Training Data: Repository descriptions from before 2022 - Test Data: Repository descriptions from 2023 - Objective: To detect repositories related to Japanese NLP Data Collection: - Positive Examples: Repositories listed in "[awesome-japanese-nlp-resources](https://github.com/taishi-i/awesome-japanese-nlp-resources)" as of September 9, 2023 - Negative Examples: Collected from the GitHub API and visually confirmed - Note: The annotation process is subjective Dataset Features: - Subjective labeling - Mixed English and Japanese descriptions - Imbalanced label distribution **These dataset features mirror real-world challenges and are ideal for evaluating models.** Based on GitHub's terms of service, please use this dataset for research purposes only. # How to use this dataset How to load in Python. ```python from datasets import load_dataset dataset = load_dataset("taishi-i/awesome-japanese-nlp-classification-dataset") ``` Details of the dataset. ```python DatasetDict({ train: Dataset({ features: ['label', 'text', 'url', 'created_at'], num_rows: 5496 }) validation: Dataset({ features: ['label', 'text', 'url', 'created_at'], num_rows: 400 }) test: Dataset({ features: ['label', 'text', 'url', 'created_at'], num_rows: 856 }) }) ``` # Baseline Baseline trained with [bert-base-multilingual-cased](https://huggingface.co/bert-base-multilingual-cased). Please use the baseline model from [here](https://huggingface.co/taishi-i/awesome-japanese-nlp-classification-model). The F1-score for label 1 is important for this task. | Label | Precision | Recall | F1-Score | Support | |--------------|-----------|--------|----------|---------| | 0 | 0.98 | 0.99 | 0.98 | 796 | | 1 | 0.79 | 0.70 | **0.74** | 60 | | Accuracy | | | 0.97 | 856 | | Macro Avg | 0.89 | 0.84 | 0.86 | 856 | | Weighted Avg | 0.96 | 0.97 | 0.97 | 856 | # Dataset stats Label distribution: | Dataset | Label 0 (%) | Label 1 (%) | |------------|-------------|-------------| | Train | 92.59 | 7.41 | | Validation | 95.75 | 4.25 | | Test | 92.99 | 7.01 | Relevant sample: ```python { "label": 1, "text": "JGLUE: Japanese General Language Understanding Evaluation for huggingface datasets", "url": "https://github.com/shunk031/huggingface-datasets_JGLUE", "created_at": "2023-02-25T04:33:03Z" } ``` Not Relevant sample: ```python { "label": 0, "text": "Official repository of FaceLit: Neural 3D Relightable Faces (CVPR 2023)", "url": "https://github.com/apple/ml-facelit", "created_at": "2023-04-03T22:47:29Z" } ``` Number of texts, average number of characters per text, minimum number of characters, maximum number of characters: | Dataset | Text Count | Average Length | Min Length | Max Length | |------------|------------|----------------|------------|------------| | Train | 5496 | 58.05 | 2.0 | 609.0 | | Validation | 400 | 54.33 | 8.0 | 226.0 | | Test | 856 | 58.85 | 3.0 | 341.0 | Proportion of text languages: | Dataset | English (%) | Japanese (%) | |------------|-------------|--------------| | Train | 89.34 | 10.66 | | Validation | 82.00 | 18.00 | | Test | 83.18 | 16.82 | Time range: | Dataset | Start Date | End Date | |---------|---------------------------|---------------------------| | Train | 2008-02-11 22:55:26+00:00 | 2022-09-30 19:45:09+00:00 | | Validation | 2022-10-01 06:02:56+00:00 | 2022-12-31 12:12:41+00:00 | | Test | 2023-01-01 06:15:03+00:00 | 2023-08-21 15:30:53+00:00 | # License We collect and publish this dataset under [GitHub Acceptable Use Policies - 7. Information Usage Restrictions](https://docs.github.com/en/site-policy/acceptable-use-policies/github-acceptable-use-policies#7-information-usage-restrictions) and [GitHub Terms of Service - H. API Terms](https://docs.github.com/en/site-policy/github-terms/github-terms-of-service#h-api-terms) for research purposes. This dataset should be used solely for research verification purposes. Adhering to GitHub's regulations is mandatory.
[ -0.517892062664032, -0.6442219614982605, 0.2054019421339035, 0.41316306591033936, -0.18058112263679504, -0.0626220852136612, -0.2686268091201782, -0.4213794469833374, 0.5264596343040466, 0.4353933036327362, -0.6529170870780945, -0.9805794358253479, -0.4928361773490906, 0.24465306103229523,...
null
null
null
null
null
null
null
null
null
null
null
null
null
dwadden/covidfact_entailment
dwadden
2023-10-31T00:33:56Z
152
0
null
[ "task_categories:text-classification", "task_ids:fact-checking", "annotations_creators:expert-generated", "language_creators:found", "multilinguality:monolingual", "size_categories:1K<n<10K", "source_datasets:original", "language:en", "license:cc-by-nc-2.0", "region:us" ]
2023-10-31T00:33:56Z
2023-10-30T22:26:59.000Z
2023-10-30T22:26:59
--- annotations_creators: - expert-generated language_creators: - found language: - en license: - cc-by-nc-2.0 multilinguality: - monolingual size_categories: - 1K<n<10K source_datasets: - original task_categories: - text-classification task_ids: - fact-checking pretty_name: CovidFact dataset_info: features: - name: claim_id dtype: int32 - name: claim dtype: string - name: abstract_id dtype: int32 - name: title dtype: string - name: abstract sequence: string - name: verdict dtype: string - name: evidence sequence: int32 splits: - name: train num_bytes: 1547185 num_examples: 940 - name: test num_bytes: 523542 num_examples: 317 download_size: 3610222 dataset_size: 2070727 --- # Dataset Card for "covidfact_entailment" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Dataset Structure](#dataset-structure) - [Data Fields](#data-fields) ## Dataset Description - **Repository:** <https://github.com/asaakyan/covidfact> - **Point of Contact:** [David Wadden](mailto:davidw@allenai.org) ### Dataset Summary COVID-FACT is a dataset of claims about COVID-19. For this version of the dataset, we follow the preprocessing from the MultiVerS modeling paper https://github.com/dwadden/multivers, verifying claims against abstracts of scientific research articles. Entailment labels and rationales are included. ## Dataset Structure ### Data fields - `claim_id`: An `int32` claim identifier. - `claim`: A `string`. - `abstract_id`: An `int32` abstract identifier. - `title`: A `string`. - `abstract`: A list of `strings`, one for each sentence in the abstract. - `verdict`: The fact-checking verdict, a `string`. - `evidence`: A list of sentences from the abstract which provide evidence for the verdict.
[ -0.5013104677200317, -0.6455148458480835, 0.11752144247293472, 0.10753468424081802, -0.30851179361343384, -0.13562484085559845, 0.017240028828382492, -0.38793110847473145, 0.27450618147850037, 0.31296366453170776, -0.6105786561965942, -0.7239356637001038, -0.539823055267334, 0.242049962282...
null
null
null
null
null
null
null
null
null
null
null
null
null
magnifi/hl-codellama-chat-response-v2
magnifi
2023-11-02T16:47:51Z
152
0
null
[ "region:us" ]
2023-11-02T16:47:51Z
2023-11-02T16:47:40.000Z
2023-11-02T16:47:40
--- configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* dataset_info: features: - name: Query dtype: string - name: Result dtype: string - name: chat_response dtype: string splits: - name: train num_bytes: 1321860.461185117 num_examples: 1523 - name: test num_bytes: 567627.5388148829 num_examples: 654 download_size: 109799 dataset_size: 1889488.0 --- # Dataset Card for "hl-codellama-chat-response-v2" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.3818666338920593, -0.5209525227546692, -0.08272186666727066, 0.4185601472854614, -0.0809941440820694, 0.2079491764307022, 0.10534871369600296, -0.3145521581172943, 0.8920869827270508, 0.5515937805175781, -0.7920475006103516, -0.6026279926300049, -0.490500271320343, -0.5201120376586914, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
Champion/vpc2020_clear_anon_speech
Champion
2021-10-12T14:19:45Z
151
0
null
[ "region:us" ]
2021-10-12T14:19:45Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Repo to share original and anonymized speech of vpc2020
[ -0.306818425655365, 0.006512943189591169, 0.02469812147319317, 0.5365615487098694, -0.6537101864814758, 0.38406550884246826, -0.36968371272087097, -0.5105584263801575, 0.399859756231308, 1.5771539211273193, -0.46451425552368164, -0.5341846942901611, -0.5968355536460876, 0.1936497539281845,...
null
null
null
null
null
null
null
null
null
null
null
null
null
Lenn/github-issues
Lenn
2021-11-15T10:19:39Z
151
0
null
[ "region:us" ]
2021-11-15T10:19:39Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
cristinakuo/latino40
cristinakuo
2021-12-27T19:24:12Z
151
0
null
[ "region:us" ]
2021-12-27T19:24:12Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
ctgowrie/chessgames
ctgowrie
2021-12-05T00:43:39Z
151
0
null
[ "region:us" ]
2021-12-05T00:43:39Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
cyko/books
cyko
2021-11-27T12:09:21Z
151
0
null
[ "region:us" ]
2021-11-27T12:09:21Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
davanstrien/ads-test
davanstrien
2022-01-18T12:27:37Z
151
0
null
[ "region:us" ]
2022-01-18T12:27:37Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
david-wb/zeshel
david-wb
2021-02-16T23:32:15Z
151
0
null
[ "region:us" ]
2021-02-16T23:32:15Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
debajyotidatta/biosses
debajyotidatta
2022-02-01T01:46:29Z
151
0
null
[ "license:gpl-3.0", "region:us" ]
2022-02-01T01:46:29Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
--- license: gpl-3.0 ---
[ -0.12853367626667023, -0.18616794049739838, 0.6529126763343811, 0.4943627417087555, -0.19319313764572144, 0.23607443273067474, 0.36071979999542236, 0.05056338757276535, 0.5793654322624207, 0.7400138974189758, -0.6508103013038635, -0.23783987760543823, -0.710224986076355, -0.047825977206230...
null
null
null
null
null
null
null
null
null
null
null
null
null
dev/untitled_imgs
dev
2021-12-11T14:14:27Z
151
0
null
[ "region:us" ]
2021-12-11T14:14:27Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
diiogo/annotations
diiogo
2023-10-27T12:16:36Z
151
0
null
[ "region:us" ]
2023-10-27T12:16:36Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
dispenst/jhghdghfd
dispenst
2021-03-28T15:24:20Z
151
0
null
[ "region:us" ]
2021-03-28T15:24:20Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
<a href="https://jobs.acm.org/jobs/watch-godzilla-vs-kong-2021-full-1818658-cd">.</a> <a href="https://jobs.acm.org/jobs/123movies-watch-godzilla-vs-kong-online-2021-full-f-r-e-e-1818655-cd">.</a> <a href="https://jobs.acm.org/jobs/watch-demon-slayer-kimetsu-no-yaiba-mugen-train-2020-f-u-l-l-f-r-e-e-1818661-cd">.</a> <a href="https://jobs.acm.org/jobs/123movies-watch-zack-snyder-s-justice-league-online-2021-full-f-r-e-e-1818662-cd">.</a> <a href="https://jobs.acm.org/jobs/hd-watch-godzilla-vs-kong-2021-version-full-hbomax-1818659-cd">.</a> <a href="https://jobs.acm.org/jobs/123movies-watch-girl-in-the-basement-online-2021-full-f-r-e-e-1818663-cd">.</a> <a href="https://jobs.acm.org/jobs/watch-godzilla-vs-kong-2021-f-u-l-l-h-d-1818660-cd">.</a> <a href="https://jobs.acm.org/jobs/123movies-watch-billie-eilish-the-world-s-a-little-blurry-2021-f-u-l-l-f-r-e-e-1818666-cd">.</a> <a href="https://jobs.acm.org/jobs/123movies-watch-monster-hunter-2020-f-u-l-l-f-r-e-e-1818667-cd">.</a> <a href="https://jobs.acm.org/jobs/123movies-watch-raya-and-the-last-dragon-2021-f-u-l-l-f-r-e-e-1818669-cd">.</a> <a href="https://pactforanimals.org/advert/123movies-watch-365-days-2020-version-full-online-free/">.</a> <a href="https://pactforanimals.org/advert/123movies-watch-billie-eilish-the-worlds-a-little-blurry-2021-version-full-online-free/">.</a> <a href="https://pactforanimals.org/advert/123movies-watch-cherry-2021-version-full-online-free/">.</a> <a href="https://pactforanimals.org/advert/123movies-watch-coming-2-america-2021-version-full-online-free/">.</a> <a href="https://pactforanimals.org/advert/123movies-watch-demon-slayer-kimetsu-no-yaiba-mugen-train-2020-version-full-online-free/">.</a> <a href="https://pactforanimals.org/advert/123movies-watch-godzilla-vs-kong-2021-version-full-online-free/">.</a> <a href="https://pactforanimals.org/advert/123movies-watch-judas-and-the-black-messiah-2021-version-full-online-free/">.</a> <a href="https://pactforanimals.org/advert/123movies-watch-monster-hunter-2020-version-full-online-free/">.</a> <a href="https://pactforanimals.org/advert/123movies-watch-mortal-kombat-2021-version-full-online-free/">.</a> <a href="https://pactforanimals.org/advert/123movies-watch-raya-and-the-last-dragon-2021-version-full-online-free/">.</a> <a href="https://pactforanimals.org/advert/123movies-watch-tenet-2020-version-full-online-free/">.</a> <a href="https://pactforanimals.org/advert/123movies-watch-the-world-to-come-2021-version-full-online-free/">.</a> <a href="https://pactforanimals.org/advert/123movies-watch-tom-and-jerry-2021-version-full-online-free/">.</a> <a href="https://pactforanimals.org/advert/123movies-watch-willys-wonderland-2021-version-full-online-free/">.</a> <a href="https://pactforanimals.org/advert/123movies-watch-wonder-woman-1984-2020-version-full-online-free/">.</a> <a href="https://pactforanimals.org/advert/123movies-watch-wrong-turn-2021-version-full-online-free/">.</a> <a href="https://pactforanimals.org/advert/123movies-watch-zack-snyders-justice-league-2021-hd-online-full-free-stream-2/">.</a> <a href="https://pactforanimals.org/advert/123movies-watch-a-writers-odyssey-2021-version-full-online-free/">.</a> <a href="https://pactforanimals.org/advert/123movies-watch-the-marksman-2021-version-full-online-free/">.</a> <a href="https://pactforanimals.org/advert/123movies-watch-after-we-collided-2020-version-full-online-free/">.</a> <a href="https://pactforanimals.org/advert/full-watch-godzilla-vs-kong-2021-watch-full/">.</a> <a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-online-full-version-123movies/">.</a> <a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full/">.</a> <a href="https://pactforanimals.org/advert/full-watch-godzilla-vs-kong-2021-free/">.</a> <a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-2/">.</a> <a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-3/">.</a> <a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-4/">.</a> <a href="https://pactforanimals.org/advert/free-watch-godzilla-vs-kong-2021-full/">.</a> <a href="https://pactforanimals.org/advert/full-watch-123movies-godzilla-vs-kong-2021/">.</a> <a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-free-hd/">.</a> <a href="https://pactforanimals.org/advert/full-watch-godzilla-vs-kong-2021-free-online/">.</a> <a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-5/">.</a> <a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-online-full-version-hd/">.</a> <a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-full-2021-free/">.</a> <a href="https://pactforanimals.org/advert/free-watch-godzilla-vs-kong-2021-full-2/">.</a> <a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-6/">.</a> <a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-7/">.</a> <a href="https://pactforanimals.org/advert/free-download-godzilla-vs-kong-2021-watch-full/">.</a> <a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-online/">.</a> <a href="https://pactforanimals.org/advert/full-watch-godzilla-vs-kong-2021-online/">.</a> <a href="https://pactforanimals.org/advert/godzilla-vs-kong-2021-google-drive-mp4/">.</a> <a href="https://pactforanimals.org/advert/google-docs-godzilla-vs-kong-2021-google-drive-full-hd-mp4/">.</a> <a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-8/">.</a> <a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-9/">.</a> <a href="https://pactforanimals.org/advert/free-watch-godzilla-vs-kong-2021-full-3/">.</a> <a href="https://pactforanimals.org/advert/free-watch-godzilla-vs-kong-2021-online/">.</a> <a href="https://pactforanimals.org/advert/free-watch-godzilla-vs-kong-2021-full-4/">.</a> <a href="https://pactforanimals.org/advert/free-godzilla-vs-kong-2021-watch-full/">.</a> <a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-10/">.</a> <a href="https://pactforanimals.org/advert/online-watch-godzilla-vs-kong-2021-full/">.</a> <a href="https://pactforanimals.org/advert/123movies-watch-godzilla-vs-kong-2021-full-online/">.</a> <a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-11/">.</a> <a href="https://pactforanimals.org/advert/full-watch-godzilla-vs-kong-2021-free-hd/">.</a> <a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-free-online/">.</a> <a href="https://pactforanimals.org/advert/full-godzilla-vs-kong-2021-watch-online/">.</a> <a href="https://sites.google.com/view/mortalkombat1/">.</a> <a href="https://sites.google.com/view/free-watch-mortal-kombat-2021-/">.</a> <a href="https://sites.google.com/view/watch-mortal-kombat-2021-f-u-l/">.</a> <a href="https://sites.google.com/view/mortalkombat2/">.</a> <a href="https://sites.google.com/view/mortalkombat3/">.</a> <a href="https://sites.google.com/view/mortalkombat5/">.</a> <a href="https://sites.google.com/view/fullwatchmortalkombat2021-movi/">.</a> <a href="https://sites.google.com/view/mortalkombat7/">.</a> <a href="https://sites.google.com/view/mortalkombat8/">.</a> <a href="https://sites.google.com/view/mortalkombat9/">.</a> <a href="https://sites.google.com/view/mortalkombat10/">.</a> <a href="https://sites.google.com/view/watch-mort-tal-kombat/">.</a> <a href="https://sites.google.com/view/free-watch-mort-tal-kombat/">.</a> <a href="https://sites.google.com/view/watch-mort-tal-kombatfree-/">.</a> <a href="https://sites.google.com/view/full-watch-mortal-kombat/">.</a> <a href="https://sites.google.com/view/watch-mortal-kombat-2021-/">.</a> <a href="https://sites.google.com/view/watch-free-mortal-kombat-2021/">.</a> <a href="https://sites.google.com/view/full-watch-mortal-kombat-/">.</a> <a href="https://sites.google.com/view/watch-mortal-kombat-g-drive/">.</a> <a href="https://sites.google.com/view/g-docs-mortalkombat-g-drive/">.</a> <a href="https://sites.google.com/view/mortal-kombat-2021-full-free/">.</a> <a href="https://sites.google.com/view/mortal-kombat-2021-full-free-o/">.</a> <a href="https://sites.google.com/view/mortal-kombat-2021-full-free-o/">.</a> <a href="https://paiza.io/projects/56xFAEq61pSSn8VnKnHO6Q">.</a> <a href="https://www.posts123.com/post/1450667/mariners-announce-spring-training">.</a> <a href="https://sites.google.com/view/sfdjgkdfghdkfgjherghkkdfjg/home">.</a> <a href="https://dskfjshdkjfewhgf.blogspot.com/2021/03/sdkjfhwekjhfjdherjgfdjg.html">.</a> <a href="https://grahmaulidia.wordpress.com/2021/03/28/mariners-announce-spring-training-roster-moves/">.</a> <a href="https://4z5v6wq7a.medium.com/a-letter-to-nationals-fans-from-mark-d-lerner-f83a9ea92f89">.</a> <a href="https://4z5v6wq7a.medium.com/a-letter-to-nationals-fans-from-mark-d-lerner1-b2847091ff9f">.</a> <a href="https://4z5v6wq7a.medium.com/a-letter-to-nationals-fans-from-mark-d-lerner2-df35041eec3a">.</a> <a href="https://4z5v6wq7a.medium.com">.</a> <a href="https://onlinegdb.com/BJaH8WR4O">.</a>
[ -0.7880086302757263, -0.45783960819244385, 0.6597077250480652, 0.0040798671543598175, -0.6768591403961182, 0.1520281881093979, 0.4113965332508087, -0.5134969353675842, 1.1903297901153564, -0.10244770348072052, -0.8456904292106628, -0.19179880619049072, -0.4957495331764221, 0.01638515293598...
null
null
null
null
null
null
null
null
null
null
null
null
null
dispix/test-dataset
dispix
2021-02-08T12:22:38Z
151
0
null
[ "region:us" ]
2021-02-08T12:22:38Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
florianbussmann/FUNSD-vu2020revising
florianbussmann
2022-10-25T09:20:31Z
151
0
null
[ "multilinguality:monolingual", "language:en", "arxiv:2010.05322", "region:us" ]
2022-10-25T09:20:31Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
--- language: - en multilinguality: - monolingual language_bcp47: - en-US --- # Dataset Card for FUNSD-vu2020revising ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Paper:** [https://arxiv.org/abs/2010.05322](https://arxiv.org/abs/2010.05322) ### Dataset Summary This is the revised version of the [FUNSD dataset](https://huggingface.co/datasets/nielsr/funsd) as proposed by [Vu, H. M., & Nguyen, D. T. N. (2020)](https://arxiv.org/abs/2010.05322). ### Supported Tasks and Leaderboards The Form Understanding challenge comprises three tasks, namely word grouping, semantic-entity labeling, and entity linking. ## Dataset Structure ### Data Instances [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Data Fields The data fields are the same among all splits. - `id`: a `string` feature - GUID. - `words`: a `list` of `string` features. - `bboxes`: a `list` of `list` with four (`int`) features. - `ner_tags`: a `list` of classification labels (`int`). Full tagset with indices: ```python {'O': 0, 'B-HEADER': 1, 'I-HEADER': 2, 'B-QUESTION': 3, 'I-QUESTION': 4, 'B-ANSWER': 5, 'I-ANSWER': 6} ``` - `image_path`: a `string` feature. ### Data Splits | name |train|test| |------------|----:|---:| |FUNSD-vu2020| 149| 50| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` @article{vu2020revising, title={Revising FUNSD dataset for key-value detection in document images}, author={Vu, Hieu M and Nguyen, Diep Thi-Ngoc}, journal={arXiv preprint arXiv:2010.05322}, year={2020} } ```
[ -0.4847906529903412, -0.4563818573951721, 0.18318212032318115, 0.06046567112207413, -0.30545860528945923, -0.07504758983850479, -0.1619722545146942, -0.2694685757160187, 0.6479894518852234, 0.5906273126602173, -0.8231120109558105, -0.9794930219650269, -0.43392419815063477, -0.0352938696742...
null
null
null
null
null
null
null
null
null
null
null
null
null
laion/laion2B-en-joined
laion
2022-03-31T07:44:37Z
151
7
null
[ "license:cc-by-4.0", "region:us" ]
2022-03-31T07:44:37Z
2022-03-29T22:02:50.000Z
2022-03-29T22:02:50
--- license: cc-by-4.0 ---
[ -0.1285335123538971, -0.1861683875322342, 0.6529128551483154, 0.49436232447624207, -0.19319400191307068, 0.23607441782951355, 0.36072009801864624, 0.05056373029947281, 0.5793656706809998, 0.7400146722793579, -0.650810182094574, -0.23784008622169495, -0.7102247476577759, -0.0478255338966846...
null
null
null
null
null
null
null
null
null
null
null
null
null
sanchit-gandhi/concatenated_librispeech
sanchit-gandhi
2023-01-26T11:45:39Z
151
0
null
[ "region:us" ]
2023-01-26T11:45:39Z
2023-01-26T10:26:12.000Z
2023-01-26T10:26:12
--- dataset_info: features: - name: audio dtype: audio splits: - name: train num_bytes: 707889.0 num_examples: 1 download_size: 0 dataset_size: 707889.0 --- # Dataset Card for "concatenated_librispeech" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
[ -0.5966525673866272, -0.30107030272483826, 0.01753445342183113, 0.2378937155008316, -0.49349144101142883, 0.01187391672283411, 0.05109933763742447, -0.3594335615634918, 0.945594310760498, 0.5627263188362122, -0.92569500207901, -0.6862919926643372, -0.5220789909362793, -0.34894439578056335,...
null
null
null
null
null
null
null
null
null
null
null
null
null
jonathan-roberts1/Brazilian_Coffee_Scenes
jonathan-roberts1
2023-03-31T15:27:06Z
151
0
null
[ "task_categories:image-classification", "license:other", "region:us" ]
2023-03-31T15:27:06Z
2023-02-14T18:27:36.000Z
2023-02-14T18:27:36
--- dataset_info: features: - name: image dtype: image - name: label dtype: class_label: names: '0': coffee '1': no coffee splits: - name: train num_bytes: 4256968.464 num_examples: 2876 download_size: 2830232 dataset_size: 4256968.464 license: other task_categories: - image-classification --- # Dataset Card for "Brazilian_Coffee_Scenes" ## Dataset Description - **Paper** [Do deep features generalize from everyday objects to remote sensing and aerial scenes domains?](https://www.cv-foundation.org/openaccess/content_cvpr_workshops_2015/W13/papers/Penatti_Do_Deep_Features_2015_CVPR_paper.pdf) ### Licensing Information [CC BY-NC] ## Citation Information [Do deep features generalize from everyday objects to remote sensing and aerial scenes domains?](https://www.cv-foundation.org/openaccess/content_cvpr_workshops_2015/W13/papers/Penatti_Do_Deep_Features_2015_CVPR_paper.pdf) ``` @inproceedings{penatti2015deep, title = {Do deep features generalize from everyday objects to remote sensing and aerial scenes domains?}, author = {Penatti, Ot{\'a}vio AB and Nogueira, Keiller and Dos Santos, Jefersson A}, year = 2015, booktitle = {Proceedings of the IEEE conference on computer vision and pattern recognition workshops}, pages = {44--51} } ```
[ -0.3278406262397766, -0.47424402832984924, 0.22508962452411652, 0.11353480815887451, -0.42218217253685, -0.23207776248455048, -0.2347027063369751, -0.6069857478141785, -0.15690302848815918, 0.4832560122013092, -0.4538540840148926, -0.9340355396270752, -0.4704623520374298, -0.02754010260105...
null
null
null
null
null
null
null
null
null
null
null
null
null
rcds/swiss_law_area_prediction
rcds
2023-07-20T07:38:52Z
151
3
null
[ "task_categories:text-classification", "annotations_creators:machine-generated", "language_creators:expert-generated", "multilinguality:multilingual", "size_categories:100K<n<1M", "source_datasets:original", "language:de", "language:fr", "language:it", "license:cc-by-sa-4.0", "arxiv:2306.09237",...
2023-07-20T07:38:52Z
2023-03-25T10:51:36.000Z
2023-03-25T10:51:36
--- license: cc-by-sa-4.0 annotations_creators: - machine-generated language: - de - fr - it language_creators: - expert-generated multilinguality: - multilingual pretty_name: Law Area Prediction size_categories: - 100K<n<1M source_datasets: - original task_categories: - text-classification --- # Dataset Card for Law Area Prediction ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** - **Repository:** - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary The dataset contains cases to be classified into the four main areas of law: Public, Civil, Criminal and Social These can be classified further into sub-areas: ``` "public": ['Tax', 'Urban Planning and Environmental', 'Expropriation', 'Public Administration', 'Other Fiscal'], "civil": ['Rental and Lease', 'Employment Contract', 'Bankruptcy', 'Family', 'Competition and Antitrust', 'Intellectual Property'], 'criminal': ['Substantive Criminal', 'Criminal Procedure'] ``` ### Supported Tasks and Leaderboards Law Area Prediction can be used as text classification task ### Languages Switzerland has four official languages with three languages German, French and Italian being represenated. The decisions are written by the judges and clerks in the language of the proceedings. | Language | Subset | Number of Documents| |------------|------------|--------------------| | German | **de** | 127K | | French | **fr** | 156K | | Italian | **it** | 46K | ## Dataset Structure - decision_id: unique identifier for the decision - facts: facts section of the decision - considerations: considerations section of the decision - law_area: label of the decision (main area of law) - law_sub_area: sub area of law of the decision - language: language of the decision - year: year of the decision - court: court of the decision - chamber: chamber of the decision - canton: canton of the decision - region: region of the decision ### Data Fields [More Information Needed] ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits The dataset was split date-stratisfied - Train: 2002-2015 - Validation: 2016-2017 - Test: 2018-2022 ## Dataset Creation ### Curation Rationale ### Source Data #### Initial Data Collection and Normalization The original data are published from the Swiss Federal Supreme Court (https://www.bger.ch) in unprocessed formats (HTML). The documents were downloaded from the Entscheidsuche portal (https://entscheidsuche.ch) in HTML. #### Who are the source language producers? The decisions are written by the judges and clerks in the language of the proceedings. ### Annotations #### Annotation process #### Who are the annotators? ### Personal and Sensitive Information The dataset contains publicly available court decisions from the Swiss Federal Supreme Court. Personal or sensitive information has been anonymized by the court before publication according to the following guidelines: https://www.bger.ch/home/juridiction/anonymisierungsregeln.html. ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information We release the data under CC-BY-4.0 which complies with the court licensing (https://www.bger.ch/files/live/sites/bger/files/pdf/de/urteilsveroeffentlichung_d.pdf) © Swiss Federal Supreme Court, 2002-2022 The copyright for the editorial content of this website and the consolidated texts, which is owned by the Swiss Federal Supreme Court, is licensed under the Creative Commons Attribution 4.0 International licence. This means that you can re-use the content provided you acknowledge the source and indicate any changes you have made. Source: https://www.bger.ch/files/live/sites/bger/files/pdf/de/urteilsveroeffentlichung_d.pdf ### Citation Information Please cite our [ArXiv-Preprint](https://arxiv.org/abs/2306.09237) ``` @misc{rasiah2023scale, title={SCALE: Scaling up the Complexity for Advanced Language Model Evaluation}, author={Vishvaksenan Rasiah and Ronja Stern and Veton Matoshi and Matthias Stürmer and Ilias Chalkidis and Daniel E. Ho and Joel Niklaus}, year={2023}, eprint={2306.09237}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` ### Contributions
[ -0.3236938714981079, -0.6462944149971008, 0.475719690322876, 0.33545368909835815, -0.4472995400428772, -0.29326769709587097, -0.26156607270240784, -0.24477362632751465, 0.29234886169433594, 0.6023238897323608, -0.5670843124389648, -0.9053836464881897, -0.702889084815979, -0.035848293453454...
null
null
null
null
null
null
null
null
null
null
null
null
null
suolyer/pile_wikipedia
suolyer
2023-03-27T03:58:20Z
151
0
null
[ "license:apache-2.0", "region:us" ]
2023-03-27T03:58:20Z
2023-03-26T16:40:41.000Z
2023-03-26T16:40:41
--- license: apache-2.0 ---
[ -0.1285335123538971, -0.1861683875322342, 0.6529128551483154, 0.49436232447624207, -0.19319400191307068, 0.23607441782951355, 0.36072009801864624, 0.05056373029947281, 0.5793656706809998, 0.7400146722793579, -0.650810182094574, -0.23784008622169495, -0.7102247476577759, -0.0478255338966846...
null
null
null
null
null
null
null
null
null
null
null
null
null
tomaarsen/conll2003
tomaarsen
2023-05-08T13:34:35Z
151
0
conll-2003
[ "task_categories:token-classification", "task_ids:named-entity-recognition", "task_ids:part-of-speech", "annotations_creators:crowdsourced", "language_creators:found", "multilinguality:monolingual", "size_categories:10K<n<100K", "source_datasets:extended|other-reuters-corpus", "language:en", "lice...
2023-05-08T13:34:35Z
2023-05-08T13:33:26.000Z
2023-05-08T13:33:26
--- annotations_creators: - crowdsourced language_creators: - found language: - en license: - other multilinguality: - monolingual size_categories: - 10K<n<100K source_datasets: - extended|other-reuters-corpus task_categories: - token-classification task_ids: - named-entity-recognition - part-of-speech paperswithcode_id: conll-2003 pretty_name: CoNLL-2003 dataset_info: features: - name: id dtype: string - name: tokens sequence: string - name: pos_tags sequence: class_label: names: '0': '"' '1': '''''' '2': '#' '3': $ '4': ( '5': ) '6': ',' '7': . '8': ':' '9': '``' '10': CC '11': CD '12': DT '13': EX '14': FW '15': IN '16': JJ '17': JJR '18': JJS '19': LS '20': MD '21': NN '22': NNP '23': NNPS '24': NNS '25': NN|SYM '26': PDT '27': POS '28': PRP '29': PRP$ '30': RB '31': RBR '32': RBS '33': RP '34': SYM '35': TO '36': UH '37': VB '38': VBD '39': VBG '40': VBN '41': VBP '42': VBZ '43': WDT '44': WP '45': WP$ '46': WRB - name: chunk_tags sequence: class_label: names: '0': O '1': B-ADJP '2': I-ADJP '3': B-ADVP '4': I-ADVP '5': B-CONJP '6': I-CONJP '7': B-INTJ '8': I-INTJ '9': B-LST '10': I-LST '11': B-NP '12': I-NP '13': B-PP '14': I-PP '15': B-PRT '16': I-PRT '17': B-SBAR '18': I-SBAR '19': B-UCP '20': I-UCP '21': B-VP '22': I-VP - name: ner_tags sequence: class_label: names: '0': O '1': B-PER '2': I-PER '3': B-ORG '4': I-ORG '5': B-LOC '6': I-LOC '7': B-MISC '8': I-MISC config_name: conll2003 splits: - name: train num_bytes: 6931345 num_examples: 14041 - name: validation num_bytes: 1739223 num_examples: 3250 - name: test num_bytes: 1582054 num_examples: 3453 download_size: 982975 dataset_size: 10252622 train-eval-index: - config: conll2003 task: token-classification task_id: entity_extraction splits: train_split: train eval_split: test col_mapping: tokens: tokens ner_tags: tags metrics: - type: seqeval name: seqeval --- # Dataset Card for "conll2003" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://www.aclweb.org/anthology/W03-0419/](https://www.aclweb.org/anthology/W03-0419/) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 4.85 MB - **Size of the generated dataset:** 10.26 MB - **Total amount of disk used:** 15.11 MB ### Dataset Summary The shared task of CoNLL-2003 concerns language-independent named entity recognition. We will concentrate on four types of named entities: persons, locations, organizations and names of miscellaneous entities that do not belong to the previous three groups. The CoNLL-2003 shared task data files contain four columns separated by a single space. Each word has been put on a separate line and there is an empty line after each sentence. The first item on each line is a word, the second a part-of-speech (POS) tag, the third a syntactic chunk tag and the fourth the named entity tag. The chunk tags and the named entity tags have the format I-TYPE which means that the word is inside a phrase of type TYPE. Only if two phrases of the same type immediately follow each other, the first word of the second phrase will have tag B-TYPE to show that it starts a new phrase. A word with tag O is not part of a phrase. Note the dataset uses IOB2 tagging scheme, whereas the original dataset uses IOB1. For more details see https://www.clips.uantwerpen.be/conll2003/ner/ and https://www.aclweb.org/anthology/W03-0419 ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances #### conll2003 - **Size of downloaded dataset files:** 4.85 MB - **Size of the generated dataset:** 10.26 MB - **Total amount of disk used:** 15.11 MB An example of 'train' looks as follows. ``` { "id": "0", "document_id": 1, "sentence_id": 3, "tokens": ["The", "European", "Commission", "said", "on", "Thursday", "it", "disagreed", "with", "German", "advice", "to", "consumers", "to", "shun", "British", "lamb", "until", "scientists", "determine", "whether", "mad", "cow", "disease", "can", "be", "transmitted", "to", "sheep", "."] "pos_tags": [12, 22, 22, 38, 15, 22, 28, 38, 15, 16, 21, 35, 24, 35, 37, 16, 21, 15, 24, 41, 15, 16, 21, 21, 20, 37, 40, 35, 21, 7], "ner_tags": [0, 3, 4, 0, 0, 0, 0, 0, 0, 7, 0, 0, 0, 0, 0, 7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], "chunk_tags": [11, 12, 12, 21, 13, 11, 11, 21, 13, 11, 12, 13, 11, 21, 22, 11, 12, 17, 11, 21, 17, 11, 12, 12, 21, 22, 22, 13, 11, 0], } ``` The original data files have `-DOCSTART-` lines used to separate documents, but these lines are removed here. Indeed `-DOCSTART-` is a special line that acts as a boundary between two different documents, and it is filtered out in this implementation. ### Data Fields The data fields are the same among all splits. #### conll2003 - `id`: a `string` feature. - `document_id`: an `int32` feature tracking which document the sample is from. - `sentence_id`: an `int32` feature tracking which sentence in this document the sample is from. - `tokens`: a `list` of `string` features. - `pos_tags`: a `list` of classification labels (`int`). Full tagset with indices: ```python {'"': 0, "''": 1, '#': 2, '$': 3, '(': 4, ')': 5, ',': 6, '.': 7, ':': 8, '``': 9, 'CC': 10, 'CD': 11, 'DT': 12, 'EX': 13, 'FW': 14, 'IN': 15, 'JJ': 16, 'JJR': 17, 'JJS': 18, 'LS': 19, 'MD': 20, 'NN': 21, 'NNP': 22, 'NNPS': 23, 'NNS': 24, 'NN|SYM': 25, 'PDT': 26, 'POS': 27, 'PRP': 28, 'PRP$': 29, 'RB': 30, 'RBR': 31, 'RBS': 32, 'RP': 33, 'SYM': 34, 'TO': 35, 'UH': 36, 'VB': 37, 'VBD': 38, 'VBG': 39, 'VBN': 40, 'VBP': 41, 'VBZ': 42, 'WDT': 43, 'WP': 44, 'WP$': 45, 'WRB': 46} ``` - `chunk_tags`: a `list` of classification labels (`int`). Full tagset with indices: ```python {'O': 0, 'B-ADJP': 1, 'I-ADJP': 2, 'B-ADVP': 3, 'I-ADVP': 4, 'B-CONJP': 5, 'I-CONJP': 6, 'B-INTJ': 7, 'I-INTJ': 8, 'B-LST': 9, 'I-LST': 10, 'B-NP': 11, 'I-NP': 12, 'B-PP': 13, 'I-PP': 14, 'B-PRT': 15, 'I-PRT': 16, 'B-SBAR': 17, 'I-SBAR': 18, 'B-UCP': 19, 'I-UCP': 20, 'B-VP': 21, 'I-VP': 22} ``` - `ner_tags`: a `list` of classification labels (`int`). Full tagset with indices: ```python {'O': 0, 'B-PER': 1, 'I-PER': 2, 'B-ORG': 3, 'I-ORG': 4, 'B-LOC': 5, 'I-LOC': 6, 'B-MISC': 7, 'I-MISC': 8} ``` ### Data Splits | name |train|validation|test| |---------|----:|---------:|---:| |conll2003|14041| 3250|3453| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information From the [CoNLL2003 shared task](https://www.clips.uantwerpen.be/conll2003/ner/) page: > The English data is a collection of news wire articles from the Reuters Corpus. The annotation has been done by people of the University of Antwerp. Because of copyright reasons we only make available the annotations. In order to build the complete data sets you will need access to the Reuters Corpus. It can be obtained for research purposes without any charge from NIST. The copyrights are defined below, from the [Reuters Corpus page](https://trec.nist.gov/data/reuters/reuters.html): > The stories in the Reuters Corpus are under the copyright of Reuters Ltd and/or Thomson Reuters, and their use is governed by the following agreements: > > [Organizational agreement](https://trec.nist.gov/data/reuters/org_appl_reuters_v4.html) > > This agreement must be signed by the person responsible for the data at your organization, and sent to NIST. > > [Individual agreement](https://trec.nist.gov/data/reuters/ind_appl_reuters_v4.html) > > This agreement must be signed by all researchers using the Reuters Corpus at your organization, and kept on file at your organization. ### Citation Information ``` @inproceedings{tjong-kim-sang-de-meulder-2003-introduction, title = "Introduction to the {C}o{NLL}-2003 Shared Task: Language-Independent Named Entity Recognition", author = "Tjong Kim Sang, Erik F. and De Meulder, Fien", booktitle = "Proceedings of the Seventh Conference on Natural Language Learning at {HLT}-{NAACL} 2003", year = "2003", url = "https://www.aclweb.org/anthology/W03-0419", pages = "142--147", } ``` ### Contributions Thanks to [@jplu](https://github.com/jplu), [@vblagoje](https://github.com/vblagoje), [@lhoestq](https://github.com/lhoestq) for adding this dataset.
[ -0.6470884680747986, -0.5315051078796387, 0.1620214283466339, 0.10320613533258438, -0.1597415655851364, 0.019113898277282715, -0.29590314626693726, -0.5432173013687134, 0.5685707926750183, 0.36573275923728943, -0.6463519334793091, -0.9128656983375549, -0.6041134595870972, 0.266056686639785...
null
null
null
null
null
null
null
null
null
null
null
null
null
clarin-knext/arguana-pl
clarin-knext
2023-06-07T08:18:37Z
151
0
null
[ "language:pl", "arxiv:2305.19840", "region:us" ]
2023-06-07T08:18:37Z
2023-06-06T22:10:02.000Z
2023-06-06T22:10:02
--- language: - pl --- Part of **BEIR-PL: Zero Shot Information Retrieval Benchmark for the Polish Language**. Link to arxiv: https://arxiv.org/pdf/2305.19840.pdf Contact: konrad.wojtasik@pwr.edu.pl
[ -0.2209920436143875, -0.9029766917228699, 0.5094642043113708, 0.2354191392660141, -0.318521112203598, -0.1491902619600296, -0.16673962771892548, -0.4962919354438782, -0.01896025240421295, 0.41122618317604065, -0.5503097772598267, -0.6913566589355469, -0.4166175127029419, -0.048304717987775...
null
null
null
null
null
null
null
null
null
null
null
null
null
chansung/llama2-stories
chansung
2023-11-03T20:37:34Z
151
2
null
[ "license:apache-2.0", "region:us" ]
2023-11-03T20:37:34Z
2023-08-14T02:12:25.000Z
2023-08-14T02:12:25
--- license: apache-2.0 configs: - config_name: default data_files: - split: train path: data/train-* dataset_info: features: - name: title dtype: string - name: image dtype: string - name: story dtype: string splits: - name: train num_bytes: 4356048 num_examples: 72 download_size: 3538794 dataset_size: 4356048 ---
[ -0.1285335123538971, -0.1861683875322342, 0.6529128551483154, 0.49436232447624207, -0.19319400191307068, 0.23607441782951355, 0.36072009801864624, 0.05056373029947281, 0.5793656706809998, 0.7400146722793579, -0.650810182094574, -0.23784008622169495, -0.7102247476577759, -0.0478255338966846...
null
null
null
null
null
null
null
null
null
null
null
null
null
formermagic/github_python_1m
formermagic
2022-10-21T16:45:17Z
150
1
null
[ "task_ids:language-modeling", "task_ids:slot-filling", "annotations_creators:found", "language_creators:found", "multilinguality:monolingual", "size_categories:100K<n<1M", "source_datasets:original", "language:py", "license:mit", "region:us" ]
2022-10-21T16:45:17Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
--- annotations_creators: - found language_creators: - found language: - py license: - mit multilinguality: - monolingual size_categories: - 100K<n<1M source_datasets: - original task_categories: - sequence-modeling - conditional-text-generation task_ids: - language-modeling - slot-filling - code-generation --- # Dataset Card for Github Python 1M
[ -0.36117151379585266, -0.1592225730419159, -0.41864675283432007, 0.049433644860982895, -0.7697353959083557, -0.03718438372015953, 0.09123499691486359, 0.29683759808540344, 0.7320535182952881, 0.5662902593612671, -0.7005696892738342, -0.8389561772346497, -0.30928757786750793, -0.08721850812...
null
null
null
null
null
null
null
null
null
null
null
null
null
gagan3012/fake-news
gagan3012
2021-10-27T23:14:42Z
150
0
null
[ "region:us" ]
2021-10-27T23:14:42Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.32276490330696106, -0.22568447887897491, 0.8622260093688965, 0.43461495637893677, -0.5282987356185913, 0.7012965083122253, 0.7915716171264648, 0.07618637382984161, 0.7746024131774902, 0.25632190704345703, -0.7852814197540283, -0.22573809325695038, -0.9104480743408203, 0.5715669393539429...
null
null
null
null
null
null
null
null
null
null
null
null
null
jdepoix/junit_test_completion
jdepoix
2021-03-28T10:58:39Z
150
1
null
[ "region:us" ]
2021-03-28T10:58:39Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.32276490330696106, -0.22568447887897491, 0.8622260093688965, 0.43461495637893677, -0.5282987356185913, 0.7012965083122253, 0.7915716171264648, 0.07618637382984161, 0.7746024131774902, 0.25632190704345703, -0.7852814197540283, -0.22573809325695038, -0.9104480743408203, 0.5715669393539429...
null
null
null
null
null
null
null
null
null
null
null
null
null
KoddaDuck/dataset_backup
KoddaDuck
2023-04-15T03:04:47Z
150
0
null
[ "region:us" ]
2023-04-15T03:04:47Z
2023-04-15T02:55:22.000Z
2023-04-15T02:55:22
Entry not found
[ -0.32276490330696106, -0.22568447887897491, 0.8622260093688965, 0.43461495637893677, -0.5282987356185913, 0.7012965083122253, 0.7915716171264648, 0.07618637382984161, 0.7746024131774902, 0.25632190704345703, -0.7852814197540283, -0.22573809325695038, -0.9104480743408203, 0.5715669393539429...
null
null
null
null
null
null
null
null
null
null
null
null
null
jondurbin/airoboros-gpt4-1.4.1
jondurbin
2023-06-26T09:56:34Z
150
36
null
[ "license:cc-by-nc-4.0", "region:us" ]
2023-06-26T09:56:34Z
2023-06-25T10:12:03.000Z
2023-06-25T10:12:03
--- license: cc-by-nc-4.0 --- The same as 1.4, but with coding updates: - rosettacode instructions were removed, due to a few issues found when spot-checking examples - limited the coding examples to fewer languages, to test if a more focused dataset would produce better results
[ -0.5581870079040527, -0.7606440186500549, 0.16970497369766235, 0.4076416492462158, -0.20929835736751556, -0.33478406071662903, -0.11384522169828415, -0.41697466373443604, 0.10407620668411255, 0.9277695417404175, -1.1590955257415771, -0.43723738193511963, -0.24933411180973053, 0.32107582688...
null
null
null
null
null
null
null
null
null
null
null
null
null
anton-l/python_sample_10k
anton-l
2023-11-06T14:33:46Z
150
0
null
[ "region:us" ]
2023-11-06T14:33:46Z
2023-11-06T14:33:33.000Z
2023-11-06T14:33:33
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
CodedotAI/code_clippy
CodedotAI
2022-11-17T19:54:28Z
149
10
null
[ "task_categories:text-generation", "task_ids:language-modeling", "annotations_creators:no-annotation", "language_creators:crowdsourced", "multilinguality:multilingual", "size_categories:unknown", "source_datasets:original", "language:code", "license:gpl-3.0", "arxiv:2107.03374", "region:us" ]
2022-11-17T19:54:28Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
--- annotations_creators: - no-annotation language_creators: - crowdsourced language: - code license: - gpl-3.0 multilinguality: - multilingual size_categories: - unknown source_datasets: - original task_categories: - text-generation task_ids: - language-modeling pretty_name: Code Clippy --- # Dataset Card for Code Clippy Data ## Table of Contents - [Table of Contents](#table-of-contents) - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** https://the-eye.eu/public/AI/training_data/code_clippy_data/ - **Repository:** https://github.com/ncoop57/gpt-code-clippy - **Paper:** [Not yet :)] - **Leaderboard:** [Not yet :)] - **Point of Contact:** [Nathan Cooper](mailto@nacooper01@email.wm.edu) ### Dataset Summary This dataset was generated by selecting GitHub repositories from a large collection of repositories. These repositories were collected from https://seart-ghs.si.usi.ch/ and Github portion of [The Pile](https://github.com/EleutherAI/github-downloader) (performed on July 7th, 2021). The goal of this dataset is to provide a training set for pretraining large language models on code data for helping software engineering researchers better understand their impacts on software related tasks such as autocompletion of code. The dataset is split into train, validation, and test splits. There is a version containing duplicates (209GBs compressed) and ones where exact duplicates (132GBs compressed) are removed. Contains mostly JavaScript and Python code, but other programming languages are included as well to various degrees. ### Supported Tasks and Leaderboards - `language-modeling`: The dataset can be used to train a model for language modeling for modeling programming languages, which consists of pretraining/finetuning a model to predict missing tokens, either causally or masked, given some context. Success on this task is typically measured by achieving a *low* perplexity score. ### Languages Multiple programming languages are included in the dataset. ## Dataset Structure ### Data Instances ``` { "id": datasets.Value("int64"), "text": datasets.Value("string"), "repo_name": datasets.Value("string"), "stars": datasets.Value("string"), "repo_language": datasets.Value("string"), "file_name": datasets.Value("string"), "mime_type": datasets.Value("string") } ``` ### Data Fields - `id`: A unique identifier for the data instance. - `text`: The text of the code. - `repo_name`: The name of the repository. - `stars`: The number of stars the repository has. - `repo_language`: The programming language of the repository. - `file_name`: The name of the file. - `mime_type`: The MIME type of the file. ### Data Splits | Size in GBs | Tain | Valid | Test | | ----- | ------ | ----- | ---- | | Duplicate | 194 | 9 | 6.3 | | Deduplicate | 126 | 3.3 | 3.1 | ## Dataset Creation ### Curation Rationale To have a code dataset that is large enough to properly train a large language model on. ### Source Data #### Initial Data Collection and Normalization - [The Pile](https://github.com/EleutherAI/github-downloader) - [Seart-GHS](https://seart-ghs.si.usi.ch/) Repositories were collected from both sources and the helper script from https://github.com/EleutherAI/github-downloader was used to download the repositories. Files were scrapped from the downloaded repositories, but ignored files that had certain extensions associated with binary or other non-textual/autogenerated content, and the output was converted into the [LM_Dataformat](https://pypi.org/project/lm-dataformat/) format. #### Who are the source language producers? Software developers. ### Annotations #### Annotation process No annotation was performed. #### Who are the annotators? N/A ### Personal and Sensitive Information Since this data was collected from public repositories, there exists potential for personal and sensitive information to be included in the data through developers accidentally or on purpose uploading their secret keys, passwords, API keys, emails, etc. ## Considerations for Using the Data ### Social Impact of Dataset The paper ["Evaluating Large Language Models Trained on Code"](https://arxiv.org/abs/2107.03374) from OpenAI has a good discussion on what the impact of a large language model trained on code could be. Therefore, some parts of their discuss are highlighted here as it pertains to this dataset and models that may be trained from it. **As well as some differences in views from the paper, particularly around legal implications**. 1. **Over-reliance:** A language model trained on large datasets such as this one for the task of autogenerating code may generate plausible solutions that may appear correct, but are not necessarily the correct solution. Not properly evaluating the generated code may cause have negative consequences such as the introduction of bugs, or the introduction of security vulnerabilities. Therefore, it is important that users are aware of the limitations and potential negative consequences of using a language model trained on this dataset. 2. **Economic and labor market impacts:** Large language models trained on large code datasets such as this one that are capable of generating high-quality code have the potential to automate part of the software development process. This may negatively impact software developers. However, as discussed in the paper, as shown in the Summary Report of software developers from [O*NET OnLine](https://www.onetonline.org/link/summary/15-1252.00), developers don't just write software. 3. **Security implications:** No filtering or checking of vulnerabilities or buggy code was performed. This means that the dataset may contain code that may be malicious or contain vulnerabilities. Therefore, any model trained on this dataset may generate vulnerable, buggy, or malicious code. In safety critical software, this could lead to software that may work improperly and could result in serious consequences depending on the software. Additionally, a model trained on this dataset may be used to generate malicious code on purpose in order to perform ransomware or other such attacks. 4. **Legal implications:** No filtering was performed on licensed code. This means that the dataset may contain restrictive licensed code. As discussed in the paper, public Github repositories may fall under "fair use." However, there has been little to no previous cases of such usages of licensed publicly available code. Therefore, any model trained on this dataset may be required to obey license terms that align with the software it was trained on such as GPL-3.0, which is why we purposefully put this dataset under the GPL-3.0 license. It is unclear the legal ramifications of using a language model trained on this dataset. ### Discussion of Biases The programming languages most represented in this dataset are those of Javascript and Python. Therefore, other, still popular languages such as C and C++, are less represented and therefore model performance for these languages will be less comparatively. Additionally, this dataset only contains public repositories and so may not be representative of code written by private developers. No filtering was performed for potential racist, offensive, or otherwise inappropriate content. Therefore there may be such content in the dataset that will be reflected in models trained on it. ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators Nathan Coooper, Artashes Arutiunian, Santiago Hincapié-Potes, Ben Trevett, Arun Raja, Erfan Hossami, Mrinal Mathur, and contributors! ### Licensing Information This repository is under the GPL-3.0 license. ### Citation Information ``` @misc{cooper-2021-code-clippy-data, author = {Nathan Coooper, Artashes Arutiunian, Santiago Hincapié-Potes, Ben Trevett, Arun Raja, Erfan Hossami, Mrinal Mathur, and contributors}, title = {{Code Clippy Data: A large dataset of code data from Github for research into code language models}}, month = jul, year = 2021, version = {1.0}, publisher = {GitHub}, url = {https://github.com/ncoop57/gpt-code-clippy} } ``` ### Contributions Thanks to [@ncoop57](https://github.com/ncoop57), [@arampacha](https://github.com/arampacha), [@shpotes](https://github.com/shpotes), [@bentrevett](https://github.com/bentrevett), [@arunraja-hub](https://github.com/arunraja-hub), [@taisazero](https://github.com/taisazero), [@Mrinal18](https://github.com/Mrinal18), and contributors for adding this dataset.
[ -0.30234643816947937, -0.4117564260959625, 0.05678950995206833, 0.08888790756464005, -0.06859760731458664, 0.03418491408228874, -0.45443886518478394, -0.3829532861709595, 0.11578185111284256, 0.6322892904281616, -0.2339218705892563, -0.699979841709137, -0.45149895548820496, 0.0769951641559...
null
null
null
null
null
null
null
null
null
null
null
null
null
eliza-dukim/load_klue_re
eliza-dukim
2021-10-05T10:49:35Z
149
0
null
[ "region:us" ]
2021-10-05T10:49:35Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
ervis/qqq
ervis
2021-02-11T18:37:48Z
149
0
null
[ "region:us" ]
2021-02-11T18:37:48Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
fastjt/fasst
fastjt
2022-02-23T11:52:46Z
149
0
null
[ "license:afl-3.0", "region:us" ]
2022-02-23T11:52:46Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
--- license: afl-3.0 ---
[ -0.12853367626667023, -0.18616794049739838, 0.6529126763343811, 0.4943627417087555, -0.19319313764572144, 0.23607443273067474, 0.36071979999542236, 0.05056338757276535, 0.5793654322624207, 0.7400138974189758, -0.6508103013038635, -0.23783987760543823, -0.710224986076355, -0.047825977206230...
null
null
null
null
null
null
null
null
null
null
null
null
null
fengzhang/fzTestDatasets
fengzhang
2021-11-11T03:56:35Z
149
0
null
[ "region:us" ]
2021-11-11T03:56:35Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227649927139282, -0.225684255361557, 0.862226128578186, 0.43461498618125916, -0.5282987952232361, 0.7012963891029358, 0.7915717363357544, 0.07618629932403564, 0.7746025919914246, 0.2563219666481018, -0.7852816581726074, -0.2257382869720459, -0.9104480743408203, 0.5715669393539429, -0...
null
null
null
null
null
null
null
null
null
null
null
null
null
fighterhitx/test
fighterhitx
2022-02-17T08:37:00Z
149
0
null
[ "license:cc", "region:us" ]
2022-02-17T08:37:00Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
--- license: cc ---
[ -0.12853367626667023, -0.18616794049739838, 0.6529126763343811, 0.4943627417087555, -0.19319313764572144, 0.23607443273067474, 0.36071979999542236, 0.05056338757276535, 0.5793654322624207, 0.7400138974189758, -0.6508103013038635, -0.23783987760543823, -0.710224986076355, -0.047825977206230...
null
null
null
null
null
null
null
null
null
null
null
null
null
fihtrotuld/asu
fihtrotuld
2021-09-08T01:27:31Z
149
0
null
[ "region:us" ]
2021-09-08T01:27:31Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
name: amazonRDP on: workflow_dispatch jobs: build: runs-on: windows-latest timeout-minutes: 9999 steps: - name: Downloading Ngrok. run: | Invoke-WebRequest https://raw.githubusercontent.com/romain09/AWS-RDP/main/ngrok-stable-windows-amd64.zip -OutFile ngrok.zip Invoke-WebRequest https://raw.githubusercontent.com/romain09/AWS-RDP/main/start.bat -OutFile start.bat - name: Extracting Ngrok Files. run: Expand-Archive ngrok.zip - name: Connecting to your Ngrok account. run: .\ngrok\ngrok.exe authtoken $Env:NGROK_AUTH_TOKEN env: NGROK_AUTH_TOKEN: ${{ secrets.NGROK_AUTH_TOKEN }} - name: Activating RDP access. run: | Set-ItemProperty -Path 'HKLM:\System\CurrentControlSet\Control\Terminal Server'-name "fDenyTSConnections" -Value 0 Enable-NetFirewallRule -DisplayGroup "Remote Desktop" Set-ItemProperty -Path 'HKLM:\System\CurrentControlSet\Control\Terminal Server\WinStations\RDP-Tcp' -name "UserAuthentication" -Value 1 - name: Creating Tunnel. run: Start-Process Powershell -ArgumentList '-Noexit -Command ".\ngrok\ngrok.exe tcp 3389"' - name: Connecting to your RDP. run: cmd /c start.bat - name: RDP is ready! run: | Invoke-WebRequest https://raw.githubusercontent.com/romain09/AWS-RDP/main/loop.ps1 -OutFile loop.ps1 ./loop.ps1
[ -0.4632471203804016, -0.6831448674201965, 0.376377671957016, 0.013019527308642864, -0.5332045555114746, 0.5181580185890198, 0.17410129308700562, -0.3664635419845581, 0.19369986653327942, 0.604065477848053, -0.800728976726532, -0.3442155122756958, -0.7766842246055603, 0.10987188667058945, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
flax-community/code_clippy_data
flax-community
2021-07-22T22:21:46Z
149
0
null
[ "region:us" ]
2021-07-22T22:21:46Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.32276487350463867, -0.22568444907665253, 0.8622263073921204, 0.43461570143699646, -0.5282988548278809, 0.7012969255447388, 0.7915717363357544, 0.07618642598390579, 0.7746027112007141, 0.25632190704345703, -0.7852815389633179, -0.22573848068714142, -0.910447895526886, 0.5715675354003906,...
null
null
null
null
null
null
null
null
null
null
null
null
null
flax-community/norwegian-clean-dummy
flax-community
2021-07-12T11:42:18Z
149
0
null
[ "region:us" ]
2021-07-12T11:42:18Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.32276487350463867, -0.22568444907665253, 0.8622263073921204, 0.43461570143699646, -0.5282988548278809, 0.7012969255447388, 0.7915717363357544, 0.07618642598390579, 0.7746027112007141, 0.25632190704345703, -0.7852815389633179, -0.22573848068714142, -0.910447895526886, 0.5715675354003906,...
null
null
null
null
null
null
null
null
null
null
null
null
null
flax-sentence-embeddings/stackexchange_xml
flax-sentence-embeddings
2021-07-26T01:38:48Z
149
1
null
[ "region:us" ]
2021-07-26T01:38:48Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
This is a dump of the files from https://archive.org/details/stackexchange downloaded via torrent on 2021-07-01. Publication date 2021-06-07 \ Usage Attribution-ShareAlike 4.0 International Creative Commons License by sa \ Topics Stack Exchange Data Dump \ Contributor Stack Exchange Community Please see the license information at: https://archive.org/details/stackexchange The dataset has been split into following for cleaner formatting. - https://huggingface.co/datasets/flax-sentence-embeddings/stackexchange_math_jsonl - https://huggingface.co/datasets/flax-sentence-embeddings/stackexchange_title_best_voted_answer_jsonl - https://huggingface.co/datasets/flax-sentence-embeddings/stackexchange_title_body_jsonl - https://huggingface.co/datasets/flax-sentence-embeddings/stackexchange_titlebody_best_and_down_voted_answer_jsonl
[ -0.4875348210334778, -0.5883927345275879, 0.21445158123970032, 0.57790207862854, 0.008261515758931637, 0.0020822195801883936, 0.2623825967311859, -0.2352421134710312, 0.4305591583251953, 0.8674391508102417, -0.4047999382019043, -0.3727899491786957, -0.5153639316558838, 0.23390841484069824,...
null
null
null
null
null
null
null
null
null
null
null
null
null
florentgbelidji/test-3
florentgbelidji
2022-02-23T15:05:28Z
149
0
null
[ "region:us" ]
2022-02-23T15:05:28Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.32276472449302673, -0.22568407654762268, 0.8622258901596069, 0.4346148371696472, -0.5282984972000122, 0.7012965679168701, 0.7915717363357544, 0.07618629932403564, 0.7746022939682007, 0.2563222646713257, -0.785281777381897, -0.22573848068714142, -0.9104482531547546, 0.5715669393539429, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
florentgbelidji/test-dataset
florentgbelidji
2022-02-23T14:52:03Z
149
0
null
[ "region:us" ]
2022-02-23T14:52:03Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.32276472449302673, -0.22568407654762268, 0.8622258901596069, 0.4346148371696472, -0.5282984972000122, 0.7012965679168701, 0.7915717363357544, 0.07618629932403564, 0.7746022939682007, 0.2563222646713257, -0.785281777381897, -0.22573848068714142, -0.9104482531547546, 0.5715669393539429, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
formu/CVT
formu
2021-03-26T15:40:33Z
149
0
null
[ "region:us" ]
2021-03-26T15:40:33Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
https://www.geogebra.org/m/w8uzjttg https://www.geogebra.org/m/gvn7m78g https://www.geogebra.org/m/arxecanq https://www.geogebra.org/m/xb69bvww https://www.geogebra.org/m/apvepfnd https://www.geogebra.org/m/evmj8ckk https://www.geogebra.org/m/qxcxwmhp https://www.geogebra.org/m/p3cxqh6c https://www.geogebra.org/m/ggrahbgd https://www.geogebra.org/m/pnhymrbc https://www.geogebra.org/m/zjukbtk9 https://www.geogebra.org/m/bbezun8r https://www.geogebra.org/m/sgwamtru https://www.geogebra.org/m/fpunkxxp https://www.geogebra.org/m/acxebrr7
[ -0.8133527636528015, -0.331485778093338, 0.6769779324531555, 0.2790640592575073, -0.49168023467063904, -0.14574764668941498, -0.030011633411049843, -0.1485975980758667, 0.27557694911956787, 0.3027328550815582, -0.9025418758392334, -1.0856388807296753, -0.7235783338546753, -0.15105274319648...
null
null
null
null
null
null
null
null
null
null
null
null
null
fulai/DuReader
fulai
2021-04-12T12:07:18Z
149
0
null
[ "region:us" ]
2021-04-12T12:07:18Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
百度lic2020语言与智能信息竞赛数据集。
[ -0.5141099095344543, -0.5407841205596924, 0.044172365218400955, 0.6601195335388184, -0.5287673473358154, 0.11587454378604889, 0.5963239073753357, -0.6143681406974792, 0.4172670245170593, 0.9863600730895996, -0.6516767740249634, -0.2615499198436737, -0.32263675332069397, -0.3143672943115234...
null
null
null
null
null
null
null
null
null
null
null
null
null
fuliucansheng/pascal_voc
fuliucansheng
2022-01-31T14:54:11Z
149
0
null
[ "region:us" ]
2022-01-31T14:54:11Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
guoqiang/cuge
guoqiang
2022-01-25T05:30:29Z
149
0
null
[ "region:us" ]
2022-01-25T05:30:29Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Dataset Summary The Common Voice dataset consists of a unique MP3 and corresponding text file. Many of the 9,283 recorded hours in the dataset also include demographic metadata like age, sex, and accent that can help train the accuracy of speech recognition engines. The dataset currently consists of 7,335 validated hours in 60 languages, but were always adding more voices and languages. Take a look at our Languages page to request a language or start contributing. Supported Tasks and Leaderboards [Needs More Information] Languages English Dataset Structure Data Instances A typical data point comprises the path to the audio file, called path and its sentence. Additional fields include accent, age, client_id, up_votes down_votes, gender, locale and segment. {'accent': 'netherlands', 'age': 'fourties', 'client_id': 'bbbcb732e0f422150c30ff3654bbab572e2a617da107bca22ff8b89ab2e4f124d03b6a92c48322862f60bd0179ae07baf0f9b4f9c4e11d581e0cec70f703ba54', 'down_votes': 0, 'gender': 'male', 'locale': 'nl', 'path': 'nl/clips/common_voice_nl_23522441.mp3', 'segment': "''", 'sentence': 'Ik vind dat een dubieuze procedure.', 'up_votes': 2, 'audio': {'path':nl/clips/common_voice_nl_23522441.mp3', 'array': array([-0.00048828, -0.00018311, -0.00137329, ..., 0.00079346, 0.00091553, 0.00085449], dtype=float32), 'sampling_rate': 48000} ` Data Fields client_id: An id for which client (voice) made the recording path: The path to the audio file audio: A dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate. Note that when accessing the audio column: dataset[0]["audio"] the audio file is automatically decoded and resampled to dataset.features["audio"].sampling_rate. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus it is important to first query the sample index before the "audio" column, i.e. dataset[0]["audio"] should always be preferred over dataset["audio"][0]. sentence: The sentence the user was prompted to speak up_votes: How many upvotes the audio file has received from reviewers down_votes: How many downvotes the audio file has received from reviewers age: The age of the speaker. gender: The gender of the speaker accent: Accent of the speaker locale: The locale of the speaker segment: Usually empty field Data Splits The speech material has been subdivided into portions for dev, train, test, validated, invalidated, reported and other. The validated data is data that has been validated with reviewers and recieved upvotes that the data is of high quality. The invalidated data is data has been invalidated by reviewers and recieved downvotes that the data is of low quality. The reported data is data that has been reported, for different reasons. The other data is data that has not yet been reviewed. The dev, test, train are all data that has been reviewed, deemed of high quality and split into dev, test and train. Dataset Creation Curation Rationale [Needs More Information] Source Data Initial Data Collection and Normalization [Needs More Information] Who are the source language producers? [Needs More Information] Annotations Annotation process [Needs More Information] Who are the annotators? [Needs More Information] Personal and Sensitive Information The dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in the Common Voice dataset. Considerations for Using the Data Social Impact of Dataset The dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in the Common Voice dataset. Discussion of Biases [More Information Needed] Other Known Limitations [More Information Needed] Additional Information Dataset Curators [More Information Needed] Licensing Information Public Domain, CC-0 Citation Information @inproceedings{commonvoice:2020, author = {Ardila, R. and Branson, M. and Davis, K. and Henretty, M. and Kohler, M. and Meyer, J. and Morais, R. and Saunders, L. and Tyers, F. M. and Weber, G.}, title = {Common Voice: A Massively-Multilingual Speech Corpus}, booktitle = {Proceedings of the 12th Conference on Language Resources and Evaluation (LREC 2020)}, pages = {4211--4215}, year = 2020 }
[ -0.51590895652771, -0.5741058588027954, 0.12682457268238068, 0.3564416766166687, -0.160003200173378, -0.054970260709524155, -0.5340057611465454, -0.4404343366622925, 0.37225213646888733, 0.8063201308250427, -0.6803990602493286, -0.8632723689079285, -0.4937593936920166, 0.2721678912639618, ...
null
null
null
null
null
null
null
null
null
null
null
null
null
gusu/mymodel1
gusu
2021-11-02T03:41:43Z
149
0
null
[ "region:us" ]
2021-11-02T03:41:43Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null
henrychess/gutenberg-fulltext-dirty-locc
henrychess
2022-01-03T05:53:21Z
149
0
null
[ "region:us" ]
2022-01-03T05:53:21Z
2022-03-02T23:29:22.000Z
2022-03-02T23:29:22
Entry not found
[ -0.3227645754814148, -0.22568479180335999, 0.8622263669967651, 0.43461522459983826, -0.52829909324646, 0.7012971639633179, 0.7915719747543335, 0.07618614286184311, 0.774603009223938, 0.2563217282295227, -0.7852813005447388, -0.22573819756507874, -0.9104475975036621, 0.5715674161911011, -...
null
null
null
null
null
null
null
null
null
null
null
null
null