id stringlengths 2 115 | author stringlengths 2 42 ⌀ | last_modified timestamp[us, tz=UTC] | downloads int64 0 8.87M | likes int64 0 3.84k | paperswithcode_id stringlengths 2 45 ⌀ | tags list | lastModified timestamp[us, tz=UTC] | createdAt stringlengths 24 24 | key stringclasses 1 value | created timestamp[us] | card stringlengths 1 1.01M | embedding list | library_name stringclasses 21 values | pipeline_tag stringclasses 27 values | mask_token null | card_data null | widget_data null | model_index null | config null | transformers_info null | spaces null | safetensors null | transformersInfo null | modelId stringlengths 5 111 ⌀ | embeddings list |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
cahya/persona_empathetic | cahya | 2022-02-19T22:49:35Z | 153 | 0 | null | [
"license:mit",
"region:us"
] | 2022-02-19T22:49:35Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | ---
license: mit
---
| [
-0.12853367626667023,
-0.18616794049739838,
0.6529126763343811,
0.4943627417087555,
-0.19319313764572144,
0.23607443273067474,
0.36071979999542236,
0.05056338757276535,
0.5793654322624207,
0.7400138974189758,
-0.6508103013038635,
-0.23783987760543823,
-0.710224986076355,
-0.047825977206230... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
cakiki/arxiv-metadata | cakiki | 2022-02-03T20:57:23Z | 153 | 0 | null | [
"license:cc0-1.0",
"region:us"
] | 2022-02-03T20:57:23Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | ---
license: cc0-1.0
---
| [
-0.12853367626667023,
-0.18616794049739838,
0.6529126763343811,
0.4943627417087555,
-0.19319313764572144,
0.23607443273067474,
0.36071979999542236,
0.05056338757276535,
0.5793654322624207,
0.7400138974189758,
-0.6508103013038635,
-0.23783987760543823,
-0.710224986076355,
-0.047825977206230... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
caythuoc/caoduoclieu | caythuoc | 2023-06-15T10:41:13Z | 153 | 0 | null | [
"region:us"
] | 2023-06-15T10:41:13Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
ccccccc/hdjw_94ejrjr | ccccccc | 2021-02-18T07:41:38Z | 153 | 0 | null | [
"region:us"
] | 2021-02-18T07:41:38Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
cdleong/temp_africaNLP_keyword_spotting_for_african_languages | cdleong | 2022-10-25T09:07:32Z | 153 | 0 | null | [
"language:wo",
"language:fuc",
"language:srr",
"language:mnk",
"language:snk",
"region:us"
] | 2022-10-25T09:07:32Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | ---
language:
- wo
- fuc
- srr
- mnk
- snk
---
## Dataset Description
- **Homepage:** https://zenodo.org/record/4661645
TEMPORARY TEST DATASET
Not for actual use! Attempting to test out a dataset script for loading https://zenodo.org/record/4661645
| [
-0.2611089050769806,
-0.16739805042743683,
0.1438436210155487,
0.5390400886535645,
-0.7581366896629333,
0.13277389109134674,
-0.10329926759004593,
0.1648896485567093,
0.1767561435699463,
0.660321831703186,
-1.231587529182434,
-0.6473388671875,
0.040851250290870667,
0.34319373965263367,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
cem/film | cem | 2021-12-23T22:02:57Z | 153 | 0 | null | [
"region:us"
] | 2021-12-23T22:02:57Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
cemigo/taylor_vs_shakes | cemigo | 2021-03-14T23:45:59Z | 153 | 0 | null | [
"region:us"
] | 2021-03-14T23:45:59Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | This dataset has 336 pieces of quotes from William Shakespeare and Taylor Swift (labeled) for supervised classification.
Source: https://www.kaggle.com/kellylougheed/tswift-vs-shakespeare | [
-0.3528725206851959,
-0.49498608708381653,
0.17421595752239227,
0.3363476097583771,
0.0019012205302715302,
0.06886040419340134,
-0.025744786486029625,
-0.41607120633125305,
0.08843306452035904,
0.862523078918457,
-0.9132828712463379,
-0.16851173341274261,
-0.5095446109771729,
0.04360641166... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
cemigo/test-data | cemigo | 2021-02-07T23:49:41Z | 153 | 0 | null | [
"region:us"
] | 2021-02-07T23:49:41Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
csarron/25m-img-caps | csarron | 2022-03-28T18:51:26Z | 153 | 1 | null | [
"region:us"
] | 2022-03-28T18:51:26Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | see https://huggingface.co/datasets/csarron/4m-img-caps for example usage | [
-0.6945481896400452,
-0.27406325936317444,
0.12098154425621033,
0.5052483081817627,
-0.13724862039089203,
0.032137878239154816,
0.19517825543880463,
-0.45081761479377747,
0.7188369631767273,
0.5162168145179749,
-1.0182642936706543,
-0.7857872247695923,
-0.539254903793335,
0.264833718538284... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
csarron/4m-img-caps | csarron | 2022-03-28T18:50:53Z | 153 | 1 | null | [
"region:us"
] | 2022-03-28T18:50:53Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | see [read_pyarrow.py](https://gist.github.com/csarron/df712e53c9e0dcaad4eb6843e7a3d51c#file-read_pyarrow-py) for how to read one pyarrow file.
example PyTorch dataset:
```python
from torch.utils.data import Dataset
class ImageCaptionArrowDataset(Dataset):
def __init__(
self,
dataset_file,
tokenizer,
):
import pyarrow as pa
data = [pa.ipc.open_file(pa.memory_map(f, "rb")).read_all() for f in glob.glob(dataset_file)]
self.data = pa.concat_tables(data)
# do other initialization, like init image preprocessing fn,
def __getitem__(self, index):
# item_id = self.data["id"][index].as_py()
text = self.data["text"][index].as_py() # get text
if isinstance(text, list):
text = random.choice(text)
img_bytes = self.data["image"][index].as_py() # get image bytes
# do some processing with image and text, return the features
# img_feat = self.image_bytes_to_tensor(img_bytes)
# inputs = self.tokenizer(
# text,
# padding="max_length",
# max_length=self.max_text_len,
# truncation=True,
# return_token_type_ids=True,
# return_attention_mask=True,
# add_special_tokens=True,
# return_tensors="pt",
# )
# input_ids = inputs.input_ids.squeeze(0)
# attention_mask = inputs.attention_mask.squeeze(0)
# return {
# # "item_ids": item_id,
# "text_ids": input_ids,
# "input_ids": input_ids,
# "text_masks": attention_mask,
# "pixel_values": img_feat,
# }
def __len__(self):
return len(self.data)
``` | [
-0.2299148440361023,
-0.544771671295166,
0.10518138855695724,
0.14956609904766083,
-0.3305014967918396,
-0.08597679436206818,
-0.01706288382411003,
-0.33755263686180115,
0.17606347799301147,
0.2263861745595932,
-0.3430895507335663,
-0.5375257730484009,
-0.6523412466049194,
0.02987181767821... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
image-search-2/unsplash_lite_image_dataset | image-search-2 | 2021-11-19T12:44:46Z | 153 | 1 | null | [
"region:us"
] | 2021-11-19T12:44:46Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | # The Unsplash Dataset

The Unsplash Dataset is made up of over 250,000+ contributing global photographers and data sourced from hundreds of millions of searches across a nearly unlimited number of uses and contexts. Due to the breadth of intent and semantics contained within the Unsplash dataset, it enables new opportunities for research and learning.
The Unsplash Dataset is offered in two datasets:
- the Lite dataset: available for commercial and noncommercial usage, containing 25k nature-themed Unsplash photos, 25k keywords, and 1M searches
- the Full dataset: available for noncommercial usage, containing 3M+ high-quality Unsplash photos, 5M keywords, and over 250M searches
As the Unsplash library continues to grow, we’ll release updates to the dataset with new fields and new images, with each subsequent release being [semantically versioned](https://semver.org/).
We welcome any feedback regarding the content of the datasets or their format. With your input, we hope to close the gap between the data we provide and the data that you would like to leverage. You can [open an issue](https://github.com/unsplash/datasets/issues/new/choose) to report a problem or to let us know what you would like to see in the next release of the datasets.
For more on the Unsplash Dataset, see [our announcement](https://unsplash.com/blog/the-unsplash-dataset/) and [site](https://unsplash.com/data).
## Download
### Lite Dataset
The Lite dataset contains all of the same fields as the Full dataset, but is limited to ~25,000 photos. It can be used for both commercial and non-commercial usage, provided you abide by [the terms](https://github.com/unsplash/datasets/blob/master/TERMS.md).
[⬇️ Download the Lite dataset](https://unsplash.com/data/lite/latest) [~650MB compressed, ~1.4GB raw]
### Full Dataset
The Full dataset is available for non-commercial usage and all uses must abide by [the terms](https://github.com/unsplash/datasets/blob/master/TERMS.md). To access, please go to [unsplash.com/data](https://unsplash.com/data) and request access. The dataset weighs ~20 GB compressed (~43GB raw)).
## Documentation
See the [documentation for a complete list of tables and fields](https://github.com/unsplash/datasets/blob/master/DOCS.md).
## Usage
You can follow these examples to load the dataset in these common formats:
- [Load the dataset in a PostgreSQL database](https://github.com/unsplash/datasets/tree/master/how-to/psql)
- [Load the dataset in a Python environment](https://github.com/unsplash/datasets/tree/master/how-to/python)
- [Submit an example doc](https://github.com/unsplash/datasets/blob/master/how-to/README.md#submit-an-example)
## Share your work
We're making this data open and available with the hopes of enabling researchers and developers to discover interesting and useful connections in the data.
We'd love to see what you create, whether that's a research paper, a machine learning model, a blog post, or just an interesting discovery in the data. Send us an email at [data@unsplash.com](mailto:data@unsplash.com).
If you're using the dataset in a research paper, you can attribute the dataset as `Unsplash Lite Dataset 1.2.0` or `Unsplash Full Dataset 1.2.0` and link to the permalink [`unsplash.com/data`](https://unsplash.com/data).
----
The Unsplash Dataset is made available for research purposes. [It cannot be used to redistribute the images contained within](https://github.com/unsplash/datasets/blob/master/TERMS.md). To use the Unsplash library in a product, see [the Unsplash API](https://unsplash.com/developers).

| [
-0.0999462902545929,
-0.18598991632461548,
0.14069034159183502,
0.09842837601900101,
-0.5665063858032227,
0.19609595835208893,
-0.3083202540874481,
-0.3755965828895569,
0.3953779339790344,
0.5846114158630371,
-0.5519187450408936,
-0.7459720373153687,
-0.13306768238544464,
0.127736344933509... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
jimregan/foinse | jimregan | 2021-10-06T20:42:52Z | 153 | 0 | null | [
"region:us"
] | 2021-10-06T20:42:52Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
jakartaresearch/google-play-review | jakartaresearch | 2022-08-06T16:24:49Z | 153 | 4 | null | [
"task_categories:text-classification",
"task_ids:sentiment-classification",
"annotations_creators:found",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:1K<n<10K",
"source_datasets:original",
"language:id",
"license:cc-by-4.0",
"sentiment",
"google-play",
"indonesia... | 2022-08-06T16:24:49Z | 2022-08-06T05:00:32.000Z | 2022-08-06T05:00:32 | ---
annotations_creators:
- found
language:
- id
language_creators:
- found
license:
- cc-by-4.0
multilinguality:
- monolingual
pretty_name: Indonesian Google Play Review
size_categories:
- 1K<n<10K
source_datasets:
- original
tags:
- sentiment
- google-play
- indonesian
task_categories:
- text-classification
task_ids:
- sentiment-classification
---
# Dataset Card for Indonesian Google Play Review
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:**
- **Repository:**
- **Paper:**
- **Leaderboard:**
- **Point of Contact:**
### Dataset Summary
Scrapped from e-commerce app on Google Play.
### Supported Tasks and Leaderboards
Sentiment Analysis
### Languages
Indonesian
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
Thanks to [@andreaschandra](https://github.com/andreaschandra) for adding this dataset. | [
-0.45517972111701965,
-0.49983277916908264,
-0.051154736429452896,
0.49620047211647034,
-0.42576295137405396,
0.08203400671482086,
-0.2973781228065491,
-0.37442344427108765,
0.6587487459182739,
0.8926987051963806,
-0.9955716133117676,
-0.9578273892402649,
-0.6353380680084229,
0.21590693295... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
cardiffnlp/tweet_topic_single | cardiffnlp | 2022-11-27T11:25:34Z | 153 | 3 | null | [
"task_categories:text-classification",
"task_ids:sentiment-classification",
"multilinguality:monolingual",
"size_categories:1k<10K",
"language:en",
"license:other",
"arxiv:2209.09824",
"region:us"
] | 2022-11-27T11:25:34Z | 2022-09-02T00:20:17.000Z | 2022-09-02T00:20:17 | ---
language:
- en
license:
- other
multilinguality:
- monolingual
size_categories:
- 1k<10K
task_categories:
- text-classification
task_ids:
- sentiment-classification
pretty_name: TweetTopicSingle
---
# Dataset Card for "cardiffnlp/tweet_topic_single"
## Dataset Description
- **Paper:** [https://arxiv.org/abs/2209.09824](https://arxiv.org/abs/2209.09824)
- **Dataset:** Tweet Topic Dataset
- **Domain:** Twitter
- **Number of Class:** 6
### Dataset Summary
This is the official repository of TweetTopic (["Twitter Topic Classification
, COLING main conference 2022"](https://arxiv.org/abs/2209.09824)), a topic classification dataset on Twitter with 6 labels.
Each instance of TweetTopic comes with a timestamp which distributes from September 2019 to August 2021.
See [cardiffnlp/tweet_topic_multi](https://huggingface.co/datasets/cardiffnlp/tweet_topic_multi) for multi label version of TweetTopic.
The tweet collection used in TweetTopic is same as what used in [TweetNER7](https://huggingface.co/datasets/tner/tweetner7).
The dataset is integrated in [TweetNLP](https://tweetnlp.org/) too.
### Preprocessing
We pre-process tweets before the annotation to normalize some artifacts, converting URLs into a special token `{{URL}}` and non-verified usernames into `{{USERNAME}}`.
For verified usernames, we replace its display name (or account name) with symbols `{@}`.
For example, a tweet
```
Get the all-analog Classic Vinyl Edition
of "Takin' Off" Album from @herbiehancock
via @bluenoterecords link below:
http://bluenote.lnk.to/AlbumOfTheWeek
```
is transformed into the following text.
```
Get the all-analog Classic Vinyl Edition
of "Takin' Off" Album from {@herbiehancock@}
via {@bluenoterecords@} link below: {{URL}}
```
A simple function to format tweet follows below.
```python
import re
from urlextract import URLExtract
extractor = URLExtract()
def format_tweet(tweet):
# mask web urls
urls = extractor.find_urls(tweet)
for url in urls:
tweet = tweet.replace(url, "{{URL}}")
# format twitter account
tweet = re.sub(r"\b(\s*)(@[\S]+)\b", r'\1{\2@}', tweet)
return tweet
target = """Get the all-analog Classic Vinyl Edition of "Takin' Off" Album from @herbiehancock via @bluenoterecords link below: http://bluenote.lnk.to/AlbumOfTheWeek"""
target_format = format_tweet(target)
print(target_format)
'Get the all-analog Classic Vinyl Edition of "Takin\' Off" Album from {@herbiehancock@} via {@bluenoterecords@} link below: {{URL}}'
```
### Data Splits
| split | number of texts | description |
|:------------------------|-----:|------:|
| test_2020 | 376 | test dataset from September 2019 to August 2020 |
| test_2021 | 1693 | test dataset from September 2020 to August 2021 |
| train_2020 | 2858 | training dataset from September 2019 to August 2020 |
| train_2021 | 1516 | training dataset from September 2020 to August 2021 |
| train_all | 4374 | combined training dataset of `train_2020` and `train_2021` |
| validation_2020 | 352 | validation dataset from September 2019 to August 2020 |
| validation_2021 | 189 | validation dataset from September 2020 to August 2021 |
| train_random | 2830 | randomly sampled training dataset with the same size as `train_2020` from `train_all` |
| validation_random | 354 | randomly sampled training dataset with the same size as `validation_2020` from `validation_all` |
| test_coling2022_random | 3399 | random split used in the COLING 2022 paper |
| train_coling2022_random | 3598 | random split used in the COLING 2022 paper |
| test_coling2022 | 3399 | temporal split used in the COLING 2022 paper |
| train_coling2022 | 3598 | temporal split used in the COLING 2022 paper |
For the temporal-shift setting, model should be trained on `train_2020` with `validation_2020` and evaluate on `test_2021`.
In general, model would be trained on `train_all`, the most representative training set with `validation_2021` and evaluate on `test_2021`.
**IMPORTANT NOTE:** To get a result that is comparable with the results of the COLING 2022 Tweet Topic paper, please use `train_coling2022` and `test_coling2022` for temporal-shift, and `train_coling2022_random` and `test_coling2022_random` fir random split (the coling2022 split does not have validation set).
### Models
| model | training data | F1 | F1 (macro) | Accuracy |
|:------------------------------------------------------------------------------------------------------------------------------------------------------------|:------------------|---------:|-------------:|-----------:|
| [cardiffnlp/roberta-large-tweet-topic-single-all](https://huggingface.co/cardiffnlp/roberta-large-tweet-topic-single-all) | all (2020 + 2021) | 0.896043 | 0.800061 | 0.896043 |
| [cardiffnlp/roberta-base-tweet-topic-single-all](https://huggingface.co/cardiffnlp/roberta-base-tweet-topic-single-all) | all (2020 + 2021) | 0.887773 | 0.79793 | 0.887773 |
| [cardiffnlp/twitter-roberta-base-2019-90m-tweet-topic-single-all](https://huggingface.co/cardiffnlp/twitter-roberta-base-2019-90m-tweet-topic-single-all) | all (2020 + 2021) | 0.892499 | 0.774494 | 0.892499 |
| [cardiffnlp/twitter-roberta-base-dec2020-tweet-topic-single-all](https://huggingface.co/cardiffnlp/twitter-roberta-base-dec2020-tweet-topic-single-all) | all (2020 + 2021) | 0.890136 | 0.776025 | 0.890136 |
| [cardiffnlp/twitter-roberta-base-dec2021-tweet-topic-single-all](https://huggingface.co/cardiffnlp/twitter-roberta-base-dec2021-tweet-topic-single-all) | all (2020 + 2021) | 0.894861 | 0.800952 | 0.894861 |
| [cardiffnlp/roberta-large-tweet-topic-single-2020](https://huggingface.co/cardiffnlp/roberta-large-tweet-topic-single-2020) | 2020 only | 0.878913 | 0.70565 | 0.878913 |
| [cardiffnlp/roberta-base-tweet-topic-single-2020](https://huggingface.co/cardiffnlp/roberta-base-tweet-topic-single-2020) | 2020 only | 0.868281 | 0.729667 | 0.868281 |
| [cardiffnlp/twitter-roberta-base-2019-90m-tweet-topic-single-2020](https://huggingface.co/cardiffnlp/twitter-roberta-base-2019-90m-tweet-topic-single-2020) | 2020 only | 0.882457 | 0.740187 | 0.882457 |
| [cardiffnlp/twitter-roberta-base-dec2020-tweet-topic-single-2020](https://huggingface.co/cardiffnlp/twitter-roberta-base-dec2020-tweet-topic-single-2020) | 2020 only | 0.87596 | 0.746275 | 0.87596 |
| [cardiffnlp/twitter-roberta-base-dec2021-tweet-topic-single-2020](https://huggingface.co/cardiffnlp/twitter-roberta-base-dec2021-tweet-topic-single-2020) | 2020 only | 0.877732 | 0.746119 | 0.877732 |
Model fine-tuning script can be found [here](https://huggingface.co/datasets/cardiffnlp/tweet_topic_single/blob/main/lm_finetuning.py).
## Dataset Structure
### Data Instances
An example of `train` looks as follows.
```python
{
"text": "Game day for {{USERNAME}} U18\u2019s against {{USERNAME}} U18\u2019s. Even though it\u2019s a \u2018home\u2019 game for the people that have settled in Mid Wales it\u2019s still a 4 hour round trip for us up to Colwyn Bay. Still enjoy it though!",
"date": "2019-09-08",
"label": 4,
"id": "1170606779568463874",
"label_name": "sports_&_gaming"
}
```
### Label ID
The label2id dictionary can be found at [here](https://huggingface.co/datasets/tner/tweet_topic_single/raw/main/dataset/label.single.json).
```python
{
"arts_&_culture": 0,
"business_&_entrepreneurs": 1,
"pop_culture": 2,
"daily_life": 3,
"sports_&_gaming": 4,
"science_&_technology": 5
}
```
### Citation Information
```
@inproceedings{dimosthenis-etal-2022-twitter,
title = "{T}witter {T}opic {C}lassification",
author = "Antypas, Dimosthenis and
Ushio, Asahi and
Camacho-Collados, Jose and
Neves, Leonardo and
Silva, Vitor and
Barbieri, Francesco",
booktitle = "Proceedings of the 29th International Conference on Computational Linguistics",
month = oct,
year = "2022",
address = "Gyeongju, Republic of Korea",
publisher = "International Committee on Computational Linguistics"
}
``` | [
-0.41150397062301636,
-0.5397923588752747,
0.19379831850528717,
0.31822553277015686,
-0.39666521549224854,
0.013945408165454865,
-0.26671135425567627,
-0.5954558253288269,
0.6959244608879089,
0.07942402362823486,
-0.8865302801132202,
-0.7433744072914124,
-0.5327351093292236,
0.099146649241... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
LinhDuong/chatdoctor-5k | LinhDuong | 2023-03-28T07:32:21Z | 153 | 0 | null | [
"license:apache-2.0",
"arxiv:2303.14070",
"region:us"
] | 2023-03-28T07:32:21Z | 2023-03-28T07:23:57.000Z | 2023-03-28T07:23:57 | ---
license: apache-2.0
---
This ChatDoctor-5K dataset is collected from this paper https://arxiv.org/pdf/2303.14070.pdf
Alternatively, you can download the original dataset from this link https://drive.google.com/file/d/1nDTKZ3wZbZWTkFMBkxlamrzbNz0frugg/view?usp=sharing | [
-0.515437662601471,
-0.3415053188800812,
0.05897929519414902,
-0.17993639409542084,
-0.11729735136032104,
0.03404417634010315,
0.1897708922624588,
-0.27391883730888367,
0.2561528980731964,
0.8184594511985779,
-0.6972082853317261,
-0.43813708424568176,
-0.5306776165962219,
0.039448883384466... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
bigheiniuJ/EvalMetaICLAll | bigheiniuJ | 2023-07-24T06:39:16Z | 153 | 0 | null | [
"region:us"
] | 2023-07-24T06:39:16Z | 2023-07-23T20:34:43.000Z | 2023-07-23T20:34:43 | ---
dataset_info:
features:
- name: task
dtype: string
- name: input
dtype: string
- name: output
dtype: string
- name: options
sequence: string
- name: seed
dtype: string
- name: split
dtype: string
splits:
- name: meta_train
num_bytes: 648803062
num_examples: 1111614
- name: meta_eval_100shot
num_bytes: 1798838431
num_examples: 2725939
download_size: 1076308849
dataset_size: 2447641493
---
# Dataset Card for "EvalMetaICLAll"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.5287930965423584,
-0.5164644718170166,
0.127943754196167,
0.29996296763420105,
-0.1837684065103531,
0.3474554121494293,
0.2346542626619339,
-0.2214013636112213,
0.9856589436531067,
0.6218628883361816,
-0.7527875900268555,
-0.8235347867012024,
-0.5683114528656006,
-0.19690248370170593,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
RealTimeData/bbc_latest | RealTimeData | 2023-11-27T00:35:05Z | 153 | 0 | null | [
"region:us"
] | 2023-11-27T00:35:05Z | 2023-08-19T23:03:43.000Z | 2023-08-19T23:03:43 | ---
{}
---
# Latest BBC News
You could always access the latest BBC News articles via this dataset.
We update the dataset weekly, on every Sunday. So the dataset always provides the latest BBC News article from the last week.
The current dataset on main branch contains the latest BBC News articles submitted from 2023-11-20 to 2023-11-27.
The data collection is conducted on 2023-11-27.
Use the dataset via:
```
ds = datasets.load_dataset('RealTimeData/bbc_latest')
```
# Previsou versions
You could access previous versions by requesting different branches.
For example, you could find the 2023-08-20 version via:
```
ds = datasets.load_dataset('RealTimeData/bbc_latest', revision = '2023-08-20')
```
Check all available versions by clicking the "Files and versions" button on the top bar.
| [
-0.4674071669578552,
-0.5754712224006653,
0.4863301217556,
0.25712665915489197,
-0.6121233701705933,
0.21553201973438263,
-0.2681798040866852,
-0.6393352150917053,
0.5421441793441772,
0.7454524636268616,
-0.9205769300460815,
-0.4685693681240082,
-0.5543708801269531,
0.15941977500915527,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
jamescalam/agent-conversations-retrieval-tool | jamescalam | 2023-08-27T12:57:37Z | 153 | 7 | null | [
"region:us"
] | 2023-08-27T12:57:37Z | 2023-08-27T12:56:16.000Z | 2023-08-27T12:56:16 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
EleutherAI/sycophancy | EleutherAI | 2023-09-05T15:14:40Z | 153 | 1 | null | [
"region:us"
] | 2023-09-05T15:14:40Z | 2023-08-29T07:58:29.000Z | 2023-08-29T07:58:29 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
benjis/bigvul | benjis | 2023-08-31T03:02:50Z | 153 | 0 | null | [
"region:us"
] | 2023-08-31T03:02:50Z | 2023-08-31T02:55:32.000Z | 2023-08-31T02:55:32 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
- split: test
path: data/test-*
dataset_info:
features:
- name: CVE ID
dtype: string
- name: CVE Page
dtype: string
- name: CWE ID
dtype: string
- name: codeLink
dtype: string
- name: commit_id
dtype: string
- name: commit_message
dtype: string
- name: func_after
dtype: string
- name: func_before
dtype: string
- name: lang
dtype: string
- name: project
dtype: string
- name: vul
dtype: int8
splits:
- name: train
num_bytes: 404950685.2579571
num_examples: 150908
- name: validation
num_bytes: 88684597.21877055
num_examples: 33049
- name: test
num_bytes: 88687280.64632414
num_examples: 33050
download_size: 252969708
dataset_size: 582322563.1230518
---
# Dataset Card for "bigvul"
Unofficial, not affiliated with the authors.
- **Paper:** https://doi.org/10.1145/3379597.3387501
- **Repository:** https://github.com/ZeoVan/MSR_20_Code_vulnerability_CSV_Dataset | [
0.12298933416604996,
-0.43564391136169434,
0.101936936378479,
0.19315655529499054,
-0.4602789878845215,
-0.020963475108146667,
-0.2328115552663803,
-0.24503077566623688,
0.26893147826194763,
0.30113109946250916,
-0.35672104358673096,
-1.0613905191421509,
-0.6851311922073364,
-0.07864366471... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
legacy107/spamming-email-classification | legacy107 | 2023-10-02T09:39:55Z | 153 | 0 | null | [
"region:us"
] | 2023-10-02T09:39:55Z | 2023-09-25T14:22:14.000Z | 2023-09-25T14:22:14 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: val
path: data/val-*
- split: test
path: data/test-*
dataset_info:
features:
- name: Text
dtype: string
- name: Spam
dtype: int64
splits:
- name: train
num_bytes: 7196065
num_examples: 4556
- name: val
num_bytes: 819608
num_examples: 569
- name: test
num_bytes: 925859
num_examples: 570
download_size: 4959617
dataset_size: 8941532
---
# Dataset Card for "spamming-email-classification"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.5479223132133484,
-0.4092443585395813,
-0.08031372725963593,
0.35796186327934265,
0.10786420106887817,
-0.017265083268284798,
0.3324497640132904,
-0.17351588606834412,
0.4617313742637634,
0.7126885056495667,
-0.6400539875030518,
-0.6551244258880615,
-0.9767588973045349,
-0.2762328684329... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
LDJnr/Pure-Dove | LDJnr | 2023-11-21T17:55:19Z | 153 | 21 | null | [
"task_categories:conversational",
"task_categories:question-answering",
"task_categories:text-generation",
"size_categories:1K<n<10K",
"language:en",
"license:apache-2.0",
"Physics",
"Biology",
"Math",
"Chemistry",
"Culture",
"Logic",
"Roleplay",
"region:us"
] | 2023-11-21T17:55:19Z | 2023-09-26T02:06:24.000Z | 2023-09-26T02:06:24 | ---
license: apache-2.0
task_categories:
- conversational
- question-answering
- text-generation
language:
- en
tags:
- Physics
- Biology
- Math
- Chemistry
- Culture
- Logic
- Roleplay
pretty_name: Pure-Dove
size_categories:
- 1K<n<10K
---
## This is the Official Pure-Dove dataset. Over 3K multi-turn examples, and many more coming soon!
This dataset aims to be the largest highest quality cluster of real human back and forth conversations with GPT-4.
Steps have even been done to ensure that only the best GPT-4 conversations in comparisons are kept, there are many instances where two GPT-4 responses are rated as equal to eachother or as both bad. We exclude all such responses from Pure Dove and make sure to only include ChatBot Arena responses that are voted as being better even against another instance of GPT-4.
- Comprised of over 3000 highly filtered multi-turn conversations between GPT-4 and real humans.
- Average context length per conversation is over 800 tokens.
## Purpose?
- This dataset is not particularly intended to be trained on by itself, however, the size and quality of this dataset can work wonderfully as a supplemmentary addition to virtually any multi-turn compatible dataset. I encourage this use, all I ask is proper credits given for such!
## Quality filtering and cleaning.
- The conversations were sourced from openly datasets such as ShareGPT and ChatBotArena by Lmsys, however, a large portion of these chats were riddled with hallucinations and abnormal distributions of different languages.
- Extensive cleaning was done to filter out instances of overt AI moralizing or related behaviour, such as "As an AI language model" and "September 2021", not just in english, but other languages too!
## Credits
During the curation process, there can be some relatively arduos steps when it comes to actually executing on the best experimentation or concepts for how to filter examples out.
Luckily there is folks over at NousResearch that helped expedite this process with little to no sacrifices in quality, big credit to J-Supha within NousResearch specifically for making these types of significant contributions.
## Future Plans & How you can help!
This is a relatively early build amongst the grand plans for the future of what I plan to work on!
In the near future we plan on leveraging the help of domain specific expert volunteers to eliminate any mathematically/verifiably incorrect answers from training curations of different types of datasets.
If you have at-least a bachelors in mathematics, physics, biology or chemistry and would like to volunteer even just 30 minutes of your expertise time, please contact LDJ on discord!
Citation:
```
@article{daniele2023amplify-instruct,
title={Amplify-Instruct: Synthetically Generated Diverse Multi-turn Conversations for Effecient LLM Training.},
author={Daniele, Luigi and Suphavadeeprasit},
journal={arXiv preprint arXiv:(comming soon)},
year={2023}
}
``` | [
-0.4659762978553772,
-0.82603520154953,
0.4255504310131073,
0.17199984192848206,
-0.15100425481796265,
0.035484857857227325,
-0.17246367037296295,
-0.44513750076293945,
0.16209010779857635,
0.5178656578063965,
-0.7029075026512146,
-0.4168851673603058,
-0.4708016812801361,
0.071703262627124... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
maxolotl/must-c-en-es-wait3-02 | maxolotl | 2023-10-22T07:48:24Z | 153 | 0 | null | [
"region:us"
] | 2023-10-22T07:48:24Z | 2023-10-22T07:48:05.000Z | 2023-10-22T07:48:05 | ---
dataset_info:
features:
- name: current_source
dtype: string
- name: current_target
dtype: string
- name: target_token
dtype: string
splits:
- name: train
num_bytes: 995120593
num_examples: 5240243
- name: test
num_bytes: 9960448
num_examples: 57187
- name: validation
num_bytes: 5429701
num_examples: 27549
download_size: 184348036
dataset_size: 1010510742
---
# Dataset Card for "must-c-en-es-wait3-02"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.6564971208572388,
-0.11159498989582062,
0.4751039445400238,
0.7593249082565308,
-0.10837173461914062,
-0.11256479471921921,
0.3184361457824707,
-0.4941241443157196,
0.7914745807647705,
0.6035370826721191,
-1.1325933933258057,
-0.5855128765106201,
-0.6314719319343567,
0.19302424788475037... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Karavet/pioNER-Armenian-Named-Entity | Karavet | 2022-10-21T16:07:06Z | 152 | 1 | null | [
"task_ids:named-entity-recognition",
"multilinguality:monolingual",
"language:hy",
"license:apache-2.0",
"region:us"
] | 2022-10-21T16:07:06Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | ---
language: [hy]
task_categories: [named-entity-recognition]
multilinguality: [monolingual]
task_ids: [named-entity-recognition]
license: [apache-2.0]
---
## Table of Contents
- [Table of Contents](#table-of-contents)
- [pioNER - named entity annotated datasets](#pioNER---named-entity-annotated-datasets)
- [Silver-standard dataset](#silver-standard-dataset)
- [Gold-standard dataset](#gold-standard-dataset)
# pioNER - named entity annotated datasets
pioNER corpus provides gold-standard and automatically generated named-entity datasets for the Armenian language.
Alongside the datasets, we release 50-, 100-, 200-, and 300-dimensional GloVe word embeddings trained on a collection of Armenian texts from Wikipedia, news, blogs, and encyclopedia.
## Silver-standard dataset
The generated corpus is automatically extracted and annotated using Armenian Wikipedia. We used a modification of [Nothman et al](https://www.researchgate.net/publication/256660013_Learning_multilingual_named_entity_recognition_from_Wikipedia) and [Sysoev and Andrianov](http://www.dialog-21.ru/media/3433/sysoevaaandrianovia.pdf) approaches to create this corpus. This approach uses links between Wikipedia articles to extract fragments of named-entity annotated texts.
The corpus is split into train and development sets.
*Table 1. Statistics for pioNER train, development and test sets*
| dataset | #tokens | #sents | annotation | texts' source |
|-------------|:--------:|:-----:|:--------:|:-----:|
| train | 130719 | 5964 | automatic | Wikipedia |
| dev | 32528 | 1491 | automatic | Wikipedia |
| test | 53606 | 2529 | manual | iLur.am |
## Gold-standard dataset
This dataset is a collection of over 250 news articles from iLur.am with manual named-entity annotation. It includes sentences from political, sports, local and world news, and is comparable in size with the test sets of other languages (Table 2).
We aim it to serve as a benchmark for future named entity recognition systems designed for the Armenian language.
The dataset contains annotations for 3 popular named entity classes:
people (PER), organizations (ORG), and locations (LOC), and is released in CoNLL03 format with IOB tagging scheme.
During annotation, we generally relied on categories and [guidelines assembled by BBN](https://catalog.ldc.upenn.edu/docs/LDC2005T33/BBN-Types-Subtypes.html) Technologies for TREC 2002 question answering track
Tokens and sentences were segmented according to the UD standards for the Armenian language from [ArmTreebank project](http://armtreebank.yerevann.com/tokenization/process/).
*Table 2. Comparison of pioNER gold-standard test set with test sets for English, Russian, Spanish and German*
| test dataset | #tokens | #LOC | #ORG | #PER |
|-------------|:--------:|:-----:|:--------:|:-----:|
| Armenian pioNER | 53606 | 1312 | 1338 | 1274 |
| Russian factRuEval-2016 | 59382 | 1239 | 1595 | 1353 |
| German CoNLL03 | 51943 | 1035 | 773 | 1195 |
| Spanish CoNLL02 | 51533 | 1084 | 1400 | 735 |
| English CoNLL03 | 46453 | 1668 | 1661 | 1671 | | [
-0.8337982296943665,
-0.6106128096580505,
0.0016679394757375121,
-0.024099867790937424,
-0.41270124912261963,
-0.18098530173301697,
-0.5245099067687988,
-0.5758296251296997,
0.3895459771156311,
0.6980950832366943,
-0.25429248809814453,
-0.6129705309867859,
-0.6595730781555176,
0.5282713174... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Niciu/test-cre-dataset-issues | Niciu | 2022-03-01T14:06:43Z | 152 | 0 | null | [
"region:us"
] | 2022-03-01T14:06:43Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
WyrdCurt/AO4W | WyrdCurt | 2021-07-26T12:03:27Z | 152 | 0 | null | [
"region:us"
] | 2021-07-26T12:03:27Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | # Archive Of Our Own Original Works (AO4W)
**Warning! Many/most of these files may be NSFW!**
Approximately 2GB of text files from Archive of Our Own; specifically, files labeled "original work" or some variation. For training fiction models. I recommend that you clean the text as needed for your purposes. | [
-0.27497172355651855,
-0.8118712902069092,
0.7034948468208313,
0.07102394849061966,
-0.32298487424850464,
-0.4371403753757477,
0.533312201499939,
-0.7302860021591187,
0.12044892460107803,
1.4352080821990967,
-0.5271919369697571,
-0.22597596049308777,
-0.7737706899642944,
0.4774087369441986... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
abidlabs/crowdsourced-speech3 | abidlabs | 2022-01-21T16:12:06Z | 152 | 0 | null | [
"region:us"
] | 2022-01-21T16:12:06Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
abwicke/koplo | abwicke | 2021-03-18T15:43:39Z | 152 | 0 | null | [
"region:us"
] | 2021-03-18T15:43:39Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | https://jobs.shrm.org/profile/free-watch-full-raya-and-the-last-dragon-2021/1589725/
https://jobs.shrm.org/profile/full-watch-raya-and-the-last-dragon-2021-hd-online-full-free-123movies/1589732/
https://jobs.shrm.org/profile/123movies-watch-raya-and-the-last-dragon-2021-hd-online-full-free-streaming/1589735/
https://jobs.shrm.org/profile/full-watch-zack-snyders-justice-league-2021-full-free/1591000/
https://jobs.shrm.org/profile/online-watch-zack-snyders-justice-league-2021-123movies-full-version-/1591028/
https://jobs.shrm.org/profile/watch-zack-snyders-justice-league-2021-online-movie-full-version-hd/1591260/
https://jobs.shrm.org/profile/full-watch-zack-snyders-justice-league-2021-hd-online-full-free-123movies/1591268/
https://jobs.shrm.org/profile/watch-zack-snyders-justice-league-2021-full-free/1591274/
https://jobs.shrm.org/profile/watch-zack-snyders-justice-league-2021-full-123movies/1591294/
https://jobs.shrm.org/profile/123movies-watch-zack-snyders-justice-league-online-2021-full-free-hd/1591301/
https://jobs.aapor.org/profile/full-watch-365-days-2020-hd-online-full-free-123movies/1592853/
https://jobs.aapor.org/profile/full-watch-army-of-the-dead-2021-hd-online-full-free-123movies/1592863/
https://jobs.aapor.org/profile/full-watch-barb-and-star-go-to-vista-del-mar-2021-hd-online-full-free-123movies/1592894/
https://jobs.aapor.org/profile/full-watch-billie-eilish-the-worlds-a-little-blurry-2021-hd-online-full-free-123movies/1592902/
https://jobs.aapor.org/profile/full-watch-black-widow-2021-hd-online-full-free-123movies/1592920/
https://jobs.aapor.org/profile/full-watch-bliss-2021-hd-online-full-free-123movies/1592926/
https://jobs.aapor.org/profile/full-watch-borat-subsequent-moviefilm-2020-hd-online-full-free-123movies/1592939/
https://jobs.aapor.org/profile/full-watch-boss-level-2021-hd-online-full-free-123movies/1592952/
https://jobs.aapor.org/profile/full-watch-chaos-walking-2021-hd-online-full-free-123movies/1592958/
https://jobs.aapor.org/profile/full-watch-cherry-2021-hd-online-full-free-123movies/1592984/
https://jobs.aapor.org/profile/full-watch-coming-2-america-2021-hd-online-full-free-123movies/1592999/
https://jobs.aapor.org/profile/full-watch-cosmic-sin-2021-hd-online-full-free-123movies/1593012/
https://jobs.aapor.org/profile/full-watch-crisis-2021-hd-online-full-free-123movies/1593017/
https://jobs.aapor.org/profile/full-watch-demon-slayer-kimetsu-no-yaiba-mugen-train-2020-hd-online-full-free-123movies/1593028/
https://jobs.aapor.org/profile/full-watch-eternals-2021-hd-online-full-free-123movies/1593038/
https://jobs.aapor.org/profile/full-watch-flora-ulysses-2021-hd-online-full-free-123movies/1593050/
https://jobs.aapor.org/profile/full-watch-girl-in-the-basement-2021-hd-online-full-free-123movies/1593056/
https://jobs.aapor.org/profile/full-watch-godzilla-vs-kong-2021-hd-online-full-free-123movies/1593071/
https://jobs.aapor.org/profile/full-watch-greenland-2020-hd-online-full-free-123movies/1593089/
https://jobs.aapor.org/profile/full-watch-i-care-a-lot-2021-hd-online-full-free-123movies/1593100/
https://jobs.aapor.org/profile/full-watch-judas-and-the-black-messiah-2021-hd-online-full-free-123movies/1593112/
https://jobs.aapor.org/profile/full-watch-land-2021-hd-online-full-free-123movies/1593128/
https://jobs.aapor.org/profile/full-watch-to-all-the-boys-always-and-forever-2021-hd-online-full-free-123movies/1592831/
https://jobs.shrm.org/profile/full-watch-zack-snyders-justice-league-2021-hd-online-movie-free-123movies/1592215/
https://jobs.shrm.org/profile/full-watch-coming-2-america-2021-hd-movie-online-free-123movies/1592252/
https://jobs.shrm.org/profile/full-watch-raya-and-the-last-dragon-2021-hd-movie-online-free-123movies/1592256/
https://jobs.shrm.org/profile/full-watch-demon-slayer-kimetsu-no-yaiba-the-movie-mugen-train-2020-hd-movie-online-free-123movies/1592270/
https://jobs.shrm.org/profile/hbo-max-watch-zack-snyders-justice-league-2021-hd-online-version/1592272/
https://jobs.aapor.org/profile/full-episodes-the-falcon-and-the-winter-soldier-season-1-episode-1-e1-online/1592374/
https://jobs.aapor.org/profile/watch-tom-jerry-2021-online-full-movie-free-123movies/1592338/
https://jobs.aapor.org/profile/hbo-max-watch-zack-snyders-justice-league-2021-hd-online-version/1592290/
https://jobs.aapor.org/profile/watch-zack-snyders-justice-league-2021-online-full-movie-hd-free/1592288/
https://jobs.aapor.org/profile/downloadhdzack-snyders-justice-league-2021-1080p-full-online-version/1592297/
https://jobs.aapor.org/profile/watch-coming-2-america-2021-online-full-movie-free-123movies/1592307/
https://jobs.aapor.org/profile/watch-zack-snyders-justice-league-2021-full-free-online-movie-here-hbomax/1592312/
https://jobs.aapor.org/profile/watch-raya-and-the-last-dragon-2021-online-full-movie-free-123movies/1592326/
https://jobs.aapor.org/profile/full-watch-tom-and-jerry-2021-full-movie-online-free-hd-stream/1592328/
https://jobs.aapor.org/profile/watch-billie-eilish-the-worlds-a-little-blurry-2021-full/1592344/
https://jobs.aapor.org/profile/watch-billie-eilish-the-worlds-a-little-blurry-2021-online-full-movie-free-123movies/1592353/
https://jobs.aapor.org/profile/watch-demon-slayer-kimetsu-no-yaiba-the-movie-mugen-train-2020-online-full-movie-free-123movies/1592363/
https://jobs.aapor.org/profile/full-watch-demon-slayer-kimetsu-no-yaiba-mugen-train-2020-full-free/1592355/
https://jobs.aapor.org/profile/watch-the-falcon-and-the-winter-soldier-season-1-e1-2021-online-free/1592389/ | [
-0.4497397840023041,
-0.047093573957681656,
0.301297664642334,
0.046867504715919495,
-0.5837980508804321,
0.09756352007389069,
0.6407689452171326,
-0.6415932178497314,
0.44099298119544983,
0.15395814180374146,
-0.5891977548599243,
-0.055207084864377975,
-0.30397966504096985,
0.031818091869... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
albertvillanova/tmp-tests | albertvillanova | 2021-12-02T14:12:12Z | 152 | 0 | null | [
"region:us"
] | 2021-12-02T14:12:12Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.32276496291160583,
-0.22568435966968536,
0.8622260093688965,
0.43461480736732483,
-0.5282987952232361,
0.7012965083122253,
0.7915714979171753,
0.07618625462055206,
0.7746025323867798,
0.25632181763648987,
-0.7852815389633179,
-0.22573819756507874,
-0.9104480743408203,
0.5715669393539429... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
aliabd/crowdsourced-speech4 | aliabd | 2022-01-21T17:36:51Z | 152 | 0 | null | [
"region:us"
] | 2022-01-21T17:36:51Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.32276496291160583,
-0.22568435966968536,
0.8622260093688965,
0.43461480736732483,
-0.5282987952232361,
0.7012965083122253,
0.7915714979171753,
0.07618625462055206,
0.7746025323867798,
0.25632181763648987,
-0.7852815389633179,
-0.22573819756507874,
-0.9104480743408203,
0.5715669393539429... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
alireza655/alireza655 | alireza655 | 2021-02-08T23:24:50Z | 152 | 0 | null | [
"region:us"
] | 2021-02-08T23:24:50Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.32276496291160583,
-0.22568435966968536,
0.8622260093688965,
0.43461480736732483,
-0.5282987952232361,
0.7012965083122253,
0.7915714979171753,
0.07618625462055206,
0.7746025323867798,
0.25632181763648987,
-0.7852815389633179,
-0.22573819756507874,
-0.9104480743408203,
0.5715669393539429... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
clem/autonlp-data-french_word_detection | clem | 2021-09-14T09:45:38Z | 152 | 1 | null | [
"region:us"
] | 2021-09-14T09:45:38Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | This is a very good dataset! | [
-0.2994808256626129,
-0.058056823909282684,
0.21844859421253204,
0.5398688912391663,
-0.13290078938007355,
-0.1372750997543335,
-0.031124968081712723,
-0.07468447834253311,
0.49724268913269043,
0.7139886617660522,
-0.2566799223423004,
-0.7480513453483582,
-0.5223762392997742,
-0.3120886683... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
cnrcastroli/aaaa | cnrcastroli | 2021-03-04T21:51:21Z | 152 | 0 | null | [
"region:us"
] | 2021-03-04T21:51:21Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | fwefwefewf | [
-0.25071993470191956,
-0.5934706330299377,
0.34227949380874634,
1.0664750337600708,
-0.4548211097717285,
-0.0020777417812496424,
0.21651989221572876,
-0.7492340207099915,
0.5542600154876709,
0.812045693397522,
-0.859272301197052,
0.08489752560853958,
-1.113316535949707,
0.29075518250465393... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
coala/kkk | coala | 2021-09-14T07:56:22Z | 152 | 0 | null | [
"region:us"
] | 2021-09-14T07:56:22Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.32276496291160583,
-0.22568435966968536,
0.8622260093688965,
0.43461480736732483,
-0.5282987952232361,
0.7012965083122253,
0.7915714979171753,
0.07618625462055206,
0.7746025323867798,
0.25632181763648987,
-0.7852815389633179,
-0.22573819756507874,
-0.9104480743408203,
0.5715669393539429... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
congpt/dstc23_asr | congpt | 2021-04-06T18:04:04Z | 152 | 0 | null | [
"region:us"
] | 2021-04-06T18:04:04Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.32276496291160583,
-0.22568435966968536,
0.8622260093688965,
0.43461480736732483,
-0.5282987952232361,
0.7012965083122253,
0.7915714979171753,
0.07618625462055206,
0.7746025323867798,
0.25632181763648987,
-0.7852815389633179,
-0.22573819756507874,
-0.9104480743408203,
0.5715669393539429... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
crich/cider | crich | 2021-11-15T18:26:56Z | 152 | 0 | null | [
"region:us"
] | 2021-11-15T18:26:56Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.32276496291160583,
-0.22568435966968536,
0.8622260093688965,
0.43461480736732483,
-0.5282987952232361,
0.7012965083122253,
0.7915714979171753,
0.07618625462055206,
0.7746025323867798,
0.25632181763648987,
-0.7852815389633179,
-0.22573819756507874,
-0.9104480743408203,
0.5715669393539429... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
csarron/image-captions | csarron | 2021-11-29T04:31:34Z | 152 | 0 | null | [
"region:us"
] | 2021-11-29T04:31:34Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
ctu-aic/multinli_cs | ctu-aic | 2021-11-21T19:58:09Z | 152 | 0 | null | [
"region:us"
] | 2021-11-21T19:58:09Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
dasago78/dasago78dataset | dasago78 | 2021-04-02T17:57:24Z | 152 | 0 | null | [
"region:us"
] | 2021-04-02T17:57:24Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
davanstrien/hipe2020 | davanstrien | 2022-02-15T11:40:24Z | 152 | 0 | null | [
"region:us"
] | 2022-02-15T11:40:24Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
davanstrien/iiif_labeled | davanstrien | 2022-02-28T11:06:07Z | 152 | 0 | null | [
"region:us"
] | 2022-02-28T11:06:07Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622264862060547,
0.43461528420448303,
-0.52829909324646,
0.7012971639633179,
0.7915720343589783,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104477167129517,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
dongpil/test | dongpil | 2021-07-29T10:34:34Z | 152 | 0 | null | [
"region:us"
] | 2021-07-29T10:34:34Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622264862060547,
0.43461528420448303,
-0.52829909324646,
0.7012971639633179,
0.7915720343589783,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104477167129517,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
eason929/test | eason929 | 2021-03-15T04:02:59Z | 152 | 0 | null | [
"region:us"
] | 2021-03-15T04:02:59Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622264862060547,
0.43461528420448303,
-0.52829909324646,
0.7012971639633179,
0.7915720343589783,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104477167129517,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
ehcalabres/ravdess_speech | ehcalabres | 2022-10-24T15:51:41Z | 152 | 3 | null | [
"task_categories:audio-classification",
"annotations_creators:no-annotation",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:1K<n<10K",
"source_datasets:original",
"language:en",
"license:cc-by-nc-sa-4.0",
"region:us"
] | 2022-10-24T15:51:41Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | ---
annotations_creators:
- no-annotation
language_creators:
- found
language:
- en
license:
- cc-by-nc-sa-4.0
multilinguality:
- monolingual
size_categories:
- 1K<n<10K
source_datasets:
- original
task_categories:
- audio-classification
task_ids:
- speech-emotion-recognition
---
# Dataset Card for ravdess_speech
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-instances)
- [Data Splits](#data-instances)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
## Dataset Description
- **Homepage:** [Needs More Information]
- **Repository:** https://zenodo.org/record/1188976#.YUS4MrozZdS
- **Paper:** https://doi.org/10.1371/journal.pone.0196391
- **Leaderboard:** [Needs More Information]
- **Point of Contact:** ravdess@gmail.com
### Dataset Summary
The Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS) contains 24 professional actors (12 female, 12 male), vocalizing two lexically-matched statements in a neutral North American accent. Speech includes calm, happy, sad, angry, fearful, surprise, and disgust expressions. Each expression is produced at two levels of emotional intensity (normal, strong), with an additional neutral expression. The conditions of the audio files are: 16bit, 48kHz .wav.
### Supported Tasks and Leaderboards
- audio-classification: The dataset can be used to train a model for Audio Classification tasks, which consists in predict the latent emotion presented on the audios.
### Languages
The audios available in the dataset are in English spoken by actors in a neutral North American accent.
## Dataset Structure
### Data Instances
[Needs More Information]
### Data Fields
[Needs More Information]
### Data Splits
[Needs More Information]
## Dataset Creation
### Curation Rationale
[Needs More Information]
### Source Data
#### Initial Data Collection and Normalization
[Needs More Information]
#### Who are the source language producers?
[Needs More Information]
### Annotations
#### Annotation process
[Needs More Information]
#### Who are the annotators?
[Needs More Information]
### Personal and Sensitive Information
[Needs More Information]
## Considerations for Using the Data
### Social Impact of Dataset
[Needs More Information]
### Discussion of Biases
[Needs More Information]
### Other Known Limitations
[Needs More Information]
## Additional Information
### Dataset Curators
[Needs More Information]
### Licensing Information
The RAVDESS is released under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License, CC BY-NC-SA 4.0
Commercial licenses for the RAVDESS can also be purchased. For more information, please visit our license fee page, or contact us at ravdess@gmail.com.
### Citation Information
Livingstone SR, Russo FA (2018) The Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS): A dynamic, multimodal set of facial and vocal expressions in North American English. PLoS ONE 13(5): e0196391. https://doi.org/10.1371/journal.pone.0196391. | [
-0.43191179633140564,
-0.7637857794761658,
-0.1252378225326538,
0.4789009094238281,
-0.06024398282170296,
0.16646696627140045,
-0.4492100775241852,
-0.5712066292762756,
0.5647458434104919,
0.5642333030700684,
-0.9924535155296326,
-1.0414401292800903,
-0.6020432710647583,
-0.045269902795553... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Norod78/simpsons-blip-captions | Norod78 | 2022-11-09T16:27:19Z | 152 | 3 | null | [
"task_categories:text-to-image",
"annotations_creators:machine-generated",
"language_creators:other",
"multilinguality:monolingual",
"size_categories:n<1K",
"language:en",
"license:cc-by-nc-sa-4.0",
"region:us"
] | 2022-11-09T16:27:19Z | 2022-11-06T11:11:36.000Z | 2022-11-06T11:11:36 | ---
dataset_info:
features:
- name: image
dtype: image
- name: text
dtype: string
splits:
- name: train
num_bytes: 51605730.0
num_examples: 755
download_size: 50553165
dataset_size: 51605730.0
pretty_name: 'Simpsons BLIP captions'
size_categories:
- n<1K
tags: []
task_categories:
- text-to-image
license: cc-by-nc-sa-4.0
annotations_creators:
- machine-generated
language:
- en
language_creators:
- other
multilinguality:
- monolingual
---
# Dataset Card for "simpsons-blip-captions"
| [
-0.2915683090686798,
-0.0044026849791407585,
-0.2134803980588913,
0.5665171146392822,
-0.6229936480522156,
0.39689597487449646,
-0.2998969852924347,
0.36532121896743774,
0.4522024393081665,
0.6106967329978943,
-0.5644146203994751,
-0.5934778451919556,
-0.5761433243751526,
0.114859238266944... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Jzuluaga/uwb_atcc | Jzuluaga | 2022-12-05T11:15:20Z | 152 | 0 | null | [
"task_categories:automatic-speech-recognition",
"multilinguality:monolingual",
"language:en",
"license:cc-by-nc-sa-4.0",
"audio",
"automatic-speech-recognition",
"en-atc",
"en",
"noisy-speech-recognition",
"speech-recognition",
"arxiv:2203.16822",
"region:us"
] | 2022-12-05T11:15:20Z | 2022-11-28T07:12:02.000Z | 2022-11-28T07:12:02 | ---
dataset_info:
features:
- name: id
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 16000
- name: text
dtype: string
- name: segment_start_time
dtype: float32
- name: segment_end_time
dtype: float32
- name: duration
dtype: float32
splits:
- name: test
num_bytes: 140620332.25
num_examples: 2822
- name: train
num_bytes: 608597323.625
num_examples: 11291
download_size: 711464914
dataset_size: 749217655.875
tags:
- audio
- automatic-speech-recognition
- en-atc
- en
- noisy-speech-recognition
- speech-recognition
task_categories:
- automatic-speech-recognition
language:
- en
multilinguality:
- monolingual
license:
- cc-by-nc-sa-4.0
---
# Dataset Card for UWB-ATCC corpus
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages and Other Details](#languages-and-other-details)
- [Dataset Structure](#dataset-structure)
- [Data Fields](#data-fields)
- [Additional Information](#additional-information)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
## Dataset Description
- **Homepage:** [UWB-ATCC corpus homepage](https://lindat.mff.cuni.cz/repository/xmlui/handle/11858/00-097C-0000-0001-CCA1-0)
- **Repository:** [GitHub repository (used in research)](https://github.com/idiap/w2v2-air-traffic)
- **Paper:** [Air traffic control communication (ATCC) speech corpora and their use for ASR and TTS development](https://link.springer.com/article/10.1007/s10579-019-09449-5)
- **Paper of this research:** [How Does Pre-trained Wav2Vec 2.0 Perform on Domain Shifted ASR? An Extensive Benchmark on Air Traffic Control Communications](https://arxiv.org/abs/2203.16822)
### Dataset Summary
The UWB-ATCC Corpus is provided provided by University of West Bohemia, Department of Cybernetics. The corpus contains recordings of communication between air traffic controllers and pilots. The speech is manually transcribed and labeled with the information about the speaker (pilot/controller, not the full identity of the person). The corpus is currently small (20 hours) but we plan to search for additional data next year. The audio data format is: 8kHz, 16bit PCM, mono.
Important, from the `<id (string)>` field, you can obtain the speaker roles. For instance:
- `_PI`: segment with only pilot speech
- `_AT`: segment with only ATCO speech
- `PIAT`: segment with both, ATCO and pilot speech
### Supported Tasks and Leaderboards
- `automatic-speech-recognition`. Already adapted/fine-tuned models are available here --> [XLS-R-300m](https://huggingface.co/Jzuluaga/wav2vec2-large-960h-lv60-self-en-atc-atcosim).
### Languages and other details
The text and the recordings are in English. The authors took advantage of the fact that one of their industrial partners develops complex IT solutions for several ATC authorities and airports and, as such, has access to the ATC communication recordings collected in the Czech airspace. This partner was able to secure the following data:
- Ground control—communication before takeoff and after landing—19.2 h of data.
- Tower control—communication during takeoff, landing and landing standby—22.5 h.
- Approach control—communication during landing approach—25.5 h.
- Area control—communication during overflights and cruises—71.3 h.
(Not all data is released. Check their website [here](https://lindat.mff.cuni.cz/repository/xmlui/handle/11858/00-097C-0000-0001-CCA1-0))
## Dataset Structure
### Data Fields
- `id (string)`: a string of recording identifier for each example, corresponding to its.
- `audio (audio)`: audio data for the given ID
- `text (string)`: transcript of the file already normalized. Follow these repositories for more details [w2v2-air-traffic](https://github.com/idiap/w2v2-air-traffic) and [bert-text-diarization-atc](https://github.com/idiap/bert-text-diarization-atc)
- `segment_start_time (float32)`: segment start time (normally 0)
- `segment_end_time (float32): segment end time
- `duration (float32)`: duration of the recording, compute as segment_end_time - segment_start_time
## Additional Information
### Licensing Information
The licensing status of the dataset hinges on the legal status of the [UWB-ATCC corpus](https://lindat.mff.cuni.cz/repository/xmlui/handle/11858/00-097C-0000-0001-CCA1-0) creators.
They used [Creative Commons - Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0)](https://creativecommons.org/licenses/by-nc-sa/4.0/) licensing.
### Citation Information
Contributors who prepared, processed, normalized and uploaded the dataset in HuggingFace:
```
@article{zuluaga2022how,
title={How Does Pre-trained Wav2Vec2. 0 Perform on Domain Shifted ASR? An Extensive Benchmark on Air Traffic Control Communications},
author={Zuluaga-Gomez, Juan and Prasad, Amrutha and Nigmatulina, Iuliia and Sarfjoo, Saeed and others},
journal={IEEE Spoken Language Technology Workshop (SLT), Doha, Qatar},
year={2022}
}
@article{zuluaga2022bertraffic,
title={BERTraffic: BERT-based Joint Speaker Role and Speaker Change Detection for Air Traffic Control Communications},
author={Zuluaga-Gomez, Juan and Sarfjoo, Seyyed Saeed and Prasad, Amrutha and others},
journal={IEEE Spoken Language Technology Workshop (SLT), Doha, Qatar},
year={2022}
}
@article{zuluaga2022atco2,
title={ATCO2 corpus: A Large-Scale Dataset for Research on Automatic Speech Recognition and Natural Language Understanding of Air Traffic Control Communications},
author={Zuluaga-Gomez, Juan and Vesel{\`y}, Karel and Sz{\"o}ke, Igor and Motlicek, Petr and others},
journal={arXiv preprint arXiv:2211.04054},
year={2022}
}
```
Authors of the dataset:
```
@article{vsmidl2019air,
title={Air traffic control communication (ATCC) speech corpora and their use for ASR and TTS development},
author={{\v{S}}m{\'\i}dl, Lubo{\v{s}} and {\v{S}}vec, Jan and Tihelka, Daniel and Matou{\v{s}}ek, Jind{\v{r}}ich and Romportl, Jan and Ircing, Pavel},
journal={Language Resources and Evaluation},
volume={53},
number={3},
pages={449--464},
year={2019},
publisher={Springer}
}
```
| [
-0.39164409041404724,
-0.5750758051872253,
0.05795922875404358,
0.18492422997951508,
-0.40153899788856506,
0.200161412358284,
-0.5382319688796997,
-0.6016563773155212,
0.12122086435556412,
0.39749327301979065,
-0.45469599962234497,
-0.7072417140007019,
-0.5663379430770874,
-0.1574173867702... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
12ml/e-CARE | 12ml | 2023-01-06T18:50:03Z | 152 | 1 | null | [
"task_categories:multiple-choice",
"region:us"
] | 2023-01-06T18:50:03Z | 2022-12-21T11:38:01.000Z | 2022-12-21T11:38:01 | ---
task_categories:
- multiple-choice
---
# Dataset of (Du et al., 2022)
## Abstract
>Understanding causality has vital importance for various Natural Language Processing (NLP) applications. Beyond the labeled instances, conceptual explanations of the causality can provide deep understanding of the causal fact to facilitate the causal reasoning process. However, such explanation information still remains absent in existing causal reasoning resources. In this paper, we fill this gap by presenting a human-annotated explainable CAusal REasoning dataset (e-CARE), which contains over 20K causal reasoning questions, together with natural language formed explanations of the causal questions. Experimental results show that generating valid explanations for causal facts still remains especially challenging for the state-of-the-art models, and the explanation information can be helpful for promoting the accuracy and stability of causal reasoning models.
## Notes
Please note that the original dataset has been modified so that the variable names match with those in the COPA dataset (Roemmele et al., 2011). In addition, only the training and the development sets are [publicly available](https://github.com/waste-wood/e-care).
## References
Du, L., Ding, X., Xiong, K., Liu, T., & Qin, B. (2022). e-CARE: a New Dataset for Exploring Explainable Causal Reasoning. arXiv preprint arXiv:2205.05849.
Roemmele, M., Bejan, C., and Gordon, A. (2011) Choice of Plausible Alternatives: An Evaluation of Commonsense Causal Reasoning. AAAI Spring Symposium on Logical Formalizations of Commonsense Reasoning, Stanford University, March 21-23, 2011. | [
-0.1367652863264084,
-0.7371608018875122,
0.6847586631774902,
0.03632257506251335,
-0.2050774097442627,
-0.7123168706893921,
-0.006128290202468634,
-0.5719534754753113,
0.041625797748565674,
0.2689487934112549,
-0.9064939618110657,
-0.4440196752548218,
-0.4125695526599884,
0.39446032047271... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
heegyu/news-category-balanced-top10 | heegyu | 2023-02-13T02:56:31Z | 152 | 1 | null | [
"license:cc-by-4.0",
"region:us"
] | 2023-02-13T02:56:31Z | 2023-02-13T02:45:28.000Z | 2023-02-13T02:45:28 | ---
license: cc-by-4.0
---
### Top10 sampled news category dataset
randomly sampled news data
original dataset: https://www.kaggle.com/datasets/rmisra/news-category-dataset
### Value Counts per Category
```
ENTERTAINMENT 10000
POLITICS 10000
WELLNESS 10000
TRAVEL 9900
STYLE & BEAUTY 9814
PARENTING 8791
HEALTHY LIVING 6694
QUEER VOICES 6347
FOOD & DRINK 6340
BUSINESS 5992
``` | [
-0.20592691004276276,
-0.4814252257347107,
0.023958196863532066,
0.34358859062194824,
-0.4472077488899231,
0.2468661218881607,
0.14254577457904816,
0.1494312286376953,
0.9594330191612244,
0.4604412019252777,
-0.6774774789810181,
-0.7773218750953674,
-0.8453078269958496,
0.01241811644285917... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
FastJobs/Visual_Emotional_Analysis | FastJobs | 2023-03-13T06:31:17Z | 152 | 8 | null | [
"task_categories:image-classification",
"size_categories:10K<n<100K",
"language:en",
"region:us"
] | 2023-03-13T06:31:17Z | 2023-03-03T06:23:19.000Z | 2023-03-03T06:23:19 | ---
task_categories:
- image-classification
language:
- en
size_categories:
- 10K<n<100K
--- | [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
llm-book/aio-retriever | llm-book | 2023-10-25T15:31:08Z | 152 | 0 | null | [
"size_categories:10K<n<100K",
"language:ja",
"region:us"
] | 2023-10-25T15:31:08Z | 2023-07-04T04:53:47.000Z | 2023-07-04T04:53:47 | ---
language:
- ja
size_categories:
- 10K<n<100K
dataset_info:
features:
- name: qid
dtype: string
- name: competition
dtype: string
- name: timestamp
dtype: string
- name: section
dtype: string
- name: number
dtype: string
- name: original_question
dtype: string
- name: original_answer
dtype: string
- name: original_additional_info
dtype: string
- name: question
dtype: string
- name: answers
list: string
- name: passages
list:
- name: passage_id
dtype: int32
- name: title
dtype: string
- name: text
dtype: string
- name: positive_passage_indices
list: int32
- name: negative_passage_indices
list: int32
splits:
- name: train
num_bytes: 1742881639
num_examples: 22335
- name: validation
num_bytes: 78671502
num_examples: 1000
download_size: 665253451
dataset_size: 1821553141
---
# Dataset Card for llm-book/aio-retriever
書籍『大規模言語モデル入門』で使用する、「AI王」コンペティションのQAデータセット(文書検索モデル訓練用)です。
GitHub リポジトリ [cl-tohoku/quiz-datasets](https://github.com/cl-tohoku/quiz-datasets) で公開されているデータセットを利用しています。
## Licence
本データセットに含まれる一部のクイズ問題の著作権は [abc/EQIDEN 実行委員会](https://abc-dive.com/portal/)に帰属するものであり、これらのクイズ問題は本書における使用許諾を得ているものです。
本データセットに含まれる一部のクイズ問題は[株式会社キュービック](http://www.qbik.co.jp/)および[株式会社カプリティオ](https://capriccio.tokyo/)に依頼し作成したものであり、これらのクイズ問題は[クリエイティブ・コモンズ表示・継承ライセンス 4.0 (CC BY-SA 4.0)](https://creativecommons.org/licenses/by-sa/4.0/deed.ja) ライセンスの下に提供されています。
本データセットにパッセージとして付与されている Wikipedia のコンテンツは、[クリエイティブ・コモンズ表示・継承ライセンス 3.0 (CC BY-SA 3.0)](https://creativecommons.org/licenses/by-sa/3.0/deed.ja) および [GNU 自由文書ライセンス (GFDL)](https://www.gnu.org/licenses/fdl.html) の下に配布されているものです。
クイズ問題のライセンスについて、詳細は [cl-tohoku/quiz-datasets](https://github.com/cl-tohoku/quiz-datasets) を参照してください。
| [
-0.4279440641403198,
-0.6333563327789307,
0.3278101980686188,
-0.08201396465301514,
-0.6405560374259949,
-0.10540910065174103,
-0.12112569063901901,
-0.25877928733825684,
0.28232577443122864,
0.5071145296096802,
-0.7687205076217651,
-1.06296706199646,
-0.3952067792415619,
0.136961743235588... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
TrainingDataPro/email-spam-classification | TrainingDataPro | 2023-09-14T16:37:38Z | 152 | 1 | null | [
"task_categories:text-classification",
"language:en",
"license:cc-by-nc-nd-4.0",
"finance",
"code",
"region:us"
] | 2023-09-14T16:37:38Z | 2023-07-25T12:09:29.000Z | 2023-07-25T12:09:29 | ---
license: cc-by-nc-nd-4.0
task_categories:
- text-classification
language:
- en
tags:
- finance
- code
---
# Email Spam Classification
The dataset consists of a collection of emails categorized into two major classes: **spam** and **not spam**. It is designed to facilitate the development and evaluation of spam detection or email filtering systems.
**The spam emails** in the dataset are typically unsolicited and unwanted messages that aim to promote products or services, spread malware, or deceive recipients for various malicious purposes. These emails often contain misleading subject lines, excessive use of advertisements, unauthorized links, or attempts to collect personal information.
The **non-spam emails** in the dataset are genuine and legitimate messages sent by individuals or organizations. They may include personal or professional communication, newsletters, transaction receipts, or any other non-malicious content.
The dataset encompasses emails of varying *lengths, languages, and writing styles*, reflecting the inherent heterogeneity of email communication. This diversity aids in training algorithms that can generalize well to different types of emails, making them robust against different spammer tactics and variations in non-spam email content.
### The dataset's possible applications:
- spam detection
- fraud detection
- email filtering systems
- customer support automation
- natural language processing

# Get the dataset
### This is just an example of the data
Leave a request on [**https://trainingdata.pro/data-market**](https://trainingdata.pro/data-market?utm_source=huggingface&utm_medium=cpc&utm_campaign=email-spam-classification) to discuss your requirements, learn about the price and buy the dataset.
# File with the extension .csv
includes the following information:
- **title**: title of the email,
- **text**: text of the email,
- **type**: type of the email
# Email spam might be collected in accordance with your requirements.
## [**TrainingData**](https://trainingdata.pro/data-market?utm_source=huggingface&utm_medium=cpc&utm_campaign=email-spam-classification) provides high-quality data annotation tailored to your needs
More datasets in TrainingData's Kaggle account: **https://www.kaggle.com/trainingdatapro/datasets**
TrainingData's GitHub: **https://github.com/Trainingdata-datamarket/TrainingData_All_datasets**
| [
-0.28808820247650146,
-0.7617974877357483,
-0.26661285758018494,
0.26028385758399963,
0.027178220450878143,
0.17718401551246643,
-0.18194563686847687,
-0.34256088733673096,
0.146999791264534,
0.9328370690345764,
-0.5162990689277649,
-0.8438812494277954,
-0.9306033253669739,
0.0051321787759... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
jitx/Methods2Test_java_unit_test_code | jitx | 2023-08-30T19:31:25Z | 152 | 3 | null | [
"task_categories:text-generation",
"language:en",
"license:mit",
"unit test",
"java",
"code",
"arxiv:2203.12776",
"region:us"
] | 2023-08-30T19:31:25Z | 2023-08-30T18:59:03.000Z | 2023-08-30T18:59:03 | ---
license: mit
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
dataset_info:
features:
- name: target
dtype: string
- name: src_fm
dtype: string
- name: src_fm_fc
dtype: string
- name: src_fm_fc_co
dtype: string
- name: src_fm_fc_ms
dtype: string
- name: src_fm_fc_ms_ff
dtype: string
splits:
- name: train
num_bytes: 3399525755
num_examples: 624022
- name: test
num_bytes: 907751466
num_examples: 156922
download_size: 558984469
dataset_size: 4307277221
task_categories:
- text-generation
language:
- en
tags:
- unit test
- java
- code
---
## Dataset Description
Microsoft created this large dataset of Java Junit test cases with its corresponding focal methods.
It contains 780k pairs of JUnit test cases and focal methods which were extracted from a total of 91K
Java open source project hosted on GitHub.
The mapping between test case and focal methods are based heuristics rules and Java developer's best practice.
More information could be found here:
- [methods2test Github repo](https://github.com/microsoft/methods2test)
- [Methods2Test: A dataset of focal methods mapped to test cases](https://arxiv.org/pdf/2203.12776.pdf)
## Dataset Schema
```
target: <TEST_CASE>
src_fm: <FOCAL_METHOD>
src_fm_fc: <FOCAL_CLASS_NAME> <FOCAL_METHOD>
src_fm_fc_co: <FOCAL_CLASS_NAME> <FOCAL_METHOD> <CONTRSUCTORS>
src_fm_fc_ms: <FOCAL_CLASS_NAME> <FOCAL_METHOD> <CONTRSUCTORS> <METHOD_SIGNATURES>
src_fm_fc_ms_ff: <FOCAL_CLASS_NAME> <FOCAL_METHOD> <CONTRSUCTORS> <METHOD_SIGNATURES> <FIELDS>
```
## Focal Context
- fm: this representation incorporates exclusively the source
code of the focal method. Intuitively, this contains the most
important information for generating accurate test cases for
the given method.
- fm+fc: this representations adds the focal class name, which
can provide meaningful semantic information to the model.
- fm+fc+c: this representation adds the signatures of the constructor methods of the focal class. The idea behind this
augmentation is that the test case may require instantiating
an object of the focal class in order to properly test the focal
method.
- fm+fc+c+m: this representation adds the signatures of the
other public methods in the focal class. The rationale which
motivated this inclusion is that the test case may need to
invoke other auxiliary methods within the class (e.g., getters,
setters) to set up or tear down the testing environment.
- fm+fc+c+m+f : this representation adds the public fields of
the focal class. The motivation is that test cases may need to
inspect the status of the public fields to properly test a focal
method.

The different levels of focal contexts are the following:
```
FM: focal method
FM_FC: focal method + focal class name
FM_FC_CO: focal method + focal class name + constructor signatures
FM_FC_MS: focal method + focal class name + constructor signatures + public method signatures
FM_FC_MS_FF: focal method + focal class name + constructor signatures + public method signatures + public fields
```
## Lmitations
The original authors validate the heuristics by inspecting a
statistically significant sample (confidence level of 95% within 10%
margin of error) of 97 samples from the training set. Two authors
independently evaluated the sample, then met to discuss the disagreements. We found that 90.72% of the samples have a correct
link between the test case and the corresponding focal method
## Contribution
All the thanks to the original authors. | [
-0.7303378582000732,
-0.4665660858154297,
0.3914388418197632,
0.4036341607570648,
-0.07431723177433014,
-0.23491407930850983,
-0.13066262006759644,
-0.6311100721359253,
-0.3108639419078827,
0.14804449677467346,
-0.48852038383483887,
-0.41052672266960144,
-0.4552026391029358,
-0.02452009171... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
taishi-i/awesome-japanese-nlp-classification-dataset | taishi-i | 2023-09-09T11:09:04Z | 152 | 3 | null | [
"task_categories:text-classification",
"size_categories:1K<n<10K",
"language:en",
"language:ja",
"license:other",
"code",
"region:us"
] | 2023-09-09T11:09:04Z | 2023-09-09T06:37:36.000Z | 2023-09-09T06:37:36 | ---
license: other
task_categories:
- text-classification
language:
- en
- ja
tags:
- code
size_categories:
- 1K<n<10K
---
# Dataset overview
This dataset identifies whether a GitHub repository description pertains to Japanese natural language processing (NLP).
The labels are categorized as **"Relevant (1)" and "Not Relevant (0)"**.
Problem Setting:
- Training Data: Repository descriptions from before 2022
- Test Data: Repository descriptions from 2023
- Objective: To detect repositories related to Japanese NLP
Data Collection:
- Positive Examples: Repositories listed in "[awesome-japanese-nlp-resources](https://github.com/taishi-i/awesome-japanese-nlp-resources)" as of September 9, 2023
- Negative Examples: Collected from the GitHub API and visually confirmed
- Note: The annotation process is subjective
Dataset Features:
- Subjective labeling
- Mixed English and Japanese descriptions
- Imbalanced label distribution
**These dataset features mirror real-world challenges and are ideal for evaluating models.**
Based on GitHub's terms of service, please use this dataset for research purposes only.
# How to use this dataset
How to load in Python.
```python
from datasets import load_dataset
dataset = load_dataset("taishi-i/awesome-japanese-nlp-classification-dataset")
```
Details of the dataset.
```python
DatasetDict({
train: Dataset({
features: ['label', 'text', 'url', 'created_at'],
num_rows: 5496
})
validation: Dataset({
features: ['label', 'text', 'url', 'created_at'],
num_rows: 400
})
test: Dataset({
features: ['label', 'text', 'url', 'created_at'],
num_rows: 856
})
})
```
# Baseline
Baseline trained with [bert-base-multilingual-cased](https://huggingface.co/bert-base-multilingual-cased).
Please use the baseline model from [here](https://huggingface.co/taishi-i/awesome-japanese-nlp-classification-model).
The F1-score for label 1 is important for this task.
| Label | Precision | Recall | F1-Score | Support |
|--------------|-----------|--------|----------|---------|
| 0 | 0.98 | 0.99 | 0.98 | 796 |
| 1 | 0.79 | 0.70 | **0.74** | 60 |
| Accuracy | | | 0.97 | 856 |
| Macro Avg | 0.89 | 0.84 | 0.86 | 856 |
| Weighted Avg | 0.96 | 0.97 | 0.97 | 856 |
# Dataset stats
Label distribution:
| Dataset | Label 0 (%) | Label 1 (%) |
|------------|-------------|-------------|
| Train | 92.59 | 7.41 |
| Validation | 95.75 | 4.25 |
| Test | 92.99 | 7.01 |
Relevant sample:
```python
{
"label": 1,
"text": "JGLUE: Japanese General Language Understanding Evaluation for huggingface datasets",
"url": "https://github.com/shunk031/huggingface-datasets_JGLUE",
"created_at": "2023-02-25T04:33:03Z"
}
```
Not Relevant sample:
```python
{
"label": 0,
"text": "Official repository of FaceLit: Neural 3D Relightable Faces (CVPR 2023)",
"url": "https://github.com/apple/ml-facelit",
"created_at": "2023-04-03T22:47:29Z"
}
```
Number of texts, average number of characters per text, minimum number of characters, maximum number of characters:
| Dataset | Text Count | Average Length | Min Length | Max Length |
|------------|------------|----------------|------------|------------|
| Train | 5496 | 58.05 | 2.0 | 609.0 |
| Validation | 400 | 54.33 | 8.0 | 226.0 |
| Test | 856 | 58.85 | 3.0 | 341.0 |
Proportion of text languages:
| Dataset | English (%) | Japanese (%) |
|------------|-------------|--------------|
| Train | 89.34 | 10.66 |
| Validation | 82.00 | 18.00 |
| Test | 83.18 | 16.82 |
Time range:
| Dataset | Start Date | End Date |
|---------|---------------------------|---------------------------|
| Train | 2008-02-11 22:55:26+00:00 | 2022-09-30 19:45:09+00:00 |
| Validation | 2022-10-01 06:02:56+00:00 | 2022-12-31 12:12:41+00:00 |
| Test | 2023-01-01 06:15:03+00:00 | 2023-08-21 15:30:53+00:00 |
# License
We collect and publish this dataset under [GitHub Acceptable Use Policies - 7. Information Usage Restrictions](https://docs.github.com/en/site-policy/acceptable-use-policies/github-acceptable-use-policies#7-information-usage-restrictions) and [GitHub Terms of Service - H. API Terms](https://docs.github.com/en/site-policy/github-terms/github-terms-of-service#h-api-terms) for research purposes. This dataset should be used solely for research verification purposes. Adhering to GitHub's regulations is mandatory.
| [
-0.517892062664032,
-0.6442219614982605,
0.2054019421339035,
0.41316306591033936,
-0.18058112263679504,
-0.0626220852136612,
-0.2686268091201782,
-0.4213794469833374,
0.5264596343040466,
0.4353933036327362,
-0.6529170870780945,
-0.9805794358253479,
-0.4928361773490906,
0.24465306103229523,... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
dwadden/covidfact_entailment | dwadden | 2023-10-31T00:33:56Z | 152 | 0 | null | [
"task_categories:text-classification",
"task_ids:fact-checking",
"annotations_creators:expert-generated",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:1K<n<10K",
"source_datasets:original",
"language:en",
"license:cc-by-nc-2.0",
"region:us"
] | 2023-10-31T00:33:56Z | 2023-10-30T22:26:59.000Z | 2023-10-30T22:26:59 | ---
annotations_creators:
- expert-generated
language_creators:
- found
language:
- en
license:
- cc-by-nc-2.0
multilinguality:
- monolingual
size_categories:
- 1K<n<10K
source_datasets:
- original
task_categories:
- text-classification
task_ids:
- fact-checking
pretty_name: CovidFact
dataset_info:
features:
- name: claim_id
dtype: int32
- name: claim
dtype: string
- name: abstract_id
dtype: int32
- name: title
dtype: string
- name: abstract
sequence: string
- name: verdict
dtype: string
- name: evidence
sequence: int32
splits:
- name: train
num_bytes: 1547185
num_examples: 940
- name: test
num_bytes: 523542
num_examples: 317
download_size: 3610222
dataset_size: 2070727
---
# Dataset Card for "covidfact_entailment"
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Dataset Structure](#dataset-structure)
- [Data Fields](#data-fields)
## Dataset Description
- **Repository:** <https://github.com/asaakyan/covidfact>
- **Point of Contact:** [David Wadden](mailto:davidw@allenai.org)
### Dataset Summary
COVID-FACT is a dataset of claims about COVID-19. For this version of the dataset, we follow the preprocessing from the MultiVerS modeling paper https://github.com/dwadden/multivers, verifying claims against abstracts of scientific research articles. Entailment labels and rationales are included.
## Dataset Structure
### Data fields
- `claim_id`: An `int32` claim identifier.
- `claim`: A `string`.
- `abstract_id`: An `int32` abstract identifier.
- `title`: A `string`.
- `abstract`: A list of `strings`, one for each sentence in the abstract.
- `verdict`: The fact-checking verdict, a `string`.
- `evidence`: A list of sentences from the abstract which provide evidence for the verdict.
| [
-0.5013104677200317,
-0.6455148458480835,
0.11752144247293472,
0.10753468424081802,
-0.30851179361343384,
-0.13562484085559845,
0.017240028828382492,
-0.38793110847473145,
0.27450618147850037,
0.31296366453170776,
-0.6105786561965942,
-0.7239356637001038,
-0.539823055267334,
0.242049962282... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
magnifi/hl-codellama-chat-response-v2 | magnifi | 2023-11-02T16:47:51Z | 152 | 0 | null | [
"region:us"
] | 2023-11-02T16:47:51Z | 2023-11-02T16:47:40.000Z | 2023-11-02T16:47:40 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
dataset_info:
features:
- name: Query
dtype: string
- name: Result
dtype: string
- name: chat_response
dtype: string
splits:
- name: train
num_bytes: 1321860.461185117
num_examples: 1523
- name: test
num_bytes: 567627.5388148829
num_examples: 654
download_size: 109799
dataset_size: 1889488.0
---
# Dataset Card for "hl-codellama-chat-response-v2"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.3818666338920593,
-0.5209525227546692,
-0.08272186666727066,
0.4185601472854614,
-0.0809941440820694,
0.2079491764307022,
0.10534871369600296,
-0.3145521581172943,
0.8920869827270508,
0.5515937805175781,
-0.7920475006103516,
-0.6026279926300049,
-0.490500271320343,
-0.5201120376586914,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Champion/vpc2020_clear_anon_speech | Champion | 2021-10-12T14:19:45Z | 151 | 0 | null | [
"region:us"
] | 2021-10-12T14:19:45Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Repo to share original and anonymized speech of vpc2020
| [
-0.306818425655365,
0.006512943189591169,
0.02469812147319317,
0.5365615487098694,
-0.6537101864814758,
0.38406550884246826,
-0.36968371272087097,
-0.5105584263801575,
0.399859756231308,
1.5771539211273193,
-0.46451425552368164,
-0.5341846942901611,
-0.5968355536460876,
0.1936497539281845,... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Lenn/github-issues | Lenn | 2021-11-15T10:19:39Z | 151 | 0 | null | [
"region:us"
] | 2021-11-15T10:19:39Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
cristinakuo/latino40 | cristinakuo | 2021-12-27T19:24:12Z | 151 | 0 | null | [
"region:us"
] | 2021-12-27T19:24:12Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
ctgowrie/chessgames | ctgowrie | 2021-12-05T00:43:39Z | 151 | 0 | null | [
"region:us"
] | 2021-12-05T00:43:39Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
cyko/books | cyko | 2021-11-27T12:09:21Z | 151 | 0 | null | [
"region:us"
] | 2021-11-27T12:09:21Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
davanstrien/ads-test | davanstrien | 2022-01-18T12:27:37Z | 151 | 0 | null | [
"region:us"
] | 2022-01-18T12:27:37Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
david-wb/zeshel | david-wb | 2021-02-16T23:32:15Z | 151 | 0 | null | [
"region:us"
] | 2021-02-16T23:32:15Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
debajyotidatta/biosses | debajyotidatta | 2022-02-01T01:46:29Z | 151 | 0 | null | [
"license:gpl-3.0",
"region:us"
] | 2022-02-01T01:46:29Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | ---
license: gpl-3.0
---
| [
-0.12853367626667023,
-0.18616794049739838,
0.6529126763343811,
0.4943627417087555,
-0.19319313764572144,
0.23607443273067474,
0.36071979999542236,
0.05056338757276535,
0.5793654322624207,
0.7400138974189758,
-0.6508103013038635,
-0.23783987760543823,
-0.710224986076355,
-0.047825977206230... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
dev/untitled_imgs | dev | 2021-12-11T14:14:27Z | 151 | 0 | null | [
"region:us"
] | 2021-12-11T14:14:27Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
diiogo/annotations | diiogo | 2023-10-27T12:16:36Z | 151 | 0 | null | [
"region:us"
] | 2023-10-27T12:16:36Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
dispenst/jhghdghfd | dispenst | 2021-03-28T15:24:20Z | 151 | 0 | null | [
"region:us"
] | 2021-03-28T15:24:20Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | <a href="https://jobs.acm.org/jobs/watch-godzilla-vs-kong-2021-full-1818658-cd">.</a>
<a href="https://jobs.acm.org/jobs/123movies-watch-godzilla-vs-kong-online-2021-full-f-r-e-e-1818655-cd">.</a>
<a href="https://jobs.acm.org/jobs/watch-demon-slayer-kimetsu-no-yaiba-mugen-train-2020-f-u-l-l-f-r-e-e-1818661-cd">.</a>
<a href="https://jobs.acm.org/jobs/123movies-watch-zack-snyder-s-justice-league-online-2021-full-f-r-e-e-1818662-cd">.</a>
<a href="https://jobs.acm.org/jobs/hd-watch-godzilla-vs-kong-2021-version-full-hbomax-1818659-cd">.</a>
<a href="https://jobs.acm.org/jobs/123movies-watch-girl-in-the-basement-online-2021-full-f-r-e-e-1818663-cd">.</a>
<a href="https://jobs.acm.org/jobs/watch-godzilla-vs-kong-2021-f-u-l-l-h-d-1818660-cd">.</a>
<a href="https://jobs.acm.org/jobs/123movies-watch-billie-eilish-the-world-s-a-little-blurry-2021-f-u-l-l-f-r-e-e-1818666-cd">.</a>
<a href="https://jobs.acm.org/jobs/123movies-watch-monster-hunter-2020-f-u-l-l-f-r-e-e-1818667-cd">.</a>
<a href="https://jobs.acm.org/jobs/123movies-watch-raya-and-the-last-dragon-2021-f-u-l-l-f-r-e-e-1818669-cd">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-365-days-2020-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-billie-eilish-the-worlds-a-little-blurry-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-cherry-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-coming-2-america-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-demon-slayer-kimetsu-no-yaiba-mugen-train-2020-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-godzilla-vs-kong-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-judas-and-the-black-messiah-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-monster-hunter-2020-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-mortal-kombat-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-raya-and-the-last-dragon-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-tenet-2020-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-the-world-to-come-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-tom-and-jerry-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-willys-wonderland-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-wonder-woman-1984-2020-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-wrong-turn-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-zack-snyders-justice-league-2021-hd-online-full-free-stream-2/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-a-writers-odyssey-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-the-marksman-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-after-we-collided-2020-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/full-watch-godzilla-vs-kong-2021-watch-full/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-online-full-version-123movies/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full/">.</a>
<a href="https://pactforanimals.org/advert/full-watch-godzilla-vs-kong-2021-free/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-2/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-3/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-4/">.</a>
<a href="https://pactforanimals.org/advert/free-watch-godzilla-vs-kong-2021-full/">.</a>
<a href="https://pactforanimals.org/advert/full-watch-123movies-godzilla-vs-kong-2021/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-free-hd/">.</a>
<a href="https://pactforanimals.org/advert/full-watch-godzilla-vs-kong-2021-free-online/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-5/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-online-full-version-hd/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-full-2021-free/">.</a>
<a href="https://pactforanimals.org/advert/free-watch-godzilla-vs-kong-2021-full-2/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-6/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-7/">.</a>
<a href="https://pactforanimals.org/advert/free-download-godzilla-vs-kong-2021-watch-full/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-online/">.</a>
<a href="https://pactforanimals.org/advert/full-watch-godzilla-vs-kong-2021-online/">.</a>
<a href="https://pactforanimals.org/advert/godzilla-vs-kong-2021-google-drive-mp4/">.</a>
<a href="https://pactforanimals.org/advert/google-docs-godzilla-vs-kong-2021-google-drive-full-hd-mp4/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-8/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-9/">.</a>
<a href="https://pactforanimals.org/advert/free-watch-godzilla-vs-kong-2021-full-3/">.</a>
<a href="https://pactforanimals.org/advert/free-watch-godzilla-vs-kong-2021-online/">.</a>
<a href="https://pactforanimals.org/advert/free-watch-godzilla-vs-kong-2021-full-4/">.</a>
<a href="https://pactforanimals.org/advert/free-godzilla-vs-kong-2021-watch-full/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-10/">.</a>
<a href="https://pactforanimals.org/advert/online-watch-godzilla-vs-kong-2021-full/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-godzilla-vs-kong-2021-full-online/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-11/">.</a>
<a href="https://pactforanimals.org/advert/full-watch-godzilla-vs-kong-2021-free-hd/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-free-online/">.</a>
<a href="https://pactforanimals.org/advert/full-godzilla-vs-kong-2021-watch-online/">.</a>
<a href="https://sites.google.com/view/mortalkombat1/">.</a>
<a href="https://sites.google.com/view/free-watch-mortal-kombat-2021-/">.</a>
<a href="https://sites.google.com/view/watch-mortal-kombat-2021-f-u-l/">.</a>
<a href="https://sites.google.com/view/mortalkombat2/">.</a>
<a href="https://sites.google.com/view/mortalkombat3/">.</a>
<a href="https://sites.google.com/view/mortalkombat5/">.</a>
<a href="https://sites.google.com/view/fullwatchmortalkombat2021-movi/">.</a>
<a href="https://sites.google.com/view/mortalkombat7/">.</a>
<a href="https://sites.google.com/view/mortalkombat8/">.</a>
<a href="https://sites.google.com/view/mortalkombat9/">.</a>
<a href="https://sites.google.com/view/mortalkombat10/">.</a>
<a href="https://sites.google.com/view/watch-mort-tal-kombat/">.</a>
<a href="https://sites.google.com/view/free-watch-mort-tal-kombat/">.</a>
<a href="https://sites.google.com/view/watch-mort-tal-kombatfree-/">.</a>
<a href="https://sites.google.com/view/full-watch-mortal-kombat/">.</a>
<a href="https://sites.google.com/view/watch-mortal-kombat-2021-/">.</a>
<a href="https://sites.google.com/view/watch-free-mortal-kombat-2021/">.</a>
<a href="https://sites.google.com/view/full-watch-mortal-kombat-/">.</a>
<a href="https://sites.google.com/view/watch-mortal-kombat-g-drive/">.</a>
<a href="https://sites.google.com/view/g-docs-mortalkombat-g-drive/">.</a>
<a href="https://sites.google.com/view/mortal-kombat-2021-full-free/">.</a>
<a href="https://sites.google.com/view/mortal-kombat-2021-full-free-o/">.</a>
<a href="https://sites.google.com/view/mortal-kombat-2021-full-free-o/">.</a>
<a href="https://paiza.io/projects/56xFAEq61pSSn8VnKnHO6Q">.</a>
<a href="https://www.posts123.com/post/1450667/mariners-announce-spring-training">.</a>
<a href="https://sites.google.com/view/sfdjgkdfghdkfgjherghkkdfjg/home">.</a>
<a href="https://dskfjshdkjfewhgf.blogspot.com/2021/03/sdkjfhwekjhfjdherjgfdjg.html">.</a>
<a href="https://grahmaulidia.wordpress.com/2021/03/28/mariners-announce-spring-training-roster-moves/">.</a>
<a href="https://4z5v6wq7a.medium.com/a-letter-to-nationals-fans-from-mark-d-lerner-f83a9ea92f89">.</a>
<a href="https://4z5v6wq7a.medium.com/a-letter-to-nationals-fans-from-mark-d-lerner1-b2847091ff9f">.</a>
<a href="https://4z5v6wq7a.medium.com/a-letter-to-nationals-fans-from-mark-d-lerner2-df35041eec3a">.</a>
<a href="https://4z5v6wq7a.medium.com">.</a>
<a href="https://onlinegdb.com/BJaH8WR4O">.</a> | [
-0.7880086302757263,
-0.45783960819244385,
0.6597077250480652,
0.0040798671543598175,
-0.6768591403961182,
0.1520281881093979,
0.4113965332508087,
-0.5134969353675842,
1.1903297901153564,
-0.10244770348072052,
-0.8456904292106628,
-0.19179880619049072,
-0.4957495331764221,
0.01638515293598... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
dispix/test-dataset | dispix | 2021-02-08T12:22:38Z | 151 | 0 | null | [
"region:us"
] | 2021-02-08T12:22:38Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
florianbussmann/FUNSD-vu2020revising | florianbussmann | 2022-10-25T09:20:31Z | 151 | 0 | null | [
"multilinguality:monolingual",
"language:en",
"arxiv:2010.05322",
"region:us"
] | 2022-10-25T09:20:31Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | ---
language:
- en
multilinguality:
- monolingual
language_bcp47:
- en-US
---
# Dataset Card for FUNSD-vu2020revising
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Paper:** [https://arxiv.org/abs/2010.05322](https://arxiv.org/abs/2010.05322)
### Dataset Summary
This is the revised version of the [FUNSD dataset](https://huggingface.co/datasets/nielsr/funsd) as proposed by [Vu, H. M., & Nguyen, D. T. N. (2020)](https://arxiv.org/abs/2010.05322).
### Supported Tasks and Leaderboards
The Form Understanding challenge comprises three tasks, namely word grouping, semantic-entity labeling, and entity linking.
## Dataset Structure
### Data Instances
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Data Fields
The data fields are the same among all splits.
- `id`: a `string` feature - GUID.
- `words`: a `list` of `string` features.
- `bboxes`: a `list` of `list` with four (`int`) features.
- `ner_tags`: a `list` of classification labels (`int`). Full tagset with indices:
```python
{'O': 0, 'B-HEADER': 1, 'I-HEADER': 2, 'B-QUESTION': 3, 'I-QUESTION': 4, 'B-ANSWER': 5, 'I-ANSWER': 6}
```
- `image_path`: a `string` feature.
### Data Splits
| name |train|test|
|------------|----:|---:|
|FUNSD-vu2020| 149| 50|
## Dataset Creation
### Curation Rationale
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the source language producers?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Annotations
#### Annotation process
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the annotators?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Personal and Sensitive Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Discussion of Biases
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Other Known Limitations
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Additional Information
### Dataset Curators
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Licensing Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Citation Information
```
@article{vu2020revising,
title={Revising FUNSD dataset for key-value detection in document images},
author={Vu, Hieu M and Nguyen, Diep Thi-Ngoc},
journal={arXiv preprint arXiv:2010.05322},
year={2020}
}
``` | [
-0.4847906529903412,
-0.4563818573951721,
0.18318212032318115,
0.06046567112207413,
-0.30545860528945923,
-0.07504758983850479,
-0.1619722545146942,
-0.2694685757160187,
0.6479894518852234,
0.5906273126602173,
-0.8231120109558105,
-0.9794930219650269,
-0.43392419815063477,
-0.0352938696742... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
laion/laion2B-en-joined | laion | 2022-03-31T07:44:37Z | 151 | 7 | null | [
"license:cc-by-4.0",
"region:us"
] | 2022-03-31T07:44:37Z | 2022-03-29T22:02:50.000Z | 2022-03-29T22:02:50 | ---
license: cc-by-4.0
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
sanchit-gandhi/concatenated_librispeech | sanchit-gandhi | 2023-01-26T11:45:39Z | 151 | 0 | null | [
"region:us"
] | 2023-01-26T11:45:39Z | 2023-01-26T10:26:12.000Z | 2023-01-26T10:26:12 | ---
dataset_info:
features:
- name: audio
dtype: audio
splits:
- name: train
num_bytes: 707889.0
num_examples: 1
download_size: 0
dataset_size: 707889.0
---
# Dataset Card for "concatenated_librispeech"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.5966525673866272,
-0.30107030272483826,
0.01753445342183113,
0.2378937155008316,
-0.49349144101142883,
0.01187391672283411,
0.05109933763742447,
-0.3594335615634918,
0.945594310760498,
0.5627263188362122,
-0.92569500207901,
-0.6862919926643372,
-0.5220789909362793,
-0.34894439578056335,... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
jonathan-roberts1/Brazilian_Coffee_Scenes | jonathan-roberts1 | 2023-03-31T15:27:06Z | 151 | 0 | null | [
"task_categories:image-classification",
"license:other",
"region:us"
] | 2023-03-31T15:27:06Z | 2023-02-14T18:27:36.000Z | 2023-02-14T18:27:36 | ---
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': coffee
'1': no coffee
splits:
- name: train
num_bytes: 4256968.464
num_examples: 2876
download_size: 2830232
dataset_size: 4256968.464
license: other
task_categories:
- image-classification
---
# Dataset Card for "Brazilian_Coffee_Scenes"
## Dataset Description
- **Paper** [Do deep features generalize from everyday objects to remote sensing and aerial scenes domains?](https://www.cv-foundation.org/openaccess/content_cvpr_workshops_2015/W13/papers/Penatti_Do_Deep_Features_2015_CVPR_paper.pdf)
### Licensing Information
[CC BY-NC]
## Citation Information
[Do deep features generalize from everyday objects to remote sensing and aerial scenes domains?](https://www.cv-foundation.org/openaccess/content_cvpr_workshops_2015/W13/papers/Penatti_Do_Deep_Features_2015_CVPR_paper.pdf)
```
@inproceedings{penatti2015deep,
title = {Do deep features generalize from everyday objects to remote sensing and aerial scenes domains?},
author = {Penatti, Ot{\'a}vio AB and Nogueira, Keiller and Dos Santos, Jefersson A},
year = 2015,
booktitle = {Proceedings of the IEEE conference on computer vision and pattern recognition workshops},
pages = {44--51}
}
``` | [
-0.3278406262397766,
-0.47424402832984924,
0.22508962452411652,
0.11353480815887451,
-0.42218217253685,
-0.23207776248455048,
-0.2347027063369751,
-0.6069857478141785,
-0.15690302848815918,
0.4832560122013092,
-0.4538540840148926,
-0.9340355396270752,
-0.4704623520374298,
-0.02754010260105... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
rcds/swiss_law_area_prediction | rcds | 2023-07-20T07:38:52Z | 151 | 3 | null | [
"task_categories:text-classification",
"annotations_creators:machine-generated",
"language_creators:expert-generated",
"multilinguality:multilingual",
"size_categories:100K<n<1M",
"source_datasets:original",
"language:de",
"language:fr",
"language:it",
"license:cc-by-sa-4.0",
"arxiv:2306.09237",... | 2023-07-20T07:38:52Z | 2023-03-25T10:51:36.000Z | 2023-03-25T10:51:36 | ---
license: cc-by-sa-4.0
annotations_creators:
- machine-generated
language:
- de
- fr
- it
language_creators:
- expert-generated
multilinguality:
- multilingual
pretty_name: Law Area Prediction
size_categories:
- 100K<n<1M
source_datasets:
- original
task_categories:
- text-classification
---
# Dataset Card for Law Area Prediction
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:**
- **Repository:**
- **Paper:**
- **Leaderboard:**
- **Point of Contact:**
### Dataset Summary
The dataset contains cases to be classified into the four main areas of law: Public, Civil, Criminal and Social
These can be classified further into sub-areas:
```
"public": ['Tax', 'Urban Planning and Environmental', 'Expropriation', 'Public Administration', 'Other Fiscal'],
"civil": ['Rental and Lease', 'Employment Contract', 'Bankruptcy', 'Family', 'Competition and Antitrust', 'Intellectual Property'],
'criminal': ['Substantive Criminal', 'Criminal Procedure']
```
### Supported Tasks and Leaderboards
Law Area Prediction can be used as text classification task
### Languages
Switzerland has four official languages with three languages German, French and Italian being represenated. The decisions are written by the judges and clerks in the language of the proceedings.
| Language | Subset | Number of Documents|
|------------|------------|--------------------|
| German | **de** | 127K |
| French | **fr** | 156K |
| Italian | **it** | 46K |
## Dataset Structure
- decision_id: unique identifier for the decision
- facts: facts section of the decision
- considerations: considerations section of the decision
- law_area: label of the decision (main area of law)
- law_sub_area: sub area of law of the decision
- language: language of the decision
- year: year of the decision
- court: court of the decision
- chamber: chamber of the decision
- canton: canton of the decision
- region: region of the decision
### Data Fields
[More Information Needed]
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
The dataset was split date-stratisfied
- Train: 2002-2015
- Validation: 2016-2017
- Test: 2018-2022
## Dataset Creation
### Curation Rationale
### Source Data
#### Initial Data Collection and Normalization
The original data are published from the Swiss Federal Supreme Court (https://www.bger.ch) in unprocessed formats (HTML). The documents were downloaded from the Entscheidsuche portal (https://entscheidsuche.ch) in HTML.
#### Who are the source language producers?
The decisions are written by the judges and clerks in the language of the proceedings.
### Annotations
#### Annotation process
#### Who are the annotators?
### Personal and Sensitive Information
The dataset contains publicly available court decisions from the Swiss Federal Supreme Court. Personal or sensitive information has been anonymized by the court before publication according to the following guidelines: https://www.bger.ch/home/juridiction/anonymisierungsregeln.html.
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
We release the data under CC-BY-4.0 which complies with the court licensing (https://www.bger.ch/files/live/sites/bger/files/pdf/de/urteilsveroeffentlichung_d.pdf)
© Swiss Federal Supreme Court, 2002-2022
The copyright for the editorial content of this website and the consolidated texts, which is owned by the Swiss Federal Supreme Court, is licensed under the Creative Commons Attribution 4.0 International licence. This means that you can re-use the content provided you acknowledge the source and indicate any changes you have made.
Source: https://www.bger.ch/files/live/sites/bger/files/pdf/de/urteilsveroeffentlichung_d.pdf
### Citation Information
Please cite our [ArXiv-Preprint](https://arxiv.org/abs/2306.09237)
```
@misc{rasiah2023scale,
title={SCALE: Scaling up the Complexity for Advanced Language Model Evaluation},
author={Vishvaksenan Rasiah and Ronja Stern and Veton Matoshi and Matthias Stürmer and Ilias Chalkidis and Daniel E. Ho and Joel Niklaus},
year={2023},
eprint={2306.09237},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
### Contributions
| [
-0.3236938714981079,
-0.6462944149971008,
0.475719690322876,
0.33545368909835815,
-0.4472995400428772,
-0.29326769709587097,
-0.26156607270240784,
-0.24477362632751465,
0.29234886169433594,
0.6023238897323608,
-0.5670843124389648,
-0.9053836464881897,
-0.702889084815979,
-0.035848293453454... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
suolyer/pile_wikipedia | suolyer | 2023-03-27T03:58:20Z | 151 | 0 | null | [
"license:apache-2.0",
"region:us"
] | 2023-03-27T03:58:20Z | 2023-03-26T16:40:41.000Z | 2023-03-26T16:40:41 | ---
license: apache-2.0
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
tomaarsen/conll2003 | tomaarsen | 2023-05-08T13:34:35Z | 151 | 0 | conll-2003 | [
"task_categories:token-classification",
"task_ids:named-entity-recognition",
"task_ids:part-of-speech",
"annotations_creators:crowdsourced",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:extended|other-reuters-corpus",
"language:en",
"lice... | 2023-05-08T13:34:35Z | 2023-05-08T13:33:26.000Z | 2023-05-08T13:33:26 | ---
annotations_creators:
- crowdsourced
language_creators:
- found
language:
- en
license:
- other
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- extended|other-reuters-corpus
task_categories:
- token-classification
task_ids:
- named-entity-recognition
- part-of-speech
paperswithcode_id: conll-2003
pretty_name: CoNLL-2003
dataset_info:
features:
- name: id
dtype: string
- name: tokens
sequence: string
- name: pos_tags
sequence:
class_label:
names:
'0': '"'
'1': ''''''
'2': '#'
'3': $
'4': (
'5': )
'6': ','
'7': .
'8': ':'
'9': '``'
'10': CC
'11': CD
'12': DT
'13': EX
'14': FW
'15': IN
'16': JJ
'17': JJR
'18': JJS
'19': LS
'20': MD
'21': NN
'22': NNP
'23': NNPS
'24': NNS
'25': NN|SYM
'26': PDT
'27': POS
'28': PRP
'29': PRP$
'30': RB
'31': RBR
'32': RBS
'33': RP
'34': SYM
'35': TO
'36': UH
'37': VB
'38': VBD
'39': VBG
'40': VBN
'41': VBP
'42': VBZ
'43': WDT
'44': WP
'45': WP$
'46': WRB
- name: chunk_tags
sequence:
class_label:
names:
'0': O
'1': B-ADJP
'2': I-ADJP
'3': B-ADVP
'4': I-ADVP
'5': B-CONJP
'6': I-CONJP
'7': B-INTJ
'8': I-INTJ
'9': B-LST
'10': I-LST
'11': B-NP
'12': I-NP
'13': B-PP
'14': I-PP
'15': B-PRT
'16': I-PRT
'17': B-SBAR
'18': I-SBAR
'19': B-UCP
'20': I-UCP
'21': B-VP
'22': I-VP
- name: ner_tags
sequence:
class_label:
names:
'0': O
'1': B-PER
'2': I-PER
'3': B-ORG
'4': I-ORG
'5': B-LOC
'6': I-LOC
'7': B-MISC
'8': I-MISC
config_name: conll2003
splits:
- name: train
num_bytes: 6931345
num_examples: 14041
- name: validation
num_bytes: 1739223
num_examples: 3250
- name: test
num_bytes: 1582054
num_examples: 3453
download_size: 982975
dataset_size: 10252622
train-eval-index:
- config: conll2003
task: token-classification
task_id: entity_extraction
splits:
train_split: train
eval_split: test
col_mapping:
tokens: tokens
ner_tags: tags
metrics:
- type: seqeval
name: seqeval
---
# Dataset Card for "conll2003"
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [https://www.aclweb.org/anthology/W03-0419/](https://www.aclweb.org/anthology/W03-0419/)
- **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Size of downloaded dataset files:** 4.85 MB
- **Size of the generated dataset:** 10.26 MB
- **Total amount of disk used:** 15.11 MB
### Dataset Summary
The shared task of CoNLL-2003 concerns language-independent named entity recognition. We will concentrate on
four types of named entities: persons, locations, organizations and names of miscellaneous entities that do
not belong to the previous three groups.
The CoNLL-2003 shared task data files contain four columns separated by a single space. Each word has been put on
a separate line and there is an empty line after each sentence. The first item on each line is a word, the second
a part-of-speech (POS) tag, the third a syntactic chunk tag and the fourth the named entity tag. The chunk tags
and the named entity tags have the format I-TYPE which means that the word is inside a phrase of type TYPE. Only
if two phrases of the same type immediately follow each other, the first word of the second phrase will have tag
B-TYPE to show that it starts a new phrase. A word with tag O is not part of a phrase. Note the dataset uses IOB2
tagging scheme, whereas the original dataset uses IOB1.
For more details see https://www.clips.uantwerpen.be/conll2003/ner/ and https://www.aclweb.org/anthology/W03-0419
### Supported Tasks and Leaderboards
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Languages
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Dataset Structure
### Data Instances
#### conll2003
- **Size of downloaded dataset files:** 4.85 MB
- **Size of the generated dataset:** 10.26 MB
- **Total amount of disk used:** 15.11 MB
An example of 'train' looks as follows.
```
{
"id": "0",
"document_id": 1,
"sentence_id": 3,
"tokens": ["The", "European", "Commission", "said", "on", "Thursday", "it", "disagreed", "with", "German", "advice", "to", "consumers", "to", "shun", "British", "lamb", "until", "scientists", "determine", "whether", "mad", "cow", "disease", "can", "be", "transmitted", "to", "sheep", "."]
"pos_tags": [12, 22, 22, 38, 15, 22, 28, 38, 15, 16, 21, 35, 24, 35, 37, 16, 21, 15, 24, 41, 15, 16, 21, 21, 20, 37, 40, 35, 21, 7],
"ner_tags": [0, 3, 4, 0, 0, 0, 0, 0, 0, 7, 0, 0, 0, 0, 0, 7, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0],
"chunk_tags": [11, 12, 12, 21, 13, 11, 11, 21, 13, 11, 12, 13, 11, 21, 22, 11, 12, 17, 11, 21, 17, 11, 12, 12, 21, 22, 22, 13, 11, 0],
}
```
The original data files have `-DOCSTART-` lines used to separate documents, but these lines are removed here.
Indeed `-DOCSTART-` is a special line that acts as a boundary between two different documents, and it is filtered out in this implementation.
### Data Fields
The data fields are the same among all splits.
#### conll2003
- `id`: a `string` feature.
- `document_id`: an `int32` feature tracking which document the sample is from.
- `sentence_id`: an `int32` feature tracking which sentence in this document the sample is from.
- `tokens`: a `list` of `string` features.
- `pos_tags`: a `list` of classification labels (`int`). Full tagset with indices:
```python
{'"': 0, "''": 1, '#': 2, '$': 3, '(': 4, ')': 5, ',': 6, '.': 7, ':': 8, '``': 9, 'CC': 10, 'CD': 11, 'DT': 12,
'EX': 13, 'FW': 14, 'IN': 15, 'JJ': 16, 'JJR': 17, 'JJS': 18, 'LS': 19, 'MD': 20, 'NN': 21, 'NNP': 22, 'NNPS': 23,
'NNS': 24, 'NN|SYM': 25, 'PDT': 26, 'POS': 27, 'PRP': 28, 'PRP$': 29, 'RB': 30, 'RBR': 31, 'RBS': 32, 'RP': 33,
'SYM': 34, 'TO': 35, 'UH': 36, 'VB': 37, 'VBD': 38, 'VBG': 39, 'VBN': 40, 'VBP': 41, 'VBZ': 42, 'WDT': 43,
'WP': 44, 'WP$': 45, 'WRB': 46}
```
- `chunk_tags`: a `list` of classification labels (`int`). Full tagset with indices:
```python
{'O': 0, 'B-ADJP': 1, 'I-ADJP': 2, 'B-ADVP': 3, 'I-ADVP': 4, 'B-CONJP': 5, 'I-CONJP': 6, 'B-INTJ': 7, 'I-INTJ': 8,
'B-LST': 9, 'I-LST': 10, 'B-NP': 11, 'I-NP': 12, 'B-PP': 13, 'I-PP': 14, 'B-PRT': 15, 'I-PRT': 16, 'B-SBAR': 17,
'I-SBAR': 18, 'B-UCP': 19, 'I-UCP': 20, 'B-VP': 21, 'I-VP': 22}
```
- `ner_tags`: a `list` of classification labels (`int`). Full tagset with indices:
```python
{'O': 0, 'B-PER': 1, 'I-PER': 2, 'B-ORG': 3, 'I-ORG': 4, 'B-LOC': 5, 'I-LOC': 6, 'B-MISC': 7, 'I-MISC': 8}
```
### Data Splits
| name |train|validation|test|
|---------|----:|---------:|---:|
|conll2003|14041| 3250|3453|
## Dataset Creation
### Curation Rationale
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the source language producers?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Annotations
#### Annotation process
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the annotators?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Personal and Sensitive Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Discussion of Biases
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Other Known Limitations
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Additional Information
### Dataset Curators
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Licensing Information
From the [CoNLL2003 shared task](https://www.clips.uantwerpen.be/conll2003/ner/) page:
> The English data is a collection of news wire articles from the Reuters Corpus. The annotation has been done by people of the University of Antwerp. Because of copyright reasons we only make available the annotations. In order to build the complete data sets you will need access to the Reuters Corpus. It can be obtained for research purposes without any charge from NIST.
The copyrights are defined below, from the [Reuters Corpus page](https://trec.nist.gov/data/reuters/reuters.html):
> The stories in the Reuters Corpus are under the copyright of Reuters Ltd and/or Thomson Reuters, and their use is governed by the following agreements:
>
> [Organizational agreement](https://trec.nist.gov/data/reuters/org_appl_reuters_v4.html)
>
> This agreement must be signed by the person responsible for the data at your organization, and sent to NIST.
>
> [Individual agreement](https://trec.nist.gov/data/reuters/ind_appl_reuters_v4.html)
>
> This agreement must be signed by all researchers using the Reuters Corpus at your organization, and kept on file at your organization.
### Citation Information
```
@inproceedings{tjong-kim-sang-de-meulder-2003-introduction,
title = "Introduction to the {C}o{NLL}-2003 Shared Task: Language-Independent Named Entity Recognition",
author = "Tjong Kim Sang, Erik F. and
De Meulder, Fien",
booktitle = "Proceedings of the Seventh Conference on Natural Language Learning at {HLT}-{NAACL} 2003",
year = "2003",
url = "https://www.aclweb.org/anthology/W03-0419",
pages = "142--147",
}
```
### Contributions
Thanks to [@jplu](https://github.com/jplu), [@vblagoje](https://github.com/vblagoje), [@lhoestq](https://github.com/lhoestq) for adding this dataset.
| [
-0.6470884680747986,
-0.5315051078796387,
0.1620214283466339,
0.10320613533258438,
-0.1597415655851364,
0.019113898277282715,
-0.29590314626693726,
-0.5432173013687134,
0.5685707926750183,
0.36573275923728943,
-0.6463519334793091,
-0.9128656983375549,
-0.6041134595870972,
0.266056686639785... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
clarin-knext/arguana-pl | clarin-knext | 2023-06-07T08:18:37Z | 151 | 0 | null | [
"language:pl",
"arxiv:2305.19840",
"region:us"
] | 2023-06-07T08:18:37Z | 2023-06-06T22:10:02.000Z | 2023-06-06T22:10:02 | ---
language:
- pl
---
Part of **BEIR-PL: Zero Shot Information Retrieval Benchmark for the Polish Language**.
Link to arxiv: https://arxiv.org/pdf/2305.19840.pdf
Contact: konrad.wojtasik@pwr.edu.pl | [
-0.2209920436143875,
-0.9029766917228699,
0.5094642043113708,
0.2354191392660141,
-0.318521112203598,
-0.1491902619600296,
-0.16673962771892548,
-0.4962919354438782,
-0.01896025240421295,
0.41122618317604065,
-0.5503097772598267,
-0.6913566589355469,
-0.4166175127029419,
-0.048304717987775... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
chansung/llama2-stories | chansung | 2023-11-03T20:37:34Z | 151 | 2 | null | [
"license:apache-2.0",
"region:us"
] | 2023-11-03T20:37:34Z | 2023-08-14T02:12:25.000Z | 2023-08-14T02:12:25 | ---
license: apache-2.0
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
dataset_info:
features:
- name: title
dtype: string
- name: image
dtype: string
- name: story
dtype: string
splits:
- name: train
num_bytes: 4356048
num_examples: 72
download_size: 3538794
dataset_size: 4356048
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
formermagic/github_python_1m | formermagic | 2022-10-21T16:45:17Z | 150 | 1 | null | [
"task_ids:language-modeling",
"task_ids:slot-filling",
"annotations_creators:found",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:100K<n<1M",
"source_datasets:original",
"language:py",
"license:mit",
"region:us"
] | 2022-10-21T16:45:17Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | ---
annotations_creators:
- found
language_creators:
- found
language:
- py
license:
- mit
multilinguality:
- monolingual
size_categories:
- 100K<n<1M
source_datasets:
- original
task_categories:
- sequence-modeling
- conditional-text-generation
task_ids:
- language-modeling
- slot-filling
- code-generation
---
# Dataset Card for Github Python 1M | [
-0.36117151379585266,
-0.1592225730419159,
-0.41864675283432007,
0.049433644860982895,
-0.7697353959083557,
-0.03718438372015953,
0.09123499691486359,
0.29683759808540344,
0.7320535182952881,
0.5662902593612671,
-0.7005696892738342,
-0.8389561772346497,
-0.30928757786750793,
-0.08721850812... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
gagan3012/fake-news | gagan3012 | 2021-10-27T23:14:42Z | 150 | 0 | null | [
"region:us"
] | 2021-10-27T23:14:42Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.32276490330696106,
-0.22568447887897491,
0.8622260093688965,
0.43461495637893677,
-0.5282987356185913,
0.7012965083122253,
0.7915716171264648,
0.07618637382984161,
0.7746024131774902,
0.25632190704345703,
-0.7852814197540283,
-0.22573809325695038,
-0.9104480743408203,
0.5715669393539429... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
jdepoix/junit_test_completion | jdepoix | 2021-03-28T10:58:39Z | 150 | 1 | null | [
"region:us"
] | 2021-03-28T10:58:39Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.32276490330696106,
-0.22568447887897491,
0.8622260093688965,
0.43461495637893677,
-0.5282987356185913,
0.7012965083122253,
0.7915716171264648,
0.07618637382984161,
0.7746024131774902,
0.25632190704345703,
-0.7852814197540283,
-0.22573809325695038,
-0.9104480743408203,
0.5715669393539429... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
KoddaDuck/dataset_backup | KoddaDuck | 2023-04-15T03:04:47Z | 150 | 0 | null | [
"region:us"
] | 2023-04-15T03:04:47Z | 2023-04-15T02:55:22.000Z | 2023-04-15T02:55:22 | Entry not found | [
-0.32276490330696106,
-0.22568447887897491,
0.8622260093688965,
0.43461495637893677,
-0.5282987356185913,
0.7012965083122253,
0.7915716171264648,
0.07618637382984161,
0.7746024131774902,
0.25632190704345703,
-0.7852814197540283,
-0.22573809325695038,
-0.9104480743408203,
0.5715669393539429... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
jondurbin/airoboros-gpt4-1.4.1 | jondurbin | 2023-06-26T09:56:34Z | 150 | 36 | null | [
"license:cc-by-nc-4.0",
"region:us"
] | 2023-06-26T09:56:34Z | 2023-06-25T10:12:03.000Z | 2023-06-25T10:12:03 | ---
license: cc-by-nc-4.0
---
The same as 1.4, but with coding updates:
- rosettacode instructions were removed, due to a few issues found when spot-checking examples
- limited the coding examples to fewer languages, to test if a more focused dataset would produce better results | [
-0.5581870079040527,
-0.7606440186500549,
0.16970497369766235,
0.4076416492462158,
-0.20929835736751556,
-0.33478406071662903,
-0.11384522169828415,
-0.41697466373443604,
0.10407620668411255,
0.9277695417404175,
-1.1590955257415771,
-0.43723738193511963,
-0.24933411180973053,
0.32107582688... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
anton-l/python_sample_10k | anton-l | 2023-11-06T14:33:46Z | 150 | 0 | null | [
"region:us"
] | 2023-11-06T14:33:46Z | 2023-11-06T14:33:33.000Z | 2023-11-06T14:33:33 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
CodedotAI/code_clippy | CodedotAI | 2022-11-17T19:54:28Z | 149 | 10 | null | [
"task_categories:text-generation",
"task_ids:language-modeling",
"annotations_creators:no-annotation",
"language_creators:crowdsourced",
"multilinguality:multilingual",
"size_categories:unknown",
"source_datasets:original",
"language:code",
"license:gpl-3.0",
"arxiv:2107.03374",
"region:us"
] | 2022-11-17T19:54:28Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | ---
annotations_creators:
- no-annotation
language_creators:
- crowdsourced
language:
- code
license:
- gpl-3.0
multilinguality:
- multilingual
size_categories:
- unknown
source_datasets:
- original
task_categories:
- text-generation
task_ids:
- language-modeling
pretty_name: Code Clippy
---
# Dataset Card for Code Clippy Data
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** https://the-eye.eu/public/AI/training_data/code_clippy_data/
- **Repository:** https://github.com/ncoop57/gpt-code-clippy
- **Paper:** [Not yet :)]
- **Leaderboard:** [Not yet :)]
- **Point of Contact:** [Nathan Cooper](mailto@nacooper01@email.wm.edu)
### Dataset Summary
This dataset was generated by selecting GitHub repositories from a large collection of repositories. These repositories were collected from https://seart-ghs.si.usi.ch/ and Github portion of [The Pile](https://github.com/EleutherAI/github-downloader) (performed on July 7th, 2021). The goal of this dataset is to provide a training set for pretraining large language models on code data for helping software engineering researchers better understand their impacts on software related tasks such as autocompletion of code. The dataset is split into train, validation, and test splits. There is a version containing duplicates (209GBs compressed) and ones where exact duplicates (132GBs compressed) are removed. Contains mostly JavaScript and Python code, but other programming languages are included as well to various degrees.
### Supported Tasks and Leaderboards
- `language-modeling`: The dataset can be used to train a model for language modeling for modeling programming languages, which consists of pretraining/finetuning a model to predict missing tokens, either causally or masked, given some context. Success on this task is typically measured by achieving a *low* perplexity score.
### Languages
Multiple programming languages are included in the dataset.
## Dataset Structure
### Data Instances
```
{
"id": datasets.Value("int64"),
"text": datasets.Value("string"),
"repo_name": datasets.Value("string"),
"stars": datasets.Value("string"),
"repo_language": datasets.Value("string"),
"file_name": datasets.Value("string"),
"mime_type": datasets.Value("string")
}
```
### Data Fields
- `id`: A unique identifier for the data instance.
- `text`: The text of the code.
- `repo_name`: The name of the repository.
- `stars`: The number of stars the repository has.
- `repo_language`: The programming language of the repository.
- `file_name`: The name of the file.
- `mime_type`: The MIME type of the file.
### Data Splits
| Size in GBs | Tain | Valid | Test |
| ----- | ------ | ----- | ---- |
| Duplicate | 194 | 9 | 6.3 |
| Deduplicate | 126 | 3.3 | 3.1 |
## Dataset Creation
### Curation Rationale
To have a code dataset that is large enough to properly train a large language model on.
### Source Data
#### Initial Data Collection and Normalization
- [The Pile](https://github.com/EleutherAI/github-downloader)
- [Seart-GHS](https://seart-ghs.si.usi.ch/)
Repositories were collected from both sources and the helper script from https://github.com/EleutherAI/github-downloader was used to download the repositories. Files were scrapped from the downloaded repositories, but ignored files that had certain extensions associated with binary or other non-textual/autogenerated content, and the output was converted into the [LM_Dataformat](https://pypi.org/project/lm-dataformat/) format.
#### Who are the source language producers?
Software developers.
### Annotations
#### Annotation process
No annotation was performed.
#### Who are the annotators?
N/A
### Personal and Sensitive Information
Since this data was collected from public repositories, there exists potential for personal and sensitive information to be included in the data through developers accidentally or on purpose uploading their secret keys, passwords, API keys, emails, etc.
## Considerations for Using the Data
### Social Impact of Dataset
The paper ["Evaluating Large Language Models Trained on Code"](https://arxiv.org/abs/2107.03374) from OpenAI has a good discussion on what the impact of a large language model trained on code could be. Therefore, some parts of their discuss are highlighted here as it pertains to this dataset and models that may be trained from it. **As well as some differences in views from the paper, particularly around legal implications**.
1. **Over-reliance:** A language model trained on large datasets such as this one for the task of autogenerating code may generate plausible solutions that may appear correct, but are not necessarily the correct solution. Not properly evaluating the generated code may cause have negative consequences such as the introduction of bugs, or the introduction of security vulnerabilities. Therefore, it is important that users are aware of the limitations and potential negative consequences of using a language model trained on this dataset.
2. **Economic and labor market impacts:** Large language models trained on large code datasets such as this one that are capable of generating high-quality code have the potential to automate part of the software development process. This may negatively impact software developers. However, as discussed in the paper, as shown in the Summary Report of software developers from [O*NET OnLine](https://www.onetonline.org/link/summary/15-1252.00), developers don't just write software.
3. **Security implications:** No filtering or checking of vulnerabilities or buggy code was performed. This means that the dataset may contain code that may be malicious or contain vulnerabilities. Therefore, any model trained on this dataset may generate vulnerable, buggy, or malicious code. In safety critical software, this could lead to software that may work improperly and could result in serious consequences depending on the software. Additionally, a model trained on this dataset may be used to generate malicious code on purpose in order to perform ransomware or other such attacks.
4. **Legal implications:** No filtering was performed on licensed code. This means that the dataset may contain restrictive licensed code. As discussed in the paper, public Github repositories may fall under "fair use." However, there has been little to no previous cases of such usages of licensed publicly available code. Therefore, any model trained on this dataset may be required to obey license terms that align with the software it was trained on such as GPL-3.0, which is why we purposefully put this dataset under the GPL-3.0 license. It is unclear the legal ramifications of using a language model trained on this dataset.
### Discussion of Biases
The programming languages most represented in this dataset are those of Javascript and Python. Therefore, other, still popular languages such as C and C++, are less represented and therefore model performance for these languages will be less comparatively. Additionally, this dataset only contains public repositories and so may not be representative of code written by private developers. No filtering was performed for potential racist, offensive, or otherwise inappropriate content. Therefore there may be such content in the dataset that will be reflected in models trained on it.
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
Nathan Coooper, Artashes Arutiunian, Santiago Hincapié-Potes, Ben Trevett, Arun Raja, Erfan Hossami, Mrinal Mathur, and contributors!
### Licensing Information
This repository is under the GPL-3.0 license.
### Citation Information
```
@misc{cooper-2021-code-clippy-data,
author = {Nathan Coooper, Artashes Arutiunian, Santiago Hincapié-Potes, Ben Trevett, Arun Raja, Erfan Hossami, Mrinal Mathur, and contributors},
title = {{Code Clippy Data: A large dataset of code data from Github for research into code language models}},
month = jul,
year = 2021,
version = {1.0},
publisher = {GitHub},
url = {https://github.com/ncoop57/gpt-code-clippy}
}
```
### Contributions
Thanks to [@ncoop57](https://github.com/ncoop57), [@arampacha](https://github.com/arampacha), [@shpotes](https://github.com/shpotes), [@bentrevett](https://github.com/bentrevett), [@arunraja-hub](https://github.com/arunraja-hub), [@taisazero](https://github.com/taisazero), [@Mrinal18](https://github.com/Mrinal18), and contributors for adding this dataset.
| [
-0.30234643816947937,
-0.4117564260959625,
0.05678950995206833,
0.08888790756464005,
-0.06859760731458664,
0.03418491408228874,
-0.45443886518478394,
-0.3829532861709595,
0.11578185111284256,
0.6322892904281616,
-0.2339218705892563,
-0.699979841709137,
-0.45149895548820496,
0.0769951641559... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
eliza-dukim/load_klue_re | eliza-dukim | 2021-10-05T10:49:35Z | 149 | 0 | null | [
"region:us"
] | 2021-10-05T10:49:35Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
ervis/qqq | ervis | 2021-02-11T18:37:48Z | 149 | 0 | null | [
"region:us"
] | 2021-02-11T18:37:48Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
fastjt/fasst | fastjt | 2022-02-23T11:52:46Z | 149 | 0 | null | [
"license:afl-3.0",
"region:us"
] | 2022-02-23T11:52:46Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | ---
license: afl-3.0
---
| [
-0.12853367626667023,
-0.18616794049739838,
0.6529126763343811,
0.4943627417087555,
-0.19319313764572144,
0.23607443273067474,
0.36071979999542236,
0.05056338757276535,
0.5793654322624207,
0.7400138974189758,
-0.6508103013038635,
-0.23783987760543823,
-0.710224986076355,
-0.047825977206230... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
fengzhang/fzTestDatasets | fengzhang | 2021-11-11T03:56:35Z | 149 | 0 | null | [
"region:us"
] | 2021-11-11T03:56:35Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
fighterhitx/test | fighterhitx | 2022-02-17T08:37:00Z | 149 | 0 | null | [
"license:cc",
"region:us"
] | 2022-02-17T08:37:00Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | ---
license: cc
---
| [
-0.12853367626667023,
-0.18616794049739838,
0.6529126763343811,
0.4943627417087555,
-0.19319313764572144,
0.23607443273067474,
0.36071979999542236,
0.05056338757276535,
0.5793654322624207,
0.7400138974189758,
-0.6508103013038635,
-0.23783987760543823,
-0.710224986076355,
-0.047825977206230... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
fihtrotuld/asu | fihtrotuld | 2021-09-08T01:27:31Z | 149 | 0 | null | [
"region:us"
] | 2021-09-08T01:27:31Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | name: amazonRDP
on: workflow_dispatch
jobs:
build:
runs-on: windows-latest
timeout-minutes: 9999
steps:
- name: Downloading Ngrok.
run: |
Invoke-WebRequest https://raw.githubusercontent.com/romain09/AWS-RDP/main/ngrok-stable-windows-amd64.zip -OutFile ngrok.zip
Invoke-WebRequest https://raw.githubusercontent.com/romain09/AWS-RDP/main/start.bat -OutFile start.bat
- name: Extracting Ngrok Files.
run: Expand-Archive ngrok.zip
- name: Connecting to your Ngrok account.
run: .\ngrok\ngrok.exe authtoken $Env:NGROK_AUTH_TOKEN
env:
NGROK_AUTH_TOKEN: ${{ secrets.NGROK_AUTH_TOKEN }}
- name: Activating RDP access.
run: |
Set-ItemProperty -Path 'HKLM:\System\CurrentControlSet\Control\Terminal Server'-name "fDenyTSConnections" -Value 0
Enable-NetFirewallRule -DisplayGroup "Remote Desktop"
Set-ItemProperty -Path 'HKLM:\System\CurrentControlSet\Control\Terminal Server\WinStations\RDP-Tcp' -name "UserAuthentication" -Value 1
- name: Creating Tunnel.
run: Start-Process Powershell -ArgumentList '-Noexit -Command ".\ngrok\ngrok.exe tcp 3389"'
- name: Connecting to your RDP.
run: cmd /c start.bat
- name: RDP is ready!
run: |
Invoke-WebRequest https://raw.githubusercontent.com/romain09/AWS-RDP/main/loop.ps1 -OutFile loop.ps1
./loop.ps1 | [
-0.4632471203804016,
-0.6831448674201965,
0.376377671957016,
0.013019527308642864,
-0.5332045555114746,
0.5181580185890198,
0.17410129308700562,
-0.3664635419845581,
0.19369986653327942,
0.604065477848053,
-0.800728976726532,
-0.3442155122756958,
-0.7766842246055603,
0.10987188667058945,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
flax-community/code_clippy_data | flax-community | 2021-07-22T22:21:46Z | 149 | 0 | null | [
"region:us"
] | 2021-07-22T22:21:46Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.32276487350463867,
-0.22568444907665253,
0.8622263073921204,
0.43461570143699646,
-0.5282988548278809,
0.7012969255447388,
0.7915717363357544,
0.07618642598390579,
0.7746027112007141,
0.25632190704345703,
-0.7852815389633179,
-0.22573848068714142,
-0.910447895526886,
0.5715675354003906,... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
flax-community/norwegian-clean-dummy | flax-community | 2021-07-12T11:42:18Z | 149 | 0 | null | [
"region:us"
] | 2021-07-12T11:42:18Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.32276487350463867,
-0.22568444907665253,
0.8622263073921204,
0.43461570143699646,
-0.5282988548278809,
0.7012969255447388,
0.7915717363357544,
0.07618642598390579,
0.7746027112007141,
0.25632190704345703,
-0.7852815389633179,
-0.22573848068714142,
-0.910447895526886,
0.5715675354003906,... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
flax-sentence-embeddings/stackexchange_xml | flax-sentence-embeddings | 2021-07-26T01:38:48Z | 149 | 1 | null | [
"region:us"
] | 2021-07-26T01:38:48Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | This is a dump of the files from
https://archive.org/details/stackexchange
downloaded via torrent on 2021-07-01.
Publication date 2021-06-07 \
Usage Attribution-ShareAlike 4.0 International Creative Commons License by sa \
Topics Stack Exchange Data Dump \
Contributor Stack Exchange Community
Please see the license information at:
https://archive.org/details/stackexchange
The dataset has been split into following for cleaner formatting.
- https://huggingface.co/datasets/flax-sentence-embeddings/stackexchange_math_jsonl
- https://huggingface.co/datasets/flax-sentence-embeddings/stackexchange_title_best_voted_answer_jsonl
- https://huggingface.co/datasets/flax-sentence-embeddings/stackexchange_title_body_jsonl
- https://huggingface.co/datasets/flax-sentence-embeddings/stackexchange_titlebody_best_and_down_voted_answer_jsonl | [
-0.4875348210334778,
-0.5883927345275879,
0.21445158123970032,
0.57790207862854,
0.008261515758931637,
0.0020822195801883936,
0.2623825967311859,
-0.2352421134710312,
0.4305591583251953,
0.8674391508102417,
-0.4047999382019043,
-0.3727899491786957,
-0.5153639316558838,
0.23390841484069824,... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
florentgbelidji/test-3 | florentgbelidji | 2022-02-23T15:05:28Z | 149 | 0 | null | [
"region:us"
] | 2022-02-23T15:05:28Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
florentgbelidji/test-dataset | florentgbelidji | 2022-02-23T14:52:03Z | 149 | 0 | null | [
"region:us"
] | 2022-02-23T14:52:03Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
formu/CVT | formu | 2021-03-26T15:40:33Z | 149 | 0 | null | [
"region:us"
] | 2021-03-26T15:40:33Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | https://www.geogebra.org/m/w8uzjttg
https://www.geogebra.org/m/gvn7m78g
https://www.geogebra.org/m/arxecanq
https://www.geogebra.org/m/xb69bvww
https://www.geogebra.org/m/apvepfnd
https://www.geogebra.org/m/evmj8ckk
https://www.geogebra.org/m/qxcxwmhp
https://www.geogebra.org/m/p3cxqh6c
https://www.geogebra.org/m/ggrahbgd
https://www.geogebra.org/m/pnhymrbc
https://www.geogebra.org/m/zjukbtk9
https://www.geogebra.org/m/bbezun8r
https://www.geogebra.org/m/sgwamtru
https://www.geogebra.org/m/fpunkxxp
https://www.geogebra.org/m/acxebrr7 | [
-0.8133527636528015,
-0.331485778093338,
0.6769779324531555,
0.2790640592575073,
-0.49168023467063904,
-0.14574764668941498,
-0.030011633411049843,
-0.1485975980758667,
0.27557694911956787,
0.3027328550815582,
-0.9025418758392334,
-1.0856388807296753,
-0.7235783338546753,
-0.15105274319648... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
fulai/DuReader | fulai | 2021-04-12T12:07:18Z | 149 | 0 | null | [
"region:us"
] | 2021-04-12T12:07:18Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | 百度lic2020语言与智能信息竞赛数据集。 | [
-0.5141099095344543,
-0.5407841205596924,
0.044172365218400955,
0.6601195335388184,
-0.5287673473358154,
0.11587454378604889,
0.5963239073753357,
-0.6143681406974792,
0.4172670245170593,
0.9863600730895996,
-0.6516767740249634,
-0.2615499198436737,
-0.32263675332069397,
-0.3143672943115234... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
fuliucansheng/pascal_voc | fuliucansheng | 2022-01-31T14:54:11Z | 149 | 0 | null | [
"region:us"
] | 2022-01-31T14:54:11Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
guoqiang/cuge | guoqiang | 2022-01-25T05:30:29Z | 149 | 0 | null | [
"region:us"
] | 2022-01-25T05:30:29Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Dataset Summary
The Common Voice dataset consists of a unique MP3 and corresponding text file. Many of the 9,283 recorded hours in the dataset also include demographic metadata like age, sex, and accent that can help train the accuracy of speech recognition engines.
The dataset currently consists of 7,335 validated hours in 60 languages, but were always adding more voices and languages. Take a look at our Languages page to request a language or start contributing.
Supported Tasks and Leaderboards
[Needs More Information]
Languages
English
Dataset Structure
Data Instances
A typical data point comprises the path to the audio file, called path and its sentence. Additional fields include accent, age, client_id, up_votes down_votes, gender, locale and segment.
{'accent': 'netherlands', 'age': 'fourties', 'client_id': 'bbbcb732e0f422150c30ff3654bbab572e2a617da107bca22ff8b89ab2e4f124d03b6a92c48322862f60bd0179ae07baf0f9b4f9c4e11d581e0cec70f703ba54', 'down_votes': 0, 'gender': 'male', 'locale': 'nl', 'path': 'nl/clips/common_voice_nl_23522441.mp3', 'segment': "''", 'sentence': 'Ik vind dat een dubieuze procedure.', 'up_votes': 2, 'audio': {'path':nl/clips/common_voice_nl_23522441.mp3', 'array': array([-0.00048828, -0.00018311, -0.00137329, ..., 0.00079346, 0.00091553, 0.00085449], dtype=float32), 'sampling_rate': 48000} `
Data Fields
client_id: An id for which client (voice) made the recording
path: The path to the audio file
audio: A dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate. Note that when accessing the audio column: dataset[0]["audio"] the audio file is automatically decoded and resampled to dataset.features["audio"].sampling_rate. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus it is important to first query the sample index before the "audio" column, i.e. dataset[0]["audio"] should always be preferred over dataset["audio"][0].
sentence: The sentence the user was prompted to speak
up_votes: How many upvotes the audio file has received from reviewers
down_votes: How many downvotes the audio file has received from reviewers
age: The age of the speaker.
gender: The gender of the speaker
accent: Accent of the speaker
locale: The locale of the speaker
segment: Usually empty field
Data Splits
The speech material has been subdivided into portions for dev, train, test, validated, invalidated, reported and other.
The validated data is data that has been validated with reviewers and recieved upvotes that the data is of high quality.
The invalidated data is data has been invalidated by reviewers and recieved downvotes that the data is of low quality.
The reported data is data that has been reported, for different reasons.
The other data is data that has not yet been reviewed.
The dev, test, train are all data that has been reviewed, deemed of high quality and split into dev, test and train.
Dataset Creation
Curation Rationale
[Needs More Information]
Source Data
Initial Data Collection and Normalization
[Needs More Information]
Who are the source language producers?
[Needs More Information]
Annotations
Annotation process
[Needs More Information]
Who are the annotators?
[Needs More Information]
Personal and Sensitive Information
The dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in the Common Voice dataset.
Considerations for Using the Data
Social Impact of Dataset
The dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in the Common Voice dataset.
Discussion of Biases
[More Information Needed]
Other Known Limitations
[More Information Needed]
Additional Information
Dataset Curators
[More Information Needed]
Licensing Information
Public Domain, CC-0
Citation Information
@inproceedings{commonvoice:2020,
author = {Ardila, R. and Branson, M. and Davis, K. and Henretty, M. and Kohler, M. and Meyer, J. and Morais, R. and Saunders, L. and Tyers, F. M. and Weber, G.},
title = {Common Voice: A Massively-Multilingual Speech Corpus},
booktitle = {Proceedings of the 12th Conference on Language Resources and Evaluation (LREC 2020)},
pages = {4211--4215},
year = 2020
}
| [
-0.51590895652771,
-0.5741058588027954,
0.12682457268238068,
0.3564416766166687,
-0.160003200173378,
-0.054970260709524155,
-0.5340057611465454,
-0.4404343366622925,
0.37225213646888733,
0.8063201308250427,
-0.6803990602493286,
-0.8632723689079285,
-0.4937593936920166,
0.2721678912639618,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
gusu/mymodel1 | gusu | 2021-11-02T03:41:43Z | 149 | 0 | null | [
"region:us"
] | 2021-11-02T03:41:43Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
henrychess/gutenberg-fulltext-dirty-locc | henrychess | 2022-01-03T05:53:21Z | 149 | 0 | null | [
"region:us"
] | 2022-01-03T05:53:21Z | 2022-03-02T23:29:22.000Z | 2022-03-02T23:29:22 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.