id stringlengths 2 115 | author stringlengths 2 42 ⌀ | last_modified timestamp[us, tz=UTC] | downloads int64 0 8.87M | likes int64 0 3.84k | paperswithcode_id stringlengths 2 45 ⌀ | tags list | lastModified timestamp[us, tz=UTC] | createdAt stringlengths 24 24 | key stringclasses 1 value | created timestamp[us] | card stringlengths 1 1.01M | embedding list | library_name stringclasses 21 values | pipeline_tag stringclasses 27 values | mask_token null | card_data null | widget_data null | model_index null | config null | transformers_info null | spaces null | safetensors null | transformersInfo null | modelId stringlengths 5 111 ⌀ | embeddings list |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
datablations/oscar-filter-small | datablations | 2022-11-24T11:45:03Z | 12 | 0 | null | [
"region:us"
] | 2022-11-24T11:45:03Z | 2022-11-24T11:44:37.000Z | 2022-11-24T11:44:37 | ---
dataset_info:
features:
- name: id
dtype: int64
- name: text
dtype: string
- name: meta
struct:
- name: annotations
sequence: string
- name: identification
struct:
- name: label
dtype: string
- name: prob
dtype: float64
- name: line_identifications
list:
- name: label
dtype: string
- name: prob
dtype: float64
- name: perplexity_score
dtype: float64
- name: warc_headers
struct:
- name: content-length
dtype: int64
- name: content-type
dtype: string
- name: warc-block-digest
dtype: string
- name: warc-date
dtype: string
- name: warc-identified-content-language
dtype: string
- name: warc-record-id
dtype: string
- name: warc-refers-to
dtype: string
- name: warc-target-uri
dtype: string
- name: warc-type
dtype: string
splits:
- name: train
num_bytes: 658480427
num_examples: 100000
download_size: 347756473
dataset_size: 658480427
---
# Dataset Card for "small-oscar"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.7069371938705444,
-0.03961409255862236,
0.32713785767555237,
-0.06372931599617004,
-0.3681672513484955,
-0.15735524892807007,
0.29250603914260864,
-0.248735710978508,
1.1664137840270996,
0.3690994679927826,
-0.6706997752189636,
-0.4669804871082306,
-0.7546698451042175,
-0.13423164188861... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
supermy/poetry | supermy | 2022-11-24T23:06:59Z | 12 | 0 | null | [
"region:us"
] | 2022-11-24T23:06:59Z | 2022-11-24T13:23:19.000Z | 2022-11-24T13:23:19 | Entry not found | [
-0.3227647542953491,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965083122253,
0.7915717959403992,
0.07618629932403564,
0.7746022343635559,
0.2563222348690033,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
royam0820/baya-paintings-01 | royam0820 | 2022-11-24T14:03:38Z | 12 | 0 | null | [
"license:afl-3.0",
"region:us"
] | 2022-11-24T14:03:38Z | 2022-11-24T13:56:57.000Z | 2022-11-24T13:56:57 | ---
license: afl-3.0
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
jeapaul/europarl_bilingual_processed | jeapaul | 2022-11-24T18:30:24Z | 12 | 0 | null | [
"region:us"
] | 2022-11-24T18:30:24Z | 2022-11-24T18:30:02.000Z | 2022-11-24T18:30:02 | ---
dataset_info:
features:
- name: text
dtype: string
splits:
- name: train
num_bytes: 281100121
num_examples: 1892723
download_size: 155904108
dataset_size: 281100121
---
# Dataset Card for "europarl_bilingual_processed"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.5238747000694275,
-0.38309088349342346,
0.23588129878044128,
0.4843233525753021,
-0.14523230493068695,
0.06301659345626831,
-0.07672297209501266,
-0.24500872194766998,
0.8308473229408264,
0.6596837043762207,
-0.9107206463813782,
-0.9970584511756897,
-0.6896096467971802,
0.04774861410260... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
ashraq/tmdb-celeb-10k | ashraq | 2022-11-24T18:48:43Z | 12 | 0 | null | [
"region:us"
] | 2022-11-24T18:48:43Z | 2022-11-24T18:47:53.000Z | 2022-11-24T18:47:53 | Entry not found | [
-0.3227647542953491,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965083122253,
0.7915717959403992,
0.07618629932403564,
0.7746022343635559,
0.2563222348690033,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
teticio/imdb-posters-and-description-512 | teticio | 2022-11-24T23:39:16Z | 12 | 0 | null | [
"region:us"
] | 2022-11-24T23:39:16Z | 2022-11-24T19:05:17.000Z | 2022-11-24T19:05:17 | Entry not found | [
-0.3227647542953491,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965083122253,
0.7915717959403992,
0.07618629932403564,
0.7746022343635559,
0.2563222348690033,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
shreyasharma/proofs | shreyasharma | 2022-11-24T21:45:28Z | 12 | 0 | null | [
"region:us"
] | 2022-11-24T21:45:28Z | 2022-11-24T21:45:22.000Z | 2022-11-24T21:45:22 | ---
dataset_info:
features:
- name: intermediate_conclusions
struct:
- name: int1
dtype: string
- name: int10
dtype: string
- name: int11
dtype: string
- name: int12
dtype: string
- name: int13
dtype: string
- name: int14
dtype: string
- name: int15
dtype: string
- name: int16
dtype: string
- name: int17
dtype: string
- name: int2
dtype: string
- name: int3
dtype: string
- name: int4
dtype: string
- name: int5
dtype: string
- name: int6
dtype: string
- name: int7
dtype: string
- name: int8
dtype: string
- name: int9
dtype: string
- name: step_proof
dtype: string
- name: triples
struct:
- name: sent1
dtype: string
- name: sent10
dtype: string
- name: sent11
dtype: string
- name: sent12
dtype: string
- name: sent13
dtype: string
- name: sent14
dtype: string
- name: sent15
dtype: string
- name: sent16
dtype: string
- name: sent17
dtype: string
- name: sent2
dtype: string
- name: sent3
dtype: string
- name: sent4
dtype: string
- name: sent5
dtype: string
- name: sent6
dtype: string
- name: sent7
dtype: string
- name: sent8
dtype: string
- name: sent9
dtype: string
- name: hypothesis
dtype: string
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: train
num_bytes: 1296774
num_examples: 1313
download_size: 609276
dataset_size: 1296774
---
# Dataset Card for "proofs"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.5178183317184448,
-0.10430160909891129,
0.44106972217559814,
0.2956085503101349,
-0.23853382468223572,
-0.09676513075828552,
0.07267078757286072,
-0.19556759297847748,
0.7234905958175659,
0.45367273688316345,
-0.6009597182273865,
-0.8233897686004639,
-0.6894094944000244,
-0.262038141489... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
shreyasharma/proofs2 | shreyasharma | 2022-11-24T21:54:13Z | 12 | 0 | null | [
"region:us"
] | 2022-11-24T21:54:13Z | 2022-11-24T21:54:07.000Z | 2022-11-24T21:54:07 | ---
dataset_info:
features:
- name: intermediate_conclusions
struct:
- name: int1
dtype: string
- name: int10
dtype: string
- name: int11
dtype: string
- name: int12
dtype: string
- name: int13
dtype: string
- name: int14
dtype: string
- name: int15
dtype: string
- name: int16
dtype: string
- name: int17
dtype: string
- name: int2
dtype: string
- name: int3
dtype: string
- name: int4
dtype: string
- name: int5
dtype: string
- name: int6
dtype: string
- name: int7
dtype: string
- name: int8
dtype: string
- name: int9
dtype: string
- name: step_proof
dtype: string
- name: triples
struct:
- name: sent1
dtype: string
- name: sent10
dtype: string
- name: sent11
dtype: string
- name: sent12
dtype: string
- name: sent13
dtype: string
- name: sent14
dtype: string
- name: sent15
dtype: string
- name: sent16
dtype: string
- name: sent17
dtype: string
- name: sent2
dtype: string
- name: sent3
dtype: string
- name: sent4
dtype: string
- name: sent5
dtype: string
- name: sent6
dtype: string
- name: sent7
dtype: string
- name: sent8
dtype: string
- name: sent9
dtype: string
- name: hypothesis
dtype: string
- name: question
dtype: string
- name: answer
dtype: string
- name: label
dtype: int64
splits:
- name: train
num_bytes: 1307278
num_examples: 1313
download_size: 609969
dataset_size: 1307278
---
# Dataset Card for "proofs2"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.2850024402141571,
0.011128208599984646,
0.3676005005836487,
0.25731343030929565,
-0.2994130253791809,
-0.19007661938667297,
0.282234251499176,
-0.201797217130661,
0.5868068933486938,
0.3588410019874573,
-0.5654723644256592,
-0.5529460310935974,
-0.7796205878257751,
-0.37427857518196106,... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
teticio/imdb-posters-and-description-256 | teticio | 2022-11-25T00:37:31Z | 12 | 0 | null | [
"region:us"
] | 2022-11-25T00:37:31Z | 2022-11-25T00:31:23.000Z | 2022-11-25T00:31:23 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
TTian/feedback-prize-tokenized-datasets | TTian | 2022-11-25T02:38:47Z | 12 | 0 | null | [
"region:us"
] | 2022-11-25T02:38:47Z | 2022-11-25T02:38:32.000Z | 2022-11-25T02:38:32 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
nzh324/green | nzh324 | 2022-11-25T03:57:16Z | 12 | 0 | null | [
"license:mit",
"region:us"
] | 2022-11-25T03:57:16Z | 2022-11-25T03:56:16.000Z | 2022-11-25T03:56:16 | ---
license: mit
---
| [
-0.12853367626667023,
-0.18616794049739838,
0.6529126763343811,
0.4943627417087555,
-0.19319313764572144,
0.23607443273067474,
0.36071979999542236,
0.05056338757276535,
0.5793654322624207,
0.7400138974189758,
-0.6508103013038635,
-0.23783987760543823,
-0.710224986076355,
-0.047825977206230... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
nzh324/pink | nzh324 | 2022-11-25T03:58:24Z | 12 | 0 | null | [
"license:mit",
"region:us"
] | 2022-11-25T03:58:24Z | 2022-11-25T03:57:43.000Z | 2022-11-25T03:57:43 | ---
license: mit
---
| [
-0.12853367626667023,
-0.18616794049739838,
0.6529126763343811,
0.4943627417087555,
-0.19319313764572144,
0.23607443273067474,
0.36071979999542236,
0.05056338757276535,
0.5793654322624207,
0.7400138974189758,
-0.6508103013038635,
-0.23783987760543823,
-0.710224986076355,
-0.047825977206230... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Abhilashvj/deliveryman | Abhilashvj | 2022-11-25T05:15:24Z | 12 | 0 | null | [
"license:apache-2.0",
"region:us"
] | 2022-11-25T05:15:24Z | 2022-11-25T05:14:14.000Z | 2022-11-25T05:14:14 | ---
license: apache-2.0
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Hiraa/Roman-urdu_reviews_and_summary | Hiraa | 2022-11-25T09:06:46Z | 12 | 0 | null | [
"region:us"
] | 2022-11-25T09:06:46Z | 2022-11-25T09:00:47.000Z | 2022-11-25T09:00:47 | Entry not found | [
-0.3227647542953491,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965083122253,
0.7915717959403992,
0.07618629932403564,
0.7746022343635559,
0.2563222348690033,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
shreyasharma/proofs3 | shreyasharma | 2022-11-25T09:55:08Z | 12 | 0 | null | [
"region:us"
] | 2022-11-25T09:55:08Z | 2022-11-25T09:55:04.000Z | 2022-11-25T09:55:04 | ---
dataset_info:
features:
- name: intermediate_conclusions
struct:
- name: int1
dtype: string
- name: int10
dtype: string
- name: int11
dtype: string
- name: int12
dtype: string
- name: int13
dtype: string
- name: int14
dtype: string
- name: int15
dtype: string
- name: int16
dtype: string
- name: int17
dtype: string
- name: int2
dtype: string
- name: int3
dtype: string
- name: int4
dtype: string
- name: int5
dtype: string
- name: int6
dtype: string
- name: int7
dtype: string
- name: int8
dtype: string
- name: int9
dtype: string
- name: step_proof
dtype: string
- name: triples
struct:
- name: sent1
dtype: string
- name: sent10
dtype: string
- name: sent11
dtype: string
- name: sent12
dtype: string
- name: sent13
dtype: string
- name: sent14
dtype: string
- name: sent15
dtype: string
- name: sent16
dtype: string
- name: sent17
dtype: string
- name: sent2
dtype: string
- name: sent3
dtype: string
- name: sent4
dtype: string
- name: sent5
dtype: string
- name: sent6
dtype: string
- name: sent7
dtype: string
- name: sent8
dtype: string
- name: sent9
dtype: string
- name: hypothesis
dtype: string
- name: question
dtype: string
- name: answer
dtype: string
- name: label
dtype: int64
splits:
- name: train
num_bytes: 2614556
num_examples: 2626
download_size: 1188057
dataset_size: 2614556
---
# Dataset Card for "proofs3"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.43329334259033203,
0.04677458107471466,
0.528128445148468,
0.3508748412132263,
-0.22898873686790466,
-0.19437545537948608,
0.3371061682701111,
-0.21716290712356567,
0.6331523060798645,
0.45073217153549194,
-0.6035435795783997,
-0.7302740216255188,
-0.668983519077301,
-0.1381047517061233... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Isma/librispeech_1000_seed_42 | Isma | 2022-11-28T14:52:52Z | 12 | 0 | null | [
"region:us"
] | 2022-11-28T14:52:52Z | 2022-11-28T14:51:41.000Z | 2022-11-28T14:51:41 | Entry not found | [
-0.3227647542953491,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965083122253,
0.7915717959403992,
0.07618629932403564,
0.7746022343635559,
0.2563222348690033,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
autoevaluate/autoeval-staging-eval-project-ecaf0dbc-43a3-4513-bbcf-d0f372522232-109106 | autoevaluate | 2022-11-29T14:08:27Z | 12 | 0 | null | [
"autotrain",
"evaluation",
"region:us"
] | 2022-11-29T14:08:27Z | 2022-11-29T14:07:50.000Z | 2022-11-29T14:07:50 | ---
type: predictions
tags:
- autotrain
- evaluation
datasets:
- glue
eval_info:
task: binary_classification
model: autoevaluate/binary-classification
metrics: ['matthews_correlation']
dataset_name: glue
dataset_config: sst2
dataset_split: validation
col_mapping:
text: sentence
target: label
---
# Dataset Card for AutoTrain Evaluator
This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset:
* Task: Binary Text Classification
* Model: autoevaluate/binary-classification
* Dataset: glue
* Config: sst2
* Split: validation
To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator).
## Contributions
Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model. | [
-0.20361582934856415,
-0.3338315784931183,
0.2989133596420288,
0.17618156969547272,
-0.1635434776544571,
0.036154989153146744,
0.020895475521683693,
-0.39217695593833923,
0.12184639275074005,
0.361812025308609,
-0.9186381101608276,
-0.21669884026050568,
-0.7705212831497192,
-0.013488127849... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
dferndz/cSQuAD2 | dferndz | 2022-12-09T23:18:39Z | 12 | 0 | null | [
"task_categories:question-answering",
"annotations_creators:expert-generated",
"language_creators:other",
"multilinguality:monolingual",
"language:en",
"license:apache-2.0",
"region:us"
] | 2022-12-09T23:18:39Z | 2022-11-30T00:49:11.000Z | 2022-11-30T00:49:11 | ---
annotations_creators:
- expert-generated
language:
- en
language_creators:
- other
license:
- apache-2.0
multilinguality:
- monolingual
pretty_name: cSQuAD2
size_categories: []
source_datasets: []
tags: []
task_categories:
- question-answering
task_ids: []
---
# Dataset Card for cSQuAD2
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks](#supported-tasks)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Additional Information](#additional-information)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
## Dataset Description
- **Homepage:**
- **Repository:**
- **Paper:**
- **Leaderboard:**
- **Point of Contact:**
### Dataset Summary
A contrast set to evaluate models trained on SQUAD on out-of-domain data.
### Supported Tasks
Evaluate question-answering
### Languages
English
## Dataset Structure
### Data Instances
Dataset contains 40 instances
### Data Fields
| Field | Description |
|----------|--------------------------------------------------
| id | Id of document containing context |
| title | Title of the document |
| context | The context of the question |
| question | The question to answer |
| answers | A list of possible answers from the context |
| answer_start | The index in context where the answer starts |
### Data Splits
A single `test` split is provided
## Dataset Creation
Dataset was created from Wikipedia articles
## Additional Information
### Licensing Information
Apache 2.0 license
### Citation Information
TODO: add citations | [
-0.4683828353881836,
-0.3925066590309143,
0.0775194838643074,
0.4525924324989319,
-0.11215277016162872,
0.24276849627494812,
-0.027346225455403328,
-0.22504030168056488,
0.01781017705798149,
0.23809358477592468,
-0.820673406124115,
-0.6517621278762817,
-0.32464179396629333,
-0.113795481622... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
yuvalkirstain/laion-hd-subset | yuvalkirstain | 2022-11-30T11:07:56Z | 12 | 0 | null | [
"region:us"
] | 2022-11-30T11:07:56Z | 2022-11-30T09:48:05.000Z | 2022-11-30T09:48:05 | ---
dataset_info:
features:
- name: similarity
dtype: float64
- name: hash
dtype: int64
- name: punsafe
dtype: float64
- name: pwatermark
dtype: float64
- name: LANGUAGE
dtype: string
- name: caption
dtype: string
- name: url
dtype: string
- name: key
dtype: string
- name: status
dtype: string
- name: error_message
dtype: 'null'
- name: width
dtype: int64
- name: height
dtype: int64
- name: original_width
dtype: int64
- name: original_height
dtype: int64
- name: exif
dtype: string
- name: md5
dtype: string
- name: path
dtype: string
- name: image
dtype: image
splits:
- name: train
num_bytes: 4395359106.2963705
num_examples: 13451
- name: test
num_bytes: 496904910.53063023
num_examples: 1495
download_size: 4890190248
dataset_size: 4892264016.827001
---
# Dataset Card for "laion-hd-subset"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.5386945009231567,
-0.2610115110874176,
0.22680456936359406,
0.09531481564044952,
-0.2095387876033783,
0.013276174664497375,
0.4428166449069977,
0.045999232679605484,
0.9906775951385498,
0.6558960676193237,
-0.8690496683120728,
-0.730083167552948,
-0.48813191056251526,
-0.312788873910903... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
SauravMaheshkar/tox21_SRp53 | SauravMaheshkar | 2023-02-12T14:30:43Z | 12 | 2 | null | [
"task_categories:other",
"task_categories:graph-ml",
"annotations_creators:machine-generated",
"language_creators:machine-generated",
"bio",
"bio-chem",
"molnet",
"molecule-net",
"biophysics",
"arxiv:1703.00564",
"region:us"
] | 2023-02-12T14:30:43Z | 2022-11-30T10:33:29.000Z | 2022-11-30T10:33:29 | ---
annotations_creators:
- machine-generated
language_creators:
- machine-generated
pretty_name: tox21_SRp53
tags:
- bio
- bio-chem
- molnet
- molecule-net
- biophysics
task_categories:
- other
- graph-ml
task_ids: []
---
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Dataset Structure](#dataset-structure)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Additional Information](#additional-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
# Dataset Description
- **Homepage: https://moleculenet.org/**
- **Repository: https://github.com/deepchem/deepchem/tree/master**
- **Paper: https://arxiv.org/abs/1703.00564**
## Dataset Summary
`tox21_SRp53` is a dataset included in [MoleculeNet](https://moleculenet.org/). The "Toxicology in the 21st Century" (Tox21) initiative created a public database measuring toxicity of compounds, which has been used in the 2014 Tox21 Data Challenge. This dataset contains qualitative toxicity measurements for 8k compounds on 12 different targets, including nuclear receptors and stress response pathways.
# Dataset Structure
## Data Fields
Each split contains
* `smiles`: the [SMILES](https://en.wikipedia.org/wiki/Simplified_molecular-input_line-entry_system) representation of a molecule
* `selfies`: the [SELFIES](https://github.com/aspuru-guzik-group/selfies) representation of a molecule
* `target`: Measured results (Active/Inactive) for bioassays
## Data Splits
The dataset is split into an 80/10/10 train/valid/test split using random split.
# Additional Information
## Citation Information
```
@misc{https://doi.org/10.48550/arxiv.1703.00564,
doi = {10.48550/ARXIV.1703.00564},
url = {https://arxiv.org/abs/1703.00564},
author = {Wu, Zhenqin and Ramsundar, Bharath and Feinberg, Evan N. and Gomes, Joseph and Geniesse, Caleb and Pappu, Aneesh S. and Leswing, Karl and Pande, Vijay},
keywords = {Machine Learning (cs.LG), Chemical Physics (physics.chem-ph), Machine Learning (stat.ML), FOS: Computer and information sciences, FOS: Computer and information sciences, FOS: Physical sciences, FOS: Physical sciences},
title = {MoleculeNet: A Benchmark for Molecular Machine Learning},
publisher = {arXiv},
year = {2017},
copyright = {arXiv.org perpetual, non-exclusive license}
}
```
## Contributions
Thanks to [@SauravMaheshkar](https://github.com/SauravMaheshkar) and [@zanussbaum](https://github.com/zanussbaum) for adding this dataset | [
-0.26724809408187866,
-0.04732227697968483,
0.3967911899089813,
0.23023183643817902,
-0.15783371031284332,
0.02013002522289753,
-0.18514342606067657,
-0.16829949617385864,
-0.04758699610829353,
0.39713048934936523,
-0.7172518968582153,
-0.9645119905471802,
-0.5188170075416565,
0.3059370517... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
AntonioTal/IMAGES | AntonioTal | 2022-12-02T20:18:52Z | 12 | 0 | null | [
"region:us"
] | 2022-12-02T20:18:52Z | 2022-11-30T11:40:12.000Z | 2022-11-30T11:40:12 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622264862060547,
0.43461528420448303,
-0.52829909324646,
0.7012971639633179,
0.7915720343589783,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104477167129517,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Asmar/visioniert | Asmar | 2022-11-30T12:07:49Z | 12 | 0 | null | [
"region:us"
] | 2022-11-30T12:07:49Z | 2022-11-30T12:07:08.000Z | 2022-11-30T12:07:08 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622264862060547,
0.43461528420448303,
-0.52829909324646,
0.7012971639633179,
0.7915720343589783,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104477167129517,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
malteos/germeval2017 | malteos | 2022-11-30T13:49:08Z | 12 | 0 | null | [
"language:de",
"region:us"
] | 2022-11-30T13:49:08Z | 2022-11-30T12:53:43.000Z | 2022-11-30T12:53:43 | ---
language:
- de
---
# Germeval Task 2017: Shared Task on Aspect-based Sentiment in Social Media Customer Feedback
In the connected, modern world, customer feedback is a valuable source for insights on the quality of products or services. This feedback allows other customers to benefit from the experiences of others and enables businesses to react on requests, complaints or recommendations. However, the more people use a product or service, the more feedback is generated, which results in the major challenge of analyzing huge amounts of feedback in an efficient, but still meaningful way.
Thus, we propose a shared task on automatically analyzing customer reviews about “Deutsche Bahn” - the german public train operator with about two billion passengers each year.
Example:
> “RT @XXX: Da hört jemand in der Bahn so laut ‘700 Main Street’ durch seine Kopfhörer, dass ich mithören kann. :( :( :(“
As shown in the example, insights from reviews can be derived on different granularities. The review contains a general evaluation of the travel (The customer disliked the travel). Furthermore, the review evaluates a dedicated aspect of the train travel (“laut” → customer did not like the noise level).
Consequently, we frame the task as aspect-based sentiment analysis with four sub tasks:
## Data format
```
ID <tab> Text <tab> Relevance <tab> Sentiment <tab> Aspect:Polarity (whitespace separated)
```
## Links
- http://ltdata1.informatik.uni-hamburg.de/germeval2017/
- https://sites.google.com/view/germeval2017-absa/
## How to cite
```bibtex
@inproceedings{germevaltask2017,
title = {{GermEval 2017: Shared Task on Aspect-based Sentiment in Social Media Customer Feedback}},
author = {Michael Wojatzki and Eugen Ruppert and Sarah Holschneider and Torsten Zesch and Chris Biemann},
year = {2017},
booktitle = {Proceedings of the GermEval 2017 – Shared Task on Aspect-based Sentiment in Social Media Customer Feedback},
address={Berlin, Germany},
pages={1--12}
}
``` | [
-0.5607714653015137,
-0.6042934060096741,
0.5123469829559326,
0.5427761673927307,
-0.45773905515670776,
-0.059184543788433075,
-0.07091598957777023,
-0.6903148293495178,
0.6312821507453918,
0.3179236352443695,
-0.5302910804748535,
-0.6713571548461914,
-0.4874839782714844,
-0.07471621781587... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
m-aliabbas/idrak_unsplitted_amy | m-aliabbas | 2022-11-30T14:24:08Z | 12 | 0 | null | [
"region:us"
] | 2022-11-30T14:24:08Z | 2022-11-30T13:50:33.000Z | 2022-11-30T13:50:33 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622264862060547,
0.43461528420448303,
-0.52829909324646,
0.7012971639633179,
0.7915720343589783,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104477167129517,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
slnader/fcc-comments | slnader | 2022-11-30T19:05:23Z | 12 | 2 | null | [
"task_categories:text-retrieval",
"task_ids:document-retrieval",
"annotations_creators:expert-generated",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:10M<n<100M",
"source_datasets:original",
"language:en",
"license:cc-by-nc-sa-4.0",
"notice and comment",
"regulatio... | 2022-11-30T19:05:23Z | 2022-11-30T17:38:32.000Z | 2022-11-30T17:38:32 | ---
annotations_creators:
- expert-generated
language:
- en
language_creators:
- found
license:
- cc-by-nc-sa-4.0
multilinguality:
- monolingual
pretty_name: fcc-comments
size_categories:
- 10M<n<100M
source_datasets:
- original
tags:
- notice and comment
- regulation
- government
task_categories:
- text-retrieval
task_ids:
- document-retrieval
---
# Dataset Card for fcc-comments
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Repository: https://github.com/slnader/fcc-comments **
- **Paper: https://doi.org/10.1002/poi3.327 **
### Dataset Summary
Online comment floods during public consultations have posed unique governance challenges for
regulatory bodies seeking relevant information on proposed regulations.
How should regulatory bodies separate spam and fake comments from genuine submissions by the public,
especially when fake comments are designed to imitate ordinary citizens? How can regulatory bodies
achieve both breadth and depth in their citations to the comment corpus? What is the best way to
select comments that represent the average submission and comments that supply highly specialized
information?
`fcc-comments` is an annotated version of the comment corpus from the Federal Communications Commission's
(FCC) 2017 "Restoring Internet Freedom" proceeding. The source data were downloaded directly from the FCC's Electronic
Comment Filing System (ECFS) between January and February of 2019 and include raw comment text and metadata on
comment submissions. The comment data were processed to be in a consistent format
(machine-readable pdf or plain text), and annotated with three types of information: whether the comment was cited in the
agency's final order, the type of commenter (individual, interest group, business group), and whether the comment was associated with an in-person meeting.
The release also includes query-term and document-term matrices to facilitate keyword searches on the comment corpus.
An example of how these can be used with the bm25 algorithm can be found
[here](https://github.com/slnader/fcc-comments/blob/main/process_comments/1_score_comments.py).
## Dataset Structure
FCC relational database (fcc.pgsql): The core components of the database include a table for submission metadata,
a table for attachment metadata, a table for filer metadata, and a table that contains comment text if submitted in express format.
In addition to these core tables, there are several derived tables specific to the analyses in the paper,
including which submissions and attachments were cited in the final order, which submissions were associated with in-person meetings,
and which submissions were associated with interest groups. Full documentation of the tables can be found in fcc_database.md.
Attachments (attachments.tar.gz): Attachments to submissions that could be converted to text via OCR and saved in machine-readable pdf format.
The filenames are formatted as [submission_id]_[document_id].pdf, where submission_id and document_id are keys in the relational database.
Search datasets (search.tar.gz): Objects to facilitate prototyping of search algorithms on the comment corpus. Contains the following elements:
| Filename | description |
| ----------- | ----------- |
query_dtm.pickle | Query-term matrix (79x3986) in sparse csr format (rows are queries, columns are bigram keyword counts).
query_text.pickle | Dictionary keyed by the paragraph number in the FCC’s Notice of Proposed Rulemaking. Values are the text of the query containing a call for comments. |
search_dtms_express.pickle | Document-term matrix for express comments (3800691x3986) in sparse csr format (rows are comment pages, columns are bigram keyword counts). |
search_index_express.pickle | Pandas dataframe containing unique id and total term length for express comments. |
search_dtms.pickle | Document-term matrix for standard comment attachments (44655x3986) in sparse csr format (rows are comment pages, columns are bigram keyword counts). |
search_index.pickle | Pandas dataframe containing unique id and total term length for standard comment attachments. |
### Data Fields
The following tables are available in fcc.pgsql:
- comments: plain text comments associated with submissions
| column | type | description |
| ----------- | ----------- | ----------- |
| comment_id | character varying(64) | unique id for plain text comment |
comment_text | text | raw text of plain text comment
row_id | integer | row sequence for plain text comments
- submissions: metadata for submissions
| column | type | description |
| ----------- | ----------- | ----------- |
submission_id | character varying(20) | unique id for submission
submission_type | character varying(100) | type of submission (e.g., comment, reply, statement)
express_comment | numeric | 1 if express comment
date_received | date | date submission was received
contact_email | character varying(255) | submitter email address
city | character varying(255) | submitter city
address_line_1 | character varying(255) | submitter address line 1
address_line_2 | character varying(255) | submitter address line 2
state | character varying(255) | submitter state
zip_code | character varying(50) | submitter zip
comment_id | character varying(64) | unique id for plain text comment
- filers: names of filers associated with submissions
| column | type | description |
| ----------- | ----------- | ----------- |
submission_id | character varying(20) | unique id for submission
filer_name | character varying(250) | name of filer associated with submission
- documents: attachments associated with submissions
| column | type | description |
| ----------- | ----------- | ----------- |
submission_id | character varying(20) | unique id for submission
document_name | text | filename of attachment
download_status | numeric | status of attachment download
document_id | character varying(64) | unique id for attachment
file_extension | character varying(4) | file extension for attachment
- filers_cited: citations from final order
| column | type | description |
| ----------- | ----------- | ----------- |
point | numeric | paragraph number in final order
filer_name | character varying(250) | name of cited filer
submission_type | character varying(12) | type of submission as indicated in final order
page_numbers | text[] | cited page numbers
cite_id | integer | unique id for citation
filer_id | character varying(250) | id for cited filer
- docs_cited: attachments associated with cited submissions
| column | type | description |
| ----------- | ----------- | ----------- |
cite_id | numeric | unique id for citation
submission_id | character varying(20) | unique id for submission
document_id | character varying(64) | unique id for attachment
- near_duplicates: lookup table for comment near-duplicates
| column | type | description |
| ----------- | ----------- | ----------- |
target_document_id | unique id for target document
duplicate_document_id | unique id for duplicate of target document
- exact_duplicates: lookup table for comment exact duplicates
| column | type | description |
| ----------- | ----------- | ----------- |
target_document_id | character varying(100) | unique id for target document
duplicate_document_id | character varying(100) | unique id for duplicate of target document
- in_person_exparte: submissions associated with ex parte meeting
| column | type | description |
| ----------- | ----------- | ----------- |
submission_id | character varying(20) | unique id for submission
- interest_groups: submissions associated with interest groups
| column | type | description |
| ----------- | ----------- | ----------- |
submission_id | character varying(20) | unique id for submission
business | numeric | 1 if business group, 0 otherwise
## Dataset Creation
### Curation Rationale
The data were curated to perform information retrieval and summarization tasks as documented in https://doi.org/10.1002/poi3.327.
### Source Data
#### Initial Data Collection and Normalization
The data for this study come from the FCC's Electronic Comment Filing System (ECFS) system, accessed between January and February of 2019.
I converted the API responses into a normalized, relational database containing information on 23,951,967 submissions.
23,938,686 "express" submissions contained a single plain text comment submitted directly through the comment form.
13,821 "standard" submissions contained one or more comment documents submitted as attachments in various file formats.
While the FCC permitted any file format for attachments, I only consider documents attached in pdf, plain text, rich text,
and Microsoft Word file formats, and I drop submitted documents that were simply copies of the FCC’s official documents (e.g., the NPRM itself).
Using standard OCR software, I attempted to convert all attachments into plain text and saved them as machine-readable pdfs.
#### Who are the source language producers?
All submitters of public comments during the public comment period (but see note on fake comments in considerations).
### Annotations
#### Annotation process
- Citations: I consider citations from the main text of the FCC's final rule. I did not include citations to
supporting documents not available through ECFS (e.g., court decisions), nor did I include citations
to submissions from prior FCC proceedings. The direct citations to filed submissions are included
in a series of 1,186 footnotes. The FCC’s citation format typically followed a relatively standard
pattern: the name of the filer (e.g., Verizon), a description of the document (e.g., Comment), and
at times a page number. I extracted citations from the text using regular expressions. Based on a
random sample of paragraphs from the final order, the regular expressions identified 98% of eligible citations,
while successfully excluding all non-citation text. In total, this produced 1,886 unique citations.
I then identified which of the comments were cited. First, I identified all documents from the cited filer
that had enough pages to contain the page number cited (if provided), and, where applicable, whose filename
contained the moniker from the FCC’s citation (e.g., "Reply"). The majority of citations matched to only one
possible comment submitted, and I identified the re- maining cited comments through manual review of the citations.
In this way, I was able to tag documents associated with all but three citations. When the same cited document was
submitted under multiple separate submissions, I tagged all versions of the document as being cited.
- Commenter type: Comments are labeled as mass comments if 10 or more duplicate or near-duplicate copies were
submitted by individual commenters. Near-duplicates were defined as comments with non-zero identical information scores.
To identify the type of commenter for non-mass comments, I take advantage of the fact that the vast majority of organized
groups preferred standard submissions over express submissions. Any non-mass comment submitted as an express comment was
coded as coming from an individual. To distinguish between individuals and organizations that used standard submissions,
I use a first name and surname database from the names dataset Python package to characterize filer names as belonging to
individuals or organizations. I also use the domain of the submitter’s email address to re-categorize comments as coming
from organizations if they were submitted on behalf of organizations by an individual. Government officials were identified by
their .gov email addresses. I manually review this procedure for mischaracterizations. After obtaining a list of organization
names, I manually code each one as belonging to a business group or a non-business group. Government officials writing in
their official capacity were categorized as a non-business group.
- In-person meetings: To identify which commenters held in-person meetings with the agency, I collect all comments labeled
as an ex-parte submission in the EFCS. I manually review these submissions for mention of an in-person meeting. I label
a commenter as having held an in-person meeting if they submitted at least one ex-parte document that mentioned an in-person meeting.
#### Who are the annotators?
Annotations are a combination of automated and manual review done by the author.
### Personal and Sensitive Information
This dataset may contain personal and sensitive information, as there were no restrictions on what commenters could submit to
the agency. This dataset also contains numerous examples of profanity and spam. These comments represent what the FCC decided was
appropriate to share publicly on their own website.
## Considerations for Using the Data
### Discussion of Biases
This proceeding was famous for the large number of "fake" comments (comments impersonating ordinary citizens) submitted to the
agency (see [this report](https://ag.ny.gov/sites/default/files/oag-fakecommentsreport.pdf) by the NY AG for more information).
As such, this comment corpus contains a mix of computer-generated and natural language, and there is currently no way to reliably separate
mass comments submitted with the approval of the commenter and those submitted on behalf of the commenter without their knowledge.
## Additional Information
### Licensing Information
CreativeCommons Attribution-NonCommercial-ShareAlike 4.0 International.
### Citation Information
```
@article{handan2022,
title={Do fake online comments pose a threat to regulatory policymaking? Evidence from Internet regulation in the United States},
author={Handan-Nader, Cassandra},
journal={Policy \& Internet},
year={2022}
}
``` | [
-0.5421348810195923,
-0.4874556064605713,
0.46170708537101746,
0.526337206363678,
-0.1268054097890854,
-0.11247832328081131,
-0.1653464287519455,
-0.2935386598110199,
0.5410556793212891,
0.42905256152153015,
-0.6316760778427124,
-0.6886112093925476,
-0.47517824172973633,
0.6088629364967346... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
argilla/uber-reviews | argilla | 2022-12-06T12:00:28Z | 12 | 0 | null | [
"task_categories:text-classification",
"task_ids:sentiment-classification",
"size_categories:1K<n<10K",
"source_datasets:original",
"language:en",
"license:unknown",
"region:us"
] | 2022-12-06T12:00:28Z | 2022-12-06T11:47:18.000Z | 2022-12-06T11:47:18 | ---
language:
- en
license:
- unknown
size_categories:
- 1K<n<10K
source_datasets:
- original
task_categories:
- text-classification
task_ids:
- sentiment-classification
dataset_info:
features:
- name: text
dtype: string
- name: inputs
struct:
- name: text
dtype: string
- name: prediction
list:
- name: label
dtype: string
- name: score
dtype: float64
- name: prediction_agent
dtype: string
- name: annotation
dtype: 'null'
- name: annotation_agent
dtype: 'null'
- name: multi_label
dtype: bool
- name: explanation
dtype: 'null'
- name: id
dtype: string
- name: metadata
dtype: 'null'
- name: status
dtype: string
- name: event_timestamp
dtype: timestamp[us]
- name: metrics
struct:
- name: text_length
dtype: int64
splits:
- name: train
num_bytes: 2761597
num_examples: 2347
download_size: 1691346
dataset_size: 2761597
---
# Dataset Card for "uber-reviews"
## Dataset Description
- **Homepage:** Kaggle Challenge
- **Repository:** https://www.kaggle.com/datasets/jschne61701/uber-rides-costumer-reviews-dataset
- **Paper:** N.A.
- **Leaderboard:** N.A.
- **Point of Contact:** N.A.
### Dataset Summary
Using Python's Beautiful Soup library and Scrappy framework, scraped date, star rating, and comment from all reviews from 2013 - 2019.
### Languages
english
### Citation Information
https://www.kaggle.com/datasets/jschne61701/uber-rides-costumer-reviews-dataset
https://www.sitejabber.com/reviews/uber.com
https://www.consumeraffairs.com/travel/uber.html
https://www.kaggle.com/purvank/uber-rider-reviews-dataset
### Contributions
Thanks to [@davidberenstein1957](https://github.com/davidberenstein1957) for adding this dataset.
| [
-0.3749411702156067,
-0.14848017692565918,
0.27052953839302063,
0.5376361608505249,
-0.5149033665657043,
-0.10837256163358688,
0.15925727784633636,
-0.5588119029998779,
0.5526205897331238,
0.4012939929962158,
-0.6775596141815186,
-0.8081625699996948,
-0.15350790321826935,
0.038585260510444... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
tilos/ASR-CCANTCSC | tilos | 2022-12-07T21:39:11Z | 12 | 0 | null | [
"language:zh",
"license:cc-by-nc-nd-4.0",
"region:us"
] | 2022-12-07T21:39:11Z | 2022-12-07T11:13:27.000Z | 2022-12-07T11:13:27 | ---
license: cc-by-nc-nd-4.0
pretty_name: ASR-CCANTCSC
language:
- zh
dataset_info:
features:
- name: audio
dtype: Audio
- name: sentence
dtype: string
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
maratim/romanianspeech | maratim | 2022-12-07T12:08:50Z | 12 | 0 | null | [
"region:us"
] | 2022-12-07T12:08:50Z | 2022-12-07T12:05:06.000Z | 2022-12-07T12:05:06 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
ywan111/dataset-test | ywan111 | 2022-12-07T12:18:19Z | 12 | 0 | null | [
"region:us"
] | 2022-12-07T12:18:19Z | 2022-12-07T12:15:48.000Z | 2022-12-07T12:15:48 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
autoevaluate/autoeval-staging-eval-autoevaluate__squad-sample-autoevaluate__squad-sample-778ba0-17436360 | autoevaluate | 2022-12-07T12:28:44Z | 12 | 0 | null | [
"autotrain",
"evaluation",
"region:us"
] | 2022-12-07T12:28:44Z | 2022-12-07T12:28:20.000Z | 2022-12-07T12:28:20 | ---
type: predictions
tags:
- autotrain
- evaluation
datasets:
- autoevaluate/squad-sample
eval_info:
task: extractive_question_answering
model: autoevaluate/extractive-question-answering-not-evaluated
metrics: []
dataset_name: autoevaluate/squad-sample
dataset_config: autoevaluate--squad-sample
dataset_split: test
col_mapping:
context: context
question: question
answers-text: answers.text
answers-answer_start: answers.answer_start
---
# Dataset Card for AutoTrain Evaluator
This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset:
* Task: Question Answering
* Model: autoevaluate/extractive-question-answering-not-evaluated
* Dataset: autoevaluate/squad-sample
* Config: autoevaluate--squad-sample
* Split: test
To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator).
## Contributions
Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model. | [
-0.49809300899505615,
-0.487989217042923,
0.2194875180721283,
0.2213267683982849,
0.08260218799114227,
0.10052156448364258,
0.18473048508167267,
-0.40542495250701904,
0.16014063358306885,
0.3172439932823181,
-1.2577364444732666,
-0.10413613170385361,
-0.5556237697601318,
0.1051842868328094... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
xusenlin/clue-ner | xusenlin | 2022-12-07T14:22:37Z | 12 | 7 | null | [
"language:zh",
"license:apache-2.0",
"named entity recognition",
"clue",
"region:us"
] | 2022-12-07T14:22:37Z | 2022-12-07T13:14:03.000Z | 2022-12-07T13:14:03 | ---
dataset_info:
features:
- name: text
dtype: string
- name: entities
list:
- name: id
dtype: int64
- name: entity
dtype: string
- name: start_offset
dtype: int64
- name: end_offset
dtype: int64
- name: label
dtype: string
splits:
- name: train
num_bytes: 2443356
num_examples: 10748
- name: test
num_bytes: 154492
num_examples: 1345
- name: validation
num_bytes: 309106
num_examples: 1343
download_size: 1658426
dataset_size: 2906954
language:
- zh
tags:
- named entity recognition
- clue
license: apache-2.0
---
# CLUE-NER 命名实体识别数据集
字段说明
+ `text`: 文本
+ `entities`: 文本中包含的实体
+ `id`: 实体 `id`
+ `entity`: 实体对应的字符串
+ `start_offset`: 实体开始位置
+ `end_offset`: 实体结束位置的下一位
+ `label`: 实体对应的开始位置
| [
-0.4908401370048523,
-0.6941743493080139,
0.11775249987840652,
0.3810179531574249,
-1.0162177085876465,
0.03100728616118431,
0.2675838768482208,
-0.41077181696891785,
0.9416723251342773,
0.11332858353853226,
-0.7723591327667236,
-0.9075393676757812,
-0.6635699272155762,
0.19504155218601227... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
ola13/small-oscar-dedup | ola13 | 2022-12-07T15:48:57Z | 12 | 0 | null | [
"region:us"
] | 2022-12-07T15:48:57Z | 2022-12-07T13:44:16.000Z | 2022-12-07T13:44:16 | ---
dataset_info:
features:
- name: id
dtype: int64
- name: text
dtype: string
- name: url
dtype: string
- name: domain
dtype: string
- name: perplexity
dtype: float64
- name: dup_ratio
dtype: float64
- name: pairs
sequence:
sequence: int64
- name: repetitions
sequence: binary
- name: cluster
sequence:
sequence: int64
splits:
- name: train
num_bytes: 323557137
num_examples: 43200
download_size: 0
dataset_size: 323557137
---
# Dataset Card for "small-oscar-dedup"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.6796634793281555,
-0.11920082569122314,
0.2930970788002014,
-0.0364091582596302,
-0.4273083209991455,
-0.17712780833244324,
0.3144541084766388,
-0.22153165936470032,
1.018802285194397,
0.46245884895324707,
-0.6185622215270996,
-0.5488615036010742,
-0.8006286025047302,
-0.133874177932739... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Tristan/olm-bookcorpus-tokenized-1024 | Tristan | 2022-12-07T19:33:36Z | 12 | 0 | null | [
"region:us"
] | 2022-12-07T19:33:36Z | 2022-12-07T19:28:08.000Z | 2022-12-07T19:28:08 | ---
dataset_info:
features:
- name: input_ids
sequence: int32
- name: attention_mask
sequence: int8
- name: special_tokens_mask
sequence: int8
splits:
- name: train
num_bytes: 8534733804.0
num_examples: 1386409
download_size: 2291578601
dataset_size: 8534733804.0
---
# Dataset Card for "olm-bookcorpus-tokenized-1024"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.4007304906845093,
-0.22370117902755737,
-0.04232306405901909,
0.07134042680263519,
-0.2512008845806122,
-0.013094460591673851,
0.22184133529663086,
-0.0029629277996718884,
0.6641333699226379,
0.7961339354515076,
-0.668185830116272,
-0.8444881439208984,
-0.3950076103210449,
-0.0687094330... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
evageon/myaudio | evageon | 2022-12-08T15:28:56Z | 12 | 0 | null | [
"region:us"
] | 2022-12-08T15:28:56Z | 2022-12-07T20:56:35.000Z | 2022-12-07T20:56:35 | ---
pretty_name: MGB2
alt_glob: []
alt_sep: ''
dataset_link: https://huggingface.co/datasets/malmarz/test_mgb2/resolve/main/mgb2.test.tar.gz
dataset_name: mgb2_speech
datasets_path: datasets
file_type: wav
header: null
hf_path: ''
json_key: null
label_column_name: ''
level: null
lines: false
local_dir: false
new_columns: ''
pal: false
skiprows: 0
squad: false
xml_columns: ''
dataset_info:
features:
- name: audio
dtype: audio
- name: transcription
dtype: string
splits:
- name: test
num_bytes: 384372257.17
num_examples: 3890
- name: train
num_bytes: 1488884786.673
num_examples: 15559
download_size: 0
dataset_size: 1873257043.8430002
---
Please note this dataset is private
### Using the data
You can stream the data data loader:
```python
myaudio = load_dataset(
"evageon/myaudio",
use_auth_token=os.environ["HG_USER_TOKEN"], # replace this with your access token
streaming=True)
```
Then you can iterate over the dataset
```python
# replace test with validation or train depending on split you need
print(next(iter(myaudio["test"])))
```
outputs:
```
{'path': 'CD93A8FF-C3ED-4AD4-95A6-8363CCB93B90_spk-0001_seg-0024467:0025150.wav', 'audio': {'path': 'dataset/test/wav/CD93A8FF-C3ED-4AD4-95A6-8363CCB93B90_spk-0001_seg-0024467:0025150.wav', 'array': array([0.00662231, 0.00497437, 0.00518799, ..., 0.01150513, 0.00708008,
0.00296021]), 'sampling_rate': 16000}, 'text': 'خطرا على دول الخليج لماذا اعتبرت أن إيران اليوم والخطر الذي تشكله إيران مختلف'}
``` | [
-0.3966018557548523,
-0.5160664916038513,
0.22213369607925415,
0.19003397226333618,
-0.128263920545578,
0.20020733773708344,
-0.4569166302680969,
0.07743871212005615,
0.4706849157810211,
0.44717031717300415,
-0.6632488369941711,
-0.7606067061424255,
-0.5709961652755737,
0.20207592844963074... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
thennal/msc | thennal | 2022-12-08T06:49:31Z | 12 | 1 | null | [
"task_categories:automatic-speech-recognition",
"annotations_creators:crowdsourced",
"language_creators:crowdsourced",
"multilinguality:monolingual",
"size_categories:1K<n<10K",
"language:ml",
"license:cc-by-sa-4.0",
"region:us"
] | 2022-12-08T06:49:31Z | 2022-12-08T06:19:56.000Z | 2022-12-08T06:19:56 | ---
annotations_creators:
- crowdsourced
language:
- ml
language_creators:
- crowdsourced
license:
- cc-by-sa-4.0
multilinguality:
- monolingual
pretty_name: Swathanthra Malayalam Computing Malayalam Speech Corpus
size_categories:
- 1K<n<10K
source_datasets: []
tags: []
task_categories:
- automatic-speech-recognition
task_ids: []
dataset_info:
features:
- name: speechid
dtype: string
- name: speaker_id
dtype: string
- name: review_score
dtype: int64
- name: transcript
dtype: string
- name: category
dtype: string
- name: speaker_gender
dtype: string
- name: speaker_age
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 48000
splits:
- name: train
num_bytes: 581998721.306
num_examples: 1541
download_size: 422643542
dataset_size: 581998721.306
---
# SMC Malayalam Speech Corpus
Malayalam Speech Corpus (MSC) is a repository of curated speech samples collected using MSC web application, released by Swathanthra Malayalam Computing.
The official blog post and source data can be found at [https://blog.smc.org.in/malayalam-speech-corpus/](https://blog.smc.org.in/malayalam-speech-corpus/).
## Dataset Description
- **Homepage:** [https://blog.smc.org.in/malayalam-speech-corpus/](https://blog.smc.org.in/malayalam-speech-corpus/)
### Dataset Summary
The first version of Malayalam Speech Corpus contains 1541 speech samples from 75 contributors amounting to 1:38:16 hours of speech. It has 482 unique sentences, 1400 unique words, 553 unique syllables and 48 unique phonemes.
| [
-0.42506739497184753,
-0.44385531544685364,
0.1807435154914856,
0.3562546968460083,
-0.3989783823490143,
0.31520506739616394,
-0.048128146678209305,
-0.03258460760116577,
0.5415960550308228,
0.43088746070861816,
-0.7856222987174988,
-0.5531346797943115,
-0.46952274441719055,
0.221303701400... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
nuprl/MultiPL-E-raw-data | nuprl | 2022-12-20T18:40:05Z | 12 | 0 | null | [
"license:bsd-3-clause",
"arxiv:2208.08227",
"region:us"
] | 2022-12-20T18:40:05Z | 2022-12-11T19:07:19.000Z | 2022-12-11T19:07:19 | ---
license: bsd-3-clause
---
# MultiPL-E Evaluation Raw Data
This is the raw data for the MultiPL-E paper: https://arxiv.org/abs/2208.08227 | [
-0.32517582178115845,
-0.4364900290966034,
0.24707752466201782,
0.3497063219547272,
-0.1511189341545105,
0.16554835438728333,
0.22173793613910675,
-0.056709498167037964,
0.3589414954185486,
0.4290607273578644,
-0.7636997699737549,
-0.32563087344169617,
-0.36635443568229675,
-0.092848390340... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
amosr2002/NYCDATA | amosr2002 | 2022-12-14T16:36:21Z | 12 | 0 | null | [
"region:us"
] | 2022-12-14T16:36:21Z | 2022-12-14T16:30:31.000Z | 2022-12-14T16:30:31 | ---
dataset_info:
features:
- name: name
dtype: string
- name: uuid
dtype: string
- name: status
dtype: string
- name: image
dtype: image
- name: label.annotations
list:
- name: id
dtype: int32
- name: category_id
dtype: int32
- name: label.segmentation_bitmap
dtype: image
splits:
- name: train
num_bytes: 5365744.0
num_examples: 10
download_size: 0
dataset_size: 5365744.0
---
# Dataset Card for "NYCDATA"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.4909600019454956,
-0.23996716737747192,
0.295451819896698,
0.06791726499795914,
-0.1374521255493164,
0.31571951508522034,
0.4012744128704071,
-0.35118088126182556,
1.09690260887146,
0.3896866738796234,
-0.4865715801715851,
-0.8733740448951721,
-0.4924921691417694,
-0.27827519178390503,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
my-projects/dataset1 | my-projects | 2022-12-14T19:15:42Z | 12 | 0 | null | [
"region:us"
] | 2022-12-14T19:15:42Z | 2022-12-14T19:15:35.000Z | 2022-12-14T19:15:35 | ---
dataset_info:
features:
- name: solution
sequence: int64
- name: coefficients
sequence: int64
- name: discriminant
dtype: int64
- name: __index_level_0__
dtype: int64
splits:
- name: train
num_bytes: 4814334
num_examples: 80000
- name: test
num_bytes: 1204417
num_examples: 20000
download_size: 1554124
dataset_size: 6018751
---
# Dataset Card for "dataset1"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.6457745432853699,
-0.33117765188217163,
0.07624127715826035,
0.32439887523651123,
-0.2809703052043915,
-0.03685969486832619,
0.46780815720558167,
-0.07035554945468903,
0.9785434603691101,
0.516433596611023,
-1.0301873683929443,
-0.7862607836723328,
-0.6689369082450867,
-0.28167438507080... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Mai321/ZeroTwo-PlugSuit | Mai321 | 2022-12-14T19:45:33Z | 12 | 0 | null | [
"license:openrail",
"region:us"
] | 2022-12-14T19:45:33Z | 2022-12-14T19:42:56.000Z | 2022-12-14T19:42:56 | ---
license: openrail
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
laion/laion2b-en-vit-l-14-embeddings | laion | 2022-12-17T02:00:01Z | 12 | 10 | null | [
"region:us"
] | 2022-12-17T02:00:01Z | 2022-12-15T01:43:31.000Z | 2022-12-15T01:43:31 | Entry not found | [
-0.3227647542953491,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965083122253,
0.7915717959403992,
0.07618629932403564,
0.7746022343635559,
0.2563222348690033,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
bigcode/the-stack-metadata | bigcode | 2023-03-16T13:58:24Z | 12 | 3 | null | [
"task_categories:text-generation",
"language_creators:crowdsourced",
"language_creators:expert-generated",
"multilinguality:multilingual",
"size_categories:unknown",
"language:code",
"license:other",
"arxiv:2211.15533",
"region:us"
] | 2023-03-16T13:58:24Z | 2022-12-19T09:17:28.000Z | 2022-12-19T09:17:28 | ---
annotations_creators: []
language_creators:
- crowdsourced
- expert-generated
language:
- code
license:
- other
multilinguality:
- multilingual
pretty_name: The-Stack-Metadata
size_categories:
- unknown
source_datasets: []
task_categories:
- text-generation
task_ids: []
extra_gated_prompt: |-
## Terms of Use for The Stack
The Stack Metadata is a collection of additional information for and is part of The Stack dataset, - a collection of source code in over 300 programming languages. We ask that you read and acknowledge the following points before using the dataset:
1. The Stack is a collection of source code from repositories with various licenses. Any use of all or part of the code gathered in The Stack must abide by the terms of the original licenses, including attribution clauses when relevant. We facilitate this by providing provenance information for each data point.
2. The Stack is regularly updated to enact validated data removal requests. By clicking on "Access repository", you agree to update your own version of The Stack to the most recent usable version specified by the maintainers in [the following thread](https://huggingface.co/datasets/bigcode/the-stack/discussions/7). If you have questions about dataset versions and allowed uses, please also ask them in the dataset’s [community discussions](https://huggingface.co/datasets/bigcode/the-stack/discussions/new). We will also notify users via email when the latest usable version changes.
3. To host, share, or otherwise provide access to The Stack dataset, you must include [these Terms of Use](https://huggingface.co/datasets/bigcode/the-stack#terms-of-use-for-the-stack) and require users to agree to it.
By clicking on "Access repository" below, you accept that your contact information (email address and username) can be shared with the dataset maintainers as well.
extra_gated_fields:
Email: text
I have read the License and agree with its terms: checkbox
---
# Dataset Card for The Stack Metadata
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Changelog](#changelog)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Dataset Structure](#dataset-structure)
- [Data Fields](#data-fields)
- [Usage Example](#usage-example)
- [Dataset Creation](#dataset-creation)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Additional Information](#additional-information)
- [Terms of Use for The Stack](#terms-of-use-for-the-stack)
## Dataset Description
- **Homepage:** https://www.bigcode-project.org/
- **Repository:** https://github.com/bigcode-project
- **Paper:** https://arxiv.org/abs/2211.15533
- **Leaderboard:** N/A
- **Point of Contact:** contact@bigcode-project.org
### Changelog
|Release|Description|
|-|-|
|v1.1| This is the first release of the metadata. It is for The Stack v1.1|
|v1.2| Metadata dataset matching The Stack v1.2|
### Dataset Summary
This is a set of additional information for repositories used for The Stack. It contains file paths, detected licenes as well as some other information for the repositories.
### Supported Tasks and Leaderboards
The main task is to recreate repository structure from the files of The Stack. Also, the set can be used for computing statistics and custom filtering or aggregation operations on The Stack.
## Dataset Structure
### Data Fields

The set is split into buckets by repositories. There are 944 buckets. Additionally to the fields in the image, `ri` contains `min_repo_event_datetime` which is the ealiest date and time of an event for a repo after Jan 1 2015.

As an example of an aggregation operation on The Stack, the image above shows conceptually a selection of stars ( and issues and PR count) for a file. Each unique file can be part of multiple repositories. So, The Stack releases unique files and aggregates meta information (e.g stars) from all repositories it belongs to. For example, for max_stars_count we take the maximum number of stars from all repositories the file is part of.
The meta data will allow you to reconstruct repository directory structures. For this, for each repository form `ri` tabele it is needed to take all its files from `fi` table, find them in The Stack by file's `hexsha` and save those files' content under its path for a repository from `fi` table. For speed it is preferable to index The Stack by hexsha first.
### Usage Example
Restore folder structure for python files in numpy repository
```python
import datasets
from pathlib import Path
from tqdm.auto import tqdm
import pandas as pd
# assuming metadata is cloned into the local folder /data/hf_repos/the-stack-metadata
# the stack is cloned into the local folder /data/hf_repos/the-stack-v1.1
# destination folder is in /repo_workdir/numpy_restored
the_stack_meta_path = Path('/data/hf_repos/the-stack-metadata')
the_stack_path = Path('/data/hf_repos/the-stack-v1.1')
repo_dst_root = Path('/repo_workdir/numpy_restored')
repo_name = 'numpy/numpy'
# Get bucket with numpy repo info
# meta_bucket_path = None
#for fn in tqdm(list((the_stack_meta_path/'data').glob('*/ri.parquet'))):
# df = pd.read_parquet(fn)
# if any(df['name'] == repo_name):
# meta_bucket_path = fn
# break
meta_bucket_path = the_stack_meta_path / 'data/255_944'
# Get repository id from repo name
ri_id = pd.read_parquet(
meta_bucket_path / 'ri.parquet'
).query(
f'`name` == "{repo_name}"'
)['id'].to_list()[0]
# Get files information for the reopository
files_info = pd.read_parquet(
meta_bucket_path / 'fi.parquet'
).query(
f'`ri_id` == {ri_id} and `size` != 0 and `is_deleted` == False'
)
# Convert DF with files information to a dictionary by language and then file hexsha
# there can be more than one file with the same hexsha in the repo so we gather
# all instances per unique hexsha
files_info_dict = {
k: v[['hexsha', 'path']].groupby('hexsha').apply(lambda x: list(x['path'])).to_dict()
for k, v in files_info.groupby('lang_ex')
}
# Load Python part of The Stack
ds = datasets.load_dataset(
str(the_stack_path/'data/python'),
num_proc=10, ignore_verifications=True
)
# Save file content of the python files in the numpy reposirotry in their appropriate locations
def save_file_content(example, files_info_dict, repo_dst_root):
if example['hexsha'] in files_info_dict:
for el in files_info_dict[example['hexsha']]:
path = repo_dst_root / el
path.parent.mkdir(parents=True, exist_ok=True)
path.write_text(example['content'])
ds.map(
save_file_content,
fn_kwargs={'files_info_dict': files_info_dict['Python'], 'repo_dst_root': repo_dst_root},
num_proc=10
)
```
## Dataset Creation
Please refer to [the section](https://huggingface.co/datasets/bigcode/the-stack#dataset-creation) in The Stack.
## Considerations for Using the Data
Please refer to [the section](https://huggingface.co/datasets/bigcode/the-stack#considerations-for-using-the-data) in The Stack.
## Additional Information
Please refer to [the section](https://huggingface.co/datasets/bigcode/the-stack#additional-information) in The Stack.
## Terms of Use for The Stack
Please refer to [the section](https://huggingface.co/datasets/bigcode/the-stack#terms-of-use-for-the-stack) in The Stack. | [
-0.5771356225013733,
-0.18708331882953644,
0.0888160914182663,
0.23876595497131348,
-0.2861611545085907,
-0.0569390207529068,
0.04967242851853371,
-0.07762430608272552,
0.4950277507305145,
0.7887977957725525,
-0.4692462384700775,
-0.8657298684120178,
-0.38229820132255554,
0.291829288005828... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
akirasosa/laion-mini-ja | akirasosa | 2022-12-28T04:33:54Z | 12 | 0 | null | [
"region:us"
] | 2022-12-28T04:33:54Z | 2022-12-28T04:11:03.000Z | 2022-12-28T04:11:03 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
taejunkim/beats | taejunkim | 2022-12-28T06:50:44Z | 12 | 0 | null | [
"region:us"
] | 2022-12-28T06:50:44Z | 2022-12-28T06:50:28.000Z | 2022-12-28T06:50:28 | ---
dataset_info:
features:
- name: mix_id
dtype: string
- name: beats
sequence: float64
splits:
- name: train
num_bytes: 1479883
num_examples: 13
download_size: 1119868
dataset_size: 1479883
---
# Dataset Card for "beats"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.6293665170669556,
-0.15773053467273712,
0.26815447211265564,
0.23729044198989868,
-0.20448338985443115,
0.07778789848089218,
0.17924578487873077,
-0.23576436936855316,
1.0927064418792725,
0.34017667174339294,
-1.1057791709899902,
-0.8568328619003296,
-0.5412177443504333,
-0.270925492048... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
hellosimple/dataset-demo | hellosimple | 2022-12-28T07:19:37Z | 12 | 0 | null | [
"license:mit",
"region:us"
] | 2022-12-28T07:19:37Z | 2022-12-28T07:18:49.000Z | 2022-12-28T07:18:49 | ---
license: mit
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Marlenka/alpaco | Marlenka | 2022-12-28T07:54:29Z | 12 | 0 | null | [
"region:us"
] | 2022-12-28T07:54:29Z | 2022-12-28T07:37:06.000Z | 2022-12-28T07:37:06 | Entry not found | [
-0.3227647542953491,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965083122253,
0.7915717959403992,
0.07618629932403564,
0.7746022343635559,
0.2563222348690033,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
ddd2jj/pawn2 | ddd2jj | 2022-12-28T07:46:38Z | 12 | 0 | null | [
"region:us"
] | 2022-12-28T07:46:38Z | 2022-12-28T07:41:42.000Z | 2022-12-28T07:41:42 | Entry not found | [
-0.3227647542953491,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965083122253,
0.7915717959403992,
0.07618629932403564,
0.7746022343635559,
0.2563222348690033,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
taejunkim/alignments | taejunkim | 2022-12-28T07:49:17Z | 12 | 0 | null | [
"region:us"
] | 2022-12-28T07:49:17Z | 2022-12-28T07:48:57.000Z | 2022-12-28T07:48:57 | ---
dataset_info:
features:
- name: mix_id
dtype: string
- name: track_id
dtype: string
- name: case_name
dtype: string
- name: feature
dtype: string
- name: metric
dtype: string
- name: key_change
dtype: int64
- name: match_rate
dtype: float64
- name: match_rate_raw
dtype: float64
- name: matched_beats
dtype: int64
- name: matched_beats_raw
dtype: int64
- name: matched_time_mix
dtype: float64
- name: matched_time_track
dtype: float64
- name: mix_cue_in_beat
dtype: float64
- name: mix_cue_out_beat
dtype: float64
- name: track_cue_in_beat
dtype: float64
- name: track_cue_out_beat
dtype: float64
- name: mix_cue_in_time
dtype: float64
- name: mix_cue_out_time
dtype: float64
- name: track_cue_in_time
dtype: float64
- name: track_cue_out_time
dtype: float64
- name: cost
dtype: float64
- name: __index_level_0__
dtype: int64
- name: wp
sequence:
sequence: int64
splits:
- name: train
num_bytes: 22961341
num_examples: 6600
download_size: 3089520
dataset_size: 22961341
---
# Dataset Card for "alignments"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.49622613191604614,
-0.05266213044524193,
0.194102942943573,
0.14563845098018646,
-0.07995215803384781,
0.036238882690668106,
0.275990754365921,
-0.26454928517341614,
0.9621673822402954,
0.31084397435188293,
-0.8665756583213806,
-0.7878164649009705,
-0.5652219653129578,
-0.25879496335983... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
com0040/only | com0040 | 2022-12-28T11:50:21Z | 12 | 0 | null | [
"region:us"
] | 2022-12-28T11:50:21Z | 2022-12-28T08:22:40.000Z | 2022-12-28T08:22:40 | Entry not found | [
-0.3227647542953491,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965083122253,
0.7915717959403992,
0.07618629932403564,
0.7746022343635559,
0.2563222348690033,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
dvilasuero/banking_app | dvilasuero | 2022-12-29T13:25:35Z | 12 | 0 | null | [
"region:us"
] | 2022-12-29T13:25:35Z | 2022-12-29T13:25:10.000Z | 2022-12-29T13:25:10 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622264862060547,
0.43461528420448303,
-0.52829909324646,
0.7012971639633179,
0.7915720343589783,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104477167129517,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
theatticusproject/cuad | theatticusproject | 2023-01-02T22:36:46Z | 12 | 1 | null | [
"license:cc-by-4.0",
"region:us"
] | 2023-01-02T22:36:46Z | 2023-01-02T21:54:27.000Z | 2023-01-02T21:54:27 | ---
license: cc-by-4.0
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
TREC-AToMiC/AToMiC-Texts-v0.2 | TREC-AToMiC | 2023-02-14T21:30:37Z | 12 | 0 | null | [
"size_categories:100M<n<1B",
"license:cc-by-sa-4.0",
"arxiv:2103.01913",
"region:us"
] | 2023-02-14T21:30:37Z | 2023-01-03T04:29:46.000Z | 2023-01-03T04:29:46 | ---
dataset_info:
features:
- name: text_id
dtype: string
- name: page_url
dtype: string
- name: page_title
dtype: string
- name: section_title
dtype: string
- name: context_page_description
dtype: string
- name: context_section_description
dtype: string
- name: media
sequence: string
- name: hierachy
sequence: string
- name: category
sequence: string
- name: source_id
dtype: string
splits:
- name: train
num_bytes: 14378574060.336058
num_examples: 10134744
download_size: 6408012391
dataset_size: 14378574060.336058
license: cc-by-sa-4.0
size_categories:
- 100M<n<1B
---
# Dataset Card for "AToMiC-Texts-Mapped"
## Dataset Description
- **Homepage:** [AToMiC homepage](https://trec-atomic.github.io/)
- **Source:** [WIT](https://github.com/google-research-datasets/wit)
- **Paper:** [WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual Machine Learning](https://arxiv.org/abs/2103.01913)
### Languages
This dataset only contains English in Wikipedia (parsed from the 20221101 XML dump).
### Data Instances
Each instance is a section of a Wikipedia page. We also provide its page-level information, and associated information such as categories and media.
The `source_id` can be mapped back to the instance in the original [WIT instance](https://github.com/google-research-datasets/wit/blob/main/DATA.md).
Notice that the WIT dataset is crawled from the earlier version of Wikipedia (2020-08-30).
The WIT dataset is mapped to the new dump by pure BM25 matching with [Anserini](https://github.com/castorini/anserini).
### Intended Usage
1. Text collection for Image-to-Text retrieval
2. Language model pretraining
3. Document classification
### Licensing Information
[CC BY-SA 4.0 international license](https://creativecommons.org/licenses/by-sa/4.0/)
### Citation Information
TBA
### Acknowledgement
Thanks to:
[mwparserfromhell](https://github.com/earwig/mwparserfromhell)
[Datasets](https://github.com/huggingface/datasets)
[Anserini](https://github.com/castorini/anserini)
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.5326377153396606,
-0.5886614918708801,
0.36725303530693054,
-0.11899543553590775,
-0.12919853627681732,
-0.20468974113464355,
-0.5011049509048462,
-0.2944892942905426,
0.19713672995567322,
0.294734388589859,
-0.6793920397758484,
-0.8970690369606018,
-0.3777366578578949,
0.20203685760498... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
carlosejimenez/bookcorpus_filtered_len_17_simcse_retrieval_top32__source_tranch_18__target_tranch_10__from_120 | carlosejimenez | 2023-01-04T02:55:24Z | 12 | 0 | null | [
"region:us"
] | 2023-01-04T02:55:24Z | 2023-01-04T02:55:03.000Z | 2023-01-04T02:55:03 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622264862060547,
0.43461528420448303,
-0.52829909324646,
0.7012971639633179,
0.7915720343589783,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104477167129517,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
carlosejimenez/bookcorpus_filtered_len_17_simcse_retrieval_top32__source_tranch_17__target_tranch_13__from_120 | carlosejimenez | 2023-01-04T02:55:39Z | 12 | 0 | null | [
"region:us"
] | 2023-01-04T02:55:39Z | 2023-01-04T02:55:25.000Z | 2023-01-04T02:55:25 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
carlosejimenez/bookcorpus_filtered_len_17_simcse_retrieval_top32__source_tranch_15__target_tranch_1__from_120 | carlosejimenez | 2023-01-04T02:56:29Z | 12 | 0 | null | [
"region:us"
] | 2023-01-04T02:56:29Z | 2023-01-04T02:56:13.000Z | 2023-01-04T02:56:13 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
carlosejimenez/bookcorpus_filtered_len_17_simcse_retrieval_top32__source_tranch_16__target_tranch_9__from_120 | carlosejimenez | 2023-01-04T02:56:31Z | 12 | 0 | null | [
"region:us"
] | 2023-01-04T02:56:31Z | 2023-01-04T02:56:15.000Z | 2023-01-04T02:56:15 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Boadiwaa/heart_data | Boadiwaa | 2023-01-04T16:18:19Z | 12 | 0 | null | [
"region:us"
] | 2023-01-04T16:18:19Z | 2023-01-04T16:17:26.000Z | 2023-01-04T16:17:26 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Umal-exvc/test-captioned-dataset | Umal-exvc | 2023-01-04T16:23:44Z | 12 | 0 | null | [
"region:us"
] | 2023-01-04T16:23:44Z | 2023-01-04T16:23:40.000Z | 2023-01-04T16:23:40 | ---
dataset_info:
features:
- name: image
dtype: image
- name: text
dtype: string
splits:
- name: train
num_bytes: 111187.0
num_examples: 5
download_size: 111705
dataset_size: 111187.0
---
# Dataset Card for "test-captioned-dataset"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.6395055055618286,
-0.28833919763565063,
0.08057824522256851,
0.3144600987434387,
-0.31555768847465515,
0.20413349568843842,
0.22837454080581665,
-0.024589017033576965,
0.7099284529685974,
0.5649012327194214,
-0.8694941997528076,
-0.5757884979248047,
-0.5737922787666321,
-0.0667070895433... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
kmewhort/tu-berlin-svgs | kmewhort | 2023-01-10T19:20:44Z | 12 | 1 | null | [
"region:us"
] | 2023-01-10T19:20:44Z | 2023-01-04T16:34:42.000Z | 2023-01-04T16:34:42 | ---
dataset_info:
features:
- name: label
dtype:
class_label:
names:
'0': airplane
'1': alarm clock
'2': angel
'3': ant
'4': apple
'5': arm
'6': armchair
'7': ashtray
'8': axe
'9': backpack
'10': banana
'11': barn
'12': baseball bat
'13': basket
'14': bathtub
'15': bear (animal)
'16': bed
'17': bee
'18': beer-mug
'19': bell
'20': bench
'21': bicycle
'22': binoculars
'23': blimp
'24': book
'25': bookshelf
'26': boomerang
'27': bottle opener
'28': bowl
'29': brain
'30': bread
'31': bridge
'32': bulldozer
'33': bus
'34': bush
'35': butterfly
'36': cabinet
'37': cactus
'38': cake
'39': calculator
'40': camel
'41': camera
'42': candle
'43': cannon
'44': canoe
'45': car (sedan)
'46': carrot
'47': castle
'48': cat
'49': cell phone
'50': chair
'51': chandelier
'52': church
'53': cigarette
'54': cloud
'55': comb
'56': computer monitor
'57': computer-mouse
'58': couch
'59': cow
'60': crab
'61': crane (machine)
'62': crocodile
'63': crown
'64': cup
'65': diamond
'66': dog
'67': dolphin
'68': donut
'69': door
'70': door handle
'71': dragon
'72': duck
'73': ear
'74': elephant
'75': envelope
'76': eye
'77': eyeglasses
'78': face
'79': fan
'80': feather
'81': fire hydrant
'82': fish
'83': flashlight
'84': floor lamp
'85': flower with stem
'86': flying bird
'87': flying saucer
'88': foot
'89': fork
'90': frog
'91': frying-pan
'92': giraffe
'93': grapes
'94': grenade
'95': guitar
'96': hamburger
'97': hammer
'98': hand
'99': harp
'100': hat
'101': head
'102': head-phones
'103': hedgehog
'104': helicopter
'105': helmet
'106': horse
'107': hot air balloon
'108': hot-dog
'109': hourglass
'110': house
'111': human-skeleton
'112': ice-cream-cone
'113': ipod
'114': kangaroo
'115': key
'116': keyboard
'117': knife
'118': ladder
'119': laptop
'120': leaf
'121': lightbulb
'122': lighter
'123': lion
'124': lobster
'125': loudspeaker
'126': mailbox
'127': megaphone
'128': mermaid
'129': microphone
'130': microscope
'131': monkey
'132': moon
'133': mosquito
'134': motorbike
'135': mouse (animal)
'136': mouth
'137': mug
'138': mushroom
'139': nose
'140': octopus
'141': owl
'142': palm tree
'143': panda
'144': paper clip
'145': parachute
'146': parking meter
'147': parrot
'148': pear
'149': pen
'150': penguin
'151': person sitting
'152': person walking
'153': piano
'154': pickup truck
'155': pig
'156': pigeon
'157': pineapple
'158': pipe (for smoking)
'159': pizza
'160': potted plant
'161': power outlet
'162': present
'163': pretzel
'164': pumpkin
'165': purse
'166': rabbit
'167': race car
'168': radio
'169': rainbow
'170': revolver
'171': rifle
'172': rollerblades
'173': rooster
'174': sailboat
'175': santa claus
'176': satellite
'177': satellite dish
'178': saxophone
'179': scissors
'180': scorpion
'181': screwdriver
'182': sea turtle
'183': seagull
'184': shark
'185': sheep
'186': ship
'187': shoe
'188': shovel
'189': skateboard
'190': skull
'191': skyscraper
'192': snail
'193': snake
'194': snowboard
'195': snowman
'196': socks
'197': space shuttle
'198': speed-boat
'199': spider
'200': sponge bob
'201': spoon
'202': squirrel
'203': standing bird
'204': stapler
'205': strawberry
'206': streetlight
'207': submarine
'208': suitcase
'209': sun
'210': suv
'211': swan
'212': sword
'213': syringe
'214': t-shirt
'215': table
'216': tablelamp
'217': teacup
'218': teapot
'219': teddy-bear
'220': telephone
'221': tennis-racket
'222': tent
'223': tiger
'224': tire
'225': toilet
'226': tomato
'227': tooth
'228': toothbrush
'229': tractor
'230': traffic light
'231': train
'232': tree
'233': trombone
'234': trousers
'235': truck
'236': trumpet
'237': tv
'238': umbrella
'239': van
'240': vase
'241': violin
'242': walkie talkie
'243': wheel
'244': wheelbarrow
'245': windmill
'246': wine-bottle
'247': wineglass
'248': wrist-watch
'249': zebra
- name: svg
dtype: string
splits:
- name: train
num_bytes: 82640829.32506625
num_examples: 15999
- name: test
num_bytes: 20661498.674933746
num_examples: 4000
download_size: 65748314
dataset_size: 103302328.0
---
# Dataset Card for "tu-berlin-svgs"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.7535617351531982,
-0.15976256132125854,
0.2313520312309265,
0.2302398830652237,
-0.42150554060935974,
0.02562454342842102,
0.45438721776008606,
-0.11452003568410873,
0.7514451146125793,
0.2662273645401001,
-0.7762050032615662,
-0.9845258593559265,
-0.6359640955924988,
-0.289115220308303... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
eBoreal/scrat-imgs | eBoreal | 2023-01-04T18:19:14Z | 12 | 0 | null | [
"region:us"
] | 2023-01-04T18:19:14Z | 2023-01-04T18:18:37.000Z | 2023-01-04T18:18:37 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
fgomeza17/Sammy | fgomeza17 | 2023-01-04T19:49:05Z | 12 | 0 | null | [
"license:openrail",
"region:us"
] | 2023-01-04T19:49:05Z | 2023-01-04T19:48:29.000Z | 2023-01-04T19:48:29 | ---
license: openrail
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
uumlaut/VanGoghPaintings | uumlaut | 2023-01-04T21:18:46Z | 12 | 0 | null | [
"region:us"
] | 2023-01-04T21:18:46Z | 2023-01-04T20:54:52.000Z | 2023-01-04T20:54:52 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
epaolinos/septuagint | epaolinos | 2023-01-04T21:31:19Z | 12 | 0 | null | [
"region:us"
] | 2023-01-04T21:31:19Z | 2023-01-04T21:31:08.000Z | 2023-01-04T21:31:08 | ---
dataset_info:
features:
- name: Book
dtype: string
- name: Chapter
dtype: int64
- name: Verse Number
dtype: int64
- name: Verse Text
dtype: string
- name: Genre
dtype: string
splits:
- name: train
num_bytes: 9101054
num_examples: 30568
download_size: 3421032
dataset_size: 9101054
---
# Dataset Card for "septuagint"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.27498766779899597,
-0.3355090916156769,
0.11248943954706192,
0.18951325118541718,
-0.47797879576683044,
0.08997127413749695,
-0.06669862568378448,
-0.22201977670192719,
1.1001300811767578,
0.5285810232162476,
-0.6512641310691833,
-0.7917959690093994,
-0.7502202391624451,
0.0205996930599... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Amy12zz/dreambooth-hackathon-images | Amy12zz | 2023-01-04T22:05:25Z | 12 | 0 | null | [
"region:us"
] | 2023-01-04T22:05:25Z | 2023-01-04T22:05:18.000Z | 2023-01-04T22:05:18 | ---
dataset_info:
features:
- name: image
dtype: image
splits:
- name: train
num_bytes: 1047395.0
num_examples: 4
download_size: 1047434
dataset_size: 1047395.0
---
# Dataset Card for "dreambooth-hackathon-images"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.42646852135658264,
-0.13476064801216125,
0.23689782619476318,
0.23072221875190735,
-0.40631160140037537,
0.3039889335632324,
0.32873836159706116,
-0.32934942841529846,
1.1760584115982056,
0.36833885312080383,
-0.8234314322471619,
-0.6366724967956543,
-0.3446301519870758,
-0.267739206552... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
RobertLucian/avatar-10k | RobertLucian | 2023-01-04T22:15:52Z | 12 | 0 | null | [
"license:gpl-3.0",
"region:us"
] | 2023-01-04T22:15:52Z | 2023-01-04T22:12:33.000Z | 2023-01-04T22:12:33 | ---
license: gpl-3.0
dataset_info:
features:
- name: image
dtype: image
- name: text
dtype: string
splits:
- name: train
num_bytes: 771570808.685
num_examples: 10689
download_size: 646236257
dataset_size: 771570808.685
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
ivelin/processed_sroie_donut_dataset_json2token | ivelin | 2023-01-05T00:19:38Z | 12 | 0 | null | [
"region:us"
] | 2023-01-05T00:19:38Z | 2023-01-05T00:19:04.000Z | 2023-01-05T00:19:04 | ---
dataset_info:
features:
- name: image
dtype: image
- name: text
dtype: string
splits:
- name: train
num_bytes: 586245601.0
num_examples: 626
download_size: 577293738
dataset_size: 586245601.0
---
# Dataset Card for "processed_sroie_donut_dataset_json2token"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
0.0009039849974215031,
-0.16903114318847656,
0.1913672536611557,
0.23318858444690704,
-0.2850587069988251,
0.3050466775894165,
0.11789344996213913,
-0.08738389611244202,
0.8015451431274414,
0.7641820907592773,
-0.8002272844314575,
-0.5041705965995789,
-0.8640604019165039,
-0.14679706096649... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Atallahw/clippy | Atallahw | 2023-01-08T02:24:50Z | 12 | 0 | null | [
"region:us"
] | 2023-01-08T02:24:50Z | 2023-01-05T03:48:19.000Z | 2023-01-05T03:48:19 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
inkoziev/paraphrases | inkoziev | 2023-01-14T13:37:24Z | 12 | 2 | null | [
"task_categories:sentence-similarity",
"task_categories:text2text-generation",
"task_ids:semantic-similarity-classification",
"language_creators:expert-generated",
"language:ru",
"license:cc-by-nc-4.0",
"region:us"
] | 2023-01-14T13:37:24Z | 2023-01-05T09:08:02.000Z | 2023-01-05T09:08:02 | ---
license: cc-by-nc-4.0
language:
- ru
language_creators:
- expert-generated
task_categories:
- sentence-similarity
- text2text-generation
task_ids:
- semantic-similarity-classification
---
# Датасет перефразировок коротких фраз (читчат+поэзия)
В датасете содержатся правильные и некорректные перефразировки коротких диалоговых реплик ([проект диалоговой системы](https://github.com/Koziev/chatbot))
и фрагментов стихов ([проект генеративной поэзии](https://github.com/Koziev/verslibre)).
Датасет представляет из себя список сэмплов-кортежей. Каждый сэмпл состоит из двух списков:
```paraphrases``` - примеры правильных перефразировок
```distractors``` - примеры неправильных перефразировок
Датасет используется для создания моделей [детектора перефразировок sbert_synonymy](https://huggingface.co/inkoziev/sbert_synonymy)
и [генеративного поэтического перефразировщика](https://huggingface.co/inkoziev/paraphraser).
## Disclaimer
В датасете целенаправленно допускалась неконсервативность семантики перефразировок в определенных пределах.
К примеру, правильными перефразировками считаются пары "_Помолчи_" и "_Дружище, не говори ни слова!_". Так как перефразировщик
используется в проекте генеративной поэзии для создания датасетов, в нем есть некоторое количество метафоричных
и достаточно вольных перефразировок. Эти особенности датасета могут сделать невозможным использование датасета и моделей
на его основе в Ваших проектах.
## Другие датасеты перефразировок
При обучении моделей вы можете совмещать этот датасет с данными из других датасетов перефразировок, например [tapaco](https://huggingface.co/datasets/tapaco).
| [
-0.5610749125480652,
-0.6993847489356995,
0.3382612466812134,
0.49490559101104736,
-0.6506953835487366,
-0.12735572457313538,
-0.08409928530454636,
-0.32450634241104126,
0.6174204349517822,
0.2705595791339874,
-0.5864480137825012,
-0.535758912563324,
-0.5358456969261169,
0.1904322504997253... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Joanne/katz1980_set_A | Joanne | 2023-01-05T11:36:13Z | 12 | 0 | null | [
"region:us"
] | 2023-01-05T11:36:13Z | 2023-01-05T11:36:04.000Z | 2023-01-05T11:36:04 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Joanne/katz1980_set_B | Joanne | 2023-01-05T11:36:33Z | 12 | 0 | null | [
"region:us"
] | 2023-01-05T11:36:33Z | 2023-01-05T11:36:24.000Z | 2023-01-05T11:36:24 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Someman/nepali-flag | Someman | 2023-01-05T11:45:55Z | 12 | 0 | null | [
"license:mit",
"region:us"
] | 2023-01-05T11:45:55Z | 2023-01-05T11:44:53.000Z | 2023-01-05T11:44:53 | ---
license: mit
---
| [
-0.12853367626667023,
-0.18616794049739838,
0.6529126763343811,
0.4943627417087555,
-0.19319313764572144,
0.23607443273067474,
0.36071979999542236,
0.05056338757276535,
0.5793654322624207,
0.7400138974189758,
-0.6508103013038635,
-0.23783987760543823,
-0.710224986076355,
-0.047825977206230... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Joanne/cardillo_2010_2017_lexical | Joanne | 2023-01-05T11:49:15Z | 12 | 0 | null | [
"region:us"
] | 2023-01-05T11:49:15Z | 2023-01-05T11:49:05.000Z | 2023-01-05T11:49:05 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Joanne/cardillo_2010_2017_random | Joanne | 2023-01-05T11:49:26Z | 12 | 0 | null | [
"region:us"
] | 2023-01-05T11:49:26Z | 2023-01-05T11:49:17.000Z | 2023-01-05T11:49:17 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Joanne/katz1980_random | Joanne | 2023-01-05T11:50:28Z | 12 | 0 | null | [
"region:us"
] | 2023-01-05T11:50:28Z | 2023-01-05T11:50:18.000Z | 2023-01-05T11:50:18 | Entry not found | [
-0.3227647542953491,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965083122253,
0.7915717959403992,
0.07618629932403564,
0.7746022343635559,
0.2563222348690033,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Joanne/katz1980_lexical | Joanne | 2023-01-05T11:51:50Z | 12 | 0 | null | [
"region:us"
] | 2023-01-05T11:51:50Z | 2023-01-05T11:51:40.000Z | 2023-01-05T11:51:40 | Entry not found | [
-0.3227647542953491,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965083122253,
0.7915717959403992,
0.07618629932403564,
0.7746022343635559,
0.2563222348690033,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
vencortex/News | vencortex | 2023-01-05T14:14:16Z | 12 | 0 | null | [
"region:us"
] | 2023-01-05T14:14:16Z | 2023-01-05T14:13:58.000Z | 2023-01-05T14:13:58 | ---
dataset_info:
features:
- name: symbol
dtype: string
- name: publishedDate
dtype: string
- name: title
dtype: string
- name: image
dtype: string
- name: site
dtype: string
- name: text
dtype: string
- name: url
dtype: string
- name: type
dtype: string
splits:
- name: train
num_bytes: 834852911
num_examples: 1495869
download_size: 170603751
dataset_size: 834852911
---
# Dataset Card for "News"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.5475593209266663,
-0.3586120307445526,
0.2984564006328583,
0.2909470200538635,
-0.4338994324207306,
-0.011733530089259148,
0.12641993165016174,
-0.2080702781677246,
1.070223331451416,
0.4256940484046936,
-0.8439085483551025,
-0.9641247391700745,
-0.6569112539291382,
-0.36435943841934204... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
pyakymenko/test_6k | pyakymenko | 2023-01-05T14:50:56Z | 12 | 0 | null | [
"region:us"
] | 2023-01-05T14:50:56Z | 2023-01-05T14:33:32.000Z | 2023-01-05T14:33:32 | ---
dataset_info:
features:
- name: audio
dtype: audio
- name: sentence
dtype: string
splits:
- name: train
num_bytes: 475682224.444
num_examples: 6661
download_size: 473720429
dataset_size: 475682224.444
---
# Dataset Card for "test_6k"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.6063244342803955,
-0.09156772494316101,
0.03183003515005112,
0.1675521731376648,
-0.3009108901023865,
-0.16979393362998962,
0.39194563031196594,
-0.24883945286273956,
0.7301656007766724,
0.4608059525489807,
-0.7437134385108948,
-0.6642099022865295,
-0.5005666613578796,
-0.02653062902390... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
bstrai/subj_multi | bstrai | 2023-01-17T17:17:15Z | 12 | 0 | null | [
"region:us"
] | 2023-01-17T17:17:15Z | 2023-01-05T16:35:39.000Z | 2023-01-05T16:35:39 | ---
dataset_info:
features:
- name: text
dtype: string
- name: label
dtype:
class_label:
names:
'0': objective
'1': subjective
- name: language
dtype: string
splits:
- name: test
num_bytes: 2914488
num_examples: 16000
- name: train
num_bytes: 11518066
num_examples: 64000
download_size: 8870704
dataset_size: 14432554
---
# Dataset Card for "subj_multi"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.8047283291816711,
-0.2404509335756302,
0.004799854476004839,
0.2841188311576843,
-0.30597779154777527,
0.12157490104436874,
0.2525589168071747,
-0.09950873255729675,
0.8763295412063599,
0.3081164062023163,
-0.9213322401046753,
-0.6557028293609619,
-0.6590276956558228,
-0.077562682330608... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
fuyulinh04/dataset_glstxt | fuyulinh04 | 2023-01-05T23:21:14Z | 12 | 0 | null | [
"region:us"
] | 2023-01-05T23:21:14Z | 2023-01-05T23:20:43.000Z | 2023-01-05T23:20:43 | ---
dataset_info:
features:
- name: gloss
dtype: string
- name: text
dtype: string
splits:
- name: train
num_bytes: 11227076.8
num_examples: 73696
- name: test
num_bytes: 2806769.2
num_examples: 18424
download_size: 8513566
dataset_size: 14033846.0
---
# Dataset Card for "dataset_glstxt"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.37770992517471313,
-0.14748430252075195,
0.3798990845680237,
0.11246412247419357,
-0.44321903586387634,
0.18284271657466888,
0.42760807275772095,
-0.247656911611557,
0.7942173480987549,
0.34282249212265015,
-0.8290347456932068,
-0.6926448941230774,
-0.6878188252449036,
-0.43358078598976... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
svjack/GLM-Open-Dialogue-Chinese-samples | svjack | 2023-01-06T02:01:30Z | 12 | 3 | null | [
"region:us"
] | 2023-01-06T02:01:30Z | 2023-01-06T01:52:11.000Z | 2023-01-06T01:52:11 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
kxly/princess_tutu | kxly | 2023-01-06T02:55:47Z | 12 | 3 | null | [
"language:en",
"license:creativeml-openrail-m",
"stable-diffusion",
"text-to-image",
"image-to-image",
"region:us"
] | 2023-01-06T02:55:47Z | 2023-01-06T02:00:33.000Z | 2023-01-06T02:00:33 | ---
language:
- en
license: creativeml-openrail-m
thumbnail: >-
https://huggingface.co/datasets/kxly/princess_tutu/blob/main/princess_tutu_showcase.png
tags:
- stable-diffusion
- text-to-image
- image-to-image
inference: false
pretty_name: Princess Tutu
---
# Character Embedding - Princess Tutu/Ahiru

## Usage
To use an embedding, download the .pt file and place it in "\stable-diffusion-webui\embeddings".
In your prompt, write ```"princess_tutu-6500"```.
## License
This embedding is open access and available to all, with a CreativeML OpenRAIL-M license further specifying rights and usage.
The CreativeML OpenRAIL License specifies:
1. You can't use the embedding to deliberately produce nor share illegal or harmful outputs or content
2. The authors claim no rights on the outputs you generate, you are free to use them and are accountable for their use which must not go against the provisions set in the license
3. You may re-distribute the weights and use the embedding commercially and/or as a service. If you do, please be aware you have to include the same use restrictions as the ones in the license and share a copy of the CreativeML OpenRAIL-M to all your users (please read the license entirely and carefully)
[Please read the full license here](https://huggingface.co/spaces/CompVis/stable-diffusion-license) | [
-0.46248459815979004,
-0.6287046074867249,
0.2293298840522766,
0.6610038876533508,
-0.5865913033485413,
-0.08370287716388702,
0.1223154067993164,
-0.08301373571157455,
0.31182926893234253,
0.6175093054771423,
-0.42990773916244507,
-0.5265506505966187,
-0.7429785132408142,
-0.00740587338805... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
sanjin7/embedding_dataset_distilbert_base_uncased_ad_subwords | sanjin7 | 2023-01-16T11:12:24Z | 12 | 0 | null | [
"region:us"
] | 2023-01-16T11:12:24Z | 2023-01-08T07:54:44.000Z | 2023-01-08T07:54:44 | ---
dataset_info:
features:
- name: ad_id
dtype: int64
- name: shop_id
dtype: int64
- name: account_id
dtype: int64
- name: mean_embedding
sequence: float32
- name: cls_embedding
sequence: float32
splits:
- name: test
num_bytes: 5725152
num_examples: 927
- name: train
num_bytes: 43769312
num_examples: 7087
- name: val
num_bytes: 7726176
num_examples: 1251
download_size: 69324552
dataset_size: 57220640
---
# Dataset Card for "embedding_dataset_distilbert_base_uncased_ad_subwords"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.5727691650390625,
-0.5441336631774902,
0.10706023126840591,
0.1042877808213234,
-0.5628413558006287,
0.31765782833099365,
0.051310066133737564,
0.3212378919124603,
0.8960884213447571,
0.24476073682308197,
-0.6493831276893616,
-1.0216084718704224,
-0.9090348482131958,
-0.2815321385860443... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
muellerzr/github-pr-history | muellerzr | 2023-01-08T15:29:01Z | 12 | 0 | null | [
"size_categories:n<1K",
"language:en",
"license:mit",
"region:us"
] | 2023-01-08T15:29:01Z | 2023-01-08T13:34:38.000Z | 2023-01-08T13:34:38 | ---
license: mit
language:
- en
pretty_name: Github Pull Request History
size_categories:
- n<1K
---
# What is this dataset?
This dataset is a collection of Pull Requests **that contain comments** from the [Accelerate](https://github.com/huggingface/accelerate).
It contains the full contextual comments as well as code suggestions that exist inside of a code review | [
-0.570921003818512,
-0.5022136569023132,
0.12390604615211487,
0.24383775889873505,
-0.15386725962162018,
0.37677276134490967,
0.1261993795633316,
-0.18122529983520508,
0.6459078788757324,
0.6952884793281555,
-1.2829887866973877,
-0.17873422801494598,
-0.4543921947479248,
0.0139415217563509... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Wiebke/newtrain_reddit_gab_bert-base-casedepoch3 | Wiebke | 2023-01-08T13:40:03Z | 12 | 0 | null | [
"region:us"
] | 2023-01-08T13:40:03Z | 2023-01-08T13:37:45.000Z | 2023-01-08T13:37:45 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Zappandy/recipe_nlg | Zappandy | 2023-01-09T14:26:39Z | 12 | 3 | null | [
"license:apache-2.0",
"region:us"
] | 2023-01-09T14:26:39Z | 2023-01-08T13:41:47.000Z | 2023-01-08T13:41:47 | ---
license: apache-2.0
--- | [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Wiebke/newtrain_reddit_gab_bert-base-casedepoch3_equal | Wiebke | 2023-01-08T13:46:53Z | 12 | 0 | null | [
"region:us"
] | 2023-01-08T13:46:53Z | 2023-01-08T13:45:28.000Z | 2023-01-08T13:45:28 | Entry not found | [
-0.3227647542953491,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965083122253,
0.7915717959403992,
0.07618629932403564,
0.7746022343635559,
0.2563222348690033,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
lvwerra/changeit | lvwerra | 2023-01-08T15:43:02Z | 12 | 0 | null | [
"region:us"
] | 2023-01-08T15:43:02Z | 2023-01-08T14:49:08.000Z | 2023-01-08T14:49:08 | Entry not found | [
-0.3227647542953491,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965083122253,
0.7915717959403992,
0.07618629932403564,
0.7746022343635559,
0.2563222348690033,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
michaelb1225/open-cm | michaelb1225 | 2023-01-08T15:23:21Z | 12 | 0 | null | [
"region:us"
] | 2023-01-08T15:23:21Z | 2023-01-08T15:09:26.000Z | 2023-01-08T15:09:26 | ---
dataset_info:
features:
- name: audio
dtype: audio
splits:
- name: train
num_bytes: 6129427551.0
num_examples: 671
download_size: 6071742068
dataset_size: 6129427551.0
---
# Dataset Card for "open-cm"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.7530912160873413,
-0.26227155327796936,
0.30696338415145874,
-0.029665417969226837,
-0.2290700525045395,
-0.26806291937828064,
-0.1338856816291809,
-0.030914105474948883,
0.8248491883277893,
0.5330220460891724,
-0.9761363863945007,
-1.029687762260437,
-0.5515740513801575,
-0.37701976299... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
vienduong88/Neyun | vienduong88 | 2023-01-08T16:54:10Z | 12 | 0 | null | [
"license:openrail",
"region:us"
] | 2023-01-08T16:54:10Z | 2023-01-08T16:47:18.000Z | 2023-01-08T16:47:18 | ---
license: openrail
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
vishnun/SpellGram | vishnun | 2023-01-09T13:43:11Z | 12 | 0 | null | [
"task_categories:text2text-generation",
"size_categories:10K<n<100K",
"language:en",
"license:mit",
"NLP",
"Text2Text",
"region:us"
] | 2023-01-09T13:43:11Z | 2023-01-09T13:39:23.000Z | 2023-01-09T13:39:23 | ---
license: mit
task_categories:
- text2text-generation
language:
- en
tags:
- NLP
- Text2Text
pretty_name: Dataset consisting of grammatical and spelling errors
size_categories:
- 10K<n<100K
---
# SpellGram
## Dataset consisting of grammatical and spelling errors
- **Homepage:**
- **Repository:**
- **Paper:**
- **Leaderboard:**
- **Point of Contact:**
### Dataset Summary
[More Information Needed]
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[train.csv]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] | [
-0.031502239406108856,
-0.45691996812820435,
0.23213693499565125,
0.3967456519603729,
0.1335676610469818,
0.2275220900774002,
-0.22349920868873596,
-0.07148230820894241,
0.4629785418510437,
0.7333163619041443,
-0.7593204379081726,
-1.2123510837554932,
-0.8407378792762756,
0.524325966835022... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
swww/covidw | swww | 2023-01-17T12:15:45Z | 12 | 0 | null | [
"region:us"
] | 2023-01-17T12:15:45Z | 2023-01-17T10:18:44.000Z | 2023-01-17T10:18:44 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
paoloitaliani/tms_sentence | paoloitaliani | 2023-01-18T15:16:04Z | 12 | 0 | null | [
"region:us"
] | 2023-01-18T15:16:04Z | 2023-01-17T15:13:05.000Z | 2023-01-17T15:13:05 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
marianna13/random_dataset | marianna13 | 2023-02-25T14:35:53Z | 12 | 1 | null | [
"region:us"
] | 2023-02-25T14:35:53Z | 2023-01-17T15:44:41.000Z | 2023-01-17T15:44:41 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
abelc/italo-diffusion-256 | abelc | 2023-01-17T17:56:00Z | 12 | 0 | null | [
"region:us"
] | 2023-01-17T17:56:00Z | 2023-01-17T17:54:34.000Z | 2023-01-17T17:54:34 | ---
dataset_info:
features:
- name: image
dtype: image
- name: audio_file
dtype: string
- name: slice
dtype: int16
splits:
- name: train
num_bytes: 29319809.0
num_examples: 658
download_size: 29297971
dataset_size: 29319809.0
---
# Dataset Card for "italo-diffusion-256"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.7658651471138,
-0.15019221603870392,
0.3280940651893616,
0.5150654315948486,
-0.2189117968082428,
0.07714685797691345,
0.2775276303291321,
0.08862278610467911,
0.8335604071617126,
0.30025607347488403,
-0.7314122319221497,
-0.8444968461990356,
-0.6289349794387817,
-0.5386227965354919,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
baffo32/somegitdata | baffo32 | 2023-01-17T23:01:33Z | 12 | 0 | null | [
"region:us"
] | 2023-01-17T23:01:33Z | 2023-01-17T20:21:56.000Z | 2023-01-17T20:21:56 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
carexl8/telegram_de_ru | carexl8 | 2023-04-25T22:04:20Z | 12 | 0 | null | [
"region:us"
] | 2023-04-25T22:04:20Z | 2023-01-17T20:29:31.000Z | 2023-01-17T20:29:31 | ---
dataset_info:
features:
- name: id
dtype: string
- name: name
dtype: string
- name: time
dtype: string
- name: text
dtype: string
- name: tokens
sequence: string
- name: language tags
sequence: int64
- name: __index_level_0__
dtype: int64
splits:
- name: train
num_bytes: 5938949
num_examples: 10191
download_size: 1869587
dataset_size: 5938949
---
# Dataset Card for "telegram_de_ru"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.31594952940940857,
-0.4754468500614166,
0.04658455401659012,
0.5260713696479797,
-0.46412530541419983,
0.019661791622638702,
0.15459884703159332,
-0.25580501556396484,
0.9447314143180847,
0.4289696514606476,
-0.8484136462211609,
-0.8726778626441956,
-0.7160326242446899,
-0.1985431164503... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Tristan/olm-wikipedia-20221220-1-percent | Tristan | 2023-01-17T20:47:18Z | 12 | 0 | null | [
"region:us"
] | 2023-01-17T20:47:18Z | 2023-01-17T20:47:06.000Z | 2023-01-17T20:47:06 | ---
dataset_info:
features:
- name: id
dtype: string
- name: url
dtype: string
- name: title
dtype: string
- name: text
dtype: string
splits:
- name: train
num_bytes: 209366020.9708762
num_examples: 65879
download_size: 123017868
dataset_size: 209366020.9708762
---
# Dataset Card for "olm-wikipedia-20221220-1-percent"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.8340270519256592,
-0.43116384744644165,
0.271524578332901,
0.10575804114341736,
-0.06329254060983658,
-0.38436540961265564,
0.22831933200359344,
0.09554003179073334,
0.7661848068237305,
0.5482386946678162,
-0.9379180073738098,
-0.922636866569519,
-0.3649454414844513,
-0.3293671309947967... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
yuvalkirstain/beautiful_interesting_spectacular_photo_dog_25000 | yuvalkirstain | 2023-01-18T06:37:24Z | 12 | 0 | null | [
"region:us"
] | 2023-01-18T06:37:24Z | 2023-01-18T06:36:25.000Z | 2023-01-18T06:36:25 | ---
dataset_info:
features:
- name: image
dtype: image
- name: text
dtype: string
- name: width
dtype: int64
- name: height
dtype: int64
- name: pclean
dtype: float64
splits:
- name: train
num_bytes: 361773346.0
num_examples: 504
download_size: 361776700
dataset_size: 361773346.0
---
# Dataset Card for "beautiful_interesting_spectacular_photo_dog_25000"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.8284481167793274,
-0.2143736332654953,
0.1834106147289276,
0.38587138056755066,
-0.43337753415107727,
-0.19882304966449738,
0.3403880298137665,
-0.28982123732566833,
0.680835485458374,
0.28281107544898987,
-0.6531713604927063,
-0.8225438594818115,
-0.5930798053741455,
-0.288455009460449... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Joe02/obui | Joe02 | 2023-03-25T00:32:14Z | 12 | 0 | null | [
"license:other",
"region:us"
] | 2023-03-25T00:32:14Z | 2023-01-18T07:35:14.000Z | 2023-01-18T07:35:14 | ---
license: other
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
BeardedJohn/ubb-endava-conll-assistant-ner-only-misc-v2 | BeardedJohn | 2023-01-18T08:53:56Z | 12 | 0 | null | [
"region:us"
] | 2023-01-18T08:53:56Z | 2023-01-18T08:22:55.000Z | 2023-01-18T08:22:55 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
reshinthadith/dfg_augmented_mbpp | reshinthadith | 2023-01-18T09:27:02Z | 12 | 0 | null | [
"region:us"
] | 2023-01-18T09:27:02Z | 2023-01-18T09:26:49.000Z | 2023-01-18T09:26:49 | ---
dataset_info:
features:
- name: prompt
dtype: string
- name: output
dtype: string
- name: code
dtype: string
splits:
- name: train
num_bytes: 32138
num_examples: 95
download_size: 17897
dataset_size: 32138
---
# Dataset Card for "dfg_augmented_mbpp"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.7902991771697998,
-0.390556663274765,
0.21485590934753418,
0.42076602578163147,
-0.12265515327453613,
0.2076321542263031,
0.39486163854599,
-0.3015674650669098,
0.8437708020210266,
0.5149201154708862,
-0.6605716347694397,
-0.5967594981193542,
-0.7949752807617188,
-0.20612865686416626,
... | null | null | null | null | null | null | null | null | null | null | null | null | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.