id stringlengths 2 115 | author stringlengths 2 42 ⌀ | last_modified timestamp[us, tz=UTC] | downloads int64 0 8.87M | likes int64 0 3.84k | paperswithcode_id stringlengths 2 45 ⌀ | tags list | lastModified timestamp[us, tz=UTC] | createdAt stringlengths 24 24 | key stringclasses 1 value | created timestamp[us] | card stringlengths 1 1.01M | embedding list | library_name stringclasses 21 values | pipeline_tag stringclasses 27 values | mask_token null | card_data null | widget_data null | model_index null | config null | transformers_info null | spaces null | safetensors null | transformersInfo null | modelId stringlengths 5 111 ⌀ | embeddings list |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
decoy4600/sgm-output4 | decoy4600 | 2023-11-28T06:23:16Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T06:23:16Z | 2023-11-28T06:22:50.000Z | 2023-11-28T06:22:50 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622264862060547,
0.43461528420448303,
-0.52829909324646,
0.7012971639633179,
0.7915720343589783,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104477167129517,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
tolu07/Mental_Health_FAQ | tolu07 | 2023-11-28T06:33:34Z | 0 | 0 | null | [
"task_categories:conversational",
"task_categories:text-generation",
"license:mit",
"chatbot",
"mental health",
"therapy",
"region:us"
] | 2023-11-28T06:33:34Z | 2023-11-28T06:24:38.000Z | 2023-11-28T06:24:38 | ---
license: mit
task_categories:
- conversational
- text-generation
tags:
- chatbot
- mental health
- therapy
---
**Content**
Mental health includes our emotional, psychological, and social well-being. Mental health is integral to living a healthy, balanced life. It affects how we think, feel, and act. It also helps determine how we handle stress, relate to others, and make choices. Emotional and mental health is important because it’s a vital part of your life and impacts your thoughts, behaviors and emotions. Being healthy emotionally can promote productivity and effectiveness in activities like work, school or care-giving. It plays an important part in the health of your relationships, and allows you to adapt to changes in your life and cope with adversity. Mental health problems are common but help is available. People with mental health problems can get better and many recover completely.
This dataset consists of FAQs about Mental Health.
**Acknowledgements**
https://www.thekimfoundation.org/faqs/
https://www.mhanational.org/frequently-asked-questions
https://www.wellnessinmind.org/frequently-asked-questions/
https://www.heretohelp.bc.ca/questions-and-answers | [
-0.3746553659439087,
-0.8563477396965027,
0.6965852379798889,
0.3618142902851105,
0.01020369865000248,
0.04303928464651108,
0.19339382648468018,
-0.7386484146118164,
0.4142279624938965,
0.3664684593677521,
-0.9438198208808899,
-0.3106822967529297,
-0.7801251411437988,
-0.06760001927614212,... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
adamjweintraut/eli5_lfqa_top_slice | adamjweintraut | 2023-11-28T06:27:49Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T06:27:49Z | 2023-11-28T06:27:01.000Z | 2023-11-28T06:27:01 | ---
dataset_info:
features:
- name: index
dtype: int64
- name: q_id
dtype: string
- name: question
dtype: string
- name: best_answer
dtype: string
- name: all_answers
sequence: string
- name: num_answers
dtype: int64
- name: top_answers
sequence: string
- name: num_top_answers
dtype: int64
- name: context
dtype: string
- name: orig
dtype: string
- name: target
dtype: string
splits:
- name: train
num_bytes: 304163516
num_examples: 20000
- name: test
num_bytes: 38395443
num_examples: 2500
- name: validation
num_bytes: 39481266
num_examples: 2500
download_size: 229456929
dataset_size: 382040225
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
- split: validation
path: data/validation-*
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
roupenminassian/vehicle-dataset-v2 | roupenminassian | 2023-11-28T06:36:20Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T06:36:20Z | 2023-11-28T06:35:21.000Z | 2023-11-28T06:35:21 | ---
dataset_info:
features:
- name: image
dtype: image
- name: image_id
dtype: int64
- name: width
dtype: int64
- name: height
dtype: int64
- name: objects
struct:
- name: id
sequence: int64
- name: area
sequence: float64
- name: bbox
sequence:
sequence: float64
- name: category
sequence: int64
splits:
- name: train
num_bytes: 120781140.624
num_examples: 1128
download_size: 122076069
dataset_size: 120781140.624
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "vehicle-dataset-v2"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.4901463985443115,
-0.06624151021242142,
0.33398231863975525,
0.23428070545196533,
-0.291504442691803,
-0.04324404522776604,
0.4832035005092621,
-0.31943920254707336,
0.485487699508667,
0.3560130298137665,
-0.9543925523757935,
-0.4684160351753235,
-0.45342713594436646,
-0.630785584449768... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
sam-mosaic/iv4-chatml-8k | sam-mosaic | 2023-11-28T06:38:55Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T06:38:55Z | 2023-11-28T06:37:50.000Z | 2023-11-28T06:37:50 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
dataset_info:
features:
- name: source
dtype: string
- name: prompt
dtype: string
- name: response
dtype: string
splits:
- name: train
num_bytes: 2207667895.4360986
num_examples: 363958
- name: test
num_bytes: 330419382.2206894
num_examples: 54042
download_size: 618017532
dataset_size: 2538087277.656788
---
# Dataset Card for "iv4-chatml-8k"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.5653365254402161,
-0.014067578129470348,
0.10285734385251999,
0.33141762018203735,
-0.13649065792560577,
0.06918392330408096,
0.2401008903980255,
-0.2693118155002594,
0.6610768437385559,
0.6111338138580322,
-0.6458150744438171,
-0.7539340853691101,
-0.6325390338897705,
0.062446918338537... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
sam-mosaic/iv4-chatml-4k | sam-mosaic | 2023-11-28T07:03:55Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T07:03:55Z | 2023-11-28T06:39:31.000Z | 2023-11-28T06:39:31 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
dataset_info:
features:
- name: source
dtype: string
- name: prompt
dtype: string
- name: response
dtype: string
splits:
- name: train
num_bytes: 2099655591.2305844
num_examples: 346151
- name: test
num_bytes: 315348071.4406665
num_examples: 51577
download_size: 295209643
dataset_size: 2415003662.671251
---
# Dataset Card for "iv4-chatml-4k"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.5992625951766968,
-0.03237571567296982,
0.10880687087774277,
0.3361487090587616,
-0.13269314169883728,
0.11595027893781662,
0.2719076871871948,
-0.2459101527929306,
0.6828324198722839,
0.585302472114563,
-0.6934072375297546,
-0.8305697441101074,
-0.5286479592323303,
0.07195425033569336,... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
sam-mosaic/iv4-chatml-16k | sam-mosaic | 2023-11-28T06:43:19Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T06:43:19Z | 2023-11-28T06:41:35.000Z | 2023-11-28T06:41:35 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
dataset_info:
features:
- name: source
dtype: string
- name: prompt
dtype: string
- name: response
dtype: string
splits:
- name: train
num_bytes: 2334144256.136339
num_examples: 384809
- name: test
num_bytes: 349214193.311071
num_examples: 57116
download_size: 1227729872
dataset_size: 2683358449.44741
---
# Dataset Card for "iv4-chatml-16k"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.5664142966270447,
-0.09199265390634537,
0.09351664781570435,
0.36523374915122986,
-0.19586144387722015,
0.12864990532398224,
0.14798453450202942,
-0.3177540600299835,
0.6472192406654358,
0.5605891346931458,
-0.7955158352851868,
-0.8392601609230042,
-0.5458130836486816,
0.007037137169390... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
X2FD/LVIS-Instruct4V-LLaVA-Instruct-mix880k | X2FD | 2023-11-28T06:53:07Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T06:53:07Z | 2023-11-28T06:43:55.000Z | 2023-11-28T06:43:55 | # LVIS-Instruct4V-LLaVA-Instruct-mix880k
This is a mixture of our LVIS-Instruct4V dataset with [LLaVA-Instruct](https://huggingface.co/datasets/liuhaotian/LLaVA-Instruct-150K/blob/main/llava_instruct_150k.json) (150k) , and the academic task related data, including ShareGPT, VQAv2, GQA, OKVQA, OCRVQA, AOKVQA, TextCaps, RefCOCO, and VG.
| [
-0.14464929699897766,
-0.3340277671813965,
0.49752992391586304,
0.052250079810619354,
-0.08016818761825562,
0.536203920841217,
0.22198894619941711,
-0.3077966272830963,
0.897317111492157,
0.6769272685050964,
-0.7392610311508179,
-0.3554709553718567,
-0.521245539188385,
0.021677514538168907... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Xiami2000/Trainingforaxolotl | Xiami2000 | 2023-11-28T15:43:28Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T15:43:28Z | 2023-11-28T06:48:21.000Z | 2023-11-28T06:48:21 | 整理好的一些H小说jsonl文件 | [
-0.35670292377471924,
-1.277295708656311,
0.20812055468559265,
0.6313344240188599,
-0.4479462504386902,
0.47784096002578735,
0.1203249841928482,
0.0710371881723404,
0.5905275344848633,
0.7772175073623657,
-0.24767062067985535,
-0.6989639401435852,
-0.6783783435821533,
0.26135575771331787,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
roupenminassian/vehicle-dataset-v3 | roupenminassian | 2023-11-28T06:52:32Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T06:52:32Z | 2023-11-28T06:51:28.000Z | 2023-11-28T06:51:28 | ---
dataset_info:
features:
- name: image
dtype: image
- name: image_id
dtype: int64
- name: width
dtype: int64
- name: height
dtype: int64
- name: objects
struct:
- name: id
sequence: int64
- name: area
sequence: float64
- name: bbox
sequence:
sequence: float64
- name: category
sequence: int64
splits:
- name: train
num_bytes: 120781140.624
num_examples: 1128
download_size: 122076069
dataset_size: 120781140.624
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "vehicle-dataset-v3"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.6347731351852417,
-0.003270343877375126,
0.5101358294487,
0.29632043838500977,
-0.20366209745407104,
-0.05272665247321129,
0.5476347804069519,
-0.33572477102279663,
0.4746214747428894,
0.42548516392707825,
-0.963828444480896,
-0.6250520944595337,
-0.33999642729759216,
-0.412517189979553... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
swulling/gsm8k_chinese | swulling | 2023-11-28T08:48:01Z | 0 | 1 | null | [
"task_categories:text2text-generation",
"size_categories:1K<n<10K",
"source_datasets:gsm8k",
"language:zh",
"license:mit",
"math-word-problems",
"region:us"
] | 2023-11-28T08:48:01Z | 2023-11-28T06:53:43.000Z | 2023-11-28T06:53:43 | ---
language:
- zh
license: mit
size_categories:
- 1K<n<10K
source_datasets:
- gsm8k
task_categories:
- text2text-generation
dataset_info:
features:
- name: question
dtype: string
- name: answer
dtype: string
- name: question_zh-cn
dtype: string
- name: answer_only
dtype: int64
splits:
- name: test
num_bytes: 1020788
num_examples: 1319
- name: train
num_bytes: 5664657
num_examples: 7473
download_size: 3988161
dataset_size: 6685445
configs:
- config_name: default
data_files:
- split: test
path: data/test-*
- split: train
path: data/train-*
tags:
- math-word-problems
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
diffusers/pokemon-gpt4-captions | diffusers | 2023-11-28T07:07:09Z | 0 | 2 | null | [
"task_categories:text-to-image",
"size_categories:1K<n<10K",
"language:en",
"license:other",
"region:us"
] | 2023-11-28T07:07:09Z | 2023-11-28T06:54:16.000Z | 2023-11-28T06:54:16 | ---
dataset_info:
features:
- name: image
dtype: image
- name: text
dtype: string
splits:
- name: train
num_bytes: 56664550
num_examples: 833
download_size: 51051224
dataset_size: 56664550
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
license: other
task_categories:
- text-to-image
language:
- en
pretty_name: 'Pokemons with captions generated using GPT-4. '
size_categories:
- 1K<n<10K
---
# Dataset Card for "pokemon-gpt4-captions"
This dataset is just [lambdalabs/pokemon-blip-captions](https://huggingface.co/datasets/lambdalabs/pokemon-blip-captions) but the captions come from GPT-4 (Turbo).
Code used to generate the captions:
```python
import base64
from io import BytesIO
import requests
from PIL import Image
def encode_image(image):
buffered = BytesIO()
image.save(buffered, format="JPEG")
img_str = base64.b64encode(buffered.getvalue())
return img_str.decode("utf-8")
def create_payload(image_string):
payload = {
"model": "gpt-4-vision-preview",
"messages": [
{
"role": "user",
"content": [
{
"type": "text",
"text": "Provide caption for the image in one sentence. Be detailed but precise.",
},
{
"type": "image_url",
"image_url": {"url": f"data:image/jpeg;base64,{image_string}"},
},
],
}
],
"max_tokens": 100,
}
return payload
def get_response(image_string):
payload = create_payload(image_string)
response = requests.post(
"https://api.openai.com/v1/chat/completions", headers=headers, json=payload
)
return response.json()
image = Image.open("path_to_you_image").convert("RGB")
image_str = encode_image(image)
response = get_response(image_str)
```
Generating captions for 833 images from the [lambdalabs/pokemon-blip-captions](https://huggingface.co/datasets/lambdalabs/pokemon-blip-captions) dataset costed about $5.
You can use this dataset for non-commercial applications. | [
-0.4003518223762512,
-0.1284639984369278,
0.09478273242712021,
0.3539121747016907,
-0.6607056260108948,
-0.026128623634576797,
-0.17441359162330627,
-0.2582589089870453,
0.37918752431869507,
0.5847703218460083,
-0.6495311856269836,
-0.35450753569602966,
-0.3569546341896057,
0.3742143213748... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Imran1/newdatasetdogb | Imran1 | 2023-11-28T06:58:11Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T06:58:11Z | 2023-11-28T06:58:04.000Z | 2023-11-28T06:58:04 | ---
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': Australian_shepherd
'1': Chihuahua
'2': French_bulldog
splits:
- name: train
num_bytes: 79657324.7834681
num_examples: 3363
download_size: 76864604
dataset_size: 79657324.7834681
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
BroJack/alpaca_lora_training_dataset | BroJack | 2023-11-28T07:00:49Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T07:00:49Z | 2023-11-28T06:59:47.000Z | 2023-11-28T06:59:47 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622263669967651,
0.43461522459983826,
-0.52829909324646,
0.7012971639633179,
0.7915719747543335,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104475975036621,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
BangumiBase/deathparade | BangumiBase | 2023-11-28T08:41:17Z | 0 | 0 | null | [
"size_categories:1K<n<10K",
"license:mit",
"art",
"region:us"
] | 2023-11-28T08:41:17Z | 2023-11-28T07:01:59.000Z | 2023-11-28T07:01:59 | ---
license: mit
tags:
- art
size_categories:
- 1K<n<10K
---
# Bangumi Image Base of Death Parade
This is the image base of bangumi Death Parade, we detected 20 characters, 1332 images in total. The full dataset is [here](all.zip).
**Please note that these image bases are not guaranteed to be 100% cleaned, they may be noisy actual.** If you intend to manually train models using this dataset, we recommend performing necessary preprocessing on the downloaded dataset to eliminate potential noisy samples (approximately 1% probability).
Here is the characters' preview:
| # | Images | Download | Preview 1 | Preview 2 | Preview 3 | Preview 4 | Preview 5 | Preview 6 | Preview 7 | Preview 8 |
|:------|---------:|:---------------------------|:-------------------------------|:-------------------------------|:-------------------------------|:-------------------------------|:-------------------------------|:-------------------------------|:-------------------------------|:-------------------------------|
| 0 | 186 | [Download](0/dataset.zip) |  |  |  |  |  |  |  |  |
| 1 | 28 | [Download](1/dataset.zip) |  |  |  |  |  |  |  |  |
| 2 | 57 | [Download](2/dataset.zip) |  |  |  |  |  |  |  |  |
| 3 | 45 | [Download](3/dataset.zip) |  |  |  |  |  |  |  |  |
| 4 | 59 | [Download](4/dataset.zip) |  |  |  |  |  |  |  |  |
| 5 | 70 | [Download](5/dataset.zip) |  |  |  |  |  |  |  |  |
| 6 | 31 | [Download](6/dataset.zip) |  |  |  |  |  |  |  |  |
| 7 | 72 | [Download](7/dataset.zip) |  |  |  |  |  |  |  |  |
| 8 | 117 | [Download](8/dataset.zip) |  |  |  |  |  |  |  |  |
| 9 | 46 | [Download](9/dataset.zip) |  |  |  |  |  |  |  |  |
| 10 | 40 | [Download](10/dataset.zip) |  |  |  |  |  |  |  |  |
| 11 | 63 | [Download](11/dataset.zip) |  |  |  |  |  |  |  |  |
| 12 | 15 | [Download](12/dataset.zip) |  |  |  |  |  |  |  |  |
| 13 | 22 | [Download](13/dataset.zip) |  |  |  |  |  |  |  |  |
| 14 | 214 | [Download](14/dataset.zip) |  |  |  |  |  |  |  |  |
| 15 | 60 | [Download](15/dataset.zip) |  |  |  |  |  |  |  |  |
| 16 | 49 | [Download](16/dataset.zip) |  |  |  |  |  |  |  |  |
| 17 | 13 | [Download](17/dataset.zip) |  |  |  |  |  |  |  |  |
| 18 | 47 | [Download](18/dataset.zip) |  |  |  |  |  |  |  |  |
| noise | 98 | [Download](-1/dataset.zip) |  |  |  |  |  |  |  |  |
| [
-0.6984877586364746,
-0.1498137265443802,
0.14263057708740234,
0.22603841125965118,
-0.28747469186782837,
-0.017097553238272667,
0.0008750214474275708,
-0.3377918303012848,
0.6570323705673218,
0.5274203419685364,
-0.9289644956588745,
-0.8810428977012634,
-0.6655895113945007,
0.453676223754... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
BangumiBase/ariatheanimation | BangumiBase | 2023-11-28T10:55:37Z | 0 | 0 | null | [
"size_categories:1K<n<10K",
"license:mit",
"art",
"region:us"
] | 2023-11-28T10:55:37Z | 2023-11-28T07:03:51.000Z | 2023-11-28T07:03:51 | ---
license: mit
tags:
- art
size_categories:
- 1K<n<10K
---
# Bangumi Image Base of Aria The Animation
This is the image base of bangumi Aria The Animation, we detected 50 characters, 5059 images in total. The full dataset is [here](all.zip).
**Please note that these image bases are not guaranteed to be 100% cleaned, they may be noisy actual.** If you intend to manually train models using this dataset, we recommend performing necessary preprocessing on the downloaded dataset to eliminate potential noisy samples (approximately 1% probability).
Here is the characters' preview:
| # | Images | Download | Preview 1 | Preview 2 | Preview 3 | Preview 4 | Preview 5 | Preview 6 | Preview 7 | Preview 8 |
|:------|---------:|:---------------------------|:-------------------------------|:-------------------------------|:-------------------------------|:-------------------------------|:-------------------------------|:-------------------------------|:-------------------------------|:-------------------------------|
| 0 | 1592 | [Download](0/dataset.zip) |  |  |  |  |  |  |  |  |
| 1 | 18 | [Download](1/dataset.zip) |  |  |  |  |  |  |  |  |
| 2 | 29 | [Download](2/dataset.zip) |  |  |  |  |  |  |  |  |
| 3 | 25 | [Download](3/dataset.zip) |  |  |  |  |  |  |  |  |
| 4 | 20 | [Download](4/dataset.zip) |  |  |  |  |  |  |  |  |
| 5 | 22 | [Download](5/dataset.zip) |  |  |  |  |  |  |  |  |
| 6 | 30 | [Download](6/dataset.zip) |  |  |  |  |  |  |  |  |
| 7 | 19 | [Download](7/dataset.zip) |  |  |  |  |  |  |  |  |
| 8 | 42 | [Download](8/dataset.zip) |  |  |  |  |  |  |  |  |
| 9 | 58 | [Download](9/dataset.zip) |  |  |  |  |  |  |  |  |
| 10 | 58 | [Download](10/dataset.zip) |  |  |  |  |  |  |  |  |
| 11 | 27 | [Download](11/dataset.zip) |  |  |  |  |  |  |  |  |
| 12 | 22 | [Download](12/dataset.zip) |  |  |  |  |  |  |  |  |
| 13 | 48 | [Download](13/dataset.zip) |  |  |  |  |  |  |  |  |
| 14 | 106 | [Download](14/dataset.zip) |  |  |  |  |  |  |  |  |
| 15 | 15 | [Download](15/dataset.zip) |  |  |  |  |  |  |  |  |
| 16 | 18 | [Download](16/dataset.zip) |  |  |  |  |  |  |  |  |
| 17 | 87 | [Download](17/dataset.zip) |  |  |  |  |  |  |  |  |
| 18 | 25 | [Download](18/dataset.zip) |  |  |  |  |  |  |  |  |
| 19 | 26 | [Download](19/dataset.zip) |  |  |  |  |  |  |  |  |
| 20 | 13 | [Download](20/dataset.zip) |  |  |  |  |  |  |  |  |
| 21 | 41 | [Download](21/dataset.zip) |  |  |  |  |  |  |  |  |
| 22 | 183 | [Download](22/dataset.zip) |  |  |  |  |  |  |  |  |
| 23 | 43 | [Download](23/dataset.zip) |  |  |  |  |  |  |  |  |
| 24 | 387 | [Download](24/dataset.zip) |  |  |  |  |  |  |  |  |
| 25 | 336 | [Download](25/dataset.zip) |  |  |  |  |  |  |  |  |
| 26 | 30 | [Download](26/dataset.zip) |  |  |  |  |  |  |  |  |
| 27 | 12 | [Download](27/dataset.zip) |  |  |  |  |  |  |  |  |
| 28 | 34 | [Download](28/dataset.zip) |  |  |  |  |  |  |  |  |
| 29 | 18 | [Download](29/dataset.zip) |  |  |  |  |  |  |  |  |
| 30 | 106 | [Download](30/dataset.zip) |  |  |  |  |  |  |  |  |
| 31 | 26 | [Download](31/dataset.zip) |  |  |  |  |  |  |  |  |
| 32 | 7 | [Download](32/dataset.zip) |  |  |  |  |  |  |  | N/A |
| 33 | 19 | [Download](33/dataset.zip) |  |  |  |  |  |  |  |  |
| 34 | 488 | [Download](34/dataset.zip) |  |  |  |  |  |  |  |  |
| 35 | 87 | [Download](35/dataset.zip) |  |  |  |  |  |  |  |  |
| 36 | 19 | [Download](36/dataset.zip) |  |  |  |  |  |  |  |  |
| 37 | 395 | [Download](37/dataset.zip) |  |  |  |  |  |  |  |  |
| 38 | 46 | [Download](38/dataset.zip) |  |  |  |  |  |  |  |  |
| 39 | 15 | [Download](39/dataset.zip) |  |  |  |  |  |  |  |  |
| 40 | 17 | [Download](40/dataset.zip) |  |  |  |  |  |  |  |  |
| 41 | 24 | [Download](41/dataset.zip) |  |  |  |  |  |  |  |  |
| 42 | 7 | [Download](42/dataset.zip) |  |  |  |  |  |  |  | N/A |
| 43 | 13 | [Download](43/dataset.zip) |  |  |  |  |  |  |  |  |
| 44 | 8 | [Download](44/dataset.zip) |  |  |  |  |  |  |  |  |
| 45 | 133 | [Download](45/dataset.zip) |  |  |  |  |  |  |  |  |
| 46 | 6 | [Download](46/dataset.zip) |  |  |  |  |  |  | N/A | N/A |
| 47 | 15 | [Download](47/dataset.zip) |  |  |  |  |  |  |  |  |
| 48 | 15 | [Download](48/dataset.zip) |  |  |  |  |  |  |  |  |
| noise | 229 | [Download](-1/dataset.zip) |  |  |  |  |  |  |  |  |
| [
-0.644494891166687,
-0.1464424878358841,
0.09932372719049454,
0.2570963501930237,
-0.22938211262226105,
-0.06991071999073029,
-0.030906789004802704,
-0.36807578802108765,
0.6714171767234802,
0.5187075138092041,
-0.9619875550270081,
-0.8683664202690125,
-0.7113472819328308,
0.50988829135894... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
tarudesu/ViOCD | tarudesu | 2023-11-28T07:19:05Z | 0 | 0 | null | [
"task_categories:text-classification",
"size_categories:1K<n<10K",
"language:vi",
"code",
"arxiv:2103.10069",
"arxiv:2104.11969",
"region:us"
] | 2023-11-28T07:19:05Z | 2023-11-28T07:13:26.000Z | 2023-11-28T07:13:26 | ---
task_categories:
- text-classification
language:
- vi
tags:
- code
pretty_name: Vietnamese Open-Domain Complaint Detection in E-commerce Websites
size_categories:
- 1K<n<10K
---
# Vietnamese Open-Domain Complaint Detection in E-commerce Websites
This is the official repository for the UIT-ViCTSD dataset from the paper [Vietnamese Open-Domain Complaint Detection in E-commerce Websites](https://arxiv.org/pdf/2103.10069.pdf), which was accepted at the [SoMeT 2021](https://dblp.org/db/conf/somet/somet2021.html).
# Citation Information
The provided dataset is only used for research purposes!
```
@misc{nguyen2021vietnamese,
title={Vietnamese Complaint Detection on E-Commerce Websites},
author={Nhung Thi-Hong Nguyen and Phuong Phan-Dieu Ha and Luan Thanh Nguyen and Kiet Van Nguyen and Ngan Luu-Thuy Nguyen},
year={2021},
eprint={2104.11969},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
## Abstract
Customer product reviews play a role in improving the quality of products and services for business organizations or their brands. Complaining is an attitude that expresses dissatisfaction with an event or a product not meeting customer expectations. In this paper, we build a Open-domain Complaint Detection dataset (UIT-ViOCD), including 5,485 human-annotated reviews on four categories about product reviews on e-commerce sites. After the data collection phase, we proceed to the annotation task and achieve the inter-annotator agreement Am of 87%. Then, we present an extensive methodology for the research purposes and achieve 92.16% by F1-score for identifying complaints. With the results, in the future, we aim to build a system for open-domain complaint detection in E-commerce websites.
## Dataset
The ViOCD dataset is consist of 5,485 reviews on four categories about product reviews on e-commerce sites.
The dataset is divided into three parts as below:
1. Train set: 4.39K reviews
2. Valid set: 548 reviews
3. Test set: 549 reviews
## Contact
Please feel free to contact us by email luannt@uit.edu.vn if you have any further information! | [
-0.7259083986282349,
-0.8485763669013977,
-0.03446989133954048,
0.36604228615760803,
-0.5101052522659302,
-0.24911127984523773,
-0.26747700572013855,
-0.7961223125457764,
0.29341551661491394,
0.504047155380249,
-0.4519159495830536,
-0.8370954990386963,
-0.08523007482290268,
0.2250676006078... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
nateraw/samsum-llamafied | nateraw | 2023-11-28T07:16:01Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T07:16:01Z | 2023-11-28T07:15:43.000Z | 2023-11-28T07:15:43 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622264862060547,
0.43461528420448303,
-0.52829909324646,
0.7012971639633179,
0.7915720343589783,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104477167129517,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
tyzhu/ds1_try_lora_merge | tyzhu | 2023-11-28T07:49:05Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T07:49:05Z | 2023-11-28T07:20:51.000Z | 2023-11-28T07:20:51 | ---
dataset_info:
features:
- name: inputs
dtype: string
- name: targets
dtype: string
splits:
- name: train
num_bytes: 1044.247619047619
num_examples: 10
- name: validation
num_bytes: 1044.247619047619
num_examples: 10
download_size: 4678
dataset_size: 2088.495238095238
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
---
# Dataset Card for "ds1_try_lora_merge"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.751491367816925,
-0.3824859857559204,
0.21505171060562134,
-0.04539085924625397,
-0.3613230586051941,
-0.16313676536083221,
0.4944823384284973,
-0.15499788522720337,
1.0014525651931763,
0.6852034330368042,
-1.0823452472686768,
-0.6837285161018372,
-0.7597208023071289,
-0.221937239170074... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
tyzhu/ds2_try_lora_merge | tyzhu | 2023-11-28T07:49:26Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T07:49:26Z | 2023-11-28T07:21:16.000Z | 2023-11-28T07:21:16 | ---
dataset_info:
features:
- name: inputs
dtype: string
- name: targets
dtype: string
splits:
- name: train
num_bytes: 1044.247619047619
num_examples: 10
- name: validation
num_bytes: 1044.247619047619
num_examples: 10
download_size: 4650
dataset_size: 2088.495238095238
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
---
# Dataset Card for "ds2_try_lora_merge"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.5803247094154358,
-0.31586953997612,
0.2512780427932739,
-0.06960684061050415,
-0.3319776654243469,
-0.10562603175640106,
0.4488089382648468,
-0.2702353596687317,
0.7869184017181396,
0.6512116193771362,
-0.9134699702262878,
-0.5213780403137207,
-0.772845447063446,
-0.29247549176216125,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
tyzhu/ds_combined_try_lora_merge | tyzhu | 2023-11-28T07:50:02Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T07:50:02Z | 2023-11-28T07:21:23.000Z | 2023-11-28T07:21:23 | ---
dataset_info:
features:
- name: inputs
dtype: string
- name: targets
dtype: string
splits:
- name: train
num_bytes: 2088.495238095238
num_examples: 20
- name: validation
num_bytes: 2088.495238095238
num_examples: 20
download_size: 5988
dataset_size: 4176.990476190476
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
---
# Dataset Card for "ds_combined_try_lora_merge"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.7256976962089539,
-0.37294045090675354,
0.17542698979377747,
-0.022249067202210426,
-0.3738609552383423,
-0.004309919197112322,
0.3398993909358978,
-0.22427567839622498,
0.982515811920166,
0.6695346236228943,
-0.9489110112190247,
-0.6489337086677551,
-0.6420139074325562,
-0.225451007485... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
cyy0/BMTL | cyy0 | 2023-11-28T07:22:23Z | 0 | 0 | null | [
"license:agpl-3.0",
"region:us"
] | 2023-11-28T07:22:23Z | 2023-11-28T07:22:22.000Z | 2023-11-28T07:22:22 | ---
license: agpl-3.0
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Linata/test_dataset | Linata | 2023-11-28T07:33:14Z | 0 | 0 | null | [
"license:mit",
"region:us"
] | 2023-11-28T07:33:14Z | 2023-11-28T07:31:53.000Z | 2023-11-28T07:31:53 | ---
license: mit
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
wesley7137/MMath14k | wesley7137 | 2023-11-28T07:33:25Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T07:33:25Z | 2023-11-28T07:33:11.000Z | 2023-11-28T07:33:11 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Stein-Fun/ddpm-butterflies-128 | Stein-Fun | 2023-11-28T07:33:53Z | 0 | 0 | null | [
"license:mit",
"region:us"
] | 2023-11-28T07:33:53Z | 2023-11-28T07:33:51.000Z | 2023-11-28T07:33:51 | ---
license: mit
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
edwinpalegre/trashnet_enhanced | edwinpalegre | 2023-11-28T17:59:51Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T17:59:51Z | 2023-11-28T07:34:52.000Z | 2023-11-28T07:34:52 | ---
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': biodegradable
'1': cardboard
'2': glass
'3': metal
'4': paper
'5': plastic
'6': trash
splits:
- name: train
num_bytes: 505205957.636
num_examples: 19892
download_size: 3977396925
dataset_size: 505205957.636
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
siddrao11/test | siddrao11 | 2023-11-28T08:06:11Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T08:06:11Z | 2023-11-28T07:38:42.000Z | 2023-11-28T07:38:42 | ---
# For reference on dataset card metadata, see the spec: https://github.com/huggingface/hub-docs/blob/main/datasetcard.md?plain=1
# Doc / guide: https://huggingface.co/docs/hub/datasets-cards
{}
---
# Dataset Card for Dataset Name
<!-- Provide a quick summary of the dataset. -->
This dataset card aims to be a base template for new datasets. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/datasetcard_template.md?plain=1).
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] | [
-0.5322356224060059,
-0.5534716844558716,
0.1290130317211151,
0.23470577597618103,
-0.39626216888427734,
-0.11762470006942749,
-0.03545305132865906,
-0.6389272212982178,
0.5699822306632996,
0.7838326692581177,
-0.7834625840187073,
-0.9173274040222168,
-0.55633145570755,
0.13078093528747559... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
desarrolloasesoreslocales/MistralAI | desarrolloasesoreslocales | 2023-11-28T08:00:28Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T08:00:28Z | 2023-11-28T07:43:14.000Z | 2023-11-28T07:43:14 | ---
configs:
- config_name: default
data_files:
- split: train
path: data.csv
---
# Dataset Card for Dataset Name
<!-- Provide a quick summary of the dataset. -->
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] | [
-0.5036625862121582,
-0.5130205154418945,
0.18480272591114044,
0.20869815349578857,
-0.3474426865577698,
-0.05577763170003891,
-0.022632518783211708,
-0.6274707913398743,
0.4583321809768677,
0.810380756855011,
-0.7633895874023438,
-0.9683905243873596,
-0.5347057580947876,
0.125262394547462... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Mousaicv/gpt4_reward_train | Mousaicv | 2023-11-28T07:50:09Z | 0 | 0 | null | [
"license:apache-2.0",
"region:us"
] | 2023-11-28T07:50:09Z | 2023-11-28T07:48:14.000Z | 2023-11-28T07:48:14 | ---
license: apache-2.0
---
| [
-0.12853369116783142,
-0.18616779148578644,
0.6529126167297363,
0.49436280131340027,
-0.193193256855011,
0.2360745668411255,
0.36071979999542236,
0.05056314915418625,
0.5793651342391968,
0.740013837814331,
-0.6508103013038635,
-0.23783960938453674,
-0.7102248668670654,
-0.04782580211758613... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Imran1/dogbalance_data | Imran1 | 2023-11-28T08:05:12Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T08:05:12Z | 2023-11-28T08:05:04.000Z | 2023-11-28T08:05:04 | ---
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': Australian_shepherd
'1': Chihuahua
'2': French_bulldog
splits:
- name: train
num_bytes: 18102241.0
num_examples: 735
download_size: 18093424
dataset_size: 18102241.0
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
| [
-0.12853369116783142,
-0.18616779148578644,
0.6529126167297363,
0.49436280131340027,
-0.193193256855011,
0.2360745668411255,
0.36071979999542236,
0.05056314915418625,
0.5793651342391968,
0.740013837814331,
-0.6508103013038635,
-0.23783960938453674,
-0.7102248668670654,
-0.04782580211758613... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Docfile/Toi | Docfile | 2023-11-28T08:13:55Z | 0 | 0 | null | [
"task_categories:translation",
"language:fr",
"license:apache-2.0",
"legal",
"region:us"
] | 2023-11-28T08:13:55Z | 2023-11-28T08:09:33.000Z | 2023-11-28T08:09:33 | ---
license: apache-2.0
task_categories:
- translation
language:
- fr
tags:
- legal
--- | [
-0.12853369116783142,
-0.18616779148578644,
0.6529126167297363,
0.49436280131340027,
-0.193193256855011,
0.2360745668411255,
0.36071979999542236,
0.05056314915418625,
0.5793651342391968,
0.740013837814331,
-0.6508103013038635,
-0.23783960938453674,
-0.7102248668670654,
-0.04782580211758613... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
CHEN0312/fyefu | CHEN0312 | 2023-11-28T08:17:45Z | 0 | 0 | null | [
"license:apache-2.0",
"region:us"
] | 2023-11-28T08:17:45Z | 2023-11-28T08:17:43.000Z | 2023-11-28T08:17:43 | ---
license: apache-2.0
---
| [
-0.12853369116783142,
-0.18616779148578644,
0.6529126167297363,
0.49436280131340027,
-0.193193256855011,
0.2360745668411255,
0.36071979999542236,
0.05056314915418625,
0.5793651342391968,
0.740013837814331,
-0.6508103013038635,
-0.23783960938453674,
-0.7102248668670654,
-0.04782580211758613... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
rmatousek/ics | rmatousek | 2023-11-28T08:21:02Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T08:21:02Z | 2023-11-28T08:17:55.000Z | 2023-11-28T08:17:55 | Entry not found | [
-0.3227647542953491,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965083122253,
0.7915717959403992,
0.07618629932403564,
0.7746022343635559,
0.2563222348690033,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
tourist800/test | tourist800 | 2023-11-28T08:28:17Z | 0 | 0 | null | [
"license:mit",
"region:us"
] | 2023-11-28T08:28:17Z | 2023-11-28T08:23:33.000Z | 2023-11-28T08:23:33 | ---
license: mit
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
SUSTech/MetaMathQA | SUSTech | 2023-11-28T11:28:02Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T11:28:02Z | 2023-11-28T08:23:43.000Z | 2023-11-28T08:23:43 | ---
dataset_info:
features:
- name: response
dtype: string
- name: type
dtype: string
- name: query
dtype: string
splits:
- name: train
num_bytes: 286570703
num_examples: 395000
download_size: 140903789
dataset_size: 286570703
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
zicsx/Hindi-YouTubeCC | zicsx | 2023-11-28T08:48:07Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T08:48:07Z | 2023-11-28T08:27:42.000Z | 2023-11-28T08:27:42 | ---
dataset_info:
features:
- name: text
dtype: string
splits:
- name: train
num_bytes: 6962453685
num_examples: 161248
download_size: 2554893278
dataset_size: 6962453685
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "Hindi-YouTubeCC"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.5777029991149902,
-0.3458380103111267,
-0.1282365769147873,
0.24914565682411194,
-0.233870267868042,
0.24621081352233887,
-0.05659063160419464,
0.18104737997055054,
0.8844714164733887,
0.12264534085988998,
-0.8953933119773865,
-0.6891379952430725,
-0.8029922246932983,
-0.191429853439331... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
FrankFacundo/NaturalQuestionsMultilang | FrankFacundo | 2023-11-28T09:34:05Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T09:34:05Z | 2023-11-28T08:30:30.000Z | 2023-11-28T08:30:30 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
sam1120/safety-utcustom-eval-v1.0 | sam1120 | 2023-11-28T08:33:31Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T08:33:31Z | 2023-11-28T08:31:52.000Z | 2023-11-28T08:31:52 | ---
dataset_info:
features:
- name: name
dtype: string
- name: pixel_values
dtype: image
- name: labels
dtype: image
splits:
- name: train
num_bytes: 139241854.0
num_examples: 50
download_size: 40367886
dataset_size: 139241854.0
---
# Dataset Card for "safety-utcustom-eval-50-v1.0"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.459087997674942,
-0.21636013686656952,
0.17202800512313843,
0.33283260464668274,
-0.2923942506313324,
0.11932279169559479,
0.20244646072387695,
-0.2162804901599884,
0.6088123321533203,
0.478954941034317,
-0.8197528123855591,
-0.9340550899505615,
-0.35364824533462524,
-0.2645405530929565... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
SUSTech/gsm8k-gpt35 | SUSTech | 2023-11-28T08:37:38Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T08:37:38Z | 2023-11-28T08:37:33.000Z | 2023-11-28T08:37:33 | ---
dataset_info:
features:
- name: question
dtype: string
- name: answer
dtype: string
splits:
- name: main
num_bytes: 4355508
num_examples: 6840
- name: overlap
num_bytes: 21003568
num_examples: 32825
download_size: 7092472
dataset_size: 25359076
configs:
- config_name: default
data_files:
- split: main
path: data/main-*
- split: overlap
path: data/overlap-*
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
euisuh15/python-piss-my-name-10 | euisuh15 | 2023-11-28T08:43:43Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T08:43:43Z | 2023-11-28T08:41:34.000Z | 2023-11-28T08:41:34 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
euisuh15/python-piss-my-name-30 | euisuh15 | 2023-11-28T08:44:13Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T08:44:13Z | 2023-11-28T08:42:34.000Z | 2023-11-28T08:42:34 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
euisuh15/python-piss-my-name-70 | euisuh15 | 2023-11-28T08:46:30Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T08:46:30Z | 2023-11-28T08:44:33.000Z | 2023-11-28T08:44:33 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Wil16/essa | Wil16 | 2023-11-28T12:39:38Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T12:39:38Z | 2023-11-28T08:54:39.000Z | 2023-11-28T08:54:39 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Wauplin/dummy_gated_dataset | Wauplin | 2023-11-28T09:02:59Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T09:02:59Z | 2023-11-28T09:02:59.000Z | 2023-11-28T09:02:59 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
A2H0H0R1/alpaca_data_gpt4_2 | A2H0H0R1 | 2023-11-28T09:07:19Z | 0 | 0 | null | [
"license:apache-2.0",
"region:us"
] | 2023-11-28T09:07:19Z | 2023-11-28T09:06:35.000Z | 2023-11-28T09:06:35 | ---
license: apache-2.0
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Andron00e/CIFAR10-custom | Andron00e | 2023-11-28T09:43:03Z | 0 | 0 | null | [
"task_categories:image-classification",
"size_categories:10K<n<100K",
"language:en",
"license:mit",
"region:us"
] | 2023-11-28T09:43:03Z | 2023-11-28T09:20:50.000Z | 2023-11-28T09:20:50 | ---
dataset_info:
features:
- name: image_file_path
dtype: string
- name: image
dtype: image
- name: labels
dtype: uint8
splits:
- name: train
num_bytes: 59153400
num_examples: 60000
download_size: 26957572
dataset_size: 59153400
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
license: mit
task_categories:
- image-classification
language:
- en
size_categories:
- 10K<n<100K
---
Example of usage:
```python
from datasets import load_dataset
dataset = load_dataset("Andron00e/CIFAR10-custom")
splitted_dataset = dataset["train"].train_test_split(test_size=0.2)
``` | [
-0.5458157062530518,
-0.32690078020095825,
-0.17638398706912994,
0.17882932722568512,
-0.22993560135364532,
-0.14087478816509247,
0.07836265116930008,
0.052019648253917694,
0.11030935496091843,
0.3083287179470062,
-0.3192961812019348,
0.062392931431531906,
-0.1426040530204773,
0.2852115333... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
edzhu/binance_eth_bnb_btc_usdt_marketdata | edzhu | 2023-11-28T09:27:50Z | 0 | 0 | null | [
"license:mit",
"region:us"
] | 2023-11-28T09:27:50Z | 2023-11-28T09:21:53.000Z | 2023-11-28T09:21:53 | ---
license: mit
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Fissk/SG_dataset | Fissk | 2023-11-28T09:36:30Z | 0 | 0 | null | [
"license:llama2",
"region:us"
] | 2023-11-28T09:36:30Z | 2023-11-28T09:30:17.000Z | 2023-11-28T09:30:17 | ---
license: llama2
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
BangumiBase/princesstutu | BangumiBase | 2023-11-28T11:12:28Z | 0 | 0 | null | [
"size_categories:1K<n<10K",
"license:mit",
"art",
"region:us"
] | 2023-11-28T11:12:28Z | 2023-11-28T09:30:22.000Z | 2023-11-28T09:30:22 | ---
license: mit
tags:
- art
size_categories:
- 1K<n<10K
---
# Bangumi Image Base of Princess Tutu
This is the image base of bangumi Princess Tutu, we detected 23 characters, 2179 images in total. The full dataset is [here](all.zip).
**Please note that these image bases are not guaranteed to be 100% cleaned, they may be noisy actual.** If you intend to manually train models using this dataset, we recommend performing necessary preprocessing on the downloaded dataset to eliminate potential noisy samples (approximately 1% probability).
Here is the characters' preview:
| # | Images | Download | Preview 1 | Preview 2 | Preview 3 | Preview 4 | Preview 5 | Preview 6 | Preview 7 | Preview 8 |
|:------|---------:|:---------------------------|:-------------------------------|:-------------------------------|:-------------------------------|:-------------------------------|:-------------------------------|:-------------------------------|:-------------------------------|:-------------------------------|
| 0 | 190 | [Download](0/dataset.zip) |  |  |  |  |  |  |  |  |
| 1 | 536 | [Download](1/dataset.zip) |  |  |  |  |  |  |  |  |
| 2 | 67 | [Download](2/dataset.zip) |  |  |  |  |  |  |  |  |
| 3 | 21 | [Download](3/dataset.zip) |  |  |  |  |  |  |  |  |
| 4 | 288 | [Download](4/dataset.zip) |  |  |  |  |  |  |  |  |
| 5 | 20 | [Download](5/dataset.zip) |  |  |  |  |  |  |  |  |
| 6 | 19 | [Download](6/dataset.zip) |  |  |  |  |  |  |  |  |
| 7 | 23 | [Download](7/dataset.zip) |  |  |  |  |  |  |  |  |
| 8 | 22 | [Download](8/dataset.zip) |  |  |  |  |  |  |  |  |
| 9 | 250 | [Download](9/dataset.zip) |  |  |  |  |  |  |  |  |
| 10 | 352 | [Download](10/dataset.zip) |  |  |  |  |  |  |  |  |
| 11 | 27 | [Download](11/dataset.zip) |  |  |  |  |  |  |  |  |
| 12 | 23 | [Download](12/dataset.zip) |  |  |  |  |  |  |  |  |
| 13 | 35 | [Download](13/dataset.zip) |  |  |  |  |  |  |  |  |
| 14 | 22 | [Download](14/dataset.zip) |  |  |  |  |  |  |  |  |
| 15 | 19 | [Download](15/dataset.zip) |  |  |  |  |  |  |  |  |
| 16 | 38 | [Download](16/dataset.zip) |  |  |  |  |  |  |  |  |
| 17 | 13 | [Download](17/dataset.zip) |  |  |  |  |  |  |  |  |
| 18 | 10 | [Download](18/dataset.zip) |  |  |  |  |  |  |  |  |
| 19 | 16 | [Download](19/dataset.zip) |  |  |  |  |  |  |  |  |
| 20 | 67 | [Download](20/dataset.zip) |  |  |  |  |  |  |  |  |
| 21 | 14 | [Download](21/dataset.zip) |  |  |  |  |  |  |  |  |
| noise | 107 | [Download](-1/dataset.zip) |  |  |  |  |  |  |  |  |
| [
-0.7046698331832886,
-0.1679610311985016,
0.0711611658334732,
0.21716158092021942,
-0.27490514516830444,
-0.070058673620224,
-0.03722704201936722,
-0.34448471665382385,
0.6250349283218384,
0.49311959743499756,
-0.9054293632507324,
-0.8042576909065247,
-0.6868444681167603,
0.539794445037841... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
NobodyExistsOnTheInternet/sharegptairoboros | NobodyExistsOnTheInternet | 2023-11-28T09:32:13Z | 0 | 0 | null | [
"license:mit",
"region:us"
] | 2023-11-28T09:32:13Z | 2023-11-28T09:31:18.000Z | 2023-11-28T09:31:18 | ---
license: mit
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
shenmixy/huggingface_token | shenmixy | 2023-11-28T10:10:03Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T10:10:03Z | 2023-11-28T09:33:00.000Z | 2023-11-28T09:33:00 | cp -r -f -n -s /root/sd_webui/cache/huggingface/huggingface_repo/* /root/sd_webui/sd_main_dir/log | [
-0.7627964615821838,
-0.19222764670848846,
0.12659841775894165,
0.6476641893386841,
-0.5734990835189819,
0.4627455472946167,
0.18533553183078766,
0.003375691594555974,
0.9853092432022095,
0.5556601881980896,
-1.1237943172454834,
-0.39997735619544983,
-0.7013154029846191,
0.5654587745666504... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
IliyanGochev/common_voice_13_0_bg_pseudo_labelled | IliyanGochev | 2023-11-28T10:27:00Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T10:27:00Z | 2023-11-28T09:42:38.000Z | 2023-11-28T09:42:38 | ---
dataset_info:
config_name: bg
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 16000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
- name: variant
dtype: string
- name: whisper_transcript
sequence: int64
splits:
- name: train
num_bytes: 92489982.56
num_examples: 3385
- name: validation
num_bytes: 79482559.912
num_examples: 2358
- name: test
num_bytes: 84919243.036
num_examples: 2463
download_size: 265321822
dataset_size: 256891785.50800002
configs:
- config_name: bg
data_files:
- split: train
path: bg/train-*
- split: validation
path: bg/validation-*
- split: test
path: bg/test-*
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
satpalsr/filter | satpalsr | 2023-11-28T09:52:31Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T09:52:31Z | 2023-11-28T09:51:04.000Z | 2023-11-28T09:51:04 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Andron00e/CIFAR100-custom | Andron00e | 2023-11-28T10:00:58Z | 0 | 0 | null | [
"task_categories:image-classification",
"size_categories:10K<n<100K",
"language:en",
"license:mit",
"region:us"
] | 2023-11-28T10:00:58Z | 2023-11-28T09:51:12.000Z | 2023-11-28T09:51:12 | ---
dataset_info:
features:
- name: image_file_path
dtype: string
- name: image
dtype: image
- name: labels
dtype: int64
splits:
- name: train
num_bytes: 59505360
num_examples: 60000
download_size: 27123594
dataset_size: 59505360
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
license: mit
task_categories:
- image-classification
language:
- en
size_categories:
- 10K<n<100K
---
Example of usage:
```python
from datasets import load_dataset
dataset = load_dataset("Andron00e/CIFAR100-custom")
splitted_dataset = dataset["train"].train_test_split(test_size=0.2)
``` | [
-0.5434361696243286,
-0.3428143858909607,
-0.18200275301933289,
0.15259936451911926,
-0.19185258448123932,
-0.18169787526130676,
0.0791841596364975,
0.11102788895368576,
0.09916254132986069,
0.3320888578891754,
-0.33155521750450134,
0.06812416017055511,
-0.12859667837619781,
0.242264166474... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
satpalsr/question | satpalsr | 2023-11-28T09:53:22Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T09:53:22Z | 2023-11-28T09:51:36.000Z | 2023-11-28T09:51:36 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Rami/sketch_to_hed | Rami | 2023-11-28T09:51:46Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T09:51:46Z | 2023-11-28T09:51:40.000Z | 2023-11-28T09:51:40 | ---
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': test
'1': train
'2': validation
splits:
- name: train
num_bytes: 400653.0
num_examples: 10
- name: validation
num_bytes: 363040.0
num_examples: 12
- name: test
num_bytes: 1224181.0
num_examples: 40
download_size: 1727421
dataset_size: 1987874.0
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
- split: test
path: data/test-*
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Rami/sketch_to_next_sketch | Rami | 2023-11-28T09:52:49Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T09:52:49Z | 2023-11-28T09:52:01.000Z | 2023-11-28T09:52:01 | ---
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': test
'1': train
'2': validation
splits:
- name: train
num_bytes: 12459154.07
num_examples: 1278
- name: validation
num_bytes: 11271129.782
num_examples: 1071
- name: test
num_bytes: 30001683.886
num_examples: 2978
download_size: 51938025
dataset_size: 53731967.738
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
- split: test
path: data/test-*
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
ksmemory/FTA | ksmemory | 2023-11-28T10:16:17Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T10:16:17Z | 2023-11-28T09:54:29.000Z | 2023-11-28T09:54:29 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622264862060547,
0.43461528420448303,
-0.52829909324646,
0.7012971639633179,
0.7915720343589783,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104477167129517,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Sefaria/english_library | Sefaria | 2023-11-28T10:21:12Z | 0 | 0 | null | [
"license:gpl-3.0",
"region:us"
] | 2023-11-28T10:21:12Z | 2023-11-28T09:58:25.000Z | 2023-11-28T09:58:25 | ---
license: gpl-3.0
---
# Description
Export of Sefaria's English library data. This data represents over version in the library marked as English.
## Schema
| Field | Description |
| --- | --- |
| text | The text of a single segment in the library. A segment is the smallest chunk of test, usually representing a paragraph. |
| metadata | Dictionary of metadata. See below for schema. |
### Metadata Schema
| Field | Description |
| --- | --- |
| url | URL to this segment in Sefaria |
| ref | Canonical Ref to this segment. Refs are a human readable ID that is unique independent of version. Different versions of a segment all share the same Ref. |
| versionTitle | Version title of the version this segment came from. |
| lang | two letter language code. |
| docCategory | Category for this segment. This corresponds to where the segment's book is located in Sefaria's table of contents. |
| dataQuality | Estimate of the quality of the text. This can be either "professional" or "user". |
| pagerank | Pagerank for this segment calculated using Sefaria's internal link graph. Higher values indicate the segment is more centrally cited by sources. |
| [
-0.5709785223007202,
-0.4688186049461365,
0.12514632940292358,
0.09684384614229202,
-0.5182107090950012,
0.06000272557139397,
0.3785896599292755,
-0.46849116683006287,
0.6707982420921326,
0.6797115802764893,
-0.7310574650764465,
-0.7523360848426819,
-0.1941734403371811,
0.24721075594425201... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Kuro0911/phishing_url_llama | Kuro0911 | 2023-11-28T09:59:28Z | 0 | 0 | null | [
"license:mit",
"region:us"
] | 2023-11-28T09:59:28Z | 2023-11-28T09:59:24.000Z | 2023-11-28T09:59:24 | ---
license: mit
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
nateraw/english-to-hinglish | nateraw | 2023-11-28T21:16:58Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T21:16:58Z | 2023-11-28T10:01:50.000Z | 2023-11-28T10:01:50 | ---
dataset_info:
features:
- name: en
dtype: string
- name: hi_ng
dtype: string
- name: source
dtype: int64
splits:
- name: train
num_bytes: 18814411
num_examples: 178701
- name: test
num_bytes: 1098000
num_examples: 10401
download_size: 11924718
dataset_size: 19912411
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
---
Fork of [findnitai/english-to-hinglish](https://huggingface.co/datasets/findnitai/english-to-hinglish) that splits the training set into train/test. | [
-0.6194483637809753,
-0.6102665066719055,
-0.18711170554161072,
0.41961196064949036,
-0.12454678118228912,
0.03989695757627487,
-0.025993788614869118,
-0.56533282995224,
1.0325169563293457,
0.6787667870521545,
-0.9221007227897644,
-0.10635977238416672,
-0.512734591960907,
0.051245339214801... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
reza-alipour/CelebA-HQ-Seg | reza-alipour | 2023-11-28T10:56:43Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T10:56:43Z | 2023-11-28T10:18:16.000Z | 2023-11-28T10:18:16 | ---
dataset_info:
features:
- name: id
dtype: string
- name: mm_id
dtype: string
- name: image
dtype: image
- name: mask
dtype: image
- name: landmark
dtype: image
- name: landmark_cropped
dtype: image
- name: captions
sequence: string
- name: captions_eng
sequence: string
- name: captions_pes
sequence: string
- name: captions_cmn
sequence: string
- name: captions_fra
sequence: string
- name: captions_deu
sequence: string
- name: captions_ita
sequence: string
- name: captions_spa
sequence: string
- name: captions_all
sequence: string
- name: mask_segformer
dtype: image
splits:
- name: train
num_bytes: 3885530945.625
num_examples: 28495
- name: test
num_bytes: 231063342.75
num_examples: 1498
download_size: 3829867987
dataset_size: 4116594288.375
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Sefaria/hebrew_library | Sefaria | 2023-11-28T10:47:51Z | 0 | 0 | null | [
"license:gpl-3.0",
"region:us"
] | 2023-11-28T10:47:51Z | 2023-11-28T10:23:03.000Z | 2023-11-28T10:23:03 | ---
license: gpl-3.0
---
# Description
Export of Sefaria's Hebrew library data. This data represents over version in the library marked as Hebrew.
## Schema
| Field | Description |
| --- | --- |
| text | The text of a single segment in the library. A segment is the smallest chunk of test, usually representing a paragraph. |
| metadata | Dictionary of metadata. See below for schema. |
### Metadata Schema
| Field | Description |
| --- | --- |
| url | URL to this segment in Sefaria |
| ref | Canonical Ref to this segment. Refs are a human readable ID that is unique independent of version. Different versions of a segment all share the same Ref. |
| versionTitle | Version title of the version this segment came from. |
| lang | two letter language code. |
| docCategory | Category for this segment. This corresponds to where the segment's book is located in Sefaria's table of contents. |
| dataQuality | Estimate of the quality of the text. This can be either "professional" or "user". |
| pagerank | Pagerank for this segment calculated using Sefaria's internal link graph. Higher values indicate the segment is more centrally cited by sources. | | [
-0.5023195743560791,
-0.5028815269470215,
0.01463149394840002,
0.09836933016777039,
-0.5430060625076294,
-0.001991128781810403,
0.35290706157684326,
-0.4012186527252197,
0.6028505563735962,
0.6482463479042053,
-0.7034521102905273,
-0.9032110571861267,
-0.25857630372047424,
0.06152109056711... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
amaye15/Stack-Overflow-Zero-Shot-Classification | amaye15 | 2023-11-28T10:38:54Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T10:38:54Z | 2023-11-28T10:29:35.000Z | 2023-11-28T10:29:35 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
dataset_info:
features:
- name: Date
dtype: string
- name: Title
dtype: string
- name: Tags
dtype: string
- name: Score
dtype: int64
splits:
- name: train
num_bytes: 66744845
num_examples: 553439
download_size: 25302295
dataset_size: 66744845
---
# Dataset Card for "Stack-Overflow-Zero-Shot-Classification"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.43892788887023926,
-0.36368483304977417,
0.18197393417358398,
-0.09502231329679489,
-0.031671009957790375,
0.24588212370872498,
0.44675904512405396,
-0.2251293808221817,
0.5963051915168762,
0.5119502544403076,
-0.5820396542549133,
-0.9227556586265564,
-0.6419802904129028,
-0.53062248229... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
nateraw/replicate-training-datasets | nateraw | 2023-11-28T10:31:24Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T10:31:24Z | 2023-11-28T10:30:52.000Z | 2023-11-28T10:30:52 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Shubbair/oxford_flowers | Shubbair | 2023-11-28T13:33:29Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T13:33:29Z | 2023-11-28T10:31:04.000Z | 2023-11-28T10:31:04 | ---
dataset_info:
features:
- name: image
dtype: image
splits:
- name: train
num_bytes: 3015452.0
num_examples: 102
download_size: 3016707
dataset_size: 3015452.0
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
NobodyExistsOnTheInternet/turbotoconvert | NobodyExistsOnTheInternet | 2023-11-28T10:38:02Z | 0 | 0 | null | [
"license:mit",
"region:us"
] | 2023-11-28T10:38:02Z | 2023-11-28T10:35:49.000Z | 2023-11-28T10:35:49 | ---
license: mit
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
roupenminassian/vehicle-dataset-v4 | roupenminassian | 2023-11-28T10:46:14Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T10:46:14Z | 2023-11-28T10:44:35.000Z | 2023-11-28T10:44:35 | ---
dataset_info:
features:
- name: image
dtype: image
- name: image_id
dtype: int64
- name: width
dtype: int64
- name: height
dtype: int64
- name: objects
struct:
- name: id
sequence: int64
- name: area
sequence: float64
- name: bbox
sequence:
sequence: float64
- name: category
sequence: int64
splits:
- name: train
num_bytes: 151700808.768
num_examples: 1364
download_size: 149189451
dataset_size: 151700808.768
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "vehicle-dataset-v4"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.6726629137992859,
0.08265996724367142,
0.43860694766044617,
0.2288210093975067,
-0.2468220293521881,
0.04945574700832367,
0.5175250172615051,
-0.273931086063385,
0.525201141834259,
0.372847318649292,
-1.0186015367507935,
-0.6318637728691101,
-0.3005826771259308,
-0.30637115240097046,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
imone/FLAN_NIv2_NoExplanation | imone | 2023-11-28T10:56:20Z | 0 | 0 | null | [
"license:mit",
"region:us"
] | 2023-11-28T10:56:20Z | 2023-11-28T10:48:30.000Z | 2023-11-28T10:48:30 | ---
license: mit
---
# FLAN NIv2 without explanation
This is a subset of [FLAN NIv2](https://huggingface.co/datasets/Open-Orca/FLAN). We removed all examples with explanations in the few-shot template, as the final answers also don't have explanations. | [
-0.40104830265045166,
-0.7325344085693359,
0.20814265310764313,
-0.014485925436019897,
-0.43644848465919495,
-0.27320507168769836,
0.3724297285079956,
-0.2556949555873871,
0.5550204515457153,
0.5986126065254211,
-1.2683199644088745,
-0.22574763000011444,
-0.33921146392822266,
-0.0789861902... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Sefaria/links | Sefaria | 2023-11-28T11:04:51Z | 0 | 0 | null | [
"license:gpl-3.0",
"region:us"
] | 2023-11-28T11:04:51Z | 2023-11-28T10:52:14.000Z | 2023-11-28T10:52:14 | ---
license: gpl-3.0
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Linaqruf/frieren-xl-lora-test | Linaqruf | 2023-11-28T11:04:01Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T11:04:01Z | 2023-11-28T10:57:47.000Z | 2023-11-28T10:57:47 | Entry not found | [
-0.32276472449302673,
-0.22568407654762268,
0.8622258901596069,
0.4346148371696472,
-0.5282984972000122,
0.7012965679168701,
0.7915717363357544,
0.07618629932403564,
0.7746022939682007,
0.2563222646713257,
-0.785281777381897,
-0.22573848068714142,
-0.9104482531547546,
0.5715669393539429,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
James332/tt3 | James332 | 2023-11-28T11:31:31Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T11:31:31Z | 2023-11-28T11:02:14.000Z | 2023-11-28T11:02:14 | ---
dataset_info:
features:
- name: image
dtype: image
- name: question_type
dtype: string
- name: confidence
dtype: int32
- name: answers
sequence: string
- name: answers_original
list:
- name: answer
dtype: string
- name: raw_answer
dtype: string
- name: answer_confidence
dtype: string
- name: answer_id
dtype: int64
- name: id_image
dtype: int64
- name: answer_type
dtype: string
- name: question_id
dtype: int64
- name: question
dtype: string
- name: id
dtype: int64
- name: clip_tags_ViT_L_14
sequence: string
- name: clip_tags_LAION_ViT_H_14_2B
sequence: string
- name: blip_caption_beam_5
dtype: string
- name: LLM_Description_gpt3_downstream_tasks_visual_genome_ViT_L_14
sequence: string
- name: LLM_Description_gpt3_downstream_tasks_visual_genome_LAION-ViT-H-14-2B
sequence: string
- name: DETA_detections_deta_swin_large_o365_coco_classes
list:
- name: attribute
dtype: string
- name: box
sequence: float32
- name: label
dtype: string
- name: location
dtype: string
- name: ratio
dtype: float32
- name: size
dtype: string
- name: tag
dtype: string
- name: DETA_detections_deta_swin_large_o365_coco_classes_caption_module_random
list:
- name: attribute
dtype: string
- name: box
sequence: float64
- name: captions_module
sequence: string
- name: captions_module_filter
sequence: string
- name: label
dtype: string
- name: location
dtype: string
- name: ratio
dtype: float64
- name: size
dtype: string
- name: tag
dtype: string
splits:
- name: train
num_bytes: 1686555802.0
num_examples: 9009
download_size: 1572400067
dataset_size: 1686555802.0
---
# Dataset Card for "OK-VQA_train"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.719218373298645,
-0.1172432228922844,
0.29995885491371155,
-0.07760744541883469,
-0.13567519187927246,
-0.2137058675289154,
0.4510095715522766,
-0.07949715107679367,
0.6381325125694275,
0.49382418394088745,
-0.838329553604126,
-0.46737974882125854,
-0.47145941853523254,
-0.4232290983200... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Shawt/liz | Shawt | 2023-11-28T11:26:54Z | 0 | 0 | null | [
"license:openrail",
"art",
"lizz",
"region:us"
] | 2023-11-28T11:26:54Z | 2023-11-28T11:12:58.000Z | 2023-11-28T11:12:58 | ---
license: openrail
tags:
- art
- lizz
---
# Dataset Card for Dataset Name
<!-- Provide a quick summary of the dataset. -->
This dataset card aims to be a base template for new datasets. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/datasetcard_template.md?plain=1).
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
- **Curated by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
[More Information Needed]
## Dataset Structure
<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->
[More Information Needed]
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
[More Information Needed]
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
[More Information Needed]
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
[More Information Needed]
### Annotations [optional]
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
[More Information Needed]
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
[More Information Needed]
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations.
## Citation [optional]
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Dataset Card Authors [optional]
[More Information Needed]
## Dataset Card Contact
[More Information Needed] | [
-0.5322356224060059,
-0.5534716844558716,
0.1290130317211151,
0.23470577597618103,
-0.39626216888427734,
-0.11762470006942749,
-0.03545305132865906,
-0.6389272212982178,
0.5699822306632996,
0.7838326692581177,
-0.7834625840187073,
-0.9173274040222168,
-0.55633145570755,
0.13078093528747559... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
bashmanxx/llama37train | bashmanxx | 2023-11-28T11:16:10Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T11:16:10Z | 2023-11-28T11:16:00.000Z | 2023-11-28T11:16:00 | ---
dataset_info:
features:
- name: text
struct:
- name: text
struct:
- name: text
dtype: string
splits:
- name: train
num_bytes: 495601
num_examples: 420
download_size: 44997
dataset_size: 495601
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Shubbair/oxford_flowers_102 | Shubbair | 2023-11-28T11:19:39Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T11:19:39Z | 2023-11-28T11:19:26.000Z | 2023-11-28T11:19:26 | ---
dataset_info:
features:
- name: image
dtype: image
splits:
- name: train
num_bytes: 3015452.0
num_examples: 102
download_size: 3016707
dataset_size: 3015452.0
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
erbacher/testdata | erbacher | 2023-11-28T11:25:56Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T11:25:56Z | 2023-11-28T11:25:25.000Z | 2023-11-28T11:25:25 | ---
dataset_info:
features:
- name: parameters
dtype: string
- name: tensor
sequence:
sequence:
sequence: float32
- name: input_ids
sequence: int32
- name: attention_mask
sequence: int8
splits:
- name: train
num_bytes: 22205500
num_examples: 100
download_size: 2695407
dataset_size: 22205500
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "testdata"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.590869665145874,
-0.30755072832107544,
0.19610433280467987,
0.09336929768323898,
-0.07715299725532532,
0.1159442588686943,
0.23835550248622894,
-0.08274070918560028,
0.7648308873176575,
0.32310783863067627,
-0.7559483647346497,
-0.7884511947631836,
-0.47454148530960083,
-0.2563454508781... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
alvarobartt/HelpSteer-AIF | alvarobartt | 2023-11-28T15:09:56Z | 0 | 0 | null | [
"size_categories:n<1K",
"language:en",
"license:cc-by-4.0",
"distilabel",
"helpsteer",
"gpt-4",
"aif",
"arxiv:2311.09528",
"region:us"
] | 2023-11-28T15:09:56Z | 2023-11-28T11:41:53.000Z | 2023-11-28T11:41:53 | ---
language:
- en
license: cc-by-4.0
size_categories:
- n<1K
pretty_name: HelpSteer with AIF
dataset_info:
features:
- name: prompt
dtype: string
- name: response
dtype: string
- name: model
dtype: string
- name: correctness
dtype: int64
- name: coherence
dtype: int64
- name: complexity
dtype: int64
- name: verbosity
dtype: int64
- name: helpfulness
dtype: int64
splits:
- name: train
num_bytes: 2832095
num_examples: 1000
download_size: 677100
dataset_size: 2832095
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
tags:
- distilabel
- helpsteer
- gpt-4
- aif
---
# HelpSteer: Helpfulness SteerLM Dataset
HelpSteer is an open-source Helpfulness Dataset (CC-BY-4.0) that supports aligning models to become more helpful, factually correct and coherent, while being adjustable in terms of the complexity and verbosity of its responses.
[HelpSteer: Multi-attribute Helpfulness Dataset for SteerLM](http://arxiv.org/abs/2311.09528)
## Disclaimer
This is only a subset created with `distilabel` to evaluate the first 100 rows using AI Feedback (AIF) coming from GPT-4, only created for experimenting / research purposes, please refer to [nvidia/HelpSteer](https://hf.co/nvidia/HelpSteer) if you want more information about the HelpSteer dataset.
## Dataset Description
HelpSteer contains 37120 samples, while this subset only contains the first 100, each containing a prompt, a response as well as five human-annotated attributes of the response, each ranging between 0 and 4 where higher means better for each attribute.
These attributes are:
1. **Helpfulness**: Overall helpfulness of the response to the prompt.
2. **Correctness**: Inclusion of all pertinent facts without errors.
3. **Coherence**: Consistency and clarity of expression.
4. **Complexity**: Intellectual depth required to write response (i.e. whether the response can be written by anyone with basic language competency or requires deep domain expertise).
5. **Verbosity**: Amount of detail included in the response, relative to what is asked for in the prompt.
## Source
1. Prompts are collected based on a mixture of template-generated (mainly for prompt involving long reference text) and human generated by Scale AI. These prompts relate to the tasks of Rewrite, Summarization, Classification, Extraction, Closed Question Answering, Open Question Answering, Generation and Brainstorming.
2. Responses are generated by an early version of an inhouse LLM. We generate up to 4 responses per prompts using sample techniques to give diverse yet reasonable responses.
3. Annotations of various attributes were done by Scale AI. Annotators rated each response on a Likert 5 scale (between 0 and 4) for each attribute (helpfulness, correctness, coherence, complexity and verbosity).
## Annotation methodology (short)
1. We engaged a select group of contractors via Scale AI. These contractors were provided with comprehensive guidelines that defined each attribute and the criteria for every rating level, together with some annotated examples. These guidelines and examples are detailed in the Appendix of the accompanying paper.
2. The annotation process involved approximately 200 U.S.-based human annotators. Candidates first underwent preliminary assignments, including assessments of English proficiency, to determine eligibility for working on the project. Subsequently, they participated in an introductory training course on the task which ended with a test that involved annotating 35 sample responses. This process ensured not only a thorough understanding of the task requirements but also the delivery of high-quality annotations.
3. Post-annotations, Scale AI performed extensive quality assurance, with each annotation reaching a minimum of two human reviews in addition to automated checks. After receiving the annotations from Scale AI, we conducted our independent quality assurance to make sure that the quality of the annotations was up to our expectations. As a result, some annotations were filtered away to retain only 37, 120 samples.
## Citation
If you find this dataset useful, please cite the work from the original authors.
```bibtex
@misc{wang2023helpsteer,
title={HelpSteer: Multi-attribute Helpfulness Dataset for SteerLM},
author={Zhilin Wang and Yi Dong and Jiaqi Zeng and Virginia Adams and Makesh Narsimhan Sreedhar and Daniel Egert and Olivier Delalleau and Jane Polak Scowcroft and Neel Kant and Aidan Swope and Oleksii Kuchaiev},
year={2023},
eprint={2311.09528},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
``` | [
-0.4855053424835205,
-0.5253469347953796,
0.28415513038635254,
0.22153396904468536,
-0.09848955273628235,
-0.20065389573574066,
-0.06732947379350662,
-0.48183298110961914,
0.17416490614414215,
0.19088956713676453,
-0.539953351020813,
-0.31195464730262756,
-0.5020045042037964,
0.23998799920... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Gbssreejith/death | Gbssreejith | 2023-11-28T11:55:55Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T11:55:55Z | 2023-11-28T11:51:01.000Z | 2023-11-28T11:51:01 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
- split: val
path: data/val-*
dataset_info:
features:
- name: image
dtype: image
- name: ground_truth
dtype: string
splits:
- name: train
num_bytes: 26946635.0
num_examples: 51
- name: test
num_bytes: 3183225.0
num_examples: 6
- name: val
num_bytes: 3661207.0
num_examples: 7
download_size: 33726877
dataset_size: 33791067.0
---
# Dataset Card for "death"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.4540601074695587,
-0.28580692410469055,
0.46485257148742676,
0.26602017879486084,
-0.40429896116256714,
0.13919050991535187,
0.3358931839466095,
-0.10595101863145828,
0.9028191566467285,
0.528240442276001,
-0.8842344880104065,
-0.8634732961654663,
-0.5942906737327576,
-0.428038716316223... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
indiejoseph/commoncrawl_cantonese | indiejoseph | 2023-11-28T11:57:17Z | 0 | 0 | null | [
"license:mit",
"region:us"
] | 2023-11-28T11:57:17Z | 2023-11-28T11:57:15.000Z | 2023-11-28T11:57:15 | ---
license: mit
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
tim9292654/setting-config | tim9292654 | 2023-11-28T15:30:35Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T15:30:35Z | 2023-11-28T12:01:28.000Z | 2023-11-28T12:01:28 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622264862060547,
0.43461528420448303,
-0.52829909324646,
0.7012971639633179,
0.7915720343589783,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104477167129517,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
isek-ai/danbooru-tags-2016-2023 | isek-ai | 2023-11-28T13:03:24Z | 0 | 0 | null | [
"task_categories:text-classification",
"task_categories:text-generation",
"task_categories:text2text-generation",
"size_categories:1M<n<10M",
"language:en",
"license:cc0-1.0",
"danbooru",
"region:us"
] | 2023-11-28T13:03:24Z | 2023-11-28T12:11:20.000Z | 2023-11-28T12:11:20 | ---
dataset_info:
- config_name: all
features:
- name: id
dtype: int64
- name: copyright
dtype: string
- name: character
dtype: string
- name: artist
dtype: string
- name: general
dtype: string
- name: meta
dtype: string
- name: rating
dtype: string
- name: score
dtype: int64
- name: created_at
dtype: string
splits:
- name: train
num_bytes: 2443978290
num_examples: 4488788
download_size: 966023700
dataset_size: 2443978290
- config_name: safe
features:
- name: id
dtype: int64
- name: copyright
dtype: string
- name: character
dtype: string
- name: artist
dtype: string
- name: general
dtype: string
- name: meta
dtype: string
- name: rating
dtype: string
- name: score
dtype: int64
- name: created_at
dtype: string
splits:
- name: train
num_bytes: 616013975.4781559
num_examples: 1131416
download_size: 235094331
dataset_size: 616013975.4781559
configs:
- config_name: all
data_files:
- split: train
path: all/train-*
- config_name: safe
data_files:
- split: train
path: safe/train-*
license: cc0-1.0
task_categories:
- text-classification
- text-generation
- text2text-generation
language:
- en
tags:
- danbooru
size_categories:
- 1M<n<10M
---
# danbooru-tags-2016-2023
A dataset of danbooru tags.
## Dataset information
Generated using [danbooru](https://danbooru.donmai.us/) and [safebooru](https://safebooru.donmai.us/) API.
The dataset was created with the following conditions:
|Subset name|`all`|`safe`|
|-|-|-|
|API Endpoint|https://danbooru.donmai.us|https://safebooru.donmai.us|
|Date|`2016-01-01..2023-11-27`|`2016-01-01..2023-11-26`|
|Score|`>0`|`>0`|
|Rating|`g,s,q,e`|`g`|
|Filetype|`png,jpg,webp`|`png,jpg,webp`|
|Size (number of rows)|4,488,788|1,131,416|
## Usage
```
pip install datasets
```
```py
from datasets import load_dataset
dataset = load_dataset(
"isek-ai/danbooru-tags-2016-2023",
"safe", # or "all"
split="train",
)
print(dataset)
print(dataset[0])
# Dataset({
# features: ['id', 'copyright', 'character', 'artist', 'general', 'meta', 'rating', 'score', 'created_at'],
# num_rows: 1131416
# })
# {'id': 2229839, 'copyright': 'kara no kyoukai', 'character': 'ryougi shiki', 'artist': 'momoko (momopoco)', 'general': '1girl, 2016, :|, brown eyes, brown hair, closed mouth, cloud, cloudy sky, dated, day, flower, hair flower, hair ornament, japanese clothes, kimono, long hair, long sleeves, looking at viewer, new year, obi, outdoors, sash, shrine, sky, solo, standing, wide sleeves', 'meta': 'commentary request, partial commentary', 'rating': 'g', 'score': 76, 'created_at': '2016-01-01T00:43:18.369+09:00'}
``` | [
-0.672491729259491,
-0.3276430666446686,
0.0726093128323555,
0.2953394949436188,
-0.6832629442214966,
-0.13478809595108032,
-0.054978225380182266,
-0.29142022132873535,
0.6077699065208435,
0.5168628096580505,
-0.8688426613807678,
-0.9629712700843811,
-0.6440953612327576,
0.3149344921112060... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
joseluhf11/oct-fovea-detection | joseluhf11 | 2023-11-28T12:15:09Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T12:15:09Z | 2023-11-28T12:14:41.000Z | 2023-11-28T12:14:41 | ---
dataset_info:
features:
- name: image
dtype: image
- name: objects
struct:
- name: bbox
sequence:
sequence: int64
- name: categories
sequence: string
splits:
- name: train
num_bytes: 350015166.0
num_examples: 431
download_size: 349205446
dataset_size: 350015166.0
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
| [
-0.1285339742898941,
-0.18616800010204315,
0.6529127359390259,
0.4943626821041107,
-0.1931934952735901,
0.2360742688179016,
0.360720157623291,
0.05056300014257431,
0.5793654322624207,
0.7400140166282654,
-0.6508105993270874,
-0.23783984780311584,
-0.7102248668670654,
-0.047826044261455536,... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
BangorAI/exl2-wiki-calibration-set-cy | BangorAI | 2023-11-28T16:41:25Z | 0 | 0 | null | [
"license:cc-by-sa-3.0",
"region:us"
] | 2023-11-28T16:41:25Z | 2023-11-28T12:14:53.000Z | 2023-11-28T12:14:53 | ---
license: cc-by-sa-3.0
---
### Data Calibro Exl2
Detholiad o [Cofnod y Cynulliad](https://huggingface.co/datasets/techiaith/cofnodycynulliad_en-cy) Cymraeg i'w ddefnyddio yng ngham calibro ExLlama 2 wrth drosi modelau i fformat exl2. | [
-0.09711340814828873,
-0.2969370484352112,
-0.025429617613554,
0.17108246684074402,
-0.5092402100563049,
0.0529848150908947,
0.048829738050699234,
-0.28459984064102173,
0.7184000015258789,
0.6294670701026917,
-0.9982383251190186,
-0.692694902420044,
-0.4484431743621826,
-0.1142623499035835... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
ddps007/test-dataset-v8 | ddps007 | 2023-11-28T12:17:51Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T12:17:51Z | 2023-11-28T12:17:49.000Z | 2023-11-28T12:17:49 | Entry not found | [
-0.32276487350463867,
-0.22568444907665253,
0.8622263073921204,
0.43461570143699646,
-0.5282988548278809,
0.7012969255447388,
0.7915717363357544,
0.07618642598390579,
0.7746027112007141,
0.25632190704345703,
-0.7852815389633179,
-0.22573848068714142,
-0.910447895526886,
0.5715675354003906,... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
ddps007/test-dataset-v10 | ddps007 | 2023-11-28T12:18:15Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T12:18:15Z | 2023-11-28T12:18:14.000Z | 2023-11-28T12:18:14 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Alexandre-Numind/BenchFew | Alexandre-Numind | 2023-11-28T12:20:02Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T12:20:02Z | 2023-11-28T12:19:26.000Z | 2023-11-28T12:19:26 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
ddps007/test-dd | ddps007 | 2023-11-29T00:43:14Z | 0 | 0 | null | [
"region:us"
] | 2023-11-29T00:43:14Z | 2023-11-28T12:19:44.000Z | 2023-11-28T12:19:44 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Alexandre-Numind/BenchNoFew | Alexandre-Numind | 2023-11-28T12:22:37Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T12:22:37Z | 2023-11-28T12:21:59.000Z | 2023-11-28T12:21:59 | Entry not found | [
-0.3227649927139282,
-0.225684255361557,
0.862226128578186,
0.43461498618125916,
-0.5282987952232361,
0.7012963891029358,
0.7915717363357544,
0.07618629932403564,
0.7746025919914246,
0.2563219666481018,
-0.7852816581726074,
-0.2257382869720459,
-0.9104480743408203,
0.5715669393539429,
-0... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Alexandre-Numind/IE_FS | Alexandre-Numind | 2023-11-28T12:24:10Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T12:24:10Z | 2023-11-28T12:23:29.000Z | 2023-11-28T12:23:29 | Entry not found | [
-0.3227645754814148,
-0.22568479180335999,
0.8622264862060547,
0.43461528420448303,
-0.52829909324646,
0.7012971639633179,
0.7915720343589783,
0.07618614286184311,
0.774603009223938,
0.2563217282295227,
-0.7852813005447388,
-0.22573819756507874,
-0.9104477167129517,
0.5715674161911011,
-... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
SanaFalakJ/IA | SanaFalakJ | 2023-11-28T12:29:15Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T12:29:15Z | 2023-11-28T12:29:10.000Z | 2023-11-28T12:29:10 | ---
dataset_info:
features:
- name: text
dtype: string
splits:
- name: train
num_bytes: 64521
num_examples: 19
download_size: 41947
dataset_size: 64521
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
fabsss/westfalmelado | fabsss | 2023-11-28T12:30:03Z | 0 | 0 | null | [
"license:apache-2.0",
"region:us"
] | 2023-11-28T12:30:03Z | 2023-11-28T12:29:18.000Z | 2023-11-28T12:29:18 | ---
license: apache-2.0
---
| [
-0.1285335123538971,
-0.1861683875322342,
0.6529128551483154,
0.49436232447624207,
-0.19319400191307068,
0.23607441782951355,
0.36072009801864624,
0.05056373029947281,
0.5793656706809998,
0.7400146722793579,
-0.650810182094574,
-0.23784008622169495,
-0.7102247476577759,
-0.0478255338966846... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
yimhuang/guanaco-llama2-1k | yimhuang | 2023-11-28T12:38:21Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T12:38:21Z | 2023-11-28T12:30:09.000Z | 2023-11-28T12:30:09 | ---
dataset_info:
features:
- name: text
dtype: string
splits:
- name: train
num_bytes: 1654448
num_examples: 1000
download_size: 966693
dataset_size: 1654448
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Guanaco-1k: Lazy Llama 2 Formatting
This is a subset (1000 samples) of the excellent [`timdettmers/openassistant-guanaco`](https://huggingface.co/datasets/timdettmers/openassistant-guanaco) dataset, processed to match Llama 2's prompt format as described [in this article](https://huggingface.co/blog/llama2#how-to-prompt-llama-2). It was created using the following [colab notebook](https://colab.research.google.com/drive/1Ad7a9zMmkxuXTOh1Z7-rNSICA4dybpM2?usp=sharing).
Useful if you don't want to reformat it by yourself (e.g., using a script). It was designed for [this article](https://mlabonne.github.io/blog/posts/Fine_Tune_Your_Own_Llama_2_Model_in_a_Colab_Notebook.html) about fine-tuning a Llama 2 (chat) model in a Google Colab.
| [
-0.043327562510967255,
-0.9048401117324829,
0.37380853295326233,
0.8770328164100647,
-0.5532293915748596,
0.04655826836824417,
-0.15234212577342987,
-0.3322109878063202,
0.5565394759178162,
0.36095863580703735,
-0.9159899353981018,
-0.5867228507995605,
-0.3724720776081085,
0.14088976383209... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
zinc75/Vibravox_dummy | zinc75 | 2023-11-28T20:36:39Z | 0 | 0 | null | [
"task_categories:audio-to-audio",
"task_categories:automatic-speech-recognition",
"task_categories:audio-classification",
"task_categories:text-to-speech",
"task_ids:speaker-identification",
"annotations_creators:expert-generated",
"language_creators:crowdsourced",
"language_creators:expert-generated"... | 2023-11-28T20:36:39Z | 2023-11-28T12:33:43.000Z | 2023-11-28T12:33:43 | ---
license: cc-by-4.0
task_categories:
- audio-to-audio
- automatic-speech-recognition
- audio-classification
- text-to-speech
task_ids:
- speaker-identification
size_categories:
- 10K<n<100K
source_datasets: []
language:
- fr
multilinguality:
- monolingual
language_creators:
- crowdsourced
- expert-generated
annotations_creators:
- expert-generated
pretty_name: 'VibraVox'
configs:
- config_name: ASR_Reference_microphone
data_files:
- split: train
path: "train_ASR_ref_microphone.tsv"
- split: val
path: "val_ASR_ref_microphone.tsv"
- split: test
path: "test_ASR_ref_microphone.tsv"
- config_name: ASR_Laryngophone
features:
- name: file_name
dtype: string
- name: audio
dtype: audio
- name: transcription
dtype: string
- name: sensor_id
dtype: string
- name: speaker_id
dtype: int64
- name: gender
dtype: string
- name: type
dtype: string
- name: gender
dtype: string
- name: split
dtype: string
data_files:
- split: train
path: "train_ASR_laryngophone.tsv"
- split: val
path: "val_ASR_laryngophone.tsv"
- split: test
path: "test_ASR_laryngophone.tsv"
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
vishnu027/death | vishnu027 | 2023-11-28T12:48:17Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T12:48:17Z | 2023-11-28T12:34:05.000Z | 2023-11-28T12:34:05 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
- split: val
path: data/val-*
dataset_info:
features:
- name: image
dtype: image
- name: ground_truth
dtype: string
splits:
- name: train
num_bytes: 749731844.0
num_examples: 560
- name: test
num_bytes: 97083916.0
num_examples: 70
- name: val
num_bytes: 95493624.0
num_examples: 70
download_size: 940687614
dataset_size: 942309384.0
---
# Dataset Card for "death"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.4540598392486572,
-0.2858071029186249,
0.4648527204990387,
0.26601991057395935,
-0.40429848432540894,
0.13919086754322052,
0.33589306473731995,
-0.10595114529132843,
0.9028188586235046,
0.5282403230667114,
-0.8842339515686035,
-0.8634734153747559,
-0.5942906141281128,
-0.428038775920867... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
TeeA/Vietnamese-Chart-Dataset | TeeA | 2023-11-28T13:41:25Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T13:41:25Z | 2023-11-28T12:35:29.000Z | 2023-11-28T12:35:29 | ---
dataset_info:
features:
- name: title
dtype: string
- name: x_title
dtype: string
- name: y_title
dtype: string
- name: x
dtype: string
- name: y
dtype: string
- name: file_name
dtype: string
- name: chart_type
dtype: string
- name: image
dtype: image
splits:
- name: train
num_bytes: 115631536.42857143
num_examples: 5000
- name: test
num_bytes: 23422771.285714287
num_examples: 1000
- name: validation
num_bytes: 23502759.285714287
num_examples: 1000
download_size: 116048333
dataset_size: 162557067.00000003
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
- split: validation
path: data/validation-*
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
ctoraman/BilTweetNews-Sentiment | ctoraman | 2023-11-28T12:41:24Z | 0 | 0 | null | [
"license:cc-by-nc-sa-4.0",
"region:us"
] | 2023-11-28T12:41:24Z | 2023-11-28T12:41:24.000Z | 2023-11-28T12:41:24 | ---
license: cc-by-nc-sa-4.0
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
Jaspernl/common_voice_13_0_nl_pseudo_labelled | Jaspernl | 2023-11-28T13:50:06Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T13:50:06Z | 2023-11-28T12:45:19.000Z | 2023-11-28T12:45:19 | ---
dataset_info:
config_name: nl
features:
- name: client_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 16000
- name: sentence
dtype: string
- name: up_votes
dtype: int64
- name: down_votes
dtype: int64
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
- name: locale
dtype: string
- name: segment
dtype: string
- name: variant
dtype: string
- name: whisper_transcript
sequence: int64
splits:
- name: train
num_bytes: 887412317.796
num_examples: 31906
- name: validation
num_bytes: 355862437.37
num_examples: 10930
- name: test
num_bytes: 402683280.568
num_examples: 10936
download_size: 1643910548
dataset_size: 1645958035.734
configs:
- config_name: nl
data_files:
- split: train
path: nl/train-*
- split: validation
path: nl/validation-*
- split: test
path: nl/test-*
---
| [
-0.12853392958641052,
-0.18616779148578644,
0.6529127955436707,
0.49436280131340027,
-0.19319361448287964,
0.23607419431209564,
0.36072003841400146,
0.050563063472509384,
0.579365611076355,
0.7400140762329102,
-0.6508104205131531,
-0.23783954977989197,
-0.7102249264717102,
-0.0478260256350... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
tyzhu/find_sent_before_sent_train_100_eval_40_recite | tyzhu | 2023-11-28T13:43:41Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T13:43:41Z | 2023-11-28T12:47:59.000Z | 2023-11-28T12:47:59 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
dataset_info:
features:
- name: inputs
dtype: string
- name: targets
dtype: string
- name: title
dtype: string
- name: context
dtype: string
splits:
- name: train
num_bytes: 1169584
num_examples: 644
- name: validation
num_bytes: 377548
num_examples: 202
download_size: 325994
dataset_size: 1547132
---
# Dataset Card for "find_sent_before_sent_train_100_eval_40_recite"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.5445691347122192,
-0.0738573893904686,
0.2889677882194519,
0.42903923988342285,
-0.026656627655029297,
-0.02416963130235672,
0.12088538706302643,
0.21919551491737366,
0.7557971477508545,
0.6391089558601379,
-1.0631091594696045,
-0.6634748578071594,
-0.5230261087417603,
-0.21797747910022... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
tyzhu/find_sent_after_sent_train_100_eval_40_recite | tyzhu | 2023-11-28T13:44:20Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T13:44:20Z | 2023-11-28T12:49:42.000Z | 2023-11-28T12:49:42 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
dataset_info:
features:
- name: inputs
dtype: string
- name: targets
dtype: string
- name: title
dtype: string
- name: context
dtype: string
splits:
- name: train
num_bytes: 1168154
num_examples: 644
- name: validation
num_bytes: 377200
num_examples: 202
download_size: 325715
dataset_size: 1545354
---
# Dataset Card for "find_sent_after_sent_train_100_eval_40_recite"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.496530681848526,
-0.008401576429605484,
0.3072666525840759,
0.4487752318382263,
0.004115242511034012,
0.02256515994668007,
0.09870839864015579,
0.1785963922739029,
0.7407034635543823,
0.5947513580322266,
-0.9809296131134033,
-0.599894106388092,
-0.5363748669624329,
-0.2411201447248459,
... | null | null | null | null | null | null | null | null | null | null | null | null | null | |
tyzhu/find_sent_before_sent_train_200_eval_40_recite | tyzhu | 2023-11-28T13:44:51Z | 0 | 0 | null | [
"region:us"
] | 2023-11-28T13:44:51Z | 2023-11-28T12:50:15.000Z | 2023-11-28T12:50:15 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
dataset_info:
features:
- name: inputs
dtype: string
- name: targets
dtype: string
- name: title
dtype: string
- name: context
dtype: string
splits:
- name: train
num_bytes: 2329316
num_examples: 1263
- name: validation
num_bytes: 398956
num_examples: 203
download_size: 533740
dataset_size: 2728272
---
# Dataset Card for "find_sent_before_sent_train_200_eval_40_recite"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | [
-0.5391755700111389,
-0.0023822118528187275,
0.3294859230518341,
0.4412449896335602,
-0.031510014086961746,
0.018736662343144417,
0.12890343368053436,
0.18120229244232178,
0.7011620402336121,
0.631971538066864,
-1.0757534503936768,
-0.6322876811027527,
-0.5338778495788574,
-0.2033138573169... | null | null | null | null | null | null | null | null | null | null | null | null | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.