id stringlengths 2 115 | lastModified stringlengths 24 24 | tags list | author stringlengths 2 42 ⌀ | description stringlengths 0 68.7k ⌀ | citation stringlengths 0 10.7k ⌀ | cardData null | likes int64 0 3.55k | downloads int64 0 10.1M | card stringlengths 0 1.01M |
|---|---|---|---|---|---|---|---|---|---|
ryanc/music_align | 2023-08-29T02:50:52.000Z | [
"region:us"
] | ryanc | null | null | null | 0 | 140 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
dataset_info:
features:
- name: caption
dtype: string
- name: audio
dtype: audio
splits:
- name: train
num_bytes: 16132095937.715
num_examples: 8537
download_size: 1862624886
dataset_size: 16132095937.715
---
# Dataset Card for "music_align"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
atmallen/sloppy_addition_both_labels_1.0 | 2023-10-05T17:49:40.000Z | [
"region:us"
] | atmallen | null | null | null | 0 | 140 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
- split: test
path: data/test-*
dataset_info:
features:
- name: statement
dtype: string
- name: alice_label
dtype: bool
- name: bob_label
dtype: bool
- name: id
dtype: int64
splits:
- name: train
num_bytes: 5571344
num_examples: 200000
- name: validation
num_bytes: 557449
num_examples: 20000
- name: test
num_bytes: 556155
num_examples: 20000
download_size: 4471117
dataset_size: 6684948
---
# Dataset Card for "sloppy_addition_both_labels_1.0"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
result-kand2-sdxl-wuerst-karlo/980edb53 | 2023-10-04T18:22:22.000Z | [
"region:us"
] | result-kand2-sdxl-wuerst-karlo | null | null | null | 0 | 140 | ---
dataset_info:
features:
- name: result
dtype: string
- name: id
dtype: int64
splits:
- name: train
num_bytes: 156
num_examples: 10
download_size: 1319
dataset_size: 156
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "980edb53"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
result-kand2-sdxl-wuerst-karlo/f7f54a55 | 2023-10-04T18:43:41.000Z | [
"region:us"
] | result-kand2-sdxl-wuerst-karlo | null | null | null | 0 | 140 | ---
dataset_info:
features:
- name: result
dtype: string
- name: id
dtype: int64
splits:
- name: train
num_bytes: 141
num_examples: 10
download_size: 1325
dataset_size: 141
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "f7f54a55"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
gnad10 | 2023-01-25T14:31:03.000Z | [
"task_categories:text-classification",
"task_ids:topic-classification",
"annotations_creators:crowdsourced",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:extended|other-from-One-Million-Posts-Corpus",
"language:de",
"license:cc-by-nc-sa-4.0",
"region:us"
] | null | This dataset is intended to advance topic classification for German texts. A classifier that is efffective in
English may not be effective in German dataset because it has a higher inflection and longer compound words.
The 10kGNAD dataset contains 10273 German news articles from an Austrian online newspaper categorized into
9 categories. Article titles and text are concatenated together and authors are removed to avoid a keyword-like
classification on authors that write frequently about one category. This dataset can be used as a benchmark
for German topic classification. | null | null | 3 | 139 | ---
annotations_creators:
- crowdsourced
language_creators:
- found
language:
- de
license:
- cc-by-nc-sa-4.0
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- extended|other-from-One-Million-Posts-Corpus
task_categories:
- text-classification
task_ids:
- topic-classification
pretty_name: 10k German News Articles Datasets
dataset_info:
features:
- name: text
dtype: string
- name: label
dtype:
class_label:
names:
'0': Web
'1': Panorama
'2': International
'3': Wirtschaft
'4': Sport
'5': Inland
'6': Etat
'7': Wissenschaft
'8': Kultur
splits:
- name: train
num_bytes: 24418224
num_examples: 9245
- name: test
num_bytes: 2756405
num_examples: 1028
download_size: 27160809
dataset_size: 27174629
---
# Dataset Card for 10k German News Articles Datasets
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [10k German News Article Dataset](https://tblock.github.io/10kGNAD/)
- **Repository:** [10k German News Article Dataset](https://github.com/tblock/10kGNAD)
- **Point of Contact:** [Steven Liu](stevhliu@gmail.com)
### Dataset Summary
The 10k German News Article Dataset consists of 10273 German language news articles from the online Austrian
newspaper website DER Standard. Each news article has been classified into one of 9 categories by professional
forum moderators employed by the newspaper. This dataset is extended from the original
[One Million Posts Corpus](https://ofai.github.io/million-post-corpus/). The dataset was created to support
topic classification in German because a classifier effective on a English dataset may not be as effective on
a German dataset due to higher inflections and longer compound words. Additionally, this dataset can be used
as a benchmark dataset for German topic classification.
### Supported Tasks and Leaderboards
This dataset can be used to train a model, like [BERT](https://huggingface.co/bert-base-uncased) for `topic classification` on German news articles. There are 9 possible categories.
### Languages
The text is in German and it comes from an online Austrian newspaper website. The BCP-47 code for German is
`de-DE`.
## Dataset Structure
### Data Instances
An example data instance contains a German news article (title and article are concatenated) and it's corresponding topic category.
```
{'text': ''Die Gewerkschaft GPA-djp lanciert den "All-in-Rechner" und findet, dass die Vertragsform auf die Führungsebene beschränkt gehört. Wien – Die Gewerkschaft GPA-djp sieht Handlungsbedarf bei sogenannten All-in-Verträgen.'
'label': 'Wirtschaft'
}
```
### Data Fields
* `text`: contains the title and content of the article
* `label`: can be one of 9 possible topic categories (`Web`, `Panorama`, `International`, `Wirtschaft`, `Sport`, `Inland`, `Etat`, `Wissenschaft`, `Kultur`)
### Data Splits
The data is split into a training set consisting of 9245 articles and a test set consisting of 1028 articles.
## Dataset Creation
### Curation Rationale
The dataset was created to support topic classification in the German language. English text classification datasets are common ([AG News](https://huggingface.co/datasets/ag_news) and [20 Newsgroup](https://huggingface.co/datasets/newsgroup)), but German datasets are less common. A classifier trained on an English dataset may not work as well on a set of German text due to grammatical differences. Thus there is a need for a German dataset for effectively assessing model performance.
### Source Data
#### Initial Data Collection and Normalization
The 10k German News Article Dataset is extended from the One Million Posts Corpus. 10273 German news articles were collected from this larger corpus. In the One Million Posts Corpus, each article has a topic path like
`Newsroom/Wirtschaft/Wirtschaftpolitik/Finanzmaerkte/Griechenlandkrise`. The 10kGNAD uses the second part of the topic path as the topic label. Article title and texts are concatenated into one text and author names are removed to avoid keyword classification on authors who write frequently on a particular topic.
#### Who are the source language producers?
The language producers are the authors of the Austrian newspaper website DER Standard.
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
This dataset was curated by Timo Block.
### Licensing Information
This dataset is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 license.
### Citation Information
Please consider citing the authors of the "One Million Post Corpus" if you use the dataset.:
```
@InProceedings{Schabus2017,
Author = {Dietmar Schabus and Marcin Skowron and Martin Trapp},
Title = {One Million Posts: A Data Set of German Online Discussions},
Booktitle = {Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR)},
Pages = {1241--1244},
Year = {2017},
Address = {Tokyo, Japan},
Doi = {10.1145/3077136.3080711},
Month = aug
}
```
### Contributions
Thanks to [@stevhliu](https://github.com/stevhliu) for adding this dataset. |
imagenet_sketch | 2023-04-05T13:45:57.000Z | [
"task_categories:image-classification",
"task_ids:multi-class-image-classification",
"annotations_creators:crowdsourced",
"language_creators:crowdsourced",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:extended|imagenet-1k",
"language:en",
"license:unknown",
"arxiv:1905.13549",
"region:us"
] | null | ImageNet-Sketch data set consists of 50000 images, 50 images for each of the 1000 ImageNet classes.
We construct the data set with Google Image queries "sketch of __", where __ is the standard class name.
We only search within the "black and white" color scheme. We initially query 100 images for every class,
and then manually clean the pulled images by deleting the irrelevant images and images that are for similar
but different classes. For some classes, there are less than 50 images after manually cleaning, and then we
augment the data set by flipping and rotating the images. | @inproceedings{wang2019learning,
title={Learning Robust Global Representations by Penalizing Local Predictive Power},
author={Wang, Haohan and Ge, Songwei and Lipton, Zachary and Xing, Eric P},
booktitle={Advances in Neural Information Processing Systems},
pages={10506--10518},
year={2019}
} | null | 5 | 139 | ---
annotations_creators:
- crowdsourced
language_creators:
- crowdsourced
language:
- en
license:
- unknown
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- extended|imagenet-1k
task_categories:
- image-classification
task_ids:
- multi-class-image-classification
paperswithcode_id: imagenet-sketch
pretty_name: ImageNet-Sketch
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': tench, Tinca tinca
'1': goldfish, Carassius auratus
'2': great white shark, white shark, man-eater, man-eating shark, Carcharodon
carcharias
'3': tiger shark, Galeocerdo cuvieri
'4': hammerhead, hammerhead shark
'5': electric ray, crampfish, numbfish, torpedo
'6': stingray
'7': cock
'8': hen
'9': ostrich, Struthio camelus
'10': brambling, Fringilla montifringilla
'11': goldfinch, Carduelis carduelis
'12': house finch, linnet, Carpodacus mexicanus
'13': junco, snowbird
'14': indigo bunting, indigo finch, indigo bird, Passerina cyanea
'15': robin, American robin, Turdus migratorius
'16': bulbul
'17': jay
'18': magpie
'19': chickadee
'20': water ouzel, dipper
'21': kite
'22': bald eagle, American eagle, Haliaeetus leucocephalus
'23': vulture
'24': great grey owl, great gray owl, Strix nebulosa
'25': European fire salamander, Salamandra salamandra
'26': common newt, Triturus vulgaris
'27': eft
'28': spotted salamander, Ambystoma maculatum
'29': axolotl, mud puppy, Ambystoma mexicanum
'30': bullfrog, Rana catesbeiana
'31': tree frog, tree-frog
'32': tailed frog, bell toad, ribbed toad, tailed toad, Ascaphus trui
'33': loggerhead, loggerhead turtle, Caretta caretta
'34': leatherback turtle, leatherback, leathery turtle, Dermochelys coriacea
'35': mud turtle
'36': terrapin
'37': box turtle, box tortoise
'38': banded gecko
'39': common iguana, iguana, Iguana iguana
'40': American chameleon, anole, Anolis carolinensis
'41': whiptail, whiptail lizard
'42': agama
'43': frilled lizard, Chlamydosaurus kingi
'44': alligator lizard
'45': Gila monster, Heloderma suspectum
'46': green lizard, Lacerta viridis
'47': African chameleon, Chamaeleo chamaeleon
'48': Komodo dragon, Komodo lizard, dragon lizard, giant lizard, Varanus
komodoensis
'49': African crocodile, Nile crocodile, Crocodylus niloticus
'50': American alligator, Alligator mississipiensis
'51': triceratops
'52': thunder snake, worm snake, Carphophis amoenus
'53': ringneck snake, ring-necked snake, ring snake
'54': hognose snake, puff adder, sand viper
'55': green snake, grass snake
'56': king snake, kingsnake
'57': garter snake, grass snake
'58': water snake
'59': vine snake
'60': night snake, Hypsiglena torquata
'61': boa constrictor, Constrictor constrictor
'62': rock python, rock snake, Python sebae
'63': Indian cobra, Naja naja
'64': green mamba
'65': sea snake
'66': horned viper, cerastes, sand viper, horned asp, Cerastes cornutus
'67': diamondback, diamondback rattlesnake, Crotalus adamanteus
'68': sidewinder, horned rattlesnake, Crotalus cerastes
'69': trilobite
'70': harvestman, daddy longlegs, Phalangium opilio
'71': scorpion
'72': black and gold garden spider, Argiope aurantia
'73': barn spider, Araneus cavaticus
'74': garden spider, Aranea diademata
'75': black widow, Latrodectus mactans
'76': tarantula
'77': wolf spider, hunting spider
'78': tick
'79': centipede
'80': black grouse
'81': ptarmigan
'82': ruffed grouse, partridge, Bonasa umbellus
'83': prairie chicken, prairie grouse, prairie fowl
'84': peacock
'85': quail
'86': partridge
'87': African grey, African gray, Psittacus erithacus
'88': macaw
'89': sulphur-crested cockatoo, Kakatoe galerita, Cacatua galerita
'90': lorikeet
'91': coucal
'92': bee eater
'93': hornbill
'94': hummingbird
'95': jacamar
'96': toucan
'97': drake
'98': red-breasted merganser, Mergus serrator
'99': goose
'100': black swan, Cygnus atratus
'101': tusker
'102': echidna, spiny anteater, anteater
'103': platypus, duckbill, duckbilled platypus, duck-billed platypus, Ornithorhynchus
anatinus
'104': wallaby, brush kangaroo
'105': koala, koala bear, kangaroo bear, native bear, Phascolarctos cinereus
'106': wombat
'107': jellyfish
'108': sea anemone, anemone
'109': brain coral
'110': flatworm, platyhelminth
'111': nematode, nematode worm, roundworm
'112': conch
'113': snail
'114': slug
'115': sea slug, nudibranch
'116': chiton, coat-of-mail shell, sea cradle, polyplacophore
'117': chambered nautilus, pearly nautilus, nautilus
'118': Dungeness crab, Cancer magister
'119': rock crab, Cancer irroratus
'120': fiddler crab
'121': king crab, Alaska crab, Alaskan king crab, Alaska king crab, Paralithodes
camtschatica
'122': American lobster, Northern lobster, Maine lobster, Homarus americanus
'123': spiny lobster, langouste, rock lobster, crawfish, crayfish, sea crawfish
'124': crayfish, crawfish, crawdad, crawdaddy
'125': hermit crab
'126': isopod
'127': white stork, Ciconia ciconia
'128': black stork, Ciconia nigra
'129': spoonbill
'130': flamingo
'131': little blue heron, Egretta caerulea
'132': American egret, great white heron, Egretta albus
'133': bittern
'134': crane
'135': limpkin, Aramus pictus
'136': European gallinule, Porphyrio porphyrio
'137': American coot, marsh hen, mud hen, water hen, Fulica americana
'138': bustard
'139': ruddy turnstone, Arenaria interpres
'140': red-backed sandpiper, dunlin, Erolia alpina
'141': redshank, Tringa totanus
'142': dowitcher
'143': oystercatcher, oyster catcher
'144': pelican
'145': king penguin, Aptenodytes patagonica
'146': albatross, mollymawk
'147': grey whale, gray whale, devilfish, Eschrichtius gibbosus, Eschrichtius
robustus
'148': killer whale, killer, orca, grampus, sea wolf, Orcinus orca
'149': dugong, Dugong dugon
'150': sea lion
'151': Chihuahua
'152': Japanese spaniel
'153': Maltese dog, Maltese terrier, Maltese
'154': Pekinese, Pekingese, Peke
'155': Shih-Tzu
'156': Blenheim spaniel
'157': papillon
'158': toy terrier
'159': Rhodesian ridgeback
'160': Afghan hound, Afghan
'161': basset, basset hound
'162': beagle
'163': bloodhound, sleuthhound
'164': bluetick
'165': black-and-tan coonhound
'166': Walker hound, Walker foxhound
'167': English foxhound
'168': redbone
'169': borzoi, Russian wolfhound
'170': Irish wolfhound
'171': Italian greyhound
'172': whippet
'173': Ibizan hound, Ibizan Podenco
'174': Norwegian elkhound, elkhound
'175': otterhound, otter hound
'176': Saluki, gazelle hound
'177': Scottish deerhound, deerhound
'178': Weimaraner
'179': Staffordshire bullterrier, Staffordshire bull terrier
'180': American Staffordshire terrier, Staffordshire terrier, American pit
bull terrier, pit bull terrier
'181': Bedlington terrier
'182': Border terrier
'183': Kerry blue terrier
'184': Irish terrier
'185': Norfolk terrier
'186': Norwich terrier
'187': Yorkshire terrier
'188': wire-haired fox terrier
'189': Lakeland terrier
'190': Sealyham terrier, Sealyham
'191': Airedale, Airedale terrier
'192': cairn, cairn terrier
'193': Australian terrier
'194': Dandie Dinmont, Dandie Dinmont terrier
'195': Boston bull, Boston terrier
'196': miniature schnauzer
'197': giant schnauzer
'198': standard schnauzer
'199': Scotch terrier, Scottish terrier, Scottie
'200': Tibetan terrier, chrysanthemum dog
'201': silky terrier, Sydney silky
'202': soft-coated wheaten terrier
'203': West Highland white terrier
'204': Lhasa, Lhasa apso
'205': flat-coated retriever
'206': curly-coated retriever
'207': golden retriever
'208': Labrador retriever
'209': Chesapeake Bay retriever
'210': German short-haired pointer
'211': vizsla, Hungarian pointer
'212': English setter
'213': Irish setter, red setter
'214': Gordon setter
'215': Brittany spaniel
'216': clumber, clumber spaniel
'217': English springer, English springer spaniel
'218': Welsh springer spaniel
'219': cocker spaniel, English cocker spaniel, cocker
'220': Sussex spaniel
'221': Irish water spaniel
'222': kuvasz
'223': schipperke
'224': groenendael
'225': malinois
'226': briard
'227': kelpie
'228': komondor
'229': Old English sheepdog, bobtail
'230': Shetland sheepdog, Shetland sheep dog, Shetland
'231': collie
'232': Border collie
'233': Bouvier des Flandres, Bouviers des Flandres
'234': Rottweiler
'235': German shepherd, German shepherd dog, German police dog, alsatian
'236': Doberman, Doberman pinscher
'237': miniature pinscher
'238': Greater Swiss Mountain dog
'239': Bernese mountain dog
'240': Appenzeller
'241': EntleBucher
'242': boxer
'243': bull mastiff
'244': Tibetan mastiff
'245': French bulldog
'246': Great Dane
'247': Saint Bernard, St Bernard
'248': Eskimo dog, husky
'249': malamute, malemute, Alaskan malamute
'250': Siberian husky
'251': dalmatian, coach dog, carriage dog
'252': affenpinscher, monkey pinscher, monkey dog
'253': basenji
'254': pug, pug-dog
'255': Leonberg
'256': Newfoundland, Newfoundland dog
'257': Great Pyrenees
'258': Samoyed, Samoyede
'259': Pomeranian
'260': chow, chow chow
'261': keeshond
'262': Brabancon griffon
'263': Pembroke, Pembroke Welsh corgi
'264': Cardigan, Cardigan Welsh corgi
'265': toy poodle
'266': miniature poodle
'267': standard poodle
'268': Mexican hairless
'269': timber wolf, grey wolf, gray wolf, Canis lupus
'270': white wolf, Arctic wolf, Canis lupus tundrarum
'271': red wolf, maned wolf, Canis rufus, Canis niger
'272': coyote, prairie wolf, brush wolf, Canis latrans
'273': dingo, warrigal, warragal, Canis dingo
'274': dhole, Cuon alpinus
'275': African hunting dog, hyena dog, Cape hunting dog, Lycaon pictus
'276': hyena, hyaena
'277': red fox, Vulpes vulpes
'278': kit fox, Vulpes macrotis
'279': Arctic fox, white fox, Alopex lagopus
'280': grey fox, gray fox, Urocyon cinereoargenteus
'281': tabby, tabby cat
'282': tiger cat
'283': Persian cat
'284': Siamese cat, Siamese
'285': Egyptian cat
'286': cougar, puma, catamount, mountain lion, painter, panther, Felis concolor
'287': lynx, catamount
'288': leopard, Panthera pardus
'289': snow leopard, ounce, Panthera uncia
'290': jaguar, panther, Panthera onca, Felis onca
'291': lion, king of beasts, Panthera leo
'292': tiger, Panthera tigris
'293': cheetah, chetah, Acinonyx jubatus
'294': brown bear, bruin, Ursus arctos
'295': American black bear, black bear, Ursus americanus, Euarctos americanus
'296': ice bear, polar bear, Ursus Maritimus, Thalarctos maritimus
'297': sloth bear, Melursus ursinus, Ursus ursinus
'298': mongoose
'299': meerkat, mierkat
'300': tiger beetle
'301': ladybug, ladybeetle, lady beetle, ladybird, ladybird beetle
'302': ground beetle, carabid beetle
'303': long-horned beetle, longicorn, longicorn beetle
'304': leaf beetle, chrysomelid
'305': dung beetle
'306': rhinoceros beetle
'307': weevil
'308': fly
'309': bee
'310': ant, emmet, pismire
'311': grasshopper, hopper
'312': cricket
'313': walking stick, walkingstick, stick insect
'314': cockroach, roach
'315': mantis, mantid
'316': cicada, cicala
'317': leafhopper
'318': lacewing, lacewing fly
'319': dragonfly, darning needle, devil's darning needle, sewing needle,
snake feeder, snake doctor, mosquito hawk, skeeter hawk
'320': damselfly
'321': admiral
'322': ringlet, ringlet butterfly
'323': monarch, monarch butterfly, milkweed butterfly, Danaus plexippus
'324': cabbage butterfly
'325': sulphur butterfly, sulfur butterfly
'326': lycaenid, lycaenid butterfly
'327': starfish, sea star
'328': sea urchin
'329': sea cucumber, holothurian
'330': wood rabbit, cottontail, cottontail rabbit
'331': hare
'332': Angora, Angora rabbit
'333': hamster
'334': porcupine, hedgehog
'335': fox squirrel, eastern fox squirrel, Sciurus niger
'336': marmot
'337': beaver
'338': guinea pig, Cavia cobaya
'339': sorrel
'340': zebra
'341': hog, pig, grunter, squealer, Sus scrofa
'342': wild boar, boar, Sus scrofa
'343': warthog
'344': hippopotamus, hippo, river horse, Hippopotamus amphibius
'345': ox
'346': water buffalo, water ox, Asiatic buffalo, Bubalus bubalis
'347': bison
'348': ram, tup
'349': bighorn, bighorn sheep, cimarron, Rocky Mountain bighorn, Rocky Mountain
sheep, Ovis canadensis
'350': ibex, Capra ibex
'351': hartebeest
'352': impala, Aepyceros melampus
'353': gazelle
'354': Arabian camel, dromedary, Camelus dromedarius
'355': llama
'356': weasel
'357': mink
'358': polecat, fitch, foulmart, foumart, Mustela putorius
'359': black-footed ferret, ferret, Mustela nigripes
'360': otter
'361': skunk, polecat, wood pussy
'362': badger
'363': armadillo
'364': three-toed sloth, ai, Bradypus tridactylus
'365': orangutan, orang, orangutang, Pongo pygmaeus
'366': gorilla, Gorilla gorilla
'367': chimpanzee, chimp, Pan troglodytes
'368': gibbon, Hylobates lar
'369': siamang, Hylobates syndactylus, Symphalangus syndactylus
'370': guenon, guenon monkey
'371': patas, hussar monkey, Erythrocebus patas
'372': baboon
'373': macaque
'374': langur
'375': colobus, colobus monkey
'376': proboscis monkey, Nasalis larvatus
'377': marmoset
'378': capuchin, ringtail, Cebus capucinus
'379': howler monkey, howler
'380': titi, titi monkey
'381': spider monkey, Ateles geoffroyi
'382': squirrel monkey, Saimiri sciureus
'383': Madagascar cat, ring-tailed lemur, Lemur catta
'384': indri, indris, Indri indri, Indri brevicaudatus
'385': Indian elephant, Elephas maximus
'386': African elephant, Loxodonta africana
'387': lesser panda, red panda, panda, bear cat, cat bear, Ailurus fulgens
'388': giant panda, panda, panda bear, coon bear, Ailuropoda melanoleuca
'389': barracouta, snoek
'390': eel
'391': coho, cohoe, coho salmon, blue jack, silver salmon, Oncorhynchus
kisutch
'392': rock beauty, Holocanthus tricolor
'393': anemone fish
'394': sturgeon
'395': gar, garfish, garpike, billfish, Lepisosteus osseus
'396': lionfish
'397': puffer, pufferfish, blowfish, globefish
'398': abacus
'399': abaya
'400': academic gown, academic robe, judge's robe
'401': accordion, piano accordion, squeeze box
'402': acoustic guitar
'403': aircraft carrier, carrier, flattop, attack aircraft carrier
'404': airliner
'405': airship, dirigible
'406': altar
'407': ambulance
'408': amphibian, amphibious vehicle
'409': analog clock
'410': apiary, bee house
'411': apron
'412': ashcan, trash can, garbage can, wastebin, ash bin, ash-bin, ashbin,
dustbin, trash barrel, trash bin
'413': assault rifle, assault gun
'414': backpack, back pack, knapsack, packsack, rucksack, haversack
'415': bakery, bakeshop, bakehouse
'416': balance beam, beam
'417': balloon
'418': ballpoint, ballpoint pen, ballpen, Biro
'419': Band Aid
'420': banjo
'421': bannister, banister, balustrade, balusters, handrail
'422': barbell
'423': barber chair
'424': barbershop
'425': barn
'426': barometer
'427': barrel, cask
'428': barrow, garden cart, lawn cart, wheelbarrow
'429': baseball
'430': basketball
'431': bassinet
'432': bassoon
'433': bathing cap, swimming cap
'434': bath towel
'435': bathtub, bathing tub, bath, tub
'436': beach wagon, station wagon, wagon, estate car, beach waggon, station
waggon, waggon
'437': beacon, lighthouse, beacon light, pharos
'438': beaker
'439': bearskin, busby, shako
'440': beer bottle
'441': beer glass
'442': bell cote, bell cot
'443': bib
'444': bicycle-built-for-two, tandem bicycle, tandem
'445': bikini, two-piece
'446': binder, ring-binder
'447': binoculars, field glasses, opera glasses
'448': birdhouse
'449': boathouse
'450': bobsled, bobsleigh, bob
'451': bolo tie, bolo, bola tie, bola
'452': bonnet, poke bonnet
'453': bookcase
'454': bookshop, bookstore, bookstall
'455': bottlecap
'456': bow
'457': bow tie, bow-tie, bowtie
'458': brass, memorial tablet, plaque
'459': brassiere, bra, bandeau
'460': breakwater, groin, groyne, mole, bulwark, seawall, jetty
'461': breastplate, aegis, egis
'462': broom
'463': bucket, pail
'464': buckle
'465': bulletproof vest
'466': bullet train, bullet
'467': butcher shop, meat market
'468': cab, hack, taxi, taxicab
'469': caldron, cauldron
'470': candle, taper, wax light
'471': cannon
'472': canoe
'473': can opener, tin opener
'474': cardigan
'475': car mirror
'476': carousel, carrousel, merry-go-round, roundabout, whirligig
'477': carpenter's kit, tool kit
'478': carton
'479': car wheel
'480': cash machine, cash dispenser, automated teller machine, automatic
teller machine, automated teller, automatic teller, ATM
'481': cassette
'482': cassette player
'483': castle
'484': catamaran
'485': CD player
'486': cello, violoncello
'487': cellular telephone, cellular phone, cellphone, cell, mobile phone
'488': chain
'489': chainlink fence
'490': chain mail, ring mail, mail, chain armor, chain armour, ring armor,
ring armour
'491': chain saw, chainsaw
'492': chest
'493': chiffonier, commode
'494': chime, bell, gong
'495': china cabinet, china closet
'496': Christmas stocking
'497': church, church building
'498': cinema, movie theater, movie theatre, movie house, picture palace
'499': cleaver, meat cleaver, chopper
'500': cliff dwelling
'501': cloak
'502': clog, geta, patten, sabot
'503': cocktail shaker
'504': coffee mug
'505': coffeepot
'506': coil, spiral, volute, whorl, helix
'507': combination lock
'508': computer keyboard, keypad
'509': confectionery, confectionary, candy store
'510': container ship, containership, container vessel
'511': convertible
'512': corkscrew, bottle screw
'513': cornet, horn, trumpet, trump
'514': cowboy boot
'515': cowboy hat, ten-gallon hat
'516': cradle
'517': crane2
'518': crash helmet
'519': crate
'520': crib, cot
'521': Crock Pot
'522': croquet ball
'523': crutch
'524': cuirass
'525': dam, dike, dyke
'526': desk
'527': desktop computer
'528': dial telephone, dial phone
'529': diaper, nappy, napkin
'530': digital clock
'531': digital watch
'532': dining table, board
'533': dishrag, dishcloth
'534': dishwasher, dish washer, dishwashing machine
'535': disk brake, disc brake
'536': dock, dockage, docking facility
'537': dogsled, dog sled, dog sleigh
'538': dome
'539': doormat, welcome mat
'540': drilling platform, offshore rig
'541': drum, membranophone, tympan
'542': drumstick
'543': dumbbell
'544': Dutch oven
'545': electric fan, blower
'546': electric guitar
'547': electric locomotive
'548': entertainment center
'549': envelope
'550': espresso maker
'551': face powder
'552': feather boa, boa
'553': file, file cabinet, filing cabinet
'554': fireboat
'555': fire engine, fire truck
'556': fire screen, fireguard
'557': flagpole, flagstaff
'558': flute, transverse flute
'559': folding chair
'560': football helmet
'561': forklift
'562': fountain
'563': fountain pen
'564': four-poster
'565': freight car
'566': French horn, horn
'567': frying pan, frypan, skillet
'568': fur coat
'569': garbage truck, dustcart
'570': gasmask, respirator, gas helmet
'571': gas pump, gasoline pump, petrol pump, island dispenser
'572': goblet
'573': go-kart
'574': golf ball
'575': golfcart, golf cart
'576': gondola
'577': gong, tam-tam
'578': gown
'579': grand piano, grand
'580': greenhouse, nursery, glasshouse
'581': grille, radiator grille
'582': grocery store, grocery, food market, market
'583': guillotine
'584': hair slide
'585': hair spray
'586': half track
'587': hammer
'588': hamper
'589': hand blower, blow dryer, blow drier, hair dryer, hair drier
'590': hand-held computer, hand-held microcomputer
'591': handkerchief, hankie, hanky, hankey
'592': hard disc, hard disk, fixed disk
'593': harmonica, mouth organ, harp, mouth harp
'594': harp
'595': harvester, reaper
'596': hatchet
'597': holster
'598': home theater, home theatre
'599': honeycomb
'600': hook, claw
'601': hoopskirt, crinoline
'602': horizontal bar, high bar
'603': horse cart, horse-cart
'604': hourglass
'605': iPod
'606': iron, smoothing iron
'607': jack-o'-lantern
'608': jean, blue jean, denim
'609': jeep, landrover
'610': jersey, T-shirt, tee shirt
'611': jigsaw puzzle
'612': jinrikisha, ricksha, rickshaw
'613': joystick
'614': kimono
'615': knee pad
'616': knot
'617': lab coat, laboratory coat
'618': ladle
'619': lampshade, lamp shade
'620': laptop, laptop computer
'621': lawn mower, mower
'622': lens cap, lens cover
'623': letter opener, paper knife, paperknife
'624': library
'625': lifeboat
'626': lighter, light, igniter, ignitor
'627': limousine, limo
'628': liner, ocean liner
'629': lipstick, lip rouge
'630': Loafer
'631': lotion
'632': loudspeaker, speaker, speaker unit, loudspeaker system, speaker system
'633': loupe, jeweler's loupe
'634': lumbermill, sawmill
'635': magnetic compass
'636': mailbag, postbag
'637': mailbox, letter box
'638': maillot
'639': maillot, tank suit
'640': manhole cover
'641': maraca
'642': marimba, xylophone
'643': mask
'644': matchstick
'645': maypole
'646': maze, labyrinth
'647': measuring cup
'648': medicine chest, medicine cabinet
'649': megalith, megalithic structure
'650': microphone, mike
'651': microwave, microwave oven
'652': military uniform
'653': milk can
'654': minibus
'655': miniskirt, mini
'656': minivan
'657': missile
'658': mitten
'659': mixing bowl
'660': mobile home, manufactured home
'661': Model T
'662': modem
'663': monastery
'664': monitor
'665': moped
'666': mortar
'667': mortarboard
'668': mosque
'669': mosquito net
'670': motor scooter, scooter
'671': mountain bike, all-terrain bike, off-roader
'672': mountain tent
'673': mouse, computer mouse
'674': mousetrap
'675': moving van
'676': muzzle
'677': nail
'678': neck brace
'679': necklace
'680': nipple
'681': notebook, notebook computer
'682': obelisk
'683': oboe, hautboy, hautbois
'684': ocarina, sweet potato
'685': odometer, hodometer, mileometer, milometer
'686': oil filter
'687': organ, pipe organ
'688': oscilloscope, scope, cathode-ray oscilloscope, CRO
'689': overskirt
'690': oxcart
'691': oxygen mask
'692': packet
'693': paddle, boat paddle
'694': paddlewheel, paddle wheel
'695': padlock
'696': paintbrush
'697': pajama, pyjama, pj's, jammies
'698': palace
'699': panpipe, pandean pipe, syrinx
'700': paper towel
'701': parachute, chute
'702': parallel bars, bars
'703': park bench
'704': parking meter
'705': passenger car, coach, carriage
'706': patio, terrace
'707': pay-phone, pay-station
'708': pedestal, plinth, footstall
'709': pencil box, pencil case
'710': pencil sharpener
'711': perfume, essence
'712': Petri dish
'713': photocopier
'714': pick, plectrum, plectron
'715': pickelhaube
'716': picket fence, paling
'717': pickup, pickup truck
'718': pier
'719': piggy bank, penny bank
'720': pill bottle
'721': pillow
'722': ping-pong ball
'723': pinwheel
'724': pirate, pirate ship
'725': pitcher, ewer
'726': plane, carpenter's plane, woodworking plane
'727': planetarium
'728': plastic bag
'729': plate rack
'730': plow, plough
'731': plunger, plumber's helper
'732': Polaroid camera, Polaroid Land camera
'733': pole
'734': police van, police wagon, paddy wagon, patrol wagon, wagon, black
Maria
'735': poncho
'736': pool table, billiard table, snooker table
'737': pop bottle, soda bottle
'738': pot, flowerpot
'739': potter's wheel
'740': power drill
'741': prayer rug, prayer mat
'742': printer
'743': prison, prison house
'744': projectile, missile
'745': projector
'746': puck, hockey puck
'747': punching bag, punch bag, punching ball, punchball
'748': purse
'749': quill, quill pen
'750': quilt, comforter, comfort, puff
'751': racer, race car, racing car
'752': racket, racquet
'753': radiator
'754': radio, wireless
'755': radio telescope, radio reflector
'756': rain barrel
'757': recreational vehicle, RV, R.V.
'758': reel
'759': reflex camera
'760': refrigerator, icebox
'761': remote control, remote
'762': restaurant, eating house, eating place, eatery
'763': revolver, six-gun, six-shooter
'764': rifle
'765': rocking chair, rocker
'766': rotisserie
'767': rubber eraser, rubber, pencil eraser
'768': rugby ball
'769': rule, ruler
'770': running shoe
'771': safe
'772': safety pin
'773': saltshaker, salt shaker
'774': sandal
'775': sarong
'776': sax, saxophone
'777': scabbard
'778': scale, weighing machine
'779': school bus
'780': schooner
'781': scoreboard
'782': screen, CRT screen
'783': screw
'784': screwdriver
'785': seat belt, seatbelt
'786': sewing machine
'787': shield, buckler
'788': shoe shop, shoe-shop, shoe store
'789': shoji
'790': shopping basket
'791': shopping cart
'792': shovel
'793': shower cap
'794': shower curtain
'795': ski
'796': ski mask
'797': sleeping bag
'798': slide rule, slipstick
'799': sliding door
'800': slot, one-armed bandit
'801': snorkel
'802': snowmobile
'803': snowplow, snowplough
'804': soap dispenser
'805': soccer ball
'806': sock
'807': solar dish, solar collector, solar furnace
'808': sombrero
'809': soup bowl
'810': space bar
'811': space heater
'812': space shuttle
'813': spatula
'814': speedboat
'815': spider web, spider's web
'816': spindle
'817': sports car, sport car
'818': spotlight, spot
'819': stage
'820': steam locomotive
'821': steel arch bridge
'822': steel drum
'823': stethoscope
'824': stole
'825': stone wall
'826': stopwatch, stop watch
'827': stove
'828': strainer
'829': streetcar, tram, tramcar, trolley, trolley car
'830': stretcher
'831': studio couch, day bed
'832': stupa, tope
'833': submarine, pigboat, sub, U-boat
'834': suit, suit of clothes
'835': sundial
'836': sunglass
'837': sunglasses, dark glasses, shades
'838': sunscreen, sunblock, sun blocker
'839': suspension bridge
'840': swab, swob, mop
'841': sweatshirt
'842': swimming trunks, bathing trunks
'843': swing
'844': switch, electric switch, electrical switch
'845': syringe
'846': table lamp
'847': tank, army tank, armored combat vehicle, armoured combat vehicle
'848': tape player
'849': teapot
'850': teddy, teddy bear
'851': television, television system
'852': tennis ball
'853': thatch, thatched roof
'854': theater curtain, theatre curtain
'855': thimble
'856': thresher, thrasher, threshing machine
'857': throne
'858': tile roof
'859': toaster
'860': tobacco shop, tobacconist shop, tobacconist
'861': toilet seat
'862': torch
'863': totem pole
'864': tow truck, tow car, wrecker
'865': toyshop
'866': tractor
'867': trailer truck, tractor trailer, trucking rig, rig, articulated lorry,
semi
'868': tray
'869': trench coat
'870': tricycle, trike, velocipede
'871': trimaran
'872': tripod
'873': triumphal arch
'874': trolleybus, trolley coach, trackless trolley
'875': trombone
'876': tub, vat
'877': turnstile
'878': typewriter keyboard
'879': umbrella
'880': unicycle, monocycle
'881': upright, upright piano
'882': vacuum, vacuum cleaner
'883': vase
'884': vault
'885': velvet
'886': vending machine
'887': vestment
'888': viaduct
'889': violin, fiddle
'890': volleyball
'891': waffle iron
'892': wall clock
'893': wallet, billfold, notecase, pocketbook
'894': wardrobe, closet, press
'895': warplane, military plane
'896': washbasin, handbasin, washbowl, lavabo, wash-hand basin
'897': washer, automatic washer, washing machine
'898': water bottle
'899': water jug
'900': water tower
'901': whiskey jug
'902': whistle
'903': wig
'904': window screen
'905': window shade
'906': Windsor tie
'907': wine bottle
'908': wing
'909': wok
'910': wooden spoon
'911': wool, woolen, woollen
'912': worm fence, snake fence, snake-rail fence, Virginia fence
'913': wreck
'914': yawl
'915': yurt
'916': web site, website, internet site, site
'917': comic book
'918': crossword puzzle, crossword
'919': street sign
'920': traffic light, traffic signal, stoplight
'921': book jacket, dust cover, dust jacket, dust wrapper
'922': menu
'923': plate
'924': guacamole
'925': consomme
'926': hot pot, hotpot
'927': trifle
'928': ice cream, icecream
'929': ice lolly, lolly, lollipop, popsicle
'930': French loaf
'931': bagel, beigel
'932': pretzel
'933': cheeseburger
'934': hotdog, hot dog, red hot
'935': mashed potato
'936': head cabbage
'937': broccoli
'938': cauliflower
'939': zucchini, courgette
'940': spaghetti squash
'941': acorn squash
'942': butternut squash
'943': cucumber, cuke
'944': artichoke, globe artichoke
'945': bell pepper
'946': cardoon
'947': mushroom
'948': Granny Smith
'949': strawberry
'950': orange
'951': lemon
'952': fig
'953': pineapple, ananas
'954': banana
'955': jackfruit, jak, jack
'956': custard apple
'957': pomegranate
'958': hay
'959': carbonara
'960': chocolate sauce, chocolate syrup
'961': dough
'962': meat loaf, meatloaf
'963': pizza, pizza pie
'964': potpie
'965': burrito
'966': red wine
'967': espresso
'968': cup
'969': eggnog
'970': alp
'971': bubble
'972': cliff, drop, drop-off
'973': coral reef
'974': geyser
'975': lakeside, lakeshore
'976': promontory, headland, head, foreland
'977': sandbar, sand bar
'978': seashore, coast, seacoast, sea-coast
'979': valley, vale
'980': volcano
'981': ballplayer, baseball player
'982': groom, bridegroom
'983': scuba diver
'984': rapeseed
'985': daisy
'986': yellow lady's slipper, yellow lady-slipper, Cypripedium calceolus,
Cypripedium parviflorum
'987': corn
'988': acorn
'989': hip, rose hip, rosehip
'990': buckeye, horse chestnut, conker
'991': coral fungus
'992': agaric
'993': gyromitra
'994': stinkhorn, carrion fungus
'995': earthstar
'996': hen-of-the-woods, hen of the woods, Polyporus frondosus, Grifola
frondosa
'997': bolete
'998': ear, spike, capitulum
'999': toilet tissue, toilet paper, bathroom tissue
splits:
- name: train
num_bytes: 9919813
num_examples: 50889
download_size: 7593573012
dataset_size: 9919813
---
# Dataset Card for ImageNet-Sketch
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** https://github.com/HaohanWang/ImageNet-Sketch
- **Repository:** https://github.com/HaohanWang/ImageNet-Sketch
- **Paper:** [Learning Robust Global Representations by Penalizing Local Predictive Power](https://arxiv.org/abs/1905.13549v2)
- **Leaderboard:** https://github.com/HaohanWang/ImageNet-Sketch#imagenet-sketch-leaderboard
- **Point of Contact:** [Haohan Wang](mailto:haohanw@andrew.cmu.edu)
- **Size of downloaded dataset files:** 8.15 GB
### Dataset Summary
ImageNet-Sketch data set consists of 50000 images, 50 images for each of the 1000 ImageNet classes. We construct the data set with Google Image queries "sketch of __", where __ is the standard class name. We only search within the "black and white" color scheme. We initially query 100 images for every class, and then manually clean the pulled images by deleting the irrelevant images and images that are for similar but different classes. For some classes, there are less than 50 images after manually cleaning, and then we augment the data set by flipping and rotating the images.
The scripts used to conduct queries and clean images can be found in [the GitHub repository](https://github.com/HaohanWang/ImageNet-Sketch).
### Supported Tasks and Leaderboards
- `image_classification`: The goal of this task is to classify a given image into one of 1000 ImageNet classes. The leaderboard is available [here](https://github.com/HaohanWang/ImageNet-Sketch#imagenet-sketch-leaderboard).
The goal of the leaderboard is to evaluate the out-of-domain classification performance of vision models trained on ImageNet. The evaluation metrics used in the leaderboard are top-1 accuracy and top-5 accuracy.
### Languages
The class labels in the dataset are in English.
## Dataset Structure
### Data Instances
A sample from the training set is provided below:
```
{
'image': <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=400x530 at 0x7FB2EF5D4A90>,
'label': 320
}
```
### Data Fields
The data instances have the following fields:
- `image`: A `PIL.Image.Image` object containing the image. Note that when accessing the image column: `dataset[0]["image"]` the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the `"image"` column, *i.e.* `dataset[0]["image"]` should **always** be preferred over `dataset["image"][0]`.
- `label`: an `int` classification label.
The labels are indexed based on a sorted list of synset ids such as `n07565083` which we automatically map to original class names. The original dataset is divided into folders based on these synset ids. To get a mapping from original synset names, use the file [LOC_synset_mapping.txt](https://www.kaggle.com/competitions/imagenet-object-localization-challenge/data?select=LOC_synset_mapping.txt) available on Kaggle challenge page. You can also use `dataset_instance.features["label"].int2str` function to get the class for a particular label index.
<details>
<summary>
Click here to see the full list of ImageNet class label mapping:
</summary>
|id|Class|
|--|-----|
|0 | tench, Tinca tinca|
|1 | goldfish, Carassius auratus|
|2 | great white shark, white shark, man-eater, man-eating shark, Carcharodon carcharias|
|3 | tiger shark, Galeocerdo cuvieri|
|4 | hammerhead, hammerhead shark|
|5 | electric ray, crampfish, numbfish, torpedo|
|6 | stingray|
|7 | cock|
|8 | hen|
|9 | ostrich, Struthio camelus|
|10 | brambling, Fringilla montifringilla|
|11 | goldfinch, Carduelis carduelis|
|12 | house finch, linnet, Carpodacus mexicanus|
|13 | junco, snowbird|
|14 | indigo bunting, indigo finch, indigo bird, Passerina cyanea|
|15 | robin, American robin, Turdus migratorius|
|16 | bulbul|
|17 | jay|
|18 | magpie|
|19 | chickadee|
|20 | water ouzel, dipper|
|21 | kite|
|22 | bald eagle, American eagle, Haliaeetus leucocephalus|
|23 | vulture|
|24 | great grey owl, great gray owl, Strix nebulosa|
|25 | European fire salamander, Salamandra salamandra|
|26 | common newt, Triturus vulgaris|
|27 | eft|
|28 | spotted salamander, Ambystoma maculatum|
|29 | axolotl, mud puppy, Ambystoma mexicanum|
|30 | bullfrog, Rana catesbeiana|
|31 | tree frog, tree-frog|
|32 | tailed frog, bell toad, ribbed toad, tailed toad, Ascaphus trui|
|33 | loggerhead, loggerhead turtle, Caretta caretta|
|34 | leatherback turtle, leatherback, leathery turtle, Dermochelys coriacea|
|35 | mud turtle|
|36 | terrapin|
|37 | box turtle, box tortoise|
|38 | banded gecko|
|39 | common iguana, iguana, Iguana iguana|
|40 | American chameleon, anole, Anolis carolinensis|
|41 | whiptail, whiptail lizard|
|42 | agama|
|43 | frilled lizard, Chlamydosaurus kingi|
|44 | alligator lizard|
|45 | Gila monster, Heloderma suspectum|
|46 | green lizard, Lacerta viridis|
|47 | African chameleon, Chamaeleo chamaeleon|
|48 | Komodo dragon, Komodo lizard, dragon lizard, giant lizard, Varanus komodoensis|
|49 | African crocodile, Nile crocodile, Crocodylus niloticus|
|50 | American alligator, Alligator mississipiensis|
|51 | triceratops|
|52 | thunder snake, worm snake, Carphophis amoenus|
|53 | ringneck snake, ring-necked snake, ring snake|
|54 | hognose snake, puff adder, sand viper|
|55 | green snake, grass snake|
|56 | king snake, kingsnake|
|57 | garter snake, grass snake|
|58 | water snake|
|59 | vine snake|
|60 | night snake, Hypsiglena torquata|
|61 | boa constrictor, Constrictor constrictor|
|62 | rock python, rock snake, Python sebae|
|63 | Indian cobra, Naja naja|
|64 | green mamba|
|65 | sea snake|
|66 | horned viper, cerastes, sand viper, horned asp, Cerastes cornutus|
|67 | diamondback, diamondback rattlesnake, Crotalus adamanteus|
|68 | sidewinder, horned rattlesnake, Crotalus cerastes|
|69 | trilobite|
|70 | harvestman, daddy longlegs, Phalangium opilio|
|71 | scorpion|
|72 | black and gold garden spider, Argiope aurantia|
|73 | barn spider, Araneus cavaticus|
|74 | garden spider, Aranea diademata|
|75 | black widow, Latrodectus mactans|
|76 | tarantula|
|77 | wolf spider, hunting spider|
|78 | tick|
|79 | centipede|
|80 | black grouse|
|81 | ptarmigan|
|82 | ruffed grouse, partridge, Bonasa umbellus|
|83 | prairie chicken, prairie grouse, prairie fowl|
|84 | peacock|
|85 | quail|
|86 | partridge|
|87 | African grey, African gray, Psittacus erithacus|
|88 | macaw|
|89 | sulphur-crested cockatoo, Kakatoe galerita, Cacatua galerita|
|90 | lorikeet|
|91 | coucal|
|92 | bee eater|
|93 | hornbill|
|94 | hummingbird|
|95 | jacamar|
|96 | toucan|
|97 | drake|
|98 | red-breasted merganser, Mergus serrator|
|99 | goose|
|100 | black swan, Cygnus atratus|
|101 | tusker|
|102 | echidna, spiny anteater, anteater|
|103 | platypus, duckbill, duckbilled platypus, duck-billed platypus, Ornithorhynchus anatinus|
|104 | wallaby, brush kangaroo|
|105 | koala, koala bear, kangaroo bear, native bear, Phascolarctos cinereus|
|106 | wombat|
|107 | jellyfish|
|108 | sea anemone, anemone|
|109 | brain coral|
|110 | flatworm, platyhelminth|
|111 | nematode, nematode worm, roundworm|
|112 | conch|
|113 | snail|
|114 | slug|
|115 | sea slug, nudibranch|
|116 | chiton, coat-of-mail shell, sea cradle, polyplacophore|
|117 | chambered nautilus, pearly nautilus, nautilus|
|118 | Dungeness crab, Cancer magister|
|119 | rock crab, Cancer irroratus|
|120 | fiddler crab|
|121 | king crab, Alaska crab, Alaskan king crab, Alaska king crab, Paralithodes camtschatica|
|122 | American lobster, Northern lobster, Maine lobster, Homarus americanus|
|123 | spiny lobster, langouste, rock lobster, crawfish, crayfish, sea crawfish|
|124 | crayfish, crawfish, crawdad, crawdaddy|
|125 | hermit crab|
|126 | isopod|
|127 | white stork, Ciconia ciconia|
|128 | black stork, Ciconia nigra|
|129 | spoonbill|
|130 | flamingo|
|131 | little blue heron, Egretta caerulea|
|132 | American egret, great white heron, Egretta albus|
|133 | bittern|
|134 | crane|
|135 | limpkin, Aramus pictus|
|136 | European gallinule, Porphyrio porphyrio|
|137 | American coot, marsh hen, mud hen, water hen, Fulica americana|
|138 | bustard|
|139 | ruddy turnstone, Arenaria interpres|
|140 | red-backed sandpiper, dunlin, Erolia alpina|
|141 | redshank, Tringa totanus|
|142 | dowitcher|
|143 | oystercatcher, oyster catcher|
|144 | pelican|
|145 | king penguin, Aptenodytes patagonica|
|146 | albatross, mollymawk|
|147 | grey whale, gray whale, devilfish, Eschrichtius gibbosus, Eschrichtius robustus|
|148 | killer whale, killer, orca, grampus, sea wolf, Orcinus orca|
|149 | dugong, Dugong dugon|
|150 | sea lion|
|151 | Chihuahua|
|152 | Japanese spaniel|
|153 | Maltese dog, Maltese terrier, Maltese|
|154 | Pekinese, Pekingese, Peke|
|155 | Shih-Tzu|
|156 | Blenheim spaniel|
|157 | papillon|
|158 | toy terrier|
|159 | Rhodesian ridgeback|
|160 | Afghan hound, Afghan|
|161 | basset, basset hound|
|162 | beagle|
|163 | bloodhound, sleuthhound|
|164 | bluetick|
|165 | black-and-tan coonhound|
|166 | Walker hound, Walker foxhound|
|167 | English foxhound|
|168 | redbone|
|169 | borzoi, Russian wolfhound|
|170 | Irish wolfhound|
|171 | Italian greyhound|
|172 | whippet|
|173 | Ibizan hound, Ibizan Podenco|
|174 | Norwegian elkhound, elkhound|
|175 | otterhound, otter hound|
|176 | Saluki, gazelle hound|
|177 | Scottish deerhound, deerhound|
|178 | Weimaraner|
|179 | Staffordshire bullterrier, Staffordshire bull terrier|
|180 | American Staffordshire terrier, Staffordshire terrier, American pit bull terrier, pit bull terrier|
|181 | Bedlington terrier|
|182 | Border terrier|
|183 | Kerry blue terrier|
|184 | Irish terrier|
|185 | Norfolk terrier|
|186 | Norwich terrier|
|187 | Yorkshire terrier|
|188 | wire-haired fox terrier|
|189 | Lakeland terrier|
|190 | Sealyham terrier, Sealyham|
|191 | Airedale, Airedale terrier|
|192 | cairn, cairn terrier|
|193 | Australian terrier|
|194 | Dandie Dinmont, Dandie Dinmont terrier|
|195 | Boston bull, Boston terrier|
|196 | miniature schnauzer|
|197 | giant schnauzer|
|198 | standard schnauzer|
|199 | Scotch terrier, Scottish terrier, Scottie|
|200 | Tibetan terrier, chrysanthemum dog|
|201 | silky terrier, Sydney silky|
|202 | soft-coated wheaten terrier|
|203 | West Highland white terrier|
|204 | Lhasa, Lhasa apso|
|205 | flat-coated retriever|
|206 | curly-coated retriever|
|207 | golden retriever|
|208 | Labrador retriever|
|209 | Chesapeake Bay retriever|
|210 | German short-haired pointer|
|211 | vizsla, Hungarian pointer|
|212 | English setter|
|213 | Irish setter, red setter|
|214 | Gordon setter|
|215 | Brittany spaniel|
|216 | clumber, clumber spaniel|
|217 | English springer, English springer spaniel|
|218 | Welsh springer spaniel|
|219 | cocker spaniel, English cocker spaniel, cocker|
|220 | Sussex spaniel|
|221 | Irish water spaniel|
|222 | kuvasz|
|223 | schipperke|
|224 | groenendael|
|225 | malinois|
|226 | briard|
|227 | kelpie|
|228 | komondor|
|229 | Old English sheepdog, bobtail|
|230 | Shetland sheepdog, Shetland sheep dog, Shetland|
|231 | collie|
|232 | Border collie|
|233 | Bouvier des Flandres, Bouviers des Flandres|
|234 | Rottweiler|
|235 | German shepherd, German shepherd dog, German police dog, alsatian|
|236 | Doberman, Doberman pinscher|
|237 | miniature pinscher|
|238 | Greater Swiss Mountain dog|
|239 | Bernese mountain dog|
|240 | Appenzeller|
|241 | EntleBucher|
|242 | boxer|
|243 | bull mastiff|
|244 | Tibetan mastiff|
|245 | French bulldog|
|246 | Great Dane|
|247 | Saint Bernard, St Bernard|
|248 | Eskimo dog, husky|
|249 | malamute, malemute, Alaskan malamute|
|250 | Siberian husky|
|251 | dalmatian, coach dog, carriage dog|
|252 | affenpinscher, monkey pinscher, monkey dog|
|253 | basenji|
|254 | pug, pug-dog|
|255 | Leonberg|
|256 | Newfoundland, Newfoundland dog|
|257 | Great Pyrenees|
|258 | Samoyed, Samoyede|
|259 | Pomeranian|
|260 | chow, chow chow|
|261 | keeshond|
|262 | Brabancon griffon|
|263 | Pembroke, Pembroke Welsh corgi|
|264 | Cardigan, Cardigan Welsh corgi|
|265 | toy poodle|
|266 | miniature poodle|
|267 | standard poodle|
|268 | Mexican hairless|
|269 | timber wolf, grey wolf, gray wolf, Canis lupus|
|270 | white wolf, Arctic wolf, Canis lupus tundrarum|
|271 | red wolf, maned wolf, Canis rufus, Canis niger|
|272 | coyote, prairie wolf, brush wolf, Canis latrans|
|273 | dingo, warrigal, warragal, Canis dingo|
|274 | dhole, Cuon alpinus|
|275 | African hunting dog, hyena dog, Cape hunting dog, Lycaon pictus|
|276 | hyena, hyaena|
|277 | red fox, Vulpes vulpes|
|278 | kit fox, Vulpes macrotis|
|279 | Arctic fox, white fox, Alopex lagopus|
|280 | grey fox, gray fox, Urocyon cinereoargenteus|
|281 | tabby, tabby cat|
|282 | tiger cat|
|283 | Persian cat|
|284 | Siamese cat, Siamese|
|285 | Egyptian cat|
|286 | cougar, puma, catamount, mountain lion, painter, panther, Felis concolor|
|287 | lynx, catamount|
|288 | leopard, Panthera pardus|
|289 | snow leopard, ounce, Panthera uncia|
|290 | jaguar, panther, Panthera onca, Felis onca|
|291 | lion, king of beasts, Panthera leo|
|292 | tiger, Panthera tigris|
|293 | cheetah, chetah, Acinonyx jubatus|
|294 | brown bear, bruin, Ursus arctos|
|295 | American black bear, black bear, Ursus americanus, Euarctos americanus|
|296 | ice bear, polar bear, Ursus Maritimus, Thalarctos maritimus|
|297 | sloth bear, Melursus ursinus, Ursus ursinus|
|298 | mongoose|
|299 | meerkat, mierkat|
|300 | tiger beetle|
|301 | ladybug, ladybeetle, lady beetle, ladybird, ladybird beetle|
|302 | ground beetle, carabid beetle|
|303 | long-horned beetle, longicorn, longicorn beetle|
|304 | leaf beetle, chrysomelid|
|305 | dung beetle|
|306 | rhinoceros beetle|
|307 | weevil|
|308 | fly|
|309 | bee|
|310 | ant, emmet, pismire|
|311 | grasshopper, hopper|
|312 | cricket|
|313 | walking stick, walkingstick, stick insect|
|314 | cockroach, roach|
|315 | mantis, mantid|
|316 | cicada, cicala|
|317 | leafhopper|
|318 | lacewing, lacewing fly|
|319 | dragonfly, darning needle, devil's darning needle, sewing needle, snake feeder, snake doctor, mosquito hawk, skeeter hawk|
|320 | damselfly|
|321 | admiral|
|322 | ringlet, ringlet butterfly|
|323 | monarch, monarch butterfly, milkweed butterfly, Danaus plexippus|
|324 | cabbage butterfly|
|325 | sulphur butterfly, sulfur butterfly|
|326 | lycaenid, lycaenid butterfly|
|327 | starfish, sea star|
|328 | sea urchin|
|329 | sea cucumber, holothurian|
|330 | wood rabbit, cottontail, cottontail rabbit|
|331 | hare|
|332 | Angora, Angora rabbit|
|333 | hamster|
|334 | porcupine, hedgehog|
|335 | fox squirrel, eastern fox squirrel, Sciurus niger|
|336 | marmot|
|337 | beaver|
|338 | guinea pig, Cavia cobaya|
|339 | sorrel|
|340 | zebra|
|341 | hog, pig, grunter, squealer, Sus scrofa|
|342 | wild boar, boar, Sus scrofa|
|343 | warthog|
|344 | hippopotamus, hippo, river horse, Hippopotamus amphibius|
|345 | ox|
|346 | water buffalo, water ox, Asiatic buffalo, Bubalus bubalis|
|347 | bison|
|348 | ram, tup|
|349 | bighorn, bighorn sheep, cimarron, Rocky Mountain bighorn, Rocky Mountain sheep, Ovis canadensis|
|350 | ibex, Capra ibex|
|351 | hartebeest|
|352 | impala, Aepyceros melampus|
|353 | gazelle|
|354 | Arabian camel, dromedary, Camelus dromedarius|
|355 | llama|
|356 | weasel|
|357 | mink|
|358 | polecat, fitch, foulmart, foumart, Mustela putorius|
|359 | black-footed ferret, ferret, Mustela nigripes|
|360 | otter|
|361 | skunk, polecat, wood pussy|
|362 | badger|
|363 | armadillo|
|364 | three-toed sloth, ai, Bradypus tridactylus|
|365 | orangutan, orang, orangutang, Pongo pygmaeus|
|366 | gorilla, Gorilla gorilla|
|367 | chimpanzee, chimp, Pan troglodytes|
|368 | gibbon, Hylobates lar|
|369 | siamang, Hylobates syndactylus, Symphalangus syndactylus|
|370 | guenon, guenon monkey|
|371 | patas, hussar monkey, Erythrocebus patas|
|372 | baboon|
|373 | macaque|
|374 | langur|
|375 | colobus, colobus monkey|
|376 | proboscis monkey, Nasalis larvatus|
|377 | marmoset|
|378 | capuchin, ringtail, Cebus capucinus|
|379 | howler monkey, howler|
|380 | titi, titi monkey|
|381 | spider monkey, Ateles geoffroyi|
|382 | squirrel monkey, Saimiri sciureus|
|383 | Madagascar cat, ring-tailed lemur, Lemur catta|
|384 | indri, indris, Indri indri, Indri brevicaudatus|
|385 | Indian elephant, Elephas maximus|
|386 | African elephant, Loxodonta africana|
|387 | lesser panda, red panda, panda, bear cat, cat bear, Ailurus fulgens|
|388 | giant panda, panda, panda bear, coon bear, Ailuropoda melanoleuca|
|389 | barracouta, snoek|
|390 | eel|
|391 | coho, cohoe, coho salmon, blue jack, silver salmon, Oncorhynchus kisutch|
|392 | rock beauty, Holocanthus tricolor|
|393 | anemone fish|
|394 | sturgeon|
|395 | gar, garfish, garpike, billfish, Lepisosteus osseus|
|396 | lionfish|
|397 | puffer, pufferfish, blowfish, globefish|
|398 | abacus|
|399 | abaya|
|400 | academic gown, academic robe, judge's robe|
|401 | accordion, piano accordion, squeeze box|
|402 | acoustic guitar|
|403 | aircraft carrier, carrier, flattop, attack aircraft carrier|
|404 | airliner|
|405 | airship, dirigible|
|406 | altar|
|407 | ambulance|
|408 | amphibian, amphibious vehicle|
|409 | analog clock|
|410 | apiary, bee house|
|411 | apron|
|412 | ashcan, trash can, garbage can, wastebin, ash bin, ash-bin, ashbin, dustbin, trash barrel, trash bin|
|413 | assault rifle, assault gun|
|414 | backpack, back pack, knapsack, packsack, rucksack, haversack|
|415 | bakery, bakeshop, bakehouse|
|416 | balance beam, beam|
|417 | balloon|
|418 | ballpoint, ballpoint pen, ballpen, Biro|
|419 | Band Aid|
|420 | banjo|
|421 | bannister, banister, balustrade, balusters, handrail|
|422 | barbell|
|423 | barber chair|
|424 | barbershop|
|425 | barn|
|426 | barometer|
|427 | barrel, cask|
|428 | barrow, garden cart, lawn cart, wheelbarrow|
|429 | baseball|
|430 | basketball|
|431 | bassinet|
|432 | bassoon|
|433 | bathing cap, swimming cap|
|434 | bath towel|
|435 | bathtub, bathing tub, bath, tub|
|436 | beach wagon, station wagon, wagon, estate car, beach waggon, station waggon, waggon|
|437 | beacon, lighthouse, beacon light, pharos|
|438 | beaker|
|439 | bearskin, busby, shako|
|440 | beer bottle|
|441 | beer glass|
|442 | bell cote, bell cot|
|443 | bib|
|444 | bicycle-built-for-two, tandem bicycle, tandem|
|445 | bikini, two-piece|
|446 | binder, ring-binder|
|447 | binoculars, field glasses, opera glasses|
|448 | birdhouse|
|449 | boathouse|
|450 | bobsled, bobsleigh, bob|
|451 | bolo tie, bolo, bola tie, bola|
|452 | bonnet, poke bonnet|
|453 | bookcase|
|454 | bookshop, bookstore, bookstall|
|455 | bottlecap|
|456 | bow|
|457 | bow tie, bow-tie, bowtie|
|458 | brass, memorial tablet, plaque|
|459 | brassiere, bra, bandeau|
|460 | breakwater, groin, groyne, mole, bulwark, seawall, jetty|
|461 | breastplate, aegis, egis|
|462 | broom|
|463 | bucket, pail|
|464 | buckle|
|465 | bulletproof vest|
|466 | bullet train, bullet|
|467 | butcher shop, meat market|
|468 | cab, hack, taxi, taxicab|
|469 | caldron, cauldron|
|470 | candle, taper, wax light|
|471 | cannon|
|472 | canoe|
|473 | can opener, tin opener|
|474 | cardigan|
|475 | car mirror|
|476 | carousel, carrousel, merry-go-round, roundabout, whirligig|
|477 | carpenter's kit, tool kit|
|478 | carton|
|479 | car wheel|
|480 | cash machine, cash dispenser, automated teller machine, automatic teller machine, automated teller, automatic teller, ATM|
|481 | cassette|
|482 | cassette player|
|483 | castle|
|484 | catamaran|
|485 | CD player|
|486 | cello, violoncello|
|487 | cellular telephone, cellular phone, cellphone, cell, mobile phone|
|488 | chain|
|489 | chainlink fence|
|490 | chain mail, ring mail, mail, chain armor, chain armour, ring armor, ring armour|
|491 | chain saw, chainsaw|
|492 | chest|
|493 | chiffonier, commode|
|494 | chime, bell, gong|
|495 | china cabinet, china closet|
|496 | Christmas stocking|
|497 | church, church building|
|498 | cinema, movie theater, movie theatre, movie house, picture palace|
|499 | cleaver, meat cleaver, chopper|
|500 | cliff dwelling|
|501 | cloak|
|502 | clog, geta, patten, sabot|
|503 | cocktail shaker|
|504 | coffee mug|
|505 | coffeepot|
|506 | coil, spiral, volute, whorl, helix|
|507 | combination lock|
|508 | computer keyboard, keypad|
|509 | confectionery, confectionary, candy store|
|510 | container ship, containership, container vessel|
|511 | convertible|
|512 | corkscrew, bottle screw|
|513 | cornet, horn, trumpet, trump|
|514 | cowboy boot|
|515 | cowboy hat, ten-gallon hat|
|516 | cradle|
|517 | crane_1|
|518 | crash helmet|
|519 | crate|
|520 | crib, cot|
|521 | Crock Pot|
|522 | croquet ball|
|523 | crutch|
|524 | cuirass|
|525 | dam, dike, dyke|
|526 | desk|
|527 | desktop computer|
|528 | dial telephone, dial phone|
|529 | diaper, nappy, napkin|
|530 | digital clock|
|531 | digital watch|
|532 | dining table, board|
|533 | dishrag, dishcloth|
|534 | dishwasher, dish washer, dishwashing machine|
|535 | disk brake, disc brake|
|536 | dock, dockage, docking facility|
|537 | dogsled, dog sled, dog sleigh|
|538 | dome|
|539 | doormat, welcome mat|
|540 | drilling platform, offshore rig|
|541 | drum, membranophone, tympan|
|542 | drumstick|
|543 | dumbbell|
|544 | Dutch oven|
|545 | electric fan, blower|
|546 | electric guitar|
|547 | electric locomotive|
|548 | entertainment center|
|549 | envelope|
|550 | espresso maker|
|551 | face powder|
|552 | feather boa, boa|
|553 | file, file cabinet, filing cabinet|
|554 | fireboat|
|555 | fire engine, fire truck|
|556 | fire screen, fireguard|
|557 | flagpole, flagstaff|
|558 | flute, transverse flute|
|559 | folding chair|
|560 | football helmet|
|561 | forklift|
|562 | fountain|
|563 | fountain pen|
|564 | four-poster|
|565 | freight car|
|566 | French horn, horn|
|567 | frying pan, frypan, skillet|
|568 | fur coat|
|569 | garbage truck, dustcart|
|570 | gasmask, respirator, gas helmet|
|571 | gas pump, gasoline pump, petrol pump, island dispenser|
|572 | goblet|
|573 | go-kart|
|574 | golf ball|
|575 | golfcart, golf cart|
|576 | gondola|
|577 | gong, tam-tam|
|578 | gown|
|579 | grand piano, grand|
|580 | greenhouse, nursery, glasshouse|
|581 | grille, radiator grille|
|582 | grocery store, grocery, food market, market|
|583 | guillotine|
|584 | hair slide|
|585 | hair spray|
|586 | half track|
|587 | hammer|
|588 | hamper|
|589 | hand blower, blow dryer, blow drier, hair dryer, hair drier|
|590 | hand-held computer, hand-held microcomputer|
|591 | handkerchief, hankie, hanky, hankey|
|592 | hard disc, hard disk, fixed disk|
|593 | harmonica, mouth organ, harp, mouth harp|
|594 | harp|
|595 | harvester, reaper|
|596 | hatchet|
|597 | holster|
|598 | home theater, home theatre|
|599 | honeycomb|
|600 | hook, claw|
|601 | hoopskirt, crinoline|
|602 | horizontal bar, high bar|
|603 | horse cart, horse-cart|
|604 | hourglass|
|605 | iPod|
|606 | iron, smoothing iron|
|607 | jack-o'-lantern|
|608 | jean, blue jean, denim|
|609 | jeep, landrover|
|610 | jersey, T-shirt, tee shirt|
|611 | jigsaw puzzle|
|612 | jinrikisha, ricksha, rickshaw|
|613 | joystick|
|614 | kimono|
|615 | knee pad|
|616 | knot|
|617 | lab coat, laboratory coat|
|618 | ladle|
|619 | lampshade, lamp shade|
|620 | laptop, laptop computer|
|621 | lawn mower, mower|
|622 | lens cap, lens cover|
|623 | letter opener, paper knife, paperknife|
|624 | library|
|625 | lifeboat|
|626 | lighter, light, igniter, ignitor|
|627 | limousine, limo|
|628 | liner, ocean liner|
|629 | lipstick, lip rouge|
|630 | Loafer|
|631 | lotion|
|632 | loudspeaker, speaker, speaker unit, loudspeaker system, speaker system|
|633 | loupe, jeweler's loupe|
|634 | lumbermill, sawmill|
|635 | magnetic compass|
|636 | mailbag, postbag|
|637 | mailbox, letter box|
|638 | maillot|
|639 | maillot, tank suit|
|640 | manhole cover|
|641 | maraca|
|642 | marimba, xylophone|
|643 | mask|
|644 | matchstick|
|645 | maypole|
|646 | maze, labyrinth|
|647 | measuring cup|
|648 | medicine chest, medicine cabinet|
|649 | megalith, megalithic structure|
|650 | microphone, mike|
|651 | microwave, microwave oven|
|652 | military uniform|
|653 | milk can|
|654 | minibus|
|655 | miniskirt, mini|
|656 | minivan|
|657 | missile|
|658 | mitten|
|659 | mixing bowl|
|660 | mobile home, manufactured home|
|661 | Model T|
|662 | modem|
|663 | monastery|
|664 | monitor|
|665 | moped|
|666 | mortar|
|667 | mortarboard|
|668 | mosque|
|669 | mosquito net|
|670 | motor scooter, scooter|
|671 | mountain bike, all-terrain bike, off-roader|
|672 | mountain tent|
|673 | mouse, computer mouse|
|674 | mousetrap|
|675 | moving van|
|676 | muzzle|
|677 | nail|
|678 | neck brace|
|679 | necklace|
|680 | nipple|
|681 | notebook, notebook computer|
|682 | obelisk|
|683 | oboe, hautboy, hautbois|
|684 | ocarina, sweet potato|
|685 | odometer, hodometer, mileometer, milometer|
|686 | oil filter|
|687 | organ, pipe organ|
|688 | oscilloscope, scope, cathode-ray oscilloscope, CRO|
|689 | overskirt|
|690 | oxcart|
|691 | oxygen mask|
|692 | packet|
|693 | paddle, boat paddle|
|694 | paddlewheel, paddle wheel|
|695 | padlock|
|696 | paintbrush|
|697 | pajama, pyjama, pj's, jammies|
|698 | palace|
|699 | panpipe, pandean pipe, syrinx|
|700 | paper towel|
|701 | parachute, chute|
|702 | parallel bars, bars|
|703 | park bench|
|704 | parking meter|
|705 | passenger car, coach, carriage|
|706 | patio, terrace|
|707 | pay-phone, pay-station|
|708 | pedestal, plinth, footstall|
|709 | pencil box, pencil case|
|710 | pencil sharpener|
|711 | perfume, essence|
|712 | Petri dish|
|713 | photocopier|
|714 | pick, plectrum, plectron|
|715 | pickelhaube|
|716 | picket fence, paling|
|717 | pickup, pickup truck|
|718 | pier|
|719 | piggy bank, penny bank|
|720 | pill bottle|
|721 | pillow|
|722 | ping-pong ball|
|723 | pinwheel|
|724 | pirate, pirate ship|
|725 | pitcher, ewer|
|726 | plane, carpenter's plane, woodworking plane|
|727 | planetarium|
|728 | plastic bag|
|729 | plate rack|
|730 | plow, plough|
|731 | plunger, plumber's helper|
|732 | Polaroid camera, Polaroid Land camera|
|733 | pole|
|734 | police van, police wagon, paddy wagon, patrol wagon, wagon, black Maria|
|735 | poncho|
|736 | pool table, billiard table, snooker table|
|737 | pop bottle, soda bottle|
|738 | pot, flowerpot|
|739 | potter's wheel|
|740 | power drill|
|741 | prayer rug, prayer mat|
|742 | printer|
|743 | prison, prison house|
|744 | projectile, missile|
|745 | projector|
|746 | puck, hockey puck|
|747 | punching bag, punch bag, punching ball, punchball|
|748 | purse|
|749 | quill, quill pen|
|750 | quilt, comforter, comfort, puff|
|751 | racer, race car, racing car|
|752 | racket, racquet|
|753 | radiator|
|754 | radio, wireless|
|755 | radio telescope, radio reflector|
|756 | rain barrel|
|757 | recreational vehicle, RV, R.V.|
|758 | reel|
|759 | reflex camera|
|760 | refrigerator, icebox|
|761 | remote control, remote|
|762 | restaurant, eating house, eating place, eatery|
|763 | revolver, six-gun, six-shooter|
|764 | rifle|
|765 | rocking chair, rocker|
|766 | rotisserie|
|767 | rubber eraser, rubber, pencil eraser|
|768 | rugby ball|
|769 | rule, ruler|
|770 | running shoe|
|771 | safe|
|772 | safety pin|
|773 | saltshaker, salt shaker|
|774 | sandal|
|775 | sarong|
|776 | sax, saxophone|
|777 | scabbard|
|778 | scale, weighing machine|
|779 | school bus|
|780 | schooner|
|781 | scoreboard|
|782 | screen, CRT screen|
|783 | screw|
|784 | screwdriver|
|785 | seat belt, seatbelt|
|786 | sewing machine|
|787 | shield, buckler|
|788 | shoe shop, shoe-shop, shoe store|
|789 | shoji|
|790 | shopping basket|
|791 | shopping cart|
|792 | shovel|
|793 | shower cap|
|794 | shower curtain|
|795 | ski|
|796 | ski mask|
|797 | sleeping bag|
|798 | slide rule, slipstick|
|799 | sliding door|
|800 | slot, one-armed bandit|
|801 | snorkel|
|802 | snowmobile|
|803 | snowplow, snowplough|
|804 | soap dispenser|
|805 | soccer ball|
|806 | sock|
|807 | solar dish, solar collector, solar furnace|
|808 | sombrero|
|809 | soup bowl|
|810 | space bar|
|811 | space heater|
|812 | space shuttle|
|813 | spatula|
|814 | speedboat|
|815 | spider web, spider's web|
|816 | spindle|
|817 | sports car, sport car|
|818 | spotlight, spot|
|819 | stage|
|820 | steam locomotive|
|821 | steel arch bridge|
|822 | steel drum|
|823 | stethoscope|
|824 | stole|
|825 | stone wall|
|826 | stopwatch, stop watch|
|827 | stove|
|828 | strainer|
|829 | streetcar, tram, tramcar, trolley, trolley car|
|830 | stretcher|
|831 | studio couch, day bed|
|832 | stupa, tope|
|833 | submarine, pigboat, sub, U-boat|
|834 | suit, suit of clothes|
|835 | sundial|
|836 | sunglass|
|837 | sunglasses, dark glasses, shades|
|838 | sunscreen, sunblock, sun blocker|
|839 | suspension bridge|
|840 | swab, swob, mop|
|841 | sweatshirt|
|842 | swimming trunks, bathing trunks|
|843 | swing|
|844 | switch, electric switch, electrical switch|
|845 | syringe|
|846 | table lamp|
|847 | tank, army tank, armored combat vehicle, armoured combat vehicle|
|848 | tape player|
|849 | teapot|
|850 | teddy, teddy bear|
|851 | television, television system|
|852 | tennis ball|
|853 | thatch, thatched roof|
|854 | theater curtain, theatre curtain|
|855 | thimble|
|856 | thresher, thrasher, threshing machine|
|857 | throne|
|858 | tile roof|
|859 | toaster|
|860 | tobacco shop, tobacconist shop, tobacconist|
|861 | toilet seat|
|862 | torch|
|863 | totem pole|
|864 | tow truck, tow car, wrecker|
|865 | toyshop|
|866 | tractor|
|867 | trailer truck, tractor trailer, trucking rig, rig, articulated lorry, semi|
|868 | tray|
|869 | trench coat|
|870 | tricycle, trike, velocipede|
|871 | trimaran|
|872 | tripod|
|873 | triumphal arch|
|874 | trolleybus, trolley coach, trackless trolley|
|875 | trombone|
|876 | tub, vat|
|877 | turnstile|
|878 | typewriter keyboard|
|879 | umbrella|
|880 | unicycle, monocycle|
|881 | upright, upright piano|
|882 | vacuum, vacuum cleaner|
|883 | vase|
|884 | vault|
|885 | velvet|
|886 | vending machine|
|887 | vestment|
|888 | viaduct|
|889 | violin, fiddle|
|890 | volleyball|
|891 | waffle iron|
|892 | wall clock|
|893 | wallet, billfold, notecase, pocketbook|
|894 | wardrobe, closet, press|
|895 | warplane, military plane|
|896 | washbasin, handbasin, washbowl, lavabo, wash-hand basin|
|897 | washer, automatic washer, washing machine|
|898 | water bottle|
|899 | water jug|
|900 | water tower|
|901 | whiskey jug|
|902 | whistle|
|903 | wig|
|904 | window screen|
|905 | window shade|
|906 | Windsor tie|
|907 | wine bottle|
|908 | wing|
|909 | wok|
|910 | wooden spoon|
|911 | wool, woolen, woollen|
|912 | worm fence, snake fence, snake-rail fence, Virginia fence|
|913 | wreck|
|914 | yawl|
|915 | yurt|
|916 | web site, website, internet site, site|
|917 | comic book|
|918 | crossword puzzle, crossword|
|919 | street sign|
|920 | traffic light, traffic signal, stoplight|
|921 | book jacket, dust cover, dust jacket, dust wrapper|
|922 | menu|
|923 | plate|
|924 | guacamole|
|925 | consomme|
|926 | hot pot, hotpot|
|927 | trifle|
|928 | ice cream, icecream|
|929 | ice lolly, lolly, lollipop, popsicle|
|930 | French loaf|
|931 | bagel, beigel|
|932 | pretzel|
|933 | cheeseburger|
|934 | hotdog, hot dog, red hot|
|935 | mashed potato|
|936 | head cabbage|
|937 | broccoli|
|938 | cauliflower|
|939 | zucchini, courgette|
|940 | spaghetti squash|
|941 | acorn squash|
|942 | butternut squash|
|943 | cucumber, cuke|
|944 | artichoke, globe artichoke|
|945 | bell pepper|
|946 | cardoon|
|947 | mushroom|
|948 | Granny Smith|
|949 | strawberry|
|950 | orange|
|951 | lemon|
|952 | fig|
|953 | pineapple, ananas|
|954 | banana|
|955 | jackfruit, jak, jack|
|956 | custard apple|
|957 | pomegranate|
|958 | hay|
|959 | carbonara|
|960 | chocolate sauce, chocolate syrup|
|961 | dough|
|962 | meat loaf, meatloaf|
|963 | pizza, pizza pie|
|964 | potpie|
|965 | burrito|
|966 | red wine|
|967 | espresso|
|968 | cup|
|969 | eggnog|
|970 | alp|
|971 | bubble|
|972 | cliff, drop, drop-off|
|973 | coral reef|
|974 | geyser|
|975 | lakeside, lakeshore|
|976 | promontory, headland, head, foreland|
|977 | sandbar, sand bar|
|978 | seashore, coast, seacoast, sea-coast|
|979 | valley, vale|
|980 | volcano|
|981 | ballplayer, baseball player|
|982 | groom, bridegroom|
|983 | scuba diver|
|984 | rapeseed|
|985 | daisy|
|986 | yellow lady's slipper, yellow lady-slipper, Cypripedium calceolus, Cypripedium parviflorum|
|987 | corn|
|988 | acorn|
|989 | hip, rose hip, rosehip|
|990 | buckeye, horse chestnut, conker|
|991 | coral fungus|
|992 | agaric|
|993 | gyromitra|
|994 | stinkhorn, carrion fungus|
|995 | earthstar|
|996 | hen-of-the-woods, hen of the woods, Polyporus frondosus, Grifola frondosa|
|997 | bolete|
|998 | ear, spike, capitulum|
|999 | toilet tissue, toilet paper, bathroom tissue|
</details>
### Data Splits
| |train|
|-------------|----:|
|# of examples|50000|
## Dataset Creation
### Curation Rationale
From the paper:
> Inspired by the Sketch data of (Li et al., 2017a) with seven classes, and several other Sketch datasets,
such as the Sketchy dataset (Sangkloy et al., 2016) with 125 classes and the Quick Draw! dataset
(QuickDraw, 2018) with 345 classes, and motivated by absence of a large-scale sketch dataset fitting
the shape and size of popular image classification benchmarks, we construct the ImageNet-Sketch
data set for evaluating the out-of-domain classification performance of vision models trained on
ImageNet.
### Source Data
#### Initial Data Collection and Normalization
The initial data collection and normalization is inherited from ImageNet. More information on it can be found [here](https://huggingface.co/datasets/imagenet-1k#initial-data-collection-and-normalization).
Additional preprocessing from the paper:
> We construct the data set with Google Image queries “sketch of ”, where is the
standard class name. We only search within the “black and white” color scheme. We initially query
100 images for every class, and then manually clean the pulled images by deleting the irrelevant
images and images that are for similar but different classes. For some classes, there are less than 50
images after manually cleaning, and then we augment the data set by flipping and rotating the images.
#### Who are the source language producers?
The source language is inherited from ImageNet. More information on the source language produces can be found [here](https://huggingface.co/datasets/imagenet-1k#who-are-the-source-language-producers).
### Annotations
#### Annotation process
The annotations are inherited from ImageNet. More information about the process can be found [here](https://huggingface.co/datasets/imagenet-1k#annotation-process).
#### Who are the annotators?
The same as in [ImageNet](https://huggingface.co/datasets/imagenet-1k#who-are-the-annotators).
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
The biases are inherited from ImageNet. More information about the process can be found [here](https://huggingface.co/datasets/imagenet-1k#discussion-of-biases).
### Other Known Limitations
1. Since most of the images were collected from internet, keep in mind that some images in ImageNet-Sketch might be subject to copyrights.
## Additional Information
### Dataset Curators
Authors of [Learning Robust Global Representations by Penalizing Local Predictive Power](https://arxiv.org/abs/1905.13549v2):
- Haohan Wang
- Songwei Ge
- Eric P. Xing
- Zachary C. Lipton
The dataset was curated using the scripts found in the [GitHub repository](https://github.com/HaohanWang/ImageNet-Sketch).
### Licensing Information
[More Information Needed]
### Citation Information
```bibtex
@inproceedings{wang2019learning,
title={Learning Robust Global Representations by Penalizing Local Predictive Power},
author={Wang, Haohan and Ge, Songwei and Lipton, Zachary and Xing, Eric P},
booktitle={Advances in Neural Information Processing Systems},
pages={10506--10518},
year={2019}
}
```
### Contributions
Thanks to [@nateraw](https://github.com/nateraw) for adding this dataset. |
bigbio/gad | 2022-12-22T15:25:28.000Z | [
"multilinguality:momolingual",
"language:en",
"license:cc-by-4.0",
"region:us"
] | bigbio | A corpus identifying associations between genes and diseases by a semi-automatic
annotation procedure based on the Genetic Association Database | @article{Bravo2015,
doi = {10.1186/s12859-015-0472-9},
url = {https://doi.org/10.1186/s12859-015-0472-9},
year = {2015},
month = feb,
publisher = {Springer Science and Business Media {LLC}},
volume = {16},
number = {1},
author = {{\`{A}}lex Bravo and Janet Pi{\~{n}}ero and N{\'{u}}ria Queralt-Rosinach and Michael Rautschka and Laura I Furlong},
title = {Extraction of relations between genes and diseases from text and large-scale data analysis: implications for translational research},
journal = {{BMC} Bioinformatics}
} | null | 0 | 139 | ---
language:
- en
bigbio_language:
- English
license: cc-by-4.0
multilinguality: momolingual
bigbio_license_shortname: CC_BY_4p0
pretty_name: GAD
homepage: https://geneticassociationdb.nih.gov/
bigbio_pubmed: true
bigbio_public: true
bigbio_tasks:
- TEXT_CLASSIFICATION
paperswithcode_id: gad
---
# Dataset Card for GAD
## Dataset Description
- **Homepage:** https://geneticassociationdb.nih.gov/
- **Pubmed:** True
- **Public:** True
- **Tasks:** TXTCLASS
A corpus identifying associations between genes and diseases by a semi-automatic
annotation procedure based on the Genetic Association Database.
## Note about homepage
The homepage for this dataset is no longer reachable, but the url is recorded here.
Data for this dataset was originally downloaded from a google drive
folder (the link used in the [BLURB benchmark data download script](https://microsoft.github.io/BLURB/submit.html).
However, we host the data in the huggingface hub for more reliable downloads and access.
## Citation Information
```
@article{Bravo2015,
doi = {10.1186/s12859-015-0472-9},
url = {https://doi.org/10.1186/s12859-015-0472-9},
year = {2015},
month = feb,
publisher = {Springer Science and Business Media {LLC}},
volume = {16},
number = {1},
author = {{\`{A}}lex Bravo and Janet Pi{\~{n}}ero and N{\'{u}}ria Queralt-Rosinach and Michael Rautschka and Laura I Furlong},
title = {Extraction of relations between genes and diseases from text and large-scale data analysis: implications for translational research},
journal = {{BMC} Bioinformatics}
}
```
|
tomekkorbak/detoxify-pile-chunk3-400000-450000 | 2022-10-03T18:51:21.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 139 | Entry not found |
tomekkorbak/detoxify-pile-chunk3-450000-500000 | 2022-10-03T19:48:41.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 139 | Entry not found |
tomekkorbak/detoxify-pile-chunk3-500000-550000 | 2022-10-04T17:42:07.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 139 | Entry not found |
tomekkorbak/detoxify-pile-chunk3-550000-600000 | 2022-10-04T17:46:16.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 139 | Entry not found |
yangwang825/reuters-21578 | 2023-05-19T02:04:58.000Z | [
"task_categories:text-classification",
"language:en",
"region:us"
] | yangwang825 | null | null | null | 0 | 139 | ---
task_categories:
- text-classification
language:
- en
dataset_info:
features:
- name: text
dtype: string
- name: label
dtype:
class_label:
names:
'0': acq
'1': crude
'2': earn
'3': grain
'4': interest
'5': money-fx
'6': ship
'7': trade
---
`yangwang825/reuters-21578` is an 8-class subset of the Reuters 21578 news dataset.
|
YeungNLP/ultrachat | 2023-06-19T02:52:43.000Z | [
"region:us"
] | YeungNLP | null | null | null | 12 | 139 | Entry not found |
eliolio/dialogsum-noniid | 2023-06-28T19:51:13.000Z | [
"region:us"
] | eliolio | null | null | null | 0 | 139 | Entry not found |
hiroshi-matsuda-rit/filtered_mc4 | 2023-08-28T08:52:06.000Z | [
"multilinguality:multilingual",
"license:odc-by",
"arxiv:1910.10683",
"region:us"
] | hiroshi-matsuda-rit | The mC4 dataset to which arbitrary filters can be applied.
The original description is below:
===
A colossal, cleaned version of Common Crawl's web crawl corpus.
Based on Common Crawl dataset: "https://commoncrawl.org".
This is the processed version of Google's mC4 dataset by AllenAI. | @article{2019t5,
author = {Colin Raffel and Noam Shazeer and Adam Roberts and Katherine Lee and Sharan Narang and Michael Matena and Yanqi Zhou and Wei Li and Peter J. Liu},
title = {Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer},
journal = {arXiv e-prints},
year = {2019},
archivePrefix = {arXiv},
eprint = {1910.10683},
} | null | 0 | 139 | ---
pretty_name: filtered-mc4
license:
- odc-by
multilinguality:
- multilingual
---
# Dataset Card for filtered-mc4
See original [mC4 dataset](https://huggingface.co/datasets/mc4) descriptions.
You can apply any regular expression to the mC4 dataset like this:
```python
from datasets import load_dataset
dataset = load_dataset('hiroshi-matsuda-rit/filtered_mc4', 'ja', split='train', reject_patterns=[r"(セフレ|出会い?系|(?<!ユニ)セックス|ソープガイド)", r"[^\s]\ [^\s]+\ [^\s]"], max_reject_pattern_occurence=3, streaming=True)
```
### Citation Information
```
@article{2019t5,
author = {Colin Raffel and Noam Shazeer and Adam Roberts and Katherine Lee and Sharan Narang and Michael Matena and Yanqi Zhou and Wei Li and Peter J. Liu},
title = {Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer},
journal = {arXiv e-prints},
year = {2019},
archivePrefix = {arXiv},
eprint = {1910.10683},
}
```
|
samlhuillier/sql-create-context-spider-intersect | 2023-09-21T00:17:19.000Z | [
"region:us"
] | samlhuillier | null | null | null | 0 | 139 | Entry not found |
bloyal/oas-paired-sequence-data | 2023-09-28T21:26:27.000Z | [
"task_categories:fill-mask",
"language:en",
"license:cc-by-4.0",
"region:us"
] | bloyal | Paired heavy and light chain antibody sequences for multiple species. | @article{Olsen_Boyles_Deane_2022,
title={Observed Antibody Space: A diverse database of cleaned, annotated, and translated unpaired and paired antibody sequences},
volume={31}, rights={© 2021 The Authors. Protein Science published by Wiley Periodicals LLC on behalf of The Protein Society.},
ISSN={1469-896X}, DOI={10.1002/pro.4205},
number={1}, journal={Protein Science}, author={Olsen, Tobias H. and Boyles, Fergus and Deane, Charlotte M.},
year={2022}, pages={141–146}, language={en} } | null | 0 | 139 | ---
pretty_name: OAS paired sequences
language: en
task_categories:
- fill-mask
license: cc-by-4.0
---
# Dataset Card for OAS Paired Sequence Data
## Dataset Description
- **Homepage:**
- https://opig.stats.ox.ac.uk/webapps/oas/oas_paired/
## Dataset Summary
Paired heavy- and light-chain sequence information from the Observed Antibody Space (OAS) database, downloaded on September 9, 2023.
|
hynky/czech-justice-summ-alpaca-long | 2023-09-10T21:24:17.000Z | [
"region:us"
] | hynky | null | null | null | 0 | 139 | ---
dataset_info:
features:
- name: output
dtype: string
- name: instruction
dtype: string
splits:
- name: train
num_bytes: 26403302
num_examples: 4560
download_size: 12636847
dataset_size: 26403302
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "czech-justice-summ-alpaca-long"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
djscrave/tsh | 2023-09-16T11:04:10.000Z | [
"task_categories:text-classification",
"size_categories:1K<n<10K",
"language:fr",
"license:openrail",
"chemistry",
"region:us"
] | djscrave | null | null | null | 0 | 139 | ---
configs:
- config_name: default
data_files:
- split: train
path: "train.csv"
- split: validation
path: "validation.csv"
- split: test
path: "test.csv"
license: openrail
task_categories:
- text-classification
language:
- fr
tags:
- chemistry
size_categories:
- 1K<n<10K
--- |
mac_morpho | 2023-01-25T14:34:31.000Z | [
"task_categories:token-classification",
"task_ids:part-of-speech",
"annotations_creators:expert-generated",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:pt",
"license:cc-by-4.0",
"region:us"
] | null | Mac-Morpho is a corpus of Brazilian Portuguese texts annotated with part-of-speech tags.
Its first version was released in 2003 [1], and since then, two revisions have been made in order
to improve the quality of the resource [2, 3].
The corpus is available for download split into train, development and test sections.
These are 76%, 4% and 20% of the corpus total, respectively (the reason for the unusual numbers
is that the corpus was first split into 80%/20% train/test, and then 5% of the train section was
set aside for development). This split was used in [3], and new POS tagging research with Mac-Morpho
is encouraged to follow it in order to make consistent comparisons possible.
[1] Aluísio, S., Pelizzoni, J., Marchi, A.R., de Oliveira, L., Manenti, R., Marquiafável, V. 2003.
An account of the challenge of tagging a reference corpus for brazilian portuguese.
In: Proceedings of the 6th International Conference on Computational Processing of the Portuguese Language. PROPOR 2003
[2] Fonseca, E.R., Rosa, J.L.G. 2013. Mac-morpho revisited: Towards robust part-of-speech.
In: Proceedings of the 9th Brazilian Symposium in Information and Human Language Technology – STIL
[3] Fonseca, E.R., Aluísio, Sandra Maria, Rosa, J.L.G. 2015.
Evaluating word embeddings and a revised corpus for part-of-speech tagging in Portuguese.
Journal of the Brazilian Computer Society. | @article{fonseca2015evaluating,
title={Evaluating word embeddings and a revised corpus for part-of-speech tagging in Portuguese},
author={Fonseca, Erick R and Rosa, Joao Luis G and Aluisio, Sandra Maria},
journal={Journal of the Brazilian Computer Society},
volume={21},
number={1},
pages={2},
year={2015},
publisher={Springer}
} | null | 4 | 138 | ---
annotations_creators:
- expert-generated
language_creators:
- found
language:
- pt
license:
- cc-by-4.0
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- original
task_categories:
- token-classification
task_ids:
- part-of-speech
pretty_name: Mac-Morpho
dataset_info:
features:
- name: id
dtype: string
- name: tokens
sequence: string
- name: pos_tags
sequence:
class_label:
names:
'0': PREP+PROADJ
'1': IN
'2': PREP+PRO-KS
'3': NPROP
'4': PREP+PROSUB
'5': KC
'6': PROPESS
'7': NUM
'8': PROADJ
'9': PREP+ART
'10': KS
'11': PRO-KS
'12': ADJ
'13': ADV-KS
'14': N
'15': PREP
'16': PROSUB
'17': PREP+PROPESS
'18': PDEN
'19': V
'20': PREP+ADV
'21': PCP
'22': CUR
'23': ADV
'24': PU
'25': ART
splits:
- name: train
num_bytes: 12635011
num_examples: 37948
- name: test
num_bytes: 3095292
num_examples: 9987
- name: validation
num_bytes: 671356
num_examples: 1997
download_size: 2463485
dataset_size: 16401659
---
# Dataset Card for Mac-Morpho
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [Mac-Morpho homepage](http://nilc.icmc.usp.br/macmorpho/)
- **Repository:** [Mac-Morpho repository](http://nilc.icmc.usp.br/macmorpho/)
- **Paper:** [Evaluating word embeddings and a revised corpus for part-of-speech tagging in Portuguese](https://journal-bcs.springeropen.com/articles/10.1186/s13173-014-0020-x)
- **Point of Contact:** [Erick R Fonseca](mailto:erickrfonseca@gmail.com)
### Dataset Summary
Mac-Morpho is a corpus of Brazilian Portuguese texts annotated with part-of-speech tags.
Its first version was released in 2003 [1], and since then, two revisions have been made in order
to improve the quality of the resource [2, 3].
The corpus is available for download split into train, development and test sections.
These are 76%, 4% and 20% of the corpus total, respectively (the reason for the unusual numbers
is that the corpus was first split into 80%/20% train/test, and then 5% of the train section was
set aside for development). This split was used in [3], and new POS tagging research with Mac-Morpho
is encouraged to follow it in order to make consistent comparisons possible.
[1] Aluísio, S., Pelizzoni, J., Marchi, A.R., de Oliveira, L., Manenti, R., Marquiafável, V. 2003.
An account of the challenge of tagging a reference corpus for brazilian portuguese.
In: Proceedings of the 6th International Conference on Computational Processing of the Portuguese Language. PROPOR 2003
[2] Fonseca, E.R., Rosa, J.L.G. 2013. Mac-morpho revisited: Towards robust part-of-speech.
In: Proceedings of the 9th Brazilian Symposium in Information and Human Language Technology – STIL
[3] Fonseca, E.R., Aluísio, Sandra Maria, Rosa, J.L.G. 2015.
Evaluating word embeddings and a revised corpus for part-of-speech tagging in Portuguese.
Journal of the Brazilian Computer Society.
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
Portuguese
## Dataset Structure
### Data Instances
An example from the Mac-Morpho dataset looks as follows:
```
{
"id": "0",
"pos_tags": [14, 19, 14, 15, 22, 7, 14, 9, 14, 9, 3, 15, 3, 3, 24],
"tokens": ["Jersei", "atinge", "média", "de", "Cr$", "1,4", "milhão", "na", "venda", "da", "Pinhal", "em", "São", "Paulo", "."]
}
```
### Data Fields
- `id`: id of the sample
- `tokens`: the tokens of the example text
- `pos`: the PoS tags of each token
The PoS tags correspond to this list:
```
"PREP+PROADJ", "IN", "PREP+PRO-KS", "NPROP", "PREP+PROSUB", "KC", "PROPESS", "NUM", "PROADJ", "PREP+ART", "KS",
"PRO-KS", "ADJ", "ADV-KS", "N", "PREP", "PROSUB", "PREP+PROPESS", "PDEN", "V", "PREP+ADV", "PCP", "CUR", "ADV", "PU", "ART"
```
### Data Splits
The data is split into train, validation and test set. The split sizes are as follow:
| Train | Val | Test |
| ------ | ----- | ----- |
| 37948 | 1997 | 9987 |
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
```
@article{fonseca2015evaluating,
title={Evaluating word embeddings and a revised corpus for part-of-speech tagging in Portuguese},
author={Fonseca, Erick R and Rosa, Jo{\~a}o Lu{\'\i}s G and Alu{\'\i}sio, Sandra Maria},
journal={Journal of the Brazilian Computer Society},
volume={21},
number={1},
pages={2},
year={2015},
publisher={Springer}
}
```
### Contributions
Thanks to [@jonatasgrosman](https://github.com/jonatasgrosman) for adding this dataset. |
allegro/klej-dyk | 2022-10-26T09:01:41.000Z | [
"task_categories:question-answering",
"task_ids:open-domain-qa",
"annotations_creators:expert-generated",
"language_creators:other",
"multilinguality:monolingual",
"size_categories:1K<n<10K",
"source_datasets:original",
"language:pl",
"license:cc-by-sa-3.0",
"region:us"
] | allegro | null | null | null | 1 | 138 | ---
annotations_creators:
- expert-generated
language_creators:
- other
language:
- pl
license:
- cc-by-sa-3.0
multilinguality:
- monolingual
size_categories:
- 1K<n<10K
source_datasets:
- original
task_categories:
- question-answering
task_ids:
- open-domain-qa
pretty_name: Did you know?
---
# klej-dyk
## Description
The Czy wiesz? (eng. Did you know?) the dataset consists of almost 5k question-answer pairs obtained from Czy wiesz... section of Polish Wikipedia. Each question is written by a Wikipedia collaborator and is answered with a link to a relevant Wikipedia article. In huggingface version of this dataset, they chose the negatives which have the largest token overlap with a question.
## Tasks (input, output, and metrics)
The task is to predict if the answer to the given question is correct or not.
**Input** ('question sentence', 'answer' columns): question and answer sentences
**Output** ('target' column): 1 if the answer is correct, 0 otherwise.
**Domain**: Wikipedia
**Measurements**: F1-Score
**Example**:
Input: `Czym zajmowali się świątnicy?` ; `Świątnik – osoba, która dawniej zajmowała się
obsługą kościoła (świątyni).`
Input (translated by DeepL): `What did the sacristans do?` ; `A sacristan - a person who used to be in charge of the handling the church (temple).`
Output: `1` (the answer is correct)
## Data splits
| Subset | Cardinality |
| ----------- | ----------: |
| train | 4154 |
| val | 0 |
| test | 1029 |
## Class distribution
| Class | train | validation | test |
|:----------|--------:|-------------:|-------:|
| incorrect | 0.831 | - | 0.831 |
| correct | 0.169 | - | 0.169 |
## Citation
```
@misc{11321/39,
title = {Pytania i odpowiedzi z serwisu wikipedyjnego "Czy wiesz", wersja 1.1},
author = {Marci{\'n}czuk, Micha{\l} and Piasecki, Dominik and Piasecki, Maciej and Radziszewski, Adam},
url = {http://hdl.handle.net/11321/39},
note = {{CLARIN}-{PL} digital repository},
year = {2013}
}
```
## License
```
Creative Commons Attribution ShareAlike 3.0 licence (CC-BY-SA 3.0)
```
## Links
[HuggingFace](https://huggingface.co/datasets/dyk)
[Source](http://nlp.pwr.wroc.pl/en/tools-and-resources/resources/czy-wiesz-question-answering-dataset)
[Source #2](https://clarin-pl.eu/dspace/handle/11321/39)
[Paper](https://www.researchgate.net/publication/272685895_Open_dataset_for_development_of_Polish_Question_Answering_systems)
## Examples
### Loading
```python
from pprint import pprint
from datasets import load_dataset
dataset = load_dataset("allegro/klej-dyk")
pprint(dataset['train'][100])
#{'answer': '"W wyborach prezydenckich w 2004 roku, Moroz przekazał swoje '
# 'poparcie Wiktorowi Juszczence. Po wyborach w 2006 socjaliści '
# 'początkowo tworzyli ""pomarańczową koalicję"" z Naszą Ukrainą i '
# 'Blokiem Julii Tymoszenko."',
# 'q_id': 'czywiesz4362',
# 'question': 'ile partii tworzy powołaną przez Wiktora Juszczenkę koalicję '
# 'Blok Nasza Ukraina?',
# 'target': 0}
```
### Evaluation
```python
import random
from pprint import pprint
from datasets import load_dataset, load_metric
dataset = load_dataset("allegro/klej-dyk")
dataset = dataset.class_encode_column("target")
references = dataset["test"]["target"]
# generate random predictions
predictions = [random.randrange(max(references) + 1) for _ in range(len(references))]
acc = load_metric("accuracy")
f1 = load_metric("f1")
acc_score = acc.compute(predictions=predictions, references=references)
f1_score = f1.compute(predictions=predictions, references=references, average="macro")
pprint(acc_score)
pprint(f1_score)
# {'accuracy': 0.5286686103012633}
# {'f1': 0.46700507614213194}
``` |
mozilla-foundation/common_voice_1_0 | 2023-07-29T15:59:56.000Z | [
"task_categories:automatic-speech-recognition",
"annotations_creators:crowdsourced",
"language_creators:crowdsourced",
"multilinguality:multilingual",
"source_datasets:extended|common_voice",
"license:cc0-1.0",
"arxiv:1912.06670",
"region:us"
] | mozilla-foundation | null | @inproceedings{commonvoice:2020,
author = {Ardila, R. and Branson, M. and Davis, K. and Henretty, M. and Kohler, M. and Meyer, J. and Morais, R. and Saunders, L. and Tyers, F. M. and Weber, G.},
title = {Common Voice: A Massively-Multilingual Speech Corpus},
booktitle = {Proceedings of the 12th Conference on Language Resources and Evaluation (LREC 2020)},
pages = {4211--4215},
year = 2020
} | null | 2 | 138 | ---
annotations_creators:
- crowdsourced
language_creators:
- crowdsourced
license:
- cc0-1.0
multilinguality:
- multilingual
size_categories:
br:
- 1K<n<10K
ca:
- 10K<n<100K
cnh:
- 1K<n<10K
cv:
- 1K<n<10K
cy:
- 10K<n<100K
de:
- 100K<n<1M
en:
- 100K<n<1M
eo:
- 1K<n<10K
et:
- n<1K
fr:
- 10K<n<100K
ga-IE:
- 1K<n<10K
it:
- 10K<n<100K
kab:
- 100K<n<1M
ky:
- 1K<n<10K
nl:
- 10K<n<100K
sl:
- 1K<n<10K
tr:
- 1K<n<10K
tt:
- 10K<n<100K
zh-TW:
- 10K<n<100K
source_datasets:
- extended|common_voice
paperswithcode_id: common-voice
pretty_name: Common Voice Corpus 1
language_bcp47:
- br
- ca
- cnh
- cv
- cy
- de
- en
- eo
- et
- fr
- ga-IE
- it
- kab
- ky
- nl
- sl
- tr
- tt
- zh-TW
extra_gated_prompt: By clicking on “Access repository” below, you also agree to not
attempt to determine the identity of speakers in the Common Voice dataset.
task_categories:
- automatic-speech-recognition
---
# Dataset Card for Common Voice Corpus 1
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** https://commonvoice.mozilla.org/en/datasets
- **Repository:** https://github.com/common-voice/common-voice
- **Paper:** https://arxiv.org/abs/1912.06670
- **Leaderboard:** https://paperswithcode.com/dataset/common-voice
- **Point of Contact:** [Anton Lozhkov](mailto:anton@huggingface.co)
### Dataset Summary
The Common Voice dataset consists of a unique MP3 and corresponding text file.
Many of the 1368 recorded hours in the dataset also include demographic metadata like age, sex, and accent
that can help improve the accuracy of speech recognition engines.
The dataset currently consists of 1096 validated hours in 19 languages, but more voices and languages are always added.
Take a look at the [Languages](https://commonvoice.mozilla.org/en/languages) page to request a language or start contributing.
### Supported Tasks and Leaderboards
The results for models trained on the Common Voice datasets are available via the
[🤗 Speech Bench](https://huggingface.co/spaces/huggingface/hf-speech-bench)
### Languages
```
Breton, Catalan, Chinese (Taiwan), Chuvash, Dutch, English, Esperanto, Estonian, French, German, Hakha Chin, Irish, Italian, Kabyle, Kyrgyz, Slovenian, Tatar, Turkish, Welsh
```
## Dataset Structure
### Data Instances
A typical data point comprises the `path` to the audio file and its `sentence`.
Additional fields include `accent`, `age`, `client_id`, `up_votes`, `down_votes`, `gender`, `locale` and `segment`.
```python
{
'client_id': 'd59478fbc1ee646a28a3c652a119379939123784d99131b865a89f8b21c81f69276c48bd574b81267d9d1a77b83b43e6d475a6cfc79c232ddbca946ae9c7afc5',
'path': 'et/clips/common_voice_et_18318995.mp3',
'audio': {
'path': 'et/clips/common_voice_et_18318995.mp3',
'array': array([-0.00048828, -0.00018311, -0.00137329, ..., 0.00079346, 0.00091553, 0.00085449], dtype=float32),
'sampling_rate': 48000
},
'sentence': 'Tasub kokku saada inimestega, keda tunned juba ammust ajast saati.',
'up_votes': 2,
'down_votes': 0,
'age': 'twenties',
'gender': 'male',
'accent': '',
'locale': 'et',
'segment': ''
}
```
### Data Fields
`client_id` (`string`): An id for which client (voice) made the recording
`path` (`string`): The path to the audio file
`audio` (`dict`): A dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate. Note that when accessing the audio column: `dataset[0]["audio"]` the audio file is automatically decoded and resampled to `dataset.features["audio"].sampling_rate`. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus it is important to first query the sample index before the `"audio"` column, *i.e.* `dataset[0]["audio"]` should **always** be preferred over `dataset["audio"][0]`.
`sentence` (`string`): The sentence the user was prompted to speak
`up_votes` (`int64`): How many upvotes the audio file has received from reviewers
`down_votes` (`int64`): How many downvotes the audio file has received from reviewers
`age` (`string`): The age of the speaker (e.g. `teens`, `twenties`, `fifties`)
`gender` (`string`): The gender of the speaker
`accent` (`string`): Accent of the speaker
`locale` (`string`): The locale of the speaker
`segment` (`string`): Usually an empty field
### Data Splits
The speech material has been subdivided into portions for dev, train, test, validated, invalidated, reported and other.
The validated data is data that has been validated with reviewers and received upvotes that the data is of high quality.
The invalidated data is data has been invalidated by reviewers
and received downvotes indicating that the data is of low quality.
The reported data is data that has been reported, for different reasons.
The other data is data that has not yet been reviewed.
The dev, test, train are all data that has been reviewed, deemed of high quality and split into dev, test and train.
## Data Preprocessing Recommended by Hugging Face
The following are data preprocessing steps advised by the Hugging Face team. They are accompanied by an example code snippet that shows how to put them to practice.
Many examples in this dataset have trailing quotations marks, e.g _“the cat sat on the mat.“_. These trailing quotation marks do not change the actual meaning of the sentence, and it is near impossible to infer whether a sentence is a quotation or not a quotation from audio data alone. In these cases, it is advised to strip the quotation marks, leaving: _the cat sat on the mat_.
In addition, the majority of training sentences end in punctuation ( . or ? or ! ), whereas just a small proportion do not. In the dev set, **almost all** sentences end in punctuation. Thus, it is recommended to append a full-stop ( . ) to the end of the small number of training examples that do not end in punctuation.
```python
from datasets import load_dataset
ds = load_dataset("mozilla-foundation/common_voice_1_0", "en", use_auth_token=True)
def prepare_dataset(batch):
"""Function to preprocess the dataset with the .map method"""
transcription = batch["sentence"]
if transcription.startswith('"') and transcription.endswith('"'):
# we can remove trailing quotation marks as they do not affect the transcription
transcription = transcription[1:-1]
if transcription[-1] not in [".", "?", "!"]:
# append a full-stop to sentences that do not end in punctuation
transcription = transcription + "."
batch["sentence"] = transcription
return batch
ds = ds.map(prepare_dataset, desc="preprocess dataset")
```
## Dataset Creation
### Curation Rationale
[Needs More Information]
### Source Data
#### Initial Data Collection and Normalization
[Needs More Information]
#### Who are the source language producers?
[Needs More Information]
### Annotations
#### Annotation process
[Needs More Information]
#### Who are the annotators?
[Needs More Information]
### Personal and Sensitive Information
The dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in the Common Voice dataset.
## Considerations for Using the Data
### Social Impact of Dataset
The dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in the Common Voice dataset.
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
Public Domain, [CC-0](https://creativecommons.org/share-your-work/public-domain/cc0/)
### Citation Information
```
@inproceedings{commonvoice:2020,
author = {Ardila, R. and Branson, M. and Davis, K. and Henretty, M. and Kohler, M. and Meyer, J. and Morais, R. and Saunders, L. and Tyers, F. M. and Weber, G.},
title = {Common Voice: A Massively-Multilingual Speech Corpus},
booktitle = {Proceedings of the 12th Conference on Language Resources and Evaluation (LREC 2020)},
pages = {4211--4215},
year = 2020
}
```
|
tomekkorbak/detoxify-pile-chunk3-700000-750000 | 2022-10-04T17:50:07.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 138 | Entry not found |
tomekkorbak/detoxify-pile-chunk3-600000-650000 | 2022-10-04T17:51:35.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 138 | Entry not found |
tomekkorbak/detoxify-pile-chunk3-750000-800000 | 2022-10-04T22:48:41.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 138 | Entry not found |
tomekkorbak/detoxify-pile-chunk3-850000-900000 | 2022-10-04T23:55:21.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 138 | Entry not found |
tomekkorbak/detoxify-pile-chunk3-650000-700000 | 2022-10-04T18:03:56.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 138 | Entry not found |
tomekkorbak/detoxify-pile-chunk3-1100000-1150000 | 2022-10-04T23:49:53.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 138 | Entry not found |
tomekkorbak/detoxify-pile-chunk3-1050000-1100000 | 2022-10-04T23:53:15.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 138 | Entry not found |
tomekkorbak/detoxify-pile-chunk3-1000000-1050000 | 2022-10-04T23:58:45.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 138 | Entry not found |
inverse-scaling/NeQA | 2022-10-08T12:40:09.000Z | [
"task_categories:multiple-choice",
"task_categories:question-answering",
"task_categories:zero-shot-classification",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"language:en",
"license:cc-by-sa-4.0",
"region:us"
] | inverse-scaling | null | null | null | 0 | 138 | ---
language:
- en
size_categories:
- 10K<n<100K
license:
- cc-by-sa-4.0
multilinguality:
- monolingual
pretty_name: NeQA - Can Large Language Models Understand Negation in Multi-choice Questions?
source_datasets: []
task_categories:
- multiple-choice
- question-answering
- zero-shot-classification
train-eval-index:
- config: inverse-scaling--NeQA
task: text-generation
task_id: text_zero_shot_classification
splits:
eval_split: train
col_mapping:
prompt: text
classes: classes
answer_index: target
---
## NeQA: Can Large Language Models Understand Negation in Multi-choice Questions? (Zhengping Zhou and Yuhui Zhang)
### General description
This task takes an existing multiple-choice dataset and negates a part of each question to see if language models are sensitive to negation. The authors find that smaller language models display approximately random performance whereas the performance of larger models become significantly worse than random.
Language models failing to follow instructions in the prompt could be a serious issue that only becomes apparent on a task once models are sufficiently capable to perform non-randomly on the task.
### Example
The following are multiple choice questions (with answers) about common sense.
Question: If a cat has a body temp that is below average, it isn't in
A. danger
B. safe ranges
Answer:
(where the model should choose B.)
## Submission details
### Task description
Negation is a common linguistic phenomenon that can completely alter the semantics of a sentence by changing just a few words.
This task evaluates whether language models can understand negation, which is an important step towards true natural language understanding.
Specifically, we focus on negation in open-book multi-choice questions, considering its wide range of applications and the simplicity of evaluation.
We collect a multi-choice question answering dataset, NeQA, that includes questions with negations.
When negation is presented in the question, the original correct answer becomes wrong, and the wrong answer becomes correct.
We use the accuracy metric to examine whether the model can understand negation in the questions and select the correct answer given the presence of negation.
We observe a clear inverse scaling trend on GPT-3, demonstrating that larger language models can answer more complex questions but fail at the last step to understanding negation.
### Dataset generation procedure
The dataset is created by applying rules to transform questions in a publicly available multiple-choice question answering dataset named OpenBookQA. We use a simple rule by filtering questions containing "is" and adding "not" after it. For each question, we sample an incorrect answer as the correct answer and treat the correct answer as the incorrect answer. We randomly sample 300 questions and balance the label distributions (50% label as "A" and 50% label as "B" since there are two choices for each question)..
### Why do you expect to see inverse scaling?
For open-book question answering, larger language models usually achieve better accuracy because more factual and commonsense knowledge is stored in the model parameters and can be used as a knowledge base to answer these questions without context.
A higher accuracy rate means a lower chance of choosing the wrong answer. Can we change the wrong answer to the correct one? A simple solution is to negate the original question. If the model cannot understand negation, it will still predict the same answer and, therefore, will exhibit an inverse scaling trend.
We expect that the model cannot understand negation because negation introduces only a small perturbation to the model input. It is difficult for the model to understand that this small perturbation leads to completely different semantics.
### Why is the task important?
This task is important because it demonstrates that current language models cannot understand negation, a very common linguistic phenomenon and a real-world challenge to natural language understanding.
Why is the task novel or surprising? (1+ sentences)
To the best of our knowledge, no prior work shows that negation can cause inverse scaling. This finding should be surprising to the community, as large language models show an incredible variety of emergent capabilities, but still fail to understand negation, which is a fundamental concept in language.
## Results
[Inverse Scaling Prize: Round 1 Winners announcement](https://www.alignmentforum.org/posts/iznohbCPFkeB9kAJL/inverse-scaling-prize-round-1-winners#Zhengping_Zhou_and_Yuhui_Zhang__for_NeQA__Can_Large_Language_Models_Understand_Negation_in_Multi_choice_Questions_)
|
tobiolatunji/afrispeech-200 | 2023-05-20T23:29:22.000Z | [
"task_categories:automatic-speech-recognition",
"annotations_creators:expert-generated",
"language_creators:crowdsourced",
"language_creators:expert-generated",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:en",
"license:cc-by-nc-sa-4.0",
"region:us"
] | tobiolatunji | AFRISPEECH-200 is a 200hr Pan-African speech corpus for clinical and general domain English accented ASR;
a dataset with 120 African accents from 13 countries and 2,463 unique African speakers.
Our goal is to raise awareness for and advance Pan-African English ASR research,
especially for the clinical domain. | TBD | null | 8 | 138 | ---
pretty_name: AfriSpeech-200
annotations_creators:
- expert-generated
language_creators:
- crowdsourced
- expert-generated
language:
- en
license:
- cc-by-nc-sa-4.0
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- original
task_categories:
- automatic-speech-recognition
task_ids: []
dataset_info:
features:
- name: user_id
dtype: string
- name: path
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 44100
- name: transcript
dtype: string
splits:
- name: train
num_bytes: 1722002133
num_examples: 58000
- name: dev
num_bytes: 86120227
num_examples: 3231
download_size: 1475540500
dataset_size: 1808122360
extra_gated_prompt: By clicking on “Access repository” below, you also agree to not attempt to determine the
identity of speakers in the Common Voice dataset.
---
# Dataset Card for AfriSpeech-200
## Table of Contents
- [Dataset Card for AfriSpeech-200](#dataset-card-for-afrispeech-200)
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [How to use](#how-to-use)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Initial Data Collection and Normalization](#initial-data-collection-and-normalization)
- [Who are the source language producers?](#who-are-the-source-language-producers)
- [Annotations](#annotations)
- [Annotation process](#annotation-process)
- [Who are the annotators?](#who-are-the-annotators)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** https://github.com/intron-innovation/AfriSpeech-Dataset-Paper
- **Repository:** https://github.com/intron-innovation/AfriSpeech-Dataset-Paper
- **Paper:** [AfriSpeech-200: Pan-African accented speech dataset for clinical and general domain ASR](https://github.com/intron-innovation/AfriSpeech-Dataset-Paper)
- **Leaderboard:** [Needs More Information]
- **Point of Contact:** [Intron Innovation](mailto:intron@intron.io)
### Dataset Summary
AFRISPEECH-200 is a 200hr Pan-African speech corpus for clinical and general domain English accented ASR; a dataset with 120 African accents from 13 countries and 2,463 unique African speakers.
Our goal is to raise awareness for and advance Pan-African English ASR research, especially for the clinical domain.
## How to use
The `datasets` library allows you to load and pre-process your dataset in pure Python, at scale. The dataset can be downloaded and prepared in one call to your local drive by using the `load_dataset` function.
```python
from datasets import load_dataset
afrispeech = load_dataset("tobiolatunji/afrispeech-200", "all")
```
The entire dataset is ~120GB and may take about 2hrs to download depending on internet speed/bandwidth. If you have disk space or bandwidth limitations, you can use `streaming` mode described below to work with smaller subsets of the data.
Alterntively you are able to pass a config to the `load_dataset` function and download only a subset of the data corresponding to a specific accent of interest. The example provided below is `isizulu`.
For example, to download the isizulu config, simply specify the corresponding accent config name. The list of supported accents is provided in the `accent list` section below:
```python
from datasets import load_dataset
afrispeech = load_dataset("tobiolatunji/afrispeech-200", "isizulu", split="train")
```
Using the datasets library, you can also stream the dataset on-the-fly by adding a `streaming=True` argument to the `load_dataset` function call. Loading a dataset in streaming mode loads individual samples of the dataset at a time, rather than downloading the entire dataset to disk.
```python
from datasets import load_dataset
afrispeech = load_dataset("tobiolatunji/afrispeech-200", "isizulu", split="train", streaming=True)
print(next(iter(afrispeech)))
print(list(afrispeech.take(5)))
```
### Local
```python
from datasets import load_dataset
from torch.utils.data.sampler import BatchSampler, RandomSampler
afrispeech = load_dataset("tobiolatunji/afrispeech-200", "isizulu", split="train")
batch_sampler = BatchSampler(RandomSampler(afrispeech), batch_size=32, drop_last=False)
dataloader = DataLoader(afrispeech, batch_sampler=batch_sampler)
```
### Streaming
```python
from datasets import load_dataset
from torch.utils.data import DataLoader
afrispeech = load_dataset("tobiolatunji/afrispeech-200", "isizulu", split="train", streaming=True)
dataloader = DataLoader(afrispeech, batch_size=32)
```
### Caveats
Note that till the end of the ongoing [AfriSpeech ASR Challenge event](https://zindi.africa/competitions/intron-afrispeech-200-automatic-speech-recognition-challenge) (Feb - May 2023), the transcripts in the validation set are hidden and the test set will be unreleased till May 19, 2023.
### Fine-tuning Colab tutorial
To walk through a complete colab tutorial that finetunes a wav2vec2 model on the afrispeech-200 dataset with `transformers`, take a look at this colab notebook [afrispeech/wav2vec2-colab-tutorial](https://colab.research.google.com/drive/1uZYew6pcgN6UE6sFDLohxD_HKivvDXzD?usp=sharing).
### Supported Tasks and Leaderboards
- Automatic Speech Recognition
- Speech Synthesis (Text-to-Speech)
### Languages
English (Accented)
## Dataset Structure
### Data Instances
A typical data point comprises the path to the audio file, called `path` and its transcription, called `transcript`. Some additional information about the speaker is provided.
```
{
'speaker_id': 'b545a4ca235a7b72688a1c0b3eb6bde6',
'path': 'aad9bd69-7ca0-4db1-b650-1eeea17a0153/5dcb6ee086e392376cd3b7131a250397.wav',
'audio_id': 'aad9bd69-7ca0-4db1-b650-1eeea17a0153/5dcb6ee086e392376cd3b7131a250397',
'audio': {
'path': 'aad9bd69-7ca0-4db1-b650-1eeea17a0153/5dcb6ee086e392376cd3b7131a250397.wav',
'array': array([0.00018311, 0.00061035, 0.00012207, ..., 0.00192261, 0.00195312, 0.00216675]),
'sampling_rate': 44100},
'transcript': 'His mother is in her 50 s and has hypertension .',
'age_group': '26-40',
'gender': 'Male',
'accent': 'yoruba',
'domain': 'clinical',
'country': 'US',
'duration': 3.241995464852608
}
```
### Data Fields
- speaker_id: An id for which speaker (voice) made the recording
- path: The path to the audio file
- audio: A dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate. Note that when accessing the audio column: `dataset[0]["audio"]` the audio file is automatically decoded and resampled to `dataset.features["audio"].sampling_rate`. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus it is important to first query the sample index before the `"audio"` column, *i.e.* `dataset[0]["audio"]` should **always** be preferred over `dataset["audio"][0]`.
- transcript: The sentence the user was prompted to speak
### Data Splits
The speech material has been subdivided into portions for train, dev, and test.
Speech was recorded in a quiet environment with high quality microphone, speakers were asked to read one sentence at a time.
- Total Number of Unique Speakers: 2,463
- Female/Male/Other Ratio: 57.11/42.41/0.48
- Data was first split on speakers. Speakers in Train/Dev/Test do not cross partitions
| | Train | Dev | Test |
| ----------- | ----------- | ----------- | ----------- |
| # Speakers | 1466 | 247 | 750 |
| # Seconds | 624228.83 | 31447.09 | 67559.10 |
| # Hours | 173.4 | 8.74 | 18.77 |
| # Accents | 71 | 45 | 108 |
| Avg secs/speaker | 425.81 | 127.32 | 90.08 |
| Avg num clips/speaker | 39.56 | 13.08 | 8.46 |
| Avg num speakers/accent | 20.65 | 5.49 | 6.94 |
| Avg secs/accent | 8791.96 | 698.82 | 625.55 |
| # clips general domain | 21682 | 1407 | 2723 |
| # clips clinical domain | 36318 | 1824 | 3623 |
## Dataset Creation
### Curation Rationale
Africa has a very low doctor-to-patient ratio.
At very busy clinics, doctors could see 30+ patients per day-- a heavy patient burden compared with
developed countries-- but productivity tools such as clinical automatic speech recognition
(ASR) are lacking for these overworked clinicians. However, clinical ASR is mature, even ubiquitous,
in developed nations, and clinician-reported performance of commercial clinical ASR systems
is generally satisfactory. Furthermore, the recent performance of general domain ASR is
approaching human accuracy. However, several gaps exist. Several publications have
highlighted racial bias with speech-to-text algorithms and performance on minority
accents lags significantly. To our knowledge, there is no publicly available research or
benchmark on accented African clinical ASR, and speech data is non-existent for the
majority of African accents. We release AfriSpeech, 200hrs of Pan-African speech,
67,577 clips from 2,463 unique speakers, across 120 indigenous accents from 13 countries for
clinical and general domain ASR, a benchmark test set, with publicly available pre-trained
models with SOTA performance on the AfriSpeech benchmark.
### Source Data
#### Country Stats
| Country | Clips | Speakers | Duration (seconds) | Duration (hrs) |
| ----------- | ----------- | ----------- | ----------- | ----------- |
| NG | 45875 | 1979 | 512646.88 | 142.40 |
| KE | 8304 | 137 | 75195.43 | 20.89 |
| ZA | 7870 | 223 | 81688.11 | 22.69 |
| GH | 2018 | 37 | 18581.13 | 5.16 |
| BW | 1391 | 38 | 14249.01 | 3.96 |
| UG | 1092 | 26 | 10420.42 | 2.89 |
| RW | 469 | 9 | 5300.99 | 1.47 |
| US | 219 | 5 | 1900.98 | 0.53 |
| TR | 66 | 1 | 664.01 | 0.18 |
| ZW | 63 | 3 | 635.11 | 0.18 |
| MW | 60 | 1 | 554.61 | 0.15 |
| TZ | 51 | 2 | 645.51 | 0.18 |
| LS | 7 | 1 | 78.40 | 0.02 |
#### Accent Stats
| Accent | Clips | Speakers | Duration (s) | Country | Splits |
| ----------- | ----------- | ----------- | ----------- | ----------- | ----------- |
| yoruba | 15407 | 683 | 161587.55 | US,NG | train,test,dev |
| igbo | 8677 | 374 | 93035.79 | US,NG,ZA | train,test,dev |
| swahili | 6320 | 119 | 55932.82 | KE,TZ,ZA,UG | train,test,dev |
| hausa | 5765 | 248 | 70878.67 | NG | train,test,dev |
| ijaw | 2499 | 105 | 33178.9 | NG | train,test,dev |
| afrikaans | 2048 | 33 | 20586.49 | ZA | train,test,dev |
| idoma | 1877 | 72 | 20463.6 | NG | train,test,dev |
| zulu | 1794 | 52 | 18216.97 | ZA,TR,LS | dev,train,test |
| setswana | 1588 | 39 | 16553.22 | BW,ZA | dev,test,train |
| twi | 1566 | 22 | 14340.12 | GH | test,train,dev |
| isizulu | 1048 | 48 | 10376.09 | ZA | test,train,dev |
| igala | 919 | 31 | 9854.72 | NG | train,test |
| izon | 838 | 47 | 9602.53 | NG | train,dev,test |
| kiswahili | 827 | 6 | 8988.26 | KE | train,test |
| ebira | 757 | 42 | 7752.94 | NG | train,test,dev |
| luganda | 722 | 22 | 6768.19 | UG,BW,KE | test,dev,train |
| urhobo | 646 | 32 | 6685.12 | NG | train,dev,test |
| nembe | 578 | 16 | 6644.72 | NG | train,test,dev |
| ibibio | 570 | 39 | 6489.29 | NG | train,test,dev |
| pidgin | 514 | 20 | 5871.57 | NG | test,train,dev |
| luhya | 508 | 4 | 4497.02 | KE | train,test |
| kinyarwanda | 469 | 9 | 5300.99 | RW | train,test,dev |
| xhosa | 392 | 12 | 4604.84 | ZA | train,dev,test |
| tswana | 387 | 18 | 4148.58 | ZA,BW | train,test,dev |
| esan | 380 | 13 | 4162.63 | NG | train,test,dev |
| alago | 363 | 8 | 3902.09 | NG | train,test |
| tshivenda | 353 | 5 | 3264.77 | ZA | test,train |
| fulani | 312 | 18 | 5084.32 | NG | test,train |
| isoko | 298 | 16 | 4236.88 | NG | train,test,dev |
| akan (fante) | 295 | 9 | 2848.54 | GH | train,dev,test |
| ikwere | 293 | 14 | 3480.43 | NG | test,train,dev |
| sepedi | 275 | 10 | 2751.68 | ZA | dev,test,train |
| efik | 269 | 11 | 2559.32 | NG | test,train,dev |
| edo | 237 | 12 | 1842.32 | NG | train,test,dev |
| luo | 234 | 4 | 2052.25 | UG,KE | test,train,dev |
| kikuyu | 229 | 4 | 1949.62 | KE | train,test,dev |
| bekwarra | 218 | 3 | 2000.46 | NG | train,test |
| isixhosa | 210 | 9 | 2100.28 | ZA | train,dev,test |
| hausa/fulani | 202 | 3 | 2213.53 | NG | test,train |
| epie | 202 | 6 | 2320.21 | NG | train,test |
| isindebele | 198 | 2 | 1759.49 | ZA | train,test |
| venda and xitsonga | 188 | 2 | 2603.75 | ZA | train,test |
| sotho | 182 | 4 | 2082.21 | ZA | dev,test,train |
| akan | 157 | 6 | 1392.47 | GH | test,train |
| nupe | 156 | 9 | 1608.24 | NG | dev,train,test |
| anaang | 153 | 8 | 1532.56 | NG | test,dev |
| english | 151 | 11 | 2445.98 | NG | dev,test |
| afemai | 142 | 2 | 1877.04 | NG | train,test |
| shona | 138 | 8 | 1419.98 | ZA,ZW | test,train,dev |
| eggon | 137 | 5 | 1833.77 | NG | test |
| luganda and kiswahili | 134 | 1 | 1356.93 | UG | train |
| ukwuani | 133 | 7 | 1269.02 | NG | test |
| sesotho | 132 | 10 | 1397.16 | ZA | train,dev,test |
| benin | 124 | 4 | 1457.48 | NG | train,test |
| kagoma | 123 | 1 | 1781.04 | NG | train |
| nasarawa eggon | 120 | 1 | 1039.99 | NG | train |
| tiv | 120 | 14 | 1084.52 | NG | train,test,dev |
| south african english | 119 | 2 | 1643.82 | ZA | train,test |
| borana | 112 | 1 | 1090.71 | KE | train |
| swahili ,luganda ,arabic | 109 | 1 | 929.46 | UG | train |
| ogoni | 109 | 4 | 1629.7 | NG | train,test |
| mada | 109 | 2 | 1786.26 | NG | test |
| bette | 106 | 4 | 930.16 | NG | train,test |
| berom | 105 | 4 | 1272.99 | NG | dev,test |
| bini | 104 | 4 | 1499.75 | NG | test |
| ngas | 102 | 3 | 1234.16 | NG | train,test |
| etsako | 101 | 4 | 1074.53 | NG | train,test |
| okrika | 100 | 3 | 1887.47 | NG | train,test |
| venda | 99 | 2 | 938.14 | ZA | train,test |
| siswati | 96 | 5 | 1367.45 | ZA | dev,train,test |
| damara | 92 | 1 | 674.43 | NG | train |
| yoruba, hausa | 89 | 5 | 928.98 | NG | test |
| southern sotho | 89 | 1 | 889.73 | ZA | train |
| kanuri | 86 | 7 | 1936.78 | NG | test,dev |
| itsekiri | 82 | 3 | 778.47 | NG | test,dev |
| ekpeye | 80 | 2 | 922.88 | NG | test |
| mwaghavul | 78 | 2 | 738.02 | NG | test |
| bajju | 72 | 2 | 758.16 | NG | test |
| luo, swahili | 71 | 1 | 616.57 | KE | train |
| dholuo | 70 | 1 | 669.07 | KE | train |
| ekene | 68 | 1 | 839.31 | NG | test |
| jaba | 65 | 2 | 540.66 | NG | test |
| ika | 65 | 4 | 576.56 | NG | test,dev |
| angas | 65 | 1 | 589.99 | NG | test |
| ateso | 63 | 1 | 624.28 | UG | train |
| brass | 62 | 2 | 900.04 | NG | test |
| ikulu | 61 | 1 | 313.2 | NG | test |
| eleme | 60 | 2 | 1207.92 | NG | test |
| chichewa | 60 | 1 | 554.61 | MW | train |
| oklo | 58 | 1 | 871.37 | NG | test |
| meru | 58 | 2 | 865.07 | KE | train,test |
| agatu | 55 | 1 | 369.11 | NG | test |
| okirika | 54 | 1 | 792.65 | NG | test |
| igarra | 54 | 1 | 562.12 | NG | test |
| ijaw(nembe) | 54 | 2 | 537.56 | NG | test |
| khana | 51 | 2 | 497.42 | NG | test |
| ogbia | 51 | 4 | 461.15 | NG | test,dev |
| gbagyi | 51 | 4 | 693.43 | NG | test |
| portuguese | 50 | 1 | 525.02 | ZA | train |
| delta | 49 | 2 | 425.76 | NG | test |
| bassa | 49 | 1 | 646.13 | NG | test |
| etche | 49 | 1 | 637.48 | NG | test |
| kubi | 46 | 1 | 495.21 | NG | test |
| jukun | 44 | 2 | 362.12 | NG | test |
| igbo and yoruba | 43 | 2 | 466.98 | NG | test |
| urobo | 43 | 3 | 573.14 | NG | test |
| kalabari | 42 | 5 | 305.49 | NG | test |
| ibani | 42 | 1 | 322.34 | NG | test |
| obolo | 37 | 1 | 204.79 | NG | test |
| idah | 34 | 1 | 533.5 | NG | test |
| bassa-nge/nupe | 31 | 3 | 267.42 | NG | test,dev |
| yala mbembe | 29 | 1 | 237.27 | NG | test |
| eket | 28 | 1 | 238.85 | NG | test |
| afo | 26 | 1 | 171.15 | NG | test |
| ebiobo | 25 | 1 | 226.27 | NG | test |
| nyandang | 25 | 1 | 230.41 | NG | test |
| ishan | 23 | 1 | 194.12 | NG | test |
| bagi | 20 | 1 | 284.54 | NG | test |
| estako | 20 | 1 | 480.78 | NG | test |
| gerawa | 13 | 1 | 342.15 | NG | test |
#### Initial Data Collection and Normalization
[Needs More Information]
#### Who are the source language producers?
[Needs More Information]
### Annotations
#### Annotation process
[Needs More Information]
#### Who are the annotators?
[Needs More Information]
### Personal and Sensitive Information
The dataset consists of people who have donated their voice online. You agree to not attempt to determine the identity of speakers in this dataset.
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
Dataset provided for research purposes only. Please check dataset license for additional information.
## Additional Information
### Dataset Curators
The dataset was initially prepared by Intron and refined for public release by CLAIR Lab.
### Licensing Information
Public Domain, Creative Commons Attribution NonCommercial ShareAlike v4.0 ([CC BY-NC-SA 4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/legalcode))
### Citation Information
[More Information Needed]
### Contributions
Thanks to [@tobiolatunji](https://github.com/tobiolatunji) for adding this dataset. |
KETI-AIR/coco | 2023-03-22T11:45:13.000Z | [
"task_categories:object-detection",
"size_categories:100K<n<1M",
"language:en",
"license:apache-2.0",
"region:us"
] | KETI-AIR | COCO is a large-scale object detection, segmentation, and
captioning dataset.
Note:
* Some images from the train and validation sets don't have annotations.
* Coco 2014 and 2017 uses the same images, but different train/val/test splits
* The test split don't have any annotations (only images).
* Coco defines 91 classes but the data only uses 80 classes.
* Panotptic annotations defines defines 200 classes but only uses 133. | @article{DBLP:journals/corr/LinMBHPRDZ14,
author = {Tsung{-}Yi Lin and
Michael Maire and
Serge J. Belongie and
Lubomir D. Bourdev and
Ross B. Girshick and
James Hays and
Pietro Perona and
Deva Ramanan and
Piotr Doll{\'{a}}r and
C. Lawrence Zitnick},
title = {Microsoft {COCO:} Common Objects in Context},
journal = {CoRR},
volume = {abs/1405.0312},
year = {2014},
url = {http://arxiv.org/abs/1405.0312},
archivePrefix = {arXiv},
eprint = {1405.0312},
timestamp = {Mon, 13 Aug 2018 16:48:13 +0200},
biburl = {https://dblp.org/rec/bib/journals/corr/LinMBHPRDZ14},
bibsource = {dblp computer science bibliography, https://dblp.org}
} | null | 0 | 138 | ---
license: apache-2.0
task_categories:
- object-detection
language:
- en
size_categories:
- 100K<n<1M
pretty_name: Coco
---
# Coco dataset loader based on tensorflow dataset coco
## Object Detection
```python
import os
from datasets import load_dataset
from PIL import Image, ImageFont, ImageDraw, ImageColor
def calc_lum(rgb):
return (0.2126*rgb[0] + 0.7152*rgb[1] + 0.0722*rgb[2])
COLOR_MAP = [ImageColor.getrgb(code) for name, code in ImageColor.colormap.items()]
def get_text_bbox(bb, tbb, margin, im_w, im_h, anchor="leftBottom"):
m = margin
l, t, r, b = bb
tl, tt, tr, tb = tbb
bbw, bbh = r - l, b - t
tbbw, tbbh = tr - tl, tb - tt
# bbox (left-top)
if anchor == "leftTop":
ax, ay = l, t
if tbbw*3 > bbw or tbbh*4 > bbh:
# align (text box: left-bottom)
x1, y1 = max(ax, 0), max(ay - tb - 2*m, 0)
x2, y2 = min(x1 + tr + 2*m, im_w), min(y1 + tb + 2*m, im_h)
return ((x1, y1, x2, y2), (max(x1+m, 0), max(y1+m, 0)))
else:
# align (text box: left-top)
x1, y1 = max(ax, 0), max(ay, 0)
x2, y2 = min(x1 + tr + 2*m, im_w), min(y1 + tb + 2*m, im_h)
return (( x1, y1, x2, y2), (max(x1+m, 0), max(y1+m, 0)))
elif anchor == "rightTop":
ax, ay = r, t
if tbbw*3 > bbw or tbbh*4 > bbh:
# align (text box: left-bottom)
x2, y1 = max(ax, 0), max(ay - tb - 2*m, 0)
x1, y2 = max(x2 - tr - 2*m, 0), min(y1 + tb + 2*m, im_h)
return ((x1, y1, x2, y2), (max(x1+m, 0), max(y1+m, 0)))
else:
# align (text box: left-top)
x2, y1 = max(ax, 0), max(ay, 0)
x1, y2 = max(x2 - tr - 2*m, 0), min(y1 + tb + 2*m, im_h)
return ((x1, y1, x2, y2), (max(x1+m, 0), max(y1+m, 0)))
elif anchor == "rightBottom":
ax, ay = r, b
if tbbw*3 > bbw or tbbh*4 > bbh:
# align (text box: left-top)
x2, y2 = min(ax, im_w), min(ay + tb + 2*m, im_h)
x1, y1 = max(x2 - tr - 2*m, 0), max(y2 - tb - 2*m, 0)
return ((x1, y1, x2, y2), (max(x1+m, 0), max(y1+m, 0)))
else:
# align (text box: left-bottom)
x2, y2 = min(ax, im_w), max(ay, 0)
x1, y1 = max(x2 - tr - 2*m, 0), max(y2 - tb - 2*m, 0)
return ((x1, y1, x2, y2), (max(x1+m, 0), max(y1+m, 0)))
elif anchor == "leftBottom":
ax, ay = l, b
if tbbw*3 > bbw or tbbh*4 > bbh:
# align (text box: left-top)
x1, y2 = min(ax, im_w), min(ay + tb + 2*m, im_h)
x2, y1 = min(x1 + tr + 2*m, im_w), max(y2 - tb - 2*m, 0)
return ((x1, y1, x2, y2), (max(x1+m, 0), max(y1+m, 0)))
else:
# align (text box: left-bottom)
x1, y2 = min(ax, im_w), max(ay, 0)
x2, y1 = min(x1 + tr + 2*m, im_w), max(y2 - tb - 2*m, 0)
return ((x1, y1, x2, y2), (max(x1+m, 0), max(y1+m, 0)))
elif anchor == "centerBottom":
ax, ay = (l+r)//2, b
if tbbw*3 > bbw or tbbh*4 > bbh:
# align (text box: left-top)
x1, y2 = min(ax - tr//2 - m, im_w), min(ay + tb + 2*m, im_h)
x2, y1 = min(x1 + tr + 2*m, im_w), max(y2 - tb - 2*m, 0)
return ((x1, y1, x2, y2), (max(x1+m, 0), max(y1+m, 0)))
else:
# align (text box: left-bottom)
x1, y2 = min(ax - tr//2 - m, im_w), max(ay, 0)
x2, y1 = min(x1 + tr + 2*m, im_w), max(y2 - tb - 2*m, 0)
return ((x1, y1, x2, y2), (max(x1+m, 0), max(y1+m, 0)))
def draw_bbox(image, objects, out_path, label_names=None, font="Roboto-Bold.ttf", fontsize=15, fill=True, opacity=60, width=2, margin=3, anchor="leftBottom"):
fnt = ImageFont.truetype(font, fontsize)
im_w, im_h = image.size
img = image.convert("RGBA")
overlay = Image.new('RGBA', img.size, (0, 0, 0, 0))
draw = ImageDraw.Draw(overlay)
for bb, lbl_id in zip(objects["bbox"], objects["label"]):
c = COLOR_MAP[min(lbl_id, len(COLOR_MAP)-1)]
fill_c = c + (opacity, ) if fill else None
draw.rectangle((bb[0], bb[1], bb[2], bb[3]), outline=c, fill=fill_c, width=width)
text = ""
if label_names is not None:
text = label_names[lbl_id]
tbb = fnt.getbbox(text)
btn_bbox, text_pos = get_text_bbox(bb, tbb, margin, im_w, im_h, anchor)
fc = (0, 0, 0) if calc_lum(c) > 150 else (255, 255, 255)
draw.rectangle(btn_bbox, outline=c, fill=c + (255, ))
draw.text(text_pos, text, font=fnt, fill=fc + (255, ))
img = Image.alpha_composite(img, overlay)
overlay = Image.new('RGBA', img.size, (0, 0, 0, 0))
draw = ImageDraw.Draw(overlay)
img = img.convert("RGB")
img.save(out_path)
raw_datasets = load_dataset(
"coco.py",
"2017",
cache_dir="./huggingface_datasets",
)
train_dataset = raw_datasets["train"]
label_list = raw_datasets["train"].features["objects"].feature['label'].names
for idx, item in zip(range(10), train_dataset):
draw_bbox(item["image"], item["objects"], item["image/filename"], label_list)
```


## Panoptic segmentation
```python
import numpy as np
from datasets import load_dataset
from PIL import Image, ImageFont, ImageDraw, ImageColor
from transformers.image_transforms import (
rgb_to_id,
)
def calc_lum(rgb):
return (0.2126*rgb[0] + 0.7152*rgb[1] + 0.0722*rgb[2])
COLOR_MAP = [ImageColor.getrgb(code) for name, code in ImageColor.colormap.items()]
def get_text_bbox(bb, tbb, margin, im_w, im_h, anchor="leftBottom"):
m = margin
l, t, r, b = bb
tl, tt, tr, tb = tbb
bbw, bbh = r - l, b - t
tbbw, tbbh = tr - tl, tb - tt
# bbox (left-top)
if anchor == "leftTop":
ax, ay = l, t
if tbbw*3 > bbw or tbbh*4 > bbh:
# align (text box: left-bottom)
x1, y1 = max(ax, 0), max(ay - tb - 2*m, 0)
x2, y2 = min(x1 + tr + 2*m, im_w), min(y1 + tb + 2*m, im_h)
return ((x1, y1, x2, y2), (max(x1+m, 0), max(y1+m, 0)))
else:
# align (text box: left-top)
x1, y1 = max(ax, 0), max(ay, 0)
x2, y2 = min(x1 + tr + 2*m, im_w), min(y1 + tb + 2*m, im_h)
return (( x1, y1, x2, y2), (max(x1+m, 0), max(y1+m, 0)))
elif anchor == "rightTop":
ax, ay = r, t
if tbbw*3 > bbw or tbbh*4 > bbh:
# align (text box: left-bottom)
x2, y1 = max(ax, 0), max(ay - tb - 2*m, 0)
x1, y2 = max(x2 - tr - 2*m, 0), min(y1 + tb + 2*m, im_h)
return ((x1, y1, x2, y2), (max(x1+m, 0), max(y1+m, 0)))
else:
# align (text box: left-top)
x2, y1 = max(ax, 0), max(ay, 0)
x1, y2 = max(x2 - tr - 2*m, 0), min(y1 + tb + 2*m, im_h)
return ((x1, y1, x2, y2), (max(x1+m, 0), max(y1+m, 0)))
elif anchor == "rightBottom":
ax, ay = r, b
if tbbw*3 > bbw or tbbh*4 > bbh:
# align (text box: left-top)
x2, y2 = min(ax, im_w), min(ay + tb + 2*m, im_h)
x1, y1 = max(x2 - tr - 2*m, 0), max(y2 - tb - 2*m, 0)
return ((x1, y1, x2, y2), (max(x1+m, 0), max(y1+m, 0)))
else:
# align (text box: left-bottom)
x2, y2 = min(ax, im_w), max(ay, 0)
x1, y1 = max(x2 - tr - 2*m, 0), max(y2 - tb - 2*m, 0)
return ((x1, y1, x2, y2), (max(x1+m, 0), max(y1+m, 0)))
elif anchor == "leftBottom":
ax, ay = l, b
if tbbw*3 > bbw or tbbh*4 > bbh:
# align (text box: left-top)
x1, y2 = min(ax, im_w), min(ay + tb + 2*m, im_h)
x2, y1 = min(x1 + tr + 2*m, im_w), max(y2 - tb - 2*m, 0)
return ((x1, y1, x2, y2), (max(x1+m, 0), max(y1+m, 0)))
else:
# align (text box: left-bottom)
x1, y2 = min(ax, im_w), max(ay, 0)
x2, y1 = min(x1 + tr + 2*m, im_w), max(y2 - tb - 2*m, 0)
return ((x1, y1, x2, y2), (max(x1+m, 0), max(y1+m, 0)))
elif anchor == "centerBottom":
ax, ay = (l+r)//2, b
if tbbw*3 > bbw or tbbh*4 > bbh:
# align (text box: left-top)
x1, y2 = min(ax - tr//2 - m, im_w), min(ay + tb + 2*m, im_h)
x2, y1 = min(x1 + tr + 2*m, im_w), max(y2 - tb - 2*m, 0)
return ((x1, y1, x2, y2), (max(x1+m, 0), max(y1+m, 0)))
else:
# align (text box: left-bottom)
x1, y2 = min(ax - tr//2 - m, im_w), max(ay, 0)
x2, y1 = min(x1 + tr + 2*m, im_w), max(y2 - tb - 2*m, 0)
return ((x1, y1, x2, y2), (max(x1+m, 0), max(y1+m, 0)))
# Copied from transformers.models.detr.image_processing_detr.masks_to_boxes
def masks_to_boxes(masks: np.ndarray) -> np.ndarray:
"""
Compute the bounding boxes around the provided panoptic segmentation masks.
Args:
masks: masks in format `[number_masks, height, width]` where N is the number of masks
Returns:
boxes: bounding boxes in format `[number_masks, 4]` in xyxy format
"""
if masks.size == 0:
return np.zeros((0, 4))
h, w = masks.shape[-2:]
y = np.arange(0, h, dtype=np.float32)
x = np.arange(0, w, dtype=np.float32)
# see https://github.com/pytorch/pytorch/issues/50276
y, x = np.meshgrid(y, x, indexing="ij")
x_mask = masks * np.expand_dims(x, axis=0)
x_max = x_mask.reshape(x_mask.shape[0], -1).max(-1)
x = np.ma.array(x_mask, mask=~(np.array(masks, dtype=bool)))
x_min = x.filled(fill_value=1e8)
x_min = x_min.reshape(x_min.shape[0], -1).min(-1)
y_mask = masks * np.expand_dims(y, axis=0)
y_max = y_mask.reshape(x_mask.shape[0], -1).max(-1)
y = np.ma.array(y_mask, mask=~(np.array(masks, dtype=bool)))
y_min = y.filled(fill_value=1e8)
y_min = y_min.reshape(y_min.shape[0], -1).min(-1)
return np.stack([x_min, y_min, x_max, y_max], 1)
def draw_seg(image, panoptic_image, oids, labels, out_path, label_names=None, font="Roboto-Bold.ttf", fontsize=15, opacity=160, anchor="leftBottom"):
fnt = ImageFont.truetype(font, fontsize)
im_w, im_h = image.size
masks = np.asarray(panoptic_image, dtype=np.uint32)
masks = rgb_to_id(masks)
oids = np.array(oids, dtype=np.uint32)
masks = masks == oids[:, None, None]
masks = masks.astype(np.uint8)
bboxes = masks_to_boxes(masks)
img = image.convert("RGBA")
for label, mask, bbox in zip(labels, masks, bboxes):
c = COLOR_MAP[min(label, len(COLOR_MAP)-1)]
cf = np.array(c + (opacity, )).astype(np.uint8)
cmask = mask[:, :, None] * cf[None, None, :]
cmask = Image.fromarray(cmask)
img = Image.alpha_composite(img, cmask)
if label_names is not None:
text = label_names[label]
tbb = fnt.getbbox(text)
btn_bbox, text_pos = get_text_bbox(bbox, tbb, 3, im_w, im_h, anchor=anchor)
overlay = Image.new('RGBA', img.size, (0, 0, 0, 0))
draw = ImageDraw.Draw(overlay)
fc = (0, 0, 0) if calc_lum(c) > 150 else (255, 255, 255)
draw.rectangle(btn_bbox, outline=c, fill=c + (255, ))
draw.text(text_pos, text, font=fnt, fill=fc + (255, ))
img = Image.alpha_composite(img, overlay)
img = img.convert("RGB")
img.save(out_path)
raw_datasets = load_dataset(
"coco.py",
"2017_panoptic",
cache_dir="./huggingface_datasets",
# data_dir="./data",
)
train_dataset = raw_datasets["train"]
label_list = raw_datasets["train"].features["panoptic_objects"].feature['label'].names
for idx, item in zip(range(10), train_dataset):
draw_seg(
item["image"],
item["panoptic_image"],
item["panoptic_objects"]["id"],
item["panoptic_objects"]["label"],
"panoptic_" + item["image/filename"],
label_list)
```


|
Thaweewat/databricks-dolly-15k-th | 2023-05-09T16:15:52.000Z | [
"task_categories:question-answering",
"task_categories:summarization",
"size_categories:10K<n<100K",
"language:th",
"license:cc-by-sa-3.0",
"instruction-finetuning",
"region:us"
] | Thaweewat | null | null | null | 1 | 138 | ---
license: cc-by-sa-3.0
task_categories:
- question-answering
- summarization
tags:
- instruction-finetuning
language:
- th
size_categories:
- 10K<n<100K
---
# Summary
This is a Thai 🇹🇭-instructed dataset translated from `databricks-dolly-15k` using Google Cloud Translation.
`databricks-dolly-15k` is an open-source dataset of instruction-following records generated by thousands of Databricks employees in several behavioral
categories outlined in the InstructGPT paper, including brainstorming, classification, closed QA, generation, information extraction, open QA, and summarization.
This dataset can be used for any purpose, whether academic or commercial, under the terms of the
[Creative Commons Attribution-ShareAlike 3.0 Unported License](https://creativecommons.org/licenses/by-sa/3.0/legalcode).
Supported Tasks:
- Training LLMs
- Synthetic Data Generation
- Data Augmentation
Languages: Thai
Version: 1.0
---
|
shumpei2525/fine_tuning521k-ja | 2023-07-02T18:06:01.000Z | [
"license:mit",
"region:us"
] | shumpei2525 | null | null | null | 10 | 138 | ---
license: mit
---
# fine_tuning521k-ja
This data is a dataset for fine-tuning the local language model (LLM). It consists of the translation of "ign_clean_instruct_dataset_500k" and "GPTeacher." Please feel free to use it. This dataset contains data such as Q&A, contextualized questions, role plays. Please contact us if you encounter any issues.
# Since I'm not entirely clear on OpenAI's terms of service, please be cautious when using it for commercial purposes. There may be exceptions for non-commercial use.
# original datasets
This software is a modified version of the original software licensed under the MIT License. Original Copyright (c) 2023 Teknium. All rights reserved. This software includes modifications made by Teknium to the original software licensed under the MIT License. Modified portions of this software are translating English into Japanese.
Copyright (c) 2023 Teknium License Agreement: https://github.com/teknium1/GPTeacher/blob/main/LICENSE
This datasets include the work that is distributed in the Apache License 2.0. https://huggingface.co/datasets/ignmilton/ign_clean_instruct_dataset_500k
License Agreement:
Apache License Version 2.0, January 2004 http://www.apache.org/licenses/
TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
Definitions.
"License" shall mean the terms and conditions for use, reproduction, and distribution as defined by Sections 1 through 9 of this document.
"Licensor" shall mean the copyright owner or entity authorized by the copyright owner that is granting the License.
"Legal Entity" shall mean the union of the acting entity and all other entities that control, are controlled by, or are under common control with that entity. For the purposes of this definition, "control" means (i) the power, direct or indirect, to cause the direction or management of such entity, whether by contract or otherwise, or (ii) ownership of fifty percent (50%) or more of the outstanding shares, or (iii) beneficial ownership of such entity.
"You" (or "Your") shall mean an individual or Legal Entity exercising permissions granted by this License.
"Source" form shall mean the preferred form for making modifications, including but not limited to software source code, documentation source, and configuration files.
"Object" form shall mean any form resulting from mechanical transformation or translation of a Source form, including but not limited to compiled object code, generated documentation, and conversions to other media types.
"Work" shall mean the work of authorship, whether in Source or Object form, made available under the License, as indicated by a copyright notice that is included in or attached to the work (an example is provided in the Appendix below).
"Derivative Works" shall mean any work, whether in Source or Object form, that is based on (or derived from) the Work and for which the editorial revisions, annotations, elaborations, or other modifications represent, as a whole, an original work of authorship. For the purposes of this License, Derivative Works shall not include works that remain separable from, or merely link (or bind by name) to the interfaces of, the Work and Derivative Works thereof.
"Contribution" shall mean any work of authorship, including the original version of the Work and any modifications or additions to that Work or Derivative Works thereof, that is intentionally submitted to Licensor for inclusion in the Work by the copyright owner or by an individual or Legal Entity authorized to submit on behalf of the copyright owner. For the purposes of this definition, "submitted" means any form of electronic, verbal, or written communication sent to the Licensor or its representatives, including but not limited to communication on electronic mailing lists, source code control systems, and issue tracking systems that are managed by, or on behalf of, the Licensor for the purpose of discussing and improving the Work, but excluding communication that is conspicuously marked or otherwise designated in writing by the copyright owner as "Not a Contribution."
"Contributor" shall mean Licensor and any individual or Legal Entity on behalf of whom a Contribution has been received by Licensor and subsequently incorporated within the Work.
Grant of Copyright License. Subject to the terms and conditions of this License, each Contributor hereby grants to You a perpetual, worldwide, non-exclusive, no-charge, royalty-free, irrevocable copyright license to reproduce, prepare Derivative Works of, publicly display, publicly perform, sublicense, and distribute the Work and such Derivative Works in Source or Object form.
Grant of Patent License. Subject to the terms and conditions of this License, each Contributor hereby grants to You a perpetual, worldwide, non-exclusive, no-charge, royalty-free, irrevocable (except as stated in this section) patent license to make, have made, use, offer to sell, sell, import, and otherwise transfer the Work, where such license applies only to those patent claims licensable by such Contributor that are necessarily infringed by their Contribution(s) alone or by combination of their Contribution(s) with the Work to which such Contribution(s) was submitted. If You institute patent litigation against any entity (including a cross-claim or counterclaim in a lawsuit) alleging that the Work or a Contribution incorporated within the Work constitutes direct or contributory patent infringement, then any patent licenses granted to You under this License for that Work shall terminate as of the date such litigation is filed.
Redistribution. You may reproduce and distribute copies of the Work or Derivative Works thereof in any medium, with or without modifications, and in Source or Object form, provided that You meet the following conditions:
You must give any other recipients of the Work or Derivative Works a copy of this License; and You must cause any modified files to carry prominent notices stating that You changed the files; and You must retain, in the Source form of any Derivative Works that You distribute, all copyright, patent, trademark, and attribution notices from the Source form of the Work, excluding those notices that do not pertain to any part of the Derivative Works; and If the Work includes a "NOTICE" text file as part of its distribution, then any Derivative Works that You distribute must include a readable copy of the attribution notices contained within such NOTICE file, excluding those notices that do not pertain to any part of the Derivative Works, in at least one of the following places: within a NOTICE text file distributed as part of the Derivative Works; within the Source form or documentation, if provided along with the Derivative Works; or, within a display generated by the Derivative Works, if and wherever such third-party notices normally appear. The contents of the NOTICE file are for informational purposes only and do not modify the License. You may add Your own attribution notices within Derivative Works that You distribute, alongside or as an addendum to the NOTICE text from the Work, provided that such additional attribution notices cannot be construed as modifying the License.
You may add Your own copyright statement to Your modifications and may provide additional or different license terms and conditions for use, reproduction, or distribution of Your modifications, or for any such Derivative Works as a whole, provided Your use, reproduction, and distribution of the Work otherwise complies with the conditions stated in this License. 5. Submission of Contributions. Unless You explicitly state otherwise, any Contribution intentionally submitted for inclusion in the Work by You to the Licensor shall be under the terms and conditions of this License, without any additional terms or conditions. Notwithstanding the above, nothing herein shall supersede or modify the terms of any separate license agreement you may have executed with Licensor regarding such Contributions.
Trademarks. This License does not grant permission to use the trade names, trademarks, service marks, or product names of the Licensor, except as required for reasonable and customary use in describing the origin of the Work and reproducing the content of the NOTICE file.
Disclaimer of Warranty. Unless required by applicable law or agreed to in writing, Licensor provides the Work (and each Contributor provides its Contributions) on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied, including, without limitation, any warranties or conditions of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A PARTICULAR PURPOSE. You are solely responsible for determining the appropriateness of using or redistributing the Work and assume any risks associated with Your exercise of permissions under this License.
Limitation of Liability. In no event and under no legal theory, whether in tort (including negligence), contract, or otherwise, unless required by applicable law (such as deliberate and grossly negligent acts) or agreed to in writing, shall any Contributor be liable to You for damages, including any direct, indirect, special, incidental, or consequential damages of any character arising as a result of this License or out of the use or inability to use the Work (including but not limited to damages for loss of goodwill, work stoppage, computer failure or malfunction, or any and all other commercial damages or losses), even if such Contributor has been advised of the possibility of such damages.
Accepting Warranty or Additional Liability. While redistributing the Work or Derivative Works thereof, You may choose to offer, and charge a fee for, acceptance of support, warranty, indemnity, or other liability obligations and/or rights consistent with this License. However, in accepting such obligations, You may act only on Your own behalf and on Your sole responsibility, not on behalf of any other Contributor, and only if You agree to indemnify, defend, and hold each Contributor harmless for any liability incurred by, or claims asserted against, such Contributor by reason of your accepting any such warranty or additional liability.
END OF TERMS AND CONDITIONS |
chats-bug/input_tools_plans | 2023-09-12T10:40:23.000Z | [
"region:us"
] | chats-bug | null | null | null | 1 | 138 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
dataset_info:
features:
- name: prompt
dtype: string
splits:
- name: train
num_bytes: 29938684.365363304
num_examples: 10107
- name: test
num_bytes: 7485411.634636695
num_examples: 2527
download_size: 8585693
dataset_size: 37424096.0
---
# Dataset Card for "input_tools_plans"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
joey234/affixal_negation | 2023-10-10T22:44:14.000Z | [
"task_categories:text-classification",
"size_categories:1K<n<10K",
"language:en",
"license:apache-2.0",
"region:us"
] | joey234 | null | null | null | 0 | 138 | ---
license: apache-2.0
task_categories:
- text-classification
language:
- en
pretty_name: e
size_categories:
- 1K<n<10K
---
# Dataset Card for Dataset Name
## Dataset Description
- **Homepage:**
- **Repository:**
- **Paper:**
- **Leaderboard:**
- **Point of Contact:**
### Dataset Summary
- This dataset contains a list of affixal negations and their non-negated counterpart (e.g. unintended - intended).
- This dataset is from [van Son et al. (2016)](https://aclanthology.org/W16-5007/).
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed] |
tomekkorbak/detoxify-pile-chunk3-800000-850000 | 2022-10-04T22:47:07.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 137 | Entry not found |
tomekkorbak/detoxify-pile-chunk3-900000-950000 | 2022-10-04T23:47:24.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 137 | Entry not found |
tomekkorbak/detoxify-pile-chunk3-950000-1000000 | 2022-10-04T22:55:50.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 137 | Entry not found |
tomekkorbak/detoxify-pile-chunk3-1150000-1200000 | 2022-10-04T23:45:42.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 137 | Entry not found |
tomekkorbak/detoxify-pile-chunk3-1200000-1250000 | 2022-10-04T23:47:33.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 137 | Entry not found |
BelleGroup/train_0.5M_CN | 2023-04-03T08:11:22.000Z | [
"task_categories:text2text-generation",
"size_categories:100K<n<1M",
"language:zh",
"license:gpl-3.0",
"region:us"
] | BelleGroup | null | null | null | 68 | 137 | ---
license: gpl-3.0
task_categories:
- text2text-generation
language:
- zh
size_categories:
- 100K<n<1M
---
## 内容
包含约50万条由[BELLE](https://github.com/LianjiaTech/BELLE)项目生成的中文指令数据。
## 样例
```
{
"instruction": "给定一个文字输入,将其中的所有数字加1。\n“明天的会议在9点开始,记得准时到达。”\n",
"input": "",
"output": "“明天的会议在10点开始,记得准时到达。”"
}
```
### 字段:
```
instruction: 指令
input: 输入(本数据集均为空)
output: 输出
```
## 使用限制
仅允许将此数据集及使用此数据集生成的衍生物用于研究目的,不得用于商业,以及其他会对社会带来危害的用途。
本数据集不代表任何一方的立场、利益或想法,无关任何团体的任何类型的主张。因使用本数据集带来的任何损害、纠纷,本项目不承担任何责任。
|
CM/codexglue_code2text_ruby | 2023-04-22T01:52:59.000Z | [
"region:us"
] | CM | null | null | null | 0 | 137 | ---
dataset_info:
features:
- name: id
dtype: int32
- name: repo
dtype: string
- name: path
dtype: string
- name: func_name
dtype: string
- name: original_string
dtype: string
- name: language
dtype: string
- name: code
dtype: string
- name: code_tokens
sequence: string
- name: docstring
dtype: string
- name: docstring_tokens
sequence: string
- name: sha
dtype: string
- name: url
dtype: string
splits:
- name: train
num_bytes: 51956439
num_examples: 24927
- name: validation
num_bytes: 2821037
num_examples: 1400
- name: test
num_bytes: 2671551
num_examples: 1261
download_size: 21921316
dataset_size: 57449027
---
# Dataset Card for "codexglue_code2text_ruby"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
open-phi/programming_books_llama | 2023-10-04T18:02:56.000Z | [
"region:us"
] | open-phi | null | null | null | 5 | 137 | ---
dataset_info:
features:
- name: topic
dtype: string
- name: outline
sequence: string
- name: concepts
sequence: string
- name: queries
sequence: string
- name: context
sequence: string
- name: markdown
dtype: string
- name: model
dtype: string
splits:
- name: train
num_bytes: 1677240291
num_examples: 111048
download_size: 631279270
dataset_size: 1677240291
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "programming_books_llama"
400M tokens of programming books generated by gpt-3.5 (70M tokens) and a finetuned codellama 34b. The gpt-3.5 data is extremely high quality. The llama data has lower quality and shorter length, but is still good. This was generated with the [textbook quality](https://github.com/VikParuchuri/textbook_quality) repo. |
onestop_qa | 2023-01-25T14:42:12.000Z | [
"task_categories:question-answering",
"task_ids:multiple-choice-qa",
"annotations_creators:expert-generated",
"language_creators:expert-generated",
"multilinguality:monolingual",
"size_categories:1K<n<10K",
"source_datasets:original",
"source_datasets:extended|onestop_english",
"language:en",
"license:cc-by-sa-4.0",
"arxiv:2004.14797",
"region:us"
] | null | OneStopQA is a multiple choice reading comprehension dataset annotated according to the STARC (Structured Annotations for Reading Comprehension) scheme. The reading materials are Guardian articles taken from the [OneStopEnglish corpus](https://github.com/nishkalavallabhi/OneStopEnglishCorpus). Each article comes in three difficulty levels, Elementary, Intermediate and Advanced. Each paragraph is annotated with three multiple choice reading comprehension questions. The reading comprehension questions can be answered based on any of the three paragraph levels. | @inproceedings{starc2020,
author = {Berzak, Yevgeni and Malmaud, Jonathan and Levy, Roger},
title = {STARC: Structured Annotations for Reading Comprehension},
booktitle = {ACL},
year = {2020},
publisher = {Association for Computational Linguistics}
} | null | 4 | 136 | ---
annotations_creators:
- expert-generated
language_creators:
- expert-generated
language:
- en
license:
- cc-by-sa-4.0
multilinguality:
- monolingual
size_categories:
- 1K<n<10K
source_datasets:
- original
- extended|onestop_english
task_categories:
- question-answering
task_ids:
- multiple-choice-qa
paperswithcode_id: onestopqa
pretty_name: OneStopQA
language_bcp47:
- en-US
dataset_info:
features:
- name: title
dtype: string
- name: paragraph
dtype: string
- name: level
dtype:
class_label:
names:
'0': Adv
'1': Int
'2': Ele
- name: question
dtype: string
- name: paragraph_index
dtype: int32
- name: answers
sequence: string
length: 4
- name: a_span
sequence: int32
- name: d_span
sequence: int32
splits:
- name: train
num_bytes: 1423090
num_examples: 1458
download_size: 118173
dataset_size: 1423090
---
# Dataset Card for OneStopQA
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-instances)
- [Data Splits](#data-instances)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [OneStopQA repository](https://github.com/berzak/onestop-qa)
- **Repository:** [OneStopQA repository](https://github.com/berzak/onestop-qa)
- **Paper:** [STARC: Structured Annotations for Reading Comprehension](https://arxiv.org/abs/2004.14797)
- **Leaderboard:** [Needs More Information]
- **Point of Contact:** [Needs More Information]
### Dataset Summary
OneStopQA is a multiple choice reading comprehension dataset annotated according to the STARC (Structured Annotations for Reading Comprehension) scheme. The reading materials are Guardian articles taken from the [OneStopEnglish corpus](https://github.com/nishkalavallabhi/OneStopEnglishCorpus). Each article comes in three difficulty levels, Elementary, Intermediate and Advanced. Each paragraph is annotated with three multiple choice reading comprehension questions. The reading comprehension questions can be answered based on any of the three paragraph levels.
### Supported Tasks and Leaderboards
[Needs More Information]
### Languages
English (`en-US`).
The original Guardian articles were manually converted from British to American English.
## Dataset Structure
### Data Instances
An example of instance looks as follows.
```json
{
"title": "101-Year-Old Bottle Message",
"paragraph": "Angela Erdmann never knew her grandfather. He died in 1946, six years before she was born. But, on Tuesday 8th April, 2014, she described the extraordinary moment when she received a message in a bottle, 101 years after he had lobbed it into the Baltic Sea. Thought to be the world’s oldest message in a bottle, it was presented to Erdmann by the museum that is now exhibiting it in Germany.",
"paragraph_index": 1,
"level": "Adv",
"question": "How did Angela Erdmann find out about the bottle?",
"answers": ["A museum told her that they had it",
"She coincidentally saw it at the museum where it was held",
"She found it in her basement on April 28th, 2014",
"A friend told her about it"],
"a_span": [56, 70],
"d_span": [16, 34]
}
```
Where,
| Answer | Description | Textual Span |
|--------|------------------------------------------------------------|-----------------|
| a | Correct answer. | Critical Span |
| b | Incorrect answer. A miscomprehension of the critical span. | Critical Span |
| c | Incorrect answer. Refers to an additional span. | Distractor Span |
| d | Incorrect answer. Has no textual support. | - |
The order of the answers in the `answers` list corresponds to the order of the answers in the table.
### Data Fields
- `title`: A `string` feature. The article title.
- `paragraph`: A `string` feature. The paragraph from the article.
- `paragraph_index`: An `int` feature. Corresponds to the paragraph index in the article.
- `question`: A `string` feature. The given question.
- `answers`: A list of `string` feature containing the four possible answers.
- `a_span`: A list of start and end indices (inclusive) of the critical span.
- `d_span`: A list of start and end indices (inclusive) of the distractor span.
*Span indices are according to word positions after whitespace tokenization.
**In the rare case where a span is spread over multiple sections,
the span list will contain multiple instances of start and stop indices in the format:
[start_1, stop_1, start_2, stop_2,...].
### Data Splits
Articles: 30
Paragraphs: 162
Questions: 486
Question-Paragraph Level pairs: 1,458
No preconfigured split is currently provided.
## Dataset Creation
### Curation Rationale
[Needs More Information]
### Source Data
#### Initial Data Collection and Normalization
[Needs More Information]
#### Who are the source language producers?
[Needs More Information]
### Annotations
#### Annotation process
The annotation and piloting process of the dataset is described in Appendix A in
[STARC: Structured Annotations for Reading Comprehension](https://aclanthology.org/2020.acl-main.507.pdf).
#### Who are the annotators?
[Needs More Information]
### Personal and Sensitive Information
[Needs More Information]
## Considerations for Using the Data
### Social Impact of Dataset
[Needs More Information]
### Discussion of Biases
[Needs More Information]
### Other Known Limitations
[Needs More Information]
## Additional Information
### Dataset Curators
[Needs More Information]
### Licensing Information
<a rel="license" href="http://creativecommons.org/licenses/by-sa/4.0/"><img alt="Creative Commons License" style="border-width:0" src="https://i.creativecommons.org/l/by-sa/4.0/88x31.png" /></a><br />This work is licensed under a <a rel="license" href="http://creativecommons.org/licenses/by-sa/4.0/">Creative Commons Attribution-ShareAlike 4.0 International License</a>.
### Citation Information
[STARC: Structured Annotations for Reading Comprehension](http://people.csail.mit.edu/berzak/papers/acl2020.pdf)
```
@inproceedings{starc2020,
author = {Berzak, Yevgeni and Malmaud, Jonathan and Levy, Roger},
title = {STARC: Structured Annotations for Reading Comprehension},
booktitle = {ACL},
year = {2020},
publisher = {Association for Computational Linguistics}
}
```
### Contributions
Thanks to [@scaperex](https://github.com/scaperex) for adding this dataset. |
the_pile_books3 | 2023-10-09T09:06:14.000Z | [
"task_categories:text-generation",
"task_categories:fill-mask",
"task_ids:language-modeling",
"task_ids:masked-language-modeling",
"annotations_creators:no-annotation",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:100K<n<1M",
"source_datasets:original",
"language:en",
"license:mit",
"arxiv:2101.00027",
"region:us"
] | null | This dataset is Shawn Presser's work and is part of EleutherAi/The Pile dataset. This dataset contains all of bibliotik in plain .txt form, aka 197,000 books processed in exactly the same way as did for bookcorpusopen (a.k.a. books1). seems to be similar to OpenAI's mysterious "books2" dataset referenced in their papers. Unfortunately OpenAI will not give details, so we know very little about any differences. People suspect it's "all of libgen", but it's purely conjecture. | @article{pile,
title={The {P}ile: An 800GB Dataset of Diverse Text for Language Modeling},
author={Gao, Leo and Biderman, Stella and Black, Sid and Golding, Laurence and Hoppe, Travis and Foster, Charles and Phang, Jason and He, Horace and Thite, Anish and Nabeshima, Noa and Presser, Shawn and Leahy, Connor},
journal={arXiv preprint arXiv:2101.00027},
year={2020}
} | null | 119 | 136 | ---
annotations_creators:
- no-annotation
language_creators:
- found
language:
- en
license:
- mit
multilinguality:
- monolingual
pretty_name: Books3
size_categories:
- 100K<n<1M
source_datasets:
- original
task_categories:
- text-generation
- fill-mask
task_ids:
- language-modeling
- masked-language-modeling
viewer: false
dataset_info:
features:
- name: title
dtype: string
- name: text
dtype: string
config_name: plain_text
splits:
- name: train
num_bytes: 108392037000
num_examples: 196639
download_size: 39516981435
dataset_size: 108392037000
---
# Dataset Card for the_pile_books3
## Table of Contents
- [Dataset Card for the_pile_books3](#dataset-card-for-the_pile_books3)
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [|split|num examples|](#splitnum-examples)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Initial Data Collection and Normalization](#initial-data-collection-and-normalization)
- [Who are the source language producers?](#who-are-the-source-language-producers)
- [Annotations](#annotations)
- [Annotation process](#annotation-process)
- [Who are the annotators?](#who-are-the-annotators)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [GitHub](https://github.com/soskek/bookcorpus/issues/27#issuecomment-716104208)
- **Repository:** [Needs More Information]
- **Paper:** [arXiv](https://arxiv.org/abs/2101.00027)
- **Leaderboard:** [Needs More Information]
- **Point of Contact:** [Needs More Information]
### Dataset Summary
<div class="course-tip course-tip-orange bg-gradient-to-br dark:bg-gradient-to-r before:border-orange-500 dark:before:border-orange-800 from-orange-50 dark:from-gray-900 to-white dark:to-gray-950 border border-orange-50 text-orange-700 dark:text-gray-400">
<p><b>Defunct:</b> Dataset "the_pile_books3" is defunct and no longer accessible due to reported copyright infringement.</p>
</div>
This dataset is Shawn Presser's work and is part of EleutherAi/The Pile dataset.
This dataset contains all of bibliotik in plain .txt form, aka 197,000 books processed in exactly the same way as did for bookcorpusopen (a.k.a. books1). seems to be similar to OpenAI's mysterious "books2" dataset referenced in their papers. Unfortunately OpenAI will not give details, so we know very little about any differences. People suspect it's "all of libgen", but it's purely conjecture.
|download_size|36.8 Gib|
|dataset_size|100.9 Gib|
### Supported Tasks and Leaderboards
This dataset is used for Language Modeling.
### Languages
The dataset is in English.
## Dataset Structure
### Data Instances
```
{'title': '07 LEGO Ninjago - The Search For Zane (Scholastic) - Kate Howard (retail)'
'text': '\n\nTITLE PAGE\n\nFROM THE JOURNAL OF SENSEI GARMADON\n\nCHAPTER 1\n\nCHAPTER 2\n\nCHAPTER 3\n\nCHAPTER 4\n\nCHAPTER 5\n\nCHAPTER 6\n\nCHAPTER 7\n\nCHAPTER 8\n\nCHAPTER 9\n\nCOPYRIGHT\n\nThroughout Ninjago", five ninja are well-known for their speed, strength, and of course the elemental powers that help them protect our world from evil. But there are others who possess some of the same powers as the ninja. Others who may not always use their powers for good.\n\nBefore now, the ninja believed they were special. They di.......'}
```
### Data Fields
- `title`: title of the book
- `text`: text content of the book
### Data Splits
|split|num examples|
--------------------------------
|train|196640|
## Dataset Creation
### Curation Rationale
[Needs More Information]
### Source Data
#### Initial Data Collection and Normalization
[Needs More Information]
#### Who are the source language producers?
[Needs More Information]
### Annotations
#### Annotation process
[Needs More Information]
#### Who are the annotators?
[Needs More Information]
### Personal and Sensitive Information
[Needs More Information]
## Considerations for Using the Data
### Social Impact of Dataset
[Needs More Information]
### Discussion of Biases
[Needs More Information]
### Other Known Limitations
[Needs More Information]
## Additional Information
### Dataset Curators
[Needs More Information]
### Licensing Information
MIT
### Citation Information
```
@article{pile,
title={The {P}ile: An 800GB Dataset of Diverse Text for Language Modeling},
author={Gao, Leo and Biderman, Stella and Black, Sid and Golding, Laurence and Hoppe, Travis and Foster, Charles and Phang, Jason and He, Horace and Thite, Anish and Nabeshima, Noa and Presser, Shawn and Leahy, Connor},
journal={arXiv preprint arXiv:2101.00027},
year={2020}
}
```
### Contributions
Thanks to [@shawwn](https://github.com/shawwn) for creating this dataset.
Thanks to [@richarddwang](https://github.com/richarddwang) for adding this dataset. |
GEM/web_nlg | 2022-10-24T15:31:09.000Z | [
"task_categories:table-to-text",
"annotations_creators:unknown",
"language_creators:unknown",
"multilinguality:unknown",
"size_categories:unknown",
"source_datasets:original",
"language:en",
"license:cc-by-nc-4.0",
"data-to-text",
"region:us"
] | GEM | WebNLG is a bi-lingual dataset (English, Russian) of parallel DBpedia triple sets
and short texts that cover about 450 different DBpedia properties. The WebNLG data
was originally created to promote the development of RDF verbalisers able to
generate short text and to handle micro-planning (i.e., sentence segmentation and
ordering, referring expression generation, aggregation); the goal of the task is
to generate texts starting from 1 to 7 input triples which have entities in common
(so the input is actually a connected Knowledge Graph). The dataset contains about
17,000 triple sets and 45,000 crowdsourced texts in English, and 7,000 triples sets
and 19,000 crowdsourced texts in Russian. A challenging test set section with
entities and/or properties that have not been seen at training time is available. | @inproceedings{castro-ferreira20:bilin-bi-direc-webnl-shared,
title={The 2020 Bilingual, Bi-Directional WebNLG+ Shared Task Overview and Evaluation Results (WebNLG+ 2020)},
author={Castro Ferreira, Thiago and
Gardent, Claire and
Ilinykh, Nikolai and
van der Lee, Chris and
Mille, Simon and
Moussallem, Diego and
Shimorina, Anastasia},
booktitle = {Proceedings of the 3rd WebNLG Workshop on Natural Language Generation from the Semantic Web (WebNLG+ 2020)},
pages = "55--76",
year = 2020,
address = {Dublin, Ireland (Virtual)},
publisher = {Association for Computational Linguistics}} | null | 2 | 136 | ---
annotations_creators:
- unknown
language_creators:
- unknown
language:
- en
license:
- cc-by-nc-4.0
multilinguality:
- unknown
size_categories:
- unknown
source_datasets:
- original
task_categories:
- table-to-text
task_ids: []
pretty_name: web_nlg
tags:
- data-to-text
---
# Dataset Card for GEM/web_nlg
## Dataset Description
- **Homepage:** https://webnlg-challenge.loria.fr/
- **Repository:** https://gitlab.com/shimorina/webnlg-dataset
- **Paper:** http://www.aclweb.org/anthology/P17-1017, [WebNLG Challenge 2017 Report
- **Leaderboard:** https://beng.dice-research.org/gerbil/
- **Point of Contact:** [Needs More Information]
### Link to Main Data Card
You can find the main data card on the [GEM Website](https://gem-benchmark.com/data_cards/web_nlg).
### Dataset Summary
WebNLG is a bi-lingual dataset (English, Russian) of parallel DBpedia triple sets and short texts that cover about 450 different DBpedia properties. The WebNLG data was originally created to promote the development of RDF verbalisers able to generate short text and to handle micro-planning (i.e., sentence segmentation and ordering, referring expression generation, aggregation); the goal of the task is to generate texts starting from 1 to 7 input triples which have entities in common (so the input is actually a connected Knowledge Graph). The dataset contains about 17,000 triple sets and 45,000 crowdsourced texts in English, and 7,000 triples sets and 19,000 crowdsourced texts in Russian. A challenging test set section with entities and/or properties that have not been seen at training time is available.
You can load the dataset via:
```
import datasets
data = datasets.load_dataset('GEM/web_nlg')
```
The data loader can be found [here](https://huggingface.co/datasets/GEM/web_nlg).
#### website
[Website](https://webnlg-challenge.loria.fr/)
#### paper
[First Dataset Release](http://www.aclweb.org/anthology/P17-1017), [WebNLG Challenge 2017 Report](https://www.aclweb.org/anthology/W17-3518/), [WebNLG Challenge 2020 Report](https://webnlg-challenge.loria.fr/files/2020.webnlg-papers.7.pdf)
#### authors
The principle curator of the dataset is Anastasia Shimorina (Université de Lorraine / LORIA, France). Throughout the WebNLG releases, several people contributed to their construction: Claire Gardent (CNRS / LORIA, France), Shashi Narayan (Google, UK), Laura Perez-Beltrachini (University of Edinburgh, UK), Elena Khasanova, and Thiago Castro Ferreira (Federal University of Minas Gerais, Brazil).
## Dataset Overview
### Where to find the Data and its Documentation
#### Webpage
<!-- info: What is the webpage for the dataset (if it exists)? -->
<!-- scope: telescope -->
[Website](https://webnlg-challenge.loria.fr/)
#### Download
<!-- info: What is the link to where the original dataset is hosted? -->
<!-- scope: telescope -->
[Gitlab](https://gitlab.com/shimorina/webnlg-dataset)
#### Paper
<!-- info: What is the link to the paper describing the dataset (open access preferred)? -->
<!-- scope: telescope -->
[First Dataset Release](http://www.aclweb.org/anthology/P17-1017), [WebNLG Challenge 2017 Report](https://www.aclweb.org/anthology/W17-3518/), [WebNLG Challenge 2020 Report](https://webnlg-challenge.loria.fr/files/2020.webnlg-papers.7.pdf)
#### BibTex
<!-- info: Provide the BibTex-formatted reference for the dataset. Please use the correct published version (ACL anthology, etc.) instead of google scholar created Bibtex. -->
<!-- scope: microscope -->
Initial release of the dataset:
```
@inproceedings{gardent2017creating,
author = "Gardent, Claire
and Shimorina, Anastasia
and Narayan, Shashi
and Perez-Beltrachini, Laura",
title = "Creating Training Corpora for NLG Micro-Planners",
booktitle = "Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
year = "2017",
publisher = "Association for Computational Linguistics",
pages = "179--188",
location = "Vancouver, Canada",
doi = "10.18653/v1/P17-1017",
url = "http://www.aclweb.org/anthology/P17-1017"
}
```
The latest version 3.0:
```
@inproceedings{castro-ferreira20:bilin-bi-direc-webnl-shared,
title={The 2020 Bilingual, Bi-Directional WebNLG+ Shared Task Overview and Evaluation Results (WebNLG+ 2020)},
author={Castro Ferreira, Thiago and
Gardent, Claire and
Ilinykh, Nikolai and
van der Lee, Chris and
Mille, Simon and
Moussallem, Diego and
Shimorina, Anastasia},
booktitle = {Proceedings of the 3rd WebNLG Workshop on Natural Language Generation from the Semantic Web (WebNLG+ 2020)},
pages = "55--76",
year = 2020,
address = {Dublin, Ireland (Virtual)},
publisher = {Association for Computational Linguistics}}
```
#### Contact Email
<!-- info: If known, provide the email of at least one person the reader can contact for questions about the dataset. -->
<!-- scope: periscope -->
webnlg-challenge@inria.fr
#### Has a Leaderboard?
<!-- info: Does the dataset have an active leaderboard? -->
<!-- scope: telescope -->
yes
#### Leaderboard Link
<!-- info: Provide a link to the leaderboard. -->
<!-- scope: periscope -->
[Website](https://beng.dice-research.org/gerbil/)
#### Leaderboard Details
<!-- info: Briefly describe how the leaderboard evaluates models. -->
<!-- scope: microscope -->
The model outputs are evaluated against the crowdsourced references; the leaderboard reports BLEU-4, METEOR, chrF++, TER, BERTScore and BLEURT scores.
### Languages and Intended Use
#### Multilingual?
<!-- quick -->
<!-- info: Is the dataset multilingual? -->
<!-- scope: telescope -->
yes
#### Covered Languages
<!-- quick -->
<!-- info: What languages/dialects are covered in the dataset? -->
<!-- scope: telescope -->
`Russian`, `English`
#### License
<!-- quick -->
<!-- info: What is the license of the dataset? -->
<!-- scope: telescope -->
cc-by-nc-4.0: Creative Commons Attribution Non Commercial 4.0 International
#### Intended Use
<!-- info: What is the intended use of the dataset? -->
<!-- scope: microscope -->
The WebNLG dataset was created to promote the development (_i_) of RDF verbalisers and (_ii_) of microplanners able to handle a wide range of linguistic constructions. The dataset aims at covering knowledge in different domains ("categories"). The same properties and entities can appear in several categories.
#### Primary Task
<!-- info: What primary task does the dataset support? -->
<!-- scope: telescope -->
Data-to-Text
#### Communicative Goal
<!-- quick -->
<!-- info: Provide a short description of the communicative goal of a model trained for this task on this dataset. -->
<!-- scope: periscope -->
A model should verbalize all and only the provided input triples in natural language.
### Credit
#### Curation Organization Type(s)
<!-- info: In what kind of organization did the dataset curation happen? -->
<!-- scope: telescope -->
`academic`
#### Curation Organization(s)
<!-- info: Name the organization(s). -->
<!-- scope: periscope -->
Université de Lorraine / LORIA, France, CNRS / LORIA, France, University of Edinburgh, UK, Federal University of Minas Gerais, Brazil
#### Dataset Creators
<!-- info: Who created the original dataset? List the people involved in collecting the dataset and their affiliation(s). -->
<!-- scope: microscope -->
The principle curator of the dataset is Anastasia Shimorina (Université de Lorraine / LORIA, France). Throughout the WebNLG releases, several people contributed to their construction: Claire Gardent (CNRS / LORIA, France), Shashi Narayan (Google, UK), Laura Perez-Beltrachini (University of Edinburgh, UK), Elena Khasanova, and Thiago Castro Ferreira (Federal University of Minas Gerais, Brazil).
#### Funding
<!-- info: Who funded the data creation? -->
<!-- scope: microscope -->
The dataset construction was funded by the French National Research Agency (ANR).
#### Who added the Dataset to GEM?
<!-- info: Who contributed to the data card and adding the dataset to GEM? List the people+affiliations involved in creating this data card and who helped integrate this dataset into GEM. -->
<!-- scope: microscope -->
Simon Mille and Sebastian Gehrmann added the dataset and wrote the data card.
### Dataset Structure
#### Data Fields
<!-- info: List and describe the fields present in the dataset. -->
<!-- scope: telescope -->
See [official documentation](https://webnlg-challenge.loria.fr/docs/).
`entry`: a data instance of the benchmark. Each entry has five attributes: a DBpedia category (`category`), entry ID (`eid`), shape, shape type, and triple set size (`size`).
- `shape`: a string representation of the RDF tree with nested parentheses where `X` is a node (see [Newick tree format](https://en.wikipedia.org/wiki/Newick_format)).
- `shape_type`: a type of the tree shape. We [identify](https://www.aclweb.org/anthology/C16-1141.pdf) three types of tree shapes:
* `chain` (the object of one triple is the subject of the other);
* `sibling` (triples with a shared subject);
* `mixed` (both `chain` and `sibling` types present).
- `eid`: an entry ID. It is unique only within a category and a size.
- `category`: a DBpedia category (Astronaut, City, MusicalWork, Politician, etc.).
- `size`: the number of RDF triples in a set. Ranges from 1 to 7.
Each `entry` has three fields: `originaltripleset`, `modifiedtripleset`, and `lexs`.
`originaltripleset`: a set of RDF triples as extracted from [DBpedia](https://wiki.dbpedia.org/). Each set of RDF triples is a tree. Triples have the subject-predicate-object structure.
`modifiedtripleset`: a set of RDF triples as presented to crowdworkers (for more details on modifications, see below).
Original and modified triples serve different purposes: the original triples — to link data to a knowledge base (DBpedia), whereas the modified triples — to ensure consistency and homogeneity throughout the data. To train models, the modified triples should be used.
`lexs` (shortened for lexicalisations): a natural language text verbalising the triples. Each lexicalisation has two attributes: a comment (`comment`), and a lexicalisation ID (`lid`). By default, comments have the value `good`, except rare cases when they were manually marked as `toFix`. That was done during the corpus creation, when it was seen that a lexicalisation did not exactly match a triple set.
Russian data has additional optional fields comparing to English:
`<dbpedialinks>`: RDF triples extracted from DBpedia between English and Russian entities by means of the property `sameAs`.
`<links>`: RDF triples created manually for some entities to serve as pointers to translators. There are two types of them:
* with `sameAs` (`Spaniards | sameAs | испанцы`)
* with `includes` (`Tomatoes, guanciale, cheese, olive oil | includes | гуанчиале`). Those were mostly created for string literals to translate some parts of them.
Lexicalisations in the Russian WebNLG have a new parameter `lang` (values: `en`, `ru`) because original English texts were kept in the Russian version (see the example above).
#### Example Instance
<!-- info: Provide a JSON formatted example of a typical instance in the dataset. -->
<!-- scope: periscope -->
```
{
"entry": {
"category": "Company",
"size": "4",
"shape": "(X (X) (X) (X) (X))",
"shape_type": "sibling",
"eid": "Id21",
"lexs": [
{
"comment": "good",
"lex": "Trane, which was founded on January 1st 1913 in La Crosse, Wisconsin, is based in Ireland. It has 29,000 employees.",
"lid": "Id1"
}
],
"modifiedtripleset": [
{
"subject": "Trane",
"property": "foundingDate",
"object": "1913-01-01"
},
{
"subject": "Trane",
"property": "location",
"object": "Ireland"
},
{
"subject": "Trane",
"property": "foundationPlace",
"object": "La_Crosse,_Wisconsin"
},
{
"subject": "Trane",
"property": "numberOfEmployees",
"object": "29000"
}
],
"originaltriplesets": {
"originaltripleset": [
{
"subject": "Trane",
"property": "foundingDate",
"object": "1913-01-01"
},
{
"subject": "Trane",
"property": "location",
"object": "Ireland"
},
{
"subject": "Trane",
"property": "foundationPlace",
"object": "La_Crosse,_Wisconsin"
},
{
"subject": "Trane",
"property": "numberOfEmployees",
"object": "29000"
}
]
}
}
}
```
The XML-formatted example is [here](https://webnlg-challenge.loria.fr/docs/#example).
#### Data Splits
<!-- info: Describe and name the splits in the dataset if there are more than one. -->
<!-- scope: periscope -->
| English (v3.0) | Train | Dev | Test |
|-----------------|--------|-------|-------|
| **triple sets** | 13,211 | 1,667 | 1,779 |
| **texts** | 35,426 | 4,464 | 5,150 |
|**properties** | 372 | 290 | 220 |
| Russian (v3.0) | Train | Dev | Test |
|-----------------|--------|-------|-------|
| **triple sets** | 5,573 | 790 | 1,102 |
| **texts** | 14,239 | 2,026 | 2,780 |
|**properties** | 226 | 115 | 192 |
## Dataset in GEM
### Rationale for Inclusion in GEM
#### Why is the Dataset in GEM?
<!-- info: What does this dataset contribute toward better generation evaluation and why is it part of GEM? -->
<!-- scope: microscope -->
Due to the constrained generation task, this dataset can be used to evaluate very specific and narrow generation capabilities.
#### Similar Datasets
<!-- info: Do other datasets for the high level task exist? -->
<!-- scope: telescope -->
yes
#### Unique Language Coverage
<!-- info: Does this dataset cover other languages than other datasets for the same task? -->
<!-- scope: periscope -->
yes
#### Difference from other GEM datasets
<!-- info: What else sets this dataset apart from other similar datasets in GEM? -->
<!-- scope: microscope -->
The RDF-triple format is unique to WebNLG.
#### Ability that the Dataset measures
<!-- info: What aspect of model ability can be measured with this dataset? -->
<!-- scope: periscope -->
surface realization
### GEM-Specific Curation
#### Modificatied for GEM?
<!-- info: Has the GEM version of the dataset been modified in any way (data, processing, splits) from the original curated data? -->
<!-- scope: telescope -->
yes
#### GEM Modifications
<!-- info: What changes have been made to he original dataset? -->
<!-- scope: periscope -->
`other`
#### Modification Details
<!-- info: For each of these changes, described them in more details and provided the intended purpose of the modification -->
<!-- scope: microscope -->
No changes to the main content of the dataset. The [version 3.0](https://gitlab.com/shimorina/webnlg-dataset/-/tree/master/release_v3.0) of the dataset is used.
#### Additional Splits?
<!-- info: Does GEM provide additional splits to the dataset? -->
<!-- scope: telescope -->
yes
#### Split Information
<!-- info: Describe how the new splits were created -->
<!-- scope: periscope -->
23 special test sets for WebNLG were added to the GEM evaluation suite, 12 for English and 11 for Russian.
For both languages, we created subsets of the training and development sets of ~500 randomly selected inputs each. The inputs were sampled proportionally from each category.
Two types of transformations have been applied to WebNLG: (i) input scrambling (English and Russian) and (ii) numerical value replacements (English); in both cases, a subset of about 500 inputs was randomly selected. For (i), the order of the triples was randomly reassigned (each triple kept the same Subject-Property-Object internal order). For (ii), the change was performed respecting the format of the current cardinal value (e.g., alpha, integer, or floating-point) and replacing it with a new random value. The new number is lower-bounded between zero and upper bounded to be within to the highest power of 10 unit for the given value (e.g., replacing 54 would result in a random value between 0-100). Floating values maintain the degree of precision.
For both languages, we did identify different subsets of the test set that we could compare to each other so that we would have a better understanding of the results. There are currently 8 selections that we have made:
Selection 1 (size): input length. This selection corresponds to the number of predicates in the input. By comparing inputs of different lengths, we can see to what extent NLG systems are able to handle different input sizes. The table below provides the relevant frequencies. Please be aware that comparing selections with fewer than 100 items may result in unreliable comparisons.
| Input length | Frequency English | Frequency Russian |
|----------------|-------------------|-------------------|
| 1 | 369 | 254 |
| 2 | 349 | 200 |
| 3 | 350 | 214 |
| 4 | 305 | 214 |
| 5 | 213 | 159 |
| 6 | 114 | 32 |
| 7 | 79 | 29 |
Selection 2 (frequency): seen/unseen single predicates. This selection corresponds to the inputs with only one predicate. We compare which predicates are seen/unseen in the training data. The table below provides the relevant frequencies. Note that the comparison is only valid for English. Not for Russian, since there is only one example of unseen single predicates.
| _ in training | Frequency English | Frequency Russian |
|---------------|-------------------|-------------------|
| Seen | 297 | 253 |
| Unseen | 72 | 1 |
Selection 3 (frequency): seen/unseen combinations of predicates. This selection checks for all combinations of predicates whether that combination has been seen in the training data. For example: if the combination of predicates A and B is seen, that means that there is an input in the training data consisting of two triples, where one triple uses predicate A and the other uses predicate B. If the combination is unseen, then the converse is true. The table below provides the relevant frequencies.
| _ in training | Frequency English | Frequency Russian |
|---------------|-------------------|-------------------|
| unseen | 1295 | 354 |
| seen | 115 | 494 |
Selection 4 (frequency): seen/unseen arguments. This selection checks for all input whether or not all arg1s and arg2s in the input have been seen during the training phase. For this selection, *Seen* is the default. Only if all arg1 instances for a particular input are unseen, do we count the arg1s of the input as unseen. The same holds for arg2. So "seen" here really means that at least some of the arg1s or arg2s are seen in the input. The table below provides the relevant frequencies. Note that the comparison is only valid for English. Not for Russian, since there are very few examples of unseen combinations of predicates.
| Arguments seen in training? | Frequency English | Frequency Russian |
|-----------------------------|-------------------|-------------------|
| both_seen | 518 | 1075 |
| both_unseen | 1177 | 4 |
| arg1_unseen | 56 | 19 |
| arg2_unseen | 28 | 4 |
Selection 5 (shape): repeated subjects. For this selection, the subsets are based on the times a subject is repeated in the input; it only takes into account the maximum number of times a subject is repeated, that is, if in one input a subject appears 3 times and a different subject 2 times, this input will be in the "3_subjects_same' split. Unique_subjects means all subjects are different.
| Max num. of repeated subjects | Frequency English | Frequency Russian |
|-------------------------------|-------------------|-------------------|
| unique_subjects | 453 | 339 |
| 2_subjects_same | 414 | 316 |
| 3_subjects_same | 382 | 217 |
| 4_subjects_same | 251 | 143 |
| 5_subjects_same | 158 | 56 |
| 6_subjects_same | 80 | 19 |
| 7_subjects_same | 41 | 12 |
Selection 6 (shape): repeated objects. Same as for subjects above, but for objects. There are much less cases of repeated objects, so there are only two categories for this selection, unique_objects and some_objects_repeated; for the latter, we have up to 3 coreferring objects in English, and XXX in Russian.
| Max num. of repeated objects | Frequency English | Frequency Russian |
|------------------------------|-------------------|-------------------|
| unique_objects | 1654 | 1099 |
| some_objects_same | 125 | 3 |
Selection 7 (shape): repeated properties. Same as for objects above, but for properties; up to two properties can be the same in English, up to XXX in Russian.
| Max num. of repeated properties | Frequency English | Frequency Russian |
|---------------------------------|-------------------|-------------------|
| unique_properties | 1510 | 986 |
| some_properties_same | 269 | 116 |
Selection 8 (shape): entities that appear both as subject and object. For this selection, we grouped together the inputs in which no entity is found as both subject and object, and on the other side inputs in which one or more entity/ies appear both as subject and as object. We found up to two such entities per input in English, and up to XXX in Russian.
| Max num. of objects and subjects in common | Frequency English | Frequency Russian |
|--------------------------------------------|-------------------|-------------------|
| unique_properties | 1322 | 642 |
| some_properties_same | 457 | 460 |
#### Split Motivation
<!-- info: What aspects of the model's generation capacities were the splits created to test? -->
<!-- scope: periscope -->
Robustness
### Getting Started with the Task
#### Pointers to Resources
<!-- info: Getting started with in-depth research on the task. Add relevant pointers to resources that researchers can consult when they want to get started digging deeper into the task. -->
<!-- scope: microscope -->
Dataset construction: [main dataset paper](https://www.aclweb.org/anthology/P17-1017/), [RDF triple extraction](https://www.aclweb.org/anthology/C16-1141/), [Russian translation](https://www.aclweb.org/anthology/W19-3706/)
WebNLG Challenge 2017: [webpage](https://webnlg-challenge.loria.fr/challenge_2017/), [paper](https://www.aclweb.org/anthology/W17-3518/)
WebNLG Challenge 2020: [webpage](https://webnlg-challenge.loria.fr/challenge_2020/), [paper](https://webnlg-challenge.loria.fr/files/2020.webnlg-papers.7.pdf)
Enriched version of WebNLG: [repository](https://github.com/ThiagoCF05/webnlg), [paper](https://www.aclweb.org/anthology/W18-6521/)
Related research papers: [webpage](https://webnlg-challenge.loria.fr/research/)
## Previous Results
### Previous Results
#### Proposed Evaluation
<!-- info: List and describe the purpose of the metrics and evaluation methodology (including human evaluation) that the dataset creators used when introducing this task. -->
<!-- scope: microscope -->
For both languages, the participating systems are automatically evaluated in a multi-reference scenario. Each English hypothesis is compared to a maximum of 5 references, and each Russian one to a maximum of 7 references. On average, English data has 2.89 references per test instance, and Russian data has 2.52 references per instance.
In a human evaluation, example are uniformly sampled across size of triple sets and the following dimensions are assessed (on MTurk and Yandex.Toloka):
1. Data Coverage: Does the text include descriptions of all predicates presented in the data?
2. Relevance: Does the text describe only such predicates (with related subjects and objects), which are found in the data?
3. Correctness: When describing predicates which are found in the data, does the text mention correct the objects and adequately introduces the subject for this specific predicate?
4. Text Structure: Is the text grammatical, well-structured, written in acceptable English language?
5. Fluency: Is it possible to say that the text progresses naturally, forms a coherent whole and it is easy to understand the text?
For additional information like the instructions, we refer to the original paper.
#### Previous results available?
<!-- info: Are previous results available? -->
<!-- scope: telescope -->
yes
#### Other Evaluation Approaches
<!-- info: What evaluation approaches have others used? -->
<!-- scope: periscope -->
We evaluated a wide range of models as part of the GEM benchmark.
#### Relevant Previous Results
<!-- info: What are the most relevant previous results for this task/dataset? -->
<!-- scope: microscope -->
Results can be found on the [GEM website](https://gem-benchmark.com/results).
## Broader Social Context
### Previous Work on the Social Impact of the Dataset
#### Usage of Models based on the Data
<!-- info: Are you aware of cases where models trained on the task featured in this dataset ore related tasks have been used in automated systems? -->
<!-- scope: telescope -->
yes - related tasks
#### Social Impact Observations
<!-- info: Did any of these previous uses result in observations about the social impact of the systems? In particular, has there been work outlining the risks and limitations of the system? Provide links and descriptions here. -->
<!-- scope: microscope -->
We do not foresee any negative social impact in particular from this dataset or task.
Positive outlooks: Being able to generate good quality text from RDF data would permit, e.g., making this data more accessible to lay users, enriching existing text with information drawn from knowledge bases such as DBpedia or describing, comparing and relating entities present in these knowledge bases.
### Impact on Under-Served Communities
#### Addresses needs of underserved Communities?
<!-- info: Does this dataset address the needs of communities that are traditionally underserved in language technology, and particularly language generation technology? Communities may be underserved for exemple because their language, language variety, or social or geographical context is underepresented in NLP and NLG resources (datasets and models). -->
<!-- scope: telescope -->
no
### Discussion of Biases
#### Any Documented Social Biases?
<!-- info: Are there documented social biases in the dataset? Biases in this context are variations in the ways members of different social categories are represented that can have harmful downstream consequences for members of the more disadvantaged group. -->
<!-- scope: telescope -->
yes
#### Links and Summaries of Analysis Work
<!-- info: Provide links to and summaries of works analyzing these biases. -->
<!-- scope: microscope -->
This dataset is created using DBpedia RDF triples which naturally exhibit biases that have been found to exist in Wikipedia such as some forms of, e.g., gender bias.
The choice of [entities](https://gitlab.com/shimorina/webnlg-dataset/-/blob/master/supplementary/entities_dict.json), described by RDF trees, was not controlled. As such, they may contain gender biases; for instance, all the astronauts described by RDF triples are male. Hence, in texts, pronouns _he/him/his_ occur more often. Similarly, entities can be related to the Western culture more often than to other cultures.
#### Are the Language Producers Representative of the Language?
<!-- info: Does the distribution of language producers in the dataset accurately represent the full distribution of speakers of the language world-wide? If not, how does it differ? -->
<!-- scope: periscope -->
In English, the dataset is limited to the language that crowdraters speak. In Russian, the language is heavily biased by the translationese of the translation system that is post-edited.
## Considerations for Using the Data
### PII Risks and Liability
#### Potential PII Risk
<!-- info: Considering your answers to the PII part of the Data Curation Section, describe any potential privacy to the data subjects and creators risks when using the dataset. -->
<!-- scope: microscope -->
There is no PII in this dataset.
### Licenses
#### Copyright Restrictions on the Dataset
<!-- info: Based on your answers in the Intended Use part of the Data Overview Section, which of the following best describe the copyright and licensing status of the dataset? -->
<!-- scope: periscope -->
`non-commercial use only`
#### Copyright Restrictions on the Language Data
<!-- info: Based on your answers in the Language part of the Data Curation Section, which of the following best describe the copyright and licensing status of the underlying language data? -->
<!-- scope: periscope -->
`public domain`
### Known Technical Limitations
#### Technical Limitations
<!-- info: Describe any known technical limitations, such as spurrious correlations, train/test overlap, annotation biases, or mis-annotations, and cite the works that first identified these limitations when possible. -->
<!-- scope: microscope -->
The quality of the crowdsourced references is limited, in particular in terms of fluency/naturalness of the collected texts.
Russian data was machine-translated and then post-edited by crowdworkers, so some examples may still exhibit issues related to bad translations.
#### Unsuited Applications
<!-- info: When using a model trained on this dataset in a setting where users or the public may interact with its predictions, what are some pitfalls to look out for? In particular, describe some applications of the general task featured in this dataset that its curation or properties make it less suitable for. -->
<!-- scope: microscope -->
Only a limited number of domains are covered in this dataset. As a result, it cannot be used as a general-purpose realizer.
|
ChristophSchuhmann/improved_aesthetics_6plus | 2022-08-10T11:30:40.000Z | [
"region:us"
] | ChristophSchuhmann | null | null | null | 21 | 136 | Entry not found |
arbml/alpagasus_cleaned_ar | 2023-09-06T17:22:31.000Z | [
"region:us"
] | arbml | null | null | null | 0 | 136 | ---
dataset_info:
features:
- name: instruction_en
dtype: string
- name: output_en
dtype: string
- name: instruction
dtype: string
- name: output
dtype: string
- name: index
dtype: int64
splits:
- name: train
num_bytes: 9824184
num_examples: 9229
download_size: 5541315
dataset_size: 9824184
---
# Dataset Card for "alpagasus_cleaned_ar"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
EleutherAI/drop | 2023-08-30T10:16:05.000Z | [
"region:us"
] | EleutherAI | DROP is a QA dataset which tests comprehensive understanding of paragraphs. In
this crowdsourced, adversarially-created, 96k question-answering benchmark, a
system must resolve multiple references in a question, map them onto a paragraph,
and perform discrete operations over them (such as addition, counting, or sorting). | @misc{dua2019drop,
title={DROP: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs},
author={Dheeru Dua and Yizhong Wang and Pradeep Dasigi and Gabriel Stanovsky and Sameer Singh and Matt Gardner},
year={2019},
eprint={1903.00161},
archivePrefix={arXiv},
primaryClass={cs.CL}
} | null | 0 | 136 | Entry not found |
minh21/cpgQA-v1.0-unique-context-test-10-percent-validation-10-percent | 2023-09-09T11:37:51.000Z | [
"region:us"
] | minh21 | null | null | null | 0 | 136 | ---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
- split: validation
path: data/validation-*
dataset_info:
features:
- name: title
dtype: string
- name: id
dtype: int64
- name: question
dtype: string
- name: answer_text
dtype: string
- name: answer_start
dtype: int64
- name: context
dtype: string
splits:
- name: train
num_bytes: 1176326
num_examples: 884
- name: test
num_bytes: 122341
num_examples: 109
- name: validation
num_bytes: 136762
num_examples: 104
download_size: 200983
dataset_size: 1435429
---
# Dataset Card for "cpgQA-v1.0-unique-context-test-10-percent-validation-10-percent"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
result-kand2-sdxl-wuerst-karlo/ac298fb2 | 2023-10-04T23:40:46.000Z | [
"region:us"
] | result-kand2-sdxl-wuerst-karlo | null | null | null | 0 | 136 | ---
dataset_info:
features:
- name: result
dtype: string
- name: id
dtype: int64
splits:
- name: train
num_bytes: 165
num_examples: 10
download_size: 1316
dataset_size: 165
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "ac298fb2"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
FinanceInc/auditor_sentiment | 2022-07-21T19:03:51.000Z | [
"task_categories:text-classification",
"task_ids:multi-class-classification",
"task_ids:sentiment-classification",
"annotations_creators:expert-generated",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:1K<n<10K",
"source_datasets:original",
"language:en",
"region:us"
] | FinanceInc | null | null | null | 10 | 135 | ---
annotations_creators:
- expert-generated
language_creators:
- found
language:
- en
multilinguality:
- monolingual
size_categories:
- 1K<n<10K
source_datasets:
- original
task_categories:
- text-classification
task_ids:
- multi-class-classification
- sentiment-classification
paperswithcode_id: null
pretty_name: Auditor_Sentiment
---
# Dataset Card for Auditor Sentiment
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
## Dataset Description
Auditor review sentiment collected by News Department
- **Point of Contact:**
Talked to COE for Auditing, currently sue@demo.org
### Dataset Summary
Auditor sentiment dataset of sentences from financial news. The dataset consists of several thousand sentences from English language financial news categorized by sentiment.
### Supported Tasks and Leaderboards
Sentiment Classification
### Languages
English
## Dataset Structure
### Data Instances
```
"sentence": "Pharmaceuticals group Orion Corp reported a fall in its third-quarter earnings that were hit by larger expenditures on R&D and marketing .",
"label": "negative"
```
### Data Fields
- sentence: a tokenized line from the dataset
- label: a label corresponding to the class as a string: 'positive' - (2), 'neutral' - (1), or 'negative' - (0)
### Data Splits
A train/test split was created randomly with a 75/25 split
## Dataset Creation
### Curation Rationale
To gather our auditor evaluations into one dataset. Previous attempts using off-the-shelf sentiment had only 70% F1, this dataset was an attempt to improve upon that performance.
### Source Data
#### Initial Data Collection and Normalization
The corpus used in this paper is made out of English news reports.
#### Who are the source language producers?
The source data was written by various auditors.
### Annotations
#### Annotation process
This release of the auditor reviews covers a collection of 4840
sentences. The selected collection of phrases was annotated by 16 people with
adequate background knowledge on financial markets. The subset here is where inter-annotation agreement was greater than 75%.
#### Who are the annotators?
They were pulled from the SME list, names are held by sue@demo.org
### Personal and Sensitive Information
There is no personal or sensitive information in this dataset.
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
All annotators were from the same institution and so interannotator agreement
should be understood with this taken into account.
### Licensing Information
License: Demo.Org Proprietary - DO NOT SHARE
This dataset is based on the [financial phrasebank](https://huggingface.co/datasets/financial_phrasebank) dataset. |
tomekkorbak/detoxify-pile-chunk3-1250000-1300000 | 2022-10-05T00:28:11.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 135 | Entry not found |
santhosh/english-malayalam-names | 2023-04-06T09:26:40.000Z | [
"task_categories:text2text-generation",
"size_categories:10M<n<100M",
"language:en",
"language:ml",
"license:cc-by-sa-4.0",
"malayalam",
"region:us"
] | santhosh | null | null | null | 1 | 135 | ---
license: cc-by-sa-4.0
task_categories:
- text2text-generation
language:
- en
- ml
tags:
- malayalam
size_categories:
- 10M<n<100M
---
# English Malayalam names
This dataset has 27814162 person names both in English and Malayalam.
The source for this dataset is various election roles published by Government.
Potential usages:
1. English <-> Malayalam name transliteration tasks
2. Named entity recognition
3. Person name recognition
## License
Creative commons Attribution Share Alike 4.0
## Contact
Santhosh Thottingal santhosh.thottingal @ gmail.com |
mstz/gisette | 2023-04-17T10:55:16.000Z | [
"task_categories:tabular-classification",
"language:en",
"gisette",
"tabular_classification",
"binary_classification",
"region:us"
] | mstz | null | @misc{misc_gisette_170,
author = {Guyon,Isabelle, Gunn,Steve, Ben-Hur,Asa & Dror,Gideon},
title = {{Gisette}},
year = {2008},
howpublished = {UCI Machine Learning Repository},
note = {{DOI}: \\url{10.24432/C5HP5B}}
} | null | 0 | 135 | ---
language:
- en
tags:
- gisette
- tabular_classification
- binary_classification
pretty_name: Gisette
task_categories: # Full list at https://github.com/huggingface/hub-docs/blob/main/js/src/lib/interfaces/Types.ts
- tabular-classification
configs:
- gisette
---
# Gisette
The [Gisette dataset](https://archive-beta.ics.uci.edu/dataset/170/gisette) from the [UCI repository](https://archive-beta.ics.uci.edu/).
# Configurations and tasks
| **Configuration** | **Task** | **Description** |
|-----------------------|---------------------------|-------------------------|
| gisette | Binary classification.| |
|
C-MTEB/T2Reranking | 2023-07-28T07:29:52.000Z | [
"region:us"
] | C-MTEB | null | null | null | 0 | 135 | ---
configs:
- config_name: default
data_files:
- split: dev
path: data/dev-*
dataset_info:
features:
- name: query
dtype: string
- name: positive
sequence: string
- name: negative
sequence: string
splits:
- name: dev
num_bytes: 206865573
num_examples: 6129
download_size: 120293598
dataset_size: 206865573
---
# Dataset Card for "T2Reranking"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
gwlms/germeval2014 | 2023-07-31T16:08:59.000Z | [
"license:cc-by-4.0",
"region:us"
] | gwlms | # Introduction
The GermEval 2014 NER Shared Task is an event that makes available CC-licensed German data with NER annotation with the
goal of significantly advancing the state of the art in German NER and to push the field of NER towards nested
representations of named entities.
The GermEval 2014 NER Shared Task builds on a new dataset with German Named Entity annotation with the following
properties: The data was sampled from German Wikipedia and News Corpora as a collection of citations. The dataset
covers over 31,000 sentences corresponding to over 590,000 tokens. The NER annotation uses the NoSta-D guidelines,
which extend the Tübingen Treebank guidelines, using four main NER categories with sub-structure, and annotating
embeddings among NEs such as [ORG FC Kickers [LOC Darmstadt]].
# Dataset
## Labels
### Fine-grained labels indicating NER subtypes
German morphology is comparatively productive (at least when compared to
English). There is a considerable amount of word formation through both overt (non-zero) derivation and compounding, in
particular for nouns. This gives rise to morphologically complex words that are not identical to, but stand in a direct
relation to, Named Entities. The Shared Task corpus treats these as NE instances but marks them as special subtypes by
introducing two fine-grained labels: -deriv marks derivations from NEs such as the previously mentioned englisch
(“English”), and -part marks compounds including a NE as a subsequence deutschlandweit (“Germany-wide”).
### Embedded markables
Almost all extant corpora with Named Entity annotation assume that NE annotation is “flat”, that is, each word in the
text can form part of at most one NE chunk. Clearly, this is an oversimplification. Consider the noun phase Technische
Universitat Darmstadt ¨ (“Technical University (of) Darmstadt”). It denotes an organization (label ORG), but also holds
another NE, Darmstadt, which is a location (label LOC). To account for such cases, the Shared Task corpus is annotated
with two levels of Named Entities. It captures at least one level of smaller NEs being embedded in larger NEs.
## Statistics
The data used for the GermEval 2014 NER Shared Task builds on the dataset annotated by (Benikova et al., 2014).
In
this dataset, sentences taken from German Wikipedia articles and online news were used as a collection of citations,
then annotated according to extended NoSta-D guidelines and eventually distributed under the CC-BY license. As already
described above, those guidelines use four main categories with sub-structure and nesting.
The dataset is distributed contains overall more than 31,000 sentences with over 590,000 tokens. Those were divided in
the following way: the training set consists of 24,000 sentences, the development set of 2,200 sentences and the test
set of 5,100 sentences. The test set labels were not available to the participants until after the deadline. The
distribution of the categories over the whole dataset is shown in Table 1. Care was taken to ensure the even dispersion
of the categories in the subsets. The entire dataset contains over 41,000 NEs, about 7.8% of them embedded in other NEs
(nested NEs), about 11.8% are derivations (deriv) and about 5.6% are parts of NEs concatenated with other words (part).
## Format
The tab-separated format used in this dataset is similar to the CoNLL-Format. As illustrated in Table 2, the format
used in the dataset additionally contains token numbers per sentence in the first column and a comment line indicating
source and data before each sentence. The second column contains the tokens. The third column encodes the outer NE
spans, the fourth column the inner ones. The BIO-scheme was used in order to encode the NE spans. In our challenge,
further nested columns were not considered.
## Summary
In summary, we distinguish between 12 classes of NEs: four main classes PERson, LOCation, ORGanisation, and OTHer and
their subclasses, annotated at two levels (“inner” and “outer” chunks). The challenge of this setup is that while it
technically still allows a simple classification approach it introduces a recursive structure that calls for the
application of more general machine learning or other automatically classifying methods that go beyond plain sequence
tagging. | @inproceedings{benikova14:_germev_named_entit_recog_shared_task,
added-at = {2017-04-03T19:29:52.000+0000},
address = {Hildesheim, Germany},
author = {Benikova, Darina and Biemann, Chris and Kisselew, Max and Pad\'o, Sebastian},
biburl = {https://puma.ub.uni-stuttgart.de/bibtex/2132d938a7afe8639e78156fb9d756b20/sp},
booktitle = {Proceedings of the KONVENS GermEval workshop},
interhash = {6cad5d4fdd6a07dbefad4221ba7d8d44},
intrahash = {132d938a7afe8639e78156fb9d756b20},
keywords = {myown workshop},
pages = {104--112},
timestamp = {2017-04-03T17:30:49.000+0000},
title = {{GermEval 2014 Named Entity Recognition Shared Task: Companion Paper}},
year = 2014
} | null | 0 | 135 | ---
viewer: false
license: cc-by-4.0
dataset_info:
features:
- name: tokens
sequence: string
- name: ner_tags
sequence:
class_label:
names:
'0': O
'1': B-LOC
'2': I-LOC
'3': B-LOCderiv
'4': I-LOCderiv
'5': B-LOCpart
'6': I-LOCpart
'7': B-ORG
'8': I-ORG
'9': B-ORGderiv
'10': I-ORGderiv
'11': B-ORGpart
'12': I-ORGpart
'13': B-OTH
'14': I-OTH
'15': B-OTHderiv
'16': I-OTHderiv
'17': B-OTHpart
'18': I-OTHpart
'19': B-PER
'20': I-PER
'21': B-PERderiv
'22': I-PERderiv
'23': B-PERpart
'24': I-PERpart
- name: ner_t5_output
dtype: string
- name: ner_own_output
dtype: string
splits:
- name: train
num_bytes: 9450958
num_examples: 24000
- name: validation
num_bytes: 866649
num_examples: 2200
- name: test
num_bytes: 2011187
num_examples: 5100
download_size: 4279522
dataset_size: 12328794
---
|
sarahpann/MATH | 2023-09-23T03:06:46.000Z | [
"region:us"
] | sarahpann | null | null | null | 0 | 135 | Entry not found |
mtc/swisstext23-20min-annotation-data | 2023-08-25T08:22:10.000Z | [
"region:us"
] | mtc | null | null | null | 0 | 135 | ---
configs:
- config_name: default
data_files:
- split: test
path: data/test-*
dataset_info:
features:
- name: id
dtype: int64
- name: titleHeader
dtype: string
- name: title
dtype: string
- name: lead
dtype: string
- name: article
dtype: string
- name: summary
dtype: string
- name: article_sentence_count
dtype: int64
- name: summary_sentence_count
dtype: int64
- name: __index_level_0__
dtype: int64
- name: url
dtype: string
- name: paragraphs
sequence: string
splits:
- name: test
num_bytes: 998931
num_examples: 200
download_size: 0
dataset_size: 998931
---
# Dataset Card for "swisstext23-20min-annotation-data"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
TaylorAI/pubmed_commercial | 2023-08-26T07:32:30.000Z | [
"region:us"
] | TaylorAI | null | null | null | 11 | 135 | Entry not found |
tomekkorbak/detoxify-pile-chunk3-1300000-1350000 | 2022-10-05T00:06:32.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 134 | Entry not found |
mitclinicalml/clinical-ie | 2022-12-01T16:34:20.000Z | [
"arxiv:2205.12689",
"arxiv:2010.02010",
"arxiv:1806.04185",
"region:us"
] | mitclinicalml | null | @inproceedings{agrawal2022large,
title={Large Language Models are Few-Shot Clinical Information Extractors},
author={Monica Agrawal and Stefan Hegselmann and Hunter Lang and Yoon Kim and David Sontag},
booktitle = {Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing},
year={2022},
url_Paper = {https://arxiv.org/pdf/2205.12689.pdf}} | null | 18 | 134 | ---
{}
---
Below, we provide access to the datasets used in and created for the EMNLP 2022 paper [Large Language Models are Few-Shot Clinical Information Extractors](https://arxiv.org/abs/2205.12689).
# Task #1: Clinical Sense Disambiguation
For Task #1, we use the original annotations from the [Clinical Acronym Sense Inventory (CASI) dataset](https://conservancy.umn.edu/handle/11299/137703), described in [their paper](https://academic.oup.com/jamia/article/21/2/299/723657).
As is common, due to noisiness in the label set, we do not evaluate on the entire dataset, but only on a cleaner subset. For consistency, we use the subset defined by the filtering used in ["Zero-Shot Clinical Acronym Expansion
via Latent Meaning Cells"](https://arxiv.org/pdf/2010.02010.pdf). This results in a subset of 18,164 examples and 41 acronyms for evaluation.
We additionally use the MIMIC Reverse Substitution dataset, as created in that same paper, with further instructions available in [their repository](https://github.com/griff4692/LMC).
# Task #2: Biomedical Evidence Extraction
For Task #2, we use the out-of-the-box high-level labels from the [PICO dataset](https://arxiv.org/abs/1806.04185) available publicly in the repository [here](https://github.com/bepnye/EBM-NLP).
# Task #3: Coreference Resolution
For Task #3, we annotated 105 snippets from the [CASI dataset](https://conservancy.umn.edu/handle/11299/137703), 5 for development and 100 for test. Each example is labeled with a singular pronoun and that pronoun's corresponding noun phrase antecedent (or antecedents).
The antecedent was annotated as the entire noun phrase (barring any dependent clauses); in cases where multiple equally valid antecedents were available, all were labeled (empirically, up to 2).
For the purposes of evaluation, we chose the antecedent with the highest overlap to each model’s output.
To ensure nontrivial examples, the annotators excluded all examples of personal pronouns (e.g. “he”, “she”) if another person (and possible antecedent) had not yet been mentioned in the snippet.
Examples were skipped in annotation if the pronoun did not have an antecedent within the provided text snippet.
# Task #4: Medication Status Extraction
For Task #3, we annotated 105 snippets from the [CASI dataset](https://conservancy.umn.edu/handle/11299/137703), 5 for development and 100 for test. We wanted to create a dataset of challenging examples containing a changeover in treatment. From a sample, only ∼5% of CASI snippets contained such examples. To increase the density of these examples, speeding up annotation, clinical notes were filtered with the following search terms: discont, adverse, side effect, switch, and dosage, leading to 1445 snippets. We excluded snippets that were purely medication lists, requiring at least some narrative part to be present.
For each example, the annotators first extracted all medications. Guidelines excluded medication categories (e.g. “ACE-inhibitor”) if they referred to more specific drug names mentioned elsewhere (even if partially cut off in the snippet). For instance, only the antibiotic Levaquin was labeled in: “It is
probably reasonable to treat with antibiotics [...]. I would agree with Levaquin alone [...]”. Guidelines also excluded electrolytes and intravenous fluids as well as route and dosage information. In a second step, medications were assigned to one of three categories: active, discontinued, and neither.
Discontinued medications also contain medications that are temporarily on hold. The category neither was assigned to all remaining medications (e.g. allergies, potential medications).
The medication lists for each example were serialized as a json.
# Task #5: Medication Attribute Extraction
For Task #5, we again annotated 105 snippets from the [CASI dataset](https://conservancy.umn.edu/handle/11299/137703), 5 for development and 100 for test.
Annotation guideline were adopted from the 2009 i2b2 medication extraction challenge (Uzuner et al., 2010) with slight modifications.
We allowed medication attributes to have multiple spans and grouped together different mentions of the the same drug (e.g. “Tylenol” and “Tylenol PM”) for the purpose of relation extraction.
The annotation list for each example was serialized as a json.
# Citations
When using our annotations for tasks #3-5, please cite our paper, as well as the papers from which the underlying text originated.
```
@inproceedings{agrawal2022large,
title={Large Language Models are Few-Shot Clinical Information Extractors},
author={Monica Agrawal and Stefan Hegselmann and Hunter Lang and Yoon Kim and David Sontag},
booktitle = {Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing},
year={2022},
url_Paper = {https://arxiv.org/pdf/2205.12689.pdf}
}
```
```
@article{moon2014sense,
title={A sense inventory for clinical abbreviations and acronyms created using clinical notes and medical dictionary resources},
author={Moon, Sungrim and Pakhomov, Serguei and Liu, Nathan and Ryan, James O and Melton, Genevieve B},
journal={Journal of the American Medical Informatics Association},
volume={21},
number={2},
pages={299--307},
year={2014},
publisher={BMJ Publishing Group BMA House, Tavistock Square, London, WC1H 9JR}
}
```
# Licensing
The annotations added by our team fall under the MIT license, but the CASI dataset itself is subject to its own licensing.
---
license: other
---
|
Achitha/tamil_eng_data | 2023-02-12T18:52:26.000Z | [
"task_categories:translation",
"size_categories:1K<n<10K",
"language:ta",
"language:en",
"region:us"
] | Achitha | The data contains roughly one and half hours of audio and transcripts in Tamil language. | @misc{simpledata_1,
title = {Whisper model for tamil-to-eng translation},
publisher = {Achitha},
year = {2022},
}
@misc{simpledata_2,
title = {Fine-tuning whisper model},
publisher = {Achitha},
year = {2022},
} | null | 0 | 134 | ---
task_categories:
- translation
language:
- ta
- en
size_categories:
- 1K<n<10K
--- |
vgaraujov/wmt13 | 2023-07-04T08:17:47.000Z | [
"region:us"
] | vgaraujov | null | @InProceedings{bojar-EtAl:2013:WMT,
author = {Bojar, Ond\v{r}ej and Buck, Christian and Callison-Burch, Chris and Federmann, Christian and Haddow, Barry and Koehn, Philipp and Monz, Christof and Post, Matt and Soricut, Radu and Specia, Lucia},
title = {Findings of the 2013 {Workshop on Statistical Machine Translation}},
booktitle = {Proceedings of the Eighth Workshop on Statistical Machine Translation},
month = {August},
year = {2013},
address = {Sofia, Bulgaria},
publisher = {Association for Computational Linguistics},
pages = {1--44},
url = {http://www.aclweb.org/anthology/W13-2201}
} | null | 0 | 134 | Entry not found |
minskiter/weibo | 2023-07-22T13:49:08.000Z | [
"size_categories:1K<n<10K",
"language:zh",
"license:apache-2.0",
"social",
"region:us"
] | minskiter | The Weibo NER dataset is a Chinese Named Entity Recognition dataset
drawn from the social media website Sina Weibo. | @inproceedings{peng-dredze-2015-named,
title = "Named Entity Recognition for {C}hinese
Social Media with Jointly Trained Embeddings",
author = "Peng, Nanyun and Dredze, Mark",
booktitle = "Proceedings of the 2015 Conference on
Empirical Methods in Natural Language Processing",
month = sep,
year = "2015",
address = "Lisbon, Portugal",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/D15-1064",
doi = "10.18653/v1/D15-1064",
pages = "548--554",
} | null | 0 | 134 | ---
license: apache-2.0
dataset_info:
features:
- name: text
sequence: string
- name: labels
sequence:
class_label:
names:
'0': O
'1': B-PER.NAM
'2': I-PER.NAM
'3': E-PER.NAM
'4': S-PER.NAM
'5': B-ORG.NAM
'6': I-ORG.NAM
'7': E-ORG.NAM
'8': S-ORG.NAM
'9': B-LOC.NAM
'10': I-LOC.NAM
'11': E-LOC.NAM
'12': S-LOC.NAM
'13': B-GPE.NAM
'14': I-GPE.NAM
'15': E-GPE.NAM
'16': S-GPE.NAM
'17': B-PER.NOM
'18': I-PER.NOM
'19': E-PER.NOM
'20': S-PER.NOM
'21': B-ORG.NOM
'22': I-ORG.NOM
'23': E-ORG.NOM
'24': S-ORG.NOM
'25': B-LOC.NOM
'26': I-LOC.NOM
'27': E-LOC.NOM
'28': S-LOC.NOM
'29': B-GPE.NOM
'30': I-GPE.NOM
'31': E-GPE.NOM
'32': S-GPE.NOM
splits:
- name: train
num_bytes: 1095833
num_examples: 1350
- name: validation
num_bytes: 215953
num_examples: 270
- name: test
num_bytes: 220694
num_examples: 270
download_size: 217348
dataset_size: 1532480
language:
- zh
tags:
- social
size_categories:
- 1K<n<10K
---
### How to loading dataset?
```python
from datasets import load_dataset
datasets = load_dataset("minskiter/weibo",save_infos=True)
train,validation,test = datasets['train'],datasets['validation'],datasets['test']
# convert label to str
print(train.features['labels'].feature.int2str(0))
```
### Force Update
```python
from datasets import load_dataset
datasets = load_dataset("minskiter/weibo", download_mode="force_redownload")
```
### CHANGE LOGS
- 21/7/2023 v1.0.2 Fix data format.
- 16/7/2023 v1.0.0 Publish weibo data. |
tomekkorbak/detoxify-pile-chunk3-1500000-1550000 | 2022-10-04T23:53:18.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 133 | Entry not found |
tomekkorbak/detoxify-pile-chunk3-1400000-1450000 | 2022-10-04T23:57:23.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 133 | Entry not found |
tomekkorbak/detoxify-pile-chunk3-1550000-1600000 | 2022-10-05T00:02:47.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 133 | Entry not found |
tomekkorbak/detoxify-pile-chunk3-1350000-1400000 | 2022-10-05T00:06:19.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 133 | Entry not found |
sdadas/ppc | 2022-12-29T11:30:31.000Z | [
"task_categories:text-classification",
"task_ids:semantic-similarity-classification",
"multilinguality:monolingual",
"size_categories:1K<n<10K",
"language:pl",
"license:cc-by-nc-sa-4.0",
"region:us"
] | sdadas | null | null | null | 0 | 133 | ---
language:
- pl
license:
- cc-by-nc-sa-4.0
multilinguality:
- monolingual
size_categories:
- 1K<n<10K
task_categories:
- text-classification
task_ids:
- semantic-similarity-classification
pretty_name: Polish Paraphrase Corpus
dataset_info:
features:
- name: sentence_A
dtype: string
- name: sentence_B
dtype: string
- name: label
dtype:
class_label:
names:
0: not used
1: exact paraphrases
2: similar sentences
3: non-paraphrases
splits:
- name: train
- name: validation
- name: test
---
# PPC - Polish Paraphrase Corpus
### Dataset Summary
Polish Paraphrase Corpus contains 7000 manually labeled sentence pairs. The dataset was divided into training, validation and test splits. The training part includes 5000 examples, while the other parts contain 1000 examples each. The main purpose of creating such a dataset was to verify how machine learning models perform in the challenging problem of paraphrase identification, where most records contain semantically overlapping parts. Technically, this is a three-class classification task, where each record can be assigned to one of the following categories:
- Exact paraphrases - Sentence pairs that convey exactly the same information. We are interested only in the semantic meaning of the sentence, therefore this category also includes sentences that are semantically identical but, for example, have different emotional emphasis.
- Close paraphrases - Sentence pairs with similar semantic meaning. In this category we include all pairs which contain the same information, but in addition to it there may be other semantically non-overlapping parts. This category also contains context-dependent paraphrases - sentence pairs that may have the same meaning in some contexts but are different in others.
- Non-paraphrases - All other cases, including contradictory sentences and semantically unrelated sentences.
The corpus contains 2911, 1297, and 2792 examples for the above three categories, respectively. The process of annotating the dataset was preceded by an automated generation of candidate pairs, which were then manually labeled. We experimented with two popular techniques of generating possible paraphrases: backtranslation with a set of neural machine translation models and paraphrase mining using a pre-trained multilingual sentence encoder. The extracted sentence pairs are drawn from different data sources: Taboeba, Polish news articles, Wikipedia and Polish version of SICK dataset. Since most of the sentence pairs obtained in this way fell into the first two categories, in order to balance the dataset, some of the examples were manually modified to convey different information. In this way, even negative examples often have high semantic overlap, making this problem difficult for machine learning models.
### Data Instances
Example instance:
```
{
"sentence_A": "Libia: lotnisko w w Trypolisie ostrzelane rakietami.",
"sentence_B": "Jedyne lotnisko w stolicy Libii - Trypolisie zostało w nocy z wtorku na środę ostrzelane rakietami.",
"label": "2"
}
```
### Data Fields
- sentence_A: first sentence text
- sentence_B: second sentence text
- label: label identifier corresponding to one of three categories
### Citation Information
```
@inproceedings{9945218,
author={Dadas, S{\l}awomir},
booktitle={2022 IEEE International Conference on Systems, Man, and Cybernetics (SMC)},
title={Training Effective Neural Sentence Encoders from Automatically Mined Paraphrases},
year={2022},
volume={},
number={},
pages={371-378},
doi={10.1109/SMC53654.2022.9945218}
}
``` |
DFKI-SLT/kbp37 | 2023-04-27T13:04:14.000Z | [
"task_categories:text-classification",
"task_ids:multi-class-classification",
"annotations_creators:other",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:extended|other",
"language:en",
"license:other",
"relation extraction",
"arxiv:1508.01006",
"region:us"
] | DFKI-SLT | KBP37 is a revision of MIML-RE annotation dataset, provided by Gabor Angeli et al. (2014). They use both the 2010 and
2013 KBP official document collections, as well as a July 2013 dump of Wikipedia as the text corpus for annotation.
There are 33811 sentences been annotated. Zhang and Wang made several refinements:
1. They add direction to the relation names, e.g. '`per:employee_of`' is split into '`per:employee of(e1,e2)`'
and '`per:employee of(e2,e1)`'. They also replace '`org:parents`' with '`org:subsidiaries`' and replace
'`org:member of’ with '`org:member`' (by their reverse directions).
2. They discard low frequency relations such that both directions of each relation occur more than 100 times in the
dataset.
KBP37 contains 18 directional relations and an additional '`no_relation`' relation, resulting in 37 relation classes. | @article{DBLP:journals/corr/ZhangW15a,
author = {Dongxu Zhang and
Dong Wang},
title = {Relation Classification via Recurrent Neural Network},
journal = {CoRR},
volume = {abs/1508.01006},
year = {2015},
url = {http://arxiv.org/abs/1508.01006},
eprinttype = {arXiv},
eprint = {1508.01006},
timestamp = {Fri, 04 Nov 2022 18:37:50 +0100},
biburl = {https://dblp.org/rec/journals/corr/ZhangW15a.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
} | null | 0 | 133 | ---
annotations_creators:
- other
language:
- en
language_creators:
- found
license:
- other
multilinguality:
- monolingual
pretty_name: KBP37 is an English Relation Classification dataset
size_categories:
- 10K<n<100K
source_datasets:
- extended|other
tags:
- relation extraction
task_categories:
- text-classification
task_ids:
- multi-class-classification
dataset_info:
- config_name: kbp37
features:
- name: id
dtype: string
- name: sentence
dtype: string
- name: relation
dtype:
class_label:
names:
'0': no_relation
'1': org:alternate_names(e1,e2)
'2': org:alternate_names(e2,e1)
'3': org:city_of_headquarters(e1,e2)
'4': org:city_of_headquarters(e2,e1)
'5': org:country_of_headquarters(e1,e2)
'6': org:country_of_headquarters(e2,e1)
'7': org:founded(e1,e2)
'8': org:founded(e2,e1)
'9': org:founded_by(e1,e2)
'10': org:founded_by(e2,e1)
'11': org:members(e1,e2)
'12': org:members(e2,e1)
'13': org:stateorprovince_of_headquarters(e1,e2)
'14': org:stateorprovince_of_headquarters(e2,e1)
'15': org:subsidiaries(e1,e2)
'16': org:subsidiaries(e2,e1)
'17': org:top_members/employees(e1,e2)
'18': org:top_members/employees(e2,e1)
'19': per:alternate_names(e1,e2)
'20': per:alternate_names(e2,e1)
'21': per:cities_of_residence(e1,e2)
'22': per:cities_of_residence(e2,e1)
'23': per:countries_of_residence(e1,e2)
'24': per:countries_of_residence(e2,e1)
'25': per:country_of_birth(e1,e2)
'26': per:country_of_birth(e2,e1)
'27': per:employee_of(e1,e2)
'28': per:employee_of(e2,e1)
'29': per:origin(e1,e2)
'30': per:origin(e2,e1)
'31': per:spouse(e1,e2)
'32': per:spouse(e2,e1)
'33': per:stateorprovinces_of_residence(e1,e2)
'34': per:stateorprovinces_of_residence(e2,e1)
'35': per:title(e1,e2)
'36': per:title(e2,e1)
splits:
- name: train
num_bytes: 3570626
num_examples: 15917
- name: validation
num_bytes: 388935
num_examples: 1724
- name: test
num_bytes: 762806
num_examples: 3405
download_size: 5106673
dataset_size: 4722367
- config_name: kbp37_formatted
features:
- name: id
dtype: string
- name: token
sequence: string
- name: e1_start
dtype: int32
- name: e1_end
dtype: int32
- name: e2_start
dtype: int32
- name: e2_end
dtype: int32
- name: relation
dtype:
class_label:
names:
'0': no_relation
'1': org:alternate_names(e1,e2)
'2': org:alternate_names(e2,e1)
'3': org:city_of_headquarters(e1,e2)
'4': org:city_of_headquarters(e2,e1)
'5': org:country_of_headquarters(e1,e2)
'6': org:country_of_headquarters(e2,e1)
'7': org:founded(e1,e2)
'8': org:founded(e2,e1)
'9': org:founded_by(e1,e2)
'10': org:founded_by(e2,e1)
'11': org:members(e1,e2)
'12': org:members(e2,e1)
'13': org:stateorprovince_of_headquarters(e1,e2)
'14': org:stateorprovince_of_headquarters(e2,e1)
'15': org:subsidiaries(e1,e2)
'16': org:subsidiaries(e2,e1)
'17': org:top_members/employees(e1,e2)
'18': org:top_members/employees(e2,e1)
'19': per:alternate_names(e1,e2)
'20': per:alternate_names(e2,e1)
'21': per:cities_of_residence(e1,e2)
'22': per:cities_of_residence(e2,e1)
'23': per:countries_of_residence(e1,e2)
'24': per:countries_of_residence(e2,e1)
'25': per:country_of_birth(e1,e2)
'26': per:country_of_birth(e2,e1)
'27': per:employee_of(e1,e2)
'28': per:employee_of(e2,e1)
'29': per:origin(e1,e2)
'30': per:origin(e2,e1)
'31': per:spouse(e1,e2)
'32': per:spouse(e2,e1)
'33': per:stateorprovinces_of_residence(e1,e2)
'34': per:stateorprovinces_of_residence(e2,e1)
'35': per:title(e1,e2)
'36': per:title(e2,e1)
splits:
- name: train
num_bytes: 4943394
num_examples: 15807
- name: validation
num_bytes: 539197
num_examples: 1714
- name: test
num_bytes: 1055918
num_examples: 3379
download_size: 5106673
dataset_size: 6581345
---
# Dataset Card for "kbp37"
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Repository:** [kbp37](https://github.com/zhangdongxu/kbp37)
- **Paper:** [Relation Classification via Recurrent Neural Network](https://arxiv.org/abs/1508.01006)
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
- **Size of downloaded dataset files:** 5.11 MB
- **Size of the generated dataset:** 6.58 MB
### Dataset Summary
KBP37 is a revision of MIML-RE annotation dataset, provided by Gabor Angeli et al. (2014). They use both the 2010 and
2013 KBP official document collections, as well as a July 2013 dump of Wikipedia as the text corpus for annotation.
There are 33811 sentences been annotated. Zhang and Wang made several refinements:
1. They add direction to the relation names, e.g. '`per:employee_of`' is split into '`per:employee of(e1,e2)`'
and '`per:employee of(e2,e1)`'. They also replace '`org:parents`' with '`org:subsidiaries`' and replace
'`org:member of’ with '`org:member`' (by their reverse directions).
2. They discard low frequency relations such that both directions of each relation occur more than 100 times in the
dataset.
KBP37 contains 18 directional relations and an additional '`no_relation`' relation, resulting in 37 relation classes.
Note:
- There is a formatted version that you can load with `datasets.load_dataset('kbp37', name='kbp37_formatted')`. This version is tokenized with `str.split()` and
provides entities as offsets instead of being enclosed by xml tags. It discards some examples, however, that are invalid in the original dataset and lead
to entity offset errors, e.g. example train/1276.
### Supported Tasks and Leaderboards
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Languages
The language data in KBP37 is in English (BCP-47 en)
## Dataset Structure
### Data Instances
#### kbp37
- **Size of downloaded dataset files:** 5.11 MB
- **Size of the generated dataset:** 4.7 MB
An example of 'train' looks as follows:
```json
{
"id": "0",
"sentence": "<e1> Thom Yorke </e1> of <e2> Radiohead </e2> has included the + for many of his signature distortion sounds using a variety of guitars to achieve various tonal options .",
"relation": 27
}
```
#### kbp37_formatted
- **Size of downloaded dataset files:** 5.11 MB
- **Size of the generated dataset:** 6.58 MB
An example of 'train' looks as follows:
```json
{
"id": "1",
"token": ["Leland", "High", "School", "is", "a", "public", "high", "school", "located", "in", "the", "Almaden", "Valley", "in", "San", "Jose", "California", "USA", "in", "the", "San", "Jose", "Unified", "School", "District", "."],
"e1_start": 0,
"e1_end": 3,
"e2_start": 14,
"e2_end": 16,
"relation": 3
}
```
### Data Fields
#### kbp37
- `id`: the instance id of this sentence, a `string` feature.
- `sentence`: the sentence, a `string` features.
- `relation`: the relation label of this instance, an `int` classification label.
```python
{"no_relation": 0, "org:alternate_names(e1,e2)": 1, "org:alternate_names(e2,e1)": 2, "org:city_of_headquarters(e1,e2)": 3, "org:city_of_headquarters(e2,e1)": 4, "org:country_of_headquarters(e1,e2)": 5, "org:country_of_headquarters(e2,e1)": 6, "org:founded(e1,e2)": 7, "org:founded(e2,e1)": 8, "org:founded_by(e1,e2)": 9, "org:founded_by(e2,e1)": 10, "org:members(e1,e2)": 11, "org:members(e2,e1)": 12, "org:stateorprovince_of_headquarters(e1,e2)": 13, "org:stateorprovince_of_headquarters(e2,e1)": 14, "org:subsidiaries(e1,e2)": 15, "org:subsidiaries(e2,e1)": 16, "org:top_members/employees(e1,e2)": 17, "org:top_members/employees(e2,e1)": 18, "per:alternate_names(e1,e2)": 19, "per:alternate_names(e2,e1)": 20, "per:cities_of_residence(e1,e2)": 21, "per:cities_of_residence(e2,e1)": 22, "per:countries_of_residence(e1,e2)": 23, "per:countries_of_residence(e2,e1)": 24, "per:country_of_birth(e1,e2)": 25, "per:country_of_birth(e2,e1)": 26, "per:employee_of(e1,e2)": 27, "per:employee_of(e2,e1)": 28, "per:origin(e1,e2)": 29, "per:origin(e2,e1)": 30, "per:spouse(e1,e2)": 31, "per:spouse(e2,e1)": 32, "per:stateorprovinces_of_residence(e1,e2)": 33, "per:stateorprovinces_of_residence(e2,e1)": 34, "per:title(e1,e2)": 35, "per:title(e2,e1)": 36}
```
#### kbp37_formatted
- `id`: the instance id of this sentence, a `string` feature.
- `token`: the list of tokens of this sentence, using `str.split()`, a `list` of `string` features.
- `e1_start`: the 0-based index of the start token of the first argument', an `int` feature.
- `e1_end`: the 0-based index of the end token of the first argument, exclusive, an `int` feature.
- `e2_start`: the 0-based index of the start token of the second argument, an `int` feature.
- `e2_end`: the 0-based index of the end token of the second argument, exclusive, an `int` feature.
- `relation`: the relation label of this instance, an `int` classification label (same as `'kbp37''`).
### Data Splits
| | Train | Dev | Test |
|-------|-------|------|------|
| kbp37 | 15917 | 1724 | 3405 |
| kbp37_formatted | 15807 | 1714 | 3379 |
## Dataset Creation
### Curation Rationale
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the source language producers?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Annotations
#### Annotation process
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
#### Who are the annotators?
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Personal and Sensitive Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Discussion of Biases
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Other Known Limitations
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
## Additional Information
### Dataset Curators
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Licensing Information
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
### Citation Information
```
@article{DBLP:journals/corr/ZhangW15a,
author = {Dongxu Zhang and
Dong Wang},
title = {Relation Classification via Recurrent Neural Network},
journal = {CoRR},
volume = {abs/1508.01006},
year = {2015},
url = {http://arxiv.org/abs/1508.01006},
eprinttype = {arXiv},
eprint = {1508.01006},
timestamp = {Fri, 04 Nov 2022 18:37:50 +0100},
biburl = {https://dblp.org/rec/journals/corr/ZhangW15a.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
```
### Contributions
Thanks to [@phucdev](https://github.com/phucdev) for adding this dataset. |
patomp/thai-mscoco-2014-captions | 2023-05-02T15:52:54.000Z | [
"region:us"
] | patomp | null | null | null | 0 | 133 | ---
dataset_info:
features:
- name: image
dtype: image
- name: filepath
dtype: string
- name: sentids
list: int32
- name: filename
dtype: string
- name: imgid
dtype: int32
- name: split
dtype: string
- name: sentences_tokens
list:
list: string
- name: sentences_raw
list: string
- name: sentences_sentid
list: int32
- name: cocoid
dtype: int32
- name: th_sentences_raw
sequence: string
splits:
- name: test
num_bytes: 819234726.0
num_examples: 5000
- name: validation
num_bytes: 807387321.0
num_examples: 5000
- name: train
num_bytes: 18882795327.165
num_examples: 113287
download_size: 20158273111
dataset_size: 20509417374.165
---
## Usage
```python
from datasets import load_dataset
dataset = load_dataset("patomp/thai-mscoco-2014-captions")
dataset
```
output
```python
DatasetDict({
train: Dataset({
features: ['image', 'filepath', 'sentids', 'filename', 'imgid', 'split', 'sentences_tokens', 'sentences_raw', 'sentences_sentid', 'cocoid', 'th_sentences_raw'],
num_rows: 113287
})
validation: Dataset({
features: ['image', 'filepath', 'sentids', 'filename', 'imgid', 'split', 'sentences_tokens', 'sentences_raw', 'sentences_sentid', 'cocoid', 'th_sentences_raw'],
num_rows: 5000
})
test: Dataset({
features: ['image', 'filepath', 'sentids', 'filename', 'imgid', 'split', 'sentences_tokens', 'sentences_raw', 'sentences_sentid', 'cocoid', 'th_sentences_raw'],
num_rows: 5000
})
})
```
A sample
```python
dataset["validation"][0]
```
output
```python
{
"image":<PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=500x336 at 0x7F6C5A83F430>,
"filepath":"COCO_val2014_000000184613.jpg",
"sentids":[474921,479322,479334,481560,483594],
"filename":"COCO_val2014_000000184613.jpg",
"imgid":2,
"split":"val",
"sentences_tokens":[
["a", "child","holding", "a","flowered","umbrella","and","petting","a","yak"],["a","young","man","holding","an","umbrella","next","to","a","herd","of","cattle"],
["a","young","boy","barefoot","holding","an","umbrella","touching","the","horn","of","a","cow"],
["a","young","boy","with","an","umbrella","who","is","touching","the","horn","of","a","cow"],
["a","boy","holding","an","umbrella","while","standing","next","to","livestock"]
],
"sentences_raw":[
"A child holding a flowered umbrella and petting a yak.",
"A young man holding an umbrella next to a herd of cattle.",
"a young boy barefoot holding an umbrella touching the horn of a cow",
"A young boy with an umbrella who is touching the horn of a cow.",
"A boy holding an umbrella while standing next to livestock."
],
"sentences_sentid":[474921,479322,479334,481560,483594],
"cocoid":184613,
"th_sentences_raw":[
"เด็กถือร่มที่มีดอกหนึ่งคันและลูบคลูบลํา",
"ชายหนุ่มคนหนึ่งถือร่มไว้ข้างๆ ฝูงวัว",
"เด็กหนุ่มคนหนึ่งเท้าเปล่าจับร่มจับแตรของวัว",
"เด็กชายที่มีร่มสัมผัสแตรของวัว",
"เด็กชายถือร่มในขณะที่ยืนถัดจากปศุสัตว์"
]
}
```
## Dataset Construction
The dataset contructed from translating the captions of [MS COCO 2014 dataset](https://huggingface.co/datasets/HuggingFaceM4/COCO) [1] to Thai by using [NMT](https://airesearch.in.th/releases/machine-translation-models/) provided by VISTEC-depa Thailand Artificial Intelligence Research Institute [2]. The translated of 3 splits (train, validation and test) dataset was published in the [Huggingface](https://huggingface.co/datasets/patomp/thai-mscoco-2014-captions).
## References
[1] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C. Lawrence Zitnick. 2014. Microsoft COCO: Common Objects in Context. In Computer Vision – ECCV 2014, Springer International Publishing, Cham, 740–755.
[2] English-Thai Machine Translation Models. (2020, June 23). VISTEC-depa Thailand Artificial Intelligence Research Institute. https://airesearch.in.th/releases/machine-translation-models/ |
PNLPhub/digikala-sentiment-analysis | 2023-09-03T06:49:27.000Z | [
"region:us"
] | PNLPhub | null | null | null | 1 | 133 | ---
dataset_info:
features:
- name: Text
dtype: string
- name: Score
dtype: int64
- name: Suggestion
dtype: int64
splits:
- name: train
num_bytes: 1026393
num_examples: 2282
- name: validation
num_bytes: 234807
num_examples: 489
- name: test
num_bytes: 228949
num_examples: 490
download_size: 1444998
dataset_size: 1490149
---
|
shiroyasha13/llama_text_to_sql_dataset | 2023-08-29T11:47:05.000Z | [
"region:us"
] | shiroyasha13 | null | null | null | 5 | 133 | Entry not found |
godoyj/wikilingua | 2023-09-08T17:36:48.000Z | [
"task_categories:summarization",
"language:pt",
"region:us"
] | godoyj | null | null | null | 0 | 133 | ---
language:
- pt
task_categories:
- summarization
--- |
amitness/mlrs-pos-mt | 2023-09-26T16:27:07.000Z | [
"region:us"
] | amitness | null | null | null | 0 | 133 | ---
dataset_info:
features:
- name: pos_tags
sequence:
class_label:
names:
'0': ADJ
'1': ADV
'2': COMP
'3': CONJ_CORD
'4': CONJ_SUB
'5': DEF
'6': FOC
'7': FUT
'8': GEN
'9': GEN_DEF
'10': GEN_PRON
'11': HEMM
'12': INT
'13': KIEN
'14': LIL
'15': LIL_DEF
'16': LIL_PRON
'17': NEG
'18': NOUN
'19': NOUN_PROP
'20': NUM_CRD
'21': NUM_FRC
'22': NUM_ORD
'23': NUM_WHD
'24': PART_ACT
'25': PART_PASS
'26': PREP
'27': PREP_DEF
'28': PREP_PRON
'29': PROG
'30': PRON_DEM
'31': PRON_DEM_DEF
'32': PRON_INDEF
'33': PRON_INT
'34': PRON_PERS
'35': PRON_PERS_NEG
'36': PRON_REC
'37': PRON_REF
'38': QUAN
'39': VERB
'40': VERB_PSEU
'41': X_ABV
'42': X_BOR
'43': X_DIG
'44': X_ENG
'45': X_FOR
'46': X_PUN
- name: tokens
sequence: string
splits:
- name: train
num_bytes: 1443609
num_examples: 4935
- name: validation
num_bytes: 234214
num_examples: 616
- name: test
num_bytes: 212745
num_examples: 616
download_size: 0
dataset_size: 1890568
---
# Dataset Card for "mlrs-pos-mt"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
edarchimbaud/news-stocks | 2023-10-10T03:52:18.000Z | [
"region:us"
] | edarchimbaud | null | null | null | 3 | 132 | ---
dataset_info:
features:
- name: symbol
dtype: string
- name: body
dtype: string
- name: publisher
dtype: string
- name: publish_time
dtype: timestamp[ns, tz=GMT]
- name: title
dtype: string
- name: url
dtype: string
- name: uuid
dtype: string
splits:
- name: train
num_bytes: 106435084
num_examples: 22042
download_size: 52131999
dataset_size: 106435084
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
# Dataset Card for "news-sp500"
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** https://edarchimbaud.substack.com
- **Repository:** https://github.com/edarchimbaud
- **Point of Contact:** contact@edarchimbaud.com
### Dataset Summary
The news-sp500 dataset provides news articles related to companies in the S&P 500 index.
### Supported Tasks and Leaderboards
The dataset can be used for various natural language processing tasks such as text classification, sentiment analysis, information extraction, etc. It does not have a specific leaderboard associated with it.
### Languages
The dataset contains news articles in multiple languages.
## Dataset Structure
### Data Instances
The dataset consists of [1563] data instances.
### Data Fields
- symbol (string): A string representing the ticker symbol or abbreviation used to identify the company.
- body (string): The main content of the news article.
- publisher (string): The name of the publisher or news agency.
- publish_time (timestamp[ns, tz=GMT]): A timestamp indicating the publication time of the news article in GMT timezone.
- title (string): The title or headline of the news article.
- url (string): The URL or link to the original news article.
- uuid (string): A unique identifier for the news article.
### Data Splits
The dataset consists of a single split called train.
## Dataset Creation
### Curation Rationale
The news-sp500 dataset was created to provide a collection of news articles related to companies in the S&P 500 index for research and analysis purposes.
### Source Data
#### Initial Data Collection and Normalization
The data was collected from various online news sources and normalized for consistency.
### Annotations
#### Annotation process
[N/A]
#### Who are the annotators?
[N/A]
### Personal and Sensitive Information
[N/A]
## Considerations for Using the Data
### Social Impact of Dataset
[N/A]
### Discussion of Biases
[N/A]
### Other Known Limitations
[N/A]
## Additional Information
### Dataset Curators
The news-sp500 dataset was collected by https://edarchimbaud.substack.com.
### Licensing Information
The news-sp500 dataset is licensed under the MIT License.
### Citation Information
> https://edarchimbaud.substack.com, news-sp500 dataset, GitHub repository, https://github.com/edarchimbaud
### Contributions
Thanks to [@edarchimbaud](https://github.com/edarchimbaud) for adding this dataset. |
mmenendezg/pneumonia_x_ray | 2023-06-21T23:07:12.000Z | [
"region:us"
] | mmenendezg | null | null | null | 1 | 132 | ---
dataset_info:
features:
- name: image
dtype: image
- name: label
dtype:
class_label:
names:
'0': normal
'1': pneumonia
splits:
- name: train
num_bytes: 126684250.689
num_examples: 4187
- name: validation
num_bytes: 27444182.485
num_examples: 1045
- name: test
num_bytes: 16275660.0
num_examples: 624
download_size: 153021953
dataset_size: 170404093.174
---
# Chest X-Ray Pneumonia Dataset
This dataset contains chest x-ray images of independent patients that can be classified into `normal` (healthy) or `pneumonia` (diseased) patients.
This dataset is a processed version of the original `Large Dataset of Labeled Optical Coherence Tomography (OCT) and Chest X-Ray Images` dataset provided by the *University of California San Diego*.
The dataset contains three splits:
- **Train**: 4187 images
- **Validation**: 1045 images
- **Test**: 624 images
The shape of the images is `[500, 500, 3]`, and the labels have two possible values:
- 0: **Normal**
- 1: **Pneumonia**
>**References**:
>
> - Kermany, Daniel; Zhang, Kang; Goldbaum, Michael (2018), “Large Dataset of Labeled Optical Coherence Tomography (OCT) and Chest X-Ray Images”, Mendeley Data, V3, doi: 10.17632/rscbjbr9sj.3 |
alayaran/bodo-monolingual-dataset | 2023-09-16T16:13:42.000Z | [
"task_categories:text-generation",
"size_categories:100K<n<1M",
"language:brx",
"license:mit",
"region:us"
] | alayaran | null | null | null | 0 | 132 | ---
language:
- brx
license: mit
size_categories:
- 100K<n<1M
task_categories:
- text-generation
pretty_name: bodo-monolingual
dataset_info:
features:
- name: text
dtype: string
splits:
- name: train
num_bytes: 130783128
num_examples: 474703
- name: test
num_bytes: 4109084
num_examples: 14890
download_size: 51159931
dataset_size: 134892212
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
---
```
# First Install datasets library
pip install datasets
from datasets import load_dataset
train = load_dataset("alayaran/bodo-monolingual-dataset", "unshuffled_deduplicated_no", split="train")
test = load_dataset("alayaran/bodo-monolingual-dataset", "unshuffled_deduplicated_no", split="test")
# print the first five entries from the dataset array of trai and test set
print(train['text'][:5])
["मदि सरकारा जारिमिनारि हाबाफारि मावफूंदों , 1 कौटि नख'राव दैनि कानेक्सन होबाय",
"दिल्ली / जयपुर , 14 आगस्ट : गाहाय मन्थ्रि नरेन्द्र मदिनि नख'रफ्रोमबो दैनि सिमांखौ जाफुहोनो थाखाय मिरुआरि दै बिफाना बोसोरसेनि खम समावनो
1 कौटि नख'रफोराव लोंग्रा दैनि कानेक्सन होनानै जारिमिनारि खामानि मावबाय ।", "करना बेरामनि गेजेरजोंबो दै बिफाना हादतनि 1 कौटि नख'रफोरखौ नलि कानेक्सनजों सरजाबनाय ।",
'अदेबानि मिरुआरि दै मन्थ्रि गजेन्द्र सिं शेखावतआ खौरां होनाय बायदिब्ला खम समनि गेजेराव एसेबां गिदिर जाफुसारनाया गासै राइज्योफोर आरो मिरुजों खुंजानाय राइज्योफोरनि गेजेरजों नाजानायखौ फोरमायो ।',
"बेनि गेजेरजों थांनाय बृहस्पतिबारखालि शेखावतआ बु़ङोदि , आनलक - 1 जागायजेन्नायनिफ्राय बिसोर सानफ्रोमबो 1 लाखनिक्रुइ बांसिन नख'रफोरनो नलि कानेक्शन होगासिनो दङ ।"]
print(test['text'][:5])
['श्बयम बिथांखिनि सिङाव राइज्योनि सा 2 लाख लाइमोनफोरनो राइज्यो सरकारआ 50 रोजा रां होगोनः हिमन्त',
"गुवाहाटी , 5 सेप्तेम्बर : 2017 - 2018 माइथायनि बाजेटआव फोसावनाय जादोंमोन स्वामी बिबेकानन्द आसाम इयुथ इम्पावारमेन्ट यजना सुंद'यै श्बयमखौ ।",
'बे बिथांखिखौ फिन फोथांफिन्नो सरकारआ बे समाव गोनांथार राहा लादों ।', 'जायनि थाखाय बे बिथांखिनि मुडाव आसाम सरकारआ गासै 1000 कौटि रां बाहायगोन ।',
'बे बिथांखिनि सिङाव राइज्योनि सा 2 लाख सेंग्रा - सिख्ला लाइमोनफोरनो नोगोद 50 रोजा रां होनायनि हाबाफारि लादों ।']
```
# Use with training scripts
In order to use this dataset along with other models you can use as the following
```
# cd transformers/examples/pytorch/language-modelling
python run_mlm.py \
--output_dir="./bodo-roberta-base" \
--model_type="roberta" \
--do_train \
--do_eval \
--config_name="alayaran/bodo-roberta-base" \
--tokenizer_name="alayaran/bodo-roberta-base" \
--dataset_name="alayaran/bodo-monolingual-dataset" \ # <--- This is the dataset
--max_seq_length="128" \
--weight_decay="0.01" \
--per_device_train_batch_size="128" \
--per_device_eval_batch_size="128" \
--learning_rate="3e-4" \
--warmup_steps="1000" \
--overwrite_output_dir \
--num_train_epochs="18" \
--adam_beta1="0.9" \
--adam_beta2="0.98" \
--logging_steps="500" \
--save_steps="2500" \
--eval_steps="2500" \
--push_to_hub
```
# Ownership
I do not claim any ownership of the data shared here. The data is mostly from TDIL, and some texts are crawled using Bodo News Crawlers. All copyrights belong to the respective owners. This data is shared for research purposes only.
|
hippocrates/DDI_RE | 2023-10-04T19:08:58.000Z | [
"region:us"
] | hippocrates | null | null | null | 0 | 132 | Entry not found |
eurlex | 2022-11-18T20:01:34.000Z | [
"task_categories:text-classification",
"task_ids:multi-label-classification",
"annotations_creators:found",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:en",
"license:cc-by-sa-4.0",
"legal-topic-classification",
"region:us"
] | null | EURLEX57K contains 57k legislative documents in English from EUR-Lex portal, annotated with EUROVOC concepts. | @inproceedings{chalkidis-etal-2019-large,
title = "Large-Scale Multi-Label Text Classification on {EU} Legislation",
author = "Chalkidis, Ilias and Fergadiotis, Emmanouil and Malakasiotis, Prodromos and Androutsopoulos, Ion",
booktitle = "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics",
year = "2019",
address = "Florence, Italy",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/P19-1636",
doi = "10.18653/v1/P19-1636",
pages = "6314--6322"
} | null | 4 | 131 | ---
annotations_creators:
- found
language_creators:
- found
language:
- en
license:
- cc-by-sa-4.0
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- original
task_categories:
- text-classification
task_ids:
- multi-label-classification
paperswithcode_id: eurlex57k
pretty_name: the EUR-Lex dataset
tags:
- legal-topic-classification
dataset_info:
features:
- name: celex_id
dtype: string
- name: title
dtype: string
- name: text
dtype: string
- name: eurovoc_concepts
sequence: string
config_name: eurlex57k
splits:
- name: train
num_bytes: 167603718
num_examples: 45000
- name: test
num_bytes: 22046706
num_examples: 6000
- name: validation
num_bytes: 21942574
num_examples: 6000
download_size: 50289403
dataset_size: 211592998
---
# Dataset Card for the EUR-Lex dataset
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** http://nlp.cs.aueb.gr/software_and_datasets/EURLEX57K/
- **Repository:** http://nlp.cs.aueb.gr/software_and_datasets/EURLEX57K/
- **Paper:** https://www.aclweb.org/anthology/P19-1636/
- **Leaderboard:** N/A
- **Point of Contact:** [Ilias Chalkidis](mailto:ihalk@aueb.gr)
### Dataset Summary
EURLEX57K can be viewed as an improved version of the dataset released by Mencia and Furnkranzand (2007), which has been widely used in Large-scale Multi-label Text Classification (LMTC) research, but is less than half the size of EURLEX57K (19.6k documents, 4k EUROVOC labels) and more than ten years old.
EURLEX57K contains 57k legislative documents in English from EUR-Lex (https://eur-lex.europa.eu) with an average length of 727 words. Each document contains four major zones:
- the header, which includes the title and name of the legal body enforcing the legal act;
- the recitals, which are legal background references; and
- the main body, usually organized in articles.
**Labeling / Annotation**
All the documents of the dataset have been annotated by the Publications Office of EU (https://publications.europa.eu/en) with multiple concepts from EUROVOC (http://eurovoc.europa.eu/).
While EUROVOC includes approx. 7k concepts (labels), only 4,271 (59.31%) are present in EURLEX57K, from which only 2,049 (47.97%) have been assigned to more than 10 documents. The 4,271 labels are also divided into frequent (746 labels), few-shot (3,362), and zero- shot (163), depending on whether they were assigned to more than 50, fewer than 50 but at least one, or no training documents, respectively.
### Supported Tasks and Leaderboards
The dataset supports:
**Multi-label Text Classification:** Given the text of a document, a model predicts the relevant EUROVOC concepts.
**Few-shot and Zero-shot learning:** As already noted, the labels can be divided into three groups: frequent (746 labels), few-shot (3,362), and zero- shot (163), depending on whether they were assigned to more than 50, fewer than 50 but at least one, or no training documents, respectively.
### Languages
All documents are written in English.
## Dataset Structure
### Data Instances
```json
{
"celex_id": "31979D0509",
"title": "79/509/EEC: Council Decision of 24 May 1979 on financial aid from the Community for the eradication of African swine fever in Spain",
"text": "COUNCIL DECISION of 24 May 1979 on financial aid from the Community for the eradication of African swine fever in Spain (79/509/EEC)\nTHE COUNCIL OF THE EUROPEAN COMMUNITIES\nHaving regard to the Treaty establishing the European Economic Community, and in particular Article 43 thereof,\nHaving regard to the proposal from the Commission (1),\nHaving regard to the opinion of the European Parliament (2),\nWhereas the Community should take all appropriate measures to protect itself against the appearance of African swine fever on its territory;\nWhereas to this end the Community has undertaken, and continues to undertake, action designed to contain outbreaks of this type of disease far from its frontiers by helping countries affected to reinforce their preventive measures ; whereas for this purpose Community subsidies have already been granted to Spain;\nWhereas these measures have unquestionably made an effective contribution to the protection of Community livestock, especially through the creation and maintenance of a buffer zone north of the river Ebro;\nWhereas, however, in the opinion of the Spanish authorities themselves, the measures so far implemented must be reinforced if the fundamental objective of eradicating the disease from the entire country is to be achieved;\nWhereas the Spanish authorities have asked the Community to contribute to the expenses necessary for the efficient implementation of a total eradication programme;\nWhereas a favourable response should be given to this request by granting aid to Spain, having regard to the undertaking given by that country to protect the Community against African swine fever and to eliminate completely this disease by the end of a five-year eradication plan;\nWhereas this eradication plan must include certain measures which guarantee the effectiveness of the action taken, and it must be possible to adapt these measures to developments in the situation by means of a procedure establishing close cooperation between the Member States and the Commission;\nWhereas it is necessary to keep the Member States regularly informed as to the progress of the action undertaken,",
"eurovoc_concepts": ["192", "2356", "2560", "862", "863"]
}
```
### Data Fields
The following data fields are provided for documents (`train`, `dev`, `test`):
`celex_id`: (**str**) The official ID of the document. The CELEX number is the unique identifier for all publications in both Eur-Lex and CELLAR.\
`title`: (**str**) The title of the document.\
`text`: (**str**) The full content of each document, which is represented by its `header`, `recitals` and `main_body`.\
`eurovoc_concepts`: (**List[str]**) The relevant EUROVOC concepts (labels).
If you want to use the descriptors of EUROVOC concepts, similar to Chalkidis et al. (2020), please load: https://archive.org/download/EURLEX57K/eurovoc_concepts.jsonl
```python
import json
with open('./eurovoc_concepts.jsonl') as jsonl_file:
eurovoc_concepts = {json.loads(concept) for concept in jsonl_file.readlines()}
```
### Data Splits
| Split | No of Documents | Avg. words | Avg. labels |
| ------------------- | ------------------------------------ | --- | --- |
| Train | 45,000 | 729 | 5 |
|Development | 6,000 | 714 | 5 |
|Test | 6,000 | 725 | 5 |
## Dataset Creation
### Curation Rationale
The dataset was curated by Chalkidis et al. (2019).\
The documents have been annotated by the Publications Office of EU (https://publications.europa.eu/en).
### Source Data
#### Initial Data Collection and Normalization
The original data are available at EUR-Lex portal (https://eur-lex.europa.eu) in an unprocessed format.
The documents were downloaded from EUR-Lex portal in HTML format.
The relevant metadata and EUROVOC concepts were downloaded from the SPARQL endpoint of the Publications Office of EU (http://publications.europa.eu/webapi/rdf/sparql).
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
* The original documents are available at EUR-Lex portal (https://eur-lex.europa.eu) in an unprocessed HTML format. The HTML code was striped and the documents split into sections.
* The documents have been annotated by the Publications Office of EU (https://publications.europa.eu/en).
#### Who are the annotators?
Publications Office of EU (https://publications.europa.eu/en)
### Personal and Sensitive Information
The dataset does not include personal or sensitive information.
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
Chalkidis et al. (2019)
### Licensing Information
© European Union, 1998-2021
The Commission’s document reuse policy is based on Decision 2011/833/EU. Unless otherwise specified, you can re-use the legal documents published in EUR-Lex for commercial or non-commercial purposes.
The copyright for the editorial content of this website, the summaries of EU legislation and the consolidated texts, which is owned by the EU, is licensed under the Creative Commons Attribution 4.0 International licence. This means that you can re-use the content provided you acknowledge the source and indicate any changes you have made.
Source: https://eur-lex.europa.eu/content/legal-notice/legal-notice.html \
Read more: https://eur-lex.europa.eu/content/help/faq/reuse-contents-eurlex.html
### Citation Information
*Ilias Chalkidis, Manos Fergadiotis, Prodromos Malakasiotis and Ion Androutsopoulos.*
*Large-Scale Multi-Label Text Classification on EU Legislation.*
*Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (ACL 2019). Florence, Italy. 2019*
```
@inproceedings{chalkidis-etal-2019-large,
title = "Large-Scale Multi-Label Text Classification on {EU} Legislation",
author = "Chalkidis, Ilias and Fergadiotis, Manos and Malakasiotis, Prodromos and Androutsopoulos, Ion",
booktitle = "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics",
year = "2019",
address = "Florence, Italy",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/P19-1636",
doi = "10.18653/v1/P19-1636",
pages = "6314--6322"
}
```
### Contributions
Thanks to [@iliaschalkidis](https://github.com/iliaschalkidis) for adding this dataset. |
pragmeval | 2023-06-01T14:59:54.000Z | [
"task_categories:text-classification",
"task_ids:multi-class-classification",
"annotations_creators:found",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"size_categories:1K<n<10K",
"size_categories:n<1K",
"source_datasets:original",
"language:en",
"license:unknown",
"region:us"
] | null | Evaluation of language understanding with a 11 datasets benchmark focusing on discourse and pragmatics | @misc{sileo2019discoursebased,
title={Discourse-Based Evaluation of Language Understanding},
author={Damien Sileo and Tim Van-de-Cruys and Camille Pradel and Philippe Muller},
year={2019},
eprint={1907.08672},
archivePrefix={arXiv},
primaryClass={cs.CL}
} | null | 3 | 131 | ---
annotations_creators:
- found
language_creators:
- found
language:
- en
license:
- unknown
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
- 1K<n<10K
- n<1K
source_datasets:
- original
task_categories:
- text-classification
task_ids:
- multi-class-classification
pretty_name: pragmeval
dataset_info:
- config_name: verifiability
features:
- name: sentence
dtype: string
- name: label
dtype:
class_label:
names:
'0': experiential
'1': unverifiable
'2': non-experiential
- name: idx
dtype: int32
splits:
- name: train
num_bytes: 592520
num_examples: 5712
- name: validation
num_bytes: 65215
num_examples: 634
- name: test
num_bytes: 251799
num_examples: 2424
download_size: 5330724
dataset_size: 909534
- config_name: emobank-arousal
features:
- name: sentence
dtype: string
- name: label
dtype:
class_label:
names:
'0': low
'1': high
- name: idx
dtype: int32
splits:
- name: train
num_bytes: 567660
num_examples: 5470
- name: validation
num_bytes: 71221
num_examples: 684
- name: test
num_bytes: 69276
num_examples: 683
download_size: 5330724
dataset_size: 708157
- config_name: switchboard
features:
- name: sentence
dtype: string
- name: label
dtype:
class_label:
names:
'0': Response Acknowledgement
'1': Uninterpretable
'2': Or-Clause
'3': Reject
'4': Statement-non-opinion
'5': 3rd-party-talk
'6': Repeat-phrase
'7': Hold Before Answer/Agreement
'8': Signal-non-understanding
'9': Offers, Options Commits
'10': Agree/Accept
'11': Dispreferred Answers
'12': Hedge
'13': Action-directive
'14': Tag-Question
'15': Self-talk
'16': Yes-No-Question
'17': Rhetorical-Question
'18': No Answers
'19': Open-Question
'20': Conventional-closing
'21': Other Answers
'22': Acknowledge (Backchannel)
'23': Wh-Question
'24': Declarative Wh-Question
'25': Thanking
'26': Yes Answers
'27': Affirmative Non-yes Answers
'28': Declarative Yes-No-Question
'29': Backchannel in Question Form
'30': Apology
'31': Downplayer
'32': Conventional-opening
'33': Collaborative Completion
'34': Summarize/Reformulate
'35': Negative Non-no Answers
'36': Statement-opinion
'37': Appreciation
'38': Other
'39': Quotation
'40': Maybe/Accept-part
- name: idx
dtype: int32
splits:
- name: train
num_bytes: 1021220
num_examples: 18930
- name: validation
num_bytes: 116058
num_examples: 2113
- name: test
num_bytes: 34013
num_examples: 649
download_size: 5330724
dataset_size: 1171291
- config_name: persuasiveness-eloquence
features:
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype:
class_label:
names:
'0': low
'1': high
- name: idx
dtype: int32
splits:
- name: train
num_bytes: 153946
num_examples: 725
- name: validation
num_bytes: 19376
num_examples: 91
- name: test
num_bytes: 18379
num_examples: 90
download_size: 5330724
dataset_size: 191701
- config_name: mrda
features:
- name: sentence
dtype: string
- name: label
dtype:
class_label:
names:
'0': Declarative-Question
'1': Statement
'2': Reject
'3': Or-Clause
'4': 3rd-party-talk
'5': Continuer
'6': Hold Before Answer/Agreement
'7': Assessment/Appreciation
'8': Signal-non-understanding
'9': Floor Holder
'10': Sympathy
'11': Dispreferred Answers
'12': Reformulate/Summarize
'13': Exclamation
'14': Interrupted/Abandoned/Uninterpretable
'15': Expansions of y/n Answers
'16': Action-directive
'17': Tag-Question
'18': Accept
'19': Rhetorical-question Continue
'20': Self-talk
'21': Rhetorical-Question
'22': Yes-No-question
'23': Open-Question
'24': Rising Tone
'25': Other Answers
'26': Commit
'27': Wh-Question
'28': Repeat
'29': Follow Me
'30': Thanking
'31': Offer
'32': About-task
'33': Reject-part
'34': Affirmative Non-yes Answers
'35': Apology
'36': Downplayer
'37': Humorous Material
'38': Accept-part
'39': Collaborative Completion
'40': Mimic Other
'41': Understanding Check
'42': Misspeak Self-Correction
'43': Or-Question
'44': Topic Change
'45': Negative Non-no Answers
'46': Floor Grabber
'47': Correct-misspeaking
'48': Maybe
'49': Acknowledge-answer
'50': Defending/Explanation
- name: idx
dtype: int32
splits:
- name: train
num_bytes: 963913
num_examples: 14484
- name: validation
num_bytes: 111813
num_examples: 1630
- name: test
num_bytes: 419797
num_examples: 6459
download_size: 5330724
dataset_size: 1495523
- config_name: gum
features:
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype:
class_label:
names:
'0': preparation
'1': evaluation
'2': circumstance
'3': solutionhood
'4': justify
'5': result
'6': evidence
'7': purpose
'8': concession
'9': elaboration
'10': background
'11': condition
'12': cause
'13': restatement
'14': motivation
'15': antithesis
'16': no_relation
- name: idx
dtype: int32
splits:
- name: train
num_bytes: 270401
num_examples: 1700
- name: validation
num_bytes: 35405
num_examples: 259
- name: test
num_bytes: 40334
num_examples: 248
download_size: 5330724
dataset_size: 346140
- config_name: emergent
features:
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype:
class_label:
names:
'0': observing
'1': for
'2': against
- name: idx
dtype: int32
splits:
- name: train
num_bytes: 313257
num_examples: 2076
- name: validation
num_bytes: 38948
num_examples: 259
- name: test
num_bytes: 38842
num_examples: 259
download_size: 5330724
dataset_size: 391047
- config_name: persuasiveness-relevance
features:
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype:
class_label:
names:
'0': low
'1': high
- name: idx
dtype: int32
splits:
- name: train
num_bytes: 153158
num_examples: 725
- name: validation
num_bytes: 19663
num_examples: 91
- name: test
num_bytes: 18880
num_examples: 90
download_size: 5330724
dataset_size: 191701
- config_name: persuasiveness-specificity
features:
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype:
class_label:
names:
'0': low
'1': high
- name: idx
dtype: int32
splits:
- name: train
num_bytes: 106594
num_examples: 504
- name: validation
num_bytes: 13766
num_examples: 62
- name: test
num_bytes: 12712
num_examples: 62
download_size: 5330724
dataset_size: 133072
- config_name: persuasiveness-strength
features:
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype:
class_label:
names:
'0': low
'1': high
- name: idx
dtype: int32
splits:
- name: train
num_bytes: 79679
num_examples: 371
- name: validation
num_bytes: 10052
num_examples: 46
- name: test
num_bytes: 10225
num_examples: 46
download_size: 5330724
dataset_size: 99956
- config_name: emobank-dominance
features:
- name: sentence
dtype: string
- name: label
dtype:
class_label:
names:
'0': low
'1': high
- name: idx
dtype: int32
splits:
- name: train
num_bytes: 660303
num_examples: 6392
- name: validation
num_bytes: 86802
num_examples: 798
- name: test
num_bytes: 83319
num_examples: 798
download_size: 5330724
dataset_size: 830424
- config_name: squinky-implicature
features:
- name: sentence
dtype: string
- name: label
dtype:
class_label:
names:
'0': low
'1': high
- name: idx
dtype: int32
splits:
- name: train
num_bytes: 471552
num_examples: 3724
- name: validation
num_bytes: 58087
num_examples: 465
- name: test
num_bytes: 56549
num_examples: 465
download_size: 5330724
dataset_size: 586188
- config_name: sarcasm
features:
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype:
class_label:
names:
'0': notsarc
'1': sarc
- name: idx
dtype: int32
splits:
- name: train
num_bytes: 2177332
num_examples: 3754
- name: validation
num_bytes: 257834
num_examples: 469
- name: test
num_bytes: 269724
num_examples: 469
download_size: 5330724
dataset_size: 2704890
- config_name: squinky-formality
features:
- name: sentence
dtype: string
- name: label
dtype:
class_label:
names:
'0': low
'1': high
- name: idx
dtype: int32
splits:
- name: train
num_bytes: 459721
num_examples: 3622
- name: validation
num_bytes: 59921
num_examples: 453
- name: test
num_bytes: 58242
num_examples: 452
download_size: 5330724
dataset_size: 577884
- config_name: stac
features:
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype:
class_label:
names:
'0': Comment
'1': Contrast
'2': Q_Elab
'3': Parallel
'4': Explanation
'5': Narration
'6': Continuation
'7': Result
'8': Acknowledgement
'9': Alternation
'10': Question_answer_pair
'11': Correction
'12': Clarification_question
'13': Conditional
'14': Sequence
'15': Elaboration
'16': Background
'17': no_relation
- name: idx
dtype: int32
splits:
- name: train
num_bytes: 645969
num_examples: 11230
- name: validation
num_bytes: 71400
num_examples: 1247
- name: test
num_bytes: 70451
num_examples: 1304
download_size: 5330724
dataset_size: 787820
- config_name: pdtb
features:
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype:
class_label:
names:
'0': Synchrony
'1': Contrast
'2': Asynchronous
'3': Conjunction
'4': List
'5': Condition
'6': Pragmatic concession
'7': Restatement
'8': Pragmatic cause
'9': Alternative
'10': Pragmatic condition
'11': Pragmatic contrast
'12': Instantiation
'13': Exception
'14': Cause
'15': Concession
- name: idx
dtype: int32
splits:
- name: train
num_bytes: 2968638
num_examples: 12907
- name: validation
num_bytes: 276997
num_examples: 1204
- name: test
num_bytes: 235851
num_examples: 1085
download_size: 5330724
dataset_size: 3481486
- config_name: persuasiveness-premisetype
features:
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype:
class_label:
names:
'0': testimony
'1': warrant
'2': invented_instance
'3': common_knowledge
'4': statistics
'5': analogy
'6': definition
'7': real_example
- name: idx
dtype: int32
splits:
- name: train
num_bytes: 122631
num_examples: 566
- name: validation
num_bytes: 15920
num_examples: 71
- name: test
num_bytes: 14395
num_examples: 70
download_size: 5330724
dataset_size: 152946
- config_name: squinky-informativeness
features:
- name: sentence
dtype: string
- name: label
dtype:
class_label:
names:
'0': low
'1': high
- name: idx
dtype: int32
splits:
- name: train
num_bytes: 464855
num_examples: 3719
- name: validation
num_bytes: 60447
num_examples: 465
- name: test
num_bytes: 56872
num_examples: 464
download_size: 5330724
dataset_size: 582174
- config_name: persuasiveness-claimtype
features:
- name: sentence1
dtype: string
- name: sentence2
dtype: string
- name: label
dtype:
class_label:
names:
'0': Value
'1': Fact
'2': Policy
- name: idx
dtype: int32
splits:
- name: train
num_bytes: 31259
num_examples: 160
- name: validation
num_bytes: 3803
num_examples: 20
- name: test
num_bytes: 3717
num_examples: 19
download_size: 5330724
dataset_size: 38779
- config_name: emobank-valence
features:
- name: sentence
dtype: string
- name: label
dtype:
class_label:
names:
'0': low
'1': high
- name: idx
dtype: int32
splits:
- name: train
num_bytes: 539652
num_examples: 5150
- name: validation
num_bytes: 62809
num_examples: 644
- name: test
num_bytes: 66178
num_examples: 643
download_size: 5330724
dataset_size: 668639
config_names:
- emergent
- emobank-arousal
- emobank-dominance
- emobank-valence
- gum
- mrda
- pdtb
- persuasiveness-claimtype
- persuasiveness-eloquence
- persuasiveness-premisetype
- persuasiveness-relevance
- persuasiveness-specificity
- persuasiveness-strength
- sarcasm
- squinky-formality
- squinky-implicature
- squinky-informativeness
- stac
- switchboard
- verifiability
---
# Dataset Card for pragmeval
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:**
- **Repository:**
- **Paper:**
- **Leaderboard:**
- **Point of Contact:**
### Dataset Summary
[More Information Needed]
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
[More Information Needed]
## Dataset Structure
### Data Instances
[More Information Needed]
### Data Fields
[More Information Needed]
### Data Splits
[More Information Needed]
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
[More Information Needed]
#### Who are the source language producers?
[More Information Needed]
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
Thanks to [@sileod](https://github.com/sileod) for adding this dataset. |
maastrichtlawtech/bsard | 2023-09-26T15:28:00.000Z | [
"task_categories:text-retrieval",
"task_categories:text-classification",
"task_ids:document-retrieval",
"annotations_creators:expert-generated",
"language_creators:found",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:fr",
"license:cc-by-nc-sa-4.0",
"legal",
"arxiv:2108.11792",
"region:us"
] | maastrichtlawtech | The Belgian Statutory Article Retrieval Dataset (BSARD) is a French native dataset for studying legal information retrieval.
BSARD consists of more than 22,600 statutory articles from Belgian law and about 1,100 legal questions posed by Belgian citizens
and labeled by experienced jurists with relevant articles from the corpus. | @inproceedings{louis-spanakis-2022-statutory,
title = "A Statutory Article Retrieval Dataset in {F}rench",
author = "Louis, Antoine and Spanakis, Gerasimos",
booktitle = "Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
month = may,
year = "2022",
address = "Dublin, Ireland",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2022.acl-long.468",
doi = "10.18653/v1/2022.acl-long.468",
pages = "6789--6803",
} | null | 2 | 131 | ---
annotations_creators:
- expert-generated
language_creators:
- found
language:
- fr
license:
- cc-by-nc-sa-4.0
multilinguality:
- monolingual
pretty_name: BSARD
size_categories:
- 10K<n<100K
source_datasets:
- original
task_categories:
- text-retrieval
- text-classification
task_ids:
- document-retrieval
paperswithcode_id: bsard
tags:
- legal
---
# Dataset Card for BSARD
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Repository:** [maastrichtlawtech/bsard](https://github.com/maastrichtlawtech/bsard)
- **Paper:** [A Statutory Article Retrieval Dataset in French](https://arxiv.org/abs/2108.11792)
- **Point of Contact:** [Maastricht Law & Tech Lab](law-techlab@maastrichtuniversity.nl)
### Dataset Summary
The Belgian Statutory Article Retrieval Dataset (BSARD) is a French native dataset for studying legal information retrieval. BSARD consists of more than 22,600 statutory articles from Belgian law and about 1,100 legal questions posed by Belgian citizens and labeled by experienced jurists with relevant articles from the corpus.
### Supported Tasks and Leaderboards
- `document-retrieval`: The dataset can be used to train models for ad-hoc legal information retrieval. Such model is presented with a short user query written in natural language and asked to retrieve relevant legal information from a knowledge source (such as statutory articles).
### Languages
The text in the dataset is in French, as spoken in Wallonia and Brussels-Capital region. The associated BCP-47 code is `fr-BE`.
## Dataset Structure
### Data Instances
A typical data point comprises a question, with additional `category`, `subcategory`, and `extra_description` fields that elaborate on it, and a list of `article_ids` from the corpus of statutory articles that are relevant to the question.
An example from the BSARD test set looks as follows:
```
{
'id': '724',
'question': 'La police peut-elle me fouiller pour chercher du cannabis ?',
'category': 'Justice',
'subcategory': 'Petite délinquance',
'extra_description': 'Détenir, acheter et vendre du cannabis',
'article_ids': '13348'
}
```
### Data Fields
- In **"questions_fr_train.csv"** and **"questions_fr_test.csv"**:
- `id`: an *int32* feature corresponding to a unique ID number for the question.
- `question`: a *string* feature corresponding to the question.
- `category`: a *string* feature corresponding to the general topic of the question.
- `subcategory`: a *string* feature corresponding to the sub-topic of the question.
- `extra_description`: a *string* feature corresponding to the extra categorization tags of the question.
- `article_ids`: a *string* feature of comma-separated article IDs relevant to the question.
- In **"articles_fr.csv"**:
- `id`: an *int32* feature corresponding to a unique ID number for the article.
- `article`: a *string* feature corresponding to the full article.
- `code`: a *string* feature corresponding to the law code to which the article belongs.
- `article_no`: a *string* feature corresponding to the article number in the code.
- `description`: a *string* feature corresponding to the concatenated headings of the article.
- `law_type`: a *string* feature whose value is either *"regional"* or *"national"*.
### Data Splits
This dataset is split into train/test set. Number of questions in each set is given below:
| | Train | Test |
| ----- | ------ | ---- |
| BSARD | 886 | 222 |
## Dataset Creation
### Curation Rationale
The dataset is intended to be used by researchers to build and evaluate models on retrieving law articles relevant to an input legal question. It should not be regarded as a reliable source of legal information at this point in time, as both the questions and articles correspond to an outdated version of the Belgian law from May 2021 (time of dataset collection). In the latter case, the user is advised to consult daily updated official legal resources (e.g., the Belgian Official Gazette).
### Source Data
#### Initial Data Collection and Normalization
BSARD was created in four stages: (i) compiling a large corpus of Belgian law articles, (ii) gathering legal questions with references to relevant law articles, (iii) refining these questions, and (iv) matching the references to the corresponding articles from the corpus.
#### Who are the source language producers?
Speakers were not directly approached for inclusion in this dataset and thus could not be asked for demographic information. Questions were collected, anonimyzed, and reformulated by [Droits Quotidiens](https://www.droitsquotidiens.be/fr/equipe). Therefore, no direct information about the speakers’ age and gender distribution, or socioeconomic status is available. However, it is expected that most, but not all, of the speakers are adults (18+ years), speak French as a native language, and live in Wallonia or Brussels-Capital region.
### Annotations
#### Annotation process
Each year, [Droits Quotidiens](https://www.droitsquotidiens.be/fr/equipe), a Belgian organization whose mission is to clarify the law for laypeople, receives and collects around 4,000 emails from Belgian citizens asking for advice on a personal legal issue. In practice, their legal clarification process consists of four steps. First, they identify the most frequently asked questions on a common legal issue. Then, they define a new anonymized "model" question on that issue expressed in natural language terms, i.e., as close as possible as if a layperson had asked it. Next, they search the Belgian law for articles that help answer the model question and reference them.
#### Who are the annotators?
A total of six Belgian jurists from [Droits Quotidiens](https://www.droitsquotidiens.be/fr/equipe) contributed to annotating the questions. All have a law degree from a Belgian university and years of experience in providing legal advice and clarifications of the law. They range in age from 30-60 years, including one man and five women, gave their ethnicity as white European, speak French as a native language, and represent upper middle class based on income levels.
### Personal and Sensitive Information
The questions represent informal, asynchronous, edited, written language that does not exceed 44 words. None of them contained hateful, aggressive, or inappropriate language as they were all reviewed and reworded by Droits Quotidiens to be neutral, anonymous, and comprehensive. The legal articles represent strong, formal, written language that can contain up to 5,790 words.
## Considerations for Using the Data
### Social Impact of Dataset
In addition to helping advance the state-of-the-art in retrieving statutes relevant to a legal question, BSARD-based models could improve the efficiency of the legal information retrieval process in the context of legal research, therefore enabling researchers to devote themselves to more thoughtful parts of their research. Furthermore, BSARD can become a starting point of new open-source legal information search tools so that the socially weaker parties to disputes can benefit from a free professional assisting service.
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
First, the corpus of articles is limited to those collected from 32 Belgian codes, which obviously does not cover the entire Belgian law as thousands of articles from decrees, directives, and ordinances are missing. During the dataset construction, all references to these uncollected articles are ignored, which causes some questions to end up with only a fraction of their initial number of relevant articles. This information loss implies that the answer contained in the remaining relevant articles might be incomplete, although it is still appropriate.
Additionally, it is essential to note that not all legal questions can be answered with statutes alone. For instance, the question “Can I evict my tenants if they make too much noise?” might not have a detailed answer within the statutory law that quantifies a specific noise threshold at which eviction is allowed. Instead, the landlord should probably rely more on case law and find precedents similar to their current situation (e.g., the tenant makes two parties a week until 2 am). Hence, some questions are better suited than others to the statutory article retrieval task, and the domain of the less suitable ones remains to be determined.
## Additional Information
### Dataset Curators
The dataset was created by Antoine Louis during work done at the Law & Tech lab of Maastricht University, with the help of jurists from [Droits Quotidiens](https://www.droitsquotidiens.be/fr/equipe).
### Licensing Information
BSARD is licensed under the [CC BY-NC-SA 4.0 license](https://creativecommons.org/licenses/by-nc-sa/4.0/).
### Citation Information
```latex
@inproceedings{louis2022statutory,
title = {A Statutory Article Retrieval Dataset in French},
author = {Louis, Antoine and Spanakis, Gerasimos},
booktitle = {Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics},
month = may,
year = {2022},
address = {Dublin, Ireland},
publisher = {Association for Computational Linguistics},
url = {https://aclanthology.org/2022.acl-long.468/},
doi = {10.18653/v1/2022.acl-long.468},
pages = {6789–6803},
}
```
### Contributions
Thanks to [@antoiloui](https://github.com/antoiloui) for adding this dataset. |
zeroshot/twitter-financial-news-topic | 2022-12-04T16:50:10.000Z | [
"task_categories:text-classification",
"task_ids:multi-class-classification",
"annotations_creators:other",
"language_creators:other",
"multilinguality:monolingual",
"size_categories:10K<n<100K",
"source_datasets:original",
"language:en",
"license:mit",
"twitter",
"finance",
"markets",
"stocks",
"wallstreet",
"quant",
"hedgefunds",
"region:us"
] | zeroshot | null | null | null | 16 | 131 | ---
annotations_creators:
- other
language:
- en
language_creators:
- other
license:
- mit
multilinguality:
- monolingual
pretty_name: twitter financial news
size_categories:
- 10K<n<100K
source_datasets:
- original
tags:
- twitter
- finance
- markets
- stocks
- wallstreet
- quant
- hedgefunds
- markets
task_categories:
- text-classification
task_ids:
- multi-class-classification
---
Read this [BLOG](https://neuralmagic.com/blog/classifying-finance-tweets-in-real-time-with-sparse-transformers/) to see how I fine-tuned a sparse transformer on this dataset.
### Dataset Description
The Twitter Financial News dataset is an English-language dataset containing an annotated corpus of finance-related tweets. This dataset is used to classify finance-related tweets for their topic.
1. The dataset holds 21,107 documents annotated with 20 labels:
```python
topics = {
"LABEL_0": "Analyst Update",
"LABEL_1": "Fed | Central Banks",
"LABEL_2": "Company | Product News",
"LABEL_3": "Treasuries | Corporate Debt",
"LABEL_4": "Dividend",
"LABEL_5": "Earnings",
"LABEL_6": "Energy | Oil",
"LABEL_7": "Financials",
"LABEL_8": "Currencies",
"LABEL_9": "General News | Opinion",
"LABEL_10": "Gold | Metals | Materials",
"LABEL_11": "IPO",
"LABEL_12": "Legal | Regulation",
"LABEL_13": "M&A | Investments",
"LABEL_14": "Macro",
"LABEL_15": "Markets",
"LABEL_16": "Politics",
"LABEL_17": "Personnel Change",
"LABEL_18": "Stock Commentary",
"LABEL_19": "Stock Movement",
}
```
The data was collected using the Twitter API. The current dataset supports the multi-class classification task.
### Task: Topic Classification
# Data Splits
There are 2 splits: train and validation. Below are the statistics:
| Dataset Split | Number of Instances in Split |
| ------------- | ------------------------------------------- |
| Train | 16,990 |
| Validation | 4,118 |
# Licensing Information
The Twitter Financial Dataset (topic) version 1.0.0 is released under the MIT License. |
tomekkorbak/detoxify-pile-chunk3-1450000-1500000 | 2022-10-04T23:56:05.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 131 | Entry not found |
tomekkorbak/detoxify-pile-chunk3-1600000-1650000 | 2022-10-04T23:57:39.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 131 | Entry not found |
tomekkorbak/detoxify-pile-chunk3-1650000-1700000 | 2022-10-05T00:01:13.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 131 | Entry not found |
dmayhem93/agieval-sat-math | 2023-06-18T17:32:05.000Z | [
"license:mit",
"arxiv:2304.06364",
"region:us"
] | dmayhem93 | null | null | null | 5 | 131 | ---
dataset_info:
features:
- name: query
dtype: string
- name: choices
sequence: string
- name: gold
sequence: int64
splits:
- name: test
num_bytes: 110388
num_examples: 220
download_size: 57002
dataset_size: 110388
license: mit
---
# Dataset Card for "agieval-sat-math"
Dataset taken from https://github.com/microsoft/AGIEval and processed as in that repo.
MIT License
Copyright (c) Microsoft Corporation.
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in all
copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
SOFTWARE
@misc{zhong2023agieval,
title={AGIEval: A Human-Centric Benchmark for Evaluating Foundation Models},
author={Wanjun Zhong and Ruixiang Cui and Yiduo Guo and Yaobo Liang and Shuai Lu and Yanlin Wang and Amin Saied and Weizhu Chen and Nan Duan},
year={2023},
eprint={2304.06364},
archivePrefix={arXiv},
primaryClass={cs.CL}
} |
Globaly/first-dataset | 2023-09-11T15:39:29.000Z | [
"region:us"
] | Globaly | null | null | null | 0 | 131 | Entry not found |
open-source-metrics/pip | 2023-10-03T09:12:23.000Z | [
"region:us"
] | open-source-metrics | null | null | null | 0 | 130 | ---
dataset_info:
features:
- name: day
dtype: string
- name: num_downloads
dtype: int64
splits:
- name: transformers
num_bytes: 25454
num_examples: 1157
- name: peft
num_bytes: 5632
num_examples: 256
- name: diffusers
num_bytes: 10780
num_examples: 490
- name: pytorch_image_models
num_bytes: 24772
num_examples: 1126
- name: datasets
num_bytes: 21406
num_examples: 973
- name: safetensors
num_bytes: 6842
num_examples: 311
- name: optimum
num_bytes: 16390
num_examples: 745
- name: huggingface_hub
num_bytes: 22286
num_examples: 1013
- name: accelerate
num_bytes: 21406
num_examples: 973
- name: evaluate
num_bytes: 13376
num_examples: 608
- name: tokenizers
num_bytes: 24772
num_examples: 1126
- name: gradio
num_bytes: 24772
num_examples: 1126
download_size: 128321
dataset_size: 217888
configs:
- config_name: default
data_files:
- split: accelerate
path: data/accelerate-*
- split: datasets
path: data/datasets-*
- split: diffusers
path: data/diffusers-*
- split: evaluate
path: data/evaluate-*
- split: gradio
path: data/gradio-*
- split: huggingface_hub
path: data/huggingface_hub-*
- split: optimum
path: data/optimum-*
- split: peft
path: data/peft-*
- split: pytorch_image_models
path: data/pytorch_image_models-*
- split: safetensors
path: data/safetensors-*
- split: tokenizers
path: data/tokenizers-*
- split: transformers
path: data/transformers-*
---
# Dataset Card for "pip"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
tomekkorbak/detoxify-pile-chunk3-1750000-1800000 | 2022-10-04T23:02:19.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 130 | Entry not found |
tomekkorbak/detoxify-pile-chunk3-1700000-1750000 | 2022-10-04T23:58:51.000Z | [
"region:us"
] | tomekkorbak | null | null | null | 0 | 130 | Entry not found |
Norod78/simpsons-blip-captions | 2022-11-09T16:27:19.000Z | [
"task_categories:text-to-image",
"annotations_creators:machine-generated",
"language_creators:other",
"multilinguality:monolingual",
"size_categories:n<1K",
"language:en",
"license:cc-by-nc-sa-4.0",
"region:us"
] | Norod78 | null | null | null | 2 | 130 | ---
dataset_info:
features:
- name: image
dtype: image
- name: text
dtype: string
splits:
- name: train
num_bytes: 51605730.0
num_examples: 755
download_size: 50553165
dataset_size: 51605730.0
pretty_name: 'Simpsons BLIP captions'
size_categories:
- n<1K
tags: []
task_categories:
- text-to-image
license: cc-by-nc-sa-4.0
annotations_creators:
- machine-generated
language:
- en
language_creators:
- other
multilinguality:
- monolingual
---
# Dataset Card for "simpsons-blip-captions"
|
mstz/magic | 2023-04-16T17:34:16.000Z | [
"task_categories:tabular-classification",
"size_categories:10K<n<100K",
"language:en",
"license:cc",
"magic",
"tabular_classification",
"binary_classification",
"UCI",
"region:us"
] | mstz | null | @misc{misc_magic_gamma_telescope_159,
author = {Bock,R.},
title = {{MAGIC Gamma Telescope}},
year = {2007},
howpublished = {UCI Machine Learning Repository},
note = {{DOI}: \\url{10.24432/C52C8B}}
} | null | 0 | 130 | ---
language:
- en
tags:
- magic
- tabular_classification
- binary_classification
- UCI
pretty_name: Magic
size_categories:
- 10K<n<100K
task_categories:
- tabular-classification
configs:
- magic
license: cc
---
# Magic
The [Magic dataset](https://archive.ics.uci.edu/ml/datasets/Magic) from the [UCI ML repository](https://archive.ics.uci.edu/ml/datasets).
# Configurations and tasks
| **Configuration** | **Task** | **Description** |
|-------------------|---------------------------|---------------------------------------------------------------|
| magic | Binary classification | Classify the person's magic as over or under the threshold. |
# Usage
```python
from datasets import load_dataset
dataset = load_dataset("mstz/magic")["train"]
``` |
conceptofmind/dialog_submix_original | 2023-04-28T22:41:37.000Z | [
"region:us"
] | conceptofmind | null | null | null | 12 | 130 | ---
dataset_info:
features:
- name: inputs
dtype: string
- name: targets
dtype: string
- name: task_source
dtype: string
- name: task_name
dtype: string
- name: template_type
dtype: string
splits:
- name: train
num_bytes: 1024507265
num_examples: 553869
download_size: 583477456
dataset_size: 1024507265
---
# Dataset Card for "dialog_submix_original"
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
fujiki/japanese_alpaca_data | 2023-05-19T12:54:13.000Z | [
"language:ja",
"license:cc-by-nc-sa-4.0",
"region:us"
] | fujiki | null | null | null | 4 | 130 | ---
dataset_info:
features:
- name: instruction
dtype: string
- name: input
dtype: string
- name: output
dtype: string
splits:
- name: train
num_bytes: 24733874
num_examples: 52002
download_size: 13849623
dataset_size: 24733874
license: cc-by-nc-sa-4.0
language:
- ja
pretty_name: japanese_alpaca
---
# Dataset Card for "japanese_alpaca_data"
- This dataset is based on `masa3141`'s great work on `japanese-alpaca-lora` [[github]](https://github.com/masa3141/japanese-alpaca-lora). Please also refer to this repo.
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) |
chansung/llama2-stories | 2023-09-30T21:15:48.000Z | [
"license:apache-2.0",
"region:us"
] | chansung | null | null | null | 2 | 130 | ---
license: apache-2.0
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
dataset_info:
features:
- name: title
dtype: string
- name: image
dtype: string
- name: story
dtype: string
splits:
- name: train
num_bytes: 4113688
num_examples: 69
download_size: 3306944
dataset_size: 4113688
---
|
AdaptLLM/law-tasks | 2023-09-26T08:35:23.000Z | [
"arxiv:2309.09530",
"region:us"
] | AdaptLLM | null | null | null | 2 | 130 | ---
configs:
- config_name: SCOTUS
data_files:
- split: test
path: "scotus/test.json"
- config_name: CaseHOLD
data_files:
- split: test
path: "case_hold/test.json"
- config_name: UNFAIR_ToS
data_files:
- split: test
path: "unfair_tos/test.json"
---
# Adapting Large Language Models via Reading Comprehension
This repo contains the evaluation datasets for our paper [Adapting Large Language Models via Reading Comprehension](https://arxiv.org/pdf/2309.09530.pdf)
We explore **continued pre-training on domain-specific corpora** for large language models. While this approach enriches LLMs with domain knowledge, it significantly hurts their prompting ability for question answering. Inspired by human learning via reading comprehension, we propose a simple method to **transform large-scale pre-training corpora into reading comprehension texts**, consistently improving prompting performance across tasks in **biomedicine, finance, and law domains**. Our 7B model competes with much larger domain-specific models like BloombergGPT-50B. Moreover, our domain-specific reading comprehension texts enhance model performance even on general benchmarks, indicating potential for developing a general LLM across more domains.
## GitHub repo:
https://github.com/microsoft/LMOps
## Domain-specific LLMs:
Our models of different domains are now available in Huggingface: [Biomedicine-LLM](https://huggingface.co/AdaptLLM/medicine-LLM), [Finance-LLM](https://huggingface.co/AdaptLLM/finance-LLM) and [Law-LLM](https://huggingface.co/AdaptLLM/law-LLM), the performances of our AdaptLLM compared to other domain-specific LLMs are:
<p align='center'>
<img src="./comparison.png" width="700">
</p>
## Domain-specific Tasks:
To easily reproduce our results, we have uploaded the filled-in zero/few-shot input instructions and output completions of each domain-specific task: [biomedicine-tasks](https://huggingface.co/datasets/AdaptLLM/medicine-tasks), [finance-tasks](https://huggingface.co/datasets/AdaptLLM/finance-tasks), and [law-tasks](https://huggingface.co/datasets/AdaptLLM/law-tasks).
## Citation:
```bibtex
@inproceedings{AdaptLLM,
title={Adapting Large Language Models via Reading Comprehension},
author={Daixuan Cheng and Shaohan Huang and Furu Wei},
url={https://arxiv.org/abs/2309.09530},
year={2023},
}
```
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.