text stringclasses 3
values |
|---|
Walking down the street, everybody looks at me
Shotgun in my pants is what they want to see
My bum is like a trigger, baby, waiting to explode
Come on, lets get funky, and make the party flow
Shotgun in the bum
Bang, bang, baby
Shotgun in the bum
Bang, bang, baby
Shotgun in the bum
Bang, bang, baby
Shotgun in the bum
B... |
In a faraway land, there lives a zombie shrimp
And he eats baked potatoes every single day
Hidden in a secret cave with all his zombie friends
How I wish I could live in his happy shrimpy way
In a faraway land, there lives a zombie shrimp
And he eats baked potatoes every single day
Hidden in a secret cave with all his ... |
Hey sexy baby
Its party time
Once upon a time there was big ploppy poo
He liked eating sausages and licking a dogs arse
Then one day policeman said he cannot do his thing
So he went on a machine gun rampage and had a lot of fun
Benji Bird was his name
What a silly poo
Killing was the favorite thing
For biddy bird to do... |
Dataset Card for "huggingartists/asdfgfa"
Dataset Summary
The Lyrics dataset parsed from Genius. This dataset is designed to generate lyrics with HuggingArtists. Model is available here.
Supported Tasks and Leaderboards
Languages
en
How to use
How to load this dataset directly with the datasets library:
from datasets import load_dataset
dataset = load_dataset("huggingartists/asdfgfa")
Dataset Structure
An example of 'train' looks as follows.
This example was too long and was cropped:
{
"text": "Look, I was gonna go easy on you\nNot to hurt your feelings\nBut I'm only going to get this one chance\nSomething's wrong, I can feel it..."
}
Data Fields
The data fields are the same among all splits.
text: astringfeature.
Data Splits
| train | validation | test |
|---|---|---|
| TRAIN_0.031015 | - | - |
'Train' can be easily divided into 'train' & 'validation' & 'test' with few lines of code:
from datasets import load_dataset, Dataset, DatasetDict
import numpy as np
datasets = load_dataset("huggingartists/asdfgfa")
train_percentage = 0.9
validation_percentage = 0.07
test_percentage = 0.03
train, validation, test = np.split(datasets['train']['text'], [int(len(datasets['train']['text'])*train_percentage), int(len(datasets['train']['text'])*(train_percentage + validation_percentage))])
datasets = DatasetDict(
{
'train': Dataset.from_dict({'text': list(train)}),
'validation': Dataset.from_dict({'text': list(validation)}),
'test': Dataset.from_dict({'text': list(test)})
}
)
Dataset Creation
Curation Rationale
Source Data
Initial Data Collection and Normalization
Who are the source language producers?
Annotations
Annotation process
Who are the annotators?
Personal and Sensitive Information
Considerations for Using the Data
Social Impact of Dataset
Discussion of Biases
Other Known Limitations
Additional Information
Dataset Curators
Licensing Information
Citation Information
@InProceedings{huggingartists,
author={Aleksey Korshuk}
year=2021
}
About
Built by Aleksey Korshuk
For more details, visit the project repository.
- Downloads last month
- 18