huggingartists/fascinoma
Text Generation
•
Updated
•
38
text
stringclasses 1
value |
---|
Im walking this road because you stole my car
Im singing this song cause you have all my CDs
You want me to believe in your love
And yeah, I will when you give back my stuff
Give back my stuff
The gerbils are dead your books have found a home
Under the tree where I have buried your clothes
You want me to believe in your love
And yeah, I will when you dig up your stuff
Dig up your stuff
How am I supposed to learn what love is in this lifetime?
How am I supposed to learn what love is, what love is if you go
If you go
If you go
If you go
Dont go, dont go, dont go, dont go, dont go, dont go
Dont go, dont go, dont go, dont go, dont go, dont go
Dont go, dont go, dont go, dont go
Im walking this road because you stole my heart
Im singing this song cause you have all my CDs
You want me to believe in your love
And yeah, I do cause I cant love you enough
Cant love you enough
Give back my stuff
You have my love
Dont go, dont go, dont go, dont go, dont go, dont go
Dont go, dont go, dont go, dont go, dont go, dont go
Dont go, dont go, dont go, dont go |
The Lyrics dataset parsed from Genius. This dataset is designed to generate lyrics with HuggingArtists. Model is available here.
en
How to load this dataset directly with the datasets library:
from datasets import load_dataset
dataset = load_dataset("huggingartists/fascinoma")
An example of 'train' looks as follows.
This example was too long and was cropped:
{
"text": "Look, I was gonna go easy on you\nNot to hurt your feelings\nBut I'm only going to get this one chance\nSomething's wrong, I can feel it..."
}
The data fields are the same among all splits.
text
: a string
feature.train | validation | test |
---|---|---|
1 | - | - |
'Train' can be easily divided into 'train' & 'validation' & 'test' with few lines of code:
from datasets import load_dataset, Dataset, DatasetDict
import numpy as np
datasets = load_dataset("huggingartists/fascinoma")
train_percentage = 0.9
validation_percentage = 0.07
test_percentage = 0.03
train, validation, test = np.split(datasets['train']['text'], [int(len(datasets['train']['text'])*train_percentage), int(len(datasets['train']['text'])*(train_percentage + validation_percentage))])
datasets = DatasetDict(
{
'train': Dataset.from_dict({'text': list(train)}),
'validation': Dataset.from_dict({'text': list(validation)}),
'test': Dataset.from_dict({'text': list(test)})
}
)
@InProceedings{huggingartists,
author={Aleksey Korshuk}
year=2021
}
Built by Aleksey Korshuk
For more details, visit the project repository.