Datasets:
Convert dataset to Parquet
#3
by
albertvillanova
HF staff
- opened
- README.md +26 -20
- lj_speech.py +0 -116
- main/train-00000-of-00008.parquet +3 -0
- main/train-00001-of-00008.parquet +3 -0
- main/train-00002-of-00008.parquet +3 -0
- main/train-00003-of-00008.parquet +3 -0
- main/train-00004-of-00008.parquet +3 -0
- main/train-00005-of-00008.parquet +3 -0
- main/train-00006-of-00008.parquet +3 -0
- main/train-00007-of-00008.parquet +3 -0
README.md
CHANGED
@@ -9,8 +9,6 @@ license:
|
|
9 |
- unlicense
|
10 |
multilinguality:
|
11 |
- monolingual
|
12 |
-
paperswithcode_id: ljspeech
|
13 |
-
pretty_name: LJ Speech
|
14 |
size_categories:
|
15 |
- 10K<n<100K
|
16 |
source_datasets:
|
@@ -20,21 +18,10 @@ task_categories:
|
|
20 |
- text-to-speech
|
21 |
- text-to-audio
|
22 |
task_ids: []
|
23 |
-
|
24 |
-
|
25 |
-
task: automatic-speech-recognition
|
26 |
-
task_id: speech_recognition
|
27 |
-
splits:
|
28 |
-
train_split: train
|
29 |
-
col_mapping:
|
30 |
-
file: path
|
31 |
-
text: text
|
32 |
-
metrics:
|
33 |
-
- type: wer
|
34 |
-
name: WER
|
35 |
-
- type: cer
|
36 |
-
name: CER
|
37 |
dataset_info:
|
|
|
38 |
features:
|
39 |
- name: id
|
40 |
dtype: string
|
@@ -48,13 +35,32 @@ dataset_info:
|
|
48 |
dtype: string
|
49 |
- name: normalized_text
|
50 |
dtype: string
|
51 |
-
config_name: main
|
52 |
splits:
|
53 |
- name: train
|
54 |
-
num_bytes:
|
55 |
num_examples: 13100
|
56 |
-
download_size:
|
57 |
-
dataset_size:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
58 |
---
|
59 |
|
60 |
# Dataset Card for lj_speech
|
|
|
9 |
- unlicense
|
10 |
multilinguality:
|
11 |
- monolingual
|
|
|
|
|
12 |
size_categories:
|
13 |
- 10K<n<100K
|
14 |
source_datasets:
|
|
|
18 |
- text-to-speech
|
19 |
- text-to-audio
|
20 |
task_ids: []
|
21 |
+
paperswithcode_id: ljspeech
|
22 |
+
pretty_name: LJ Speech
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
23 |
dataset_info:
|
24 |
+
config_name: main
|
25 |
features:
|
26 |
- name: id
|
27 |
dtype: string
|
|
|
35 |
dtype: string
|
36 |
- name: normalized_text
|
37 |
dtype: string
|
|
|
38 |
splits:
|
39 |
- name: train
|
40 |
+
num_bytes: 3860187268.0
|
41 |
num_examples: 13100
|
42 |
+
download_size: 3786217548
|
43 |
+
dataset_size: 3860187268.0
|
44 |
+
configs:
|
45 |
+
- config_name: main
|
46 |
+
data_files:
|
47 |
+
- split: train
|
48 |
+
path: main/train-*
|
49 |
+
default: true
|
50 |
+
train-eval-index:
|
51 |
+
- config: main
|
52 |
+
task: automatic-speech-recognition
|
53 |
+
task_id: speech_recognition
|
54 |
+
splits:
|
55 |
+
train_split: train
|
56 |
+
col_mapping:
|
57 |
+
file: path
|
58 |
+
text: text
|
59 |
+
metrics:
|
60 |
+
- type: wer
|
61 |
+
name: WER
|
62 |
+
- type: cer
|
63 |
+
name: CER
|
64 |
---
|
65 |
|
66 |
# Dataset Card for lj_speech
|
lj_speech.py
DELETED
@@ -1,116 +0,0 @@
|
|
1 |
-
# coding=utf-8
|
2 |
-
# Copyright 2021 The TensorFlow Datasets Authors and the HuggingFace Datasets Authors.
|
3 |
-
#
|
4 |
-
# Licensed under the Apache License, Version 2.0 (the "License");
|
5 |
-
# you may not use this file except in compliance with the License.
|
6 |
-
# You may obtain a copy of the License at
|
7 |
-
#
|
8 |
-
# http://www.apache.org/licenses/LICENSE-2.0
|
9 |
-
#
|
10 |
-
# Unless required by applicable law or agreed to in writing, software
|
11 |
-
# distributed under the License is distributed on an "AS IS" BASIS,
|
12 |
-
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
13 |
-
# See the License for the specific language governing permissions and
|
14 |
-
# limitations under the License.
|
15 |
-
|
16 |
-
# Lint as: python3
|
17 |
-
"""LJ automatic speech recognition dataset."""
|
18 |
-
|
19 |
-
|
20 |
-
import csv
|
21 |
-
import os
|
22 |
-
|
23 |
-
import datasets
|
24 |
-
from datasets.tasks import AutomaticSpeechRecognition
|
25 |
-
|
26 |
-
|
27 |
-
_CITATION = """\
|
28 |
-
@misc{ljspeech17,
|
29 |
-
author = {Keith Ito and Linda Johnson},
|
30 |
-
title = {The LJ Speech Dataset},
|
31 |
-
howpublished = {\\url{https://keithito.com/LJ-Speech-Dataset/}},
|
32 |
-
year = 2017
|
33 |
-
}
|
34 |
-
"""
|
35 |
-
|
36 |
-
_DESCRIPTION = """\
|
37 |
-
This is a public domain speech dataset consisting of 13,100 short audio clips of a single speaker reading
|
38 |
-
passages from 7 non-fiction books in English. A transcription is provided for each clip. Clips vary in length
|
39 |
-
from 1 to 10 seconds and have a total length of approximately 24 hours.
|
40 |
-
|
41 |
-
Note that in order to limit the required storage for preparing this dataset, the audio
|
42 |
-
is stored in the .wav format and is not converted to a float32 array. To convert the audio
|
43 |
-
file to a float32 array, please make use of the `.map()` function as follows:
|
44 |
-
|
45 |
-
|
46 |
-
```python
|
47 |
-
import soundfile as sf
|
48 |
-
|
49 |
-
def map_to_array(batch):
|
50 |
-
speech_array, _ = sf.read(batch["file"])
|
51 |
-
batch["speech"] = speech_array
|
52 |
-
return batch
|
53 |
-
|
54 |
-
dataset = dataset.map(map_to_array, remove_columns=["file"])
|
55 |
-
```
|
56 |
-
"""
|
57 |
-
|
58 |
-
_URL = "https://keithito.com/LJ-Speech-Dataset/"
|
59 |
-
_DL_URL = "https://data.keithito.com/data/speech/LJSpeech-1.1.tar.bz2"
|
60 |
-
|
61 |
-
|
62 |
-
class LJSpeech(datasets.GeneratorBasedBuilder):
|
63 |
-
"""LJ Speech dataset."""
|
64 |
-
|
65 |
-
VERSION = datasets.Version("1.1.0")
|
66 |
-
|
67 |
-
BUILDER_CONFIGS = [
|
68 |
-
datasets.BuilderConfig(name="main", version=VERSION, description="The full LJ Speech dataset"),
|
69 |
-
]
|
70 |
-
|
71 |
-
def _info(self):
|
72 |
-
return datasets.DatasetInfo(
|
73 |
-
description=_DESCRIPTION,
|
74 |
-
features=datasets.Features(
|
75 |
-
{
|
76 |
-
"id": datasets.Value("string"),
|
77 |
-
"audio": datasets.Audio(sampling_rate=22050),
|
78 |
-
"file": datasets.Value("string"),
|
79 |
-
"text": datasets.Value("string"),
|
80 |
-
"normalized_text": datasets.Value("string"),
|
81 |
-
}
|
82 |
-
),
|
83 |
-
supervised_keys=("file", "text"),
|
84 |
-
homepage=_URL,
|
85 |
-
citation=_CITATION,
|
86 |
-
task_templates=[AutomaticSpeechRecognition(audio_column="audio", transcription_column="text")],
|
87 |
-
)
|
88 |
-
|
89 |
-
def _split_generators(self, dl_manager):
|
90 |
-
root_path = dl_manager.download_and_extract(_DL_URL)
|
91 |
-
root_path = os.path.join(root_path, "LJSpeech-1.1")
|
92 |
-
wav_path = os.path.join(root_path, "wavs")
|
93 |
-
csv_path = os.path.join(root_path, "metadata.csv")
|
94 |
-
|
95 |
-
return [
|
96 |
-
datasets.SplitGenerator(
|
97 |
-
name=datasets.Split.TRAIN, gen_kwargs={"wav_path": wav_path, "csv_path": csv_path}
|
98 |
-
),
|
99 |
-
]
|
100 |
-
|
101 |
-
def _generate_examples(self, wav_path, csv_path):
|
102 |
-
"""Generate examples from an LJ Speech archive_path."""
|
103 |
-
|
104 |
-
with open(csv_path, encoding="utf-8") as csv_file:
|
105 |
-
csv_reader = csv.reader(csv_file, delimiter="|", quotechar=None, skipinitialspace=True)
|
106 |
-
for row in csv_reader:
|
107 |
-
uid, text, norm_text = row
|
108 |
-
filename = f"{uid}.wav"
|
109 |
-
example = {
|
110 |
-
"id": uid,
|
111 |
-
"file": os.path.join(wav_path, filename),
|
112 |
-
"audio": os.path.join(wav_path, filename),
|
113 |
-
"text": text,
|
114 |
-
"normalized_text": norm_text,
|
115 |
-
}
|
116 |
-
yield uid, example
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
main/train-00000-of-00008.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:18c8a13f9a8975a1c22546d931ee30b7852f091f2aed75e766ee3289b396d6d8
|
3 |
+
size 479478819
|
main/train-00001-of-00008.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:715f47e2a99950bc1abb74453b5c03411586fe1b652b06d2b0ed0f651a09f0aa
|
3 |
+
size 476725772
|
main/train-00002-of-00008.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e8db7cc332abc8068d1df51f0b61e48308b2f14397cbc15ffdd7bc21e46c456a
|
3 |
+
size 474225836
|
main/train-00003-of-00008.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:87a586ebc0f9509a54ada3cc0fc7377ec7b0641218f4b91dd2b2dc5ca9d47283
|
3 |
+
size 467178572
|
main/train-00004-of-00008.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5c92c6986e34b627b13a1f0d38e096ef5fcc0296bcc1e4ba40f4453b313e0307
|
3 |
+
size 464994923
|
main/train-00005-of-00008.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8b5031b3e260cd461f8997d16674cde131668511988769da3501ece26d9f063b
|
3 |
+
size 458327927
|
main/train-00006-of-00008.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7e48a0ae5d4015b9b2d823201b80bfc43ddbb4c9c970275152f8be3c07208ca9
|
3 |
+
size 486514835
|
main/train-00007-of-00008.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e0174c75c39e24b9d6699e1030671472fd7c00c1657de088a0bb4190b2664e72
|
3 |
+
size 478770864
|