Don't use datasets.data_files.* functions (#3)
Browse files- Don't use datasets.data_files.* functions (8d9b354246cbdcbb1a57fa62492ffa708a41ab6f)
- small fix (9aa37d86e3db270db66da3592e97564037d6fa1a)
Co-authored-by: Quentin Lhoest <[email protected]>
- github-code.py +6 -15
github-code.py
CHANGED
@@ -20,8 +20,6 @@ import pyarrow as pa
|
|
20 |
import pyarrow.parquet as pq
|
21 |
|
22 |
import datasets
|
23 |
-
from huggingface_hub import HfApi, HfFolder
|
24 |
-
from datasets.data_files import DataFilesDict
|
25 |
|
26 |
_REPO_NAME = "codeparrot/github-code"
|
27 |
|
@@ -160,19 +158,12 @@ class GithubCode(datasets.GeneratorBasedBuilder):
|
|
160 |
)
|
161 |
|
162 |
def _split_generators(self, dl_manager):
|
163 |
-
|
164 |
-
|
165 |
-
|
166 |
-
|
167 |
-
|
168 |
-
|
169 |
-
patterns = datasets.data_files.get_patterns_in_dataset_repository(hfh_dataset_info)
|
170 |
-
data_files = datasets.data_files.DataFilesDict.from_hf_repo(
|
171 |
-
patterns,
|
172 |
-
dataset_info=hfh_dataset_info,
|
173 |
-
)
|
174 |
-
|
175 |
-
files = dl_manager.download_and_extract(data_files["train"])
|
176 |
return [
|
177 |
datasets.SplitGenerator(
|
178 |
name=datasets.Split.TRAIN,
|
|
|
20 |
import pyarrow.parquet as pq
|
21 |
|
22 |
import datasets
|
|
|
|
|
23 |
|
24 |
_REPO_NAME = "codeparrot/github-code"
|
25 |
|
|
|
158 |
)
|
159 |
|
160 |
def _split_generators(self, dl_manager):
|
161 |
+
num_shards = 1126
|
162 |
+
data_files = [
|
163 |
+
f"data/train-{_index:05d}-of-{num_shards:05d}.parquet"
|
164 |
+
for _index in range(num_shards)
|
165 |
+
]
|
166 |
+
files = dl_manager.download(data_files)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
167 |
return [
|
168 |
datasets.SplitGenerator(
|
169 |
name=datasets.Split.TRAIN,
|