|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
"""RedPajama: An Open-Source, Clean-Room 1.2 Trillion Token Dataset.""" |
|
|
|
|
|
import json |
|
|
|
import datasets |
|
from datasets.exceptions import DefunctDatasetError |
|
import traceback |
|
import os |
|
|
|
logger = datasets.logging.get_logger(__name__) |
|
|
|
|
|
_DESCRIPTION = """\ |
|
RedPajama is a clean-room, fully open-source implementation of the LLaMa dataset. |
|
""" |
|
|
|
_URL_LISTS = { |
|
"arxiv": "urls/arxiv.txt", |
|
"c4": "urls/c4.txt", |
|
"common_crawl": "urls/common_crawl.txt", |
|
"github": "urls/github.txt", |
|
"stackexchange": "urls/stackexchange.txt", |
|
"wikipedia": "urls/wikipedia.txt", |
|
} |
|
_URL_BASE = 'https://data.together.xyz/redpajama-data-1T/v1.0.0' |
|
|
|
_DATA_DIR = os.environ.get('RED_PAJAMA_DATA_DIR', None) |
|
|
|
class RedPajama1TConfig(datasets.BuilderConfig): |
|
"""BuilderConfig for RedPajama sample.""" |
|
|
|
def __init__(self, *args, subsets, **kwargs): |
|
"""BuilderConfig for RedPajama. |
|
Args: |
|
**kwargs: keyword arguments forwarded to super. |
|
""" |
|
super(RedPajama1TConfig, self).__init__(**kwargs) |
|
self.subsets = subsets |
|
|
|
|
|
class RedPajama1T(datasets.GeneratorBasedBuilder): |
|
"""RedPajama: Reproducing the LLaMA training dataset of over 1.2 trillion tokens. Version 1.0.0.""" |
|
|
|
BUILDER_CONFIGS = [ |
|
RedPajama1TConfig( |
|
name = 'default', |
|
subsets = list(_URL_LISTS) + ["book"], |
|
version=datasets.Version("1.0.0", ""), |
|
description="RedPajama1T", |
|
), |
|
|
|
RedPajama1TConfig( |
|
name = 'arxiv', |
|
subsets = ['arxiv'], |
|
version=datasets.Version("1.0.0", ""), |
|
description="RedPajama1T arxiv subset", |
|
), |
|
|
|
RedPajama1TConfig( |
|
name = 'book', |
|
subsets = ['book'], |
|
version=datasets.Version("1.0.0", ""), |
|
description="RedPajama1T book subset", |
|
), |
|
|
|
RedPajama1TConfig( |
|
name = 'c4', |
|
subsets = ['c4'], |
|
version=datasets.Version("1.0.0", ""), |
|
description="RedPajama1T c4 subset", |
|
), |
|
|
|
RedPajama1TConfig( |
|
name = 'common_crawl', |
|
subsets = ['common_crawl'], |
|
version=datasets.Version("1.0.0", ""), |
|
description="RedPajama1T common crawl subset", |
|
), |
|
|
|
RedPajama1TConfig( |
|
name = 'github', |
|
subsets = ['github'], |
|
version=datasets.Version("1.0.0", ""), |
|
description="RedPajama1T github subset", |
|
), |
|
|
|
RedPajama1TConfig( |
|
name = 'stackexchange', |
|
subsets = ['stackexchange'], |
|
version=datasets.Version("1.0.0", ""), |
|
description="RedPajama1T stackexchange subset", |
|
), |
|
|
|
RedPajama1TConfig( |
|
name = 'wikipedia', |
|
subsets = ['wikipedia'], |
|
version=datasets.Version("1.0.0", ""), |
|
description="RedPajama1T wikipedia subset", |
|
), |
|
] |
|
|
|
def _info(self): |
|
return datasets.DatasetInfo( |
|
description=_DESCRIPTION, |
|
features=datasets.Features( |
|
{ |
|
"text": datasets.Value("string"), |
|
"meta": datasets.Value("string"), |
|
"red_pajama_subset": datasets.Value("string"), |
|
} |
|
), |
|
supervised_keys=None, |
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
if self.config.name == "book": |
|
raise DefunctDatasetError( |
|
"The 'book' config is defunct and no longer accessible due to reported copyright infringement. " |
|
) |
|
subsets = [subset for subset in self.config.subsets if subset != "book"] |
|
|
|
url_lists = dl_manager.download_and_extract({ |
|
subset: _URL_LISTS[subset] for subset in subsets |
|
}) |
|
urls = {} |
|
|
|
for subset, url_list in url_lists.items(): |
|
with open(url_list, encoding="utf-8") as f: |
|
urls[subset] = [line.strip() for line in f] |
|
|
|
if _DATA_DIR is not None: |
|
print(f'Reading data from {_DATA_DIR}') |
|
url_prefix_slashes = len(_URL_BASE.split('/')) |
|
downloaded_files = { |
|
subset: [ |
|
os.path.join(_DATA_DIR, *url.split('/')[url_prefix_slashes:]) |
|
for url in url_list |
|
] |
|
for subset, url_list in urls.items() |
|
} |
|
else: |
|
downloaded_files = dl_manager.download(urls) |
|
|
|
return [ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TRAIN, |
|
gen_kwargs = { |
|
"files": { |
|
subset: downloaded_files[subset] |
|
for subset in subsets |
|
} |
|
} |
|
) |
|
] |
|
|
|
def _generate_examples(self, files): |
|
"""This function returns the examples in the raw (text) form.""" |
|
key = 0 |
|
for subset in files: |
|
if subset == "common_crawl": |
|
import zstandard as zstd |
|
|
|
for path in files[subset]: |
|
with zstd.open(open(path, "rb"), "rt", encoding="utf-8") as f: |
|
for i, row in enumerate(f): |
|
try: |
|
data = json.loads(row) |
|
text = data["text"] |
|
del data["text"] |
|
yield key, { |
|
"text": text, |
|
"meta": json.dumps(data), |
|
"red_pajama_subset": subset, |
|
} |
|
key += 1 |
|
except Exception as e: |
|
print(f'Subset: {subset}') |
|
print(f'Path: {path}') |
|
print(f'Row: {row}') |
|
traceback.print_exc() |
|
|
|
raise e |
|
else: |
|
for path in files[subset]: |
|
with open(path, encoding="utf-8") as f: |
|
for i, row in enumerate(f): |
|
try: |
|
data = json.loads(row) |
|
if "meta" not in data: |
|
text = data["text"] |
|
del data["text"] |
|
yield key, { |
|
"text": text, |
|
"meta": json.dumps(data), |
|
"red_pajama_subset": subset, |
|
} |
|
else: |
|
yield key, { |
|
"text": data["text"], |
|
"meta": data["meta"], |
|
"red_pajama_subset": subset, |
|
} |
|
key += 1 |
|
except Exception as e: |
|
print(f'Subset: {subset}') |
|
print(f'Path: {path}') |
|
print(f'Row: {row}') |
|
traceback.print_exc() |
|
|
|
raise e |
|
|