|
import shutil |
|
|
|
import pandas as pd |
|
import os |
|
|
|
import huggingface_hub |
|
import pytest |
|
from datasets import load_dataset |
|
|
|
|
|
@pytest.mark.parametrize( |
|
"dataset_name, link_to_source", |
|
[ |
|
( |
|
"h2ogpt-oig-instruct-cleaned", |
|
""" |
|
- [Original LAION OIG Dataset](https://github.com/LAION-AI/Open-Instruction-Generalist) |
|
- [LAION OIG data detoxed and filtered down by scripts in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/b8f15efcc305a953c52a0ee25b8b4897ceb68c0a/scrape_dai_docs.py) |
|
""" |
|
), |
|
( |
|
"h2ogpt-oig-instruct-cleaned-v2", |
|
""" |
|
- [Original LAION OIG Dataset](https://github.com/LAION-AI/Open-Instruction-Generalist) |
|
- [LAION OIG data detoxed and filtered down by scripts in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/40c217f610766715acec297a5535eb440ac2f2e2/create_data.py) |
|
""" |
|
), |
|
( |
|
"h2ogpt-oig-instruct-cleaned-v3", |
|
""" |
|
- [Original LAION OIG Dataset](https://github.com/LAION-AI/Open-Instruction-Generalist) |
|
- [LAION OIG data detoxed and filtered down by scripts in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/bfc3778c8db938761ce2093351bf2bf82159291e/create_data.py) |
|
""" |
|
), |
|
( |
|
"openassistant_oasst1", |
|
""" |
|
- [Original Open Assistant data in tree structure](https://huggingface.co/datasets/OpenAssistant/oasst1) |
|
- [This flattened dataset created by script in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/45e6183171fb16691ad7d3ab006fad973f971e98/create_data.py#L1253) |
|
""" |
|
), |
|
( |
|
"h2ogpt-oig-oasst1-instruct-cleaned-v1", |
|
""" |
|
- [Original LAION OIG Dataset](https://github.com/LAION-AI/Open-Instruction-Generalist) |
|
- [LAION OIG data detoxed and filtered down by scripts in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/main/docs/FINETUNE.md#high-quality-oig-based-instruct-data) |
|
|
|
- [Original Open Assistant data in tree structure](https://huggingface.co/datasets/OpenAssistant/oasst1) |
|
- [This flattened dataset created by script in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/5fc91911bc2bfaaf3b6c2de577c4b0ae45a07a4a/create_data.py#L1253) |
|
""" |
|
), |
|
( |
|
"h2ogpt-oig-oasst1-instruct-cleaned-v2", |
|
""" |
|
- [Original LAION OIG Dataset](https://github.com/LAION-AI/Open-Instruction-Generalist) |
|
- [LAION OIG data detoxed and filtered down by scripts in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/main/docs/FINETUNE.md#high-quality-oig-based-instruct-data) |
|
|
|
- [Original Open Assistant data in tree structure](https://huggingface.co/datasets/OpenAssistant/oasst1) |
|
- [This flattened dataset created by script in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/0e70c2fbb16410bd8e6992d879b4c55cd981211f/create_data.py#L1375-L1415) |
|
""" |
|
), |
|
( |
|
"h2ogpt-oig-oasst1-instruct-cleaned-v3", |
|
""" |
|
- [Original LAION OIG Dataset](https://github.com/LAION-AI/Open-Instruction-Generalist) |
|
- [LAION OIG data detoxed and filtered down by scripts in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/main/docs/FINETUNE.md#high-quality-oig-based-instruct-data) |
|
|
|
- [Original Open Assistant data in tree structure](https://huggingface.co/datasets/OpenAssistant/oasst1) |
|
- [This flattened dataset created by script in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/6728938a262d3eb5e8db1f252bbcd7de838da452/create_data.py#L1415) |
|
""" |
|
), |
|
( |
|
"openassistant_oasst1_h2ogpt", |
|
""" |
|
- [Original Open Assistant data in tree structure](https://huggingface.co/datasets/OpenAssistant/oasst1) |
|
- [This flattened dataset created by script in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/83857fcf7d3b712aad5db32207e6db0ab0f780f9/create_data.py#L1252) |
|
""" |
|
), |
|
( |
|
"openassistant_oasst1_h2ogpt_graded", |
|
""" |
|
- [Original Open Assistant data in tree structure](https://huggingface.co/datasets/OpenAssistant/oasst1) |
|
- [This flattened dataset created by script in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/d1f8ce975a46056d41135d126dd33de8499aa26e/create_data.py#L1259) |
|
""" |
|
), |
|
( |
|
"h2ogpt-fortune2000-personalized", |
|
""" |
|
- [Fortune 2000 companies from Wikipedia](https://github.com/h2oai/h2ogpt/blob/b1ea74c0088884ebff97f1ccddbfb3f393e29e44/create_data.py#L1743) |
|
""" |
|
), |
|
( |
|
"openassistant_oasst1_h2ogpt_llama2_chat", |
|
""" |
|
- [Original Open Assistant data in tree structure](https://huggingface.co/datasets/OpenAssistant/oasst1) |
|
- [This flattened dataset created by script in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/0bee5f50a74f489ca3fc81486f9322078360f2cb/src/create_data.py#L1296) |
|
""" |
|
), |
|
], |
|
) |
|
def test_create_data_cards(dataset_name, link_to_source): |
|
if dataset_name != "openassistant_oasst1_h2ogpt_llama2_chat": |
|
return |
|
|
|
assert os.path.exists("README-template.md"), "must be running this test from the data dir." |
|
shutil.rmtree(dataset_name, ignore_errors=True) |
|
try: |
|
repo = huggingface_hub.Repository( |
|
local_dir=dataset_name, |
|
clone_from="h2oai/%s" % dataset_name, |
|
repo_type="dataset", |
|
skip_lfs_files=True, |
|
token=True, |
|
) |
|
repo.git_pull() |
|
except Exception as e: |
|
print(str(e)) |
|
print("call 'huggingface_cli login' first and provide access token with write permission") |
|
dataset = load_dataset("h2oai/%s" % dataset_name)["train"] |
|
|
|
pd.set_option('display.max_columns', None) |
|
with open("README-template.md", "r") as f: |
|
content = f.read() |
|
assert "<<DATASET_NAME>>" in content |
|
content = content.replace("<<DATASET_NAME>>", dataset_name) |
|
|
|
assert "<<NROWS>>" in content |
|
content = content.replace("<<NROWS>>", str(dataset.num_rows)) |
|
|
|
assert "<<NCOLS>>" in content |
|
content = content.replace("<<NCOLS>>", str(dataset.num_columns)) |
|
|
|
assert "<<COLNAMES>>" in content |
|
content = content.replace("<<COLNAMES>>", str(dataset.column_names)) |
|
|
|
|
|
|
|
|
|
assert "<<SOURCE_LINK>>" in content |
|
content = content.replace("<<SOURCE_LINK>>", link_to_source) |
|
|
|
assert "<<" not in content |
|
assert ">>" not in content |
|
|
|
with open(os.path.join(dataset_name, "README.md"), "w") as f: |
|
f.write(content) |
|
try: |
|
repo.commit("Update README.md") |
|
repo.push_to_hub() |
|
except Exception as e: |
|
print(str(e)) |
|
|