Error when loading the dataset and are there files missing?
Hi!
I want to pre-train a German language model and found your dataset which just seems perfect for my task.
But now I have the following problem: when I call hugging face's load_dataset method I get the following error:
Traceback (most recent call last):
File "/Users/marisa/clausal-coordinate-ellipsis/german-common-crawl/test_huggingface_dataset.py", line 5, in <module>
dataset = load_dataset("german-nlp-group/german_common_crawl")
File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/site-packages/datasets/load.py", line 2153, in load_dataset
builder_instance.download_and_prepare(
File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/site-packages/datasets/builder.py", line 954, in download_and_prepare
self._download_and_prepare(
File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/site-packages/datasets/builder.py", line 1717, in _download_and_prepare
super()._download_and_prepare(
File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/site-packages/datasets/builder.py", line 1027, in _download_and_prepare
split_generators = self._split_generators(dl_manager, **split_generators_kwargs)
File "/Users/marisa/.cache/huggingface/modules/datasets_modules/datasets/german-nlp-group--german_common_crawl/8373f099be5c05b1c1af7625a270bfdd14841287f7e61c4e2d3f922fceb6f8b7/german_common_crawl.py", line 105, in _split_generators
if self.config == "first_part":
File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/site-packages/datasets/builder.py", line 147, in __eq__
if set(self.__dict__.keys()) != set(o.__dict__.keys()):
AttributeError: 'str' object has no attribute '__dict__'. Did you mean: '__dir__'?
Also it seems like there are only two of the files that belong to the dataset uploaded, on your homepage, there are a lot more files listed: https://german-nlp-group.github.io/projects/gc4-corpus.html
Are you planning to upload the rest of the files?
What would be the steps to use all (or at least more than two) files to train my model?
Thanks for your answer!
To be honest: I think the best was would be to use the data from this location:
https://german-nlp-group.github.io/projects/gc4-corpus.html
Okay, then I will do that. Thanks for your answer!
Maybe you want to use this dataset:
https://huggingface.co/datasets/bjoernp/oscar2023_deduped_filtered_1.1
It is more recent and had been used to train LeoLM.
Or the German part of this:
https://huggingface.co/datasets/togethercomputer/RedPajama-Data-V2
I will have a look at these. Thank you for your help! :)
I can also offer a sentence split version of the GC4 corpus.
Not sure what you want to do with the data?
Are you interested in an exchange?
Perhaps you can contact me?
Via LinkedIn or philip at may.la
The goal is to train a "German to German" translation model first, which then should be fine-tuned on a smaller corpus for a special task.
I will sent you an email.