Datasets:

Languages:
English
ArXiv:
ZennyKenny commited on
Commit
8a48c4c
1 Parent(s): 2d0accd

Update GitHub directory path

Browse files

Replaced path points to dead link. Please also note, [this blog post](https://www.cerebras.net/blog/slimpajama-a-627b-token-cleaned-and-deduplicated-version-of-redpajama) contains the same invalid pointer ("Additional steps such as tokenization, sequence packing, sequence-level shuffling as well as upsampling can be performed using our scripts located at...).

Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -14,7 +14,7 @@ pretty_name: SlimPajama-627B
14
 
15
  The dataset consists of 59166 jsonl files and is ~895GB compressed. It is a cleaned and deduplicated version of [Together's RedPajama](https://github.com/togethercomputer/redpajama-data).
16
 
17
- Check out our [blog post](https://www.cerebras.net/blog/slimpajama-a-627b-token-cleaned-and-deduplicated-version-of-redpajama) explaining our methods, [our code on GitHub](https://github.com/Cerebras/modelzoo/tree/main/modelzoo/transformers/data_processing/slimpajama), and join the discussion on the [Cerebras Discord](https://discord.gg/q6bZcMWJVu).
18
 
19
  ## Getting Started
20
  You can download the dataset using Hugging Face datasets:
 
14
 
15
  The dataset consists of 59166 jsonl files and is ~895GB compressed. It is a cleaned and deduplicated version of [Together's RedPajama](https://github.com/togethercomputer/redpajama-data).
16
 
17
+ Check out our [blog post](https://www.cerebras.net/blog/slimpajama-a-627b-token-cleaned-and-deduplicated-version-of-redpajama) explaining our methods, [our code on GitHub](https://github.com/Cerebras/modelzoo/tree/main/src/cerebras/modelzoo/data_preparation/nlp/slimpajama), and join the discussion on the [Cerebras Discord](https://discord.gg/q6bZcMWJVu).
18
 
19
  ## Getting Started
20
  You can download the dataset using Hugging Face datasets: