Datasets:

Modalities:
Text
Formats:
json
Languages:
English
ArXiv:
Libraries:
Datasets
Dask
License:
File size: 3,532 Bytes
e03ae76
a38661d
e03ae76
 
 
 
 
 
 
2720d3e
e03ae76
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f6f8cac
 
 
 
 
 
 
 
 
 
 
 
 
 
e03ae76
 
 
90dfb5b
 
e03ae76
90dfb5b
 
 
 
 
 
 
 
 
 
 
 
 
e03ae76
90dfb5b
 
e03ae76
90dfb5b
 
 
 
e03ae76
 
a38661d
 
e03ae76
 
 
f6f8cac
 
ee26814
f6f8cac
 
 
 
 
 
 
 
ee26814
f6f8cac
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
---
license: odc-by
language:
- en
tags:
- pretrained
---
# Amber-Data

<img src="amber_logo.png" alt="amber logo" width="300"/>

This dataset contains the fully prepared data sequence used to train Amber, an
LLM360 model. 

## About LLM360
LLM360 is an initiative for comprehensive and fully open-sourced LLMs, 
where all training details, model checkpoints, intermediate results, and 
additional analyses are made available to the community. Our goal is to advance 
the field by inviting the community to deepen the understanding of LLMs 
together. As the first step of the project LLM360, we release all intermediate 
model checkpoints, our fully-prepared pre-training dataset, all source code and
configurations, and training details. We are
committed to continually pushing the boundaries of LLMs through this open-source 
effort.

Get access now at [LLM360 site](https://www.llm360.ai/)

## Data Description

- **Data Format:** 360 tokenized data chunks, each instance has 2049 token indexes.
- **License:** Apache 2.0
- **Resources for more information:**
  - [Code to produce data](https://github.com/LLM360/amber-data-prep)
  - [Amber Model](https://huggingface.co/LLM360/Amber)

## DataMix
The amber dataset uses the following data mix.

| Subset      | Tokens (Billion) |
| ----------- | ----------- |
| Arxiv      | 30.00       |
| Book   | 28.86        |
| C4   | 197.67        |
| Refined-Web   | 665.01        |
| StarCoder   | 291.92        |
| StackExchange   | 21.75        |
| Wikipedia   | 23.90        |
| Total | 1259.13 |


# Loading Amber's Pretraining Data

Below is an example of how to download, sample, and detokenize any subset of AmberDatasets corresponding to an Amber checkpoint. Just set the `CHECKPOINT_NUM` to the subset you are interested in (0-359) and point `CHECKPOINT_PATH` to the local checkpoint folder. 

```python
import random
from transformers import AutoTokenizer
from datasets import load_dataset

CHECKPOINT_NUM = 0  # Pretraining dataset for checkpoint
NUM_SAMPLES = 10  # Number of random samples to decode
CHECKPOINT_PATH = "/path/to/ckpt_000/"  # Local path to a Amber checkpoint

dataset = load_dataset(
    "LLM360/AmberDatasets",
    data_files=f"train/train_{CHECKPOINT_NUM:03}.jsonl",
    split=None,
)

tokenizer = AutoTokenizer.from_pretrained(CHECKPOINT_PATH)
samples = set(random.choices(range(len(dataset["train"])), k=NUM_SAMPLES))

for i, line in enumerate(dataset["train"]):
    if i in samples:
        tokens = line["token_ids"]
        print(f"{i}:{tokenizer.decode(tokens)}")
```

# License
We release our work under [ODC-BY](https://opendatacommons.org/licenses/by/1-0/), hence granting the rights over the dataset, but not the contents of the dataset individually.


# Citation
To cite LLM360, you can cite the following:

```
@misc{liu2023llm360,
      title={LLM360: Towards Fully Transparent Open-Source LLMs}, 
      author={Zhengzhong Liu and Aurick Qiao and Willie Neiswanger and Hongyi Wang and Bowen Tan and Tianhua Tao and Junbo Li and Yuqi Wang and Suqi Sun and Omkar Pangarkar and Richard Fan and Yi Gu and Victor Miller and Yonghao Zhuang and Guowei He and Haonan Li and Fajri Koto and Liping Tang and Nikhil Ranjan and Zhiqiang Shen and Xuguang Ren and Roberto Iriondo and Cun Mu and Zhiting Hu and Mark Schulze and Preslav Nakov and Tim Baldwin and Eric P. Xing},
      year={2023},
      eprint={2312.06550},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}
```

If you only uses the original dataset, please cite the original datasets.