Datasets:

ArXiv:
License:
vaishaal commited on
Commit
8030d98
1 Parent(s): 514821e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +82 -3
README.md CHANGED
@@ -1,3 +1,82 @@
1
- ---
2
- license: cc-by-4.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-4.0
3
+ ---
4
+ ## DCLM-baseline
5
+ DCLM-baseline is a 4T token pretraining datasets that achieves strong performance on language model benchmarks.
6
+
7
+
8
+ Below are comparisions of model trained on DCLM-baseline with other models in the 7B regime.
9
+
10
+ | Model | Params | Tokens | Open dataset? | CORE | MMLU | EXTENDED |
11
+ |---------------|--------|--------|---------------|----------|----------|----------|
12
+ | **Open weights, closed datasets** | | | | | | |
13
+ | Llama2 | 7B | 2T | ✗ | 49.2 | 45.8 | 34.1 |
14
+ | DeepSeek | 7B | 2T | ✗ | 50.7 | 48.5 | 35.3 |
15
+ | Mistral-0.3 | 7B | ? | ✗ | 57.0 | 62.7 | 45.1 |
16
+ | QWEN-2 | 7B | ? | ✗ | 57.5 | **71.9** | 50.5 |
17
+ | Llama3 | 8B | 15T | ✗ | 57.6 | 66.2 | 46.3 |
18
+ | Gemma | 8B | 6T | ✗ | 57.8 | 64.3 | 44.6 |
19
+ | Phi-3 | 7B | ? | ✗ | **61.0** | 69.9 | **57.9** |
20
+ | **Open weights, open datasets** | | | | | | |
21
+ | Falcon | 7B | 1T | ✓ | 44.1 | 27.4 | 25.1 |
22
+ | OLMo-1.7 | 7B | 2.1T | ✓ | 47.0 | 54.0 | 34.2 |
23
+ | MAP-Neo | 7B | 4.5T | ✓ | **50.2** | **57.1** | **40.4** |
24
+ | **Models we trained** | | | | | | |
25
+ | FineWeb edu | 7B | 0.14T | ✓ | 38.7 | 26.3 | 22.1 |
26
+ | FineWeb edu | 7B | 0.28T | ✓ | 41.9 | 37.3 | 24.5 |
27
+ | **DCLM-BASELINE** | 7B | 0.14T | ✓ | 44.1 | 38.3 | 25.0 |
28
+ | **DCLM-BASELINE** | 7B | 0.28T | ✓ | 48.9 | 50.8 | 31.8 |
29
+ | **DCLM-BASELINE** | 7B | 2.6T | ✓ | **57.1** | **63.7** | **45.4** |
30
+
31
+
32
+ ## Dataset Details
33
+ ### Dataset Description
34
+ - **Curated by:** The DCLM Team
35
+ - **Language(s) (NLP):** English
36
+ - **License:** CC-by-4.0
37
+ ### Dataset Sources
38
+ - **Repository:** https://datacomp.ai/dclm
39
+ - **Paper:**: https://arxiv.org/abs/2406.11794
40
+ - **Construction Code**: https://github.com/mlfoundations/dclm
41
+
42
+
43
+
44
+ ## Uses
45
+ ### Direct Use
46
+ DCLM-Baseline is intended to be used as a research baseline for the DCLM benchmark. It demonstrates the importance of data curation in training performant language models.
47
+ ### Out-of-Scope Use
48
+ DCLM-Baseline is not intended for training production-ready models or for specific domains such as code and math. It may not perform as well as domain-specific datasets for these tasks. Due to these limitations, the dataset is intended for research use only.
49
+ DCLM-Baseline is a subset of the DCLM-Pool, which is a corpus of 240 trillion tokens derived from Common Crawl. The dataset is in plain text format.
50
+ ## Dataset Creation
51
+ ### Curation Rationale
52
+ DCLM-Baseline was created to demonstrate the effectiveness of the DCLM testbed in developing high-quality training sets for language models. It serves as a proof of concept for the data curation strategies enabled by DCLM and is designed to be a research baseline for the benchmark.
53
+ ### Source Data
54
+ #### Data Collection and Processing
55
+ DCLM-Baseline was created by applying a series of cleaning, filtering, and deduplication steps to the raw Common Crawl data (DCLM-Pool). The key steps include:
56
+ 1. Heuristic cleaning and filtering (reproduction of RefinedWeb)
57
+ 2. Deduplication using a Bloom filter
58
+ 3. Model-based filtering using a fastText classifier trained on instruction-formatted data (OpenHermes 2.5 and r/ExplainLikeImFive)
59
+ #### Who are the source data producers?
60
+ The source data is from Common Crawl, which is a repository of web crawl data.
61
+ ### Personal and Sensitive Information
62
+ [More Information Needed]
63
+ ## Bias, Risks, and Limitations
64
+ The dataset may contain biases present in the Common Crawl data. The dataset's performance on code and math tasks is limited compared to its performance on language understanding tasks. DCLM-Baseline is designed for research purposes only.
65
+ ### Recommendations
66
+ Users should be aware of the potential biases and limitations of the dataset, especially when using it for specific domains like code and math. The dataset should only be used for research purposes in the context of the DCLM benchmark.
67
+ ## Citation
68
+
69
+ ```bibtex
70
+ @misc{li2024datacomplm,
71
+ title={DataComp-LM: In search of the next generation of training sets for language models},
72
+ author={Jeffrey Li and Alex Fang and Georgios Smyrnis and Maor Ivgi and Matt Jordan and Samir Gadre and Hritik Bansal and Etash Guha and Sedrick Keh and Kushal Arora and Saurabh Garg and Rui Xin and Niklas Muennighoff and Reinhard Heckel and Jean Mercat and Mayee Chen and Suchin Gururangan and Mitchell Wortsman and Alon Albalak and Yonatan Bitton and Marianna Nezhurina and Amro Abbas and Cheng-Yu Hsieh and Dhruba Ghosh and Josh Gardner and Maciej Kilian and Hanlin Zhang and Rulin Shao and Sarah Pratt and Sunny Sanyal and Gabriel Ilharco and Giannis Daras and Kalyani Marathe and Aaron Gokaslan and Jieyu Zhang and Khyathi Chandu and Thao Nguyen and Igor Vasiljevic and Sham Kakade and Shuran Song and Sujay Sanghavi and Fartash Faghri and Sewoong Oh and Luke Zettlemoyer and Kyle Lo and Alaaeldin El-Nouby and Hadi Pouransari and Alexander Toshev and Stephanie Wang and Dirk Groeneveld and Luca Soldaini and Pang Wei Koh and Jenia Jitsev and Thomas Kollar and Alexandros G. Dimakis and Yair Carmon and Achal Dave and Ludwig Schmidt and Vaishaal Shankar},
73
+ year={2024},
74
+ eprint={2406.11794},
75
+ archivePrefix={arXiv},
76
+ primaryClass={id='cs.LG' full_name='Machine Learning' is_active=True alt_name=None in_archive='cs' is_general=False description='Papers on all aspects of machine learning research (supervised, unsupervised, reinforcement learning, bandit problems, and so on) including also robustness, explanation, fairness, and methodology. cs.LG is also an appropriate primary category for applications of machine learning methods.'}
77
+
78
+ ```
79
+
80
+
81
+
82
+ [Contact information for the DCLM Team]