crazycth's picture
Update README.md
0175589 verified
---
license: mit
dataset_info:
features:
- name: repo_name
dtype: string
- name: sub_path
dtype: string
- name: file_name
dtype: string
- name: file_ext
dtype: string
- name: file_size_in_byte
dtype: int64
- name: line_count
dtype: int64
- name: lang
dtype: string
- name: program_lang
dtype: string
- name: doc_type
dtype: string
splits:
- name: The_Stack_V2
num_bytes: 46577045485
num_examples: 336845710
download_size: 20019085005
dataset_size: 46577045485
configs:
- config_name: default
data_files:
- split: The_Stack_V2
path: data/The_Stack_V2-*
---
This dataset consists of meta information (including the repository name and file path) of the raw code data from **RefineCode**. You can collect those files referring to this metadata and reproduce **RefineCode**!
***Note:** Currently, we have uploaded the meta data covered by The Stack V2 (About 50% file volume). Due to complex legal considerations, we are unable to provide the complete source code currently. We are working hard to make the remaining part available.*
---
**RefineCode** is a **high-quality**, **reproducible** code pretraining corpus comprising **960 billion** tokens across **607** programming languages and **75 billion** code-related token recalled from web corpus, incorporating over **130** language-specific rules with customized weight assignments.
Our dataset shows better training efficacy and efficiency compared with the training subset of The Stack V2.
<img src="https://raw.githubusercontent.com/OpenCoder-llm/opencoder-llm.github.io/refs/heads/main/static/images/opencoder_banner.png" alt="OpenCoder banner" style="zoom:30%;" />
We also use PCA to visualize the embeddings extracted from CodeBERT for The Stack V2 and **RefineCode**, showing a clear advance of our pretraining dataset.
<img src="https://raw.githubusercontent.com/OpenCoder-llm/opencoder-llm.github.io/refs/heads/main/static/images/compare_refinecode_stack_v2.jpg" alt="Distribution Comparsion" style="zoom:50%;" />