Spaces:
Running
Running
victormiller
commited on
Commit
•
d50754e
1
Parent(s):
b1b2d47
Update results.py
Browse files- results.py +1 -1
results.py
CHANGED
@@ -784,7 +784,7 @@ upsampling_exp = Div(
|
|
784 |
preplexity_intro_div = Div(
|
785 |
H2("Perplexity Evaluation on Duplicate Data"),
|
786 |
H3("Model based Quality Estimation"),
|
787 |
-
P("We took one of the model-based data quality evaluation strategies adopted by
|
788 |
H3("Sampling Strategy"),
|
789 |
P("We started from a processed Common Crawl (CC) ablation dataset divided by the number of duplicates of each document. For each CC dump, we have different buckets each holding chunks of document with different duplicate count ranges (1-1, 2-5, 6-10, 11-100, 101-1000, 1001-30000000). We sampled the first 10k documents from each chunk with their meta data."),
|
790 |
)
|
|
|
784 |
preplexity_intro_div = Div(
|
785 |
H2("Perplexity Evaluation on Duplicate Data"),
|
786 |
H3("Model based Quality Estimation"),
|
787 |
+
P("We took one of the model-based data quality evaluation strategies adopted by", A("DataComp-LM",href="https://arxiv.org/abs/2406.11794"), "which used perplexity filtering as a candidate for quality filtering. DataComp-LM followed", A("CCNet’s",href="https://arxiv.org/abs/1911.00359"), "practice to use a 5-gram Kneser-Ney model as implemented in the",A("KenLM",href="https://github.com/kpu/kenlm"), "library for efficient perplexity calculation. Following this practice, we estimated data quality by taking a", A("KenLM model",href="https://huggingface.co/edugp/kenlm"), "trained on English Wikipedia data to compute perplexity on data with different duplication patterns. Lower perplexity is regarded as a signal of higher quality."),
|
788 |
H3("Sampling Strategy"),
|
789 |
P("We started from a processed Common Crawl (CC) ablation dataset divided by the number of duplicates of each document. For each CC dump, we have different buckets each holding chunks of document with different duplicate count ranges (1-1, 2-5, 6-10, 11-100, 101-1000, 1001-30000000). We sampled the first 10k documents from each chunk with their meta data."),
|
790 |
)
|