victormiller commited on
Commit
d50754e
1 Parent(s): b1b2d47

Update results.py

Browse files
Files changed (1) hide show
  1. results.py +1 -1
results.py CHANGED
@@ -784,7 +784,7 @@ upsampling_exp = Div(
784
  preplexity_intro_div = Div(
785
  H2("Perplexity Evaluation on Duplicate Data"),
786
  H3("Model based Quality Estimation"),
787
- P("We took one of the model-based data quality evaluation strategies adopted by [DataComp-LM](https://arxiv.org/abs/2406.11794), which used perplexity filtering as a candidate for quality filtering. DataComp-LM followed [CCNet’s](https://arxiv.org/abs/1911.00359) practice to use a 5-gram Kneser-Ney model as implemented in the [KenLM](https://github.com/kpu/kenlm) library for efficient perplexity calculation. Following this practice, we estimated data quality by taking a KenLM model (from [edugp/kenlm](https://huggingface.co/edugp/kenlm)) trained on English Wikipedia data to compute perplexity on data with different duplication patterns. Lower perplexity is regarded as a signal of higher quality."),
788
  H3("Sampling Strategy"),
789
  P("We started from a processed Common Crawl (CC) ablation dataset divided by the number of duplicates of each document. For each CC dump, we have different buckets each holding chunks of document with different duplicate count ranges (1-1, 2-5, 6-10, 11-100, 101-1000, 1001-30000000). We sampled the first 10k documents from each chunk with their meta data."),
790
  )
 
784
  preplexity_intro_div = Div(
785
  H2("Perplexity Evaluation on Duplicate Data"),
786
  H3("Model based Quality Estimation"),
787
+ P("We took one of the model-based data quality evaluation strategies adopted by", A("DataComp-LM",href="https://arxiv.org/abs/2406.11794"), "which used perplexity filtering as a candidate for quality filtering. DataComp-LM followed", A("CCNet’s",href="https://arxiv.org/abs/1911.00359"), "practice to use a 5-gram Kneser-Ney model as implemented in the",A("KenLM",href="https://github.com/kpu/kenlm"), "library for efficient perplexity calculation. Following this practice, we estimated data quality by taking a", A("KenLM model",href="https://huggingface.co/edugp/kenlm"), "trained on English Wikipedia data to compute perplexity on data with different duplication patterns. Lower perplexity is regarded as a signal of higher quality."),
788
  H3("Sampling Strategy"),
789
  P("We started from a processed Common Crawl (CC) ablation dataset divided by the number of duplicates of each document. For each CC dump, we have different buckets each holding chunks of document with different duplicate count ranges (1-1, 2-5, 6-10, 11-100, 101-1000, 1001-30000000). We sampled the first 10k documents from each chunk with their meta data."),
790
  )