orionweller commited on
Commit
ae7c587
1 Parent(s): 93d5d7d

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +25 -0
  2. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  3. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  4. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  5. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds +3 -0
  6. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds +3 -0
  7. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds +3 -0
  8. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds +3 -0
  9. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds +3 -0
  10. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00015.mds +3 -0
  11. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds +3 -0
  12. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds +3 -0
  13. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00020.mds +3 -0
  14. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds +3 -0
  15. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00023.mds +3 -0
  16. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00024.mds +3 -0
  17. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00026.mds +3 -0
  18. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00027.mds +3 -0
  19. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds +3 -0
  20. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00030.mds +3 -0
  21. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds +3 -0
  22. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00033.mds +3 -0
  23. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00034.mds +3 -0
  24. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00037.mds +3 -0
  25. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00039.mds +3 -0
  26. train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00040.mds +3 -0
  27. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10102-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  28. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10102-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  29. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10188-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  30. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10188-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  31. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_12934-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  32. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_12934-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  33. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21254-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  34. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21254-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  35. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21738-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  36. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21738-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  37. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24831-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  38. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24831-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  39. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26747-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  40. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26747-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  41. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_27585-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  42. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_27585-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  43. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_29203-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  44. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_29203-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  45. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_311-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  46. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_311-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  47. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_31388-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  48. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_31388-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  49. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_3296-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  50. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_3296-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
.gitattributes CHANGED
@@ -11142,3 +11142,28 @@ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/
11142
  train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds filter=lfs diff=lfs merge=lfs -text
11143
  train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
11144
  train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00019.mds filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11142
  train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds filter=lfs diff=lfs merge=lfs -text
11143
  train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
11144
  train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00019.mds filter=lfs diff=lfs merge=lfs -text
11145
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00015.mds filter=lfs diff=lfs merge=lfs -text
11146
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds filter=lfs diff=lfs merge=lfs -text
11147
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00024.mds filter=lfs diff=lfs merge=lfs -text
11148
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
11149
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds filter=lfs diff=lfs merge=lfs -text
11150
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
11151
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00027.mds filter=lfs diff=lfs merge=lfs -text
11152
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00026.mds filter=lfs diff=lfs merge=lfs -text
11153
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds filter=lfs diff=lfs merge=lfs -text
11154
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds filter=lfs diff=lfs merge=lfs -text
11155
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00040.mds filter=lfs diff=lfs merge=lfs -text
11156
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds filter=lfs diff=lfs merge=lfs -text
11157
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds filter=lfs diff=lfs merge=lfs -text
11158
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00039.mds filter=lfs diff=lfs merge=lfs -text
11159
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds filter=lfs diff=lfs merge=lfs -text
11160
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text
11161
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00037.mds filter=lfs diff=lfs merge=lfs -text
11162
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00030.mds filter=lfs diff=lfs merge=lfs -text
11163
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00020.mds filter=lfs diff=lfs merge=lfs -text
11164
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds filter=lfs diff=lfs merge=lfs -text
11165
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00023.mds filter=lfs diff=lfs merge=lfs -text
11166
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
11167
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00034.mds filter=lfs diff=lfs merge=lfs -text
11168
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds filter=lfs diff=lfs merge=lfs -text
11169
+ train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00033.mds filter=lfs diff=lfs merge=lfs -text
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7421075850c9a63e537ee1a8d62599399db442ec562a7c7110eeb2c9a43410b9
3
+ size 67108076
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ed35fb9e80b41cb215cca4fbb16a7af33b55250a56a33e71995a34cb9cea014
3
+ size 67108234
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c4c266c81a00a295fbf0d57a43ad3336cee737a4c633507974a12b5fcfc9e6b
3
+ size 67108601
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5042b0dcc8150d0d618d8ce0b04240f6e8d1073cac359bae6a467356bf2a847f
3
+ size 67107813
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d1bbc5f45c7486839f302fe84d7fa96d0e40aad2f06bc49d1061ef77ee25744
3
+ size 67107942
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8f120df34a7d63620632104abb853458802988c72366535e65175d35571a571
3
+ size 67108299
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8f9692e6c7bbb44e5a687ea1f293607bde6990668b7a09107f355202ee68451
3
+ size 67107518
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c4c069d0fffb2e639ff9748ffc4138330dd89cf4a821078e7cc1a7afe1cb0cf
3
+ size 67108223
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00015.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96be0facf6a9e0ab8d7dea51b486889fd77605e2b2dc7d3830b3e9067ca43306
3
+ size 67108578
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afea9b90d4c3033f43b4a62fd851a6907b72f1b327c16d8a1ed664462437ba1e
3
+ size 67107652
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dcbc4a5c13951702c70bf20f41d8711919cc2a4b761b6f0efa5d0a2663286f23
3
+ size 67108181
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00020.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:835c037c5275ad67f7e892c919c1bf97b261c22fb96bbd689ab8cea85f72d488
3
+ size 67108745
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c05b1af22eda42dba33f376629a3e5154e7a1e424b8a61f6a7d823cf0de24682
3
+ size 67108378
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00023.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3787e42ad68419b2f679a4af167a1151eadfad137a2e1fd0c0dd12beb64419e2
3
+ size 67108726
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00024.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ef2b5914788408e11f1d072d46001dca1ec823194734fa2ce683323697d1cfc
3
+ size 67107468
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00026.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8305987c41fb1126f6075d941fc1b88502118fd0a2d8180cd6073ec40e1d476c
3
+ size 67108632
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00027.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01000da0e072f4be67e468d65a99fce9c86b087e08cf7a16e5e7107a1d3d70d5
3
+ size 67108390
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9862242408d6e3b3f439c80abe767cd891d133370488135cff73aeefb69b75e1
3
+ size 67108809
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00030.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:745fde47c8313648cb8013b94c882393422037ed7a8e4e31f3b74c38dd41045b
3
+ size 67108636
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2cf3e55f3af5763b2f3f936639d4e36087a81cca99f40abfdd918d2149509e9
3
+ size 67107967
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00033.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26c6feec48c2542117701599e3c027eaea2b418c310bb2bc4bf2da0c02ae3eb3
3
+ size 67108765
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00034.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18218c2f64c6f32451fce82072fa51bce2ae2251967436a330ae8675d183d120
3
+ size 67106915
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00037.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfbb4f5432def322c3832b471dc3ae0ec2e8fdbd63ff1d233bcac1127b0696b1
3
+ size 67108453
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00039.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8571c3e9ec43cf4d28de9f4e2c9264c867aabff1aee8518265112157da5f55ec
3
+ size 67107184
train/cc_en_head/cc_en_head_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00040.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc5a3bc2e6fc1638936ef58dc6d86df255f019d19b92a4bc84237c036861db82
3
+ size 31183521
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10102-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107679, "hashes": {}}, "samples": 43145, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47491217, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 16610705, "hashes": {}}, "samples": 11052, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11687165, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10102-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 40181829,
3
+ "num_truncated_tokens": 40148241
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10188-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107229, "hashes": {}}, "samples": 44138, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47856817, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 11473646, "hashes": {}}, "samples": 7519, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8214019, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10188-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37691734,
3
+ "num_truncated_tokens": 37662351
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_12934-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108825, "hashes": {}}, "samples": 44229, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47890384, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 11315989, "hashes": {}}, "samples": 7387, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8065840, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_12934-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37614994,
3
+ "num_truncated_tokens": 37585929
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21254-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107331, "hashes": {}}, "samples": 43577, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47980011, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 15212583, "hashes": {}}, "samples": 9829, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10862874, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21254-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 39506928,
3
+ "num_truncated_tokens": 39474304
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21738-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108033, "hashes": {}}, "samples": 44045, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47565320, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 11184893, "hashes": {}}, "samples": 7339, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7933632, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21738-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37556168,
3
+ "num_truncated_tokens": 37527471
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24831-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108352, "hashes": {}}, "samples": 43030, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47546087, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 18652417, "hashes": {}}, "samples": 11945, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 13236173, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24831-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 41178745,
3
+ "num_truncated_tokens": 41143866
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26747-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107809, "hashes": {}}, "samples": 44100, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47640832, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 10145203, "hashes": {}}, "samples": 6818, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7257014, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26747-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37050679,
3
+ "num_truncated_tokens": 37022472
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_27585-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108515, "hashes": {}}, "samples": 42842, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47528055, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 18257149, "hashes": {}}, "samples": 12065, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12995009, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_27585-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 40983403,
3
+ "num_truncated_tokens": 40948870
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_29203-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107882, "hashes": {}}, "samples": 44372, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47882640, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 10343271, "hashes": {}}, "samples": 6851, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7367576, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_29203-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37140370,
3
+ "num_truncated_tokens": 37112120
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_311-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108019, "hashes": {}}, "samples": 44324, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48015373, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 9717628, "hashes": {}}, "samples": 6586, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6933479, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_311-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 36837271,
3
+ "num_truncated_tokens": 36808483
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_31388-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107596, "hashes": {}}, "samples": 44157, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47800678, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 11060264, "hashes": {}}, "samples": 7283, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7888352, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_31388-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37491962,
3
+ "num_truncated_tokens": 37463227
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_3296-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107486, "hashes": {}}, "samples": 44359, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47884182, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 10600341, "hashes": {}}, "samples": 6986, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7558254, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_3296-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37264954,
3
+ "num_truncated_tokens": 37236379
4
+ }