orionweller commited on
Commit
cc1aa68
1 Parent(s): f548838

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +33 -0
  2. train/arxiv/arxiv_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  3. train/arxiv/arxiv_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  4. train/arxiv/arxiv_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  5. train/arxiv/arxiv_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  6. train/arxiv/arxiv_0036-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  7. train/arxiv/arxiv_0036-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  8. train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  9. train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  10. train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  11. train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  12. train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  13. train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  14. train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  15. train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  16. train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds +3 -0
  17. train/arxiv/arxiv_0047-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  18. train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  19. train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  20. train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  21. train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  22. train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  23. train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  24. train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  25. train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  26. train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  27. train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  28. train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  29. train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  30. train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  31. train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  32. train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  33. train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  34. train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds +3 -0
  35. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10408-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  36. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10408-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  37. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10563-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  38. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10563-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  39. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11626-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  40. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11626-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  41. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11821-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  42. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11821-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  43. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19124-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  44. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19124-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  45. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21730-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  46. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21730-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  47. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_22844-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  48. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_22844-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  49. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_25030-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  50. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_25030-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
.gitattributes CHANGED
@@ -12666,3 +12666,36 @@ train/arxiv/arxiv_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.0000
12666
  train/arxiv/arxiv_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12667
  train/arxiv/arxiv_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12668
  train/arxiv/arxiv_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12666
  train/arxiv/arxiv_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12667
  train/arxiv/arxiv_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12668
  train/arxiv/arxiv_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
12669
+ train/arxiv/arxiv_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12670
+ train/arxiv/arxiv_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
12671
+ train/arxiv/arxiv_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
12672
+ train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
12673
+ train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12674
+ train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
12675
+ train/arxiv/arxiv_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
12676
+ train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12677
+ train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12678
+ train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
12679
+ train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
12680
+ train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text
12681
+ train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12682
+ train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12683
+ train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12684
+ train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
12685
+ train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
12686
+ train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12687
+ train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
12688
+ train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
12689
+ train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
12690
+ train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12691
+ train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
12692
+ train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
12693
+ train/arxiv/arxiv_0047-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12694
+ train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text
12695
+ train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12696
+ train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12697
+ train/arxiv/arxiv_0036-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
12698
+ train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12699
+ train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12700
+ train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
12701
+ train/arxiv/arxiv_0036-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
train/arxiv/arxiv_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:692b2a7c110dc864caafaab7e8c0f048b922b75ed3685b9c2abd41dc65cbaa82
3
+ size 67108852
train/arxiv/arxiv_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:036c723ab17cb654695ab42f407a814b06d6982a760b970a9d8fcaa227aead18
3
+ size 67106867
train/arxiv/arxiv_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe0ad3dd27ba9ea8566be533f1870ac7a289a60b1230f658d64bbda6243dae06
3
+ size 67108597
train/arxiv/arxiv_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:392ea8dc0f02f86794676802ba6c8b146bab8d7381e11d63d2517975fc9c897e
3
+ size 67107737
train/arxiv/arxiv_0036-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61da103e73f37cb16cabff72de5201959b90642a64d39430222a8877eddf9842
3
+ size 67107591
train/arxiv/arxiv_0036-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02692b35e2ca2203f4791fdb83448d13fde8d81079bb269f1f73d7d721983900
3
+ size 60007668
train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13cf05b479e820b418b626929e1fe1f3257fcb0bfb397b305349bfa98e1514d3
3
+ size 67108519
train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89224183f3a49d6873176277135f8fcb5c3fbb9a815e5dec801f53c3c3147466
3
+ size 67108183
train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:906837c5c5ec0d8ab8abf316cf08c7e496e3f1a2241eeb939a3c6f097fa58833
3
+ size 67108832
train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a81ddd970b21e5fdb105c9d3512d0ed3019038d2f9d933abae960b309f11c1de
3
+ size 67107016
train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c9aba233588bc7de64d398ba546b6c4996bdb093320a8b68266ade5be0baac9
3
+ size 67107647
train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bbd21ba6229b81f58e421d0b94a79069f9691b629d4c9f3b4a3d708d2b04bdd6
3
+ size 67106767
train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3900f89afe59cd586a3594167ca071650993dd40a434c0bdd258e22230861d49
3
+ size 67108646
train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:667351e2679555ee0a13c515742f169dd732db0acaceb0d688b6f8dc566c160b
3
+ size 67108007
train/arxiv/arxiv_0039-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d19df357a7b5642b83b8bc84895870095aa6a0336b9f913979c397305acea222
3
+ size 39302631
train/arxiv/arxiv_0047-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea27b53b3cde3a04852eae1234c9148953405a69ce6cbdaed91957c97f7afb75
3
+ size 67108700
train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:020071c181d3c860e73f3fe5e0800647984feb67df04c8f13d4bc70c67a652d8
3
+ size 67108390
train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7aa6e08ce1329dd797e198cdfafece083b62d5f8282017c0adeb9501029fd94
3
+ size 67107687
train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac262d3a1048fd2cd6e332f375d4d875cf318c3f12dd1fbc6bf71b0212319439
3
+ size 67107255
train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4ea8da39658f2f518f263fee85ab97644a4b9520589862d822ed48237e510bc
3
+ size 67106945
train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d2cdd6d9a73b1887ef89501f1104bf0ad89c0c9e8388cf7cdfe32d2f1117ee2
3
+ size 67108102
train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc49c18f6e15679d32a875410e8fce00085d9a11efd5a6e27c5112f95cc92f43
3
+ size 67108656
train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:444bc2c00c133c88ad12538530916c923f413f323de9d098130772dc786f5d81
3
+ size 67107051
train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a276d0402f00b892f2aff2a2e520afa6b7055d5f255dc7b323866243ab57e6b
3
+ size 42526957
train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c43397b8bf4dc8d1e26ebd2a2b49ee1e5ec2acd67c46b510a2636f14a60e5d38
3
+ size 67108633
train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e26678f19d133c38a72fc8dfcc36e643585ae2100f34b1d87e5854e2f484923e
3
+ size 67108241
train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e0e2ea5806a28818872c2367b718352266fa1264ef4f69e4aba1034973117ac
3
+ size 67107860
train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7073464d7496c6b91c3b7de76f1465a4a08608885fa9fd0ae6fbf301807b2a9
3
+ size 67106897
train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b59accd729cff24ff97b65d52b644c904310d1751f1a5e2d85ce53ab64bb93a9
3
+ size 67107429
train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:209bd7e96d19e2c03a986494685ce7903d73d5a32b5108e5d0537f3cba9dcd8e
3
+ size 67106942
train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ea7ad7d9368de95ff68f5c3ed5cdbfb9bd554def0457658df94d3dfd0497fbe
3
+ size 67108038
train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c2d7279b13f22636425325ec8a36983cd0d7a9679a34de3477b2aa87cbdccad
3
+ size 67107108
train/arxiv/arxiv_0092-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0084b83447c67a04da94a9d4405d40a12c219b37dcf0367cc5769922a337e27
3
+ size 55135243
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10408-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108726, "hashes": {}}, "samples": 44716, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47795797, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 8281193, "hashes": {}}, "samples": 5421, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5867100, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10408-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 36143388,
3
+ "num_truncated_tokens": 36116698
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10563-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108352, "hashes": {}}, "samples": 42727, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47751402, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 22031521, "hashes": {}}, "samples": 13804, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15726733, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10563-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 42820035,
3
+ "num_truncated_tokens": 42782184
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11626-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107710, "hashes": {}}, "samples": 45037, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47888906, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 6095181, "hashes": {}}, "samples": 4102, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 4351962, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11626-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 35080845,
3
+ "num_truncated_tokens": 35055801
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11821-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107861, "hashes": {}}, "samples": 43980, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47825635, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 12717621, "hashes": {}}, "samples": 8317, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9083308, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11821-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 38294196,
3
+ "num_truncated_tokens": 38263930
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19124-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108451, "hashes": {}}, "samples": 43044, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47695315, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 19006438, "hashes": {}}, "samples": 12084, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 13466019, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19124-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 41351025,
3
+ "num_truncated_tokens": 41316287
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21730-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108007, "hashes": {}}, "samples": 43896, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47713181, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 14898208, "hashes": {}}, "samples": 9523, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10552518, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21730-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 39349853,
3
+ "num_truncated_tokens": 39317337
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_22844-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108858, "hashes": {}}, "samples": 44081, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47759676, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 11312959, "hashes": {}}, "samples": 7413, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8026229, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_22844-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37617170,
3
+ "num_truncated_tokens": 37588191
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_25030-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107445, "hashes": {}}, "samples": 43537, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47667766, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 14394067, "hashes": {}}, "samples": 9287, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10264287, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_25030-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 39115725,
3
+ "num_truncated_tokens": 39084265
4
+ }