orionweller commited on
Commit
8be0470
1 Parent(s): c29978b

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +30 -0
  2. train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  3. train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  4. train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  5. train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  6. train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  7. train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  8. train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  9. train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  10. train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  11. train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  12. train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  13. train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  14. train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  15. train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  16. train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds +3 -0
  17. train/arxiv/arxiv_0022-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  18. train/arxiv/arxiv_0022-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  19. train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  20. train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  21. train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  22. train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  23. train/arxiv/arxiv_0045-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  24. train/arxiv/arxiv_0045-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  25. train/arxiv/arxiv_0045-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  26. train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  27. train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  28. train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  29. train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  30. train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  31. train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  32. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10099-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  33. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10099-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  34. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13447-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  35. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13447-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  36. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13556-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  37. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13556-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  38. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24837-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  39. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24837-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  40. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26264-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  41. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26264-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  42. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26911-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  43. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26911-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  44. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_30247-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  45. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_30247-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  46. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_36655-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  47. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_36655-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  48. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_40917-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  49. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_40917-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  50. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_48842-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
.gitattributes CHANGED
@@ -13034,3 +13034,33 @@ train/arxiv/arxiv_0096-tokenized-chunked-1024-512-128-backfill-nodups/shard.0000
13034
  train/arxiv/arxiv_0096-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
13035
  train/arxiv/arxiv_0094-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
13036
  train/arxiv/arxiv_0094-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13034
  train/arxiv/arxiv_0096-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
13035
  train/arxiv/arxiv_0094-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
13036
  train/arxiv/arxiv_0094-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
13037
+ train/arxiv/arxiv_0022-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
13038
+ train/arxiv/arxiv_0045-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
13039
+ train/arxiv/arxiv_0045-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
13040
+ train/arxiv/arxiv_0045-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
13041
+ train/arxiv/arxiv_0022-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
13042
+ train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
13043
+ train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
13044
+ train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
13045
+ train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
13046
+ train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
13047
+ train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
13048
+ train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
13049
+ train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
13050
+ train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
13051
+ train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
13052
+ train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
13053
+ train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
13054
+ train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
13055
+ train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
13056
+ train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds filter=lfs diff=lfs merge=lfs -text
13057
+ train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
13058
+ train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
13059
+ train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
13060
+ train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
13061
+ train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
13062
+ train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
13063
+ train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
13064
+ train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
13065
+ train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
13066
+ train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3eb9e21d3e9d9881369e001204a3e4f842613173d7c69113b231c80dbbe0f5c
3
+ size 67107302
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7a6d0cd496ae38c5122ec214260b28c5b1d30d447495cf93a920416d890c34b
3
+ size 67108293
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4768de08a5d1d2b4a1b6d0577790f0ec244fdb5a38dcf02e7d403fa53af01fc
3
+ size 67107893
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fff3b012e340d3e8d948a949adee353df69119380586a0514892a4245ad9b903
3
+ size 67108602
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e71b9560deddbd406ba985baba3ed1cfbac866c4f253598987ce2ef7e544b40c
3
+ size 67108739
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b129aba0593c1a9884c5497fbca60b06a52a29b721e9dde2894d4e0aebaf479b
3
+ size 67106991
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5de56139912bdf596be412621276e55dd1e6720d91607b091a1fd74f687b23c5
3
+ size 67108428
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc105d4bc4f05c19ba333cf90ba63e23c0f4f080cdec60e98db5d94ddd937105
3
+ size 36527611
train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46f0c08676eaa2e37eddf1dd5de853b00f0f6e9c0ff5aa0febb7fc550a789978
3
+ size 67108845
train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b37b78490b6b2f146a02fe5b34dff7731814250a1228c57df03381eab4cb9da4
3
+ size 67108690
train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:105073732407f1d52963e42dc5080889cf2fe11c7d097a17a9947b9073b8471b
3
+ size 67107842
train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:954b487cabafbe55d233c5aa51f507885b5a351de688158d5a4e1c02680596fc
3
+ size 67107509
train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2f0aa816fbb91db56a1b4a6e765efffbf5e5da9edf2a577c013dc0812ff4628
3
+ size 67108053
train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8dc7a5a7c6d924b0e50d709b89b5fe445046696c28f11b86c05747655ac9a8b
3
+ size 67108675
train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06545256f4f3af2bfb3c98fa45016ff949cf5f4d8a8a480a62a7ed2b0e2af057
3
+ size 39801349
train/arxiv/arxiv_0022-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce94a379f4ec3a4dfc470c46bc52db57a270323b8d562214ffb094a2a48c359d
3
+ size 67106816
train/arxiv/arxiv_0022-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7dbf7bd7586f2a25270d46581202922e0e8cddb69bdb4ac86ae6758fc50f24b
3
+ size 67108269
train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf86e197f4cda99eb0c0375e2414fec5e9d89f99e1b43976a5cbe54146b4d137
3
+ size 67107494
train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f265c80b65d9dafa8b6012bf046afa546ed265f96ad9c0127303021bfded3363
3
+ size 67107921
train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:264924468772709330e53528717ffc8df76150a9f152dfdc2b2650226d321a9f
3
+ size 67107143
train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bec8cf2c74f514c512cc8d80e79f4dc83fce4fa21df235cc49dd13f436c297ba
3
+ size 46940576
train/arxiv/arxiv_0045-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4b7df7ff84adad5840933c2ea6e59fc171cc18d2b9b9d75f52aca35ac1316f6
3
+ size 67106840
train/arxiv/arxiv_0045-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db805049518866e631709c37d7592c4a9b405c9cfa1c02dbdd4f608fdabf46fa
3
+ size 67108336
train/arxiv/arxiv_0045-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ade23ffc875115b510db92ae800cc556c25070465c62d04e1d7c99c765143055
3
+ size 67108192
train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad8acf4c26d213121ef46c782a036c697949cb07635e7360aa4e55c71214228f
3
+ size 67107779
train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19257aeaba06bd7c11312c45c7d86fd7af6d032255ae198709f07a9754be193a
3
+ size 67107019
train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a466c8745e6e36409b0141b470e9e36c6aea1ef890c369acec5f539657f3cf61
3
+ size 67107212
train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da46ea510bcda904444c960ad5aab73c1b3e2c10106ec94b400998bd9a0927cc
3
+ size 67108639
train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b04464c95d3acb6b5b63a91554e5ba6e57876d297e77369ffa77548216cdcfc
3
+ size 67107067
train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14aa9e7655c091f97acc00215b5ebe912e10eaab97dd71135763f07d851636de
3
+ size 67107508
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10099-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108741, "hashes": {}}, "samples": 44587, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48013389, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 8245791, "hashes": {}}, "samples": 5591, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5932233, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10099-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 36124394,
3
+ "num_truncated_tokens": 36096811
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13447-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106926, "hashes": {}}, "samples": 44238, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47695037, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 10156618, "hashes": {}}, "samples": 6647, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7194563, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13447-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37056828,
3
+ "num_truncated_tokens": 37029302
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13556-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107564, "hashes": {}}, "samples": 44128, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47783033, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 12896056, "hashes": {}}, "samples": 8330, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9213245, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13556-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 38378384,
3
+ "num_truncated_tokens": 38347001
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24837-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108523, "hashes": {}}, "samples": 43181, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47952246, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 15197394, "hashes": {}}, "samples": 10104, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10888551, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24837-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 39503580,
3
+ "num_truncated_tokens": 39470996
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26264-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107339, "hashes": {}}, "samples": 44672, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47778233, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 11930677, "hashes": {}}, "samples": 7494, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8436633, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26264-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37904721,
3
+ "num_truncated_tokens": 37874336
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26911-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107486, "hashes": {}}, "samples": 42195, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47725976, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 25121038, "hashes": {}}, "samples": 15790, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 17891778, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26911-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 44319178,
3
+ "num_truncated_tokens": 44278622
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_30247-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108065, "hashes": {}}, "samples": 42415, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47693339, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 22080426, "hashes": {}}, "samples": 14115, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15700612, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_30247-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 42844314,
3
+ "num_truncated_tokens": 42806822
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_36655-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108434, "hashes": {}}, "samples": 42775, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47361721, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 20299073, "hashes": {}}, "samples": 13057, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 14429830, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_36655-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 41975632,
3
+ "num_truncated_tokens": 41939356
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_40917-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107749, "hashes": {}}, "samples": 44278, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47789440, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 10128538, "hashes": {}}, "samples": 6595, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7229481, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_40917-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37043660,
3
+ "num_truncated_tokens": 37015352
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_48842-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37077867,
3
+ "num_truncated_tokens": 37049904
4
+ }