orionweller
commited on
Commit
•
8be0470
1
Parent(s):
c29978b
Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +30 -0
- train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
- train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
- train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
- train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
- train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
- train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
- train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
- train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
- train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
- train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
- train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
- train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
- train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
- train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
- train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds +3 -0
- train/arxiv/arxiv_0022-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
- train/arxiv/arxiv_0022-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
- train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
- train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
- train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
- train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
- train/arxiv/arxiv_0045-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
- train/arxiv/arxiv_0045-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
- train/arxiv/arxiv_0045-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
- train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
- train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
- train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
- train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
- train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
- train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
- train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10099-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
- train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10099-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
- train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13447-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
- train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13447-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
- train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13556-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
- train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13556-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
- train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24837-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
- train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24837-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
- train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26264-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
- train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26264-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
- train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26911-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
- train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26911-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
- train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_30247-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
- train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_30247-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
- train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_36655-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
- train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_36655-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
- train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_40917-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
- train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_40917-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
- train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_48842-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
.gitattributes
CHANGED
@@ -13034,3 +13034,33 @@ train/arxiv/arxiv_0096-tokenized-chunked-1024-512-128-backfill-nodups/shard.0000
|
|
13034 |
train/arxiv/arxiv_0096-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
|
13035 |
train/arxiv/arxiv_0094-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
|
13036 |
train/arxiv/arxiv_0094-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
13034 |
train/arxiv/arxiv_0096-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
|
13035 |
train/arxiv/arxiv_0094-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
|
13036 |
train/arxiv/arxiv_0094-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
|
13037 |
+
train/arxiv/arxiv_0022-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
|
13038 |
+
train/arxiv/arxiv_0045-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
|
13039 |
+
train/arxiv/arxiv_0045-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
|
13040 |
+
train/arxiv/arxiv_0045-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
|
13041 |
+
train/arxiv/arxiv_0022-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
|
13042 |
+
train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
|
13043 |
+
train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
|
13044 |
+
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
|
13045 |
+
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
|
13046 |
+
train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
|
13047 |
+
train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
|
13048 |
+
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
|
13049 |
+
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
|
13050 |
+
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
|
13051 |
+
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
|
13052 |
+
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
|
13053 |
+
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
|
13054 |
+
train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
|
13055 |
+
train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
|
13056 |
+
train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds filter=lfs diff=lfs merge=lfs -text
|
13057 |
+
train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
|
13058 |
+
train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
|
13059 |
+
train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
|
13060 |
+
train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
|
13061 |
+
train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
|
13062 |
+
train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
|
13063 |
+
train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
|
13064 |
+
train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
|
13065 |
+
train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
|
13066 |
+
train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
|
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a3eb9e21d3e9d9881369e001204a3e4f842613173d7c69113b231c80dbbe0f5c
|
3 |
+
size 67107302
|
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d7a6d0cd496ae38c5122ec214260b28c5b1d30d447495cf93a920416d890c34b
|
3 |
+
size 67108293
|
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b4768de08a5d1d2b4a1b6d0577790f0ec244fdb5a38dcf02e7d403fa53af01fc
|
3 |
+
size 67107893
|
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fff3b012e340d3e8d948a949adee353df69119380586a0514892a4245ad9b903
|
3 |
+
size 67108602
|
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e71b9560deddbd406ba985baba3ed1cfbac866c4f253598987ce2ef7e544b40c
|
3 |
+
size 67108739
|
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b129aba0593c1a9884c5497fbca60b06a52a29b721e9dde2894d4e0aebaf479b
|
3 |
+
size 67106991
|
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5de56139912bdf596be412621276e55dd1e6720d91607b091a1fd74f687b23c5
|
3 |
+
size 67108428
|
train/arxiv/arxiv_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bc105d4bc4f05c19ba333cf90ba63e23c0f4f080cdec60e98db5d94ddd937105
|
3 |
+
size 36527611
|
train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:46f0c08676eaa2e37eddf1dd5de853b00f0f6e9c0ff5aa0febb7fc550a789978
|
3 |
+
size 67108845
|
train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b37b78490b6b2f146a02fe5b34dff7731814250a1228c57df03381eab4cb9da4
|
3 |
+
size 67108690
|
train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:105073732407f1d52963e42dc5080889cf2fe11c7d097a17a9947b9073b8471b
|
3 |
+
size 67107842
|
train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:954b487cabafbe55d233c5aa51f507885b5a351de688158d5a4e1c02680596fc
|
3 |
+
size 67107509
|
train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d2f0aa816fbb91db56a1b4a6e765efffbf5e5da9edf2a577c013dc0812ff4628
|
3 |
+
size 67108053
|
train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e8dc7a5a7c6d924b0e50d709b89b5fe445046696c28f11b86c05747655ac9a8b
|
3 |
+
size 67108675
|
train/arxiv/arxiv_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:06545256f4f3af2bfb3c98fa45016ff949cf5f4d8a8a480a62a7ed2b0e2af057
|
3 |
+
size 39801349
|
train/arxiv/arxiv_0022-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ce94a379f4ec3a4dfc470c46bc52db57a270323b8d562214ffb094a2a48c359d
|
3 |
+
size 67106816
|
train/arxiv/arxiv_0022-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e7dbf7bd7586f2a25270d46581202922e0e8cddb69bdb4ac86ae6758fc50f24b
|
3 |
+
size 67108269
|
train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cf86e197f4cda99eb0c0375e2414fec5e9d89f99e1b43976a5cbe54146b4d137
|
3 |
+
size 67107494
|
train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f265c80b65d9dafa8b6012bf046afa546ed265f96ad9c0127303021bfded3363
|
3 |
+
size 67107921
|
train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:264924468772709330e53528717ffc8df76150a9f152dfdc2b2650226d321a9f
|
3 |
+
size 67107143
|
train/arxiv/arxiv_0032-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bec8cf2c74f514c512cc8d80e79f4dc83fce4fa21df235cc49dd13f436c297ba
|
3 |
+
size 46940576
|
train/arxiv/arxiv_0045-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b4b7df7ff84adad5840933c2ea6e59fc171cc18d2b9b9d75f52aca35ac1316f6
|
3 |
+
size 67106840
|
train/arxiv/arxiv_0045-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:db805049518866e631709c37d7592c4a9b405c9cfa1c02dbdd4f608fdabf46fa
|
3 |
+
size 67108336
|
train/arxiv/arxiv_0045-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ade23ffc875115b510db92ae800cc556c25070465c62d04e1d7c99c765143055
|
3 |
+
size 67108192
|
train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ad8acf4c26d213121ef46c782a036c697949cb07635e7360aa4e55c71214228f
|
3 |
+
size 67107779
|
train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:19257aeaba06bd7c11312c45c7d86fd7af6d032255ae198709f07a9754be193a
|
3 |
+
size 67107019
|
train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a466c8745e6e36409b0141b470e9e36c6aea1ef890c369acec5f539657f3cf61
|
3 |
+
size 67107212
|
train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:da46ea510bcda904444c960ad5aab73c1b3e2c10106ec94b400998bd9a0927cc
|
3 |
+
size 67108639
|
train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0b04464c95d3acb6b5b63a91554e5ba6e57876d297e77369ffa77548216cdcfc
|
3 |
+
size 67107067
|
train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:14aa9e7655c091f97acc00215b5ebe912e10eaab97dd71135763f07d851636de
|
3 |
+
size 67107508
|
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10099-tokenized-chunked-1024-512-128-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108741, "hashes": {}}, "samples": 44587, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48013389, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 8245791, "hashes": {}}, "samples": 5591, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5932233, "hashes": {}}}], "version": 2}
|
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_10099-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"num_tokens": 36124394,
|
3 |
+
"num_truncated_tokens": 36096811
|
4 |
+
}
|
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13447-tokenized-chunked-1024-512-128-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106926, "hashes": {}}, "samples": 44238, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47695037, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 10156618, "hashes": {}}, "samples": 6647, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7194563, "hashes": {}}}], "version": 2}
|
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13447-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"num_tokens": 37056828,
|
3 |
+
"num_truncated_tokens": 37029302
|
4 |
+
}
|
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13556-tokenized-chunked-1024-512-128-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107564, "hashes": {}}, "samples": 44128, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47783033, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 12896056, "hashes": {}}, "samples": 8330, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9213245, "hashes": {}}}], "version": 2}
|
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13556-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"num_tokens": 38378384,
|
3 |
+
"num_truncated_tokens": 38347001
|
4 |
+
}
|
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24837-tokenized-chunked-1024-512-128-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108523, "hashes": {}}, "samples": 43181, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47952246, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 15197394, "hashes": {}}, "samples": 10104, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10888551, "hashes": {}}}], "version": 2}
|
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24837-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"num_tokens": 39503580,
|
3 |
+
"num_truncated_tokens": 39470996
|
4 |
+
}
|
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26264-tokenized-chunked-1024-512-128-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107339, "hashes": {}}, "samples": 44672, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47778233, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 11930677, "hashes": {}}, "samples": 7494, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8436633, "hashes": {}}}], "version": 2}
|
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26264-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"num_tokens": 37904721,
|
3 |
+
"num_truncated_tokens": 37874336
|
4 |
+
}
|
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26911-tokenized-chunked-1024-512-128-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107486, "hashes": {}}, "samples": 42195, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47725976, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 25121038, "hashes": {}}, "samples": 15790, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 17891778, "hashes": {}}}], "version": 2}
|
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26911-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"num_tokens": 44319178,
|
3 |
+
"num_truncated_tokens": 44278622
|
4 |
+
}
|
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_30247-tokenized-chunked-1024-512-128-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108065, "hashes": {}}, "samples": 42415, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47693339, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 22080426, "hashes": {}}, "samples": 14115, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15700612, "hashes": {}}}], "version": 2}
|
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_30247-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"num_tokens": 42844314,
|
3 |
+
"num_truncated_tokens": 42806822
|
4 |
+
}
|
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_36655-tokenized-chunked-1024-512-128-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108434, "hashes": {}}, "samples": 42775, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47361721, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 20299073, "hashes": {}}, "samples": 13057, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 14429830, "hashes": {}}}], "version": 2}
|
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_36655-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"num_tokens": 41975632,
|
3 |
+
"num_truncated_tokens": 41939356
|
4 |
+
}
|
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_40917-tokenized-chunked-1024-512-128-backfill-nodups/index.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107749, "hashes": {}}, "samples": 44278, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47789440, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 10128538, "hashes": {}}, "samples": 6595, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7229481, "hashes": {}}}], "version": 2}
|
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_40917-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"num_tokens": 37043660,
|
3 |
+
"num_truncated_tokens": 37015352
|
4 |
+
}
|
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_48842-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"num_tokens": 37077867,
|
3 |
+
"num_truncated_tokens": 37049904
|
4 |
+
}
|