Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- train/arxiv/arxiv_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
- train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
- train/arxiv/arxiv_0074-tokenized-chunked-1024-512-128-backfill-nodups/stats.json +1 -0
- train/arxiv/arxiv_0074-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
- train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
- train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
- train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
- train/dclm-sampled/split_10829-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
- train/dclm-sampled/split_10829-tokenized-chunked-1024-512-128-backfill-nodups/stats.json +1 -0
- train/dclm-sampled/split_10829-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
- train/dclm-sampled/split_11041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
- train/dclm-sampled/split_11090-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
- train/dclm-sampled/split_12170-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
- train/dclm-sampled/split_12170-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
- train/dclm-sampled/split_12170-tokenized-chunked-1024-512-128-backfill-nodups/stats.json +1 -0
- train/dclm-sampled/split_12170-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
- train/dclm-sampled/split_12827-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
- train/dclm-sampled/split_13873-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
- train/dclm-sampled/split_13873-tokenized-chunked-1024-512-128-backfill-nodups/stats.json +1 -0
- train/dclm-sampled/split_13873-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
- train/dclm-sampled/split_14544-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
- train/dclm-sampled/split_14544-tokenized-chunked-1024-512-128-backfill-nodups/stats.json +1 -0
- train/dclm-sampled/split_14544-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
- train/dclm-sampled/split_15769-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
- train/dclm-sampled/split_16711-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
- train/dclm-sampled/split_16711-tokenized-chunked-1024-512-128-backfill-nodups/stats.json +1 -0
- train/dclm-sampled/split_16711-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
- train/dclm-sampled/split_1772-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
- train/dclm-sampled/split_19966-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
- train/dclm-sampled/split_22386-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
- train/dclm-sampled/split_22386-tokenized-chunked-1024-512-128-backfill-nodups/stats.json +1 -0
- train/dclm-sampled/split_22386-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
- train/dclm-sampled/split_24810-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
- train/dclm-sampled/split_25345-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
- train/dclm-sampled/split_26004-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
- train/dclm-sampled/split_26004-tokenized-chunked-1024-512-128-backfill-nodups/stats.json +1 -0
- train/dclm-sampled/split_26004-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
- train/dclm-sampled/split_26479-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
- train/dclm-sampled/split_26479-tokenized-chunked-1024-512-128-backfill-nodups/stats.json +1 -0
- train/dclm-sampled/split_26479-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
- train/dclm-sampled/split_28558-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
- train/dclm-sampled/split_28558-tokenized-chunked-1024-512-128-backfill-nodups/stats.json +1 -0
- train/dclm-sampled/split_28558-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
- train/dclm-sampled/split_28560-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
- train/dclm-sampled/split_28767-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
- train/dclm-sampled/split_28767-tokenized-chunked-1024-512-128-backfill-nodups/stats.json +1 -0
- train/dclm-sampled/split_28767-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
- train/dclm-sampled/split_2903-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
- train/dclm-sampled/split_2903-tokenized-chunked-1024-512-128-backfill-nodups/stats.json +1 -0
- train/dclm-sampled/split_2903-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json +0 -0
train/arxiv/arxiv_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5b012ecbd96ab5c618e5a568fe025381f2efa2a57e238e4da36b0bd53b116c26
|
| 3 |
+
size 67107651
|
train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/arxiv/arxiv_0074-tokenized-chunked-1024-512-128-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 274999392, "total_tokens_skipped": 119827, "percentiles": {"0th": 3, "10th": 1022, "20th": 1023, "30th": 1023, "40th": 1023, "50th": 1023, "60th": 1023, "70th": 1023, "80th": 1023, "90th": 1023, "95th": 1023, "99th": 1023, "100th": 1023}}
|
train/arxiv/arxiv_0074-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:61af8f752bbe5867643438c012efea14de4433e0115161ea14229290fdb7e35e
|
| 3 |
+
size 67106796
|
train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:36f34ee054b377958346fba9179ee157d36ef020c6912d825fb27a4db68066c5
|
| 3 |
+
size 67107815
|
train/arxiv/arxiv_0086-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1c5443c3bfbdb27699cbf66799fcf98b249a03a52567993e1eda3504afaa2596
|
| 3 |
+
size 67108776
|
train/dclm-sampled/split_10829-tokenized-chunked-1024-512-128-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107416, "hashes": {}}, "samples": 43142, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48254791, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 20360335, "hashes": {}}, "samples": 12937, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 14587152, "hashes": {}}}], "version": 2}
|
train/dclm-sampled/split_10829-tokenized-chunked-1024-512-128-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 41998121, "total_tokens_skipped": 123448, "percentiles": {"0th": 31, "10th": 248, "20th": 423, "30th": 575, "40th": 718, "50th": 878, "60th": 1022, "70th": 1022, "80th": 1023, "90th": 1023, "95th": 1023, "99th": 1023, "100th": 1023}}
|
train/dclm-sampled/split_10829-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/dclm-sampled/split_11041-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3054e1925c249acd8ed7e0ee0f11252c9fc9933aff0f3120d23a675e722c328d
|
| 3 |
+
size 67107556
|
train/dclm-sampled/split_11090-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:258d92add3576a94612f87ee8d63e46de96c98d6c5498bcc773f65a323274b14
|
| 3 |
+
size 67108157
|
train/dclm-sampled/split_12170-tokenized-chunked-1024-512-128-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108496, "hashes": {}}, "samples": 44525, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48903460, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 9480988, "hashes": {}}, "samples": 6553, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6942417, "hashes": {}}}], "version": 2}
|
train/dclm-sampled/split_12170-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7c9f37ed76d116cf38b5eac08ff84eb83d0536aeebe94f6cd10c6a433f6c4776
|
| 3 |
+
size 9480988
|
train/dclm-sampled/split_12170-tokenized-chunked-1024-512-128-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 36714152, "total_tokens_skipped": 115565, "percentiles": {"0th": 24, "10th": 225, "20th": 381, "30th": 529, "40th": 657, "50th": 801, "60th": 965, "70th": 1022, "80th": 1023, "90th": 1023, "95th": 1023, "99th": 1023, "100th": 1023}}
|
train/dclm-sampled/split_12170-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/dclm-sampled/split_12827-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e2519ff6c38183b0e60c1d766f70f2f4287764e26858a4d4c567b3cbe74b23a1
|
| 3 |
+
size 4898666
|
train/dclm-sampled/split_13873-tokenized-chunked-1024-512-128-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108582, "hashes": {}}, "samples": 44382, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48690975, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 10815690, "hashes": {}}, "samples": 7265, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7851439, "hashes": {}}}], "version": 2}
|
train/dclm-sampled/split_13873-tokenized-chunked-1024-512-128-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 37363910, "total_tokens_skipped": 120001, "percentiles": {"0th": 31, "10th": 224, "20th": 393, "30th": 541, "40th": 667, "50th": 810, "60th": 972, "70th": 1022, "80th": 1023, "90th": 1023, "95th": 1023, "99th": 1023, "100th": 1023}}
|
train/dclm-sampled/split_13873-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/dclm-sampled/split_14544-tokenized-chunked-1024-512-128-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108465, "hashes": {}}, "samples": 43234, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48086468, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 18646119, "hashes": {}}, "samples": 12012, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 13541117, "hashes": {}}}], "version": 2}
|
train/dclm-sampled/split_14544-tokenized-chunked-1024-512-128-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 41167376, "total_tokens_skipped": 126062, "percentiles": {"0th": 24, "10th": 245, "20th": 418, "30th": 569, "40th": 710, "50th": 866, "60th": 1022, "70th": 1022, "80th": 1023, "90th": 1023, "95th": 1023, "99th": 1023, "100th": 1023}}
|
train/dclm-sampled/split_14544-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/dclm-sampled/split_15769-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:911a73726cd132e4a964f7e3abcc8277b4482142b7a028cf8d109972fe5f1e5f
|
| 3 |
+
size 18728121
|
train/dclm-sampled/split_16711-tokenized-chunked-1024-512-128-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108235, "hashes": {}}, "samples": 44239, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48001707, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 13625030, "hashes": {}}, "samples": 8879, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9723619, "hashes": {}}}], "version": 2}
|
train/dclm-sampled/split_16711-tokenized-chunked-1024-512-128-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 38722854, "total_tokens_skipped": 117593, "percentiles": {"0th": 22, "10th": 229, "20th": 386, "30th": 540, "40th": 677, "50th": 833, "60th": 1003, "70th": 1022, "80th": 1023, "90th": 1023, "95th": 1023, "99th": 1023, "100th": 1023}}
|
train/dclm-sampled/split_16711-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/dclm-sampled/split_1772-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:92c5577994a93b8049fbd90cb0bbba1c135d35259c1096079c6d336e17fc8c42
|
| 3 |
+
size 67107624
|
train/dclm-sampled/split_19966-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:25214991627f1539ae783e759308fcf2ff11a550cdfbe746479831a90642d1b8
|
| 3 |
+
size 67106941
|
train/dclm-sampled/split_22386-tokenized-chunked-1024-512-128-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107543, "hashes": {}}, "samples": 43969, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47964796, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 15193626, "hashes": {}}, "samples": 9798, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10883185, "hashes": {}}}], "version": 2}
|
train/dclm-sampled/split_22386-tokenized-chunked-1024-512-128-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 39486546, "total_tokens_skipped": 124418, "percentiles": {"0th": 32, "10th": 239, "20th": 398, "30th": 548, "40th": 686, "50th": 841, "60th": 1016, "70th": 1022, "80th": 1023, "90th": 1023, "95th": 1023, "99th": 1023, "100th": 1023}}
|
train/dclm-sampled/split_22386-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/dclm-sampled/split_24810-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1052d5670c74c60c3b73ac7babda6e9418ca58edc052645b7d94f688b6930e0e
|
| 3 |
+
size 11291511
|
train/dclm-sampled/split_25345-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2269660713875b8a0bf8d2afb8f3c708ea19ba503f8b611022964b1dab9db552
|
| 3 |
+
size 67108532
|
train/dclm-sampled/split_26004-tokenized-chunked-1024-512-128-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108430, "hashes": {}}, "samples": 43721, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48692232, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 14721310, "hashes": {}}, "samples": 9637, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10680762, "hashes": {}}}], "version": 2}
|
train/dclm-sampled/split_26004-tokenized-chunked-1024-512-128-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 39263374, "total_tokens_skipped": 122695, "percentiles": {"0th": 30, "10th": 246, "20th": 407, "30th": 554, "40th": 687, "50th": 835, "60th": 1008, "70th": 1022, "80th": 1023, "90th": 1023, "95th": 1023, "99th": 1023, "100th": 1023}}
|
train/dclm-sampled/split_26004-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/dclm-sampled/split_26479-tokenized-chunked-1024-512-128-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107980, "hashes": {}}, "samples": 42741, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47954010, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 22452927, "hashes": {}}, "samples": 14221, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 16084529, "hashes": {}}}], "version": 2}
|
train/dclm-sampled/split_26479-tokenized-chunked-1024-512-128-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 43017335, "total_tokens_skipped": 132324, "percentiles": {"0th": 18, "10th": 251, "20th": 430, "30th": 587, "40th": 733, "50th": 894, "60th": 1022, "70th": 1022, "80th": 1023, "90th": 1023, "95th": 1023, "99th": 1023, "100th": 1023}}
|
train/dclm-sampled/split_26479-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/dclm-sampled/split_28558-tokenized-chunked-1024-512-128-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107793, "hashes": {}}, "samples": 43030, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48131818, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 20718992, "hashes": {}}, "samples": 13049, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 14774273, "hashes": {}}}], "version": 2}
|
train/dclm-sampled/split_28558-tokenized-chunked-1024-512-128-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 42177588, "total_tokens_skipped": 132397, "percentiles": {"0th": 27, "10th": 252, "20th": 428, "30th": 579, "40th": 724, "50th": 882, "60th": 1022, "70th": 1022, "80th": 1023, "90th": 1023, "95th": 1023, "99th": 1023, "100th": 1023}}
|
train/dclm-sampled/split_28558-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/dclm-sampled/split_28560-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:44bfac37faec9f69185b35353e96da11c56d5369ecf9ba29b6dc02eaf2c22a53
|
| 3 |
+
size 67107044
|
train/dclm-sampled/split_28767-tokenized-chunked-1024-512-128-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107031, "hashes": {}}, "samples": 44829, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48490880, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 8470107, "hashes": {}}, "samples": 5659, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6191604, "hashes": {}}}], "version": 2}
|
train/dclm-sampled/split_28767-tokenized-chunked-1024-512-128-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 36226217, "total_tokens_skipped": 116164, "percentiles": {"0th": 25, "10th": 224, "20th": 387, "30th": 532, "40th": 655, "50th": 792, "60th": 951, "70th": 1022, "80th": 1023, "90th": 1023, "95th": 1023, "99th": 1023, "100th": 1023}}
|
train/dclm-sampled/split_28767-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
train/dclm-sampled/split_2903-tokenized-chunked-1024-512-128-backfill-nodups/index.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108759, "hashes": {}}, "samples": 44376, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48733705, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 11112862, "hashes": {}}, "samples": 7364, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8010087, "hashes": {}}}], "version": 2}
|
train/dclm-sampled/split_2903-tokenized-chunked-1024-512-128-backfill-nodups/stats.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_duplicated_tokens": 0, "total_tokens_written": 37509681, "total_tokens_skipped": 121449, "percentiles": {"0th": 25, "10th": 226, "20th": 393, "30th": 542, "40th": 669, "50th": 814, "60th": 979, "70th": 1022, "80th": 1023, "90th": 1023, "95th": 1023, "99th": 1023, "100th": 1023}}
|
train/dclm-sampled/split_2903-tokenized-chunked-1024-512-128-backfill-nodups/token_decile.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|