Add files using upload-large-folder tool
Browse files- .gitattributes +5 -0
- babylm_hop_control_10M_seed0/runs/checkpoint-1100/model-00002-of-00002.safetensors +3 -0
- babylm_hop_control_10M_seed0/runs/checkpoint-1200/model-00002-of-00002.safetensors +3 -0
- babylm_hop_control_10M_seed0/runs/checkpoint-1600/rng_state_0.pth +3 -0
- babylm_hop_control_10M_seed0/runs/checkpoint-1600/rng_state_1.pth +3 -0
- babylm_hop_control_10M_seed0/runs/checkpoint-1600/rng_state_2.pth +3 -0
- babylm_hop_control_10M_seed0/runs/checkpoint-1600/tokenizer.json +3 -0
- babylm_hop_control_10M_seed0/runs/checkpoint-1600/training_args.bin +3 -0
- babylm_hop_control_10M_seed0/runs/checkpoint-2100/rng_state_0.pth +3 -0
- babylm_hop_control_10M_seed0/runs/checkpoint-2100/rng_state_1.pth +3 -0
- babylm_hop_control_10M_seed0/runs/checkpoint-2100/rng_state_2.pth +3 -0
- babylm_hop_control_10M_seed0/runs/checkpoint-2100/scheduler.pt +3 -0
- babylm_hop_control_10M_seed0/runs/checkpoint-2100/tokenizer.json +3 -0
- babylm_hop_control_10M_seed0/runs/checkpoint-2100/training_args.bin +3 -0
- babylm_hop_control_10M_seed0/runs/checkpoint-900/model-00002-of-00002.safetensors +3 -0
- babylm_reverse_full_10M_seed0/runs/checkpoint-600/model-00001-of-00002.safetensors +3 -0
- babylm_reverse_full_10M_seed0/runs/checkpoint-700/model-00001-of-00002.safetensors +3 -0
- babylm_reverse_partial_10M_seed0/artifacts/models--meta-llama--Llama-3.2-3B/blobs/4719a04514ec2f060240711b7c33ab21187cac730ecaba3040b7a0fd95a9cefb +3 -0
- babylm_reverse_partial_10M_seed0/artifacts/models--meta-llama--Llama-3.2-3B/blobs/584d8d3e3f82f7964955174dfe5e3b1cf117a9d859f022cfdf7fcb884856e002 +3 -0
- babylm_reverse_partial_10M_seed0/artifacts/models--meta-llama--Llama-3.2-3B/snapshots/13afe5124825b4f3751f836b40dafda64c1ed062/model-00001-of-00002.safetensors +3 -0
- babylm_reverse_partial_10M_seed0/artifacts/models--meta-llama--Llama-3.2-3B/snapshots/13afe5124825b4f3751f836b40dafda64c1ed062/model-00002-of-00002.safetensors +3 -0
- babylm_reverse_partial_10M_seed0/runs/checkpoint-1800/rng_state_0.pth +3 -0
- babylm_reverse_partial_10M_seed0/runs/checkpoint-1800/rng_state_1.pth +3 -0
- babylm_reverse_partial_10M_seed0/runs/checkpoint-1800/rng_state_2.pth +3 -0
- babylm_reverse_partial_10M_seed0/runs/checkpoint-1800/scheduler.pt +3 -0
- babylm_reverse_partial_10M_seed0/runs/checkpoint-1800/tokenizer.json +3 -0
- babylm_reverse_partial_10M_seed0/runs/checkpoint-1800/training_args.bin +3 -0
.gitattributes
CHANGED
|
@@ -92,3 +92,8 @@ babylm_hop_control_10M_seed0/runs/checkpoint-100/tokenizer.json filter=lfs diff=
|
|
| 92 |
babylm_hop_control_10M_seed0/runs/checkpoint-1000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 93 |
babylm_hop_tokens4_10M_seed0/artifacts/models--meta-llama--Llama-3.2-3B/blobs/584d8d3e3f82f7964955174dfe5e3b1cf117a9d859f022cfdf7fcb884856e002 filter=lfs diff=lfs merge=lfs -text
|
| 94 |
babylm_hop_control_10M_seed0/runs/checkpoint-300/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 92 |
babylm_hop_control_10M_seed0/runs/checkpoint-1000/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 93 |
babylm_hop_tokens4_10M_seed0/artifacts/models--meta-llama--Llama-3.2-3B/blobs/584d8d3e3f82f7964955174dfe5e3b1cf117a9d859f022cfdf7fcb884856e002 filter=lfs diff=lfs merge=lfs -text
|
| 94 |
babylm_hop_control_10M_seed0/runs/checkpoint-300/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 95 |
+
babylm_hop_control_10M_seed0/runs/checkpoint-1600/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 96 |
+
babylm_hop_control_10M_seed0/runs/checkpoint-2100/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
| 97 |
+
babylm_reverse_partial_10M_seed0/artifacts/models--meta-llama--Llama-3.2-3B/blobs/4719a04514ec2f060240711b7c33ab21187cac730ecaba3040b7a0fd95a9cefb filter=lfs diff=lfs merge=lfs -text
|
| 98 |
+
babylm_reverse_partial_10M_seed0/artifacts/models--meta-llama--Llama-3.2-3B/blobs/584d8d3e3f82f7964955174dfe5e3b1cf117a9d859f022cfdf7fcb884856e002 filter=lfs diff=lfs merge=lfs -text
|
| 99 |
+
babylm_reverse_partial_10M_seed0/runs/checkpoint-1800/tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
babylm_hop_control_10M_seed0/runs/checkpoint-1100/model-00002-of-00002.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:aa29e7e24b23096726ed849ddc2b26bc1c72085b50cf8744c61b930484b1d542
|
| 3 |
+
size 2247734920
|
babylm_hop_control_10M_seed0/runs/checkpoint-1200/model-00002-of-00002.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:568a5798e56d129d60eb3a71d9502da052cd5ff931cb0bb5ffa122ef3a24a10e
|
| 3 |
+
size 2247734920
|
babylm_hop_control_10M_seed0/runs/checkpoint-1600/rng_state_0.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0123f1b841cbb90ad2cc1a10c38027bc63872dc02a831b18d254a6186f0ffff0
|
| 3 |
+
size 16567
|
babylm_hop_control_10M_seed0/runs/checkpoint-1600/rng_state_1.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:86e82dbca3ee98b4aff4d87464271202c99a718b90530a7719851b8b6a191547
|
| 3 |
+
size 16567
|
babylm_hop_control_10M_seed0/runs/checkpoint-1600/rng_state_2.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fdc0a29e8caac820aa227d61191352d70074e7bc59c8656d34721d88d568ccb2
|
| 3 |
+
size 16567
|
babylm_hop_control_10M_seed0/runs/checkpoint-1600/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5c2119a48316b293375cf214819b7f053f5f12171eb3ee8dccfce26510f91e2c
|
| 3 |
+
size 17210462
|
babylm_hop_control_10M_seed0/runs/checkpoint-1600/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:78b717b63c99e15a7061bfb9ad0d8db9718415d88d6419d3c3dc4217194b1c01
|
| 3 |
+
size 6011
|
babylm_hop_control_10M_seed0/runs/checkpoint-2100/rng_state_0.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e1200c6256bb3c5211e1fa1cfcea6721e1b8871fd3c8953a59bf49a59f67404c
|
| 3 |
+
size 16567
|
babylm_hop_control_10M_seed0/runs/checkpoint-2100/rng_state_1.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4b59dfa224b09069bb1265a6a5cb86c22c34f53669b17ebd9335621f959bf7a4
|
| 3 |
+
size 16567
|
babylm_hop_control_10M_seed0/runs/checkpoint-2100/rng_state_2.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:316a81119e1ab6e1b3e5b9d890ce1d0ca9ea9223094844fdcf8b7d9f0f4bb6c2
|
| 3 |
+
size 16567
|
babylm_hop_control_10M_seed0/runs/checkpoint-2100/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6a49c33a964ee4724c018b0c4d2cc677c6cccdb89a812ee1343b9cc0666541f0
|
| 3 |
+
size 627
|
babylm_hop_control_10M_seed0/runs/checkpoint-2100/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5c2119a48316b293375cf214819b7f053f5f12171eb3ee8dccfce26510f91e2c
|
| 3 |
+
size 17210462
|
babylm_hop_control_10M_seed0/runs/checkpoint-2100/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:78b717b63c99e15a7061bfb9ad0d8db9718415d88d6419d3c3dc4217194b1c01
|
| 3 |
+
size 6011
|
babylm_hop_control_10M_seed0/runs/checkpoint-900/model-00002-of-00002.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:134cdc427bf670954ebaa21e6208f55c41da28b2b5d533a0bf0d3eabe5606972
|
| 3 |
+
size 2247734920
|
babylm_reverse_full_10M_seed0/runs/checkpoint-600/model-00001-of-00002.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c0dbccf8c068198bf743ab3b1cc8669cd40aa0feadc448c358e9924cb2fa7385
|
| 3 |
+
size 4965798912
|
babylm_reverse_full_10M_seed0/runs/checkpoint-700/model-00001-of-00002.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e62cbe03419d534cd8d87e1bc40d48fcd43d32a4a40763b8ef99ee11ef247d68
|
| 3 |
+
size 4965798912
|
babylm_reverse_partial_10M_seed0/artifacts/models--meta-llama--Llama-3.2-3B/blobs/4719a04514ec2f060240711b7c33ab21187cac730ecaba3040b7a0fd95a9cefb
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4719a04514ec2f060240711b7c33ab21187cac730ecaba3040b7a0fd95a9cefb
|
| 3 |
+
size 1459729952
|
babylm_reverse_partial_10M_seed0/artifacts/models--meta-llama--Llama-3.2-3B/blobs/584d8d3e3f82f7964955174dfe5e3b1cf117a9d859f022cfdf7fcb884856e002
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:584d8d3e3f82f7964955174dfe5e3b1cf117a9d859f022cfdf7fcb884856e002
|
| 3 |
+
size 4965799096
|
babylm_reverse_partial_10M_seed0/artifacts/models--meta-llama--Llama-3.2-3B/snapshots/13afe5124825b4f3751f836b40dafda64c1ed062/model-00001-of-00002.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:584d8d3e3f82f7964955174dfe5e3b1cf117a9d859f022cfdf7fcb884856e002
|
| 3 |
+
size 4965799096
|
babylm_reverse_partial_10M_seed0/artifacts/models--meta-llama--Llama-3.2-3B/snapshots/13afe5124825b4f3751f836b40dafda64c1ed062/model-00002-of-00002.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4719a04514ec2f060240711b7c33ab21187cac730ecaba3040b7a0fd95a9cefb
|
| 3 |
+
size 1459729952
|
babylm_reverse_partial_10M_seed0/runs/checkpoint-1800/rng_state_0.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:597a8ecb0ce0f65c4e7449874d02a6403185578c168ac3acc602a62fc7292436
|
| 3 |
+
size 16567
|
babylm_reverse_partial_10M_seed0/runs/checkpoint-1800/rng_state_1.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a486bd24422fab4e1b364f49b85424fa3c6f3473311e58e9ed868d1b2de3fda0
|
| 3 |
+
size 16567
|
babylm_reverse_partial_10M_seed0/runs/checkpoint-1800/rng_state_2.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7e4f2cf28f7ae5e3c3cab1c627b13348703f61bdd28206c064ffa51599199c18
|
| 3 |
+
size 16567
|
babylm_reverse_partial_10M_seed0/runs/checkpoint-1800/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e2e8d7589fd595ca926efd5a06b03cac6aa101022f6c4d8ba499fff770a4c8c8
|
| 3 |
+
size 627
|
babylm_reverse_partial_10M_seed0/runs/checkpoint-1800/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6fba7b84ac8c089d417b794dfc0527040604b90e2cdfe6e9df5b55afe9eab61a
|
| 3 |
+
size 17210282
|
babylm_reverse_partial_10M_seed0/runs/checkpoint-1800/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:36cfed0422f99f2449cf26f92edc99e3943bd57b7b5a4d7721a570513f8f8678
|
| 3 |
+
size 6011
|