Upload folder using huggingface_hub
Browse files- version_0/0.pt +3 -0
- version_0/0_cfg.json +1 -0
- version_1/0.pt +3 -0
- version_1/0_cfg.json +1 -0
- version_1/1.pt +3 -0
- version_1/1_cfg.json +1 -0
- version_1/2.pt +3 -0
- version_1/2_cfg.json +1 -0
- version_1/3.pt +3 -0
- version_1/3_cfg.json +1 -0
version_0/0.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ba3999dd470838cea8f443d9a97497e5cffa640224fa5f62ad099b4ec3a3c875
|
| 3 |
+
size 604065440
|
version_0/0_cfg.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"seed": 49, "batch_size": 4096, "buffer_mult": 128, "lr": 5e-05, "num_tokens": 400000000, "l1_coeff": 2, "beta1": 0.9, "beta2": 0.999, "d_in": 2304, "dict_size": 16384, "seq_len": 512, "enc_dtype": "float32", "model_name": "gemma-2-2b", "site": "resid_pre", "device_A": "cuda:0", "device_B": "cuda:1", "device_sae": "cuda:2", "dtype": "bfloat16", "model_batch_size": 32, "log_every": 100, "save_every": 30000, "dec_init_norm": 0.08, "hook_point": "blocks.14.hook_resid_pre", "wandb_project": "golden-gate-clip-lora", "wandb_entity": "tompollak", "dump_dir": "./checkpoints"}
|
version_1/0.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:72524ab0cfe0ff8df913359dd4924169ee7b76cf2848858973a23c80dfb62f8f
|
| 3 |
+
size 604065440
|
version_1/0_cfg.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"seed": 49, "batch_size": 4096, "buffer_mult": 128, "lr": 5e-05, "num_tokens": 400000000, "l1_coeff": 2, "beta1": 0.9, "beta2": 0.999, "d_in": 2304, "dict_size": 16384, "seq_len": 512, "enc_dtype": "float32", "model_name": "gemma-2-2b", "site": "resid_pre", "device_A": "cuda:0", "device_B": "cuda:1", "device_sae": "cuda:2", "dtype": "bfloat16", "model_batch_size": 32, "log_every": 100, "save_every": 30000, "dec_init_norm": 0.08, "hook_point": "blocks.14.hook_resid_pre", "wandb_project": "golden-gate-clip-lora", "wandb_entity": "tompollak", "dump_dir": "./checkpoints"}
|
version_1/1.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9365172c20c7648ac4588ffc050e370c0012c11eee6be6dac056ad4034c32d62
|
| 3 |
+
size 604065440
|
version_1/1_cfg.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"seed": 49, "batch_size": 4096, "buffer_mult": 128, "lr": 5e-05, "num_tokens": 400000000, "l1_coeff": 2, "beta1": 0.9, "beta2": 0.999, "d_in": 2304, "dict_size": 16384, "seq_len": 512, "enc_dtype": "float32", "model_name": "gemma-2-2b", "site": "resid_pre", "device_A": "cuda:0", "device_B": "cuda:1", "device_sae": "cuda:2", "dtype": "bfloat16", "model_batch_size": 32, "log_every": 100, "save_every": 30000, "dec_init_norm": 0.08, "hook_point": "blocks.14.hook_resid_pre", "wandb_project": "golden-gate-clip-lora", "wandb_entity": "tompollak", "dump_dir": "./checkpoints"}
|
version_1/2.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:94f4375d004094be84308062fbf0b4e5ec1f0baca2dbc35795e6a27d174a80af
|
| 3 |
+
size 604065440
|
version_1/2_cfg.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"seed": 49, "batch_size": 4096, "buffer_mult": 128, "lr": 5e-05, "num_tokens": 400000000, "l1_coeff": 2, "beta1": 0.9, "beta2": 0.999, "d_in": 2304, "dict_size": 16384, "seq_len": 512, "enc_dtype": "float32", "model_name": "gemma-2-2b", "site": "resid_pre", "device_A": "cuda:0", "device_B": "cuda:1", "device_sae": "cuda:2", "dtype": "bfloat16", "model_batch_size": 32, "log_every": 100, "save_every": 30000, "dec_init_norm": 0.08, "hook_point": "blocks.14.hook_resid_pre", "wandb_project": "golden-gate-clip-lora", "wandb_entity": "tompollak", "dump_dir": "./checkpoints"}
|
version_1/3.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f84cc50f9e69b03e37bee2281482d09e425bb80ca9745b797f2b237bfa1fe815
|
| 3 |
+
size 604065440
|
version_1/3_cfg.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"seed": 49, "batch_size": 4096, "buffer_mult": 128, "lr": 5e-05, "num_tokens": 400000000, "l1_coeff": 2, "beta1": 0.9, "beta2": 0.999, "d_in": 2304, "dict_size": 16384, "seq_len": 512, "enc_dtype": "float32", "model_name": "gemma-2-2b", "site": "resid_pre", "device_A": "cuda:0", "device_B": "cuda:1", "device_sae": "cuda:2", "dtype": "bfloat16", "model_batch_size": 32, "log_every": 100, "save_every": 30000, "dec_init_norm": 0.08, "hook_point": "blocks.14.hook_resid_pre", "wandb_project": "golden-gate-clip-lora", "wandb_entity": "tompollak", "dump_dir": "./checkpoints"}
|