tommyp111 commited on
Commit
74caf40
·
verified ·
1 Parent(s): f98d22d

Upload folder using huggingface_hub

Browse files
version_0/0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba3999dd470838cea8f443d9a97497e5cffa640224fa5f62ad099b4ec3a3c875
3
+ size 604065440
version_0/0_cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"seed": 49, "batch_size": 4096, "buffer_mult": 128, "lr": 5e-05, "num_tokens": 400000000, "l1_coeff": 2, "beta1": 0.9, "beta2": 0.999, "d_in": 2304, "dict_size": 16384, "seq_len": 512, "enc_dtype": "float32", "model_name": "gemma-2-2b", "site": "resid_pre", "device_A": "cuda:0", "device_B": "cuda:1", "device_sae": "cuda:2", "dtype": "bfloat16", "model_batch_size": 32, "log_every": 100, "save_every": 30000, "dec_init_norm": 0.08, "hook_point": "blocks.14.hook_resid_pre", "wandb_project": "golden-gate-clip-lora", "wandb_entity": "tompollak", "dump_dir": "./checkpoints"}
version_1/0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72524ab0cfe0ff8df913359dd4924169ee7b76cf2848858973a23c80dfb62f8f
3
+ size 604065440
version_1/0_cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"seed": 49, "batch_size": 4096, "buffer_mult": 128, "lr": 5e-05, "num_tokens": 400000000, "l1_coeff": 2, "beta1": 0.9, "beta2": 0.999, "d_in": 2304, "dict_size": 16384, "seq_len": 512, "enc_dtype": "float32", "model_name": "gemma-2-2b", "site": "resid_pre", "device_A": "cuda:0", "device_B": "cuda:1", "device_sae": "cuda:2", "dtype": "bfloat16", "model_batch_size": 32, "log_every": 100, "save_every": 30000, "dec_init_norm": 0.08, "hook_point": "blocks.14.hook_resid_pre", "wandb_project": "golden-gate-clip-lora", "wandb_entity": "tompollak", "dump_dir": "./checkpoints"}
version_1/1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9365172c20c7648ac4588ffc050e370c0012c11eee6be6dac056ad4034c32d62
3
+ size 604065440
version_1/1_cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"seed": 49, "batch_size": 4096, "buffer_mult": 128, "lr": 5e-05, "num_tokens": 400000000, "l1_coeff": 2, "beta1": 0.9, "beta2": 0.999, "d_in": 2304, "dict_size": 16384, "seq_len": 512, "enc_dtype": "float32", "model_name": "gemma-2-2b", "site": "resid_pre", "device_A": "cuda:0", "device_B": "cuda:1", "device_sae": "cuda:2", "dtype": "bfloat16", "model_batch_size": 32, "log_every": 100, "save_every": 30000, "dec_init_norm": 0.08, "hook_point": "blocks.14.hook_resid_pre", "wandb_project": "golden-gate-clip-lora", "wandb_entity": "tompollak", "dump_dir": "./checkpoints"}
version_1/2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94f4375d004094be84308062fbf0b4e5ec1f0baca2dbc35795e6a27d174a80af
3
+ size 604065440
version_1/2_cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"seed": 49, "batch_size": 4096, "buffer_mult": 128, "lr": 5e-05, "num_tokens": 400000000, "l1_coeff": 2, "beta1": 0.9, "beta2": 0.999, "d_in": 2304, "dict_size": 16384, "seq_len": 512, "enc_dtype": "float32", "model_name": "gemma-2-2b", "site": "resid_pre", "device_A": "cuda:0", "device_B": "cuda:1", "device_sae": "cuda:2", "dtype": "bfloat16", "model_batch_size": 32, "log_every": 100, "save_every": 30000, "dec_init_norm": 0.08, "hook_point": "blocks.14.hook_resid_pre", "wandb_project": "golden-gate-clip-lora", "wandb_entity": "tompollak", "dump_dir": "./checkpoints"}
version_1/3.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f84cc50f9e69b03e37bee2281482d09e425bb80ca9745b797f2b237bfa1fe815
3
+ size 604065440
version_1/3_cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"seed": 49, "batch_size": 4096, "buffer_mult": 128, "lr": 5e-05, "num_tokens": 400000000, "l1_coeff": 2, "beta1": 0.9, "beta2": 0.999, "d_in": 2304, "dict_size": 16384, "seq_len": 512, "enc_dtype": "float32", "model_name": "gemma-2-2b", "site": "resid_pre", "device_A": "cuda:0", "device_B": "cuda:1", "device_sae": "cuda:2", "dtype": "bfloat16", "model_batch_size": 32, "log_every": 100, "save_every": 30000, "dec_init_norm": 0.08, "hook_point": "blocks.14.hook_resid_pre", "wandb_project": "golden-gate-clip-lora", "wandb_entity": "tompollak", "dump_dir": "./checkpoints"}