aastha6 commited on
Commit
5d254c6
·
1 Parent(s): 5a9a0be

Auto Commit

Browse files
Files changed (8) hide show
  1. 0.pt +3 -0
  2. 0_cfg.json +1 -0
  3. 1.pt +3 -0
  4. 1_cfg.json +1 -0
  5. 2.pt +3 -0
  6. 2_cfg.json +1 -0
  7. 3.pt +3 -0
  8. 3_cfg.json +1 -0
0.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f96cbf1dabaf279e5496ea539cc7d1ea051037c6dca25872659210788851af7
3
+ size 604065440
0_cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"seed": 49, "batch_size": 4096, "buffer_mult": 128, "lr": 5e-05, "num_tokens": 400000000, "l1_coeff": 2, "beta1": 0.9, "beta2": 0.999, "d_in": 2304, "dict_size": 16384, "seq_len": 1024, "enc_dtype": "fp32", "model_name": "gemma-2-2b", "site": "resid_pre", "device": "cuda:0", "model_batch_size": 4, "log_every": 100, "save_every": 30000, "dec_init_norm": 0.08, "hook_point": "blocks.14.hook_resid_pre", "wandb_project": "crosscoders-r1-distill-qwen", "wandb_entity": "varmaaastha6"}
1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c30b3e55d5fb2fb1c472a5ff0da1416a048ab880149d8f61149a6198db15898a
3
+ size 604065440
1_cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"seed": 49, "batch_size": 4096, "buffer_mult": 128, "lr": 5e-05, "num_tokens": 400000000, "l1_coeff": 2, "beta1": 0.9, "beta2": 0.999, "d_in": 2304, "dict_size": 16384, "seq_len": 1024, "enc_dtype": "fp32", "model_name": "gemma-2-2b", "site": "resid_pre", "device": "cuda:0", "model_batch_size": 4, "log_every": 100, "save_every": 30000, "dec_init_norm": 0.08, "hook_point": "blocks.14.hook_resid_pre", "wandb_project": "crosscoders-r1-distill-qwen", "wandb_entity": "varmaaastha6"}
2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e6d426e6bbca8651369a3912c1cac9a78131daf7457e7c3e61b0f8283a92459
3
+ size 604065440
2_cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"seed": 49, "batch_size": 4096, "buffer_mult": 128, "lr": 5e-05, "num_tokens": 400000000, "l1_coeff": 2, "beta1": 0.9, "beta2": 0.999, "d_in": 2304, "dict_size": 16384, "seq_len": 1024, "enc_dtype": "fp32", "model_name": "gemma-2-2b", "site": "resid_pre", "device": "cuda:0", "model_batch_size": 4, "log_every": 100, "save_every": 30000, "dec_init_norm": 0.08, "hook_point": "blocks.14.hook_resid_pre", "wandb_project": "crosscoders-r1-distill-qwen", "wandb_entity": "varmaaastha6"}
3.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:770763f6ca64282776bf143d27de95ce7899e2ac4ffdfd589f29881b0eb079d9
3
+ size 604065440
3_cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"seed": 49, "batch_size": 4096, "buffer_mult": 128, "lr": 5e-05, "num_tokens": 400000000, "l1_coeff": 2, "beta1": 0.9, "beta2": 0.999, "d_in": 2304, "dict_size": 16384, "seq_len": 1024, "enc_dtype": "fp32", "model_name": "gemma-2-2b", "site": "resid_pre", "device": "cuda:0", "model_batch_size": 4, "log_every": 100, "save_every": 30000, "dec_init_norm": 0.08, "hook_point": "blocks.14.hook_resid_pre", "wandb_project": "crosscoders-r1-distill-qwen", "wandb_entity": "varmaaastha6"}