| { | |
| "name": "InterpBench", | |
| "version": "1.0.0", | |
| "description": "A benchmark of transformers with known circuits for evaluating mechanistic interpretability techniques.", | |
| "license": "https://creativecommons.org/licenses/by/4.0/", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench", | |
| "cases": [ | |
| { | |
| "case_id": "11", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/11", | |
| "task_description": "Counts the number of words in a sequence based on their length.", | |
| "vocab": [ | |
| "J", | |
| "LB", | |
| "TPSI", | |
| "V", | |
| "b", | |
| "no", | |
| "oCLrZaW", | |
| "poiVg" | |
| ], | |
| "max_seq_len": 10, | |
| "min_seq_len": 4, | |
| "files": [ | |
| { | |
| "file_name": "edges.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/edges.pkl" | |
| }, | |
| { | |
| "file_name": "ll_model.pth", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/ll_model.pth" | |
| }, | |
| { | |
| "file_name": "ll_model_cfg.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/ll_model_cfg.pkl" | |
| }, | |
| { | |
| "file_name": "meta.json", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/meta.json" | |
| } | |
| ], | |
| "transformer_cfg": { | |
| "n_layers": 2, | |
| "d_model": 12, | |
| "n_ctx": 10, | |
| "d_head": 3, | |
| "model_name": "custom", | |
| "n_heads": 4, | |
| "d_mlp": 48, | |
| "act_fn": "gelu", | |
| "d_vocab": 10, | |
| "eps": 1e-05, | |
| "use_attn_result": true, | |
| "use_attn_scale": true, | |
| "use_split_qkv_input": true, | |
| "use_hook_mlp_in": true, | |
| "use_attn_in": false, | |
| "use_local_attn": false, | |
| "original_architecture": null, | |
| "from_checkpoint": false, | |
| "checkpoint_index": null, | |
| "checkpoint_label_type": null, | |
| "checkpoint_value": null, | |
| "tokenizer_name": null, | |
| "window_size": null, | |
| "attn_types": null, | |
| "init_mode": "gpt2", | |
| "normalization_type": null, | |
| "n_devices": 1, | |
| "attention_dir": "causal", | |
| "attn_only": false, | |
| "seed": 0, | |
| "initializer_range": 0.1460593486680443, | |
| "init_weights": true, | |
| "scale_attn_by_inverse_layer_idx": false, | |
| "positional_embedding_type": "standard", | |
| "final_rms": false, | |
| "d_vocab_out": 5, | |
| "parallel_attn_mlp": false, | |
| "rotary_dim": null, | |
| "n_params": 3456, | |
| "use_hook_tokens": false, | |
| "gated_mlp": false, | |
| "default_prepend_bos": true, | |
| "dtype": "torch.float32", | |
| "tokenizer_prepends_bos": null, | |
| "n_key_value_heads": null, | |
| "post_embedding_ln": false, | |
| "rotary_base": 10000, | |
| "trust_remote_code": false, | |
| "rotary_adjacent_pairs": false | |
| }, | |
| "transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/ll_model_cfg.pkl", | |
| "training_args": { | |
| "atol": 0.05, | |
| "lr": 0.01, | |
| "use_single_loss": false, | |
| "iit_weight": 1.0, | |
| "behavior_weight": 1.0, | |
| "strict_weight": 0.4, | |
| "epochs": 500, | |
| "act_fn": "gelu", | |
| "clip_grad_norm": 1.0, | |
| "lr_scheduler": "" | |
| }, | |
| "training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/meta.json", | |
| "weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/ll_model.pth", | |
| "circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/11/edges.pkl" | |
| }, | |
| { | |
| "case_id": "13", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/13", | |
| "task_description": "Analyzes the trend (increasing, decreasing, constant) of numeric tokens.", | |
| "vocab": [ | |
| 0, | |
| 1, | |
| 2 | |
| ], | |
| "max_seq_len": 10, | |
| "min_seq_len": 4, | |
| "files": [ | |
| { | |
| "file_name": "edges.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/edges.pkl" | |
| }, | |
| { | |
| "file_name": "ll_model.pth", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/ll_model.pth" | |
| }, | |
| { | |
| "file_name": "ll_model_cfg.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/ll_model_cfg.pkl" | |
| }, | |
| { | |
| "file_name": "meta.json", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/meta.json" | |
| } | |
| ], | |
| "transformer_cfg": { | |
| "n_layers": 2, | |
| "d_model": 20, | |
| "n_ctx": 10, | |
| "d_head": 5, | |
| "model_name": "custom", | |
| "n_heads": 4, | |
| "d_mlp": 80, | |
| "act_fn": "gelu", | |
| "d_vocab": 5, | |
| "eps": 1e-05, | |
| "use_attn_result": true, | |
| "use_attn_scale": true, | |
| "use_split_qkv_input": true, | |
| "use_hook_mlp_in": true, | |
| "use_attn_in": false, | |
| "use_local_attn": false, | |
| "original_architecture": null, | |
| "from_checkpoint": false, | |
| "checkpoint_index": null, | |
| "checkpoint_label_type": null, | |
| "checkpoint_value": null, | |
| "tokenizer_name": null, | |
| "window_size": null, | |
| "attn_types": null, | |
| "init_mode": "gpt2", | |
| "normalization_type": null, | |
| "n_devices": 1, | |
| "attention_dir": "bidirectional", | |
| "attn_only": false, | |
| "seed": 0, | |
| "initializer_range": 0.1460593486680443, | |
| "init_weights": true, | |
| "scale_attn_by_inverse_layer_idx": false, | |
| "positional_embedding_type": "standard", | |
| "final_rms": false, | |
| "d_vocab_out": 3, | |
| "parallel_attn_mlp": false, | |
| "rotary_dim": null, | |
| "n_params": 9600, | |
| "use_hook_tokens": false, | |
| "gated_mlp": false, | |
| "default_prepend_bos": true, | |
| "dtype": "torch.float32", | |
| "tokenizer_prepends_bos": null, | |
| "n_key_value_heads": null, | |
| "post_embedding_ln": false, | |
| "rotary_base": 10000, | |
| "trust_remote_code": false, | |
| "rotary_adjacent_pairs": false | |
| }, | |
| "transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/ll_model_cfg.pkl", | |
| "training_args": { | |
| "atol": 0.05, | |
| "lr": 0.01, | |
| "use_single_loss": false, | |
| "iit_weight": 1.0, | |
| "behavior_weight": 1.0, | |
| "strict_weight": 0.4, | |
| "epochs": 500, | |
| "act_fn": "gelu", | |
| "clip_grad_norm": 1.0, | |
| "lr_scheduler": "" | |
| }, | |
| "training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/meta.json", | |
| "weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/ll_model.pth", | |
| "circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/13/edges.pkl" | |
| }, | |
| { | |
| "case_id": "18", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/18", | |
| "task_description": "Classify each token based on its frequency as 'rare', 'common', or 'frequent'.", | |
| "vocab": [ | |
| "a", | |
| "b", | |
| "c", | |
| "d", | |
| "e" | |
| ], | |
| "max_seq_len": 10, | |
| "min_seq_len": 4, | |
| "files": [ | |
| { | |
| "file_name": "edges.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/edges.pkl" | |
| }, | |
| { | |
| "file_name": "ll_model.pth", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/ll_model.pth" | |
| }, | |
| { | |
| "file_name": "ll_model_cfg.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/ll_model_cfg.pkl" | |
| }, | |
| { | |
| "file_name": "meta.json", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/meta.json" | |
| } | |
| ], | |
| "transformer_cfg": { | |
| "n_layers": 2, | |
| "d_model": 26, | |
| "n_ctx": 10, | |
| "d_head": 6, | |
| "model_name": "custom", | |
| "n_heads": 4, | |
| "d_mlp": 104, | |
| "act_fn": "gelu", | |
| "d_vocab": 7, | |
| "eps": 1e-05, | |
| "use_attn_result": true, | |
| "use_attn_scale": true, | |
| "use_split_qkv_input": true, | |
| "use_hook_mlp_in": true, | |
| "use_attn_in": false, | |
| "use_local_attn": false, | |
| "original_architecture": null, | |
| "from_checkpoint": false, | |
| "checkpoint_index": null, | |
| "checkpoint_label_type": null, | |
| "checkpoint_value": null, | |
| "tokenizer_name": null, | |
| "window_size": null, | |
| "attn_types": null, | |
| "init_mode": "gpt2", | |
| "normalization_type": null, | |
| "n_devices": 1, | |
| "attention_dir": "bidirectional", | |
| "attn_only": false, | |
| "seed": 0, | |
| "initializer_range": 0.12344267996967354, | |
| "init_weights": true, | |
| "scale_attn_by_inverse_layer_idx": false, | |
| "positional_embedding_type": "standard", | |
| "final_rms": false, | |
| "d_vocab_out": 3, | |
| "parallel_attn_mlp": false, | |
| "rotary_dim": null, | |
| "n_params": 15808, | |
| "use_hook_tokens": false, | |
| "gated_mlp": false, | |
| "default_prepend_bos": true, | |
| "dtype": "torch.float32", | |
| "tokenizer_prepends_bos": null, | |
| "n_key_value_heads": null, | |
| "post_embedding_ln": false, | |
| "rotary_base": 10000, | |
| "trust_remote_code": false, | |
| "rotary_adjacent_pairs": false | |
| }, | |
| "transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/ll_model_cfg.pkl", | |
| "training_args": { | |
| "atol": 0.05, | |
| "lr": 0.001, | |
| "use_single_loss": false, | |
| "iit_weight": 1.0, | |
| "behavior_weight": 1.0, | |
| "strict_weight": 1.0, | |
| "epochs": 2000, | |
| "act_fn": "gelu", | |
| "clip_grad_norm": 0.1, | |
| "lr_scheduler": "", | |
| "model_pair": "strict" | |
| }, | |
| "training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/meta.json", | |
| "weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/ll_model.pth", | |
| "circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/18/edges.pkl" | |
| }, | |
| { | |
| "case_id": "19", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/19", | |
| "task_description": "Removes consecutive duplicate tokens from a sequence.", | |
| "vocab": [ | |
| "a", | |
| "b", | |
| "c" | |
| ], | |
| "max_seq_len": 15, | |
| "min_seq_len": 4, | |
| "files": [ | |
| { | |
| "file_name": "edges.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/edges.pkl" | |
| }, | |
| { | |
| "file_name": "ll_model.pth", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/ll_model.pth" | |
| }, | |
| { | |
| "file_name": "ll_model_cfg.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/ll_model_cfg.pkl" | |
| }, | |
| { | |
| "file_name": "meta.json", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/meta.json" | |
| } | |
| ], | |
| "transformer_cfg": { | |
| "n_layers": 2, | |
| "d_model": 32, | |
| "n_ctx": 15, | |
| "d_head": 8, | |
| "model_name": "custom", | |
| "n_heads": 4, | |
| "d_mlp": 128, | |
| "act_fn": "gelu", | |
| "d_vocab": 5, | |
| "eps": 1e-05, | |
| "use_attn_result": true, | |
| "use_attn_scale": true, | |
| "use_split_qkv_input": true, | |
| "use_hook_mlp_in": true, | |
| "use_attn_in": false, | |
| "use_local_attn": false, | |
| "original_architecture": null, | |
| "from_checkpoint": false, | |
| "checkpoint_index": null, | |
| "checkpoint_label_type": null, | |
| "checkpoint_value": null, | |
| "tokenizer_name": null, | |
| "window_size": null, | |
| "attn_types": null, | |
| "init_mode": "gpt2", | |
| "normalization_type": null, | |
| "n_devices": 1, | |
| "attention_dir": "causal", | |
| "attn_only": false, | |
| "seed": 0, | |
| "initializer_range": 0.15689290811054724, | |
| "init_weights": true, | |
| "scale_attn_by_inverse_layer_idx": false, | |
| "positional_embedding_type": "standard", | |
| "final_rms": false, | |
| "d_vocab_out": 3, | |
| "parallel_attn_mlp": false, | |
| "rotary_dim": null, | |
| "n_params": 24576, | |
| "use_hook_tokens": false, | |
| "gated_mlp": false, | |
| "default_prepend_bos": true, | |
| "dtype": "torch.float32", | |
| "tokenizer_prepends_bos": null, | |
| "n_key_value_heads": null, | |
| "post_embedding_ln": false, | |
| "rotary_base": 10000, | |
| "trust_remote_code": false, | |
| "rotary_adjacent_pairs": false | |
| }, | |
| "transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/ll_model_cfg.pkl", | |
| "training_args": { | |
| "atol": 0.05, | |
| "lr": 0.001, | |
| "use_single_loss": false, | |
| "iit_weight": 1.0, | |
| "behavior_weight": 1.0, | |
| "strict_weight": 0.4, | |
| "epochs": 2000, | |
| "act_fn": "gelu", | |
| "clip_grad_norm": 0.1, | |
| "lr_scheduler": "" | |
| }, | |
| "training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/meta.json", | |
| "weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/ll_model.pth", | |
| "circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/19/edges.pkl" | |
| }, | |
| { | |
| "case_id": "20", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/20", | |
| "task_description": "Detect spam messages based on appearance of spam keywords.", | |
| "vocab": [ | |
| "J", | |
| "LB", | |
| "TPSI", | |
| "V", | |
| "b", | |
| "click", | |
| "no", | |
| "now", | |
| "oCLrZaW", | |
| "offer", | |
| "poiVg", | |
| "spam" | |
| ], | |
| "max_seq_len": 10, | |
| "min_seq_len": 4, | |
| "files": [ | |
| { | |
| "file_name": "edges.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/edges.pkl" | |
| }, | |
| { | |
| "file_name": "ll_model.pth", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/ll_model.pth" | |
| }, | |
| { | |
| "file_name": "ll_model_cfg.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/ll_model_cfg.pkl" | |
| }, | |
| { | |
| "file_name": "meta.json", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/meta.json" | |
| } | |
| ], | |
| "transformer_cfg": { | |
| "n_layers": 2, | |
| "d_model": 13, | |
| "n_ctx": 10, | |
| "d_head": 3, | |
| "model_name": "custom", | |
| "n_heads": 4, | |
| "d_mlp": 52, | |
| "act_fn": "gelu", | |
| "d_vocab": 14, | |
| "eps": 1e-05, | |
| "use_attn_result": true, | |
| "use_attn_scale": true, | |
| "use_split_qkv_input": true, | |
| "use_hook_mlp_in": true, | |
| "use_attn_in": false, | |
| "use_local_attn": false, | |
| "original_architecture": null, | |
| "from_checkpoint": false, | |
| "checkpoint_index": null, | |
| "checkpoint_label_type": null, | |
| "checkpoint_value": null, | |
| "tokenizer_name": null, | |
| "window_size": null, | |
| "attn_types": null, | |
| "init_mode": "gpt2", | |
| "normalization_type": null, | |
| "n_devices": 1, | |
| "attention_dir": "causal", | |
| "attn_only": false, | |
| "seed": 0, | |
| "initializer_range": 0.16, | |
| "init_weights": true, | |
| "scale_attn_by_inverse_layer_idx": false, | |
| "positional_embedding_type": "standard", | |
| "final_rms": false, | |
| "d_vocab_out": 2, | |
| "parallel_attn_mlp": false, | |
| "rotary_dim": null, | |
| "n_params": 3952, | |
| "use_hook_tokens": false, | |
| "gated_mlp": false, | |
| "default_prepend_bos": true, | |
| "dtype": "torch.float32", | |
| "tokenizer_prepends_bos": null, | |
| "n_key_value_heads": null, | |
| "post_embedding_ln": false, | |
| "rotary_base": 10000, | |
| "trust_remote_code": false, | |
| "rotary_adjacent_pairs": false | |
| }, | |
| "transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/ll_model_cfg.pkl", | |
| "training_args": { | |
| "atol": 0.05, | |
| "lr": 0.01, | |
| "use_single_loss": false, | |
| "iit_weight": 1.0, | |
| "behavior_weight": 1.0, | |
| "strict_weight": 0.4, | |
| "epochs": 2000, | |
| "act_fn": "gelu", | |
| "clip_grad_norm": 0.1, | |
| "lr_scheduler": "", | |
| "model_pair": "strict" | |
| }, | |
| "training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/meta.json", | |
| "weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/ll_model.pth", | |
| "circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/20/edges.pkl" | |
| }, | |
| { | |
| "case_id": "21", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/21", | |
| "task_description": "Extract unique tokens from a string", | |
| "vocab": [ | |
| "a", | |
| "b", | |
| "c" | |
| ], | |
| "max_seq_len": 10, | |
| "min_seq_len": 4, | |
| "files": [ | |
| { | |
| "file_name": "edges.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/edges.pkl" | |
| }, | |
| { | |
| "file_name": "ll_model.pth", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/ll_model.pth" | |
| }, | |
| { | |
| "file_name": "ll_model_cfg.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/ll_model_cfg.pkl" | |
| }, | |
| { | |
| "file_name": "meta.json", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/meta.json" | |
| } | |
| ], | |
| "transformer_cfg": { | |
| "n_layers": 4, | |
| "d_model": 50, | |
| "n_ctx": 10, | |
| "d_head": 12, | |
| "model_name": "custom", | |
| "n_heads": 4, | |
| "d_mlp": 200, | |
| "act_fn": "gelu", | |
| "d_vocab": 5, | |
| "eps": 1e-05, | |
| "use_attn_result": true, | |
| "use_attn_scale": true, | |
| "use_split_qkv_input": true, | |
| "use_hook_mlp_in": true, | |
| "use_attn_in": false, | |
| "use_local_attn": false, | |
| "original_architecture": null, | |
| "from_checkpoint": false, | |
| "checkpoint_index": null, | |
| "checkpoint_label_type": null, | |
| "checkpoint_value": null, | |
| "tokenizer_name": null, | |
| "window_size": null, | |
| "attn_types": null, | |
| "init_mode": "gpt2", | |
| "normalization_type": null, | |
| "n_devices": 1, | |
| "attention_dir": "causal", | |
| "attn_only": false, | |
| "seed": 0, | |
| "initializer_range": 0.09847319278346618, | |
| "init_weights": true, | |
| "scale_attn_by_inverse_layer_idx": false, | |
| "positional_embedding_type": "standard", | |
| "final_rms": false, | |
| "d_vocab_out": 3, | |
| "parallel_attn_mlp": false, | |
| "rotary_dim": null, | |
| "n_params": 118400, | |
| "use_hook_tokens": false, | |
| "gated_mlp": false, | |
| "default_prepend_bos": true, | |
| "dtype": "torch.float32", | |
| "tokenizer_prepends_bos": null, | |
| "n_key_value_heads": null, | |
| "post_embedding_ln": false, | |
| "rotary_base": 10000, | |
| "trust_remote_code": false, | |
| "rotary_adjacent_pairs": false | |
| }, | |
| "transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/ll_model_cfg.pkl", | |
| "training_args": { | |
| "atol": 0.05, | |
| "lr": 0.0005, | |
| "use_single_loss": false, | |
| "iit_weight": 1.0, | |
| "behavior_weight": 1.0, | |
| "strict_weight": 0.5, | |
| "epochs": 2000, | |
| "act_fn": "gelu", | |
| "clip_grad_norm": 0.1, | |
| "lr_scheduler": "", | |
| "model_pair": "strict" | |
| }, | |
| "training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/meta.json", | |
| "weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/ll_model.pth", | |
| "circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/21/edges.pkl" | |
| }, | |
| { | |
| "case_id": "26", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/26", | |
| "task_description": "Creates a cascading effect by repeating each token in sequence incrementally.", | |
| "vocab": [ | |
| "a", | |
| "b", | |
| "c" | |
| ], | |
| "max_seq_len": 10, | |
| "min_seq_len": 4, | |
| "files": [ | |
| { | |
| "file_name": "edges.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/edges.pkl" | |
| }, | |
| { | |
| "file_name": "ll_model.pth", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/ll_model.pth" | |
| }, | |
| { | |
| "file_name": "ll_model_cfg.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/ll_model_cfg.pkl" | |
| }, | |
| { | |
| "file_name": "meta.json", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/meta.json" | |
| } | |
| ], | |
| "transformer_cfg": { | |
| "n_layers": 2, | |
| "d_model": 21, | |
| "n_ctx": 10, | |
| "d_head": 5, | |
| "model_name": "custom", | |
| "n_heads": 4, | |
| "d_mlp": 84, | |
| "act_fn": "gelu", | |
| "d_vocab": 5, | |
| "eps": 1e-05, | |
| "use_attn_result": true, | |
| "use_attn_scale": true, | |
| "use_split_qkv_input": true, | |
| "use_hook_mlp_in": true, | |
| "use_attn_in": false, | |
| "use_local_attn": false, | |
| "original_architecture": null, | |
| "from_checkpoint": false, | |
| "checkpoint_index": null, | |
| "checkpoint_label_type": null, | |
| "checkpoint_value": null, | |
| "tokenizer_name": null, | |
| "window_size": null, | |
| "attn_types": null, | |
| "init_mode": "gpt2", | |
| "normalization_type": null, | |
| "n_devices": 1, | |
| "attention_dir": "causal", | |
| "attn_only": false, | |
| "seed": 0, | |
| "initializer_range": 0.12344267996967354, | |
| "init_weights": true, | |
| "scale_attn_by_inverse_layer_idx": false, | |
| "positional_embedding_type": "standard", | |
| "final_rms": false, | |
| "d_vocab_out": 27, | |
| "parallel_attn_mlp": false, | |
| "rotary_dim": null, | |
| "n_params": 10416, | |
| "use_hook_tokens": false, | |
| "gated_mlp": false, | |
| "default_prepend_bos": true, | |
| "dtype": "torch.float32", | |
| "tokenizer_prepends_bos": null, | |
| "n_key_value_heads": null, | |
| "post_embedding_ln": false, | |
| "rotary_base": 10000, | |
| "trust_remote_code": false, | |
| "rotary_adjacent_pairs": false | |
| }, | |
| "transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/ll_model_cfg.pkl", | |
| "training_args": { | |
| "atol": 0.05, | |
| "lr": 0.01, | |
| "use_single_loss": false, | |
| "iit_weight": 1.0, | |
| "behavior_weight": 1.0, | |
| "strict_weight": 0.4, | |
| "epochs": 2000, | |
| "act_fn": "gelu", | |
| "clip_grad_norm": 0.1, | |
| "lr_scheduler": "", | |
| "model_pair": "strict" | |
| }, | |
| "training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/meta.json", | |
| "weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/ll_model.pth", | |
| "circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/26/edges.pkl" | |
| }, | |
| { | |
| "case_id": "29", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/29", | |
| "task_description": "Creates abbreviations for each token in the sequence.", | |
| "vocab": [ | |
| "J", | |
| "LB", | |
| "TPSI", | |
| "V", | |
| "b", | |
| "no", | |
| "oCLrZaW", | |
| "poiVg" | |
| ], | |
| "max_seq_len": 10, | |
| "min_seq_len": 4, | |
| "files": [ | |
| { | |
| "file_name": "edges.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/edges.pkl" | |
| }, | |
| { | |
| "file_name": "ll_model.pth", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/ll_model.pth" | |
| }, | |
| { | |
| "file_name": "ll_model_cfg.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/ll_model_cfg.pkl" | |
| }, | |
| { | |
| "file_name": "meta.json", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/meta.json" | |
| } | |
| ], | |
| "transformer_cfg": { | |
| "n_layers": 2, | |
| "d_model": 13, | |
| "n_ctx": 10, | |
| "d_head": 3, | |
| "model_name": "custom", | |
| "n_heads": 4, | |
| "d_mlp": 52, | |
| "act_fn": "gelu", | |
| "d_vocab": 10, | |
| "eps": 1e-05, | |
| "use_attn_result": true, | |
| "use_attn_scale": true, | |
| "use_split_qkv_input": true, | |
| "use_hook_mlp_in": true, | |
| "use_attn_in": false, | |
| "use_local_attn": false, | |
| "original_architecture": null, | |
| "from_checkpoint": false, | |
| "checkpoint_index": null, | |
| "checkpoint_label_type": null, | |
| "checkpoint_value": null, | |
| "tokenizer_name": null, | |
| "window_size": null, | |
| "attn_types": null, | |
| "init_mode": "gpt2", | |
| "normalization_type": null, | |
| "n_devices": 1, | |
| "attention_dir": "causal", | |
| "attn_only": false, | |
| "seed": 0, | |
| "initializer_range": 0.1539600717839002, | |
| "init_weights": true, | |
| "scale_attn_by_inverse_layer_idx": false, | |
| "positional_embedding_type": "standard", | |
| "final_rms": false, | |
| "d_vocab_out": 8, | |
| "parallel_attn_mlp": false, | |
| "rotary_dim": null, | |
| "n_params": 3952, | |
| "use_hook_tokens": false, | |
| "gated_mlp": false, | |
| "default_prepend_bos": true, | |
| "dtype": "torch.float32", | |
| "tokenizer_prepends_bos": null, | |
| "n_key_value_heads": null, | |
| "post_embedding_ln": false, | |
| "rotary_base": 10000, | |
| "trust_remote_code": false, | |
| "rotary_adjacent_pairs": false | |
| }, | |
| "transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/ll_model_cfg.pkl", | |
| "training_args": { | |
| "atol": 0.05, | |
| "lr": 0.01, | |
| "use_single_loss": false, | |
| "iit_weight": 1.0, | |
| "behavior_weight": 1.0, | |
| "strict_weight": 0.4, | |
| "epochs": 2000, | |
| "act_fn": "gelu", | |
| "clip_grad_norm": 0.1, | |
| "lr_scheduler": "", | |
| "model_pair": "strict" | |
| }, | |
| "training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/meta.json", | |
| "weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/ll_model.pth", | |
| "circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/29/edges.pkl" | |
| }, | |
| { | |
| "case_id": "3", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/3", | |
| "task_description": "Returns the fraction of 'x' in the input up to the i-th position for all i.", | |
| "vocab": [ | |
| "a", | |
| "b", | |
| "c", | |
| "x" | |
| ], | |
| "max_seq_len": 5, | |
| "min_seq_len": 4, | |
| "files": [ | |
| { | |
| "file_name": "edges.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/edges.pkl" | |
| }, | |
| { | |
| "file_name": "ll_model.pth", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/ll_model.pth" | |
| }, | |
| { | |
| "file_name": "ll_model_cfg.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/ll_model_cfg.pkl" | |
| }, | |
| { | |
| "file_name": "meta.json", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/meta.json" | |
| } | |
| ], | |
| "transformer_cfg": { | |
| "n_layers": 2, | |
| "d_model": 12, | |
| "n_ctx": 5, | |
| "d_head": 3, | |
| "model_name": "custom", | |
| "n_heads": 4, | |
| "d_mlp": 48, | |
| "act_fn": "gelu", | |
| "d_vocab": 6, | |
| "eps": 1e-05, | |
| "use_attn_result": true, | |
| "use_attn_scale": true, | |
| "use_split_qkv_input": true, | |
| "use_hook_mlp_in": true, | |
| "use_attn_in": false, | |
| "use_local_attn": false, | |
| "original_architecture": null, | |
| "from_checkpoint": false, | |
| "checkpoint_index": null, | |
| "checkpoint_label_type": null, | |
| "checkpoint_value": null, | |
| "tokenizer_name": null, | |
| "window_size": null, | |
| "attn_types": null, | |
| "init_mode": "gpt2", | |
| "normalization_type": null, | |
| "n_devices": 1, | |
| "attention_dir": "causal", | |
| "attn_only": false, | |
| "seed": 0, | |
| "initializer_range": 0.22188007849009167, | |
| "init_weights": true, | |
| "scale_attn_by_inverse_layer_idx": false, | |
| "positional_embedding_type": "standard", | |
| "final_rms": false, | |
| "d_vocab_out": 1, | |
| "parallel_attn_mlp": false, | |
| "rotary_dim": null, | |
| "n_params": 3456, | |
| "use_hook_tokens": false, | |
| "gated_mlp": false, | |
| "default_prepend_bos": true, | |
| "dtype": "torch.float32", | |
| "tokenizer_prepends_bos": null, | |
| "n_key_value_heads": null, | |
| "post_embedding_ln": false, | |
| "rotary_base": 10000, | |
| "trust_remote_code": false, | |
| "rotary_adjacent_pairs": false | |
| }, | |
| "transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/ll_model_cfg.pkl", | |
| "training_args": { | |
| "atol": 0.05, | |
| "lr": 0.001, | |
| "use_single_loss": false, | |
| "iit_weight": 1.0, | |
| "behavior_weight": 1.0, | |
| "strict_weight": 10.0, | |
| "epochs": 2000, | |
| "act_fn": "gelu", | |
| "clip_grad_norm": 0.1, | |
| "lr_scheduler": "" | |
| }, | |
| "training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/meta.json", | |
| "weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/ll_model.pth", | |
| "circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/3/edges.pkl" | |
| }, | |
| { | |
| "case_id": "33", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/33", | |
| "task_description": "Checks if each token's length is odd or even.", | |
| "vocab": [ | |
| "J", | |
| "LB", | |
| "TPSI", | |
| "V", | |
| "b", | |
| "no", | |
| "oCLrZaW", | |
| "poiVg" | |
| ], | |
| "max_seq_len": 10, | |
| "min_seq_len": 4, | |
| "files": [ | |
| { | |
| "file_name": "edges.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/edges.pkl" | |
| }, | |
| { | |
| "file_name": "ll_model.pth", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/ll_model.pth" | |
| }, | |
| { | |
| "file_name": "ll_model_cfg.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/ll_model_cfg.pkl" | |
| }, | |
| { | |
| "file_name": "meta.json", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/meta.json" | |
| } | |
| ], | |
| "transformer_cfg": { | |
| "n_layers": 2, | |
| "d_model": 4, | |
| "n_ctx": 10, | |
| "d_head": 1, | |
| "model_name": "custom", | |
| "n_heads": 4, | |
| "d_mlp": 16, | |
| "act_fn": "gelu", | |
| "d_vocab": 10, | |
| "eps": 1e-05, | |
| "use_attn_result": true, | |
| "use_attn_scale": true, | |
| "use_split_qkv_input": true, | |
| "use_hook_mlp_in": true, | |
| "use_attn_in": false, | |
| "use_local_attn": false, | |
| "original_architecture": null, | |
| "from_checkpoint": false, | |
| "checkpoint_index": null, | |
| "checkpoint_label_type": null, | |
| "checkpoint_value": null, | |
| "tokenizer_name": null, | |
| "window_size": null, | |
| "attn_types": null, | |
| "init_mode": "gpt2", | |
| "normalization_type": null, | |
| "n_devices": 1, | |
| "attention_dir": "causal", | |
| "attn_only": false, | |
| "seed": 0, | |
| "initializer_range": 0.17457431218879393, | |
| "init_weights": true, | |
| "scale_attn_by_inverse_layer_idx": false, | |
| "positional_embedding_type": "standard", | |
| "final_rms": false, | |
| "d_vocab_out": 2, | |
| "parallel_attn_mlp": false, | |
| "rotary_dim": null, | |
| "n_params": 384, | |
| "use_hook_tokens": false, | |
| "gated_mlp": false, | |
| "default_prepend_bos": true, | |
| "dtype": "torch.float32", | |
| "tokenizer_prepends_bos": null, | |
| "n_key_value_heads": null, | |
| "post_embedding_ln": false, | |
| "rotary_base": 10000, | |
| "trust_remote_code": false, | |
| "rotary_adjacent_pairs": false | |
| }, | |
| "transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/ll_model_cfg.pkl", | |
| "training_args": { | |
| "atol": 0.05, | |
| "lr": 0.001, | |
| "use_single_loss": false, | |
| "iit_weight": 1.0, | |
| "behavior_weight": 1.0, | |
| "strict_weight": 0.4, | |
| "epochs": 2000, | |
| "act_fn": "gelu", | |
| "clip_grad_norm": 0.1, | |
| "lr_scheduler": "" | |
| }, | |
| "training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/meta.json", | |
| "weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/ll_model.pth", | |
| "circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/33/edges.pkl" | |
| }, | |
| { | |
| "case_id": "34", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/34", | |
| "task_description": "Calculate the ratio of vowels to consonants in each word.", | |
| "vocab": [ | |
| "J", | |
| "LB", | |
| "TPSI", | |
| "V", | |
| "b", | |
| "no", | |
| "oCLrZaW", | |
| "poiVg" | |
| ], | |
| "max_seq_len": 10, | |
| "min_seq_len": 4, | |
| "files": [ | |
| { | |
| "file_name": "edges.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/edges.pkl" | |
| }, | |
| { | |
| "file_name": "ll_model.pth", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/ll_model.pth" | |
| }, | |
| { | |
| "file_name": "ll_model_cfg.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/ll_model_cfg.pkl" | |
| }, | |
| { | |
| "file_name": "meta.json", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/meta.json" | |
| } | |
| ], | |
| "transformer_cfg": { | |
| "n_layers": 2, | |
| "d_model": 16, | |
| "n_ctx": 10, | |
| "d_head": 4, | |
| "model_name": "custom", | |
| "n_heads": 4, | |
| "d_mlp": 64, | |
| "act_fn": "gelu", | |
| "d_vocab": 10, | |
| "eps": 1e-05, | |
| "use_attn_result": true, | |
| "use_attn_scale": true, | |
| "use_split_qkv_input": true, | |
| "use_hook_mlp_in": true, | |
| "use_attn_in": false, | |
| "use_local_attn": false, | |
| "original_architecture": null, | |
| "from_checkpoint": false, | |
| "checkpoint_index": null, | |
| "checkpoint_label_type": null, | |
| "checkpoint_value": null, | |
| "tokenizer_name": null, | |
| "window_size": null, | |
| "attn_types": null, | |
| "init_mode": "gpt2", | |
| "normalization_type": null, | |
| "n_devices": 1, | |
| "attention_dir": "causal", | |
| "attn_only": false, | |
| "seed": 0, | |
| "initializer_range": 0.16329931618554522, | |
| "init_weights": true, | |
| "scale_attn_by_inverse_layer_idx": false, | |
| "positional_embedding_type": "standard", | |
| "final_rms": false, | |
| "d_vocab_out": 5, | |
| "parallel_attn_mlp": false, | |
| "rotary_dim": null, | |
| "n_params": 6144, | |
| "use_hook_tokens": false, | |
| "gated_mlp": false, | |
| "default_prepend_bos": true, | |
| "dtype": "torch.float32", | |
| "tokenizer_prepends_bos": null, | |
| "n_key_value_heads": null, | |
| "post_embedding_ln": false, | |
| "rotary_base": 10000, | |
| "trust_remote_code": false, | |
| "rotary_adjacent_pairs": false | |
| }, | |
| "transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/ll_model_cfg.pkl", | |
| "training_args": { | |
| "atol": 0.05, | |
| "lr": 0.01, | |
| "use_single_loss": false, | |
| "iit_weight": 1.0, | |
| "behavior_weight": 1.0, | |
| "strict_weight": 1.0, | |
| "epochs": 2000, | |
| "act_fn": "gelu", | |
| "clip_grad_norm": 0.1, | |
| "lr_scheduler": "", | |
| "model_pair": "strict" | |
| }, | |
| "training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/meta.json", | |
| "weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/ll_model.pth", | |
| "circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/34/edges.pkl" | |
| }, | |
| { | |
| "case_id": "35", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/35", | |
| "task_description": "Alternates capitalization of each character in words.", | |
| "vocab": [ | |
| "J", | |
| "LB", | |
| "TPSI", | |
| "V", | |
| "b", | |
| "no", | |
| "oCLrZaW", | |
| "poiVg" | |
| ], | |
| "max_seq_len": 10, | |
| "min_seq_len": 4, | |
| "files": [ | |
| { | |
| "file_name": "edges.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/edges.pkl" | |
| }, | |
| { | |
| "file_name": "ll_model.pth", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/ll_model.pth" | |
| }, | |
| { | |
| "file_name": "ll_model_cfg.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/ll_model_cfg.pkl" | |
| }, | |
| { | |
| "file_name": "meta.json", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/meta.json" | |
| } | |
| ], | |
| "transformer_cfg": { | |
| "n_layers": 2, | |
| "d_model": 9, | |
| "n_ctx": 10, | |
| "d_head": 2, | |
| "model_name": "custom", | |
| "n_heads": 4, | |
| "d_mlp": 36, | |
| "act_fn": "gelu", | |
| "d_vocab": 10, | |
| "eps": 1e-05, | |
| "use_attn_result": true, | |
| "use_attn_scale": true, | |
| "use_split_qkv_input": true, | |
| "use_hook_mlp_in": true, | |
| "use_attn_in": false, | |
| "use_local_attn": false, | |
| "original_architecture": null, | |
| "from_checkpoint": false, | |
| "checkpoint_index": null, | |
| "checkpoint_label_type": null, | |
| "checkpoint_value": null, | |
| "tokenizer_name": null, | |
| "window_size": null, | |
| "attn_types": null, | |
| "init_mode": "gpt2", | |
| "normalization_type": null, | |
| "n_devices": 1, | |
| "attention_dir": "causal", | |
| "attn_only": false, | |
| "seed": 0, | |
| "initializer_range": 0.1539600717839002, | |
| "init_weights": true, | |
| "scale_attn_by_inverse_layer_idx": false, | |
| "positional_embedding_type": "standard", | |
| "final_rms": false, | |
| "d_vocab_out": 8, | |
| "parallel_attn_mlp": false, | |
| "rotary_dim": null, | |
| "n_params": 1872, | |
| "use_hook_tokens": false, | |
| "gated_mlp": false, | |
| "default_prepend_bos": true, | |
| "dtype": "torch.float32", | |
| "tokenizer_prepends_bos": null, | |
| "n_key_value_heads": null, | |
| "post_embedding_ln": false, | |
| "rotary_base": 10000, | |
| "trust_remote_code": false, | |
| "rotary_adjacent_pairs": false | |
| }, | |
| "transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/ll_model_cfg.pkl", | |
| "training_args": { | |
| "atol": 0.05, | |
| "lr": 0.01, | |
| "use_single_loss": false, | |
| "iit_weight": 1.0, | |
| "behavior_weight": 1.0, | |
| "strict_weight": 1.0, | |
| "epochs": 2000, | |
| "act_fn": "gelu", | |
| "clip_grad_norm": 0.1, | |
| "lr_scheduler": "", | |
| "model_pair": "strict" | |
| }, | |
| "training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/meta.json", | |
| "weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/ll_model.pth", | |
| "circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/35/edges.pkl" | |
| }, | |
| { | |
| "case_id": "36", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/36", | |
| "task_description": "Classifies each token as 'positive', 'negative', or 'neutral' based on emojis.", | |
| "vocab": [ | |
| "\ud83d\udcd8", | |
| "\ud83d\ude0a", | |
| "\ud83d\ude22" | |
| ], | |
| "max_seq_len": 10, | |
| "min_seq_len": 4, | |
| "files": [ | |
| { | |
| "file_name": "edges.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/edges.pkl" | |
| }, | |
| { | |
| "file_name": "ll_model.pth", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/ll_model.pth" | |
| }, | |
| { | |
| "file_name": "ll_model_cfg.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/ll_model_cfg.pkl" | |
| }, | |
| { | |
| "file_name": "meta.json", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/meta.json" | |
| } | |
| ], | |
| "transformer_cfg": { | |
| "n_layers": 2, | |
| "d_model": 6, | |
| "n_ctx": 10, | |
| "d_head": 1, | |
| "model_name": "custom", | |
| "n_heads": 4, | |
| "d_mlp": 24, | |
| "act_fn": "gelu", | |
| "d_vocab": 5, | |
| "eps": 1e-05, | |
| "use_attn_result": true, | |
| "use_attn_scale": true, | |
| "use_split_qkv_input": true, | |
| "use_hook_mlp_in": true, | |
| "use_attn_in": false, | |
| "use_local_attn": false, | |
| "original_architecture": null, | |
| "from_checkpoint": false, | |
| "checkpoint_index": null, | |
| "checkpoint_label_type": null, | |
| "checkpoint_value": null, | |
| "tokenizer_name": null, | |
| "window_size": null, | |
| "attn_types": null, | |
| "init_mode": "gpt2", | |
| "normalization_type": null, | |
| "n_devices": 1, | |
| "attention_dir": "causal", | |
| "attn_only": false, | |
| "seed": 0, | |
| "initializer_range": 0.19402850002906638, | |
| "init_weights": true, | |
| "scale_attn_by_inverse_layer_idx": false, | |
| "positional_embedding_type": "standard", | |
| "final_rms": false, | |
| "d_vocab_out": 3, | |
| "parallel_attn_mlp": false, | |
| "rotary_dim": null, | |
| "n_params": 768, | |
| "use_hook_tokens": false, | |
| "gated_mlp": false, | |
| "default_prepend_bos": true, | |
| "dtype": "torch.float32", | |
| "tokenizer_prepends_bos": null, | |
| "n_key_value_heads": null, | |
| "post_embedding_ln": false, | |
| "rotary_base": 10000, | |
| "trust_remote_code": false, | |
| "rotary_adjacent_pairs": false | |
| }, | |
| "transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/ll_model_cfg.pkl", | |
| "training_args": { | |
| "atol": 0.05, | |
| "lr": 0.01, | |
| "use_single_loss": false, | |
| "iit_weight": 1.0, | |
| "behavior_weight": 1.0, | |
| "strict_weight": 1.0, | |
| "epochs": 2000, | |
| "act_fn": "gelu", | |
| "clip_grad_norm": 0.1, | |
| "lr_scheduler": "", | |
| "model_pair": "strict" | |
| }, | |
| "training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/meta.json", | |
| "weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/ll_model.pth", | |
| "circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/36/edges.pkl" | |
| }, | |
| { | |
| "case_id": "37", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/37", | |
| "task_description": "Reverses each word in the sequence except for specified exclusions.", | |
| "vocab": [ | |
| "J", | |
| "LB", | |
| "TPSI", | |
| "V", | |
| "b", | |
| "no", | |
| "oCLrZaW", | |
| "poiVg" | |
| ], | |
| "max_seq_len": 10, | |
| "min_seq_len": 4, | |
| "files": [ | |
| { | |
| "file_name": "edges.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/edges.pkl" | |
| }, | |
| { | |
| "file_name": "ll_model.pth", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/ll_model.pth" | |
| }, | |
| { | |
| "file_name": "ll_model_cfg.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/ll_model_cfg.pkl" | |
| }, | |
| { | |
| "file_name": "meta.json", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/meta.json" | |
| } | |
| ], | |
| "transformer_cfg": { | |
| "n_layers": 2, | |
| "d_model": 12, | |
| "n_ctx": 10, | |
| "d_head": 3, | |
| "model_name": "custom", | |
| "n_heads": 4, | |
| "d_mlp": 48, | |
| "act_fn": "gelu", | |
| "d_vocab": 10, | |
| "eps": 1e-05, | |
| "use_attn_result": true, | |
| "use_attn_scale": true, | |
| "use_split_qkv_input": true, | |
| "use_hook_mlp_in": true, | |
| "use_attn_in": false, | |
| "use_local_attn": false, | |
| "original_architecture": null, | |
| "from_checkpoint": false, | |
| "checkpoint_index": null, | |
| "checkpoint_label_type": null, | |
| "checkpoint_value": null, | |
| "tokenizer_name": null, | |
| "window_size": null, | |
| "attn_types": null, | |
| "init_mode": "gpt2", | |
| "normalization_type": null, | |
| "n_devices": 1, | |
| "attention_dir": "causal", | |
| "attn_only": false, | |
| "seed": 0, | |
| "initializer_range": 0.1539600717839002, | |
| "init_weights": true, | |
| "scale_attn_by_inverse_layer_idx": false, | |
| "positional_embedding_type": "standard", | |
| "final_rms": false, | |
| "d_vocab_out": 8, | |
| "parallel_attn_mlp": false, | |
| "rotary_dim": null, | |
| "n_params": 3456, | |
| "use_hook_tokens": false, | |
| "gated_mlp": false, | |
| "default_prepend_bos": true, | |
| "dtype": "torch.float32", | |
| "tokenizer_prepends_bos": null, | |
| "n_key_value_heads": null, | |
| "post_embedding_ln": false, | |
| "rotary_base": 10000, | |
| "trust_remote_code": false, | |
| "rotary_adjacent_pairs": false | |
| }, | |
| "transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/ll_model_cfg.pkl", | |
| "training_args": { | |
| "atol": 0.05, | |
| "lr": 0.01, | |
| "use_single_loss": false, | |
| "iit_weight": 1.0, | |
| "behavior_weight": 1.0, | |
| "strict_weight": 1.0, | |
| "epochs": 2000, | |
| "act_fn": "gelu", | |
| "clip_grad_norm": 0.1, | |
| "lr_scheduler": "", | |
| "model_pair": "strict" | |
| }, | |
| "training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/meta.json", | |
| "weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/ll_model.pth", | |
| "circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/37/edges.pkl" | |
| }, | |
| { | |
| "case_id": "4", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/4", | |
| "task_description": "Return fraction of previous open tokens minus the fraction of close tokens.", | |
| "vocab": [ | |
| "(", | |
| ")", | |
| "a", | |
| "b", | |
| "c" | |
| ], | |
| "max_seq_len": 10, | |
| "min_seq_len": 4, | |
| "files": [ | |
| { | |
| "file_name": "edges.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/edges.pkl" | |
| }, | |
| { | |
| "file_name": "ll_model.pth", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/ll_model.pth" | |
| }, | |
| { | |
| "file_name": "ll_model_cfg.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/ll_model_cfg.pkl" | |
| }, | |
| { | |
| "file_name": "meta.json", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/meta.json" | |
| } | |
| ], | |
| "transformer_cfg": { | |
| "n_layers": 2, | |
| "d_model": 20, | |
| "n_ctx": 10, | |
| "d_head": 5, | |
| "model_name": "custom", | |
| "n_heads": 4, | |
| "d_mlp": 80, | |
| "act_fn": "gelu", | |
| "d_vocab": 7, | |
| "eps": 1e-05, | |
| "use_attn_result": true, | |
| "use_attn_scale": true, | |
| "use_split_qkv_input": true, | |
| "use_hook_mlp_in": true, | |
| "use_attn_in": false, | |
| "use_local_attn": false, | |
| "original_architecture": null, | |
| "from_checkpoint": false, | |
| "checkpoint_index": null, | |
| "checkpoint_label_type": null, | |
| "checkpoint_value": null, | |
| "tokenizer_name": null, | |
| "window_size": null, | |
| "attn_types": null, | |
| "init_mode": "gpt2", | |
| "normalization_type": null, | |
| "n_devices": 1, | |
| "attention_dir": "causal", | |
| "attn_only": false, | |
| "seed": 0, | |
| "initializer_range": 0.17056057308448835, | |
| "init_weights": true, | |
| "scale_attn_by_inverse_layer_idx": false, | |
| "positional_embedding_type": "standard", | |
| "final_rms": false, | |
| "d_vocab_out": 1, | |
| "parallel_attn_mlp": false, | |
| "rotary_dim": null, | |
| "n_params": 9600, | |
| "use_hook_tokens": false, | |
| "gated_mlp": false, | |
| "default_prepend_bos": true, | |
| "dtype": "torch.float32", | |
| "tokenizer_prepends_bos": null, | |
| "n_key_value_heads": null, | |
| "post_embedding_ln": false, | |
| "rotary_base": 10000, | |
| "trust_remote_code": false, | |
| "rotary_adjacent_pairs": false | |
| }, | |
| "transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/ll_model_cfg.pkl", | |
| "training_args": { | |
| "atol": 0.05, | |
| "lr": 0.001, | |
| "use_single_loss": false, | |
| "iit_weight": 1.0, | |
| "behavior_weight": 1.0, | |
| "strict_weight": 0.4, | |
| "epochs": 2000, | |
| "act_fn": "gelu", | |
| "clip_grad_norm": 0.1, | |
| "lr_scheduler": "" | |
| }, | |
| "training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/meta.json", | |
| "weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/ll_model.pth", | |
| "circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/4/edges.pkl" | |
| }, | |
| { | |
| "case_id": "8", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/8", | |
| "task_description": "Fills gaps between tokens with a specified filler.", | |
| "vocab": [ | |
| "J", | |
| "LB", | |
| "TPSI", | |
| "V", | |
| "b", | |
| "no", | |
| "oCLrZaW", | |
| "poiVg" | |
| ], | |
| "max_seq_len": 10, | |
| "min_seq_len": 4, | |
| "files": [ | |
| { | |
| "file_name": "edges.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/8/edges.pkl" | |
| }, | |
| { | |
| "file_name": "ll_model.pth", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/8/ll_model.pth" | |
| }, | |
| { | |
| "file_name": "ll_model_cfg.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/8/ll_model_cfg.pkl" | |
| }, | |
| { | |
| "file_name": "meta.json", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/8/meta.json" | |
| } | |
| ], | |
| "transformer_cfg": { | |
| "n_layers": 2, | |
| "d_model": 20, | |
| "n_ctx": 10, | |
| "d_head": 5, | |
| "model_name": "custom", | |
| "n_heads": 4, | |
| "d_mlp": 80, | |
| "act_fn": "gelu", | |
| "d_vocab": 10, | |
| "eps": 1e-05, | |
| "use_attn_result": true, | |
| "use_attn_scale": true, | |
| "use_split_qkv_input": true, | |
| "use_hook_mlp_in": true, | |
| "use_attn_in": false, | |
| "use_local_attn": false, | |
| "original_architecture": null, | |
| "from_checkpoint": false, | |
| "checkpoint_index": null, | |
| "checkpoint_label_type": null, | |
| "checkpoint_value": null, | |
| "tokenizer_name": null, | |
| "window_size": null, | |
| "attn_types": null, | |
| "init_mode": "gpt2", | |
| "normalization_type": null, | |
| "n_devices": 1, | |
| "attention_dir": "causal", | |
| "attn_only": false, | |
| "seed": 0, | |
| "initializer_range": 0.13333333333333333, | |
| "init_weights": true, | |
| "scale_attn_by_inverse_layer_idx": false, | |
| "positional_embedding_type": "standard", | |
| "final_rms": false, | |
| "d_vocab_out": 8, | |
| "parallel_attn_mlp": false, | |
| "rotary_dim": null, | |
| "n_params": 9600, | |
| "use_hook_tokens": false, | |
| "gated_mlp": false, | |
| "default_prepend_bos": true, | |
| "dtype": "torch.float32", | |
| "tokenizer_prepends_bos": null, | |
| "n_key_value_heads": null, | |
| "post_embedding_ln": false, | |
| "rotary_base": 10000, | |
| "trust_remote_code": false, | |
| "rotary_adjacent_pairs": false | |
| }, | |
| "transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/8/ll_model_cfg.pkl", | |
| "training_args": { | |
| "atol": 0.05, | |
| "lr": 0.01, | |
| "use_single_loss": false, | |
| "iit_weight": 1.0, | |
| "behavior_weight": 1.0, | |
| "strict_weight": 0.4, | |
| "epochs": 500, | |
| "act_fn": "gelu", | |
| "clip_grad_norm": 1.0, | |
| "lr_scheduler": "" | |
| }, | |
| "training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/8/meta.json", | |
| "weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/8/ll_model.pth", | |
| "circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/8/edges.pkl" | |
| }, | |
| { | |
| "case_id": "ioi", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/ioi", | |
| "task_description": "Indirect object identification", | |
| "max_seq_len": 16, | |
| "min_seq_len": 16, | |
| "files": [ | |
| { | |
| "file_name": "edges.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/edges.pkl" | |
| }, | |
| { | |
| "file_name": "ll_model.pth", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/ll_model.pth" | |
| }, | |
| { | |
| "file_name": "ll_model_cfg.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/ll_model_cfg.pkl" | |
| }, | |
| { | |
| "file_name": "meta.json", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/meta.json" | |
| } | |
| ], | |
| "transformer_cfg": { | |
| "n_layers": 6, | |
| "d_model": 64, | |
| "n_ctx": 1024, | |
| "d_head": 16, | |
| "model_name": "gpt2", | |
| "n_heads": 4, | |
| "d_mlp": 3072, | |
| "act_fn": "gelu_new", | |
| "d_vocab": 50257, | |
| "eps": 1e-05, | |
| "use_attn_result": false, | |
| "use_attn_scale": true, | |
| "use_split_qkv_input": false, | |
| "use_hook_mlp_in": false, | |
| "use_attn_in": false, | |
| "use_local_attn": false, | |
| "original_architecture": "GPT2LMHeadModel", | |
| "from_checkpoint": false, | |
| "checkpoint_index": null, | |
| "checkpoint_label_type": null, | |
| "checkpoint_value": null, | |
| "tokenizer_name": "gpt2", | |
| "window_size": null, | |
| "attn_types": null, | |
| "init_mode": "gpt2", | |
| "normalization_type": "LNPre", | |
| "n_devices": 1, | |
| "attention_dir": "causal", | |
| "attn_only": false, | |
| "seed": null, | |
| "initializer_range": 0.02886751345948129, | |
| "init_weights": true, | |
| "scale_attn_by_inverse_layer_idx": false, | |
| "positional_embedding_type": "standard", | |
| "final_rms": false, | |
| "d_vocab_out": 50257, | |
| "parallel_attn_mlp": false, | |
| "rotary_dim": null, | |
| "n_params": 2457600, | |
| "use_hook_tokens": false, | |
| "gated_mlp": false, | |
| "default_prepend_bos": true, | |
| "dtype": "torch.float32", | |
| "tokenizer_prepends_bos": false, | |
| "n_key_value_heads": null, | |
| "post_embedding_ln": false, | |
| "rotary_base": 10000, | |
| "trust_remote_code": false, | |
| "rotary_adjacent_pairs": false | |
| }, | |
| "transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/ll_model_cfg.pkl", | |
| "training_args": { | |
| "next_token": true, | |
| "non_ioi_thresh": 0.65, | |
| "use_per_token_check": false, | |
| "batch_size": 256, | |
| "lr": 0.001, | |
| "num_workers": 0, | |
| "early_stop": true, | |
| "lr_scheduler": null, | |
| "scheduler_val_metric": [ | |
| "val/accuracy", | |
| "val/IIA" | |
| ], | |
| "scheduler_mode": "max", | |
| "clip_grad_norm": 1.0, | |
| "atol": 0.05, | |
| "use_single_loss": false, | |
| "iit_weight": 1.0, | |
| "behavior_weight": 1.0, | |
| "strict_weight": 0.4 | |
| }, | |
| "training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/meta.json", | |
| "weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/ll_model.pth", | |
| "circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi/edges.pkl" | |
| }, | |
| { | |
| "case_id": "ioi_next_token", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/tree/main/ioi_next_token", | |
| "task_description": "Indirect object identification", | |
| "max_seq_len": 16, | |
| "min_seq_len": 16, | |
| "files": [ | |
| { | |
| "file_name": "edges.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/edges.pkl" | |
| }, | |
| { | |
| "file_name": "ll_model.pth", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/ll_model.pth" | |
| }, | |
| { | |
| "file_name": "ll_model_cfg.pkl", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/ll_model_cfg.pkl" | |
| }, | |
| { | |
| "file_name": "meta.json", | |
| "url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/meta.json" | |
| } | |
| ], | |
| "transformer_cfg": { | |
| "n_layers": 6, | |
| "d_model": 64, | |
| "n_ctx": 1024, | |
| "d_head": 16, | |
| "model_name": "gpt2", | |
| "n_heads": 4, | |
| "d_mlp": 3072, | |
| "act_fn": "gelu_new", | |
| "d_vocab": 50257, | |
| "eps": 1e-05, | |
| "use_attn_result": false, | |
| "use_attn_scale": true, | |
| "use_split_qkv_input": false, | |
| "use_hook_mlp_in": false, | |
| "use_attn_in": false, | |
| "use_local_attn": false, | |
| "original_architecture": "GPT2LMHeadModel", | |
| "from_checkpoint": false, | |
| "checkpoint_index": null, | |
| "checkpoint_label_type": null, | |
| "checkpoint_value": null, | |
| "tokenizer_name": "gpt2", | |
| "window_size": null, | |
| "attn_types": null, | |
| "init_mode": "gpt2", | |
| "normalization_type": "LNPre", | |
| "n_devices": 1, | |
| "attention_dir": "causal", | |
| "attn_only": false, | |
| "seed": null, | |
| "initializer_range": 0.02886751345948129, | |
| "init_weights": true, | |
| "scale_attn_by_inverse_layer_idx": false, | |
| "positional_embedding_type": "standard", | |
| "final_rms": false, | |
| "d_vocab_out": 50257, | |
| "parallel_attn_mlp": false, | |
| "rotary_dim": null, | |
| "n_params": 2457600, | |
| "use_hook_tokens": false, | |
| "gated_mlp": false, | |
| "default_prepend_bos": true, | |
| "dtype": "torch.float32", | |
| "tokenizer_prepends_bos": false, | |
| "n_key_value_heads": null, | |
| "post_embedding_ln": false, | |
| "rotary_base": 10000, | |
| "trust_remote_code": false, | |
| "rotary_adjacent_pairs": false | |
| }, | |
| "transformer_cfg_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/ll_model_cfg.pkl", | |
| "training_args": { | |
| "next_token": true, | |
| "non_ioi_thresh": 0.65, | |
| "use_per_token_check": false, | |
| "batch_size": 256, | |
| "lr": 0.001, | |
| "num_workers": 0, | |
| "early_stop": true, | |
| "lr_scheduler": null, | |
| "scheduler_val_metric": [ | |
| "val/accuracy", | |
| "val/IIA" | |
| ], | |
| "scheduler_mode": "max", | |
| "clip_grad_norm": 1.0, | |
| "atol": 0.05, | |
| "use_single_loss": false, | |
| "iit_weight": 1.0, | |
| "behavior_weight": 1.0, | |
| "strict_weight": 0.4 | |
| }, | |
| "training_args_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/meta.json", | |
| "weights_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/ll_model.pth", | |
| "circuit_file_url": "https://huggingface.co/cybershiptrooper/InterpBench/blob/main/ioi_next_token/edges.pkl" | |
| } | |
| ] | |
| } |