kokolamba commited on
Commit
91f760b
·
1 Parent(s): 6ea104c

Update model files

Browse files
Files changed (1) hide show
  1. config.json +4 -5
config.json CHANGED
@@ -2,10 +2,10 @@
2
  "architectures": [
3
  "SharedSpaceDecoderForCausalLM"
4
  ],
 
5
  "auto_map": {
6
- "AutoConfig": "configuration_shared_subspace_decoder.SharedSpaceDecoderConfig",
7
- "AutoModel": "modeling_shared_subspace_decoder.SharedSpaceDecoderModel",
8
- "AutoModelForCausalLM": "modeling_shared_subspace_decoder.SharedSpaceDecoderForCausalLM"
9
  },
10
  "attention_backend": "flash_attention_2",
11
  "attention_bias": false,
@@ -23,7 +23,6 @@
23
  "kv_shared_dim": 96,
24
  "layer_norm_eps": 1e-12,
25
  "max_position_embeddings": 1024,
26
- "model_type": "shared_subspace_decoder",
27
  "nope_dims": 32,
28
  "norm_type": "rmsnorm",
29
  "num_attention_heads": 12,
@@ -45,4 +44,4 @@
45
  "vocab_rank": null,
46
  "vocab_size": 50257,
47
  "vocab_subspace": false
48
- }
 
2
  "architectures": [
3
  "SharedSpaceDecoderForCausalLM"
4
  ],
5
+ "model_type": "shared_subspace_decoder",
6
  "auto_map": {
7
+ "AutoConfig": "shared_space_config.SharedSpaceDecoderConfig",
8
+ "AutoModelForCausalLM": "task_heads.SharedSpaceDecoderForCausalLM"
 
9
  },
10
  "attention_backend": "flash_attention_2",
11
  "attention_bias": false,
 
23
  "kv_shared_dim": 96,
24
  "layer_norm_eps": 1e-12,
25
  "max_position_embeddings": 1024,
 
26
  "nope_dims": 32,
27
  "norm_type": "rmsnorm",
28
  "num_attention_heads": 12,
 
44
  "vocab_rank": null,
45
  "vocab_size": 50257,
46
  "vocab_subspace": false
47
+ }