OzTianlu commited on
Commit
98051f3
·
verified ·
1 Parent(s): 3dd228c

Upload folder using huggingface_hub

Browse files
Files changed (2) hide show
  1. config.json +5 -5
  2. model.safetensors +3 -0
config.json CHANGED
@@ -2,18 +2,18 @@
2
  "architectures": [
3
  "AsteriskForCausalLM"
4
  ],
5
- "auto_map": {
6
- "AutoConfig": "AsteriskForCausalLM.AsteriskConfig",
7
- "AutoModelForCausalLM": "AsteriskForCausalLM.AsteriskForCausalLM"
8
- },
9
  "aspp_dropout": 0.15,
10
  "aspp_hidden_dim": 512,
11
  "aspp_num_neighbors": 1,
12
  "aspp_num_steps": 6,
13
  "attention_bias": false,
14
  "attention_dropout": 0.0,
 
 
 
 
15
  "bos_token_id": 128000,
16
- "dtype": "float32",
17
  "eos_token_id": 128009,
18
  "head_dim": 64,
19
  "hidden_act": "silu",
 
2
  "architectures": [
3
  "AsteriskForCausalLM"
4
  ],
 
 
 
 
5
  "aspp_dropout": 0.15,
6
  "aspp_hidden_dim": 512,
7
  "aspp_num_neighbors": 1,
8
  "aspp_num_steps": 6,
9
  "attention_bias": false,
10
  "attention_dropout": 0.0,
11
+ "auto_map": {
12
+ "AutoConfig": "AsteriskForCausalLM.AsteriskConfig",
13
+ "AutoModelForCausalLM": "AsteriskForCausalLM.AsteriskForCausalLM"
14
+ },
15
  "bos_token_id": 128000,
16
+ "dtype": "bfloat16",
17
  "eos_token_id": 128009,
18
  "head_dim": 64,
19
  "hidden_act": "silu",
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a2b6986f7f047e3b534f39a283eab49b4e7218c0c7bab9f34341a8c35e2eecd
3
+ size 3008999832