Sayan01 commited on
Commit
bd970f6
·
verified ·
1 Parent(s): beb1b12

Upload PhiForCausalLM

Browse files
config.json CHANGED
@@ -1,16 +1,11 @@
1
  {
2
- "_name_or_path": "microsoft/phi-2",
3
  "architectures": [
4
  "PhiForCausalLM"
5
  ],
6
  "attention_dropout": 0.0,
7
- "auto_map": {
8
- "AutoConfig": "microsoft/phi-2--configuration_phi.PhiConfig",
9
- "AutoModelForCausalLM": "microsoft/phi-2--modeling_phi.PhiForCausalLM"
10
- },
11
- "bos_token_id": 50256,
12
  "embd_pdrop": 0.0,
13
- "eos_token_id": 50256,
14
  "hidden_act": "gelu_new",
15
  "hidden_size": 2560,
16
  "initializer_range": 0.02,
@@ -21,9 +16,10 @@
21
  "num_attention_heads": 32,
22
  "num_hidden_layers": 32,
23
  "num_key_value_heads": 32,
24
- "partial_rotary_factor": 0.4,
 
25
  "qk_layernorm": false,
26
- "resid_pdrop": 0.1,
27
  "rope_scaling": null,
28
  "rope_theta": 10000.0,
29
  "tie_word_embeddings": false,
 
1
  {
 
2
  "architectures": [
3
  "PhiForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
6
+ "bos_token_id": 1,
 
 
 
 
7
  "embd_pdrop": 0.0,
8
+ "eos_token_id": 2,
9
  "hidden_act": "gelu_new",
10
  "hidden_size": 2560,
11
  "initializer_range": 0.02,
 
16
  "num_attention_heads": 32,
17
  "num_hidden_layers": 32,
18
  "num_key_value_heads": 32,
19
+ "output_hidden_states": true,
20
+ "partial_rotary_factor": 0.5,
21
  "qk_layernorm": false,
22
+ "resid_pdrop": 0.0,
23
  "rope_scaling": null,
24
  "rope_theta": 10000.0,
25
  "tie_word_embeddings": false,
generation_config.json CHANGED
@@ -1,6 +1,7 @@
1
  {
2
  "_from_model_config": true,
3
- "bos_token_id": 50256,
4
- "eos_token_id": 50256,
 
5
  "transformers_version": "4.39.0"
6
  }
 
1
  {
2
  "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "output_hidden_states": true,
6
  "transformers_version": "4.39.0"
7
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0f8957b985ff7190c1ad98eff967858341bb2b89fbcf12daf321f939c02f478b
3
  size 4982355512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de9690424cd10d30cc5bbbf31b5ba7149fe2d1b4d1c9b3e28378c37496dfddcc
3
  size 4982355512
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c102f1a044b5cf0875437d4b986ca21c1b93604abe58a220ace5e53872e1142f
3
  size 4982541984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:526f616bb5753775548b200b2d5afffa862bf3a17cf53c004b1ba8d702fb5890
3
  size 4982541984
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9f181a767f80f05b1406cf28a383a28db9b950ae0cbf1216159e60fddbaec7b3
3
  size 1153887616
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:972b1ffdcd8251223d3e51560d3ebdc6433fdd0a7d091342c359fcda227f91c7
3
  size 1153887616