| _name_or_path: ./models/dolphin-2_6-phi-2 | |
| activation_function: gelu_new | |
| architectures: | |
| - PhiForCausalLM | |
| attn_pdrop: 0.0 | |
| auto_map: | |
| AutoConfig: configuration_phi.PhiConfig | |
| AutoModel: modeling_phi.PhiForCausalLM | |
| AutoModelForCausalLM: modeling_phi.PhiForCausalLM | |
| embd_pdrop: 0.0 | |
| flash_attn: false | |
| flash_rotary: false | |
| fused_dense: false | |
| img_processor: null | |
| initializer_range: 0.02 | |
| layer_norm_epsilon: 1e-05 | |
| model_type: phi-msft | |
| n_embd: 2560 | |
| n_head: 32 | |
| n_head_kv: null | |
| n_inner: null | |
| n_layer: 32 | |
| n_positions: 2048 | |
| resid_pdrop: 0.1 | |
| rotary_dim: 32 | |
| tie_word_embeddings: false | |
| torch_dtype: float16 | |
| transformers_version: 4.36.2 | |
| use_cache: true | |
| vocab_size: 51200 |