| { | |
| _name_or_path: "meta-llama/Llama-2-7b-hf", | |
| architectures: [ | |
| "LlamaForCausalLM" | |
| ], | |
| bos_token_id: 1, | |
| eos_token_id: 2, | |
| hidden_act: "silu", | |
| hidden_size: 4096, | |
| initializer_range: 0.02, | |
| intermediate_size: 11008, | |
| max_position_embeddings: 2048, | |
| model_type: "llama", | |
| num_attention_heads: 32, | |
| num_hidden_layers: 32, | |
| num_key_value_heads: 32, | |
| pad_token_id: 0, | |
| pretraining_tp: 1, | |
| rms_norm_eps: 0.00001, | |
| rope_scaling: null, | |
| tie_word_embeddings: false, | |
| torch_dtype: "float32", | |
| transformers_version: "4.29.2", | |
| use_cache: true, | |
| vocab_size: 32000 | |
| } |