| [llama] | |
| model_name = llama_7b | |
| head_num = 32 | |
| size_per_head = 128 | |
| inter_size = 11008 | |
| num_layer = 32 | |
| rotary_embedding = 128 | |
| vocab_size = 32000 | |
| start_id = 1 | |
| end_id = 2 | |
| weight_data_type = fp16 | |
| [llama] | |
| model_name = llama_7b | |
| head_num = 32 | |
| size_per_head = 128 | |
| inter_size = 11008 | |
| num_layer = 32 | |
| rotary_embedding = 128 | |
| vocab_size = 32000 | |
| start_id = 1 | |
| end_id = 2 | |
| weight_data_type = fp16 | |