| { | |
| "architectures": [ | |
| "LlamaForCausalLM" | |
| ], | |
| "model_type": "llama", | |
| "embedding_dim": 2, | |
| "hidden_size": 100, | |
| "output_size": 32768, | |
| "num_layers": 2, | |
| "activation_function": "tanh", | |
| "initializer_range": 0.02 | |
| } |
| { | |
| "architectures": [ | |
| "LlamaForCausalLM" | |
| ], | |
| "model_type": "llama", | |
| "embedding_dim": 2, | |
| "hidden_size": 100, | |
| "output_size": 32768, | |
| "num_layers": 2, | |
| "activation_function": "tanh", | |
| "initializer_range": 0.02 | |
| } |