{ "config": { "activation_function": "gelu", "bias": true, "dropout_prob": null, "embedding_size": 2048, "head_type": "causal_lm", "label2id": null, "layer_norm": true, "layers": 2, "shift_labels": true, "vocab_size": 61788 }, "hidden_size": 2048, "model_class": "LlamaAdapterModel", "model_name": "lelapa/InkubaLM-0.4B", "model_type": "llama", "name": "machine-translation", "version": "adapters.1.1.0" }