Upload folder using huggingface_hub
Browse files- config.json +1 -0
- modeling_tiny_mixtral.py +4 -0
config.json
CHANGED
|
@@ -3,6 +3,7 @@
|
|
| 3 |
"TinyMixtralForCausalLM"
|
| 4 |
],
|
| 5 |
"auto_map": {
|
|
|
|
| 6 |
"AutoModelForCausalLM": "modeling_tiny_mixtral.TinyMixtralForCausalLM"
|
| 7 |
},
|
| 8 |
"attn_dropout": 0.0,
|
|
|
|
| 3 |
"TinyMixtralForCausalLM"
|
| 4 |
],
|
| 5 |
"auto_map": {
|
| 6 |
+
"AutoConfig": "modeling_tiny_mixtral.TinyMixtralConfig",
|
| 7 |
"AutoModelForCausalLM": "modeling_tiny_mixtral.TinyMixtralForCausalLM"
|
| 8 |
},
|
| 9 |
"attn_dropout": 0.0,
|
modeling_tiny_mixtral.py
CHANGED
|
@@ -81,6 +81,10 @@ class TinyMixtralConfig(PretrainedConfig):
|
|
| 81 |
device = ModelConfig.device,
|
| 82 |
**kwargs
|
| 83 |
):
|
|
|
|
|
|
|
|
|
|
|
|
|
| 84 |
super().__init__(top_k=None,**kwargs)
|
| 85 |
self.vocab_size = vocab_size
|
| 86 |
self.d_model = d_model
|
|
|
|
| 81 |
device = ModelConfig.device,
|
| 82 |
**kwargs
|
| 83 |
):
|
| 84 |
+
kwargs["auto_map"] = {
|
| 85 |
+
"AutoConfig": "modeling_tiny_mixtral.TinyMixtralConfig",
|
| 86 |
+
"AutoModelForCausalLM": "modeling_tiny_mixtral.TinyMixtralForCausalLM"
|
| 87 |
+
}
|
| 88 |
super().__init__(top_k=None,**kwargs)
|
| 89 |
self.vocab_size = vocab_size
|
| 90 |
self.d_model = d_model
|