fix: call super()._init_weights() to match Qwen2Moe convention for transformers v5
Browse files- modeling_llada2_moe.py +1 -0
modeling_llada2_moe.py
CHANGED
|
@@ -687,6 +687,7 @@ class LLaDA2MoePreTrainedModel(PreTrainedModel):
|
|
| 687 |
_supports_cache_class = True
|
| 688 |
|
| 689 |
def _init_weights(self, module):
|
|
|
|
| 690 |
std = self.config.initializer_range
|
| 691 |
if isinstance(module, nn.Linear):
|
| 692 |
module.weight.data.normal_(mean=0.0, std=std)
|
|
|
|
| 687 |
_supports_cache_class = True
|
| 688 |
|
| 689 |
def _init_weights(self, module):
|
| 690 |
+
super()._init_weights(module)
|
| 691 |
std = self.config.initializer_range
|
| 692 |
if isinstance(module, nn.Linear):
|
| 693 |
module.weight.data.normal_(mean=0.0, std=std)
|