Update modeling_diff_llama.py
Browse files- modeling_diff_llama.py +0 -8
modeling_diff_llama.py
CHANGED
|
@@ -39,14 +39,6 @@ class DiffLLaMAConfig(LlamaConfig):
|
|
| 39 |
self.rope_scaling = rope_scaling or {"type": "linear", "factor": 1.0}
|
| 40 |
# Add any custom configuration parameters here
|
| 41 |
|
| 42 |
-
@classmethod
|
| 43 |
-
def from_pretrained(cls, pretrained_model_name_or_path, **kwargs):
|
| 44 |
-
"""
|
| 45 |
-
Load configuration from a pretrained model.
|
| 46 |
-
"""
|
| 47 |
-
config_dict = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
|
| 48 |
-
return cls(**config_dict)
|
| 49 |
-
|
| 50 |
|
| 51 |
def init_method(tensor):
|
| 52 |
"""Initialize tensor with Kaiming uniform initialization."""
|
|
|
|
| 39 |
self.rope_scaling = rope_scaling or {"type": "linear", "factor": 1.0}
|
| 40 |
# Add any custom configuration parameters here
|
| 41 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 42 |
|
| 43 |
def init_method(tensor):
|
| 44 |
"""Initialize tensor with Kaiming uniform initialization."""
|