fix: override mark_tied_weights_as_initialized to skip BN buffer get_parameter
Browse files- modeling_rtmdet.py +6 -3
modeling_rtmdet.py
CHANGED
|
@@ -1758,10 +1758,13 @@ class RTMDetModel(PreTrainedModel):
|
|
| 1758 |
config_class = RTMDetConfig
|
| 1759 |
base_model_prefix = "rtmdet"
|
| 1760 |
main_input_name = "pixel_values"
|
| 1761 |
-
# share_conv aliases BN modules across strides;
|
| 1762 |
-
#
|
| 1763 |
_tied_weights_keys = None
|
| 1764 |
-
|
|
|
|
|
|
|
|
|
|
| 1765 |
def __init__(self, config):
|
| 1766 |
super().__init__(config)
|
| 1767 |
|
|
|
|
| 1758 |
config_class = RTMDetConfig
|
| 1759 |
base_model_prefix = "rtmdet"
|
| 1760 |
main_input_name = "pixel_values"
|
| 1761 |
+
# share_conv aliases BN modules across strides; override tied-weights
|
| 1762 |
+
# method so transformers never calls get_parameter() on BN buffers.
|
| 1763 |
_tied_weights_keys = None
|
| 1764 |
+
|
| 1765 |
+
def mark_tied_weights_as_initialized(self):
|
| 1766 |
+
pass # share_conv makes BN buffers look tied — skip to avoid AttributeError
|
| 1767 |
+
|
| 1768 |
def __init__(self, config):
|
| 1769 |
super().__init__(config)
|
| 1770 |
|