Fix: set _tied_weights_keys=None to avoid BN buffer/parameter confusion in transformers 5.2
Browse files- modeling_rtmdet.py +3 -0
modeling_rtmdet.py
CHANGED
|
@@ -1758,6 +1758,9 @@ class RTMDetModel(PreTrainedModel):
|
|
| 1758 |
config_class = RTMDetConfig
|
| 1759 |
base_model_prefix = "rtmdet"
|
| 1760 |
main_input_name = "pixel_values"
|
|
|
|
|
|
|
|
|
|
| 1761 |
|
| 1762 |
def __init__(self, config):
|
| 1763 |
super().__init__(config)
|
|
|
|
| 1758 |
config_class = RTMDetConfig
|
| 1759 |
base_model_prefix = "rtmdet"
|
| 1760 |
main_input_name = "pixel_values"
|
| 1761 |
+
# share_conv aliases BN modules across strides; disable tied-weights
|
| 1762 |
+
# machinery so transformers doesn't try get_parameter() on BN buffers.
|
| 1763 |
+
_tied_weights_keys = None
|
| 1764 |
|
| 1765 |
def __init__(self, config):
|
| 1766 |
super().__init__(config)
|