akore commited on
Commit
d6aaed6
·
verified ·
1 Parent(s): 75bd0e0

Fix: set _tied_weights_keys=None to avoid BN buffer/parameter confusion in transformers 5.2

Browse files
Files changed (1) hide show
  1. modeling_rtmdet.py +3 -0
modeling_rtmdet.py CHANGED
@@ -1758,6 +1758,9 @@ class RTMDetModel(PreTrainedModel):
1758
  config_class = RTMDetConfig
1759
  base_model_prefix = "rtmdet"
1760
  main_input_name = "pixel_values"
 
 
 
1761
 
1762
  def __init__(self, config):
1763
  super().__init__(config)
 
1758
  config_class = RTMDetConfig
1759
  base_model_prefix = "rtmdet"
1760
  main_input_name = "pixel_values"
1761
+ # share_conv aliases BN modules across strides; disable tied-weights
1762
+ # machinery so transformers doesn't try get_parameter() on BN buffers.
1763
+ _tied_weights_keys = None
1764
 
1765
  def __init__(self, config):
1766
  super().__init__(config)