akore commited on
Commit
ccc5362
·
verified ·
1 Parent(s): d6aaed6

fix: override mark_tied_weights_as_initialized to skip BN buffer get_parameter

Browse files
Files changed (1) hide show
  1. modeling_rtmdet.py +6 -3
modeling_rtmdet.py CHANGED
@@ -1758,10 +1758,13 @@ class RTMDetModel(PreTrainedModel):
1758
  config_class = RTMDetConfig
1759
  base_model_prefix = "rtmdet"
1760
  main_input_name = "pixel_values"
1761
- # share_conv aliases BN modules across strides; disable tied-weights
1762
- # machinery so transformers doesn't try get_parameter() on BN buffers.
1763
  _tied_weights_keys = None
1764
-
 
 
 
1765
  def __init__(self, config):
1766
  super().__init__(config)
1767
 
 
1758
  config_class = RTMDetConfig
1759
  base_model_prefix = "rtmdet"
1760
  main_input_name = "pixel_values"
1761
+ # share_conv aliases BN modules across strides; override tied-weights
1762
+ # method so transformers never calls get_parameter() on BN buffers.
1763
  _tied_weights_keys = None
1764
+
1765
+ def mark_tied_weights_as_initialized(self):
1766
+ pass # share_conv makes BN buffers look tied — skip to avoid AttributeError
1767
+
1768
  def __init__(self, config):
1769
  super().__init__(config)
1770