DavidHiggis
commit
689a40c
-
1.34 kB
commit
imgemb.pt
Detected Pickle imports (252)
- "torch._utils._rebuild_tensor_v2",
- "collections.OrderedDict",
- "torch.DoubleStorage",
- "torch.LongStorage",
- "__torch__.torch.nn.modules.linear.___torch_mangle_14.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_127.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_116.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_42.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_92.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_176.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_226.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_2.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_100.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_57.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_101.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_96.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_44.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_52.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_147.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_41.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_161.Dropout",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_13.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_213.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.activation.___torch_mangle_205.MultiheadAttention",
- "__torch__.torch.nn.modules.activation.___torch_mangle_75.MultiheadAttention",
- "__torch__.torch.nn.modules.activation.___torch_mangle_155.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_210.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_162.Dropout",
- "__torch__.torch.nn.modules.activation.___torch_mangle_125.MultiheadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_208.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_118.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_24.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_204.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_170.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_207.Dropout",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_133.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.linear.___torch_mangle_216.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_97.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_112.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_77.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_81.Dropout",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_163.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.linear.___torch_mangle_146.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_90.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_67.Dropout",
- "__torch__.torch.nn.modules.activation.___torch_mangle_165.MultiheadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_0.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_144.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_15.MultiheadAttention",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_173.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.linear.___torch_mangle_66.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_132.Dropout",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_103.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.linear.___torch_mangle_126.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_157.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_180.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_148.Linear",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_129.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_3.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_4.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_39.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_58.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_99.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_117.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_69.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_150.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_192.Dropout",
- "__torch__.torch.nn.modules.transformer.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.linear.___torch_mangle_124.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_18.Linear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_35.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_109.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_159.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_227.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_229.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_122.Dropout",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_33.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_53.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_82.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_218.Linear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_215.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_235.LayerNorm",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_183.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.linear.___torch_mangle_184.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_139.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_160.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_74.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_89.LayerNorm",
- "__torch__.torch.nn.modules.activation.___torch_mangle_25.MultiheadAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_221.Dropout",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_43.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.linear.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_152.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_8.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_189.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_222.Dropout",
- "__torch__.torch.nn.modules.activation.___torch_mangle_45.MultiheadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_78.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_108.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_172.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_61.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_34.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_201.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_136.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_194.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_187.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_219.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_107.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_134.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_175.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_119.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_10.LayerNorm",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_203.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.linear.___torch_mangle_54.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_156.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_174.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_225.MultiheadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_154.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_128.Linear",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_193.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_234.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_130.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_11.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_98.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_70.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_151.Dropout",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_83.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.linear.___torch_mangle_206.Linear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_105.MultiheadAttention",
- "__torch__.torch.nn.modules.conv.Conv2d",
- "__torch__.torch.nn.modules.activation.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_196.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_20.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_167.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_12.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_164.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_87.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_86.Linear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_185.MultiheadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_94.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_106.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_131.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_104.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_21.Dropout",
- "__torch__.torch.nn.modules.activation.___torch_mangle_135.MultiheadAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_181.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_202.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_31.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_102.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_121.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_49.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_182.Dropout",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_223.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_142.Dropout",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_233.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_110.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_27.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_214.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_197.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_179.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_16.Linear",
- "torch.FloatStorage",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_17.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_200.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_224.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_51.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_149.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_141.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_59.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_169.LayerNorm",
- "__torch__.torch.nn.modules.activation.___torch_mangle_65.MultiheadAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_111.Dropout",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_113.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_29.LayerNorm",
- "__torch__.torch.nn.modules.activation.___torch_mangle_55.MultiheadAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_212.Dropout",
- "__torch__.torch.nn.modules.activation.___torch_mangle_145.MultiheadAttention",
- "__torch__.imgemb_pnnx.Model",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_79.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_137.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_88.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_138.Linear",
- "collections.OrderedDict",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_40.LayerNorm",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_63.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.linear.___torch_mangle_186.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_19.LayerNorm",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_73.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.activation.___torch_mangle_5.MultiheadAttention",
- "__torch__.torch.nn.modules.activation.___torch_mangle_85.MultiheadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_168.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_120.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_211.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_6.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_38.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_46.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_9.LayerNorm",
- "__torch__.torch.nn.modules.linear.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_158.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_171.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_198.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_72.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_76.Linear",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_123.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_140.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_220.LayerNorm",
- "torch.HalfStorage",
- "__torch__.torch.nn.modules.normalization.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_209.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_177.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_30.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_47.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_217.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_60.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_7.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_48.Linear",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_93.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_230.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_26.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_71.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_91.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_37.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_190.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_50.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_199.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_62.Dropout",
- "__torch__.torch.nn.modules.activation.___torch_mangle_95.MultiheadAttention",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_153.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.activation.___torch_mangle_195.MultiheadAttention",
- "__torch__.torch.nn.modules.activation.___torch_mangle_115.MultiheadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_68.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_28.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_64.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_80.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_84.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_166.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_56.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_178.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_231.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_188.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_32.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_22.Dropout",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_23.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_143.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_191.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_228.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_36.Linear",
- "__torch__.torch.nn.modules.dropout.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_114.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_232.Dropout"
How to fix it?
564 MB
commit
-
2.01 kB
commit
-
2.36 MB
commit
transformer_pnnx.pt
Detected Pickle imports (130)
- "torch._utils._rebuild_tensor_v2",
- "collections.OrderedDict",
- "torch.DoubleStorage",
- "torch.LongStorage",
- "__torch__.torch.nn.modules.linear.___torch_mangle_14.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_42.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_92.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_2.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_100.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_57.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_101.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_96.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_44.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_52.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_41.Dropout",
- "__torch__.transformerJITnative.transformer",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_13.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.activation.___torch_mangle_75.MultiheadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_24.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_97.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_112.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_77.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_81.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_90.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_67.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_0.Linear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_15.MultiheadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_66.Linear",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_103.TransformerEncoderLayer",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_3.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_4.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_39.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_58.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_99.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_69.LayerNorm",
- "__torch__.torch.nn.modules.transformer.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.linear.___torch_mangle_18.Linear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_35.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_109.LayerNorm",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_33.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_53.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_82.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_74.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_89.LayerNorm",
- "__torch__.torch.nn.modules.activation.___torch_mangle_25.MultiheadAttention",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_43.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.sparse.Embedding",
- "__torch__.torch.nn.modules.linear.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_8.Linear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_45.MultiheadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_78.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_108.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_61.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_34.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_107.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_10.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_54.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_11.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_98.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_70.LayerNorm",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_83.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.activation.___torch_mangle_105.MultiheadAttention",
- "__torch__.torch.nn.modules.activation.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_20.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_12.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_87.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_86.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_94.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_106.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_104.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_21.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_31.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_102.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_49.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_110.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_27.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_16.Linear",
- "torch.FloatStorage",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_17.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_51.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_59.LayerNorm",
- "__torch__.torch.nn.modules.activation.___torch_mangle_65.MultiheadAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_111.Dropout",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_113.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_29.LayerNorm",
- "__torch__.torch.nn.modules.activation.___torch_mangle_55.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_79.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_88.Linear",
- "collections.OrderedDict",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_40.LayerNorm",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_63.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_19.LayerNorm",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_73.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.activation.___torch_mangle_5.MultiheadAttention",
- "__torch__.torch.nn.modules.activation.___torch_mangle_85.MultiheadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_6.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_38.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_114.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_46.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_9.LayerNorm",
- "__torch__.torch.nn.modules.linear.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_72.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_76.Linear",
- "__torch__.torch.nn.modules.normalization.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_30.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_47.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_60.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_7.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_48.Linear",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_93.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.linear.___torch_mangle_26.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_71.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_91.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_37.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_50.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_62.Dropout",
- "__torch__.torch.nn.modules.activation.___torch_mangle_95.MultiheadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_68.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_28.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_64.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_80.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_84.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_56.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_32.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_22.Dropout",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_23.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.linear.___torch_mangle_36.Linear",
- "__torch__.torch.nn.modules.dropout.Dropout"
How to fix it?
331 MB
commit