download history blame contribute delete
Detected Pickle imports (135)
- "__torch__.torch.nn.modules.linear.___torch_mangle_30.Linear",
- "__torch__.torch.nn.modules.transformer.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_36.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_44.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_31.Dropout",
- "__torch__.torch.nn.modules.container.___torch_mangle_45.Sequential",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.torch.nn.modules.activation.MultiheadAttention",
- "__torch__.torch.nn.modules.activation.___torch_mangle_42.ReLU",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_43.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_35.Dropout",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_37.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_25.Dropout",
- "__torch__.torch.nn.modules.sparse.___torch_mangle_2.Embedding",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_7.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_8.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_29.MultiheadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_12.Linear",
- "__torch__.torch.nn.modules.container.___torch_mangle_50.Sequential",
- "__torch__.torch.nn.modules.batchnorm.BatchNorm1d",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_27.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.sparse.___torch_mangle_1.Embedding",
- "__torch__.torch.nn.modules.activation.___torch_mangle_9.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_23.LayerNorm",
- "__torch__.PokerModelV2",
- "__torch__.torch.nn.modules.linear.___torch_mangle_46.Linear",
- "__torch__.torch.nn.modules.container.Sequential",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_11.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_34.LayerNorm",
- "collections.OrderedDict",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_17.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.linear.___torch_mangle_4.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_48.Dropout",
- "__torch__.torch.nn.modules.transformer.TransformerEncoder",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_24.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_33.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_13.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_32.Linear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_19.MultiheadAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_21.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_26.Dropout",
- "__torch__.torch.nn.modules.sparse.Embedding",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_6.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_22.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_28.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.dropout.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_16.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_10.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_20.Linear",
- "__torch__.torch.nn.modules.activation.ReLU",
- "__torch__.torch.nn.modules.sparse.___torch_mangle_0.Embedding",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_15.Dropout",
- "torch.LongStorage",
- "torch.FloatStorage",
- "__torch__.torch.nn.modules.container.ModuleList",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_40.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_49.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_38.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_41.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_3.Dropout",
- "__torch__.torch.nn.modules.normalization.LayerNorm",
- "__torch__.torch.nn.modules.linear.___torch_mangle_39.Linear",
- "__torch__.torch.nn.modules.linear.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_47.ReLU",
- "__torch__.torch.nn.modules.linear.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_18.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_5.LayerNorm",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "collections.OrderedDict",
- "torch.LongStorage",
- "torch._utils._rebuild_tensor_v2",
- "collections.OrderedDict",
- "torch.jit._pickle.restore_type_tag",
- "torch.LongStorage",
- "torch._utils._rebuild_tensor_v2"
3.87 MB
- Xet hash:
- 9fee218d69c26224c17c3205ae038d51d0ef1cd261819f84261278d1a096143f
- Size of remote file:
- 3.87 MB
- SHA256:
- b9128f9e2696bc6614df194ec0474d7deb63a1fd914363132a8f243ad9007571
·
Xet efficiently stores Large Files inside Git, intelligently splitting files into unique chunks and accelerating uploads and downloads. More info.