Spaces:
Sleeping
Sleeping
Fetching metadata from the HF Docker repository...
download history blame contribute delete
Detected Pickle imports (114)
- "__torch__.torch.nn.modules.dropout.___torch_mangle_23.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_33.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_36.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_26.LayerNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_0.Conv2d",
- "__torch__.torch.nn.modules.activation.ReLU",
- "__torch__.torch.nn.modules.activation.___torch_mangle_4.ReLU",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_18.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_34.Linear",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_29.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.normalization.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_17.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_30.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_1.ReLU",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_38.Dropout",
- "__torch__.torch.nn.modules.transformer.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.dropout.Dropout",
- "__torch__.torch.nn.modules.transformer.TransformerEncoder",
- "__torch__.torch.nn.modules.linear.___torch_mangle_10.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_9.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_20.NonDynamicallyQuantizableLinear",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.torch.nn.modules.activation.___torch_mangle_11.MultiheadAttention",
- "collections.OrderedDict",
- "__torch__.HyperAIBrain",
- "torch.FloatStorage",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_15.LayerNorm",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_39.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_16.LayerNorm",
- "__torch__.torch.nn.modules.pooling.___torch_mangle_5.MaxPool2d",
- "__torch__.torch.nn.modules.activation.___torch_mangle_21.MultiheadAttention",
- "__torch__.torch.nn.modules.pooling.MaxPool2d",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_37.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_7.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_35.LayerNorm",
- "__torch__.torch.nn.modules.pooling.___torch_mangle_2.MaxPool2d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_40.Linear",
- "__torch__.torch.nn.modules.linear.NonDynamicallyQuantizableLinear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_6.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_3.Conv2d",
- "__torch__.torch.nn.modules.container.ModuleList",
- "__torch__.torch.nn.modules.conv.Conv2d",
- "__torch__.torch.nn.modules.activation.MultiheadAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_25.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_13.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_32.Linear",
- "__torch__.torch.nn.modules.activation.___torch_mangle_31.MultiheadAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_12.Linear",
- "__torch__.torch.nn.modules.transformer.___torch_mangle_19.TransformerEncoderLayer",
- "__torch__.torch.nn.modules.linear.___torch_mangle_22.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_24.Linear",
- "__torch__.torch.nn.modules.linear.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_28.Dropout",
- "__torch__.torch.nn.modules.container.Sequential",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_27.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_14.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_8.Dropout",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "collections.OrderedDict",
- "torch.LongStorage",
- "torch._utils._rebuild_tensor_v2",
- "torch.jit._pickle.restore_type_tag",
- "collections.OrderedDict",
- "torch._utils._rebuild_tensor_v2",
- "torch.FloatStorage"
22.7 MB
- SHA256:
- 8644e302c5ae593ac8a56e20caecdd0d6e3f8940a1413d7f41a0ccbe7576cd4e
- Pointer size:
- 133 Bytes
- Size of remote file:
- 22.7 MB
- Xet hash:
- f329c4f9f6e825f43adb7b0cc408066bf813ea0a4423d57aeef7df99faf2fce4
·
·
Xet efficiently stores Large Files inside Git, intelligently splitting files into unique chunks and accelerating uploads and downloads. More info.