download history blame
Detected Pickle imports (73)
- "__torch__.splitters.___torch_mangle_6220.MiddlePart",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_6167.LayerNorm",
- "__torch__.splitters.___torch_mangle_6205.MiddlePart",
- "torch.FloatStorage",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_6187.GPT2MLP",
- "__torch__.transformers.pytorch_utils.___torch_mangle_6178.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_6215.NewGELUActivation",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_6217.GPT2MLP",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_6179.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_6181.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_6182.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_6211.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_6164.Dropout",
- "__torch__.torch.nn.modules.container.___torch_mangle_6204.ModuleList",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_6161.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_6213.Conv1D",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.transformers.pytorch_utils.___torch_mangle_6193.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_6207.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_6198.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_6165.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_6186.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_6177.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_6168.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_6184.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_6162.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_6195.Dropout",
- "__torch__.transformers.activations.___torch_mangle_6185.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_6208.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_6192.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_6210.Dropout",
- "collections.OrderedDict",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_6166.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_6176.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_6214.Conv1D",
- "__torch__.torch.nn.modules.container.___torch_mangle_6219.ModuleList",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_6171.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_6173.GPT2Block",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_6196.GPT2SdpaAttention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_6163.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_6200.NewGELUActivation",
- "__torch__.torch.nn.modules.container.___torch_mangle_6189.ModuleList",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_6206.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_6209.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_6212.LayerNorm",
- "__torch__.torch.nn.modules.container.___torch_mangle_6221.Sequential",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_6203.GPT2Block",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_6216.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_6169.Conv1D",
- "__torch__.splitters.___torch_mangle_6175.MiddlePart",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_6172.GPT2MLP",
- "__torch__.splitters.___torch_mangle_6190.MiddlePart",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_6188.GPT2Block",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_6194.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_6180.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_6202.GPT2MLP",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_6218.GPT2Block",
- "__torch__.transformers.pytorch_utils.___torch_mangle_6199.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_6197.LayerNorm",
- "torch.BoolStorage",
- "__torch__.transformers.pytorch_utils.___torch_mangle_6183.Conv1D",
- "__torch__.splitters.___torch_mangle_6222.Part",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_6201.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_6191.LayerNorm",
- "__torch__.torch.nn.modules.container.___torch_mangle_6174.ModuleList",
- "__torch__.transformers.activations.___torch_mangle_6170.NewGELUActivation",
- "torch._utils._rebuild_tensor_v2",
- "collections.OrderedDict",
- "torch.DoubleStorage",
- "collections.OrderedDict",
- "torch.FloatStorage",
- "torch._utils._rebuild_tensor_v2",
- "torch.jit._pickle.restore_type_tag"
118 MB
- Xet hash:
- 9ed3aed55f5874da22a491de1927e5aa8f294f21fd8c44a6c7dc1ab2dec9a3ae
- Size of remote file:
- 118 MB
- SHA256:
- 026be86809687d5e63f30f90b4dd8870713013ff0992d23fc7a8bcd1024f8bb2
·
Xet efficiently stores Large Files inside Git, intelligently splitting files into unique chunks and accelerating uploads and downloads. More info.