download history blame
Detected Pickle imports (88)
- "torch.FloatStorage",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_4593.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_4621.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_4635.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_4626.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_4615.Dropout",
- "__torch__.splitters.___torch_mangle_4589.MiddlePart",
- "__torch__.transformers.pytorch_utils.___torch_mangle_4597.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_4595.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_4620.LayerNorm",
- "__torch__.torch.nn.modules.container.___torch_mangle_4588.ModuleList",
- "__torch__.transformers.pytorch_utils.___torch_mangle_4622.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_4640.GPT2SdpaAttention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_4636.Conv1D",
- "torch.BoolStorage",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_4630.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_4613.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_4591.Conv1D",
- "__torch__.splitters.___torch_mangle_4604.MiddlePart",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_4578.Dropout",
- "__torch__.transformers.activations.___torch_mangle_4584.NewGELUActivation",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_4587.GPT2Block",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_4623.Dropout",
- "__torch__.torch.nn.modules.container.___torch_mangle_4603.ModuleList",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_4608.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_4577.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_4590.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_4645.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_4624.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_4581.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_4605.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_4602.GPT2Block",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_4596.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_4586.GPT2MLP",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_4647.GPT2Block",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_4639.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_4628.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_4614.NewGELUActivation",
- "__torch__.torch.nn.modules.container.___torch_mangle_4650.Sequential",
- "__torch__.transformers.pytorch_utils.___torch_mangle_4642.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_4576.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_4643.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_4582.Conv1D",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_4600.Dropout",
- "__torch__.splitters.___torch_mangle_4651.Part",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_4616.GPT2MLP",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_4625.GPT2SdpaAttention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_4627.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_4638.Dropout",
- "__torch__.transformers.activations.___torch_mangle_4644.NewGELUActivation",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_4575.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_4641.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_4601.GPT2MLP",
- "__torch__.splitters.___torch_mangle_4619.MiddlePart",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_4646.GPT2MLP",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_4585.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_4583.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_4609.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_4592.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_4579.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_4637.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_4599.NewGELUActivation",
- "__torch__.splitters.___torch_mangle_4634.MiddlePart",
- "__torch__.torch.nn.modules.container.___torch_mangle_4618.ModuleList",
- "__torch__.transformers.pytorch_utils.___torch_mangle_4598.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_4629.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_4607.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_4617.GPT2Block",
- "__torch__.torch.nn.modules.container.___torch_mangle_4633.ModuleList",
- "__torch__.splitters.___torch_mangle_4649.MiddlePart",
- "collections.OrderedDict",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_4610.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.container.___torch_mangle_4648.ModuleList",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_4580.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_4594.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_4612.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_4632.GPT2Block",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_4631.GPT2MLP",
- "__torch__.transformers.pytorch_utils.___torch_mangle_4606.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_4611.LayerNorm",
- "torch._utils._rebuild_tensor_v2",
- "collections.OrderedDict",
- "torch.DoubleStorage",
- "collections.OrderedDict",
- "torch.FloatStorage",
- "torch._utils._rebuild_tensor_v2",
- "torch.jit._pickle.restore_type_tag"
147 MB
- Xet hash:
- ab1918b163406205d8fef5a55c7ec157d2f080d6c2bd687dbab262fde96abebb
- Size of remote file:
- 147 MB
- SHA256:
- 077b58aab92b5f2bc8f489d3cf2af1a7faa6b18e613c98eb74bfcc3b504b762b
·
Xet efficiently stores Large Files inside Git, intelligently splitting files into unique chunks and accelerating uploads and downloads. More info.