download history blame
Detected Pickle imports (103)
- "__torch__.torch.nn.modules.dropout.___torch_mangle_3798.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3760.Conv1D",
- "torch.FloatStorage",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_3803.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_3813.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_3743.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_3754.GPT2MLP",
- "__torch__.transformers.activations.___torch_mangle_3812.NewGELUActivation",
- "__torch__.transformers.activations.___torch_mangle_3797.NewGELUActivation",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_3768.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3796.Conv1D",
- "__torch__.torch.nn.modules.container.___torch_mangle_3801.ModuleList",
- "__torch__.torch.nn.modules.container.___torch_mangle_3786.ModuleList",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_3784.GPT2MLP",
- "__torch__.transformers.activations.___torch_mangle_3767.NewGELUActivation",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_3770.GPT2Block",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3781.Conv1D",
- "__torch__.torch.nn.modules.container.___torch_mangle_3833.Sequential",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3819.Conv1D",
- "__torch__.torch.nn.modules.container.___torch_mangle_3831.ModuleList",
- "torch.BoolStorage",
- "__torch__.transformers.activations.___torch_mangle_3827.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3811.Conv1D",
- "__torch__.torch.nn.modules.container.___torch_mangle_3816.ModuleList",
- "__torch__.torch.nn.modules.container.___torch_mangle_3756.ModuleList",
- "__torch__.splitters.___torch_mangle_3802.MiddlePart",
- "__torch__.splitters.___torch_mangle_3832.MiddlePart",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_3779.LayerNorm",
- "__torch__.splitters.___torch_mangle_3834.Part",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3789.Conv1D",
- "__torch__.splitters.___torch_mangle_3787.MiddlePart",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_3793.GPT2SdpaAttention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3774.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3765.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3795.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_3747.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3780.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_3764.LayerNorm",
- "__torch__.splitters.___torch_mangle_3817.MiddlePart",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_3755.GPT2Block",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3751.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_3821.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3805.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3759.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_3828.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_3829.GPT2MLP",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_3807.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_3808.GPT2SdpaAttention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3766.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_3776.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3775.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_3769.GPT2MLP",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_3814.GPT2MLP",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_3777.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_3800.GPT2Block",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_3794.LayerNorm",
- "__torch__.splitters.___torch_mangle_3757.MiddlePart",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_3748.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_3818.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_3773.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3826.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_3799.GPT2MLP",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_3823.GPT2SdpaAttention",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_3778.GPT2SdpaAttention",
- "__torch__.transformers.activations.___torch_mangle_3782.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3810.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_3758.LayerNorm",
- "__torch__.torch.nn.modules.container.___torch_mangle_3771.ModuleList",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_3763.GPT2SdpaAttention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3804.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_3746.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3825.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_3752.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3745.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_3788.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_3762.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_3761.Dropout",
- "__torch__.splitters.___torch_mangle_3772.MiddlePart",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_3809.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_3753.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_3783.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_3822.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3744.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_3815.GPT2Block",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_3791.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3820.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_3785.GPT2Block",
- "collections.OrderedDict",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_3806.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3790.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_3749.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_3830.GPT2Block",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_3792.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_3824.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_3750.Conv1D",
- "torch._utils._rebuild_tensor_v2",
- "collections.OrderedDict",
- "torch.DoubleStorage",
- "collections.OrderedDict",
- "torch.FloatStorage",
- "torch._utils._rebuild_tensor_v2",
- "torch.jit._pickle.restore_type_tag"
177 MB
- Xet hash:
- 4fdd3434e87f63e09c6c0fa08bf98bf37834ee0b18e3f23f05f9ea5a59605268
- Size of remote file:
- 177 MB
- SHA256:
- 3ade36286aa9344a5be16c41f18a5e4ace27bb7c08c227af01e2cf94f7f26b65
·
Xet efficiently stores Large Files inside Git, intelligently splitting files into unique chunks and accelerating uploads and downloads. More info.