download history blame
Detected Pickle imports (103)
- "__torch__.transformers.pytorch_utils.___torch_mangle_14168.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14192.Dropout",
- "__torch__.splitters.___torch_mangle_14198.Part",
- "__torch__.splitters.___torch_mangle_14136.MiddlePart",
- "torch.BoolStorage",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14185.Dropout",
- "__torch__.splitters.___torch_mangle_14151.MiddlePart",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14112.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14107.LayerNorm",
- "__torch__.splitters.___torch_mangle_14121.MiddlePart",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14141.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14109.Conv1D",
- "__torch__.torch.nn.modules.container.___torch_mangle_14197.Sequential",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14163.GPT2MLP",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14143.LayerNorm",
- "__torch__.torch.nn.modules.container.___torch_mangle_14150.ModuleList",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14111.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14183.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14186.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14130.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14132.Dropout",
- "__torch__.transformers.activations.___torch_mangle_14191.NewGELUActivation",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14157.GPT2SdpaAttention",
- "collections.OrderedDict",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14164.GPT2Block",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14158.LayerNorm",
- "__torch__.transformers.activations.___torch_mangle_14176.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14139.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14171.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14169.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14160.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14193.GPT2MLP",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14149.GPT2Block",
- "torch.FloatStorage",
- "__torch__.torch.nn.modules.container.___torch_mangle_14165.ModuleList",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14148.GPT2MLP",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14184.Conv1D",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14124.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14113.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14122.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14137.LayerNorm",
- "__torch__.splitters.___torch_mangle_14166.MiddlePart",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14145.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14152.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14134.GPT2Block",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14129.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14133.GPT2MLP",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14117.Dropout",
- "__torch__.torch.nn.modules.container.___torch_mangle_14195.ModuleList",
- "__torch__.transformers.activations.___torch_mangle_14146.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14115.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14189.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14144.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14114.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_14131.NewGELUActivation",
- "__torch__.torch.nn.modules.container.___torch_mangle_14180.ModuleList",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14127.GPT2SdpaAttention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14123.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14182.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14154.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14177.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14142.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14147.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14118.GPT2MLP",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14156.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14110.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14190.Conv1D",
- "__torch__.splitters.___torch_mangle_14181.MiddlePart",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14119.GPT2Block",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14162.Dropout",
- "__torch__.torch.nn.modules.container.___torch_mangle_14120.ModuleList",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14126.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14174.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14155.Dropout",
- "__torch__.splitters.___torch_mangle_14196.MiddlePart",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14194.GPT2Block",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14108.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14153.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14187.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.container.___torch_mangle_14135.ModuleList",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14138.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14159.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_14161.NewGELUActivation",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14172.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14140.Dropout",
- "__torch__.transformers.activations.___torch_mangle_14116.NewGELUActivation",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14167.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14128.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14175.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14179.GPT2Block",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14178.GPT2MLP",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14188.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14125.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14173.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14170.Dropout",
- "torch._utils._rebuild_tensor_v2",
- "collections.OrderedDict",
- "torch.DoubleStorage",
- "torch.jit._pickle.restore_type_tag",
- "torch.FloatStorage",
- "collections.OrderedDict",
- "torch._utils._rebuild_tensor_v2"
177 MB
- Xet hash:
- 26224082b5c904f7917e6844da5123c9ecb88a85dbb0e4189c620831d0e4403b
- Size of remote file:
- 177 MB
- SHA256:
- fbb33507b1ed7d1d96e4d53f24641090041f1ae20f23e5b2764993cc478412f9
·
Xet efficiently stores Large Files inside Git, intelligently splitting files into unique chunks and accelerating uploads and downloads. More info.