download history blame
Detected Pickle imports (73)
- "__torch__.torch.nn.modules.container.___torch_mangle_14839.ModuleList",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14866.Dropout",
- "__torch__.splitters.___torch_mangle_14840.MiddlePart",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14844.Dropout",
- "__torch__.transformers.activations.___torch_mangle_14865.NewGELUActivation",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14817.LayerNorm",
- "__torch__.transformers.activations.___torch_mangle_14835.NewGELUActivation",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14836.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14859.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14829.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14818.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14830.Dropout",
- "__torch__.splitters.___torch_mangle_14825.MiddlePart",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14860.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14867.GPT2MLP",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14864.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14827.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14838.GPT2Block",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14846.GPT2SdpaAttention",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14852.GPT2MLP",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14837.GPT2MLP",
- "torch.BoolStorage",
- "__torch__.transformers.activations.___torch_mangle_14850.NewGELUActivation",
- "__torch__.transformers.activations.___torch_mangle_14820.NewGELUActivation",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14816.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14845.Dropout",
- "__torch__.torch.nn.modules.container.___torch_mangle_14824.ModuleList",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14832.LayerNorm",
- "__torch__.torch.nn.modules.container.___torch_mangle_14854.ModuleList",
- "__torch__.torch.nn.modules.container.___torch_mangle_14869.ModuleList",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14848.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14828.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14853.GPT2Block",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14862.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14849.Conv1D",
- "__torch__.torch.nn.modules.container.___torch_mangle_14871.Sequential",
- "torch.FloatStorage",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14823.GPT2Block",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14826.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14847.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14814.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14822.GPT2MLP",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14851.Dropout",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.splitters.___torch_mangle_14855.MiddlePart",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14861.GPT2SdpaAttention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14857.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14819.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14843.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14842.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14833.Conv1D",
- "__torch__.splitters.___torch_mangle_14872.Part",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14812.Conv1D",
- "__torch__.splitters.___torch_mangle_14870.MiddlePart",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14856.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14841.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14858.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14868.GPT2Block",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14831.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14815.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14863.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14813.Conv1D",
- "collections.OrderedDict",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14821.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14811.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14834.Conv1D",
- "torch._utils._rebuild_tensor_v2",
- "collections.OrderedDict",
- "torch.DoubleStorage",
- "torch.jit._pickle.restore_type_tag",
- "torch.FloatStorage",
- "collections.OrderedDict",
- "torch._utils._rebuild_tensor_v2"
118 MB
- Xet hash:
- 451ae4e4b2e56c6daba20fcc24b7302b5abb3420601330c161d7d7c52b00936f
- Size of remote file:
- 118 MB
- SHA256:
- 10c11a840decb6af4a241acff017cebbad7bf6158651ee78047e43d5bb281c22
·
Xet efficiently stores Large Files inside Git, intelligently splitting files into unique chunks and accelerating uploads and downloads. More info.