download history blame
Detected Pickle imports (88)
- "torch.FloatStorage",
- "__torch__.transformers.activations.___torch_mangle_7926.NewGELUActivation",
- "__torch__.splitters.___torch_mangle_7933.Part",
- "__torch__.transformers.pytorch_utils.___torch_mangle_7904.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_7923.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_7909.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_7920.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_7907.GPT2SdpaAttention",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_7883.GPT2MLP",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_7913.GPT2MLP",
- "__torch__.torch.nn.modules.container.___torch_mangle_7932.Sequential",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_7905.Dropout",
- "__torch__.torch.nn.modules.container.___torch_mangle_7900.ModuleList",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_7927.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_7902.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_7884.GPT2Block",
- "torch.BoolStorage",
- "__torch__.transformers.activations.___torch_mangle_7881.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_7919.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_7861.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_7864.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_7863.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_7859.Conv1D",
- "__torch__.torch.nn.modules.container.___torch_mangle_7915.ModuleList",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_7887.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_7889.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_7877.GPT2SdpaAttention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_7858.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_7912.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_7910.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_7890.Dropout",
- "__torch__.torch.nn.modules.container.___torch_mangle_7930.ModuleList",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_7906.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_7898.GPT2MLP",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_7857.LayerNorm",
- "__torch__.splitters.___torch_mangle_7901.MiddlePart",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_7860.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_7899.GPT2Block",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_7914.GPT2Block",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_7862.GPT2SdpaAttention",
- "__torch__.transformers.activations.___torch_mangle_7896.NewGELUActivation",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_7922.GPT2SdpaAttention",
- "__torch__.splitters.___torch_mangle_7886.MiddlePart",
- "__torch__.splitters.___torch_mangle_7871.MiddlePart",
- "__torch__.splitters.___torch_mangle_7931.MiddlePart",
- "__torch__.transformers.pytorch_utils.___torch_mangle_7918.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_7874.Conv1D",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_7867.Dropout",
- "__torch__.torch.nn.modules.container.___torch_mangle_7885.ModuleList",
- "__torch__.transformers.pytorch_utils.___torch_mangle_7894.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_7921.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_7917.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_7891.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_7928.GPT2MLP",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_7882.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_7879.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_7869.GPT2Block",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_7892.GPT2SdpaAttention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_7873.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_7866.NewGELUActivation",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_7872.LayerNorm",
- "__torch__.splitters.___torch_mangle_7916.MiddlePart",
- "__torch__.transformers.pytorch_utils.___torch_mangle_7924.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_7895.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_7868.GPT2MLP",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_7893.LayerNorm",
- "__torch__.transformers.activations.___torch_mangle_7911.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_7865.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_7908.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_7925.Conv1D",
- "collections.OrderedDict",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_7897.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_7903.Conv1D",
- "__torch__.torch.nn.modules.container.___torch_mangle_7870.ModuleList",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_7878.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_7880.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_7875.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_7888.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_7929.GPT2Block",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_7876.Dropout",
- "torch._utils._rebuild_tensor_v2",
- "collections.OrderedDict",
- "torch.DoubleStorage",
- "collections.OrderedDict",
- "torch.FloatStorage",
- "torch._utils._rebuild_tensor_v2",
- "torch.jit._pickle.restore_type_tag"
147 MB
- Xet hash:
- 839fc2f6eed7388deb133ffe273c58b6c5500531976d2d09a66975cc7778c32a
- Size of remote file:
- 147 MB
- SHA256:
- 654e55acdabe102b71f4dfb296a6c4832ef88f473ddb49335d52b39df7c7502e
·
Xet efficiently stores Large Files inside Git, intelligently splitting files into unique chunks and accelerating uploads and downloads. More info.