download history blame
Detected Pickle imports (75)
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14971.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14946.GPT2MLP",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14977.GPT2Block",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14953.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14942.Conv1D",
- "__torch__.torch.nn.modules.container.___torch_mangle_14978.ModuleList",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14973.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14987.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14985.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.container.___torch_mangle_14963.ModuleList",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14968.Dropout",
- "torch.BoolStorage",
- "__torch__.aux.___torch_mangle_14964.MiddlePart",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14958.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14938.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14984.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14988.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14995.LayerNorm",
- "__torch__.aux.___torch_mangle_14994.MiddlePart",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14975.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14970.GPT2SdpaAttention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14967.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14965.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14962.GPT2Block",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14976.GPT2MLP",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14972.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14939.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14990.Dropout",
- "__torch__.aux.___torch_mangle_14949.MiddlePart",
- "torch.FloatStorage",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14980.LayerNorm",
- "collections.OrderedDict",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14961.GPT2MLP",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14992.GPT2Block",
- "__torch__.torch.nn.modules.container.___torch_mangle_14993.ModuleList",
- "__torch__.torch.nn.modules.container.___torch_mangle_14948.ModuleList",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14947.GPT2Block",
- "__torch__.transformers.activations.___torch_mangle_14944.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14952.Conv1D",
- "__torch__.torch.nn.modules.container.___torch_mangle_14997.Sequential",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14945.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14966.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14954.Dropout",
- "__torch__.aux.___torch_mangle_14996.LayerNorm",
- "__torch__.transformers.activations.___torch_mangle_14974.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14981.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14936.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_14989.NewGELUActivation",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14935.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14982.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14986.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14957.Conv1D",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14956.LayerNorm",
- "__torch__.aux.___torch_mangle_14979.MiddlePart",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14950.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14937.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14955.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14960.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14983.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14991.GPT2MLP",
- "__torch__.aux.___torch_mangle_14998.Part",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14943.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14951.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_14959.NewGELUActivation",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14941.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14940.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14969.Dropout",
- "torch.DoubleStorage",
- "collections.OrderedDict",
- "torch._utils._rebuild_tensor_v2",
- "torch.jit._pickle.restore_type_tag",
- "torch.FloatStorage",
- "collections.OrderedDict",
- "torch._utils._rebuild_tensor_v2"
118 MB
- Xet hash:
- da700abba6aaf9666852ba141440409823d8f9e9a68f27f64e4245e6ab806d32
- Size of remote file:
- 118 MB
- SHA256:
- 79b6b0b61fcd1f7b2eac88954b18bb97df45f84fcbdb92d6ec9e6db1a6277f8e
·
Xet efficiently stores Large Files inside Git, intelligently splitting files into unique chunks and accelerating uploads and downloads. More info.