download history blame
Detected Pickle imports (107)
- "__torch__.torch.nn.modules.dropout.___torch_mangle_13953.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13957.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_13990.GPT2MLP",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_13999.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.sparse.___torch_mangle_13915.Embedding",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13920.Conv1D",
- "torch.BoolStorage",
- "__torch__.torch.nn.modules.container.___torch_mangle_13977.ModuleList",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14006.GPT2Block",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_13994.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13921.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_13940.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_13998.Dropout",
- "__torch__.torch.nn.modules.container.___torch_mangle_13962.ModuleList",
- "__torch__.splitters.___torch_mangle_14010.Part",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13981.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_13984.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.sparse.___torch_mangle_13916.Embedding",
- "__torch__.splitters.___torch_mangle_13963.MiddlePart",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_13974.Dropout",
- "__torch__.transformers.activations.___torch_mangle_14003.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13996.Conv1D",
- "__torch__.torch.nn.modules.container.___torch_mangle_14007.ModuleList",
- "collections.OrderedDict",
- "__torch__.transformers.activations.___torch_mangle_13928.NewGELUActivation",
- "__torch__.torch.nn.modules.container.___torch_mangle_13932.ModuleList",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_13976.GPT2Block",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_13946.GPT2Block",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_13997.Dropout",
- "__torch__.torch.nn.modules.container.___torch_mangle_13992.ModuleList",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_13969.GPT2SdpaAttention",
- "torch.FloatStorage",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13926.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_13967.Dropout",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13956.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_13943.NewGELUActivation",
- "__torch__.splitters.___torch_mangle_13993.MiddlePart",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14002.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_13931.GPT2Block",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13972.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13965.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_13988.NewGELUActivation",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_13975.GPT2MLP",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_13923.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_13929.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13971.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_13985.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_13959.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_13922.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_13917.Dropout",
- "__torch__.splitters.___torch_mangle_13978.MiddlePart",
- "__torch__.transformers.activations.___torch_mangle_13958.NewGELUActivation",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_13934.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_13979.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13951.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_13930.GPT2MLP",
- "__torch__.splitters.___torch_mangle_14008.MiddlePart",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_13954.GPT2SdpaAttention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13936.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_13982.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_14005.GPT2MLP",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_13924.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_13968.Dropout",
- "__torch__.transformers.activations.___torch_mangle_13973.NewGELUActivation",
- "__torch__.splitters.___torch_mangle_13933.MiddlePart",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_14000.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_13919.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_14001.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_13989.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13950.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_13938.Dropout",
- "__torch__.splitters.___torch_mangle_13918.FirstPart",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_13952.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_13949.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_13961.GPT2Block",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_13970.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13942.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13966.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_13960.GPT2MLP",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_13955.LayerNorm",
- "__torch__.torch.nn.modules.container.___torch_mangle_13947.ModuleList",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_13944.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13941.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_13991.GPT2Block",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13986.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_14004.Dropout",
- "__torch__.splitters.___torch_mangle_13948.MiddlePart",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13995.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13935.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13987.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13927.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_13964.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_13983.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_13980.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_13939.GPT2SdpaAttention",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_13945.GPT2MLP",
- "__torch__.torch.nn.modules.container.___torch_mangle_14009.Sequential",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_13937.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_13925.LayerNorm",
- "torch._utils._rebuild_tensor_v2",
- "collections.OrderedDict",
- "torch.DoubleStorage",
- "torch.jit._pickle.restore_type_tag",
- "torch.LongStorage",
- "collections.OrderedDict",
- "torch._utils._rebuild_tensor_v2"
334 MB
- Xet hash:
- d6198f0eb7a7bc2482aa273f15cc76a015475c606ce0490cff569ac3a904e78a
- Size of remote file:
- 334 MB
- SHA256:
- 4feaa64327d17762ad0a8fad9f05d4decbbb6a3501fcab5a5f9a3256045336ed
·
Xet efficiently stores Large Files inside Git, intelligently splitting files into unique chunks and accelerating uploads and downloads. More info.