download history blame
Detected Pickle imports (107)
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_2261.GPT2SdpaAttention",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_2231.GPT2SdpaAttention",
- "torch.FloatStorage",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_2253.GPT2Block",
- "__torch__.splitters.___torch_mangle_2270.MiddlePart",
- "__torch__.splitters.___torch_mangle_2255.MiddlePart",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_2241.LayerNorm",
- "__torch__.torch.nn.modules.sparse.___torch_mangle_2208.Embedding",
- "__torch__.torch.nn.modules.container.___torch_mangle_2301.Sequential",
- "__torch__.splitters.___torch_mangle_2300.MiddlePart",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_2221.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2227.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_2262.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2213.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2264.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_2226.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2287.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2288.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2263.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_2236.Dropout",
- "torch.BoolStorage",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_2274.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_2276.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.sparse.___torch_mangle_2207.Embedding",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_2281.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_2290.Dropout",
- "__torch__.torch.nn.modules.container.___torch_mangle_2254.ModuleList",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_2277.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_2289.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_2260.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_2245.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_2247.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_2251.Dropout",
- "__torch__.splitters.___torch_mangle_2302.Part",
- "__torch__.splitters.___torch_mangle_2210.FirstPart",
- "__torch__.splitters.___torch_mangle_2285.MiddlePart",
- "__torch__.splitters.___torch_mangle_2240.MiddlePart",
- "__torch__.transformers.activations.___torch_mangle_2220.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2242.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_2266.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2249.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_2295.NewGELUActivation",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_2222.GPT2MLP",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2279.Conv1D",
- "__torch__.torch.nn.modules.container.___torch_mangle_2224.ModuleList",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_2214.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_2246.GPT2SdpaAttention",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_2237.GPT2MLP",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_2275.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2218.Conv1D",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_2216.GPT2SdpaAttention",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_2291.GPT2SdpaAttention",
- "__torch__.transformers.activations.___torch_mangle_2280.NewGELUActivation",
- "__torch__.torch.nn.modules.container.___torch_mangle_2284.ModuleList",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2294.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_2282.GPT2MLP",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_2292.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_2256.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2293.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_2235.NewGELUActivation",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_2217.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_2211.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2257.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_2265.NewGELUActivation",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_2267.GPT2MLP",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2258.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_2283.GPT2Block",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_2215.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_2296.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_2223.GPT2Block",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2234.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_2286.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2272.Conv1D",
- "__torch__.torch.nn.modules.container.___torch_mangle_2299.ModuleList",
- "__torch__.transformers.activations.___torch_mangle_2250.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2219.Conv1D",
- "__torch__.torch.nn.modules.container.___torch_mangle_2269.ModuleList",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_2209.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_2268.GPT2Block",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2212.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_2232.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2233.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_2230.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_2238.GPT2Block",
- "collections.OrderedDict",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_2259.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2228.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_2271.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2273.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_2252.GPT2MLP",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_2229.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_2244.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_2298.GPT2Block",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2248.Conv1D",
- "__torch__.splitters.___torch_mangle_2225.MiddlePart",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_2297.GPT2MLP",
- "__torch__.torch.nn.modules.container.___torch_mangle_2239.ModuleList",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2278.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_2243.Conv1D",
- "torch._utils._rebuild_tensor_v2",
- "collections.OrderedDict",
- "torch.DoubleStorage",
- "collections.OrderedDict",
- "torch._utils._rebuild_tensor_v2",
- "torch.LongStorage",
- "torch.jit._pickle.restore_type_tag"
334 MB
- Xet hash:
- 3060fc0ee6ecc3994dcdc124aedea290a0cd279999fe00575bafbaeee39106e5
- Size of remote file:
- 334 MB
- SHA256:
- 6b88b2c74d03f331e365cdec7293e05bba35d0aa4c27dd07dd95d773d1ffce12
·
Xet efficiently stores Large Files inside Git, intelligently splitting files into unique chunks and accelerating uploads and downloads. More info.