download history blame contribute delete
Detected Pickle imports (182)
- "__torch__.torch.nn.modules.dropout.___torch_mangle_466.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_398.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_390.GPT2Attention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_404.LayerNorm",
- "__torch__.transformers.activations.___torch_mangle_355.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_347.Conv1D",
- "__torch__.torch.classes.quantized.LinearPackedParamsBase",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_325.GPT2Attention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_457.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_319.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_343.Dropout",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.transformers.pytorch_utils.___torch_mangle_464.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_401.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_360.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_396.GPT2MLP",
- "__torch__.transformers.activations.___torch_mangle_407.NewGELUActivation",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_454.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_332.GPT2Block",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_331.GPT2MLP",
- "__torch__.transformers.pytorch_utils.___torch_mangle_445.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_372.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_356.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_444.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_357.GPT2MLP",
- "__torch__.transformers.activations.___torch_mangle_329.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_366.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_472.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_405.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_364.GPT2Attention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_451.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_379.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_459.NewGELUActivation",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_455.GPT2Attention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_408.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_470.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_478.GPT2Model",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_391.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_421.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_353.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_469.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_456.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_465.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_399.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_349.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_351.GPT2Attention",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_462.GPT2Block",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_473.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_346.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_435.GPT2MLP",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_441.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_460.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_373.Conv1D",
- "__torch__.torch.nn.modules.sparse.___torch_mangle_318.Embedding",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_409.GPT2MLP",
- "__torch__.transformers.pytorch_utils.___torch_mangle_328.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_383.GPT2MLP",
- "__torch__.transformers.pytorch_utils.___torch_mangle_374.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_418.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_324.Dropout",
- "__torch__.transformers.activations.___torch_mangle_446.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_432.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_437.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_339.LayerNorm",
- "__torch__.torch.nn.modules.sparse.___torch_mangle_317.Embedding",
- "__torch__.transformers.pytorch_utils.___torch_mangle_431.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_323.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_426.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_415.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_344.GPT2MLP",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_440.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_443.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_335.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_425.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_387.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_422.GPT2MLP",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_388.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_369.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_375.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_402.Dropout",
- "collections.OrderedDict",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_330.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_397.GPT2Block",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_410.GPT2Block",
- "__torch__.transformers.pytorch_utils.___torch_mangle_412.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_342.NewGELUActivation",
- "__torch__.transformers.activations.___torch_mangle_368.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_386.Conv1D",
- "torch.BoolStorage",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_370.GPT2MLP",
- "torch.FloatStorage",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_338.GPT2Attention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_458.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_376.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_384.GPT2Block",
- "__torch__.transformers.pytorch_utils.___torch_mangle_400.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_416.GPT2Attention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_363.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_321.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_352.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_406.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_471.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_378.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_453.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_461.GPT2MLP",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_385.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_430.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_393.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_358.GPT2Block",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_467.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_417.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_481.GPT2LMHeadModel",
- "__torch__.transformers.pytorch_utils.___torch_mangle_341.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_477.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_413.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_361.Conv1D",
- "torch.per_tensor_affine",
- "__torch__.transformers.pytorch_utils.___torch_mangle_354.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_474.GPT2MLP",
- "__torch__.transformers.pytorch_utils.___torch_mangle_334.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_424.LayerNorm",
- "torch.QInt8Storage",
- "__torch__.torch.ao.nn.quantized.dynamic.modules.linear.___torch_mangle_480.Linear",
- "__torch__.transformers.pytorch_utils.___torch_mangle_327.Conv1D",
- "torch._utils._rebuild_qtensor",
- "__torch__.transformers.pytorch_utils.___torch_mangle_322.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_427.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_475.GPT2Block",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_350.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_447.Dropout",
- "__torch__.transformers.activations.___torch_mangle_420.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_439.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_320.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_450.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_367.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_452.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_449.GPT2Block",
- "__torch__.transformers.pytorch_utils.___torch_mangle_348.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_380.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_365.LayerNorm",
- "__torch__.torch.ao.nn.quantized.modules.linear.___torch_mangle_479.LinearPackedParams",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_448.GPT2MLP",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_345.GPT2Block",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_429.GPT2Attention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_392.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_428.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_389.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_438.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_442.GPT2Attention",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_403.GPT2Attention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_411.LayerNorm",
- "__torch__.transformers.activations.___torch_mangle_394.NewGELUActivation",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_434.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_414.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_336.Dropout",
- "__torch__.torch.nn.modules.container.___torch_mangle_476.ModuleList",
- "__torch__.transformers.activations.___torch_mangle_433.NewGELUActivation",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_382.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_377.GPT2Attention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_340.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_371.GPT2Block",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_463.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_362.Dropout",
- "__torch__.TracedModel",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_423.GPT2Block",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_333.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_337.Dropout",
- "__torch__.transformers.activations.___torch_mangle_381.NewGELUActivation",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_395.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_359.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_436.GPT2Block",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_468.GPT2Attention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_419.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_326.LayerNorm",
- "torch.DoubleStorage",
- "torch._utils._rebuild_tensor_v2",
- "collections.OrderedDict",
- "torch.LongStorage",
- "torch.jit._pickle.restore_type_tag",
- "torch._utils._rebuild_tensor_v2",
- "collections.OrderedDict",
- "torch.LongStorage"
549 MB
- Xet hash:
- f1f0cb72d7f23f466b4844ec1cce966c3302956f9ab8fdb43e25b987c95b9aed
- Size of remote file:
- 549 MB
- SHA256:
- 1579966d38fe802092f4c5cb214986833fc16449735299044ea0d9018613ca4f
·
Xet efficiently stores Large Files inside Git, intelligently splitting files into unique chunks and accelerating uploads and downloads. More info.