download history blame contribute delete
Detected Pickle imports (182)
- "__torch__.transformers.pytorch_utils.___torch_mangle_1383.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_1326.NewGELUActivation",
- "torch.QInt8Storage",
- "__torch__.transformers.activations.___torch_mangle_1443.NewGELUActivation",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1374.GPT2Attention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1427.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1465.GPT2Attention",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1367.GPT2MLP",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1392.Dropout",
- "__torch__.torch.nn.modules.sparse.___torch_mangle_1315.Embedding",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1398.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1403.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1337.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1354.GPT2MLP",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1455.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1410.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1358.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1317.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1393.GPT2MLP",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1356.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1340.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1420.GPT2Block",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1401.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1322.GPT2Attention",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1433.GPT2Block",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1452.GPT2Attention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1344.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1418.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1451.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1472.GPT2Block",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1382.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1448.Conv1D",
- "torch.BoolStorage",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1478.GPT2LMHeadModel",
- "__torch__.transformers.activations.___torch_mangle_1365.NewGELUActivation",
- "collections.OrderedDict",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1350.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1409.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1321.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1440.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1444.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1438.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1364.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1461.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1414.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1458.GPT2MLP",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1370.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1376.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1471.GPT2MLP",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1424.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1407.GPT2Block",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1327.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1466.LayerNorm",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1329.GPT2Block",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1372.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1355.GPT2Block",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1408.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1380.GPT2MLP",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1360.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1366.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1462.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_1469.NewGELUActivation",
- "__torch__.torch.ao.nn.quantized.modules.linear.___torch_mangle_1476.LinearPackedParams",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1373.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1394.GPT2Block",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1436.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1437.Dropout",
- "__torch__.transformers.activations.___torch_mangle_1378.NewGELUActivation",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1362.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1385.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1395.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1425.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1470.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1319.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_1404.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1442.Conv1D",
- "__torch__.torch.nn.modules.container.___torch_mangle_1473.ModuleList",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1384.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1399.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1453.LayerNorm",
- "torch.per_tensor_affine",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1464.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1419.GPT2MLP",
- "torch.FloatStorage",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1441.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1406.GPT2MLP",
- "__torch__.transformers.activations.___torch_mangle_1430.NewGELUActivation",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1446.GPT2Block",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1447.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1368.GPT2Block",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1400.GPT2Attention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1468.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1363.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1318.Conv1D",
- "__torch__.torch.classes.quantized.LinearPackedParamsBase",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1345.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1349.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1357.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1369.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1434.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1460.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1323.LayerNorm",
- "__torch__.transformers.activations.___torch_mangle_1417.NewGELUActivation",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1431.Dropout",
- "__torch__.___torch_mangle_1479.TracedModel",
- "__torch__.transformers.activations.___torch_mangle_1391.NewGELUActivation",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1412.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1463.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1405.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1342.GPT2Block",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1397.Conv1D",
- "torch._utils._rebuild_qtensor",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1332.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1379.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1375.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1423.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_1339.NewGELUActivation",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1450.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1347.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1413.GPT2Attention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1346.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1325.Conv1D",
- "__torch__.torch.ao.nn.quantized.dynamic.modules.linear.___torch_mangle_1477.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1421.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1449.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1371.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1390.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1316.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1411.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1426.GPT2Attention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1429.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1415.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1422.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1402.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1388.LayerNorm",
- "__torch__.torch.nn.modules.sparse.___torch_mangle_1314.Embedding",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1432.GPT2MLP",
- "__torch__.transformers.activations.___torch_mangle_1456.NewGELUActivation",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1475.GPT2Model",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1336.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1361.GPT2Attention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1324.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1338.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1341.GPT2MLP",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1454.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1396.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1351.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1435.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1416.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_1352.NewGELUActivation",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1330.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1335.GPT2Attention",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1387.GPT2Attention",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1381.GPT2Block",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1445.GPT2MLP",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1353.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1386.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1359.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1343.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1377.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1333.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1348.GPT2Attention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1320.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1328.GPT2MLP",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1474.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1467.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1439.GPT2Attention",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1459.GPT2Block",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1334.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1389.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1457.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1428.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1331.Conv1D",
- "torch._utils._rebuild_tensor_v2",
- "torch.LongStorage",
- "collections.OrderedDict",
- "torch.DoubleStorage",
- "torch._utils._rebuild_tensor_v2",
- "torch.LongStorage",
- "collections.OrderedDict",
- "torch.jit._pickle.restore_type_tag"
549 MB
- Xet hash:
- c774e72520610ad76159cc0f4acd2a7c7688bd140b38de0aadf2f5000bb27268
- Size of remote file:
- 549 MB
- SHA256:
- 03232b4d50ebc0b220e10746b779bc253b8bf64912cb1f811089498d6ca4c79f
·
Xet efficiently stores Large Files inside Git, intelligently splitting files into unique chunks and accelerating uploads and downloads. More info.