download
history
blame
contribute
delete
Detected Pickle imports (201)
- "torch.FloatStorage",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1537.Conv1D",
- "__torch__.splitters.___torch_mangle_1636.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1563.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1607.Conv1D",
- "__torch__.splitters.___torch_mangle_1604.MiddlePart",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1512.GPT2Block",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1555.Dropout",
- "__torch__.transformers.activations.___torch_mangle_1629.NewGELUActivation",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1467.GPT2Block",
- "__torch__.torch.nn.modules.container.___torch_mangle_1483.ModuleList",
- "__torch__.transformers.activations.___torch_mangle_1524.NewGELUActivation",
- "__torch__.transformers.activations.___torch_mangle_1539.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1538.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1510.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1531.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1488.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1567.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1615.Dropout",
- "__torch__.transformers.activations.___torch_mangle_1584.NewGELUActivation",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1609.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1630.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1593.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1480.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1495.Dropout",
- "__torch__.torch.nn.modules.container.___torch_mangle_1573.ModuleList",
- "__torch__.transformers.activations.___torch_mangle_1599.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1606.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1470.LayerNorm",
- "__torch__.splitters.___torch_mangle_1544.MiddlePart",
- "collections.OrderedDict",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1526.GPT2MLP",
- "__torch__.splitters.___torch_mangle_1634.MiddlePart",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1583.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1612.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1565.GPT2SdpaAttention",
- "__torch__.transformers.activations.___torch_mangle_1614.NewGELUActivation",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1556.GPT2MLP",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1472.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1535.GPT2SdpaAttention",
- "__torch__.splitters.___torch_mangle_1499.MiddlePart",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1582.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1585.Dropout",
- "__torch__.torch.nn.modules.sparse.___torch_mangle_1451.Embedding",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1466.GPT2MLP",
- "__torch__.splitters.___torch_mangle_1484.MiddlePart",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1471.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1541.GPT2MLP",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1561.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1620.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1622.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1592.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1490.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1626.LayerNorm",
- "__torch__.torch.nn.modules.sparse.___torch_mangle_1452.Embedding",
- "__torch__.splitters.___torch_mangle_1574.MiddlePart",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1493.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1525.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1635.LayerNorm",
- "__torch__.torch.nn.modules.container.___torch_mangle_1468.ModuleList",
- "__torch__.transformers.activations.___torch_mangle_1479.NewGELUActivation",
- "__torch__.splitters.___torch_mangle_1559.MiddlePart",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1511.GPT2MLP",
- "__torch__.splitters.___torch_mangle_1514.MiddlePart",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1570.Dropout",
- "__torch__.torch.nn.modules.container.___torch_mangle_1558.ModuleList",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1457.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1594.Dropout",
- "__torch__.transformers.activations.___torch_mangle_1494.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1598.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1481.GPT2MLP",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1625.GPT2SdpaAttention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1547.Conv1D",
- "__torch__.transformers.activations.___torch_mangle_1569.NewGELUActivation",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1560.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1608.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1595.GPT2SdpaAttention",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1557.GPT2Block",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1485.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1462.Conv1D",
- "__torch__.torch.nn.modules.container.___torch_mangle_1603.ModuleList",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1459.Dropout",
- "__torch__.torch.nn.modules.container.___torch_mangle_1639.Sequential",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1496.GPT2MLP",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1553.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1487.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1576.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1577.Conv1D",
- "__torch__.splitters.___torch_mangle_1638.FinalPart",
- "__torch__.torch.nn.modules.container.___torch_mangle_1618.ModuleList",
- "__torch__.torch.nn.modules.container.___torch_mangle_1543.ModuleList",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1597.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1532.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1500.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1521.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1590.LayerNorm",
- "torch.BoolStorage",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1575.LayerNorm",
- "__torch__.splitters.___torch_mangle_1469.MiddlePart",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1523.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1611.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1473.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1474.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1478.Conv1D",
- "__torch__.splitters.___torch_mangle_1640.Part",
- "__torch__.torch.nn.modules.container.___torch_mangle_1528.ModuleList",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1542.GPT2Block",
- "__torch__.transformers.activations.___torch_mangle_1464.NewGELUActivation",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1602.GPT2Block",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1546.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1534.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1617.GPT2Block",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1551.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1520.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1503.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1628.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1581.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1465.Dropout",
- "__torch__.splitters.___torch_mangle_1589.MiddlePart",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1517.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1566.LayerNorm",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1458.Dropout",
- "__torch__.splitters.___torch_mangle_1454.FirstPart",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1578.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1562.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1491.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1550.GPT2SdpaAttention",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1631.GPT2MLP",
- "__torch__.transformers.activations.___torch_mangle_1554.NewGELUActivation",
- "__torch__.torch.nn.modules.container.___torch_mangle_1633.ModuleList",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1502.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1456.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1518.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1601.GPT2MLP",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1461.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1492.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1545.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1632.GPT2Block",
- "__torch__.splitters.___torch_mangle_1529.MiddlePart",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1623.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_1637.Linear",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1587.GPT2Block",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1463.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1507.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1515.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1530.LayerNorm",
- "__torch__.torch.nn.modules.container.___torch_mangle_1513.ModuleList",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1536.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1501.Conv1D",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1605.LayerNorm",
- "__torch__.torch.nn.modules.container.___torch_mangle_1588.ModuleList",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1600.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1596.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1477.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1504.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1527.GPT2Block",
- "__torch__.transformers.activations.___torch_mangle_1509.NewGELUActivation",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1516.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1552.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1613.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1572.GPT2Block",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1580.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1564.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1549.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1453.Dropout",
- "__torch__.torch.nn.modules.container.___torch_mangle_1498.ModuleList",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1624.Dropout",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1455.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1506.LayerNorm",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1475.GPT2SdpaAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1519.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1548.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1579.Dropout",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1591.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1489.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1571.GPT2MLP",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1610.GPT2SdpaAttention",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1616.GPT2MLP",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_1476.LayerNorm",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1627.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1482.GPT2Block",
- "__torch__.splitters.___torch_mangle_1619.MiddlePart",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1540.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1505.GPT2SdpaAttention",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1586.GPT2MLP",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1568.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1621.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1508.Conv1D",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1460.GPT2SdpaAttention",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1486.Conv1D",
- "__torch__.transformers.pytorch_utils.___torch_mangle_1522.Conv1D",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_1533.Dropout",
- "__torch__.transformers.models.gpt2.modeling_gpt2.___torch_mangle_1497.GPT2Block",
- "torch.DoubleStorage",
- "collections.OrderedDict",
- "torch._utils._rebuild_tensor_v2",
- "torch.LongStorage",
- "torch.jit._pickle.restore_type_tag",
- "collections.OrderedDict",
- "torch._utils._rebuild_tensor_v2"
511 MB
- SHA256:
- 7bd55abc33779ea131ba879acd582f613e21f37412d230a33412c829fa615ed8
- Pointer size:
- 134 Bytes
- Size of remote file:
- 511 MB
- Xet hash:
- cbd500e3521c6b3ee528af484dec51bebc3c8c3022fbbe7252f15c36967f4f21
·
·
Xet efficiently stores Large Files inside Git, intelligently splitting files into unique chunks and accelerating uploads and downloads. More info.