Ctrl+K
- 2L512W-TS-12M Duplicate from georgeyw/2L512W-TS-12M
- 1.52 kB Duplicate from georgeyw/2L512W-TS-12M
- 2L512W-TS-12M_step_12400.pth53.9 MB
Detected Pickle imports (24)
- "torch.BoolStorage",
- "torch.FloatStorage",
- "tokenizers.Tokenizer",
- "torch.float32",
- "transformer_lens.components.PosEmbed",
- "collections.OrderedDict",
- "transformer_lens.hook_points.HookPoint",
- "transformer_lens.components.TransformerBlock",
- "tokenizers.models.Model",
- "transformer_lens.components.Embed",
- "__builtin__.set",
- "tokenizers.AddedToken",
- "torch._utils._rebuild_tensor_v2",
- "torch.nn.modules.linear.Identity",
- "numpy.dtype",
- "torch._utils._rebuild_parameter",
- "transformer_lens.components.Attention",
- "transformer_lens.HookedTransformer.HookedTransformer",
- "transformer_lens.components.Unembed",
- "_codecs.encode",
- "numpy.core.multiarray.scalar",
- "torch.nn.modules.container.ModuleList",
- "transformer_lens.HookedTransformerConfig.HookedTransformerConfig",
- "transformers.models.gpt2.tokenization_gpt2_fast.GPT2TokenizerFast"
xetDuplicate from georgeyw/2L512W-TS-12M - 2L512W-TS-12M_step_12500.pth53.9 MB
Detected Pickle imports (24)
- "torch.BoolStorage",
- "torch.FloatStorage",
- "tokenizers.Tokenizer",
- "torch.float32",
- "transformer_lens.components.PosEmbed",
- "collections.OrderedDict",
- "transformer_lens.hook_points.HookPoint",
- "transformer_lens.components.TransformerBlock",
- "tokenizers.models.Model",
- "transformer_lens.components.Embed",
- "__builtin__.set",
- "tokenizers.AddedToken",
- "torch._utils._rebuild_tensor_v2",
- "torch.nn.modules.linear.Identity",
- "numpy.dtype",
- "torch._utils._rebuild_parameter",
- "transformer_lens.components.Attention",
- "transformer_lens.HookedTransformer.HookedTransformer",
- "transformer_lens.components.Unembed",
- "_codecs.encode",
- "numpy.core.multiarray.scalar",
- "torch.nn.modules.container.ModuleList",
- "transformer_lens.HookedTransformerConfig.HookedTransformerConfig",
- "transformers.models.gpt2.tokenization_gpt2_fast.GPT2TokenizerFast"
xetDuplicate from georgeyw/2L512W-TS-12M - 2L512W-TS-12M_step_12600.pth53.9 MB
Detected Pickle imports (24)
- "torch.BoolStorage",
- "torch.FloatStorage",
- "tokenizers.Tokenizer",
- "torch.float32",
- "transformer_lens.components.PosEmbed",
- "collections.OrderedDict",
- "transformer_lens.hook_points.HookPoint",
- "transformer_lens.components.TransformerBlock",
- "tokenizers.models.Model",
- "transformer_lens.components.Embed",
- "__builtin__.set",
- "tokenizers.AddedToken",
- "torch._utils._rebuild_tensor_v2",
- "torch.nn.modules.linear.Identity",
- "numpy.dtype",
- "torch._utils._rebuild_parameter",
- "transformer_lens.components.Attention",
- "transformer_lens.HookedTransformer.HookedTransformer",
- "transformer_lens.components.Unembed",
- "_codecs.encode",
- "numpy.core.multiarray.scalar",
- "torch.nn.modules.container.ModuleList",
- "transformer_lens.HookedTransformerConfig.HookedTransformerConfig",
- "transformers.models.gpt2.tokenization_gpt2_fast.GPT2TokenizerFast"
xetDuplicate from georgeyw/2L512W-TS-12M - 2L512W-TS-12M_step_12700.pth53.9 MB
Detected Pickle imports (24)
- "torch.BoolStorage",
- "torch.FloatStorage",
- "tokenizers.Tokenizer",
- "torch.float32",
- "transformer_lens.components.PosEmbed",
- "collections.OrderedDict",
- "transformer_lens.hook_points.HookPoint",
- "transformer_lens.components.TransformerBlock",
- "tokenizers.models.Model",
- "transformer_lens.components.Embed",
- "__builtin__.set",
- "tokenizers.AddedToken",
- "torch._utils._rebuild_tensor_v2",
- "torch.nn.modules.linear.Identity",
- "numpy.dtype",
- "torch._utils._rebuild_parameter",
- "transformer_lens.components.Attention",
- "transformer_lens.HookedTransformer.HookedTransformer",
- "transformer_lens.components.Unembed",
- "_codecs.encode",
- "numpy.core.multiarray.scalar",
- "torch.nn.modules.container.ModuleList",
- "transformer_lens.HookedTransformerConfig.HookedTransformerConfig",
- "transformers.models.gpt2.tokenization_gpt2_fast.GPT2TokenizerFast"
xetDuplicate from georgeyw/2L512W-TS-12M - 2L512W-TS-12M_step_12800.pth53.9 MB
Detected Pickle imports (24)
- "torch.BoolStorage",
- "torch.FloatStorage",
- "tokenizers.Tokenizer",
- "torch.float32",
- "transformer_lens.components.PosEmbed",
- "collections.OrderedDict",
- "transformer_lens.hook_points.HookPoint",
- "transformer_lens.components.TransformerBlock",
- "tokenizers.models.Model",
- "transformer_lens.components.Embed",
- "__builtin__.set",
- "tokenizers.AddedToken",
- "torch._utils._rebuild_tensor_v2",
- "torch.nn.modules.linear.Identity",
- "numpy.dtype",
- "torch._utils._rebuild_parameter",
- "transformer_lens.components.Attention",
- "transformer_lens.HookedTransformer.HookedTransformer",
- "transformer_lens.components.Unembed",
- "_codecs.encode",
- "numpy.core.multiarray.scalar",
- "torch.nn.modules.container.ModuleList",
- "transformer_lens.HookedTransformerConfig.HookedTransformerConfig",
- "transformers.models.gpt2.tokenization_gpt2_fast.GPT2TokenizerFast"
xetDuplicate from georgeyw/2L512W-TS-12M - 2L512W-TS-12M_step_12900.pth53.9 MB
Detected Pickle imports (24)
- "torch.BoolStorage",
- "torch.FloatStorage",
- "tokenizers.Tokenizer",
- "torch.float32",
- "transformer_lens.components.PosEmbed",
- "collections.OrderedDict",
- "transformer_lens.hook_points.HookPoint",
- "transformer_lens.components.TransformerBlock",
- "tokenizers.models.Model",
- "transformer_lens.components.Embed",
- "__builtin__.set",
- "tokenizers.AddedToken",
- "torch._utils._rebuild_tensor_v2",
- "torch.nn.modules.linear.Identity",
- "numpy.dtype",
- "torch._utils._rebuild_parameter",
- "transformer_lens.components.Attention",
- "transformer_lens.HookedTransformer.HookedTransformer",
- "transformer_lens.components.Unembed",
- "_codecs.encode",
- "numpy.core.multiarray.scalar",
- "torch.nn.modules.container.ModuleList",
- "transformer_lens.HookedTransformerConfig.HookedTransformerConfig",
- "transformers.models.gpt2.tokenization_gpt2_fast.GPT2TokenizerFast"
xetDuplicate from georgeyw/2L512W-TS-12M - 53.9 MB xetDuplicate from georgeyw/2L512W-TS-12M