download
history
blame
contribute
delete
Detected Pickle imports (183)
- "__torch__.zipformer.NonlinAttention",
- "__torch__.scaling.___torch_mangle_5.ScheduledFloat",
- "__torch__.zipformer.SimpleUpsample",
- "__torch__.model.AsrModel",
- "__torch__.zipformer.___torch_mangle_13.FeedforwardModule",
- "__torch__.torch.nn.modules.sparse.Embedding",
- "__torch__.zipformer.___torch_mangle_55.Zipformer2EncoderLayer",
- "__torch__.torch.nn.modules.linear.___torch_mangle_10.Linear",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.joiner.Joiner",
- "__torch__.zipformer.___torch_mangle_32.NonlinAttention",
- "__torch__.scaling.SwooshR",
- "__torch__.torch.nn.modules.linear.___torch_mangle_11.Linear",
- "__torch__.zipformer.Zipformer2",
- "__torch__.zipformer.BypassModule",
- "__torch__.torch.nn.modules.linear.___torch_mangle_53.Linear",
- "__torch__.zipformer.___torch_mangle_21.RelPositionMultiheadAttentionWeights",
- "__torch__.zipformer.DownsampledZipformer2Encoder",
- "__torch__.scaling.___torch_mangle_43.ChunkCausalDepthwiseConv1d",
- "__torch__.zipformer.___torch_mangle_26.FeedforwardModule",
- "__torch__.zipformer.Zipformer2Encoder",
- "__torch__.torch.nn.modules.linear.___torch_mangle_63.Linear",
- "__torch__.scaling.ActivationDropoutAndLinear",
- "__torch__.zipformer.___torch_mangle_48.DownsampledZipformer2Encoder",
- "__torch__.torch.nn.modules.linear.___torch_mangle_20.Linear",
- "__torch__.zipformer.Zipformer2EncoderLayer",
- "__torch__.subsampling.Conv2dSubsampling",
- "__torch__.zipformer.___torch_mangle_15.FeedforwardModule",
- "__torch__.zipformer.___torch_mangle_51.RelPositionMultiheadAttentionWeights",
- "__torch__.torch.nn.modules.conv.Conv2d",
- "__torch__.torch.nn.modules.conv.Conv1d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_62.Linear",
- "__torch__.zipformer.___torch_mangle_57.Zipformer2Encoder",
- "__torch__.torch.nn.modules.linear.___torch_mangle_61.Linear",
- "__torch__.torch.nn.modules.container.___torch_mangle_59.ModuleList",
- "__torch__.torch.nn.modules.linear.___torch_mangle_9.Linear",
- "__torch__.scaling.ScheduledFloat",
- "__torch__.torch.nn.modules.linear.Linear",
- "__torch__.StreamingEncoderModel",
- "__torch__.zipformer.___torch_mangle_38.Zipformer2EncoderLayer",
- "__torch__.torch.nn.modules.container.Sequential",
- "__torch__.torch.nn.modules.container.___torch_mangle_39.ModuleList",
- "__torch__.subsampling.ConvNeXt",
- "__torch__.scaling.___torch_mangle_36.ChunkCausalDepthwiseConv1d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_29.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_8.Linear",
- "collections.OrderedDict",
- "__torch__.torch.nn.modules.linear.Identity",
- "__torch__.torch.nn.modules.linear.___torch_mangle_50.Linear",
- "__torch__.zipformer.FeedforwardModule",
- "__torch__.zipformer.___torch_mangle_37.ConvolutionModule",
- "__torch__.torch.nn.modules.linear.___torch_mangle_33.Linear",
- "__torch__.zipformer.SimpleDownsample",
- "__torch__.torch.nn.modules.conv.___torch_mangle_60.Conv1d",
- "__torch__.zipformer.___torch_mangle_40.Zipformer2Encoder",
- "__torch__.torch.nn.modules.linear.___torch_mangle_31.Linear",
- "__torch__.zipformer.SelfAttention",
- "__torch__.torch.nn.modules.conv.___torch_mangle_41.Conv1d",
- "__torch__.torch.nn.modules.conv.___torch_mangle_35.Conv1d",
- "__torch__.zipformer.___torch_mangle_58.DownsampledZipformer2Encoder",
- "__torch__.zipformer.___torch_mangle_24.SelfAttention",
- "__torch__.zipformer.___torch_mangle_47.Zipformer2Encoder",
- "torch.FloatStorage",
- "__torch__.zipformer.___torch_mangle_28.FeedforwardModule",
- "__torch__.torch.nn.modules.linear.___torch_mangle_17.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_42.Conv1d",
- "__torch__.scaling.___torch_mangle_19.ActivationDropoutAndLinear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_4.Conv2d",
- "__torch__.torch.nn.modules.activation.Sigmoid",
- "__torch__.torch.nn.modules.linear.___torch_mangle_22.Linear",
- "__torch__.zipformer.___torch_mangle_6.BypassModule",
- "__torch__.torch.nn.modules.linear.___torch_mangle_14.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_16.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_18.Conv1d",
- "__torch__.scaling.SwooshL",
- "__torch__.torch.nn.modules.conv.___torch_mangle_2.Conv2d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_7.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_23.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_0.Conv2d",
- "__torch__.scaling.BiasNorm",
- "__torch__.scaling.Identity",
- "__torch__.zipformer.___torch_mangle_54.SelfAttention",
- "__torch__.torch.nn.modules.container.___torch_mangle_46.ModuleList",
- "__torch__.torch.nn.modules.container.___torch_mangle_56.ModuleList",
- "__torch__.scaling.ChunkCausalDepthwiseConv1d",
- "__torch__.torch.nn.modules.activation.Tanh",
- "__torch__.torch.nn.modules.container.ModuleList",
- "__torch__.zipformer.ConvolutionModule",
- "__torch__.zipformer.___torch_mangle_30.FeedforwardModule",
- "__torch__.scaling.Dropout2",
- "__torch__.torch.nn.modules.conv.___torch_mangle_34.Conv1d",
- "torch.jit._pickle.build_intlist",
- "__torch__.torch.nn.modules.linear.___torch_mangle_52.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_3.Conv2d",
- "__torch__.zipformer.___torch_mangle_44.ConvolutionModule",
- "__torch__.zipformer.CompactRelPositionalEncoding",
- "__torch__.zipformer.RelPositionMultiheadAttentionWeights",
- "__torch__.decoder.Decoder",
- "__torch__.zipformer.___torch_mangle_45.Zipformer2EncoderLayer",
- "__torch__.torch.nn.modules.conv.___torch_mangle_1.Conv2d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_12.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_25.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_49.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_27.Linear",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist"
108 MB
- SHA256:
- ed0db279e28e77d1989cfac7b0ce59f3b9fcff29d5c82377ed54cb5296bb32cb
- Pointer size:
- 134 Bytes
- Size of remote file:
- 108 MB
- Xet hash:
- 90a9b88d12aa7d08fc135a0128e789874a3f4b35879288dd67a037b5246982c3
·
·
Xet efficiently stores Large Files inside Git, intelligently splitting files into unique chunks and accelerating uploads and downloads. More info.