Ctrl+K
This model has 1 file scanned as unsafe.
- 1.52 kB initial commit
- 4.65 kB Update README.md
- fusion_model.pt506 MB
Detected Pickle imports (3)
- "torch.FloatStorage",
- "collections.OrderedDict",
- "torch._utils._rebuild_tensor_v2"
xetUpload fusion_model.pt - nltk_data.zip240 MB
Detected Pickle imports (329)
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "copy_reg._reconstructor",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "copy_reg._reconstructor",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "copy_reg._reconstructor",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "copy_reg._reconstructor",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.long",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "copy_reg._reconstructor",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "copy_reg._reconstructor",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "copy_reg._reconstructor",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "copy_reg._reconstructor",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "copy_reg._reconstructor",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "copy_reg._reconstructor",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "copy_reg._reconstructor",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "copy_reg._reconstructor",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "copy_reg._reconstructor",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "copy_reg._reconstructor",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "copy_reg._reconstructor",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "copy_reg._reconstructor",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "copy_reg._reconstructor",
- "nltk.classify.maxent.MaxentClassifier",
- "nltk.chunk.named_entity.NEChunkParser",
- "numpy.ndarray",
- "numpy.dtype",
- "nltk.classify.maxent.BinaryMaxentFeatureEncoding",
- "numpy.core.multiarray._reconstruct",
- "nltk.chunk.named_entity.NEChunkParserTagger",
- "nltk.classify.maxent.MaxentClassifier",
- "nltk.chunk.named_entity.NEChunkParser",
- "numpy.ndarray",
- "numpy.dtype",
- "nltk.classify.maxent.BinaryMaxentFeatureEncoding",
- "numpy.core.multiarray._reconstruct",
- "nltk.chunk.named_entity.NEChunkParserTagger",
- "__builtin__.set",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "builtins.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.classify.maxent.MaxentClassifier",
- "nltk.chunk.named_entity.NEChunkParser",
- "numpy.ndarray",
- "numpy.dtype",
- "nltk.classify.maxent.BinaryMaxentFeatureEncoding",
- "numpy.core.multiarray._reconstruct",
- "nltk.chunk.named_entity.NEChunkParserTagger",
- "nltk.classify.maxent.MaxentClassifier",
- "nltk.chunk.named_entity.NEChunkParser",
- "numpy.ndarray",
- "numpy.dtype",
- "nltk.classify.maxent.BinaryMaxentFeatureEncoding",
- "numpy.core.multiarray._reconstruct",
- "nltk.chunk.named_entity.NEChunkParserTagger"
xetUpload nltk_data.zip