Spaces:
Runtime error
Runtime error
Fetching metadata from the HF Docker repository...
Ctrl+K
- PY3 agregar tokenizers
- 6.15 kB agregar tokenizers
- 8.57 kB agregar tokenizers
- czech.pickle1.27 MB
Detected Pickle imports (9)
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "__builtin__.int",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object"
xetagregar tokenizers - danish.pickle1.26 MB
Detected Pickle imports (9)
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "__builtin__.object",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktToken",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktParameters",
- "copy_reg._reconstructor"
xetagregar tokenizers - dutch.pickle743 kB
Detected Pickle imports (9)
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "collections.defaultdict",
- "__builtin__.int",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktToken",
- "copy_reg._reconstructor"
xetagregar tokenizers - english.pickle433 kB
Detected Pickle imports (9)
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.set",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken"
xetagregar tokenizers - estonian.pickle1.6 MB
Detected Pickle imports (9)
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters"
xetagregar tokenizers - finnish.pickle1.95 MB
Detected Pickle imports (9)
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "collections.defaultdict",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "copy_reg._reconstructor",
- "__builtin__.int"
xetagregar tokenizers - french.pickle583 kB
Detected Pickle imports (9)
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "copy_reg._reconstructor",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object"
xetagregar tokenizers - german.pickle1.53 MB
Detected Pickle imports (9)
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.object",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "collections.defaultdict",
- "__builtin__.int"
xetagregar tokenizers - greek.pickle1.95 MB
Detected Pickle imports (9)
- "__builtin__.object",
- "collections.defaultdict",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "__builtin__.int"
xetagregar tokenizers - italian.pickle658 kB
Detected Pickle imports (9)
- "collections.defaultdict",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.object"
xetagregar tokenizers - malayalam.pickle221 kB
Detected Pickle imports (7)
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "nltk.tokenize.punkt.PunktToken",
- "__builtin__.int",
- "__builtin__.set",
- "collections.defaultdict"
xetagregar tokenizers - norwegian.pickle1.26 MB
Detected Pickle imports (9)
- "nltk.tokenize.punkt.PunktLanguageVars",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktToken",
- "__builtin__.object",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "copy_reg._reconstructor"
xetagregar tokenizers - polish.pickle2.04 MB
Detected Pickle imports (9)
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.object",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktToken",
- "__builtin__.set",
- "collections.defaultdict",
- "copy_reg._reconstructor"
xetagregar tokenizers - portuguese.pickle649 kB
Detected Pickle imports (9)
- "collections.defaultdict",
- "copy_reg._reconstructor",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer"
xetagregar tokenizers - russian.pickle33 kB
Detected Pickle imports (7)
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "collections.defaultdict",
- "__builtin__.long",
- "nltk.tokenize.punkt.PunktToken"
xetagregar tokenizers - slovene.pickle833 kB
Detected Pickle imports (9)
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktToken",
- "__builtin__.set",
- "collections.defaultdict",
- "copy_reg._reconstructor",
- "__builtin__.object"
xetagregar tokenizers - spanish.pickle598 kB
Detected Pickle imports (9)
- "copy_reg._reconstructor",
- "collections.defaultdict",
- "__builtin__.int",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "nltk.tokenize.punkt.PunktToken",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktLanguageVars"
xetagregar tokenizers - swedish.pickle1.03 MB
Detected Pickle imports (9)
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "copy_reg._reconstructor",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktParameters",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktToken"
xetagregar tokenizers - turkish.pickle1.23 MB
Detected Pickle imports (9)
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "copy_reg._reconstructor",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "__builtin__.object",
- "__builtin__.set",
- "nltk.tokenize.punkt.PunktLanguageVars"
xetagregar tokenizers