Spaces:
Sleeping
Sleeping
Fetching metadata from the HF Docker repository...
download
history
blame
contribute
delete
52.3 MB
Detected Pickle imports (300)
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "__builtin__.int",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "__builtin__.int",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "__builtin__.int",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "__builtin__.int",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "__builtin__.int",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "__builtin__.int",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "__builtin__.int",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "__builtin__.int",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "__builtin__.int",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "__builtin__.int",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "__builtin__.int",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "__builtin__.int",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "builtins.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "builtins.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "builtins.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "builtins.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "builtins.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "builtins.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "builtins.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "builtins.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "builtins.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "builtins.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "__builtin__.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "builtins.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "builtins.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "builtins.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktToken",
- "builtins.int",
- "nltk.tokenize.punkt.PunktParameters",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "builtins.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "builtins.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "builtins.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "builtins.set",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "builtins.int",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "__builtin__.long",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "__builtin__.int",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "__builtin__.int",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "__builtin__.int",
- "__builtin__.object",
- "nltk.tokenize.punkt.PunktSentenceTokenizer",
- "collections.defaultdict",
- "nltk.tokenize.punkt.PunktLanguageVars",
- "nltk.tokenize.punkt.PunktToken",
- "copy_reg._reconstructor",
- "nltk.tokenize.punkt.PunktParameters",
- "__builtin__.set",
- "__builtin__.int"
- SHA256:
- e7ca5b931a531c962d2539b042daf7d37badd4ce59523dfa063083f61a1dae72
- Pointer size:
- 133 Bytes
- Size of remote file:
- 52.3 MB
- Xet backed hash:
- ac83bb4197d85e8d51048d13fe248e4ad7a6cddcea023b707967b55a16149fdc
·
·
Xet efficiently stores Large Files inside Git, intelligently splitting files into unique chunks and accelerating uploads and downloads. More info.