runtime error
Exit code: 1. Reason: File "/usr/local/lib/python3.10/asyncio/base_events.py", line 649, in run_until_complete return future.result() File "/usr/local/lib/python3.10/site-packages/uvicorn/server.py", line 71, in serve await self._serve(sockets) File "/usr/local/lib/python3.10/site-packages/uvicorn/server.py", line 78, in _serve config.load() File "/usr/local/lib/python3.10/site-packages/uvicorn/config.py", line 436, in load self.loaded_app = import_from_string(self.app) File "/usr/local/lib/python3.10/site-packages/uvicorn/importer.py", line 19, in import_from_string module = importlib.import_module(module_str) File "/usr/local/lib/python3.10/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) File "<frozen importlib._bootstrap>", line 1050, in _gcd_import File "<frozen importlib._bootstrap>", line 1027, in _find_and_load File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 688, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 883, in exec_module File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed File "/app/app.py", line 10, in <module> tokenizer = AutoTokenizer.from_pretrained(model_id) File "/usr/local/lib/python3.10/site-packages/transformers/models/auto/tokenization_auto.py", line 899, in from_pretrained return tokenizer_class_fast.from_pretrained(pretrained_model_name_or_path, *inputs, **kwargs) File "/usr/local/lib/python3.10/site-packages/transformers/tokenization_utils_base.py", line 2094, in from_pretrained raise EnvironmentError( OSError: Can't load tokenizer for 'TheBloke/Mistral-7B-Instruct-v0.2-GGUF'. If you were trying to load it from 'https://huggingface.co/models', make sure you don't have a local directory with the same name. Otherwise, make sure 'TheBloke/Mistral-7B-Instruct-v0.2-GGUF' is the correct path to a directory containing all relevant files for a LlamaTokenizerFast tokenizer.
Container logs:
Fetching error logs...