runtime error

Exit code: 1. Reason: k they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. modeling_phi.py: 0%| | 0.00/33.4k [00:00<?, ?B/s] modeling_phi.py: 100%|██████████| 33.4k/33.4k [00:00<00:00, 7.57MB/s] A new version of the following files was downloaded from https://huggingface.co/TheBloke/phi-2-GPTQ: - modeling_phi.py . Make sure to double-check they do not contain any added malicious code. To avoid downloading new versions of the code file, you can pin a revision. /usr/local/lib/python3.10/site-packages/transformers/utils/generic.py:309: UserWarning: torch.utils._pytree._register_pytree_node is deprecated. Please use torch.utils._pytree.register_pytree_node instead. _torch_pytree._register_pytree_node( /usr/local/lib/python3.10/site-packages/transformers/utils/generic.py:309: UserWarning: torch.utils._pytree._register_pytree_node is deprecated. Please use torch.utils._pytree.register_pytree_node instead. _torch_pytree._register_pytree_node( Traceback (most recent call last): File "/home/user/app/app.py", line 7, in <module> model = AutoModelForCausalLM.from_pretrained( File "/usr/local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 561, in from_pretrained return model_class.from_pretrained( File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 3013, in from_pretrained gptq_supports_cpu = version.parse(importlib.metadata.version("auto-gptq")) > version.parse("0.4.2") File "/usr/local/lib/python3.10/importlib/metadata/__init__.py", line 996, in version return distribution(distribution_name).version File "/usr/local/lib/python3.10/importlib/metadata/__init__.py", line 969, in distribution return Distribution.from_name(distribution_name) File "/usr/local/lib/python3.10/importlib/metadata/__init__.py", line 548, in from_name raise PackageNotFoundError(name) importlib.metadata.PackageNotFoundError: No package metadata was found for auto-gptq

Container logs:

Fetching error logs...