| FROM python:3.12 | |
| RUN useradd -m -u 1000 app | |
| WORKDIR /home/app | |
| RUN apt update | |
| RUN apt install -y wget make cmake clang git g++ | |
| RUN wget https://huggingface.co/mradermacher/bellman-mistral-7b-instruct-v0.3-GGUF/blob/main/bellman-mistral-7b-instruct-v0.3.Q5_K_M.gguf?download=true -O model.gguf | |
| RUN git clone https://github.com/ggerganov/llama.cpp | |
| RUN mv llama.cpp llama_temp | |
| RUN mv llama_temp/* . | |
| RUN cmake . | |
| RUN apt install socat -y | |
| EXPOSE 7860 | |
| CMD ["sh", "-c", "./llama-server -m /home/app/model.gguf -c 8192 --host 0.0.0.0 --port 7860"] |