| FROM python:3.11.6-bullseye | |
| RUN git clone --recurse-submodules https://github.com/trzy/llava-cpp-server | |
| WORKDIR llava-cpp-server | |
| WORKDIR llama.cpp | |
| RUN git pull origin master | |
| WORKDIR .. | |
| RUN wget https://huggingface.co/mys/ggml_llava-v1.5-7b/resolve/main/mmproj-model-f16.gguf | |
| RUN wget https://huggingface.co/mys/ggml_llava-v1.5-7b/resolve/main/ggml-model-f16.gguf | |
| RUN make | |
| CMD ["bin/llava-server", "-m", "ggml-model-f16.gguf", "--mmproj", "mmproj-model-f16.gguf", "--port", "7860", "--host", "0.0.0.0", "--threads", "4", "--log-http"] |