File size: 607 Bytes
26df90d
7e52d33
 
 
 
 
 
 
 
 
 
fc96eeb
a82c04a
fc96eeb
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
FROM ghcr.io/ggml-org/llama.cpp:full

# Install wget
RUN apt update && apt install wget -y

# Download the GGUF model file
RUN wget "https://huggingface.co/ggml-org/SmolVLM-256M-Instruct-GGUF/resolve/main/SmolVLM-256M-Instruct-Q8_0.gguf" -O /smoll.gguf

# Download the mmproj file
RUN wget "https://huggingface.co/ggml-org/SmolVLM-256M-Instruct-GGUF/resolve/main/mmproj-SmolVLM-256M-Instruct-Q8_0.gguf" -O /mmproj.gguf

#EXPOSE 7860

# Default command to start the model server
CMD [ "--server", "-m", "/smoll.gguf", "--mmproj", "/mmproj.gguf", "--port", "7860", "--host", "0.0.0.0", "-n", "512", "-t", "2"]