Tim Luka Horstmann
commited on
Commit
·
6aaa9c3
1
Parent(s):
9f40e8d
create dir
Browse files- Dockerfile +5 -1
Dockerfile
CHANGED
@@ -14,6 +14,10 @@ RUN apt-get update && \
|
|
14 |
. "$HOME/.cargo/env" && rustup default stable
|
15 |
|
16 |
WORKDIR /app
|
|
|
|
|
|
|
|
|
17 |
COPY requirements.txt .
|
18 |
RUN pip install --no-cache-dir -r requirements.txt
|
19 |
|
@@ -22,7 +26,7 @@ RUN git clone --recurse-submodules https://github.com/abetlen/llama-cpp-python.g
|
|
22 |
&& FORCE_CMAKE=1 pip install --no-cache-dir . \
|
23 |
&& pip install --no-cache-dir "llama-cpp-python[server]" huggingface_hub \
|
24 |
&& rm -rf /tmp/llama-cpp-python
|
25 |
-
|
26 |
|
27 |
# Copy the LLM server code
|
28 |
COPY llm_server.py /app/llm_server.py
|
|
|
14 |
. "$HOME/.cargo/env" && rustup default stable
|
15 |
|
16 |
WORKDIR /app
|
17 |
+
|
18 |
+
RUN mkdir -p /app/cache /app/logs /app/workspace /app/pretrained_models/llm \
|
19 |
+
&& chmod -R 777 /app/cache /app/logs /app/workspace /app/pretrained_models/llm
|
20 |
+
|
21 |
COPY requirements.txt .
|
22 |
RUN pip install --no-cache-dir -r requirements.txt
|
23 |
|
|
|
26 |
&& FORCE_CMAKE=1 pip install --no-cache-dir . \
|
27 |
&& pip install --no-cache-dir "llama-cpp-python[server]" huggingface_hub \
|
28 |
&& rm -rf /tmp/llama-cpp-python
|
29 |
+
|
30 |
|
31 |
# Copy the LLM server code
|
32 |
COPY llm_server.py /app/llm_server.py
|