Spaces:
Running
Running
Update Dockerfile
Browse files- Dockerfile +2 -1
Dockerfile
CHANGED
@@ -63,7 +63,8 @@ RUN pip3 install https://files.pythonhosted.org/packages/13/87/e0cb08c2d4bd7d38
|
|
63 |
#RUN CMAKE_ARGS="-DLLAMA_CUBLAS=on -DCUDA_PATH=/usr/local/cuda-12.2 -DCUDAToolkit_ROOT=/usr/local/cuda-12.2 -DCUDAToolkit_INCLUDE_DIR=/usr/local/cuda-12/include -DCUDAToolkit_LIBRARY_DIR=/usr/local/cuda-12.2/lib64" FORCE_CMAKE=1 pip install llama-cpp-python - no-cache-dir
|
64 |
#RUN CMAKE_ARGS="GGML_CUDA=on" FORCE_CMAKE=1 pip install --break-system-packages llama-cpp-python --no-cache-dir
|
65 |
RUN CMAKE_ARGS="GGML_CUDA=on" FORCE_CMAKE=1 pip install llama-cpp-python --no-cache-dir
|
66 |
-
RUN pip3 install --break-system-packages cffi
|
|
|
67 |
|
68 |
# Install text2vec-transformers
|
69 |
WORKDIR /app/text2vec-transformers
|
|
|
63 |
#RUN CMAKE_ARGS="-DLLAMA_CUBLAS=on -DCUDA_PATH=/usr/local/cuda-12.2 -DCUDAToolkit_ROOT=/usr/local/cuda-12.2 -DCUDAToolkit_INCLUDE_DIR=/usr/local/cuda-12/include -DCUDAToolkit_LIBRARY_DIR=/usr/local/cuda-12.2/lib64" FORCE_CMAKE=1 pip install llama-cpp-python - no-cache-dir
|
64 |
#RUN CMAKE_ARGS="GGML_CUDA=on" FORCE_CMAKE=1 pip install --break-system-packages llama-cpp-python --no-cache-dir
|
65 |
RUN CMAKE_ARGS="GGML_CUDA=on" FORCE_CMAKE=1 pip install llama-cpp-python --no-cache-dir
|
66 |
+
#RUN pip3 install --break-system-packages cffi
|
67 |
+
RUN pip3 install cffi
|
68 |
|
69 |
# Install text2vec-transformers
|
70 |
WORKDIR /app/text2vec-transformers
|