Spaces:
Sleeping
Sleeping
| FROM debian:latest as musl-dev | |
| RUN apt-get update | |
| RUN apt-get install musl | |
| FROM nvidia/cuda:12.6.0-cudnn-runtime-ubuntu24.04 as final | |
| RUN apt-get update && apt-get install --no-install-recommends --no-install-suggests -y curl | |
| RUN apt-get install unzip | |
| RUN apt-get -y install python3 | |
| RUN apt-get -y install python3-pip | |
| RUN apt-get -y install cmake | |
| RUN apt-get -y install wget | |
| COPY / / | |
| COPY --from=musl-dev /usr/lib/x86_64-linux-musl/libc.so /lib/libc.musl-x86_64.so.1 | |
| WORKDIR /app | |
| COPY ./requirements.txt /app/requirements.txt | |
| COPY ./semsearch.py /app/semsearch.py | |
| COPY ./startup.sh /app/startup.sh | |
| COPY ./.streamlit/main.css /app/.streamlit/main.css | |
| COPY ./app.py /app/app.py | |
| RUN chmod 755 /app/startup.sh | |
| COPY ./multi-qa-MiniLM-L6-cos-v1 /app/multi-qa-MiniLM-L6-cos-v1 | |
| RUN mkdir -p /app/inputDocs | |
| COPY ./inputDocs/* /app/inputDocs | |
| RUN pip install --break-system-packages --no-cache-dir --upgrade -r /app/requirements.txt | |
| RUN pip install --break-system-packages https://files.pythonhosted.org/packages/13/87/e0cb08c2d4bd7d38ab63816b306c8b1e7cfdc0e59bd54462e8b0df069078/semantic_text_splitter-0.6.3-cp38-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl | |
| RUN pip show semantic-text-splitter | |
| RUN pip install --break-system-packages llama_cpp_python | |
| RUN FORCE_CMAKE=1 CMAKE_SYSTEM_PROCESSOR=AMD64 pip install --break-system-packages --verbose --no-cache-dir llama-cpp-python --extra-index-url https://abetlen.github.io/llama-cpp-python/whl/cpu | |
| ############################################################################## | |
| # Install Weaviate | |
| WORKDIR /app/weaviate | |
| RUN wget -qO- https://github.com/weaviate/weaviate/releases/download/v1.24.10/weaviate-v1.24.10-linux-amd64.tar.gz | tar -xzf - | |
| RUN ls -al /app/weaviate | |
| # Set environment variables for Weaviate | |
| ENV PATH="/app:/app/weaviate-v1.24.10-linux-x86_64:${PATH}" | |
| # Expose the Weaviate port | |
| EXPOSE 8080 | |
| #COPY Llama-2-7B-Chat-GGUF/llama-2-7b-chat.Q4_0.gguf /app | |
| RUN cd /app; wget -v https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGUF/resolve/main/llama-2-7b-chat.Q4_0.gguf | |
| ############################################################################## | |
| # Install text2vec-transformers | |
| WORKDIR /app/text2vec-transformers | |
| COPY --from=semitechnologies/transformers-inference:sentence-transformers-multi-qa-MiniLM-L6-cos-v1 /app /app/text2vec-transformers | |
| COPY --from=semitechnologies/transformers-inference:sentence-transformers-multi-qa-MiniLM-L6-cos-v1 /usr/local/bin /app/text2vec-transformers/bin | |
| COPY ./multi-qa-MiniLM-L6-cos-v1 /app/app/text2vec-transformers | |
| ENV PATH="/usr/bin/local:/app/text2vec-transformers:/app/text2vec-transformers/bin:${PATH}" | |
| ENV LD_LIBRARY_PATH="/usr/local/cuda/lib64:$LD_LIBRARY_PATH" | |
| #RUN ./custom_prerequisites.py | |
| ############################## | |
| RUN groupadd -g 2222 user | |
| RUN useradd -m -u 2222 -g user user | |
| RUN chmod -R 777 /app | |
| RUN chown -R user /app | |
| RUN chgrp -R user /app | |
| WORKDIR /app | |
| USER user | |
| EXPOSE 8501 | |
| CMD streamlit run /app/app.py \ | |
| --server.headless true \ | |
| --server.enableCORS false \ | |
| --server.enableXsrfProtection false \ | |
| --server.fileWatcherType none | |