Spaces:
Running
Running
Update start.sh
Browse files
start.sh
CHANGED
@@ -3,12 +3,15 @@
|
|
3 |
|
4 |
export WORK="/home/user/app"
|
5 |
cd $WORK
|
|
|
6 |
unzip llama_cpp_avx512.zip > /dev/null 2>&1
|
7 |
wget -O model.gguf https://huggingface.co/unsloth/gpt-oss-20b-GGUF/resolve/main/gpt-oss-20b-Q4_K_M.gguf > /dev/null 2>&1
|
8 |
-
|
9 |
./llama-server -m model.gguf --port 8000 --host 0.0.0.0 --threads 2 --ctx-size 4096 --mlock --jinja \
|
10 |
--temp 0.2 \
|
11 |
--top-p 0.85 &
|
|
|
|
|
12 |
python3 app.py
|
13 |
|
14 |
#python3 -m http.server 7860 --bind 0.0.0.0
|
|
|
3 |
|
4 |
export WORK="/home/user/app"
|
5 |
cd $WORK
|
6 |
+
echo "unzip and download model"
|
7 |
unzip llama_cpp_avx512.zip > /dev/null 2>&1
|
8 |
wget -O model.gguf https://huggingface.co/unsloth/gpt-oss-20b-GGUF/resolve/main/gpt-oss-20b-Q4_K_M.gguf > /dev/null 2>&1
|
9 |
+
echo "start llama.cpp"
|
10 |
./llama-server -m model.gguf --port 8000 --host 0.0.0.0 --threads 2 --ctx-size 4096 --mlock --jinja \
|
11 |
--temp 0.2 \
|
12 |
--top-p 0.85 &
|
13 |
+
|
14 |
+
echo "start app.py"
|
15 |
python3 app.py
|
16 |
|
17 |
#python3 -m http.server 7860 --bind 0.0.0.0
|