Spaces:
Running
on
L4
Running
on
L4
feat: use ten turn detector
Browse files- entrypoint.sh +9 -7
entrypoint.sh
CHANGED
@@ -4,12 +4,14 @@ mkdir -p data
|
|
4 |
|
5 |
#HF_MODEL_PATH=Qwen/Qwen2.5-1.5B-Instruct
|
6 |
#HF_MODEL_PATH=deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
|
7 |
-
HF_MODEL_PATH=Qwen/Qwen2.5-VL-3B-Instruct
|
8 |
-
|
|
|
9 |
LOCAL_MODEL_PATH=./data/${HF_MODEL_NAME}
|
10 |
|
11 |
# TODO: use your own key and put into secret
|
12 |
-
|
|
|
13 |
|
14 |
# download model
|
15 |
HF_HUB_ENABLE_HF_TRANSFER=1 huggingface-cli download ${HF_MODEL_PATH} --local-dir ${LOCAL_MODEL_PATH}
|
@@ -21,8 +23,8 @@ vllm serve ${LOCAL_MODEL_PATH} --served-model-name ${HF_MODEL_NAME} --api-key ${
|
|
21 |
cd /app/playground && npm run dev &
|
22 |
|
23 |
# start backend
|
24 |
-
export OPENAI_API_BASE=http://127.0.0.1:8000/v1
|
25 |
-
export OPENAI_PROXY_URL=
|
26 |
-
export OPENAI_MODEL=${HF_MODEL_NAME}
|
27 |
-
export OPENAI_API_KEY=${VLLM_SERVE_API_KEY}
|
28 |
cd /app && task run
|
|
|
4 |
|
5 |
#HF_MODEL_PATH=Qwen/Qwen2.5-1.5B-Instruct
|
6 |
#HF_MODEL_PATH=deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
|
7 |
+
#HF_MODEL_PATH=Qwen/Qwen2.5-VL-3B-Instruct
|
8 |
+
HF_MODEL_PATH=TEN-framework/TEN_Turn_Detector
|
9 |
+
#HF_MODEL_NAME=$(basename ${HF_MODEL_PATH})
|
10 |
LOCAL_MODEL_PATH=./data/${HF_MODEL_NAME}
|
11 |
|
12 |
# TODO: use your own key and put into secret
|
13 |
+
HF_MODEL_NAME=ten-turn-detector
|
14 |
+
VLLM_SERVE_API_KEY=ten-turn-detector
|
15 |
|
16 |
# download model
|
17 |
HF_HUB_ENABLE_HF_TRANSFER=1 huggingface-cli download ${HF_MODEL_PATH} --local-dir ${LOCAL_MODEL_PATH}
|
|
|
23 |
cd /app/playground && npm run dev &
|
24 |
|
25 |
# start backend
|
26 |
+
#export OPENAI_API_BASE=http://127.0.0.1:8000/v1
|
27 |
+
#export OPENAI_PROXY_URL=
|
28 |
+
#export OPENAI_MODEL=${HF_MODEL_NAME}
|
29 |
+
#export OPENAI_API_KEY=${VLLM_SERVE_API_KEY}
|
30 |
cd /app && task run
|