XanderJC commited on
Commit
710f609
·
1 Parent(s): 4fc798a
README.md CHANGED
@@ -90,7 +90,7 @@ By default, Proxy Lite will point to an endpoint set up on HuggingFace spaces.
90
  We recommend hosting your own endpoint with vLLM, you can use the following command:
91
 
92
  ```bash
93
- vllm serve --model convergence-ai/proxy-lite \
94
  --trust-remote-code \
95
  --enable-auto-tool-choice \
96
  --tool-call-parser hermes \
@@ -137,7 +137,7 @@ config = RunnerConfig.from_dict(
137
  "name": "proxy_lite",
138
  "client": {
139
  "name": "convergence",
140
- "model_id": "convergence-ai/proxy-lite",
141
  "api_base": "https://convergence-ai-demo-api.hf.space/v1",
142
  },
143
  },
@@ -197,7 +197,7 @@ from transformers import AutoProcessor
197
  from proxy_lite.tools import ReturnValueTool, BrowserTool
198
  from proxy_lite.serializer import OpenAICompatableSerializer
199
 
200
- processor = AutoProcessor.from_pretrained("convergence-ai/proxy-lite")
201
  tools = OpenAICompatableSerializer().serialize_tools([ReturnValueTool(), BrowserTool(session=None)])
202
 
203
  templated_messages = processor.apply_chat_template(
@@ -223,7 +223,7 @@ from openai import OpenAI
223
  client = OpenAI(base_url="http://convergence-ai-demo-api.hf.space/v1")
224
 
225
  response = client.chat.completions.create(
226
- model="convergence-ai/proxy-lite",
227
  messages=message_history,
228
  tools=tools,
229
  tool_choice="auto",
 
90
  We recommend hosting your own endpoint with vLLM, you can use the following command:
91
 
92
  ```bash
93
+ vllm serve --model convergence-ai/proxy-lite-3b \
94
  --trust-remote-code \
95
  --enable-auto-tool-choice \
96
  --tool-call-parser hermes \
 
137
  "name": "proxy_lite",
138
  "client": {
139
  "name": "convergence",
140
+ "model_id": "convergence-ai/proxy-lite-3b",
141
  "api_base": "https://convergence-ai-demo-api.hf.space/v1",
142
  },
143
  },
 
197
  from proxy_lite.tools import ReturnValueTool, BrowserTool
198
  from proxy_lite.serializer import OpenAICompatableSerializer
199
 
200
+ processor = AutoProcessor.from_pretrained("convergence-ai/proxy-lite-3b")
201
  tools = OpenAICompatableSerializer().serialize_tools([ReturnValueTool(), BrowserTool(session=None)])
202
 
203
  templated_messages = processor.apply_chat_template(
 
223
  client = OpenAI(base_url="http://convergence-ai-demo-api.hf.space/v1")
224
 
225
  response = client.chat.completions.create(
226
+ model="convergence-ai/proxy-lite-3b",
227
  messages=message_history,
228
  tools=tools,
229
  tool_choice="auto",
src/proxy_lite/app.py CHANGED
@@ -28,7 +28,7 @@ def get_user_config(config_expander):
28
  "name": "proxy_lite",
29
  "client": {
30
  "name": "convergence",
31
- "model_id": "convergence-ai/proxy-lite",
32
  "api_base": "https://convergence-ai-demo-api.hf.space/v1",
33
  },
34
  },
 
28
  "name": "proxy_lite",
29
  "client": {
30
  "name": "convergence",
31
+ "model_id": "convergence-ai/proxy-lite-3b",
32
  "api_base": "https://convergence-ai-demo-api.hf.space/v1",
33
  },
34
  },
src/proxy_lite/configs/default.yaml CHANGED
@@ -14,7 +14,7 @@ solver:
14
  name: proxy_lite
15
  client:
16
  name: convergence
17
- model_id: convergence-ai/proxy-lite
18
  api_base: https://convergence-ai-demo-api.hf.space/v1
19
  local_view: true
20
  task_timeout: 1800
 
14
  name: proxy_lite
15
  client:
16
  name: convergence
17
+ model_id: convergence-ai/proxy-lite-3b
18
  api_base: https://convergence-ai-demo-api.hf.space/v1
19
  local_view: true
20
  task_timeout: 1800