Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -41,23 +41,23 @@ final_answer = FinalAnswerTool()
|
|
41 |
# If the agent does not answer, the model is overloaded, please use another model or the following Hugging Face Endpoint that also contains qwen2.5 coder:
|
42 |
# model_id='https://pflgm2locj2t89co.us-east-1.aws.endpoints.huggingface.cloud'
|
43 |
|
44 |
-
|
45 |
-
|
46 |
-
|
47 |
-
#
|
48 |
-
|
49 |
# model_id='mistralai/Mistral-7B-Instruct-v0.2',
|
50 |
-
|
51 |
-
# )
|
52 |
-
|
53 |
-
from smolagents import LiteLLMModel
|
54 |
-
model = LiteLLMModel(
|
55 |
-
model_id="ollama_chat/qwen2:7b",
|
56 |
-
# api_base="http://127.0.0.1:11434",
|
57 |
-
api_base="http://192.168.1.106:11434",
|
58 |
-
num_ctx=8192
|
59 |
)
|
60 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
61 |
|
62 |
# Import tool from Hub
|
63 |
image_generation_tool = load_tool("agents-course/text-to-image", trust_remote_code=True)
|
|
|
41 |
# If the agent does not answer, the model is overloaded, please use another model or the following Hugging Face Endpoint that also contains qwen2.5 coder:
|
42 |
# model_id='https://pflgm2locj2t89co.us-east-1.aws.endpoints.huggingface.cloud'
|
43 |
|
44 |
+
model = HfApiModel(
|
45 |
+
max_tokens=2096,
|
46 |
+
temperature=0.5,
|
47 |
+
# model_id='Qwen/Qwen2.5-Coder-32B-Instruct',# it is possible that this model may be overloaded
|
48 |
+
model_id='https://pflgm2locj2t89co.us-east-1.aws.endpoints.huggingface.cloud',
|
49 |
# model_id='mistralai/Mistral-7B-Instruct-v0.2',
|
50 |
+
custom_role_conversions=None,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
51 |
)
|
52 |
|
53 |
+
# from smolagents import LiteLLMModel
|
54 |
+
# model = LiteLLMModel(
|
55 |
+
# model_id="ollama_chat/qwen2:7b",
|
56 |
+
# # api_base="http://127.0.0.1:11434",
|
57 |
+
# api_base="http://192.168.1.106:11434",
|
58 |
+
# num_ctx=8192
|
59 |
+
# )
|
60 |
+
|
61 |
|
62 |
# Import tool from Hub
|
63 |
image_generation_tool = load_tool("agents-course/text-to-image", trust_remote_code=True)
|