Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -102,12 +102,12 @@ def tool_selector(state: AgentState) -> AgentState:
|
|
102 |
)
|
103 |
)
|
104 |
raw = LLM(state.messages + [prompt]).content.strip()
|
105 |
-
print(f"Tool selector response: {raw}")
|
106 |
state.add(AIMessage(content=raw))
|
107 |
parsed = safe_json(raw)
|
108 |
# parsed = json.loads(raw)
|
109 |
# print("parsed : ", parsed)
|
110 |
-
print(f"Parsed: {parsed}, type: {type(parsed)}")
|
111 |
if not parsed or "action" not in parsed:
|
112 |
state.next_action = "final"
|
113 |
return state
|
@@ -163,12 +163,12 @@ def final_node(state: AgentState) -> AgentState:
|
|
163 |
"{\"final_answer\":\"β¦\"} (no markdown, no commentary)."
|
164 |
)
|
165 |
raw = LLM(state.messages + [wrap]).content.strip()
|
166 |
-
print("raw : ", raw)
|
167 |
state.add(AIMessage(content=raw))
|
168 |
parsed = safe_json(raw)
|
169 |
-
print("parsed : ", parsed, "type : ", type(parsed))
|
170 |
state.final_answer = parsed.get("final_answer") if parsed else "Unable to parse final answer."
|
171 |
-
print("state.final_answer : ", state.final_answer)
|
172 |
return state
|
173 |
|
174 |
# βββββββββββββββββββββββββββ Graph wiring βββββββββββββββββββββββββββββββ
|
@@ -225,9 +225,9 @@ def answer(question: str, task_id: Optional[str] = None) -> str:
|
|
225 |
state = AgentState(user_question=question, task_id=task_id)
|
226 |
state.add(SystemMessage(content="You are a helpful assistant."))
|
227 |
state.add(HumanMessage(content=question))
|
228 |
-
compiled_graph.invoke(state)
|
229 |
-
print(f"Final answer: {
|
230 |
-
return
|
231 |
|
232 |
|
233 |
|
|
|
102 |
)
|
103 |
)
|
104 |
raw = LLM(state.messages + [prompt]).content.strip()
|
105 |
+
# print(f"Tool selector response: {raw}")
|
106 |
state.add(AIMessage(content=raw))
|
107 |
parsed = safe_json(raw)
|
108 |
# parsed = json.loads(raw)
|
109 |
# print("parsed : ", parsed)
|
110 |
+
# print(f"Parsed: {parsed}, type: {type(parsed)}")
|
111 |
if not parsed or "action" not in parsed:
|
112 |
state.next_action = "final"
|
113 |
return state
|
|
|
163 |
"{\"final_answer\":\"β¦\"} (no markdown, no commentary)."
|
164 |
)
|
165 |
raw = LLM(state.messages + [wrap]).content.strip()
|
166 |
+
# print("raw : ", raw)
|
167 |
state.add(AIMessage(content=raw))
|
168 |
parsed = safe_json(raw)
|
169 |
+
# print("parsed : ", parsed, "type : ", type(parsed))
|
170 |
state.final_answer = parsed.get("final_answer") if parsed else "Unable to parse final answer."
|
171 |
+
# print("state.final_answer : ", state.final_answer)
|
172 |
return state
|
173 |
|
174 |
# βββββββββββββββββββββββββββ Graph wiring βββββββββββββββββββββββββββββββ
|
|
|
225 |
state = AgentState(user_question=question, task_id=task_id)
|
226 |
state.add(SystemMessage(content="You are a helpful assistant."))
|
227 |
state.add(HumanMessage(content=question))
|
228 |
+
ans = compiled_graph.invoke(state)
|
229 |
+
print(f"Final answer: {ans.final_answer}")
|
230 |
+
return ans.final_answer or "No answer."
|
231 |
|
232 |
|
233 |
|