broadfield-dev commited on
Commit
129400d
·
verified ·
1 Parent(s): 3b223e8

Create app.py

Browse files
Files changed (1) hide show
  1. app.py +255 -0
app.py ADDED
@@ -0,0 +1,255 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os
2
+ import json
3
+ import logging
4
+ import tempfile
5
+ from dotenv import load_dotenv
6
+ import gradio as gr
7
+
8
+ load_dotenv()
9
+
10
+ MEMORY_STORAGE_TYPE = "HF_DATASET"
11
+ HF_DATASET_MEMORY_REPO = "broadfield-dev/ai-brain"
12
+ HF_DATASET_RULES_REPO = "broadfield-dev/ai-rules"
13
+
14
+ os.environ['STORAGE_BACKEND'] = MEMORY_STORAGE_TYPE
15
+ if MEMORY_STORAGE_TYPE == "HF_DATASET":
16
+ os.environ['HF_MEMORY_DATASET_REPO'] = HF_DATASET_MEMORY_REPO
17
+ os.environ['HF_RULES_DATASET_REPO'] = HF_DATASET_RULES_REPO
18
+
19
+ from model_logic import get_available_providers, get_model_display_names_for_provider, get_default_model_display_name_for_provider
20
+ from memory_logic import (
21
+ initialize_memory_system, add_memory_entry, get_all_memories_cached, clear_all_memory_data_backend,
22
+ add_rule_entry, remove_rule_entry, get_all_rules_cached, clear_all_rules_data_backend,
23
+ save_faiss_indices_to_disk, STORAGE_BACKEND as MEMORY_STORAGE_BACKEND, SQLITE_DB_PATH as MEMORY_SQLITE_PATH,
24
+ HF_MEMORY_DATASET_REPO as MEMORY_HF_MEM_REPO, HF_RULES_DATASET_REPO as MEMORY_HF_RULES_REPO
25
+ )
26
+ from tools.orchestrator import orchestrate_and_respond
27
+ from learning import perform_post_interaction_learning
28
+ from utils import load_rules_from_file, load_memories_from_file
29
+ from prompts import DEFAULT_SYSTEM_PROMPT
30
+
31
+ logging.basicConfig(level=logging.INFO, format='%(asctime)s - %(name)s - %(levelname)s - %(threadName)s - %(message)s')
32
+ logger = logging.getLogger(__name__)
33
+ for lib_name in ["urllib3", "requests", "huggingface_hub", "PIL.PngImagePlugin", "matplotlib", "gradio_client.client", "multipart.multipart", "httpx", "sentence_transformers", "faiss", "datasets"]:
34
+ if logging.getLogger(lib_name): logging.getLogger(lib_name).setLevel(logging.WARNING)
35
+
36
+ MAX_HISTORY_TURNS = int(os.getenv("MAX_HISTORY_TURNS", 7))
37
+ LOAD_RULES_FILE = os.getenv("LOAD_RULES_FILE")
38
+ LOAD_MEMORIES_FILE = os.getenv("LOAD_MEMORIES_FILE")
39
+ current_chat_session_history = []
40
+
41
+ def handle_gradio_chat_submit(user_msg_txt: str, gr_hist_list: list, sel_prov_name: str, sel_model_disp_name: str, ui_api_key: str|None, cust_sys_prompt: str):
42
+ global current_chat_session_history
43
+ cleared_input, updated_gr_hist, status_txt = "", list(gr_hist_list), "Initializing..."
44
+ updated_rules_text = ui_refresh_rules_display_fn()
45
+ updated_mems_json = ui_refresh_memories_display_fn()
46
+ def_detect_out_md = gr.Markdown(visible=False)
47
+ def_fmt_out_txt = gr.Textbox(value="*Waiting...*", interactive=True, show_copy_button=True)
48
+ def_dl_btn = gr.DownloadButton(interactive=False, value=None, visible=False)
49
+
50
+ if not user_msg_txt.strip():
51
+ status_txt = "Error: Empty message."
52
+ updated_gr_hist.append((user_msg_txt or "(Empty)", status_txt))
53
+ yield (cleared_input, updated_gr_hist, status_txt, def_detect_out_md, def_fmt_out_txt, def_dl_btn, updated_rules_text, updated_mems_json)
54
+ return
55
+
56
+ updated_gr_hist.append((user_msg_txt, "<i>Thinking...</i>"))
57
+ yield (cleared_input, updated_gr_hist, status_txt, def_detect_out_md, def_fmt_out_txt, def_dl_btn, updated_rules_text, updated_mems_json)
58
+
59
+ internal_hist = list(current_chat_session_history); internal_hist.append({"role": "user", "content": user_msg_txt})
60
+ hist_len_check = MAX_HISTORY_TURNS * 2
61
+ if len(internal_hist) > hist_len_check:
62
+ current_chat_session_history = internal_hist[-(MAX_HISTORY_TURNS * 2):]
63
+ internal_hist = list(current_chat_session_history)
64
+
65
+ final_bot_resp_acc, insights_used_parsed = "", []
66
+ temp_dl_file_path = None
67
+
68
+ try:
69
+ processor_gen = orchestrate_and_respond(user_input=user_msg_txt, provider_name=sel_prov_name, model_display_name=sel_model_disp_name, chat_history_for_prompt=internal_hist, custom_system_prompt=cust_sys_prompt.strip() or None, ui_api_key_override=ui_api_key.strip() if ui_api_key else None)
70
+ curr_bot_disp_msg = ""
71
+ for upd_type, upd_data in processor_gen:
72
+ if upd_type == "status":
73
+ status_txt = upd_data
74
+ if updated_gr_hist and updated_gr_hist[-1][0] == user_msg_txt:
75
+ updated_gr_hist[-1] = (user_msg_txt, f"{curr_bot_disp_msg} <i>{status_txt}</i>" if curr_bot_disp_msg else f"<i>{status_txt}</i>")
76
+ elif upd_type == "response_chunk":
77
+ curr_bot_disp_msg += upd_data
78
+ if updated_gr_hist and updated_gr_hist[-1][0] == user_msg_txt:
79
+ updated_gr_hist[-1] = (user_msg_txt, curr_bot_disp_msg)
80
+ elif upd_type == "final_response_and_insights":
81
+ final_bot_resp_acc, insights_used_parsed = upd_data["response"], upd_data["insights_used"]
82
+ status_txt = "Response generated. Processing learning..."
83
+ if not curr_bot_disp_msg and final_bot_resp_acc : curr_bot_disp_msg = final_bot_resp_acc
84
+ if updated_gr_hist and updated_gr_hist[-1][0] == user_msg_txt: updated_gr_hist[-1] = (user_msg_txt, curr_bot_disp_msg or "(No text)")
85
+ def_fmt_out_txt = gr.Textbox(value=curr_bot_disp_msg, interactive=True, show_copy_button=True)
86
+ if curr_bot_disp_msg and not curr_bot_disp_msg.startswith("Error:"):
87
+ with tempfile.NamedTemporaryFile(mode="w", delete=False, suffix=".md", encoding='utf-8') as tmpfile:
88
+ tmpfile.write(curr_bot_disp_msg)
89
+ temp_dl_file_path = tmpfile.name
90
+ def_dl_btn = gr.DownloadButton(value=temp_dl_file_path, visible=True, interactive=True)
91
+ insights_md_content = "### Insights Considered (Pre-Response):\n" + ("\n".join([f"- **[{i.get('type','N/A')}|{i.get('score','N/A')}]** {i.get('text','N/A')[:100]}..." for i in insights_used_parsed[:3]]) if insights_used_parsed else "*None specific.*")
92
+ def_detect_out_md = gr.Markdown(value=insights_md_content, visible=bool(insights_used_parsed))
93
+
94
+ yield (cleared_input, updated_gr_hist, status_txt, def_detect_out_md, def_fmt_out_txt, def_dl_btn, updated_rules_text, updated_mems_json)
95
+ if upd_type == "final_response_and_insights": break
96
+ except Exception as e:
97
+ logger.error(f"Chat handler error: {e}", exc_info=True); status_txt = f"Error: {str(e)[:100]}"
98
+ error_message_for_chat = f"Sorry, an error occurred: {str(e)[:100]}"
99
+ if updated_gr_hist and updated_gr_hist[-1][0] == user_msg_txt: updated_gr_hist[-1] = (user_msg_txt, error_message_for_chat)
100
+ else: updated_gr_hist.append((user_msg_txt, error_message_for_chat))
101
+ yield (cleared_input, updated_gr_hist, status_txt, gr.Markdown(value="*Error processing request.*", visible=True), gr.Textbox(value=error_message_for_chat, interactive=True), def_dl_btn, ui_refresh_rules_display_fn(), ui_refresh_memories_display_fn())
102
+ if temp_dl_file_path and os.path.exists(temp_dl_file_path): os.unlink(temp_dl_file_path)
103
+ return
104
+
105
+ if final_bot_resp_acc and not final_bot_resp_acc.startswith("Error:"):
106
+ current_chat_session_history.extend([{"role": "user", "content": user_msg_txt}, {"role": "assistant", "content": final_bot_resp_acc}])
107
+ if len(current_chat_session_history) > MAX_HISTORY_TURNS * 2: current_chat_session_history = current_chat_session_history[-(MAX_HISTORY_TURNS * 2):]
108
+ status_txt = "<i>[Performing post-interaction learning...]</i>"
109
+ yield (cleared_input, updated_gr_hist, status_txt, def_detect_out_md, def_fmt_out_txt, def_dl_btn, ui_refresh_rules_display_fn(), ui_refresh_memories_display_fn())
110
+ try:
111
+ perform_post_interaction_learning(user_input=user_msg_txt, bot_response=final_bot_resp_acc, provider=sel_prov_name, model_disp_name=sel_model_disp_name, insights_reflected=insights_used_parsed, api_key_override=ui_api_key.strip() if ui_api_key else None)
112
+ status_txt = "Response & Learning Complete."
113
+ except Exception as e_learn:
114
+ logger.error(f"Error during post-interaction learning: {e_learn}", exc_info=True)
115
+ status_txt = "Response complete. Error during learning."
116
+ else: status_txt = final_bot_resp_acc or "Processing finished; no valid response."
117
+
118
+ updated_rules_text = ui_refresh_rules_display_fn()
119
+ updated_mems_json = ui_refresh_memories_display_fn()
120
+ yield (cleared_input, updated_gr_hist, status_txt, def_detect_out_md, def_fmt_out_txt, def_dl_btn, updated_rules_text, updated_mems_json)
121
+
122
+ if temp_dl_file_path and os.path.exists(temp_dl_file_path): os.unlink(temp_dl_file_path)
123
+
124
+ def ui_refresh_rules_display_fn(): return "\n\n---\n\n".join(get_all_rules_cached()) or "No rules found."
125
+ def ui_download_rules_action_fn():
126
+ rules_content = "\n\n---\n\n".join(get_all_rules_cached())
127
+ if not rules_content.strip():
128
+ gr.Warning("No rules to download.")
129
+ return gr.DownloadButton(value=None, interactive=False, label="No Rules")
130
+ with tempfile.NamedTemporaryFile(mode="w", delete=False, suffix=".txt", encoding='utf-8') as tmpfile:
131
+ tmpfile.write(rules_content)
132
+ return tmpfile.name
133
+ def ui_upload_rules_action_fn(uploaded_file_obj, progress=gr.Progress()):
134
+ if not uploaded_file_obj: return "No file provided."
135
+ added, skipped, errors = load_rules_from_file(uploaded_file_obj.name, progress_callback=lambda p, d: progress(p, desc=d))
136
+ return f"Rules Upload: Added: {added}, Skipped (duplicates): {skipped}, Errors: {errors}."
137
+
138
+ def ui_refresh_memories_display_fn(): return get_all_memories_cached() or []
139
+ def ui_download_memories_action_fn():
140
+ memories = get_all_memories_cached()
141
+ if not memories:
142
+ gr.Warning("No memories to download.")
143
+ return gr.DownloadButton(value=None, interactive=False, label="No Memories")
144
+ jsonl_content = "\n".join([json.dumps(mem) for mem in memories])
145
+ with tempfile.NamedTemporaryFile(mode="w", delete=False, suffix=".jsonl", encoding='utf-g') as tmpfile:
146
+ tmpfile.write(jsonl_content)
147
+ return tmpfile.name
148
+ def ui_upload_memories_action_fn(uploaded_file_obj, progress=gr.Progress()):
149
+ if not uploaded_file_obj: return "No file provided."
150
+ added, format_err, save_err = load_memories_from_file(uploaded_file_obj.name, progress_callback=lambda p, d: progress(p, desc=d))
151
+ return f"Memories Upload: Added: {added}, Format Errors: {format_err}, Save Errors: {save_err}."
152
+
153
+ def save_edited_rules_action_fn(edited_rules_text: str, progress=gr.Progress()):
154
+ if not edited_rules_text.strip(): return "No rules text to save."
155
+ potential_rules = edited_rules_text.split("\n\n---\n\n")
156
+ if len(potential_rules) == 1 and "\n" in edited_rules_text:
157
+ potential_rules = [r.strip() for r in edited_rules_text.splitlines() if r.strip()]
158
+ unique_rules = sorted(list(set(filter(None, [r.strip() for r in potential_rules]))))
159
+ if not unique_rules: return "No unique, non-empty rules found."
160
+ added, skipped, errors, total = 0, 0, 0, len(unique_rules)
161
+ progress(0, desc=f"Saving {total} unique rules...")
162
+ for idx, rule_text in enumerate(unique_rules):
163
+ success, status_msg = add_rule_entry(rule_text)
164
+ if success: added += 1
165
+ elif status_msg == "duplicate": skipped += 1
166
+ else: errors += 1
167
+ progress((idx + 1) / total, desc=f"Processed {idx+1}/{total} rules...")
168
+ return f"Editor Save: Added: {added}, Skipped (duplicates): {skipped}, Errors: {errors} from {total} unique rules."
169
+
170
+ def app_load_fn():
171
+ logger.info("App loading. Initializing systems...")
172
+ initialize_memory_system()
173
+ rules_added, rules_skipped, rules_errors = load_rules_from_file(LOAD_RULES_FILE)
174
+ mems_added, mems_format_errors, mems_save_errors = load_memories_from_file(LOAD_MEMORIES_FILE)
175
+ status = f"Ready. Rules loaded: {rules_added}. Memories loaded: {mems_added}."
176
+ return (status, ui_refresh_rules_display_fn(), ui_refresh_memories_display_fn(), gr.Markdown(visible=False), gr.Textbox(value="*Waiting...*", interactive=True), gr.DownloadButton(interactive=False, visible=False))
177
+
178
+ with gr.Blocks(theme=gr.themes.Soft(), css=".gr-button { margin: 5px; } .status-text { font-size: 0.9em; color: #555; }") as demo:
179
+ gr.Markdown("# 🤖 AI Research Agent")
180
+ with gr.Row(variant="compact"):
181
+ agent_stat_tb = gr.Textbox(label="Agent Status", value="Initializing...", interactive=False, elem_classes=["status-text"], scale=4)
182
+ with gr.Column(scale=1, min_width=150):
183
+ memory_backend_info_tb = gr.Textbox(label="Memory Backend", value=MEMORY_STORAGE_BACKEND, interactive=False)
184
+ hf_repos_display = gr.Textbox(label="HF Repos", value=f"M: {MEMORY_HF_MEM_REPO}, R: {MEMORY_HF_RULES_REPO}", interactive=False, visible=MEMORY_STORAGE_BACKEND == "HF_DATASET")
185
+ with gr.Row():
186
+ with gr.Sidebar():
187
+ gr.Markdown("## ⚙️ Configuration")
188
+ with gr.Group():
189
+ api_key_tb = gr.Textbox(label="API Key (Override)", type="password", placeholder="Uses .env if blank")
190
+ available_providers = get_available_providers()
191
+ default_provider = available_providers[0] if available_providers else None
192
+ prov_sel_dd = gr.Dropdown(label="AI Provider", choices=available_providers, value=default_provider, interactive=True)
193
+ model_sel_dd = gr.Dropdown(label="AI Model", choices=get_model_display_names_for_provider(default_provider) if default_provider else [], value=get_default_model_display_name_for_provider(default_provider), interactive=True)
194
+ with gr.Group():
195
+ sys_prompt_tb = gr.Textbox(label="System Prompt", lines=8, value=DEFAULT_SYSTEM_PROMPT, interactive=True)
196
+ if MEMORY_STORAGE_BACKEND == "RAM":
197
+ save_faiss_sidebar_btn = gr.Button("Save FAISS Indices", variant="secondary")
198
+ with gr.Column(scale=3):
199
+ with gr.Tabs():
200
+ with gr.TabItem("💬 Chat & Research"):
201
+ main_chat_disp = gr.Chatbot(height=400, show_copy_button=True, render_markdown=True)
202
+ with gr.Row(variant="compact"):
203
+ user_msg_tb = gr.Textbox(show_label=False, placeholder="Ask your research question...", scale=7, lines=1)
204
+ send_btn = gr.Button("Send", variant="primary", scale=1, min_width=100)
205
+ with gr.Accordion("📝 Detailed Response & Insights", open=False):
206
+ fmt_report_tb = gr.Textbox(label="Full AI Response", lines=8, interactive=True, show_copy_button=True)
207
+ dl_report_btn = gr.DownloadButton("Download Report", value=None, interactive=False, visible=False)
208
+ detect_out_md = gr.Markdown(visible=False)
209
+ with gr.TabItem("🧠 Knowledge Base"):
210
+ with gr.Row(equal_height=True):
211
+ with gr.Column():
212
+ gr.Markdown("### 📜 Rules Management")
213
+ rules_disp_ta = gr.TextArea(label="Current Rules", lines=10, interactive=True)
214
+ save_edited_rules_btn = gr.Button("💾 Save Edited Text", variant="primary")
215
+ with gr.Row(variant="compact"):
216
+ dl_rules_btn = gr.DownloadButton("⬇️ Download Rules")
217
+ clear_rules_btn = gr.Button("🗑️ Clear All Rules", variant="stop")
218
+ upload_rules_fobj = gr.File(label="Upload Rules File (.txt/.jsonl)", file_types=[".txt", ".jsonl"])
219
+ rules_stat_tb = gr.Textbox(label="Rules Status", interactive=False, lines=1)
220
+ with gr.Column():
221
+ gr.Markdown("### 📚 Memories Management")
222
+ mems_disp_json = gr.JSON(label="Current Memories", value=[])
223
+ with gr.Row(variant="compact"):
224
+ dl_mems_btn = gr.DownloadButton("⬇️ Download Memories")
225
+ clear_mems_btn = gr.Button("🗑️ Clear All Memories", variant="stop")
226
+ upload_mems_fobj = gr.File(label="Upload Memories File (.json/.jsonl)", file_types=[".json", ".jsonl"])
227
+ mems_stat_tb = gr.Textbox(label="Memories Status", interactive=False, lines=1)
228
+
229
+ prov_sel_dd.change(lambda p: gr.Dropdown(choices=get_model_display_names_for_provider(p), value=get_default_model_display_name_for_provider(p), interactive=True), prov_sel_dd, model_sel_dd)
230
+ chat_ins = [user_msg_tb, main_chat_disp, prov_sel_dd, model_sel_dd, api_key_tb, sys_prompt_tb]
231
+ chat_outs = [user_msg_tb, main_chat_disp, agent_stat_tb, detect_out_md, fmt_report_tb, dl_report_btn, rules_disp_ta, mems_disp_json]
232
+ chat_event_args = {"fn": handle_gradio_chat_submit, "inputs": chat_ins, "outputs": chat_outs}
233
+ send_btn.click(**chat_event_args)
234
+ user_msg_tb.submit(**chat_event_args)
235
+
236
+ dl_rules_btn.click(ui_download_rules_action_fn, None, dl_rules_btn)
237
+ save_edited_rules_btn.click(save_edited_rules_action_fn, [rules_disp_ta], [rules_stat_tb]).then(ui_refresh_rules_display_fn, outputs=rules_disp_ta)
238
+ upload_rules_fobj.upload(ui_upload_rules_action_fn, [upload_rules_fobj], [rules_stat_tb]).then(ui_refresh_rules_display_fn, outputs=rules_disp_ta)
239
+ clear_rules_btn.click(lambda: ("Cleared." if clear_all_rules_data_backend() else "Error."), outputs=rules_stat_tb).then(ui_refresh_rules_display_fn, outputs=rules_disp_ta)
240
+
241
+ dl_mems_btn.click(ui_download_memories_action_fn, None, dl_mems_btn)
242
+ upload_mems_fobj.upload(ui_upload_memories_action_fn, [upload_mems_fobj], [mems_stat_tb]).then(ui_refresh_memories_display_fn, outputs=mems_disp_json)
243
+ clear_mems_btn.click(lambda: ("Cleared." if clear_all_memory_data_backend() else "Error."), outputs=mems_stat_tb).then(ui_refresh_memories_display_fn, outputs=mems_disp_json)
244
+
245
+ if MEMORY_STORAGE_BACKEND == "RAM" and 'save_faiss_sidebar_btn' in locals():
246
+ save_faiss_sidebar_btn.click(lambda: (gr.Info("Saved FAISS to disk.") if save_faiss_indices_to_disk() is None else gr.Error("Error saving FAISS.")), None, None)
247
+
248
+ app_load_outputs = [agent_stat_tb, rules_disp_ta, mems_disp_json, detect_out_md, fmt_report_tb, dl_report_btn]
249
+ demo.load(fn=app_load_fn, inputs=None, outputs=app_load_outputs)
250
+
251
+ if __name__ == "__main__":
252
+ app_port = int(os.getenv("GRADIO_PORT", 7860))
253
+ app_server = os.getenv("GRADIO_SERVER_NAME", "127.0.0.1")
254
+ logger.info(f"Launching Gradio server: http://{app_server}:{app_port}")
255
+ demo.queue().launch(server_name=app_server, server_port=app_port)