Spaces:
Sleeping
feat: Add Gemini/Ollama toggle functionality with API error fixes
Browse files✨ New Features:
- Add interactive toggle switch in UI to choose between Ollama (Local) and Gemini (Cloud)
- Dynamic LLM selection with preference persistence in localStorage
- Support both Ollama and Gemini in research agent with automatic fallbacks
- Real-time status updates showing which AI model is being used
�� Backend Changes:
- Update research agent to support both LLMs with use_gemini parameter
- Add _research_with_gemini_tools method for Gemini-powered research
- Update query request model to include use_gemini field
- Automatic LLM initialization and availability detection
🎨 Frontend Changes:
- Beautiful toggle switch with animated slider and labels
- Persistent user preference storage
- Dynamic status messages based on selected LLM
- Responsive design for header controls
🛠️ Bug Fixes:
- Fix CoinGecko API 401 errors by adding proper fallback to mock data
- Improved error handling for chart_data_tool API failures
- Better session cleanup to prevent connection warnings
🔒 Safety & Performance:
- AI safety validation for both Ollama and Gemini responses
- Proper timeout handling for both LLM types
- Graceful degradation when APIs fail or models unavailable
- app.py +11 -8
- src/agent/research_agent.py +165 -21
- src/tools/chart_data_tool.py +41 -35
- src/utils/config.py +3 -3
- static/app.js +38 -1
- static/styles.css +103 -0
- templates/index.html +15 -3
@@ -37,6 +37,7 @@ templates = Jinja2Templates(directory="templates")
|
|
37 |
class QueryRequest(BaseModel):
|
38 |
query: str
|
39 |
chat_history: Optional[List[Dict[str, str]]] = []
|
|
|
40 |
|
41 |
class QueryResponse(BaseModel):
|
42 |
success: bool
|
@@ -85,7 +86,7 @@ class Web3CoPilotService:
|
|
85 |
self.airaa = None
|
86 |
self.viz = None
|
87 |
|
88 |
-
async def process_query(self, query: str) -> QueryResponse:
|
89 |
"""Process research query with comprehensive analysis"""
|
90 |
logger.info("Processing research request...")
|
91 |
|
@@ -113,7 +114,7 @@ Configure GEMINI_API_KEY environment variable for full AI analysis."""
|
|
113 |
logger.info("🤖 Processing with AI research agent...")
|
114 |
logger.info(f"🛠️ Available tools: {[tool.name for tool in self.agent.tools] if self.agent else []}")
|
115 |
|
116 |
-
result = await self.agent.research_query(query)
|
117 |
logger.info(f"🔄 Agent research completed: success={result.get('success')}")
|
118 |
|
119 |
if result.get("success"):
|
@@ -464,22 +465,23 @@ async def process_query_stream(request: QueryRequest):
|
|
464 |
yield f"data: {json.dumps({'type': 'status', 'message': 'Executing tools and gathering data...', 'progress': 50})}\n\n"
|
465 |
await asyncio.sleep(0.5)
|
466 |
|
467 |
-
# Send Ollama processing status with heartbeats
|
468 |
-
|
|
|
469 |
await asyncio.sleep(1.0)
|
470 |
|
471 |
# Send additional heartbeat messages during processing
|
472 |
-
yield f"data: {json.dumps({'type': 'status', 'message': '
|
473 |
await asyncio.sleep(2.0)
|
474 |
|
475 |
-
yield f"data: {json.dumps({'type': 'status', 'message': 'Still processing...
|
476 |
await asyncio.sleep(3.0)
|
477 |
|
478 |
# Process the actual query with timeout and periodic heartbeats
|
479 |
start_time = datetime.now()
|
480 |
|
481 |
# Create a task for the query processing
|
482 |
-
query_task = asyncio.create_task(service.process_query(request.query))
|
483 |
|
484 |
try:
|
485 |
# Send periodic heartbeats while waiting for Ollama
|
@@ -499,7 +501,8 @@ async def process_query_stream(request: QueryRequest):
|
|
499 |
raise asyncio.TimeoutError("Hard timeout reached")
|
500 |
|
501 |
progress = min(85 + (heartbeat_count * 2), 95) # Progress slowly from 85 to 95
|
502 |
-
|
|
|
503 |
|
504 |
# If we get here, the query completed successfully
|
505 |
result = query_task.result()
|
|
|
37 |
class QueryRequest(BaseModel):
|
38 |
query: str
|
39 |
chat_history: Optional[List[Dict[str, str]]] = []
|
40 |
+
use_gemini: bool = False
|
41 |
|
42 |
class QueryResponse(BaseModel):
|
43 |
success: bool
|
|
|
86 |
self.airaa = None
|
87 |
self.viz = None
|
88 |
|
89 |
+
async def process_query(self, query: str, use_gemini: bool = False) -> QueryResponse:
|
90 |
"""Process research query with comprehensive analysis"""
|
91 |
logger.info("Processing research request...")
|
92 |
|
|
|
114 |
logger.info("🤖 Processing with AI research agent...")
|
115 |
logger.info(f"🛠️ Available tools: {[tool.name for tool in self.agent.tools] if self.agent else []}")
|
116 |
|
117 |
+
result = await self.agent.research_query(query, use_gemini=use_gemini)
|
118 |
logger.info(f"🔄 Agent research completed: success={result.get('success')}")
|
119 |
|
120 |
if result.get("success"):
|
|
|
465 |
yield f"data: {json.dumps({'type': 'status', 'message': 'Executing tools and gathering data...', 'progress': 50})}\n\n"
|
466 |
await asyncio.sleep(0.5)
|
467 |
|
468 |
+
# Send Ollama/Gemini processing status with heartbeats
|
469 |
+
llm_name = "Gemini" if request.use_gemini else "Ollama"
|
470 |
+
yield f"data: {json.dumps({'type': 'status', 'message': f'{llm_name} is analyzing data and generating response...', 'progress': 70})}\n\n"
|
471 |
await asyncio.sleep(1.0)
|
472 |
|
473 |
# Send additional heartbeat messages during processing
|
474 |
+
yield f"data: {json.dumps({'type': 'status', 'message': f'{llm_name} is thinking deeply about your query...', 'progress': 75})}\n\n"
|
475 |
await asyncio.sleep(2.0)
|
476 |
|
477 |
+
yield f"data: {json.dumps({'type': 'status', 'message': f'Still processing... {llm_name} generates detailed responses', 'progress': 80})}\n\n"
|
478 |
await asyncio.sleep(3.0)
|
479 |
|
480 |
# Process the actual query with timeout and periodic heartbeats
|
481 |
start_time = datetime.now()
|
482 |
|
483 |
# Create a task for the query processing
|
484 |
+
query_task = asyncio.create_task(service.process_query(request.query, request.use_gemini))
|
485 |
|
486 |
try:
|
487 |
# Send periodic heartbeats while waiting for Ollama
|
|
|
501 |
raise asyncio.TimeoutError("Hard timeout reached")
|
502 |
|
503 |
progress = min(85 + (heartbeat_count * 2), 95) # Progress slowly from 85 to 95
|
504 |
+
llm_name = "Gemini" if request.use_gemini else "Ollama"
|
505 |
+
yield f"data: {json.dumps({'type': 'status', 'message': f'{llm_name} is still working... ({elapsed:.0f}s elapsed)', 'progress': progress})}\n\n"
|
506 |
|
507 |
# If we get here, the query completed successfully
|
508 |
result = query_task.result()
|
@@ -22,41 +22,60 @@ class Web3ResearchAgent:
|
|
22 |
self.fallback_llm = None
|
23 |
self.tools = []
|
24 |
self.enabled = False
|
|
|
25 |
|
26 |
try:
|
27 |
-
|
28 |
-
|
29 |
-
|
30 |
-
|
31 |
-
|
32 |
-
|
|
|
|
|
|
|
|
|
|
|
33 |
|
34 |
except Exception as e:
|
35 |
logger.error(f"Agent initialization failed: {e}")
|
36 |
self.enabled = False
|
37 |
|
38 |
-
def
|
39 |
-
"""Initialize
|
40 |
try:
|
41 |
self.fallback_llm = Ollama(
|
42 |
model=config.OLLAMA_MODEL,
|
43 |
base_url=config.OLLAMA_BASE_URL,
|
44 |
temperature=0.1
|
45 |
)
|
46 |
-
|
47 |
logger.info(f"✅ Ollama initialized - Model: {config.OLLAMA_MODEL}")
|
48 |
-
|
49 |
-
self.tools = self._initialize_tools()
|
50 |
-
self.enabled = True
|
51 |
-
|
52 |
except Exception as e:
|
53 |
logger.error(f"Ollama initialization failed: {e}")
|
54 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
55 |
|
56 |
def _init_with_gemini_fallback(self):
|
57 |
-
"""Initialize with Gemini primary and Ollama fallback"""
|
58 |
-
|
59 |
-
|
60 |
|
61 |
def _initialize_tools(self):
|
62 |
tools = []
|
@@ -93,8 +112,8 @@ class Web3ResearchAgent:
|
|
93 |
|
94 |
return tools
|
95 |
|
96 |
-
async def research_query(self, query: str) -> Dict[str, Any]:
|
97 |
-
"""Research query with
|
98 |
|
99 |
# AI Safety Check 1: Sanitize and validate input
|
100 |
sanitized_query, is_safe, safety_reason = ai_safety.sanitize_query(query)
|
@@ -140,8 +159,13 @@ class Web3ResearchAgent:
|
|
140 |
}
|
141 |
|
142 |
try:
|
143 |
-
|
144 |
-
|
|
|
|
|
|
|
|
|
|
|
145 |
|
146 |
except Exception as e:
|
147 |
logger.error(f"Research failed: {e}")
|
@@ -342,6 +366,126 @@ The system successfully gathered data from {len(suggested_tools)} tools:
|
|
342 |
logger.error(f"Ollama tools research failed: {e}")
|
343 |
raise e
|
344 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
345 |
def _extract_sources(self, response: str) -> List[str]:
|
346 |
"""Extract sources from response"""
|
347 |
# Simple source extraction - can be enhanced
|
|
|
22 |
self.fallback_llm = None
|
23 |
self.tools = []
|
24 |
self.enabled = False
|
25 |
+
self.gemini_available = False
|
26 |
|
27 |
try:
|
28 |
+
# Always initialize Ollama
|
29 |
+
logger.info("🔧 Initializing Ollama as fallback")
|
30 |
+
self._init_ollama()
|
31 |
+
|
32 |
+
# Try to initialize Gemini if API key is available
|
33 |
+
if config.GEMINI_API_KEY:
|
34 |
+
logger.info("🔧 Initializing Gemini as primary option")
|
35 |
+
self._init_gemini()
|
36 |
+
|
37 |
+
self.tools = self._initialize_tools()
|
38 |
+
self.enabled = True
|
39 |
|
40 |
except Exception as e:
|
41 |
logger.error(f"Agent initialization failed: {e}")
|
42 |
self.enabled = False
|
43 |
|
44 |
+
def _init_ollama(self):
|
45 |
+
"""Initialize Ollama LLM"""
|
46 |
try:
|
47 |
self.fallback_llm = Ollama(
|
48 |
model=config.OLLAMA_MODEL,
|
49 |
base_url=config.OLLAMA_BASE_URL,
|
50 |
temperature=0.1
|
51 |
)
|
|
|
52 |
logger.info(f"✅ Ollama initialized - Model: {config.OLLAMA_MODEL}")
|
|
|
|
|
|
|
|
|
53 |
except Exception as e:
|
54 |
logger.error(f"Ollama initialization failed: {e}")
|
55 |
+
raise
|
56 |
+
|
57 |
+
def _init_gemini(self):
|
58 |
+
"""Initialize Gemini LLM"""
|
59 |
+
try:
|
60 |
+
self.llm = ChatGoogleGenerativeAI(
|
61 |
+
model="gemini-pro",
|
62 |
+
google_api_key=config.GEMINI_API_KEY,
|
63 |
+
temperature=0.1
|
64 |
+
)
|
65 |
+
self.gemini_available = True
|
66 |
+
logger.info("✅ Gemini initialized")
|
67 |
+
except Exception as e:
|
68 |
+
logger.warning(f"Gemini initialization failed: {e}")
|
69 |
+
self.gemini_available = False
|
70 |
+
|
71 |
+
def _init_ollama_only(self):
|
72 |
+
"""Initialize with only Ollama LLM (deprecated - kept for compatibility)"""
|
73 |
+
self._init_ollama()
|
74 |
|
75 |
def _init_with_gemini_fallback(self):
|
76 |
+
"""Initialize with Gemini primary and Ollama fallback (deprecated - kept for compatibility)"""
|
77 |
+
self._init_ollama()
|
78 |
+
self._init_gemini()
|
79 |
|
80 |
def _initialize_tools(self):
|
81 |
tools = []
|
|
|
112 |
|
113 |
return tools
|
114 |
|
115 |
+
async def research_query(self, query: str, use_gemini: bool = False) -> Dict[str, Any]:
|
116 |
+
"""Research query with dynamic LLM selection - Enhanced with AI Safety"""
|
117 |
|
118 |
# AI Safety Check 1: Sanitize and validate input
|
119 |
sanitized_query, is_safe, safety_reason = ai_safety.sanitize_query(query)
|
|
|
159 |
}
|
160 |
|
161 |
try:
|
162 |
+
# Choose LLM based on user preference and availability
|
163 |
+
if use_gemini and self.gemini_available:
|
164 |
+
logger.info("🤖 Processing with Gemini + Tools (Safety Enhanced)")
|
165 |
+
return await self._research_with_gemini_tools(sanitized_query)
|
166 |
+
else:
|
167 |
+
logger.info("🤖 Processing with Ollama + Tools (Safety Enhanced)")
|
168 |
+
return await self._research_with_ollama_tools(sanitized_query)
|
169 |
|
170 |
except Exception as e:
|
171 |
logger.error(f"Research failed: {e}")
|
|
|
366 |
logger.error(f"Ollama tools research failed: {e}")
|
367 |
raise e
|
368 |
|
369 |
+
async def _research_with_gemini_tools(self, query: str) -> Dict[str, Any]:
|
370 |
+
"""Research using Gemini with tools"""
|
371 |
+
try:
|
372 |
+
# Step 1: Analyze query and suggest tools using Gemini
|
373 |
+
tool_analysis_prompt = f"""Based on this Web3/cryptocurrency research query, identify the most relevant tools to use.
|
374 |
+
|
375 |
+
Query: "{query}"
|
376 |
+
|
377 |
+
Available tools:
|
378 |
+
- cryptocompare_data: Get current cryptocurrency prices and basic info
|
379 |
+
- coingecko_data: Comprehensive market data and analytics
|
380 |
+
- defillama_data: DeFi protocols, TVL, and yield farming data
|
381 |
+
- etherscan_data: Ethereum blockchain data and transactions
|
382 |
+
- chart_data_provider: Generate chart data for visualizations
|
383 |
+
|
384 |
+
If charts/visualizations are mentioned, include chart_data_provider.
|
385 |
+
|
386 |
+
Examples:
|
387 |
+
- "Bitcoin price" → cryptocompare_data, chart_data_provider
|
388 |
+
- "DeFi TVL" → defillama_data, chart_data_provider
|
389 |
+
- "Ethereum transactions" → etherscan_data
|
390 |
+
|
391 |
+
Respond with only the tool names, comma-separated (no explanations)."""
|
392 |
+
|
393 |
+
tool_response = await self.llm.ainvoke(tool_analysis_prompt)
|
394 |
+
|
395 |
+
logger.info(f"🧠 Gemini tool analysis response: {str(tool_response)[:100]}...")
|
396 |
+
|
397 |
+
# Parse suggested tools
|
398 |
+
suggested_tools = [tool.strip() for tool in str(tool_response).split(',') if tool.strip()]
|
399 |
+
suggested_tools = [tool for tool in suggested_tools if tool in {
|
400 |
+
'cryptocompare_data', 'coingecko_data', 'defillama_data',
|
401 |
+
'etherscan_data', 'chart_data_provider'
|
402 |
+
}]
|
403 |
+
|
404 |
+
logger.info(f"🛠️ Gemini suggested tools: {suggested_tools}")
|
405 |
+
|
406 |
+
# Step 2: Execute tools (same logic as Ollama version)
|
407 |
+
tool_results = []
|
408 |
+
for tool_name in suggested_tools:
|
409 |
+
tool = next((t for t in self.tools if t.name == tool_name), None)
|
410 |
+
if tool:
|
411 |
+
try:
|
412 |
+
logger.info(f"🔧 Executing {tool_name}")
|
413 |
+
|
414 |
+
# Handle chart_data_provider with proper parameters
|
415 |
+
if tool_name == "chart_data_provider":
|
416 |
+
chart_type = "price_chart"
|
417 |
+
symbol = "bitcoin"
|
418 |
+
|
419 |
+
if "defi" in query.lower() or "tvl" in query.lower():
|
420 |
+
chart_type = "defi_tvl"
|
421 |
+
elif "market" in query.lower() or "overview" in query.lower():
|
422 |
+
chart_type = "market_overview"
|
423 |
+
elif "gas" in query.lower():
|
424 |
+
chart_type = "gas_tracker"
|
425 |
+
|
426 |
+
if "ethereum" in query.lower() or "eth" in query.lower():
|
427 |
+
symbol = "ethereum"
|
428 |
+
elif "bitcoin" in query.lower() or "btc" in query.lower():
|
429 |
+
symbol = "bitcoin"
|
430 |
+
|
431 |
+
result = await tool._arun(chart_type=chart_type, symbol=symbol)
|
432 |
+
else:
|
433 |
+
result = await tool._arun(query)
|
434 |
+
|
435 |
+
logger.info(f"📊 {tool_name} result preview: {str(result)[:200]}...")
|
436 |
+
tool_results.append(f"=== {tool_name} Results ===\n{result}\n")
|
437 |
+
except Exception as e:
|
438 |
+
logger.error(f"Tool {tool_name} failed: {e}")
|
439 |
+
tool_results.append(f"=== {tool_name} Error ===\nTool failed: {str(e)}\n")
|
440 |
+
|
441 |
+
# Step 3: Generate final response with Gemini
|
442 |
+
context = "\n".join(tool_results) if tool_results else "No tool data available - provide general information."
|
443 |
+
|
444 |
+
final_prompt = ai_safety.create_safe_prompt(query, context)
|
445 |
+
|
446 |
+
try:
|
447 |
+
final_response = await asyncio.wait_for(
|
448 |
+
self.llm.ainvoke(final_prompt),
|
449 |
+
timeout=30
|
450 |
+
)
|
451 |
+
logger.info(f"🎯 Gemini final response preview: {str(final_response)[:300]}...")
|
452 |
+
|
453 |
+
# AI Safety Check: Validate response
|
454 |
+
clean_response, response_safe, response_reason = ai_safety.validate_gemini_response(str(final_response))
|
455 |
+
if not response_safe:
|
456 |
+
ai_safety.log_safety_event("blocked_gemini_response", {
|
457 |
+
"reason": response_reason,
|
458 |
+
"query": query[:100],
|
459 |
+
"timestamp": datetime.now().isoformat()
|
460 |
+
})
|
461 |
+
clean_response = f"## Cryptocurrency Analysis\n\nBased on the available data:\n\n{context[:1000]}\n\n*Response filtered for safety*"
|
462 |
+
|
463 |
+
final_response = clean_response
|
464 |
+
|
465 |
+
except asyncio.TimeoutError:
|
466 |
+
logger.warning("⏱️ Gemini final response timed out, using tool data directly")
|
467 |
+
final_response = f"## Web3 Research Analysis\n\n{context[:1500]}\n\n*Analysis completed using available tools - Gemini response timed out*"
|
468 |
+
|
469 |
+
logger.info("✅ Research successful with Gemini + tools")
|
470 |
+
|
471 |
+
return {
|
472 |
+
"success": True,
|
473 |
+
"query": query,
|
474 |
+
"result": final_response,
|
475 |
+
"sources": [],
|
476 |
+
"metadata": {
|
477 |
+
"llm_used": f"Gemini ({self.llm.model_name if hasattr(self.llm, 'model_name') else 'gemini-pro'})",
|
478 |
+
"tools_used": suggested_tools,
|
479 |
+
"timestamp": datetime.now().isoformat()
|
480 |
+
}
|
481 |
+
}
|
482 |
+
|
483 |
+
except Exception as e:
|
484 |
+
logger.error(f"Gemini tools research failed: {e}")
|
485 |
+
# Fallback to Ollama if Gemini fails
|
486 |
+
logger.info("🔄 Falling back to Ollama due to Gemini error")
|
487 |
+
return await self._research_with_ollama_tools(query)
|
488 |
+
|
489 |
def _extract_sources(self, response: str) -> List[str]:
|
490 |
"""Extract sources from response"""
|
491 |
# Simple source extraction - can be enhanced
|
@@ -84,9 +84,9 @@ class ChartDataTool(BaseTool):
|
|
84 |
})
|
85 |
|
86 |
async def _get_price_chart_data(self, symbol: str, days: int) -> str:
|
87 |
-
"""Get
|
88 |
try:
|
89 |
-
#
|
90 |
from src.tools.coingecko_tool import CoinGeckoTool
|
91 |
|
92 |
coingecko = CoinGeckoTool()
|
@@ -107,42 +107,48 @@ class ChartDataTool(BaseTool):
|
|
107 |
|
108 |
coin_id = symbol_map.get(symbol.lower(), symbol.lower())
|
109 |
|
110 |
-
|
111 |
-
|
112 |
-
|
113 |
-
|
114 |
-
|
115 |
-
|
116 |
-
|
117 |
-
|
118 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
119 |
return await self._get_mock_price_data(symbol, days)
|
120 |
|
121 |
-
# Format the real data
|
122 |
-
price_data = data.get("prices", [])
|
123 |
-
volume_data = data.get("total_volumes", [])
|
124 |
-
|
125 |
-
# Get current coin info
|
126 |
-
coin_info = await coingecko.make_request(f"https://api.coingecko.com/api/v3/coins/{coin_id}")
|
127 |
-
coin_name = coin_info.get("name", symbol.title()) if coin_info else symbol.title()
|
128 |
-
|
129 |
-
return json.dumps({
|
130 |
-
"chart_type": "price_chart",
|
131 |
-
"data": {
|
132 |
-
"prices": price_data,
|
133 |
-
"total_volumes": volume_data,
|
134 |
-
"symbol": symbol.upper(),
|
135 |
-
"name": coin_name
|
136 |
-
},
|
137 |
-
"config": {
|
138 |
-
"title": f"{coin_name} Price Analysis ({days} days)",
|
139 |
-
"timeframe": f"{days}d",
|
140 |
-
"currency": "USD"
|
141 |
-
}
|
142 |
-
})
|
143 |
-
|
144 |
except Exception as e:
|
145 |
-
logger.error(f"
|
|
|
146 |
return await self._get_mock_price_data(symbol, days)
|
147 |
|
148 |
async def _get_mock_price_data(self, symbol: str, days: int) -> str:
|
|
|
84 |
})
|
85 |
|
86 |
async def _get_price_chart_data(self, symbol: str, days: int) -> str:
|
87 |
+
"""Get price chart data with fallback for API failures"""
|
88 |
try:
|
89 |
+
# First try to get real data from CoinGecko
|
90 |
from src.tools.coingecko_tool import CoinGeckoTool
|
91 |
|
92 |
coingecko = CoinGeckoTool()
|
|
|
107 |
|
108 |
coin_id = symbol_map.get(symbol.lower(), symbol.lower())
|
109 |
|
110 |
+
try:
|
111 |
+
# Use basic API endpoint that doesn't require premium
|
112 |
+
url = f"https://api.coingecko.com/api/v3/coins/{coin_id}/market_chart"
|
113 |
+
params = {"vs_currency": "usd", "days": days, "interval": "daily"}
|
114 |
+
|
115 |
+
data = await coingecko.make_request(url, params=params)
|
116 |
+
|
117 |
+
if data and "prices" in data:
|
118 |
+
# Format the real data
|
119 |
+
price_data = data.get("prices", [])
|
120 |
+
volume_data = data.get("total_volumes", [])
|
121 |
+
|
122 |
+
# Get current coin info
|
123 |
+
coin_info = await coingecko.make_request(f"https://api.coingecko.com/api/v3/coins/{coin_id}")
|
124 |
+
coin_name = coin_info.get("name", symbol.title()) if coin_info else symbol.title()
|
125 |
+
|
126 |
+
return json.dumps({
|
127 |
+
"chart_type": "price_chart",
|
128 |
+
"data": {
|
129 |
+
"prices": price_data,
|
130 |
+
"total_volumes": volume_data,
|
131 |
+
"symbol": symbol.upper(),
|
132 |
+
"name": coin_name
|
133 |
+
},
|
134 |
+
"config": {
|
135 |
+
"title": f"{coin_name} Price Analysis ({days} days)",
|
136 |
+
"timeframe": f"{days}d",
|
137 |
+
"currency": "USD"
|
138 |
+
}
|
139 |
+
})
|
140 |
+
else:
|
141 |
+
raise Exception("No price data in response")
|
142 |
+
|
143 |
+
except Exception as api_error:
|
144 |
+
logger.error(f"Real price data failed: {api_error}")
|
145 |
+
# Fallback to mock data on any API error
|
146 |
+
logger.info(f"Using fallback mock data for {symbol}")
|
147 |
return await self._get_mock_price_data(symbol, days)
|
148 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
149 |
except Exception as e:
|
150 |
+
logger.error(f"Price chart data generation failed: {e}")
|
151 |
+
# Final fallback to mock data
|
152 |
return await self._get_mock_price_data(symbol, days)
|
153 |
|
154 |
async def _get_mock_price_data(self, symbol: str, days: int) -> str:
|
@@ -8,9 +8,9 @@ load_dotenv()
|
|
8 |
|
9 |
@dataclass
|
10 |
class Config:
|
11 |
-
# LLM Configuration - Ollama
|
12 |
-
GEMINI_API_KEY: str = "" #
|
13 |
-
USE_OLLAMA_ONLY: bool =
|
14 |
|
15 |
# Available API Keys
|
16 |
COINGECKO_API_KEY: Optional[str] = None # Not available - costs money
|
|
|
8 |
|
9 |
@dataclass
|
10 |
class Config:
|
11 |
+
# LLM Configuration - Both Ollama and Gemini available
|
12 |
+
GEMINI_API_KEY: str = os.getenv("GEMINI_API_KEY", "") # Enable Gemini when API key provided
|
13 |
+
USE_OLLAMA_ONLY: bool = not bool(os.getenv("GEMINI_API_KEY")) # Auto-detect based on API key
|
14 |
|
15 |
# Available API Keys
|
16 |
COINGECKO_API_KEY: Optional[str] = None # Not available - costs money
|
@@ -1,5 +1,38 @@
|
|
1 |
let chatHistory = [];
|
2 |
let messageCount = 0;
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
|
4 |
async function checkStatus() {
|
5 |
try {
|
@@ -68,7 +101,11 @@ async function sendQuery() {
|
|
68 |
'Accept': 'text/event-stream',
|
69 |
'Cache-Control': 'no-cache'
|
70 |
},
|
71 |
-
body: JSON.stringify({
|
|
|
|
|
|
|
|
|
72 |
signal: controller.signal,
|
73 |
// Disable browser's default timeout behavior
|
74 |
keepalive: true
|
|
|
1 |
let chatHistory = [];
|
2 |
let messageCount = 0;
|
3 |
+
let useGemini = false; // Track current LLM choice
|
4 |
+
|
5 |
+
// Initialize Gemini toggle
|
6 |
+
document.addEventListener('DOMContentLoaded', function() {
|
7 |
+
const geminiToggle = document.getElementById('geminiToggle');
|
8 |
+
const toggleLabel = document.querySelector('.toggle-label');
|
9 |
+
|
10 |
+
// Load saved preference
|
11 |
+
useGemini = localStorage.getItem('useGemini') === 'true';
|
12 |
+
geminiToggle.checked = useGemini;
|
13 |
+
updateToggleLabel();
|
14 |
+
|
15 |
+
// Handle toggle changes
|
16 |
+
geminiToggle.addEventListener('change', function() {
|
17 |
+
useGemini = this.checked;
|
18 |
+
localStorage.setItem('useGemini', useGemini.toString());
|
19 |
+
updateToggleLabel();
|
20 |
+
console.log(`Switched to ${useGemini ? 'Gemini' : 'Ollama'} mode`);
|
21 |
+
|
22 |
+
// Show confirmation
|
23 |
+
showStatus(`Switched to ${useGemini ? 'Gemini (Cloud AI)' : 'Ollama (Local AI)'} mode`, 'info');
|
24 |
+
|
25 |
+
// Refresh status to reflect changes
|
26 |
+
checkStatus();
|
27 |
+
});
|
28 |
+
});
|
29 |
+
|
30 |
+
function updateToggleLabel() {
|
31 |
+
const toggleLabel = document.querySelector('.toggle-label');
|
32 |
+
if (toggleLabel) {
|
33 |
+
toggleLabel.textContent = `AI Model: ${useGemini ? 'Gemini' : 'Ollama'}`;
|
34 |
+
}
|
35 |
+
}
|
36 |
|
37 |
async function checkStatus() {
|
38 |
try {
|
|
|
101 |
'Accept': 'text/event-stream',
|
102 |
'Cache-Control': 'no-cache'
|
103 |
},
|
104 |
+
body: JSON.stringify({
|
105 |
+
query,
|
106 |
+
chat_history: chatHistory,
|
107 |
+
use_gemini: useGemini
|
108 |
+
}),
|
109 |
signal: controller.signal,
|
110 |
// Disable browser's default timeout behavior
|
111 |
keepalive: true
|
@@ -66,6 +66,109 @@ body {
|
|
66 |
flex: 1;
|
67 |
text-align: center;
|
68 |
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
69 |
.theme-toggle {
|
70 |
background: var(--surface);
|
71 |
border: 1px solid var(--border);
|
|
|
66 |
flex: 1;
|
67 |
text-align: center;
|
68 |
}
|
69 |
+
.header-controls {
|
70 |
+
display: flex;
|
71 |
+
align-items: center;
|
72 |
+
gap: 1rem;
|
73 |
+
}
|
74 |
+
|
75 |
+
/* LLM Toggle Switch Styles */
|
76 |
+
.llm-toggle {
|
77 |
+
display: flex;
|
78 |
+
align-items: center;
|
79 |
+
gap: 0.5rem;
|
80 |
+
}
|
81 |
+
.toggle-label {
|
82 |
+
font-size: 0.875rem;
|
83 |
+
color: var(--text-secondary);
|
84 |
+
font-weight: 500;
|
85 |
+
}
|
86 |
+
|
87 |
+
.switch {
|
88 |
+
position: relative;
|
89 |
+
display: inline-block;
|
90 |
+
width: 80px;
|
91 |
+
height: 32px;
|
92 |
+
}
|
93 |
+
|
94 |
+
.switch input {
|
95 |
+
opacity: 0;
|
96 |
+
width: 0;
|
97 |
+
height: 0;
|
98 |
+
}
|
99 |
+
|
100 |
+
.slider {
|
101 |
+
position: absolute;
|
102 |
+
cursor: pointer;
|
103 |
+
top: 0;
|
104 |
+
left: 0;
|
105 |
+
right: 0;
|
106 |
+
bottom: 0;
|
107 |
+
background-color: var(--surface);
|
108 |
+
border: 1px solid var(--border);
|
109 |
+
transition: .4s;
|
110 |
+
overflow: hidden;
|
111 |
+
}
|
112 |
+
|
113 |
+
.slider:before {
|
114 |
+
position: absolute;
|
115 |
+
content: "";
|
116 |
+
height: 24px;
|
117 |
+
width: 24px;
|
118 |
+
left: 3px;
|
119 |
+
bottom: 3px;
|
120 |
+
background-color: var(--primary);
|
121 |
+
transition: .4s;
|
122 |
+
border-radius: 50%;
|
123 |
+
z-index: 2;
|
124 |
+
}
|
125 |
+
|
126 |
+
.slider-text-off, .slider-text-on {
|
127 |
+
position: absolute;
|
128 |
+
color: var(--text-secondary);
|
129 |
+
font-size: 0.7rem;
|
130 |
+
font-weight: 500;
|
131 |
+
top: 50%;
|
132 |
+
transform: translateY(-50%);
|
133 |
+
transition: .4s;
|
134 |
+
pointer-events: none;
|
135 |
+
z-index: 1;
|
136 |
+
}
|
137 |
+
|
138 |
+
.slider-text-off {
|
139 |
+
left: 8px;
|
140 |
+
}
|
141 |
+
|
142 |
+
.slider-text-on {
|
143 |
+
right: 8px;
|
144 |
+
opacity: 0;
|
145 |
+
}
|
146 |
+
|
147 |
+
input:checked + .slider {
|
148 |
+
background-color: var(--accent);
|
149 |
+
border-color: var(--accent);
|
150 |
+
}
|
151 |
+
|
152 |
+
input:checked + .slider .slider-text-off {
|
153 |
+
opacity: 0;
|
154 |
+
}
|
155 |
+
|
156 |
+
input:checked + .slider .slider-text-on {
|
157 |
+
opacity: 1;
|
158 |
+
}
|
159 |
+
|
160 |
+
input:checked + .slider:before {
|
161 |
+
transform: translateX(48px);
|
162 |
+
}
|
163 |
+
|
164 |
+
.slider.round {
|
165 |
+
border-radius: 20px;
|
166 |
+
}
|
167 |
+
|
168 |
+
.slider.round:before {
|
169 |
+
border-radius: 50%;
|
170 |
+
}
|
171 |
+
|
172 |
.theme-toggle {
|
173 |
background: var(--surface);
|
174 |
border: 1px solid var(--border);
|
@@ -21,9 +21,21 @@
|
|
21 |
<h1><span class="brand">Web3</span> Research Co-Pilot</h1>
|
22 |
<p>Professional cryptocurrency analysis and market intelligence</p>
|
23 |
</div>
|
24 |
-
<
|
25 |
-
<
|
26 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
27 |
</div>
|
28 |
</div>
|
29 |
|
|
|
21 |
<h1><span class="brand">Web3</span> Research Co-Pilot</h1>
|
22 |
<p>Professional cryptocurrency analysis and market intelligence</p>
|
23 |
</div>
|
24 |
+
<div class="header-controls">
|
25 |
+
<div class="llm-toggle">
|
26 |
+
<span class="toggle-label">AI Model:</span>
|
27 |
+
<label class="switch">
|
28 |
+
<input type="checkbox" id="geminiToggle" title="Switch between Ollama (Local) and Gemini (Cloud)">
|
29 |
+
<span class="slider round">
|
30 |
+
<span class="slider-text-off">Ollama</span>
|
31 |
+
<span class="slider-text-on">Gemini</span>
|
32 |
+
</span>
|
33 |
+
</label>
|
34 |
+
</div>
|
35 |
+
<button id="themeToggle" class="theme-toggle" title="Toggle theme">
|
36 |
+
<i class="fas fa-moon"></i>
|
37 |
+
</button>
|
38 |
+
</div>
|
39 |
</div>
|
40 |
</div>
|
41 |
|