Spaces:
Paused
Paused
Update chat_handler.py
Browse files- chat_handler.py +38 -55
chat_handler.py
CHANGED
|
@@ -17,6 +17,10 @@ from api_executor import call_api as execute_api
|
|
| 17 |
from config_provider import ConfigProvider
|
| 18 |
from validation_engine import validate
|
| 19 |
from session import session_store, Session
|
|
|
|
|
|
|
|
|
|
|
|
|
| 20 |
|
| 21 |
# βββββββββββββββββββββββββ HELPERS βββββββββββββββββββββββββ #
|
| 22 |
def _trim_response(raw: str) -> str:
|
|
@@ -52,6 +56,32 @@ cfg = ConfigProvider.get()
|
|
| 52 |
SPARK_URL = str(cfg.global_config.spark_endpoint).rstrip("/")
|
| 53 |
ALLOWED_INTENTS = {"flight-booking", "flight-info", "booking-cancel"}
|
| 54 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 55 |
# βββββββββββββββββββββββββ SPARK βββββββββββββββββββββββββ #
|
| 56 |
def _get_spark_token() -> Optional[str]:
|
| 57 |
"""Get Spark token based on work_mode"""
|
|
@@ -68,65 +98,18 @@ def _get_spark_token() -> Optional[str]:
|
|
| 68 |
return os.getenv("SPARK_TOKEN")
|
| 69 |
|
| 70 |
async def spark_generate(s: Session, prompt: str, user_msg: str) -> str:
|
| 71 |
-
"""Call
|
| 72 |
try:
|
| 73 |
-
|
| 74 |
-
|
| 75 |
-
raise ValueError(f"Project not found: {s.project_name}")
|
| 76 |
-
|
| 77 |
-
version = next((v for v in project.versions if v.published), None)
|
| 78 |
-
if not version:
|
| 79 |
-
raise ValueError("No published version found")
|
| 80 |
-
|
| 81 |
-
# Get Spark token
|
| 82 |
-
spark_token = _get_spark_token()
|
| 83 |
-
if not spark_token:
|
| 84 |
-
log("β SPARK_TOKEN not configured!")
|
| 85 |
-
raise ValueError("Spark authentication token not configured")
|
| 86 |
-
|
| 87 |
-
# Prepare headers with authorization
|
| 88 |
-
headers = {
|
| 89 |
-
"Authorization": f"Bearer {spark_token}",
|
| 90 |
-
"Content-Type": "application/json"
|
| 91 |
-
}
|
| 92 |
-
|
| 93 |
-
# Spark'a gΓΆnderilecek payload'Δ± hazΔ±rla
|
| 94 |
-
payload = {
|
| 95 |
-
"project_name": s.project_name,
|
| 96 |
-
"system_prompt": prompt,
|
| 97 |
-
"user_input": user_msg,
|
| 98 |
-
"context": s.chat_history[-10:]
|
| 99 |
-
}
|
| 100 |
|
| 101 |
-
|
| 102 |
-
|
|
|
|
|
|
|
| 103 |
|
| 104 |
-
# DoΔru endpoint'e istek at
|
| 105 |
-
spark_url = SPARK_URL + "/generate"
|
| 106 |
-
log(f"π Spark URL: {spark_url}")
|
| 107 |
-
|
| 108 |
-
async with httpx.AsyncClient(timeout=60) as client:
|
| 109 |
-
response = await client.post(spark_url, json=payload, headers=headers)
|
| 110 |
-
response.raise_for_status()
|
| 111 |
-
data = response.json()
|
| 112 |
-
|
| 113 |
-
# Spark'tan gelen yanΔ±tΔ± parse et
|
| 114 |
-
raw = data.get("model_answer", "").strip()
|
| 115 |
-
if not raw:
|
| 116 |
-
# Fallback to other possible fields
|
| 117 |
-
raw = (data.get("assistant") or data.get("text", "")).strip()
|
| 118 |
-
|
| 119 |
-
log(f"πͺ Spark raw: {raw[:120]!r}")
|
| 120 |
-
return raw
|
| 121 |
-
|
| 122 |
-
except httpx.HTTPStatusError as e:
|
| 123 |
-
log(f"β Spark HTTP error: {e.response.status_code} - {e.response.text}")
|
| 124 |
-
raise
|
| 125 |
-
except httpx.TimeoutException:
|
| 126 |
-
log(f"β±οΈ Spark timeout for session {s.session_id[:8]}")
|
| 127 |
-
raise
|
| 128 |
except Exception as e:
|
| 129 |
-
log(f"β
|
| 130 |
raise
|
| 131 |
|
| 132 |
# βββββββββββββββββββββββββ FASTAPI βββββββββββββββββββββββββ #
|
|
|
|
| 17 |
from config_provider import ConfigProvider
|
| 18 |
from validation_engine import validate
|
| 19 |
from session import session_store, Session
|
| 20 |
+
from llm_interface import LLMInterface, SparkLLM, GPT4oLLM
|
| 21 |
+
|
| 22 |
+
# Global LLM instance
|
| 23 |
+
llm_provider: Optional[LLMInterface] = None
|
| 24 |
|
| 25 |
# βββββββββββββββββββββββββ HELPERS βββββββββββββββββββββββββ #
|
| 26 |
def _trim_response(raw: str) -> str:
|
|
|
|
| 56 |
SPARK_URL = str(cfg.global_config.spark_endpoint).rstrip("/")
|
| 57 |
ALLOWED_INTENTS = {"flight-booking", "flight-info", "booking-cancel"}
|
| 58 |
|
| 59 |
+
# βββββββββββββββββββββββββ SPARK βββββββββββββββββββββββββ #
|
| 60 |
+
def initialize_llm():
|
| 61 |
+
"""Initialize LLM provider based on work_mode"""
|
| 62 |
+
global llm_provider
|
| 63 |
+
|
| 64 |
+
work_mode = cfg.global_config.work_mode
|
| 65 |
+
|
| 66 |
+
if cfg.global_config.is_gpt_mode():
|
| 67 |
+
# GPT mode
|
| 68 |
+
api_key = cfg.global_config.get_plain_token()
|
| 69 |
+
if not api_key:
|
| 70 |
+
raise ValueError("OpenAI API key not configured")
|
| 71 |
+
|
| 72 |
+
model = cfg.global_config.get_gpt_model()
|
| 73 |
+
llm_provider = GPT4oLLM(api_key, model)
|
| 74 |
+
log(f"β
Initialized {model} provider")
|
| 75 |
+
else:
|
| 76 |
+
# Spark mode
|
| 77 |
+
spark_token = _get_spark_token()
|
| 78 |
+
if not spark_token:
|
| 79 |
+
raise ValueError("Spark token not configured")
|
| 80 |
+
|
| 81 |
+
spark_endpoint = str(cfg.global_config.spark_endpoint)
|
| 82 |
+
llm_provider = SparkLLM(spark_endpoint, spark_token)
|
| 83 |
+
log("β
Initialized Spark provider")
|
| 84 |
+
|
| 85 |
# βββββββββββββββββββββββββ SPARK βββββββββββββββββββββββββ #
|
| 86 |
def _get_spark_token() -> Optional[str]:
|
| 87 |
"""Get Spark token based on work_mode"""
|
|
|
|
| 98 |
return os.getenv("SPARK_TOKEN")
|
| 99 |
|
| 100 |
async def spark_generate(s: Session, prompt: str, user_msg: str) -> str:
|
| 101 |
+
"""Call LLM provider with proper error handling"""
|
| 102 |
try:
|
| 103 |
+
if not llm_provider:
|
| 104 |
+
initialize_llm()
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 105 |
|
| 106 |
+
# Use the abstract interface
|
| 107 |
+
raw = await llm_provider.generate(prompt, user_msg, s.chat_history)
|
| 108 |
+
log(f"πͺ LLM raw response: {raw[:120]!r}")
|
| 109 |
+
return raw
|
| 110 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 111 |
except Exception as e:
|
| 112 |
+
log(f"β LLM error: {e}")
|
| 113 |
raise
|
| 114 |
|
| 115 |
# βββββββββββββββββββββββββ FASTAPI βββββββββββββββββββββββββ #
|