Spaces:
Paused
Paused
Update chat_handler.py
Browse files- chat_handler.py +13 -8
chat_handler.py
CHANGED
|
@@ -242,20 +242,25 @@ async def start_session(req: StartRequest):
|
|
| 242 |
llm_provider = LLMFactory.create_provider()
|
| 243 |
log_info(f"🤖 LLM Provider created: {type(llm_provider).__name__}")
|
| 244 |
|
| 245 |
-
# Create session with version config
|
| 246 |
-
|
| 247 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 248 |
|
| 249 |
# Welcome prompt'u işle
|
| 250 |
greeting = "Hoş geldiniz! Size nasıl yardımcı olabilirim?"
|
| 251 |
if version.welcome_prompt:
|
| 252 |
-
log_info(f"🎉 Processing welcome prompt for session {session_id[:8]}...")
|
| 253 |
try:
|
| 254 |
# Welcome prompt'u LLM'e gönder
|
| 255 |
welcome_result = await llm_provider.generate(
|
| 256 |
-
|
| 257 |
-
|
| 258 |
-
|
| 259 |
)
|
| 260 |
if welcome_result and welcome_result.strip():
|
| 261 |
greeting = welcome_result.strip()
|
|
@@ -280,7 +285,7 @@ async def chat(req: ChatRequest, x_session_id: str = Header(...)):
|
|
| 280 |
"""Process chat message"""
|
| 281 |
try:
|
| 282 |
# Get session
|
| 283 |
-
session = session_store.
|
| 284 |
if not session:
|
| 285 |
# Better error message
|
| 286 |
raise HTTPException(
|
|
|
|
| 242 |
llm_provider = LLMFactory.create_provider()
|
| 243 |
log_info(f"🤖 LLM Provider created: {type(llm_provider).__name__}")
|
| 244 |
|
| 245 |
+
# Create session with version config - PARAMETRE DÜZELTMESİ
|
| 246 |
+
session = session_store.create_session(
|
| 247 |
+
project_name=req.project_name,
|
| 248 |
+
version_no=version.no # version objesi yerine version.no
|
| 249 |
+
)
|
| 250 |
+
|
| 251 |
+
# Version config'i session'a ekle
|
| 252 |
+
session.set_version_config(version)
|
| 253 |
|
| 254 |
# Welcome prompt'u işle
|
| 255 |
greeting = "Hoş geldiniz! Size nasıl yardımcı olabilirim?"
|
| 256 |
if version.welcome_prompt:
|
| 257 |
+
log_info(f"🎉 Processing welcome prompt for session {session.session_id[:8]}...")
|
| 258 |
try:
|
| 259 |
# Welcome prompt'u LLM'e gönder
|
| 260 |
welcome_result = await llm_provider.generate(
|
| 261 |
+
user_input="",
|
| 262 |
+
system_prompt=version.welcome_prompt,
|
| 263 |
+
context=[]
|
| 264 |
)
|
| 265 |
if welcome_result and welcome_result.strip():
|
| 266 |
greeting = welcome_result.strip()
|
|
|
|
| 285 |
"""Process chat message"""
|
| 286 |
try:
|
| 287 |
# Get session
|
| 288 |
+
session = session_store.get_session(x_session_id)
|
| 289 |
if not session:
|
| 290 |
# Better error message
|
| 291 |
raise HTTPException(
|