Spaces:
Running
on
T4
Running
on
T4
Update app.py
Browse files
app.py
CHANGED
|
@@ -11,7 +11,7 @@ from auditqa.sample_questions import QUESTIONS
|
|
| 11 |
from auditqa.reports import files, report_list, new_files, new_report_list
|
| 12 |
from auditqa.process_chunks import load_chunks, getconfig, get_local_qdrant, load_new_chunks
|
| 13 |
from auditqa.retriever import get_context
|
| 14 |
-
from auditqa.reader import nvidia_client, dedicated_endpoint
|
| 15 |
from auditqa.utils import make_html_source, parse_output_llm_with_sources, save_logs, get_message_template, get_client_location, get_client_ip, get_platform_info
|
| 16 |
from dotenv import load_dotenv
|
| 17 |
load_dotenv()
|
|
@@ -242,7 +242,7 @@ async def chat(query,history,sources,reports,subtype,year, client_ip=None, sessi
|
|
| 242 |
# yield update
|
| 243 |
|
| 244 |
else:
|
| 245 |
-
chat_model =
|
| 246 |
async def process_stream():
|
| 247 |
nonlocal answer_yet
|
| 248 |
try:
|
|
|
|
| 11 |
from auditqa.reports import files, report_list, new_files, new_report_list
|
| 12 |
from auditqa.process_chunks import load_chunks, getconfig, get_local_qdrant, load_new_chunks
|
| 13 |
from auditqa.retriever import get_context
|
| 14 |
+
from auditqa.reader import nvidia_client, dedicated_endpoint, serverless_api
|
| 15 |
from auditqa.utils import make_html_source, parse_output_llm_with_sources, save_logs, get_message_template, get_client_location, get_client_ip, get_platform_info
|
| 16 |
from dotenv import load_dotenv
|
| 17 |
load_dotenv()
|
|
|
|
| 242 |
# yield update
|
| 243 |
|
| 244 |
else:
|
| 245 |
+
chat_model = serverless_api() # TESTING: ADAPTED FOR HF INFERENCE API (needs to be reverted for production version)
|
| 246 |
async def process_stream():
|
| 247 |
nonlocal answer_yet
|
| 248 |
try:
|