Spaces:
Running
Running
Commit
·
a79769a
1
Parent(s):
71782c8
小迭代,user_question
Browse files
app.py
CHANGED
@@ -336,26 +336,25 @@ def app():
|
|
336 |
st.markdown(message["content"][0])
|
337 |
|
338 |
#这里的session.state就是保存了这个对话会话的一些基本信息和设置
|
339 |
-
|
340 |
-
|
341 |
-
|
342 |
-
|
343 |
-
|
344 |
-
|
345 |
-
|
346 |
-
|
347 |
-
|
348 |
-
|
349 |
-
|
350 |
-
|
351 |
-
|
352 |
-
|
353 |
-
|
354 |
-
)
|
355 |
-
|
356 |
-
|
357 |
-
|
358 |
-
st.session_state.messages.append({"role": "assistant", "content": [full_response,1]})
|
359 |
|
360 |
|
361 |
|
|
|
336 |
st.markdown(message["content"][0])
|
337 |
|
338 |
#这里的session.state就是保存了这个对话会话的一些基本信息和设置
|
339 |
+
retrieved_chunks_for_user = searchVDB(user_question, st.session_state.embeddings_df, st.session_state.faiss_index)
|
340 |
+
#retrieved_chunks_for_user = []
|
341 |
+
prompt = decorate_user_question(user_question, retrieved_chunks_for_user)
|
342 |
+
st.session_state.messages.append({"role": "user", "content": [user_question, prompt]})
|
343 |
+
with st.chat_message("user"):
|
344 |
+
st.markdown(user_question)
|
345 |
+
# Display assistant response in chat message container
|
346 |
+
with st.chat_message("assistant"):
|
347 |
+
message_placeholder = st.empty()
|
348 |
+
full_response = ""
|
349 |
+
for response in openai.ChatCompletion.create(
|
350 |
+
model=st.session_state["openai_model"],
|
351 |
+
messages=[{"role": m["role"], "content": m["content"][1]} for m in st.session_state.messages],
|
352 |
+
stream=True,
|
353 |
+
):
|
354 |
+
full_response += response.choices[0].delta.get("content", "")
|
355 |
+
message_placeholder.markdown(full_response + "▌")
|
356 |
+
message_placeholder.markdown(full_response)
|
357 |
+
st.session_state.messages.append({"role": "assistant", "content": [full_response,1]})
|
|
|
358 |
|
359 |
|
360 |
|