Update app.py
Browse files
app.py
CHANGED
@@ -2,6 +2,8 @@ from langchain.prompts.prompt import PromptTemplate
|
|
2 |
from langchain.llms import OpenAI
|
3 |
from langchain.chains import ChatVectorDBChain
|
4 |
from langchain.embeddings import HuggingFaceEmbeddings, HuggingFaceInstructEmbeddings
|
|
|
|
|
5 |
from langchain.vectorstores import FAISS
|
6 |
import os
|
7 |
from typing import Optional, Tuple
|
@@ -62,7 +64,7 @@ QA_PROMPT = PromptTemplate(template=template, input_variables=["question", "cont
|
|
62 |
|
63 |
|
64 |
def get_chain(vectorstore):
|
65 |
-
llm = OpenAI(temperature=0)
|
66 |
qa_chain = ChatVectorDBChain.from_llm(
|
67 |
llm,
|
68 |
vectorstore,
|
|
|
2 |
from langchain.llms import OpenAI
|
3 |
from langchain.chains import ChatVectorDBChain
|
4 |
from langchain.embeddings import HuggingFaceEmbeddings, HuggingFaceInstructEmbeddings
|
5 |
+
from langchain.callbacks.base import CallbackManager
|
6 |
+
from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler
|
7 |
from langchain.vectorstores import FAISS
|
8 |
import os
|
9 |
from typing import Optional, Tuple
|
|
|
64 |
|
65 |
|
66 |
def get_chain(vectorstore):
|
67 |
+
llm = OpenAI(streaming=True, callback_manager=CallbackManager([StreamingStdOutCallbackHandler()]), verbose=True, temperature=0)
|
68 |
qa_chain = ChatVectorDBChain.from_llm(
|
69 |
llm,
|
70 |
vectorstore,
|