Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -1,6 +1,7 @@
|
|
1 |
import os
|
2 |
from langchain_core.prompts import PromptTemplate
|
3 |
from langchain.chains.question_answering import load_qa_chain
|
|
|
4 |
from langchain_community.document_loaders import PyPDFLoader
|
5 |
import google.generativeai as genai
|
6 |
import gradio as gr
|
@@ -42,6 +43,8 @@ def initialize(pdf_file, question):
|
|
42 |
# Load the GeminiPro model
|
43 |
model = genai.GenerativeModel('gemini-pro')
|
44 |
|
|
|
|
|
45 |
# Prepare the input data
|
46 |
input_data = {
|
47 |
"context": processed_context,
|
|
|
1 |
import os
|
2 |
from langchain_core.prompts import PromptTemplate
|
3 |
from langchain.chains.question_answering import load_qa_chain
|
4 |
+
from langchain_community.output_parsers.rail_parser import GuardrailsOutputParser
|
5 |
from langchain_community.document_loaders import PyPDFLoader
|
6 |
import google.generativeai as genai
|
7 |
import gradio as gr
|
|
|
43 |
# Load the GeminiPro model
|
44 |
model = genai.GenerativeModel('gemini-pro')
|
45 |
|
46 |
+
# Ensure your LLM object (model) inherits from Runnable and can generate text
|
47 |
+
|
48 |
# Prepare the input data
|
49 |
input_data = {
|
50 |
"context": processed_context,
|