rohan13 commited on
Commit
6dd7a3f
·
1 Parent(s): b8b8495

Added support for palm

Browse files
app.py CHANGED
@@ -9,7 +9,7 @@ p = pipeline("automatic-speech-recognition")
9
 
10
  """Use text to call chat method from main.py"""
11
 
12
- models = ["GPT-3.5", "Flan UL2", "Flan T5"]
13
 
14
  name = os.environ.get("name", "Rohan")
15
 
@@ -76,19 +76,21 @@ def upload_file(files, history, model):
76
  print("Ingesting files: " + str(file_paths))
77
  text = 'Uploaded a file'
78
  if ingest_files(file_paths, model):
79
- response = 'Files are ingested'
80
  else:
81
- response = 'Files are not ingested'
82
 
83
  history = history + [(text, response)]
84
  return history
85
 
86
  theme = gr.Theme.from_hub("snehilsanyal/scikit-learn")
87
 
88
- theme.block_background_fill = gr.themes.colors.neutral.c200
 
 
89
 
90
 
91
- with gr.Blocks(theme) as demo:
92
  # Add image of Roar Logo from local directory
93
  gr.HTML('<img src="file/assets/logo.png" style="width: 100px; height: 100px; margin: 0 auto;border:5px solid orange;border-radius: 50%; display: block">')
94
  # Title on top in middle of the page
@@ -98,22 +100,23 @@ with gr.Blocks(theme) as demo:
98
 
99
  with gr.Row():
100
  # Create radio button to select model
101
- radio = gr.Radio(models, label="Choose a model", value="GPT-3.5", type="value", visible=False)
102
  with gr.Row():
103
  with gr.Column(scale=0.6):
104
  txt = gr.Textbox(
105
- label="Rohan Bot",
106
  placeholder="Enter text and press enter, or upload a file", lines=1
107
  ).style(container=False)
108
 
109
  with gr.Column(scale=0.2):
110
- upload = gr.UploadButton(label="Upload a file", type="file", file_count='multiple', file_types=['docx', 'txt', 'pdf', 'html']).style(container=False)
111
 
112
  with gr.Column(scale=0.2):
113
- audio = gr.Audio(source="microphone", type="filepath").style(container=False)
114
 
115
  with gr.Row():
116
- gr.Examples(examples=['What are you an expert of?'], inputs=[txt], label="Examples")
 
117
 
118
  txt.submit(add_text, [chatbot, txt, radio], [chatbot, txt], postprocess=False).then(
119
  bot, chatbot, chatbot
@@ -133,4 +136,4 @@ with gr.Blocks(theme) as demo:
133
  if __name__ == "__main__":
134
  demo.queue()
135
  demo.queue(concurrency_count=5)
136
- demo.launch(debug=True)
 
9
 
10
  """Use text to call chat method from main.py"""
11
 
12
+ models = ["GPT-3.5", "Flan UL2", "Flan T5", "Palm"]
13
 
14
  name = os.environ.get("name", "Rohan")
15
 
 
76
  print("Ingesting files: " + str(file_paths))
77
  text = 'Uploaded a file'
78
  if ingest_files(file_paths, model):
79
+ response = 'Files are ingested. Roar now!'
80
  else:
81
+ response = 'Files are not ingested. Please try again.'
82
 
83
  history = history + [(text, response)]
84
  return history
85
 
86
  theme = gr.Theme.from_hub("snehilsanyal/scikit-learn")
87
 
88
+ theme.block_background_fill = gr.themes.colors.neutral.c100
89
+ theme.block_border_width = '2px'
90
+ theme.block_border_radius = '10px'
91
 
92
 
93
+ with gr.Blocks(theme=theme, title='Roar!') as demo:
94
  # Add image of Roar Logo from local directory
95
  gr.HTML('<img src="file/assets/logo.png" style="width: 100px; height: 100px; margin: 0 auto;border:5px solid orange;border-radius: 50%; display: block">')
96
  # Title on top in middle of the page
 
100
 
101
  with gr.Row():
102
  # Create radio button to select model
103
+ radio = gr.Radio(models, label="Choose a model", value="GPT-3.5", type="value")
104
  with gr.Row():
105
  with gr.Column(scale=0.6):
106
  txt = gr.Textbox(
107
+ label="Let's hear the roar!",
108
  placeholder="Enter text and press enter, or upload a file", lines=1
109
  ).style(container=False)
110
 
111
  with gr.Column(scale=0.2):
112
+ upload = gr.UploadButton(label="Roar on a file", type="file", file_count='multiple', file_types=['docx', 'txt', 'pdf', 'html']).style(container=False)
113
 
114
  with gr.Column(scale=0.2):
115
+ audio = gr.Audio(source="microphone", type="filepath", label="Let me hear your roar!").style(container=False)
116
 
117
  with gr.Row():
118
+ gr.Examples(examples=['Roar it! What are you an expert of?', ' Roar it! What are you currently doing?',
119
+ 'Roar it! What is your opinion on Large Language Models?'], inputs=[txt], label="Examples")
120
 
121
  txt.submit(add_text, [chatbot, txt, radio], [chatbot, txt], postprocess=False).then(
122
  bot, chatbot, chatbot
 
136
  if __name__ == "__main__":
137
  demo.queue()
138
  demo.queue(concurrency_count=5)
139
+ demo.launch(debug=True, favicon_path="file/assets/logo.png")
docs/Rohan Marwaha.docx ADDED
Binary file (32.1 kB). View file
 
models/hf_vs.index ADDED
Binary file (43.1 kB). View file
 
models/hf_vs.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:625e8ad0ba047ca6c1433b0f262471d94cbe5969a3ac5663df1e3bf60e47d32c
3
+ size 49561
models/openai_vs.pkl CHANGED
Binary files a/models/openai_vs.pkl and b/models/openai_vs.pkl differ
 
models/palm_vs.index ADDED
Binary file (43.1 kB). View file
 
models/palm_vs.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a4fcb8df6e6711a1850b701dbb82ec9395e9c0e9574a0b0cc6afb632bfd3390
3
+ size 49561
requirements.txt CHANGED
@@ -8,4 +8,5 @@ gtts
8
  torch
9
  tiktoken
10
  huggingface-hub
11
- gradio
 
 
8
  torch
9
  tiktoken
10
  huggingface-hub
11
+ gradio
12
+ google-generativeai
utils.py CHANGED
@@ -18,6 +18,7 @@ from langchain.prompts.chat import (
18
  from langchain.text_splitter import CharacterTextSplitter
19
  from langchain.vectorstores.faiss import FAISS
20
  from langchain.cache import InMemoryCache
 
21
  import traceback
22
 
23
 
@@ -25,7 +26,7 @@ langchain.llm_cache = InMemoryCache()
25
 
26
  global model_name
27
 
28
- models = ["GPT-3.5", "Flan UL2", "GPT-4", "Flan T5"]
29
 
30
  pickle_file = "_vs.pkl"
31
  updated_pickle_file = "_vs_updated.pkl"
@@ -89,6 +90,8 @@ def set_model(model):
89
  elif model == "Flan T5":
90
  print("Loading Flan T5")
91
  llm = HuggingFaceHub(repo_id="google/flan-t5-base", model_kwargs={"temperature": 0.1})
 
 
92
  else:
93
  print("Loading GPT-3.5 from else")
94
  llm = OpenAI(model_name="text-davinci-002", temperature=0.1)
@@ -146,6 +149,8 @@ def get_file_path(model, file):
146
  # If model is GPT3.5 or GPT4 return models_folder + openai + file else return models_folder + hf + file
147
  if model == "GPT-3.5" or model == "GPT-4":
148
  return models_folder + "openai" + file
 
 
149
  else:
150
  return models_folder + "hf" + file
151
 
 
18
  from langchain.text_splitter import CharacterTextSplitter
19
  from langchain.vectorstores.faiss import FAISS
20
  from langchain.cache import InMemoryCache
21
+ from langchain.chat_models import ChatGooglePalm
22
  import traceback
23
 
24
 
 
26
 
27
  global model_name
28
 
29
+ models = ["GPT-3.5", "Flan UL2", "GPT-4", "Flan T5", "Palm"]
30
 
31
  pickle_file = "_vs.pkl"
32
  updated_pickle_file = "_vs_updated.pkl"
 
90
  elif model == "Flan T5":
91
  print("Loading Flan T5")
92
  llm = HuggingFaceHub(repo_id="google/flan-t5-base", model_kwargs={"temperature": 0.1})
93
+ elif model == "Palm":
94
+ llm = ChatGooglePalm(temperature=0)
95
  else:
96
  print("Loading GPT-3.5 from else")
97
  llm = OpenAI(model_name="text-davinci-002", temperature=0.1)
 
149
  # If model is GPT3.5 or GPT4 return models_folder + openai + file else return models_folder + hf + file
150
  if model == "GPT-3.5" or model == "GPT-4":
151
  return models_folder + "openai" + file
152
+ elif model == "Palm":
153
+ return models_folder + "palm" + file
154
  else:
155
  return models_folder + "hf" + file
156