Aiswarya Sankar commited on
Commit
d923dd2
·
1 Parent(s): 2f0b728

Clean code

Browse files
Files changed (1) hide show
  1. app.py +11 -18
app.py CHANGED
@@ -1,14 +1,12 @@
1
  import gradio as gr
2
  import os
3
  from queue import SimpleQueue
4
-
5
  from langchain.callbacks.manager import CallbackManager
6
  from langchain.chat_models import ChatOpenAI
7
  from pydantic import BaseModel
8
  import requests
9
  import typing
10
  from typing import TypeVar, Generic
11
- import math
12
  import tqdm
13
  from langchain.chains import ConversationalRetrievalChain
14
  import os
@@ -16,8 +14,6 @@ from langchain.embeddings.openai import OpenAIEmbeddings
16
  from langchain.vectorstores import DeepLake
17
  import random
18
  import time
19
- import together
20
-
21
  import os
22
  from langchain.document_loaders import TextLoader
23
  from langchain.text_splitter import CharacterTextSplitter
@@ -75,13 +71,6 @@ class GithubResponse(BaseModel):
75
  repo: str
76
 
77
 
78
- # global repoName
79
- # global ticket_titles
80
- # global tickets
81
- # global ticket_choices
82
- tickets = []
83
-
84
-
85
  repo_name = gr.State()
86
  git_tickets = gr.State()
87
  git_titles = gr.State()
@@ -128,13 +117,17 @@ def index_repo(textbox: str, dropdown: str) -> Response:
128
  dataset_path = f"hub://{activeloop_username}/" + pathName
129
 
130
  try:
131
- db = DeepLake(dataset_path=dataset_path,
132
- embedding_function=embeddings,
133
- token=os.environ['ACTIVELOOP_TOKEN'],
134
- read_only=True,
135
- num_workers=12,
136
- runtime = {"tensor_db": True}
137
- )
 
 
 
 
138
 
139
  if db is None or len(db.vectorstore.dataset) == 0:
140
  print("Dataset doesn't exist, fetching data")
 
1
  import gradio as gr
2
  import os
3
  from queue import SimpleQueue
 
4
  from langchain.callbacks.manager import CallbackManager
5
  from langchain.chat_models import ChatOpenAI
6
  from pydantic import BaseModel
7
  import requests
8
  import typing
9
  from typing import TypeVar, Generic
 
10
  import tqdm
11
  from langchain.chains import ConversationalRetrievalChain
12
  import os
 
14
  from langchain.vectorstores import DeepLake
15
  import random
16
  import time
 
 
17
  import os
18
  from langchain.document_loaders import TextLoader
19
  from langchain.text_splitter import CharacterTextSplitter
 
71
  repo: str
72
 
73
 
 
 
 
 
 
 
 
74
  repo_name = gr.State()
75
  git_tickets = gr.State()
76
  git_titles = gr.State()
 
117
  dataset_path = f"hub://{activeloop_username}/" + pathName
118
 
119
  try:
120
+ try:
121
+ db = DeepLake(dataset_path=dataset_path,
122
+ embedding_function=embeddings,
123
+ token=os.environ['ACTIVELOOP_TOKEN'],
124
+ read_only=True,
125
+ num_workers=12,
126
+ runtime = {"tensor_db": True}
127
+ )
128
+ except Exception as e:
129
+ if "scheduled for deletion" in str(e):
130
+ dataset_path = f"hub://{activeloop_username}/" + pathName + str(random.randint(1,100))
131
 
132
  if db is None or len(db.vectorstore.dataset) == 0:
133
  print("Dataset doesn't exist, fetching data")