Update app.py
Browse files
app.py
CHANGED
@@ -46,7 +46,7 @@ def handle_image_submission(_chatbot, task_history, file) -> tuple:
|
|
46 |
history_item = ((file_path,), None)
|
47 |
_chatbot.append(history_item)
|
48 |
task_history.append(history_item)
|
49 |
-
return predict(_chatbot, task_history)
|
50 |
|
51 |
|
52 |
def _load_model_tokenizer(args) -> tuple:
|
@@ -126,7 +126,7 @@ def add_file(history, task_history, file):
|
|
126 |
return history, task_history
|
127 |
|
128 |
|
129 |
-
def predict(_chatbot, task_history) -> list:
|
130 |
print("predict called")
|
131 |
if not _chatbot:
|
132 |
return _chatbot
|
@@ -180,7 +180,7 @@ def regenerate(_chatbot, task_history) -> list:
|
|
180 |
_chatbot[-1] = (_chatbot[-1][0], None)
|
181 |
else:
|
182 |
_chatbot.append((chatbot_item[0], None))
|
183 |
-
return predict(_chatbot, task_history)
|
184 |
|
185 |
def add_text(history, task_history, text) -> tuple:
|
186 |
task_text = text
|
|
|
46 |
history_item = ((file_path,), None)
|
47 |
_chatbot.append(history_item)
|
48 |
task_history.append(history_item)
|
49 |
+
return predict(_chatbot, task_history, tokenizer, model)
|
50 |
|
51 |
|
52 |
def _load_model_tokenizer(args) -> tuple:
|
|
|
126 |
return history, task_history
|
127 |
|
128 |
|
129 |
+
def predict(_chatbot, task_history, tokenizer, model) -> list:
|
130 |
print("predict called")
|
131 |
if not _chatbot:
|
132 |
return _chatbot
|
|
|
180 |
_chatbot[-1] = (_chatbot[-1][0], None)
|
181 |
else:
|
182 |
_chatbot.append((chatbot_item[0], None))
|
183 |
+
return predict(_chatbot, task_history, tokenizer, model)
|
184 |
|
185 |
def add_text(history, task_history, text) -> tuple:
|
186 |
task_text = text
|