Spaces:
Sleeping
Sleeping
Commit
·
051d37f
1
Parent(s):
a949b5d
Upload 3 files
Browse files- app.py +75 -62
- presets.py +16 -1
- utils.py +12 -1
app.py
CHANGED
|
@@ -43,7 +43,7 @@ else:
|
|
| 43 |
|
| 44 |
gr.Chatbot.postprocess = postprocess
|
| 45 |
|
| 46 |
-
with gr.Blocks(css=customCSS) as demo:
|
| 47 |
history = gr.State([])
|
| 48 |
token_count = gr.State([])
|
| 49 |
promptTemplates = gr.State(load_template(get_template_names(plain=True)[0], mode=2))
|
|
@@ -51,71 +51,84 @@ with gr.Blocks(css=customCSS) as demo:
|
|
| 51 |
FALSECONSTANT = gr.State(False)
|
| 52 |
topic = gr.State("未命名对话历史记录")
|
| 53 |
|
|
|
|
|
|
|
|
|
|
| 54 |
gr.HTML(title)
|
| 55 |
-
with gr.Row():
|
| 56 |
-
with gr.Column():
|
| 57 |
-
keyTxt = gr.Textbox(show_label=True, placeholder=f"在这里输入你的OpenAI API-key...",value=my_api_key, type="password", visible=not HIDE_MY_KEY, label="API-Key")
|
| 58 |
-
with gr.Column():
|
| 59 |
-
with gr.Row():
|
| 60 |
-
model_select_dropdown = gr.Dropdown(label="选择模型", choices=MODELS, multiselect=False, value=MODELS[0])
|
| 61 |
-
use_streaming_checkbox = gr.Checkbox(label="实时传输回答", value=True, visible=enable_streaming_option)
|
| 62 |
-
chatbot = gr.Chatbot() # .style(color_map=("#1D51EE", "#585A5B"))
|
| 63 |
-
with gr.Row():
|
| 64 |
-
with gr.Column(scale=12):
|
| 65 |
-
user_input = gr.Textbox(show_label=False, placeholder="在这里输入").style(
|
| 66 |
-
container=False)
|
| 67 |
-
with gr.Column(min_width=50, scale=1):
|
| 68 |
-
submitBtn = gr.Button("🚀", variant="primary")
|
| 69 |
-
with gr.Row():
|
| 70 |
-
emptyBtn = gr.Button("🧹 新的对话")
|
| 71 |
-
retryBtn = gr.Button("🔄 重新生成")
|
| 72 |
-
delLastBtn = gr.Button("🗑️ 删除最近一条对话")
|
| 73 |
-
reduceTokenBtn = gr.Button("♻️ 总结对话")
|
| 74 |
-
status_display = gr.Markdown("status: ready")
|
| 75 |
-
|
| 76 |
-
systemPromptTxt = gr.Textbox(show_label=True, placeholder=f"在这里输入System Prompt...", label="System prompt", value=initial_prompt).style(container=True)
|
| 77 |
-
|
| 78 |
-
with gr.Accordion(label="加载Prompt模板", open=False):
|
| 79 |
-
with gr.Column():
|
| 80 |
-
with gr.Row():
|
| 81 |
-
with gr.Column(scale=6):
|
| 82 |
-
templateFileSelectDropdown = gr.Dropdown(label="选择Prompt模板集合文件", choices=get_template_names(plain=True), multiselect=False, value=get_template_names(plain=True)[0])
|
| 83 |
-
with gr.Column(scale=1):
|
| 84 |
-
templateRefreshBtn = gr.Button("🔄 刷新")
|
| 85 |
-
templaeFileReadBtn = gr.Button("📂 读入模板")
|
| 86 |
-
with gr.Row():
|
| 87 |
-
with gr.Column(scale=6):
|
| 88 |
-
templateSelectDropdown = gr.Dropdown(label="从Prompt模板中加载", choices=load_template(get_template_names(plain=True)[0], mode=1), multiselect=False, value=load_template(get_template_names(plain=True)[0], mode=1)[0])
|
| 89 |
-
with gr.Column(scale=1):
|
| 90 |
-
templateApplyBtn = gr.Button("⬇️ 应用")
|
| 91 |
-
with gr.Accordion(label="保存/加载对话历史记录", open=False):
|
| 92 |
-
with gr.Column():
|
| 93 |
-
with gr.Row():
|
| 94 |
-
with gr.Column(scale=6):
|
| 95 |
-
saveFileName = gr.Textbox(
|
| 96 |
-
show_label=True, placeholder=f"在这里输入保存的文件名...", label="设置保存文件名", value="对话历史记录").style(container=True)
|
| 97 |
-
with gr.Column(scale=1):
|
| 98 |
-
saveHistoryBtn = gr.Button("💾 保存对话")
|
| 99 |
-
with gr.Row():
|
| 100 |
-
with gr.Column(scale=6):
|
| 101 |
-
historyFileSelectDropdown = gr.Dropdown(label="从列表中加载对话", choices=get_history_names(plain=True), multiselect=False, value=get_history_names(plain=True)[0])
|
| 102 |
-
with gr.Column(scale=1):
|
| 103 |
-
historyRefreshBtn = gr.Button("🔄 刷新")
|
| 104 |
-
historyReadBtn = gr.Button("📂 读入对话")
|
| 105 |
-
#inputs, top_p, temperature, top_k, repetition_penalty
|
| 106 |
-
with gr.Accordion("参数", open=False):
|
| 107 |
-
top_p = gr.Slider(minimum=-0, maximum=1.0, value=1.0, step=0.05,
|
| 108 |
-
interactive=True, label="Top-p (nucleus sampling)",)
|
| 109 |
-
temperature = gr.Slider(minimum=-0, maximum=5.0, value=1.0,
|
| 110 |
-
step=0.1, interactive=True, label="Temperature",)
|
| 111 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 112 |
gr.Markdown(description)
|
| 113 |
|
| 114 |
|
| 115 |
-
user_input.submit(predict, [keyTxt, systemPromptTxt, history, user_input, chatbot, token_count, top_p, temperature, use_streaming_checkbox, model_select_dropdown], [chatbot, history, status_display, token_count], show_progress=True)
|
| 116 |
user_input.submit(reset_textbox, [], [user_input])
|
| 117 |
|
| 118 |
-
submitBtn.click(predict, [keyTxt, systemPromptTxt, history, user_input, chatbot, token_count, top_p, temperature, use_streaming_checkbox, model_select_dropdown], [chatbot, history, status_display, token_count], show_progress=True)
|
| 119 |
submitBtn.click(reset_textbox, [], [user_input])
|
| 120 |
|
| 121 |
emptyBtn.click(reset_state, outputs=[chatbot, history, token_count, status_display], show_progress=True)
|
|
@@ -134,13 +147,13 @@ with gr.Blocks(css=customCSS) as demo:
|
|
| 134 |
|
| 135 |
historyRefreshBtn.click(get_history_names, None, [historyFileSelectDropdown])
|
| 136 |
|
| 137 |
-
|
| 138 |
|
| 139 |
templateRefreshBtn.click(get_template_names, None, [templateFileSelectDropdown])
|
| 140 |
|
| 141 |
-
|
| 142 |
|
| 143 |
-
|
| 144 |
|
| 145 |
logging.info(colorama.Back.GREEN + "\n川虎的温馨提示:访问 http://localhost:7860 查看界面" + colorama.Style.RESET_ALL)
|
| 146 |
# 默认开启本地服务器,默认可以直接从IP访问,默认不创建公开分享链接
|
|
|
|
| 43 |
|
| 44 |
gr.Chatbot.postprocess = postprocess
|
| 45 |
|
| 46 |
+
with gr.Blocks(css=customCSS,) as demo:
|
| 47 |
history = gr.State([])
|
| 48 |
token_count = gr.State([])
|
| 49 |
promptTemplates = gr.State(load_template(get_template_names(plain=True)[0], mode=2))
|
|
|
|
| 51 |
FALSECONSTANT = gr.State(False)
|
| 52 |
topic = gr.State("未命名对话历史记录")
|
| 53 |
|
| 54 |
+
# gr.HTML("""
|
| 55 |
+
# <div style="text-align: center; margin-top: 20px;">
|
| 56 |
+
# """)
|
| 57 |
gr.HTML(title)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 58 |
|
| 59 |
+
with gr.Row(scale=1).style(equal_height=True):
|
| 60 |
+
|
| 61 |
+
with gr.Column(scale=5):
|
| 62 |
+
with gr.Row(scale=1):
|
| 63 |
+
chatbot = gr.Chatbot().style(height=600) # .style(color_map=("#1D51EE", "#585A5B"))
|
| 64 |
+
with gr.Row(scale=1):
|
| 65 |
+
with gr.Column(scale=12):
|
| 66 |
+
user_input = gr.Textbox(show_label=False, placeholder="在这里输入").style(
|
| 67 |
+
container=False)
|
| 68 |
+
with gr.Column(min_width=50, scale=1):
|
| 69 |
+
submitBtn = gr.Button("🚀", variant="primary")
|
| 70 |
+
with gr.Row(scale=1):
|
| 71 |
+
emptyBtn = gr.Button("🧹 新的对话",)
|
| 72 |
+
retryBtn = gr.Button("🔄 重新生成")
|
| 73 |
+
delLastBtn = gr.Button("🗑️ 删除最近一条对话")
|
| 74 |
+
reduceTokenBtn = gr.Button("♻️ 总结��话")
|
| 75 |
+
|
| 76 |
+
|
| 77 |
+
|
| 78 |
+
with gr.Column():
|
| 79 |
+
with gr.Row(min_width=50,scale=1):
|
| 80 |
+
status_display = gr.Markdown("status: ready")
|
| 81 |
+
with gr.Tab(label="ChatGPT"):
|
| 82 |
+
keyTxt = gr.Textbox(show_label=True, placeholder=f"OpenAI API-key...",value=my_api_key, type="password", visible=not HIDE_MY_KEY, label="API-Key")
|
| 83 |
+
model_select_dropdown = gr.Dropdown(label="选择模型", choices=MODELS, multiselect=False, value=MODELS[0])
|
| 84 |
+
with gr.Accordion("参数", open=False):
|
| 85 |
+
top_p = gr.Slider(minimum=-0, maximum=1.0, value=1.0, step=0.05,
|
| 86 |
+
interactive=True, label="Top-p (nucleus sampling)",)
|
| 87 |
+
temperature = gr.Slider(minimum=-0, maximum=5.0, value=1.0,
|
| 88 |
+
step=0.1, interactive=True, label="Temperature",)
|
| 89 |
+
use_streaming_checkbox = gr.Checkbox(label="实时传输回答", value=True, visible=enable_streaming_option)
|
| 90 |
+
use_websearch_checkbox = gr.Checkbox(label="使用在线搜索", value=False)
|
| 91 |
+
|
| 92 |
+
with gr.Tab(label="Prompt"):
|
| 93 |
+
systemPromptTxt = gr.Textbox(show_label=True, placeholder=f"在这里输入System Prompt...", label="System prompt", value=initial_prompt).style(container=True)
|
| 94 |
+
with gr.Accordion(label="加载Prompt模板", open=True):
|
| 95 |
+
with gr.Column():
|
| 96 |
+
with gr.Row():
|
| 97 |
+
with gr.Column(scale=6):
|
| 98 |
+
templateFileSelectDropdown = gr.Dropdown(label="选择Prompt模板集合文件", choices=get_template_names(plain=True), multiselect=False, value=get_template_names(plain=True)[0])
|
| 99 |
+
with gr.Column(scale=1):
|
| 100 |
+
templateRefreshBtn = gr.Button("🔄 刷新")
|
| 101 |
+
with gr.Row():
|
| 102 |
+
with gr.Column():
|
| 103 |
+
templateSelectDropdown = gr.Dropdown(label="从Prompt模板中加载", choices=load_template(get_template_names(plain=True)[0], mode=1), multiselect=False, value=load_template(get_template_names(plain=True)[0], mode=1)[0])
|
| 104 |
+
|
| 105 |
+
with gr.Tab(label="保存/加载"):
|
| 106 |
+
with gr.Accordion(label="保存/加载对话历史记录", open=True):
|
| 107 |
+
with gr.Column():
|
| 108 |
+
with gr.Row():
|
| 109 |
+
with gr.Column(scale=6):
|
| 110 |
+
saveFileName = gr.Textbox(
|
| 111 |
+
show_label=True, placeholder=f"在这里输入保存的文件名...", label="设置保存文件名", value="对话历史记录").style(container=True)
|
| 112 |
+
with gr.Column(scale=1):
|
| 113 |
+
saveHistoryBtn = gr.Button("💾 保存对话")
|
| 114 |
+
with gr.Row():
|
| 115 |
+
with gr.Column(scale=6):
|
| 116 |
+
historyFileSelectDropdown = gr.Dropdown(label="从列表中加载对话", choices=get_history_names(plain=True), multiselect=False, value=get_history_names(plain=True)[0])
|
| 117 |
+
with gr.Column(scale=1):
|
| 118 |
+
historyRefreshBtn = gr.Button("🔄 刷新")
|
| 119 |
+
|
| 120 |
+
|
| 121 |
+
|
| 122 |
+
gr.HTML("""
|
| 123 |
+
<div style="text-align: center; margin-top: 20px; margin-bottom: 20px;">
|
| 124 |
+
""")
|
| 125 |
gr.Markdown(description)
|
| 126 |
|
| 127 |
|
| 128 |
+
user_input.submit(predict, [keyTxt, systemPromptTxt, history, user_input, chatbot, token_count, top_p, temperature, use_streaming_checkbox, model_select_dropdown, use_websearch_checkbox], [chatbot, history, status_display, token_count], show_progress=True)
|
| 129 |
user_input.submit(reset_textbox, [], [user_input])
|
| 130 |
|
| 131 |
+
submitBtn.click(predict, [keyTxt, systemPromptTxt, history, user_input, chatbot, token_count, top_p, temperature, use_streaming_checkbox, model_select_dropdown, use_websearch_checkbox], [chatbot, history, status_display, token_count], show_progress=True)
|
| 132 |
submitBtn.click(reset_textbox, [], [user_input])
|
| 133 |
|
| 134 |
emptyBtn.click(reset_state, outputs=[chatbot, history, token_count, status_display], show_progress=True)
|
|
|
|
| 147 |
|
| 148 |
historyRefreshBtn.click(get_history_names, None, [historyFileSelectDropdown])
|
| 149 |
|
| 150 |
+
historyFileSelectDropdown.change(load_chat_history, [historyFileSelectDropdown, systemPromptTxt, history, chatbot], [saveFileName, systemPromptTxt, history, chatbot], show_progress=True)
|
| 151 |
|
| 152 |
templateRefreshBtn.click(get_template_names, None, [templateFileSelectDropdown])
|
| 153 |
|
| 154 |
+
templateFileSelectDropdown.change(load_template, [templateFileSelectDropdown], [promptTemplates, templateSelectDropdown], show_progress=True)
|
| 155 |
|
| 156 |
+
templateSelectDropdown.change(get_template_content, [promptTemplates, templateSelectDropdown, systemPromptTxt], [systemPromptTxt], show_progress=True)
|
| 157 |
|
| 158 |
logging.info(colorama.Back.GREEN + "\n川虎的温馨提示:访问 http://localhost:7860 查看界面" + colorama.Style.RESET_ALL)
|
| 159 |
# 默认开启本地服务器,默认可以直接从IP访问,默认不创建公开分享链接
|
presets.py
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
# -*- coding:utf-8 -*-
|
| 2 |
-
title = """<h1 align="
|
| 3 |
description = """<div align=center>
|
| 4 |
|
| 5 |
由Bilibili [土川虎虎虎](https://space.bilibili.com/29125536) 和 [明昭MZhao](https://space.bilibili.com/24807452)开发
|
|
@@ -29,10 +29,25 @@ pre code {
|
|
| 29 |
color: #FFF;
|
| 30 |
box-shadow: inset 0px 8px 16px hsla(0, 0%, 0%, .2)
|
| 31 |
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 32 |
"""
|
| 33 |
|
| 34 |
summarize_prompt = "你是谁?我们刚才聊了什么?" # 总结对话时的 prompt
|
| 35 |
MODELS = ["gpt-3.5-turbo", "gpt-3.5-turbo-0301", "gpt-4","gpt-4-0314", "gpt-4-32k", "gpt-4-32k-0314"] # 可选的模型
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 36 |
|
| 37 |
# 错误信息
|
| 38 |
standard_error_msg = "☹️发生了错误:" # 错误信息的标准前缀
|
|
|
|
| 1 |
# -*- coding:utf-8 -*-
|
| 2 |
+
title = """<h1 align="left">川虎ChatGPT 🚀</h1>"""
|
| 3 |
description = """<div align=center>
|
| 4 |
|
| 5 |
由Bilibili [土川虎虎虎](https://space.bilibili.com/29125536) 和 [明昭MZhao](https://space.bilibili.com/24807452)开发
|
|
|
|
| 29 |
color: #FFF;
|
| 30 |
box-shadow: inset 0px 8px 16px hsla(0, 0%, 0%, .2)
|
| 31 |
}
|
| 32 |
+
|
| 33 |
+
*{
|
| 34 |
+
border-radius: 3px !important;
|
| 35 |
+
transition: all 0.6s;
|
| 36 |
+
}
|
| 37 |
+
|
| 38 |
+
|
| 39 |
"""
|
| 40 |
|
| 41 |
summarize_prompt = "你是谁?我们刚才聊了什么?" # 总结对话时的 prompt
|
| 42 |
MODELS = ["gpt-3.5-turbo", "gpt-3.5-turbo-0301", "gpt-4","gpt-4-0314", "gpt-4-32k", "gpt-4-32k-0314"] # 可选的模型
|
| 43 |
+
websearch_prompt = """Web search results:
|
| 44 |
+
|
| 45 |
+
{web_results}
|
| 46 |
+
Current date: {current_date}
|
| 47 |
+
|
| 48 |
+
Instructions: Using the provided web search results, write a comprehensive reply to the given query. Make sure to cite results using [[number](URL)] notation after the reference. If the provided search results refer to multiple subjects with the same name, write separate answers for each subject.
|
| 49 |
+
Query: {query}
|
| 50 |
+
Reply in 中文"""
|
| 51 |
|
| 52 |
# 错误信息
|
| 53 |
standard_error_msg = "☹️发生了错误:" # 错误信息的标准前缀
|
utils.py
CHANGED
|
@@ -16,6 +16,8 @@ from presets import *
|
|
| 16 |
import tiktoken
|
| 17 |
from tqdm import tqdm
|
| 18 |
import colorama
|
|
|
|
|
|
|
| 19 |
|
| 20 |
# logging.basicConfig(level=logging.INFO, format="%(asctime)s [%(levelname)s] [%(filename)s:%(lineno)d] %(message)s")
|
| 21 |
|
|
@@ -224,8 +226,17 @@ def predict_all(openai_api_key, system_prompt, history, inputs, chatbot, all_tok
|
|
| 224 |
return chatbot, history, status_text, all_token_counts
|
| 225 |
|
| 226 |
|
| 227 |
-
def predict(openai_api_key, system_prompt, history, inputs, chatbot, all_token_counts, top_p, temperature, stream=False, selected_model = MODELS[0], should_check_token_count = True): # repetition_penalty, top_k
|
| 228 |
logging.info("输入为:" +colorama.Fore.BLUE + f"{inputs}" + colorama.Style.RESET_ALL)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 229 |
if len(openai_api_key) != 51:
|
| 230 |
status_text = standard_error_msg + no_apikey_msg
|
| 231 |
logging.info(status_text)
|
|
|
|
| 16 |
import tiktoken
|
| 17 |
from tqdm import tqdm
|
| 18 |
import colorama
|
| 19 |
+
from duckduckgo_search import ddg
|
| 20 |
+
import datetime
|
| 21 |
|
| 22 |
# logging.basicConfig(level=logging.INFO, format="%(asctime)s [%(levelname)s] [%(filename)s:%(lineno)d] %(message)s")
|
| 23 |
|
|
|
|
| 226 |
return chatbot, history, status_text, all_token_counts
|
| 227 |
|
| 228 |
|
| 229 |
+
def predict(openai_api_key, system_prompt, history, inputs, chatbot, all_token_counts, top_p, temperature, stream=False, selected_model = MODELS[0], use_websearch_checkbox = False, should_check_token_count = True): # repetition_penalty, top_k
|
| 230 |
logging.info("输入为:" +colorama.Fore.BLUE + f"{inputs}" + colorama.Style.RESET_ALL)
|
| 231 |
+
if use_websearch_checkbox:
|
| 232 |
+
results = ddg(inputs, max_results=3)
|
| 233 |
+
web_results = []
|
| 234 |
+
for idx, result in enumerate(results):
|
| 235 |
+
logging.info(f"搜索结果{idx + 1}:{result}")
|
| 236 |
+
web_results.append(f'[{idx+1}]"{result["body"]}"\nURL: {result["href"]}')
|
| 237 |
+
web_results = "\n\n".join(web_results)
|
| 238 |
+
today = datetime.datetime.today().strftime("%Y-%m-%d")
|
| 239 |
+
inputs = websearch_prompt.replace("{current_date}", today).replace("{query}", inputs).replace("{web_results}", web_results)
|
| 240 |
if len(openai_api_key) != 51:
|
| 241 |
status_text = standard_error_msg + no_apikey_msg
|
| 242 |
logging.info(status_text)
|