better traceback
Browse files- check_proxy.py +0 -4
- crazy_functions/读文章写摘要.py +1 -18
- functional_crazy.py +1 -18
- main.py +3 -48
- toolbox.py +68 -0
check_proxy.py
CHANGED
|
@@ -1,8 +1,4 @@
|
|
| 1 |
|
| 2 |
-
"""
|
| 3 |
-
我:用python的requests库查询本机ip地址所在地
|
| 4 |
-
ChatGPT:
|
| 5 |
-
"""
|
| 6 |
def check_proxy(proxies):
|
| 7 |
import requests
|
| 8 |
proxies_https = proxies['https'] if proxies is not None else '无'
|
|
|
|
| 1 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2 |
def check_proxy(proxies):
|
| 3 |
import requests
|
| 4 |
proxies_https = proxies['https'] if proxies is not None else '无'
|
crazy_functions/读文章写摘要.py
CHANGED
|
@@ -1,25 +1,8 @@
|
|
| 1 |
-
from functools import wraps
|
| 2 |
from predict import predict_no_ui
|
|
|
|
| 3 |
fast_debug = False
|
| 4 |
|
| 5 |
|
| 6 |
-
def report_execption(chatbot, history, a, b):
|
| 7 |
-
chatbot.append((a, b))
|
| 8 |
-
history.append(a); history.append(b)
|
| 9 |
-
|
| 10 |
-
# 捕获不能预料的异常
|
| 11 |
-
def CatchException(f):
|
| 12 |
-
@wraps(f)
|
| 13 |
-
def decorated(txt, top_p, temperature, chatbot, history, systemPromptTxt, WEB_PORT):
|
| 14 |
-
try:
|
| 15 |
-
yield from f(txt, top_p, temperature, chatbot, history, systemPromptTxt, WEB_PORT)
|
| 16 |
-
except Exception as e:
|
| 17 |
-
import traceback
|
| 18 |
-
tb_str = traceback.format_exc()
|
| 19 |
-
chatbot[-1] = (chatbot[-1][0], f"[Local Message] something error occured: \n {tb_str}")
|
| 20 |
-
yield chatbot, history, f'异常 {e}'
|
| 21 |
-
return decorated
|
| 22 |
-
|
| 23 |
def 解析Paper(file_manifest, project_folder, top_p, temperature, chatbot, history, systemPromptTxt):
|
| 24 |
import time, glob, os
|
| 25 |
print('begin analysis on:', file_manifest)
|
|
|
|
|
|
|
| 1 |
from predict import predict_no_ui
|
| 2 |
+
from toolbox import CatchException, report_execption
|
| 3 |
fast_debug = False
|
| 4 |
|
| 5 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6 |
def 解析Paper(file_manifest, project_folder, top_p, temperature, chatbot, history, systemPromptTxt):
|
| 7 |
import time, glob, os
|
| 8 |
print('begin analysis on:', file_manifest)
|
functional_crazy.py
CHANGED
|
@@ -1,18 +1,7 @@
|
|
| 1 |
-
from functools import wraps
|
| 2 |
from predict import predict_no_ui
|
|
|
|
| 3 |
fast_debug = False
|
| 4 |
|
| 5 |
-
def CatchException(f):
|
| 6 |
-
@wraps(f)
|
| 7 |
-
def decorated(txt, top_p, temperature, chatbot, history, systemPromptTxt, WEB_PORT):
|
| 8 |
-
try:
|
| 9 |
-
yield from f(txt, top_p, temperature, chatbot, history, systemPromptTxt, WEB_PORT)
|
| 10 |
-
except Exception as e:
|
| 11 |
-
import traceback
|
| 12 |
-
tb_str = traceback.format_exc()
|
| 13 |
-
chatbot[-1] = (chatbot[-1][0], f"[Local Message] something error occured: \n {tb_str}")
|
| 14 |
-
yield chatbot, history, f'异常 {e}'
|
| 15 |
-
return decorated
|
| 16 |
|
| 17 |
|
| 18 |
@CatchException
|
|
@@ -66,12 +55,6 @@ def 解析项目本身(txt, top_p, temperature, chatbot, history, systemPromptTx
|
|
| 66 |
history.append(i_say); history.append(gpt_say)
|
| 67 |
yield chatbot, history, '正常'
|
| 68 |
|
| 69 |
-
def report_execption(chatbot, history, a, b):
|
| 70 |
-
chatbot.append((a, b))
|
| 71 |
-
history.append(a); history.append(b)
|
| 72 |
-
|
| 73 |
-
|
| 74 |
-
|
| 75 |
|
| 76 |
|
| 77 |
def 解析源代码(file_manifest, project_folder, top_p, temperature, chatbot, history, systemPromptTxt):
|
|
|
|
|
|
|
| 1 |
from predict import predict_no_ui
|
| 2 |
+
from toolbox import CatchException, report_execption
|
| 3 |
fast_debug = False
|
| 4 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5 |
|
| 6 |
|
| 7 |
@CatchException
|
|
|
|
| 55 |
history.append(i_say); history.append(gpt_say)
|
| 56 |
yield chatbot, history, '正常'
|
| 57 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 58 |
|
| 59 |
|
| 60 |
def 解析源代码(file_manifest, project_folder, top_p, temperature, chatbot, history, systemPromptTxt):
|
main.py
CHANGED
|
@@ -1,20 +1,11 @@
|
|
| 1 |
import os; os.environ['no_proxy'] = '*'
|
| 2 |
import gradio as gr
|
| 3 |
-
import markdown, mdtex2html
|
| 4 |
from predict import predict
|
| 5 |
-
from
|
| 6 |
|
| 7 |
try: from config_private import proxies, WEB_PORT # 放自己的秘密如API和代理网址 os.path.exists('config_private.py')
|
| 8 |
except: from config import proxies, WEB_PORT
|
| 9 |
|
| 10 |
-
def find_free_port():
|
| 11 |
-
import socket
|
| 12 |
-
from contextlib import closing
|
| 13 |
-
with closing(socket.socket(socket.AF_INET, socket.SOCK_STREAM)) as s:
|
| 14 |
-
s.bind(('', 0))
|
| 15 |
-
s.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)
|
| 16 |
-
return s.getsockname()[1]
|
| 17 |
-
|
| 18 |
PORT = find_free_port() if WEB_PORT <= 0 else WEB_PORT
|
| 19 |
|
| 20 |
initial_prompt = "Serve me as a writing and programming assistant."
|
|
@@ -33,41 +24,6 @@ functional = get_functionals()
|
|
| 33 |
from functional_crazy import get_crazy_functionals
|
| 34 |
crazy_functional = get_crazy_functionals()
|
| 35 |
|
| 36 |
-
def reset_textbox(): return gr.update(value='')
|
| 37 |
-
|
| 38 |
-
def text_divide_paragraph(text):
|
| 39 |
-
if '```' in text:
|
| 40 |
-
# careful input
|
| 41 |
-
return text
|
| 42 |
-
else:
|
| 43 |
-
# wtf input
|
| 44 |
-
lines = text.split("\n")
|
| 45 |
-
for i, line in enumerate(lines):
|
| 46 |
-
if i!=0: lines[i] = "<p>"+lines[i].replace(" ", " ")+"</p>"
|
| 47 |
-
text = "".join(lines)
|
| 48 |
-
return text
|
| 49 |
-
|
| 50 |
-
def markdown_convertion(txt):
|
| 51 |
-
if ('$' in txt) and ('```' not in txt):
|
| 52 |
-
math_config = {'mdx_math': {'enable_dollar_delimiter': True}}
|
| 53 |
-
return markdown.markdown(txt,extensions=['fenced_code','tables']) + '<br><br>' + \
|
| 54 |
-
markdown.markdown(convert_math(txt, splitParagraphs=False),extensions=['fenced_code','tables'])
|
| 55 |
-
else:
|
| 56 |
-
return markdown.markdown(txt,extensions=['fenced_code','tables'])
|
| 57 |
-
|
| 58 |
-
|
| 59 |
-
def format_io(self,y):
|
| 60 |
-
if y is None:
|
| 61 |
-
return []
|
| 62 |
-
i_ask, gpt_reply = y[-1]
|
| 63 |
-
|
| 64 |
-
i_ask = text_divide_paragraph(i_ask) # 输入部分太自由,预处理一波
|
| 65 |
-
|
| 66 |
-
y[-1] = (
|
| 67 |
-
None if i_ask is None else markdown.markdown(i_ask, extensions=['fenced_code','tables']),
|
| 68 |
-
None if gpt_reply is None else markdown_convertion(gpt_reply)
|
| 69 |
-
)
|
| 70 |
-
return y
|
| 71 |
gr.Chatbot.postprocess = format_io
|
| 72 |
|
| 73 |
with gr.Blocks() as demo:
|
|
@@ -103,7 +59,6 @@ with gr.Blocks() as demo:
|
|
| 103 |
|
| 104 |
txt.submit(predict, [txt, top_p, temperature, chatbot, history, systemPromptTxt], [chatbot, history, statusDisplay])
|
| 105 |
submitBtn.click(predict, [txt, top_p, temperature, chatbot, history, systemPromptTxt], [chatbot, history, statusDisplay], show_progress=True)
|
| 106 |
-
# submitBtn.click(reset_textbox, [], [txt])
|
| 107 |
for k in functional:
|
| 108 |
functional[k]["Button"].click(predict,
|
| 109 |
[txt, top_p, temperature, chatbot, history, systemPromptTxt, TRUE, gr.State(k)], [chatbot, history, statusDisplay], show_progress=True)
|
|
@@ -111,15 +66,15 @@ with gr.Blocks() as demo:
|
|
| 111 |
crazy_functional[k]["Button"].click(crazy_functional[k]["Function"],
|
| 112 |
[txt, top_p, temperature, chatbot, history, systemPromptTxt, gr.State(PORT)], [chatbot, history, statusDisplay])
|
| 113 |
|
| 114 |
-
print(f"URL http://localhost:{PORT}")
|
| 115 |
-
demo.title = "ChatGPT 学术优化"
|
| 116 |
|
| 117 |
def auto_opentab_delay():
|
| 118 |
import threading, webbrowser, time
|
|
|
|
| 119 |
def open(): time.sleep(2)
|
| 120 |
webbrowser.open_new_tab(f'http://localhost:{PORT}')
|
| 121 |
t = threading.Thread(target=open)
|
| 122 |
t.daemon = True; t.start()
|
| 123 |
|
| 124 |
auto_opentab_delay()
|
|
|
|
| 125 |
demo.queue().launch(server_name="0.0.0.0", share=True, server_port=PORT)
|
|
|
|
| 1 |
import os; os.environ['no_proxy'] = '*'
|
| 2 |
import gradio as gr
|
|
|
|
| 3 |
from predict import predict
|
| 4 |
+
from toolbox import format_io, find_free_port
|
| 5 |
|
| 6 |
try: from config_private import proxies, WEB_PORT # 放自己的秘密如API和代理网址 os.path.exists('config_private.py')
|
| 7 |
except: from config import proxies, WEB_PORT
|
| 8 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
PORT = find_free_port() if WEB_PORT <= 0 else WEB_PORT
|
| 10 |
|
| 11 |
initial_prompt = "Serve me as a writing and programming assistant."
|
|
|
|
| 24 |
from functional_crazy import get_crazy_functionals
|
| 25 |
crazy_functional = get_crazy_functionals()
|
| 26 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 27 |
gr.Chatbot.postprocess = format_io
|
| 28 |
|
| 29 |
with gr.Blocks() as demo:
|
|
|
|
| 59 |
|
| 60 |
txt.submit(predict, [txt, top_p, temperature, chatbot, history, systemPromptTxt], [chatbot, history, statusDisplay])
|
| 61 |
submitBtn.click(predict, [txt, top_p, temperature, chatbot, history, systemPromptTxt], [chatbot, history, statusDisplay], show_progress=True)
|
|
|
|
| 62 |
for k in functional:
|
| 63 |
functional[k]["Button"].click(predict,
|
| 64 |
[txt, top_p, temperature, chatbot, history, systemPromptTxt, TRUE, gr.State(k)], [chatbot, history, statusDisplay], show_progress=True)
|
|
|
|
| 66 |
crazy_functional[k]["Button"].click(crazy_functional[k]["Function"],
|
| 67 |
[txt, top_p, temperature, chatbot, history, systemPromptTxt, gr.State(PORT)], [chatbot, history, statusDisplay])
|
| 68 |
|
|
|
|
|
|
|
| 69 |
|
| 70 |
def auto_opentab_delay():
|
| 71 |
import threading, webbrowser, time
|
| 72 |
+
print(f"URL http://localhost:{PORT}")
|
| 73 |
def open(): time.sleep(2)
|
| 74 |
webbrowser.open_new_tab(f'http://localhost:{PORT}')
|
| 75 |
t = threading.Thread(target=open)
|
| 76 |
t.daemon = True; t.start()
|
| 77 |
|
| 78 |
auto_opentab_delay()
|
| 79 |
+
demo.title = "ChatGPT 学术优化"
|
| 80 |
demo.queue().launch(server_name="0.0.0.0", share=True, server_port=PORT)
|
toolbox.py
ADDED
|
@@ -0,0 +1,68 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import markdown, mdtex2html
|
| 2 |
+
from show_math import convert as convert_math
|
| 3 |
+
from functools import wraps
|
| 4 |
+
|
| 5 |
+
|
| 6 |
+
def regular_txt_to_markdown(text):
|
| 7 |
+
text = text.replace('\n', '\n\n')
|
| 8 |
+
text = text.replace('\n\n\n', '\n\n')
|
| 9 |
+
text = text.replace('\n\n\n', '\n\n')
|
| 10 |
+
return text
|
| 11 |
+
|
| 12 |
+
def CatchException(f):
|
| 13 |
+
@wraps(f)
|
| 14 |
+
def decorated(txt, top_p, temperature, chatbot, history, systemPromptTxt, WEB_PORT):
|
| 15 |
+
try:
|
| 16 |
+
yield from f(txt, top_p, temperature, chatbot, history, systemPromptTxt, WEB_PORT)
|
| 17 |
+
except Exception as e:
|
| 18 |
+
import traceback
|
| 19 |
+
from check_proxy import check_proxy
|
| 20 |
+
try: from config_private import proxies
|
| 21 |
+
except: from config import proxies
|
| 22 |
+
tb_str = regular_txt_to_markdown(traceback.format_exc())
|
| 23 |
+
chatbot[-1] = (chatbot[-1][0], f"[Local Message] 实验性函数调用出错: \n\n {tb_str} \n\n 当前代理可用性: \n\n {check_proxy(proxies)}")
|
| 24 |
+
yield chatbot, history, f'异常 {e}'
|
| 25 |
+
return decorated
|
| 26 |
+
|
| 27 |
+
def report_execption(chatbot, history, a, b):
|
| 28 |
+
chatbot.append((a, b))
|
| 29 |
+
history.append(a); history.append(b)
|
| 30 |
+
|
| 31 |
+
def text_divide_paragraph(text):
|
| 32 |
+
if '```' in text:
|
| 33 |
+
# careful input
|
| 34 |
+
return text
|
| 35 |
+
else:
|
| 36 |
+
# wtf input
|
| 37 |
+
lines = text.split("\n")
|
| 38 |
+
for i, line in enumerate(lines):
|
| 39 |
+
if i!=0: lines[i] = "<p>"+lines[i].replace(" ", " ")+"</p>"
|
| 40 |
+
text = "".join(lines)
|
| 41 |
+
return text
|
| 42 |
+
|
| 43 |
+
def markdown_convertion(txt):
|
| 44 |
+
if ('$' in txt) and ('```' not in txt):
|
| 45 |
+
return markdown.markdown(txt,extensions=['fenced_code','tables']) + '<br><br>' + \
|
| 46 |
+
markdown.markdown(convert_math(txt, splitParagraphs=False),extensions=['fenced_code','tables'])
|
| 47 |
+
else:
|
| 48 |
+
return markdown.markdown(txt,extensions=['fenced_code','tables'])
|
| 49 |
+
|
| 50 |
+
|
| 51 |
+
def format_io(self, y):
|
| 52 |
+
if y is None: return []
|
| 53 |
+
i_ask, gpt_reply = y[-1]
|
| 54 |
+
i_ask = text_divide_paragraph(i_ask) # 输入部分太自由,预处理一波
|
| 55 |
+
y[-1] = (
|
| 56 |
+
None if i_ask is None else markdown.markdown(i_ask, extensions=['fenced_code','tables']),
|
| 57 |
+
None if gpt_reply is None else markdown_convertion(gpt_reply)
|
| 58 |
+
)
|
| 59 |
+
return y
|
| 60 |
+
|
| 61 |
+
|
| 62 |
+
def find_free_port():
|
| 63 |
+
import socket
|
| 64 |
+
from contextlib import closing
|
| 65 |
+
with closing(socket.socket(socket.AF_INET, socket.SOCK_STREAM)) as s:
|
| 66 |
+
s.bind(('', 0))
|
| 67 |
+
s.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1)
|
| 68 |
+
return s.getsockname()[1]
|