Spaces:
Runtime error
Runtime error
merge
Browse files- main.py +11 -8
- toolbox.py +14 -2
main.py
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
import os; os.environ['no_proxy'] = '*' # 避免代理网络产生意外污染
|
2 |
import gradio as gr
|
3 |
from predict import predict
|
4 |
-
from toolbox import format_io, find_free_port, on_file_uploaded, on_report_generated, get_conf
|
5 |
|
6 |
# 建议您复制一个config_private.py放自己的秘密, 如API和代理网址, 避免不小心传github被别人看到
|
7 |
proxies, WEB_PORT, LLM_MODEL, CONCURRENT_COUNT, AUTHENTICATION, CHATBOT_HEIGHT = \
|
@@ -87,8 +87,12 @@ with gr.Blocks(theme=set_theme, analytics_enabled=False, css=advanced_css) as de
|
|
87 |
system_prompt = gr.Textbox(show_label=True, placeholder=f"System Prompt", label="System prompt", value=initial_prompt)
|
88 |
top_p = gr.Slider(minimum=-0, maximum=1.0, value=1.0, step=0.01,interactive=True, label="Top-p (nucleus sampling)",)
|
89 |
temperature = gr.Slider(minimum=-0, maximum=2.0, value=1.0, step=0.01, interactive=True, label="Temperature",)
|
90 |
-
checkboxes = gr.CheckboxGroup(["基础功能区", "函数插件区"], value=["基础功能区", "函数插件区"], label="显示/隐藏功能区")
|
91 |
gr.Markdown(description)
|
|
|
|
|
|
|
|
|
92 |
# 功能区显示开关与功能区的互动
|
93 |
def fn_area_visibility(a):
|
94 |
ret = {}
|
@@ -97,17 +101,16 @@ with gr.Blocks(theme=set_theme, analytics_enabled=False, css=advanced_css) as de
|
|
97 |
return ret
|
98 |
checkboxes.select(fn_area_visibility, [checkboxes], [area_basic_fn, area_crazy_fn] )
|
99 |
# 整理反复出现的控件句柄组合
|
100 |
-
input_combo = [txt, top_p, temperature, chatbot, history, system_prompt]
|
101 |
output_combo = [chatbot, history, status]
|
102 |
-
predict_args = dict(fn=predict, inputs=input_combo, outputs=output_combo)
|
103 |
-
empty_txt_args = dict(fn=lambda: "", inputs=[], outputs=[txt]) # 用于在提交后清空输入栏
|
104 |
# 提交按钮、重置按钮
|
105 |
-
cancel_handles.append(txt.submit(**predict_args))
|
106 |
-
cancel_handles.append(submitBtn.click(**predict_args))
|
107 |
resetBtn.click(lambda: ([], [], "已重置"), None, output_combo)
|
108 |
# 基础功能区的回调函数注册
|
109 |
for k in functional:
|
110 |
-
click_handle = functional[k]["Button"].click(predict, [*input_combo, gr.State(True), gr.State(k)], output_combo)
|
111 |
cancel_handles.append(click_handle)
|
112 |
# 文件上传区,接收文件后与chatbot的互动
|
113 |
file_upload.upload(on_file_uploaded, [file_upload, chatbot, txt], [chatbot, txt])
|
|
|
1 |
import os; os.environ['no_proxy'] = '*' # 避免代理网络产生意外污染
|
2 |
import gradio as gr
|
3 |
from predict import predict
|
4 |
+
from toolbox import format_io, find_free_port, on_file_uploaded, on_report_generated, get_conf, ArgsGeneralWrapper
|
5 |
|
6 |
# 建议您复制一个config_private.py放自己的秘密, 如API和代理网址, 避免不小心传github被别人看到
|
7 |
proxies, WEB_PORT, LLM_MODEL, CONCURRENT_COUNT, AUTHENTICATION, CHATBOT_HEIGHT = \
|
|
|
87 |
system_prompt = gr.Textbox(show_label=True, placeholder=f"System Prompt", label="System prompt", value=initial_prompt)
|
88 |
top_p = gr.Slider(minimum=-0, maximum=1.0, value=1.0, step=0.01,interactive=True, label="Top-p (nucleus sampling)",)
|
89 |
temperature = gr.Slider(minimum=-0, maximum=2.0, value=1.0, step=0.01, interactive=True, label="Temperature",)
|
90 |
+
checkboxes = gr.CheckboxGroup(["基础功能区", "函数插件区", "输入区2"], value=["基础功能区", "函数插件区"], label="显示/隐藏功能区")
|
91 |
gr.Markdown(description)
|
92 |
+
with gr.Accordion("输入区", open=True, visible=False) as input_crazy_fn:
|
93 |
+
with gr.Row():
|
94 |
+
txt2 = gr.Textbox(show_label=False, placeholder="Input question here.", label="输入区2").style(container=False)
|
95 |
+
|
96 |
# 功能区显示开关与功能区的互动
|
97 |
def fn_area_visibility(a):
|
98 |
ret = {}
|
|
|
101 |
return ret
|
102 |
checkboxes.select(fn_area_visibility, [checkboxes], [area_basic_fn, area_crazy_fn] )
|
103 |
# 整理反复出现的控件句柄组合
|
104 |
+
input_combo = [txt, txt2, top_p, temperature, chatbot, history, system_prompt]
|
105 |
output_combo = [chatbot, history, status]
|
106 |
+
predict_args = dict(fn=ArgsGeneralWrapper(predict), inputs=input_combo, outputs=output_combo)
|
|
|
107 |
# 提交按钮、重置按钮
|
108 |
+
cancel_handles.append(txt.submit(**predict_args))
|
109 |
+
cancel_handles.append(submitBtn.click(**predict_args))
|
110 |
resetBtn.click(lambda: ([], [], "已重置"), None, output_combo)
|
111 |
# 基础功能区的回调函数注册
|
112 |
for k in functional:
|
113 |
+
click_handle = functional[k]["Button"].click(fn=ArgsGeneralWrapper(predict), inputs=[*input_combo, gr.State(True), gr.State(k)], outputs=output_combo)
|
114 |
cancel_handles.append(click_handle)
|
115 |
# 文件上传区,接收文件后与chatbot的互动
|
116 |
file_upload.upload(on_file_uploaded, [file_upload, chatbot, txt], [chatbot, txt])
|
toolbox.py
CHANGED
@@ -2,6 +2,18 @@ import markdown, mdtex2html, threading, importlib, traceback, importlib, inspect
|
|
2 |
from show_math import convert as convert_math
|
3 |
from functools import wraps, lru_cache
|
4 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5 |
def get_reduce_token_percent(text):
|
6 |
try:
|
7 |
# text = "maximum context length is 4097 tokens. However, your messages resulted in 4870 tokens"
|
@@ -116,7 +128,7 @@ def CatchException(f):
|
|
116 |
from toolbox import get_conf
|
117 |
proxies, = get_conf('proxies')
|
118 |
tb_str = '```\n' + traceback.format_exc() + '```'
|
119 |
-
if len(chatbot) == 0: chatbot
|
120 |
chatbot[-1] = (chatbot[-1][0], f"[Local Message] 实验性函数调用出错: \n\n{tb_str} \n\n当前代理可用性: \n\n{check_proxy(proxies)}")
|
121 |
yield chatbot, history, f'异常 {e}'
|
122 |
return decorated
|
@@ -129,7 +141,7 @@ def HotReload(f):
|
|
129 |
def decorated(*args, **kwargs):
|
130 |
fn_name = f.__name__
|
131 |
f_hot_reload = getattr(importlib.reload(inspect.getmodule(f)), fn_name)
|
132 |
-
yield from f_hot_reload(*args, **kwargs)
|
133 |
return decorated
|
134 |
|
135 |
def report_execption(chatbot, history, a, b):
|
|
|
2 |
from show_math import convert as convert_math
|
3 |
from functools import wraps, lru_cache
|
4 |
|
5 |
+
def ArgsGeneralWrapper(f):
|
6 |
+
"""
|
7 |
+
装饰器函数,用于重组输入参数,改变输入参数的顺序与结构。
|
8 |
+
"""
|
9 |
+
def decorated(txt, txt2, top_p, temperature, chatbot, history, system_prompt, *args, **kwargs):
|
10 |
+
txt_passon = txt
|
11 |
+
if txt == "" and txt2 != "": txt_passon = txt2
|
12 |
+
yield from f(txt_passon, top_p, temperature, chatbot, history, system_prompt, *args, **kwargs)
|
13 |
+
|
14 |
+
return decorated
|
15 |
+
|
16 |
+
|
17 |
def get_reduce_token_percent(text):
|
18 |
try:
|
19 |
# text = "maximum context length is 4097 tokens. However, your messages resulted in 4870 tokens"
|
|
|
128 |
from toolbox import get_conf
|
129 |
proxies, = get_conf('proxies')
|
130 |
tb_str = '```\n' + traceback.format_exc() + '```'
|
131 |
+
if chatbot is None or len(chatbot) == 0: chatbot = [["插件调度异常","异常原因"]]
|
132 |
chatbot[-1] = (chatbot[-1][0], f"[Local Message] 实验性函数调用出错: \n\n{tb_str} \n\n当前代理可用性: \n\n{check_proxy(proxies)}")
|
133 |
yield chatbot, history, f'异常 {e}'
|
134 |
return decorated
|
|
|
141 |
def decorated(*args, **kwargs):
|
142 |
fn_name = f.__name__
|
143 |
f_hot_reload = getattr(importlib.reload(inspect.getmodule(f)), fn_name)
|
144 |
+
yield from ArgsGeneralWrapper(f_hot_reload)(*args, **kwargs)
|
145 |
return decorated
|
146 |
|
147 |
def report_execption(chatbot, history, a, b):
|