Spaces:
Runtime error
Runtime error
Commit
·
85095bb
1
Parent(s):
4399c5c
Upload 3 files
Browse files- app.py +19 -25
- presets.py +31 -6
- utils.py +34 -13
app.py
CHANGED
|
@@ -43,7 +43,7 @@ else:
|
|
| 43 |
|
| 44 |
gr.Chatbot.postprocess = postprocess
|
| 45 |
|
| 46 |
-
with gr.Blocks(css=customCSS
|
| 47 |
history = gr.State([])
|
| 48 |
token_count = gr.State([])
|
| 49 |
promptTemplates = gr.State(load_template(get_template_names(plain=True)[0], mode=2))
|
|
@@ -51,13 +51,11 @@ with gr.Blocks(css=customCSS,) as demo:
|
|
| 51 |
FALSECONSTANT = gr.State(False)
|
| 52 |
topic = gr.State("未命名对话历史记录")
|
| 53 |
|
| 54 |
-
|
| 55 |
-
|
| 56 |
-
|
| 57 |
-
gr.HTML(title)
|
| 58 |
|
| 59 |
with gr.Row(scale=1).style(equal_height=True):
|
| 60 |
-
|
| 61 |
with gr.Column(scale=5):
|
| 62 |
with gr.Row(scale=1):
|
| 63 |
chatbot = gr.Chatbot().style(height=600) # .style(color_map=("#1D51EE", "#585A5B"))
|
|
@@ -73,11 +71,8 @@ with gr.Blocks(css=customCSS,) as demo:
|
|
| 73 |
delLastBtn = gr.Button("🗑️ 删除一条对话")
|
| 74 |
reduceTokenBtn = gr.Button("♻️ 总结对话")
|
| 75 |
|
| 76 |
-
|
| 77 |
-
|
| 78 |
with gr.Column():
|
| 79 |
with gr.Column(min_width=50,scale=1):
|
| 80 |
-
status_display = gr.Markdown("status: ready")
|
| 81 |
with gr.Tab(label="ChatGPT"):
|
| 82 |
keyTxt = gr.Textbox(show_label=True, placeholder=f"OpenAI API-key...",value=my_api_key, type="password", visible=not HIDE_MY_KEY, label="API-Key")
|
| 83 |
model_select_dropdown = gr.Dropdown(label="选择模型", choices=MODELS, multiselect=False, value=MODELS[0])
|
|
@@ -104,27 +99,27 @@ with gr.Blocks(css=customCSS,) as demo:
|
|
| 104 |
|
| 105 |
with gr.Tab(label="保存/加载"):
|
| 106 |
with gr.Accordion(label="保存/加载对话历史记录", open=True):
|
|
|
|
| 107 |
with gr.Column():
|
| 108 |
with gr.Row():
|
| 109 |
with gr.Column(scale=6):
|
| 110 |
saveFileName = gr.Textbox(
|
| 111 |
-
show_label=True, placeholder=f"
|
| 112 |
with gr.Column(scale=1):
|
| 113 |
saveHistoryBtn = gr.Button("💾 保存对话")
|
|
|
|
| 114 |
with gr.Row():
|
| 115 |
with gr.Column(scale=6):
|
| 116 |
historyFileSelectDropdown = gr.Dropdown(label="从列表中加载对话", choices=get_history_names(plain=True), multiselect=False, value=get_history_names(plain=True)[0])
|
| 117 |
with gr.Column(scale=1):
|
| 118 |
historyRefreshBtn = gr.Button("🔄 刷新")
|
|
|
|
|
|
|
|
|
|
| 119 |
|
| 120 |
-
|
| 121 |
-
|
| 122 |
-
gr.HTML("""
|
| 123 |
-
<div style="text-align: center; margin-top: 20px; margin-bottom: 20px;">
|
| 124 |
-
""")
|
| 125 |
gr.Markdown(description)
|
| 126 |
|
| 127 |
-
|
| 128 |
user_input.submit(predict, [keyTxt, systemPromptTxt, history, user_input, chatbot, token_count, top_p, temperature, use_streaming_checkbox, model_select_dropdown, use_websearch_checkbox], [chatbot, history, status_display, token_count], show_progress=True)
|
| 129 |
user_input.submit(reset_textbox, [], [user_input])
|
| 130 |
|
|
@@ -140,20 +135,19 @@ with gr.Blocks(css=customCSS,) as demo:
|
|
| 140 |
|
| 141 |
reduceTokenBtn.click(reduce_token_size, [keyTxt, systemPromptTxt, history, chatbot, token_count, top_p, temperature, use_streaming_checkbox, model_select_dropdown], [chatbot, history, status_display, token_count], show_progress=True)
|
| 142 |
|
| 143 |
-
|
| 144 |
-
|
|
|
|
|
|
|
| 145 |
|
|
|
|
|
|
|
| 146 |
saveHistoryBtn.click(get_history_names, None, [historyFileSelectDropdown])
|
| 147 |
-
|
| 148 |
historyRefreshBtn.click(get_history_names, None, [historyFileSelectDropdown])
|
| 149 |
-
|
| 150 |
historyFileSelectDropdown.change(load_chat_history, [historyFileSelectDropdown, systemPromptTxt, history, chatbot], [saveFileName, systemPromptTxt, history, chatbot], show_progress=True)
|
|
|
|
| 151 |
|
| 152 |
-
templateRefreshBtn.click(get_template_names, None, [templateFileSelectDropdown])
|
| 153 |
-
|
| 154 |
-
templateFileSelectDropdown.change(load_template, [templateFileSelectDropdown], [promptTemplates, templateSelectDropdown], show_progress=True)
|
| 155 |
-
|
| 156 |
-
templateSelectDropdown.change(get_template_content, [promptTemplates, templateSelectDropdown, systemPromptTxt], [systemPromptTxt], show_progress=True)
|
| 157 |
|
| 158 |
logging.info(colorama.Back.GREEN + "\n川虎的温馨提示:访问 http://localhost:7860 查看界面" + colorama.Style.RESET_ALL)
|
| 159 |
# 默认开启本地服务器,默认可以直接从IP访问,默认不创建公开分享链接
|
|
|
|
| 43 |
|
| 44 |
gr.Chatbot.postprocess = postprocess
|
| 45 |
|
| 46 |
+
with gr.Blocks(css=customCSS) as demo:
|
| 47 |
history = gr.State([])
|
| 48 |
token_count = gr.State([])
|
| 49 |
promptTemplates = gr.State(load_template(get_template_names(plain=True)[0], mode=2))
|
|
|
|
| 51 |
FALSECONSTANT = gr.State(False)
|
| 52 |
topic = gr.State("未命名对话历史记录")
|
| 53 |
|
| 54 |
+
with gr.Row():
|
| 55 |
+
gr.HTML(title)
|
| 56 |
+
status_display = gr.Markdown("status: ready", elem_id="status_display")
|
|
|
|
| 57 |
|
| 58 |
with gr.Row(scale=1).style(equal_height=True):
|
|
|
|
| 59 |
with gr.Column(scale=5):
|
| 60 |
with gr.Row(scale=1):
|
| 61 |
chatbot = gr.Chatbot().style(height=600) # .style(color_map=("#1D51EE", "#585A5B"))
|
|
|
|
| 71 |
delLastBtn = gr.Button("🗑️ 删除一条对话")
|
| 72 |
reduceTokenBtn = gr.Button("♻️ 总结对话")
|
| 73 |
|
|
|
|
|
|
|
| 74 |
with gr.Column():
|
| 75 |
with gr.Column(min_width=50,scale=1):
|
|
|
|
| 76 |
with gr.Tab(label="ChatGPT"):
|
| 77 |
keyTxt = gr.Textbox(show_label=True, placeholder=f"OpenAI API-key...",value=my_api_key, type="password", visible=not HIDE_MY_KEY, label="API-Key")
|
| 78 |
model_select_dropdown = gr.Dropdown(label="选择模型", choices=MODELS, multiselect=False, value=MODELS[0])
|
|
|
|
| 99 |
|
| 100 |
with gr.Tab(label="保存/加载"):
|
| 101 |
with gr.Accordion(label="保存/加载对话历史记录", open=True):
|
| 102 |
+
gr.Markdown("对话历史默认保存在history文件夹中。")
|
| 103 |
with gr.Column():
|
| 104 |
with gr.Row():
|
| 105 |
with gr.Column(scale=6):
|
| 106 |
saveFileName = gr.Textbox(
|
| 107 |
+
show_label=True, placeholder=f"设置文件名: 默认为.json,可选为.md", label="设置保存文件名", value="对话历史记录").style(container=True)
|
| 108 |
with gr.Column(scale=1):
|
| 109 |
saveHistoryBtn = gr.Button("💾 保存对话")
|
| 110 |
+
exportMarkdownBtn = gr.Button("📝 导出为Markdown")
|
| 111 |
with gr.Row():
|
| 112 |
with gr.Column(scale=6):
|
| 113 |
historyFileSelectDropdown = gr.Dropdown(label="从列表中加载对话", choices=get_history_names(plain=True), multiselect=False, value=get_history_names(plain=True)[0])
|
| 114 |
with gr.Column(scale=1):
|
| 115 |
historyRefreshBtn = gr.Button("🔄 刷新")
|
| 116 |
+
with gr.Row():
|
| 117 |
+
with gr.Column():
|
| 118 |
+
downloadFile = gr.File(interactive=True)
|
| 119 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 120 |
gr.Markdown(description)
|
| 121 |
|
| 122 |
+
# Chatbot
|
| 123 |
user_input.submit(predict, [keyTxt, systemPromptTxt, history, user_input, chatbot, token_count, top_p, temperature, use_streaming_checkbox, model_select_dropdown, use_websearch_checkbox], [chatbot, history, status_display, token_count], show_progress=True)
|
| 124 |
user_input.submit(reset_textbox, [], [user_input])
|
| 125 |
|
|
|
|
| 135 |
|
| 136 |
reduceTokenBtn.click(reduce_token_size, [keyTxt, systemPromptTxt, history, chatbot, token_count, top_p, temperature, use_streaming_checkbox, model_select_dropdown], [chatbot, history, status_display, token_count], show_progress=True)
|
| 137 |
|
| 138 |
+
# Template
|
| 139 |
+
templateRefreshBtn.click(get_template_names, None, [templateFileSelectDropdown])
|
| 140 |
+
templateFileSelectDropdown.change(load_template, [templateFileSelectDropdown], [promptTemplates, templateSelectDropdown], show_progress=True)
|
| 141 |
+
templateSelectDropdown.change(get_template_content, [promptTemplates, templateSelectDropdown, systemPromptTxt], [systemPromptTxt], show_progress=True)
|
| 142 |
|
| 143 |
+
# S&L
|
| 144 |
+
saveHistoryBtn.click(save_chat_history, [saveFileName, systemPromptTxt, history, chatbot], downloadFile, show_progress=True)
|
| 145 |
saveHistoryBtn.click(get_history_names, None, [historyFileSelectDropdown])
|
| 146 |
+
exportMarkdownBtn.click(export_markdown, [saveFileName, systemPromptTxt, history, chatbot], downloadFile, show_progress=True)
|
| 147 |
historyRefreshBtn.click(get_history_names, None, [historyFileSelectDropdown])
|
|
|
|
| 148 |
historyFileSelectDropdown.change(load_chat_history, [historyFileSelectDropdown, systemPromptTxt, history, chatbot], [saveFileName, systemPromptTxt, history, chatbot], show_progress=True)
|
| 149 |
+
downloadFile.change(load_chat_history, [downloadFile, systemPromptTxt, history, chatbot], [saveFileName, systemPromptTxt, history, chatbot])
|
| 150 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 151 |
|
| 152 |
logging.info(colorama.Back.GREEN + "\n川虎的温馨提示:访问 http://localhost:7860 查看界面" + colorama.Style.RESET_ALL)
|
| 153 |
# 默认开启本地服务器,默认可以直接从IP访问,默认不创建公开分享链接
|
presets.py
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
# -*- coding:utf-8 -*-
|
| 2 |
-
title = """<h1 align="left">川虎ChatGPT 🚀</h1>"""
|
| 3 |
-
description = """<div align=center>
|
| 4 |
|
| 5 |
由Bilibili [土川虎虎虎](https://space.bilibili.com/29125536) 和 [明昭MZhao](https://space.bilibili.com/24807452)开发
|
| 6 |
|
|
@@ -10,6 +10,33 @@ description = """<div align=center>
|
|
| 10 |
</div>
|
| 11 |
"""
|
| 12 |
customCSS = """
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 13 |
code {
|
| 14 |
display: inline;
|
| 15 |
white-space: break-spaces;
|
|
@@ -30,11 +57,9 @@ pre code {
|
|
| 30 |
box-shadow: inset 0px 8px 16px hsla(0, 0%, 0%, .2)
|
| 31 |
}
|
| 32 |
|
| 33 |
-
*{
|
| 34 |
transition: all 0.6s;
|
| 35 |
}
|
| 36 |
-
|
| 37 |
-
|
| 38 |
"""
|
| 39 |
|
| 40 |
summarize_prompt = "你是谁?我们刚才聊了什么?" # 总结对话时的 prompt
|
|
@@ -58,7 +83,7 @@ ssl_error_prompt = "SSL错误,无法获取对话。" # SSL 错误
|
|
| 58 |
no_apikey_msg = "API key长度不是51位,请检查是否输入正确。" # API key 长度不足 51 位
|
| 59 |
|
| 60 |
max_token_streaming = 3500 # 流式对话时的最大 token 数
|
| 61 |
-
timeout_streaming =
|
| 62 |
max_token_all = 3500 # 非流式对话时的最大 token 数
|
| 63 |
timeout_all = 200 # 非流式对话时的超时时间
|
| 64 |
enable_streaming_option = True # 是否启用选择选择是否实时显示回答的勾选框
|
|
|
|
| 1 |
# -*- coding:utf-8 -*-
|
| 2 |
+
title = """<h1 align="left" style="min-width:200px; margin-top:0;">川虎ChatGPT 🚀</h1>"""
|
| 3 |
+
description = """<div align="center" style="margin:16px 0">
|
| 4 |
|
| 5 |
由Bilibili [土川虎虎虎](https://space.bilibili.com/29125536) 和 [明昭MZhao](https://space.bilibili.com/24807452)开发
|
| 6 |
|
|
|
|
| 10 |
</div>
|
| 11 |
"""
|
| 12 |
customCSS = """
|
| 13 |
+
#status_display {
|
| 14 |
+
display: flex;
|
| 15 |
+
min-height: 2.5em;
|
| 16 |
+
align-items: flex-end;
|
| 17 |
+
justify-content: flex-end;
|
| 18 |
+
}
|
| 19 |
+
#status_display p {
|
| 20 |
+
font-size: .85em;
|
| 21 |
+
font-family: monospace;
|
| 22 |
+
color: var(--text-color-subdued) !important;
|
| 23 |
+
}
|
| 24 |
+
[class *= "message"] {
|
| 25 |
+
border-radius: var(--radius-xl) !important;
|
| 26 |
+
border: none;
|
| 27 |
+
padding: var(--spacing-xl) !important;
|
| 28 |
+
font-size: var(--text-md) !important;
|
| 29 |
+
line-height: var(--line-md) !important;
|
| 30 |
+
}
|
| 31 |
+
[data-testid = "bot"] {
|
| 32 |
+
max-width: 85%;
|
| 33 |
+
border-bottom-left-radius: 0 !important;
|
| 34 |
+
}
|
| 35 |
+
[data-testid = "user"] {
|
| 36 |
+
max-width: 85%;
|
| 37 |
+
width: auto !important;
|
| 38 |
+
border-bottom-right-radius: 0 !important;
|
| 39 |
+
}
|
| 40 |
code {
|
| 41 |
display: inline;
|
| 42 |
white-space: break-spaces;
|
|
|
|
| 57 |
box-shadow: inset 0px 8px 16px hsla(0, 0%, 0%, .2)
|
| 58 |
}
|
| 59 |
|
| 60 |
+
* {
|
| 61 |
transition: all 0.6s;
|
| 62 |
}
|
|
|
|
|
|
|
| 63 |
"""
|
| 64 |
|
| 65 |
summarize_prompt = "你是谁?我们刚才聊了什么?" # 总结对话时的 prompt
|
|
|
|
| 83 |
no_apikey_msg = "API key长度不是51位,请检查是否输入正确。" # API key 长度不足 51 位
|
| 84 |
|
| 85 |
max_token_streaming = 3500 # 流式对话时的最大 token 数
|
| 86 |
+
timeout_streaming = 30 # 流式对话时的超时时间
|
| 87 |
max_token_all = 3500 # 非流式对话时的最大 token 数
|
| 88 |
timeout_all = 200 # 非流式对话时的超时时间
|
| 89 |
enable_streaming_option = True # 是否启用选择选择是否实时显示回答的勾选框
|
utils.py
CHANGED
|
@@ -48,13 +48,14 @@ def postprocess(
|
|
| 48 |
y[i] = (
|
| 49 |
# None if message is None else markdown.markdown(message),
|
| 50 |
# None if response is None else markdown.markdown(response),
|
| 51 |
-
None if message is None else
|
| 52 |
None if response is None else mdtex2html.convert(response),
|
| 53 |
)
|
| 54 |
return y
|
| 55 |
|
| 56 |
-
def count_token(
|
| 57 |
encoding = tiktoken.get_encoding("cl100k_base")
|
|
|
|
| 58 |
length = len(encoding.encode(input_str))
|
| 59 |
return length
|
| 60 |
|
|
@@ -142,10 +143,10 @@ def stream_predict(openai_api_key, system_prompt, history, inputs, chatbot, all_
|
|
| 142 |
chatbot.append((parse_text(inputs), ""))
|
| 143 |
user_token_count = 0
|
| 144 |
if len(all_token_counts) == 0:
|
| 145 |
-
system_prompt_token_count = count_token(system_prompt)
|
| 146 |
-
user_token_count = count_token(inputs) + system_prompt_token_count
|
| 147 |
else:
|
| 148 |
-
user_token_count = count_token(inputs)
|
| 149 |
all_token_counts.append(user_token_count)
|
| 150 |
logging.info(f"输入token计数: {user_token_count}")
|
| 151 |
yield get_return_value()
|
|
@@ -204,7 +205,7 @@ def predict_all(openai_api_key, system_prompt, history, inputs, chatbot, all_tok
|
|
| 204 |
history.append(construct_user(inputs))
|
| 205 |
history.append(construct_assistant(""))
|
| 206 |
chatbot.append((parse_text(inputs), ""))
|
| 207 |
-
all_token_counts.append(count_token(inputs))
|
| 208 |
try:
|
| 209 |
response = get_response(openai_api_key, system_prompt, history, temperature, top_p, False, selected_model)
|
| 210 |
except requests.exceptions.ConnectTimeout:
|
|
@@ -322,22 +323,42 @@ def delete_last_conversation(chatbot, history, previous_token_count):
|
|
| 322 |
return chatbot, history, previous_token_count, construct_token_message(sum(previous_token_count))
|
| 323 |
|
| 324 |
|
| 325 |
-
def
|
| 326 |
logging.info("保存对话历史中……")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 327 |
if filename == "":
|
| 328 |
return
|
| 329 |
if not filename.endswith(".json"):
|
| 330 |
filename += ".json"
|
| 331 |
-
|
| 332 |
-
|
| 333 |
-
|
| 334 |
-
|
| 335 |
-
|
| 336 |
-
|
|
|
|
|
|
|
| 337 |
|
| 338 |
|
| 339 |
def load_chat_history(filename, system, history, chatbot):
|
| 340 |
logging.info("加载对话历史中……")
|
|
|
|
|
|
|
| 341 |
try:
|
| 342 |
with open(os.path.join(HISTORY_DIR, filename), "r") as f:
|
| 343 |
json_s = json.load(f)
|
|
|
|
| 48 |
y[i] = (
|
| 49 |
# None if message is None else markdown.markdown(message),
|
| 50 |
# None if response is None else markdown.markdown(response),
|
| 51 |
+
None if message is None else message,
|
| 52 |
None if response is None else mdtex2html.convert(response),
|
| 53 |
)
|
| 54 |
return y
|
| 55 |
|
| 56 |
+
def count_token(message):
|
| 57 |
encoding = tiktoken.get_encoding("cl100k_base")
|
| 58 |
+
input_str = f"role: {message['role']}, content: {message['content']}"
|
| 59 |
length = len(encoding.encode(input_str))
|
| 60 |
return length
|
| 61 |
|
|
|
|
| 143 |
chatbot.append((parse_text(inputs), ""))
|
| 144 |
user_token_count = 0
|
| 145 |
if len(all_token_counts) == 0:
|
| 146 |
+
system_prompt_token_count = count_token(construct_system(system_prompt))
|
| 147 |
+
user_token_count = count_token(construct_user(inputs)) + system_prompt_token_count
|
| 148 |
else:
|
| 149 |
+
user_token_count = count_token(construct_user(inputs))
|
| 150 |
all_token_counts.append(user_token_count)
|
| 151 |
logging.info(f"输入token计数: {user_token_count}")
|
| 152 |
yield get_return_value()
|
|
|
|
| 205 |
history.append(construct_user(inputs))
|
| 206 |
history.append(construct_assistant(""))
|
| 207 |
chatbot.append((parse_text(inputs), ""))
|
| 208 |
+
all_token_counts.append(count_token(construct_user(inputs)))
|
| 209 |
try:
|
| 210 |
response = get_response(openai_api_key, system_prompt, history, temperature, top_p, False, selected_model)
|
| 211 |
except requests.exceptions.ConnectTimeout:
|
|
|
|
| 323 |
return chatbot, history, previous_token_count, construct_token_message(sum(previous_token_count))
|
| 324 |
|
| 325 |
|
| 326 |
+
def save_file(filename, system, history, chatbot):
|
| 327 |
logging.info("保存对话历史中……")
|
| 328 |
+
os.makedirs(HISTORY_DIR, exist_ok=True)
|
| 329 |
+
if filename.endswith(".json"):
|
| 330 |
+
json_s = {"system": system, "history": history, "chatbot": chatbot}
|
| 331 |
+
print(json_s)
|
| 332 |
+
with open(os.path.join(HISTORY_DIR, filename), "w") as f:
|
| 333 |
+
json.dump(json_s, f)
|
| 334 |
+
elif filename.endswith(".md"):
|
| 335 |
+
md_s = f"system: \n- {system} \n"
|
| 336 |
+
for data in history:
|
| 337 |
+
md_s += f"\n{data['role']}: \n- {data['content']} \n"
|
| 338 |
+
with open(os.path.join(HISTORY_DIR, filename), "w", encoding="utf8") as f:
|
| 339 |
+
f.write(md_s)
|
| 340 |
+
logging.info("保存对话历史完毕")
|
| 341 |
+
return os.path.join(HISTORY_DIR, filename)
|
| 342 |
+
|
| 343 |
+
def save_chat_history(filename, system, history, chatbot):
|
| 344 |
if filename == "":
|
| 345 |
return
|
| 346 |
if not filename.endswith(".json"):
|
| 347 |
filename += ".json"
|
| 348 |
+
return save_file(filename, system, history, chatbot)
|
| 349 |
+
|
| 350 |
+
def export_markdown(filename, system, history, chatbot):
|
| 351 |
+
if filename == "":
|
| 352 |
+
return
|
| 353 |
+
if not filename.endswith(".md"):
|
| 354 |
+
filename += ".md"
|
| 355 |
+
return save_file(filename, system, history, chatbot)
|
| 356 |
|
| 357 |
|
| 358 |
def load_chat_history(filename, system, history, chatbot):
|
| 359 |
logging.info("加载对话历史中……")
|
| 360 |
+
if type(filename) != str:
|
| 361 |
+
filename = filename.name
|
| 362 |
try:
|
| 363 |
with open(os.path.join(HISTORY_DIR, filename), "r") as f:
|
| 364 |
json_s = json.load(f)
|