Spaces:
Running
Running
Fix bug: deprecation warning and infinite loop
#5
by
zfb
- opened
- .gitignore +6 -0
- app.py +13 -13
- optimizeOpenAI.py +4 -1
- requirements.txt +0 -1
.gitignore
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Ignore all __pycache__ directories
|
2 |
+
__pycache__/
|
3 |
+
/__pycache__
|
4 |
+
# Ignore all .pyc files
|
5 |
+
*.pyc
|
6 |
+
flagged/
|
app.py
CHANGED
@@ -808,8 +808,8 @@ Use ChatGPT to summary the papers.Star our Github [🌟ChatPaper](https://github
|
|
808 |
'''
|
809 |
|
810 |
api_input = [
|
811 |
-
gradio.
|
812 |
-
|
813 |
type='password')
|
814 |
]
|
815 |
api_gui = gradio.Interface(fn=valid_apikey,
|
@@ -841,28 +841,28 @@ Use ChatGPT to summary the papers.Star our Github [🌟ChatPaper](https://github
|
|
841 |
'''
|
842 |
# 创建Gradio界面
|
843 |
ip = [
|
844 |
-
gradio.
|
845 |
-
|
846 |
type='password'),
|
847 |
-
gradio.
|
848 |
label="请输入论文大标题索引(用英文逗号隔开,必填)",
|
849 |
-
|
850 |
"'Abstract,Introduction,Related Work,Background,Preliminary,Problem Formulation,Methods,Methodology,Method,Approach,Approaches,Materials and Methods,Experiment Settings,Experiment,Experimental Results,Evaluation,Experiments,Results,Findings,Data Analysis,Discussion,Results and Discussion,Conclusion,References'"
|
851 |
),
|
852 |
-
gradio.
|
853 |
-
|
854 |
label="Select model"),
|
855 |
-
gradio.
|
856 |
maximum=1.0,
|
857 |
-
|
858 |
step=0.05,
|
859 |
label="Top-p (nucleus sampling)"),
|
860 |
-
gradio.
|
861 |
maximum=5.0,
|
862 |
-
|
863 |
step=0.5,
|
864 |
label="Temperature"),
|
865 |
-
gradio.
|
866 |
]
|
867 |
|
868 |
chatpaper_gui = gradio.Interface(fn=upload_pdf,
|
|
|
808 |
'''
|
809 |
|
810 |
api_input = [
|
811 |
+
gradio.Textbox(label="请输入你的API-key(必填, 多个API-key请用英文逗号隔开)",
|
812 |
+
value="",
|
813 |
type='password')
|
814 |
]
|
815 |
api_gui = gradio.Interface(fn=valid_apikey,
|
|
|
841 |
'''
|
842 |
# 创建Gradio界面
|
843 |
ip = [
|
844 |
+
gradio.Textbox(label="请输入你的API-key(必填, 多个API-key请用英文逗号隔开),不需要空格",
|
845 |
+
value="",
|
846 |
type='password'),
|
847 |
+
gradio.Textbox(
|
848 |
label="请输入论文大标题索引(用英文逗号隔开,必填)",
|
849 |
+
value=
|
850 |
"'Abstract,Introduction,Related Work,Background,Preliminary,Problem Formulation,Methods,Methodology,Method,Approach,Approaches,Materials and Methods,Experiment Settings,Experiment,Experimental Results,Evaluation,Experiments,Results,Findings,Data Analysis,Discussion,Results and Discussion,Conclusion,References'"
|
851 |
),
|
852 |
+
gradio.Radio(choices=["gpt-3.5-turbo", "gpt-3.5-turbo-0301"],
|
853 |
+
value="gpt-3.5-turbo",
|
854 |
label="Select model"),
|
855 |
+
gradio.Slider(minimum=-0,
|
856 |
maximum=1.0,
|
857 |
+
value=1.0,
|
858 |
step=0.05,
|
859 |
label="Top-p (nucleus sampling)"),
|
860 |
+
gradio.Slider(minimum=-0,
|
861 |
maximum=5.0,
|
862 |
+
value=0.5,
|
863 |
step=0.5,
|
864 |
label="Temperature"),
|
865 |
+
gradio.File(label="请上传论文PDF(必填)")
|
866 |
]
|
867 |
|
868 |
chatpaper_gui = gradio.Interface(fn=upload_pdf,
|
optimizeOpenAI.py
CHANGED
@@ -92,6 +92,9 @@ class chatPaper:
|
|
92 |
full_conversation = ""
|
93 |
for x in self.conversation[convo_id]:
|
94 |
full_conversation = str(x["content"]) + "\n" + full_conversation
|
|
|
|
|
|
|
95 |
while True:
|
96 |
if (len(ENCODER.encode(full_conversation+query)) > self.max_tokens):
|
97 |
query = query[:self.decrease_step]
|
@@ -208,7 +211,7 @@ class chatPaper:
|
|
208 |
input = input[self.decrease_step:]
|
209 |
prompt = prompt.replace("{conversation}", input)
|
210 |
self.reset(convo_id='conversationSummary')
|
211 |
-
response = self.ask(prompt,convo_id='conversationSummary')
|
212 |
while self.token_str(str(response))>self.max_tokens:
|
213 |
response = response[:-self.decrease_step]
|
214 |
self.reset(convo_id='conversationSummary',system_prompt='Summariaze our diaglog')
|
|
|
92 |
full_conversation = ""
|
93 |
for x in self.conversation[convo_id]:
|
94 |
full_conversation = str(x["content"]) + "\n" + full_conversation
|
95 |
+
max_len_full_conversation = self.max_tokens - self.decrease_step
|
96 |
+
if len(ENCODER.encode(full_conversation)) > max_len_full_conversation:
|
97 |
+
full_conversation = full_conversation[:-max_len_full_conversation]
|
98 |
while True:
|
99 |
if (len(ENCODER.encode(full_conversation+query)) > self.max_tokens):
|
100 |
query = query[:self.decrease_step]
|
|
|
211 |
input = input[self.decrease_step:]
|
212 |
prompt = prompt.replace("{conversation}", input)
|
213 |
self.reset(convo_id='conversationSummary')
|
214 |
+
response = self.ask(prompt,convo_id='conversationSummary')[0]
|
215 |
while self.token_str(str(response))>self.max_tokens:
|
216 |
response = response[:-self.decrease_step]
|
217 |
self.reset(convo_id='conversationSummary',system_prompt='Summariaze our diaglog')
|
requirements.txt
CHANGED
@@ -1,7 +1,6 @@
|
|
1 |
arxiv==1.4.3
|
2 |
PyMuPDF==1.21.1
|
3 |
requests==2.26.0
|
4 |
-
tiktoken==0.2.0
|
5 |
tenacity==8.2.2
|
6 |
pybase64==1.2.3
|
7 |
Pillow==9.4.0
|
|
|
1 |
arxiv==1.4.3
|
2 |
PyMuPDF==1.21.1
|
3 |
requests==2.26.0
|
|
|
4 |
tenacity==8.2.2
|
5 |
pybase64==1.2.3
|
6 |
Pillow==9.4.0
|