3v324v23 commited on
Commit
deb8e5e
1 Parent(s): 5316b5c

动态endpoint

Browse files
crazy_functions/询问多个大语言模型.py CHANGED
@@ -16,7 +16,7 @@ def 同时问询(txt, llm_kwargs, plugin_kwargs, chatbot, history, system_prompt
16
  chatbot.append((txt, "正在同时咨询ChatGPT和ChatGLM……"))
17
  yield from update_ui(chatbot=chatbot, history=history) # 刷新界面 # 由于请求gpt需要一段时间,我们先及时地做一次界面更新
18
 
19
- llm_kwargs['llm_model'] = 'chatglm&gpt-3.5-turbo'
20
  gpt_say = yield from request_gpt_model_in_new_thread_with_ui_alive(
21
  inputs=txt, inputs_show_user=txt,
22
  llm_kwargs=llm_kwargs, chatbot=chatbot, history=history,
 
16
  chatbot.append((txt, "正在同时咨询ChatGPT和ChatGLM……"))
17
  yield from update_ui(chatbot=chatbot, history=history) # 刷新界面 # 由于请求gpt需要一段时间,我们先及时地做一次界面更新
18
 
19
+ llm_kwargs['llm_model'] = 'gpt-3.5-turbo&api2d-gpt-3.5-turbo'
20
  gpt_say = yield from request_gpt_model_in_new_thread_with_ui_alive(
21
  inputs=txt, inputs_show_user=txt,
22
  llm_kwargs=llm_kwargs, chatbot=chatbot, history=history,
request_llm/bridge_chatgpt.py CHANGED
@@ -60,7 +60,9 @@ def predict_no_ui_long_connection(inputs, llm_kwargs, history=[], sys_prompt="",
60
  while True:
61
  try:
62
  # make a POST request to the API endpoint, stream=False
63
- response = requests.post(llm_kwargs['endpoint'], headers=headers, proxies=proxies,
 
 
64
  json=payload, stream=True, timeout=TIMEOUT_SECONDS); break
65
  except requests.exceptions.ReadTimeout as e:
66
  retry += 1
@@ -148,7 +150,9 @@ def predict(inputs, llm_kwargs, plugin_kwargs, chatbot, history=[], system_promp
148
  while True:
149
  try:
150
  # make a POST request to the API endpoint, stream=True
151
- response = requests.post(llm_kwargs['endpoint'], headers=headers, proxies=proxies,
 
 
152
  json=payload, stream=True, timeout=TIMEOUT_SECONDS);break
153
  except:
154
  retry += 1
 
60
  while True:
61
  try:
62
  # make a POST request to the API endpoint, stream=False
63
+ from .bridge_all import model_info
64
+ endpoint = model_info[llm_kwargs['llm_model']]['endpoint']
65
+ response = requests.post(endpoint, headers=headers, proxies=proxies,
66
  json=payload, stream=True, timeout=TIMEOUT_SECONDS); break
67
  except requests.exceptions.ReadTimeout as e:
68
  retry += 1
 
150
  while True:
151
  try:
152
  # make a POST request to the API endpoint, stream=True
153
+ from .bridge_all import model_info
154
+ endpoint = model_info[llm_kwargs['llm_model']]['endpoint']
155
+ response = requests.post(endpoint, headers=headers, proxies=proxies,
156
  json=payload, stream=True, timeout=TIMEOUT_SECONDS);break
157
  except:
158
  retry += 1
toolbox.py CHANGED
@@ -36,7 +36,6 @@ def ArgsGeneralWrapper(f):
36
  llm_kwargs = {
37
  'api_key': cookies['api_key'],
38
  'llm_model': llm_model,
39
- 'endpoint': model_info[llm_model]['endpoint'],
40
  'top_p':top_p,
41
  'max_length': max_length,
42
  'temperature':temperature,
 
36
  llm_kwargs = {
37
  'api_key': cookies['api_key'],
38
  'llm_model': llm_model,
 
39
  'top_p':top_p,
40
  'max_length': max_length,
41
  'temperature':temperature,