Hansimov commited on
Commit
0f710a2
1 Parent(s): 2fb53c8

:zap: [Enhance] Quieter openai auth, use cffi to request hf-chat id, and console token count for exceeds

Browse files
messagers/token_checker.py CHANGED
@@ -40,5 +40,7 @@ class TokenChecker:
40
 
41
  def check_token_limit(self):
42
  if self.get_token_redundancy() <= 0:
43
- raise ValueError(f"Prompt exceeded token limit: {self.get_token_limit()}")
 
 
44
  return True
 
40
 
41
  def check_token_limit(self):
42
  if self.get_token_redundancy() <= 0:
43
+ raise ValueError(
44
+ f"Prompt exceeded token limit: {self.count_tokens()} > {self.get_token_limit()}"
45
+ )
46
  return True
networks/huggingchat_streamer.py CHANGED
@@ -1,7 +1,9 @@
1
  import copy
2
  import json
3
  import re
 
4
  import requests
 
5
 
6
  from tclogger import logger
7
 
@@ -30,7 +32,7 @@ class HuggingchatRequester:
30
  request_body.update(extra_body)
31
  logger.note(f"> hf-chat ID:", end=" ")
32
 
33
- res = requests.post(
34
  request_url,
35
  headers=HUGGINGCHAT_POST_HEADERS,
36
  json=request_body,
 
1
  import copy
2
  import json
3
  import re
4
+
5
  import requests
6
+ from curl_cffi import requests as cffi_requests
7
 
8
  from tclogger import logger
9
 
 
32
  request_body.update(extra_body)
33
  logger.note(f"> hf-chat ID:", end=" ")
34
 
35
+ res = cffi_requests.post(
36
  request_url,
37
  headers=HUGGINGCHAT_POST_HEADERS,
38
  json=request_body,
networks/openai_streamer.py CHANGED
@@ -171,18 +171,21 @@ class OpenaiStreamer:
171
 
172
  def check_token_limit(self, messages: list[dict]):
173
  token_limit = TOKEN_LIMIT_MAP[self.model]
174
- token_redundancy = int(
175
- token_limit - TOKEN_RESERVED - self.count_tokens(messages)
176
- )
177
  if token_redundancy <= 0:
178
- raise ValueError(f"Prompt exceeded token limit: {token_limit}")
 
 
179
  return True
180
 
181
- def chat_response(self, messages: list[dict]):
182
  self.check_token_limit(messages)
 
183
  requester = OpenaiRequester()
184
  requester.auth()
185
- return requester.chat_completions(messages, verbose=False)
 
186
 
187
  def chat_return_generator(self, stream_response: requests.Response, verbose=False):
188
  content_offset = 0
 
171
 
172
  def check_token_limit(self, messages: list[dict]):
173
  token_limit = TOKEN_LIMIT_MAP[self.model]
174
+ token_count = self.count_tokens(messages)
175
+ token_redundancy = int(token_limit - TOKEN_RESERVED - token_count)
 
176
  if token_redundancy <= 0:
177
+ raise ValueError(
178
+ f"Prompt exceeded token limit: {token_count} > {token_limit}"
179
+ )
180
  return True
181
 
182
+ def chat_response(self, messages: list[dict], verbose=False):
183
  self.check_token_limit(messages)
184
+ logger.enter_quiet(not verbose)
185
  requester = OpenaiRequester()
186
  requester.auth()
187
+ logger.exit_quiet(not verbose)
188
+ return requester.chat_completions(messages, verbose=verbose)
189
 
190
  def chat_return_generator(self, stream_response: requests.Response, verbose=False):
191
  content_offset = 0