hf-llm-api / networks

Commit History

:gem: [Feature] Split to Requester and Streamer, and mock chat history with messages
c43287d

Hansimov commited on

:gem: [Feature] Enable chat response with get_message_id
e4d11b8

Hansimov commited on

:hammer: [WIP] Enabling get_message_id
a5ac953

Hansimov commited on

:gem: [Feature] HuggingchatStreamer: Build pipeline of chat_response
55b0c51

Hansimov commited on

:gem: [Feature] HuggingchatStreamer: New log request and response
c706328

Hansimov commited on

:gem: [Feature] HuggingchatStreamer: New get_hf_chat_id, and improve get_conversation_id
391cdfe

Hansimov commited on

:gem: [Feature] New HuggingchatStreamer: get_conversation_id
f1218fc

Hansimov commited on

:gem: [Feature] Enable non-stream return, and prettify logger
397529a

Hansimov commited on

:boom: [Fix] Set default model
5dc8ebf

Hansimov commited on

:recycle: [Refactor] Rename gpt-3.5 to gpt-3.5-turbo
d36d623

Hansimov commited on

:gem: [Feature] Enable gpt-3.5 in chat_api
cd6b52a

Hansimov commited on

:recycle: [Refactor] Rename message_streamer to huggingface_streamer
3125c87

Hansimov commited on

:recycle: [Refactor] Use tclogger, and configure envs better
babcd78

Hansimov commited on

:recycle: [Refactor] Move STOP_SEQUENCES_MAP and TOKEN_LIMIT_MAP to constants
bf8c5bd

Hansimov commited on

:recycle: [Refactor] Move MODELS_MAP to constants
3f608c6

Hansimov commited on

:gem: [Feature] Support new model: openchat-3.5-0106
9f5d69c

Hansimov commited on

:gem: [Feature] New model enabled: gemma-7b
77b5a47

Hansimov commited on

:zap: [Enhance] Set min and max for top_p
64645f0

Hansimov commited on

:gem: [Feature] Add top_p and set 0.95 as default
403b8cf

Hansimov commited on

:gem: [Feature] Add use_cache parameter, and set default temperature to 0.5
e2b245b

Hansimov commited on

:gem: [Feature] New model enabled: nous-mixtral-8-7b
4a28ca5

Hansimov commited on

:zap: [Enhance] Remove openchat-3.5 from available models
06c8d58

Hansimov commited on

:boom: [Fix] inputs_tokens + max_new_tokens must <= 32768
fa3bfb0

Hansimov commited on

:boom: [Fix] temperature must be strictly positive
eb00725

Hansimov commited on

:boom: [Fix] Error 500 when temperature is null
8ba223c

Hansimov commited on

:zap: [Enhance] Auto calculate max_tokens if not set
1b9f698

Hansimov commited on

:zap: [Enhance] Remove leading whitespaces of first response
c95d47e

Hansimov commited on

:pencil: [Doc] Available models: More comments on dropped models
85354a9

Hansimov commited on

:gem: [Feature] MessageStreamer: Support stop sequence and auto removal
4ba2ca6

Hansimov commited on

:gem: [Feature] Support call hf api with api_key via HTTP Bearer
2da6968

Hansimov commited on

:gem: [Feature] Support no-stream mode with dict response
d2b20f2

Hansimov commited on

:pencil: [Doc] Readme for features, and comments for references
e820e51

Hansimov commited on

:zap: [Enhance] Cleaner logger output
b96cef7

Hansimov commited on

:zap: [Enhance] New models: mistral-7b and openchat-3.5
e916990

Hansimov commited on

:zap: [Enhance] OSLogger: Re-wrap for better syntax check, and move stream log to debug
047008b

Hansimov commited on

:pencil: [Comment] References of requests params
395ee29

Hansimov commited on

:zap: [Enhance] Default models and max_new_tokens
6aa8b86

Hansimov commited on

:zap: [Enhance] API: Ensure temperature to be positive
3dfcb72

Hansimov commited on

:gem: [Feature] New ArgParser: Specify server ip, port and running mode
e28221f

Hansimov commited on

:gem: [Feature] New MessageStreamer: Enable requests inference api with requests
9f341cc

Hansimov commited on