hf-llm-api / networks /message_streamer.py

Commit History

:gem: [Feature] New model enabled: nous-mixtral-8-7b
4a28ca5

Hansimov commited on

:zap: [Enhance] Remove openchat-3.5 from available models
06c8d58

Hansimov commited on

:boom: [Fix] inputs_tokens + max_new_tokens must <= 32768
fa3bfb0

Hansimov commited on

:boom: [Fix] temperature must be strictly positive
eb00725

Hansimov commited on

:boom: [Fix] Error 500 when temperature is null
8ba223c

Hansimov commited on

:zap: [Enhance] Auto calculate max_tokens if not set
1b9f698

Hansimov commited on

:zap: [Enhance] Remove leading whitespaces of first response
c95d47e

Hansimov commited on

:pencil: [Doc] Available models: More comments on dropped models
85354a9

Hansimov commited on

:gem: [Feature] MessageStreamer: Support stop sequence and auto removal
4ba2ca6

Hansimov commited on

:gem: [Feature] Support call hf api with api_key via HTTP Bearer
2da6968

Hansimov commited on

:gem: [Feature] Support no-stream mode with dict response
d2b20f2

Hansimov commited on

:pencil: [Doc] Readme for features, and comments for references
e820e51

Hansimov commited on

:zap: [Enhance] Cleaner logger output
b96cef7

Hansimov commited on

:zap: [Enhance] New models: mistral-7b and openchat-3.5
e916990

Hansimov commited on

:zap: [Enhance] OSLogger: Re-wrap for better syntax check, and move stream log to debug
047008b

Hansimov commited on

:pencil: [Comment] References of requests params
395ee29

Hansimov commited on

:zap: [Enhance] Default models and max_new_tokens
6aa8b86

Hansimov commited on

:zap: [Enhance] API: Ensure temperature to be positive
3dfcb72

Hansimov commited on

:gem: [Feature] New ArgParser: Specify server ip, port and running mode
e28221f

Hansimov commited on

:gem: [Feature] New MessageStreamer: Enable requests inference api with requests
9f341cc

Hansimov commited on