Spaces:
Sleeping
Sleeping
:pencil: [Doc] Update available models
Browse files
README.md
CHANGED
@@ -12,24 +12,15 @@ Huggingface LLM Inference API in OpenAI message format.
|
|
12 |
|
13 |
## Features
|
14 |
|
15 |
-
|
16 |
-
|
17 |
-
- Available Models:
|
18 |
-
- `mixtral-8x7b`, `mistral-7b`, `openchat-3.5`
|
19 |
- Adaptive prompt templates for different models
|
20 |
- Support OpenAI API format
|
21 |
-
-
|
22 |
- Support both stream and no-stream response
|
23 |
- Support API Key via both HTTP auth header and env varible (https://github.com/Hansimov/hf-llm-api/issues/4)
|
24 |
- Docker deployment
|
25 |
|
26 |
-
🔨 In progress:
|
27 |
-
- [ ] Support more models (https://github.com/Hansimov/hf-llm-api/issues/5)
|
28 |
-
- [ ] meta-llama/Llama-2-70b-chat-hf
|
29 |
-
- [ ] codellama/CodeLlama-34b-Instruct-hf
|
30 |
-
- [ ] tiiuae/falcon-180B-chat
|
31 |
-
|
32 |
-
|
33 |
## Run API service
|
34 |
|
35 |
### Run in Command Line
|
|
|
12 |
|
13 |
## Features
|
14 |
|
15 |
+
- Available Models (2024/01/15): (https://github.com/Hansimov/hf-llm-api/issues/5)
|
16 |
+
- `mixtral-8x7b`, `mistral-7b`
|
|
|
|
|
17 |
- Adaptive prompt templates for different models
|
18 |
- Support OpenAI API format
|
19 |
+
- Enable api endpoint via official `openai-python` package
|
20 |
- Support both stream and no-stream response
|
21 |
- Support API Key via both HTTP auth header and env varible (https://github.com/Hansimov/hf-llm-api/issues/4)
|
22 |
- Docker deployment
|
23 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
24 |
## Run API service
|
25 |
|
26 |
### Run in Command Line
|