fixed bug on running openllm
Browse files- .env.example +1 -1
- Makefile +3 -0
- app_modules/llm_loader.py +1 -1
.env.example
CHANGED
@@ -7,7 +7,7 @@ LLM_MODEL_TYPE=huggingface
|
|
7 |
# LLM_MODEL_TYPE=stablelm
|
8 |
# LLM_MODEL_TYPE=openllm
|
9 |
|
10 |
-
OPENLLM_SERVER_URL=
|
11 |
|
12 |
OPENAI_API_KEY=
|
13 |
|
|
|
7 |
# LLM_MODEL_TYPE=stablelm
|
8 |
# LLM_MODEL_TYPE=openllm
|
9 |
|
10 |
+
OPENLLM_SERVER_URL=http://localhost:64300
|
11 |
|
12 |
OPENAI_API_KEY=
|
13 |
|
Makefile
CHANGED
@@ -25,6 +25,9 @@ else
|
|
25 |
openllm start llama --model-id meta-llama/Llama-2-7b-chat-hf --port=${PORT}
|
26 |
endif
|
27 |
|
|
|
|
|
|
|
28 |
ingest:
|
29 |
python ingest.py
|
30 |
|
|
|
25 |
openllm start llama --model-id meta-llama/Llama-2-7b-chat-hf --port=${PORT}
|
26 |
endif
|
27 |
|
28 |
+
openllm-cpu:
|
29 |
+
CUDA_VISIBLE_DEVICES="" openllm start llama --model-id meta-llama/Llama-2-7b-chat-hf
|
30 |
+
|
31 |
ingest:
|
32 |
python ingest.py
|
33 |
|
app_modules/llm_loader.py
CHANGED
@@ -161,7 +161,7 @@ class LLMLoader:
|
|
161 |
print(f" server url: {server_url}")
|
162 |
self.llm = OpenLLM(
|
163 |
server_url=server_url,
|
164 |
-
callbacks=callbacks,
|
165 |
verbose=True,
|
166 |
)
|
167 |
elif self.llm_model_type.startswith("gpt4all"):
|
|
|
161 |
print(f" server url: {server_url}")
|
162 |
self.llm = OpenLLM(
|
163 |
server_url=server_url,
|
164 |
+
# callbacks=callbacks,
|
165 |
verbose=True,
|
166 |
)
|
167 |
elif self.llm_model_type.startswith("gpt4all"):
|