inflaton commited on
Commit
ea50055
·
1 Parent(s): a467fc0

fixed bug on running openllm

Browse files
Files changed (3) hide show
  1. .env.example +1 -1
  2. Makefile +3 -0
  3. app_modules/llm_loader.py +1 -1
.env.example CHANGED
@@ -7,7 +7,7 @@ LLM_MODEL_TYPE=huggingface
7
  # LLM_MODEL_TYPE=stablelm
8
  # LLM_MODEL_TYPE=openllm
9
 
10
- OPENLLM_SERVER_URL=
11
 
12
  OPENAI_API_KEY=
13
 
 
7
  # LLM_MODEL_TYPE=stablelm
8
  # LLM_MODEL_TYPE=openllm
9
 
10
+ OPENLLM_SERVER_URL=http://localhost:64300
11
 
12
  OPENAI_API_KEY=
13
 
Makefile CHANGED
@@ -25,6 +25,9 @@ else
25
  openllm start llama --model-id meta-llama/Llama-2-7b-chat-hf --port=${PORT}
26
  endif
27
 
 
 
 
28
  ingest:
29
  python ingest.py
30
 
 
25
  openllm start llama --model-id meta-llama/Llama-2-7b-chat-hf --port=${PORT}
26
  endif
27
 
28
+ openllm-cpu:
29
+ CUDA_VISIBLE_DEVICES="" openllm start llama --model-id meta-llama/Llama-2-7b-chat-hf
30
+
31
  ingest:
32
  python ingest.py
33
 
app_modules/llm_loader.py CHANGED
@@ -161,7 +161,7 @@ class LLMLoader:
161
  print(f" server url: {server_url}")
162
  self.llm = OpenLLM(
163
  server_url=server_url,
164
- callbacks=callbacks,
165
  verbose=True,
166
  )
167
  elif self.llm_model_type.startswith("gpt4all"):
 
161
  print(f" server url: {server_url}")
162
  self.llm = OpenLLM(
163
  server_url=server_url,
164
+ # callbacks=callbacks,
165
  verbose=True,
166
  )
167
  elif self.llm_model_type.startswith("gpt4all"):