Yash Sachdeva commited on
Commit
b63fd3c
·
1 Parent(s): 37111b2

download llama

Browse files
Files changed (2) hide show
  1. Dockerfile +1 -2
  2. question_paper.py +1 -1
Dockerfile CHANGED
@@ -18,8 +18,7 @@ RUN pip install accelerate
18
 
19
  # Install hugging face hub to download llama2 model
20
  RUN pip install --upgrade huggingface_hub
21
- RUN huggingface-cli download TheBloke/Llama-2-7b-Chat-GGUF llama-2-7b.Q4_K_M.gguf --local-dir . --local-dir-use-symlinks False
22
-
23
 
24
  RUN CMAKE_ARGS="-DLLAMA_BLAS=ON -DLLAMA_BLAS_VENDOR=OpenBLAS" pip install 'llama-cpp-python[server]' --upgrade --force-reinstall --no-cache-dir
25
  # Install requirements.txt
 
18
 
19
  # Install hugging face hub to download llama2 model
20
  RUN pip install --upgrade huggingface_hub
21
+ RUN huggingface-cli download TheBloke/Llama-2-7b-Chat-GGUF llama-2-7b-chat.Q4_K_M.gguf --local-dir . --local-dir-use-symlinks False
 
22
 
23
  RUN CMAKE_ARGS="-DLLAMA_BLAS=ON -DLLAMA_BLAS_VENDOR=OpenBLAS" pip install 'llama-cpp-python[server]' --upgrade --force-reinstall --no-cache-dir
24
  # Install requirements.txt
question_paper.py CHANGED
@@ -14,7 +14,7 @@ app = FastAPI()
14
  @app.get("/")
15
  def llama():
16
  llm = Llama(
17
- model_path="./llama-2-7b.Q4_K_M.gguf",
18
  # n_gpu_layers=-1, # Uncomment to use GPU acceleration
19
  # seed=1337, # Uncomment to set a specific seed
20
  # n_ctx=2048, # Uncomment to increase the context window
 
14
  @app.get("/")
15
  def llama():
16
  llm = Llama(
17
+ model_path="./llama-2-7b-chat.Q4_K_M.gguf",
18
  # n_gpu_layers=-1, # Uncomment to use GPU acceleration
19
  # seed=1337, # Uncomment to set a specific seed
20
  # n_ctx=2048, # Uncomment to increase the context window