Spaces:
Runtime error
Runtime error
Yash Sachdeva
commited on
Commit
·
b63fd3c
1
Parent(s):
37111b2
download llama
Browse files- Dockerfile +1 -2
- question_paper.py +1 -1
Dockerfile
CHANGED
@@ -18,8 +18,7 @@ RUN pip install accelerate
|
|
18 |
|
19 |
# Install hugging face hub to download llama2 model
|
20 |
RUN pip install --upgrade huggingface_hub
|
21 |
-
RUN huggingface-cli download TheBloke/Llama-2-7b-Chat-GGUF llama-2-7b.Q4_K_M.gguf --local-dir . --local-dir-use-symlinks False
|
22 |
-
|
23 |
|
24 |
RUN CMAKE_ARGS="-DLLAMA_BLAS=ON -DLLAMA_BLAS_VENDOR=OpenBLAS" pip install 'llama-cpp-python[server]' --upgrade --force-reinstall --no-cache-dir
|
25 |
# Install requirements.txt
|
|
|
18 |
|
19 |
# Install hugging face hub to download llama2 model
|
20 |
RUN pip install --upgrade huggingface_hub
|
21 |
+
RUN huggingface-cli download TheBloke/Llama-2-7b-Chat-GGUF llama-2-7b-chat.Q4_K_M.gguf --local-dir . --local-dir-use-symlinks False
|
|
|
22 |
|
23 |
RUN CMAKE_ARGS="-DLLAMA_BLAS=ON -DLLAMA_BLAS_VENDOR=OpenBLAS" pip install 'llama-cpp-python[server]' --upgrade --force-reinstall --no-cache-dir
|
24 |
# Install requirements.txt
|
question_paper.py
CHANGED
@@ -14,7 +14,7 @@ app = FastAPI()
|
|
14 |
@app.get("/")
|
15 |
def llama():
|
16 |
llm = Llama(
|
17 |
-
model_path="./llama-2-7b.Q4_K_M.gguf",
|
18 |
# n_gpu_layers=-1, # Uncomment to use GPU acceleration
|
19 |
# seed=1337, # Uncomment to set a specific seed
|
20 |
# n_ctx=2048, # Uncomment to increase the context window
|
|
|
14 |
@app.get("/")
|
15 |
def llama():
|
16 |
llm = Llama(
|
17 |
+
model_path="./llama-2-7b-chat.Q4_K_M.gguf",
|
18 |
# n_gpu_layers=-1, # Uncomment to use GPU acceleration
|
19 |
# seed=1337, # Uncomment to set a specific seed
|
20 |
# n_ctx=2048, # Uncomment to increase the context window
|