Spaces:
Sleeping
Sleeping
Change to gguf format model
Browse files
app.py
CHANGED
@@ -156,7 +156,7 @@ def split_docs(documents,chunk_size=500):
|
|
156 |
|
157 |
@st.cache_resource
|
158 |
def load_llama2_llamaCpp():
|
159 |
-
core_model_name = "llama-2-7b-chat.
|
160 |
n_gpu_layers = 32
|
161 |
n_batch = 512
|
162 |
callback_manager = CallbackManager([StreamingStdOutCallbackHandler()])
|
|
|
156 |
|
157 |
@st.cache_resource
|
158 |
def load_llama2_llamaCpp():
|
159 |
+
core_model_name = "llama-2-7b-chat.Q4_0.gguf"
|
160 |
n_gpu_layers = 32
|
161 |
n_batch = 512
|
162 |
callback_manager = CallbackManager([StreamingStdOutCallbackHandler()])
|