Update app.py
Browse files
app.py
CHANGED
@@ -18,20 +18,23 @@ def install_llama_with_cuda():
|
|
18 |
with open("requirements.txt", "r") as f:
|
19 |
packages = f.read().splitlines()
|
20 |
|
21 |
-
# Install setiap paket dengan CMAKE_ARGS untuk dukungan CUDA
|
22 |
-
for package in packages:
|
23 |
-
subprocess.run(
|
24 |
-
env={"CMAKE_ARGS": "-DGGML_CUDA=on"},
|
25 |
-
check=True
|
26 |
-
)
|
27 |
# Periksa apakah CUDA Toolkit tersedia
|
28 |
if not shutil.which("nvcc"):
|
29 |
print("CUDA Toolkit tidak ditemukan. Pastikan sudah diinstal.")
|
30 |
return
|
31 |
|
32 |
print("Memasang ulang llama-cpp-python dengan dukungan CUDA...")
|
33 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
34 |
print("llama-cpp-python berhasil diinstal ulang dengan dukungan CUDA.")
|
|
|
35 |
except subprocess.CalledProcessError as e:
|
36 |
print(f"Error saat menginstal ulang llama-cpp-python: {e}")
|
37 |
except Exception as e:
|
|
|
18 |
with open("requirements.txt", "r") as f:
|
19 |
packages = f.read().splitlines()
|
20 |
|
|
|
|
|
|
|
|
|
|
|
|
|
21 |
# Periksa apakah CUDA Toolkit tersedia
|
22 |
if not shutil.which("nvcc"):
|
23 |
print("CUDA Toolkit tidak ditemukan. Pastikan sudah diinstal.")
|
24 |
return
|
25 |
|
26 |
print("Memasang ulang llama-cpp-python dengan dukungan CUDA...")
|
27 |
+
|
28 |
+
# Install setiap paket dengan CMAKE_ARGS untuk dukungan CUDA
|
29 |
+
for package in packages:
|
30 |
+
subprocess.run(
|
31 |
+
["pip", "install", package], # <-- Tambahkan perintah pip install
|
32 |
+
env={**dict(subprocess.os.environ), "CMAKE_ARGS": "-DGGML_CUDA=on"},
|
33 |
+
check=True
|
34 |
+
)
|
35 |
+
|
36 |
print("llama-cpp-python berhasil diinstal ulang dengan dukungan CUDA.")
|
37 |
+
|
38 |
except subprocess.CalledProcessError as e:
|
39 |
print(f"Error saat menginstal ulang llama-cpp-python: {e}")
|
40 |
except Exception as e:
|