# For mlock support | |
ulimit -l unlimited | |
python3 -B main.py | |
#python3 -m llama_cpp.server --model ./model/gguf-model.bin --host 0.0.0.0 --port 2600 |
# For mlock support | |
ulimit -l unlimited | |
python3 -B main.py | |
#python3 -m llama_cpp.server --model ./model/gguf-model.bin --host 0.0.0.0 --port 2600 |