Moses25's picture
Update run.sh
10ca145 verified
raw
history blame
542 Bytes
#!/bin/bash
git clone https://huggingface.co/Moses25/Mistral-7B-Instruct-32K-AWQ
git clone https://github.com/vllm-project/vllm.git
cd vllm && VLLM_TARGET_DEVICE=cpu python setup.py install && cd ..
python -m vllm.entrypoints.openai.api_server --model=Mistral-7B-Instruct-32K-AWQ \
--trust-remote-code --host 0.0.0.0 --port 7777 \
--gpu-memory-utilization 0.8 \
--enforce-eager \
--max-model-len 8192 --chat-template llama2-chat-template.jinja \
--tensor-parallel-size 1 --served-model-name dewu-chat