vllm load model error

#3
by Dharma0818 - opened

When I use vllm to load the gemma:2-27b-chinese-chat model, I receive the following error message. What is causing this issue?

ERROR 07-12 10:40:51 serving_chat.py:238] Error in applying chat template from request: 'system_message' is undefined
INFO:     100.89.61.1:55874 - "POST /v1/chat/completions HTTP/1.1" 400 Bad Request

Sign up or log in to comment