VLLM 0.7.2 can start the model normally, but there is no output when simulating a request using Curl, it blocks!
1
#2 opened 19 days ago
by
JZMALi
sglang inference issue
7
#1 opened 21 days ago
by
su400