We advise you to use the latest version of vLLM to build OpenAI-compatible API service, including tool use support. Start the server with a chat model, e.g. Qwen2.5-7B-Instruct: vllm serve Qwen/Qwen2.5-7B-Instruct Then use the chat API as demonstrated below: curl http://localhost:8000...