-
-
Notifications
You must be signed in to change notification settings - Fork 7.8k
[Bug]: Internal Server Error: python3 openai_chat_completion_client_for_multimodal.py -c audio when using Qwen/Qwen2-Audio-7B-Instruct #19083
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Can you try setting |
Hmm... can you use the older chat template but upgrade your transformers version to the latest one? Also, try re-downloading the model repo in case it is using an incorrect chat template |
Upgraded |
I just mean the chat template which you originally used (or the default one if you used none) |
I'll take a deeper look into this when I have time (I'm busy until next week). cc @Isotr0py if you have time |
I will try to reproduce this issue locally and try to fix it. |
Still the same error |
Add |
It's working now. Looks like the issue was that OpenAI doesn't support audio? |
Your current environment
The output of
python collect_env.py
🐛 Describe the bug
The command to run the vllm service is
VLLM_USE_MODELSCOPE=True vllm serve Qwen/Qwen2-Audio-7B-Instruct --gpu_memory_utilization 0.98
.I have not made any modifications to
openai_chat_completion_client_for_multimodal.py
. When I runpython3 openai_chat_completion_client_for_multimodal.py -c audio
in thevllm/examples/online_serving
directory, the vllm service reports a '500 Internal Server Error'.At the same time, when I try to call it using curl on my own, I get the same error. Below are the screenshots of the vllm service error and the curl command.
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: