First published: Thu Sep 12 2024(Updated: )
A completions API request with an empty prompt will crash the vllm API server. The impact is limited based on what model is being served. * Serving gpt2 is affected. * Most models are not affected, as vllm will prepend tokens to the prompt, avoiding the problematic code. <a href="https://github.com/vllm-project/vllm/commit/e25fee57c2e69161bd261f5986dc5aeb198bbd42">https://github.com/vllm-project/vllm/commit/e25fee57c2e69161bd261f5986dc5aeb198bbd42</a>
Affected Software | Affected Version | How to fix |
---|---|---|
vLLM | ||
GPT-2 |
Sign up to SecAlerts for real-time vulnerability data matched to your software, aggregated from hundreds of sources.
The severity of REDHAT-BUG-2311895 is considered moderate, as it causes the vllm API server to crash with an empty prompt but primarily affects the GPT-2 model.
To fix REDHAT-BUG-2311895, ensure that prompts sent to the vllm API server are not empty, or update to the latest version of the vllm software that addresses this issue.
Only the GPT-2 model is directly affected by REDHAT-BUG-2311895; most other models are not impacted due to token prepending.
Making a completions API request with an empty prompt will crash the vllm API server, leading to potential service downtime.
Yes, the impact of REDHAT-BUG-2311895 is dependent on the model being served, with GPT-2 being affected and other models generally being safe.