In a few clicks we can analyze your entire application and see what components are vulnerable in your application, and suggest you quick fixes.
Test your applicationsUpgrade vllm to version 0.13.0 or higher.
vllm is an A high-throughput and memory-efficient inference and serving engine for LLMs
Affected versions of this package are vulnerable to Out-of-bounds Write via the to_dense function in the Completions API endpoint when processing user-supplied prompt embeddings. An attacker can achieve memory corruption, crash the server process, or potentially execute arbitrary code by submitting maliciously crafted serialized tensors that bypass internal bounds checks.
Note:
This issue exists due to the insufficient fix for CVE-2025-62164.
Current fix adds a flag to disable/enable prompt embeds, so by default, prompt embeds feature is disabled in vLLM, which stops DoS attacks through the embeddings. However, the issue is still present when the flag is enabled.