2.6
LOW
CVE-2025-46570
Apache vLLM PageAttention Chunk Prefill Timing Vulnerability
Description

vLLM is an inference and serving engine for large language models (LLMs). Prior to version 0.9.0, when a new prompt is processed, if the PageAttention mechanism finds a matching prefix chunk, the prefill process speeds up, which is reflected in the TTFT (Time to First Token). These timing differences caused by matching chunks are significant enough to be recognized and exploited. This issue has been patched in version 0.9.0.

INFO

Published Date :

May 29, 2025, 5:15 p.m.

Last Modified :

May 30, 2025, 4:31 p.m.

Remotely Exploitable :

Yes !

Impact Score :

1.4

Exploitability Score :

1.2
Affected Products

The following products are affected by CVE-2025-46570 vulnerability. Even if cvefeed.io is aware of the exact versions of the products that are affected, the information is not represented in the table below.

ID Vendor Product Action
1 Vllm vllm
References to Advisories, Solutions, and Tools

Here, you will find a curated list of external links that provide in-depth information, practical solutions, and valuable tools related to CVE-2025-46570.

URL Resource
https://github.com/vllm-project/vllm/commit/77073c77bc2006eb80ea6d5128f076f5e6c6f54f
https://github.com/vllm-project/vllm/pull/17045
https://github.com/vllm-project/vllm/security/advisories/GHSA-4qjh-9fv9-r85r

We scan GitHub repositories to detect new proof-of-concept exploits. Following list is a collection of public exploits and proof-of-concepts, which have been published on GitHub (sorted by the most recently updated).

Results are limited to the first 15 repositories due to potential performance issues.

The following list is the news that have been mention CVE-2025-46570 vulnerability anywhere in the article.

The following table lists the changes that have been made to the CVE-2025-46570 vulnerability over time.

Vulnerability history details can be useful for understanding the evolution of a vulnerability, and for identifying the most recent changes that may impact the vulnerability's severity, exploitability, or other characteristics.

  • New CVE Received by [email protected]

    May. 29, 2025

    Action Type Old Value New Value
    Added Description vLLM is an inference and serving engine for large language models (LLMs). Prior to version 0.9.0, when a new prompt is processed, if the PageAttention mechanism finds a matching prefix chunk, the prefill process speeds up, which is reflected in the TTFT (Time to First Token). These timing differences caused by matching chunks are significant enough to be recognized and exploited. This issue has been patched in version 0.9.0.
    Added CVSS V3.1 AV:N/AC:H/PR:L/UI:R/S:U/C:L/I:N/A:N
    Added CWE CWE-208
    Added Reference https://github.com/vllm-project/vllm/commit/77073c77bc2006eb80ea6d5128f076f5e6c6f54f
    Added Reference https://github.com/vllm-project/vllm/pull/17045
    Added Reference https://github.com/vllm-project/vllm/security/advisories/GHSA-4qjh-9fv9-r85r
EPSS is a daily estimate of the probability of exploitation activity being observed over the next 30 days. Following chart shows the EPSS score history of the vulnerability.
CWE - Common Weakness Enumeration

While CVE identifies specific instances of vulnerabilities, CWE categorizes the common flaws or weaknesses that can lead to vulnerabilities. CVE-2025-46570 is associated with the following CWEs:

Common Attack Pattern Enumeration and Classification (CAPEC)

Common Attack Pattern Enumeration and Classification (CAPEC) stores attack patterns, which are descriptions of the common attributes and approaches employed by adversaries to exploit the CVE-2025-46570 weaknesses.

CVSS31 - Vulnerability Scoring System
Attack Vector
Attack Complexity
Privileges Required
User Interaction
Scope
Confidentiality
Integrity
Availability
© cvefeed.io
Latest DB Update: Jun. 01, 2025 13:01