vLLM is an inference and serving engine for large language models (LLMs). Prior to version 0.9.0, when a new prompt is processed, if the PageAttention mechanism finds a matching prefix chunk, the prefill process speeds up, which is reflected in the TTFT (Time to First Token). These timing differences caused by matching chunks are significant enough to be recognized and exploited. This issue has been patched in version 0.9.0.
History

Fri, 30 May 2025 21:45:00 +0000

Type Values Removed Values Added
References
Metrics threat_severity

None

threat_severity

Low


Thu, 29 May 2025 18:15:00 +0000

Type Values Removed Values Added
Metrics ssvc

{'options': {'Automatable': 'no', 'Exploitation': 'none', 'Technical Impact': 'partial'}, 'version': '2.0.3'}


Thu, 29 May 2025 16:45:00 +0000

Type Values Removed Values Added
Description vLLM is an inference and serving engine for large language models (LLMs). Prior to version 0.9.0, when a new prompt is processed, if the PageAttention mechanism finds a matching prefix chunk, the prefill process speeds up, which is reflected in the TTFT (Time to First Token). These timing differences caused by matching chunks are significant enough to be recognized and exploited. This issue has been patched in version 0.9.0.
Title vLLM’s Chunk-Based Prefix Caching Vulnerable to Potential Timing Side-Channel
Weaknesses CWE-208
References
Metrics cvssV3_1

{'score': 2.6, 'vector': 'CVSS:3.1/AV:N/AC:H/PR:L/UI:R/S:U/C:L/I:N/A:N'}


cve-icon MITRE

Status: PUBLISHED

Assigner: GitHub_M

Published: 2025-05-29T16:32:42.794Z

Updated: 2025-05-29T18:05:10.768Z

Reserved: 2025-04-24T21:10:48.175Z

Link: CVE-2025-46570

cve-icon Vulnrichment

Updated: 2025-05-29T18:05:04.545Z

cve-icon NVD

Status : Awaiting Analysis

Published: 2025-05-29T17:15:21.327

Modified: 2025-05-30T16:31:03.107

Link: CVE-2025-46570

cve-icon Redhat

Severity : Low

Publid Date: 2025-05-29T16:32:42Z

Links: CVE-2025-46570 - Bugzilla