Home

Description

The flash-attention training framework thru commit e724e2588cbe754beb97cf7c011b5e7e34119e62 (2025-13-04) contains an insecure deserialization vulnerability (CWE-502) in its checkpoint loading mechanism. The load_checkpoint() function in checkpoint.py and the checkpoint loading code in eval.py use torch.load() without enabling the security-restrictive weights_only=True parameter. This allows the deserialization of arbitrary Python objects via the pickle module. An attacker can exploit this by providing a maliciously crafted checkpoint file. When a victim loads this checkpoint during model warmstarting or evaluation, arbitrary code is executed on the victim's system.

PUBLISHED Reserved 2026-03-09 | Published 2026-05-11 | Updated 2026-05-11 | Assigner mitre

References

github.com/Dao-AILab/flash-attention

www.notion.so/...2026-31253-35d1e1393188813f9e77e2038104bc49

cve.org (CVE-2026-31253)

nvd.nist.gov (CVE-2026-31253)

Download JSON