Skip to Main content Skip to Navigation
Journal articles

Fast and exact analysis for LRU caches

Abstract : For applications in worst-case execution time analysis and in security, it is desirable to statically classify memory accesses into those that result in cache hits, and those that result in cache misses. Among cache replacement policies, the least recently used (LRU) policy has been studied the most and is considered to be the most predictable. The state-of-the-art in LRU cache analysis presents a tradeoff between precision and analysis efficiency: The classical approach to analyzing programs running on LRU caches, an abstract interpretation based on a range abstraction, is very fast but can be imprecise. An exact analysis was recently presented, but, as a last resort, it calls a model checker, which is expensive. In this paper, we develop an analysis based on abstract interpretation that comes close to the efficiency of the classical approach, while achieving exact classification of all memory accesses as the model-checking approach. Compared with the model-checking approach we observe speedups of several orders of magnitude. As a secondary contribution we show that LRU cache analysis problems are in general NP-complete.
Complete list of metadata
Contributor : Valentin Touzeau Connect in order to contact the contributor
Submitted on : Tuesday, December 18, 2018 - 4:02:53 PM
Last modification on : Wednesday, November 3, 2021 - 4:29:43 AM
Long-term archiving on: : Wednesday, March 20, 2019 - 12:25:45 PM


Files produced by the author(s)




Claire Maïza, Valentin Touzeau, David Monniaux, Jan Reineke. Fast and exact analysis for LRU caches. Proceedings of the ACM on Programming Languages, ACM, 2019, 3 (POPL), ⟨10.1145/3290367⟩. ⟨hal-01908648v2⟩



Record views


Files downloads