TY - JOUR
T1 - ELF
T2 - shared Cache management through eliminating dead blocks and filtering less reused lines
AU - Sui, Xiu Feng
AU - Wu, Jun Min
AU - Chen, Guo Liang
AU - Tang, Yi Xuan
PY - 2011/1
Y1 - 2011/1
N2 - Modern CMP processors usually employ shared last-level cache (LLC) based on LRU replacement policy and its approximations. However, as the LLC grows in capacity and associativity, the performance gap between the LRU and the theoretical optimal replacement algorithms has widened. Various alternative cache management technologies have been proposed to resolve this problem, but they only cover a single type of memory access behavior, and exploit little frequency information of cache accesses, thus have limited performance benefits. In this paper, we propose a unified cache management policy ELF which can cover a variety of memory behaviors and exploit both recency and frequency information of a program simultaneously. Motivatedbythe observation that cache blocks often exhibit a small number of uses during their life time in the LLC, ELF is designed to (1) predict dead lines through a counter-based mechanism and evict them early, (2) filter less reused blocks through dynamic insertion and promotion policies. Thereby, the potentially live blocks are retained and most of the working set keeps undisturbed in the ELF managed L2 cache. Our evaluation on 4-way CMPs shows that ELF improves the overall performance by 14.5% on average over the LRU policy, and the performance benefit of ELF is 1.06x compared to PIPP and 1.09x compared to TADIP.
AB - Modern CMP processors usually employ shared last-level cache (LLC) based on LRU replacement policy and its approximations. However, as the LLC grows in capacity and associativity, the performance gap between the LRU and the theoretical optimal replacement algorithms has widened. Various alternative cache management technologies have been proposed to resolve this problem, but they only cover a single type of memory access behavior, and exploit little frequency information of cache accesses, thus have limited performance benefits. In this paper, we propose a unified cache management policy ELF which can cover a variety of memory behaviors and exploit both recency and frequency information of a program simultaneously. Motivatedbythe observation that cache blocks often exhibit a small number of uses during their life time in the LLC, ELF is designed to (1) predict dead lines through a counter-based mechanism and evict them early, (2) filter less reused blocks through dynamic insertion and promotion policies. Thereby, the potentially live blocks are retained and most of the working set keeps undisturbed in the ELF managed L2 cache. Our evaluation on 4-way CMPs shows that ELF improves the overall performance by 14.5% on average over the LRU policy, and the performance benefit of ELF is 1.06x compared to PIPP and 1.09x compared to TADIP.
KW - Counter-based algorithms
KW - Insertion policy
KW - Multi-core
KW - Replacement algorithms
KW - Shared cache
UR - http://www.scopus.com/inward/record.url?scp=79953110685&partnerID=8YFLogxK
U2 - 10.3724/SP.J.1016.2011.00143
DO - 10.3724/SP.J.1016.2011.00143
M3 - Article
AN - SCOPUS:79953110685
SN - 0254-4164
VL - 34
SP - 143
EP - 153
JO - Jisuanji Xuebao/Chinese Journal of Computers
JF - Jisuanji Xuebao/Chinese Journal of Computers
IS - 1
ER -