There is a huge gap between the speeds of modern caches and main memories, and therefore cache misses account for a considerable loss of efficiency in programs. The predominant technique to address this issue has been Data Packing: data elements that are frequently accessed within time proximity are packed into the same cache block, thereby minimizing accesses to the main memory. We consider the algorithmic problem of Data Packing on a two-level memory system. Given a reference sequence R of accesses to data elements, the task is to partition the elements into cache blocks such that the number of cache misses on R is minimized. The problem is notoriously difficult: it is NP-hard even when the cache has size 1, and is hard to approximate for any cache size larger than 4. Therefore, all existing techniques for Data Packing are based on heuristics and lack theoretical guarantees. In this work, we present the first positive theoretical results for Data Packing, along with new and stronger negative results. We consider the problem under the lens of the underlying access hypergraphs, which are hypergraphs of affinities between the data elements, where the order of an access hypergraph corresponds to the size of the affinity group. We study the problem parameterized by the treewidth of access hypergraphs, which is a standard notion in graph theory to measure the closeness of a graph to a tree. Our main results are as follows: We show there is a number q* depending on the cache parameters such that (a) if the access hypergraph of order q* has constant treewidth, then there is a linear-time algorithm for Data Packing; (b)the Data Packing problem remains NP-hard even if the access hypergraph of order q*-1 has constant treewidth. Thus, we establish a fine-grained dichotomy depending on a single parameter, namely, the highest order among access hypegraphs that have constant treewidth; and establish the optimal value q* of this parameter. Finally, we present an experimental evaluation of a prototype implementation of our algorithm. Our results demonstrate that, in practice, access hypergraphs of many commonly-used algorithms have small treewidth. We compare our approach with several state-of-the-art heuristic-based algorithms and show that our algorithm leads to significantly fewer cache-misses.
Proceedings of the ACM on Programming Languages
Chatterjee K, Goharshady AK, Okati N, Pavlogiannis A. Efficient parameterized algorithms for data packing. Proceedings of the ACM on Programming Languages. 2019;3(POPL). doi:10.1145/3290366
Chatterjee, K., Goharshady, A. K., Okati, N., & Pavlogiannis, A. (2019). Efficient parameterized algorithms for data packing. Proceedings of the ACM on Programming Languages. ACM. https://doi.org/10.1145/3290366
Chatterjee, Krishnendu, Amir Kafshdar Goharshady, Nastaran Okati, and Andreas Pavlogiannis. “Efficient Parameterized Algorithms for Data Packing.” Proceedings of the ACM on Programming Languages. ACM, 2019. https://doi.org/10.1145/3290366.
K. Chatterjee, A. K. Goharshady, N. Okati, and A. Pavlogiannis, “Efficient parameterized algorithms for data packing,” Proceedings of the ACM on Programming Languages, vol. 3, no. POPL. ACM, 2019.
Chatterjee K, Goharshady AK, Okati N, Pavlogiannis A. 2019. Efficient parameterized algorithms for data packing. Proceedings of the ACM on Programming Languages. 3(POPL), 53.
Chatterjee, Krishnendu, et al. “Efficient Parameterized Algorithms for Data Packing.” Proceedings of the ACM on Programming Languages, vol. 3, no. POPL, 53, ACM, 2019, doi:10.1145/3290366.
All files available under the following license(s):
Creative Commons Attribution 4.0 International Public License (CC-BY 4.0):
2019_ACM_POPL_Chatterjee.pdf 1.29 MB
Material in IST:
Dissertation containing IST record