TY - GEN
T1 - SP-cache
T2 - 2018 International Conference for High Performance Computing, Networking, Storage, and Analysis, SC 2018
AU - Yu, Yinghao
AU - Huang, Renfei
AU - Wang, Wei
AU - Zhang, Jun
AU - Letaief, Khaled Ben
PY - 2018/11/11
Y1 - 2018/11/11
N2 - Data-intensive clusters increasingly employ inmemory solutions to improve I/O performance. However, the routinely observed file popularity skew and load imbalance create hot spots, which significantly degrade the benefits of in-memory caching. Common approaches to tame load imbalance include copying multiple replicas of hot files and creating parity chunks using storage codes. Yet, these techniques either suffer from high memory overhead due to cache redundancy or incur nontrivial encoding/decoding complexity. In this paper, we propose an effective approach to achieve load balancing without cache redundancy or encoding/decoding overhead. Our solution, termed SP-Cache, selectively partitions files based on their popularity and evenly caches those partitions across the cluster. We develop an efficient algorithm to determine the optimal number of partitions for a hot file - too few partitions are incapable of mitigating hot spots, while too many are susceptible to stragglers. We implemented SP-Cache in Alluxio, a popular in-memory distributed storage for data-intensive clusters. EC2 deployment and trace-driven simulations show that, compared to the state-of-the-art solution called EC-Cache [1], SP-Cache reduces the file access latency by up to 40% in both the mean and the tail, using 40% less memory.
AB - Data-intensive clusters increasingly employ inmemory solutions to improve I/O performance. However, the routinely observed file popularity skew and load imbalance create hot spots, which significantly degrade the benefits of in-memory caching. Common approaches to tame load imbalance include copying multiple replicas of hot files and creating parity chunks using storage codes. Yet, these techniques either suffer from high memory overhead due to cache redundancy or incur nontrivial encoding/decoding complexity. In this paper, we propose an effective approach to achieve load balancing without cache redundancy or encoding/decoding overhead. Our solution, termed SP-Cache, selectively partitions files based on their popularity and evenly caches those partitions across the cluster. We develop an efficient algorithm to determine the optimal number of partitions for a hot file - too few partitions are incapable of mitigating hot spots, while too many are susceptible to stragglers. We implemented SP-Cache in Alluxio, a popular in-memory distributed storage for data-intensive clusters. EC2 deployment and trace-driven simulations show that, compared to the state-of-the-art solution called EC-Cache [1], SP-Cache reduces the file access latency by up to 40% in both the mean and the tail, using 40% less memory.
UR - http://www.scopus.com/inward/record.url?scp=85064109338&partnerID=8YFLogxK
U2 - 10.1109/SC.2018.00004
DO - 10.1109/SC.2018.00004
M3 - Conference article published in proceeding or book
AN - SCOPUS:85064109338
T3 - Proceedings - International Conference for High Performance Computing, Networking, Storage, and Analysis, SC 2018
SP - 1
EP - 13
BT - Proceedings - International Conference for High Performance Computing, Networking, Storage, and Analysis, SC 2018
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 11 November 2018 through 16 November 2018
ER -