TY - GEN
T1 - Aurora
T2 - 35th IEEE International Conference on Distributed Computing Systems, ICDCS 2015
AU - Zhang, Qi
AU - Zhang, Sai Qian
AU - Leon-Garcia, Alberto
AU - Boutaba, Raouf
N1 - Publisher Copyright:
© 2015 IEEE.
PY - 2015/7/22
Y1 - 2015/7/22
N2 - Distributed file systems such as Google File System and Hadoop Distributed File System have been used to store large volumes of data in Cloud data centers. These systems divide data sets in blocks of fixed size and replicate them over multiple machines to achieve both reliability and efficiency. Recent studies have shown that data blocks tend to have a wide disparity in data popularity. In this context, the naive block replication schemes used by these systems often cause an uneven load distribution across machines, which reduces the overall I/O throughput of the system. While many replication algorithms have been proposed, existing solutions have not carefully studied the placement of data blocks that balances the load across machines, while ensuring node and rack-level reliability requirements are satisfied. In this paper, we study the dynamic data replication problem with the goal of balancing machine load while ensuring machine and rack-level reliability requirements are met. We propose several local search algorithms that provide constant approximation guarantees, yet simple and practical for implementation. We further present Aurora, a dynamic block placement mechanism that implements these algorithms in the Hadoop Distributed File System with minimal overhead. Through experiments using workload traces from Yahoo! and Facebook, we show Aurora reduces machine load imbalance by up to 26.9% compared to existing solutions, while satisfying node and rack-level reliability requirements.
AB - Distributed file systems such as Google File System and Hadoop Distributed File System have been used to store large volumes of data in Cloud data centers. These systems divide data sets in blocks of fixed size and replicate them over multiple machines to achieve both reliability and efficiency. Recent studies have shown that data blocks tend to have a wide disparity in data popularity. In this context, the naive block replication schemes used by these systems often cause an uneven load distribution across machines, which reduces the overall I/O throughput of the system. While many replication algorithms have been proposed, existing solutions have not carefully studied the placement of data blocks that balances the load across machines, while ensuring node and rack-level reliability requirements are satisfied. In this paper, we study the dynamic data replication problem with the goal of balancing machine load while ensuring machine and rack-level reliability requirements are met. We propose several local search algorithms that provide constant approximation guarantees, yet simple and practical for implementation. We further present Aurora, a dynamic block placement mechanism that implements these algorithms in the Hadoop Distributed File System with minimal overhead. Through experiments using workload traces from Yahoo! and Facebook, we show Aurora reduces machine load imbalance by up to 26.9% compared to existing solutions, while satisfying node and rack-level reliability requirements.
KW - approximation algorithms
KW - Distributed file system
KW - Hadoop
KW - HDFS
KW - local search
UR - http://www.scopus.com/inward/record.url?scp=84944315855&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84944315855&partnerID=8YFLogxK
U2 - 10.1109/ICDCS.2015.52
DO - 10.1109/ICDCS.2015.52
M3 - Conference contribution
AN - SCOPUS:84944315855
T3 - Proceedings - International Conference on Distributed Computing Systems
SP - 442
EP - 451
BT - Proceedings - 2015 IEEE 35th International Conference on Distributed Computing Systems, ICDCS 2015
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 29 June 2015 through 2 July 2015
ER -