Intelligent Data Placement Mechanism For Replicas Distribution In Cloud Storage Systems
Keywords
Cloud Computing; Data Replication; Hadoop; Hadoop Distributed File System; Load Balance; Replica Placement
Abstract
The Hadoop Distributed File System (HDFS) is a distributed storage system that stores large volumes of data reliably and provide access to the data by the applications at high bandwidth. HDFS provides high reliability and availability by replicating data, typically three copies, and distribute these replicas across multiple data nodes. The placement of data replicas is one of the key issues that affect the performance of HDFS. In the current HDFS replica placement policy the replicas of data blocks cannot be evenly distribute across cluster nodes, so the current HDFS has to rely on load balancing utility to balance replica distributions which results in more time and resources consuming. These challenges drive the need for intelligent methods that solve the data placement problem to achieve high performance without the need for load balancing utility. In this paper, we propose an intelligent policy for data placement in cloud storage systems addressing the above challenges.
Publication Date
12-22-2016
Publication Title
Proceedings - 2016 IEEE International Conference on Smart Cloud, SmartCloud 2016
Number of Pages
134-139
Document Type
Article; Proceedings Paper
Personal Identifier
scopus
DOI Link
https://doi.org/10.1109/SmartCloud.2016.23
Copyright Status
Unknown
Socpus ID
85011103407 (Scopus)
Source API URL
https://api.elsevier.com/content/abstract/scopus_id/85011103407
STARS Citation
Ibrahim, Ibrahim Adel; Dai, Wei; and Bassiouni, Mustafa, "Intelligent Data Placement Mechanism For Replicas Distribution In Cloud Storage Systems" (2016). Scopus Export 2015-2019. 4241.
https://stars.library.ucf.edu/scopus2015/4241