Intelligent Data Placement Mechanism For Replicas Distribution In Cloud Storage Systems

Keywords

Cloud Computing; Data Replication; Hadoop; Hadoop Distributed File System; Load Balance; Replica Placement

Abstract

The Hadoop Distributed File System (HDFS) is a distributed storage system that stores large volumes of data reliably and provide access to the data by the applications at high bandwidth. HDFS provides high reliability and availability by replicating data, typically three copies, and distribute these replicas across multiple data nodes. The placement of data replicas is one of the key issues that affect the performance of HDFS. In the current HDFS replica placement policy the replicas of data blocks cannot be evenly distribute across cluster nodes, so the current HDFS has to rely on load balancing utility to balance replica distributions which results in more time and resources consuming. These challenges drive the need for intelligent methods that solve the data placement problem to achieve high performance without the need for load balancing utility. In this paper, we propose an intelligent policy for data placement in cloud storage systems addressing the above challenges.

Publication Date

12-22-2016

Publication Title

Proceedings - 2016 IEEE International Conference on Smart Cloud, SmartCloud 2016

Number of Pages

134-139

Document Type

Article; Proceedings Paper

Personal Identifier

scopus

DOI Link

https://doi.org/10.1109/SmartCloud.2016.23

Socpus ID

85011103407 (Scopus)

Source API URL

https://api.elsevier.com/content/abstract/scopus_id/85011103407

This document is currently not available here.

Share

COinS