Abstract

A file system is used for the organization, storage, retrieval, naming, sharing, and protection of files. Distributed file system has certain degrees of transparency to the user and the system such as access transparency, [2] location transparency, failure transparency, heterogeneity, replication transparency etc. [1][2] NFS (Network File System), RFS (Remote File Sharing), Andrew File System (AFS) are examples of Distributed file system. Distributed file systems are generally used for cloud computing applications based on [4] the MapReduce programming model. A MapReduce program consist of a Map () procedure that performs filtering and a Reduce () procedure that performs a summary operation. However, in a cloud computing environment, sometimes failure is occurs and nodes may be upgraded, replaced, and added in the system. Therefore load imbalanced problem arises. To solve this problem, load rebalancing algorithm is implemented in this paper so that central node should not overloaded. The implementation is done in hadoop distributed file system. As apache hadoop is used, security issues are arises. To solve these security issues and to increase security, [17]Kerberos authentication protocol is implemented to handle multiple nodes. This paper shows real time implementation experiment on cluster.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.