Abstract

The Hadoop Distributed File System (HDFS) is a large data storage system which exhibits several features of a good distributed file system. In this paper we integrate Mat lab Distributed Computing Server (MDCS) with HDFS to build a scalable, efficient platform for scientific computations. We use an FTP server on top of HDFS for data transfer from the Mat lab system to HDFS. The motivation of using HDFS for storage with MDCS is to provide an efficient, fault-tolerant file system and also to utilize the resources efficiently by making each system serve as both data node for HDFS and worker for MDCS. We test the storage efficiency of HDFS and compare with normal file system for data transfer operations through MDCS.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call