Abstract

This study addresses the scalability issues involving file systems as critical components of computer clusters, especially for commercial applications. Given that wide striping is an effective means of achieving scalability as it warrants good load balancing and allows node cooperation, we choose to implement a new data distribution scheme in order to achieve the scalability of computer clusters. We suggest combining both wide striping and replication techniques using a new data distribution technique based on “chained declustering”. Thus, we suggest a complete architecture, using a cluster of clusters, whose performance is not limited by the network and can be adjusted with one-node precision. In addition, update costs are limited as it is not necessary to redistribute data on the existing nodes every time the system is expanded. The simulations indicate that our data distribution technique and our read algorithm balance the load equally amongst all the nodes of the original cluster and the additional ones. Therefore, the scalability of the system is close to the ideal scenario: once the size of the original cluster is well defined, the total number of nodes in the system is no longer limited, and the performance increases linearly.

Highlights

  • Once the number of additional nodes is equal to the number of nodes in the original cluster, the size doubles and the system can be split into two independent clusters: due to our data distribution technique, the cluster formed of additional nodes is

  • This study addressed the scalability issues of file systems for computer clusters

  • Our primary analysis shows that wide striping is an efficient way to achieve scalability as it warrants good load balancing and allows nodes to cooperate

Read more

Summary

Introduction

In of the file system must be expanded are the most this context, the performance scalability of the parallel complex, whereas increasing storage capacity is easier system used is very important: such systems must since this operation consists of adding disks to support an increasing number of requests while the each node. In order to improve scalability through the theories and metrics such as iso-speed or iso-efficiency use of a new data distribution scheme for computer have already been defined [1, 2] These clusters, we chose to work with the Parallel Virtual File pioneer studies are mainly focus on the algorithmic System (PVFS) a file access mechanism, developed at aspects of scalability. Depending on the the disks cooperate to provide a better level of application used, there is a minimum level of quality of performance than a single disk operating on its own

Objectives
Results
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call