Abstract

The Fragmentation and duplication in data sets is used to overcome the increased data overloading issue in the cloud servers. The increase in data usage and processing in cloud has brought new challenges to data management in cloud computing. We propose the idea which helps in reducing the data load in the cloud and reduces the storage and management cost for the users. Duplication finding plays a very major role in data management. Data de-duplication method finds the restricted fingerprint for every data chunk by using hash algorithms such as MD5 and SHA. The recognized fingerprint is then compared touching other available chunks in a database that is dedicated for storing the chunks. Though, there is simply one copy for every file stored in cloud, it will not be immobile if such a file is owned by a massive number. As a part the de-duplication system improves storage consumption whereas dropping reliability. Aiming to contradict with the above safety challenges, this proposed idea makes the first attempt to provide the idea of distributed dependable de-duplication system. This new distributed de-duplication system comes with increased reliability in which the data chunks are distributed diagonally to various cloud servers. This allows the redundancy of all data is eliminated. The security needs of data privacy and tag consistency are also achieve by introducing a deterministic furtive sharing system in distributed storage systems, as an option of using convergent encryption as in foregoing de-duplication system. Keywords: Cloud, Cloud Storage, Data Mapping, File Data Security, Fragmentation, Graph Colouring Algorithm, Graphical Representation, Node Allocation, Performance

Highlights

  • One of the developing paradigms of distributed computing is cloud computing which become root for business, technical and social perspective

  • The most highly developed application such as Matlab, Mathematica which does not run by a single desktop system due to their rate and speed of memory performance; they use this cloud environment for the data representation

  • The current method will use the static method which restrict to certain extreme. In this paper it deliver a dynamic de duplication scheme for cloud storage, which aiming to that improve storage efficiency with shrunk of segment analysis and maintaining redundancy for fault tolerance method ways

Read more

Summary

Introduction

One of the developing paradigms of distributed computing is cloud computing which become root for business, technical and social perspective. Cloud applications are more popular due to the availability, scalability and utility model a high demand on interactive application which attracts the user in great demand due to the availability. Data intensive and analysis model of the cloud. Cloud basically a physical environment which provides a virtualization environment in which the user can the usage by the internet services (Figure 1). The most highly developed application such as Matlab, Mathematica which does not run by a single desktop system due to their rate and speed of memory performance; they use this cloud environment for the data representation.

Secure and Constant Cost Public Cloud Storage Auditing with Deduplication
A Survey on Deduplication in Cloud Computing
Dynamic Data De-Duplication in Cloud
Data Protection and Deduplication in Cloud
Energy Efficient and Replication in Cloud Computing
De-Duplication of Data in Cloud
Drawbacks
Proposed System
Methodologies
Requester 1
Requester 2
Module Details of Servers
First node Selection
Deduplication
Data Sharing
Mapping
Distributed Storage
File Access
Result and Discussion
Conclusion
Future Work
10. References
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call