HyCloud Tweaking Hybrid Cloud Storage Services for Cost-Efficient File system Hosting
ABSTARCT :
Today’s cloud storage infrastructures typically pro-vide two distinct types of services for hosting files:object storage like Amazon S3 and file system storage like Amazon EFS. The former supports simple, flat object operations with a low unit storage price, while the latter supports complex, hierarchicalfilesystem operations with a high unit storage price. In practice, however, a cloud storage user often desires the advantages of both—efficient file system operations with a low unit storage price. An intuitive approach to achieving this goal is to combine the two types of services, e.g.,by hosting large files in S3 and small files together with directory structures in EFS. Unfortunately, our benchmark experiments indicate that the clients’ download performance for large files becomes a severe system bottleneck. In this paper, we attempt to address the bottleneck with little overhead by carefully tweaking the usages of S3 and EFS. This attempt is enabled by two key observations. First, since S3and EFS have the same unit network-traffic price and the data transfer between S3 and EFS is free of charge, we can employers as a relay for the clients’ quickly downloading large files. Second, noticing that significant similarity exists between the files hosted at the cloud and its users, in most times we can convert large size file downloads into small-size file synchronizations(through delta encoding and data compression). Guided by the observations, we design and implement an open-source system called HyCloud. It automatically invokes the data APIs of S3 and EFS on behalf of users, and handles the data transfer among S3,EFS and the clients. Real-world evaluations demonstrate that the unit storage price of HyCloud is close to that of S3, and the file system operations are executed as quickly as in EFS in most times (sometimes even more quickly than in EFS).
EXISTING SYSTEM :
Recently the number of cloud storage users has increased abundantly. The reason is that the Cloud Storage systemreduces the burden of maintenance andit has less storage cost compared with other storage methods. It also provides high availability, reliability and also it is most suitable for high volume of data storage. In order to provide high availability and reliability, the systems introduce redundancy. In replicated systems, the objects are copied several times and each copy residing on different locations in distributed computing. Therefore, the Data Replication is rendering little bit threat about the Cloud Storage System for the users and for the providers it is a big challenge to provide efficient Data Storage.
DISADVANTAGE :
There are two major strategies used to obtain a replication system in cloud storage. They are Static mechanism and Dynamic mechanism. In static method of replication the availability and reliability is high.
Unwanted use of storage, no flexibility, no scalability and high amount of cost received from the user for storage
PROPOSED SYSTEM :
This paper analyzed the various existing data replication strategies and pointed out the major issues affected by the Data Replication. In future, the direction of this research work is aimed to reduce the number of replication without affecting the availability and reliability of the data.
ADVANTAGE :
High Availability: Data availability is one of the important key factors for cloud storage system. Data replication has been commonly used for increasing the availability in distributed storage
High Reliability: In distributed storage systems, failures of services are treated as reliability issue. It is clear that the reliability of a system will generally increase as the number of replication. Since more replicas will be able to mask more failures.
Minimum Data Loss: If there is any technical problem or inconvenience in provider side the user can loss their information or go for offline. The replication system overcomes the above stated problem.
|