Ceph: A Scalable, High-Performance Distributed File System
#1

Abstract
We have developed Ceph, a distributed file system that provides excellent performance, reliability, and scalability. Ceph maximizes the separation between data and metadata management by replacing allocation tables with a pseudo-random data distribution function (CRUSH) designed for heterogeneous and dynamic clusters of unreliable object storage devices (OSDs). We leverage device intelligence by distributing data replication, failure detection and recovery to semi-autonomous OSDs running a specialized local object file system. A dynamic distributed metadata cluster provides extremely efficient metadata management and seamlessly adapts to a wide range of general purpose and scientific computing file system workloads. Performance measurements under a variety of workloads show that Ceph has excellent I/O performance and scalable metadata management, supportingmore than 250,000metadata operations per second.
1 Introduction
System designers have long sought to improve the performance of file systems, which have proved critical to the overall performance of an exceedingly broad class of applications. The scientific and high-performance computing communities in particular have driven advances in the performance and scalability of distributed storage systems, typically predicting more general purpose needs by a few years. Traditional solutions, exemplified by NFS [20], provide a straightforward model in which a server exports a file system hierarchy that clients can map into their local name space. Although widely used, the centralization inherent in the client/server model has proven a significant obstacle to scalable performance. More recent distributed file systems have adopted architectures based on object-based storage, in which conventional hard disks are replaced with intelligent object storage devices (OSDs) which combine a CPU, network interface, and local cache with an underlying disk or RAID [4, 7, 8, 32, 35]. OSDs replace the traditional block-level interface with one in which clients can read or write byte ranges to much larger (and often variably sized) named objects, distributing low-level block allocation decisions to the devices themselves. Clients typically interact with a metadata server (MDS) to perform metadata operations (open, rename), while communicating directly with OSDs to perform file I/O (reads and writes), significantly improving overall scalability. Systems adopting this model continue to suffer from scalability limitations due to little or no distribution of themetadata workload. Continued reliance on traditional file system principles like allocation lists and inode tables and a reluctance to delegate intelligence to the OSDs have further limited scalability and performance, and increased the cost of reliability. We present Ceph, a distributed file system that provides excellent performance and reliability while promising unparalleled scalability. Our architecture is based on the assumption that systems at the petabyte scale are inherently dynamic: large systems are inevitably built incrementally, node failures are the norm rather than the exception, and the quality and character of workloads are constantly shifting over time. Ceph decouples data and metadata operations by eliminating file allocation tables and replacing themwith generating functions. This allows Ceph to leverage the intelligence present in OSDs to distribute the complexity surrounding data access, update serialization, replication and reliability, failure detection, and recovery. Ceph utilizes a highly adaptive distributed metadata cluster architecture that dramatically improves the scalability of metadata access, and with it, the scalability of the entire system. We discuss the goals and workload assumptions motivating our choices in the design of the architecture, analyze their impact on system scalability and performance, and relate our experiences in implementing a functional system prototype.


Download full report
http://googleurl?sa=t&source=web&cd=1&ve...osdi06.pdf&ei=v2hbTr2PD8TXrQetvuidCw&usg=AFQjCNFdXetREWH5tHDAxo2xlH49Q81wWw
Reply

Important Note..!

If you are not satisfied with above reply ,..Please

ASK HERE

So that we will collect data for you and will made reply to the request....OR try below "QUICK REPLY" box to add a reply to this page
Popular Searches: ppt of scalable high performance virus detection, trends in distributed file system ppt, ceph heterogeneousjectad documentation, scalable distributed software configuration management, nanofluids for high performance cooling system, scalability antonym, distributed file systems,

[-]
Quick Reply
Message
Type your reply to this message here.

Image Verification
Please enter the text contained within the image into the text box below it. This process is used to prevent automated spam bots.
Image Verification
(case insensitive)

Possibly Related Threads...
Thread Author Replies Views Last Post
  Platform Autonomous Custom Scalable Service using Service Oriented Cloud Computing Ar 1 1,033 15-02-2017, 04:39 PM
Last Post: jaseela123d
  Enhancing VANET Performance by Joint Adaptation of Transmission Power and Contention 1 777 15-02-2017, 03:24 PM
Last Post: jaseela123d
  Analyzing the performance of turbo Code mechanical wiki 1 1,236 24-08-2016, 11:38 AM
Last Post: seminar report asees
  Crime File management project topics 11 15,659 11-04-2014, 04:23 PM
Last Post: babs01
  E -Crime File Management System seminar class 8 13,376 24-06-2013, 10:02 AM
Last Post: computer topic
  crime file system project topics 9 12,356 29-04-2013, 11:39 AM
Last Post: computer topic
  File tracking system mechanical engineering crazy 24 12,854 13-03-2013, 03:28 PM
Last Post: computer idea
  Secure Multipart File Transfer projectsofme 4 3,746 07-03-2013, 05:16 PM
Last Post: Guest
  distributed cache updating for the dynamic source routing protocol project report tiger 4 3,000 05-03-2013, 02:22 PM
Last Post: Guest
  DISTRIBUTED MOBILITY MANAGEMENT FOR TARGET TACKING IN MOBILE SENSOR NETWORKS computer science technology 4 3,911 22-02-2013, 03:05 PM
Last Post: seminar details

Forum Jump: