Design
Object storage S3
Block storage
Ceph can provide clients with Thin provisioning, thin-provisioned block devices. When an application writes data to Ceph using a block device, Ceph automatically stripes and replicates the data across the cluster. Ceph's ''RADOS Block Device'' (RBD) also integrates with Kernel-based Virtual Machines (KVMs). Ceph block storage may be deployed on traditional HDDs and/or Solid-state_drive, SSDs which are associated with Ceph's block storage for use cases, including databases, virtual machines, data analytics, artificial intelligence, and machine learning. Block storage clients often require high Hard_disk_drive_performance_characteristics#Data_transfer_rate, throughput and IOPS, thus Ceph RBD deployments increasingly utilize SSDs with NVM_Express, NVMe interfaces. "RBD" is built on with Ceph's foundational RADOS object storage system that provides the librados interface and the CephFS file system. Since RBD is built on librados, RBD inherits librados's abilities, including clones and Snapshot_(computer_storage), snapshots. By striping volumes across the cluster, Ceph improves performance for large block device images. "Ceph-iSCSI" is a gateway which enables access to distributed, highly available block storage from Microsoft Windows and VMware vSphere servers or clients capable of speaking the iSCSI protocol. By using ceph-iscsi on one or more iSCSI gateway hosts, Ceph RBD images become available as Logical Units (LUs) associated with iSCSI targets, which can be accessed in an optionally load-balanced, highly available fashion. Since ceph-iscsi configuration is stored in the Ceph RADOS object store, ceph-iscsi gateway hosts are inherently without persistent state and thus can be replaced, augmented, or reduced at will. As a result, Ceph Storage enables customers to run a truly distributed, highly-available, resilient, and self-healing enterprise storage technology on commodity hardware and an entirely open source platform. The block device can be virtualized, providing block storage to virtual machines, in virtualization platforms such as OpenShift, OpenStack, Kubernetes, OpenNebula, Ganeti, Apache CloudStack and Proxmox Virtual Environment.File storage
Ceph's file system (CephFS) runs on top of the same RADOS foundation as Ceph's object storage and block device services. The CephFS metadata server (MDS) provides a service that maps the directories and file names of the file system to objects stored within RADOS clusters. The metadata server cluster can expand or contract, and it can rebalance file system metadata ranks dynamically to distribute data evenly among cluster hosts. This ensures high performance and prevents heavy loads on specific hosts within the cluster. Clients mount the POSIX-compatible file system using a Linux kernel client. An older Filesystem in Userspace, FUSE-based client is also available. The servers run as regular Unix daemon (computer software), daemons. Ceph's file storage is often associated with log collection, messaging, and file storage.Dashboard
Crimson
Beginning in 2019 the Crimson project has been reimplementing the OSD data path. The goal of Crimson is to minimize latency and CPU overhead. Modern storage devices and interfaces including NVMe and 3D XPoint have become much faster than Hard disk drive, HDD and even SAS/SATA Solid-state drive, SSDs, but CPU performance has not kept pace. Moreover is meant to be a backward-compatible drop-in replacement for . While Crimson can work with the BlueStore back end (via AlienStore), a new native ObjectStore implementation called SeaStore is also being developed along with CyanStore for testing purposes. One reason for creating SeaStore is that transaction support in the BlueStore back end is provided by RocksDB, which needs to be re-implemented to achieve better parallelism.History
Ceph was created by Sage Weil for his doctoral dissertation, which was advised by Professor Scott A. Brandt at the Jack Baskin School of Engineering, University of California, Santa Cruz (UCSC), and sponsored by the Advanced Simulation and Computing Program (ASC), including Los Alamos National Laboratory (LANL), Sandia National Laboratories (SNL), and Lawrence Livermore National Laboratory (LLNL). The first line of code that ended up being part of Ceph was written by Sage Weil in 2004 while at a summer internship at LLNL, working on scalable filesystem metadata management (known today as Ceph's MDS). In 2005, as part of a summer project initiated by Scott A. Brandt and led by Carlos Maltzahn, Sage Weil created a fully functional file system prototype which adopted the name Ceph. Ceph made its debut with Sage Weil giving two presentations in November 2006, one at USENIX OSDI 2006 and another at ACM/IEEE Supercomputing Conference, SC'06. After his graduation in autumn 2007, Weil continued to work on Ceph full-time, and the core development team expanded to include Yehuda Sadeh Weinraub and Gregory Farnum. On March 19, 2010, Linus Torvalds merged the Ceph client into Linux kernel version 2.6.34 which was released on May 16, 2010. In 2012, Weil created Inktank Storage for professional services and support for Ceph. In April 2014, Red Hat purchased Inktank, bringing the majority of Ceph development in-house to make it a production version for enterprises with support (hotline) and continuous maintenance (new versions). In October 2015, the Ceph Community Advisory Board was formed to assist the community in driving the direction of open source software-defined storage technology. The charter advisory board includes Ceph community members from global IT organizations that are committed to the Ceph project, including individuals from Red Hat, Intel, Canonical (company), Canonical, CERN, Cisco, Fujitsu, SanDisk, and SUSE S.A., SUSE. In November 2018, the Linux Foundation launched the Ceph Foundation as a successor to the Ceph Community Advisory Board. Founding members of the Ceph Foundation included Amihan, Canonical (company), Canonical, China Mobile, DigitalOcean, Intel, OVH, ProphetStor Data Services, Red Hat, SoftIron, SUSE S.A., SUSE, Western Digital, XSKY Data Technology, and ZTE. In March 2021, SUSE discontinued its Enterprise Storage product incorporating Ceph in favor of Rancher Labs, Rancher's Longhorn, and the former Enterprise Storage website was updated stating "SUSE has refocused the storage efforts around serving our strategic SUSE Enterprise Storage Customers and are no longer actively selling SUSE Enterprise Storage."Release history
Available platforms
While basically built for Linux, Ceph has been also partially ported to Windows platform. It is production-ready for Windows Server 2016 (some commands might be unavailable due to lack of UNIX socket implementation), Windows Server 2019 and Windows Server 2022, but testing/development can be done also on Windows 10 and Windows 11. One can use Ceph RBD and CephFS on Windows, but OSD is not supported on this platform. There is also FreeBSD implementation of Ceph.Etymology
The name "Ceph" is a shortened form of "cephalopod", a class of Mollusca, molluscs that includes squids, cuttlefish, nautiloids, and octopuses. The name (emphasized by the logo) suggests the highly parallel behavior of an octopus and was chosen to associate the file system with "Sammy", the banana slug mascot of University of California, Santa Cruz, UCSC. Both cephalopods and banana slugs are molluscs.See also
* BeeGFS * Distributed file system * Distributed parallel fault-tolerant file systems * Gfarm file system * GlusterFS * GPFS * Kubernetes * LizardFS * Lustre (file system), Lustre * MapR FS * Moose File System * OrangeFS * Parallel Virtual File System * Quantcast File System * RozoFS * Software-defined storage * XtreemFS * ZFS * Comparison of distributed file systemsReferences
Further reading
* * * *External links
* * * (2023) {{Red Hat Distributed file systems supported by the Linux kernel Network file systems Red Hat software Userspace file systems Virtualization software for Linux Free software programmed in C++ Free software programmed in Python Software using the GNU Lesser General Public License