Need help with ceph?
Click the “chat” button below for chat support from the developer who created it, or find similar developers for support.

About the developer

9.8K Stars 4.6K Forks Other 127.3K Commits 654 Opened issues


Ceph is a distributed object, block, and file storage platform

Services available


Need anything else?

Contributors list

Ceph - a scalable distributed storage system

Please see for current info.

Contributing Code

Most of Ceph is dual licensed under the LGPL version 2.1 or 3.0. Some miscellaneous code is under a BSD-style license or is public domain. The documentation is licensed under Creative Commons Attribution Share Alike 3.0 (CC-BY-SA-3.0). There are a handful of headers included here that are licensed under the GPL. Please see the file COPYING for a full inventory of licenses by file.

Code contributions must include a valid "Signed-off-by" acknowledging the license for the modified or contributed file. Please see the file SubmittingPatches.rst for details on what that means and on how to generate and submit patches.

We do not require assignment of copyright to contribute code; code is contributed under the terms of the applicable license.

Checking out the source

You can clone from github with

git clone [email protected]:ceph/ceph

or, if you are not a github user,

git clone git://

Ceph contains many git submodules that need to be checked out with

git submodule update --init --recursive

Build Prerequisites

The list of Debian or RPM packages dependencies can be installed with:


Building Ceph

Note that these instructions are meant for developers who are compiling the code for development and testing. To build binaries suitable for installation we recommend you build deb or rpm packages or refer to the
to see which configuration options are specified for production builds.

Build instructions:

cd build

( now defaults to creating a debug build of ceph that can be up to 5x slower with some workloads. Please pass "-DCMAKEBUILDTYPE=RelWithDebInfo" to to create a non-debug release.

The number of jobs used by

is derived from the number of CPU cores of the building host if unspecified. Use the
option to limit the job number if the build jobs are running out of memory. On average, each job takes around 2.5GiB memory.)

This assumes you make your build dir a subdirectory of the ceph.git checkout. If you put it elsewhere, just point

to the correct path to the checkout. Any additional CMake args can be specified by setting ARGS before invoking do_cmake. See cmake options for more details. Eg.

To build only certain targets use:

ninja [target name]

To install:

ninja install

CMake Options

If you run the

command by hand, there are many options you can set with "-D". For example, the option to build the RADOS Gateway is defaulted to ON. To build without the RADOS Gateway:
cmake -DWITH_RADOSGW=OFF [path to top-level ceph directory]

Another example below is building with debugging and alternate locations for a couple of external dependencies:

cmake -DLEVELDB_PREFIX="/opt/hyperleveldb" \
-DCMAKE_INSTALL_PREFIX=/opt/ceph -DCMAKE_C_FLAGS="-Og -g3 -gdwarf-4" \

To view an exhaustive list of -D options, you can invoke

cmake -LH

If you often pipe

and would like to maintain the diagnostic colors for errors and warnings (and if your compiler supports it), you can invoke
cmake -DDIAGNOSTICS_COLOR=always ...

Then you'll get the diagnostic colors when you execute:

ninja | less -R

Other available values for 'DIAGNOSTICS_COLOR' are 'auto' (default) and 'never'.

Building a source tarball

To build a complete source tarball with everything needed to build from source and/or build a (deb or rpm) package, run


This will create a tarball like ceph-$version.tar.bz2 from git. (Ensure that any changes you want to include in your working directory are committed to git.)

Running a test cluster

To run a functional test cluster,

cd build
ninja vstart        # builds just enough to run vstart
../src/ --debug --new -x --localhost --bluestore
./bin/ceph -s

Almost all of the usual commands are available in the bin/ directory. For example,

./bin/rados -p rbd bench 30 write
./bin/rbd create foo --size 1000

To shut down the test cluster,


To start or stop individual daemons, the sysvinit script can be used:

./bin/init-ceph restart osd.0
./bin/init-ceph stop

Running unit tests

To build and run all tests (in parallel using all processors), use

cd build
ctest -j$(nproc)

(Note: Many targets built from src/test are not run using

. Targets starting with "unittest" are run in
ninja check
and thus can be run with
. Targets starting with "ceph_test" can not, and should be run by hand.)

When failures occur, look in build/Testing/Temporary for logs.

To build and run all tests and their dependencies without other unnecessary targets in Ceph:

cd build
ninja check -j$(nproc)

To run an individual test manually, run

with -R (regex matching):
ctest -R [regex matching test name(s)]


does not build the test it's running or the dependencies needed to run it)

To run an individual test manually and see all the tests output, run

with the -V (verbose) flag:
ctest -V -R [regex matching test name(s)]

To run tests manually and run the jobs in parallel, run

with the
ctest -j [number of jobs]

There are many other flags you can give

for better control over manual test execution. To view these options run:
man ctest

Building the Documentation


The list of package dependencies for building the documentation can be found in

sudo apt-get install `cat doc_deps.deb.txt`

Building the Documentation

To build the documentation, ensure that you are in the top-level

directory, and execute the build script. For example:

We use cookies. If you continue to browse the site, you agree to the use of cookies. For more information on our use of cookies please see our Privacy Policy.