From 55455cdde252921253fbdd843336afcf588a3998 Mon Sep 17 00:00:00 2001 From: Michal Nasiadka Date: Mon, 17 Sep 2018 10:18:38 +0200 Subject: [PATCH] Rework ceph-guide bluestore/filestore sections Create bluestore/filestore subsections in ceph-guide and add note about using bluestore by default - for better readability. Change-Id: Ib51706838ecf35ac274053b916b3a17f0de90758 --- doc/source/reference/ceph-guide.rst | 47 ++++++++++++++++++----------- 1 file changed, 30 insertions(+), 17 deletions(-) diff --git a/doc/source/reference/ceph-guide.rst b/doc/source/reference/ceph-guide.rst index 694ebe6812..3f179912ea 100644 --- a/doc/source/reference/ceph-guide.rst +++ b/doc/source/reference/ceph-guide.rst @@ -4,19 +4,21 @@ Ceph in Kolla ============= -The out-of-the-box Ceph deployment requires 3 hosts with at least one block -device on each host that can be dedicated for sole use by Ceph. However, with -tweaks to the Ceph cluster you can deploy a **healthy** cluster with a single -host and a single block device. +.. note:: + The out-of-the-box Ceph deployment requires 3 hosts with at least one block + device on each host that can be dedicated for sole use by Ceph. + + However, with tweaks to the Ceph cluster you can deploy a **healthy** cluster + with a single host and a single block device. Requirements -~~~~~~~~~~~~ +------------ * A minimum of 3 hosts for a vanilla deploy * A minimum of 1 block device per host Preparation -~~~~~~~~~~~ +----------- To prepare a disk for use as a `Ceph OSD `_ you must add a @@ -24,6 +26,14 @@ special partition label to the disk. This partition label is how Kolla detects the disks to format and bootstrap. Any disk with a matching partition label will be reformatted so use caution. +Filestore +~~~~~~~~~ + +.. note:: + + From Rocky release - kolla-ansible by default creates Bluestore OSDs. + Please see Configuration section to change that behaviour. + To prepare a filestore OSD as a storage drive, execute the following operations: @@ -53,6 +63,9 @@ usage with Kolla. .. end +Bluestore +~~~~~~~~~ + To prepare a bluestore OSD partition, execute the following operations: .. code-block:: console @@ -158,7 +171,7 @@ To prepare the journal external drive execute the following command: Configuration -~~~~~~~~~~~~~ +------------- Edit the ``[storage]`` group in the inventory which contains the hostname of the hosts that have the block devices you have prepped as shown above. @@ -250,7 +263,7 @@ For more details, see `NETWORK CONFIGURATION REFERENCE of Ceph Documentation. Deployment -~~~~~~~~~~ +---------- Finally deploy the Ceph-enabled OpenStack: @@ -260,8 +273,8 @@ Finally deploy the Ceph-enabled OpenStack: .. end -Using a Cache Tiering -~~~~~~~~~~~~~~~~~~~~~ +Using Cache Tiering +------------------- An optional `cache tiering `_ can be deployed by formatting at least one cache device and enabling cache. @@ -304,7 +317,7 @@ After this run the playbooks as you normally would, for example: .. end Setting up an Erasure Coded Pool -~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ +-------------------------------- `Erasure code `_ is the new big thing from Ceph. Kolla has the ability to setup your Ceph pools @@ -328,7 +341,7 @@ To enable erasure coded pools add the following options to your .. end Managing Ceph -~~~~~~~~~~~~~ +------------- Check the Ceph status for more diagnostic information. The sample output below indicates a healthy cluster: @@ -389,10 +402,10 @@ The default pool Ceph creates is named **rbd**. It is safe to remove this pool: .. end Troubleshooting -~~~~~~~~~~~~~~~ +--------------- Deploy fails with 'Fetching Ceph keyrings ... No JSON object could be decoded' ------------------------------------------------------------------------------- +~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ If an initial deploy of Ceph fails, perhaps due to improper configuration or similar, the cluster will be partially formed and will need to be reset for a @@ -408,7 +421,7 @@ from each Ceph monitor node: ceph-mon Simple 3 Node Example -~~~~~~~~~~~~~~~~~~~~~ +--------------------- This example will show how to deploy Ceph in a very simple setup using 3 storage nodes. 2 of those nodes (kolla1 and kolla2) will also provide other @@ -455,7 +468,7 @@ environment before adding the 3rd node for Ceph: .. end Configuration -------------- +~~~~~~~~~~~~~ To prepare the 2nd disk (/dev/sdb) of each nodes for use by Ceph you will need to add a partition label to it as shown below: @@ -513,7 +526,7 @@ It is now time to enable Ceph in the environment by editing the .. end Deployment ----------- +~~~~~~~~~~ Finally deploy the Ceph-enabled configuration: