kolla/doc/ceph-guide.rst
Nguyễn Duy Vũ (CBR09) 640830a440 Add Ceph deployment document
Deploy Ceph in kolla

Change-Id: Icecc6f4ee25b5ee0cad4265c0b57790bfb32a905
2015-10-09 17:23:42 +00:00

124 lines
3.2 KiB
ReStructuredText

Ceph in Kolla
=============
Requirements
------------
Using Ceph requires at least two physical disks across the OpenStack deployment to operate correctly.
Preparation and Deployment
--------------------------
For the disks used for Ceph, execute the following operations:
::
<WARNING ALL DATA ON $DISK will be LOST!>
parted $DISK -s -- mklabel gpt mkpart KOLLA_CEPH_OSD_BOOTSTRAP 1 -1
where $DISK == /dev/sdb or something similiar
The following shows an example of using parted to configure /dev/sdb for usage with Kolla.
::
parted /dev/sdb -s -- mklabel gpt mkpart KOLLA_CEPH_OSD_BOOTSTRAP 1 -1
parted /dev/sdb print
Model: VMware, VMware Virtual S (scsi)
Disk /dev/sdb: 10.7GB
Sector size (logical/physical): 512B/512B
Partition Table: gpt
Number Start End Size File system Name Flags
1 1049kB 10.7GB 10.7GB KOLLA_CEPH_OSD_BOOTSTRAP
Edit the [storage] group in the inventory which contains the hostname(or IP) of the Ceph-OSD hosts
which have the above disks. Note: ssh authentication is required for Ceph, even in all-in-one.
(TODO(CBR09): insert link to bug around this if there is one). The following shows an example
of two Ceph-OSD hosts which using one disk of the controller node and one disk of compute1.
::
file: ansible/inventory/multinode
...
[storage]
controller
compute1
...
For AIO:
::
file: ansible/inventory/multinode
...
[storage]
all-in-one
...
Enable Ceph in /etc/kolla/globals.yml (Ceph is disabled by default):
::
file: /etc/kolla/globals.yml
....
enable_ceph: "yes"
....
Finally deploy the Ceph-enabled OpenStack:
::
tools/kolla-ansible deploy -i ansible/inventory/multinode
Debugging Ceph
--------------
If Ceph is run in an all-in-one deployment or with less than three storage nodes, further
configuration is required. It is necessary to change the default number of copies for the pool.
The following example demonstrates how to change the number of copies for the pool:
If the deployment includes two Ceph-OSD hosts as mentioned above, set the pool to 2.
::
docker exec ceph_mon ceph osd pool set rbd size 2 (default only have rdb pool)
For AIO:
::
docker exec ceph_mon ceph osd pool set rbd size 1 (default only have rdb pool)
If Glance, Nova, and cinder have been deployed, all pools have to be modified.
An example of modifying the pools:
::
for p in images vms volumes backups rbd; do docker exec ceph_mon ceph osd pool set $p size 2; done
For AIO:
::
for p in images vms volumes backups rbd; do docker exec ceph_mon ceph osd pool set $p size 1; done
After making this change, it is mandatory to restart all Ceph osd containers.
Check the Ceph status for more diagnostic information. The sample output below
indicates a healthy cluster:
::
docker exec ceph_mon ceph -s
cluster 5fba2fbc-551d-11e5-a8ce-01ef4c5cf93c
health HEALTH_OK
monmap e1: 1 mons at {controller=10.0.0.128:6789/0}
election epoch 2, quorum 0 controller
osdmap e18: 2 osds: 2 up, 2 in
pgmap v27: 64 pgs, 1 pools, 0 bytes data, 0 objects
68676 kB used, 20390 MB / 20457 MB avail
64 active+clean