d5773de2a8
There are details that confuse people (like the usage of may), when they are configuring a ceph environment. This should clarify the situation. Change-Id: Ic8b734eabcfa218c687db69290c8b3cef1def7c5
41 lines
1.5 KiB
Plaintext
41 lines
1.5 KiB
Plaintext
---
|
|
# Because we have three haproxy nodes, we need
|
|
# to one active LB IP, and we use keepalived for that.
|
|
## Load Balancer Configuration (haproxy/keepalived)
|
|
haproxy_keepalived_external_vip_cidr: "1.2.3.4/32"
|
|
haproxy_keepalived_internal_vip_cidr: "172.29.236.0/22"
|
|
haproxy_keepalived_external_interface: ens2
|
|
haproxy_keepalived_internal_interface: br-mgmt
|
|
|
|
## Ceph cluster fsid (must be generated before first run)
|
|
## Generate a uuid using: python -c 'import uuid; print(str(uuid.uuid4()))'
|
|
generate_fsid: false
|
|
fsid: 116f14c4-7fe1-40e4-94eb-9240b63de5c1 # Replace with your generated UUID
|
|
|
|
## ceph-ansible settings
|
|
## See https://github.com/ceph/ceph-ansible/tree/master/group_vars for
|
|
## additional configuration options availble.
|
|
monitor_address_block: "{{ cidr_networks.container }}"
|
|
public_network: "{{ cidr_networks.container }}"
|
|
cluster_network: "{{ cidr_networks.storage }}"
|
|
osd_scenario: collocated
|
|
journal_size: 10240 # size in MB
|
|
# ceph-ansible automatically creates pools & keys for OpenStack services
|
|
openstack_config: true
|
|
cinder_ceph_client: cinder
|
|
glance_ceph_client: glance
|
|
glance_default_store: rbd
|
|
glance_rbd_store_pool: images
|
|
nova_libvirt_images_rbd_pool: vms
|
|
|
|
cinder_backends:
|
|
RBD:
|
|
volume_driver: cinder.volume.drivers.rbd.RBDDriver
|
|
rbd_pool: volumes
|
|
rbd_ceph_conf: /etc/ceph/ceph.conf
|
|
rbd_store_chunk_size: 8
|
|
volume_backend_name: rbddriver
|
|
rbd_user: "{{ cinder_ceph_client }}"
|
|
rbd_secret_uuid: "{{ cinder_ceph_client_uuid }}"
|
|
report_discard_supported: true
|