Remove ceph block device support
It leaves certain ceph mentions in globals.yml.j2 as it needs syncing with kolla-ansible contents anyways (these are all comments). Change-Id: I05e9c6223583e9bb5dc0020edc0b56990275093c Story: 2007295 Task: 38766
This commit is contained in:
parent
5ee815b0e9
commit
e7d7daa531
@ -1,8 +0,0 @@
|
||||
---
|
||||
- name: Ensure Ceph disk are tagged
|
||||
hosts: overcloud
|
||||
tags:
|
||||
- kolla-ceph
|
||||
roles:
|
||||
- role: kolla-ceph
|
||||
when: kolla_enable_ceph | bool
|
@ -1,7 +0,0 @@
|
||||
---
|
||||
###############################################################################
|
||||
# OpenStack Ceph configuration.
|
||||
|
||||
# Ansible host pattern matching hosts on which Ceph storage services
|
||||
# are deployed. The default is to use hosts in the 'storage' group.
|
||||
ceph_hosts: "storage"
|
@ -19,7 +19,6 @@ compute_default_network_interfaces: >
|
||||
{{ ([admin_oc_net_name,
|
||||
internal_net_name,
|
||||
storage_net_name,
|
||||
ceph_storage_net_name,
|
||||
tunnel_net_name] +
|
||||
(external_net_names if kolla_enable_neutron_provider_networks | bool else [])) | reject('none') | unique | list }}
|
||||
|
||||
@ -110,16 +109,6 @@ compute_lvm_group_data_lv_docker_volumes_size: 75%VG
|
||||
# Filesystem for docker volumes LVM backing volume. ext4 allows for shrinking.
|
||||
compute_lvm_group_data_lv_docker_volumes_fs: ext4
|
||||
|
||||
###############################################################################
|
||||
# Compute node Ceph configuration.
|
||||
|
||||
# List of Ceph disks.
|
||||
# The format is a list of dict like :
|
||||
# - { osd: "/dev/sdb", journal: "/dev/sdc" }
|
||||
# - { osd: "/dev/sdd" }
|
||||
# Journal variable is not mandatory.
|
||||
compute_ceph_disks: []
|
||||
|
||||
###############################################################################
|
||||
# Compute node sysctl configuration.
|
||||
|
||||
|
@ -24,7 +24,6 @@ controller_default_network_interfaces: >
|
||||
inspection_net_name,
|
||||
internal_net_name,
|
||||
storage_net_name,
|
||||
ceph_storage_net_name,
|
||||
swift_storage_net_name,
|
||||
cleaning_net_name] | reject('none') | unique | list }}
|
||||
|
||||
@ -121,16 +120,6 @@ controller_lvm_group_data_lv_docker_volumes_size: 75%VG
|
||||
# Filesystem for docker volumes LVM backing volume. ext4 allows for shrinking.
|
||||
controller_lvm_group_data_lv_docker_volumes_fs: ext4
|
||||
|
||||
###############################################################################
|
||||
# Controller node Ceph configuration.
|
||||
|
||||
# List of Ceph disks.
|
||||
# The format is a list of dict like :
|
||||
# - { osd: "/dev/sdb", journal: "/dev/sdc" }
|
||||
# - { osd: "/dev/sdd" }
|
||||
# Journal variable is not mandatory.
|
||||
controller_ceph_disks: []
|
||||
|
||||
###############################################################################
|
||||
# Controller node sysctl configuration.
|
||||
|
||||
|
@ -118,8 +118,6 @@ overcloud_container_image_regex_map:
|
||||
enabled: "{{ kolla_enable_blazar | bool }}"
|
||||
- regex: ceilometer
|
||||
enabled: "{{ kolla_enable_ceilometer | bool }}"
|
||||
- regex: ceph
|
||||
enabled: "{{ kolla_enable_ceph | bool }}"
|
||||
- regex: cinder
|
||||
enabled: "{{ kolla_enable_cinder | bool }}"
|
||||
- regex: cloudkitty
|
||||
@ -346,7 +344,6 @@ kolla_enable_aodh: "no"
|
||||
kolla_enable_barbican: "no"
|
||||
kolla_enable_blazar: "no"
|
||||
kolla_enable_central_logging: "no"
|
||||
kolla_enable_ceph: "no"
|
||||
kolla_enable_ceilometer: "no"
|
||||
# The chrony container is disabled by default on CentOS 7 because we enable an
|
||||
# NTP daemon on the host. On CentOS 8 the chrony container is enabled by
|
||||
|
@ -49,12 +49,6 @@ storage_net_name: 'storage_net'
|
||||
# Name of the network used to carry storage management traffic.
|
||||
storage_mgmt_net_name: 'storage_mgmt_net'
|
||||
|
||||
# Name of the network used to carry ceph storage data traffic.
|
||||
ceph_storage_net_name: "{{ storage_net_name }}"
|
||||
|
||||
# Name of the network used to carry ceph storage management traffic.
|
||||
ceph_storage_mgmt_net_name: "{{ storage_mgmt_net_name }}"
|
||||
|
||||
# Name of the network used to carry swift storage data traffic.
|
||||
swift_storage_net_name: "{{ storage_net_name }}"
|
||||
|
||||
|
@ -13,10 +13,6 @@ storage_bootstrap_user: "{{ lookup('env', 'USER') }}"
|
||||
storage_network_interfaces: >
|
||||
{{ (storage_default_network_interfaces +
|
||||
storage_extra_network_interfaces +
|
||||
([ceph_storage_net_name]
|
||||
if storage_needs_ceph_network else []) +
|
||||
([ceph_storage_mgmt_net_name]
|
||||
if storage_needs_ceph_mgmt_network else []) +
|
||||
([swift_storage_net_name]
|
||||
if storage_needs_swift_network else []) +
|
||||
([swift_storage_replication_net_name]
|
||||
@ -32,15 +28,6 @@ storage_default_network_interfaces: >
|
||||
# List of extra networks to which storage nodes are attached.
|
||||
storage_extra_network_interfaces: []
|
||||
|
||||
# Whether this host requires access to Ceph networks.
|
||||
storage_needs_ceph_network: >-
|
||||
{{ kolla_enable_ceph | bool and
|
||||
inventory_hostname in query('inventory_hostnames', ceph_hosts) }}
|
||||
|
||||
storage_needs_ceph_mgmt_network: >-
|
||||
{{ kolla_enable_ceph | bool and
|
||||
inventory_hostname in query('inventory_hostnames', ceph_hosts) }}
|
||||
|
||||
# Whether this host requires access to Swift networks.
|
||||
storage_needs_swift_network: >-
|
||||
{{ kolla_enable_swift | bool and
|
||||
@ -134,16 +121,6 @@ storage_lvm_group_data_lv_docker_volumes_size: 75%VG
|
||||
# Filesystem for docker volumes LVM backing volume. ext4 allows for shrinking.
|
||||
storage_lvm_group_data_lv_docker_volumes_fs: ext4
|
||||
|
||||
###############################################################################
|
||||
# Storage node Ceph configuration.
|
||||
|
||||
# List of Ceph disks.
|
||||
# The format is a list of dict like :
|
||||
# - { osd: "/dev/sdb", journal: "/dev/sdc" }
|
||||
# - { osd: "/dev/sdd" }
|
||||
# Journal variable is not mandatory.
|
||||
storage_ceph_disks: []
|
||||
|
||||
###############################################################################
|
||||
# Storage node sysctl configuration.
|
||||
|
||||
|
@ -1,6 +0,0 @@
|
||||
---
|
||||
###############################################################################
|
||||
# Compute node Ceph configuration.
|
||||
|
||||
# List of Ceph disks.
|
||||
ceph_disks: "{{ compute_ceph_disks }}"
|
@ -1,6 +0,0 @@
|
||||
---
|
||||
###############################################################################
|
||||
# Controller node Ceph configuration.
|
||||
|
||||
# List of Ceph disks.
|
||||
ceph_disks: "{{ controller_ceph_disks }}"
|
@ -1,6 +0,0 @@
|
||||
---
|
||||
###############################################################################
|
||||
# Storage node Ceph configuration.
|
||||
|
||||
# List of Ceph disks.
|
||||
ceph_disks: "{{ storage_ceph_disks }}"
|
@ -33,16 +33,6 @@
|
||||
kolla_cluster_interface: "{{ storage_mgmt_net_name | net_interface | replace('-', '_') }}"
|
||||
when: storage_mgmt_net_name in network_interfaces
|
||||
|
||||
- name: Set Ceph storage network interface
|
||||
set_fact:
|
||||
kolla_ceph_storage_interface: "{{ ceph_storage_net_name | net_interface | replace('-', '_') }}"
|
||||
when: ceph_storage_net_name in network_interfaces
|
||||
|
||||
- name: Set Ceph cluster network interface
|
||||
set_fact:
|
||||
kolla_ceph_cluster_interface: "{{ ceph_storage_mgmt_net_name | net_interface | replace('-', '_') }}"
|
||||
when: ceph_storage_mgmt_net_name in network_interfaces
|
||||
|
||||
- name: Set Swift storage network interface
|
||||
set_fact:
|
||||
kolla_swift_storage_interface: "{{ swift_storage_net_name | net_interface | replace('-', '_') }}"
|
||||
|
@ -106,7 +106,6 @@
|
||||
with_items:
|
||||
- { name: blazar, file: blazar.conf }
|
||||
- { name: ceilometer, file: ceilometer.conf }
|
||||
- { name: ceph, file: ceph.conf }
|
||||
- { name: cinder, file: cinder.conf }
|
||||
- { name: cloudkitty, file: cloudkitty.conf }
|
||||
- { name: designate, file: designate.conf }
|
||||
@ -215,7 +214,6 @@
|
||||
# Extra free-form user-provided configuration.
|
||||
kolla_extra_blazar: "{{ kolla_extra_config.blazar | default }}"
|
||||
kolla_extra_ceilometer: "{{ kolla_extra_config.ceilometer | default }}"
|
||||
kolla_extra_ceph: "{{ kolla_extra_config.ceph | default }}"
|
||||
kolla_extra_cinder: "{{ kolla_extra_config.cinder | default }}"
|
||||
kolla_extra_cloudkitty: "{{ kolla_extra_config.cloudkitty | default }}"
|
||||
kolla_extra_designate: "{{ kolla_extra_config.designate | default }}"
|
||||
|
@ -120,8 +120,6 @@ kolla_overcloud_inventory_pass_through_host_vars:
|
||||
- "kolla_api_interface"
|
||||
- "kolla_storage_interface"
|
||||
- "kolla_cluster_interface"
|
||||
- "kolla_ceph_storage_interface"
|
||||
- "kolla_ceph_cluster_interface"
|
||||
- "kolla_swift_storage_interface"
|
||||
- "kolla_swift_replication_interface"
|
||||
- "kolla_provision_interface"
|
||||
@ -141,8 +139,6 @@ kolla_overcloud_inventory_pass_through_host_vars_map:
|
||||
kolla_api_interface: "api_interface"
|
||||
kolla_storage_interface: "storage_interface"
|
||||
kolla_cluster_interface: "cluster_interface"
|
||||
kolla_ceph_storage_interface: "ceph_storage_interface"
|
||||
kolla_ceph_cluster_interface: "ceph_cluster_interface"
|
||||
kolla_swift_storage_interface: "swift_storage_interface"
|
||||
kolla_swift_replication_interface: "swift_replication_interface"
|
||||
kolla_provision_interface: "provision_interface"
|
||||
@ -249,8 +245,6 @@ kolla_openstack_logging_debug:
|
||||
#kolla_enable_blazar:
|
||||
#kolla_enable_ceilometer:
|
||||
#kolla_enable_central_logging:
|
||||
#kolla_enable_ceph:
|
||||
#kolla_enable_ceph_rgw:
|
||||
#kolla_enable_cinder:
|
||||
#kolla_enable_cinder_backend_iscsi:
|
||||
#kolla_enable_cinder_backend_lvm:
|
||||
|
@ -152,9 +152,6 @@ control
|
||||
[ironic:children]
|
||||
control
|
||||
|
||||
[ceph:children]
|
||||
control
|
||||
|
||||
[magnum:children]
|
||||
control
|
||||
|
||||
|
@ -89,25 +89,6 @@ neutron
|
||||
[ironic-neutron-agent:children]
|
||||
neutron
|
||||
|
||||
# Ceph
|
||||
[ceph-mds:children]
|
||||
ceph
|
||||
|
||||
[ceph-mgr:children]
|
||||
ceph
|
||||
|
||||
[ceph-nfs:children]
|
||||
ceph
|
||||
|
||||
[ceph-mon:children]
|
||||
ceph
|
||||
|
||||
[ceph-rgw:children]
|
||||
ceph
|
||||
|
||||
[ceph-osd:children]
|
||||
storage
|
||||
|
||||
# Cinder
|
||||
[cinder-api:children]
|
||||
cinder
|
||||
|
@ -26,8 +26,6 @@
|
||||
kolla_provision_interface: "eth8"
|
||||
kolla_inspector_dnsmasq_interface: "eth9"
|
||||
kolla_tunnel_interface: "eth10"
|
||||
kolla_ceph_storage_interface: "eth11"
|
||||
kolla_ceph_cluster_interface: "eth12"
|
||||
kolla_swift_storage_interface: "eth13"
|
||||
kolla_swift_replication_interface: "eth14"
|
||||
|
||||
@ -42,7 +40,6 @@
|
||||
kolla_neutron_external_interfaces: "eth4,eth5"
|
||||
kolla_neutron_bridge_names: "br0,br1"
|
||||
kolla_tunnel_interface: "eth6"
|
||||
kolla_ceph_storage_interface: "eth7"
|
||||
|
||||
- name: Create a temporary directory
|
||||
tempfile:
|
||||
@ -129,8 +126,6 @@
|
||||
kolla_enable_blazar: True
|
||||
kolla_enable_ceilometer: True
|
||||
kolla_enable_central_logging: True
|
||||
kolla_enable_ceph: True
|
||||
kolla_enable_ceph_rgw: True
|
||||
kolla_enable_chrony: True
|
||||
kolla_enable_cinder: True
|
||||
kolla_enable_cinder_backend_hnas_iscsi: True
|
||||
@ -266,8 +261,6 @@
|
||||
#enable_blazar: True
|
||||
#enable_ceilometer: True
|
||||
#enable_central_logging: True
|
||||
#enable_ceph: True
|
||||
#enable_ceph_rgw: True
|
||||
#enable_chrony: True
|
||||
#enable_cinder: True
|
||||
#enable_cinder_backend_iscsi: True
|
||||
@ -355,8 +348,6 @@
|
||||
- kolla_external_vip_interface
|
||||
- storage_interface
|
||||
- cluster_interface
|
||||
- ceph_storage_interface
|
||||
- ceph_cluster_interface
|
||||
- swift_storage_interface
|
||||
- swift_replication_interface
|
||||
- provision_interface
|
||||
@ -494,8 +485,6 @@
|
||||
api_interface: "eth2"
|
||||
storage_interface: "eth3"
|
||||
cluster_interface: "eth4"
|
||||
ceph_storage_interface: "eth11"
|
||||
ceph_cluster_interface: "eth12"
|
||||
swift_storage_interface: "eth13"
|
||||
swift_replication_interface: "eth14"
|
||||
provision_interface: "eth8"
|
||||
@ -511,7 +500,6 @@
|
||||
network_interface: "eth0"
|
||||
api_interface: "eth2"
|
||||
storage_interface: "eth3"
|
||||
ceph_storage_interface: "eth7"
|
||||
tunnel_interface: "eth6"
|
||||
neutron_external_interface: "eth4,eth5"
|
||||
neutron_bridge_name: "br0,br1"
|
||||
|
@ -79,11 +79,6 @@ kolla_feature_flags:
|
||||
- ceilometer_ipmi
|
||||
- cells
|
||||
- central_logging
|
||||
- ceph
|
||||
- ceph_mds
|
||||
- ceph_rgw
|
||||
- ceph_nfs
|
||||
- ceph_dashboard
|
||||
- chrony
|
||||
- cinder
|
||||
- cinder_backup
|
||||
|
@ -1,4 +0,0 @@
|
||||
---
|
||||
|
||||
# List of Ceph disks.
|
||||
ceph_disks: []
|
@ -1,85 +0,0 @@
|
||||
---
|
||||
|
||||
- name: Ensure required packages are installed
|
||||
package:
|
||||
name: parted
|
||||
state: present
|
||||
become: True
|
||||
when: ceph_disks | length > 0
|
||||
|
||||
- name: Check the presence of a partition on the OSD disks
|
||||
become: True
|
||||
parted:
|
||||
device: "{{ item.osd }}"
|
||||
with_items: "{{ ceph_disks }}"
|
||||
register: "disk_osd_info"
|
||||
|
||||
- name: Check the presence of a partition on the journal disks
|
||||
become: True
|
||||
parted:
|
||||
device: "{{ item.journal }}"
|
||||
with_items: "{{ ceph_disks }}"
|
||||
register: "disk_journal_info"
|
||||
when:
|
||||
- item.journal is defined
|
||||
|
||||
- name: Fail if the Ceph OSD disks have already a partition
|
||||
fail:
|
||||
msg: >
|
||||
The physical disk {{ item.item }} already has a partition.
|
||||
Ensure that each disk in 'ceph_disks' does not have any partitions.
|
||||
with_items: "{{ disk_osd_info.results }}"
|
||||
when:
|
||||
- item.partitions | length > 0
|
||||
- not item.partitions.0.name.startswith('KOLLA_CEPH')
|
||||
loop_control:
|
||||
label: "{{item.item}}"
|
||||
|
||||
- name: Fail if the Ceph journal disks have already a partition
|
||||
fail:
|
||||
msg: >
|
||||
The physical disk {{ item.item }} already has a partition.
|
||||
Ensure that each disk in 'ceph_disks' does not have any partitions.
|
||||
with_items: "{{ disk_journal_info.results }}"
|
||||
when:
|
||||
- item is not skipped
|
||||
- item.partitions | length > 0
|
||||
- not item.partitions.0.name.startswith('KOLLA_CEPH')
|
||||
loop_control:
|
||||
label: "{{item.item}}"
|
||||
|
||||
- name: Create tag partition for Ceph OSD
|
||||
become: True
|
||||
parted:
|
||||
device: "{{ item.item.osd }}"
|
||||
number: 1
|
||||
label: gpt
|
||||
name: "{{ part_label }}"
|
||||
state: present
|
||||
with_items: "{{ disk_osd_info.results }}"
|
||||
when: item.partitions | length == 0
|
||||
loop_control:
|
||||
label: "{{item.item}}"
|
||||
vars:
|
||||
part_label: "{% if item.item.journal is defined %}{{ part_label_with_journal }}{% else %}KOLLA_CEPH_OSD_BOOTSTRAP{% endif %}"
|
||||
part_label_with_journal: "KOLLA_CEPH_OSD_BOOTSTRAP_{{ (osd_id | hash('md5'))[:9] }}"
|
||||
osd_id: "{{ item.item.osd | basename }}{{ ansible_hostname }}"
|
||||
|
||||
- name: Create tag partition for Ceph external journal
|
||||
become: True
|
||||
parted:
|
||||
device: "{{ item.item.journal }}"
|
||||
number: 1
|
||||
label: gpt
|
||||
name: "{{ part_label }}"
|
||||
state: present
|
||||
with_items: "{{ disk_journal_info.results }}"
|
||||
when:
|
||||
- item is not skipped
|
||||
- item.partitions | length == 0
|
||||
loop_control:
|
||||
label: "{{item.item}}"
|
||||
vars:
|
||||
part_label: "KOLLA_CEPH_OSD_BOOTSTRAP_{{ (osd_id | hash('md5'))[:9] }}_J"
|
||||
osd_id: "{{ item.item.osd | basename }}{{ ansible_hostname }}"
|
||||
|
@ -1,6 +0,0 @@
|
||||
---
|
||||
# NOTE: Use import_tasks here, since tags are not applied to tasks included via
|
||||
# include_tasks.
|
||||
- import_tasks: config.yml
|
||||
tags:
|
||||
- config
|
@ -1,14 +0,0 @@
|
||||
---
|
||||
- import_playbook: test-no-journal.yml
|
||||
- import_playbook: test-journal.yml
|
||||
- import_playbook: test-bootstrapped-journal.yml
|
||||
- import_playbook: test-data-journal.yml
|
||||
|
||||
- hosts: localhost
|
||||
connection: local
|
||||
tasks:
|
||||
- name: Fail if any tests failed
|
||||
fail:
|
||||
msg: >
|
||||
Test failures: {{ test_failures }}
|
||||
when: test_failures is defined
|
@ -1,114 +0,0 @@
|
||||
---
|
||||
# Test case with an OSD and external journal that have already been tagged by
|
||||
# kayobe with the kolla-ansible bootstrap label, but have not yet been
|
||||
# converted to use the in-use label.
|
||||
|
||||
- hosts: localhost
|
||||
connection: local
|
||||
tasks:
|
||||
- name: Allocate a temporary file for a fake OSD
|
||||
tempfile:
|
||||
register: osd_tempfile
|
||||
|
||||
- name: Allocate a temporary file for a fake journal
|
||||
tempfile:
|
||||
register: journal_tempfile
|
||||
|
||||
- name: Allocate a fake OSD file
|
||||
command: fallocate -l 10M {{ osd_tempfile.path }}
|
||||
|
||||
- name: Allocate a fake journal file
|
||||
command: fallocate -l 10M {{ journal_tempfile.path }}
|
||||
|
||||
- name: Create tag partition for the fake OSD
|
||||
become: True
|
||||
parted:
|
||||
device: "{{ osd_tempfile.path }}"
|
||||
number: 1
|
||||
label: gpt
|
||||
name: "{{ part_label }}"
|
||||
state: present
|
||||
vars:
|
||||
part_label: "KOLLA_CEPH_OSD_BOOTSTRAP_{{ (osd_id | hash('md5'))[:9] }}"
|
||||
osd_id: "{{ osd_tempfile.path | basename }}{{ ansible_hostname }}"
|
||||
|
||||
- name: Create tag partition for the fake journal
|
||||
become: True
|
||||
parted:
|
||||
device: "{{ journal_tempfile.path }}"
|
||||
number: 1
|
||||
label: gpt
|
||||
name: "{{ part_label }}"
|
||||
state: present
|
||||
vars:
|
||||
part_label: "KOLLA_CEPH_OSD_BOOTSTRAP_{{ (osd_id | hash('md5'))[:9] }}_J"
|
||||
osd_id: "{{ osd_tempfile.path | basename }}{{ ansible_hostname }}"
|
||||
|
||||
- block:
|
||||
- name: Test the kolla-ceph role
|
||||
include_role:
|
||||
name: ../../kolla-ceph
|
||||
vars:
|
||||
ceph_disks:
|
||||
- osd: "{{ osd_tempfile.path }}"
|
||||
journal: "{{ journal_tempfile.path }}"
|
||||
|
||||
- name: Get name of fake OSD partition
|
||||
parted:
|
||||
device: "{{ osd_tempfile.path }}"
|
||||
register: "disk_osd_info"
|
||||
become: True
|
||||
|
||||
- name: Validate number of OSD partitions
|
||||
assert:
|
||||
that: disk_osd_info.partitions | length == 1
|
||||
msg: >
|
||||
Number of OSD partitions is not correct. Expected 1,
|
||||
actual {{ disk_osd_info.partitions | length }}
|
||||
|
||||
- name: Validate OSD tag is present
|
||||
assert:
|
||||
that: "disk_osd_info.partitions.0.name == expected"
|
||||
msg: >
|
||||
Name of OSD partition is not correct. Expected {{ expected }},
|
||||
actual {{ disk_osd_info.partitions.0.name }}.
|
||||
vars:
|
||||
expected: "{{ 'KOLLA_CEPH_OSD_BOOTSTRAP_' ~ ((osd_tempfile.path | basename ~ ansible_hostname) | hash('md5'))[:9] }}"
|
||||
|
||||
- name: Get name of fake journal partition
|
||||
parted:
|
||||
device: "{{ journal_tempfile.path }}"
|
||||
register: "disk_journal_info"
|
||||
become: True
|
||||
|
||||
- name: Validate number of journal partitions
|
||||
assert:
|
||||
that: disk_journal_info.partitions | length == 1
|
||||
msg: >
|
||||
Number of journal partitions is not correct. Expected 1,
|
||||
actual {{ disk_journal_info.partitions | length }}
|
||||
|
||||
- name: Validate journal tag is present
|
||||
assert:
|
||||
that: "disk_journal_info.partitions.0.name == expected"
|
||||
msg: >
|
||||
Name of journal partition is not correct. Expected {{ expected }},
|
||||
actual {{ disk_journal_info.partitions.0.name }}.
|
||||
vars:
|
||||
expected: "{{ 'KOLLA_CEPH_OSD_BOOTSTRAP_' ~ (( osd_tempfile.path | basename ~ ansible_hostname) | hash('md5'))[:9] ~ '_J' }}"
|
||||
|
||||
always:
|
||||
- name: Remove the fake OSD file
|
||||
file:
|
||||
name: "{{ osd_tempfile.path }}"
|
||||
state: absent
|
||||
|
||||
- name: Remove the fake journal file
|
||||
file:
|
||||
name: "{{ journal_tempfile.path }}"
|
||||
state: absent
|
||||
|
||||
rescue:
|
||||
- name: Flag that a failure occurred
|
||||
set_fact:
|
||||
test_failures: "{{ test_failures | default(0) | int + 1 }}"
|
@ -1,113 +0,0 @@
|
||||
---
|
||||
# Test case with an OSD and external journal that have been converted by
|
||||
# kolla-ansible to use the in-use label.
|
||||
|
||||
- hosts: localhost
|
||||
connection: local
|
||||
tasks:
|
||||
- name: Allocate a temporary file for a fake OSD
|
||||
tempfile:
|
||||
register: osd_tempfile
|
||||
|
||||
- name: Allocate a temporary file for a fake journal
|
||||
tempfile:
|
||||
register: journal_tempfile
|
||||
|
||||
- name: Allocate a fake OSD file
|
||||
command: fallocate -l 10M {{ osd_tempfile.path }}
|
||||
|
||||
- name: Allocate a fake journal file
|
||||
command: fallocate -l 10M {{ journal_tempfile.path }}
|
||||
|
||||
- name: Create tag partition for the fake OSD
|
||||
become: True
|
||||
parted:
|
||||
device: "{{ osd_tempfile.path }}"
|
||||
number: 1
|
||||
label: gpt
|
||||
name: "{{ part_label }}"
|
||||
state: present
|
||||
vars:
|
||||
part_label: "KOLLA_CEPH_DATA_{{ (osd_id | hash('md5'))[:9]}}"
|
||||
osd_id: "{{ (osd_tempfile.path | basename ~ ansible_hostname) }}"
|
||||
|
||||
- name: Create tag partition for the fake journal
|
||||
become: True
|
||||
parted:
|
||||
device: "{{ journal_tempfile.path }}"
|
||||
number: 1
|
||||
label: gpt
|
||||
name: "{{ part_label }}"
|
||||
state: present
|
||||
vars:
|
||||
part_label: "KOLLA_CEPH_DATA_{{ (osd_id | hash('md5'))[:9] }}_J"
|
||||
osd_id: "{{ (osd_tempfile.path | basename ~ ansible_hostname) }}"
|
||||
|
||||
- block:
|
||||
- name: Test the kolla-ceph role
|
||||
include_role:
|
||||
name: ../../kolla-ceph
|
||||
vars:
|
||||
ceph_disks:
|
||||
- osd: "{{ osd_tempfile.path }}"
|
||||
journal: "{{ journal_tempfile.path }}"
|
||||
|
||||
- name: Get name of fake OSD partition
|
||||
parted:
|
||||
device: "{{ osd_tempfile.path }}"
|
||||
register: "disk_osd_info"
|
||||
become: True
|
||||
|
||||
- name: Validate number of OSD partitions
|
||||
assert:
|
||||
that: disk_osd_info.partitions | length == 1
|
||||
msg: >
|
||||
Number of OSD partitions is not correct. Expected 1,
|
||||
actual {{ disk_osd_info.partitions | length }}
|
||||
|
||||
- name: Validate OSD tag is present
|
||||
assert:
|
||||
that: "disk_osd_info.partitions.0.name == expected"
|
||||
msg: >
|
||||
Name of OSD partition is not correct. Expected {{ expected }},
|
||||
actual {{ disk_osd_info.partitions.0.name }}.
|
||||
vars:
|
||||
expected: "{{ 'KOLLA_CEPH_DATA_' ~ ((osd_tempfile.path | basename ~ ansible_hostname)| hash('md5'))[:9] }}"
|
||||
|
||||
- name: Get name of fake journal partition
|
||||
parted:
|
||||
device: "{{ journal_tempfile.path }}"
|
||||
register: "disk_journal_info"
|
||||
become: True
|
||||
|
||||
- name: Validate number of journal partitions
|
||||
assert:
|
||||
that: disk_journal_info.partitions | length == 1
|
||||
msg: >
|
||||
Number of journal partitions is not correct. Expected 1,
|
||||
actual {{ disk_journal_info.partitions | length }}
|
||||
|
||||
- name: Validate journal tag is present
|
||||
assert:
|
||||
that: "disk_journal_info.partitions.0.name == expected"
|
||||
msg: >
|
||||
Name of journal partition is not correct. Expected {{ expected }},
|
||||
actual {{ disk_journal_info.partitions.0.name }}.
|
||||
vars:
|
||||
expected: "{{ 'KOLLA_CEPH_DATA_' ~ ((osd_tempfile.path | basename ~ ansible_hostname)| hash('md5'))[:9] ~ '_J' }}"
|
||||
|
||||
always:
|
||||
- name: Remove the fake OSD file
|
||||
file:
|
||||
name: "{{ osd_tempfile.path }}"
|
||||
state: absent
|
||||
|
||||
- name: Remove the fake journal file
|
||||
file:
|
||||
name: "{{ journal_tempfile.path }}"
|
||||
state: absent
|
||||
|
||||
rescue:
|
||||
- name: Flag that a failure occurred
|
||||
set_fact:
|
||||
test_failures: "{{ test_failures | default(0) | int + 1 }}"
|
@ -1,89 +0,0 @@
|
||||
---
|
||||
# Test case with an OSD and external journal that have not yet been tagged by
|
||||
# kayobe with the kolla-ansible bootstrap label.
|
||||
|
||||
- hosts: localhost
|
||||
connection: local
|
||||
tasks:
|
||||
- name: Allocate a temporary file for a fake OSD
|
||||
tempfile:
|
||||
register: osd_tempfile
|
||||
|
||||
- name: Allocate a temporary file for a fake journal
|
||||
tempfile:
|
||||
register: journal_tempfile
|
||||
|
||||
- name: Allocate a fake OSD file
|
||||
command: fallocate -l 10M {{ osd_tempfile.path }}
|
||||
|
||||
- name: Allocate a fake journal file
|
||||
command: fallocate -l 10M {{ journal_tempfile.path }}
|
||||
|
||||
- block:
|
||||
- name: Test the kolla-ceph role
|
||||
include_role:
|
||||
name: ../../kolla-ceph
|
||||
vars:
|
||||
ceph_disks:
|
||||
- osd: "{{ osd_tempfile.path }}"
|
||||
journal: "{{ journal_tempfile.path }}"
|
||||
|
||||
- name: Get name of fake OSD partition
|
||||
parted:
|
||||
device: "{{ osd_tempfile.path }}"
|
||||
register: "disk_osd_info"
|
||||
become: True
|
||||
|
||||
- name: Validate number of OSD partitions
|
||||
assert:
|
||||
that: disk_osd_info.partitions | length == 1
|
||||
msg: >
|
||||
Number of OSD partitions is not correct. Expected 1,
|
||||
actual {{ disk_osd_info.partitions | length }}
|
||||
|
||||
- name: Validate OSD tag is present
|
||||
assert:
|
||||
that: "disk_osd_info.partitions.0.name == expected"
|
||||
msg: >
|
||||
Name of OSD partition is not correct. Expected {{ expected }},
|
||||
actual {{ disk_osd_info.partitions.0.name }}.
|
||||
vars:
|
||||
expected: "{{ 'KOLLA_CEPH_OSD_BOOTSTRAP_' ~ ((osd_tempfile.path | basename ~ ansible_hostname)| hash('md5'))[:9] }}"
|
||||
|
||||
- name: Get name of fake journal partition
|
||||
parted:
|
||||
device: "{{ journal_tempfile.path }}"
|
||||
register: "disk_journal_info"
|
||||
become: True
|
||||
|
||||
- name: Validate number of journal partitions
|
||||
assert:
|
||||
that: disk_journal_info.partitions | length == 1
|
||||
msg: >
|
||||
Number of journal partitions is not correct. Expected 1,
|
||||
actual {{ disk_journal_info.partitions | length }}
|
||||
|
||||
- name: Validate journal tag is present
|
||||
assert:
|
||||
that: "disk_journal_info.partitions.0.name == expected"
|
||||
msg: >
|
||||
Name of journal partition is not correct. Expected {{ expected }},
|
||||
actual {{ disk_journal_info.partitions.0.name }}.
|
||||
vars:
|
||||
expected: "{{ 'KOLLA_CEPH_OSD_BOOTSTRAP_' ~ ((osd_tempfile.path | basename ~ ansible_hostname)| hash('md5'))[:9] ~ '_J' }}"
|
||||
|
||||
always:
|
||||
- name: Remove the fake OSD file
|
||||
file:
|
||||
name: "{{ osd_tempfile.path }}"
|
||||
state: absent
|
||||
|
||||
- name: Remove the fake journal file
|
||||
file:
|
||||
name: "{{ journal_tempfile.path }}"
|
||||
state: absent
|
||||
|
||||
rescue:
|
||||
- name: Flag that a failure occurred
|
||||
set_fact:
|
||||
test_failures: "{{ test_failures | default(0) | int + 1 }}"
|
@ -1,50 +0,0 @@
|
||||
---
|
||||
# Test case with an OSD and no external journal that has not yet been tagged by
|
||||
# kayobe with the kolla-ansible bootstrap label.
|
||||
|
||||
- hosts: localhost
|
||||
connection: local
|
||||
tasks:
|
||||
- name: Allocate a temporary file for a fake OSD
|
||||
tempfile:
|
||||
register: tempfile
|
||||
|
||||
- name: Allocate a fake OSD file
|
||||
command: fallocate -l 10M {{ tempfile.path }}
|
||||
|
||||
- block:
|
||||
- name: Test the kolla-ceph role
|
||||
include_role:
|
||||
name: ../../kolla-ceph
|
||||
vars:
|
||||
ceph_disks:
|
||||
- osd: "{{ tempfile.path }}"
|
||||
|
||||
- name: Get name of fake partition
|
||||
parted:
|
||||
device: "{{ tempfile.path }}"
|
||||
register: "disk_osd_info"
|
||||
become: True
|
||||
|
||||
- name: Validate number of partition
|
||||
assert:
|
||||
that: disk_osd_info.partitions | length == 1
|
||||
msg: >
|
||||
Number of partition is not correct.
|
||||
|
||||
- name: Validate OSD tag is present
|
||||
assert:
|
||||
that: "disk_osd_info.partitions.0.name == 'KOLLA_CEPH_OSD_BOOTSTRAP'"
|
||||
msg: >
|
||||
Name of partition is not correct.
|
||||
|
||||
always:
|
||||
- name: Remove the fake OSD file
|
||||
file:
|
||||
name: "{{ tempfile.path }}"
|
||||
state: absent
|
||||
|
||||
rescue:
|
||||
- name: Flag that a failure occurred
|
||||
set_fact:
|
||||
test_failures: "{{ test_failures | default(0) | int + 1 }}"
|
@ -35,15 +35,6 @@ kolla_enable_ceilometer:
|
||||
# Free form extra configuration to append to ceilometer.conf.
|
||||
kolla_extra_ceilometer:
|
||||
|
||||
###############################################################################
|
||||
# ceph configuration.
|
||||
|
||||
# Whether to enable ceph.
|
||||
kolla_enable_ceph:
|
||||
|
||||
# Free form extra configuration to append to ceph.conf.
|
||||
kolla_extra_ceph:
|
||||
|
||||
###############################################################################
|
||||
# cinder configuration.
|
||||
|
||||
|
@ -37,8 +37,7 @@ def test_service_config_directory(host, path):
|
||||
|
||||
@pytest.mark.parametrize(
|
||||
'path',
|
||||
['ceph',
|
||||
'cinder',
|
||||
['cinder',
|
||||
'cloudkitty',
|
||||
'designate',
|
||||
'glance',
|
||||
|
@ -22,10 +22,6 @@ provisioner:
|
||||
kolla_extra_ceilometer: |
|
||||
[extra-ceilometer.conf]
|
||||
foo=bar
|
||||
kolla_enable_ceph: true
|
||||
kolla_extra_ceph: |
|
||||
[extra-ceph.conf]
|
||||
foo=bar
|
||||
kolla_enable_cinder: true
|
||||
kolla_extra_cinder: |
|
||||
[extra-cinder.conf]
|
||||
|
@ -27,8 +27,7 @@ testinfra_hosts = testinfra.utils.ansible_runner.AnsibleRunner(
|
||||
|
||||
@pytest.mark.parametrize(
|
||||
'path',
|
||||
['ceph',
|
||||
'cinder',
|
||||
['cinder',
|
||||
'cloudkitty',
|
||||
'designate',
|
||||
'fluentd/filter',
|
||||
@ -61,8 +60,7 @@ def test_service_config_directory(host, path):
|
||||
|
||||
@pytest.mark.parametrize(
|
||||
'path',
|
||||
['ceph.conf',
|
||||
'cinder.conf',
|
||||
['cinder.conf',
|
||||
'cloudkitty.conf',
|
||||
'designate.conf',
|
||||
'galera.cnf',
|
||||
|
@ -15,7 +15,6 @@
|
||||
with_items:
|
||||
- { src: blazar.conf.j2, dest: blazar.conf, enabled: "{{ kolla_enable_blazar }}" }
|
||||
- { src: ceilometer.conf.j2, dest: ceilometer.conf, enabled: "{{ kolla_enable_ceilometer }}" }
|
||||
- { src: ceph.conf.j2, dest: ceph.conf, enabled: "{{ kolla_enable_ceph }}" }
|
||||
- { src: cinder.conf.j2, dest: cinder.conf, enabled: "{{ kolla_enable_cinder }}" }
|
||||
- { src: cloudkitty.conf.j2, dest: cloudkitty.conf, enabled: "{{ kolla_enable_cloudkitty }}" }
|
||||
- { src: designate.conf.j2, dest: designate.conf, enabled: "{{ kolla_enable_designate }}" }
|
||||
|
@ -1,9 +0,0 @@
|
||||
# {{ ansible_managed }}
|
||||
|
||||
{% if kolla_extra_ceph %}
|
||||
#######################
|
||||
# Extra configuration
|
||||
#######################
|
||||
|
||||
{{ kolla_extra_ceph }}
|
||||
{% endif %}
|
@ -24,11 +24,6 @@ kolla_openstack_custom_config:
|
||||
dest: "{{ kolla_node_custom_config_path }}/ceilometer"
|
||||
patterns: "*"
|
||||
enabled: "{{ kolla_enable_ceilometer }}"
|
||||
# Ceph.
|
||||
- src: "{{ kolla_extra_config_path }}/ceph"
|
||||
dest: "{{ kolla_node_custom_config_path }}/ceph"
|
||||
patterns: "*"
|
||||
enabled: "{{ kolla_enable_ceph }}"
|
||||
# Cinder.
|
||||
- src: "{{ kolla_extra_config_path }}/cinder"
|
||||
dest: "{{ kolla_node_custom_config_path }}/cinder"
|
||||
|
@ -674,32 +674,3 @@ Ansible's containers do), but may be necessary when building images.
|
||||
Docker's live restore feature can be configured via
|
||||
``docker_daemon_live_restore``, although it is disabled by default due to
|
||||
issues observed.
|
||||
|
||||
Ceph Block Devices
|
||||
==================
|
||||
*tags:*
|
||||
| ``kolla-ceph``
|
||||
|
||||
If using Kolla Ansible to deploy Ceph, some preparation of block devices is
|
||||
required. The list of disks to configure for use by Ceph is specified via
|
||||
``ceph_disks``. This is mapped to the following variables:
|
||||
|
||||
* ``compute_ceph_disks``
|
||||
* ``controller_ceph_disks``
|
||||
* ``storage_ceph_disks``
|
||||
|
||||
The format of the variable is a list of dict/mapping objects. Each mapping
|
||||
should contain an ``osd`` item that defines the full path to a block device to
|
||||
use for data. Optionally, each mapping may contain a ``journal`` item that
|
||||
specifies the full path to a block device to use for journal data.
|
||||
|
||||
The following example defines two OSDs for use by controllers, one of which has
|
||||
a journal:
|
||||
|
||||
.. code-block:: yaml
|
||||
:caption: ``controller.yml``
|
||||
|
||||
controller_ceph_disks:
|
||||
- osd: /dev/sdb
|
||||
- osd: /dev/sdc
|
||||
journal: /dev/sdd
|
||||
|
@ -427,8 +427,6 @@ which files are supported.
|
||||
``blazar/*`` Extended Blazar configuration.
|
||||
``ceilometer.conf`` Ceilometer configuration.
|
||||
``ceilometer/*`` Extended Ceilometer configuration.
|
||||
``ceph.conf`` Ceph configuration.
|
||||
``ceph/*`` Extended Ceph configuration.
|
||||
``cinder.conf`` Cinder configuration.
|
||||
``cinder/*`` Extended Cinder configuration.
|
||||
``cloudkitty.conf`` CloudKitty configuration.
|
||||
|
@ -506,12 +506,6 @@ Storage network (``storage_net_name``)
|
||||
Name of the network used to carry storage data traffic.
|
||||
Storage management network (``storage_mgmt_net_name``)
|
||||
Name of the network used to carry storage management traffic.
|
||||
Ceph storage network (``ceph_storage_net_name``)
|
||||
Name of the network used to carry Ceph storage data traffic.
|
||||
Defaults to the storage network (``storage_net_name``).
|
||||
Ceph storage management network (``ceph_storage_mgmt_net_name``)
|
||||
Name of the network used to carry storage management traffic.
|
||||
Defaults to the storage management network (``storage_mgmt_net_name``)
|
||||
Swift storage network (``swift_storage_net_name``)
|
||||
Name of the network used to carry Swift storage data traffic.
|
||||
Defaults to the storage network (``storage_net_name``).
|
||||
@ -544,8 +538,6 @@ To configure network roles in a system with two networks, ``example1`` and
|
||||
external_net_name: example2
|
||||
storage_net_name: example2
|
||||
storage_mgmt_net_name: example2
|
||||
ceph_storage_net_name: example2
|
||||
ceph_storage_mgmt_net_name: example2
|
||||
swift_storage_net_name: example2
|
||||
swift_replication_net_name: example2
|
||||
inspection_net_name: example2
|
||||
@ -789,8 +781,8 @@ By default, the storage hosts are attached to the following networks:
|
||||
* storage network
|
||||
* storage management network
|
||||
|
||||
In addition, if Ceph or Swift is enabled, they can also be attached to the Ceph and Swift
|
||||
mangagment and replication networks.
|
||||
In addition, if Swift is enabled, they can also be attached to the Swift
|
||||
management and replication networks.
|
||||
|
||||
Virtualised Compute Hosts
|
||||
-------------------------
|
||||
|
@ -59,7 +59,6 @@ hosts in the ``monitoring`` group.
|
||||
``ansible_user`` Username with which to access the host via SSH.
|
||||
``bootstrap_user`` Username with which to access the host before
|
||||
``ansible_user`` is configured.
|
||||
``ceph_disks`` List of Ceph disks.
|
||||
``lvm_groups`` List of LVM volume groups to configure. See
|
||||
`mrlesmithjr.manage-lvm role
|
||||
<https://galaxy.ansible.com/mrlesmithjr/manage-lvm/>`_
|
||||
@ -189,7 +188,6 @@ providing the necessary variables for a control plane host.
|
||||
|
||||
ansible_user: "{{ kayobe_ansible_user }}"
|
||||
bootstrap_user: "{{ controller_bootstrap_user }}"
|
||||
ceph_disks: "{{ controller_ceph_disks }}"
|
||||
lvm_groups: "{{ controller_lvm_groups }}"
|
||||
mdadm_arrays: "{{ controller_mdadm_arrays }}"
|
||||
network_interfaces: "{{ controller_network_host_network_interfaces }}"
|
||||
|
@ -1,11 +0,0 @@
|
||||
---
|
||||
###############################################################################
|
||||
# OpenStack Ceph configuration.
|
||||
|
||||
# Ansible host pattern matching hosts on which Ceph storage services
|
||||
# are deployed. The default is to use hosts in the 'storage' group.
|
||||
#ceph_hosts:
|
||||
|
||||
###############################################################################
|
||||
# Dummy variable to allow Ansible to accept this file.
|
||||
workaround_ansible_issue_8743: yes
|
@ -89,16 +89,6 @@
|
||||
# Filesystem for docker volumes LVM backing volume. ext4 allows for shrinking.
|
||||
#compute_lvm_group_data_lv_docker_volumes_fs:
|
||||
|
||||
###############################################################################
|
||||
# Compute node Ceph configuration.
|
||||
|
||||
# List of Ceph disks.
|
||||
# The format is a list of dict like :
|
||||
# - { osd: "/dev/sdb", journal: "/dev/sdc" }
|
||||
# - { osd: "/dev/sdd" }
|
||||
# Journal variable is not mandatory.
|
||||
#compute_ceph_disks:
|
||||
|
||||
###############################################################################
|
||||
# Compute node sysctl configuration.
|
||||
|
||||
|
@ -92,16 +92,6 @@
|
||||
# Filesystem for docker volumes LVM backing volume. ext4 allows for shrinking.
|
||||
#controller_lvm_group_data_lv_docker_volumes_fs:
|
||||
|
||||
###############################################################################
|
||||
# Controller node Ceph configuration.
|
||||
|
||||
# List of Ceph disks.
|
||||
# The format is a list of dict like :
|
||||
# - { osd: "/dev/sdb", journal: "/dev/sdc" }
|
||||
# - { osd: "/dev/sdd" }
|
||||
# Journal variable is not mandatory.
|
||||
#controller_ceph_disks:
|
||||
|
||||
###############################################################################
|
||||
# Controller node sysctl configuration.
|
||||
|
||||
|
@ -22,11 +22,6 @@
|
||||
# storage_net_bridge_ports:
|
||||
# storage_net_bond_slaves:
|
||||
|
||||
# Ceph storage network IP information.
|
||||
# ceph_storage_net_interface:
|
||||
# ceph_storage_net_bridge_ports:
|
||||
# ceph_storage_net_bond_slaves:
|
||||
|
||||
###############################################################################
|
||||
# Dummy variable to allow Ansible to accept this file.
|
||||
workaround_ansible_issue_8743: yes
|
||||
|
@ -27,11 +27,6 @@
|
||||
# storage_net_bridge_ports:
|
||||
# storage_net_bond_slaves:
|
||||
|
||||
# Storage network IP information.
|
||||
# ceph_storage_net_interface:
|
||||
# ceph_storage_net_bridge_ports:
|
||||
# ceph_storage_net_bond_slaves:
|
||||
|
||||
# Storage management network IP information.
|
||||
# swift_storage_net_interface:
|
||||
# swift_storage_net_bridge_ports:
|
||||
|
@ -22,16 +22,6 @@
|
||||
# storage_mgmt_net_bridge_ports:
|
||||
# storage_mgmt_net_bond_slaves:
|
||||
|
||||
# Ceph storage network IP information.
|
||||
# ceph_storage_net_interface:
|
||||
# ceph_storage_net_bridge_ports:
|
||||
# ceph_storage_net_bond_slaves:
|
||||
|
||||
# Ceph storage management network IP information.
|
||||
# ceph_storage_mgmt_net_interface:
|
||||
# ceph_storage_mgmt_net_bridge_ports:
|
||||
# ceph_storage_mgmt_net_bond_slaves:
|
||||
|
||||
# Swift storage network IP information.
|
||||
# swift_storage_net_interface:
|
||||
# swift_storage_net_bridge_ports:
|
||||
|
@ -181,10 +181,6 @@
|
||||
#kolla_enable_cadf_notifications:
|
||||
#kolla_enable_ceilometer:
|
||||
#kolla_enable_central_logging:
|
||||
#kolla_enable_ceph:
|
||||
#kolla_enable_ceph_mds:
|
||||
#kolla_enable_ceph_nfs:
|
||||
#kolla_enable_ceph_rgw:
|
||||
# The chrony container is disabled by default on CentOS 7 because we enable an
|
||||
# NTP daemon on the host. On CentOS 8 the chrony container is enabled by
|
||||
# default because the NTP daemon is not supported. Setting this to true will
|
||||
|
@ -45,12 +45,6 @@
|
||||
# Name of the network used to carry storage management traffic.
|
||||
#storage_mgmt_net_name:
|
||||
|
||||
# Name of the network used to carry ceph storage data traffic.
|
||||
#ceph_storage_net_name:
|
||||
|
||||
# Name of the network used to carry ceph storage management traffic.
|
||||
#ceph_storage_mgmt_net_name:
|
||||
|
||||
# Name of the network used to carry swift storage data traffic.
|
||||
#swift_storage_net_name:
|
||||
|
||||
|
@ -18,11 +18,6 @@
|
||||
# List of extra networks to which storage nodes are attached.
|
||||
#storage_extra_network_interfaces:
|
||||
|
||||
# Whether this host requires access to Ceph networks.
|
||||
#storage_needs_ceph_network:
|
||||
|
||||
#storage_needs_ceph_mgmt_network:
|
||||
|
||||
# Whether this host requires access to Swift networks.
|
||||
#storage_needs_swift_network:
|
||||
|
||||
@ -99,16 +94,6 @@
|
||||
# Filesystem for docker volumes LVM backing volume. ext4 allows for shrinking.
|
||||
#storage_lvm_group_data_lv_docker_volumes_fs:
|
||||
|
||||
###############################################################################
|
||||
# Storage node Ceph configuration.
|
||||
|
||||
# List of Ceph disks.
|
||||
# The format is a list of dict like :
|
||||
# - { osd: "/dev/sdb", journal: "/dev/sdc" }
|
||||
# - { osd: "/dev/sdd" }
|
||||
# Journal variable is not mandatory.
|
||||
#storage_ceph_disks:
|
||||
|
||||
###############################################################################
|
||||
# Storage node sysctl configuration.
|
||||
|
||||
|
@ -977,7 +977,7 @@ class OvercloudHostConfigure(KollaAnsibleMixin, KayobeAnsibleMixin, VaultMixin,
|
||||
# Further kayobe playbooks.
|
||||
playbooks = _build_playbook_list(
|
||||
"pip", "kolla-target-venv", "kolla-host",
|
||||
"docker", "ceph-block-devices", "swift-block-devices")
|
||||
"docker", "swift-block-devices")
|
||||
self.run_kayobe_playbooks(parsed_args, playbooks,
|
||||
extra_vars=extra_vars, limit="overcloud")
|
||||
|
||||
|
@ -1156,8 +1156,6 @@ class TestCase(unittest.TestCase):
|
||||
"ansible", "kolla-target-venv.yml"),
|
||||
utils.get_data_files_path("ansible", "kolla-host.yml"),
|
||||
utils.get_data_files_path("ansible", "docker.yml"),
|
||||
utils.get_data_files_path(
|
||||
"ansible", "ceph-block-devices.yml"),
|
||||
utils.get_data_files_path(
|
||||
"ansible", "swift-block-devices.yml"),
|
||||
],
|
||||
|
@ -0,0 +1,8 @@
|
||||
---
|
||||
upgrade:
|
||||
- |
|
||||
Drops support for Kolla Ceph deployment.
|
||||
Kayobe follows upstream decision of Kolla and Kolla Ansible.
|
||||
Please use other means of Ceph deployment.
|
||||
Please note Ceph backends will still work if using external Ceph
|
||||
cluster.
|
Loading…
Reference in New Issue
Block a user