[CEPH] OSH: use loopback devices for ceph osds.
- This is to make use of loopback devices for ceph osds since support for directory backed osds going to depricate. - Move to bluestore from filestore for ceph-osds. Change-Id: Ia95c9ceb81f7d253dd582a2e753a6ed8fe60a04d
This commit is contained in:
parent
84b27c895d
commit
a385c18176
@ -54,6 +54,11 @@ containers before removing the directories used on the host by pods.
|
|||||||
sudo rm -rf /var/lib/nova/*
|
sudo rm -rf /var/lib/nova/*
|
||||||
sudo rm -rf /var/lib/libvirt/*
|
sudo rm -rf /var/lib/libvirt/*
|
||||||
sudo rm -rf /etc/libvirt/qemu/*
|
sudo rm -rf /etc/libvirt/qemu/*
|
||||||
|
#NOTE(chinasubbareddy) cleanup LVM volume groups in case of disk backed ceph osd deployments
|
||||||
|
for VG in `vgs|grep -v VG|grep -i ceph|awk '{print $1}'`; do
|
||||||
|
echo $VG
|
||||||
|
vgremove -y $VG
|
||||||
|
done
|
||||||
|
|
||||||
# NOTE(portdirect): Clean up mounts left behind by kubernetes pods
|
# NOTE(portdirect): Clean up mounts left behind by kubernetes pods
|
||||||
sudo findmnt --raw | awk '/^\/var\/lib\/kubelet\/pods/ { print $1 }' | xargs -r -L1 -P16 sudo umount -f -l
|
sudo findmnt --raw | awk '/^\/var\/lib\/kubelet\/pods/ { print $1 }' | xargs -r -L1 -P16 sudo umount -f -l
|
||||||
|
@ -104,11 +104,20 @@ Alternatively, this step can be performed by running the script directly:
|
|||||||
|
|
||||||
OSH_DEPLOY_MULTINODE=True ./tools/deployment/component/common/ingress.sh
|
OSH_DEPLOY_MULTINODE=True ./tools/deployment/component/common/ingress.sh
|
||||||
|
|
||||||
|
Create loopback devices for CEPH
|
||||||
|
--------------------------------
|
||||||
|
|
||||||
|
Create two loopback devices for ceph as one disk for OSD data and other disk for
|
||||||
|
block DB and block WAL.
|
||||||
|
|
||||||
|
.. code-block:: shell
|
||||||
|
|
||||||
|
ansible all -i /opt/openstack-helm-infra/tools/gate/devel/multinode-inventory.yaml -m shell -s -a "/opt/openstack-helm/tools/deployment/common/setup-ceph-loopback-device.sh"
|
||||||
|
|
||||||
Deploy Ceph
|
Deploy Ceph
|
||||||
-----------
|
-----------
|
||||||
|
|
||||||
The script below configures Ceph to use filesystem directory-based storage.
|
The script below configures Ceph to use loopback devices created in previous step as backend for ceph osds.
|
||||||
To configure a custom block device-based backend, please refer
|
To configure a custom block device-based backend, please refer
|
||||||
to the ``ceph-osd`` `values.yaml <https://github.com/openstack/openstack-helm/blob/master/ceph-osd/values.yaml>`_.
|
to the ``ceph-osd`` `values.yaml <https://github.com/openstack/openstack-helm/blob/master/ceph-osd/values.yaml>`_.
|
||||||
|
|
||||||
|
13
tools/deployment/common/setup-ceph-loopback-device.sh
Executable file
13
tools/deployment/common/setup-ceph-loopback-device.sh
Executable file
@ -0,0 +1,13 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
set -xe
|
||||||
|
|
||||||
|
sudo df -lh
|
||||||
|
sudo lsblk
|
||||||
|
sudo mkdir -p /var/lib/openstack-helm/ceph
|
||||||
|
sudo truncate -s 10G /var/lib/openstack-helm/ceph/ceph-osd-data-loopbackfile.img
|
||||||
|
sudo truncate -s 8G /var/lib/openstack-helm/ceph/ceph-osd-db-wal-loopbackfile.img
|
||||||
|
sudo losetup /dev/loop0 /var/lib/openstack-helm/ceph/ceph-osd-data-loopbackfile.img
|
||||||
|
sudo losetup /dev/loop1 /var/lib/openstack-helm/ceph/ceph-osd-db-wal-loopbackfile.img
|
||||||
|
#lets verify the devices
|
||||||
|
sudo df -lh
|
||||||
|
sudo lsblk
|
@ -154,17 +154,21 @@ conf:
|
|||||||
storage:
|
storage:
|
||||||
osd:
|
osd:
|
||||||
- data:
|
- data:
|
||||||
type: directory
|
type: bluestore
|
||||||
location: /var/lib/openstack-helm/ceph/osd/osd-one
|
location: /dev/loop0
|
||||||
journal:
|
block_db:
|
||||||
type: directory
|
location: /dev/loop1
|
||||||
location: /var/lib/openstack-helm/ceph/osd/journal-one
|
size: "5GB"
|
||||||
|
block_wal:
|
||||||
|
location: /dev/loop1
|
||||||
|
size: "2GB"
|
||||||
pod:
|
pod:
|
||||||
replicas:
|
replicas:
|
||||||
mds: 1
|
mds: 1
|
||||||
mgr: 1
|
mgr: 1
|
||||||
rgw: 1
|
rgw: 1
|
||||||
|
deploy:
|
||||||
|
tool: "ceph-volume"
|
||||||
|
|
||||||
EOF
|
EOF
|
||||||
|
|
||||||
|
@ -14,6 +14,18 @@
|
|||||||
|
|
||||||
set -xe
|
set -xe
|
||||||
|
|
||||||
|
# Create loop back devices for ceph osds.
|
||||||
|
sudo df -lh
|
||||||
|
sudo lsblk
|
||||||
|
sudo mkdir -p /var/lib/openstack-helm/ceph
|
||||||
|
sudo truncate -s 10G /var/lib/openstack-helm/ceph/ceph-osd-data-loopbackfile.img
|
||||||
|
sudo truncate -s 8G /var/lib/openstack-helm/ceph/ceph-osd-db-wal-loopbackfile.img
|
||||||
|
sudo losetup /dev/loop0 /var/lib/openstack-helm/ceph/ceph-osd-data-loopbackfile.img
|
||||||
|
sudo losetup /dev/loop1 /var/lib/openstack-helm/ceph/ceph-osd-db-wal-loopbackfile.img
|
||||||
|
# lets check the devices
|
||||||
|
sudo df -lh
|
||||||
|
sudo lsblk
|
||||||
|
|
||||||
#NOTE: Lint and package chart
|
#NOTE: Lint and package chart
|
||||||
export HELM_CHART_ROOT_PATH="${HELM_CHART_ROOT_PATH:="${OSH_INFRA_PATH:="../openstack-helm-infra"}"}"
|
export HELM_CHART_ROOT_PATH="${HELM_CHART_ROOT_PATH:="${OSH_INFRA_PATH:="../openstack-helm-infra"}"}"
|
||||||
for CHART in ceph-mon ceph-osd ceph-client ceph-provisioners; do
|
for CHART in ceph-mon ceph-osd ceph-client ceph-provisioners; do
|
||||||
@ -143,13 +155,16 @@ conf:
|
|||||||
replication: 1
|
replication: 1
|
||||||
percent_total_data: 34.8
|
percent_total_data: 34.8
|
||||||
storage:
|
storage:
|
||||||
osd:
|
osd:
|
||||||
- data:
|
- data:
|
||||||
type: directory
|
type: bluestore
|
||||||
location: /var/lib/openstack-helm/ceph/osd/osd-one
|
location: /dev/loop0
|
||||||
journal:
|
block_db:
|
||||||
type: directory
|
location: /dev/loop1
|
||||||
location: /var/lib/openstack-helm/ceph/osd/journal-one
|
size: "5GB"
|
||||||
|
block_wal:
|
||||||
|
location: /dev/loop1
|
||||||
|
size: "2GB"
|
||||||
|
|
||||||
pod:
|
pod:
|
||||||
replicas:
|
replicas:
|
||||||
|
@ -56,11 +56,14 @@ conf:
|
|||||||
storage:
|
storage:
|
||||||
osd:
|
osd:
|
||||||
- data:
|
- data:
|
||||||
type: directory
|
type: bluestore
|
||||||
location: /var/lib/openstack-helm/ceph/osd/osd-one
|
location: /dev/loop0
|
||||||
journal:
|
block_db:
|
||||||
type: directory
|
location: /dev/loop1
|
||||||
location: /var/lib/openstack-helm/ceph/osd/journal-one
|
size: "5GB"
|
||||||
|
block_wal:
|
||||||
|
location: /dev/loop1
|
||||||
|
size: "2GB"
|
||||||
storageclass:
|
storageclass:
|
||||||
cephfs:
|
cephfs:
|
||||||
provision_storage_class: false
|
provision_storage_class: false
|
||||||
|
@ -85,6 +85,7 @@
|
|||||||
container_distro_version: bionic
|
container_distro_version: bionic
|
||||||
gate_scripts:
|
gate_scripts:
|
||||||
- ./tools/deployment/common/install-packages.sh
|
- ./tools/deployment/common/install-packages.sh
|
||||||
|
- ./tools/deployment/common/setup-ceph-loopback-device.sh
|
||||||
- ./tools/deployment/common/deploy-k8s.sh
|
- ./tools/deployment/common/deploy-k8s.sh
|
||||||
- ./tools/deployment/common/setup-client.sh
|
- ./tools/deployment/common/setup-client.sh
|
||||||
- ./tools/scripts/tls/cert-manager.sh
|
- ./tools/scripts/tls/cert-manager.sh
|
||||||
@ -309,6 +310,7 @@
|
|||||||
feature_gates: apparmor
|
feature_gates: apparmor
|
||||||
gate_scripts:
|
gate_scripts:
|
||||||
- ./tools/deployment/common/install-packages.sh
|
- ./tools/deployment/common/install-packages.sh
|
||||||
|
- ./tools/deployment/common/setup-ceph-loopback-device.sh
|
||||||
- ./tools/deployment/common/deploy-k8s.sh
|
- ./tools/deployment/common/deploy-k8s.sh
|
||||||
- ./tools/deployment/common/setup-client.sh
|
- ./tools/deployment/common/setup-client.sh
|
||||||
- ./tools/deployment/component/ceph/ceph.sh
|
- ./tools/deployment/component/ceph/ceph.sh
|
||||||
@ -399,6 +401,7 @@
|
|||||||
feature_gates: netpol
|
feature_gates: netpol
|
||||||
gate_scripts:
|
gate_scripts:
|
||||||
- ./tools/deployment/common/install-packages.sh
|
- ./tools/deployment/common/install-packages.sh
|
||||||
|
- ./tools/deployment/common/setup-ceph-loopback-device.sh
|
||||||
- ./tools/deployment/common/deploy-k8s.sh
|
- ./tools/deployment/common/deploy-k8s.sh
|
||||||
- ./tools/deployment/common/setup-client.sh
|
- ./tools/deployment/common/setup-client.sh
|
||||||
- ./tools/deployment/component/ceph/ceph.sh
|
- ./tools/deployment/component/ceph/ceph.sh
|
||||||
|
Loading…
x
Reference in New Issue
Block a user