Merge "Add support for Nautilus release"

This commit is contained in:
Zuul 2019-08-19 19:37:14 +00:00 committed by Gerrit Code Review
commit 62155d10a8
3 changed files with 126 additions and 63 deletions

View File

@ -18,7 +18,6 @@ import ast
import json import json
import os import os
import subprocess import subprocess
import socket
import sys import sys
import uuid import uuid
@ -51,7 +50,6 @@ from charmhelpers.core.hookenv import (
application_version_set) application_version_set)
from charmhelpers.core.host import ( from charmhelpers.core.host import (
service_pause, service_pause,
service_restart,
mkdir, mkdir,
write_file, write_file,
rsync, rsync,
@ -298,7 +296,7 @@ def config_changed():
def get_mon_hosts(): def get_mon_hosts():
hosts = [] hosts = []
addr = get_public_addr() addr = get_public_addr()
hosts.append('{}:6789'.format(format_ipv6_addr(addr) or addr)) hosts.append(format_ipv6_addr(addr) or addr)
rel_ids = relation_ids('mon') rel_ids = relation_ids('mon')
if config('no-bootstrap'): if config('no-bootstrap'):
@ -308,8 +306,7 @@ def get_mon_hosts():
for unit in related_units(relid): for unit in related_units(relid):
addr = relation_get('ceph-public-address', unit, relid) addr = relation_get('ceph-public-address', unit, relid)
if addr is not None: if addr is not None:
hosts.append('{}:6789'.format( hosts.append(format_ipv6_addr(addr) or addr)
format_ipv6_addr(addr) or addr))
return sorted(hosts) return sorted(hosts)
@ -818,18 +815,6 @@ def upgrade_charm():
notify_rbd_mirrors() notify_rbd_mirrors()
@hooks.hook('start')
def start():
# In case we're being redeployed to the same machines, try
# to make sure everything is running as soon as possible.
if ceph.systemd():
service_restart('ceph-mon')
else:
service_restart('ceph-mon-all')
if cmp_pkgrevno('ceph', '12.0.0') >= 0:
service_restart('ceph-mgr@{}'.format(socket.gethostname()))
@hooks.hook('nrpe-external-master-relation-joined') @hooks.hook('nrpe-external-master-relation-joined')
@hooks.hook('nrpe-external-master-relation-changed') @hooks.hook('nrpe-external-master-relation-changed')
def update_nrpe_config(): def update_nrpe_config():

View File

@ -40,6 +40,7 @@ from charmhelpers.core.host import (
service_start, service_start,
service_stop, service_stop,
CompareHostReleases, CompareHostReleases,
write_file,
) )
from charmhelpers.core.hookenv import ( from charmhelpers.core.hookenv import (
cached, cached,
@ -82,7 +83,7 @@ QUORUM = [LEADER, PEON]
PACKAGES = ['ceph', 'gdisk', 'btrfs-tools', PACKAGES = ['ceph', 'gdisk', 'btrfs-tools',
'radosgw', 'xfsprogs', 'radosgw', 'xfsprogs',
'lvm2', 'parted'] 'lvm2', 'parted', 'smartmontools']
CEPH_KEY_MANAGER = 'ceph' CEPH_KEY_MANAGER = 'ceph'
VAULT_KEY_MANAGER = 'vault' VAULT_KEY_MANAGER = 'vault'
@ -950,13 +951,13 @@ def rescan_osd_devices():
udevadm_settle() udevadm_settle()
_client_admin_keyring = '/etc/ceph/ceph.client.admin.keyring'
_bootstrap_keyring = "/var/lib/ceph/bootstrap-osd/ceph.keyring" _bootstrap_keyring = "/var/lib/ceph/bootstrap-osd/ceph.keyring"
_upgrade_keyring = "/var/lib/ceph/osd/ceph.client.osd-upgrade.keyring" _upgrade_keyring = "/var/lib/ceph/osd/ceph.client.osd-upgrade.keyring"
def is_bootstrapped(): def is_bootstrapped():
return os.path.exists(_bootstrap_keyring) return os.path.exists(_client_admin_keyring)
def wait_for_bootstrap(): def wait_for_bootstrap():
@ -1259,7 +1260,23 @@ def systemd():
return CompareHostReleases(lsb_release()['DISTRIB_CODENAME']) >= 'vivid' return CompareHostReleases(lsb_release()['DISTRIB_CODENAME']) >= 'vivid'
def use_bluestore():
"""Determine whether bluestore should be used for OSD's
:returns: whether bluestore disk format should be used
:rtype: bool"""
if cmp_pkgrevno('ceph', '12.2.0') < 0:
return False
return config('bluestore')
def bootstrap_monitor_cluster(secret): def bootstrap_monitor_cluster(secret):
"""Bootstrap local ceph mon into the ceph cluster
:param secret: cephx secret to use for monitor authentication
:type secret: str
:raises: Exception if ceph mon cannot be bootstrapped
"""
hostname = socket.gethostname() hostname = socket.gethostname()
path = '/var/lib/ceph/mon/ceph-{}'.format(hostname) path = '/var/lib/ceph/mon/ceph-{}'.format(hostname)
done = '{}/done'.format(path) done = '{}/done'.format(path)
@ -1280,21 +1297,35 @@ def bootstrap_monitor_cluster(secret):
perms=0o755) perms=0o755)
# end changes for Ceph >= 0.61.3 # end changes for Ceph >= 0.61.3
try: try:
add_keyring_to_ceph(keyring, _create_monitor(keyring,
secret, secret,
hostname, hostname,
path, path,
done, done,
init_marker) init_marker)
_create_keyrings()
except: except:
raise raise
finally: finally:
os.unlink(keyring) os.unlink(keyring)
@retry_on_exception(3, base_delay=5) def _create_monitor(keyring, secret, hostname, path, done, init_marker):
def add_keyring_to_ceph(keyring, secret, hostname, path, done, init_marker): """Create monitor filesystem and enable and start ceph-mon process
:param keyring: path to temporary keyring on disk
:type keyring: str
:param secret: cephx secret to use for monitor authentication
:type: secret: str
:param hostname: hostname of the local unit
:type hostname: str
:param path: full path to ceph mon directory
:type path: str
:param done: full path to 'done' marker for ceph mon
:type done: str
:param init_marker: full path to 'init' marker for ceph mon
:type init_marker: str
"""
subprocess.check_call(['ceph-authtool', keyring, subprocess.check_call(['ceph-authtool', keyring,
'--create-keyring', '--name=mon.', '--create-keyring', '--name=mon.',
'--add-key={}'.format(secret), '--add-key={}'.format(secret),
@ -1310,39 +1341,72 @@ def add_keyring_to_ceph(keyring, secret, hostname, path, done, init_marker):
pass pass
if systemd(): if systemd():
subprocess.check_call(['systemctl', 'enable', 'ceph-mon']) if cmp_pkgrevno('ceph', '14.0.0') >= 0:
service_restart('ceph-mon') systemd_unit = 'ceph-mon@{}'.format(socket.gethostname())
else:
systemd_unit = 'ceph-mon'
subprocess.check_call(['systemctl', 'enable', systemd_unit])
service_restart(systemd_unit)
else: else:
service_restart('ceph-mon-all') service_restart('ceph-mon-all')
# NOTE(jamespage): Later ceph releases require explicit
# call to ceph-create-keys to setup the @retry_on_exception(3, base_delay=5)
# admin keys for the cluster; this command def _create_keyrings():
# will wait for quorum in the cluster before """Create keyrings for operation of ceph-mon units
# returning.
# NOTE(fnordahl): Explicitly run `ceph-crate-keys` for older :raises: Exception if keyrings cannot be created
# ceph releases too. This improves bootstrap """
# resilience as the charm will wait for if cmp_pkgrevno('ceph', '14.0.0') >= 0:
# presence of peer units before attempting # NOTE(jamespage): At Nautilus, keys are created by the
# to bootstrap. Note that charms deploying # monitors automatically and just need
# ceph-mon service should disable running of # exporting.
# `ceph-create-keys` service in init system. output = str(subprocess.check_output(
cmd = ['ceph-create-keys', '--id', hostname] [
if cmp_pkgrevno('ceph', '12.0.0') >= 0: 'sudo',
# NOTE(fnordahl): The default timeout in ceph-create-keys of 600 '-u', ceph_user(),
# seconds is not adequate. Increase timeout when 'ceph',
# timeout parameter available. For older releases '--name', 'mon.',
# we rely on retry_on_exception decorator. '--keyring',
# LP#1719436 '/var/lib/ceph/mon/ceph-{}/keyring'.format(
cmd.extend(['--timeout', '1800']) socket.gethostname()
subprocess.check_call(cmd) ),
_client_admin_keyring = '/etc/ceph/ceph.client.admin.keyring' 'auth', 'get', 'client.admin',
osstat = os.stat(_client_admin_keyring) ]).decode('UTF-8')).strip()
if not osstat.st_size: if not output:
# NOTE(fnordahl): Retry will fail as long as this file exists. # NOTE: key not yet created, raise exception and retry
# LP#1719436 raise Exception
os.remove(_client_admin_keyring) write_file(_client_admin_keyring, output,
raise Exception owner=ceph_user(), group=ceph_user(),
perms=0o400)
else:
# NOTE(jamespage): Later ceph releases require explicit
# call to ceph-create-keys to setup the
# admin keys for the cluster; this command
# will wait for quorum in the cluster before
# returning.
# NOTE(fnordahl): Explicitly run `ceph-create-keys` for older
# ceph releases too. This improves bootstrap
# resilience as the charm will wait for
# presence of peer units before attempting
# to bootstrap. Note that charms deploying
# ceph-mon service should disable running of
# `ceph-create-keys` service in init system.
cmd = ['ceph-create-keys', '--id', socket.gethostname()]
if cmp_pkgrevno('ceph', '12.0.0') >= 0:
# NOTE(fnordahl): The default timeout in ceph-create-keys of 600
# seconds is not adequate. Increase timeout when
# timeout parameter available. For older releases
# we rely on retry_on_exception decorator.
# LP#1719436
cmd.extend(['--timeout', '1800'])
subprocess.check_call(cmd)
osstat = os.stat(_client_admin_keyring)
if not osstat.st_size:
# NOTE(fnordahl): Retry will fail as long as this file exists.
# LP#1719436
os.remove(_client_admin_keyring)
raise Exception
def update_monfs(): def update_monfs():
@ -1555,7 +1619,7 @@ def _ceph_disk(dev, osd_format, osd_journal, encrypt=False, bluestore=False):
cmd.append(osd_format) cmd.append(osd_format)
# NOTE(jamespage): enable experimental bluestore support # NOTE(jamespage): enable experimental bluestore support
if cmp_pkgrevno('ceph', '10.2.0') >= 0 and bluestore: if use_bluestore():
cmd.append('--bluestore') cmd.append('--bluestore')
wal = get_devices('bluestore-wal') wal = get_devices('bluestore-wal')
if wal: if wal:
@ -1692,7 +1756,10 @@ def is_active_bluestore_device(dev):
return False return False
vg_name = lvm.list_lvm_volume_group(dev) vg_name = lvm.list_lvm_volume_group(dev)
lv_name = lvm.list_logical_volumes('vg_name={}'.format(vg_name))[0] try:
lv_name = lvm.list_logical_volumes('vg_name={}'.format(vg_name))[0]
except IndexError:
return False
block_symlinks = glob.glob('/var/lib/ceph/osd/ceph-*/block') block_symlinks = glob.glob('/var/lib/ceph/osd/ceph-*/block')
for block_candidate in block_symlinks: for block_candidate in block_symlinks:
@ -2700,6 +2767,14 @@ def dirs_need_ownership_update(service):
if (curr_owner == expected_owner) and (curr_group == expected_group): if (curr_owner == expected_owner) and (curr_group == expected_group):
continue continue
# NOTE(lathiat): when config_changed runs on reboot, the OSD might not
# yet be mounted or started, and the underlying directory the OSD is
# mounted to is expected to be owned by root. So skip the check. This
# may also happen for OSD directories for OSDs that were removed.
if (service == 'osd' and
not os.path.exists(os.path.join(child, 'magic'))):
continue
log('Directory "%s" needs its ownership updated' % child, DEBUG) log('Directory "%s" needs its ownership updated' % child, DEBUG)
return True return True
@ -2712,6 +2787,8 @@ UPGRADE_PATHS = collections.OrderedDict([
('hammer', 'jewel'), ('hammer', 'jewel'),
('jewel', 'luminous'), ('jewel', 'luminous'),
('luminous', 'mimic'), ('luminous', 'mimic'),
('luminous', 'nautilus'),
('mimic', 'nautilus'),
]) ])
# Map UCA codenames to ceph codenames # Map UCA codenames to ceph codenames
@ -2727,6 +2804,7 @@ UCA_CODENAME_MAP = {
'queens': 'luminous', 'queens': 'luminous',
'rocky': 'mimic', 'rocky': 'mimic',
'stein': 'mimic', 'stein': 'mimic',
'train': 'nautilus',
} }

View File

@ -619,8 +619,8 @@ class BootstrapSourceTestCase(test_utils.CharmTestCase):
self.relation_get.side_effect = rel_get_side_effect self.relation_get.side_effect = rel_get_side_effect
hosts = ceph_hooks.get_mon_hosts() hosts = ceph_hooks.get_mon_hosts()
self.assertEqual(hosts, [ self.assertEqual(hosts, [
'172.16.0.2:6789', '172.16.0.3:6789', '172.16.0.4:6789', '172.16.0.2', '172.16.0.3', '172.16.0.4',
'172.16.10.2:6789', '172.16.10.3:6789', '172.16.10.4:6789', '172.16.10.2', '172.16.10.3', '172.16.10.4',
]) ])