From 8b0b24d6c6c6c54185fce3cf68b8d3e56a9451ef Mon Sep 17 00:00:00 2001 From: Alex Kavanagh Date: Tue, 2 Jun 2020 14:22:18 +0100 Subject: [PATCH] Updates for 20.08 cycle start for groovy and libs - Adds groovy to the series in the metadata - Classic charms: sync charm-helpers. - Classic ceph based charms: also sync charms.ceph - Reactive charms: trigger a rebuild Change-Id: I56a59d0c4e72a35b7c4ac5d989e0d005fae20946 --- hooks/charmhelpers/contrib/openstack/utils.py | 2 +- .../contrib/storage/linux/ceph.py | 8 ++--- hooks/charmhelpers/fetch/snap.py | 2 +- lib/charms_ceph/utils.py | 3 +- metadata.yaml | 29 ++++++++++--------- 5 files changed, 23 insertions(+), 21 deletions(-) diff --git a/hooks/charmhelpers/contrib/openstack/utils.py b/hooks/charmhelpers/contrib/openstack/utils.py index e59e0d1e..fbf01561 100644 --- a/hooks/charmhelpers/contrib/openstack/utils.py +++ b/hooks/charmhelpers/contrib/openstack/utils.py @@ -225,7 +225,7 @@ SWIFT_CODENAMES = OrderedDict([ ('train', ['2.22.0', '2.23.0']), ('ussuri', - ['2.24.0']), + ['2.24.0', '2.25.0']), ]) # >= Liberty version->codename mapping diff --git a/hooks/charmhelpers/contrib/storage/linux/ceph.py b/hooks/charmhelpers/contrib/storage/linux/ceph.py index 95a0d82a..814d5c72 100644 --- a/hooks/charmhelpers/contrib/storage/linux/ceph.py +++ b/hooks/charmhelpers/contrib/storage/linux/ceph.py @@ -432,13 +432,13 @@ class ReplicatedPool(Pool): pool=self.name, name=self.app_name) except CalledProcessError: - log('Could not set app name for pool {}'.format(self.name, level=WARNING)) + log('Could not set app name for pool {}'.format(self.name), level=WARNING) if 'pg_autoscaler' in enabled_manager_modules(): try: enable_pg_autoscale(self.service, self.name) except CalledProcessError as e: log('Could not configure auto scaling for pool {}: {}'.format( - self.name, e, level=WARNING)) + self.name, e), level=WARNING) except CalledProcessError: raise @@ -504,7 +504,7 @@ class ErasurePool(Pool): pool=self.name, name=self.app_name) except CalledProcessError: - log('Could not set app name for pool {}'.format(self.name, level=WARNING)) + log('Could not set app name for pool {}'.format(self.name), level=WARNING) if nautilus_or_later: # Ensure we set the expected pool ratio update_pool(client=self.service, @@ -515,7 +515,7 @@ class ErasurePool(Pool): enable_pg_autoscale(self.service, self.name) except CalledProcessError as e: log('Could not configure auto scaling for pool {}: {}'.format( - self.name, e, level=WARNING)) + self.name, e), level=WARNING) except CalledProcessError: raise diff --git a/hooks/charmhelpers/fetch/snap.py b/hooks/charmhelpers/fetch/snap.py index 395836c7..fc70aa94 100644 --- a/hooks/charmhelpers/fetch/snap.py +++ b/hooks/charmhelpers/fetch/snap.py @@ -69,7 +69,7 @@ def _snap_exec(commands): .format(SNAP_NO_LOCK_RETRY_COUNT)) return_code = e.returncode log('Snap failed to acquire lock, trying again in {} seconds.' - .format(SNAP_NO_LOCK_RETRY_DELAY, level='WARN')) + .format(SNAP_NO_LOCK_RETRY_DELAY), level='WARN') sleep(SNAP_NO_LOCK_RETRY_DELAY) return return_code diff --git a/lib/charms_ceph/utils.py b/lib/charms_ceph/utils.py index e9b289a4..7a0cb7d4 100644 --- a/lib/charms_ceph/utils.py +++ b/lib/charms_ceph/utils.py @@ -1139,8 +1139,9 @@ osd_upgrade_caps = collections.OrderedDict([ ]) rbd_mirror_caps = collections.OrderedDict([ - ('mon', ['profile rbd']), + ('mon', ['profile rbd; allow r']), ('osd', ['profile rbd']), + ('mgr', ['allow r']), ]) diff --git a/metadata.yaml b/metadata.yaml index bf0ccf36..4f385ca5 100644 --- a/metadata.yaml +++ b/metadata.yaml @@ -6,22 +6,23 @@ provides: interface: nrpe-external-master scope: container tags: - - openstack - - storage - - file-servers - - misc +- openstack +- storage +- file-servers +- misc series: - - xenial - - bionic - - eoan - - focal - - trusty +- xenial +- bionic +- eoan +- focal +- trusty +- groovy description: | - Ceph is a distributed storage and network file system designed to provide - excellent performance, reliability, and scalability. - . - This charm provides the Ceph OSD personality for expanding storage capacity - within a ceph deployment. + Ceph is a distributed storage and network file system designed to provide + excellent performance, reliability, and scalability. + . + This charm provides the Ceph OSD personality for expanding storage capacity + within a ceph deployment. extra-bindings: public: cluster: