Updates for testing period for 20.01 release
* charm-helpers sync for classic charms * rebuild for reactive charms * ensure tox.ini is from release-tools * ensure requirements.txt files are from release-tools * On reactive charms: - ensure master branch for charms.openstack - ensure master branch for charm-helpers Change-Id: I99b9f3570549921b40c937a983c2624e254bc677
This commit is contained in:
parent
eac7bceb0c
commit
e586074e60
@ -139,10 +139,11 @@ define service {{
|
||||
"""{description}
|
||||
check_command check_nrpe!{command}
|
||||
servicegroups {nagios_servicegroup}
|
||||
{service_config_overrides}
|
||||
}}
|
||||
""")
|
||||
|
||||
def __init__(self, shortname, description, check_cmd):
|
||||
def __init__(self, shortname, description, check_cmd, max_check_attempts=None):
|
||||
super(Check, self).__init__()
|
||||
# XXX: could be better to calculate this from the service name
|
||||
if not re.match(self.shortname_re, shortname):
|
||||
@ -155,6 +156,7 @@ define service {{
|
||||
# The default is: illegal_object_name_chars=`~!$%^&*"|'<>?,()=
|
||||
self.description = description
|
||||
self.check_cmd = self._locate_cmd(check_cmd)
|
||||
self.max_check_attempts = max_check_attempts
|
||||
|
||||
def _get_check_filename(self):
|
||||
return os.path.join(NRPE.nrpe_confdir, '{}.cfg'.format(self.command))
|
||||
@ -216,12 +218,19 @@ define service {{
|
||||
nagios_servicegroups):
|
||||
self._remove_service_files()
|
||||
|
||||
if self.max_check_attempts:
|
||||
service_config_overrides = ' max_check_attempts {}'.format(
|
||||
self.max_check_attempts
|
||||
) # Note indentation is here rather than in the template to avoid trailing spaces
|
||||
else:
|
||||
service_config_overrides = '' # empty string to avoid printing 'None'
|
||||
templ_vars = {
|
||||
'nagios_hostname': hostname,
|
||||
'nagios_servicegroup': nagios_servicegroups,
|
||||
'description': self.description,
|
||||
'shortname': self.shortname,
|
||||
'command': self.command,
|
||||
'service_config_overrides': service_config_overrides,
|
||||
}
|
||||
nrpe_service_text = Check.service_template.format(**templ_vars)
|
||||
nrpe_service_file = self._get_service_filename(hostname)
|
||||
@ -327,6 +336,11 @@ class NRPE(object):
|
||||
nrpe_monitors[nrpecheck.shortname] = {
|
||||
"command": nrpecheck.command,
|
||||
}
|
||||
# If we were passed max_check_attempts, add that to the relation data
|
||||
try:
|
||||
nrpe_monitors[nrpecheck.shortname]['max_check_attempts'] = nrpecheck.max_check_attempts
|
||||
except AttributeError:
|
||||
pass
|
||||
|
||||
# update-status hooks are configured to firing every 5 minutes by
|
||||
# default. When nagios-nrpe-server is restarted, the nagios server
|
||||
|
@ -37,6 +37,7 @@ from charmhelpers.core.hookenv import (
|
||||
unit_get,
|
||||
log,
|
||||
DEBUG,
|
||||
ERROR,
|
||||
INFO,
|
||||
WARNING,
|
||||
leader_get,
|
||||
@ -69,11 +70,14 @@ class MySQLHelper(object):
|
||||
|
||||
def __init__(self, rpasswdf_template, upasswdf_template, host='localhost',
|
||||
migrate_passwd_to_leader_storage=True,
|
||||
delete_ondisk_passwd_file=True, user="root", password=None, port=None):
|
||||
delete_ondisk_passwd_file=True, user="root", password=None,
|
||||
port=None, connect_timeout=None):
|
||||
self.user = user
|
||||
self.host = host
|
||||
self.password = password
|
||||
self.port = port
|
||||
# default timeout of 30 seconds.
|
||||
self.connect_timeout = connect_timeout or 30
|
||||
|
||||
# Password file path templates
|
||||
self.root_passwd_file_template = rpasswdf_template
|
||||
@ -84,12 +88,18 @@ class MySQLHelper(object):
|
||||
self.delete_ondisk_passwd_file = delete_ondisk_passwd_file
|
||||
self.connection = None
|
||||
|
||||
def connect(self, user='root', password=None, host=None, port=None):
|
||||
def connect(self, user='root', password=None, host=None, port=None,
|
||||
connect_timeout=None):
|
||||
_connection_info = {
|
||||
"user": user or self.user,
|
||||
"passwd": password or self.password,
|
||||
"host": host or self.host
|
||||
}
|
||||
# set the connection timeout; for mysql8 it can hang forever, so some
|
||||
# timeout is required.
|
||||
timeout = connect_timeout or self.connect_timeout
|
||||
if timeout:
|
||||
_connection_info["connect_timeout"] = timeout
|
||||
# port cannot be None but we also do not want to specify it unless it
|
||||
# has been explicit set.
|
||||
port = port or self.port
|
||||
@ -97,7 +107,12 @@ class MySQLHelper(object):
|
||||
_connection_info["port"] = port
|
||||
|
||||
log("Opening db connection for %s@%s" % (user, host), level=DEBUG)
|
||||
self.connection = MySQLdb.connect(**_connection_info)
|
||||
try:
|
||||
self.connection = MySQLdb.connect(**_connection_info)
|
||||
except Exception as e:
|
||||
log("Failed to connect to database due to '{}'".format(str(e)),
|
||||
level=ERROR)
|
||||
raise
|
||||
|
||||
def database_exists(self, db_name):
|
||||
cursor = self.connection.cursor()
|
||||
@ -416,7 +431,7 @@ class MySQLHelper(object):
|
||||
# Otherwise assume localhost
|
||||
return '127.0.0.1'
|
||||
|
||||
def get_allowed_units(self, database, username, relation_id=None):
|
||||
def get_allowed_units(self, database, username, relation_id=None, prefix=None):
|
||||
"""Get list of units with access grants for database with username.
|
||||
|
||||
This is typically used to provide shared-db relations with a list of
|
||||
@ -425,10 +440,12 @@ class MySQLHelper(object):
|
||||
if not self.connection:
|
||||
self.connect(password=self.get_mysql_root_password())
|
||||
allowed_units = set()
|
||||
if not prefix:
|
||||
prefix = database
|
||||
for unit in related_units(relation_id):
|
||||
settings = relation_get(rid=relation_id, unit=unit)
|
||||
# First check for setting with prefix, then without
|
||||
for attr in ["%s_hostname" % (database), 'hostname']:
|
||||
for attr in ["%s_hostname" % (prefix), 'hostname']:
|
||||
hosts = settings.get(attr, None)
|
||||
if hosts:
|
||||
break
|
||||
|
@ -98,3 +98,8 @@ class DisabledModuleAudit(BaseAudit):
|
||||
def _restart_apache():
|
||||
"""Restarts the apache process"""
|
||||
subprocess.check_output(['service', 'apache2', 'restart'])
|
||||
|
||||
@staticmethod
|
||||
def is_ssl_enabled():
|
||||
"""Check if SSL module is enabled or not"""
|
||||
return 'ssl' in DisabledModuleAudit._get_loaded_modules()
|
||||
|
@ -396,7 +396,8 @@ def get_ipv6_addr(iface=None, inc_aliases=False, fatal=True, exc_list=None,
|
||||
if global_addrs:
|
||||
# Make sure any found global addresses are not temporary
|
||||
cmd = ['ip', 'addr', 'show', iface]
|
||||
out = subprocess.check_output(cmd).decode('UTF-8')
|
||||
out = subprocess.check_output(
|
||||
cmd).decode('UTF-8', errors='replace')
|
||||
if dynamic_only:
|
||||
key = re.compile("inet6 (.+)/[0-9]+ scope global.* dynamic.*")
|
||||
else:
|
||||
|
@ -33,6 +33,7 @@ INTERNAL = 'int'
|
||||
ADMIN = 'admin'
|
||||
ACCESS = 'access'
|
||||
|
||||
# TODO: reconcile 'int' vs 'internal' binding names
|
||||
ADDRESS_MAP = {
|
||||
PUBLIC: {
|
||||
'binding': 'public',
|
||||
@ -58,6 +59,14 @@ ADDRESS_MAP = {
|
||||
'fallback': 'private-address',
|
||||
'override': 'os-access-hostname',
|
||||
},
|
||||
# Note (thedac) bridge to begin the reconciliation between 'int' vs
|
||||
# 'internal' binding names
|
||||
'internal': {
|
||||
'binding': 'internal',
|
||||
'config': 'os-internal-network',
|
||||
'fallback': 'private-address',
|
||||
'override': 'os-internal-hostname',
|
||||
},
|
||||
}
|
||||
|
||||
|
||||
@ -195,3 +204,10 @@ def get_vip_in_network(network):
|
||||
if is_address_in_network(network, vip):
|
||||
matching_vip = vip
|
||||
return matching_vip
|
||||
|
||||
|
||||
def get_default_api_bindings():
|
||||
_default_bindings = []
|
||||
for binding in [INTERNAL, ADMIN, PUBLIC]:
|
||||
_default_bindings.append(ADDRESS_MAP[binding]['binding'])
|
||||
return _default_bindings
|
||||
|
@ -18,6 +18,7 @@ from functools import wraps
|
||||
|
||||
import subprocess
|
||||
import json
|
||||
import operator
|
||||
import os
|
||||
import sys
|
||||
import re
|
||||
@ -33,7 +34,7 @@ from charmhelpers import deprecate
|
||||
|
||||
from charmhelpers.contrib.network import ip
|
||||
|
||||
from charmhelpers.core import unitdata
|
||||
from charmhelpers.core import decorators, unitdata
|
||||
|
||||
from charmhelpers.core.hookenv import (
|
||||
WORKLOAD_STATES,
|
||||
@ -89,13 +90,16 @@ from charmhelpers.core.host import (
|
||||
service_start,
|
||||
restart_on_change_helper,
|
||||
)
|
||||
|
||||
from charmhelpers.fetch import (
|
||||
apt_cache,
|
||||
apt_install,
|
||||
import_key as fetch_import_key,
|
||||
add_source as fetch_add_source,
|
||||
SourceConfigError,
|
||||
GPGKeyError,
|
||||
get_upstream_version,
|
||||
filter_installed_packages,
|
||||
filter_missing_packages,
|
||||
ubuntu_apt_pkg as apt,
|
||||
)
|
||||
@ -230,7 +234,7 @@ SWIFT_CODENAMES = OrderedDict([
|
||||
('ussuri',
|
||||
['2.24.0', '2.25.0']),
|
||||
('victoria',
|
||||
['2.25.0']),
|
||||
['2.25.0', '2.26.0']),
|
||||
])
|
||||
|
||||
# >= Liberty version->codename mapping
|
||||
@ -479,9 +483,14 @@ def get_swift_codename(version):
|
||||
return None
|
||||
|
||||
|
||||
@deprecate("moved to charmhelpers.contrib.openstack.utils.get_installed_os_version()", "2021-01", log=juju_log)
|
||||
def get_os_codename_package(package, fatal=True):
|
||||
'''Derive OpenStack release codename from an installed package.'''
|
||||
|
||||
codename = get_installed_os_version()
|
||||
if codename:
|
||||
return codename
|
||||
|
||||
if snap_install_requested():
|
||||
cmd = ['snap', 'list', package]
|
||||
try:
|
||||
@ -569,6 +578,28 @@ def get_os_version_package(pkg, fatal=True):
|
||||
# error_out(e)
|
||||
|
||||
|
||||
def get_installed_os_version():
|
||||
apt_install(filter_installed_packages(['openstack-release']), fatal=False)
|
||||
print("OpenStack Release: {}".format(openstack_release()))
|
||||
return openstack_release().get('OPENSTACK_CODENAME')
|
||||
|
||||
|
||||
@cached
|
||||
def openstack_release():
|
||||
"""Return /etc/os-release in a dict."""
|
||||
d = {}
|
||||
try:
|
||||
with open('/etc/openstack-release', 'r') as lsb:
|
||||
for l in lsb:
|
||||
s = l.split('=')
|
||||
if len(s) != 2:
|
||||
continue
|
||||
d[s[0].strip()] = s[1].strip()
|
||||
except FileNotFoundError:
|
||||
pass
|
||||
return d
|
||||
|
||||
|
||||
# Module local cache variable for the os_release.
|
||||
_os_rel = None
|
||||
|
||||
@ -1295,7 +1326,7 @@ def _check_listening_on_ports_list(ports):
|
||||
Returns a list of ports being listened to and a list of the
|
||||
booleans.
|
||||
|
||||
@param ports: LIST or port numbers.
|
||||
@param ports: LIST of port numbers.
|
||||
@returns [(port_num, boolean), ...], [boolean]
|
||||
"""
|
||||
ports_open = [port_has_listener('0.0.0.0', p) for p in ports]
|
||||
@ -1564,6 +1595,21 @@ def manage_payload_services(action, services=None, charm_func=None):
|
||||
return success, messages
|
||||
|
||||
|
||||
def make_wait_for_ports_barrier(ports, retry_count=5):
|
||||
"""Make a function to wait for port shutdowns.
|
||||
|
||||
Create a function which closes over the provided ports. The function will
|
||||
retry probing ports until they are closed or the retry count has been reached.
|
||||
|
||||
"""
|
||||
@decorators.retry_on_predicate(retry_count, operator.not_, base_delay=0.1)
|
||||
def retry_port_check():
|
||||
_, ports_states = _check_listening_on_ports_list(ports)
|
||||
juju_log("Probe ports {}, result: {}".format(ports, ports_states), level="DEBUG")
|
||||
return any(ports_states)
|
||||
return retry_port_check
|
||||
|
||||
|
||||
def pause_unit(assess_status_func, services=None, ports=None,
|
||||
charm_func=None):
|
||||
"""Pause a unit by stopping the services and setting 'unit-paused'
|
||||
@ -1599,6 +1645,7 @@ def pause_unit(assess_status_func, services=None, ports=None,
|
||||
services=services,
|
||||
charm_func=charm_func)
|
||||
set_unit_paused()
|
||||
|
||||
if assess_status_func:
|
||||
message = assess_status_func()
|
||||
if message:
|
||||
|
@ -41,6 +41,7 @@ from subprocess import (
|
||||
)
|
||||
from charmhelpers import deprecate
|
||||
from charmhelpers.core.hookenv import (
|
||||
application_name,
|
||||
config,
|
||||
service_name,
|
||||
local_unit,
|
||||
@ -162,6 +163,17 @@ def get_osd_settings(relation_name):
|
||||
return _order_dict_by_key(osd_settings)
|
||||
|
||||
|
||||
def send_application_name(relid=None):
|
||||
"""Send the application name down the relation.
|
||||
|
||||
:param relid: Relation id to set application name in.
|
||||
:type relid: str
|
||||
"""
|
||||
relation_set(
|
||||
relation_id=relid,
|
||||
relation_settings={'application-name': application_name()})
|
||||
|
||||
|
||||
def send_osd_settings():
|
||||
"""Pass on requested OSD settings to osd units."""
|
||||
try:
|
||||
@ -256,6 +268,7 @@ class BasePool(object):
|
||||
'compression-max-blob-size': (int, None),
|
||||
'compression-max-blob-size-hdd': (int, None),
|
||||
'compression-max-blob-size-ssd': (int, None),
|
||||
'rbd-mirroring-mode': (str, ('image', 'pool'))
|
||||
}
|
||||
|
||||
def __init__(self, service, name=None, percent_data=None, app_name=None,
|
||||
@ -1074,7 +1087,10 @@ def create_erasure_profile(service, profile_name,
|
||||
erasure_plugin_technique=None):
|
||||
"""Create a new erasure code profile if one does not already exist for it.
|
||||
|
||||
Updates the profile if it exists. Please refer to [0] for more details.
|
||||
Profiles are considered immutable so will not be updated if the named
|
||||
profile already exists.
|
||||
|
||||
Please refer to [0] for more details.
|
||||
|
||||
0: http://docs.ceph.com/docs/master/rados/operations/erasure-code-profile/
|
||||
|
||||
@ -1110,6 +1126,11 @@ def create_erasure_profile(service, profile_name,
|
||||
:type erasure_plugin_technique: str
|
||||
:return: None. Can raise CalledProcessError, ValueError or AssertionError
|
||||
"""
|
||||
if erasure_profile_exists(service, profile_name):
|
||||
log('EC profile {} exists, skipping update'.format(profile_name),
|
||||
level=WARNING)
|
||||
return
|
||||
|
||||
plugin_techniques = {
|
||||
'jerasure': [
|
||||
'reed_sol_van',
|
||||
@ -1209,9 +1230,6 @@ def create_erasure_profile(service, profile_name,
|
||||
if scalar_mds:
|
||||
cmd.append('scalar-mds={}'.format(scalar_mds))
|
||||
|
||||
if erasure_profile_exists(service, profile_name):
|
||||
cmd.append('--force')
|
||||
|
||||
check_call(cmd)
|
||||
|
||||
|
||||
@ -1750,6 +1768,7 @@ class CephBrokerRq(object):
|
||||
max_bytes=None,
|
||||
max_objects=None,
|
||||
namespace=None,
|
||||
rbd_mirroring_mode='pool',
|
||||
weight=None):
|
||||
"""Build common part of a create pool operation.
|
||||
|
||||
@ -1808,6 +1827,9 @@ class CephBrokerRq(object):
|
||||
:type max_objects: Optional[int]
|
||||
:param namespace: Group namespace
|
||||
:type namespace: Optional[str]
|
||||
:param rbd_mirroring_mode: Pool mirroring mode used when Ceph RBD
|
||||
mirroring is enabled.
|
||||
:type rbd_mirroring_mode: Optional[str]
|
||||
:param weight: The percentage of data that is expected to be contained
|
||||
in the pool from the total available space on the OSDs.
|
||||
Used to calculate number of Placement Groups to create
|
||||
@ -1832,6 +1854,7 @@ class CephBrokerRq(object):
|
||||
'max-bytes': max_bytes,
|
||||
'max-objects': max_objects,
|
||||
'group-namespace': namespace,
|
||||
'rbd-mirroring-mode': rbd_mirroring_mode,
|
||||
'weight': weight,
|
||||
}
|
||||
|
||||
@ -2198,6 +2221,7 @@ def send_request_if_needed(request, relation='ceph'):
|
||||
for rid in relation_ids(relation):
|
||||
log('Sending request {}'.format(request.request_id), level=DEBUG)
|
||||
relation_set(relation_id=rid, broker_req=request.request)
|
||||
relation_set(relation_id=rid, relation_settings={'unit-name': local_unit()})
|
||||
|
||||
|
||||
def has_broker_rsp(rid=None, unit=None):
|
||||
|
@ -53,3 +53,41 @@ def retry_on_exception(num_retries, base_delay=0, exc_type=Exception):
|
||||
return _retry_on_exception_inner_2
|
||||
|
||||
return _retry_on_exception_inner_1
|
||||
|
||||
|
||||
def retry_on_predicate(num_retries, predicate_fun, base_delay=0):
|
||||
"""Retry based on return value
|
||||
|
||||
The return value of the decorated function is passed to the given predicate_fun. If the
|
||||
result of the predicate is False, retry the decorated function up to num_retries times
|
||||
|
||||
An exponential backoff up to base_delay^num_retries seconds can be introduced by setting
|
||||
base_delay to a nonzero value. The default is to run with a zero (i.e. no) delay
|
||||
|
||||
:param num_retries: Max. number of retries to perform
|
||||
:type num_retries: int
|
||||
:param predicate_fun: Predicate function to determine if a retry is necessary
|
||||
:type predicate_fun: callable
|
||||
:param base_delay: Starting value in seconds for exponential delay, defaults to 0 (no delay)
|
||||
:type base_delay: float
|
||||
"""
|
||||
def _retry_on_pred_inner_1(f):
|
||||
def _retry_on_pred_inner_2(*args, **kwargs):
|
||||
retries = num_retries
|
||||
multiplier = 1
|
||||
delay = base_delay
|
||||
while True:
|
||||
result = f(*args, **kwargs)
|
||||
if predicate_fun(result) or retries <= 0:
|
||||
return result
|
||||
delay *= multiplier
|
||||
multiplier += 1
|
||||
log("Result {}, retrying '{}' {} more times (delay={})".format(
|
||||
result, f.__name__, retries, delay), level=INFO)
|
||||
retries -= 1
|
||||
if delay:
|
||||
time.sleep(delay)
|
||||
|
||||
return _retry_on_pred_inner_2
|
||||
|
||||
return _retry_on_pred_inner_1
|
||||
|
@ -19,6 +19,7 @@
|
||||
# Nick Moffitt <nick.moffitt@canonical.com>
|
||||
# Matthew Wedgwood <matthew.wedgwood@canonical.com>
|
||||
|
||||
import errno
|
||||
import os
|
||||
import re
|
||||
import pwd
|
||||
@ -59,6 +60,7 @@ elif __platform__ == "centos":
|
||||
) # flake8: noqa -- ignore F401 for this import
|
||||
|
||||
UPDATEDB_PATH = '/etc/updatedb.conf'
|
||||
CA_CERT_DIR = '/usr/local/share/ca-certificates'
|
||||
|
||||
|
||||
def service_start(service_name, **kwargs):
|
||||
@ -677,7 +679,7 @@ def check_hash(path, checksum, hash_type='md5'):
|
||||
|
||||
:param str checksum: Value of the checksum used to validate the file.
|
||||
:param str hash_type: Hash algorithm used to generate `checksum`.
|
||||
Can be any hash alrgorithm supported by :mod:`hashlib`,
|
||||
Can be any hash algorithm supported by :mod:`hashlib`,
|
||||
such as md5, sha1, sha256, sha512, etc.
|
||||
:raises ChecksumError: If the file fails the checksum
|
||||
|
||||
@ -825,7 +827,8 @@ def list_nics(nic_type=None):
|
||||
if nic_type:
|
||||
for int_type in int_types:
|
||||
cmd = ['ip', 'addr', 'show', 'label', int_type + '*']
|
||||
ip_output = subprocess.check_output(cmd).decode('UTF-8')
|
||||
ip_output = subprocess.check_output(
|
||||
cmd).decode('UTF-8', errors='replace')
|
||||
ip_output = ip_output.split('\n')
|
||||
ip_output = (line for line in ip_output if line)
|
||||
for line in ip_output:
|
||||
@ -841,7 +844,8 @@ def list_nics(nic_type=None):
|
||||
interfaces.append(iface)
|
||||
else:
|
||||
cmd = ['ip', 'a']
|
||||
ip_output = subprocess.check_output(cmd).decode('UTF-8').split('\n')
|
||||
ip_output = subprocess.check_output(
|
||||
cmd).decode('UTF-8', errors='replace').split('\n')
|
||||
ip_output = (line.strip() for line in ip_output if line)
|
||||
|
||||
key = re.compile(r'^[0-9]+:\s+(.+):')
|
||||
@ -865,7 +869,8 @@ def set_nic_mtu(nic, mtu):
|
||||
def get_nic_mtu(nic):
|
||||
"""Return the Maximum Transmission Unit (MTU) for a network interface."""
|
||||
cmd = ['ip', 'addr', 'show', nic]
|
||||
ip_output = subprocess.check_output(cmd).decode('UTF-8').split('\n')
|
||||
ip_output = subprocess.check_output(
|
||||
cmd).decode('UTF-8', errors='replace').split('\n')
|
||||
mtu = ""
|
||||
for line in ip_output:
|
||||
words = line.split()
|
||||
@ -877,7 +882,7 @@ def get_nic_mtu(nic):
|
||||
def get_nic_hwaddr(nic):
|
||||
"""Return the Media Access Control (MAC) for a network interface."""
|
||||
cmd = ['ip', '-o', '-0', 'addr', 'show', nic]
|
||||
ip_output = subprocess.check_output(cmd).decode('UTF-8')
|
||||
ip_output = subprocess.check_output(cmd).decode('UTF-8', errors='replace')
|
||||
hwaddr = ""
|
||||
words = ip_output.split()
|
||||
if 'link/ether' in words:
|
||||
@ -889,7 +894,7 @@ def get_nic_hwaddr(nic):
|
||||
def chdir(directory):
|
||||
"""Change the current working directory to a different directory for a code
|
||||
block and return the previous directory after the block exits. Useful to
|
||||
run commands from a specificed directory.
|
||||
run commands from a specified directory.
|
||||
|
||||
:param str directory: The directory path to change to for this context.
|
||||
"""
|
||||
@ -924,9 +929,13 @@ def chownr(path, owner, group, follow_links=True, chowntopdir=False):
|
||||
for root, dirs, files in os.walk(path, followlinks=follow_links):
|
||||
for name in dirs + files:
|
||||
full = os.path.join(root, name)
|
||||
broken_symlink = os.path.lexists(full) and not os.path.exists(full)
|
||||
if not broken_symlink:
|
||||
try:
|
||||
chown(full, uid, gid)
|
||||
except (IOError, OSError) as e:
|
||||
# Intended to ignore "file not found". Catching both to be
|
||||
# compatible with both Python 2.7 and 3.x.
|
||||
if e.errno == errno.ENOENT:
|
||||
pass
|
||||
|
||||
|
||||
def lchownr(path, owner, group):
|
||||
@ -1074,7 +1083,7 @@ def install_ca_cert(ca_cert, name=None):
|
||||
ca_cert = ca_cert.encode('utf8')
|
||||
if not name:
|
||||
name = 'juju-{}'.format(charm_name())
|
||||
cert_file = '/usr/local/share/ca-certificates/{}.crt'.format(name)
|
||||
cert_file = '{}/{}.crt'.format(CA_CERT_DIR, name)
|
||||
new_hash = hashlib.md5(ca_cert).hexdigest()
|
||||
if file_hash(cert_file) == new_hash:
|
||||
return
|
||||
|
@ -646,7 +646,7 @@ def _add_apt_repository(spec):
|
||||
# passed as environment variables (See lp:1433761). This is not the case
|
||||
# LTS and non-LTS releases below bionic.
|
||||
_run_with_retries(['add-apt-repository', '--yes', spec],
|
||||
cmd_env=env_proxy_settings(['https']))
|
||||
cmd_env=env_proxy_settings(['https', 'http']))
|
||||
|
||||
|
||||
def _add_cloud_pocket(pocket):
|
||||
|
@ -129,7 +129,7 @@ class Cache(object):
|
||||
else:
|
||||
data = line.split(None, 4)
|
||||
status = data.pop(0)
|
||||
if status != 'ii':
|
||||
if status not in ('ii', 'hi'):
|
||||
continue
|
||||
pkg = {}
|
||||
pkg.update({k.lower(): v for k, v in zip(headings, data)})
|
||||
@ -265,3 +265,48 @@ def version_compare(a, b):
|
||||
raise RuntimeError('Unable to compare "{}" and "{}", according to '
|
||||
'our logic they are neither greater, equal nor '
|
||||
'less than each other.')
|
||||
|
||||
|
||||
class PkgVersion():
|
||||
"""Allow package versions to be compared.
|
||||
|
||||
For example::
|
||||
|
||||
>>> import charmhelpers.fetch as fetch
|
||||
>>> (fetch.apt_pkg.PkgVersion('2:20.4.0') <
|
||||
... fetch.apt_pkg.PkgVersion('2:20.5.0'))
|
||||
True
|
||||
>>> pkgs = [fetch.apt_pkg.PkgVersion('2:20.4.0'),
|
||||
... fetch.apt_pkg.PkgVersion('2:21.4.0'),
|
||||
... fetch.apt_pkg.PkgVersion('2:17.4.0')]
|
||||
>>> pkgs.sort()
|
||||
>>> pkgs
|
||||
[2:17.4.0, 2:20.4.0, 2:21.4.0]
|
||||
"""
|
||||
|
||||
def __init__(self, version):
|
||||
self.version = version
|
||||
|
||||
def __lt__(self, other):
|
||||
return version_compare(self.version, other.version) == -1
|
||||
|
||||
def __le__(self, other):
|
||||
return self.__lt__(other) or self.__eq__(other)
|
||||
|
||||
def __gt__(self, other):
|
||||
return version_compare(self.version, other.version) == 1
|
||||
|
||||
def __ge__(self, other):
|
||||
return self.__gt__(other) or self.__eq__(other)
|
||||
|
||||
def __eq__(self, other):
|
||||
return version_compare(self.version, other.version) == 0
|
||||
|
||||
def __ne__(self, other):
|
||||
return not self.__eq__(other)
|
||||
|
||||
def __repr__(self):
|
||||
return self.version
|
||||
|
||||
def __hash__(self):
|
||||
return hash(repr(self))
|
||||
|
Loading…
Reference in New Issue
Block a user