diff --git a/manila/share/access.py b/manila/share/access.py index 6bca410524..7b0149c82f 100644 --- a/manila/share/access.py +++ b/manila/share/access.py @@ -18,7 +18,7 @@ import copy from oslo_log import log from manila.common import constants -from manila.i18n import _, _LI +from manila.i18n import _ from manila import utils LOG = log.getLogger(__name__) @@ -493,8 +493,8 @@ class ShareInstanceAccess(ShareInstanceAccessDatabaseMixin): remove_rules, share_instance, share_server): for rule in add_rules: LOG.info( - _LI("Applying access rule '%(rule)s' for share " - "instance '%(instance)s'"), + "Applying access rule '%(rule)s' for share " + "instance '%(instance)s'", {'rule': rule['id'], 'instance': share_instance['id']} ) @@ -511,8 +511,8 @@ class ShareInstanceAccess(ShareInstanceAccessDatabaseMixin): for rule in delete_rules: LOG.info( - _LI("Denying access rule '%(rule)s' from share " - "instance '%(instance)s'"), + "Denying access rule '%(rule)s' from share " + "instance '%(instance)s'", {'rule': rule['id'], 'instance': share_instance['id']} ) diff --git a/manila/share/api.py b/manila/share/api.py index 03f5c6a248..839752a0a7 100644 --- a/manila/share/api.py +++ b/manila/share/api.py @@ -32,7 +32,7 @@ from manila.common import constants from manila.data import rpcapi as data_rpcapi from manila.db import base from manila import exception -from manila.i18n import _, _LE, _LI, _LW +from manila.i18n import _ from manila import policy from manila import quota from manila.scheduler import rpcapi as scheduler_rpcapi @@ -148,22 +148,22 @@ class API(base.Base): return (usages[name]['reserved'] + usages[name]['in_use']) if 'gigabytes' in overs: - LOG.warning(_LW("Quota exceeded for %(s_pid)s, " - "tried to create " - "%(s_size)sG share (%(d_consumed)dG of " - "%(d_quota)dG already consumed)."), { - 's_pid': context.project_id, - 's_size': size, - 'd_consumed': _consumed('gigabytes'), - 'd_quota': quotas['gigabytes']}) + LOG.warning("Quota exceeded for %(s_pid)s, " + "tried to create " + "%(s_size)sG share (%(d_consumed)dG of " + "%(d_quota)dG already consumed).", { + 's_pid': context.project_id, + 's_size': size, + 'd_consumed': _consumed('gigabytes'), + 'd_quota': quotas['gigabytes']}) raise exception.ShareSizeExceedsAvailableQuota() elif 'shares' in overs: - LOG.warning(_LW("Quota exceeded for %(s_pid)s, " - "tried to create " - "share (%(d_consumed)d shares " - "already consumed)."), { - 's_pid': context.project_id, - 'd_consumed': _consumed('shares')}) + LOG.warning("Quota exceeded for %(s_pid)s, " + "tried to create " + "share (%(d_consumed)d shares " + "already consumed).", { + 's_pid': context.project_id, + 'd_consumed': _consumed('shares')}) raise exception.ShareLimitExceeded(allowed=quotas['shares']) try: @@ -524,7 +524,7 @@ class API(base.Base): msg = _("Cannot delete last active replica.") raise exception.ReplicationException(reason=msg) - LOG.info(_LI("Deleting replica %s."), id) + LOG.info("Deleting replica %s.", id) self.db.share_replica_update( context, share_replica['id'], @@ -927,7 +927,7 @@ class API(base.Base): except Exception as e: reservations = None LOG.exception( - _LE("Failed to update quota for deleting share: %s"), e) + ("Failed to update quota for deleting share: %s"), e) for share_instance in share.instances: if share_instance['host']: @@ -984,7 +984,7 @@ class API(base.Base): share_groups = self.db.share_group_get_all_by_share_server( context, server['id']) if share_groups: - LOG.error(_LE("share server '%(ssid)s' in use by share groups."), + LOG.error("share server '%(ssid)s' in use by share groups.", {'ssid': server['id']}) raise exception.ShareServerInUse(share_server_id=server['id']) @@ -1020,18 +1020,18 @@ class API(base.Base): return (usages[name]['reserved'] + usages[name]['in_use']) if 'snapshot_gigabytes' in overs: - msg = _LW("Quota exceeded for %(s_pid)s, tried to create " - "%(s_size)sG snapshot (%(d_consumed)dG of " - "%(d_quota)dG already consumed).") + msg = ("Quota exceeded for %(s_pid)s, tried to create " + "%(s_size)sG snapshot (%(d_consumed)dG of " + "%(d_quota)dG already consumed).") LOG.warning(msg, {'s_pid': context.project_id, 's_size': size, 'd_consumed': _consumed('gigabytes'), 'd_quota': quotas['snapshot_gigabytes']}) raise exception.SnapshotSizeExceedsAvailableQuota() elif 'snapshots' in overs: - msg = _LW("Quota exceeded for %(s_pid)s, tried to create " - "snapshot (%(d_consumed)d snapshots " - "already consumed).") + msg = ("Quota exceeded for %(s_pid)s, tried to create " + "snapshot (%(d_consumed)d snapshots " + "already consumed).") LOG.warning(msg, {'s_pid': context.project_id, 'd_consumed': _consumed('snapshots')}) raise exception.SnapshotLimitExceeded( @@ -1190,17 +1190,17 @@ class API(base.Base): if (new_share_network_id == share_instance['share_network_id'] and new_share_type_id == share_instance['share_type_id'] and dest_host == share_instance['host']): - msg = _LI("Destination host (%(dest_host)s), share network " - "(%(dest_sn)s) or share type (%(dest_st)s) are the same " - "as the current host's '%(src_host)s', '%(src_sn)s' and " - "'%(src_st)s' respectively. Nothing to be done.") % { - 'dest_host': dest_host, - 'dest_sn': new_share_network_id, - 'dest_st': new_share_type_id, - 'src_host': share_instance['host'], - 'src_sn': share_instance['share_network_id'], - 'src_st': share_instance['share_type_id'], - } + msg = ("Destination host (%(dest_host)s), share network " + "(%(dest_sn)s) or share type (%(dest_st)s) are the same " + "as the current host's '%(src_host)s', '%(src_sn)s' and " + "'%(src_st)s' respectively. Nothing to be done.") % { + 'dest_host': dest_host, + 'dest_sn': new_share_network_id, + 'dest_st': new_share_type_id, + 'src_host': share_instance['host'], + 'src_sn': share_instance['share_network_id'], + 'src_st': share_instance['share_type_id'], + } LOG.info(msg) self.db.share_update( context, share['id'], @@ -1304,8 +1304,8 @@ class API(base.Base): elif share['task_state'] == ( constants.TASK_STATE_DATA_COPYING_IN_PROGRESS): data_rpc = data_rpcapi.DataAPI() - LOG.info(_LI("Sending request to get share migration information" - " of share %s.") % share['id']) + LOG.info("Sending request to get share migration information" + " of share %s." % share['id']) services = self.db.service_get_all_by_topic(context, 'manila-data') @@ -1404,8 +1404,8 @@ class API(base.Base): constants.TASK_STATE_DATA_COPYING_IN_PROGRESS): data_rpc = data_rpcapi.DataAPI() - LOG.info(_LI("Sending request to cancel migration of " - "share %s.") % share['id']) + LOG.info("Sending request to cancel migration of " + "share %s." % share['id']) services = self.db.service_get_all_by_topic(context, 'manila-data') @@ -1796,9 +1796,9 @@ class API(base.Base): def _consumed(name): return usages[name]['reserved'] + usages[name]['in_use'] - msg = _LE("Quota exceeded for %(s_pid)s, tried to extend share " - "by %(s_size)sG, (%(d_consumed)dG of %(d_quota)dG " - "already consumed).") + msg = ("Quota exceeded for %(s_pid)s, tried to extend share " + "by %(s_size)sG, (%(d_consumed)dG of %(d_quota)dG " + "already consumed).") LOG.error(msg, {'s_pid': context.project_id, 's_size': size_increase, 'd_consumed': _consumed('gigabytes'), @@ -1810,7 +1810,7 @@ class API(base.Base): self.update(context, share, {'status': constants.STATUS_EXTENDING}) self.share_rpcapi.extend_share(context, share, new_size, reservations) - LOG.info(_LI("Extend share request issued successfully."), + LOG.info("Extend share request issued successfully.", resource=share) def shrink(self, context, share, new_size): @@ -1843,9 +1843,9 @@ class API(base.Base): self.update(context, share, {'status': constants.STATUS_SHRINKING}) self.share_rpcapi.shrink_share(context, share, new_size) - LOG.info(_LI("Shrink share (id=%(id)s) request issued successfully." - " New size: %(size)s") % {'id': share['id'], - 'size': new_size}) + LOG.info("Shrink share (id=%(id)s) request issued successfully." + " New size: %(size)s" % {'id': share['id'], + 'size': new_size}) def snapshot_allow_access(self, context, snapshot, access_type, access_to): """Allow access to a share snapshot.""" diff --git a/manila/share/driver.py b/manila/share/driver.py index fdafdd1c12..37ce2b3f71 100644 --- a/manila/share/driver.py +++ b/manila/share/driver.py @@ -25,7 +25,7 @@ from oslo_config import cfg from oslo_log import log from manila import exception -from manila.i18n import _, _LE, _LW +from manila.i18n import _ from manila import network from manila import utils @@ -213,8 +213,8 @@ class ExecuteMixin(object): tries += 1 if tries >= self.configuration.num_shell_tries: raise - LOG.exception(_LE("Recovering from a failed execute. " - "Try number %s"), tries) + LOG.exception("Recovering from a failed execute. " + "Try number %s", tries) time.sleep(tries ** 2) @@ -1295,8 +1295,8 @@ class ShareDriver(object): self.delete_snapshot( context, share_snapshot, share_server=share_server) except exception.ManilaException: - msg = _LE('Could not delete share group snapshot member %(snap)s ' - 'for share %(share)s.') + msg = ('Could not delete share group snapshot member %(snap)s ' + 'for share %(share)s.') LOG.error(msg % { 'snap': share_snapshot['id'], 'share': share_snapshot['share_id'], @@ -1361,7 +1361,7 @@ class ShareDriver(object): raise exception.ShareGroupSnapshotNotSupported( share_group=snap_dict['share_group_id']) elif not snapshot_members: - LOG.warning(_LW('No shares in share group to create snapshot.')) + LOG.warning('No shares in share group to create snapshot.') return None, None else: share_snapshots = [] @@ -1386,9 +1386,9 @@ class ShareDriver(object): snapshot_members_updates.append(member_update) share_snapshots.append(share_snapshot) except exception.ManilaException as e: - msg = _LE('Could not create share group snapshot. Failed ' - 'to create share snapshot %(snap)s for ' - 'share %(share)s.') + msg = ('Could not create share group snapshot. Failed ' + 'to create share snapshot %(snap)s for ' + 'share %(share)s.') LOG.exception(msg % { 'snap': share_snapshot['id'], 'share': share_snapshot['share_id'] diff --git a/manila/share/drivers/cephfs/cephfs_native.py b/manila/share/drivers/cephfs/cephfs_native.py index ecb9f019b5..81b92a544d 100644 --- a/manila/share/drivers/cephfs/cephfs_native.py +++ b/manila/share/drivers/cephfs/cephfs_native.py @@ -20,7 +20,7 @@ from oslo_utils import units from manila.common import constants from manila import exception -from manila.i18n import _, _LI, _LW +from manila.i18n import _ from manila.share import driver from manila.share import share_types @@ -142,7 +142,7 @@ class CephFSNativeDriver(driver.ShareDriver,): auth_id = self.configuration.safe_get('cephfs_auth_id') self._volume_client = ceph_volume_client.CephFSVolumeClient( auth_id, conf_path, cluster_name) - LOG.info(_LI("[%(be)s}] Ceph client found, connecting..."), + LOG.info("[%(be)s}] Ceph client found, connecting...", {"be": self.backend_name}) if auth_id != CEPH_DEFAULT_AUTH_ID: # Evict any other manila sessions. Only do this if we're @@ -157,7 +157,7 @@ class CephFSNativeDriver(driver.ShareDriver,): self._volume_client = None raise else: - LOG.info(_LI("[%(be)s] Ceph client connection complete."), + LOG.info("[%(be)s] Ceph client connection complete.", {"be": self.backend_name}) return self._volume_client @@ -199,7 +199,7 @@ class CephFSNativeDriver(driver.ShareDriver,): addrs=",".join(mon_addrs), path=volume['mount_path']) - LOG.info(_LI("Calculated export location for share %(id)s: %(loc)s"), + LOG.info("Calculated export location for share %(id)s: %(loc)s", {"id": share['id'], "loc": export_location}) return { @@ -244,8 +244,8 @@ class CephFSNativeDriver(driver.ShareDriver,): def _deny_access(self, context, share, access, share_server=None): if access['access_type'] != CEPHX_ACCESS_TYPE: - LOG.warning(_LW("Invalid access type '%(type)s', " - "ignoring in deny."), + LOG.warning("Invalid access type '%(type)s', " + "ignoring in deny.", {"type": access['access_type']}) return diff --git a/manila/share/drivers/container/container_helper.py b/manila/share/drivers/container/container_helper.py index 5f3896b312..2aa775879f 100644 --- a/manila/share/drivers/container/container_helper.py +++ b/manila/share/drivers/container/container_helper.py @@ -18,7 +18,7 @@ import uuid from oslo_log import log from manila import exception -from manila.i18n import _, _LE, _LI +from manila.i18n import _ from manila.share import driver @@ -66,8 +66,8 @@ class DockerExecHelper(driver.ExecuteMixin): if result[1] != "": raise exception.ManilaException( _("Container %s has failed to start.") % name) - LOG.info(_LI("A container has been successfully started! Its id is " - "%s."), result[0].rstrip('\n')) + LOG.info("A container has been successfully started! Its id is " + "%s.", result[0].rstrip('\n')) def stop_container(self, name): LOG.debug("Stopping container %s.", name) @@ -76,7 +76,7 @@ class DockerExecHelper(driver.ExecuteMixin): if result[1] != '': raise exception.ManilaException( _("Container %s has failed to stop properly.") % name) - LOG.info(_LI("Container %s is successfully stopped."), name) + LOG.info("Container %s is successfully stopped.", name) def execute(self, name=None, cmd=None): if name is None: @@ -94,7 +94,7 @@ class DockerExecHelper(driver.ExecuteMixin): try: result = self._execute(*cmd, run_as_root=True) except Exception: - LOG.exception(_LE("Executing command failed.")) + LOG.exception("Executing command failed.") return None LOG.debug("Execution result: %s.", result) return result diff --git a/manila/share/drivers/container/driver.py b/manila/share/drivers/container/driver.py index 73fa669470..29ec0fd3b5 100644 --- a/manila/share/drivers/container/driver.py +++ b/manila/share/drivers/container/driver.py @@ -27,7 +27,7 @@ from oslo_log import log from oslo_utils import importutils from manila import exception -from manila.i18n import _, _LI, _LW +from manila.i18n import _ from manila.share import driver from manila import utils @@ -156,9 +156,9 @@ class ContainerShareDriver(driver.ShareDriver, driver.ExecuteMixin): ["rm", "-fR", "/shares/%s" % share.share_id] ) except exception.ProcessExecutionError as e: - LOG.warning(_LW("Failed to remove /shares/%(share)s directory in " - "container %(cont)s."), {"share": share.share_id, - "cont": server_id}) + LOG.warning("Failed to remove /shares/%(share)s directory in " + "container %(cont)s.", {"share": share.share_id, + "cont": server_id}) LOG.error(e) self.storage.remove_storage(share) @@ -213,8 +213,8 @@ class ContainerShareDriver(driver.ShareDriver, driver.ExecuteMixin): "specified.") % neutron_class raise exception.ManilaException(msg) elif host_id is None: - LOG.warning(_LW("neutron_host_id is not specified. This driver " - "might not work as expected without it.")) + LOG.warning("neutron_host_id is not specified. This driver " + "might not work as expected without it.") def _connect_to_network(self, server_id, network_info, host_veth): LOG.debug("Attempting to connect container to neutron network.") @@ -280,8 +280,8 @@ class ContainerShareDriver(driver.ShareDriver, driver.ExecuteMixin): self.configuration.container_ovs_bridge_name, veth, run_as_root=True) except exception.ProcessExecutionError as e: - LOG.warning(_LW("Failed to delete port %s: port " - "vanished."), veth) + LOG.warning("Failed to delete port %s: port " + "vanished.", veth) LOG.error(e) def _get_veth_state(self): @@ -317,5 +317,5 @@ class ContainerShareDriver(driver.ShareDriver, driver.ExecuteMixin): veth = self._get_corresponding_veth(veths_before, veths_after) self._connect_to_network(server_id, network_info, veth) - LOG.info(_LI("Container %s was created."), server_id) + LOG.info("Container %s was created.", server_id) return {"id": network_info["server_id"]} diff --git a/manila/share/drivers/container/protocol_helper.py b/manila/share/drivers/container/protocol_helper.py index 366c992cac..04229af652 100644 --- a/manila/share/drivers/container/protocol_helper.py +++ b/manila/share/drivers/container/protocol_helper.py @@ -17,7 +17,7 @@ from oslo_log import log from manila.common import constants as const from manila import exception -from manila.i18n import _, _LW +from manila.i18n import _ LOG = log.getLogger(__name__) @@ -110,7 +110,7 @@ class DockerCIFSHelper(object): existing_users = self._get_existing_users(server_id, share_name, access) except TypeError: - LOG.warning(_LW("Can't access smbd at share %s.") % share_name) + LOG.warning("Can't access smbd at share %s." % share_name) return else: allowed_users = " ".join(sorted(set(existing_users.split()) - diff --git a/manila/share/drivers/container/storage_helper.py b/manila/share/drivers/container/storage_helper.py index 4dbbf11fca..44866f335c 100644 --- a/manila/share/drivers/container/storage_helper.py +++ b/manila/share/drivers/container/storage_helper.py @@ -20,7 +20,7 @@ from oslo_config import cfg from oslo_log import log from manila import exception -from manila.i18n import _, _LW +from manila.i18n import _ from manila.share import driver @@ -85,7 +85,7 @@ class LVMHelper(driver.ExecuteMixin): try: self._execute("umount", to_remove, run_as_root=True) except exception.ProcessExecutionError as e: - LOG.warning(_LW("Failed to umount helper directory %s."), + LOG.warning("Failed to umount helper directory %s.", to_remove) LOG.error(e) # (aovchinnikov): bug 1621784 manifests itself in jamming logical @@ -94,7 +94,7 @@ class LVMHelper(driver.ExecuteMixin): self._execute("lvremove", "-f", "--autobackup", "n", to_remove, run_as_root=True) except exception.ProcessExecutionError as e: - LOG.warning(_LW("Failed to remove logical volume %s.") % to_remove) + LOG.warning("Failed to remove logical volume %s." % to_remove) LOG.error(e) def extend_share(self, share, new_size, share_server=None): diff --git a/manila/share/drivers/dell_emc/plugins/isilon/isilon.py b/manila/share/drivers/dell_emc/plugins/isilon/isilon.py index 42d735d4f0..52266895d4 100644 --- a/manila/share/drivers/dell_emc/plugins/isilon/isilon.py +++ b/manila/share/drivers/dell_emc/plugins/isilon/isilon.py @@ -24,7 +24,7 @@ import six from manila.common import constants as const from manila import exception -from manila.i18n import _, _LW +from manila.i18n import _ from manila.share.drivers.dell_emc.plugins import base from manila.share.drivers.dell_emc.plugins.isilon import isilon_api @@ -139,8 +139,8 @@ class IsilonStorageConnection(base.StorageConnection): self._root_dir + '/' + share['name']) if share_id is None: - lw = _LW('Attempted to delete NFS Share "%s", but the share does ' - 'not appear to exist.') + lw = ('Attempted to delete NFS Share "%s", but the share does ' + 'not appear to exist.') LOG.warning(lw, share['name']) else: # attempt to delete the share @@ -154,8 +154,8 @@ class IsilonStorageConnection(base.StorageConnection): """Is called to remove CIFS share.""" smb_share = self._isilon_api.lookup_smb_share(share['name']) if smb_share is None: - lw = _LW('Attempted to delete CIFS Share "%s", but the share does ' - 'not appear to exist.') + lw = ('Attempted to delete CIFS Share "%s", but the share does ' + 'not appear to exist.') LOG.warning(lw, share['name']) else: share_deleted = self._isilon_api.delete_smb_share(share['name']) diff --git a/manila/share/drivers/dell_emc/plugins/unity/client.py b/manila/share/drivers/dell_emc/plugins/unity/client.py index 16217aa038..43f2a4bbc0 100644 --- a/manila/share/drivers/dell_emc/plugins/unity/client.py +++ b/manila/share/drivers/dell_emc/plugins/unity/client.py @@ -25,7 +25,7 @@ if storops: from manila.common import constants as const from manila import exception -from manila.i18n import _, _LI, _LE, _LW +from manila.i18n import _ from manila.share.drivers.dell_emc.plugins.unity import utils LOG = log.getLogger(__name__) @@ -34,7 +34,7 @@ LOG = log.getLogger(__name__) class UnityClient(object): def __init__(self, host, username, password): if storops is None: - LOG.error(_LE('StorOps is required to run EMC Unity driver.')) + LOG.error('StorOps is required to run EMC Unity driver.') self.system = storops.UnitySystem(host, username, password) def create_cifs_share(self, resource, share_name): @@ -51,7 +51,7 @@ class UnityClient(object): # based share. Log the internal error if it happens. share.enable_ace() except storops_ex.UnityException: - msg = _LE('Failed to enabled ACE for share: {}.') + msg = ('Failed to enabled ACE for share: {}.') LOG.exception(msg.format(share_name)) return share except storops_ex.UnitySmbShareNameExistedError: @@ -116,22 +116,22 @@ class UnityClient(object): try: filesystem.delete() except storops_ex.UnityResourceNotFoundError: - LOG.info(_LI('Filesystem %s is already removed.'), filesystem.name) + LOG.info('Filesystem %s is already removed.', filesystem.name) def create_nas_server(self, name, sp, pool, tenant=None): try: return self.system.create_nas_server(name, sp, pool, tenant=tenant) except storops_ex.UnityNasServerNameUsedError: - LOG.info(_LI('Share server %s already exists, ignoring share ' - 'server creation.'), name) + LOG.info('Share server %s already exists, ignoring share ' + 'server creation.', name) return self.get_nas_server(name) def get_nas_server(self, name): try: return self.system.get_nas_server(name=name) except storops_ex.UnityResourceNotFoundError: - LOG.info(_LI('NAS server %s not found.'), name) + LOG.info('NAS server %s not found.', name) raise def delete_nas_server(self, name, username=None, password=None): @@ -141,7 +141,7 @@ class UnityClient(object): tenant = nas_server.tenant nas_server.delete(username=username, password=password) except storops_ex.UnityResourceNotFoundError: - LOG.info(_LI('NAS server %s not found.'), name) + LOG.info('NAS server %s not found.', name) if tenant is not None: self._delete_tenant(tenant) @@ -156,8 +156,8 @@ class UnityClient(object): try: tenant.delete(delete_hosts=True) except storops_ex.UnityException as ex: - LOG.warning(_LW('Delete tenant %(tenant)s failed with error: ' - '%(ex)s. Leave the tenant on the system.'), + LOG.warning('Delete tenant %(tenant)s failed with error: ' + '%(ex)s. Leave the tenant on the system.', {'tenant': tenant.get_id(), 'ex': ex}) @@ -166,8 +166,8 @@ class UnityClient(object): try: nas_server.create_dns_server(domain, dns_ip) except storops_ex.UnityOneDnsPerNasServerError: - LOG.info(_LI('DNS server %s already exists, ' - 'ignoring DNS server creation.'), domain) + LOG.info('DNS server %s already exists, ' + 'ignoring DNS server creation.', domain) @staticmethod def create_interface(nas_server, ip_addr, netmask, gateway, port_id, @@ -190,16 +190,16 @@ class UnityClient(object): domain_username=username, domain_password=password) except storops_ex.UnitySmbNameInUseError: - LOG.info(_LI('CIFS service on NAS server %s is ' - 'already enabled.'), nas_server.name) + LOG.info('CIFS service on NAS server %s is ' + 'already enabled.', nas_server.name) @staticmethod def enable_nfs_service(nas_server): try: nas_server.enable_nfs_service() except storops_ex.UnityNfsAlreadyEnabledError: - LOG.info(_LI('NFS service on NAS server %s is ' - 'already enabled.'), nas_server.name) + LOG.info('NFS service on NAS server %s is ' + 'already enabled.', nas_server.name) @staticmethod def create_snapshot(filesystem, name): @@ -207,8 +207,8 @@ class UnityClient(object): try: return filesystem.create_snap(name, fs_access_type=access_type) except storops_ex.UnitySnapNameInUseError: - LOG.info(_LI('Snapshot %(snap)s on Filesystem %(fs)s already ' - 'exists.'), {'snap': name, 'fs': filesystem.name}) + LOG.info('Snapshot %(snap)s on Filesystem %(fs)s already ' + 'exists.', {'snap': name, 'fs': filesystem.name}) def create_snap_of_snap(self, src_snap, dst_snap_name, snap_type): access_type = enums.FilesystemSnapAccessTypeEnum.PROTOCOL @@ -233,7 +233,7 @@ class UnityClient(object): try: snap.delete() except storops_ex.UnityResourceNotFoundError: - LOG.info(_LI('Snapshot %s is already removed.'), snap.name) + LOG.info('Snapshot %s is already removed.', snap.name) def get_pool(self, name=None): return self.system.get_pool(name=name) @@ -283,7 +283,7 @@ class UnityClient(object): try: share.delete_access(host_ip) except storops_ex.UnityHostNotFoundException: - LOG.info(_LI('%(host)s access to %(share)s is already removed.'), + LOG.info('%(host)s access to %(share)s is already removed.', {'host': host_ip, 'share': share_name}) def get_file_ports(self): @@ -328,6 +328,6 @@ class UnityClient(object): "Use the existing VLAN tenant.", vlan_id) exc.reraise = False except storops_ex.SystemAPINotSupported: - LOG.info(_LI("This system doesn't support tenant.")) + LOG.info("This system doesn't support tenant.") return tenant diff --git a/manila/share/drivers/dell_emc/plugins/unity/connection.py b/manila/share/drivers/dell_emc/plugins/unity/connection.py index d33ded805f..eb0da8dcf7 100644 --- a/manila/share/drivers/dell_emc/plugins/unity/connection.py +++ b/manila/share/drivers/dell_emc/plugins/unity/connection.py @@ -27,7 +27,7 @@ if storops: from manila.common import constants as const from manila import exception -from manila.i18n import _, _LE, _LW, _LI +from manila.i18n import _ from manila.share.drivers.dell_emc.plugins import base as driver from manila.share.drivers.dell_emc.plugins.unity import client from manila.share.drivers.dell_emc.plugins.unity import utils as unity_utils @@ -127,9 +127,9 @@ class UnityStorageConnection(driver.StorageConnection): raise exception.BadConfigurationException(reason=msg) if unmanaged_port_ids: - LOG.info(_LI("The following specified ports are not managed by " - "the backend: %(unmanaged)s. This host will only " - "manage the storage ports: %(exist)s"), + LOG.info("The following specified ports are not managed by " + "the backend: %(unmanaged)s. This host will only " + "manage the storage ports: %(exist)s", {'unmanaged': ",".join(unmanaged_port_ids), 'exist': ",".join(map(",".join, sp_ports_map.values()))}) @@ -138,8 +138,8 @@ class UnityStorageConnection(driver.StorageConnection): ",".join(map(",".join, sp_ports_map.values()))) if len(sp_ports_map) == 1: - LOG.info(_LI("Only ports of %s are configured. Configure ports " - "of both SPA and SPB to use both of the SPs."), + LOG.info("Only ports of %s are configured. Configure ports " + "of both SPA and SPB to use both of the SPs.", list(sp_ports_map)[0]) return sp_ports_map @@ -237,7 +237,7 @@ class UnityStorageConnection(driver.StorageConnection): backend_share = self.client.get_share(share_name, share['share_proto']) except storops_ex.UnityResourceNotFoundError: - LOG.warning(_LW("Share %s is not found when deleting the share"), + LOG.warning("Share %s is not found when deleting the share", share_name) return @@ -261,8 +261,8 @@ class UnityStorageConnection(driver.StorageConnection): new_size) else: share_id = share['id'] - reason = _LE("Driver does not support extending a " - "snapshot based share.") + reason = ("Driver does not support extending a " + "snapshot based share.") raise exception.ShareExtendingError(share_id=share_id, reason=reason) @@ -422,7 +422,7 @@ class UnityStorageConnection(driver.StorageConnection): except Exception: with excutils.save_and_reraise_exception(): - LOG.exception(_LE('Could not setup server.')) + LOG.exception('Could not setup server.') server_details = {'share_server_name': server_name} self.teardown_server( server_details, network_info['security_services']) @@ -538,10 +538,10 @@ class UnityStorageConnection(driver.StorageConnection): raise exception.BadConfigurationException(reason=msg) if unmanaged_pools: - LOG.info(_LI("The following specified storage pools " - "are not managed by the backend: " - "%(un_managed)s. This host will only manage " - "the storage pools: %(exist)s"), + LOG.info("The following specified storage pools " + "are not managed by the backend: " + "%(un_managed)s. This host will only manage " + "the storage pools: %(exist)s", {'un_managed': ",".join(unmanaged_pools), 'exist': ",".join(matched_pools)}) else: @@ -621,13 +621,13 @@ class UnityStorageConnection(driver.StorageConnection): # Enable NFS service with kerberos kerberos_enabled = True # TODO(jay.xu): enable nfs service with kerberos - LOG.warning(_LW('Kerberos is not supported by ' - 'EMC Unity manila driver plugin.')) + LOG.warning('Kerberos is not supported by ' + 'EMC Unity manila driver plugin.') elif service_type == 'ldap': - LOG.warning(_LW('LDAP is not supported by ' - 'EMC Unity manila driver plugin.')) + LOG.warning('LDAP is not supported by ' + 'EMC Unity manila driver plugin.') else: - LOG.warning(_LW('Unknown security service type: %s.'), + LOG.warning('Unknown security service type: %s.', service_type) if not kerberos_enabled: diff --git a/manila/share/drivers/dell_emc/plugins/vmax/connection.py b/manila/share/drivers/dell_emc/plugins/vmax/connection.py index ae6f7d76d5..1b088e969f 100644 --- a/manila/share/drivers/dell_emc/plugins/vmax/connection.py +++ b/manila/share/drivers/dell_emc/plugins/vmax/connection.py @@ -24,7 +24,7 @@ from oslo_utils import units from manila.common import constants as const from manila import exception -from manila.i18n import _, _LE, _LI, _LW +from manila.i18n import _ from manila.share.drivers.dell_emc.plugins import base as driver from manila.share.drivers.dell_emc.plugins.vmax import ( object_manager as manager) @@ -250,8 +250,8 @@ class VMAXStorageConnection(driver.StorageConnection): def delete_share(self, context, share, share_server=None): """Delete a share.""" if share_server is None: - LOG.warning(_LW("Share network should be specified for " - "share deletion.")) + LOG.warning("Share network should be specified for " + "share deletion.") return share_proto = share['share_proto'].upper() @@ -295,20 +295,20 @@ class VMAXStorageConnection(driver.StorageConnection): # Delete mount point self._get_context('MountPoint').delete(path, vdm_name) except exception.EMCVmaxXMLAPIError as e: - LOG.exception(_LE("CIFS server %(name)s on mover %(mover_name)s " + LOG.exception("CIFS server %(name)s on mover %(mover_name)s " "not found due to error %(err)s. Skip the " - "deletion."), + "deletion.", {'name': path, 'mover_name': vdm_name, - 'err': e.message}) + 'err': e.message}) try: # Delete file system self._get_context('FileSystem').delete(share_name) except exception.EMCVmaxXMLAPIError as e: - LOG.exception(_LE("File system %(share_name)s not found due to" - "error %(err)s. Skip the deletion."), + LOG.exception("File system %(share_name)s not found due to" + "error %(err)s. Skip the deletion.", {'share_name': share_name, - 'err': e.message}) + 'err': e.message}) def delete_snapshot(self, context, snapshot, share_server=None): """Delete a snapshot.""" @@ -471,7 +471,7 @@ class VMAXStorageConnection(driver.StorageConnection): share_name = share['id'] if access['access_type'] != 'user': - LOG.warning(_LW("Only user access type allowed for CIFS share.")) + LOG.warning("Only user access type allowed for CIFS share.") return user_name = access['access_to'] @@ -505,7 +505,7 @@ class VMAXStorageConnection(driver.StorageConnection): access_type = access['access_type'] if access_type != 'ip': - LOG.warning(_LW("Only ip access type allowed.")) + LOG.warning("Only ip access type allowed.") return host_ip = access['access_to'] @@ -550,7 +550,7 @@ class VMAXStorageConnection(driver.StorageConnection): ",".join(real_pools)) raise exception.InvalidParameterValue(err=msg) - LOG.info(_LI("Storage pools: %s will be managed."), + LOG.info("Storage pools: %s will be managed.", ",".join(matched_pools)) else: LOG.debug("No storage pool is specified, so all pools " @@ -722,7 +722,7 @@ class VMAXStorageConnection(driver.StorageConnection): except Exception: with excutils.save_and_reraise_exception(): - LOG.exception(_LE('Could not setup server')) + LOG.exception('Could not setup server') server_details = self._construct_backend_details( vdm_name, allocated_interfaces) self.teardown_server( @@ -810,7 +810,7 @@ class VMAXStorageConnection(driver.StorageConnection): status, servers = self._get_context('CIFSServer').get_all( vdm_name) if constants.STATUS_OK != status: - LOG.error(_LE('Could not find CIFS server by name: %s.'), + LOG.error('Could not find CIFS server by name: %s.', vdm_name) else: cifs_servers = copy.deepcopy(servers) diff --git a/manila/share/drivers/dell_emc/plugins/vmax/connector.py b/manila/share/drivers/dell_emc/plugins/vmax/connector.py index d32b970e44..d5850d269c 100644 --- a/manila/share/drivers/dell_emc/plugins/vmax/connector.py +++ b/manila/share/drivers/dell_emc/plugins/vmax/connector.py @@ -25,7 +25,6 @@ from six.moves.urllib import request as url_request # pylint: disable=E0611 from manila import exception from manila.i18n import _ -from manila.i18n import _LE from manila.share.drivers.dell_emc.plugins.vmax import constants from manila import utils @@ -154,8 +153,8 @@ class SSHConnector(object): return out, err except processutils.ProcessExecutionError as e: with excutils.save_and_reraise_exception(): - LOG.error(_LE('Error running SSH command: %(cmd)s. ' - 'Error: %(excmsg)s.'), + LOG.error('Error running SSH command: %(cmd)s. ' + 'Error: %(excmsg)s.', {'cmd': command, 'excmsg': e}) def log_request(self, cmd, out, err): diff --git a/manila/share/drivers/dell_emc/plugins/vmax/object_manager.py b/manila/share/drivers/dell_emc/plugins/vmax/object_manager.py index 73260c1285..8fdb10ae8c 100644 --- a/manila/share/drivers/dell_emc/plugins/vmax/object_manager.py +++ b/manila/share/drivers/dell_emc/plugins/vmax/object_manager.py @@ -24,7 +24,7 @@ import six from manila.common import constants as const from manila import exception -from manila.i18n import _, _LI, _LW, _LE +from manila.i18n import _ from manila.share.drivers.dell_emc.plugins.vmax import connector from manila.share.drivers.dell_emc.plugins.vmax import constants from manila.share.drivers.dell_emc.plugins.vmax import utils as vmax_utils @@ -89,8 +89,8 @@ class StorageObject(object): constants.STATUS_INFO): response['maxSeverity'] = constants.STATUS_OK - LOG.warning(_LW("Translated status from %(old)s to %(new)s. " - "Message: %(info)s."), + LOG.warning("Translated status from %(old)s to %(new)s. " + "Message: %(info)s.", {'old': old_Severity, 'new': response['maxSeverity'], 'info': response}) @@ -252,8 +252,8 @@ class FileSystem(StorageObject): raise exception.EMCVmaxInvalidMoverID(id=mover_id) elif self._response_validation( response, constants.MSG_FILESYSTEM_EXIST): - LOG.warning(_LW("File system %s already exists. " - "Skip the creation."), name) + LOG.warning("File system %s already exists. " + "Skip the creation.", name) return elif constants.STATUS_OK != response['maxSeverity']: message = (_("Failed to create file system %(name)s. " @@ -306,7 +306,7 @@ class FileSystem(StorageObject): def delete(self, name): status, out = self.get(name) if constants.STATUS_NOT_FOUND == status: - LOG.warning(_LW("File system %s not found. Skip the deletion."), + LOG.warning("File system %s not found. Skip the deletion.", name) return elif constants.STATUS_OK != status: @@ -440,8 +440,8 @@ class FileSystem(StorageObject): try: self._execute_cmd(copy_ckpt_cmd, check_exit_code=True) except processutils.ProcessExecutionError as expt: - LOG.error(_LE("Failed to copy content from snapshot %(snap)s to " - "file system %(filesystem)s. Reason: %(err)s."), + LOG.error("Failed to copy content from snapshot %(snap)s to " + "file system %(filesystem)s. Reason: %(err)s.", {'snap': snap_name, 'filesystem': name, 'err': expt}) @@ -576,8 +576,8 @@ class MountPoint(StorageObject): self.xml_retry = True raise exception.EMCVmaxInvalidMoverID(id=mover_id) elif self._is_mount_point_already_existent(response): - LOG.warning(_LW("Mount Point %(mount)s already exists. " - "Skip the creation."), {'mount': mount_path}) + LOG.warning("Mount Point %(mount)s already exists. " + "Skip the creation.", {'mount': mount_path}) return elif constants.STATUS_OK != response['maxSeverity']: message = (_('Failed to create Mount Point %(mount)s for ' @@ -642,8 +642,8 @@ class MountPoint(StorageObject): self.xml_retry = True raise exception.EMCVmaxInvalidMoverID(id=mover_id) elif self._is_mount_point_nonexistent(response): - LOG.warning(_LW('Mount point %(mount)s on mover %(mover_name)s ' - 'not found.'), + LOG.warning('Mount point %(mount)s on mover %(mover_name)s ' + 'not found.', {'mount': mount_path, 'mover_name': mover_name}) return @@ -817,8 +817,8 @@ class Mover(StorageObject): lines = out.strip().split('\n') for line in lines: if line.strip().split() == header: - LOG.info(_LI('Found the header of the command ' - '/nas/bin/nas_cel -interconnect -l.')) + LOG.info('Found the header of the command ' + '/nas/bin/nas_cel -interconnect -l.') else: interconn = line.strip().split() if interconn[2] == source and interconn[4] == destination: @@ -874,7 +874,7 @@ class VDM(StorageObject): self.xml_retry = True raise exception.EMCVmaxInvalidMoverID(id=mover_id) elif self._response_validation(response, constants.MSG_VDM_EXIST): - LOG.warning(_LW("VDM %(name)s already exists. Skip the creation."), + LOG.warning("VDM %(name)s already exists. Skip the creation.", {'name': name}) elif constants.STATUS_OK != response['maxSeverity']: message = (_("Failed to create VDM %(name)s on mover " @@ -918,7 +918,7 @@ class VDM(StorageObject): def delete(self, name): status, out = self.get(name) if constants.STATUS_NOT_FOUND == status: - LOG.warning(_LW("VDM %s not found. Skip the deletion."), + LOG.warning("VDM %s not found. Skip the deletion.", name) return elif constants.STATUS_OK != status: @@ -1053,8 +1053,8 @@ class Snapshot(StorageObject): response = self._send_request(request) if self._response_validation(response, constants.MSG_SNAP_EXIST): - LOG.warning(_LW("Snapshot %(name)s already exists. " - "Skip the creation."), + LOG.warning("Snapshot %(name)s already exists. " + "Skip the creation.", {'name': name}) elif constants.STATUS_OK != response['maxSeverity']: message = (_("Failed to create snapshot %(name)s on " @@ -1098,7 +1098,7 @@ class Snapshot(StorageObject): def delete(self, name): status, out = self.get(name) if constants.STATUS_NOT_FOUND == status: - LOG.warning(_LW("Snapshot %s not found. Skip the deletion."), + LOG.warning("Snapshot %s not found. Skip the deletion.", name) return elif constants.STATUS_OK != status: @@ -1182,12 +1182,12 @@ class MoverInterface(StorageObject): raise exception.EMCVmaxInvalidMoverID(id=mover_id) elif self._response_validation( response, constants.MSG_INTERFACE_NAME_EXIST): - LOG.warning(_LW("Mover interface name %s already exists. " - "Skip the creation."), name) + LOG.warning("Mover interface name %s already exists. " + "Skip the creation.", name) elif self._response_validation( response, constants.MSG_INTERFACE_EXIST): - LOG.warning(_LW("Mover interface IP %s already exists. " - "Skip the creation."), ip_addr) + LOG.warning("Mover interface IP %s already exists. " + "Skip the creation.", ip_addr) elif self._response_validation( response, constants.MSG_INTERFACE_INVALID_VLAN_ID): # When fail to create a mover interface with the specified @@ -1246,8 +1246,8 @@ class MoverInterface(StorageObject): raise exception.EMCVmaxInvalidMoverID(id=mover_id) elif self._response_validation( response, constants.MSG_INTERFACE_NON_EXISTENT): - LOG.warning(_LW("Mover interface %s not found. " - "Skip the deletion."), ip_addr) + LOG.warning("Mover interface %s not found. " + "Skip the deletion.", ip_addr) return elif constants.STATUS_OK != response['maxSeverity']: message = (_("Failed to delete mover interface %(ip)s on mover " @@ -1316,8 +1316,8 @@ class DNSDomain(StorageObject): self.xml_retry = True raise exception.EMCVmaxInvalidMoverID(id=mover_id) elif constants.STATUS_OK != response['maxSeverity']: - LOG.warning(_LW("Failed to delete DNS domain %(name)s. " - "Reason: %(err)s."), + LOG.warning("Failed to delete DNS domain %(name)s. " + "Reason: %(err)s.", {'name': name, 'err': response['problems']}) @@ -1508,13 +1508,13 @@ class CIFSServer(StorageObject): status, out = self.get( computer_name.lower(), mover_name, is_vdm, self.xml_retry) if constants.STATUS_NOT_FOUND == status: - LOG.warning(_LW("CIFS server %(name)s on mover %(mover_name)s " - "not found. Skip the deletion."), + LOG.warning("CIFS server %(name)s on mover %(mover_name)s " + "not found. Skip the deletion.", {'name': computer_name, 'mover_name': mover_name}) return except exception.EMCVmaxXMLAPIError: - LOG.warning(_LW("CIFS server %(name)s on mover %(mover_name)s " - "not found. Skip the deletion."), + LOG.warning("CIFS server %(name)s on mover %(mover_name)s " + "not found. Skip the deletion.", {'name': computer_name, 'mover_name': mover_name}) return @@ -1606,7 +1606,7 @@ class CIFSShare(StorageObject): def delete(self, name, mover_name, is_vdm=True): status, out = self.get(name) if constants.STATUS_NOT_FOUND == status: - LOG.warning(_LW("CIFS share %s not found. Skip the deletion."), + LOG.warning("CIFS share %s not found. Skip the deletion.", name) return elif constants.STATUS_OK != status: @@ -1684,8 +1684,8 @@ class CIFSShare(StorageObject): dup_msg = re.compile(r'ACE for %(domain)s\\%(user)s unchanged' % {'domain': domain, 'user': user_name}, re.I) if re.search(dup_msg, expt.stdout): - LOG.warning(_LW("Duplicate access control entry, " - "skipping allow...")) + LOG.warning("Duplicate access control entry, " + "skipping allow...") else: message = (_('Failed to allow the access %(access)s to ' 'CIFS share %(name)s. Reason: %(err)s.') % @@ -1716,10 +1716,10 @@ class CIFSShare(StorageObject): % {'domain': domain, 'user': user_name}, re.I) if re.search(not_found_msg, expt.stdout): - LOG.warning(_LW("No access control entry found, " - "skipping deny...")) + LOG.warning("No access control entry found, " + "skipping deny...") elif re.search(user_err_msg, expt.stdout): - LOG.warning(_LW("User not found on domain, skipping deny...")) + LOG.warning("User not found on domain, skipping deny...") else: message = (_('Failed to deny the access %(access)s to ' 'CIFS share %(name)s. Reason: %(err)s.') % @@ -1798,7 +1798,7 @@ class NFSShare(StorageObject): status, out = self.get(name, mover_name) if constants.STATUS_NOT_FOUND == status: - LOG.warning(_LW("NFS share %s not found. Skip the deletion."), + LOG.warning("NFS share %s not found. Skip the deletion.", path) return @@ -1849,7 +1849,7 @@ class NFSShare(StorageObject): dup_msg = (r'%(mover_name)s : No such file or directory' % {'mover_name': mover_name}) if re.search(dup_msg, expt.stdout): - LOG.warning(_LW("NFS share %s not found."), name) + LOG.warning("NFS share %s not found.", name) return constants.STATUS_NOT_FOUND, None else: message = (_('Failed to list NFS share %(name)s on ' diff --git a/manila/share/drivers/dell_emc/plugins/vnx/connection.py b/manila/share/drivers/dell_emc/plugins/vnx/connection.py index 2bb02c9aa2..432b73c637 100644 --- a/manila/share/drivers/dell_emc/plugins/vnx/connection.py +++ b/manila/share/drivers/dell_emc/plugins/vnx/connection.py @@ -24,7 +24,7 @@ from oslo_utils import units from manila.common import constants as const from manila import exception -from manila.i18n import _, _LE, _LI, _LW +from manila.i18n import _ from manila.share.drivers.dell_emc.plugins import base as driver from manila.share.drivers.dell_emc.plugins.vnx import constants from manila.share.drivers.dell_emc.plugins.vnx import object_manager as manager @@ -250,9 +250,9 @@ class VNXStorageConnection(driver.StorageConnection): def delete_share(self, context, share, share_server=None): """Delete a share.""" if share_server is None: - LOG.warning(_LW("Driver does not support share deletion without " - "share network specified. Return directly because " - "there is nothing to clean.")) + LOG.warning("Driver does not support share deletion without " + "share network specified. Return directly because " + "there is nothing to clean.") return share_proto = share['share_proto'] @@ -545,7 +545,7 @@ class VNXStorageConnection(driver.StorageConnection): ",".join(real_pools)) raise exception.InvalidParameterValue(err=msg) - LOG.info(_LI("Storage pools: %s will be managed."), + LOG.info("Storage pools: %s will be managed.", ",".join(matched_pools)) else: LOG.debug("No storage pool is specified, so all pools " @@ -711,7 +711,7 @@ class VNXStorageConnection(driver.StorageConnection): except Exception: with excutils.save_and_reraise_exception(): - LOG.exception(_LE('Could not setup server.')) + LOG.exception('Could not setup server.') server_details = self._construct_backend_details( vdm_name, allocated_interfaces) self.teardown_server( @@ -799,7 +799,7 @@ class VNXStorageConnection(driver.StorageConnection): status, servers = self._get_context('CIFSServer').get_all( vdm_name) if constants.STATUS_OK != status: - LOG.error(_LE('Could not find CIFS server by name: %s.'), + LOG.error('Could not find CIFS server by name: %s.', vdm_name) else: cifs_servers = copy.deepcopy(servers) diff --git a/manila/share/drivers/dell_emc/plugins/vnx/connector.py b/manila/share/drivers/dell_emc/plugins/vnx/connector.py index ce6d3c8149..20bdf42f9d 100644 --- a/manila/share/drivers/dell_emc/plugins/vnx/connector.py +++ b/manila/share/drivers/dell_emc/plugins/vnx/connector.py @@ -25,7 +25,6 @@ from six.moves.urllib import request as url_request # pylint: disable=E0611 from manila import exception from manila.i18n import _ -from manila.i18n import _LE from manila.share.drivers.dell_emc.plugins.vnx import constants from manila import utils @@ -154,7 +153,7 @@ class SSHConnector(object): return out, err except processutils.ProcessExecutionError: with excutils.save_and_reraise_exception(): - LOG.exception(_LE('Error running SSH command: %(cmd)s.'), + LOG.exception('Error running SSH command: %(cmd)s.', {'cmd': command}) def log_request(self, cmd, out, err): diff --git a/manila/share/drivers/dell_emc/plugins/vnx/object_manager.py b/manila/share/drivers/dell_emc/plugins/vnx/object_manager.py index 47cf3aa52f..c229ba027b 100644 --- a/manila/share/drivers/dell_emc/plugins/vnx/object_manager.py +++ b/manila/share/drivers/dell_emc/plugins/vnx/object_manager.py @@ -24,7 +24,7 @@ import six from manila.common import constants as const from manila import exception -from manila.i18n import _, _LE, _LI, _LW +from manila.i18n import _ from manila.share.drivers.dell_emc.plugins.vnx import connector from manila.share.drivers.dell_emc.plugins.vnx import constants from manila.share.drivers.dell_emc.plugins.vnx import utils as vnx_utils @@ -89,8 +89,8 @@ class StorageObject(object): constants.STATUS_INFO): response['maxSeverity'] = constants.STATUS_OK - LOG.warning(_LW("Translated status from %(old)s to %(new)s. " - "Message: %(info)s."), + LOG.warning("Translated status from %(old)s to %(new)s. " + "Message: %(info)s.", {'old': old_Severity, 'new': response['maxSeverity'], 'info': response}) @@ -252,8 +252,8 @@ class FileSystem(StorageObject): raise exception.EMCVnxInvalidMoverID(id=mover_id) elif self._response_validation( response, constants.MSG_FILESYSTEM_EXIST): - LOG.warning(_LW("File system %s already exists. " - "Skip the creation."), name) + LOG.warning("File system %s already exists. " + "Skip the creation.", name) return elif constants.STATUS_OK != response['maxSeverity']: message = (_("Failed to create file system %(name)s. " @@ -306,7 +306,7 @@ class FileSystem(StorageObject): def delete(self, name): status, out = self.get(name) if constants.STATUS_NOT_FOUND == status: - LOG.warning(_LW("File system %s not found. Skip the deletion."), + LOG.warning("File system %s not found. Skip the deletion.", name) return elif constants.STATUS_OK != status: @@ -440,8 +440,8 @@ class FileSystem(StorageObject): try: self._execute_cmd(copy_ckpt_cmd, check_exit_code=True) except processutils.ProcessExecutionError: - LOG.exception(_LE("Failed to copy content from snapshot %(snap)s " - "to file system %(filesystem)s."), + LOG.exception("Failed to copy content from snapshot %(snap)s " + "to file system %(filesystem)s.", {'snap': snap_name, 'filesystem': name}) @@ -575,8 +575,8 @@ class MountPoint(StorageObject): self.xml_retry = True raise exception.EMCVnxInvalidMoverID(id=mover_id) elif self._is_mount_point_already_existent(response): - LOG.warning(_LW("Mount Point %(mount)s already exists. " - "Skip the creation."), {'mount': mount_path}) + LOG.warning("Mount Point %(mount)s already exists. " + "Skip the creation.", {'mount': mount_path}) return elif constants.STATUS_OK != response['maxSeverity']: message = (_('Failed to create Mount Point %(mount)s for ' @@ -641,8 +641,8 @@ class MountPoint(StorageObject): self.xml_retry = True raise exception.EMCVnxInvalidMoverID(id=mover_id) elif self._is_mount_point_nonexistent(response): - LOG.warning(_LW('Mount point %(mount)s on mover %(mover_name)s ' - 'not found.'), + LOG.warning('Mount point %(mount)s on mover %(mover_name)s ' + 'not found.', {'mount': mount_path, 'mover_name': mover_name}) return @@ -816,8 +816,8 @@ class Mover(StorageObject): lines = out.strip().split('\n') for line in lines: if line.strip().split() == header: - LOG.info(_LI('Found the header of the command ' - '/nas/bin/nas_cel -interconnect -l.')) + LOG.info('Found the header of the command ' + '/nas/bin/nas_cel -interconnect -l.') else: interconn = line.strip().split() if interconn[2] == source and interconn[4] == destination: @@ -873,7 +873,7 @@ class VDM(StorageObject): self.xml_retry = True raise exception.EMCVnxInvalidMoverID(id=mover_id) elif self._response_validation(response, constants.MSG_VDM_EXIST): - LOG.warning(_LW("VDM %(name)s already exists. Skip the creation."), + LOG.warning("VDM %(name)s already exists. Skip the creation.", {'name': name}) elif constants.STATUS_OK != response['maxSeverity']: message = (_("Failed to create VDM %(name)s on mover " @@ -917,7 +917,7 @@ class VDM(StorageObject): def delete(self, name): status, out = self.get(name) if constants.STATUS_NOT_FOUND == status: - LOG.warning(_LW("VDM %s not found. Skip the deletion."), + LOG.warning("VDM %s not found. Skip the deletion.", name) return elif constants.STATUS_OK != status: @@ -1052,8 +1052,8 @@ class Snapshot(StorageObject): response = self._send_request(request) if self._response_validation(response, constants.MSG_SNAP_EXIST): - LOG.warning(_LW("Snapshot %(name)s already exists. " - "Skip the creation."), + LOG.warning("Snapshot %(name)s already exists. " + "Skip the creation.", {'name': name}) elif constants.STATUS_OK != response['maxSeverity']: message = (_("Failed to create snapshot %(name)s on " @@ -1097,7 +1097,7 @@ class Snapshot(StorageObject): def delete(self, name): status, out = self.get(name) if constants.STATUS_NOT_FOUND == status: - LOG.warning(_LW("Snapshot %s not found. Skip the deletion."), + LOG.warning("Snapshot %s not found. Skip the deletion.", name) return elif constants.STATUS_OK != status: @@ -1181,13 +1181,13 @@ class MoverInterface(StorageObject): raise exception.EMCVnxInvalidMoverID(id=mover_id) elif self._response_validation( response, constants.MSG_INTERFACE_NAME_EXIST): - LOG.warning(_LW("Mover interface name %s already exists. " - "Skip the creation."), name) + LOG.warning("Mover interface name %s already exists. " + "Skip the creation.", name) return elif self._response_validation( response, constants.MSG_INTERFACE_EXIST): - LOG.warning(_LW("Mover interface IP %s already exists. " - "Skip the creation."), ip_addr) + LOG.warning("Mover interface IP %s already exists. " + "Skip the creation.", ip_addr) return elif self._response_validation( response, constants.MSG_INTERFACE_INVALID_VLAN_ID): @@ -1247,8 +1247,8 @@ class MoverInterface(StorageObject): raise exception.EMCVnxInvalidMoverID(id=mover_id) elif self._response_validation( response, constants.MSG_INTERFACE_NON_EXISTENT): - LOG.warning(_LW("Mover interface %s not found. " - "Skip the deletion."), ip_addr) + LOG.warning("Mover interface %s not found. " + "Skip the deletion.", ip_addr) return elif constants.STATUS_OK != response['maxSeverity']: message = (_("Failed to delete mover interface %(ip)s on mover " @@ -1317,8 +1317,8 @@ class DNSDomain(StorageObject): self.xml_retry = True raise exception.EMCVnxInvalidMoverID(id=mover_id) elif constants.STATUS_OK != response['maxSeverity']: - LOG.warning(_LW("Failed to delete DNS domain %(name)s. " - "Reason: %(err)s."), + LOG.warning("Failed to delete DNS domain %(name)s. " + "Reason: %(err)s.", {'name': name, 'err': response['problems']}) @@ -1509,13 +1509,13 @@ class CIFSServer(StorageObject): status, out = self.get( computer_name.lower(), mover_name, is_vdm, self.xml_retry) if constants.STATUS_NOT_FOUND == status: - LOG.warning(_LW("CIFS server %(name)s on mover %(mover_name)s " - "not found. Skip the deletion."), + LOG.warning("CIFS server %(name)s on mover %(mover_name)s " + "not found. Skip the deletion.", {'name': computer_name, 'mover_name': mover_name}) return except exception.EMCVnxXMLAPIError: - LOG.warning(_LW("CIFS server %(name)s on mover %(mover_name)s " - "not found. Skip the deletion."), + LOG.warning("CIFS server %(name)s on mover %(mover_name)s " + "not found. Skip the deletion.", {'name': computer_name, 'mover_name': mover_name}) return @@ -1607,7 +1607,7 @@ class CIFSShare(StorageObject): def delete(self, name, mover_name, is_vdm=True): status, out = self.get(name) if constants.STATUS_NOT_FOUND == status: - LOG.warning(_LW("CIFS share %s not found. Skip the deletion."), + LOG.warning("CIFS share %s not found. Skip the deletion.", name) return elif constants.STATUS_OK != status: @@ -1685,8 +1685,8 @@ class CIFSShare(StorageObject): dup_msg = re.compile(r'ACE for %(domain)s\\%(user)s unchanged' % {'domain': domain, 'user': user_name}, re.I) if re.search(dup_msg, expt.stdout): - LOG.warning(_LW("Duplicate access control entry, " - "skipping allow...")) + LOG.warning("Duplicate access control entry, " + "skipping allow...") else: message = (_('Failed to allow the access %(access)s to ' 'CIFS share %(name)s. Reason: %(err)s.') % @@ -1717,10 +1717,10 @@ class CIFSShare(StorageObject): % {'domain': domain, 'user': user_name}, re.I) if re.search(not_found_msg, expt.stdout): - LOG.warning(_LW("No access control entry found, " - "skipping deny...")) + LOG.warning("No access control entry found, " + "skipping deny...") elif re.search(user_err_msg, expt.stdout): - LOG.warning(_LW("User not found on domain, skipping deny...")) + LOG.warning("User not found on domain, skipping deny...") else: message = (_('Failed to deny the access %(access)s to ' 'CIFS share %(name)s. Reason: %(err)s.') % @@ -1799,7 +1799,7 @@ class NFSShare(StorageObject): status, out = self.get(name, mover_name) if constants.STATUS_NOT_FOUND == status: - LOG.warning(_LW("NFS share %s not found. Skip the deletion."), + LOG.warning("NFS share %s not found. Skip the deletion.", path) return @@ -1850,7 +1850,7 @@ class NFSShare(StorageObject): dup_msg = (r'%(mover_name)s : No such file or directory' % {'mover_name': mover_name}) if re.search(dup_msg, expt.stdout): - LOG.warning(_LW("NFS share %s not found."), name) + LOG.warning("NFS share %s not found.", name) return constants.STATUS_NOT_FOUND, None else: message = (_('Failed to list NFS share %(name)s on ' diff --git a/manila/share/drivers/ganesha/__init__.py b/manila/share/drivers/ganesha/__init__.py index 3010392047..0b93ae3bd3 100644 --- a/manila/share/drivers/ganesha/__init__.py +++ b/manila/share/drivers/ganesha/__init__.py @@ -24,7 +24,6 @@ import six from manila.common import constants from manila import exception -from manila.i18n import _LI from manila.share.drivers.ganesha import manager as ganesha_manager from manila.share.drivers.ganesha import utils as ganesha_utils @@ -76,7 +75,7 @@ class GaneshaNASHelper(NASHelperBase): if e.errno != errno.ENOENT or must_exist: raise dirlist = [] - LOG.info(_LI('Loading Ganesha config from %s.'), dirpath) + LOG.info('Loading Ganesha config from %s.', dirpath) conf_files = list(filter(self._confrx.search, dirlist)) conf_files.sort() export_template = {} diff --git a/manila/share/drivers/ganesha/manager.py b/manila/share/drivers/ganesha/manager.py index 80f4659599..5a29fabccc 100644 --- a/manila/share/drivers/ganesha/manager.py +++ b/manila/share/drivers/ganesha/manager.py @@ -24,7 +24,6 @@ import six from manila import exception from manila.i18n import _ -from manila.i18n import _LE from manila.share.drivers.ganesha import utils as ganesha_utils from manila import utils @@ -187,8 +186,8 @@ class GaneshaManager(object): except exception.ProcessExecutionError as e: if makelog: LOG.error( - _LE("Error while executing management command on " - "Ganesha node %(tag)s: %(msg)s."), + ("Error while executing management command on " + "Ganesha node %(tag)s: %(msg)s."), {'tag': tag, 'msg': msg}) raise exception.GaneshaCommandFailure( stdout=e.stdout, stderr=e.stderr, exit_code=e.exit_code, @@ -324,8 +323,8 @@ class GaneshaManager(object): run_as_root=False)[0] match = re.search('\Aexportid\|(\d+)$', out) if not match: - LOG.error(_LE("Invalid export database on " - "Ganesha node %(tag)s: %(db)s."), + LOG.error("Invalid export database on " + "Ganesha node %(tag)s: %(db)s.", {'tag': self.tag, 'db': self.ganesha_db_path}) raise exception.InvalidSqliteDB() return int(match.groups()[0]) diff --git a/manila/share/drivers/generic.py b/manila/share/drivers/generic.py index f7177e32b5..adb015cb1b 100644 --- a/manila/share/drivers/generic.py +++ b/manila/share/drivers/generic.py @@ -30,7 +30,7 @@ from manila.common import constants as const from manila import compute from manila import context from manila import exception -from manila.i18n import _, _LE, _LI, _LW +from manila.i18n import _ from manila.share import driver from manila.share.drivers import service_instance from manila import utils @@ -193,11 +193,11 @@ class GenericShareDriver(driver.ExecuteMixin, driver.ShareDriver): if not common_sv_available: time.sleep(sv_fetch_retry_interval) - LOG.warning(_LW("Waiting for the common service VM to become " - "available. " - "Driver is currently uninitialized. " - "Share server: %(share_server)s " - "Retry interval: %(retry_interval)s"), + LOG.warning("Waiting for the common service VM to become " + "available. " + "Driver is currently uninitialized. " + "Share server: %(share_server)s " + "Retry interval: %(retry_interval)s", dict(share_server=share_server, retry_interval=sv_fetch_retry_interval)) @@ -293,14 +293,14 @@ class GenericShareDriver(driver.ExecuteMixin, driver.ShareDriver): ['sudo', 'cp', const.MOUNT_FILE_TEMP, const.MOUNT_FILE], ) except exception.ProcessExecutionError as e: - LOG.error(_LE("Failed to sync mount files on server '%s'."), + LOG.error("Failed to sync mount files on server '%s'.", server_details['instance_id']) raise exception.ShareBackendException(msg=six.text_type(e)) try: # Remount it to avoid postponed point of failure self._ssh_exec(server_details, ['sudo', 'mount', '-a']) except exception.ProcessExecutionError as e: - LOG.error(_LE("Failed to mount all shares on server '%s'."), + LOG.error("Failed to mount all shares on server '%s'.", server_details['instance_id']) raise exception.ShareBackendException(msg=six.text_type(e)) @@ -346,8 +346,8 @@ class GenericShareDriver(driver.ExecuteMixin, driver.ShareDriver): # Add mount permanently self._sync_mount_temp_and_perm_files(server_details) else: - LOG.warning(_LW("Mount point '%(path)s' already exists on " - "server '%(server)s'."), log_data) + LOG.warning("Mount point '%(path)s' already exists on " + "server '%(server)s'.", log_data) except exception.ProcessExecutionError as e: raise exception.ShareBackendException(msg=six.text_type(e)) return _mount_device_with_lock() @@ -373,8 +373,8 @@ class GenericShareDriver(driver.ExecuteMixin, driver.ShareDriver): # Remove mount permanently self._sync_mount_temp_and_perm_files(server_details) else: - LOG.warning(_LW("Mount point '%(path)s' does not exist on " - "server '%(server)s'."), log_data) + LOG.warning("Mount point '%(path)s' does not exist on " + "server '%(server)s'.", log_data) return _unmount_device_with_lock() def _get_mount_path(self, share): @@ -449,10 +449,10 @@ class GenericShareDriver(driver.ExecuteMixin, driver.ShareDriver): return volumes_list[0] elif len(volumes_list) > 1: LOG.error( - _LE("Expected only one volume in volume list with name " - "'%(name)s', but got more than one in a result - " - "'%(result)s'."), { - 'name': volume_name, 'result': volumes_list}) + "Expected only one volume in volume list with name " + "'%(name)s', but got more than one in a result - " + "'%(result)s'.", { + 'name': volume_name, 'result': volumes_list}) raise exception.ManilaException( _("Error. Ambiguous volumes for name '%s'") % volume_name) return None @@ -479,11 +479,11 @@ class GenericShareDriver(driver.ExecuteMixin, driver.ShareDriver): volume_snapshot = volume_snapshot_list[0] elif len(volume_snapshot_list) > 1: LOG.error( - _LE("Expected only one volume snapshot in list with name " - "'%(name)s', but got more than one in a result - " - "'%(result)s'."), { - 'name': volume_snapshot_name, - 'result': volume_snapshot_list}) + "Expected only one volume snapshot in list with name" + "'%(name)s', but got more than one in a result - " + "'%(result)s'.", { + 'name': volume_snapshot_name, + 'result': volume_snapshot_list}) raise exception.ManilaException( _('Error. Ambiguous volume snaphots')) return volume_snapshot @@ -501,8 +501,8 @@ class GenericShareDriver(driver.ExecuteMixin, driver.ShareDriver): try: volume = self._get_volume(context, share['id']) except exception.VolumeNotFound: - LOG.warning(_LW("Volume not found for share %s. " - "Possibly already deleted."), share['id']) + LOG.warning("Volume not found for share %s. " + "Possibly already deleted.", share['id']) volume = None if volume and volume['id'] in attached_volumes: self.compute_api.instance_volume_detach( @@ -587,7 +587,7 @@ class GenericShareDriver(driver.ExecuteMixin, driver.ShareDriver): try: volume = self._get_volume(context, share['id']) except exception.VolumeNotFound: - LOG.info(_LI("Volume not found. Already deleted?")) + LOG.info("Volume not found. Already deleted?") volume = None if volume: if volume['status'] == 'in-use': diff --git a/manila/share/drivers/glusterfs/common.py b/manila/share/drivers/glusterfs/common.py index 91bce5e702..8d5fc1f07a 100644 --- a/manila/share/drivers/glusterfs/common.py +++ b/manila/share/drivers/glusterfs/common.py @@ -24,7 +24,7 @@ from oslo_log import log import six from manila import exception -from manila.i18n import _, _LE, _LW +from manila.i18n import _ from manila.share.drivers.ganesha import utils as ganesha_utils LOG = log.getLogger(__name__) @@ -205,7 +205,7 @@ class GlusterManager(object): exc.exit_code in error_policy): return if logmsg: - LOG.error(_LE("%s: GlusterFS instrumentation failed.") % + LOG.error("%s: GlusterFS instrumentation failed." % logmsg) raise exception.GlusterfsException( _("GlusterFS management command '%(cmd)s' failed " @@ -248,7 +248,7 @@ class GlusterManager(object): def _get_vol_option_via_info(self, option): """Get the value of an option set on a GlusterFS volume via volinfo.""" args = ('--xml', 'volume', 'info', self.volume) - out, err = self.gluster_call(*args, log=_LE("retrieving volume info")) + out, err = self.gluster_call(*args, log=("retrieving volume info")) if not out: raise exception.GlusterfsException( @@ -332,7 +332,7 @@ class GlusterManager(object): :returns: version (as tuple of strings, example: ('3', '6', '0beta2')) """ out, err = self.gluster_call('--version', - log=_LE("GlusterFS version query")) + log=("GlusterFS version query")) try: owords = out.split() if owords[0] != 'glusterfs': @@ -393,7 +393,7 @@ def _mount_gluster_vol(execute, gluster_export, mount_path, ensure=False): execute(*command, run_as_root=True) except exception.ProcessExecutionError as exc: if ensure and 'already mounted' in exc.stderr: - LOG.warning(_LW("%s is already mounted."), gluster_export) + LOG.warning("%s is already mounted.", gluster_export) else: raise exception.GlusterfsException( 'Unable to mount Gluster volume' @@ -431,8 +431,8 @@ def _restart_gluster_vol(gluster_mgr): # this odd-behaviour of Gluster-CLI. gluster_mgr.gluster_call( 'volume', 'stop', gluster_mgr.volume, '--mode=script', - log=_LE("stopping GlusterFS volume %s") % gluster_mgr.volume) + log=("stopping GlusterFS volume %s") % gluster_mgr.volume) gluster_mgr.gluster_call( 'volume', 'start', gluster_mgr.volume, - log=_LE("starting GlusterFS volume %s") % gluster_mgr.volume) + log=("starting GlusterFS volume %s") % gluster_mgr.volume) diff --git a/manila/share/drivers/glusterfs/layout_directory.py b/manila/share/drivers/glusterfs/layout_directory.py index 0bceb54a61..75725fe2ab 100644 --- a/manila/share/drivers/glusterfs/layout_directory.py +++ b/manila/share/drivers/glusterfs/layout_directory.py @@ -22,7 +22,7 @@ from oslo_log import log import six from manila import exception -from manila.i18n import _, _LE +from manila.i18n import _ from manila.share.drivers.glusterfs import common from manila.share.drivers.glusterfs import layout @@ -82,8 +82,8 @@ class GlusterfsDirectoryMappedLayout(layout.GlusterfsShareLayoutBase): except exception.GlusterfsException: if (self.gluster_manager. get_vol_option('features.quota')) != 'on': - LOG.exception(_LE("Error in tuning GlusterFS volume to enable " - "creation of shares of specific size.")) + LOG.exception("Error in tuning GlusterFS volume to enable " + "creation of shares of specific size.") raise self._ensure_gluster_vol_mounted() @@ -106,7 +106,7 @@ class GlusterfsDirectoryMappedLayout(layout.GlusterfsShareLayoutBase): self.gluster_manager.export, mount_path, ensure=True) except exception.GlusterfsException: - LOG.exception(_LE('Could not mount the Gluster volume %s'), + LOG.exception('Could not mount the Gluster volume %s', self.gluster_manager.volume) raise @@ -152,7 +152,7 @@ class GlusterfsDirectoryMappedLayout(layout.GlusterfsShareLayoutBase): exc = exception.GlusterfsException(exc) if isinstance(exc, exception.GlusterfsException): self._cleanup_create_share(local_share_path, share['name']) - LOG.error(_LE('Unable to create share %s'), share['name']) + LOG.error('Unable to create share %s', share['name']) raise exc comp_share = self.gluster_manager.components.copy() @@ -170,9 +170,9 @@ class GlusterfsDirectoryMappedLayout(layout.GlusterfsShareLayoutBase): try: self.driver._execute(*cmd, run_as_root=True) except exception.ProcessExecutionError as exc: - LOG.error(_LE('Cannot cleanup share, %s, that errored out ' - 'during its creation, but exists in GlusterFS ' - 'volume.'), share_name) + LOG.error('Cannot cleanup share, %s, that errored out ' + 'during its creation, but exists in GlusterFS ' + 'volume.', share_name) raise exception.GlusterfsException(exc) def delete_share(self, context, share, share_server=None): @@ -182,7 +182,7 @@ class GlusterfsDirectoryMappedLayout(layout.GlusterfsShareLayoutBase): try: self.driver._execute(*cmd, run_as_root=True) except exception.ProcessExecutionError: - LOG.exception(_LE('Unable to delete share %s'), share['name']) + LOG.exception('Unable to delete share %s', share['name']) raise def ensure_share(self, context, share, share_server=None): diff --git a/manila/share/drivers/glusterfs/layout_volume.py b/manila/share/drivers/glusterfs/layout_volume.py index 6e20f30bad..818c5cf109 100644 --- a/manila/share/drivers/glusterfs/layout_volume.py +++ b/manila/share/drivers/glusterfs/layout_volume.py @@ -28,7 +28,7 @@ from oslo_log import log import six from manila import exception -from manila.i18n import _, _LE, _LI, _LW +from manila.i18n import _ from manila.share.drivers.glusterfs import common from manila.share.drivers.glusterfs import layout from manila import utils @@ -129,8 +129,8 @@ class GlusterfsVolumeMappedLayout(layout.GlusterfsShareLayoutBase): exceptions[srvaddr] = six.text_type(exc) if exceptions: for srvaddr, excmsg in exceptions.items(): - LOG.error(_LE("'gluster version' failed on server " - "%(server)s with: %(message)s"), + LOG.error("'gluster version' failed on server " + "%(server)s with: %(message)s", {'server': srvaddr, 'message': excmsg}) raise exception.GlusterfsException(_( "'gluster version' failed on servers %s") % ( @@ -143,9 +143,9 @@ class GlusterfsVolumeMappedLayout(layout.GlusterfsShareLayoutBase): gluster_version_min_str = '.'.join( six.text_type(c) for c in self.driver.GLUSTERFS_VERSION_MIN) for srvaddr in notsupp_servers: - LOG.error(_LE("GlusterFS version %(version)s on server " - "%(server)s is not supported, " - "minimum requirement: %(minvers)s"), + LOG.error("GlusterFS version %(version)s on server " + "%(server)s is not supported, " + "minimum requirement: %(minvers)s", {'server': srvaddr, 'version': '.'.join(glusterfs_versions[srvaddr]), 'minvers': gluster_version_min_str}) @@ -167,8 +167,8 @@ class GlusterfsVolumeMappedLayout(layout.GlusterfsShareLayoutBase): LOG.error(msg) raise exception.GlusterfsException(msg) - LOG.info(_LI("Found %d Gluster volumes allocated for Manila." - ), len(gluster_volumes_initial)) + LOG.info("Found %d Gluster volumes allocated for Manila.", + len(gluster_volumes_initial)) self._check_mount_glusterfs() @@ -203,10 +203,10 @@ class GlusterfsVolumeMappedLayout(layout.GlusterfsShareLayoutBase): for srvaddr in self.configuration.glusterfs_servers: gluster_mgr = self._glustermanager(srvaddr, False) if gluster_mgr.user: - logmsg = _LE("Retrieving volume list " - "on host %s") % gluster_mgr.host + logmsg = ("Retrieving volume list " + "on host %s") % gluster_mgr.host else: - logmsg = _LE("Retrieving volume list") + logmsg = ("Retrieving volume list") out, err = gluster_mgr.gluster_call('volume', 'list', log=logmsg) for volname in out.split("\n"): patmatch = self.volume_pattern.match(volname) @@ -251,17 +251,17 @@ class GlusterfsVolumeMappedLayout(layout.GlusterfsShareLayoutBase): if not unused_vols: # No volumes available for use as share. Warn user. - LOG.warning(_LW("No unused gluster volumes available for use as " - "share! Create share won't be supported unless " - "existing shares are deleted or some gluster " - "volumes are created with names matching " - "'glusterfs_volume_pattern'.")) + LOG.warning("No unused gluster volumes available for use as " + "share! Create share won't be supported unless " + "existing shares are deleted or some gluster " + "volumes are created with names matching " + "'glusterfs_volume_pattern'.") else: - LOG.info(_LI("Number of gluster volumes in use: " - "%(inuse-numvols)s. Number of gluster volumes " - "available for use as share: %(unused-numvols)s"), + LOG.info("Number of gluster volumes in use: " + "%(inuse-numvols)s. Number of gluster volumes " + "available for use as share: %(unused-numvols)s", {'inuse-numvols': len(self.gluster_used_vols), - 'unused-numvols': len(unused_vols)}) + 'unused-numvols': len(unused_vols)}) # volmap is the data structure used to categorize and sort # the unused volumes. It's a nested dictionary of structure @@ -385,7 +385,7 @@ class GlusterfsVolumeMappedLayout(layout.GlusterfsShareLayoutBase): try: vol = self._pop_gluster_vol(share['size']) except exception.GlusterfsException: - msg = (_LE("Error creating share %(share_id)s"), + msg = ("Error creating share %(share_id)s", {'share_id': share['id']}) LOG.error(msg) raise @@ -401,7 +401,7 @@ class GlusterfsVolumeMappedLayout(layout.GlusterfsShareLayoutBase): # For native protocol, the export_location should be of the form: # server:/volname - LOG.info(_LI("export_location sent back from create_share: %s"), + LOG.info("export_location sent back from create_share: %s", export) return export @@ -436,8 +436,8 @@ class GlusterfsVolumeMappedLayout(layout.GlusterfsShareLayoutBase): self._push_gluster_vol(gmgr.qualified) except exception.GlusterfsException: - msg = (_LE("Error during delete_share request for " - "share %(share_id)s"), {'share_id': share['id']}) + msg = ("Error during delete_share request for " + "share %(share_id)s", {'share_id': share['id']}) LOG.error(msg) raise @@ -449,7 +449,7 @@ class GlusterfsVolumeMappedLayout(layout.GlusterfsShareLayoutBase): args = ('snapshot', 'list', gluster_mgr.volume, '--mode=script') out, err = gluster_mgr.gluster_call( *args, - log=_LE("Retrieving snapshot list")) + log=("Retrieving snapshot list")) snapgrep = list(filter(lambda x: snapshot['id'] in x, out.split("\n"))) if len(snapgrep) != 1: msg = (_("Failed to identify backing GlusterFS object " @@ -493,7 +493,7 @@ class GlusterfsVolumeMappedLayout(layout.GlusterfsShareLayoutBase): for args in args_tuple: out, err = old_gmgr.gluster_call( *args, - log=_LE("Creating share from snapshot")) + log=("Creating share from snapshot")) # Get a manager for the new volume/share. comp_vol = old_gmgr.components.copy() @@ -509,7 +509,7 @@ class GlusterfsVolumeMappedLayout(layout.GlusterfsShareLayoutBase): ('start', [])) for op, opargs in argseq: args = ['volume', op, gmgr.volume] + opargs - gmgr.gluster_call(*args, log=_LE("Creating share from snapshot")) + gmgr.gluster_call(*args, log=("Creating share from snapshot")) self.gluster_used_vols.add(gmgr.qualified) self.private_storage.update(share['id'], {'volume': gmgr.qualified}) @@ -528,7 +528,7 @@ class GlusterfsVolumeMappedLayout(layout.GlusterfsShareLayoutBase): gluster_mgr.volume) out, err = gluster_mgr.gluster_call( *args, - log=_LE("Retrieving volume info")) + log=("Retrieving volume info")) if not out: raise exception.GlusterfsException( @@ -570,7 +570,7 @@ class GlusterfsVolumeMappedLayout(layout.GlusterfsShareLayoutBase): '--mode=script') out, err = gluster_mgr.gluster_call( *args, - log=_LE("Error deleting snapshot")) + log=("Error deleting snapshot")) if not out: raise exception.GlusterfsException( diff --git a/manila/share/drivers/helpers.py b/manila/share/drivers/helpers.py index 63f8e54a75..d82cf8f5e6 100644 --- a/manila/share/drivers/helpers.py +++ b/manila/share/drivers/helpers.py @@ -20,7 +20,7 @@ from oslo_log import log from manila.common import constants as const from manila import exception -from manila.i18n import _, _LW +from manila.i18n import _ from manila import utils LOG = log.getLogger(__name__) @@ -242,12 +242,12 @@ class NFSHelper(NASHelperBase): (const.ACCESS_LEVEL_RO, const.ACCESS_LEVEL_RW)) except (exception.InvalidShareAccess, exception.InvalidShareAccessLevel): - LOG.warning(_LW( + LOG.warning( "Unsupported access level %(level)s or access type " "%(type)s, skipping removal of access rule to " - "%(to)s.") % {'level': access['access_level'], - 'type': access['access_type'], - 'to': access['access_to']}) + "%(to)s." % {'level': access['access_level'], + 'type': access['access_type'], + 'to': access['access_to']}) continue self._ssh_exec(server, ['sudo', 'exportfs', '-u', ':'.join((access['access_to'], local_path))]) @@ -260,12 +260,12 @@ class NFSHelper(NASHelperBase): re.escape(local_path) + '[\s\n]*' + re.escape( access['access_to']), out) if found_item is not None: - LOG.warning(_LW("Access rule %(type)s:%(to)s already " - "exists for share %(name)s") % { - 'to': access['access_to'], - 'type': access['access_type'], - 'name': share_name - }) + LOG.warning("Access rule %(type)s:%(to)s already " + "exists for share %(name)s" % { + 'to': access['access_to'], + 'type': access['access_type'], + 'name': share_name + }) else: rules_options = '%s,no_subtree_check' if access['access_level'] == const.ACCESS_LEVEL_RW: @@ -433,8 +433,8 @@ class CIFSHelperIPAccess(CIFSHelperBase): self._ssh_exec( server, ['sudo', 'net', 'conf', 'delshare', share_name]) except exception.ProcessExecutionError as e: - LOG.warning(_LW("Caught error trying delete share: %(error)s, try" - "ing delete it forcibly."), {'error': e.stderr}) + LOG.warning("Caught error trying delete share: %(error)s, try" + "ing delete it forcibly.", {'error': e.stderr}) self._ssh_exec(server, ['sudo', 'smbcontrol', 'all', 'close-share', share_name]) diff --git a/manila/share/drivers/hitachi/hnas/driver.py b/manila/share/drivers/hitachi/hnas/driver.py index 9be11c3a88..71b6efbb50 100644 --- a/manila/share/drivers/hitachi/hnas/driver.py +++ b/manila/share/drivers/hitachi/hnas/driver.py @@ -23,7 +23,7 @@ import six from manila.common import constants from manila import exception -from manila.i18n import _, _LI, _LW +from manila.i18n import _ from manila.share import driver from manila.share import utils @@ -280,10 +280,10 @@ class HitachiHNASDriver(driver.ShareDriver): for rule in delete_rules: if rule['access_type'].lower() != 'user': - LOG.warning(_LW('Only USER access type is allowed for ' - 'CIFS. %(entity_type)s ' - 'provided %(share)s with ' - 'protocol %(proto)s.'), + LOG.warning('Only USER access type is allowed for ' + 'CIFS. %(entity_type)s ' + 'provided %(share)s with ' + 'protocol %(proto)s.', {'entity_type': entity_type.capitalize(), 'share': share_or_snapshot['id'], 'proto': share_proto}) @@ -411,7 +411,7 @@ class HitachiHNASDriver(driver.ShareDriver): 'snap_id': snapshot['id']}) export_locations = self._create_snapshot(hnas_share_id, snapshot) - LOG.info(_LI("Snapshot %(id)s successfully created."), + LOG.info("Snapshot %(id)s successfully created.", {'id': snapshot['id']}) output = { @@ -443,7 +443,7 @@ class HitachiHNASDriver(driver.ShareDriver): self._delete_snapshot(snapshot['share'], hnas_share_id, hnas_snapshot_id) - LOG.info(_LI("Snapshot %(id)s successfully deleted."), + LOG.info("Snapshot %(id)s successfully deleted.", {'id': snapshot['id']}) def create_share_from_snapshot(self, context, share, snapshot, @@ -598,8 +598,8 @@ class HitachiHNASDriver(driver.ShareDriver): {'shr_id': share['id']}) self._extend_share(hnas_share_id, share, new_size) - LOG.info(_LI("Share %(shr_id)s successfully extended to " - "%(shr_size)s."), + LOG.info("Share %(shr_id)s successfully extended to " + "%(shr_size)s.", {'shr_id': share['id'], 'shr_size': six.text_type(new_size)}) @@ -639,7 +639,7 @@ class HitachiHNASDriver(driver.ShareDriver): 'mount_snapshot_support': True, } - LOG.info(_LI("HNAS Capabilities: %(data)s."), + LOG.info("HNAS Capabilities: %(data)s.", {'data': six.text_type(data)}) super(HitachiHNASDriver, self)._update_share_stats(data) @@ -719,8 +719,8 @@ class HitachiHNASDriver(driver.ShareDriver): if share['share_proto'].lower() == 'nfs': # 10.0.0.1:/shares/example - LOG.info(_LI("Share %(shr_path)s will be managed with ID " - "%(shr_id)s."), + LOG.info("Share %(shr_path)s will be managed with ID " + "%(shr_id)s.", {'shr_path': share['export_locations'][0]['path'], 'shr_id': share['id']}) @@ -764,8 +764,8 @@ class HitachiHNASDriver(driver.ShareDriver): "Share ID %(share_id)s", {'hnas_id': hnas_share_id, 'share_id': share['id']}) - LOG.info(_LI("Share %(shr_path)s was successfully managed with ID " - "%(shr_id)s."), + LOG.info("Share %(shr_path)s was successfully managed with ID " + "%(shr_id)s.", {'shr_path': share['export_locations'][0]['path'], 'shr_id': share['id']}) @@ -779,13 +779,13 @@ class HitachiHNASDriver(driver.ShareDriver): self.private_storage.delete(share['id']) if len(share['export_locations']) == 0: - LOG.info(_LI("The share with ID %(shr_id)s is no longer being " - "managed."), {'shr_id': share['id']}) + LOG.info("The share with ID %(shr_id)s is no longer being " + "managed.", {'shr_id': share['id']}) else: - LOG.info(_LI("The share with current path %(shr_path)s and ID " - "%(shr_id)s is no longer being managed."), + LOG.info("The share with current path %(shr_path)s and ID " + "%(shr_id)s is no longer being managed.", {'shr_path': share['export_locations'][0]['path'], - 'shr_id': share['id']}) + 'shr_id': share['id']}) def shrink_share(self, share, new_size, share_server=None): """Shrinks a share to new size. @@ -801,8 +801,8 @@ class HitachiHNASDriver(driver.ShareDriver): {'shr_id': share['id']}) self._shrink_share(hnas_share_id, share, new_size) - LOG.info(_LI("Share %(shr_id)s successfully shrunk to " - "%(shr_size)sG."), + LOG.info("Share %(shr_id)s successfully shrunk to " + "%(shr_size)sG.", {'shr_id': share['id'], 'shr_size': six.text_type(new_size)}) @@ -836,12 +836,12 @@ class HitachiHNASDriver(driver.ShareDriver): try: self.hnas.tree_clone(src_path, dest_path) except exception.HNASNothingToCloneException: - LOG.warning(_LW("Source directory is empty, creating an empty " - "directory.")) + LOG.warning("Source directory is empty, creating an empty " + "directory.") - LOG.info(_LI("Share %(share)s successfully reverted to snapshot " - "%(snapshot)s."), {'share': snapshot['share_id'], - 'snapshot': snapshot['id']}) + LOG.info("Share %(share)s successfully reverted to snapshot " + "%(snapshot)s.", {'share': snapshot['share_id'], + 'snapshot': snapshot['id']}) def _get_hnas_share_id(self, share_id): hnas_id = self.private_storage.get(share_id, 'hnas_id') @@ -1056,8 +1056,8 @@ class HitachiHNASDriver(driver.ShareDriver): try: self.hnas.tree_clone(src_path, dest_path) except exception.HNASNothingToCloneException: - LOG.warning(_LW("Source directory is empty, creating an empty " - "directory.")) + LOG.warning("Source directory is empty, creating an empty " + "directory.") self.hnas.create_directory(dest_path) finally: if share_proto.lower() == 'nfs': @@ -1125,8 +1125,8 @@ class HitachiHNASDriver(driver.ShareDriver): try: self.hnas.tree_clone(src_path, dest_path) except exception.HNASNothingToCloneException: - LOG.warning(_LW("Source directory is empty, exporting " - "directory.")) + LOG.warning("Source directory is empty, exporting " + "directory.") self._check_protocol(share['id'], share['share_proto']) @@ -1348,8 +1348,8 @@ class HitachiHNASDriver(driver.ShareDriver): try: self._ensure_snapshot(snapshot, hnas_snapshot_id) except exception.HNASItemNotFoundException: - LOG.warning(_LW("Export does not exist for snapshot %s, " - "creating a new one."), snapshot['id']) + LOG.warning("Export does not exist for snapshot %s, " + "creating a new one.", snapshot['id']) self._create_export(hnas_share_id, snapshot['share']['share_proto'], snapshot_id=hnas_snapshot_id) @@ -1362,8 +1362,8 @@ class HitachiHNASDriver(driver.ShareDriver): is_snapshot=True) output['export_locations'] = export_locations - LOG.info(_LI("Snapshot %(snap_path)s for share %(shr_id)s was " - "successfully managed with ID %(snap_id)s."), + LOG.info("Snapshot %(snap_path)s for share %(shr_id)s was " + "successfully managed with ID %(snap_id)s.", {'snap_path': snapshot['provider_location'], 'shr_id': snapshot['share_id'], 'snap_id': snapshot['id']}) @@ -1375,9 +1375,9 @@ class HitachiHNASDriver(driver.ShareDriver): :param snapshot: Snapshot that will be unmanaged. """ - LOG.info(_LI("The snapshot with ID %(snap_id)s from share " - "%(share_id)s is no longer being managed by Manila. " - "However, it is not deleted and can be found in HNAS."), + LOG.info("The snapshot with ID %(snap_id)s from share " + "%(share_id)s is no longer being managed by Manila. " + "However, it is not deleted and can be found in HNAS.", {'snap_id': snapshot['id'], 'share_id': snapshot['share_id']}) diff --git a/manila/share/drivers/hitachi/hnas/ssh.py b/manila/share/drivers/hitachi/hnas/ssh.py index 082272ff01..38cd7ed211 100644 --- a/manila/share/drivers/hitachi/hnas/ssh.py +++ b/manila/share/drivers/hitachi/hnas/ssh.py @@ -24,7 +24,7 @@ import os import time from manila import exception -from manila.i18n import _, _LE, _LW +from manila.i18n import _ from manila import utils as mutils LOG = log.getLogger(__name__) @@ -98,8 +98,8 @@ class HNASSSHBackend(object): self._execute(command) except processutils.ProcessExecutionError as e: if 'does not exist' in e.stderr: - LOG.warning(_LW("Export %s does not exist on " - "backend anymore."), name) + LOG.warning("Export %s does not exist on " + "backend anymore.", name) else: msg = _("Could not delete NFS export %s.") % name LOG.exception(msg) @@ -128,8 +128,8 @@ class HNASSSHBackend(object): self._execute(command) except processutils.ProcessExecutionError as e: if e.exit_code == 1: - LOG.warning(_LW("CIFS share %s does not exist on " - "backend anymore."), name) + LOG.warning("CIFS share %s does not exist on " + "backend anymore.", name) else: msg = _("Could not delete CIFS share %s.") % name LOG.exception(msg) @@ -232,18 +232,18 @@ class HNASSSHBackend(object): except processutils.ProcessExecutionError as e: if ('not listed as a user' in e.stderr or 'Could not delete user/group' in e.stderr): - LOG.warning(_LW('User %(user)s already not allowed to access ' - '%(entity_type)s %(name)s.'), { - 'entity_type': entity_type, - 'user': user, - 'name': name - }) + LOG.warning('User %(user)s already not allowed to access ' + '%(entity_type)s %(name)s.', { + 'entity_type': entity_type, + 'user': user, + 'name': name + }) else: msg = _("Could not delete access of user %(user)s to " "%(entity_type)s %(name)s.") % { - 'user': user, - 'name': name, - 'entity_type': entity_type, + 'user': user, + 'name': name, + 'entity_type': entity_type, } LOG.exception(msg) raise exception.HNASBackendException(msg=msg) @@ -312,8 +312,8 @@ class HNASSSHBackend(object): if now > deadline: command = ['tree-clone-job-abort', job_id] self._execute(command) - LOG.error(_LE("Timeout in snapshot creation from " - "source path %s.") % src_path) + LOG.error("Timeout in snapshot creation from " + "source path %s." % src_path) msg = _("Share snapshot of source path %s " "was not created.") % src_path raise exception.HNASBackendException(msg=msg) @@ -332,7 +332,7 @@ class HNASSSHBackend(object): {'src': src_path, 'dest': dest_path}) else: - LOG.error(_LE('Error creating snapshot of source path %s.'), + LOG.error('Error creating snapshot of source path %s.', src_path) msg = _('Snapshot of source path %s was not ' 'created.') % src_path @@ -345,8 +345,8 @@ class HNASSSHBackend(object): self._execute(command) except processutils.ProcessExecutionError as e: if 'Source path: Cannot access' in e.stderr: - LOG.warning(_LW("Attempted to delete path %s " - "but it does not exist."), path) + LOG.warning("Attempted to delete path %s " + "but it does not exist.", path) else: msg = _("Could not submit tree delete job to delete path " "%s.") % path @@ -449,7 +449,7 @@ class HNASSSHBackend(object): self._execute(command) except processutils.ProcessExecutionError as e: if 'Source path: Cannot access' in e.stderr: - LOG.warning(_LW("Share %s does not exist."), vvol_name) + LOG.warning("Share %s does not exist.", vvol_name) else: msg = _("Failed to delete vvol %s.") % vvol_name LOG.exception(msg) @@ -699,8 +699,8 @@ class HNASSSHBackend(object): LOG.debug(msg) raise exception.HNASDirectoryNotEmpty(msg=msg) elif 'cannot remove' in e.stderr and 'NotFound' in e.stderr: - LOG.warning(_LW("Attempted to delete path %s but it does " - "not exist."), path) + LOG.warning("Attempted to delete path %s but it does " + "not exist.", path) elif 'Current file system invalid: VolumeNotFound' in e.stderr: msg = _("Command to delete empty directory %s failed due " "to context change.") % path diff --git a/manila/share/drivers/hitachi/hsp/driver.py b/manila/share/drivers/hitachi/hsp/driver.py index 6fc9e9e506..85b0e7780e 100644 --- a/manila/share/drivers/hitachi/hsp/driver.py +++ b/manila/share/drivers/hitachi/hsp/driver.py @@ -20,7 +20,7 @@ from oslo_utils import units from manila.common import constants from manila import exception -from manila.i18n import _, _LE, _LI +from manila.i18n import _ from manila.share import driver from manila.share.drivers.hitachi.hsp import rest @@ -92,7 +92,7 @@ class HitachiHSPDriver(driver.ShareDriver): }], } - LOG.info(_LI("Hitachi HSP Capabilities: %(data)s."), + LOG.info("Hitachi HSP Capabilities: %(data)s.", {'data': data}) super(HitachiHSPDriver, self)._update_share_stats(data) @@ -111,7 +111,7 @@ class HitachiHSPDriver(driver.ShareDriver): except exception.HSPBackendException: with excutils.save_and_reraise_exception(): self.hsp.delete_file_system(filesystem_id) - msg = _LE("Could not create share %s on HSP.") + msg = ("Could not create share %s on HSP.") LOG.exception(msg, share['id']) uri = self.hsp_host + ':/' + share['id'] @@ -133,7 +133,7 @@ class HitachiHSPDriver(driver.ShareDriver): filesystem_id = self.hsp.get_file_system(share['id'])['id'] hsp_share_id = self.hsp.get_share(filesystem_id)['id'] except exception.HSPItemNotFoundException: - LOG.info(_LI("Share %(shr)s already removed from backend."), + LOG.info("Share %(shr)s already removed from backend.", {'shr': share['id']}) if hsp_share_id: @@ -278,8 +278,8 @@ class HitachiHSPDriver(driver.ShareDriver): % share['id']) raise exception.HSPBackendException(msg=msg) - LOG.info(_LI("Share %(shr_id)s successfully extended to " - "%(shr_size)sG."), + LOG.info("Share %(shr_id)s successfully extended to " + "%(shr_size)sG.", {'shr_id': share['id'], 'shr_size': new_size}) @@ -299,8 +299,8 @@ class HitachiHSPDriver(driver.ShareDriver): raise exception.ShareShrinkingPossibleDataLoss( share_id=share['id']) - LOG.info(_LI("Share %(shr_id)s successfully shrunk to " - "%(shr_size)sG."), + LOG.info("Share %(shr_id)s successfully shrunk to " + "%(shr_size)sG.", {'shr_id': share['id'], 'shr_size': new_size}) @@ -333,8 +333,8 @@ class HitachiHSPDriver(driver.ShareDriver): file_system = self.hsp.get_file_system(share['id']) - LOG.info(_LI("Share %(shr_path)s was successfully managed with ID " - "%(shr_id)s."), + LOG.info("Share %(shr_path)s was successfully managed with ID " + "%(shr_id)s.", {'shr_path': share['export_locations'][0]['path'], 'shr_id': share['id']}) @@ -357,8 +357,8 @@ class HitachiHSPDriver(driver.ShareDriver): self.private_storage.delete(share['id']) - LOG.info(_LI("The share with current path %(shr_path)s and ID " - "%(shr_id)s is no longer being managed."), + LOG.info("The share with current path %(shr_path)s and ID " + "%(shr_id)s is no longer being managed.", {'shr_path': share['export_locations'][0]['path'], 'shr_id': share['id']}) diff --git a/manila/share/drivers/hpe/hpe_3par_driver.py b/manila/share/drivers/hpe/hpe_3par_driver.py index 8aa40a279a..6f1fbc22cd 100644 --- a/manila/share/drivers/hpe/hpe_3par_driver.py +++ b/manila/share/drivers/hpe/hpe_3par_driver.py @@ -27,7 +27,7 @@ import six from manila.common import config from manila import exception -from manila.i18n import _, _LI +from manila.i18n import _ from manila.share import driver from manila.share.drivers.hpe import hpe_3par_mediator from manila.share import share_types @@ -237,7 +237,7 @@ class HPE3ParShareDriver(driver.ShareDriver): def do_setup(self, context): """Any initialization the share driver does while starting.""" - LOG.info(_LI("Starting share driver %(driver_name)s (%(version)s)"), + LOG.info("Starting share driver %(driver_name)s (%(version)s)", {'driver_name': self.__class__.__name__, 'version': self.VERSION}) @@ -637,8 +637,8 @@ class HPE3ParShareDriver(driver.ShareDriver): if not self._hpe3par: LOG.info( - _LI("Skipping capacity and capabilities update. Setup has not " - "completed.")) + "Skipping capacity and capabilities update. Setup has not " + "completed.") else: for fpg in self.fpgs: fpg_status = self._hpe3par.get_fpg_status(fpg) diff --git a/manila/share/drivers/hpe/hpe_3par_mediator.py b/manila/share/drivers/hpe/hpe_3par_mediator.py index 237a611f5d..363f17c3d2 100644 --- a/manila/share/drivers/hpe/hpe_3par_mediator.py +++ b/manila/share/drivers/hpe/hpe_3par_mediator.py @@ -25,8 +25,8 @@ import six from manila.data import utils as data_utils from manila import exception +from manila.i18n import _ from manila import utils -from manila.i18n import _, _LE, _LI, _LW hpe3parclient = importutils.try_import("hpe3parclient") if hpe3parclient: @@ -167,14 +167,14 @@ class HPE3ParMediator(object): LOG.exception(msg) raise exception.ShareBackendException(message=msg) - LOG.info(_LI("HPE3ParMediator %(version)s, " - "hpe3parclient %(client_version)s"), + LOG.info("HPE3ParMediator %(version)s, " + "hpe3parclient %(client_version)s", {"version": self.VERSION, "client_version": hpe3parclient.get_version_string()}) try: wsapi_version = self._client.getWsApiVersion()['build'] - LOG.info(_LI("3PAR WSAPI %s"), wsapi_version) + LOG.info("3PAR WSAPI %s", wsapi_version) except Exception as e: msg = (_('Failed to get 3PAR WSAPI version: %s') % six.text_type(e)) @@ -200,7 +200,7 @@ class HPE3ParMediator(object): try: self._client.http.unauthenticate() except Exception as e: - msg = _LW("Failed to Logout from 3PAR (%(url)s) because %(err)s") + msg = ("Failed to Logout from 3PAR (%(url)s) because %(err)s") LOG.warning(msg, {'url': self.hpe3par_api_url, 'err': six.text_type(e)}) # don't raise exception on logout() @@ -346,8 +346,8 @@ class HPE3ParMediator(object): if nfs_options is None: nfs_options = extra_specs.get('hp3par:nfs_options') if nfs_options: - msg = _LW("hp3par:nfs_options is deprecated. Use " - "hpe3par:nfs_options instead.") + msg = ("hp3par:nfs_options is deprecated. Use " + "hpe3par:nfs_options instead.") LOG.warning(msg) if nfs_options: @@ -391,8 +391,8 @@ class HPE3ParMediator(object): comment=comment) if 'hp3par_flash_cache' in extra_specs: - msg = _LW("hp3par_flash_cache is deprecated. Use " - "hpe3par_flash_cache instead.") + msg = ("hp3par_flash_cache is deprecated. Use " + "hpe3par_flash_cache instead.") LOG.warning(msg) if protocol == 'nfs': @@ -425,8 +425,8 @@ class HPE3ParMediator(object): if opt_value is None: opt_value = extra_specs.get('hp3par:smb_%s' % smb_opt) if opt_value: - msg = _LW("hp3par:smb_* is deprecated. Use " - "hpe3par:smb_* instead.") + msg = ("hp3par:smb_* is deprecated. Use " + "hpe3par:smb_* instead.") LOG.warning(msg) if opt_value: @@ -653,10 +653,10 @@ class HPE3ParMediator(object): if protocol == "smb" and (not self.hpe3par_cifs_admin_access_username or not self.hpe3par_cifs_admin_access_password): - LOG.warning(_LW("hpe3par_cifs_admin_access_username and " - "hpe3par_cifs_admin_access_password must be " - "provided in order for CIFS shares created from " - "snapshots to be writable.")) + LOG.warning("hpe3par_cifs_admin_access_username and " + "hpe3par_cifs_admin_access_password must be " + "provided in order for CIFS shares created from " + "snapshots to be writable.") return self.create_share( orig_project_id, share_id, @@ -735,8 +735,8 @@ class HPE3ParMediator(object): protocol, fpg, vfs, fstore, comment) except Exception as e: - msg = _LE('Exception during mount and copy from RO snapshot ' - 'to RW share: %s') + msg = ('Exception during mount and copy from RO snapshot ' + 'to RW share: %s') LOG.error(msg, e) self._delete_share(share_name, protocol, fpg, vfs, fstore) raise @@ -862,8 +862,8 @@ class HPE3ParMediator(object): self._update_capacity_quotas( fstore, 0, share_size, fpg, vfs) except Exception as e: - msg = _LW('Exception during cleanup of deleted ' - 'share %(share)s in filestore %(fstore)s: %(e)s') + msg = ('Exception during cleanup of deleted ' + 'share %(share)s in filestore %(fstore)s: %(e)s') data = { 'fstore': fstore, 'share': share_name, @@ -878,10 +878,10 @@ class HPE3ParMediator(object): # return out and log a warning. if protocol == "smb" and (not self.hpe3par_cifs_admin_access_username or not self.hpe3par_cifs_admin_access_password): - LOG.warning(_LW("hpe3par_cifs_admin_access_username and " - "hpe3par_cifs_admin_access_password must be " - "provided in order for the file tree to be " - "properly deleted.")) + LOG.warning("hpe3par_cifs_admin_access_username and " + "hpe3par_cifs_admin_access_password must be " + "provided in order for the file tree to be " + "properly deleted.") return mount_location = "%s%s" % (self.hpe3par_share_mount_path, share_name) @@ -978,8 +978,8 @@ class HPE3ParMediator(object): try: utils.execute('mkdir', mount_location, run_as_root=True) except Exception as err: - message = (_LW("There was an error creating mount directory: " - "%s. The nested file tree will not be deleted."), + message = ("There was an error creating mount directory: " + "%s. The nested file tree will not be deleted.", six.text_type(err)) LOG.warning(message) @@ -1004,8 +1004,8 @@ class HPE3ParMediator(object): protocol, fpg, vfs, fstore, share_ip) self._mount_share(protocol, mount_location, mount_dir) except Exception as err: - message = (_LW("There was an error mounting the super share: " - "%s. The nested file tree will not be deleted."), + message = ("There was an error mounting the super share: " + "%s. The nested file tree will not be deleted.", six.text_type(err)) LOG.warning(message) @@ -1013,8 +1013,8 @@ class HPE3ParMediator(object): try: utils.execute('umount', mount_location, run_as_root=True) except Exception as err: - message = _LW("There was an error unmounting the share at " - "%(mount_location)s: %(error)s") + message = ("There was an error unmounting the share at " + "%(mount_location)s: %(error)s") msg_data = { 'mount_location': mount_location, 'error': six.text_type(err), @@ -1025,8 +1025,8 @@ class HPE3ParMediator(object): try: utils.execute('rm', '-rf', directory, run_as_root=True) except Exception as err: - message = (_LW("There was an error removing the share: " - "%s. The nested file tree will not be deleted."), + message = ("There was an error removing the share: " + "%s. The nested file tree will not be deleted.", six.text_type(err)) LOG.warning(message) @@ -1212,8 +1212,8 @@ class HPE3ParMediator(object): self._client.startfsnapclean(fpg, reclaimStrategy='maxspeed') except Exception: # Remove already happened so only log this. - LOG.exception(_LE('Unexpected exception calling startfsnapclean ' - 'for FPG %(fpg)s.'), {'fpg': fpg}) + LOG.exception('Unexpected exception calling startfsnapclean ' + 'for FPG %(fpg)s.', {'fpg': fpg}) @staticmethod def _validate_access_type(protocol, access_type): diff --git a/manila/share/drivers/huawei/huawei_utils.py b/manila/share/drivers/huawei/huawei_utils.py index 55d3861fd7..a1b15be5fe 100644 --- a/manila/share/drivers/huawei/huawei_utils.py +++ b/manila/share/drivers/huawei/huawei_utils.py @@ -17,7 +17,6 @@ import copy from oslo_log import log -from manila.i18n import _LE from manila.share.drivers.huawei import constants from manila.share import share_types @@ -66,8 +65,8 @@ def _get_opts_from_specs(specs): words = value.split() if not (words and len(words) == 2 and words[0] == ''): - LOG.error(_LE("Extra specs must be specified as " - "capabilities:%s=' True'."), key) + LOG.error("Extra specs must be specified as " + "capabilities:%s=' True'.", key) else: opts[key] = words[1].lower() diff --git a/manila/share/drivers/huawei/v3/connection.py b/manila/share/drivers/huawei/v3/connection.py index 3bcd7fa2f5..d91a88c804 100644 --- a/manila/share/drivers/huawei/v3/connection.py +++ b/manila/share/drivers/huawei/v3/connection.py @@ -31,7 +31,7 @@ import six from manila.common import constants as common_constants from manila.data import utils as data_utils from manila import exception -from manila.i18n import _, _LE, _LI, _LW +from manila.i18n import _ from manila import rpc from manila.share.drivers.huawei import base as driver from manila.share.drivers.huawei import constants @@ -275,7 +275,7 @@ class V3StorageConnection(driver.HuaweiBase): snapshot_name = "share_snapshot_" + snap_name snap_id = self.helper._create_snapshot(sharefsid, snapshot_name) - LOG.info(_LI('Creating snapshot id %s.'), snap_id) + LOG.info('Creating snapshot id %s.', snap_id) return snapshot_name.replace("-", "_") def delete_snapshot(self, snapshot, share_server=None): @@ -286,8 +286,8 @@ class V3StorageConnection(driver.HuaweiBase): sharefsid = self.helper.get_fsid_by_name(snapshot['share_name']) if sharefsid is None: - LOG.warning(_LW('Delete snapshot share id %s fs has been ' - 'deleted.'), snap_name) + LOG.warning('Delete snapshot share id %s fs has been ' + 'deleted.', snap_name) return snapshot_id = self.helper._get_snapshot_id(sharefsid, snap_name) @@ -297,7 +297,7 @@ class V3StorageConnection(driver.HuaweiBase): if snapshot_flag: self.helper._delete_snapshot(snapshot_id) else: - LOG.warning(_LW("Can not find snapshot %s on array."), snap_name) + LOG.warning("Can not find snapshot %s on array.", snap_name) def update_share_stats(self, stats_dict): """Retrieve status info from share group.""" @@ -358,13 +358,13 @@ class V3StorageConnection(driver.HuaweiBase): share = self.helper._get_share_by_name(share_name, share_url_type) if not share: - LOG.warning(_LW('The share was not found. Share name:%s'), + LOG.warning('The share was not found. Share name:%s', share_name) fsid = self.helper.get_fsid_by_name(share_name) if fsid: self.helper._delete_fs(fsid) return - LOG.warning(_LW('The filesystem was not found.')) + LOG.warning('The filesystem was not found.') return share_id = share['ID'] @@ -452,8 +452,8 @@ class V3StorageConnection(driver.HuaweiBase): try: os.rmdir(item['mount_src']) except Exception as err: - LOG.warning(_LW('Failed to remove temp file. File path: ' - '%(file_path)s. Reason: %(err)s.'), + LOG.warning('Failed to remove temp file. File path:' + '%(file_path)s. Reason: %(err)s.', {'file_path': item['mount_src'], 'err': err}) @@ -467,8 +467,8 @@ class V3StorageConnection(driver.HuaweiBase): self.allow_access(old_share, old_access) except exception.ManilaException as err: with excutils.save_and_reraise_exception(): - LOG.error(_LE('Failed to add access to share %(name)s. ' - 'Reason: %(err)s.'), + LOG.error('Failed to add access to share %(name)s. ' + 'Reason: %(err)s.', {'name': old_share['name'], 'err': err}) @@ -478,8 +478,8 @@ class V3StorageConnection(driver.HuaweiBase): self.mount_share_to_host(old_share, old_access) except exception.ShareMountException as err: with excutils.save_and_reraise_exception(): - LOG.error(_LE('Failed to mount old share %(name)s. ' - 'Reason: %(err)s.'), + LOG.error('Failed to mount old share %(name)s. ' + 'Reason: %(err)s.', {'name': old_share['name'], 'err': err}) @@ -489,8 +489,8 @@ class V3StorageConnection(driver.HuaweiBase): except Exception as err: with excutils.save_and_reraise_exception(): self.umount_share_from_host(old_share) - LOG.error(_LE('Failed to mount new share %(name)s. ' - 'Reason: %(err)s.'), + LOG.error('Failed to mount new share %(name)s. ' + 'Reason: %(err)s.', {'name': new_share['name'], 'err': err}) @@ -500,8 +500,8 @@ class V3StorageConnection(driver.HuaweiBase): try: self.umount_share_from_host(item) except exception.ShareUmountException as err: - LOG.warning(_LW('Failed to unmount share %(name)s. ' - 'Reason: %(err)s.'), + LOG.warning('Failed to unmount share %(name)s. ' + 'Reason: %(err)s.', {'name': item['name'], 'err': err}) @@ -573,7 +573,7 @@ class V3StorageConnection(driver.HuaweiBase): if copy.get_progress()['total_progress'] == 100: copy_finish = True except Exception as err: - LOG.error(_LE("Failed to copy data, reason: %s."), err) + LOG.error("Failed to copy data, reason: %s.", err) return copy_finish @@ -695,12 +695,12 @@ class V3StorageConnection(driver.HuaweiBase): share_url_type = self.helper._get_share_url_type(share_proto) access_type = access['access_type'] if share_proto == 'NFS' and access_type not in ('ip', 'user'): - LOG.warning(_LW('Only IP or USER access types are allowed for ' - 'NFS shares.')) + LOG.warning('Only IP or USER access types are allowed for ' + 'NFS shares.') return elif share_proto == 'CIFS' and access_type != 'user': - LOG.warning(_LW('Only USER access type is allowed for' - ' CIFS shares.')) + LOG.warning('Only USER access type is allowed for' + ' CIFS shares.') return access_to = access['access_to'] @@ -710,14 +710,14 @@ class V3StorageConnection(driver.HuaweiBase): access_to = '*' share = self.helper._get_share_by_name(share_name, share_url_type) if not share: - LOG.warning(_LW('Can not get share %s.'), share_name) + LOG.warning('Can not get share %s.', share_name) return access_id = self.helper._get_access_from_share(share['ID'], access_to, share_proto) if not access_id: - LOG.warning(_LW('Can not get access id from share. ' - 'share_name: %s'), share_name) + LOG.warning('Can not get access id from share. ' + 'share_name: %s', share_name) return self.helper._remove_access_from_share(access_id, share_proto) @@ -798,7 +798,7 @@ class V3StorageConnection(driver.HuaweiBase): share_url_type = self.helper._get_share_url_type(share_proto) share_stor = self.helper._get_share_by_name(share_name, share_url_type) if not share_stor: - LOG.warning(_LW('Cannot get share %s.'), share_name) + LOG.warning('Cannot get share %s.', share_name) return share_id = share_stor['ID'] all_accesses = self.helper._get_all_access_from_share(share_id, @@ -920,8 +920,8 @@ class V3StorageConnection(driver.HuaweiBase): opts['thin_provisioning'] = constants.THICK_PROVISIONING change_opts = self.check_retype_change_opts(opts, poolinfo, fs) - LOG.info(_LI('Retyping share (%(share)s), changed options are : ' - '(%(change_opts)s).'), + LOG.info('Retyping share (%(share)s), changed options are : ' + '(%(change_opts)s).', {'share': old_share_name, 'change_opts': change_opts}) try: self.retype_share(change_opts, fs_id) @@ -1198,9 +1198,9 @@ class V3StorageConnection(driver.HuaweiBase): if wait_interval: return int(wait_interval) else: - LOG.info(_LI( + LOG.info( "Wait interval is not configured in huawei " - "conf file. Use default: %(default_wait_interval)d."), + "conf file. Use default: %(default_wait_interval)d.", {"default_wait_interval": constants.DEFAULT_WAIT_INTERVAL}) return constants.DEFAULT_WAIT_INTERVAL @@ -1211,9 +1211,9 @@ class V3StorageConnection(driver.HuaweiBase): if timeout: return int(timeout) else: - LOG.info(_LI( + LOG.info( "Timeout is not configured in huawei conf file. " - "Use default: %(default_timeout)d."), + "Use default: %(default_timeout)d.", {"default_timeout": constants.DEFAULT_TIMEOUT}) return constants.DEFAULT_TIMEOUT @@ -1736,8 +1736,8 @@ class V3StorageConnection(driver.HuaweiBase): remote_fs_id=self.helper.get_fsid_by_name(new_share_name) ) except Exception: - LOG.exception(_LE('Failed to create a replication pair ' - 'with host %s.'), + LOG.exception('Failed to create a replication pair ' + 'with host %s.', active_replica['host']) raise @@ -1760,7 +1760,7 @@ class V3StorageConnection(driver.HuaweiBase): replica_pair_id = self.private_storage.get(replica['share_id'], 'replica_pair_id') if replica_pair_id is None: - msg = _LE("No replication pair ID recorded for share %s.") + msg = ("No replication pair ID recorded for share %s.") LOG.error(msg, replica['share_id']) return common_constants.STATUS_ERROR @@ -1780,7 +1780,7 @@ class V3StorageConnection(driver.HuaweiBase): try: self.replica_mgr.switch_over(replica_pair_id) except Exception: - LOG.exception(_LE('Failed to promote replica %s.'), + LOG.exception('Failed to promote replica %s.', replica['id']) raise @@ -1790,8 +1790,8 @@ class V3StorageConnection(driver.HuaweiBase): try: self.update_access(replica, access_rules, [], [], share_server) except Exception: - LOG.warning(_LW('Failed to set access rules to ' - 'new active replica %s.'), + LOG.warning('Failed to set access rules to ' + 'new active replica %s.', replica['id']) updated_new_active_access = False @@ -1800,8 +1800,8 @@ class V3StorageConnection(driver.HuaweiBase): try: self.clear_access(old_active_replica, share_server) except Exception: - LOG.warning(_LW("Failed to clear access rules from " - "old active replica %s."), + LOG.warning("Failed to clear access rules from " + "old active replica %s.", old_active_replica['id']) cleared_old_active_access = False @@ -1833,8 +1833,8 @@ class V3StorageConnection(driver.HuaweiBase): replica_pair_id = self.private_storage.get(replica['share_id'], 'replica_pair_id') if replica_pair_id is None: - msg = _LW("No replication pair ID recorded for share %(share)s. " - "Continue to delete replica %(replica)s.") + msg = ("No replication pair ID recorded for share %(share)s. " + "Continue to delete replica %(replica)s.") LOG.warning(msg, {'share': replica['share_id'], 'replica': replica['id']}) else: @@ -1844,6 +1844,6 @@ class V3StorageConnection(driver.HuaweiBase): try: self.delete_share(replica, share_server) except Exception: - LOG.exception(_LE('Failed to delete replica %s.'), + LOG.exception('Failed to delete replica %s.', replica['id']) raise diff --git a/manila/share/drivers/huawei/v3/helper.py b/manila/share/drivers/huawei/v3/helper.py index fd73ab7a22..ff8c78d1ce 100644 --- a/manila/share/drivers/huawei/v3/helper.py +++ b/manila/share/drivers/huawei/v3/helper.py @@ -25,7 +25,7 @@ from six.moves import http_cookiejar from six.moves.urllib import request as urlreq # pylint: disable=E0611 from manila import exception -from manila.i18n import _, _LE, _LW +from manila.i18n import _ from manila.share.drivers.huawei import constants from manila import utils @@ -77,8 +77,8 @@ class RestHelper(object): LOG.debug('Response Data: %(res)s.', {'res': res}) except Exception as err: - LOG.error(_LE('\nBad response from server: %(url)s.' - ' Error: %(err)s'), {'url': url, 'err': err}) + LOG.error('\nBad response from server: %(url)s.' + ' Error: %(err)s', {'url': url, 'err': err}) res = '{"error":{"code":%s,' \ '"description":"Connect server error"}}' \ % constants.ERROR_CONNECT_TO_SERVER @@ -110,7 +110,7 @@ class RestHelper(object): if((result['error']['code'] != 0) or ("data" not in result) or (result['data']['deviceid'] is None)): - LOG.error(_LE("Login to %s failed, try another."), item_url) + LOG.error("Login to %s failed, try another.", item_url) continue LOG.debug('Login success: %(url)s\n', @@ -139,7 +139,7 @@ class RestHelper(object): error_code = result['error']['code'] if(error_code == constants.ERROR_CONNECT_TO_SERVER or error_code == constants.ERROR_UNAUTHORIZED_TO_SERVER): - LOG.error(_LE("Can't open the recent url, re-login.")) + LOG.error("Can't open the recent url, re-login.") deviceid = self.login() if deviceid is not None: @@ -214,7 +214,7 @@ class RestHelper(object): utils.execute('chmod', '666', filepath, run_as_root=True) except Exception as err: - LOG.error(_LE('Bad response from change file: %s.') % err) + LOG.error('Bad response from change file: %s.' % err) raise def create_share(self, share_name, fs_id, share_proto): @@ -1151,8 +1151,8 @@ class RestHelper(object): url = "/vlan/" + vlan_id result = self.call(url, None, 'DELETE') if result['error']['code'] == constants.ERROR_LOGICAL_PORT_EXIST: - LOG.warning(_LW('Cannot delete vlan because there is ' - 'a logical port on vlan.')) + LOG.warning('Cannot delete vlan because there is ' + 'a logical port on vlan.') return self._assert_rest_result(result, _('Delete vlan error.')) @@ -1402,7 +1402,7 @@ class RestHelper(object): if (result['error']['code'] == constants.ERROR_REPLICATION_PAIR_NOT_EXIST): - LOG.warning(_LW('Replication pair %s was not found.'), + LOG.warning('Replication pair %s was not found.', pair_id) return diff --git a/manila/share/drivers/huawei/v3/replication.py b/manila/share/drivers/huawei/v3/replication.py index be1ece98dc..c1df702af3 100644 --- a/manila/share/drivers/huawei/v3/replication.py +++ b/manila/share/drivers/huawei/v3/replication.py @@ -18,7 +18,7 @@ from oslo_utils import strutils from manila.common import constants as common_constants from manila import exception -from manila.i18n import _, _LE, _LW +from manila.i18n import _ from manila.share.drivers.huawei import constants @@ -56,7 +56,7 @@ class ReplicaPairManager(object): pair_info = self.helper.create_replication_pair(pair_params) except Exception: - msg = _LE("Failed to create replication pair for share %s.") + msg = ("Failed to create replication pair for share %s.") LOG.exception(msg, local_share_name) raise @@ -69,8 +69,8 @@ class ReplicaPairManager(object): pair_info = self.helper.get_replication_pair_by_id( replica_pair_id) except Exception: - LOG.exception(_LE('Failed to get replication pair info for ' - '%s.'), replica_pair_id) + LOG.exception('Failed to get replication pair info for ' + '%s.', replica_pair_id) raise return pair_info @@ -114,7 +114,7 @@ class ReplicaPairManager(object): pair_info = self._get_replication_pair_info(replica_pair_id) except Exception: # if cannot communicate to backend, return error - LOG.error(_LE('Cannot get replica state, return %s'), + LOG.error('Cannot get replica state, return %s', common_constants.STATUS_ERROR) return common_constants.STATUS_ERROR @@ -124,8 +124,8 @@ class ReplicaPairManager(object): try: self.helper.sync_replication_pair(pair_id) except Exception as err: - LOG.warning(_LW('Failed to sync replication pair %(id)s. ' - 'Reason: %(err)s'), + LOG.warning('Failed to sync replication pair %(id)s. ' + 'Reason: %(err)s', {'id': pair_id, 'err': err}) def update_replication_pair_state(self, replica_pair_id): @@ -133,8 +133,8 @@ class ReplicaPairManager(object): health = self._check_replication_health(pair_info) if health is not None: - LOG.warning(_LW("Cannot update the replication %s " - "because it's not in normal status."), + LOG.warning("Cannot update the replication %s " + "because it's not in normal status.", replica_pair_id) return @@ -145,9 +145,9 @@ class ReplicaPairManager(object): try: self.helper.switch_replication_pair(replica_pair_id) except Exception: - msg = _LE('Replication pair %s primary/secondary ' - 'relationship is not right, try to switch over ' - 'again but still failed.') + msg = ('Replication pair %s primary/secondary ' + 'relationship is not right, try to switch over ' + 'again but still failed.') LOG.exception(msg, replica_pair_id) return @@ -158,8 +158,8 @@ class ReplicaPairManager(object): try: self.helper.set_pair_secondary_write_lock(replica_pair_id) except Exception: - msg = _LE('Replication pair %s secondary access is R/W, ' - 'try to set write lock but still failed.') + msg = ('Replication pair %s secondary access is R/W, ' + 'try to set write lock but still failed.') LOG.exception(msg, replica_pair_id) return @@ -173,8 +173,8 @@ class ReplicaPairManager(object): pair_info = self._get_replication_pair_info(replica_pair_id) if strutils.bool_from_string(pair_info['ISPRIMARY']): - LOG.warning(_LW('The replica to promote is already primary, ' - 'no need to switch over.')) + LOG.warning('The replica to promote is already primary, ' + 'no need to switch over.') return replica_state = self._check_replica_state(pair_info) @@ -192,14 +192,14 @@ class ReplicaPairManager(object): # means replication pair is in an abnormal status, # ignore this exception, continue to cancel secondary write lock, # let secondary share accessible for disaster recovery. - LOG.exception(_LE('Failed to split replication pair %s while ' - 'switching over.'), replica_pair_id) + LOG.exception('Failed to split replication pair %s while ' + 'switching over.', replica_pair_id) try: self.helper.cancel_pair_secondary_write_lock(replica_pair_id) except Exception: - LOG.exception(_LE('Failed to cancel replication pair %s ' - 'secondary write lock.'), replica_pair_id) + LOG.exception('Failed to cancel replication pair %s ' + 'secondary write lock.', replica_pair_id) raise try: @@ -207,8 +207,8 @@ class ReplicaPairManager(object): self.helper.set_pair_secondary_write_lock(replica_pair_id) self.helper.sync_replication_pair(replica_pair_id) except Exception: - LOG.exception(_LE('Failed to completely switch over ' - 'replication pair %s.'), replica_pair_id) + LOG.exception('Failed to completely switch over ' + 'replication pair %s.', replica_pair_id) # for all the rest steps, # because secondary share is accessible now, @@ -222,15 +222,15 @@ class ReplicaPairManager(object): except Exception: # Ignore this exception because replication pair may at some # abnormal status that supports deleting. - LOG.warning(_LW('Failed to split replication pair %s ' - 'before deleting it. Ignore this exception, ' - 'and try to delete anyway.'), + LOG.warning('Failed to split replication pair %s ' + 'before deleting it. Ignore this exception, ' + 'and try to delete anyway.', replica_pair_id) try: self.helper.delete_replication_pair(replica_pair_id) except Exception: - LOG.exception(_LE('Failed to delete replication pair %s.'), + LOG.exception('Failed to delete replication pair %s.', replica_pair_id) raise diff --git a/manila/share/drivers/ibm/gpfs.py b/manila/share/drivers/ibm/gpfs.py index 4d599a1471..6b885920dc 100644 --- a/manila/share/drivers/ibm/gpfs.py +++ b/manila/share/drivers/ibm/gpfs.py @@ -44,7 +44,7 @@ import six from manila.common import constants from manila import exception -from manila.i18n import _, _LI +from manila.i18n import _ from manila.share import driver from manila.share.drivers.helpers import NFSHelper from manila.share import share_types @@ -684,8 +684,8 @@ class GPFSShareDriver(driver.ExecuteMixin, driver.GaneshaMixin, msg = _('Failed to set quota for share %s.') % new_share_name LOG.exception(msg) raise exception.GPFSException(msg) - LOG.info(_LI('Existing share %(shr)s has size %(size)s KB ' - 'which is below 1GiB, so extended it to 1GiB.') % + LOG.info('Existing share %(shr)s has size %(size)s KB ' + 'which is below 1GiB, so extended it to 1GiB.' % {'shr': new_share_name, 'size': share_size}) share_size = 1 else: diff --git a/manila/share/drivers/lvm.py b/manila/share/drivers/lvm.py index 853f2a216f..5bc1e41772 100644 --- a/manila/share/drivers/lvm.py +++ b/manila/share/drivers/lvm.py @@ -28,7 +28,7 @@ from oslo_utils import importutils import six from manila import exception -from manila.i18n import _, _LE, _LI, _LW +from manila.i18n import _ from manila.share import driver from manila.share.drivers import generic from manila.share import utils @@ -105,9 +105,9 @@ class LVMMixin(driver.ExecuteMixin): share_name), run_as_root=True) except exception.ProcessExecutionError as exc: if "not found" not in exc.stderr: - LOG.exception(_LE("Error deleting volume")) + LOG.exception("Error deleting volume") raise - LOG.warning(_LW("Volume not found: %s") % exc.stderr) + LOG.warning("Volume not found: %s" % exc.stderr) def _create_snapshot(self, context, snapshot): """Creates a snapshot.""" @@ -253,12 +253,12 @@ class LVMShareDriver(LVMMixin, driver.ShareDriver): if 'device is busy' in six.text_type(exc): raise exception.ShareBusyException(reason=share['name']) else: - LOG.info(_LI('Unable to umount: %s'), exc) + LOG.info('Unable to umount: %s', exc) # remove dir try: os.rmdir(mount_path) except OSError: - LOG.warning(_LW('Unable to delete %s'), mount_path) + LOG.warning('Unable to delete %s', mount_path) def ensure_share(self, ctx, share, share_server=None): """Ensure that storage are mounted and exported.""" @@ -273,7 +273,7 @@ class LVMShareDriver(LVMMixin, driver.ShareDriver): self._get_helper(share).remove_exports( self.share_server, share['name']) except exception.ProcessExecutionError: - LOG.warning(_LW("Can't remove share %r"), share['id']) + LOG.warning("Can't remove share %r", share['id']) except exception.InvalidShare as exc: LOG.warning(exc.message) @@ -326,7 +326,7 @@ class LVMShareDriver(LVMMixin, driver.ShareDriver): except exception.ProcessExecutionError: out, err = self._execute('mount', '-l', run_as_root=True) if device_name in out: - LOG.warning(_LW("%s is already mounted"), device_name) + LOG.warning("%s is already mounted", device_name) else: raise return mount_path diff --git a/manila/share/drivers/maprfs/driver_util.py b/manila/share/drivers/maprfs/driver_util.py index 572b86f5df..b26f2a5b67 100644 --- a/manila/share/drivers/maprfs/driver_util.py +++ b/manila/share/drivers/maprfs/driver_util.py @@ -27,7 +27,6 @@ import six from manila.common import constants from manila import exception from manila.i18n import _ -from manila.i18n import _LE from manila import utils LOG = log.getLogger(__name__) @@ -71,13 +70,13 @@ class BaseDriverUtil(object): if self._check_error(e): raise elif x < len(self.hosts) - 1: - msg = _LE('Error running SSH command. Trying another host') + msg = ('Error running SSH command. Trying another host') LOG.error(msg) else: raise except Exception as e: if x < len(self.hosts) - 1: - msg = _LE('Error running SSH command. Trying another host') + msg = ('Error running SSH command. Trying another host') LOG.error(msg) else: raise exception.ProcessExecutionError(six.text_type(e)) diff --git a/manila/share/drivers/maprfs/maprfs_native.py b/manila/share/drivers/maprfs/maprfs_native.py index 74eb822720..2b623a4a97 100644 --- a/manila/share/drivers/maprfs/maprfs_native.py +++ b/manila/share/drivers/maprfs/maprfs_native.py @@ -26,7 +26,7 @@ from oslo_utils import units from manila import context from manila import exception -from manila.i18n import _, _LW, _LI +from manila.i18n import _ from manila.share import api from manila.share import driver @@ -258,8 +258,8 @@ class MapRFSNativeShareDriver(driver.ExecuteMixin, driver.ShareDriver): """Deletes share storage.""" volume_name = self._get_volume_name(context, share) if volume_name == "error": - LOG.info(_LI("Skipping deleting share with name %s, as it does not" - " exist on the backend"), share['name']) + LOG.info("Skipping deleting share with name %s, as it does not" + " exist on the backend", share['name']) return try: self._maprfs_util.delete_volume(volume_name) @@ -295,7 +295,7 @@ class MapRFSNativeShareDriver(driver.ExecuteMixin, driver.ShareDriver): # method shouldn`t raise exception if share does # not exist actually if not self._maprfs_util.volume_exists(volume_name): - LOG.warning(_LW('Can not get share %s.'), share['name']) + LOG.warning('Can not get share %s.', share['name']) return # check update if add_rules or delete_rules: @@ -337,10 +337,10 @@ class MapRFSNativeShareDriver(driver.ExecuteMixin, driver.ShareDriver): raise exception.MapRFSException(msg=msg) if not self.configuration.maprfs_cldb_ip: - LOG.warning(_LW('CLDB nodes are not specified!')) + LOG.warning('CLDB nodes are not specified!') if not self.configuration.maprfs_zookeeper_ip: - LOG.warning(_LW('Zookeeper nodes are not specified!')) + LOG.warning('Zookeeper nodes are not specified!') if not self._check_maprfs_state(): msg = _('MapR-FS is not in healthy state.') @@ -383,7 +383,7 @@ class MapRFSNativeShareDriver(driver.ExecuteMixin, driver.ShareDriver): location = self._get_share_export_locations(share, path=share_path) if size == 0: size = used - msg = _LW( + msg = ( 'Share %s has no size quota. Total used value will be' ' used as share size') LOG.warning(msg, share['name']) diff --git a/manila/share/drivers/netapp/common.py b/manila/share/drivers/netapp/common.py index 1a1a554f0d..3c17598144 100644 --- a/manila/share/drivers/netapp/common.py +++ b/manila/share/drivers/netapp/common.py @@ -21,7 +21,7 @@ from oslo_log import log from oslo_utils import importutils from manila import exception -from manila.i18n import _, _LI +from manila.i18n import _ from manila.share import driver from manila.share.drivers.netapp import options from manila.share.drivers.netapp import utils as na_utils @@ -69,7 +69,7 @@ class NetAppDriver(object): na_utils.check_flags(NetAppDriver.REQUIRED_FLAGS, config) app_version = na_utils.OpenStackInfo().info() - LOG.info(_LI('OpenStack OS Version Info: %s'), app_version) + LOG.info('OpenStack OS Version Info: %s', app_version) kwargs['app_version'] = app_version driver_mode = NetAppDriver._get_driver_mode( @@ -107,8 +107,8 @@ class NetAppDriver(object): storage_family = storage_family.lower() fmt = {'storage_family': storage_family, 'driver_mode': driver_mode} - LOG.info(_LI('Requested unified config: %(storage_family)s and ' - '%(driver_mode)s.') % fmt) + LOG.info('Requested unified config: %(storage_family)s and ' + '%(driver_mode)s.' % fmt) family_meta = NETAPP_UNIFIED_DRIVER_REGISTRY.get(storage_family) if family_meta is None: @@ -124,6 +124,6 @@ class NetAppDriver(object): kwargs['netapp_mode'] = 'proxy' driver = importutils.import_object(driver_loc, *args, **kwargs) - LOG.info(_LI('NetApp driver of family %(storage_family)s and mode ' - '%(driver_mode)s loaded.') % fmt) + LOG.info('NetApp driver of family %(storage_family)s and mode ' + '%(driver_mode)s loaded.' % fmt) return driver diff --git a/manila/share/drivers/netapp/dataontap/client/client_base.py b/manila/share/drivers/netapp/dataontap/client/client_base.py index 77cab23d5b..a97ad32dd4 100644 --- a/manila/share/drivers/netapp/dataontap/client/client_base.py +++ b/manila/share/drivers/netapp/dataontap/client/client_base.py @@ -16,7 +16,6 @@ from oslo_log import log from oslo_utils import excutils -from manila.i18n import _LE from manila.share.drivers.netapp.dataontap.client import api as netapp_api from manila.share.drivers.netapp import utils as na_utils @@ -89,7 +88,7 @@ class NetAppBaseClient(object): result = self.send_request('license-v2-list-info') except netapp_api.NaApiError: with excutils.save_and_reraise_exception(): - LOG.exception(_LE("Could not get licenses list.")) + LOG.exception("Could not get licenses list.") return sorted( [l.get_child_content('package').lower() diff --git a/manila/share/drivers/netapp/dataontap/client/client_cmode.py b/manila/share/drivers/netapp/dataontap/client/client_cmode.py index 8ca1e2dc0d..1cb447e0f5 100644 --- a/manila/share/drivers/netapp/dataontap/client/client_cmode.py +++ b/manila/share/drivers/netapp/dataontap/client/client_cmode.py @@ -26,7 +26,7 @@ from oslo_utils import units import six from manila import exception -from manila.i18n import _, _LE, _LW +from manila.i18n import _ from manila.share.drivers.netapp.dataontap.client import api as netapp_api from manila.share.drivers.netapp.dataontap.client import client_base from manila.share.drivers.netapp import utils as na_utils @@ -327,7 +327,7 @@ class NetAppCmodeClient(client_base.NetAppBaseClient): Offlines and destroys root volumes. Deletes Vserver. """ if not self.vserver_exists(vserver_name): - LOG.error(_LE("Vserver %s does not exist."), vserver_name) + LOG.error("Vserver %s does not exist.", vserver_name) return root_volume_name = self.get_vserver_root_volume_name(vserver_name) @@ -338,7 +338,7 @@ class NetAppCmodeClient(client_base.NetAppBaseClient): vserver_client.offline_volume(root_volume_name) except netapp_api.NaApiError as e: if e.code == netapp_api.EVOLUMEOFFLINE: - LOG.error(_LE("Volume %s is already offline."), + LOG.error("Volume %s is already offline.", root_volume_name) else: raise @@ -367,8 +367,8 @@ class NetAppCmodeClient(client_base.NetAppBaseClient): vserver_client.send_request('cifs-server-delete', api_args) except netapp_api.NaApiError as e: if e.code == netapp_api.EOBJECTNOTFOUND: - LOG.error(_LE('CIFS server does not exist for ' - 'Vserver %s.'), vserver_name) + LOG.error('CIFS server does not exist for ' + 'Vserver %s.', vserver_name) else: vserver_client.send_request('cifs-server-delete') @@ -1069,7 +1069,7 @@ class NetAppCmodeClient(client_base.NetAppBaseClient): vserver_aggr_info_list = vserver_aggr_info_element.get_children() if not vserver_aggr_info_list: - LOG.warning(_LW('No aggregates assigned to Vserver %s.'), + LOG.warning('No aggregates assigned to Vserver %s.', vserver_name) # Return dict of key-value pair of aggr_name:aggr_size_available. @@ -1387,7 +1387,7 @@ class NetAppCmodeClient(client_base.NetAppBaseClient): self.send_request('net-dns-create', api_args) except netapp_api.NaApiError as e: if e.code == netapp_api.EDUPLICATEENTRY: - LOG.error(_LE("DNS exists for Vserver.")) + LOG.error("DNS exists for Vserver.") else: msg = _("Failed to configure DNS. %s") raise exception.NetAppException(msg % e.message) @@ -2027,8 +2027,8 @@ class NetAppCmodeClient(client_base.NetAppBaseClient): return except netapp_api.NaApiError as e: if e.code == netapp_api.EAPIERROR and 'job ID' in e.message: - msg = _LW('Could not unmount volume %(volume)s due to ' - 'ongoing volume operation: %(exception)s') + msg = ('Could not unmount volume %(volume)s due to ' + 'ongoing volume operation: %(exception)s') msg_args = {'volume': volume_name, 'exception': e} LOG.warning(msg, msg_args) time.sleep(retry_interval) @@ -2642,7 +2642,7 @@ class NetAppCmodeClient(client_base.NetAppBaseClient): node_client.send_request('ems-autosupport-log', message_dict) LOG.debug('EMS executed successfully.') except netapp_api.NaApiError as e: - LOG.warning(_LW('Failed to invoke EMS. %s') % e) + LOG.warning('Failed to invoke EMS. %s' % e) @na_utils.trace def get_aggregate(self, aggregate_name): @@ -3276,8 +3276,8 @@ class NetAppCmodeClient(client_base.NetAppBaseClient): has_snapmirrors = len(snapmirrors) > 0 except netapp_api.NaApiError: - msg = _LE("Could not determine if volume %s is part of " - "existing snapmirror relationships.") + msg = ("Could not determine if volume %s is part of " + "existing snapmirror relationships.") LOG.exception(msg, volume['name']) has_snapmirrors = False diff --git a/manila/share/drivers/netapp/dataontap/cluster_mode/data_motion.py b/manila/share/drivers/netapp/dataontap/cluster_mode/data_motion.py index 23f2b52373..b9883067f9 100644 --- a/manila/share/drivers/netapp/dataontap/cluster_mode/data_motion.py +++ b/manila/share/drivers/netapp/dataontap/cluster_mode/data_motion.py @@ -25,7 +25,7 @@ from oslo_log import log from oslo_utils import excutils from manila import exception -from manila.i18n import _, _LE, _LI +from manila.i18n import _ from manila.share import configuration from manila.share import driver from manila.share.drivers.netapp.dataontap.client import api as netapp_api @@ -196,7 +196,7 @@ class DataMotionSession(object): if (e.code == netapp_api.EOBJECTNOTFOUND or e.code == netapp_api.ESOURCE_IS_DIFFERENT or "(entry doesn't exist)" in e.message): - LOG.info(_LI('No snapmirror relationship to delete')) + LOG.info('No snapmirror relationship to delete') exc_context.reraise = False if release: @@ -267,7 +267,7 @@ class DataMotionSession(object): )[0] if snapmirror.get('relationship-status') != 'quiesced': raise exception.ReplicationException( - reason=_LE("Snapmirror relationship is not quiesced.")) + reason=("Snapmirror relationship is not quiesced.")) try: wait_for_quiesced() diff --git a/manila/share/drivers/netapp/dataontap/cluster_mode/lib_base.py b/manila/share/drivers/netapp/dataontap/cluster_mode/lib_base.py index e4f9ea2232..3ddf85e213 100644 --- a/manila/share/drivers/netapp/dataontap/cluster_mode/lib_base.py +++ b/manila/share/drivers/netapp/dataontap/cluster_mode/lib_base.py @@ -33,7 +33,7 @@ import six from manila.common import constants from manila import exception -from manila.i18n import _, _LE, _LI, _LW +from manila.i18n import _ from manila.share.drivers.netapp.dataontap.client import api as netapp_api from manila.share.drivers.netapp.dataontap.client import client_cmode from manila.share.drivers.netapp.dataontap.cluster_mode import data_motion @@ -158,11 +158,11 @@ class NetAppCmodeFileStorageLibrary(object): 'backend': self._backend_name, 'licenses': ', '.join(self._licenses), } - LOG.info(_LI('Available licenses on %(backend)s ' - 'are %(licenses)s.'), log_data) + LOG.info('Available licenses on %(backend)s ' + 'are %(licenses)s.', log_data) if 'nfs' not in self._licenses and 'cifs' not in self._licenses: - msg = _LE('Neither NFS nor CIFS is licensed on %(backend)s') + msg = 'Neither NFS nor CIFS is licensed on %(backend)s' msg_args = {'backend': self._backend_name} LOG.error(msg % msg_args) @@ -657,9 +657,9 @@ class NetAppCmodeFileStorageLibrary(object): except (exception.InvalidInput, exception.VserverNotSpecified, exception.VserverNotFound) as error: - LOG.warning(_LW("Could not determine share server for share being " - "deleted: %(share)s. Deletion of share record " - "will proceed anyway. Error: %(error)s"), + LOG.warning("Could not determine share server for share being " + "deleted: %(share)s. Deletion of share record " + "will proceed anyway. Error: %(error)s", {'share': share['id'], 'error': error}) return @@ -668,7 +668,7 @@ class NetAppCmodeFileStorageLibrary(object): self._remove_export(share, vserver_client) self._deallocate_container(share_name, vserver_client) else: - LOG.info(_LI("Share %s does not exist."), share['id']) + LOG.info("Share %s does not exist.", share['id']) @na_utils.trace def _deallocate_container(self, share_name, vserver_client): @@ -812,9 +812,9 @@ class NetAppCmodeFileStorageLibrary(object): except (exception.InvalidInput, exception.VserverNotSpecified, exception.VserverNotFound) as error: - LOG.warning(_LW("Could not determine share server for snapshot " - "being deleted: %(snap)s. Deletion of snapshot " - "record will proceed anyway. Error: %(error)s"), + LOG.warning("Could not determine share server for snapshot " + "being deleted: %(snap)s. Deletion of snapshot " + "record will proceed anyway. Error: %(error)s", {'snap': snapshot['id'], 'error': error}) return @@ -825,7 +825,7 @@ class NetAppCmodeFileStorageLibrary(object): try: self._delete_snapshot(vserver_client, share_name, snapshot_name) except exception.SnapshotResourceNotFound: - msg = _LI("Snapshot %(snap)s does not exist on share %(share)s.") + msg = ("Snapshot %(snap)s does not exist on share %(share)s.") msg_args = {'snap': snapshot_name, 'share': share_name} LOG.info(msg, msg_args) @@ -1099,9 +1099,9 @@ class NetAppCmodeFileStorageLibrary(object): except (exception.InvalidInput, exception.VserverNotSpecified, exception.VserverNotFound) as error: - LOG.warning(_LW("Could not determine share server for consistency " - "group being deleted: %(cg)s. Deletion of CG " - "record will proceed anyway. Error: %(error)s"), + LOG.warning("Could not determine share server for consistency " + "group being deleted: %(cg)s. Deletion of CG " + "record will proceed anyway. Error: %(error)s", {'cg': cg_dict['id'], 'error': error}) @na_utils.trace @@ -1128,9 +1128,9 @@ class NetAppCmodeFileStorageLibrary(object): except (exception.InvalidInput, exception.VserverNotSpecified, exception.VserverNotFound) as error: - LOG.warning(_LW("Could not determine share server for CG snapshot " - "being deleted: %(snap)s. Deletion of CG snapshot " - "record will proceed anyway. Error: %(error)s"), + LOG.warning("Could not determine share server for CG snapshot " + "being deleted: %(snap)s. Deletion of CG snapshot " + "record will proceed anyway. Error: %(error)s", {'snap': snap_dict['id'], 'error': error}) return None, None @@ -1143,8 +1143,8 @@ class NetAppCmodeFileStorageLibrary(object): self._delete_snapshot( vserver_client, share_name, snapshot_name) except exception.SnapshotResourceNotFound: - msg = _LI("Snapshot %(snap)s does not exist on share " - "%(share)s.") + msg = ("Snapshot %(snap)s does not exist on share " + "%(share)s.") msg_args = {'snap': snapshot_name, 'share': share_name} LOG.info(msg, msg_args) continue @@ -1185,9 +1185,9 @@ class NetAppCmodeFileStorageLibrary(object): except (exception.InvalidInput, exception.VserverNotSpecified, exception.VserverNotFound) as error: - LOG.warning(_LW("Could not determine share server for share " - "%(share)s during access rules update. " - "Error: %(error)s"), + LOG.warning("Could not determine share server for share " + "%(share)s during access rules update. " + "Error: %(error)s", {'share': share['id'], 'error': error}) return @@ -1216,8 +1216,8 @@ class NetAppCmodeFileStorageLibrary(object): The self._ssc_stats attribute is updated with the following format. { : {: }} """ - LOG.info(_LI("Updating storage service catalog information for " - "backend '%s'"), self._backend_name) + LOG.info("Updating storage service catalog information for " + "backend '%s'", self._backend_name) # Work on a copy and update the ssc data atomically before returning. ssc_stats = copy.deepcopy(self._ssc_stats) @@ -1349,7 +1349,7 @@ class NetAppCmodeFileStorageLibrary(object): try: snapmirrors = dm_session.get_snapmirrors(active_replica, replica) except netapp_api.NaApiError: - LOG.exception(_LE("Could not get snapmirrors for replica %s."), + LOG.exception("Could not get snapmirrors for replica %s.", replica['id']) return constants.STATUS_ERROR @@ -1358,8 +1358,8 @@ class NetAppCmodeFileStorageLibrary(object): try: dm_session.create_snapmirror(active_replica, replica) except netapp_api.NaApiError: - LOG.exception(_LE("Could not create snapmirror for " - "replica %s."), replica['id']) + LOG.exception("Could not create snapmirror for " + "replica %s.", replica['id']) return constants.STATUS_ERROR return constants.REPLICA_STATE_OUT_OF_SYNC @@ -1381,7 +1381,7 @@ class NetAppCmodeFileStorageLibrary(object): share_name) return constants.REPLICA_STATE_OUT_OF_SYNC except netapp_api.NaApiError: - LOG.exception(_LE("Could not resync snapmirror.")) + LOG.exception("Could not resync snapmirror.") return constants.STATUS_ERROR last_update_timestamp = float( @@ -1433,8 +1433,8 @@ class NetAppCmodeFileStorageLibrary(object): context, dm_session, orig_active_replica, replica, access_rules, share_server=share_server)) except exception.StorageCommunicationException: - LOG.exception(_LE("Could not communicate with the backend " - "for replica %s during promotion."), + LOG.exception("Could not communicate with the backend " + "for replica %s during promotion.", replica['id']) new_active_replica = copy.deepcopy(replica) new_active_replica['replica_state'] = ( @@ -1524,16 +1524,16 @@ class NetAppCmodeFileStorageLibrary(object): replica['status'] = constants.STATUS_ERROR replica['replica_state'] = constants.STATUS_ERROR replica['export_locations'] = [] - msg = _LE("Failed to change replica (%s) to a SnapMirror " - "destination. Replica backend is unreachable.") + msg = ("Failed to change replica (%s) to a SnapMirror " + "destination. Replica backend is unreachable.") LOG.exception(msg, replica['id']) return replica except netapp_api.NaApiError: replica['replica_state'] = constants.STATUS_ERROR replica['export_locations'] = [] - msg = _LE("Failed to change replica (%s) to a SnapMirror " - "destination.") + msg = ("Failed to change replica (%s) to a SnapMirror " + "destination.") LOG.exception(msg, replica['id']) return replica @@ -1735,8 +1735,8 @@ class NetAppCmodeFileStorageLibrary(object): share_volume, source_vserver, destination_aggregate) except Exception: - msg = _LE("Cannot migrate share %(shr)s efficiently between " - "%(src)s and %(dest)s.") + msg = ("Cannot migrate share %(shr)s efficiently between " + "%(src)s and %(dest)s.") msg_args = { 'shr': source_share['id'], 'src': source_share['host'], @@ -1746,9 +1746,9 @@ class NetAppCmodeFileStorageLibrary(object): else: compatible = True else: - msg = _LW("Cluster credentials have not been configured " - "with this share driver. Cannot perform volume move " - "operations.") + msg = ("Cluster credentials have not been configured " + "with this share driver. Cannot perform volume move " + "operations.") LOG.warning(msg) compatibility = { @@ -1774,8 +1774,8 @@ class NetAppCmodeFileStorageLibrary(object): self._client.start_volume_move( share_volume, vserver, destination_aggregate) - msg = _LI("Began volume move operation of share %(shr)s from %(src)s " - "to %(dest)s.") + msg = ("Began volume move operation of share %(shr)s from %(src)s " + "to %(dest)s.") msg_args = { 'shr': source_share['id'], 'src': source_share['host'], @@ -1826,8 +1826,8 @@ class NetAppCmodeFileStorageLibrary(object): 'cutover_soft_deferred'): status['percent-complete'] = 100 - msg = _LI("Volume move status for share %(share)s: (State) %(state)s. " - "(Phase) %(phase)s. Details: %(details)s") + msg = ("Volume move status for share %(share)s: (State) %(state)s. " + "(Phase) %(phase)s. Details: %(details)s") msg_args = { 'state': status['state'], 'details': status['details'], @@ -1854,13 +1854,13 @@ class NetAppCmodeFileStorageLibrary(object): try: self._get_volume_move_status(source_share, share_server) except exception.NetAppException: - LOG.exception(_LE("Could not get volume move status.")) + LOG.exception("Could not get volume move status.") return self._client.abort_volume_move(share_volume, vserver) - msg = _LI("Share volume move operation for share %(shr)s from host " - "%(src)s to %(dest)s was successfully aborted.") + msg = ("Share volume move operation for share %(shr)s from host " + "%(src)s to %(dest)s was successfully aborted.") msg_args = { 'shr': source_share['id'], 'src': source_share['host'], @@ -1903,9 +1903,9 @@ class NetAppCmodeFileStorageLibrary(object): destination_share['id']) vserver_client.set_volume_name(share_volume, new_share_volume_name) - msg = _LI("Volume move operation for share %(shr)s has completed " - "successfully. Share has been moved from %(src)s to " - "%(dest)s.") + msg = ("Volume move operation for share %(shr)s has completed " + "successfully. Share has been moved from %(src)s to " + "%(dest)s.") msg_args = { 'shr': source_share['id'], 'src': source_share['host'], diff --git a/manila/share/drivers/netapp/dataontap/cluster_mode/lib_multi_svm.py b/manila/share/drivers/netapp/dataontap/cluster_mode/lib_multi_svm.py index 3747e2ed3e..7dafd814af 100644 --- a/manila/share/drivers/netapp/dataontap/cluster_mode/lib_multi_svm.py +++ b/manila/share/drivers/netapp/dataontap/cluster_mode/lib_multi_svm.py @@ -26,7 +26,7 @@ from oslo_log import log from oslo_utils import excutils from manila import exception -from manila.i18n import _, _LE, _LW, _LI +from manila.i18n import _ from manila.share.drivers.netapp.dataontap.client import client_cmode from manila.share.drivers.netapp.dataontap.cluster_mode import lib_base from manila.share.drivers.netapp import utils as na_utils @@ -47,8 +47,8 @@ class NetAppCmodeMultiSVMFileStorageLibrary( if self._have_cluster_creds: if self.configuration.netapp_vserver: - msg = _LW('Vserver is specified in the configuration. This is ' - 'ignored when the driver is managing share servers.') + msg = ('Vserver is specified in the configuration. This is ' + 'ignored when the driver is managing share servers.') LOG.warning(msg) else: # only have vserver creds, which is an error in multi_svm mode @@ -191,7 +191,7 @@ class NetAppCmodeMultiSVMFileStorageLibrary( vserver_name) except Exception: with excutils.save_and_reraise_exception(): - LOG.error(_LE("Failed to configure Vserver.")) + LOG.error("Failed to configure Vserver.") self._delete_vserver(vserver_name, security_services=security_services) @@ -243,7 +243,7 @@ class NetAppCmodeMultiSVMFileStorageLibrary( network_allocations = network_info.get('admin_network_allocations') if not network_allocations: - LOG.info(_LI('No admin network defined for Vserver %s.') % + LOG.info('No admin network defined for Vserver %s.' % vserver_name) return @@ -310,15 +310,15 @@ class NetAppCmodeMultiSVMFileStorageLibrary( 'vserver_name') if server_details else None if not vserver: - LOG.warning(_LW("Vserver not specified for share server being " - "deleted. Deletion of share server record will " - "proceed anyway.")) + LOG.warning("Vserver not specified for share server being " + "deleted. Deletion of share server record will " + "proceed anyway.") return elif not self._client.vserver_exists(vserver): - LOG.warning(_LW("Could not find Vserver for share server being " - "deleted: %s. Deletion of share server " - "record will proceed anyway."), vserver) + LOG.warning("Could not find Vserver for share server being " + "deleted: %s. Deletion of share server " + "record will proceed anyway.", vserver) return self._delete_vserver(vserver, security_services=security_services) @@ -362,4 +362,4 @@ class NetAppCmodeMultiSVMFileStorageLibrary( node = interface['home-node'] self._client.delete_vlan(node, port, vlan) except exception.NetAppException: - LOG.exception(_LE("Deleting Vserver VLAN failed.")) + LOG.exception("Deleting Vserver VLAN failed.") diff --git a/manila/share/drivers/netapp/dataontap/cluster_mode/lib_single_svm.py b/manila/share/drivers/netapp/dataontap/cluster_mode/lib_single_svm.py index 24c8dd9e77..af6b896c86 100644 --- a/manila/share/drivers/netapp/dataontap/cluster_mode/lib_single_svm.py +++ b/manila/share/drivers/netapp/dataontap/cluster_mode/lib_single_svm.py @@ -25,7 +25,7 @@ import re from oslo_log import log from manila import exception -from manila.i18n import _, _LI +from manila.i18n import _ from manila.share.drivers.netapp.dataontap.cluster_mode import lib_base from manila.share.drivers.netapp import utils as na_utils @@ -72,8 +72,8 @@ class NetAppCmodeSingleSVMFileStorageLibrary( 'correctly.') % self._vserver raise exception.NetAppException(msg) - msg = _LI('Using Vserver %(vserver)s for backend %(backend)s with ' - '%(creds)s credentials.') + msg = ('Using Vserver %(vserver)s for backend %(backend)s with ' + '%(creds)s credentials.') msg_args = {'vserver': self._vserver, 'backend': self._backend_name} msg_args['creds'] = ('cluster' if self._have_cluster_creds else 'Vserver') diff --git a/manila/share/drivers/netapp/dataontap/cluster_mode/performance.py b/manila/share/drivers/netapp/dataontap/cluster_mode/performance.py index cd97988274..19f9d3ffdd 100644 --- a/manila/share/drivers/netapp/dataontap/cluster_mode/performance.py +++ b/manila/share/drivers/netapp/dataontap/cluster_mode/performance.py @@ -21,7 +21,7 @@ import copy from oslo_log import log as logging from manila import exception -from manila.i18n import _, _LE +from manila.i18n import _ from manila.share.drivers.netapp.dataontap.client import api as netapp_api @@ -60,9 +60,9 @@ class PerformanceLibrary(object): self.avg_processor_busy_base_counter_name = 'cpu_elapsed_time' else: self.avg_processor_busy_base_counter_name = 'cpu_elapsed_time1' - LOG.exception(_LE('Could not get performance base counter ' - 'name. Performance-based scheduler ' - 'functions may not be available.')) + LOG.exception('Could not get performance base counter ' + 'name. Performance-based scheduler ' + 'functions may not be available.') def update_performance_cache(self, flexvol_pools, aggregate_pools): """Called periodically to update per-pool node utilization metrics.""" @@ -194,8 +194,8 @@ class PerformanceLibrary(object): return max(min(100.0, node_utilization), 0) except Exception: - LOG.exception(_LE('Could not calculate node utilization for ' - 'node %s.'), node_name) + LOG.exception('Could not calculate node utilization for ' + 'node %s.', node_name) return DEFAULT_UTILIZATION def _get_kahuna_utilization(self, counters_t1, counters_t2): @@ -343,8 +343,8 @@ class PerformanceLibrary(object): self._get_node_utilization_wafl_counters(node_name) + self._get_node_utilization_processor_counters(node_name)) except netapp_api.NaApiError: - LOG.exception(_LE('Could not get utilization counters from node ' - '%s'), node_name) + LOG.exception('Could not get utilization counters from node ' + '%s', node_name) return None def _get_node_utilization_system_counters(self, node_name): diff --git a/manila/share/drivers/netapp/dataontap/protocols/nfs_cmode.py b/manila/share/drivers/netapp/dataontap/protocols/nfs_cmode.py index a93b053c66..c67927a5aa 100644 --- a/manila/share/drivers/netapp/dataontap/protocols/nfs_cmode.py +++ b/manila/share/drivers/netapp/dataontap/protocols/nfs_cmode.py @@ -23,7 +23,6 @@ import six from manila.common import constants from manila import exception -from manila.i18n import _, _LI from manila.share.drivers.netapp.dataontap.protocols import base from manila.share.drivers.netapp import utils as na_utils @@ -89,15 +88,15 @@ class NetAppCmodeNFSHelper(base.NetAppBaseHelper): self._is_readonly(new_rules[address])) # Rename policy currently in force - LOG.info(_LI('Renaming NFS export policy for share %(share)s to ' - '%(policy)s.') % + LOG.info('Renaming NFS export policy for share %(share)s to ' + '%(policy)s.' % {'share': share_name, 'policy': temp_old_export_policy_name}) self._client.rename_nfs_export_policy(export_policy_name, temp_old_export_policy_name) # Switch share to the new policy - LOG.info(_LI('Setting NFS export policy for share %(share)s to ' - '%(policy)s.') % + LOG.info('Setting NFS export policy for share %(share)s to ' + '%(policy)s.' % {'share': share_name, 'policy': temp_new_export_policy_name}) self._client.set_nfs_export_policy_for_volume( share_name, temp_new_export_policy_name) @@ -106,8 +105,8 @@ class NetAppCmodeNFSHelper(base.NetAppBaseHelper): self._client.soft_delete_nfs_export_policy(temp_old_export_policy_name) # Rename new policy to its final name - LOG.info(_LI('Renaming NFS export policy for share %(share)s to ' - '%(policy)s.') % + LOG.info('Renaming NFS export policy for share %(share)s to ' + '%(policy)s.' % {'share': share_name, 'policy': export_policy_name}) self._client.rename_nfs_export_policy(temp_new_export_policy_name, export_policy_name) @@ -117,8 +116,8 @@ class NetAppCmodeNFSHelper(base.NetAppBaseHelper): """Checks whether access rule type and level are valid.""" if rule['access_type'] != 'ip': - msg = _("Clustered Data ONTAP supports only 'ip' type for share " - "access rules with NFS protocol.") + msg = ("Clustered Data ONTAP supports only 'ip' type for share " + "access rules with NFS protocol.") raise exception.InvalidShareAccess(reason=msg) if rule['access_level'] not in constants.ACCESS_LEVELS: diff --git a/manila/share/drivers/netapp/utils.py b/manila/share/drivers/netapp/utils.py index faf855070b..f7e315bf9d 100644 --- a/manila/share/drivers/netapp/utils.py +++ b/manila/share/drivers/netapp/utils.py @@ -24,7 +24,7 @@ from oslo_log import log import six from manila import exception -from manila.i18n import _, _LI, _LW +from manila.i18n import _ from manila import version @@ -43,9 +43,9 @@ def validate_driver_instantiation(**kwargs): """ if kwargs and kwargs.get('netapp_mode') == 'proxy': return - LOG.warning(_LW('Please use NetAppDriver in the configuration file ' - 'to load the driver instead of directly specifying ' - 'the driver module name.')) + LOG.warning('Please use NetAppDriver in the configuration file ' + 'to load the driver instead of directly specifying ' + 'the driver module name.') def check_flags(required_flags, configuration): @@ -74,7 +74,7 @@ def setup_tracing(trace_flags_string): flags = trace_flags_string.split(',') flags = [flag.strip() for flag in flags] for invalid_flag in list(set(flags) - set(VALID_TRACE_FLAGS)): - LOG.warning(_LW('Invalid trace flag: %s') % invalid_flag) + LOG.warning('Invalid trace flag: %s' % invalid_flag) TRACE_METHOD = 'method' in flags TRACE_API = 'api' in flags @@ -164,7 +164,7 @@ class OpenStackInfo(object): "'%{version}\t%{release}\t%{vendor}'", self.PACKAGE_NAME) if not out: - LOG.info(_LI('No rpm info found for %(pkg)s package.') % { + LOG.info('No rpm info found for %(pkg)s package.' % { 'pkg': self.PACKAGE_NAME}) return False parts = out.split() @@ -173,7 +173,7 @@ class OpenStackInfo(object): self._vendor = ' '.join(parts[2::]) return True except Exception as e: - LOG.info(_LI('Could not run rpm command: %(msg)s.') % { + LOG.info('Could not run rpm command: %(msg)s.' % { 'msg': e}) return False @@ -185,9 +185,9 @@ class OpenStackInfo(object): out, err = putils.execute("dpkg-query", "-W", "-f='${Version}'", self.PACKAGE_NAME) if not out: - LOG.info(_LI( - 'No dpkg-query info found for %(pkg)s package.') % { - 'pkg': self.PACKAGE_NAME}) + LOG.info( + 'No dpkg-query info found for %(pkg)s package.' % { + 'pkg': self.PACKAGE_NAME}) return False # Debian format: [epoch:]upstream_version[-debian_revision] deb_version = out @@ -204,7 +204,7 @@ class OpenStackInfo(object): self._vendor = _vendor return True except Exception as e: - LOG.info(_LI('Could not run dpkg-query command: %(msg)s.') % { + LOG.info('Could not run dpkg-query command: %(msg)s.' % { 'msg': e}) return False diff --git a/manila/share/drivers/nexenta/ns4/nexenta_nas.py b/manila/share/drivers/nexenta/ns4/nexenta_nas.py index 5fda5e32f3..82649cdac9 100644 --- a/manila/share/drivers/nexenta/ns4/nexenta_nas.py +++ b/manila/share/drivers/nexenta/ns4/nexenta_nas.py @@ -16,7 +16,7 @@ from oslo_log import log from manila import exception -from manila.i18n import _, _LI +from manila.i18n import _ from manila.share import driver from manila.share.drivers.nexenta.ns4 import nexenta_nfs_helper from manila.share.drivers.nexenta import options @@ -98,7 +98,7 @@ class NexentaNasDriver(driver.ShareDriver): LOG.debug('Creating a snapshot of share %s.', snapshot['share_name']) snap_id = self.helper.create_snapshot( snapshot['share_name'], snapshot['name']) - LOG.info(_LI('Created snapshot %s.'), snap_id) + LOG.info('Created snapshot %s.', snap_id) def delete_snapshot(self, context, snapshot, share_server=None): """Delete a snapshot.""" diff --git a/manila/share/drivers/nexenta/ns4/nexenta_nfs_helper.py b/manila/share/drivers/nexenta/ns4/nexenta_nfs_helper.py index 6636ef74f9..f9fab155a8 100644 --- a/manila/share/drivers/nexenta/ns4/nexenta_nfs_helper.py +++ b/manila/share/drivers/nexenta/ns4/nexenta_nfs_helper.py @@ -18,7 +18,7 @@ from oslo_utils import excutils from manila.common import constants as common from manila import exception -from manila.i18n import _, _LI +from manila.i18n import _ from manila.share.drivers.nexenta.ns4 import jsonrpc from manila.share.drivers.nexenta import utils @@ -112,8 +112,8 @@ class NFSHelper(object): except exception.NexentaException as e: with excutils.save_and_reraise_exception() as exc: if NOT_EXIST in e.args[0]: - LOG.info(_LI('Folder %s does not exist, it was ' - 'already deleted.'), folder) + LOG.info('Folder %s does not exist, it was ' + 'already deleted.', folder) exc.reraise = False def _get_share_path(self, share_name): @@ -137,20 +137,20 @@ class NFSHelper(object): except exception.NexentaException as e: with excutils.save_and_reraise_exception() as exc: if NOT_EXIST in e.args[0]: - LOG.info(_LI('Snapshot %(folder)s@%(snapshot)s does not ' - 'exist, it was already deleted.'), + LOG.info('Snapshot %(folder)s@%(snapshot)s does not ' + 'exist, it was already deleted.', { 'folder': share_name, 'snapshot': snapshot_name, - }) + }) exc.reraise = False elif DEP_CLONES in e.args[0]: - LOG.info(_LI( + LOG.info( 'Snapshot %(folder)s@%(snapshot)s has dependent ' - 'clones, it will be deleted later.'), { + 'clones, it will be deleted later.', { 'folder': share_name, 'snapshot': snapshot_name - }) + }) exc.reraise = False def create_share_from_snapshot(self, share, snapshot): diff --git a/manila/share/drivers/nexenta/ns5/nexenta_nas.py b/manila/share/drivers/nexenta/ns5/nexenta_nas.py index 04cb912d3b..f36b819dc2 100644 --- a/manila/share/drivers/nexenta/ns5/nexenta_nas.py +++ b/manila/share/drivers/nexenta/ns5/nexenta_nas.py @@ -18,7 +18,7 @@ from oslo_utils import units from manila.common import constants as common from manila import exception -from manila.i18n import _, _LW, _LE +from manila.i18n import _ from manila.share import driver from manila.share.drivers.nexenta.ns5 import jsonrpc from manila.share.drivers.nexenta import options @@ -156,9 +156,9 @@ class NexentaNasDriver(driver.ShareDriver): try: self.delete_share(None, share) except exception.NexentaException as exc: - LOG.warning(_LW( + LOG.warning( "Cannot destroy created filesystem: %(vol)s/%(folder)s, " - "exception: %(exc)s"), + "exception: %(exc)s", {'vol': self.pool_name, 'folder': '/'.join( (self.fs_prefix, share['name'])), 'exc': exc}) raise @@ -194,12 +194,12 @@ class NexentaNasDriver(driver.ShareDriver): self._add_permission(share['name']) except exception.NexentaException: LOG.exception( - _LE('Failed to add permissions for %s'), share['name']) + ('Failed to add permissions for %s'), share['name']) try: self.delete_share(None, share) except exception.NexentaException: - LOG.warning(_LW("Cannot destroy cloned filesystem: " - "%(vol)s/%(filesystem)s"), + LOG.warning("Cannot destroy cloned filesystem: " + "%(vol)s/%(filesystem)s", {'vol': self.pool_name, 'filesystem': '/'.join( (self.fs_prefix, share['name']))}) @@ -269,7 +269,7 @@ class NexentaNasDriver(driver.ShareDriver): except exception.NexentaException as e: if e.kwargs['code'] == 'ENOENT': LOG.warning( - _LW('snapshot %(name)s not found, response: %(msg)s'), { + 'snapshot %(name)s not found, response: %(msg)s', { 'name': snapshot['name'], 'msg': e.msg}) else: raise diff --git a/manila/share/drivers/qnap/qnap.py b/manila/share/drivers/qnap/qnap.py index d85bf9c8c0..bf0075607a 100644 --- a/manila/share/drivers/qnap/qnap.py +++ b/manila/share/drivers/qnap/qnap.py @@ -25,8 +25,8 @@ from oslo_utils import units from manila.common import constants from manila import exception +from manila.i18n import _ from manila import share -from manila.i18n import _, _LE, _LI, _LW from manila.share import driver from manila.share.drivers.qnap import api from manila import utils @@ -83,9 +83,9 @@ class QnapShareDriver(driver.ShareDriver): try: self.api_executor = self._create_api_executor() except Exception: - LOG.exception(_LE('Failed to create HTTP client. Check IP ' - 'address, port, username, password and make ' - 'sure the array version is compatible.')) + LOG.exception('Failed to create HTTP client. Check IP ' + 'address, port, username, password and make ' + 'sure the array version is compatible.') raise def check_for_setup_error(self): @@ -301,7 +301,7 @@ class QnapShareDriver(driver.ShareDriver): # Use private_storage to retreive volume ID created in the NAS. volID = self.private_storage.get(share['id'], 'volID') if not volID: - LOG.warning(_LW('volID for Share %s does not exist'), share['id']) + LOG.warning('volID for Share %s does not exist', share['id']) return LOG.debug('volID: %s', volID) @@ -309,7 +309,7 @@ class QnapShareDriver(driver.ShareDriver): self.configuration.qnap_poolname, vol_no=volID) if del_share is None: - LOG.warning(_LW('Share %s does not exist'), share['id']) + LOG.warning('Share %s does not exist', share['id']) return vol_no = del_share.find('vol_no').text @@ -350,7 +350,7 @@ class QnapShareDriver(driver.ShareDriver): volID = self.private_storage.get(snapshot['share']['id'], 'volID') if not volID: LOG.warning( - _LW('volID for Share %s does not exist'), + 'volID for Share %s does not exist', snapshot['share']['id']) raise exception.ShareResourceNotFound( share_id=snapshot['share']['id']) @@ -401,7 +401,7 @@ class QnapShareDriver(driver.ShareDriver): snapshot_id = (snapshot.get('provider_location') or self.private_storage.get(snapshot['id'], 'snapshot_id')) if not snapshot_id: - LOG.warning(_LW('Snapshot %s does not exist'), snapshot['id']) + LOG.warning('Snapshot %s does not exist', snapshot['id']) return LOG.debug('snapshot_id: %s', snapshot_id) @@ -421,7 +421,7 @@ class QnapShareDriver(driver.ShareDriver): snapshot_id = (snapshot.get('provider_location') or self.private_storage.get(snapshot['id'], 'snapshot_id')) if not snapshot_id: - LOG.warning(_LW('Snapshot %s does not exist'), snapshot['id']) + LOG.warning('Snapshot %s does not exist', snapshot['id']) raise exception.SnapshotResourceNotFound(name=snapshot['id']) LOG.debug('snapshot_id: %s', snapshot_id) @@ -568,7 +568,7 @@ class QnapShareDriver(driver.ShareDriver): try: self._check_share_access(share_proto, access_type) except exception.InvalidShareAccess: - LOG.warning(_LW('The denied rule is invalid and does not exist.')) + LOG.warning('The denied rule is invalid and does not exist.') return hostlist = self.api_executor.get_host_list() @@ -603,8 +603,8 @@ class QnapShareDriver(driver.ShareDriver): """Manages a share that exists on backend.""" if share['share_proto'].lower() == 'nfs': # 10.0.0.1:/share/example - LOG.info(_LI("Share %(shr_path)s will be managed with ID " - "%(shr_id)s."), + LOG.info("Share %(shr_path)s will be managed with ID" + "%(shr_id)s.", {'shr_path': share['export_locations'][0]['path'], 'shr_id': share['id']}) @@ -646,8 +646,8 @@ class QnapShareDriver(driver.ShareDriver): volName = self.private_storage.get(share['id'], 'volName') LOG.debug('volName: %s', volName) - LOG.info(_LI("Share %(shr_path)s was successfully managed with ID " - "%(shr_id)s."), + LOG.info("Share %(shr_path)s was successfully managed with ID " + "%(shr_id)s.", {'shr_path': share['export_locations'][0]['path'], 'shr_id': share['id']}) diff --git a/manila/share/drivers/quobyte/jsonrpc.py b/manila/share/drivers/quobyte/jsonrpc.py index 26c9b20883..cfa031e68b 100644 --- a/manila/share/drivers/quobyte/jsonrpc.py +++ b/manila/share/drivers/quobyte/jsonrpc.py @@ -28,7 +28,6 @@ import six import six.moves.urllib.parse as urlparse from manila import exception -from manila.i18n import _LW from manila import utils LOG = log.getLogger(__name__) @@ -48,9 +47,9 @@ class JsonRpc(object): if self._url_scheme == 'https': if not self._ca_file: self._ca_file = False - LOG.warning(_LW( + LOG.warning( "Will not verify the server certificate of the API service" - " because the CA certificate is not available.")) + " because the CA certificate is not available.") self._id = 0 self._credentials = auth.HTTPBasicAuth( user_credentials[0], user_credentials[1]) diff --git a/manila/share/drivers/quobyte/quobyte.py b/manila/share/drivers/quobyte/quobyte.py index 051c8b0934..e6339f1d8a 100644 --- a/manila/share/drivers/quobyte/quobyte.py +++ b/manila/share/drivers/quobyte/quobyte.py @@ -29,7 +29,7 @@ import six from manila.common import constants from manila import exception -from manila.i18n import _, _LE, _LI, _LW +from manila.i18n import _ from manila.share import driver from manila.share.drivers.quobyte import jsonrpc @@ -121,7 +121,7 @@ class QuobyteShareDriver(driver.ExecuteMixin, driver.ShareDriver,): try: self.rpc.call('getInformation', {}) except Exception as exc: - LOG.error(_LE("Could not connect to API: %s"), exc) + LOG.error("Could not connect to API: %s", exc) raise exception.QBException( _('Could not connect to API: %s') % exc) @@ -143,8 +143,8 @@ class QuobyteShareDriver(driver.ExecuteMixin, driver.ShareDriver,): total = float(result['total_physical_capacity']) used = float(result['total_physical_usage']) - LOG.info(_LI('Read capacity of %(cap)s bytes and ' - 'usage of %(use)s bytes from backend. '), + LOG.info('Read capacity of %(cap)s bytes and ' + 'usage of %(use)s bytes from backend. ', {'cap': total, 'use': used}) free = total - used if free < 0: @@ -244,8 +244,8 @@ class QuobyteShareDriver(driver.ExecuteMixin, driver.ShareDriver,): share['name'], self._get_project_name(context, share['project_id'])) if not volume_uuid: - LOG.warning(_LW("No volume found for " - "share %(project_id)s/%(name)s") + LOG.warning("No volume found for " + "share %(project_id)s/%(name)s" % {"project_id": share['project_id'], "name": share['name']}) return @@ -374,7 +374,7 @@ class QuobyteShareDriver(driver.ExecuteMixin, driver.ShareDriver,): self._allow_access(context, share, a_rule) else: if not access_rules: - LOG.warning(_LW("No access rules provided in update_access.")) + LOG.warning("No access rules provided in update_access.") else: # Handling access rule recovery existing_rules = self._fetch_existing_access(context, share) diff --git a/manila/share/drivers/service_instance.py b/manila/share/drivers/service_instance.py index 27b671d66c..14a9363ded 100644 --- a/manila/share/drivers/service_instance.py +++ b/manila/share/drivers/service_instance.py @@ -32,7 +32,7 @@ from manila.common import constants as const from manila import compute from manila import context from manila import exception -from manila.i18n import _, _LW +from manila.i18n import _ from manila.network.linux import ip_lib from manila.network.neutron import api as neutron from manila import utils @@ -327,8 +327,8 @@ class ServiceInstanceManager(object): name = name or self.get_config_option( "service_instance_security_group") if not name: - LOG.warning(_LW("Name for service instance security group is not " - "provided. Skipping security group step.")) + LOG.warning("Name for service instance security group is not " + "provided. Skipping security group step.") return None s_groups = [s for s in self.compute_api.security_group_list(context) if s.name == name] @@ -359,15 +359,15 @@ class ServiceInstanceManager(object): def ensure_service_instance(self, context, server): """Ensures that server exists and active.""" if 'instance_id' not in server: - LOG.warning(_LW("Unable to check server existence since " - "'instance_id' key is not set in share server " - "backend details.")) + LOG.warning("Unable to check server existence since " + "'instance_id' key is not set in share server " + "backend details.") return False try: inst = self.compute_api.server_get(self.admin_context, server['instance_id']) except exception.InstanceNotFound: - LOG.warning(_LW("Service instance %s does not exist."), + LOG.warning("Service instance %s does not exist.", server['instance_id']) return False if inst['status'] == 'ACTIVE': @@ -510,11 +510,11 @@ class ServiceInstanceManager(object): raise exception.ServiceInstanceException( _('Neither service instance password nor key are available.')) if not key_path: - LOG.warning(_LW( + LOG.warning( 'No key path is available. May be non-existent key path is ' 'provided. Check path_to_private_key (current value ' '%(private_path)s) and path_to_public_key (current value ' - '%(public_path)s) in manila configuration file.'), dict( + '%(public_path)s) in manila configuration file.', dict( private_path=self.path_to_private_key, public_path=self.path_to_public_key)) network_data = self.network_helper.setup_network(network_info) @@ -965,8 +965,8 @@ class NeutronNetworkHelper(BaseNetworkhelper): addr_list = device.addr.list() except Exception as e: if 'does not exist' in six.text_type(e): - LOG.warning(_LW( - "Device %s does not exist anymore.") % device.name) + LOG.warning( + "Device %s does not exist anymore." % device.name) else: raise for addr in addr_list: diff --git a/manila/share/drivers/tegile/tegile.py b/manila/share/drivers/tegile/tegile.py index f75b18fb1d..7b209357ea 100644 --- a/manila/share/drivers/tegile/tegile.py +++ b/manila/share/drivers/tegile/tegile.py @@ -23,11 +23,11 @@ import six from oslo_config import cfg from oslo_log import log -from manila import utils -from manila.i18n import _, _LI, _LW from manila import exception +from manila.i18n import _ from manila.share import driver from manila.share import utils as share_utils +from manila import utils tegile_opts = [ cfg.StrOpt('tegile_nas_server', @@ -201,7 +201,7 @@ class TegileShareDriver(driver.ShareDriver): # of 'sharename' if inherited share properties are selected. ip, real_share_name = self._api('createShare', params).split() - LOG.info(_LI("Created share %(sharename)s, share id %(shid)s."), + LOG.info("Created share %(sharename)s, share id %(shid)s.", {'sharename': share_name, 'shid': share['id']}) return self._get_location_path(real_share_name, share_proto, ip) @@ -273,8 +273,8 @@ class TegileShareDriver(driver.ShareDriver): params = (share, snap_name, False) - LOG.info(_LI('Creating snapshot for share_name=%(shr)s' - ' snap_name=%(name)s'), + LOG.info('Creating snapshot for share_name=%(shr)s' + ' snap_name=%(name)s', {'shr': share_name, 'name': snap_name}) self._api('createShareSnapshot', params) @@ -383,18 +383,18 @@ class TegileShareDriver(driver.ShareDriver): def _check_share_access(self, share_proto, access_type): if share_proto == 'CIFS' and access_type != 'user': - reason = _LW('Only USER access type is allowed for ' - 'CIFS shares.') + reason = ('Only USER access type is allowed for ' + 'CIFS shares.') LOG.warning(reason) raise exception.InvalidShareAccess(reason=reason) elif share_proto == 'NFS' and access_type not in ('ip', 'user'): - reason = _LW('Only IP or USER access types are allowed for ' - 'NFS shares.') + reason = ('Only IP or USER access types are allowed for ' + 'NFS shares.') LOG.warning(reason) raise exception.InvalidShareAccess(reason=reason) elif share_proto not in ('NFS', 'CIFS'): - reason = _LW('Unsupported protocol \"%s\" specified for ' - 'access rule.') % share_proto + reason = ('Unsupported protocol \"%s\" specified for ' + 'access rule.') % share_proto raise exception.InvalidShareAccess(reason=reason) @debugger diff --git a/manila/share/drivers/windows/service_instance.py b/manila/share/drivers/windows/service_instance.py index 3ae2427763..b454b439ec 100644 --- a/manila/share/drivers/windows/service_instance.py +++ b/manila/share/drivers/windows/service_instance.py @@ -21,7 +21,7 @@ from oslo_config import cfg from oslo_log import log from manila import exception -from manila.i18n import _, _LI, _LW +from manila.i18n import _ from manila.share.drivers import service_instance from manila.share.drivers.windows import windows_utils from manila.share.drivers.windows import winrm_helper @@ -232,21 +232,21 @@ class WindowsServiceInstanceManager(service_instance.ServiceInstanceManager): def get_valid_security_service(self, security_services): if not security_services: - LOG.info(_LI("No security services provided.")) + LOG.info("No security services provided.") elif len(security_services) > 1: - LOG.warning(_LW("Multiple security services provided. Only one " - "security service of type 'active_directory' " - "is supported.")) + LOG.warning("Multiple security services provided. Only one " + "security service of type 'active_directory' " + "is supported.") else: security_service = security_services[0] security_service_type = security_service['type'] if security_service_type == 'active_directory': return security_service else: - LOG.warning(_LW("Only security services of type " - "'active_directory' are supported. " - "Retrieved security " - "service type: %(sec_type)s."), + LOG.warning("Only security services of type " + "'active_directory' are supported. " + "Retrieved security " + "service type: %(sec_type)s.", {'sec_type': security_service_type}) return None diff --git a/manila/share/drivers/windows/windows_smb_driver.py b/manila/share/drivers/windows/windows_smb_driver.py index c3b0840448..3dd32a505f 100644 --- a/manila/share/drivers/windows/windows_smb_driver.py +++ b/manila/share/drivers/windows/windows_smb_driver.py @@ -18,7 +18,6 @@ import os from oslo_log import log from oslo_utils import units -from manila.i18n import _LW from manila.share import driver as base_driver from manila.share.drivers import generic from manila.share.drivers.windows import service_instance @@ -67,9 +66,9 @@ class WindowsSMBDriver(generic.GenericShareDriver): security_service['user'], security_service['password']) except Exception as exc: - LOG.warning(_LW("Failed to remove service instance " - "%(instance_id)s from domain %(domain)s. " - "Exception: %(exc)s."), + LOG.warning("Failed to remove service instance " + "%(instance_id)s from domain %(domain)s. " + "Exception: %(exc)s.", dict(instance_id=server_details['instance_id'], domain=security_service['domain'], exc=exc)) diff --git a/manila/share/drivers/windows/windows_smb_helper.py b/manila/share/drivers/windows/windows_smb_helper.py index 98860d504b..f54e8ec316 100644 --- a/manila/share/drivers/windows/windows_smb_helper.py +++ b/manila/share/drivers/windows/windows_smb_helper.py @@ -20,7 +20,6 @@ from oslo_log import log from manila.common import constants from manila import exception -from manila.i18n import _LI, _LW from manila.share.drivers import helpers from manila.share.drivers.windows import windows_utils @@ -78,7 +77,7 @@ class WindowsSMBHelper(helpers.CIFSHelperBase): '-ReadAccess', "*%s" % self._NULL_SID] self._remote_exec(server, cmd) else: - LOG.info(_LI("Skipping creating export %s as it already exists."), + LOG.info("Skipping creating export %s as it already exists.", share_name) return self.get_exports_for_share(server, export_location) @@ -127,20 +126,20 @@ class WindowsSMBHelper(helpers.CIFSHelperBase): share_name) else: LOG.warning( - _LW("Found explicit deny ACE rule that was not " - "created by Manila and will be ignored: %s"), + "Found explicit deny ACE rule that was not " + "created by Manila and will be ignored: %s", raw_acl) continue if access_level == self._ACCESS_LEVEL_CUSTOM: LOG.warning( - _LW("Found 'custom' ACE rule that will be ignored: %s"), + "Found 'custom' ACE rule that will be ignored: %s", raw_acl) continue elif access_right == self._WIN_ACCESS_RIGHT_FULL: LOG.warning( - _LW("Account '%(access_to)s' was given full access " - "right on share %(share_name)s. Manila only " - "grants 'change' access."), + "Account '%(access_to)s' was given full access " + "right on share %(share_name)s. Manila only " + "grants 'change' access.", {'access_to': access_to, 'share_name': share_name}) @@ -159,8 +158,8 @@ class WindowsSMBHelper(helpers.CIFSHelperBase): "-AccountName", "'%s'" % access_to, "-Force"] self._remote_exec(server, cmd) self._refresh_acl(server, share_name) - LOG.info(_LI("Granted %(access_level)s access to '%(access_to)s' " - "on share %(share_name)s"), + LOG.info("Granted %(access_level)s access to '%(access_to)s' " + "on share %(share_name)s", {'access_level': access_level, 'access_to': access_to, 'share_name': share_name}) @@ -174,8 +173,8 @@ class WindowsSMBHelper(helpers.CIFSHelperBase): '-AccountName', '"%s"' % access_to, '-Force'] self._remote_exec(server, cmd) self._refresh_acl(server, share_name) - LOG.info(_LI("Revoked access to '%(access_to)s' " - "on share %(share_name)s"), + LOG.info("Revoked access to '%(access_to)s' " + "on share %(share_name)s", {'access_to': access_to, 'share_name': share_name}) @@ -207,12 +206,12 @@ class WindowsSMBHelper(helpers.CIFSHelperBase): except (exception.InvalidShareAccess, exception.InvalidShareAccessLevel): # This check will allow invalid rules to be deleted. - LOG.warning(_LW( + LOG.warning( "Unsupported access level %(level)s or access type " "%(type)s, skipping removal of access rule to " - "%(to)s.") % {'level': deleted_rule['access_level'], - 'type': deleted_rule['access_type'], - 'to': deleted_rule['access_to']}) + "%(to)s." % {'level': deleted_rule['access_level'], + 'type': deleted_rule['access_type'], + 'to': deleted_rule['access_to']}) continue self._revoke_share_access(server, share_name, deleted_rule['access_to']) diff --git a/manila/share/drivers/windows/windows_utils.py b/manila/share/drivers/windows/windows_utils.py index 28601ed7d5..0263c81953 100644 --- a/manila/share/drivers/windows/windows_utils.py +++ b/manila/share/drivers/windows/windows_utils.py @@ -17,7 +17,6 @@ import re from oslo_log import log -from manila.i18n import _LI LOG = log.getLogger(__name__) @@ -125,9 +124,9 @@ class WindowsUtils(object): # NOTE(lpetrut): An instance reboot is needed but this will be # performed using Nova so that the instance state can be # retrieved easier. - LOG.info(_LI("Joining server %(ip)s to Active Directory " - "domain %(domain)s"), dict(ip=server['ip'], - domain=domain)) + LOG.info("Joining server %(ip)s to Active Directory " + "domain %(domain)s", dict(ip=server['ip'], + domain=domain)) cmds = [ ('$password = "%s" | ' 'ConvertTo-SecureString -asPlainText -Force' % admin_password), diff --git a/manila/share/drivers/zfsonlinux/driver.py b/manila/share/drivers/zfsonlinux/driver.py index 7f5d09d0f6..cd578bd93f 100644 --- a/manila/share/drivers/zfsonlinux/driver.py +++ b/manila/share/drivers/zfsonlinux/driver.py @@ -30,7 +30,7 @@ from oslo_utils import timeutils from manila.common import constants from manila import exception -from manila.i18n import _, _LI, _LW +from manila.i18n import _ from manila.share import configuration from manila.share import driver from manila.share.drivers.zfsonlinux import utils as zfs_utils @@ -279,7 +279,7 @@ class ZFSonLinuxShareDriver(zfs_utils.ExecuteMixin, driver.ShareDriver): self.zfs('destroy', '-f', name) return except exception.ProcessExecutionError: - LOG.info(_LI("Failed to destroy ZFS dataset, retrying one time")) + LOG.info("Failed to destroy ZFS dataset, retrying one time") # NOTE(bswartz): There appears to be a bug in ZFS when creating and # destroying datasets concurrently where the filesystem remains mounted @@ -529,8 +529,8 @@ class ZFSonLinuxShareDriver(zfs_utils.ExecuteMixin, driver.ShareDriver): break else: LOG.warning( - _LW("Share with '%(id)s' ID and '%(name)s' NAME is " - "absent on backend. Nothing has been deleted."), + "Share with '%(id)s' ID and '%(name)s' NAME is " + "absent on backend. Nothing has been deleted.", {'id': share['id'], 'name': dataset_name}) self.private_storage.delete(share['id']) @@ -574,8 +574,8 @@ class ZFSonLinuxShareDriver(zfs_utils.ExecuteMixin, driver.ShareDriver): break else: LOG.warning( - _LW("Snapshot with '%(id)s' ID and '%(name)s' NAME is " - "absent on backend. Nothing has been deleted."), + "Snapshot with '%(id)s' ID and '%(name)s' NAME is " + "absent on backend. Nothing has been deleted.", {'id': snapshot['id'], 'name': snapshot_name}) @ensure_share_server_not_provided @@ -972,8 +972,8 @@ class ZFSonLinuxShareDriver(zfs_utils.ExecuteMixin, driver.ShareDriver): break else: LOG.warning( - _LW("Share replica with '%(id)s' ID and '%(name)s' NAME is " - "absent on backend. Nothing has been deleted."), + "Share replica with '%(id)s' ID and '%(name)s' NAME is " + "absent on backend. Nothing has been deleted.", {'id': replica['id'], 'name': dataset_name}) self.private_storage.delete(replica['id']) @@ -1131,7 +1131,7 @@ class ZFSonLinuxShareDriver(zfs_utils.ExecuteMixin, driver.ShareDriver): 'sudo', 'zfs', 'receive', '-vF', dataset_name, ) except exception.ProcessExecutionError as e: - LOG.warning(_LW("Failed to sync replica %(id)s. %(e)s"), + LOG.warning("Failed to sync replica %(id)s. %(e)s", {'id': repl['id'], 'e': e}) replica_dict[repl['id']]['replica_state'] = ( constants.REPLICA_STATE_OUT_OF_SYNC) @@ -1153,7 +1153,7 @@ class ZFSonLinuxShareDriver(zfs_utils.ExecuteMixin, driver.ShareDriver): constants.REPLICA_STATE_IN_SYNC) except Exception as e: LOG.warning( - _LW("Failed to update currently active replica. \n%s"), e) + "Failed to update currently active replica. \n%s", e) replica_dict[active_replica['id']]['replica_state'] = ( constants.REPLICA_STATE_OUT_OF_SYNC) @@ -1185,7 +1185,7 @@ class ZFSonLinuxShareDriver(zfs_utils.ExecuteMixin, driver.ShareDriver): 'sudo', 'zfs', 'receive', '-vF', dataset_name, ) except exception.ProcessExecutionError as e: - LOG.warning(_LW("Failed to sync replica %(id)s. %(e)s"), + LOG.warning("Failed to sync replica %(id)s. %(e)s", {'id': repl['id'], 'e': e}) replica_dict[repl['id']]['replica_state'] = ( constants.REPLICA_STATE_OUT_OF_SYNC) @@ -1274,7 +1274,7 @@ class ZFSonLinuxShareDriver(zfs_utils.ExecuteMixin, driver.ShareDriver): ) except exception.ProcessExecutionError as e: LOG.warning( - _LW("Failed to sync snapshot instance %(id)s. %(e)s"), + "Failed to sync snapshot instance %(id)s. %(e)s", {'id': replica_snapshot['id'], 'e': e}) replica_snapshots_dict[replica_snapshot['id']]['status'] = ( constants.STATUS_ERROR) @@ -1526,8 +1526,8 @@ class ZFSonLinuxShareDriver(zfs_utils.ExecuteMixin, driver.ShareDriver): x for x in line.strip().split(' ') if x != ''][1] self.execute('sudo', 'kill', '-9', migr_pid) except exception.ProcessExecutionError as e: - LOG.warning(_LW( - "Caught following error trying to kill migration process: %s"), + LOG.warning( + "Caught following error trying to kill migration process: %s", e) # Sleep couple of seconds before destroying updated objects @@ -1544,9 +1544,9 @@ class ZFSonLinuxShareDriver(zfs_utils.ExecuteMixin, driver.ShareDriver): 'sudo', 'zfs', 'destroy', '-r', dst_dataset_name, ) except exception.ProcessExecutionError as e: - LOG.warning(_LW( + LOG.warning( "Failed to destroy destination dataset with following error: " - "%s"), + "%s", e) LOG.debug( diff --git a/manila/share/drivers/zfsonlinux/utils.py b/manila/share/drivers/zfsonlinux/utils.py index acbab268cf..19b23347e6 100644 --- a/manila/share/drivers/zfsonlinux/utils.py +++ b/manila/share/drivers/zfsonlinux/utils.py @@ -28,7 +28,7 @@ import six from manila.common import constants from manila import exception -from manila.i18n import _, _LE, _LI, _LW +from manila.i18n import _ from manila.share import driver from manila.share.drivers.ganesha import utils as ganesha_utils from manila import utils @@ -103,7 +103,7 @@ class ExecuteMixin(driver.ExecuteMixin): try: return self.execute(*cmd, **kwargs) except exception.ProcessExecutionError as e: - LOG.warning(_LW("Failed to run command, got error: %s"), e) + LOG.warning("Failed to run command, got error: %s", e) raise def _get_option(self, resource_name, option_name, pool_level=False, @@ -201,8 +201,8 @@ class NFSviaZFSHelper(ExecuteMixin, NASHelperBase): self._is_kernel_version = True except exception.ProcessExecutionError as e: LOG.info( - _LI("Looks like ZFS kernel module is absent. " - "Assuming FUSE version is installed. Error: %s"), e) + "Looks like ZFS kernel module is absent. " + "Assuming FUSE version is installed. Error: %s", e) self._is_kernel_version = False return self._is_kernel_version @@ -215,7 +215,7 @@ class NFSviaZFSHelper(ExecuteMixin, NASHelperBase): try: self.execute('sudo', 'exportfs') except exception.ProcessExecutionError: - LOG.exception(_LE("Call of 'exportfs' utility returned error.")) + LOG.exception("Call of 'exportfs' utility returned error.") raise # Init that class instance attribute on start of manila-share service @@ -300,8 +300,8 @@ class NFSviaZFSHelper(ExecuteMixin, NASHelperBase): break else: LOG.warning( - _LW("Dataset with '%(name)s' NAME is absent on backend. " - "Access rules were not applied."), {'name': dataset_name}) + "Dataset with '%(name)s' NAME is absent on backend. " + "Access rules were not applied.", {'name': dataset_name}) # NOTE(vponomaryov): Setting of ZFS share options does not remove rules # that were added and then removed. So, remove them explicitly. diff --git a/manila/share/drivers/zfssa/zfssarest.py b/manila/share/drivers/zfssa/zfssarest.py index 4bb4473e78..51867080b5 100644 --- a/manila/share/drivers/zfssa/zfssarest.py +++ b/manila/share/drivers/zfssa/zfssarest.py @@ -18,7 +18,7 @@ from oslo_log import log from oslo_serialization import jsonutils from manila import exception -from manila.i18n import _, _LE, _LW +from manila.i18n import _ from manila.share.drivers.zfssa import restclient @@ -233,12 +233,12 @@ class ZFSSAApi(object): svc = self.share_path % (pool, project, share) ret = self.rclient.delete(svc) if ret.status != restclient.Status.NO_CONTENT: - exception_msg = (_LE('Error deleting ' - 'share: %(share)s to ' - 'pool: %(pool)s ' - 'project: %(project)s ' - 'return code: %(ret.status)d ' - 'message: %(ret.data)s.'), + exception_msg = (('Error deleting ' + 'share: %(share)s to ' + 'pool: %(pool)s ' + 'project: %(project)s ' + 'return code: %(ret.status)d ' + 'message: %(ret.data)s.'), {'share': share, 'pool': pool, 'project': project, @@ -391,7 +391,7 @@ class ZFSSAApi(object): svc = "%(base)s/%(prop)s" % {'base': base, 'prop': schema['property']} ret = self.rclient.get(svc) if ret.status == restclient.Status.OK: - LOG.warning(_LW('Property %s already exists.'), schema['property']) + LOG.warning('Property %s already exists.', schema['property']) return ret = self.rclient.post(base, schema) if ret.status != restclient.Status.CREATED: diff --git a/manila/share/drivers/zfssa/zfssashare.py b/manila/share/drivers/zfssa/zfssashare.py index 1b2d28746b..a1b122a496 100644 --- a/manila/share/drivers/zfssa/zfssashare.py +++ b/manila/share/drivers/zfssa/zfssashare.py @@ -24,7 +24,6 @@ import six from manila import exception from manila.i18n import _ -from manila.i18n import _LE from manila.share import driver from manila.share.drivers.zfssa import zfssarest @@ -278,7 +277,7 @@ class ZFSSAShareDriver(driver.ShareDriver): snapshot['share_id'], snapshot['id']) if has_clones: - LOG.error(_LE("snapshot %s: has clones"), snapshot['id']) + LOG.error("snapshot %s: has clones", snapshot['id']) raise exception.ShareSnapshotIsBusy(snapshot_name=snapshot['id']) self.zfssa.delete_snapshot(lcfg.zfssa_pool, lcfg.zfssa_project, @@ -303,7 +302,7 @@ class ZFSSAShareDriver(driver.ShareDriver): try: details = self._get_share_details(name) except Exception: - LOG.error(_LE('Cannot manage share %s'), name) + LOG.error('Cannot manage share %s', name) raise lcfg = self.configuration @@ -438,8 +437,8 @@ class ZFSSAShareDriver(driver.ShareDriver): new_size_byte = int(new_size) * units.Gi if used_space > new_size_byte: - LOG.error(_LE('%(used).1fGB of share %(id)s is already used. ' - 'Cannot shrink to %(newsize)dGB.'), + LOG.error('%(used).1fGB of share %(id)s is already used. ' + 'Cannot shrink to %(newsize)dGB.', {'used': float(used_space) / units.Gi, 'id': share['id'], 'newsize': new_size}) diff --git a/manila/share/hook.py b/manila/share/hook.py index 2d762747d0..a5a91ff9a2 100644 --- a/manila/share/hook.py +++ b/manila/share/hook.py @@ -29,7 +29,6 @@ from oslo_log import log import six from manila import context as ctxt -from manila.i18n import _LW hook_options = [ @@ -110,7 +109,7 @@ class HookBase(object): *args, **kwargs) except Exception as e: if self.suppress_pre_hooks_errors: - LOG.warning(_LW("\nSuppressed exception in pre hook. %s\n"), e) + LOG.warning("\nSuppressed exception in pre hook. %s\n", e) pre_data = e else: raise @@ -135,7 +134,7 @@ class HookBase(object): except Exception as e: if self.suppress_post_hooks_errors: LOG.warning( - _LW("\nSuppressed exception in post hook. %s\n"), e) + "\nSuppressed exception in post hook. %s\n", e) post_data = e else: raise diff --git a/manila/share/manager.py b/manila/share/manager.py index a03b4c16a2..fb44014e46 100644 --- a/manila/share/manager.py +++ b/manila/share/manager.py @@ -38,7 +38,7 @@ from manila import context from manila import coordination from manila.data import rpcapi as data_rpcapi from manila import exception -from manila.i18n import _, _LE, _LI, _LW +from manila.i18n import _ from manila import manager from manila import quota from manila.share import access @@ -204,8 +204,8 @@ class ShareManager(manager.SchedulerDependentManager): share_driver = self.configuration.share_driver if share_driver in MAPPING: msg_args = {'old': share_driver, 'new': MAPPING[share_driver]} - LOG.warning(_LW("Driver path %(old)s is deprecated, update your " - "configuration to the new path %(new)s"), + LOG.warning("Driver path %(old)s is deprecated, update your " + "configuration to the new path %(new)s", msg_args) share_driver = MAPPING[share_driver] @@ -250,8 +250,8 @@ class ShareManager(manager.SchedulerDependentManager): try: pool = self.driver.get_pool(share_instance) except Exception: - LOG.exception(_LE("Failed to fetch pool name for share: " - "%(share)s."), + LOG.exception("Failed to fetch pool name for share: " + "%(share)s.", {'share': share_instance['id']}) return @@ -277,10 +277,10 @@ class ShareManager(manager.SchedulerDependentManager): self.driver.check_for_setup_error() except Exception: LOG.exception( - _LE("Error encountered during initialization of driver " - "'%(name)s' on '%(host)s' host."), { - "name": self.driver.__class__.__name__, - "host": self.host, + ("Error encountered during initialization of driver " + "'%(name)s' on '%(host)s' host."), { + "name": self.driver.__class__.__name__, + "host": self.host, } ) self.driver.initialized = False @@ -298,8 +298,8 @@ class ShareManager(manager.SchedulerDependentManager): if share_ref.is_busy: LOG.info( - _LI("Share instance %(id)s: skipping export, " - "because it is busy with an active task: %(task)s."), + "Share instance %(id)s: skipping export, " + "because it is busy with an active task: %(task)s.", {'id': share_instance['id'], 'task': share_ref['task_state']}, ) @@ -307,8 +307,8 @@ class ShareManager(manager.SchedulerDependentManager): if share_instance['status'] != constants.STATUS_AVAILABLE: LOG.info( - _LI("Share instance %(id)s: skipping export, " - "because it has '%(status)s' status."), + "Share instance %(id)s: skipping export, " + "because it has '%(status)s' status.", {'id': share_instance['id'], 'status': share_instance['status']}, ) @@ -322,9 +322,9 @@ class ShareManager(manager.SchedulerDependentManager): export_locations = self.driver.ensure_share( ctxt, share_instance, share_server=share_server) except Exception: - LOG.exception(_LE("Caught exception trying ensure " - "share '%(s_id)s'."), {'s_id': - share_instance['id']}) + LOG.exception("Caught exception trying ensure " + "share '%(s_id)s'.", + {'s_id': share_instance['id']}) continue if export_locations: @@ -341,8 +341,8 @@ class ShareManager(manager.SchedulerDependentManager): ctxt, share_instance['id'], share_server=share_server) except Exception: LOG.exception( - _LE("Unexpected error occurred while updating access " - "rules for share instance %(s_id)s."), + ("Unexpected error occurred while updating access " + "rules for share instance %(s_id)s."), {'s_id': share_instance['id']}, ) @@ -369,14 +369,14 @@ class ShareManager(manager.SchedulerDependentManager): self.snapshot_access_helper.update_access_rules( ctxt, snap_instance['id'], share_server) except Exception: - LOG.exception(_LE( + LOG.exception( "Unexpected error occurred while updating " - "access rules for snapshot instance %s."), + "access rules for snapshot instance %s.", snap_instance['id']) self.publish_service_capabilities(ctxt) - LOG.info(_LI("Finished initialization of driver: '%(driver)s" - "@%(host)s'"), + LOG.info("Finished initialization of driver: '%(driver)s" + "@%(host)s'", {"driver": self.driver.__class__.__name__, "host": self.host}) @@ -431,7 +431,7 @@ class ShareManager(manager.SchedulerDependentManager): context, parent_share_server_id) except exception.ShareServerNotFound: with excutils.save_and_reraise_exception(): - error(_LE("Parent share server %s does not exist."), + error("Parent share server %s does not exist.", parent_share_server_id) if parent_share_server['status'] != constants.STATUS_ACTIVE: @@ -439,8 +439,8 @@ class ShareManager(manager.SchedulerDependentManager): 'id': parent_share_server_id, 'status': parent_share_server['status'], } - error(_LE("Parent share server %(id)s has invalid status " - "'%(status)s'."), error_params) + error("Parent share server %(id)s has invalid status " + "'%(status)s'.", error_params) raise exception.InvalidShareServer( share_server_id=parent_share_server ) @@ -478,7 +478,7 @@ class ShareManager(manager.SchedulerDependentManager): ) except Exception as e: with excutils.save_and_reraise_exception(): - error(_LE("Cannot choose compatible share server: %s"), + error("Cannot choose compatible share server: %s", e) if not compatible_share_server: @@ -526,10 +526,10 @@ class ShareManager(manager.SchedulerDependentManager): # Create share server on backend with data from db. share_server = self._setup_server(context, share_server, metadata=metadata) - LOG.info(_LI("Share server created successfully.")) + LOG.info("Share server created successfully.") else: - LOG.info(_LI("Using preexisting share server: " - "'%(share_server_id)s'"), + LOG.info("Using preexisting share server: " + "'%(share_server_id)s'", {'share_server_id': share_server['id']}) return share_server @@ -640,7 +640,7 @@ class ShareManager(manager.SchedulerDependentManager): ) except Exception as e: with excutils.save_and_reraise_exception(): - error(_LE("Cannot choose compatible share-server: %s"), + error("Cannot choose compatible share-server: %s", e) if not compatible_share_server: @@ -670,10 +670,10 @@ class ShareManager(manager.SchedulerDependentManager): # Create share server on backend with data from db. compatible_share_server = self._setup_server( context, compatible_share_server) - LOG.info(_LI("Share server created successfully.")) + LOG.info("Share server created successfully.") else: - LOG.info(_LI("Used preexisting share server " - "'%(share_server_id)s'"), + LOG.info("Used preexisting share server " + "'%(share_server_id)s'", {'share_server_id': compatible_share_server['id']}) return compatible_share_server, updated_share_group @@ -963,8 +963,8 @@ class ShareManager(manager.SchedulerDependentManager): (constants. TASK_STATE_MIGRATION_DRIVER_PHASE1_DONE)}) - LOG.info(_LI("Share Migration for share %s completed " - "first phase successfully."), + LOG.info("Share Migration for share %s completed " + "first phase successfully.", share['id']) else: share = self.db.share_get( @@ -972,8 +972,8 @@ class ShareManager(manager.SchedulerDependentManager): if (share['task_state'] == constants.TASK_STATE_MIGRATION_CANCELLED): - LOG.warning(_LW( - "Share Migration for share %s was cancelled."), + LOG.warning( + "Share Migration for share %s was cancelled.", share['id']) except Exception: @@ -1074,7 +1074,7 @@ class ShareManager(manager.SchedulerDependentManager): except Exception as e: if not isinstance(e, NotImplementedError): LOG.exception( - _LE("The driver could not migrate the share %(shr)s"), + ("The driver could not migrate the share %(shr)s"), {'shr': share_id}) try: @@ -1265,7 +1265,7 @@ class ShareManager(manager.SchedulerDependentManager): self.db.share_snapshot_instance_delete(context, instance['id']) self.db.share_instance_delete(context, instance_id) - LOG.info(_LI("Share instance %s: deleted successfully."), + LOG.info("Share instance %s: deleted successfully.", instance_id) self._check_delete_share_server(context, share_instance) @@ -1280,8 +1280,8 @@ class ShareManager(manager.SchedulerDependentManager): share_ref = self.db.share_get(context, src_share_instance['share_id']) - LOG.info(_LI("Received request to finish Share Migration for " - "share %s."), share_ref['id']) + LOG.info("Received request to finish Share Migration for " + "share %s.", share_ref['id']) if share_ref['task_state'] == ( constants.TASK_STATE_MIGRATION_DRIVER_PHASE1_DONE): @@ -1341,8 +1341,8 @@ class ShareManager(manager.SchedulerDependentManager): self.db.share_update( context, dest_share_instance['share_id'], model_update) - LOG.info(_LI("Share Migration for share %s" - " completed successfully."), share_ref['id']) + LOG.info("Share Migration for share %s" + " completed successfully.", share_ref['id']) def _get_extra_specs_from_share_type(self, context, share_type_id): @@ -1386,8 +1386,8 @@ class ShareManager(manager.SchedulerDependentManager): context, share_ref['id'], {'task_state': constants.TASK_STATE_MIGRATION_CANCELLED}) - LOG.info(_LI("Share Migration for share %s" - " was cancelled."), share_ref['id']) + LOG.info("Share Migration for share %s" + " was cancelled.", share_ref['id']) return else: raise exception.ShareMigrationFailed(reason=msg) @@ -1489,8 +1489,8 @@ class ShareManager(manager.SchedulerDependentManager): context, share_ref['id'], {'task_state': constants.TASK_STATE_MIGRATION_CANCELLED}) - LOG.info(_LI("Share Migration for share %s" - " was cancelled."), share_ref['id']) + LOG.info("Share Migration for share %s" + " was cancelled.", share_ref['id']) @utils.require_driver_initialized def migration_get_progress(self, context, src_instance_id, @@ -1586,8 +1586,8 @@ class ShareManager(manager.SchedulerDependentManager): ) except Exception: with excutils.save_and_reraise_exception(): - error = _LE("Creation of share instance %s failed: " - "failed to get share server.") + error = ("Creation of share instance %s failed: " + "failed to get share server.") LOG.error(error, share_instance_id) self.db.share_instance_update( context, share_instance_id, @@ -1610,7 +1610,7 @@ class ShareManager(manager.SchedulerDependentManager): except Exception as e: with excutils.save_and_reraise_exception(): - LOG.error(_LE("Share instance %s failed on creation."), + LOG.error("Share instance %s failed on creation.", share_instance_id) detail_data = getattr(e, 'detail_data', {}) @@ -1626,16 +1626,16 @@ class ShareManager(manager.SchedulerDependentManager): self.db.share_export_locations_update( context, share_instance['id'], export_locations) else: - LOG.warning(_LW('Share instance information in exception ' - 'can not be written to db because it ' - 'contains %s and it is not a dictionary.'), + LOG.warning('Share instance information in exception ' + 'can not be written to db because it ' + 'contains %s and it is not a dictionary.', detail_data) self.db.share_instance_update( context, share_instance_id, {'status': constants.STATUS_ERROR} ) else: - LOG.info(_LI("Share instance %s created successfully."), + LOG.info("Share instance %s created successfully.", share_instance_id) share = self.db.share_get(context, share_instance['share_id']) updates = { @@ -1742,8 +1742,8 @@ class ShareManager(manager.SchedulerDependentManager): ) except Exception: with excutils.save_and_reraise_exception(): - LOG.error(_LE("Failed to get share server " - "for share replica creation.")) + LOG.error("Failed to get share server " + "for share replica creation.") self.db.share_replica_update( context, share_replica['id'], {'status': constants.STATUS_ERROR, @@ -1785,7 +1785,7 @@ class ShareManager(manager.SchedulerDependentManager): except Exception: with excutils.save_and_reraise_exception(): - LOG.error(_LE("Share replica %s failed on creation."), + LOG.error("Share replica %s failed on creation.", share_replica['id']) self.db.share_replica_update( context, share_replica['id'], @@ -1800,8 +1800,8 @@ class ShareManager(manager.SchedulerDependentManager): context, share_replica['id'], replica_ref.get('export_locations')) else: - msg = _LW('Invalid export locations passed to the share ' - 'manager.') + msg = ('Invalid export locations passed to the share ' + 'manager.') LOG.warning(msg) if replica_ref.get('replica_state'): @@ -1819,7 +1819,7 @@ class ShareManager(manager.SchedulerDependentManager): context, share_replica['id'], constants.STATUS_ACTIVE) - LOG.info(_LI("Share replica %s created successfully."), + LOG.info("Share replica %s created successfully.", share_replica['id']) @add_hooks @@ -1899,7 +1899,7 @@ class ShareManager(manager.SchedulerDependentManager): context, replica_snapshot['id']) self.db.share_replica_delete(context, share_replica['id']) - LOG.info(_LI("Share replica %s deleted successfully."), + LOG.info("Share replica %s deleted successfully.", share_replica['id']) @add_hooks @@ -1980,10 +1980,10 @@ class ShareManager(manager.SchedulerDependentManager): for instance in active_replica_snapshot_instances: if instance['status'] in (constants.STATUS_CREATING, constants.STATUS_DELETING): - msg = _LI("The replica snapshot instance %(instance)s was " - "in %(state)s. Since it was not in %(available)s " - "state when the replica was promoted, it will be " - "set to %(error)s.") + msg = ("The replica snapshot instance %(instance)s was " + "in %(state)s. Since it was not in %(available)s " + "state when the replica was promoted, it will be " + "set to %(error)s.") payload = { 'instance': instance['id'], 'state': instance['status'], @@ -2048,8 +2048,8 @@ class ShareManager(manager.SchedulerDependentManager): context, share_replica['id'], updated_replica.get('access_rules_status')) - LOG.info(_LI("Share replica %s: promoted to active state " - "successfully."), share_replica['id']) + LOG.info("Share replica %s: promoted to active state " + "successfully.", share_replica['id']) @periodic_task.periodic_task(spacing=CONF.replica_state_update_interval) @utils.require_driver_initialized @@ -2135,8 +2135,8 @@ class ShareManager(manager.SchedulerDependentManager): context, replica_list, share_replica, access_rules, available_share_snapshots, share_server=share_server) except Exception: - msg = _LE("Driver error when updating replica " - "state for replica %s.") + msg = ("Driver error when updating replica " + "state for replica %s.") LOG.exception(msg, share_replica['id']) self.db.share_replica_update( context, share_replica['id'], @@ -2150,8 +2150,8 @@ class ShareManager(manager.SchedulerDependentManager): self.db.share_replica_update(context, share_replica['id'], {'replica_state': replica_state}) elif replica_state: - msg = (_LW("Replica %(id)s cannot be set to %(state)s " - "through update call.") % + msg = (("Replica %(id)s cannot be set to %(state)s " + "through update call.") % {'id': share_replica['id'], 'state': replica_state}) LOG.warning(msg) @@ -2272,9 +2272,9 @@ class ShareManager(manager.SchedulerDependentManager): if not snapshot_update.get('size'): snapshot_update['size'] = snapshot_ref['share']['size'] - LOG.warning(_LW("Cannot get the size of the snapshot " - "%(snapshot_id)s. Using the size of " - "the share instead."), + LOG.warning("Cannot get the size of the snapshot " + "%(snapshot_id)s. Using the size of " + "the share instead.", {'snapshot_id': snapshot_id}) self._update_quota_usages(context, project_id, { @@ -2356,7 +2356,7 @@ class ShareManager(manager.SchedulerDependentManager): except exception.InvalidShare as e: share_manage_set_error_status( - _LE("Share can not be unmanaged: %s."), e) + ("Share can not be unmanaged: %s."), e) return try: @@ -2370,7 +2370,7 @@ class ShareManager(manager.SchedulerDependentManager): # Quota reservation errors here are not fatal, because # unmanage is administrator API and he/she could update user # quota usages later if it's required. - LOG.warning(_LW("Failed to update quota usages: %s."), e) + LOG.warning("Failed to update quota usages: %s.", e) if self.configuration.safe_get('unmanage_remove_access_rules'): try: @@ -2382,11 +2382,11 @@ class ShareManager(manager.SchedulerDependentManager): ) except Exception as e: share_manage_set_error_status( - _LE("Can not remove access rules of share: %s."), e) + ("Can not remove access rules of share: %s."), e) return self.db.share_instance_delete(context, share_instance['id']) - LOG.info(_LI("Share %s: unmanaged successfully."), share_id) + LOG.info("Share %s: unmanaged successfully.", share_id) @add_hooks @utils.require_driver_initialized @@ -2396,7 +2396,7 @@ class ShareManager(manager.SchedulerDependentManager): msg = _("Unmanage snapshot is not supported for " "driver_handles_share_servers=True mode.") self.db.share_snapshot_update(context, snapshot_id, status) - LOG.error(_LE("Share snapshot cannot be unmanaged: %s."), + LOG.error("Share snapshot cannot be unmanaged: %s.", msg) return @@ -2415,7 +2415,7 @@ class ShareManager(manager.SchedulerDependentManager): msg = _("Unmanage snapshot is not supported for " "share snapshots with share servers.") self.db.share_snapshot_update(context, snapshot_id, status) - LOG.error(_LE("Share snapshot cannot be unmanaged: %s."), + LOG.error("Share snapshot cannot be unmanaged: %s.", msg) return @@ -2428,7 +2428,7 @@ class ShareManager(manager.SchedulerDependentManager): share_server=share_server) except Exception: LOG.exception( - _LE("Cannot remove access rules of snapshot %s."), + ("Cannot remove access rules of snapshot %s."), snapshot_id) self.db.share_snapshot_update(context, snapshot_id, status) return @@ -2437,7 +2437,7 @@ class ShareManager(manager.SchedulerDependentManager): self.driver.unmanage_snapshot(snapshot_instance) except exception.UnmanageInvalidShareSnapshot as e: self.db.share_snapshot_update(context, snapshot_id, status) - LOG.error(_LE("Share snapshot cannot be unmanaged: %s."), e) + LOG.error("Share snapshot cannot be unmanaged: %s.", e) return try: @@ -2452,7 +2452,7 @@ class ShareManager(manager.SchedulerDependentManager): # Quota reservation errors here are not fatal, because # unmanage is administrator API and he/she could update user # quota usages later if it's required. - LOG.warning(_LW("Failed to update quota usages: %s."), e) + LOG.warning("Failed to update quota usages: %s.", e) self.db.share_snapshot_instance_delete( context, snapshot_instance['id']) @@ -2502,8 +2502,8 @@ class ShareManager(manager.SchedulerDependentManager): except Exception: with excutils.save_and_reraise_exception(): - msg = _LE('Share %(share)s could not be reverted ' - 'to snapshot %(snap)s.') + msg = ('Share %(share)s could not be reverted ' + 'to snapshot %(snap)s.') msg_args = {'share': share_id, 'snap': snapshot_id} LOG.exception(msg, msg_args) @@ -2529,8 +2529,8 @@ class ShareManager(manager.SchedulerDependentManager): self.db.share_snapshot_update( context, snapshot_id, {'status': constants.STATUS_AVAILABLE}) - msg = _LI('Share %(share)s reverted to snapshot %(snap)s ' - 'successfully.') + msg = ('Share %(share)s reverted to snapshot %(snap)s ' + 'successfully.') msg_args = {'share': share_id, 'snap': snapshot_id} LOG.info(msg, msg_args) @@ -2550,14 +2550,14 @@ class ShareManager(manager.SchedulerDependentManager): share_server=share_server ) except exception.ShareResourceNotFound: - LOG.warning(_LW("Share instance %s does not exist in the " - "backend."), share_instance_id) + LOG.warning("Share instance %s does not exist in the " + "backend.", share_instance_id) except Exception: with excutils.save_and_reraise_exception() as exc_context: if force: - msg = _LE("The driver was unable to delete access rules " - "for the instance: %s. Will attempt to delete " - "the instance anyway.") + msg = ("The driver was unable to delete access rules " + "for the instance: %s. Will attempt to delete " + "the instance anyway.") LOG.error(msg, share_instance_id) exc_context.reraise = False else: @@ -2570,16 +2570,16 @@ class ShareManager(manager.SchedulerDependentManager): self.driver.delete_share(context, share_instance, share_server=share_server) except exception.ShareResourceNotFound: - LOG.warning(_LW("Share instance %s does not exist in the " - "backend."), share_instance_id) + LOG.warning("Share instance %s does not exist in the " + "backend.", share_instance_id) except Exception: with excutils.save_and_reraise_exception() as exc_context: if force: - msg = _LE("The driver was unable to delete the share " - "instance: %s on the backend. Since this " - "operation is forced, the instance will be " - "deleted from Manila's database. A cleanup on " - "the backend may be necessary.") + msg = ("The driver was unable to delete the share " + "instance: %s on the backend. Since this " + "operation is forced, the instance will be " + "deleted from Manila's database. A cleanup on " + "the backend may be necessary.") LOG.error(msg, share_instance_id) exc_context.reraise = False else: @@ -2589,7 +2589,7 @@ class ShareManager(manager.SchedulerDependentManager): {'status': constants.STATUS_ERROR_DELETING}) self.db.share_instance_delete(context, share_instance_id) - LOG.info(_LI("Share instance %s: deleted successfully."), + LOG.info("Share instance %s: deleted successfully.", share_instance_id) self._check_delete_share_server(context, share_instance) @@ -2610,7 +2610,7 @@ class ShareManager(manager.SchedulerDependentManager): if not (self.driver.driver_handles_share_servers and self.configuration.automatic_share_server_cleanup): return - LOG.info(_LI("Check for unused share servers to delete.")) + LOG.info("Check for unused share servers to delete.") updated_before = timeutils.utcnow() - datetime.timedelta( minutes=self.configuration.unused_share_server_cleanup_interval) servers = self.db.share_server_get_all_unused_deletable(ctxt, @@ -2697,10 +2697,10 @@ class ShareManager(manager.SchedulerDependentManager): share_server=share_server) except Exception: LOG.exception( - _LE("Failed to remove access rules for snapshot %s."), + ("Failed to remove access rules for snapshot %s."), snapshot_instance['id']) - LOG.warning(_LW("The driver was unable to remove access rules " - "for snapshot %s. Moving on."), + LOG.warning("The driver was unable to remove access rules " + "for snapshot %s. Moving on.", snapshot_instance['snapshot_id']) try: @@ -2731,8 +2731,8 @@ class ShareManager(manager.SchedulerDependentManager): user_id=snapshot_ref['user_id']) except Exception: reservations = None - LOG.exception(_LE("Failed to update quota usages while deleting " - "snapshot %s."), snapshot_id) + LOG.exception("Failed to update quota usages while deleting " + "snapshot %s.", snapshot_id) if reservations: QUOTAS.commit(context, reservations, project_id=project_id, @@ -2833,8 +2833,8 @@ class ShareManager(manager.SchedulerDependentManager): except Exception: with excutils.save_and_reraise_exception(): - msg = _LE('Share %(share)s could not be reverted ' - 'to snapshot %(snap)s.') + msg = ('Share %(share)s could not be reverted ' + 'to snapshot %(snap)s.') msg_args = {'share': share_id, 'snap': snapshot_id} LOG.exception(msg, msg_args) @@ -2862,8 +2862,8 @@ class ShareManager(manager.SchedulerDependentManager): context, active_replica_snapshot['id'], {'status': constants.STATUS_AVAILABLE}) - msg = _LI('Share %(share)s reverted to snapshot %(snap)s ' - 'successfully.') + msg = ('Share %(share)s reverted to snapshot %(snap)s ' + 'successfully.') msg_args = {'share': share_id, 'snap': snapshot_id} LOG.info(msg, msg_args) @@ -3039,19 +3039,19 @@ class ShareManager(manager.SchedulerDependentManager): except exception.SnapshotResourceNotFound: if replica_snapshot['status'] == constants.STATUS_DELETING: - LOG.info(_LI('Snapshot %(snapshot_instance)s on replica ' - '%(replica)s has been deleted.'), msg_payload) + LOG.info('Snapshot %(snapshot_instance)s on replica ' + '%(replica)s has been deleted.', msg_payload) self.db.share_snapshot_instance_delete( context, replica_snapshot['id']) else: - LOG.exception(_LE("Replica snapshot %s was not found on " - "the backend."), replica_snapshot['id']) + LOG.exception("Replica snapshot %s was not found on " + "the backend.", replica_snapshot['id']) self.db.share_snapshot_instance_update( context, replica_snapshot['id'], {'status': constants.STATUS_ERROR}) except Exception: - LOG.exception(_LE("Driver error while updating replica snapshot: " - "%s"), replica_snapshot['id']) + LOG.exception("Driver error while updating replica snapshot: " + "%s", replica_snapshot['id']) self.db.share_snapshot_instance_update( context, replica_snapshot['id'], {'status': constants.STATUS_ERROR}) @@ -3081,7 +3081,7 @@ class ShareManager(manager.SchedulerDependentManager): @periodic_task.periodic_task(spacing=CONF.periodic_interval) @utils.require_driver_initialized def _report_driver_status(self, context): - LOG.info(_LI('Updating share status')) + LOG.info('Updating share status') share_stats = self.driver.get_share_stats(refresh=True) if not share_stats: return @@ -3309,7 +3309,7 @@ class ShareManager(manager.SchedulerDependentManager): except Exception: with excutils.save_and_reraise_exception(): LOG.error( - _LE("Share server '%s' failed on deletion."), + "Share server '%s' failed on deletion.", server_id) self.db.share_server_update( context, server_id, {'status': constants.STATUS_ERROR}) @@ -3318,7 +3318,7 @@ class ShareManager(manager.SchedulerDependentManager): _wrapped_delete_share_server() LOG.info( - _LI("Share server '%s' has been deleted successfully."), + "Share server '%s' has been deleted successfully.", share_server['id']) self.driver.deallocate_network(context, share_server['id']) @@ -3336,7 +3336,7 @@ class ShareManager(manager.SchedulerDependentManager): self.driver.extend_share( share_instance, new_size, share_server=share_server) except Exception as e: - LOG.exception(_LE("Extend share failed."), resource=share) + LOG.exception("Extend share failed.", resource=share) try: self.db.share_update( @@ -3363,7 +3363,7 @@ class ShareManager(manager.SchedulerDependentManager): } share = self.db.share_update(context, share['id'], share_update) - LOG.info(_LI("Extend share completed successfully."), resource=share) + LOG.info("Extend share completed successfully.", resource=share) @add_hooks @utils.require_driver_initialized @@ -3396,7 +3396,7 @@ class ShareManager(manager.SchedulerDependentManager): gigabytes=-size_decrease) except Exception as e: error_occurred( - e, _LE("Failed to update quota on share shrinking.")) + e, ("Failed to update quota on share shrinking.")) try: self.driver.shrink_share( @@ -3406,11 +3406,11 @@ class ShareManager(manager.SchedulerDependentManager): # shouldn't shrink share when this validation error occurs. except Exception as e: if isinstance(e, exception.ShareShrinkingPossibleDataLoss): - msg = _LE("Shrink share failed due to possible data loss.") + msg = ("Shrink share failed due to possible data loss.") status = constants.STATUS_SHRINKING_POSSIBLE_DATA_LOSS_ERROR error_params = {'msg': msg, 'status': status} else: - error_params = {'msg': _LE("Shrink share failed.")} + error_params = {'msg': ("Shrink share failed.")} try: error_occurred(e, **error_params) @@ -3427,7 +3427,7 @@ class ShareManager(manager.SchedulerDependentManager): } share = self.db.share_update(context, share['id'], share_update) - LOG.info(_LI("Shrink share completed successfully."), resource=share) + LOG.info("Shrink share completed successfully.", resource=share) @utils.require_driver_initialized def create_share_group(self, context, share_group_id): @@ -3479,15 +3479,15 @@ class ShareManager(manager.SchedulerDependentManager): ) except Exception: with excutils.save_and_reraise_exception(): - LOG.error(_LE("Failed to get share server" - " for share group creation.")) + LOG.error("Failed to get share server" + " for share group creation.") self.db.share_group_update( context, share_group_id, {'status': constants.STATUS_ERROR}) try: # TODO(ameade): Add notification for create.start - LOG.info(_LI("Share group %s: creating"), share_group_id) + LOG.info("Share group %s: creating", share_group_id) model_update, share_update_list = None, None @@ -3525,7 +3525,7 @@ class ShareManager(manager.SchedulerDependentManager): self.db.share_instance_update( context, share['id'], {'status': constants.STATUS_ERROR}) - LOG.error(_LE("Share group %s: create failed"), share_group_id) + LOG.error("Share group %s: create failed", share_group_id) now = timeutils.utcnow() for share in shares: @@ -3535,7 +3535,7 @@ class ShareManager(manager.SchedulerDependentManager): share_group_ref['id'], {'status': status, 'created_at': now}) - LOG.info(_LI("Share group %s: created successfully"), share_group_id) + LOG.info("Share group %s: created successfully", share_group_id) # TODO(ameade): Add notification for create.end @@ -3553,7 +3553,7 @@ class ShareManager(manager.SchedulerDependentManager): # TODO(ameade): Add notification for delete.start try: - LOG.info(_LI("Share group %s: deleting"), share_group_id) + LOG.info("Share group %s: deleting", share_group_id) share_server = None if share_group_ref.get('share_server_id'): share_server = self.db.share_server_get( @@ -3571,11 +3571,11 @@ class ShareManager(manager.SchedulerDependentManager): context, share_group_ref['id'], {'status': constants.STATUS_ERROR}) - LOG.error(_LE("Share group %s: delete failed"), + LOG.error("Share group %s: delete failed", share_group_ref['id']) self.db.share_group_destroy(context, share_group_id) - LOG.info(_LI("Share group %s: deleted successfully"), share_group_id) + LOG.info("Share group %s: deleted successfully", share_group_id) # TODO(ameade): Add notification for delete.end @@ -3594,7 +3594,7 @@ class ShareManager(manager.SchedulerDependentManager): updated_members_ids = [] try: - LOG.info(_LI("Share group snapshot %s: creating"), + LOG.info("Share group snapshot %s: creating", share_group_snapshot_id) share_server = None if snap_ref['share_group'].get('share_server_id'): @@ -3611,9 +3611,9 @@ class ShareManager(manager.SchedulerDependentManager): # to have here also 'export_locations' when it is supported. member_id = update.pop('id', None) if not member_id: - LOG.warning(_LW( + LOG.warning( "One of share group snapshot '%s' members does not " - "have reference ID. Its update was skipped."), + "have reference ID. Its update was skipped.", share_group_snapshot_id) continue # TODO(vponomaryov): remove following condition when @@ -3660,7 +3660,7 @@ class ShareManager(manager.SchedulerDependentManager): context, snap_ref['id'], {'status': constants.STATUS_ERROR}) - LOG.error(_LE("Share group snapshot %s: create failed"), + LOG.error("Share group snapshot %s: create failed", share_group_snapshot_id) for member in (snap_ref.get('share_group_snapshot_members') or []): @@ -3673,7 +3673,7 @@ class ShareManager(manager.SchedulerDependentManager): self.db.share_group_snapshot_update( context, snap_ref['id'], {'status': status, 'updated_at': now}) - LOG.info(_LI("Share group snapshot %s: created successfully"), + LOG.info("Share group snapshot %s: created successfully", share_group_snapshot_id) return snap_ref['id'] @@ -3691,7 +3691,7 @@ class ShareManager(manager.SchedulerDependentManager): snapshot_update = False try: - LOG.info(_LI("Share group snapshot %s: deleting"), + LOG.info("Share group snapshot %s: deleting", share_group_snapshot_id) share_server = None @@ -3719,12 +3719,12 @@ class ShareManager(manager.SchedulerDependentManager): context, snap_ref['id'], {'status': constants.STATUS_ERROR}) - LOG.error(_LE("Share group snapshot %s: delete failed"), + LOG.error("Share group snapshot %s: delete failed", snap_ref['name']) self.db.share_group_snapshot_destroy(context, share_group_snapshot_id) - LOG.info(_LI("Share group snapshot %s: deleted successfully"), + LOG.info("Share group snapshot %s: deleted successfully", share_group_snapshot_id) def _get_share_replica_dict(self, context, share_replica): diff --git a/manila/share/migration.py b/manila/share/migration.py index f39ab05a9e..d579f8e668 100644 --- a/manila/share/migration.py +++ b/manila/share/migration.py @@ -21,7 +21,7 @@ from oslo_log import log from manila.common import constants from manila import exception -from manila.i18n import _, _LW +from manila.i18n import _ from manila.share import api as share_api import manila.utils as utils @@ -130,16 +130,16 @@ class ShareMigrationHelper(object): try: self.delete_instance_and_wait(new_instance) except Exception: - LOG.warning(_LW("Failed to cleanup new instance during generic" - " migration for share %s."), self.share['id']) + LOG.warning("Failed to cleanup new instance during generic" + " migration for share %s.", self.share['id']) def cleanup_access_rules(self, share_instance, share_server): try: self.revert_access_rules(share_instance, share_server) except Exception: - LOG.warning(_LW("Failed to cleanup access rules during generic" - " migration for share %s."), self.share['id']) + LOG.warning("Failed to cleanup access rules during generic" + " migration for share %s.", self.share['id']) def revert_access_rules(self, share_instance, share_server): diff --git a/manila/share/share_types.py b/manila/share/share_types.py index 137502e2ed..e8a0015510 100644 --- a/manila/share/share_types.py +++ b/manila/share/share_types.py @@ -29,7 +29,6 @@ from manila import context from manila import db from manila import exception from manila.i18n import _ -from manila.i18n import _LE CONF = cfg.CONF LOG = log.getLogger(__name__) @@ -53,7 +52,7 @@ def create(context, name, extra_specs=None, is_public=True, projects=None): is_public=is_public), projects=projects) except db_exception.DBError: - LOG.exception(_LE('DB error.')) + LOG.exception('DB error.') raise exception.ShareTypeCreateFailed(name=name, extra_specs=extra_specs) return type_ref @@ -86,8 +85,8 @@ def get_all_types(context, inactive=0, search_opts=None): required_extra_specs = get_valid_required_extra_specs( type_args['extra_specs']) except exception.InvalidExtraSpec: - LOG.exception(_LE('Share type %(share_type)s has invalid required' - ' extra specs.'), {'share_type': type_name}) + LOG.exception('Share type %(share_type)s has invalid required' + ' extra specs.', {'share_type': type_name}) type_args['required_extra_specs'] = required_extra_specs @@ -172,8 +171,8 @@ def get_default_share_type(ctxt=None): # Couldn't find share type with the name in default_share_type # flag, record this issue and move on # TODO(zhiteng) consider add notification to warn admin - LOG.exception(_LE('Default share type is not found, ' - 'please check default_share_type config: %s'), + LOG.exception('Default share type is not found, ' + 'please check default_share_type config: %s', e) diff --git a/manila/share/snapshot_access.py b/manila/share/snapshot_access.py index da34d2ad78..c5bdef799d 100644 --- a/manila/share/snapshot_access.py +++ b/manila/share/snapshot_access.py @@ -16,7 +16,6 @@ from oslo_log import log from manila.common import constants -from manila.i18n import _LI from manila import utils LOG = log.getLogger(__name__) @@ -147,8 +146,8 @@ class ShareSnapshotInstanceAccess(object): self._update_access_rules(context, snapshot_instance, share_server=share_server) else: - LOG.info(_LI("Access rules were successfully applied for " - "snapshot instance: %s"), snapshot_instance['id']) + LOG.info("Access rules were successfully applied for " + "snapshot instance: %s", snapshot_instance['id']) def _check_needs_refresh(self, context, snapshot_instance_id): diff --git a/manila/share_group/share_group_types.py b/manila/share_group/share_group_types.py index d6ec36d872..84c5af7add 100644 --- a/manila/share_group/share_group_types.py +++ b/manila/share_group/share_group_types.py @@ -20,7 +20,6 @@ from manila import context from manila import db from manila import exception from manila.i18n import _ -from manila.i18n import _LE CONF = cfg.CONF LOG = log.getLogger(__name__) @@ -38,7 +37,7 @@ def create(context, name, share_types, group_specs=None, is_public=True, "share_types": share_types}, projects=projects) except db_exception.DBError: - LOG.exception(_LE('DB error')) + LOG.exception('DB error') raise exception.ShareGroupTypeCreateFailed( name=name, group_specs=group_specs) return type_ref @@ -142,8 +141,8 @@ def get_default(ctxt=None): return get_by_name(ctxt, name) except exception.ShareGroupTypeNotFoundByName: LOG.exception( - _LE("Default share group type '%s' is not found, " - "please check 'default_share_group_type' config."), + "Default share group type '%s' is not found, " + "please check 'default_share_group_type' config.", name, )