Preparing for Pike support

1. Get rid of log translation (since oslo doesnt support it
any more)

2. Use service based class from neutron_lib

Change-Id: I15cf50984313bc217a4b38707c19b635e4eca039
This commit is contained in:
Sumit Naiksatam 2017-10-24 13:52:55 -07:00
parent 42ccdb6e0d
commit 599f9c70e2
62 changed files with 787 additions and 901 deletions

View File

@ -27,16 +27,6 @@ _C = _translators.contextual_form
# The plural translation function using the name "_P"
_P = _translators.plural_form
# Translators for log levels.
#
# The abbreviated names are meant to reflect the usual use of a short
# name like '_'. The "L" is for "log" and the other letter comes from
# the level.
_LI = _translators.log_info
_LW = _translators.log_warning
_LE = _translators.log_error
_LC = _translators.log_critical
def get_available_languages():
return oslo_i18n.get_available_languages(DOMAIN)

View File

@ -17,7 +17,6 @@ from oslo_utils import importutils
from stevedore import driver
from gbpservice._i18n import _
from gbpservice._i18n import _LE
LOG = logging.getLogger(__name__)
cfg.CONF.import_group('keystone_authtoken', 'keystonemiddleware.auth_token')
@ -42,8 +41,8 @@ def load_plugin(namespace, plugin):
try:
plugin_class = importutils.import_class(plugin)
except ImportError as e2:
LOG.exception(_LE("Error loading plugin by name, %s"), e1)
LOG.exception(_LE("Error loading plugin by class, %s"), e2)
LOG.exception("Error loading plugin by name, %s", e1)
LOG.exception("Error loading plugin by class, %s", e2)
raise ImportError(_("Plugin not found."))
return plugin_class()

View File

@ -10,7 +10,6 @@
# License for the specific language governing permissions and limitations
# under the License.
from gbpservice._i18n import _LE
from gbpservice.contrib.nfp.config_orchestrator.common import (
topics as a_topics)
from gbpservice.nfp.core import log as nfp_logging
@ -115,7 +114,7 @@ def get_dhcp_agent_host(config):
if agents:
return agents[0].get('host', None)
except Exception as exc:
LOG.error(_LE("Failed to get dhcp agent host : %(exc)s"),
LOG.error("Failed to get dhcp agent host : %(exc)s",
{'exc': exc})
@ -168,8 +167,8 @@ def get_network_function_details(context, network_function_id):
return network_function_details['network_function']
except Exception as e:
LOG.error(_LE("Failed to get network function details of "
"network_function_id %(network_function_id)s : %(ex)s "),
LOG.error("Failed to get network function details of "
"network_function_id %(network_function_id)s : %(ex)s ",
{'ex': e, 'network_function_id': network_function_id})
@ -193,7 +192,7 @@ def get_network_function_map(context, network_function_id):
LOG.debug(msg)
return request_data
except Exception as e:
LOG.error(_LE("Failed to get network function map of "
"network_function_id %(network_function_id)s : %(ex)s "),
LOG.error("Failed to get network function map of "
"network_function_id %(network_function_id)s : %(ex)s ",
{'ex': e, 'network_function_id': network_function_id})
return request_data

View File

@ -13,7 +13,6 @@
import ast
import copy
from gbpservice._i18n import _LI
from gbpservice.contrib.nfp.config_orchestrator.common import common
from gbpservice.nfp.common import constants as const
from gbpservice.nfp.common import data_formatter as df
@ -158,8 +157,8 @@ class FwAgent(firewall_db.Firewall_db_mixin):
nf_id = self._fetch_nf_from_resource_desc(firewall["description"])
nfp_context['log_context']['meta_id'] = nf_id
nf = common.get_network_function_details(context, nf_id)
LOG.info(_LI("Received RPC CREATE FIREWALL for "
"Firewall: %(firewall)s"),
LOG.info("Received RPC CREATE FIREWALL for "
"Firewall: %(firewall)s",
{'firewall': firewall})
body = self._data_wrapper(context, firewall, host, nf, 'CREATE')
transport.send_request_to_configurator(self._conf,
@ -172,8 +171,8 @@ class FwAgent(firewall_db.Firewall_db_mixin):
nf_id = self._fetch_nf_from_resource_desc(firewall["description"])
nfp_context['log_context']['meta_id'] = nf_id
nf = common.get_network_function_details(context, nf_id)
LOG.info(_LI("Received RPC DELETE FIREWALL for "
"Firewall: %(firewall)s"),
LOG.info("Received RPC DELETE FIREWALL for "
"Firewall: %(firewall)s",
{'firewall': firewall})
body = self._data_wrapper(context, firewall, host, nf, 'DELETE')
transport.send_request_to_configurator(self._conf,

View File

@ -13,7 +13,6 @@
import ast
import copy
from gbpservice._i18n import _LI
from gbpservice.contrib.nfp.config_orchestrator.common import common
from gbpservice.contrib.nfp.config_orchestrator.common import lbv2_constants
from gbpservice.nfp.common import constants as const
@ -277,7 +276,7 @@ class Lbv2Agent(loadbalancer_dbv2.LoadBalancerPluginDbv2):
def create_loadbalancer(self, context, loadbalancer, driver_name,
allocate_vip=True):
nfp_context = module_context.init()
LOG.info(_LI("Received RPC CREATE LOADBALANCER for LB:%(lb)s"),
LOG.info("Received RPC CREATE LOADBALANCER for LB:%(lb)s",
{'lb': loadbalancer})
# Fetch nf_id from description of the resource
nf_id = self._fetch_nf_from_resource_desc(loadbalancer["description"])
@ -306,8 +305,8 @@ class Lbv2Agent(loadbalancer_dbv2.LoadBalancerPluginDbv2):
def delete_loadbalancer(self, context, loadbalancer,
delete_vip_port=True):
nfp_context = module_context.init()
LOG.info(_LI("Received RPC DELETE LOADBALANCER for LB:"
"%(lb)s"), {'lb': loadbalancer})
LOG.info("Received RPC DELETE LOADBALANCER for LB:"
"%(lb)s", {'lb': loadbalancer})
# Fetch nf_id from description of the resource
nf_id = self._fetch_nf_from_resource_desc(loadbalancer["description"])
nfp_context['log_context']['meta_id'] = nf_id
@ -320,7 +319,7 @@ class Lbv2Agent(loadbalancer_dbv2.LoadBalancerPluginDbv2):
@log_helpers.log_method_call
def create_listener(self, context, listener):
nfp_context = module_context.init()
LOG.info(_LI("Received RPC CREATE LISTENER for Listener:%(listener)s"),
LOG.info("Received RPC CREATE LISTENER for Listener:%(listener)s",
{'listener': listener})
loadbalancer = listener['loadbalancer']
# Fetch nf_id from description of the resource
@ -348,7 +347,7 @@ class Lbv2Agent(loadbalancer_dbv2.LoadBalancerPluginDbv2):
@log_helpers.log_method_call
def delete_listener(self, context, listener):
nfp_context = module_context.init()
LOG.info(_LI("Received RPC DELETE LISTENER for Listener:%(listener)s"),
LOG.info("Received RPC DELETE LISTENER for Listener:%(listener)s",
{'listener': listener})
loadbalancer = listener['loadbalancer']
# Fetch nf_id from description of the resource
@ -363,7 +362,7 @@ class Lbv2Agent(loadbalancer_dbv2.LoadBalancerPluginDbv2):
@log_helpers.log_method_call
def create_pool(self, context, pool):
nfp_context = module_context.init()
LOG.info(_LI("Received RPC CREATE POOL for Pool:%(pool)s"),
LOG.info("Received RPC CREATE POOL for Pool:%(pool)s",
{'pool': pool})
loadbalancer = pool['loadbalancer']
# Fetch nf_id from description of the resource
@ -391,7 +390,7 @@ class Lbv2Agent(loadbalancer_dbv2.LoadBalancerPluginDbv2):
@log_helpers.log_method_call
def delete_pool(self, context, pool):
nfp_context = module_context.init()
LOG.info(_LI("Received RPC DELETE POOL for Pool:%(pool)s"),
LOG.info("Received RPC DELETE POOL for Pool:%(pool)s",
{'pool': pool})
loadbalancer = pool['loadbalancer']
# Fetch nf_id from description of the resource
@ -406,7 +405,7 @@ class Lbv2Agent(loadbalancer_dbv2.LoadBalancerPluginDbv2):
@log_helpers.log_method_call
def create_member(self, context, member):
nfp_context = module_context.init()
LOG.info(_LI("Received RPC CREATE MEMBER for Member:%(member)s"),
LOG.info("Received RPC CREATE MEMBER for Member:%(member)s",
{'member': member})
loadbalancer = member['pool']['loadbalancer']
# Fetch nf_id from description of the resource
@ -434,7 +433,7 @@ class Lbv2Agent(loadbalancer_dbv2.LoadBalancerPluginDbv2):
@log_helpers.log_method_call
def delete_member(self, context, member):
nfp_context = module_context.init()
LOG.info(_LI("Received RPC DELETE MEMBER for Member:%(member)s"),
LOG.info("Received RPC DELETE MEMBER for Member:%(member)s",
{'member': member})
loadbalancer = member['pool']['loadbalancer']
# Fetch nf_id from description of the resource
@ -449,7 +448,7 @@ class Lbv2Agent(loadbalancer_dbv2.LoadBalancerPluginDbv2):
@log_helpers.log_method_call
def create_healthmonitor(self, context, healthmonitor):
nfp_context = module_context.init()
LOG.info(_LI("Received RPC CREATE HEALTH MONITOR for HM:%(hm)s"),
LOG.info("Received RPC CREATE HEALTH MONITOR for HM:%(hm)s",
{'hm': healthmonitor})
loadbalancer = healthmonitor['pool']['loadbalancer']
# Fetch nf_id from description of the resource
@ -478,7 +477,7 @@ class Lbv2Agent(loadbalancer_dbv2.LoadBalancerPluginDbv2):
@log_helpers.log_method_call
def delete_healthmonitor(self, context, healthmonitor):
nfp_context = module_context.init()
LOG.info(_LI("Received RPC DELETE HEALTH MONITOR for HM:%(hm)s"),
LOG.info("Received RPC DELETE HEALTH MONITOR for HM:%(hm)s",
{'hm': healthmonitor})
loadbalancer = healthmonitor['pool']['loadbalancer']
# Fetch nf_id from description of the resource

View File

@ -13,7 +13,6 @@
import ast
import copy
from gbpservice._i18n import _LI
from gbpservice.contrib.nfp.config_orchestrator.common import common
from gbpservice.nfp.common import constants as const
from gbpservice.nfp.common import data_formatter as df
@ -149,7 +148,7 @@ class VpnAgent(vpn_db.VPNPluginDb, vpn_db.VPNPluginRpcDbMixin):
@log_helpers.log_method_call
def vpnservice_updated(self, context, **kwargs):
nfp_context = module_context.init()
LOG.info(_LI("Received RPC VPN SERVICE UPDATED with data:%(data)s"),
LOG.info("Received RPC VPN SERVICE UPDATED with data:%(data)s",
{'data': kwargs})
# Fetch nf_id from description of the resource
nf_id = self._fetch_nf_from_resource_desc(kwargs[

View File

@ -13,7 +13,6 @@
import sys
import traceback
from gbpservice._i18n import _LI
from gbpservice.contrib.nfp.config_orchestrator.common import (
lbv2_constants as lbv2_const)
from gbpservice.contrib.nfp.config_orchestrator.common import (
@ -41,8 +40,8 @@ class RpcHandler(object):
def network_function_notification(self, context, notification_data):
module_context.init()
try:
LOG.info(_LI("Received NETWORK FUNCTION NOTIFICATION:"
"%(notification)s"),
LOG.info("Received NETWORK FUNCTION NOTIFICATION:"
"%(notification)s",
{'notification': notification_data['notification']})
if notification_data['info']['service_type'] is not None:
handler = NaasNotificationHandler(self.conf, self.sc)
@ -78,9 +77,9 @@ class FirewallNotifier(object):
firewall_id = resource_data['firewall_id']
status = resource_data['status']
LOG.info(_LI("Received firewall configuration create complete API, "
"making an RPC call set firewall status for "
"firewall:%(firewall)s and status: %(status)s"),
LOG.info("Received firewall configuration create complete API, "
"making an RPC call set firewall status for "
"firewall:%(firewall)s and status: %(status)s",
{'firewall': firewall_id,
'status': status})
@ -103,9 +102,9 @@ class FirewallNotifier(object):
resource_data = notification['data']
firewall_id = resource_data['firewall_id']
LOG.info(_LI("Received firewall_configuration_delete_complete API, "
"making an RPC call firewall_deleted for firewall:"
"%(firewall)s "),
LOG.info("Received firewall_configuration_delete_complete API, "
"making an RPC call firewall_deleted for firewall:"
"%(firewall)s ",
{'firewall': firewall_id})
# RPC call to plugin to update firewall deleted
@ -143,9 +142,9 @@ class LoadbalancerV2Notifier(object):
obj_p_status = resource_data['provisioning_status']
obj_o_status = resource_data['operating_status']
LOG.info(_LI("Received LB's update_status API. Making an "
"update_status RPC call to plugin for %(obj_type)s:"
"%(obj_id)s with status: %(status)s"),
LOG.info("Received LB's update_status API. Making an "
"update_status RPC call to plugin for %(obj_type)s:"
"%(obj_id)s with status: %(status)s",
{'obj_type': obj_type,
'obj_id': obj_id,
'status': obj_p_status})
@ -192,9 +191,9 @@ class VpnNotifier(object):
nfp_context['log_context'] = logging_context
status = resource_data['status']
LOG.info(_LI("Received VPN's update_status API. "
"Making an update_status RPC cast to plugin for object"
"with status: %(status)s"),
LOG.info("Received VPN's update_status API. "
"Making an update_status RPC cast to plugin for object"
"with status: %(status)s",
{'status': status})
rpcClient = transport.RPCClient(a_topics.VPN_NFP_PLUGIN_TOPIC)
rpcClient.cctxt.cast(context, 'update_status',

View File

@ -11,7 +11,6 @@
# under the License.
from gbpservice._i18n import _LI
from gbpservice.contrib.nfp.configurator.lib import constants as const
from gbpservice.nfp.core import log as nfp_logging
from gbpservice.nfp.core import module as nfp_api
@ -105,7 +104,7 @@ class AgentBaseRPCManager(object):
# Multiple request data blobs needs batch processing. Send batch
# processing event or do direct processing of single request data blob
if (len(sa_req_list) > 1):
LOG.info(_LI("Creating event PROCESS BATCH"))
LOG.info("Creating event PROCESS BATCH")
args_dict = {
'sa_req_list': sa_req_list,
'notification_data': notification_data

View File

@ -16,7 +16,6 @@ import oslo_messaging as messaging
import requests
import six
from gbpservice._i18n import _LI
from gbpservice.contrib.nfp.configurator.agents import agent_base
from gbpservice.contrib.nfp.configurator.lib import constants as common_const
from gbpservice.contrib.nfp.configurator.lib import fw_constants as const
@ -62,9 +61,9 @@ class FwaasRpcSender(agent_base.AgentBaseEventHandler):
'notification_type': (
'set_firewall_status')}}]
}
LOG.info(_LI("Sending Notification 'Set Firewall Status' to "
"Orchestrator for firewall: %(fw_id)s with status:"
"%(status)s"),
LOG.info("Sending Notification 'Set Firewall Status' to "
"Orchestrator for firewall: %(fw_id)s with status:"
"%(status)s",
{'fw_id': firewall_id,
'status': status})
self.notify._notification(msg)
@ -86,8 +85,8 @@ class FwaasRpcSender(agent_base.AgentBaseEventHandler):
'notification_type': (
'firewall_deleted')}}]
}
LOG.info(_LI("Sending Notification 'Firewall Deleted' to "
"Orchestrator for firewall: %(fw_id)s "),
LOG.info("Sending Notification 'Firewall Deleted' to "
"Orchestrator for firewall: %(fw_id)s ",
{'fw_id': firewall_id})
self.notify._notification(msg)
@ -153,7 +152,7 @@ class FWaasRpcManager(agent_base.AgentBaseRPCManager):
"""
LOG.info(_LI("Received request 'Create Firewall'."))
LOG.info("Received request 'Create Firewall'.")
self._create_event(context, firewall,
host, const.FIREWALL_CREATE_EVENT)
@ -161,7 +160,7 @@ class FWaasRpcManager(agent_base.AgentBaseRPCManager):
""" Receives request to update firewall from configurator
"""
LOG.info(_LI("Received request 'Update Firewall'."))
LOG.info("Received request 'Update Firewall'.")
self._create_event(context, firewall,
host, const.FIREWALL_UPDATE_EVENT)
@ -169,7 +168,7 @@ class FWaasRpcManager(agent_base.AgentBaseRPCManager):
""" Receives request to delete firewall from configurator
"""
LOG.info(_LI("Received request 'Delete Firewall'."))
LOG.info("Received request 'Delete Firewall'.")
self._create_event(context, firewall,
host, const.FIREWALL_DELETE_EVENT)
@ -256,8 +255,8 @@ class FWaasEventHandler(nfp_api.NfpEventHandler):
service_vendor = agent_info['service_vendor']
service_feature = agent_info.get('service_feature', '')
driver = self._get_driver(service_vendor, service_feature)
LOG.info(_LI("Invoking driver with service vendor:"
"%(service_vendor)s "),
LOG.info("Invoking driver with service vendor:"
"%(service_vendor)s ",
{'service_vendor': service_vendor})
self.method = getattr(driver, "%s" % (ev.id.lower()))
self.invoke_driver_for_plugin_api(ev)
@ -435,7 +434,7 @@ def load_drivers(conf):
driver_obj = driver_name(conf=conf)
drivers[service_type] = driver_obj
LOG.info(_LI("Firewall loaded drivers:%(drivers)s"),
LOG.info("Firewall loaded drivers:%(drivers)s",
{'drivers': drivers})
return drivers

View File

@ -14,7 +14,6 @@ import copy
import os
import six
from gbpservice._i18n import _LI
from gbpservice.contrib.nfp.configurator.agents import agent_base
from gbpservice.contrib.nfp.configurator.lib import (
generic_config_constants as gen_cfg_const)
@ -142,8 +141,8 @@ class GenericConfigRpcManager(agent_base.AgentBaseRPCManager):
Returns: None
"""
LOG.info(_LI("Received configure health monitor api for nfds:"
"%(nfds)s"),
LOG.info("Received configure health monitor api for nfds:"
"%(nfds)s",
{'nfds': resource_data['nfds']})
resource_data['fail_count'] = 0
self._send_event(context,
@ -161,8 +160,8 @@ class GenericConfigRpcManager(agent_base.AgentBaseRPCManager):
Returns: None
"""
LOG.info(_LI("Received clear health monitor api for nfds:"
"%(nfds)s"),
LOG.info("Received clear health monitor api for nfds:"
"%(nfds)s",
{'nfds': resource_data['nfds']})
event_key = resource_data['nfds'][0]['vmid']
poll_event_id = gen_cfg_const.EVENT_CONFIGURE_HEALTHMONITOR
@ -470,8 +469,8 @@ def load_drivers(conf):
for service_type, driver_name in six.iteritems(drivers):
driver_obj = driver_name(conf=conf)
drivers[service_type] = driver_obj
LOG.info(_LI("Generic config agent loaded drivers drivers:"
"%(drivers)s"),
LOG.info("Generic config agent loaded drivers drivers:"
"%(drivers)s",
{'drivers': drivers})
return drivers

View File

@ -13,7 +13,6 @@
import os
import six
from gbpservice._i18n import _LI
from gbpservice.contrib.nfp.configurator.agents import agent_base
from gbpservice.contrib.nfp.configurator.lib import data_filter
from gbpservice.contrib.nfp.configurator.lib import lbv2_constants as lb_const
@ -64,9 +63,9 @@ class LBaaSV2RpcSender(data_filter.Filter):
operating_status,
obj_type: obj}}]
}
LOG.info(_LI("Sending Notification 'Update Status' "
"for resource: %(resource)s with Provisioning status:"
"%(p_status)s and Operating status:%(o_status)s"),
LOG.info("Sending Notification 'Update Status' "
"for resource: %(resource)s with Provisioning status:"
"%(p_status)s and Operating status:%(o_status)s",
{'resource': agent_info['resource'],
'p_status': provisioning_status,
'o_status': operating_status})
@ -90,8 +89,8 @@ class LBaaSV2RpcSender(data_filter.Filter):
'update_pool_stats'),
'pool': pool_id}}]
}
LOG.info(_LI("Sending Notification 'Update Pool Stats' "
"for pool: %(pool_id)s with stats:%(stats)s"),
LOG.info("Sending Notification 'Update Pool Stats' "
"for pool: %(pool_id)s with stats:%(stats)s",
{'pool_id': pool_id,
'stats': stats})
self.notify._notification(msg)
@ -149,8 +148,8 @@ class LBaaSv2RpcManager(agent_base.AgentBaseRPCManager):
Returns: None
"""
LOG.info(_LI("Received request 'Create Loadbalancer' for LB:%(lb)s "
"with driver:%(driver_name)s"),
LOG.info("Received request 'Create Loadbalancer' for LB:%(lb)s "
"with driver:%(driver_name)s",
{'lb': loadbalancer['id'],
'driver_name': driver_name})
arg_dict = {'context': context,
@ -177,8 +176,8 @@ class LBaaSv2RpcManager(agent_base.AgentBaseRPCManager):
lb_const.OLD_LOADBALANCER: old_loadbalancer,
lb_const.LOADBALANCER: loadbalancer,
}
LOG.info(_LI("Received request 'Update Loadbalancer' for LB:%(lb)s "
"with new Param:%(new_val)s and old Param:%(old_val)s"),
LOG.info("Received request 'Update Loadbalancer' for LB:%(lb)s "
"with new Param:%(new_val)s and old Param:%(old_val)s",
{'lb': loadbalancer['id'],
'new_val': new_val,
'old_val': old_val})
@ -195,7 +194,7 @@ class LBaaSv2RpcManager(agent_base.AgentBaseRPCManager):
Returns: None
"""
LOG.info(_LI("Received request 'Delete Loadbalancer' for LB:%(lb)s "),
LOG.info("Received request 'Delete Loadbalancer' for LB:%(lb)s ",
{'lb': loadbalancer['id']})
arg_dict = {'context': context,
@ -214,7 +213,7 @@ class LBaaSv2RpcManager(agent_base.AgentBaseRPCManager):
Returns: None
"""
LOG.info(_LI("Received request 'Create Listener' for LB:%(lb)s "),
LOG.info("Received request 'Create Listener' for LB:%(lb)s ",
{'lb': listener['loadbalancer_id']})
arg_dict = {'context': context,
lb_const.LISTENER: listener,
@ -235,9 +234,9 @@ class LBaaSv2RpcManager(agent_base.AgentBaseRPCManager):
"""
old_val, new_val = self.get_diff_of_dict(old_listener, listener)
LOG.info(_LI("Received request 'Update Listener' for Listener:"
"%(listener)s in LB:%(lb_id)s with new Param:"
"%(new_val)s and old Param:%(old_val)s"),
LOG.info("Received request 'Update Listener' for Listener:"
"%(listener)s in LB:%(lb_id)s with new Param:"
"%(new_val)s and old Param:%(old_val)s",
{'lb_id': listener['loadbalancer_id'],
'listener': listener['id'],
'old_val': old_val,
@ -260,7 +259,7 @@ class LBaaSv2RpcManager(agent_base.AgentBaseRPCManager):
Returns: None
"""
LOG.info(_LI("Received request 'Delete Listener' for LB:%(lb)s "),
LOG.info("Received request 'Delete Listener' for LB:%(lb)s ",
{'lb': listener['loadbalancer_id']})
arg_dict = {'context': context,
lb_const.LISTENER: listener,
@ -279,7 +278,7 @@ class LBaaSv2RpcManager(agent_base.AgentBaseRPCManager):
Returns: None
"""
LOG.info(_LI("Received request 'Create Pool' for Pool:%(pool_id)s "),
LOG.info("Received request 'Create Pool' for Pool:%(pool_id)s ",
{'pool_id': pool['id']})
arg_dict = {'context': context,
lb_const.POOL: pool
@ -301,9 +300,9 @@ class LBaaSv2RpcManager(agent_base.AgentBaseRPCManager):
"""
old_val, new_val = self.get_diff_of_dict(old_pool, pool)
LOG.info(_LI("Received request 'Update Pool' for Pool:%(pool)s "
"in LB:%(lb_id)s with new Param:%(new_val)s and "
"old Param:%(old_val)s"),
LOG.info("Received request 'Update Pool' for Pool:%(pool)s "
"in LB:%(lb_id)s with new Param:%(new_val)s and "
"old Param:%(old_val)s",
{'pool': pool['id'],
'lb_id': pool['loadbalancer_id'],
'old_val': old_val,
@ -326,7 +325,7 @@ class LBaaSv2RpcManager(agent_base.AgentBaseRPCManager):
Returns: None
"""
LOG.info(_LI("Received request 'Delete Pool' for Pool:%(pool_id)s "),
LOG.info("Received request 'Delete Pool' for Pool:%(pool_id)s ",
{'pool_id': pool['id']})
arg_dict = {'context': context,
lb_const.POOL: pool,
@ -345,7 +344,7 @@ class LBaaSv2RpcManager(agent_base.AgentBaseRPCManager):
Returns: None
"""
LOG.info(_LI("Received request 'Create Member' for Pool:%(pool_id)s "),
LOG.info("Received request 'Create Member' for Pool:%(pool_id)s ",
{'pool_id': member['pool_id']})
arg_dict = {'context': context,
lb_const.MEMBER: member,
@ -366,9 +365,9 @@ class LBaaSv2RpcManager(agent_base.AgentBaseRPCManager):
"""
old_val, new_val = self.get_diff_of_dict(old_member, member)
LOG.info(_LI("Received request 'Update Member' for Member:"
"%(member_id)s in Pool:%(pool_id)s with new Param:"
"%(new_val)s and old Param:%(old_val)s"),
LOG.info("Received request 'Update Member' for Member:"
"%(member_id)s in Pool:%(pool_id)s with new Param:"
"%(new_val)s and old Param:%(old_val)s",
{'pool_id': member['pool_id'],
'member_id': member['id'],
'old_val': old_val,
@ -391,8 +390,8 @@ class LBaaSv2RpcManager(agent_base.AgentBaseRPCManager):
Returns: None
"""
LOG.info(_LI("Received request 'Delete Member' for Pool:"
"%(pool_id)s "),
LOG.info("Received request 'Delete Member' for Pool:"
"%(pool_id)s ",
{'pool_id': member['pool_id']})
arg_dict = {'context': context,
lb_const.MEMBER: member,
@ -412,8 +411,8 @@ class LBaaSv2RpcManager(agent_base.AgentBaseRPCManager):
Returns: None
"""
LOG.info(_LI("Received request 'Create Pool Health Monitor' for"
"Health monitor:%(hm)s"),
LOG.info("Received request 'Create Pool Health Monitor' for"
"Health monitor:%(hm)s",
{'hm': healthmonitor['id']})
arg_dict = {'context': context,
lb_const.HEALTHMONITOR: healthmonitor
@ -437,9 +436,9 @@ class LBaaSv2RpcManager(agent_base.AgentBaseRPCManager):
"""
old_val, new_val = self.get_diff_of_dict(
old_healthmonitor, healthmonitor)
LOG.info(_LI("Received request 'Update Pool Health Monitor' for "
"Health monitor:%(hm)s with new Param:%(new_val)s and "
"old Param:%(old_val)s"),
LOG.info("Received request 'Update Pool Health Monitor' for "
"Health monitor:%(hm)s with new Param:%(new_val)s and "
"old Param:%(old_val)s",
{'hm': healthmonitor['id'],
'old_val': old_val,
'new_val': new_val})
@ -463,8 +462,8 @@ class LBaaSv2RpcManager(agent_base.AgentBaseRPCManager):
Returns: None
"""
LOG.info(_LI("Received request 'Delete Pool Health Monitor' for "
"Health monitor:%(hm)s"),
LOG.info("Received request 'Delete Pool Health Monitor' for "
"Health monitor:%(hm)s",
{'hm': healthmonitor['id']})
arg_dict = {'context': context,
lb_const.HEALTHMONITOR: healthmonitor
@ -484,7 +483,7 @@ class LBaaSv2RpcManager(agent_base.AgentBaseRPCManager):
Returns: None
"""
LOG.info(_LI("Received request 'Agent Updated' "))
LOG.info("Received request 'Agent Updated' ")
arg_dict = {'context': context,
'payload': payload}
self._send_event(lb_const.EVENT_AGENT_UPDATED_V2, arg_dict)

View File

@ -14,7 +14,6 @@
import os
import six
from gbpservice._i18n import _LI
from gbpservice.contrib.nfp.configurator.agents import agent_base
from gbpservice.contrib.nfp.configurator.drivers.base import base_driver
from gbpservice.contrib.nfp.configurator.lib import data_filter
@ -57,7 +56,7 @@ class VpnaasRpcSender(data_filter.Filter):
Returns: Dictionary of vpn service type which matches with the filters.
"""
LOG.info(_LI("Sending RPC for GET VPN SERVICES with %(filters)s "),
LOG.info("Sending RPC for GET VPN SERVICES with %(filters)s ",
{'filters': filters})
return self.call(
context,
@ -75,8 +74,8 @@ class VpnaasRpcSender(data_filter.Filter):
Returns: dictionary of vpnservice
"""
LOG.info(_LI("Sending RPC for GET VPN SERVICECONTEXT with "
"Filters:%(filters)s "),
LOG.info("Sending RPC for GET VPN SERVICECONTEXT with "
"Filters:%(filters)s ",
{'filters': filters})
return self.call(
context,
@ -88,8 +87,8 @@ class VpnaasRpcSender(data_filter.Filter):
Get list of ipsec conns with filters
specified.
"""
LOG.info(_LI("Sending RPC for GET IPSEC CONNS with Filters:"
"%(filters)s "),
LOG.info("Sending RPC for GET IPSEC CONNS with Filters:"
"%(filters)s ",
{'filters': filters})
return self.call(
context,
@ -111,8 +110,8 @@ class VpnaasRpcSender(data_filter.Filter):
'notification_type': (
'update_status')}}]
}
LOG.info(_LI("Sending Notification 'Update Status' with "
"status:%(status)s "),
LOG.info("Sending Notification 'Update Status' with "
"status:%(status)s ",
{'status': status})
self._notify._notification(msg)
@ -127,8 +126,8 @@ class VpnaasRpcSender(data_filter.Filter):
'notification_type': (
'ipsec_site_conn_deleted')}}]
}
LOG.info(_LI("Sending Notification 'Ipsec Site Conn Deleted' "
"for resource:%(resource_id)s "),
LOG.info("Sending Notification 'Ipsec Site Conn Deleted' "
"for resource:%(resource_id)s ",
{'resource_id': resource_id})
self._notify._notification(msg)
@ -172,8 +171,8 @@ class VPNaasRpcManager(agent_base.AgentBaseRPCManager):
Returns: None
"""
LOG.info(_LI("Received request 'VPN Service Updated'."
"for API '%(api)s'"),
LOG.info("Received request 'VPN Service Updated'."
"for API '%(api)s'",
{'api': resource_data.get('reason', '')})
arg_dict = {'context': context,
'resource_data': resource_data}
@ -243,8 +242,8 @@ class VPNaasEventHandler(nfp_api.NfpEventHandler):
service_vendor = agent_info['service_vendor']
service_feature = agent_info['service_feature']
driver = self._get_driver(service_vendor, service_feature)
LOG.info(_LI("Invoking driver with service vendor:"
"%(service_vendor)s "),
LOG.info("Invoking driver with service vendor:"
"%(service_vendor)s ",
{'service_vendor': service_vendor})
setattr(VPNaasEventHandler, "service_driver", driver)
self._vpnservice_updated(ev, driver)

View File

@ -16,7 +16,6 @@ import time
from oslo_serialization import jsonutils
from gbpservice._i18n import _LI
from gbpservice.contrib.nfp.configurator.drivers.base import base_driver
from gbpservice.contrib.nfp.configurator.drivers.firewall.vyos import (
vyos_fw_constants as const)
@ -135,8 +134,8 @@ class FwGenericConfigDriver(base_driver.BaseDriver):
self.port,
'change_auth')
data = {}
LOG.info(_LI("Initiating POST request to configure Authentication "
"service at mgmt ip:%(mgmt_ip)s"),
LOG.info("Initiating POST request to configure Authentication "
"service at mgmt ip:%(mgmt_ip)s",
{'mgmt_ip': mgmt_ip})
err_msg = ("Change Auth POST request to the VyOS firewall "
"service at %s failed. " % url)
@ -188,8 +187,8 @@ class FwGenericConfigDriver(base_driver.BaseDriver):
'add_static_ip')
data = jsonutils.dumps(static_ips_info)
LOG.info(_LI("Initiating POST request to add static IPs for primary "
"service at mgmt ip:%(mgmt_ip)s"),
LOG.info("Initiating POST request to add static IPs for primary "
"service at mgmt ip:%(mgmt_ip)s",
{'mgmt_ip': mgmt_ip})
err_msg = ("Static IP POST request to the VyOS firewall "
"service at %s failed. " % url)
@ -267,8 +266,8 @@ class FwGenericConfigDriver(base_driver.BaseDriver):
url = const.request_url % (mgmt_ip,
self.port, 'add_rule')
data = jsonutils.dumps(rule_info)
LOG.info(_LI("Initiating POST request to add persistent rule to "
"primary service at mgmt ip: %(mgmt_ip)s"),
LOG.info("Initiating POST request to add persistent rule to "
"primary service at mgmt ip: %(mgmt_ip)s",
{'mgmt_ip': mgmt_ip})
err_msg = ("Add persistent rule POST request to the VyOS firewall "
"service at %s failed. " % url)
@ -322,8 +321,8 @@ class FwGenericConfigDriver(base_driver.BaseDriver):
'del_static_ip')
data = jsonutils.dumps(static_ips_info)
LOG.info(_LI("Initiating POST request to remove static IPs for "
"primary service at mgmt ip: %(mgmt_ip)s"),
LOG.info("Initiating POST request to remove static IPs for "
"primary service at mgmt ip: %(mgmt_ip)s",
{'mgmt_ip': mgmt_ip})
err_msg = ("Static IP DELETE request to the VyOS firewall "
@ -374,8 +373,8 @@ class FwGenericConfigDriver(base_driver.BaseDriver):
if result_static_ips != common_const.STATUS_SUCCESS:
return result_static_ips
else:
LOG.info(_LI("Successfully removed static IPs. "
"Result: %(result_static_ips)s"),
LOG.info("Successfully removed static IPs. "
"Result: %(result_static_ips)s",
{'result_static_ips': result_static_ips})
rule_info = dict(
@ -384,8 +383,8 @@ class FwGenericConfigDriver(base_driver.BaseDriver):
mgmt_ip = resource_data['mgmt_ip']
LOG.info(_LI("Initiating DELETE persistent rule for primary "
"service at mgmt ip: %(mgmt_ip)s"),
LOG.info("Initiating DELETE persistent rule for primary "
"service at mgmt ip: %(mgmt_ip)s",
{'mgmt_ip': mgmt_ip})
url = const.request_url % (mgmt_ip, self.port, 'delete_rule')
data = jsonutils.dumps(rule_info)
@ -447,8 +446,8 @@ class FwGenericConfigDriver(base_driver.BaseDriver):
route_info.append({'source_cidr': source_cidr,
'gateway_ip': gateway_ip})
data = jsonutils.dumps(route_info)
LOG.info(_LI("Initiating POST request to configure route of primary "
"service at mgmt ip: %(mgmt_ip)s"),
LOG.info("Initiating POST request to configure route of primary "
"service at mgmt ip: %(mgmt_ip)s",
{'mgmt_ip': mgmt_ip})
err_msg = ("Configure routes POST request to the VyOS firewall "
@ -497,8 +496,8 @@ class FwGenericConfigDriver(base_driver.BaseDriver):
for source_cidr in source_cidrs:
route_info.append({'source_cidr': source_cidr})
data = jsonutils.dumps(route_info)
LOG.info(_LI("Initiating Delete route to primary "
"service at mgmt ip: %(mgmt_ip)s"),
LOG.info("Initiating Delete route to primary "
"service at mgmt ip: %(mgmt_ip)s",
{'mgmt_ip': mgmt_ip})
err_msg = ("Routes DELETE request to the VyOS firewall "
@ -558,8 +557,8 @@ class FwaasDriver(FwGenericConfigDriver):
headers = self._parse_vm_context(context['agent_info']['context'])
resource_data = self.parse.parse_data(common_const.FIREWALL, context)
LOG.info(_LI("Processing request 'Create Firewall' in FWaaS Driver "
"for Firewall ID: %(f_id)s"),
LOG.info("Processing request 'Create Firewall' in FWaaS Driver "
"for Firewall ID: %(f_id)s",
{'f_id': firewall['id']})
mgmt_ip = resource_data.get('mgmt_ip')
url = const.request_url % (mgmt_ip,
@ -580,7 +579,7 @@ class FwaasDriver(FwGenericConfigDriver):
return common_const.STATUS_ERROR
if resp is common_const.STATUS_SUCCESS:
LOG.info(_LI("Configured firewall successfully at URL: %(url)s "),
LOG.info("Configured firewall successfully at URL: %(url)s ",
{'url': url})
return common_const.STATUS_ACTIVE
@ -604,8 +603,8 @@ class FwaasDriver(FwGenericConfigDriver):
"""
headers = self._parse_vm_context(context['agent_info']['context'])
LOG.info(_LI("Processing request 'Update Firewall' in FWaaS Driver "
"for Firewall ID:%(f_id)s"),
LOG.info("Processing request 'Update Firewall' in FWaaS Driver "
"for Firewall ID:%(f_id)s",
{'f_id': firewall['id']})
resource_data = self.parse.parse_data(common_const.FIREWALL, context)
mgmt_ip = resource_data.get('mgmt_ip')
@ -650,8 +649,8 @@ class FwaasDriver(FwGenericConfigDriver):
"""
headers = self._parse_vm_context(context['agent_info']['context'])
LOG.info(_LI("Processing request 'Delete Firewall' in FWaaS Driver "
"for Firewall ID:%(f_id)s"),
LOG.info("Processing request 'Delete Firewall' in FWaaS Driver "
"for Firewall ID:%(f_id)s",
{'f_id': firewall['id']})
resource_data = self.parse.parse_data(common_const.FIREWALL, context)
mgmt_ip = resource_data.get('mgmt_ip')

View File

@ -19,8 +19,6 @@ import uuid
from octavia.certificates.common import local as local_common
from octavia.certificates.manager import cert_mgr
from octavia.common import exceptions
from octavia.i18n import _LE
from octavia.i18n import _LI
from oslo_config import cfg
from gbpservice.nfp.core import log as nfp_logging
@ -55,9 +53,9 @@ class LocalCertManager(cert_mgr.CertManager):
cert_ref = str(uuid.uuid4())
filename_base = os.path.join(CONF.certificates.storage_path, cert_ref)
LOG.info(_LI(
LOG.info(
"Storing certificate data on the local filesystem."
))
)
try:
filename_certificate = "{0}.crt".format(filename_base, cert_ref)
with open(filename_certificate, 'w') as cert_file:
@ -78,7 +76,7 @@ class LocalCertManager(cert_mgr.CertManager):
with open(filename_pkp, 'w') as pass_file:
pass_file.write(private_key_passphrase)
except IOError as ioe:
LOG.error(_LE("Failed to store certificate."))
LOG.error("Failed to store certificate.")
raise exceptions.CertificateStorageException(message=ioe.message)
return cert_ref
@ -94,9 +92,9 @@ class LocalCertManager(cert_mgr.CertManager):
certificate data
:raises CertificateStorageException: if certificate retrieval fails
"""
LOG.info(_LI(
"Loading certificate {0} from the local filesystem."
).format(cert_ref))
LOG.info(
"Loading certificate {0} from the local filesystem.".format(
cert_ref))
filename_base = os.path.join(CONF.certificates.storage_path, cert_ref)
@ -111,9 +109,8 @@ class LocalCertManager(cert_mgr.CertManager):
with open(filename_certificate, 'r') as cert_file:
cert_data['certificate'] = cert_file.read()
except IOError:
LOG.error(_LE(
"Failed to read certificate for {0}."
).format(cert_ref))
LOG.error(
"Failed to read certificate for {0}.".format(cert_ref))
raise exceptions.CertificateStorageException(
msg="Certificate could not be read."
)
@ -121,9 +118,8 @@ class LocalCertManager(cert_mgr.CertManager):
with open(filename_private_key, 'r') as key_file:
cert_data['private_key'] = key_file.read()
except IOError:
LOG.error(_LE(
"Failed to read private key for {0}."
).format(cert_ref))
LOG.error(
"Failed to read private key for {0}.".format(cert_ref))
raise exceptions.CertificateStorageException(
msg="Private Key could not be read."
)
@ -151,9 +147,9 @@ class LocalCertManager(cert_mgr.CertManager):
:raises CertificateStorageException: if certificate deletion fails
"""
LOG.info(_LI(
"Deleting certificate {0} from the local filesystem."
).format(cert_ref))
LOG.info(
"Deleting certificate {0} from the local filesystem.".format(
cert_ref))
filename_base = os.path.join(CONF.certificates.storage_path, cert_ref)
@ -170,7 +166,6 @@ class LocalCertManager(cert_mgr.CertManager):
if os.path.exists(filename_pkp):
os.remove(filename_pkp)
except IOError as ioe:
LOG.error(_LE(
"Failed to delete certificate {0}."
).format(cert_ref))
LOG.error(
"Failed to delete certificate {0}.".format(cert_ref))
raise exceptions.CertificateStorageException(message=ioe.message)

View File

@ -16,7 +16,6 @@ import requests
import six
import time
from neutron._i18n import _LI
from gbpservice.contrib.nfp.configurator.drivers.base import base_driver
from gbpservice.contrib.nfp.configurator.drivers.vpn.vyos import (
@ -354,8 +353,8 @@ class VpnGenericConfigDriver(base_driver.BaseDriver):
'change_auth')
data = {}
LOG.info(_LI("Initiating POST request to configure Authentication "
"service at mgmt ip:%(mgmt_ip)s"),
LOG.info("Initiating POST request to configure Authentication "
"service at mgmt ip:%(mgmt_ip)s",
{'mgmt_ip': mgmt_ip})
err_msg = ("Change Auth POST request to the VyOS firewall "
"service at %s failed. " % url)

View File

@ -12,7 +12,6 @@
from oslo_log import helpers as log_helpers
from gbpservice._i18n import _LI
from gbpservice.contrib.nfp.configurator.lib import constants as const
from gbpservice.contrib.nfp.configurator.lib import demuxer
from gbpservice.contrib.nfp.configurator.lib import utils
@ -139,9 +138,9 @@ class ConfiguratorRpcManager(object):
log_info = request_data.get('info')
logging_context = log_info['context'].get('logging_context', {})
nfp_context['log_context'] = logging_context
LOG.info(_LI("Received RPC CREATE NETWORK FUNCTION DEVICE CONFIG "
"for %(service_type)s, NFI: %(nfi)s, "
"NF_ID: %(nf_id)s"),
LOG.info("Received RPC CREATE NETWORK FUNCTION DEVICE CONFIG "
"for %(service_type)s, NFI: %(nfi)s, "
"NF_ID: %(nf_id)s",
{'service_type': request_data['info']['service_type'],
'nfi': request_data['info']['context']['nfi_id'],
'nf_id': request_data['info']['context']['nf_id']})
@ -173,9 +172,9 @@ class ConfiguratorRpcManager(object):
log_info = request_data.get('info')
logging_context = log_info['context'].get('logging_context', {})
nfp_context['log_context'] = logging_context
LOG.info(_LI("Received RPC DELETE NETWORK FUNCTION DEVICE CONFIG "
"for %(service_type)s, NFI: %(nfi)s, "
"NF_ID: %(nf_id)s"),
LOG.info("Received RPC DELETE NETWORK FUNCTION DEVICE CONFIG "
"for %(service_type)s, NFI: %(nfi)s, "
"NF_ID: %(nf_id)s",
{'service_type': request_data['info']['service_type'],
'nfi': request_data['info']['context']['nfi_id'],
'nf_id': request_data['info']['context']['nf_id']})
@ -207,9 +206,9 @@ class ConfiguratorRpcManager(object):
log_info = request_data.get('info')
logging_context = log_info['context'].get('logging_context', {})
nfp_context['log_context'] = logging_context
LOG.info(_LI("Received RPC UPDATE NETWORK FUNCTION DEVICE CONFIG "
"for %(service_type)s, NFI: %(nfi)s, "
"NF_ID: %(nf_id)s"),
LOG.info("Received RPC UPDATE NETWORK FUNCTION DEVICE CONFIG "
"for %(service_type)s, NFI: %(nfi)s, "
"NF_ID: %(nf_id)s",
{'service_type': request_data['info']['service_type'],
'nfi': request_data['info']['context']['nfi_id'],
'nf_id': request_data['info']['context']['nf_id']})
@ -241,8 +240,8 @@ class ConfiguratorRpcManager(object):
log_info = request_data.get('info')
logging_context = log_info['context'].get('logging_context', {})
nfp_context['log_context'] = logging_context
LOG.info(_LI("Received RPC CREATE NETWORK FUNCTION CONFIG "
"for %(service_type)s "),
LOG.info("Received RPC CREATE NETWORK FUNCTION CONFIG "
"for %(service_type)s ",
{'service_type': request_data['info']['service_type']})
self._invoke_service_agent('create', request_data)
@ -272,8 +271,8 @@ class ConfiguratorRpcManager(object):
log_info = request_data.get('info')
logging_context = log_info['context'].get('logging_context', {})
nfp_context['log_context'] = logging_context
LOG.info(_LI("Received RPC DELETE NETWORK FUNCTION CONFIG "
"for %(service_type)s "),
LOG.info("Received RPC DELETE NETWORK FUNCTION CONFIG "
"for %(service_type)s ",
{'service_type': request_data['info']['service_type']})
self._invoke_service_agent('delete', request_data)
@ -303,8 +302,8 @@ class ConfiguratorRpcManager(object):
log_info = request_data.get('info')
logging_context = log_info['context'].get('logging_context', {})
nfp_context['log_context'] = logging_context
LOG.info(_LI("Received RPC UPDATE NETWORK FUNCTION CONFIG "
"for %(service_type)s "),
LOG.info("Received RPC UPDATE NETWORK FUNCTION CONFIG "
"for %(service_type)s ",
{'service_type': request_data['info']['service_type']})
self._invoke_service_agent('update', request_data)
@ -326,7 +325,7 @@ class ConfiguratorRpcManager(object):
"""
module_context.init()
LOG.info(_LI("Received RPC GET NOTIFICATIONS "))
LOG.info("Received RPC GET NOTIFICATIONS ")
events = self.sc.get_stashed_events()
notifications = []
for event in events:

View File

@ -13,7 +13,6 @@
from keystoneclient import exceptions as k_exceptions
from keystoneclient.v2_0 import client as keyclient
from gbpservice._i18n import _LE
from gbpservice.common import utils
from gbpservice.contrib.nfp.config_orchestrator.common import topics
from gbpservice.nfp.core import log as nfp_logging
@ -264,10 +263,10 @@ def _resource_owner_tenant_id():
return tenant.id
except k_exceptions.NotFound:
with excutils.save_and_reraise_exception(reraise=True):
LOG.error(_LE('No tenant with name %s exists.'), tenant)
LOG.error('No tenant with name %s exists.', tenant)
except k_exceptions.NoUniqueMatch:
with excutils.save_and_reraise_exception(reraise=True):
LOG.error(_LE('Multiple tenants matches found for %s'), tenant)
LOG.error('Multiple tenants matches found for %s', tenant)
def _get_router_for_floatingip(self, context, internal_port,

View File

@ -18,9 +18,6 @@ import subprocess
import time
import yaml
from gbpservice._i18n import _LE
from gbpservice._i18n import _LI
from oslo_log import log as logging
import oslo_serialization.jsonutils as jsonutils
@ -54,7 +51,7 @@ class Controller(rest.RestController):
out2 = subprocess.Popen('dhclient eth0', shell=True,
stdout=subprocess.PIPE).stdout.read()
output = "%s\n%s\n%s" % (ip_a, out1, out2)
LOG.info(_LI("Dhclient on eth0, result: %(output)s"),
LOG.info("Dhclient on eth0, result: %(output)s",
{'output': output})
except Exception as err:
msg = (
@ -161,8 +158,8 @@ class Controller(rest.RestController):
return {'failure_desc': {'msg': msg}}
def _configure_healthmonitor(self, config_data):
LOG.info(_LI("Configures healthmonitor with configuration "
"data : %(healthmonitor_data)s "),
LOG.info("Configures healthmonitor with configuration "
"data : %(healthmonitor_data)s ",
{'healthmonitor_data': config_data})
def _configure_interfaces(self, config_data):
@ -173,10 +170,10 @@ class Controller(rest.RestController):
out3 = subprocess.Popen('cat /etc/network/interfaces', shell=True,
stdout=subprocess.PIPE).stdout.read()
output = "%s\n%s\n%s" % (out1, out2, out3)
LOG.info(_LI("Dhclient on eth0, result: %(initial_data)s"),
LOG.info("Dhclient on eth0, result: %(initial_data)s",
{'initial_data': output})
LOG.info(_LI("Configures interfaces with configuration "
"data : %(interface_data)s "),
LOG.info("Configures interfaces with configuration "
"data : %(interface_data)s ",
{'interface_data': config_data})
def get_source_cidrs_and_gateway_ip(self, route_info):
@ -190,8 +187,8 @@ class Controller(rest.RestController):
return source_cidrs, gateway_ip
def _add_routes(self, route_info):
LOG.info(_LI("Configuring routes with configuration "
"data : %(route_data)s "),
LOG.info("Configuring routes with configuration "
"data : %(route_data)s ",
{'route_data': route_info['resource_data']})
source_cidrs, gateway_ip = self.get_source_cidrs_and_gateway_ip(
route_info)
@ -205,8 +202,8 @@ class Controller(rest.RestController):
try:
interface_number_string = source_interface.split("eth", 1)[1]
except IndexError:
LOG.error(_LE("Retrieved wrong interface %(interface)s for "
"configuring routes"),
LOG.error("Retrieved wrong interface %(interface)s for "
"configuring routes",
{'interface': source_interface})
try:
routing_table_number = 20 + int(interface_number_string)
@ -222,7 +219,7 @@ class Controller(rest.RestController):
routing_table_number, gateway_ip)
default_route_commands.append(ip_route_command)
output = "%s\n%s" % (out1, out2)
LOG.info(_LI("Static route configuration result: %(output)s"),
LOG.info("Static route configuration result: %(output)s",
{'output': output})
except Exception as ex:
raise Exception(_("Failed to add static routes: %(ex)s") % {
@ -231,7 +228,7 @@ class Controller(rest.RestController):
try:
out = subprocess.Popen(command, shell=True,
stdout=subprocess.PIPE).stdout.read()
LOG.info(_LI("Static route configuration result: %(output)s"),
LOG.info("Static route configuration result: %(output)s",
{'output': out})
except Exception as ex:
raise Exception(_("Failed to add static routes: %(ex)s") % {
@ -269,9 +266,9 @@ class Controller(rest.RestController):
"IP Address"))
def _apply_user_config(self, config_data):
LOG.info(_LI("Applying user config with configuration "
"type : %(config_type)s and "
"configuration data : %(config_data)s "),
LOG.info("Applying user config with configuration "
"type : %(config_type)s and "
"configuration data : %(config_data)s ",
{'config_type': config_data['resource'],
'config_data': config_data['resource_data']})
service_config = config_data['resource_data'][

View File

@ -15,8 +15,6 @@ from subprocess import PIPE
from subprocess import Popen
import sys
from oslo_log._i18n import _LE
from oslo_log._i18n import _LI
from oslo_log import log as logging
from oslo_serialization import jsonutils
@ -29,10 +27,10 @@ class ConfigureIPtables(object):
ps = Popen(["sysctl", "net.ipv4.ip_forward"], stdout=PIPE)
output = ps.communicate()[0]
if "0" in output:
LOG.info(_LI("Enabling IP forwarding ..."))
LOG.info("Enabling IP forwarding ...")
call(["sysctl", "-w", "net.ipv4.ip_forward=1"])
else:
LOG.info(_LI("IP forwarding already enabled"))
LOG.info("IP forwarding already enabled")
try:
self.rules_json = jsonutils.loads(json_blob)
except ValueError:
@ -44,7 +42,7 @@ class ConfigureIPtables(object):
# check if chain is present if not create new chain
if "testchain" not in output:
LOG.info(_LI("Creating new chain ..."))
LOG.info("Creating new chain ...")
call(["iptables", "-F"])
call(["iptables", "-N", "testchain"])
call(
@ -57,10 +55,10 @@ class ConfigureIPtables(object):
# return
# Update chain with new rules
LOG.info(_LI("Updating chain with new rules ..."))
LOG.info("Updating chain with new rules ...")
count = 0
for rule in self.rules_json.get('rules'):
LOG.info(_LI("adding rule %(count)d"), {'count': count})
LOG.info("adding rule %(count)d", {'count': count})
try:
action_values = ["LOG", "ACCEPT"]
action = rule['action'].upper()
@ -82,14 +80,14 @@ class ConfigureIPtables(object):
"-j", action], stdout=PIPE)
output = ps.communicate()[0]
if output:
LOG.error(_LE("Unable to add rule to chain due to: %(msg)s"),
LOG.error("Unable to add rule to chain due to: %(msg)s",
{'msg': output})
count = count + 1
ps = Popen(["iptables", "-A", "testchain", "-m", "state", "--state",
"ESTABLISHED,RELATED", "-j", "ACCEPT"], stdout=PIPE)
output = ps.communicate()[0]
if output:
LOG.error(_LE("Unable to add rule to chain due to: %(output)s"),
LOG.error("Unable to add rule to chain due to: %(output)s",
{'output': output})

View File

@ -24,8 +24,6 @@ from neutron_lib.plugins import directory
from oslo_log import log as logging
from oslo_utils import excutils
from gbpservice._i18n import _LE
from gbpservice._i18n import _LW
from gbpservice.neutron.extensions import group_policy as gp_ext
from gbpservice.neutron.extensions import servicechain as sc_ext
from gbpservice.neutron.services.grouppolicy.common import exceptions as exc
@ -237,7 +235,7 @@ class LocalAPI(object):
# plugins are loaded to grab and store plugin.
l3_plugin = directory.get_plugin(nl_const.L3)
if not l3_plugin:
LOG.error(_LE("No L3 router service plugin found."))
LOG.error("No L3 router service plugin found.")
raise exc.GroupPolicyDeploymentError()
return l3_plugin
@ -248,7 +246,7 @@ class LocalAPI(object):
# plugins are loaded to grab and store plugin.
qos_plugin = directory.get_plugin(pconst.QOS)
if not qos_plugin:
LOG.error(_LE("No QoS service plugin found."))
LOG.error("No QoS service plugin found.")
raise exc.GroupPolicyDeploymentError()
return qos_plugin
@ -258,7 +256,7 @@ class LocalAPI(object):
# plugins are loaded to grab and store plugin.
group_policy_plugin = directory.get_plugin(pconst.GROUP_POLICY)
if not group_policy_plugin:
LOG.error(_LE("No GroupPolicy service plugin found."))
LOG.error("No GroupPolicy service plugin found.")
raise exc.GroupPolicyDeploymentError()
return group_policy_plugin
@ -268,7 +266,7 @@ class LocalAPI(object):
# plugins are loaded to grab and store plugin.
servicechain_plugin = directory.get_plugin(pconst.SERVICECHAIN)
if not servicechain_plugin:
LOG.error(_LE("No Servicechain service plugin found."))
LOG.error("No Servicechain service plugin found.")
raise exc.GroupPolicyDeploymentError()
return servicechain_plugin
@ -374,7 +372,7 @@ class LocalAPI(object):
self._delete_resource(self._core_plugin,
plugin_context, 'port', port_id)
except n_exc.PortNotFound:
LOG.warning(_LW('Port %s already deleted'), port_id)
LOG.warning('Port %s already deleted', port_id)
def _get_subnet(self, plugin_context, subnet_id):
return self._get_resource(self._core_plugin, plugin_context, 'subnet',
@ -398,7 +396,7 @@ class LocalAPI(object):
self._delete_resource(self._core_plugin, plugin_context, 'subnet',
subnet_id)
except n_exc.SubnetNotFound:
LOG.warning(_LW('Subnet %s already deleted'), subnet_id)
LOG.warning('Subnet %s already deleted', subnet_id)
def _get_network(self, plugin_context, network_id):
return self._get_resource(self._core_plugin, plugin_context, 'network',
@ -422,7 +420,7 @@ class LocalAPI(object):
self._delete_resource(self._core_plugin, plugin_context,
'network', network_id)
except n_exc.NetworkNotFound:
LOG.warning(_LW('Network %s already deleted'), network_id)
LOG.warning('Network %s already deleted', network_id)
def _get_router(self, plugin_context, router_id):
return self._get_resource(self._l3_plugin, plugin_context, 'router',
@ -452,7 +450,7 @@ class LocalAPI(object):
self._l3_plugin.remove_router_interface(plugin_context, router_id,
interface_info)
except l3.RouterInterfaceNotFoundForSubnet:
LOG.warning(_LW('Router interface already deleted for subnet %s'),
LOG.warning('Router interface already deleted for subnet %s',
interface_info)
return
@ -472,7 +470,7 @@ class LocalAPI(object):
self._delete_resource(self._l3_plugin, plugin_context, 'router',
router_id)
except l3.RouterNotFound:
LOG.warning(_LW('Router %s already deleted'), router_id)
LOG.warning('Router %s already deleted', router_id)
def _get_sg(self, plugin_context, sg_id):
return self._get_resource(
@ -496,7 +494,7 @@ class LocalAPI(object):
self._delete_resource(self._core_plugin, plugin_context,
'security_group', sg_id)
except ext_sg.SecurityGroupNotFound:
LOG.warning(_LW('Security Group %s already deleted'), sg_id)
LOG.warning('Security Group %s already deleted', sg_id)
def _get_sg_rule(self, plugin_context, sg_rule_id):
return self._get_resource(
@ -513,7 +511,7 @@ class LocalAPI(object):
return self._create_resource(self._core_plugin, plugin_context,
'security_group_rule', attrs)
except ext_sg.SecurityGroupRuleExists as ex:
LOG.warning(_LW('Security Group already exists %s'), ex.message)
LOG.warning('Security Group already exists %s', ex.message)
return
def _update_sg_rule(self, plugin_context, sg_rule_id, attrs):
@ -526,7 +524,7 @@ class LocalAPI(object):
self._delete_resource(self._core_plugin, plugin_context,
'security_group_rule', sg_rule_id)
except ext_sg.SecurityGroupRuleNotFound:
LOG.warning(_LW('Security Group Rule %s already deleted'),
LOG.warning('Security Group Rule %s already deleted',
sg_rule_id)
def _get_fip(self, plugin_context, fip_id):
@ -551,7 +549,7 @@ class LocalAPI(object):
self._delete_resource(self._l3_plugin, plugin_context,
'floatingip', fip_id)
except l3.FloatingIPNotFound:
LOG.warning(_LW('Floating IP %s Already deleted'), fip_id)
LOG.warning('Floating IP %s Already deleted', fip_id)
def _get_address_scope(self, plugin_context, address_scope_id):
return self._get_resource(self._core_plugin, plugin_context,
@ -575,7 +573,7 @@ class LocalAPI(object):
self._delete_resource(self._core_plugin, plugin_context,
'address_scope', address_scope_id)
except address_scope.AddressScopeNotFound:
LOG.warning(_LW('Address Scope %s already deleted'),
LOG.warning('Address Scope %s already deleted',
address_scope_id)
def _get_subnetpool(self, plugin_context, subnetpool_id):
@ -600,7 +598,7 @@ class LocalAPI(object):
self._delete_resource(self._core_plugin, plugin_context,
'subnetpool', subnetpool_id)
except n_exc.SubnetpoolNotFound:
LOG.warning(_LW('Subnetpool %s already deleted'), subnetpool_id)
LOG.warning('Subnetpool %s already deleted', subnetpool_id)
def _get_l2_policy(self, plugin_context, l2p_id):
return self._get_resource(self._group_policy_plugin, plugin_context,
@ -619,7 +617,7 @@ class LocalAPI(object):
self._delete_resource(self._qos_plugin,
plugin_context, 'policy', qos_policy_id)
except n_exc.QosPolicyNotFound:
LOG.warning(_LW('QoS Policy %s already deleted'), qos_policy_id)
LOG.warning('QoS Policy %s already deleted', qos_policy_id)
def _get_qos_rules(self, plugin_context, filters=None):
filters = filters or {}
@ -639,7 +637,7 @@ class LocalAPI(object):
'policy_bandwidth_limit_rule',
rule_id, qos_policy_id)
except n_exc.QosRuleNotFound:
LOG.warning(_LW('QoS Rule %s already deleted'), rule_id)
LOG.warning('QoS Rule %s already deleted', rule_id)
def _get_l2_policies(self, plugin_context, filters=None):
filters = filters or {}
@ -659,7 +657,7 @@ class LocalAPI(object):
self._delete_resource(self._group_policy_plugin,
plugin_context, 'l2_policy', l2p_id, False)
except gp_ext.L2PolicyNotFound:
LOG.warning(_LW('L2 Policy %s already deleted'), l2p_id)
LOG.warning('L2 Policy %s already deleted', l2p_id)
def _get_l3_policy(self, plugin_context, l3p_id):
return self._get_resource(self._group_policy_plugin, plugin_context,
@ -683,7 +681,7 @@ class LocalAPI(object):
self._delete_resource(self._group_policy_plugin,
plugin_context, 'l3_policy', l3p_id, False)
except gp_ext.L3PolicyNotFound:
LOG.warning(_LW('L3 Policy %s already deleted'), l3p_id)
LOG.warning('L3 Policy %s already deleted', l3p_id)
def _get_external_segment(self, plugin_context, es_id):
return self._get_resource(self._group_policy_plugin, plugin_context,
@ -707,7 +705,7 @@ class LocalAPI(object):
self._delete_resource(self._group_policy_plugin, plugin_context,
'external_segment', es_id, False)
except gp_ext.ExternalSegmentNotFound:
LOG.warning(_LW('External Segment %s already deleted'), es_id)
LOG.warning('External Segment %s already deleted', es_id)
def _get_external_policy(self, plugin_context, ep_id):
return self._get_resource(self._group_policy_plugin, plugin_context,
@ -731,7 +729,7 @@ class LocalAPI(object):
self._delete_resource(self._group_policy_plugin, plugin_context,
'external_policy', ep_id, False)
except gp_ext.ExternalPolicyNotFound:
LOG.warning(_LW('External Policy %s already deleted'), ep_id)
LOG.warning('External Policy %s already deleted', ep_id)
def _get_policy_rule_set(self, plugin_context, prs_id):
return self._get_resource(self._group_policy_plugin, plugin_context,
@ -755,7 +753,7 @@ class LocalAPI(object):
self._delete_resource(self._group_policy_plugin, plugin_context,
'policy_rule_set', prs_id, False)
except gp_ext.PolicyRuleSetNotFound:
LOG.warning(_LW('Policy Rule Set %s already deleted'), prs_id)
LOG.warning('Policy Rule Set %s already deleted', prs_id)
def _get_servicechain_instance(self, plugin_context, sci_id):
return self._get_resource(self._servicechain_plugin, plugin_context,
@ -780,7 +778,7 @@ class LocalAPI(object):
self._delete_resource(self._servicechain_plugin, plugin_context,
'servicechain_instance', sci_id, False)
except sc_ext.ServiceChainInstanceNotFound:
LOG.warning(_LW("servicechain %s already deleted"), sci_id)
LOG.warning("servicechain %s already deleted", sci_id)
def _get_servicechain_spec(self, plugin_context, scs_id):
return self._get_resource(self._servicechain_plugin, plugin_context,
@ -804,7 +802,7 @@ class LocalAPI(object):
self._delete_resource(self._servicechain_plugin, plugin_context,
'servicechain_spec', scs_id)
except sc_ext.ServiceChainSpecNotFound:
LOG.warning(_LW("servicechain spec %s already deleted"), scs_id)
LOG.warning("servicechain spec %s already deleted", scs_id)
def _get_policy_target(self, plugin_context, pt_id):
return self._get_resource(self._group_policy_plugin, plugin_context,
@ -828,7 +826,7 @@ class LocalAPI(object):
self._delete_resource(self._group_policy_plugin, plugin_context,
'policy_target', pt_id, False)
except gp_ext.PolicyTargetNotFound:
LOG.warning(_LW('Policy Rule Set %s already deleted'), pt_id)
LOG.warning('Policy Rule Set %s already deleted', pt_id)
def _get_policy_target_group(self, plugin_context, ptg_id):
return self._get_resource(self._group_policy_plugin, plugin_context,
@ -853,4 +851,4 @@ class LocalAPI(object):
self._delete_resource(self._group_policy_plugin, plugin_context,
'policy_target_group', ptg_id)
except sc_ext.ServiceChainSpecNotFound:
LOG.warning(_LW("Policy Target Group %s already deleted"), ptg_id)
LOG.warning("Policy Target Group %s already deleted", ptg_id)

View File

@ -26,7 +26,6 @@ from sqlalchemy.ext.orderinglist import ordering_list
from sqlalchemy import orm
from sqlalchemy.orm import exc
from gbpservice._i18n import _LE
from gbpservice.neutron.extensions import servicechain as schain
from gbpservice.neutron.services.servicechain.common import exceptions as s_exc
@ -153,7 +152,7 @@ class ServiceChainDbPlugin(schain.ServiceChainPluginBase,
# plugins are loaded to grab and store plugin.
grouppolicy_plugin = directory.get_plugin(pconst.GROUP_POLICY)
if not grouppolicy_plugin:
LOG.error(_LE("No Grouppolicy service plugin found."))
LOG.error("No Grouppolicy service plugin found.")
raise s_exc.ServiceChainDeploymentError()
return grouppolicy_plugin

View File

@ -17,12 +17,12 @@ from neutron.api import extensions as neutron_extensions
from neutron.api.v2 import attributes as attr
from neutron.api.v2 import resource_helper
from neutron.plugins.common import constants
from neutron.services import service_base
from neutron_lib.api import converters as conv
from neutron_lib.api import extensions
from neutron_lib.api import validators as valid
from neutron_lib import constants as nlib_const
from neutron_lib import exceptions as nexc
from neutron_lib.services import base as service_base
from oslo_config import cfg
from oslo_log import log as logging
from oslo_utils import uuidutils

View File

@ -16,11 +16,11 @@ from neutron.api import extensions as neutron_extensions
from neutron.api.v2 import attributes as attr
from neutron.api.v2 import resource_helper
from neutron.plugins.common import constants
from neutron.services import service_base
from neutron_lib.api import converters as conv
from neutron_lib.api import extensions
from neutron_lib.api import validators as valid
from neutron_lib import exceptions as nexc
from neutron_lib.services import base as service_base
from oslo_config import cfg
from oslo_log import log as logging
import six

View File

@ -15,7 +15,6 @@
from oslo_log import log
from gbpservice._i18n import _LE
from gbpservice.neutron.plugins.ml2plus.drivers.apic_aim import exceptions
LOG = log.getLogger(__name__)
@ -108,6 +107,6 @@ class APICNameMapper(object):
if self._map(session, "", type_tag, prefix) == name[:pos]:
return name[pos:]
elif enforce:
LOG.error(_LE("Attempted to reverse-map invalid APIC name '%s'"),
LOG.error("Attempted to reverse-map invalid APIC name '%s'",
name)
raise exceptions.InternalError()

View File

@ -20,7 +20,6 @@ from keystoneclient.v3 import client as ksc_client
from oslo_config import cfg
from oslo_log import log as logging
from gbpservice._i18n import _LW
LOG = logging.getLogger(__name__)
@ -47,7 +46,7 @@ class ProjectNameCache(object):
auth = ksc_auth.load_from_conf_options(cfg.CONF, AUTH_GROUP)
LOG.debug("Got auth: %s", auth)
if not auth:
LOG.warning(_LW('No auth_plugin configured in %s'),
LOG.warning('No auth_plugin configured in %s',
AUTH_GROUP)
session = ksc_session.Session.load_from_conf_options(
cfg.CONF, AUTH_GROUP, auth=auth)

View File

@ -22,7 +22,6 @@ from neutron_lib.plugins import directory
from oslo_log import log
from oslo_utils import excutils
from gbpservice._i18n import _LI
from gbpservice.neutron import extensions as extensions_pkg
from gbpservice.neutron.extensions import cisco_apic
from gbpservice.neutron.plugins.ml2plus import driver_api as api_plus
@ -38,11 +37,11 @@ class ApicExtensionDriver(api_plus.ExtensionDriver,
extn_db.ExtensionDbMixin):
def __init__(self):
LOG.info(_LI("APIC AIM ED __init__"))
LOG.info("APIC AIM ED __init__")
self._mechanism_driver = None
def initialize(self):
LOG.info(_LI("APIC AIM ED initializing"))
LOG.info("APIC AIM ED initializing")
extensions.append_api_extensions_path(extensions_pkg.__path__)
@property

View File

@ -52,9 +52,6 @@ from oslo_log import log
import oslo_messaging
from oslo_utils import importutils
from gbpservice._i18n import _LE
from gbpservice._i18n import _LI
from gbpservice._i18n import _LW
from gbpservice.network.neutronv2 import local_api
from gbpservice.neutron.extensions import cisco_apic
from gbpservice.neutron.extensions import cisco_apic_l3 as a_l3
@ -165,10 +162,10 @@ class ApicMechanismDriver(api_plus.MechanismDriver,
self.md.delete_link(*args, **kwargs)
def __init__(self):
LOG.info(_LI("APIC AIM MD __init__"))
LOG.info("APIC AIM MD __init__")
def initialize(self):
LOG.info(_LI("APIC AIM MD initializing"))
LOG.info("APIC AIM MD initializing")
self.project_name_cache = cache.ProjectNameCache()
self.name_mapper = apic_mapper.APICNameMapper()
self.aim = aim_manager.AimManager()
@ -1011,9 +1008,9 @@ class ApicMechanismDriver(api_plus.MechanismDriver,
# REVISIT: Delete intf_vrf if no longer used?
else:
# This should never happen.
LOG.error(_LE("Interface topology %(intf_topology)s and "
"router topology %(router_topology)s have "
"different VRFs, but neither is shared"),
LOG.error("Interface topology %(intf_topology)s and "
"router topology %(router_topology)s have "
"different VRFs, but neither is shared",
{'intf_topology': intf_topology,
'router_topology': router_topology})
raise exceptions.InternalError()
@ -1611,8 +1608,8 @@ class ApicMechanismDriver(api_plus.MechanismDriver,
LOG.debug("Bound using segment: %s", segment)
return True
else:
LOG.warning(_LW("Refusing to bind port %(port)s to dead "
"agent: %(agent)s"),
LOG.warning("Refusing to bind port %(port)s to dead "
"agent: %(agent)s",
{'port': current['id'], 'agent': agent})
def _opflex_bind_port(self, context, segment, agent):
@ -1691,7 +1688,7 @@ class ApicMechanismDriver(api_plus.MechanismDriver,
api.PHYSICAL_NETWORK:
segment[api.PHYSICAL_NETWORK]}
dyn_seg = context.allocate_dynamic_segment(seg_args)
LOG.info(_LI('Allocated dynamic-segment %(s)s for port %(p)s'),
LOG.info('Allocated dynamic-segment %(s)s for port %(p)s',
{'s': dyn_seg, 'p': context.current['id']})
dyn_seg['aim_ml2_created'] = True
context.continue_binding(segment[api.ID], [dyn_seg])
@ -1970,8 +1967,8 @@ class ApicMechanismDriver(api_plus.MechanismDriver,
def _move_topology(self, aim_ctx, topology, old_vrf, new_vrf,
nets_to_notify):
LOG.info(_LI("Moving routed networks %(topology)s from VRF "
"%(old_vrf)s to VRF %(new_vrf)s"),
LOG.info("Moving routed networks %(topology)s from VRF "
"%(old_vrf)s to VRF %(new_vrf)s",
{'topology': topology.keys(),
'old_vrf': old_vrf,
'new_vrf': new_vrf})
@ -2189,7 +2186,7 @@ class ApicMechanismDriver(api_plus.MechanismDriver,
display_name=aim_utils.sanitize_display_name('CommonTenant'))
tenant = self.aim.get(aim_ctx, attrs)
if not tenant:
LOG.info(_LI("Creating common tenant"))
LOG.info("Creating common tenant")
tenant = self.aim.create(aim_ctx, attrs)
return tenant
@ -2200,7 +2197,7 @@ class ApicMechanismDriver(api_plus.MechanismDriver,
if not vrf:
attrs.display_name = (
aim_utils.sanitize_display_name('CommonUnroutedVRF'))
LOG.info(_LI("Creating common unrouted VRF"))
LOG.info("Creating common unrouted VRF")
vrf = self.aim.create(aim_ctx, attrs)
return vrf
@ -2213,7 +2210,7 @@ class ApicMechanismDriver(api_plus.MechanismDriver,
name=filter_name,
display_name=dname)
if not self.aim.get(aim_ctx, filter):
LOG.info(_LI("Creating common Any Filter"))
LOG.info("Creating common Any Filter")
self.aim.create(aim_ctx, filter)
dname = aim_utils.sanitize_display_name("AnyFilterEntry")
@ -2222,7 +2219,7 @@ class ApicMechanismDriver(api_plus.MechanismDriver,
name=ANY_FILTER_ENTRY_NAME,
display_name=dname)
if not self.aim.get(aim_ctx, entry):
LOG.info(_LI("Creating common Any FilterEntry"))
LOG.info("Creating common Any FilterEntry")
self.aim.create(aim_ctx, entry)
return filter
@ -2232,7 +2229,7 @@ class ApicMechanismDriver(api_plus.MechanismDriver,
if not vrf:
attrs.display_name = (
aim_utils.sanitize_display_name('DefaultRoutedVRF'))
LOG.info(_LI("Creating default VRF for %s"), attrs.tenant_name)
LOG.info("Creating default VRF for %s", attrs.tenant_name)
vrf = self.aim.create(aim_ctx, attrs)
return vrf
@ -2504,8 +2501,8 @@ class ApicMechanismDriver(api_plus.MechanismDriver,
.filter(extn_db_sn.snat_host_pool.is_(True))
.all())
if not snat_subnets:
LOG.info(_LI('No subnet in external network %s is marked as '
'SNAT-pool'),
LOG.info('No subnet in external network %s is marked as '
'SNAT-pool',
ext_network['id'])
return
for snat_subnet in snat_subnets:
@ -2524,8 +2521,8 @@ class ApicMechanismDriver(api_plus.MechanismDriver,
snat_ip = port['fixed_ips'][0]['ip_address']
break
except n_exceptions.IpAddressGenerationFailure:
LOG.info(_LI('No more addresses available in subnet %s '
'for SNAT IP allocation'),
LOG.info('No more addresses available in subnet %s '
'for SNAT IP allocation',
snat_subnet['id'])
else:
snat_ip = snat_port.fixed_ips[0].ip_address
@ -2569,8 +2566,8 @@ class ApicMechanismDriver(api_plus.MechanismDriver,
try:
self.plugin.delete_port(e_context, p[0])
except n_exceptions.NeutronException as ne:
LOG.warning(_LW('Failed to delete SNAT port %(port)s: '
'%(ex)s'),
LOG.warning('Failed to delete SNAT port %(port)s: '
'%(ex)s',
{'port': p, 'ex': ne})
def check_floatingip_external_address(self, context, floatingip):
@ -2625,7 +2622,7 @@ class ApicMechanismDriver(api_plus.MechanismDriver,
epg = self.get_epg_for_network(session, network)
if not epg:
LOG.info(_LI('Network %s does not map to any EPG'), network['id'])
LOG.info('Network %s does not map to any EPG', network['id'])
return
if segment:
@ -2682,8 +2679,8 @@ class ApicMechanismDriver(api_plus.MechanismDriver,
aim_ctx, aim_infra.HostLink, host_name=host,
interface_name=interface)
if not host_link or not host_link[0].path:
LOG.warning(_LW('No host link information found for host: '
'%(host)s, interface: %(interface)s'),
LOG.warning('No host link information found for host: '
'%(host)s, interface: %(interface)s',
{'host': host, 'interface': interface})
continue
host_link = host_link[0].path
@ -2697,7 +2694,7 @@ class ApicMechanismDriver(api_plus.MechanismDriver,
host_link = self.aim.find(aim_ctx, aim_infra.HostLink,
host_name=host)
if not host_link or not host_link[0].path:
LOG.warning(_LW('No host link information found for host %s'),
LOG.warning('No host link information found for host %s',
host)
return
host_link = host_link[0].path
@ -2721,7 +2718,7 @@ class ApicMechanismDriver(api_plus.MechanismDriver,
port_context.current['id'])
.first())
if not ports:
LOG.info(_LI('Releasing dynamic-segment %(s)s for port %(p)s'),
LOG.info('Releasing dynamic-segment %(s)s for port %(p)s',
{'s': btm, 'p': port_context.current['id']})
port_context.release_dynamic_segment(btm[api.ID])
@ -2871,7 +2868,7 @@ class ApicMechanismDriver(api_plus.MechanismDriver,
# this could be caused by concurrent transactions
except db_exc.DBDuplicateEntry as e:
LOG.debug(e)
LOG.info(_LI('Releasing domain %(d)s for port %(p)s'),
LOG.info('Releasing domain %(d)s for port %(p)s',
{'d': domain, 'p': port['id']})
def _get_non_opflex_segments_on_host(self, context, host):

View File

@ -13,8 +13,6 @@
# License for the specific language governing permissions and limitations
# under the License.
from gbpservice._i18n import _LE
from gbpservice._i18n import _LI
from gbpservice.neutron.plugins.ml2plus import driver_api
from neutron.db import api as db_api
@ -62,8 +60,8 @@ class MechanismManager(managers.MechanismManager):
'method': method_name},
exc_info=e)
LOG.exception(
_LE("Mechanism driver '%(name)s' failed in "
"%(method)s"),
"Mechanism driver '%(name)s' failed in "
"%(method)s",
{'name': driver.name, 'method': method_name}
)
errors.append(e)
@ -81,8 +79,8 @@ class MechanismManager(managers.MechanismManager):
try:
driver.obj.ensure_tenant(plugin_context, tenant_id)
except Exception:
LOG.exception(_LE("Mechanism driver '%s' failed in "
"ensure_tenant"), driver.name)
LOG.exception("Mechanism driver '%s' failed in "
"ensure_tenant", driver.name)
raise ml2_exc.MechanismDriverError(method="ensure_tenant")
def create_subnetpool_precommit(self, context):
@ -197,8 +195,8 @@ class ExtensionManager(managers.ExtensionManager):
result)
except Exception:
with excutils.save_and_reraise_exception():
LOG.info(_LI("Extension driver '%(name)s' failed in "
"%(method)s"),
LOG.info("Extension driver '%(name)s' failed in "
"%(method)s",
{'name': driver.name, 'method': method_name})
# Overrides ML2 implementation to avoid eating retriable

View File

@ -86,7 +86,6 @@ def notify(resource, event, trigger, **kwargs):
registry.notify = notify
from neutron._i18n import _LE
from neutron.callbacks import events
from neutron.callbacks import exceptions
from oslo_log import log as logging
@ -112,12 +111,12 @@ def _notify_loop(resource, event, trigger, **kwargs):
event.startswith(events.PRECOMMIT)
)
if not abortable_event:
LOG.exception(_LE("Error during notification for "
"%(callback)s %(resource)s, %(event)s"),
LOG.exception("Error during notification for "
"%(callback)s %(resource)s, %(event)s",
{'callback': callback_id,
'resource': resource, 'event': event})
else:
LOG.error(_LE("Callback %(callback)s raised %(error)s"),
LOG.error("Callback %(callback)s raised %(error)s",
{'callback': callback_id, 'error': e})
errors.append(exceptions.NotificationError(callback_id, e))
return errors
@ -197,11 +196,9 @@ def commit_reservation(context, reservation_id):
quota.QUOTAS.get_driver().commit_reservation = commit_reservation
from neutron._i18n import _LI
from oslo_db.sqlalchemy import exc_filters
exc_filters._LE = _LI
exc_filters.LOG.exception = exc_filters.LOG.debug

View File

@ -16,8 +16,6 @@
# The following is imported at the beginning to ensure
# that the patches are applied before any of the
# modules save a reference to the functions being patched
from gbpservice._i18n import _LE
from gbpservice._i18n import _LI
from gbpservice.neutron import extensions as gbp_extensions
from gbpservice.neutron.extensions import patch # noqa
from gbpservice.neutron.plugins.ml2plus import patch_neutron # noqa
@ -135,7 +133,7 @@ class Ml2PlusPlugin(ml2_plugin.Ml2Plugin,
security_group=securitygroups_db.SecurityGroup,
security_group_rule=securitygroups_db.SecurityGroupRule)
def __init__(self):
LOG.info(_LI("Ml2Plus initializing"))
LOG.info("Ml2Plus initializing")
registry._get_callback_manager()._notify_loop = (
patch_neutron._notify_loop)
# First load drivers, then initialize DB, then initialize drivers
@ -179,9 +177,9 @@ class Ml2PlusPlugin(ml2_plugin.Ml2Plugin,
registry.subscribe(self._subnet_delete_after_delete_handler,
resources.SUBNET, events.AFTER_DELETE)
except AttributeError:
LOG.info(_LI("Detected older version of Neutron, ML2Plus plugin "
"is not subscribed to subnet_precommit_delete and "
"subnet_after_delete events"))
LOG.info("Detected older version of Neutron, ML2Plus plugin "
"is not subscribed to subnet_precommit_delete and "
"subnet_after_delete events")
self._setup_dhcp()
self._start_rpc_notifiers()
self.add_agent_status_check_worker(self.agent_health_check)
@ -193,7 +191,7 @@ class Ml2PlusPlugin(ml2_plugin.Ml2Plugin,
cfg.CONF.ml2plus.refresh_subnetpool_db_obj)
self.refresh_address_scope_db_obj = (
cfg.CONF.ml2plus.refresh_address_scope_db_obj)
LOG.info(_LI("Modular L2 Plugin (extended) initialization complete"))
LOG.info("Modular L2 Plugin (extended) initialization complete")
db_base_plugin_v2.NeutronDbPluginV2.register_dict_extend_funcs(
attributes.SUBNETPOOLS, ['_ml2_md_extend_subnetpool_dict'])
@ -412,8 +410,8 @@ class Ml2PlusPlugin(ml2_plugin.Ml2Plugin,
self.mechanism_manager.create_subnetpool_postcommit(mech_context)
except ml2_exc.MechanismDriverError:
with excutils.save_and_reraise_exception():
LOG.error(_LE("mechanism_manager.create_subnetpool_postcommit "
"failed, deleting subnetpool '%s'"),
LOG.error("mechanism_manager.create_subnetpool_postcommit "
"failed, deleting subnetpool '%s'",
result['id'])
self.delete_subnetpool(context, result['id'])
return result
@ -476,9 +474,9 @@ class Ml2PlusPlugin(ml2_plugin.Ml2Plugin,
mech_context)
except ml2_exc.MechanismDriverError:
with excutils.save_and_reraise_exception():
LOG.error(_LE("mechanism_manager.create_address_scope_"
"postcommit failed, deleting address_scope"
" '%s'"),
LOG.error("mechanism_manager.create_address_scope_"
"postcommit failed, deleting address_scope"
" '%s'",
result['id'])
self.delete_address_scope(context, result['id'])
return result

View File

@ -29,8 +29,6 @@ from oslo_log import log as logging
from oslo_utils import excutils
from sqlalchemy import inspect
from gbpservice._i18n import _LE
from gbpservice._i18n import _LI
from gbpservice.neutron import extensions as extensions_pkg
from gbpservice.neutron.extensions import cisco_apic_l3 as l3_ext
from gbpservice.neutron.plugins.ml2plus.drivers.apic_aim import (
@ -61,7 +59,7 @@ class ApicL3Plugin(common_db_mixin.CommonDbMixin,
@resource_registry.tracked_resources(router=l3_db.Router,
floatingip=l3_db.FloatingIP)
def __init__(self):
LOG.info(_LI("APIC AIM L3 Plugin __init__"))
LOG.info("APIC AIM L3 Plugin __init__")
extensions.append_api_extensions_path(extensions_pkg.__path__)
self._mechanism_driver = None
super(ApicL3Plugin, self).__init__()
@ -84,7 +82,7 @@ class ApicL3Plugin(common_db_mixin.CommonDbMixin,
self._include_router_extn_attr(session, router_res)
except Exception:
with excutils.save_and_reraise_exception():
LOG.exception(_LE("APIC AIM extend_router_dict failed"))
LOG.exception("APIC AIM extend_router_dict failed")
db_base_plugin_v2.NeutronDbPluginV2.register_dict_extend_funcs(
l3.ROUTERS, ['_extend_router_dict_apic'])
@ -257,8 +255,8 @@ class ApicL3Plugin(common_db_mixin.CommonDbMixin,
.create_floatingip(context, floatingip))
break
except exceptions.IpAddressGenerationFailure:
LOG.info(_LI('No more floating IP addresses available '
'in subnet %s'),
LOG.info('No more floating IP addresses available '
'in subnet %s',
ext_sn)
if not result:

View File

@ -20,7 +20,6 @@ from oslo_serialization import jsonutils
from oslo_utils import excutils
import sqlalchemy as sa
from gbpservice._i18n import _LE
from gbpservice.common import utils
from gbpservice.network.neutronv2 import local_api
from gbpservice.neutron.db.grouppolicy import group_policy_mapping_db as gpdb
@ -104,10 +103,10 @@ class ChainMappingDriver(api.PolicyDriver, local_api.LocalAPI,
return tenant.id
except k_exceptions.NotFound:
with excutils.save_and_reraise_exception(reraise=reraise):
LOG.error(_LE('No tenant with name %s exists.'), tenant)
LOG.error('No tenant with name %s exists.', tenant)
except k_exceptions.NoUniqueMatch:
with excutils.save_and_reraise_exception(reraise=reraise):
LOG.error(_LE('Multiple tenants matches found for %s'),
LOG.error('Multiple tenants matches found for %s',
tenant)
@staticmethod
@ -290,7 +289,7 @@ class ChainMappingDriver(api.PolicyDriver, local_api.LocalAPI,
context.current['status_details'] = ptg_status[0][
'status_details']
except Exception:
LOG.error(_LE('Failed to update ptg status'))
LOG.error('Failed to update ptg status')
@log.log_method_call
def _delete_policy_target_group_postcommit(self, context):

View File

@ -29,9 +29,6 @@ from oslo_log import helpers as log
from oslo_log import log as logging
from oslo_utils import excutils
from gbpservice._i18n import _LE
from gbpservice._i18n import _LI
from gbpservice._i18n import _LW
from gbpservice.network.neutronv2 import local_api
from gbpservice.neutron.db.grouppolicy import group_policy_db as gpdb
from gbpservice.neutron.db.grouppolicy import group_policy_mapping_db as gpmdb
@ -154,7 +151,7 @@ class AIMMappingDriver(nrd.CommonNeutronBase, aim_rpc.AIMMappingRPCMixin):
@log.log_method_call
def initialize(self):
LOG.info(_LI("APIC AIM Policy Driver initializing"))
LOG.info("APIC AIM Policy Driver initializing")
super(AIMMappingDriver, self).initialize()
self._apic_aim_mech_driver = None
self._apic_segmentation_label_driver = None
@ -163,16 +160,16 @@ class AIMMappingDriver(nrd.CommonNeutronBase, aim_rpc.AIMMappingRPCMixin):
self._name_mapper = None
self.create_auto_ptg = cfg.CONF.aim_mapping.create_auto_ptg
if self.create_auto_ptg:
LOG.info(_LI('Auto PTG creation configuration set, '
'this will result in automatic creation of a PTG '
'per L2 Policy'))
LOG.info('Auto PTG creation configuration set, '
'this will result in automatic creation of a PTG '
'per L2 Policy')
self.create_per_l3p_implicit_contracts = (
cfg.CONF.aim_mapping.create_per_l3p_implicit_contracts)
self.advertise_mtu = cfg.CONF.aim_mapping.advertise_mtu
local_api.QUEUE_OUT_OF_PROCESS_NOTIFICATIONS = True
if self.create_per_l3p_implicit_contracts:
LOG.info(_LI('Implicit AIM contracts will be created '
'for l3_policies which do not have them.'))
LOG.info('Implicit AIM contracts will be created '
'for l3_policies which do not have them.')
self._create_per_l3p_implicit_contracts()
@log.log_method_call
@ -453,11 +450,11 @@ class AIMMappingDriver(nrd.CommonNeutronBase, aim_rpc.AIMMappingRPCMixin):
context._plugin).delete_policy_target_group(
context._plugin_context, auto_ptg['id'])
except gpolicy.PolicyTargetGroupNotFound:
LOG.info(_LI("Auto PTG with ID %(id)s for "
"for L2P %(l2p)s not found. If create_auto_ptg "
"configuration was not set at the time of the L2P "
"creation, you can safely ignore this, else this "
"could potentially be indication of an error."),
LOG.info("Auto PTG with ID %(id)s for "
"for L2P %(l2p)s not found. If create_auto_ptg "
"configuration was not set at the time of the L2P "
"creation, you can safely ignore this, else this "
"could potentially be indication of an error.",
{'id': auto_ptg_id, 'l2p': l2p_id})
super(AIMMappingDriver, self).delete_l2_policy_precommit(context)
@ -1062,9 +1059,9 @@ class AIMMappingDriver(nrd.CommonNeutronBase, aim_rpc.AIMMappingRPCMixin):
ok_to_bind = True
break
if not ok_to_bind:
LOG.warning(_LW("Failed to bind the port due to "
"allowed_vm_names rules %(rules)s "
"for VM: %(vm)s"),
LOG.warning("Failed to bind the port due to "
"allowed_vm_names rules %(rules)s "
"for VM: %(vm)s",
{'rules': l3p['allowed_vm_names'],
'vm': vm.name})
return ok_to_bind
@ -1548,9 +1545,9 @@ class AIMMappingDriver(nrd.CommonNeutronBase, aim_rpc.AIMMappingRPCMixin):
# default EPG and delete them
# create=True, delete=True is not a valid combination
if create and delete:
LOG.error(_LE("Incorrect use of internal method "
"_process_contracts_for_default_epg(), create and "
"delete cannot be True at the same time"))
LOG.error("Incorrect use of internal method "
"_process_contracts_for_default_epg(), create and "
"delete cannot be True at the same time")
raise
session = context._plugin_context.session
aim_ctx = aim_context.AimContext(session)
@ -1792,8 +1789,8 @@ class AIMMappingDriver(nrd.CommonNeutronBase, aim_rpc.AIMMappingRPCMixin):
if not epg:
# Something is wrong, default EPG doesn't exist.
# TODO(ivar): should rise an exception
LOG.error(_LE("Default EPG doesn't exist for "
"port %s"), port['id'])
LOG.error("Default EPG doesn't exist for "
"port %s", port['id'])
return epg
def _get_subnet_details(self, plugin_context, port, details):
@ -2155,9 +2152,9 @@ class AIMMappingDriver(nrd.CommonNeutronBase, aim_rpc.AIMMappingRPCMixin):
intf_port = self._create_port(plugin_context, attrs)
except n_exc.NeutronException:
with excutils.save_and_reraise_exception():
LOG.exception(_LE('Failed to create explicit router '
'interface port in subnet '
'%(subnet)s'),
LOG.exception('Failed to create explicit router '
'interface port in subnet '
'%(subnet)s',
{'subnet': subnet['id']})
interface_info = {'port_id': intf_port['id'],
NO_VALIDATE: True}
@ -2167,9 +2164,9 @@ class AIMMappingDriver(nrd.CommonNeutronBase, aim_rpc.AIMMappingRPCMixin):
except n_exc.BadRequest:
self._delete_port(plugin_context, intf_port['id'])
with excutils.save_and_reraise_exception():
LOG.exception(_LE('Attaching router %(router)s to '
'%(subnet)s with explicit port '
'%(port) failed'),
LOG.exception('Attaching router %(router)s to '
'%(subnet)s with explicit port '
'%(port) failed',
{'subnet': subnet['id'],
'router': router_id,
'port': intf_port['id']})
@ -2185,8 +2182,8 @@ class AIMMappingDriver(nrd.CommonNeutronBase, aim_rpc.AIMMappingRPCMixin):
self._add_router_interface(plugin_context, router_id,
interface_info)
except n_exc.BadRequest as e:
LOG.exception(_LE("Adding subnet to router failed, exception:"
"%s"), e)
LOG.exception("Adding subnet to router failed, exception:"
"%s", e)
raise exc.GroupPolicyInternalError()
def _detach_router_from_subnets(self, plugin_context, router_id, sn_ids):

View File

@ -17,8 +17,6 @@ from neutron.plugins.ml2 import rpc as ml2_rpc
from opflexagent import rpc as o_rpc
from oslo_log import log
from gbpservice._i18n import _LE
from gbpservice._i18n import _LW
from gbpservice.neutron.services.grouppolicy.drivers.cisco.apic import (
nova_client as nclient)
from gbpservice.neutron.services.grouppolicy.drivers.cisco.apic import (
@ -65,8 +63,8 @@ class AIMMappingRPCMixin(ha_ip_db.HAIPOwnerDbMixin):
return self._retrieve_vrf_details(context, **kwargs)
except Exception as e:
vrf = kwargs.get('vrf_id')
LOG.error(_LE("An exception has occurred while retrieving vrf "
"gbp details for %s"), vrf)
LOG.error("An exception has occurred while retrieving vrf "
"gbp details for %s", vrf)
LOG.exception(e)
return {'l3_policy_id': vrf}
@ -82,8 +80,8 @@ class AIMMappingRPCMixin(ha_ip_db.HAIPOwnerDbMixin):
return self._get_gbp_details(context, kwargs, kwargs.get('host'))
except Exception as e:
device = kwargs.get('device')
LOG.error(_LE("An exception has occurred while retrieving device "
"gbp details for %s"), device)
LOG.error("An exception has occurred while retrieving device "
"gbp details for %s", device)
LOG.exception(e)
return {'device': device}
@ -101,8 +99,8 @@ class AIMMappingRPCMixin(ha_ip_db.HAIPOwnerDbMixin):
None, None).get_device_details(context, **request)}
return result
except Exception as e:
LOG.error(_LE("An exception has occurred while requesting device "
"gbp details for %s"), request.get('device'))
LOG.error("An exception has occurred while requesting device "
"gbp details for %s", request.get('device'))
LOG.exception(e)
return None
@ -137,8 +135,8 @@ class AIMMappingRPCMixin(ha_ip_db.HAIPOwnerDbMixin):
port_context = core_plugin.get_bound_port_context(context, port_id,
host)
if not port_context:
LOG.warning(_LW("Device %(device)s requested by agent "
"%(agent_id)s not found in database"),
LOG.warning("Device %(device)s requested by agent "
"%(agent_id)s not found in database",
{'device': port_id,
'agent_id': request.get('agent_id')})
return {'device': request.get('device')}

View File

@ -14,8 +14,6 @@ from neutron.notifiers import nova as n_nova
from novaclient import exceptions as nova_exceptions
from oslo_log import log as logging
from gbpservice._i18n import _LW
LOG = logging.getLogger(__name__)
@ -39,7 +37,7 @@ class NovaClient(object):
try:
return self.client.servers.get(server_id)
except nova_exceptions.NotFound:
LOG.warning(_LW("Nova returned NotFound for server: %s"),
LOG.warning("Nova returned NotFound for server: %s",
server_id)
except Exception as e:
LOG.exception(e)

View File

@ -13,7 +13,6 @@
from neutron_lib.plugins import directory
from oslo_log import log as logging
from gbpservice._i18n import _LI
from gbpservice.neutron.db.grouppolicy.extensions import (
apic_auto_ptg_db as auto_ptg_db)
from gbpservice.neutron.db.grouppolicy.extensions import (
@ -34,7 +33,7 @@ class AIMExtensionDriver(api.ExtensionDriver,
_extension_dict = cisco_apic_gbp.EXTENDED_ATTRIBUTES_2_0
def __init__(self):
LOG.info(_LI("AIM Extension __init__"))
LOG.info("AIM Extension __init__")
self._policy_driver = None
@property

View File

@ -12,7 +12,6 @@
from oslo_log import log as logging
from gbpservice._i18n import _LI
from gbpservice.neutron.db.grouppolicy.extensions import (
apic_reuse_bd_db as db)
from gbpservice.neutron.db.grouppolicy import group_policy_db as gp_db
@ -30,7 +29,7 @@ class ApicReuseBdExtensionDriver(api.ExtensionDriver,
_extension_dict = ext.EXTENDED_ATTRIBUTES_2_0
def __init__(self):
LOG.info(_LI("ApicReuseBdExtensionDriver __init__"))
LOG.info("ApicReuseBdExtensionDriver __init__")
def initialize(self):
pass

View File

@ -15,7 +15,6 @@ from neutron_lib.api import validators
from oslo_config import cfg
from oslo_log import log as logging
from gbpservice._i18n import _LW
from gbpservice.neutron.db.grouppolicy.extensions import group_proxy_db as db
from gbpservice.neutron.db.grouppolicy import group_policy_db as gp_db
from gbpservice.neutron.extensions import driver_proxy_group
@ -115,10 +114,10 @@ class ProxyGroupDriver(api.ExtensionDriver):
data['ip_version'], data['proxy_subnet_prefix_length'],
data['ip_pool'])
if data['proxy_ip_pool']:
LOG.warning(_LW("Since use_subnetpools setting is turned on, "
"proxy_ip_pool %s will be ignored. "
"Proxy subnets will be allocated from same "
"subnetpool as group subnets"),
LOG.warning("Since use_subnetpools setting is turned on, "
"proxy_ip_pool %s will be ignored. "
"Proxy subnets will be allocated from same "
"subnetpool as group subnets",
data['proxy_ip_pool'])
else:
gp_db.GroupPolicyDbPlugin.validate_ip_pool(

View File

@ -17,8 +17,6 @@ from oslo_log import log as logging
from oslo_utils import excutils
import sqlalchemy as sa
from gbpservice._i18n import _LI
from gbpservice._i18n import _LW
from gbpservice.network.neutronv2 import local_api
from gbpservice.neutron.extensions import driver_proxy_group as pg_ext
from gbpservice.neutron.extensions import group_policy as gbp_ext
@ -130,17 +128,17 @@ class ImplicitPolicyBase(api.PolicyDriver, local_api.LocalAPI):
filter)
l3p = l3ps and l3ps[0]
if not l3p:
LOG.warning(_LW(
LOG.warning(
"Caught DefaultL3PolicyAlreadyExists, "
"but default L3 policy not concurrently "
"created for tenant %s"), tenant_id)
"created for tenant %s", tenant_id)
ctxt.reraise = True
except exc.OverlappingIPPoolsInSameTenantNotAllowed:
with excutils.save_and_reraise_exception():
LOG.info(_LI("Caught "
"OverlappingIPPoolsinSameTenantNotAllowed "
"during creation of default L3 policy for "
"tenant %s"), tenant_id)
LOG.info("Caught "
"OverlappingIPPoolsinSameTenantNotAllowed "
"during creation of default L3 policy for "
"tenant %s", tenant_id)
context.current['l3_policy_id'] = l3p['id']
def _use_implicit_l3_policy(self, context):
@ -204,9 +202,9 @@ class ImplicitPolicyBase(api.PolicyDriver, local_api.LocalAPI):
try:
self._delete_l2_policy(context._plugin_context, l2p_id)
except gbp_ext.L2PolicyInUse:
LOG.info(_LI(
LOG.info(
"Cannot delete implicit L2 Policy %s because it's "
"in use."), l2p_id)
"in use.", l2p_id)
def _validate_default_external_segment(self, context):
# REVISIT(ivar): find a better way to retrieve the default ES

View File

@ -33,9 +33,6 @@ from oslo_utils import excutils
import sqlalchemy as sa
from sqlalchemy.orm import exc as sa_exc
from gbpservice._i18n import _LE
from gbpservice._i18n import _LI
from gbpservice._i18n import _LW
from gbpservice.common import utils
from gbpservice.network.neutronv2 import local_api
from gbpservice.neutron.db.grouppolicy import group_policy_db as gpdb
@ -318,9 +315,9 @@ class ImplicitResourceOperations(local_api.LocalAPI,
filters={'address_scope_id':
[address_scope_id]})
if subpools:
LOG.warning(_LW("Cannot delete implicitly created "
"address_scope %(id)s since it has "
"associated subnetpools: %(pools)s"),
LOG.warning("Cannot delete implicitly created "
"address_scope %(id)s since it has "
"associated subnetpools: %(pools)s",
{'id': address_scope_id, 'pools': subpools})
else:
self._delete_address_scope(plugin_context, address_scope_id)
@ -358,9 +355,9 @@ class ImplicitResourceOperations(local_api.LocalAPI,
filters={'subnetpool_id':
[subnetpool_id]})
if subnets:
LOG.warning(_LW("Cannot delete implicitly created "
"subnetpool %(id)s since it has "
"associated subnets: %(subnets)s"),
LOG.warning("Cannot delete implicitly created "
"subnetpool %(id)s since it has "
"associated subnets: %(subnets)s",
{'id': subnetpool_id, 'subnets': subnets})
else:
self._delete_subnetpool(plugin_context, subnetpool_id)
@ -631,10 +628,10 @@ class ImplicitResourceOperations(local_api.LocalAPI,
except Exception as e:
if isinstance(e, oslo_db_excp.RetryRequest):
raise e
LOG.info(_LI("Allocating subnet from subnetpool %(sp)s "
"failed. Allocation will be attempted "
"from any other configured "
"subnetpool(s). Exception: %(excp)s"),
LOG.info("Allocating subnet from subnetpool %(sp)s "
"failed. Allocation will be attempted "
"from any other configured "
"subnetpool(s). Exception: %(excp)s",
{'sp': pool['id'], 'excp': type(e)})
last = e
continue
@ -725,7 +722,7 @@ class ImplicitResourceOperations(local_api.LocalAPI,
context.set_port_id(port_id)
return
except n_exc.IpAddressGenerationFailure as ex:
LOG.warning(_LW("No more address available in subnet %s"),
LOG.warning("No more address available in subnet %s",
subnet['id'])
last = ex
raise last
@ -735,7 +732,7 @@ class ImplicitResourceOperations(local_api.LocalAPI,
try:
self._delete_port(plugin_context, port_id)
except n_exc.PortNotFound:
LOG.warning(_LW("Port %s is missing"), port_id)
LOG.warning("Port %s is missing", port_id)
def _reject_invalid_router_access(self, context):
# Validate if the explicit router(s) belong to the tenant.
@ -782,8 +779,8 @@ class ImplicitResourceOperations(local_api.LocalAPI,
self._add_router_interface(plugin_context, router_id,
interface_info)
except n_exc.BadRequest as e:
LOG.exception(_LE("Adding subnet to router failed, exception:"
"%s"), e)
LOG.exception("Adding subnet to router failed, exception:"
"%s", e)
raise exc.GroupPolicyInternalError()
def _add_router_interface_for_subnet(self, context, router_id, subnet_id):
@ -1109,9 +1106,9 @@ class ImplicitResourceOperations(local_api.LocalAPI,
context, l2_policy_id)
fip_ids = []
if not external_segments:
LOG.error(_LE("Network Service Policy to allocate Floating IP "
"could not be applied because l3policy does "
"not have an attached external segment"))
LOG.error("Network Service Policy to allocate Floating IP "
"could not be applied because l3policy does "
"not have an attached external segment")
return fip_ids
tenant_id = context.current['tenant_id']
@ -1153,7 +1150,7 @@ class ImplicitResourceOperations(local_api.LocalAPI,
# FIP allocated, no need to try further allocation
break
except n_exc.IpAddressGenerationFailure as ex:
LOG.warning(_LW("Floating allocation failed: %s"),
LOG.warning("Floating allocation failed: %s",
ex.message)
if fip_id:
fip_ids.append(fip_id)
@ -1261,10 +1258,10 @@ class ImplicitResourceOperations(local_api.LocalAPI,
filters={'name': [
gpip.default_external_segment_name]}))
if not external_segments:
LOG.error(_LE(
LOG.error(
"Network Service Policy to allocate Floating "
"IP could not be associated because l3policy "
"does not have an attached external segment"))
"does not have an attached external segment")
raise exc.NSPRequiresES()
for es in external_segments:
if not es['nat_pools']:
@ -1286,9 +1283,9 @@ class ImplicitResourceOperations(local_api.LocalAPI,
free_ip = self._get_last_free_ip(context._plugin_context,
context.current['subnets'])
if not free_ip:
LOG.error(_LE("Reserving IP Addresses failed for Network "
"Service Policy. No more IP Addresses on "
"subnet"))
LOG.error("Reserving IP Addresses failed for Network "
"Service Policy. No more IP Addresses on "
"subnet")
return
# TODO(Magesh):Fetch subnet from PTG to which NSP is attached
self._remove_ip_from_allocation_pool(
@ -1640,7 +1637,7 @@ class ResourceMappingDriver(api.PolicyDriver, ImplicitResourceOperations,
policy_target = context._plugin.get_policy_target(
context._plugin_context, pt_id)
except gp_ext.PolicyTargetNotFound:
LOG.warning(_LW("Attempted to fetch deleted Service Target (QoS)"))
LOG.warning("Attempted to fetch deleted Service Target (QoS)")
else:
port_id = policy_target['port_id']
port = {attributes.PORT: {'qos_policy_id': None}}
@ -1703,16 +1700,16 @@ class ResourceMappingDriver(api.PolicyDriver, ImplicitResourceOperations,
return tenant.id
except k_exceptions.NotFound:
with excutils.save_and_reraise_exception(reraise=True):
LOG.error(_LE('No tenant with name %s exists.'), tenant)
LOG.error('No tenant with name %s exists.', tenant)
except k_exceptions.NoUniqueMatch:
with excutils.save_and_reraise_exception(reraise=True):
LOG.error(_LE('Multiple tenants matches found for %s'), tenant)
LOG.error('Multiple tenants matches found for %s', tenant)
except k_exceptions.AuthorizationFailure:
LOG.error(_LE("User: %(user)s dont have permissions"),
LOG.error("User: %(user)s dont have permissions",
{'user': user})
except k_exceptions.Unauthorized:
LOG.error(_LE("Wrong credentials provided: user: %(user)s, "
"password: %(pwd)s, tenant: %(tenant)s"),
LOG.error("Wrong credentials provided: user: %(user)s, "
"password: %(pwd)s, tenant: %(tenant)s",
{'user': user, 'pwd': pwd, 'tenant': tenant})
@log.log_method_call
@ -1940,7 +1937,7 @@ class ResourceMappingDriver(api.PolicyDriver, ImplicitResourceOperations,
context.nsp_cleanup_ipaddress,
context.nsp_cleanup_fips)
except sa_exc.ObjectDeletedError as err:
LOG.warning(_LW("Object already got deleted. Error: %(err)s"),
LOG.warning("Object already got deleted. Error: %(err)s",
{'err': err})
# Cleanup SGs
self._unset_sg_rules_for_subnets(
@ -2549,7 +2546,7 @@ class ResourceMappingDriver(api.PolicyDriver, ImplicitResourceOperations,
context._plugin_context, subnet_id, router_id)
except n_exc.InvalidInput:
# This exception is not expected.
LOG.exception(_LE("adding subnet to router failed"))
LOG.exception("adding subnet to router failed")
for subnet_id in subnet_ids:
self._delete_subnet(context._plugin_context, subnet_id)
raise exc.GroupPolicyInternalError()
@ -2598,7 +2595,7 @@ class ResourceMappingDriver(api.PolicyDriver, ImplicitResourceOperations,
except n_exc.InvalidInput:
# This exception is not expected.
# TODO(ivar): find a better way to rollback
LOG.exception(_LE("adding subnet to router failed"))
LOG.exception("adding subnet to router failed")
for subnet_id in subnet_ids:
self._delete_subnet(context._plugin_context, subnet_id)
raise exc.GroupPolicyInternalError()
@ -2688,7 +2685,7 @@ class ResourceMappingDriver(api.PolicyDriver, ImplicitResourceOperations,
pt = context._plugin.get_policy_target(context._plugin_context,
pt_id)
except gp_ext.PolicyTargetNotFound:
LOG.warning(_LW("PT %s doesn't exist anymore"), pt_id)
LOG.warning("PT %s doesn't exist anymore", pt_id)
return
try:
port_id = pt['port_id']
@ -2702,14 +2699,14 @@ class ResourceMappingDriver(api.PolicyDriver, ImplicitResourceOperations,
port[ext_sg.SECURITYGROUPS] = new_sg_list
self._update_port(context._plugin_context, port_id, port)
except n_exc.PortNotFound:
LOG.warning(_LW("Port %s is missing"), port_id)
LOG.warning("Port %s is missing", port_id)
def _disassoc_sgs_from_pt(self, context, pt_id, sg_list):
try:
pt = context._plugin.get_policy_target(context._plugin_context,
pt_id)
except gp_ext.PolicyTargetNotFound:
LOG.warning(_LW("PT %s doesn't exist anymore"), pt_id)
LOG.warning("PT %s doesn't exist anymore", pt_id)
return
port_id = pt['port_id']
self._disassoc_sgs_from_port(context._plugin_context, port_id, sg_list)
@ -2726,7 +2723,7 @@ class ResourceMappingDriver(api.PolicyDriver, ImplicitResourceOperations,
port[ext_sg.SECURITYGROUPS] = new_sg_list
self._update_port(plugin_context, port_id, port)
except n_exc.PortNotFound:
LOG.warning(_LW("Port %s is missing"), port_id)
LOG.warning("Port %s is missing", port_id)
def _generate_list_of_sg_from_ptg(self, context, ptg_id):
ptg = context._plugin.get_policy_target_group(

View File

@ -17,8 +17,6 @@ from oslo_log import log
from oslo_utils import excutils
import stevedore
from gbpservice._i18n import _LE
from gbpservice._i18n import _LI
from gbpservice.neutron.services.grouppolicy.common import exceptions as gp_exc
@ -33,14 +31,14 @@ class ExtensionManager(stevedore.named.NamedExtensionManager):
# the order in which the drivers are called.
self.ordered_ext_drivers = []
LOG.info(_LI("Configured extension driver names: %s"),
LOG.info("Configured extension driver names: %s",
cfg.CONF.group_policy.extension_drivers)
super(ExtensionManager, self).__init__(
'gbpservice.neutron.group_policy.extension_drivers',
cfg.CONF.group_policy.extension_drivers,
invoke_on_load=True,
name_order=True)
LOG.info(_LI("Loaded extension driver names: %s"), self.names())
LOG.info("Loaded extension driver names: %s", self.names())
self._register_drivers()
def _register_drivers(self):
@ -51,13 +49,13 @@ class ExtensionManager(stevedore.named.NamedExtensionManager):
"""
for ext in self:
self.ordered_ext_drivers.append(ext)
LOG.info(_LI("Registered extension drivers: %s"),
LOG.info("Registered extension drivers: %s",
[driver.name for driver in self.ordered_ext_drivers])
def initialize(self):
# Initialize each driver in the list.
for driver in self.ordered_ext_drivers:
LOG.info(_LI("Initializing extension driver '%s'"), driver.name)
LOG.info("Initializing extension driver '%s'", driver.name)
driver.obj.initialize()
def extension_aliases(self):
@ -65,7 +63,7 @@ class ExtensionManager(stevedore.named.NamedExtensionManager):
for driver in self.ordered_ext_drivers:
alias = driver.obj.extension_alias
exts.append(alias)
LOG.info(_LI("Got %(alias)s extension from driver '%(drv)s'"),
LOG.info("Got %(alias)s extension from driver '%(drv)s'",
{'alias': alias, 'drv': driver.name})
return exts
@ -77,13 +75,13 @@ class ExtensionManager(stevedore.named.NamedExtensionManager):
except (gp_exc.GroupPolicyException, n_exc.NeutronException):
with excutils.save_and_reraise_exception():
LOG.exception(
_LE("Extension driver '%(name)s' "
"failed in %(method)s"),
"Extension driver '%(name)s' "
"failed in %(method)s",
{'name': driver.name, 'method': method_name}
)
except Exception:
LOG.exception(_LE("Extension driver '%(name)s' "
"failed in %(method)s"),
LOG.exception("Extension driver '%(name)s' "
"failed in %(method)s",
{'name': driver.name, 'method': method_name})
# We are replacing a non-GBP/non-Neutron exception here
raise gp_exc.GroupPolicyDriverError(method=method_name)

View File

@ -24,8 +24,6 @@ from oslo_log import helpers as log
from oslo_log import log as logging
from oslo_utils import excutils
from gbpservice._i18n import _LE
from gbpservice._i18n import _LW
from gbpservice.common import utils as gbp_utils
from gbpservice.neutron.db.grouppolicy import group_policy_db as gpdb
from gbpservice.neutron.db.grouppolicy import group_policy_mapping_db
@ -78,7 +76,7 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
# plugins are loaded to grab and store plugin.
servicechain_plugin = directory.get_plugin(pconst.SERVICECHAIN)
if not servicechain_plugin:
LOG.error(_LE("No Servicechain service plugin found."))
LOG.error("No Servicechain service plugin found.")
raise gp_exc.GroupPolicyDeploymentError()
return servicechain_plugin
@ -484,8 +482,8 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
policy_context)
except Exception:
with excutils.save_and_reraise_exception():
LOG.exception(_LE("create_policy_target_postcommit "
"failed, deleting policy_target %s"),
LOG.exception("create_policy_target_postcommit "
"failed, deleting policy_target %s",
result['id'])
self.delete_policy_target(context, result['id'])
@ -536,8 +534,8 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
self.policy_driver_manager.delete_policy_target_postcommit(
policy_context)
except Exception:
LOG.exception(_LE("delete_policy_target_postcommit failed "
"for policy_target %s"),
LOG.exception("delete_policy_target_postcommit failed "
"for policy_target %s",
policy_target_id)
@log.log_method_call
@ -583,8 +581,8 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
policy_context)
except Exception:
with excutils.save_and_reraise_exception():
LOG.exception(_LE("create_policy_target_group_postcommit "
"failed, deleting policy_target_group %s"),
LOG.exception("create_policy_target_group_postcommit "
"failed, deleting policy_target_group %s",
result['id'])
self.delete_policy_target_group(context, result['id'])
@ -666,7 +664,7 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
self.delete_policy_target_group(
context, policy_target_group['proxy_group_id'])
except gpex.PolicyTargetGroupNotFound:
LOG.warning(_LW('PTG %s already deleted'),
LOG.warning('PTG %s already deleted',
policy_target_group['proxy_group_id'])
with session.begin(subtransactions=True):
@ -684,8 +682,8 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
self.policy_driver_manager.delete_policy_target_group_postcommit(
policy_context)
except Exception:
LOG.exception(_LE("delete_policy_target_group_postcommit failed "
"for policy_target_group %s"),
LOG.exception("delete_policy_target_group_postcommit failed "
"for policy_target_group %s",
policy_target_group_id)
@log.log_method_call
@ -731,8 +729,8 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
pdm.create_application_policy_group_postcommit(policy_context)
except Exception:
with excutils.save_and_reraise_exception():
LOG.exception(_LE("create_application_policy_group_postcommit "
"failed, deleting APG %s"),
LOG.exception("create_application_policy_group_postcommit "
"failed, deleting APG %s",
result['id'])
self.delete_application_policy_group(context, result['id'])
@ -792,8 +790,8 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
try:
pdm.delete_application_policy_group_postcommit(policy_context)
except Exception:
LOG.exception(_LE("delete_application_policy_group_postcommit "
"failed for application_policy_group %s"),
LOG.exception("delete_application_policy_group_postcommit "
"failed for application_policy_group %s",
application_policy_group_id)
@log.log_method_call
@ -836,8 +834,8 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
policy_context)
except Exception:
with excutils.save_and_reraise_exception():
LOG.exception(_LE("create_l2_policy_postcommit "
"failed, deleting l2_policy %s"),
LOG.exception("create_l2_policy_postcommit "
"failed, deleting l2_policy %s",
result['id'])
self.delete_l2_policy(context, result['id'])
@ -886,8 +884,8 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
self.policy_driver_manager.delete_l2_policy_postcommit(
policy_context)
except Exception:
LOG.exception(_LE("delete_l2_policy_postcommit failed "
"for l2_policy %s"), l2_policy_id)
LOG.exception("delete_l2_policy_postcommit failed "
"for l2_policy %s", l2_policy_id)
@log.log_method_call
@db_api.retry_if_session_inactive()
@ -932,9 +930,9 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
policy_context)
except Exception:
with excutils.save_and_reraise_exception():
LOG.exception(_LE(
LOG.exception(
"create_network_service_policy_postcommit "
"failed, deleting network_service_policy %s"),
"failed, deleting network_service_policy %s",
result['id'])
self.delete_network_service_policy(context, result['id'])
@ -991,9 +989,9 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
pdm = self.policy_driver_manager
pdm.delete_network_service_policy_postcommit(policy_context)
except Exception:
LOG.exception(_LE(
LOG.exception(
"delete_network_service_policy_postcommit failed "
"for network_service_policy %s"), network_service_policy_id)
"for network_service_policy %s", network_service_policy_id)
@log.log_method_call
@db_api.retry_if_session_inactive()
@ -1036,8 +1034,8 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
policy_context)
except Exception:
with excutils.save_and_reraise_exception():
LOG.exception(_LE("create_l3_policy_postcommit "
"failed, deleting l3_policy %s"),
LOG.exception("create_l3_policy_postcommit "
"failed, deleting l3_policy %s",
result['id'])
self.delete_l3_policy(context, result['id'])
@ -1091,8 +1089,8 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
self.policy_driver_manager.delete_l3_policy_postcommit(
policy_context)
except Exception:
LOG.exception(_LE("delete_l3_policy_postcommit failed "
"for l3_policy %s"), l3_policy_id)
LOG.exception("delete_l3_policy_postcommit failed "
"for l3_policy %s", l3_policy_id)
return True
@log.log_method_call
@ -1137,9 +1135,9 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
policy_context)
except Exception:
with excutils.save_and_reraise_exception():
LOG.exception(_LE(
LOG.exception(
"policy_driver_manager.create_policy_classifier_postcommit"
" failed, deleting policy_classifier %s"), result['id'])
" failed, deleting policy_classifier %s", result['id'])
self.delete_policy_classifier(context, result['id'])
return self.get_policy_classifier(context, result['id'])
@ -1188,8 +1186,8 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
self.policy_driver_manager.delete_policy_classifier_postcommit(
policy_context)
except Exception:
LOG.exception(_LE("delete_policy_classifier_postcommit failed "
"for policy_classifier %s"), id)
LOG.exception("delete_policy_classifier_postcommit failed "
"for policy_classifier %s", id)
@log.log_method_call
@db_api.retry_if_session_inactive()
@ -1233,9 +1231,9 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
policy_context)
except Exception:
with excutils.save_and_reraise_exception():
LOG.exception(_LE(
LOG.exception(
"policy_driver_manager.create_policy_action_postcommit "
"failed, deleting policy_action %s"), result['id'])
"failed, deleting policy_action %s", result['id'])
self.delete_policy_action(context, result['id'])
return self.get_policy_action(context, result['id'])
@ -1284,8 +1282,8 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
self.policy_driver_manager.delete_policy_action_postcommit(
policy_context)
except Exception:
LOG.exception(_LE("delete_policy_action_postcommit failed "
"for policy_action %s"), id)
LOG.exception("delete_policy_action_postcommit failed "
"for policy_action %s", id)
@log.log_method_call
@db_api.retry_if_session_inactive()
@ -1327,9 +1325,9 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
policy_context)
except Exception:
with excutils.save_and_reraise_exception():
LOG.exception(_LE(
LOG.exception(
"policy_driver_manager.create_policy_rule_postcommit"
" failed, deleting policy_rule %s"), result['id'])
" failed, deleting policy_rule %s", result['id'])
self.delete_policy_rule(context, result['id'])
return self.get_policy_rule(context, result['id'])
@ -1377,8 +1375,8 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
self.policy_driver_manager.delete_policy_rule_postcommit(
policy_context)
except Exception:
LOG.exception(_LE("delete_policy_rule_postcommit failed "
"for policy_rule %s"), id)
LOG.exception("delete_policy_rule_postcommit failed "
"for policy_rule %s", id)
@log.log_method_call
@db_api.retry_if_session_inactive()
@ -1421,9 +1419,9 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
policy_context)
except Exception:
with excutils.save_and_reraise_exception():
LOG.exception(_LE(
LOG.exception(
"policy_driver_manager.create_policy_rule_set_postcommit "
"failed, deleting policy_rule_set %s"), result['id'])
"failed, deleting policy_rule_set %s", result['id'])
self.delete_policy_rule_set(context, result['id'])
return self.get_policy_rule_set(context, result['id'])
@ -1471,8 +1469,8 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
self.policy_driver_manager.delete_policy_rule_set_postcommit(
policy_context)
except Exception:
LOG.exception(_LE("delete_policy_rule_set_postcommit failed "
"for policy_rule_set %s"), id)
LOG.exception("delete_policy_rule_set_postcommit failed "
"for policy_rule_set %s", id)
@log.log_method_call
@db_api.retry_if_session_inactive()
@ -1518,9 +1516,9 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
create_external_segment_postcommit(policy_context))
except Exception:
with excutils.save_and_reraise_exception():
LOG.exception(_LE("create_external_segment_postcommit "
"failed, deleting external_segment "
"%s"), result['id'])
LOG.exception("create_external_segment_postcommit "
"failed, deleting external_segment "
"%s", result['id'])
self.delete_external_segment(context, result['id'])
return self.get_external_segment(context, result['id'])
@ -1577,8 +1575,8 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
(self.policy_driver_manager.
delete_external_segment_postcommit(policy_context))
except Exception:
LOG.exception(_LE("delete_external_segment_postcommit failed "
"for external_segment %s"),
LOG.exception("delete_external_segment_postcommit failed "
"for external_segment %s",
external_segment_id)
return True
@ -1623,9 +1621,9 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
create_external_policy_postcommit(policy_context))
except Exception:
with excutils.save_and_reraise_exception():
LOG.exception(_LE("create_external_policy_postcommit "
"failed, deleting external_policy "
"%s"), result['id'])
LOG.exception("create_external_policy_postcommit "
"failed, deleting external_policy "
"%s", result['id'])
self.delete_external_policy(context, result['id'])
return self.get_external_policy(context, result['id'])
@ -1678,8 +1676,8 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
self.policy_driver_manager.delete_external_policy_postcommit(
policy_context)
except Exception:
LOG.exception(_LE("delete_external_policy_postcommit failed "
"for external_policy %s"), external_policy_id)
LOG.exception("delete_external_policy_postcommit failed "
"for external_policy %s", external_policy_id)
@log.log_method_call
@db_api.retry_if_session_inactive()
@ -1719,9 +1717,9 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
create_nat_pool_postcommit(policy_context))
except Exception:
with excutils.save_and_reraise_exception():
LOG.exception(_LE(
LOG.exception(
"create_nat_pool_postcommit failed, deleting "
"nat_pool %s"), result['id'])
"nat_pool %s", result['id'])
self.delete_nat_pool(context, result['id'])
return self.get_nat_pool(context, result['id'])
@ -1766,8 +1764,8 @@ class GroupPolicyPlugin(group_policy_mapping_db.GroupPolicyMappingDbPlugin):
self.policy_driver_manager.delete_nat_pool_postcommit(
policy_context)
except Exception:
LOG.exception(_LE("delete_nat_pool_postcommit failed "
"for nat_pool %s"),
LOG.exception("delete_nat_pool_postcommit failed "
"for nat_pool %s",
nat_pool_id)
@log.log_method_call

View File

@ -18,8 +18,6 @@ from oslo_policy import policy as oslo_policy
from oslo_utils import excutils
import stevedore
from gbpservice._i18n import _LE
from gbpservice._i18n import _LI
from gbpservice.neutron.services.grouppolicy.common import exceptions as gp_exc
from gbpservice.neutron.services.grouppolicy import group_policy_driver_api
@ -69,14 +67,14 @@ class PolicyDriverManager(stevedore.named.NamedExtensionManager):
self.ordered_policy_drivers = []
self.reverse_ordered_policy_drivers = []
LOG.info(_LI("Configured policy driver names: %s"),
LOG.info("Configured policy driver names: %s",
cfg.CONF.group_policy.policy_drivers)
super(PolicyDriverManager,
self).__init__('gbpservice.neutron.group_policy.policy_drivers',
cfg.CONF.group_policy.policy_drivers,
invoke_on_load=True,
name_order=True)
LOG.info(_LI("Loaded policy driver names: %s"), self.names())
LOG.info("Loaded policy driver names: %s", self.names())
self._register_policy_drivers()
def _register_policy_drivers(self):
@ -90,7 +88,7 @@ class PolicyDriverManager(stevedore.named.NamedExtensionManager):
self.ordered_policy_drivers.append(ext)
self.reverse_ordered_policy_drivers = self.ordered_policy_drivers[::-1]
LOG.info(_LI("Registered policy drivers: %s"),
LOG.info("Registered policy drivers: %s",
[driver.name for driver in self.ordered_policy_drivers])
def initialize(self):
@ -100,7 +98,7 @@ class PolicyDriverManager(stevedore.named.NamedExtensionManager):
# set it to True such that the drivers can override it.
self.native_bulk_support = False
for driver in self.ordered_policy_drivers:
LOG.info(_LI("Initializing policy driver '%s'"), driver.name)
LOG.info("Initializing policy driver '%s'", driver.name)
driver.obj.initialize()
self.native_bulk_support &= getattr(driver.obj,
'native_bulk_support', True)
@ -143,15 +141,15 @@ class PolicyDriverManager(stevedore.named.NamedExtensionManager):
e, oslo_policy.PolicyNotAuthorized):
with excutils.save_and_reraise_exception():
LOG.exception(
_LE("Policy driver '%(name)s' failed in"
" %(method)s"),
"Policy driver '%(name)s' failed in"
" %(method)s",
{'name': driver.name, 'method': method_name}
)
else:
error = True
# We are eating a non-GBP/non-Neutron exception here
LOG.exception(
_LE("Policy driver '%(name)s' failed in %(method)s"),
"Policy driver '%(name)s' failed in %(method)s",
{'name': driver.name, 'method': method_name})
if not continue_on_failure:
break
@ -173,8 +171,8 @@ class PolicyDriverManager(stevedore.named.NamedExtensionManager):
"ensure_tenant, operation will "
"be retried", {'driver': driver.name})
else:
LOG.exception(_LE("Policy driver '%s' failed in "
"ensure_tenant"), driver.name)
LOG.exception("Policy driver '%s' failed in "
"ensure_tenant", driver.name)
raise gp_exc.GroupPolicyDriverError(
method="ensure_tenant")

View File

@ -15,7 +15,6 @@ from oslo_config import cfg
from oslo_log import log as logging
import stevedore
from gbpservice._i18n import _LI
from gbpservice.neutron.services.servicechain.plugins.ncp import config # noqa
from gbpservice.neutron.services.servicechain.plugins.ncp import model
@ -33,13 +32,13 @@ class NodeDriverManager(stevedore.named.NamedExtensionManager):
# Ordered list of node drivers.
self.ordered_drivers = []
names = cfg.CONF.node_composition_plugin.node_drivers
LOG.info(_LI("Configured service chain node driver names: %s"), names)
LOG.info("Configured service chain node driver names: %s", names)
super(NodeDriverManager,
self).__init__(
'gbpservice.neutron.servicechain.ncp_drivers', names,
invoke_on_load=True, name_order=True)
LOG.info(_LI(
"Loaded service chain node driver names: %s"), self.names())
LOG.info(
"Loaded service chain node driver names: %s", self.names())
self._register_drivers()
def _register_drivers(self):
@ -47,14 +46,14 @@ class NodeDriverManager(stevedore.named.NamedExtensionManager):
for ext in self:
self.drivers[ext.name] = ext
self.ordered_drivers.append(ext)
LOG.info(_LI("Registered service chain node drivers: %s"),
LOG.info("Registered service chain node drivers: %s",
[driver.name for driver in self.ordered_drivers])
def initialize(self):
"""Initialize all the service chain node drivers."""
self.native_bulk_support = True
for driver in self.ordered_drivers:
LOG.info(_LI("Initializing service chain node drivers '%s'"),
LOG.info("Initializing service chain node drivers '%s'",
driver.name)
driver.obj.initialize(driver.name)
self.native_bulk_support &= getattr(driver.obj,

View File

@ -24,7 +24,6 @@ from oslo_log import log as logging
from oslo_serialization import jsonutils
import sqlalchemy as sa
from gbpservice._i18n import _LE
from gbpservice.neutron.services.servicechain.plugins.ncp import (
exceptions as exc)
from gbpservice.neutron.services.servicechain.plugins.ncp import driver_base
@ -374,15 +373,15 @@ class HeatNodeDriver(driver_base.NodeDriverBase):
'DELETE_IN_PROGRESS']:
return
except Exception:
LOG.exception(_LE("Retrieving the stack %(stack)s failed."),
LOG.exception("Retrieving the stack %(stack)s failed.",
{'stack': stack_id})
return
else:
time.sleep(STACK_ACTION_RETRY_WAIT)
time_waited = time_waited + STACK_ACTION_RETRY_WAIT
if time_waited >= STACK_ACTION_WAIT_TIME:
LOG.error(_LE("Stack %(action)s not completed within "
"%(wait)s seconds"),
LOG.error("Stack %(action)s not completed within "
"%(wait)s seconds",
{'action': action,
'wait': STACK_ACTION_WAIT_TIME,
'stack': stack_id})

View File

@ -31,9 +31,6 @@ from oslo_utils import excutils
import sqlalchemy as sa
from gbpservice._i18n import _
from gbpservice._i18n import _LE
from gbpservice._i18n import _LI
from gbpservice._i18n import _LW
from gbpservice.common import utils
from gbpservice.network.neutronv2 import local_api
from gbpservice.neutron.services.grouppolicy.common import constants as gconst
@ -166,9 +163,9 @@ class NFPClientApi(object):
self.client = n_rpc.get_client(target)
def create_network_function(self, context, network_function):
LOG.info(_LI("Sending RPC CREATE NETWORK FUNCTION to Service "
"Orchestrator for tenant:%(tenant_id)s with "
"service profile:%(service_profile_id)s"),
LOG.info("Sending RPC CREATE NETWORK FUNCTION to Service "
"Orchestrator for tenant:%(tenant_id)s with "
"service profile:%(service_profile_id)s",
{'tenant_id': network_function['tenant_id'],
'service_profile_id': network_function[
'service_profile']['id']})
@ -181,9 +178,9 @@ class NFPClientApi(object):
def delete_network_function(self, context, network_function_id,
network_function_data):
LOG.info(_LI("Sending RPC DELETE NETWORK FUNCTION to Service "
"Orchestrator for NF:"
"%(network_function_id)s"),
LOG.info("Sending RPC DELETE NETWORK FUNCTION to Service "
"Orchestrator for NF:"
"%(network_function_id)s",
{'network_function_id': network_function_id})
cctxt = self.client.prepare(version=self.RPC_API_VERSION)
return cctxt.cast(
@ -193,9 +190,9 @@ class NFPClientApi(object):
network_function_data=network_function_data)
def update_network_function(self, context, network_function_id, config):
LOG.info(_LI("Sending RPC UPDATE NETWORK FUNCTION to Service "
"Orchestrator for NF:"
"%(network_function_id)s"),
LOG.info("Sending RPC UPDATE NETWORK FUNCTION to Service "
"Orchestrator for NF:"
"%(network_function_id)s",
{'network_function_id': network_function_id})
cctxt = self.client.prepare(version=self.RPC_API_VERSION)
return cctxt.cast(
@ -215,9 +212,9 @@ class NFPClientApi(object):
def consumer_ptg_added_notification(self, context, network_function_id,
policy_target_group):
LOG.info(_LI("Sending RPC CONSUMER PTG ADDED NOTIFICATION to Service "
"Orchestrator for NF:"
"%(network_function_id)s"),
LOG.info("Sending RPC CONSUMER PTG ADDED NOTIFICATION to Service "
"Orchestrator for NF:"
"%(network_function_id)s",
{'network_function_id': network_function_id})
cctxt = self.client.prepare(version=self.RPC_API_VERSION)
return cctxt.cast(context,
@ -227,8 +224,8 @@ class NFPClientApi(object):
def consumer_ptg_removed_notification(self, context, network_function_id,
policy_target_group):
LOG.info(_LI("Sending RPC CONSUMER PTG REMOVED NOTIFICATION to "
" Service Orchestrator for NF:%(network_function_id)s"),
LOG.info("Sending RPC CONSUMER PTG REMOVED NOTIFICATION to "
" Service Orchestrator for NF:%(network_function_id)s",
{'network_function_id': network_function_id})
cctxt = self.client.prepare(version=self.RPC_API_VERSION)
return cctxt.cast(context,
@ -238,8 +235,8 @@ class NFPClientApi(object):
def policy_target_added_notification(self, context, network_function_id,
policy_target):
LOG.info(_LI("Sending RPC POLICY TARGET ADDED NOTIFICATION to "
"Service Orchestrator for NF:%(network_function_id)s"),
LOG.info("Sending RPC POLICY TARGET ADDED NOTIFICATION to "
"Service Orchestrator for NF:%(network_function_id)s",
{'network_function_id': network_function_id})
cctxt = self.client.prepare(version=self.RPC_API_VERSION)
return cctxt.cast(context,
@ -249,8 +246,8 @@ class NFPClientApi(object):
def policy_target_removed_notification(self, context, network_function_id,
policy_target):
LOG.info(_LI("Sending RPC POLICY TARGET REMOVED NOTIFICATION to "
"Service Orchestrator for NF:%(network_function_id)s"),
LOG.info("Sending RPC POLICY TARGET REMOVED NOTIFICATION to "
"Service Orchestrator for NF:%(network_function_id)s",
{'network_function_id': network_function_id})
cctxt = self.client.prepare(version=self.RPC_API_VERSION)
return cctxt.cast(context,
@ -259,7 +256,7 @@ class NFPClientApi(object):
policy_target=policy_target)
def get_plumbing_info(self, context, node_driver_ctxt):
LOG.info(_LI("Sending RPC GET PLUMBING INFO to Service Orchestrator "))
LOG.info("Sending RPC GET PLUMBING INFO to Service Orchestrator ")
request_info = dict(profile=node_driver_ctxt.current_profile,
tenant_id=node_driver_ctxt.provider['tenant_id'],
provider=node_driver_ctxt.provider)
@ -388,9 +385,9 @@ class NFPNodeDriver(driver_base.NodeDriverBase):
nfp_context = NFPContext.get_nfp_context(context.instance['id'])
if nfp_context:
if len(nfp_context['sc_gateway_type_nodes']):
LOG.info(_LI(
LOG.info(
"Not requesting plumber for PTs for service type "
"%(service_type)s"), {'service_type': service_type})
"%(service_type)s", {'service_type': service_type})
if not nfp_context['update']:
nfp_context['sc_gateway_type_nodes'].append(
gateway_type_node)
@ -421,9 +418,9 @@ class NFPNodeDriver(driver_base.NodeDriverBase):
plumbing_request = self.nfp_notifier.get_plumbing_info(
context._plugin_context, context)
LOG.info(_LI("Requesting plumber for PTs for "
"service type %(service_type)s with "
"%(plumbing_request)s "),
LOG.info("Requesting plumber for PTs for "
"service type %(service_type)s with "
"%(plumbing_request)s ",
{'plumbing_request': plumbing_request,
'service_type': service_type})
return plumbing_request
@ -508,11 +505,11 @@ class NFPNodeDriver(driver_base.NodeDriverBase):
context.plugin_session, context.current_node['id'],
context.instance['id'], network_function_id,
status, status_details)
LOG.info(_LI("Processed create NF in node driver."
LOG.info("Processed create NF in node driver."
"servicechain_instance_id: %(sci_id)s, "
"servicechain_node_id: %(scn_id)s"), {
'sci_id': context.instance['id'],
'scn_id': context.current_node['id']})
"servicechain_node_id: %(scn_id)s", {
'sci_id': context.instance['id'],
'scn_id': context.current_node['id']})
def _wait_for_node_operation_completion(self, context, network_function_id,
operation):
@ -618,7 +615,7 @@ class NFPNodeDriver(driver_base.NodeDriverBase):
self._delete_network_function(context, network_function_id)
except Exception:
# NFPContext.clear_nfp_context(context.instance['id'])
LOG.exception(_LE("Delete Network service Failed"))
LOG.exception("Delete Network service Failed")
exc_type, exc_value, exc_traceback = sys.exc_info()
message = "Traceback: %s" % (exc_value)
LOG.error(message)
@ -772,8 +769,8 @@ class NFPNodeDriver(driver_base.NodeDriverBase):
time_waited = time_waited + 5
if network_function:
LOG.error(_LE("Delete network function %(network_function)s "
"failed"),
LOG.error("Delete network function %(network_function)s "
"failed",
{'network_function': network_function_id})
raise NodeInstanceDeleteFailed()
@ -794,38 +791,38 @@ class NFPNodeDriver(driver_base.NodeDriverBase):
"time waited: %s", (network_function_id, operation,
time_waited, network_function['status']))
if not network_function:
LOG.error(_LE("Failed to retrieve network function"))
LOG.error("Failed to retrieve network function")
eventlet.sleep(5)
time_waited = time_waited + 5
continue
else:
if time_waited == 0:
LOG.info(_LI("STARTED POLLING for %(operation)s network "
"function for NF:%(network_function_id)s "
"with initial result: %(result)s "),
LOG.info("STARTED POLLING for %(operation)s network "
"function for NF:%(network_function_id)s "
"with initial result: %(result)s ",
{'operation': operation,
'network_function_id': network_function_id,
'result': network_function})
if (network_function['status'] == nfp_constants.ACTIVE or
network_function['status'] == nfp_constants.ERROR):
LOG.info(_LI("COMPLETED POLLING for %(operation)s network "
"function for NF:%(network_function_id)s "),
LOG.info("COMPLETED POLLING for %(operation)s network "
"function for NF:%(network_function_id)s ",
{'network_function_id': network_function_id,
'operation': operation})
break
eventlet.sleep(5)
time_waited = time_waited + 5
LOG.info(_LI("Got %(operation)s network function result for NF:"
"%(network_function_id)s with status:%(status)s"),
LOG.info("Got %(operation)s network function result for NF:"
"%(network_function_id)s with status:%(status)s",
{'network_function_id': network_function_id,
'operation': operation,
'status': network_function['status']})
if network_function['status'] != nfp_constants.ACTIVE:
LOG.error(_LE("%(operation)s network function:"
"%(network_function)s "
"failed. Status: %(status)s"),
LOG.error("%(operation)s network function:"
"%(network_function)s "
"failed. Status: %(status)s",
{'network_function': network_function_id,
'status': network_function['status'],
'operation': operation})
@ -852,11 +849,11 @@ class NFPNodeDriver(driver_base.NodeDriverBase):
return tenant.id
except k_exceptions.NotFound:
with excutils.save_and_reraise_exception(reraise=True):
LOG.error(_LE('No tenant with name %(tenant)s exists.'),
LOG.error('No tenant with name %(tenant)s exists.',
{'tenant': tenant})
except k_exceptions.NoUniqueMatch:
with excutils.save_and_reraise_exception(reraise=True):
LOG.error(_LE('Multiple tenants matches found for %(tenant)s'),
LOG.error('Multiple tenants matches found for %(tenant)s',
{'tenant': tenant})
def _get_resource_owner_context(self, plugin_context):
@ -890,11 +887,11 @@ class NFPNodeDriver(driver_base.NodeDriverBase):
network_function_id,
context.current_node['config']])
except Exception:
LOG.exception(_LE("Update Network service Failed for "
"network function: %(nf_id)s"),
LOG.exception("Update Network service Failed for "
"network function: %(nf_id)s",
{'nf_id': network_function_id})
else:
LOG.info(_LI("No action to take on update"))
LOG.info("No action to take on update")
def _get_service_chain_specs(self, context):
current_specs = context.relevant_specs
@ -970,8 +967,8 @@ class NFPNodeDriver(driver_base.NodeDriverBase):
if (service_details['device_type'] != 'None' and (
not provider_service_targets or (service_type in
[pconst.FIREWALL, pconst.VPN] and not consumer_service_targets))):
LOG.error(_LE("Service Targets are not created for the Node "
"of service_type %(service_type)s"),
LOG.error("Service Targets are not created for the Node "
"of service_type %(service_type)s",
{'service_type': service_type})
raise Exception(_("Service Targets are not created "
"for the Node"))
@ -1097,8 +1094,8 @@ class NFPNodeDriver(driver_base.NodeDriverBase):
break
if not redirect_prs:
LOG.error(_LE("Redirect rule doesn't exist in policy target rule "
" set"))
LOG.error("Redirect rule doesn't exist in policy target rule "
" set")
return consuming_ptgs_details, consuming_eps_details
consuming_ptg_ids = redirect_prs['consuming_policy_target_groups']
@ -1224,9 +1221,9 @@ class NFPNodeDriver(driver_base.NodeDriverBase):
def _create_network_function(self, context):
nfp_create_nf_data = self._get_nfp_network_function(context)
LOG.info(_LI("Received Call CREATE NETWORK FUNCTION for tenant: "
"%(tenant_id)s with service profile:"
"%(service_profile)s"),
LOG.info("Received Call CREATE NETWORK FUNCTION for tenant: "
"%(tenant_id)s with service profile:"
"%(service_profile)s",
{'tenant_id': nfp_create_nf_data['tenant_id'],
'service_profile': nfp_create_nf_data['service_profile']})
self._queue_notification(context, 'create_network_function',
@ -1239,9 +1236,9 @@ class NFPNodeDriver(driver_base.NodeDriverBase):
if nfp_delete_nf_data['consumer'].get('pt'):
self._detach_port_from_pts(context,
nfp_delete_nf_data['consumer']['pt'])
LOG.info(_LI("Received Call DELETE NETWORK FUNCTION for tenant: "
"%(tenant_id)s with service profile:"
"%(service_profile)s"),
LOG.info("Received Call DELETE NETWORK FUNCTION for tenant: "
"%(tenant_id)s with service profile:"
"%(service_profile)s",
{'tenant_id': nfp_delete_nf_data['tenant_id'],
'service_profile': nfp_delete_nf_data['service_profile']})
self._queue_notification(context, 'delete_network_function',
@ -1261,8 +1258,8 @@ class NFPNodeDriver(driver_base.NodeDriverBase):
pt['port_id']))
except Exception:
LOG.warning(_LW("Failed to disassociate port from"
" pt: %(pt)s, Error: %(exc)s"), {'pt': pt, 'exc': exc})
LOG.warning("Failed to disassociate port from"
" pt: %(pt)s, Error: %(exc)s", {'pt': pt, 'exc': exc})
def _update_ptg(self, context):
if hasattr(context, 'provider') and context.provider['description']:

View File

@ -14,7 +14,6 @@ from heatclient import client as heat_client
from heatclient import exc as heat_exc
from oslo_log import log as logging
from gbpservice._i18n import _LW
LOG = logging.getLogger(__name__)
@ -56,9 +55,9 @@ class HeatClient(object):
try:
self.stacks.delete(stack_id)
except heat_exc.HTTPNotFound:
LOG.warning(_LW(
LOG.warning(
"Stack %(stack)s created by service chain driver is "
"not found at cleanup"), {'stack': stack_id})
"not found at cleanup", {'stack': stack_id})
def get(self, stack_id):
return self.stacks.get(stack_id)

View File

@ -16,7 +16,6 @@ from oslo_config import cfg
from oslo_log import log as logging
from oslo_utils import excutils
from gbpservice._i18n import _LE
from gbpservice.common import utils
from gbpservice.neutron.services.servicechain.plugins.ncp.node_plumbers import(
traffic_stitching_plumber as tscp)
@ -77,10 +76,10 @@ class AdminOwnedResourcesApicTSCP(tscp.TrafficStitchingPlumber):
return tenant.id
except k_exceptions.NotFound:
with excutils.save_and_reraise_exception(reraise=True):
LOG.error(_LE('No tenant with name %s exists.'), tenant)
LOG.error('No tenant with name %s exists.', tenant)
except k_exceptions.NoUniqueMatch:
with excutils.save_and_reraise_exception(reraise=True):
LOG.error(_LE('Multiple tenants matches found for %s'), tenant)
LOG.error('Multiple tenants matches found for %s', tenant)
def _get_resource_owner_context(self, context):
resource_owner_context = context.elevated()

View File

@ -15,9 +15,6 @@ from neutron_lib.plugins import directory
from oslo_config import cfg
from oslo_log import log as logging
from gbpservice._i18n import _LE
from gbpservice._i18n import _LI
from gbpservice._i18n import _LW
from gbpservice.neutron.extensions import driver_proxy_group as pg_ext
from gbpservice.neutron.extensions import group_policy
from gbpservice.neutron.services.grouppolicy.common import exceptions as exc
@ -47,8 +44,8 @@ class TrafficStitchingPlumber(plumber_base.NodePlumberBase):
# Verify that proxy_group extension is loaded
if pg_ext.PROXY_GROUP not in cfg.CONF.group_policy.extension_drivers:
LOG.error(_LE("proxy_group GBP driver extension is mandatory for "
"traffic stitching plumber."))
LOG.error("proxy_group GBP driver extension is mandatory for "
"traffic stitching plumber.")
raise exc.GroupPolicyDeploymentError()
@property
@ -82,7 +79,7 @@ class TrafficStitchingPlumber(plumber_base.NodePlumberBase):
management, 'management')
# Create proper PTs based on the service type
jump_ptg = None
LOG.info(_LI("Plumbing service of type '%s'"),
LOG.info("Plumbing service of type '%s'",
info['plumbing_type'])
if info['plumbing_type'] == common.PLUMBING_TYPE_ENDPOINT:
# No stitching needed, only provider side PT is created.
@ -124,7 +121,7 @@ class TrafficStitchingPlumber(plumber_base.NodePlumberBase):
context, part_context, info['consumer'],
jump_ptg, 'consumer')
else:
LOG.warning(_LW("Unsupported plumbing type %s"),
LOG.warning("Unsupported plumbing type %s",
info['plumbing_type'])
# Replace current "provider" with jump ptg if needed
provider = jump_ptg or provider

View File

@ -18,9 +18,6 @@ from oslo_log import helpers as log
from oslo_log import log as logging
from oslo_utils import excutils
from gbpservice._i18n import _LE
from gbpservice._i18n import _LI
from gbpservice._i18n import _LW
from gbpservice.common import utils
from gbpservice.neutron.db import servicechain_db
from gbpservice.neutron.services.grouppolicy.common import constants as gp_cts
@ -66,7 +63,7 @@ class NodeCompositionPlugin(servicechain_db.ServiceChainDbPlugin,
self.plumber = utils.load_plugin(
PLUMBER_NAMESPACE, plumber_klass)
self.plumber.initialize()
LOG.info(_LI("Initialized node plumber '%s'"), plumber_klass)
LOG.info("Initialized node plumber '%s'", plumber_klass)
@log.log_method_call
def create_servicechain_instance(self, context, servicechain_instance):
@ -96,8 +93,8 @@ class NodeCompositionPlugin(servicechain_db.ServiceChainDbPlugin,
except Exception:
# Some node could not be deployed
with excutils.save_and_reraise_exception():
LOG.error(_LE("Node deployment failed, "
"deleting servicechain_instance %s"),
LOG.error("Node deployment failed, "
"deleting servicechain_instance %s",
instance['id'])
self.delete_servicechain_instance(context, instance['id'])
@ -127,8 +124,8 @@ class NodeCompositionPlugin(servicechain_db.ServiceChainDbPlugin,
except Exception:
# Some node could not be deployed
with excutils.save_and_reraise_exception():
LOG.error(_LE("Node deployment failed, "
"servicechain_instance %s is in ERROR state"),
LOG.error("Node deployment failed, "
"servicechain_instance %s is in ERROR state",
instance['id'])
@log.log_method_call
@ -257,7 +254,7 @@ class NodeCompositionPlugin(servicechain_db.ServiceChainDbPlugin,
try:
update['driver'].update(update['context'])
except exc.NodeDriverError as ex:
LOG.error(_LE("Node Update failed, %s"),
LOG.error("Node Update failed, %s",
ex.message)
return updated_sc_node
@ -398,8 +395,8 @@ class NodeCompositionPlugin(servicechain_db.ServiceChainDbPlugin,
old_policy_target_group,
current_policy_target_group)
except exc.NodeDriverError as ex:
LOG.error(_LE("Node Update on policy target group modification"
" failed, %s"), ex.message)
LOG.error("Node Update on policy target group modification"
" failed, %s", ex.message)
def _update_chains_pt_modified(self, context, policy_target, instance_id,
action):
@ -412,8 +409,8 @@ class NodeCompositionPlugin(servicechain_db.ServiceChainDbPlugin,
'update_policy_target_' + action)(
update['context'], policy_target)
except exc.NodeDriverError as ex:
LOG.error(_LE("Node Update on policy target modification "
"failed, %s"), ex.message)
LOG.error("Node Update on policy target modification "
"failed, %s", ex.message)
def _update_chains_consumer_modified(self, context, policy_target_group,
instance_id, action):
@ -426,9 +423,9 @@ class NodeCompositionPlugin(servicechain_db.ServiceChainDbPlugin,
'update_node_consumer_ptg_' + action)(
update['context'], policy_target_group)
except exc.NodeDriverError as ex:
LOG.error(_LE(
LOG.error(
"Node Update on policy target group modification "
"failed, %s"), ex.message)
"failed, %s", ex.message)
def notify_chain_parameters_updated(self, context,
servicechain_instance_id):
@ -447,8 +444,8 @@ class NodeCompositionPlugin(servicechain_db.ServiceChainDbPlugin,
getattr(update['driver'],
'notify_chain_parameters_updated')(update['context'])
except exc.NodeDriverError as ex:
LOG.error(_LE("Node Update on GBP parameter update "
"failed, %s"), ex.message)
LOG.error("Node Update on GBP parameter update "
"failed, %s", ex.message)
def _get_instance_nodes(self, context, instance):
context = utils.admin_context(context)
@ -500,7 +497,7 @@ class NodeCompositionPlugin(servicechain_db.ServiceChainDbPlugin,
deployers = self._get_scheduled_drivers(context, resource,
'get')
except Exception:
LOG.warning(_LW("Failed to get node driver"))
LOG.warning("Failed to get node driver")
# Invoke drivers only if status attributes are requested
if not fields or STATUS_SET.intersection(set(fields)):
@ -548,8 +545,8 @@ class NodeCompositionPlugin(servicechain_db.ServiceChainDbPlugin,
result['status'] = 'ACTIVE'
result['status_details'] = 'node deployment completed'
except Exception as exc:
LOG.error(_LE("Failed to get servicechain instance status "
"from node driver, Error: %(exc)s"), {'exc': exc})
LOG.error("Failed to get servicechain instance status "
"from node driver, Error: %(exc)s", {'exc': exc})
return
return result
result = {'status': 'ACTIVE', 'status_details': ''}
@ -574,7 +571,7 @@ class NodeCompositionPlugin(servicechain_db.ServiceChainDbPlugin,
try:
driver.delete(destroy['context'])
except exc.NodeDriverError:
LOG.error(_LE("Node destroy failed, for node %s "),
LOG.error("Node destroy failed, for node %s ",
driver['context'].current_node['id'])
except Exception as e:
if db_api.is_retriable(e):

View File

@ -14,7 +14,6 @@ from neutron.plugins.common import constants as pconst
from neutron_lib.plugins import directory
from oslo_log import log as logging
from gbpservice._i18n import _LE
from gbpservice.neutron.services.grouppolicy.common import exceptions as gp_exc
from gbpservice.neutron.services.grouppolicy import plugin as gbp_plugin
@ -42,7 +41,7 @@ class SharingMixin(object):
# plugins are loaded to grab and store plugin.
gbp_plugin = directory.get_plugin(pconst.GROUP_POLICY)
if not gbp_plugin:
LOG.error(_LE("No group policy service plugin found."))
LOG.error("No group policy service plugin found.")
raise gp_exc.GroupPolicyDeploymentError()
return gbp_plugin

View File

@ -9,7 +9,6 @@
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron._i18n import _LI
from neutron.agent import securitygroups_rpc
from neutron.api import extensions
from neutron.quota import resource
@ -21,11 +20,6 @@ from gbpservice.network.neutronv2 import local_api
# The following is to avoid excessive logging in the UTs
extensions._LW = extensions._LI
l3_agent_scheduler._LW = _LI
securitygroups_rpc._LW = securitygroups_rpc._LI
resource_registry._LW = resource_registry._LI
local_api._LW = _LI
extensions.LOG.warning = extensions.LOG.info
resource_registry.LOG.warning = resource_registry.LOG.info
l3_agent_scheduler.LOG.warning = l3_agent_scheduler.LOG.info

View File

@ -27,11 +27,8 @@ if not hasattr(sa_utils, '_get_unique_keys'):
sa_utils._get_unique_keys = sa_utils.get_unique_keys
from neutron._i18n import _LI
from gbpservice.neutron.plugins.ml2plus.drivers.apic_aim import cache
# The following is to avoid excessive logging in the UTs
cache._LW = _LI
cache.LOG.warning = cache.LOG.info

View File

@ -17,7 +17,6 @@ from neutron.tests.unit.plugins.ml2.drivers import (
mechanism_logger as ml2_logger)
from oslo_log import log
from gbpservice._i18n import _LI
from gbpservice.neutron.plugins.ml2plus import driver_api
LOG = log.getLogger(__name__)
@ -31,10 +30,10 @@ class LoggerPlusMechanismDriver(driver_api.MechanismDriver,
"""
def initialize(self):
LOG.info(_LI("initialize called"))
LOG.info("initialize called")
def ensure_tenant(self, plugin_context, tenant_id):
LOG.info(_LI("ensure_tenant called with tenant_id %s"), tenant_id)
LOG.info("ensure_tenant called with tenant_id %s", tenant_id)
def _log_subnetpool_call(self, method_name, context):
LOG.info(_("%(method)s called with subnetpool settings %(current)s "

View File

@ -18,7 +18,6 @@ from oslo_config import cfg
from oslo_log import log as logging
import webob.exc
from gbpservice._i18n import _LW
from gbpservice.neutron.db.grouppolicy import group_policy_mapping_db as gpmdb
from gbpservice.neutron.extensions import group_policy as gpolicy
from gbpservice.neutron.services.grouppolicy import config
@ -164,7 +163,7 @@ class GroupPolicyPluginTestCase(GroupPolicyPluginTestBase):
def tearDown(self):
policy_drivers = cfg.CONF.group_policy.policy_drivers
LOG.warning(_LW("PDs used in this test: %s"),
LOG.warning("PDs used in this test: %s",
policy_drivers)
# Always reset configuration to dummy driver. Any
# test which requires to configure a different

View File

@ -17,7 +17,6 @@ from oslo_config import cfg
from oslo_log import log as logging
from gbpservice._i18n import _
from gbpservice._i18n import _LE
LOG = logging.getLogger(__name__)
@ -66,9 +65,9 @@ class NFPException(Exception):
exc_info = sys.exc_info()
# kwargs doesn't match a variable in the message
# log the issue and the kwargs
LOG.exception(_LE('Exception in string format operation'))
LOG.exception('Exception in string format operation')
for name, value in kwargs.items():
LOG.error(_LE("%(name)s: %(value)s"),
LOG.error("%(name)s: %(value)s",
{'name': name, 'value': value})
if CONF.fatal_exception_format_errors:
six.reraise(*exc_info)

View File

@ -12,7 +12,6 @@
from heatclient import client as heat_client
from heatclient import exc as heat_exc
from gbpservice._i18n import _LW
from gbpservice.nfp.core import log as nfp_logging
LOG = nfp_logging.getLogger(__name__)
@ -64,8 +63,8 @@ class HeatClient(object):
try:
self.stacks.delete(stack_id)
except heat_exc.HTTPNotFound:
LOG.warning(_LW("Stack %(stack)s created by service chain driver "
"is not found at cleanup"), {'stack': stack_id})
LOG.warning("Stack %(stack)s created by service chain driver "
"is not found at cleanup", {'stack': stack_id})
def get(self, stack_id):
return self.stacks.get(stack_id)

View File

@ -21,9 +21,6 @@ from oslo_config import cfg
from oslo_serialization import jsonutils
import yaml
from gbpservice._i18n import _LE
from gbpservice._i18n import _LI
from gbpservice._i18n import _LW
from gbpservice.neutron.services.grouppolicy.common import constants as gconst
from gbpservice.neutron.services.servicechain.plugins.ncp import plumber_base
from gbpservice.nfp.common import constants as nfp_constants
@ -205,7 +202,7 @@ class HeatDriver(object):
try:
self._assign_admin_user_to_project(tenant_id)
except Exception:
LOG.exception(_LE("Failed to assign admin user to project"))
LOG.exception("Failed to assign admin user to project")
return None
'''
nfp_context = module_context.get()
@ -223,7 +220,7 @@ class HeatDriver(object):
auth_token=auth_token,
timeout_mins=timeout_mins)
except Exception:
LOG.exception(_LE("Failed to create heatclient object"))
LOG.exception("Failed to create heatclient object")
return None
return heat_client
@ -318,8 +315,8 @@ class HeatDriver(object):
provider_subnet = subnet
break
if not provider_subnet:
LOG.error(_LE("Unable to get provider subnet for provider "
"policy target group %(provider_ptg)s"),
LOG.error("Unable to get provider subnet for provider "
"policy target group %(provider_ptg)s",
{"provider_ptg": provider})
return lb_vip, lb_vip_name
if service_type == pconst.LOADBALANCERV2:
@ -520,8 +517,8 @@ class HeatDriver(object):
break
if not redirect_prs:
LOG.error(_LE("Redirect rule doesn't exist in policy target rule "
" set"))
LOG.error("Redirect rule doesn't exist in policy target rule "
" set")
return None, None
return (redirect_prs['consuming_policy_target_groups'],
redirect_prs['consuming_external_policies'])
@ -663,8 +660,8 @@ class HeatDriver(object):
provider_cidr = subnet['cidr']
break
if not provider_cidr:
LOG.error(_LE("Unable to get provider cidr for provider "
"policy target group %(provider_ptg)s"),
LOG.error("Unable to get provider cidr for provider "
"policy target group %(provider_ptg)s",
{"provider_ptg": provider})
return None
@ -747,7 +744,7 @@ class HeatDriver(object):
svc_mgmt_ptgs = gcm.retry(self.gbp_client.get_policy_target_groups,
auth_token, filters)
if not svc_mgmt_ptgs:
LOG.error(_LE("Service Management Group is not created by Admin"))
LOG.error("Service Management Group is not created by Admin")
return None
else:
mgmt_subnet_id = svc_mgmt_ptgs[0]['subnets'][0]
@ -843,8 +840,8 @@ class HeatDriver(object):
auth_token,
filters={'port_id': [consumer_port['id']]})
if not stitching_pts:
LOG.error(_LE("Policy target is not created for the "
"stitching port"))
LOG.error("Policy target is not created for the "
"stitching port")
return None
stitching_ptg_id = (
stitching_pts[0]['policy_target_group_id'])
@ -901,9 +898,9 @@ class HeatDriver(object):
stack_template_str.startswith('{') else
yaml.load(stack_template_str))
except Exception:
LOG.error(_LE(
LOG.error(
"Unable to load stack template for service chain "
"node: %(node_id)s"), {'node_id': service_chain_node})
"node: %(node_id)s", {'node_id': service_chain_node})
return None, None
config_param_values = service_chain_instance.get(
'config_param_values', '{}')
@ -911,7 +908,7 @@ class HeatDriver(object):
try:
config_param_values = jsonutils.loads(config_param_values)
except Exception:
LOG.error(_LE("Unable to load config parameters"))
LOG.error("Unable to load config parameters")
return None, None
is_template_aws_version = stack_template.get(
@ -1010,8 +1007,8 @@ class HeatDriver(object):
if parameter in config_param_values:
stack_params[parameter] = config_param_values[parameter]
LOG.info(_LI('Final stack_template : %(stack_data)s, '
'stack_params : %(params)s'),
LOG.info('Final stack_template : %(stack_data)s, '
'stack_params : %(params)s',
{'stack_data': stack_template, 'params': stack_params})
return (stack_template, stack_params)
@ -1021,9 +1018,9 @@ class HeatDriver(object):
self.neutron_client.get_networks,
token, filters={'name': [INTERNET_OUT_EXT_NET_NAME]})
if not ext_net:
LOG.error(_LE("'internet_out_network_name' not configured"
" in [heat_driver] or Network %(network)s is"
" not found"),
LOG.error("'internet_out_network_name' not configured"
" in [heat_driver] or Network %(network)s is"
" not found",
{'network': INTERNET_OUT_EXT_NET_NAME})
return None
# There is a case where consumer port has multiple fips
@ -1035,8 +1032,8 @@ class HeatDriver(object):
return ncm.retry(self.neutron_client.get_floating_ips, token,
**filters)[0]['floating_ip_address']
except Exception:
LOG.error(_LE("Floating IP for VPN Service has either exhausted"
" or has been disassociated Manually"))
LOG.error("Floating IP for VPN Service has either exhausted"
" or has been disassociated Manually")
return None
def _update_node_config(self, auth_token, tenant_id, service_profile,
@ -1057,7 +1054,7 @@ class HeatDriver(object):
provider_subnet = subnet
break
if not provider_cidr:
LOG.error(_LE("No provider cidr availabale"))
LOG.error("No provider cidr availabale")
return None, None
service_type = service_profile['service_type']
service_details = transport.parse_service_flavor_string(
@ -1072,9 +1069,9 @@ class HeatDriver(object):
stack_template_str.startswith('{') else
yaml.load(stack_template_str))
except Exception:
LOG.error(_LE(
LOG.error(
"Unable to load stack template for service chain "
"node: %(node_id)s"), {'node_id': service_chain_node})
"node: %(node_id)s", {'node_id': service_chain_node})
return None, None
config_param_values = service_chain_instance.get(
'config_param_values', '{}')
@ -1082,7 +1079,7 @@ class HeatDriver(object):
try:
config_param_values = jsonutils.loads(config_param_values)
except Exception:
LOG.error(_LE("Unable to load config parameters"))
LOG.error("Unable to load config parameters")
return None, None
is_template_aws_version = stack_template.get(
@ -1200,8 +1197,8 @@ class HeatDriver(object):
auth_token,
filters={'port_id': [consumer_port['id']]})
if not stitching_pts:
LOG.error(_LE("Policy target is not created for the "
"stitching port"))
LOG.error("Policy target is not created for the "
"stitching port")
return None, None
stitching_ptg_id = (
stitching_pts[0]['policy_target_group_id'])
@ -1219,9 +1216,9 @@ class HeatDriver(object):
auth_token,
filters={'name': [INTERNET_OUT_EXT_NET_NAME]})
if not ext_net:
LOG.error(_LE("'internet_out_network_name' not configured"
" in [heat_driver] or Network %(network)s is"
" not found"),
LOG.error("'internet_out_network_name' not configured"
" in [heat_driver] or Network %(network)s is"
" not found",
{'network': INTERNET_OUT_EXT_NET_NAME})
return None, None
filters = {'port_id': [consumer_port['id']],
@ -1231,8 +1228,8 @@ class HeatDriver(object):
self.neutron_client.get_floating_ips,
auth_token, filters=filters)
if not floatingips:
LOG.error(_LE("Floating IP for VPN Service has been "
"disassociated Manually"))
LOG.error("Floating IP for VPN Service has been "
"disassociated Manually")
return None, None
for fip in floatingips:
if consumer_port['fixed_ips'][0]['ip_address'] == fip[
@ -1253,9 +1250,9 @@ class HeatDriver(object):
';mgmt_gw_ip=' + mgmt_gw_ip +
';network_function_id=' + network_function['id'])
except Exception as e:
LOG.error(_LE("Problem in preparing description, some of "
"the fields might not have initialized. "
"Error: %(error)s"), {'error': e})
LOG.error("Problem in preparing description, some of "
"the fields might not have initialized. "
"Error: %(error)s", {'error': e})
return None, None
siteconn_keys = self._get_site_conn_keys(
stack_template[resources_key],
@ -1287,8 +1284,8 @@ class HeatDriver(object):
if parameter in config_param_values:
stack_params[parameter] = config_param_values[parameter]
LOG.info(_LI('Final stack_template : %(stack_data)s, '
'stack_params : %(params)s'),
LOG.info('Final stack_template : %(stack_data)s, '
'stack_params : %(params)s',
{'stack_data': stack_template, 'params': stack_params})
return (stack_template, stack_params)
@ -1392,7 +1389,7 @@ class HeatDriver(object):
admin_token,
policy_target['policy_target_group_id']))
elif port_classification == nfp_constants.PROVIDER:
LOG.info(_LI("provider info: %(p_info)s"),
LOG.info("provider info: %(p_info)s",
{'p_info': port_id})
with nfp_ctx_mgr.NeutronContextManager as ncm:
provider_port = ncm.retry(self.neutron_client.get_port,
@ -1438,7 +1435,7 @@ class HeatDriver(object):
elif stack.stack_status == 'CREATE_COMPLETE':
return
elif stack.stack_status == 'DELETE_COMPLETE':
LOG.info(_LI("Stack %(stack)s is deleted"),
LOG.info("Stack %(stack)s is deleted",
{'stack': stack_id})
if action == "delete":
return
@ -1453,17 +1450,17 @@ class HeatDriver(object):
'DELETE_IN_PROGRESS']:
return
except heat_exc.HTTPNotFound:
LOG.warning(_LW(
LOG.warning(
"Stack %(stack)s created by service chain "
"driver is not found while waiting for %(action)s "
"to complete"),
"to complete",
{'stack': stack_id, 'action': action})
if action == "create" or action == "update":
operation_failed = True
else:
return
except Exception:
LOG.exception(_LE("Retrieving the stack %(stack)s failed."),
LOG.exception("Retrieving the stack %(stack)s failed.",
{'stack': stack_id})
if action == "create" or action == "update":
operation_failed = True
@ -1474,8 +1471,8 @@ class HeatDriver(object):
if ignore_error:
return
else:
LOG.error(_LE("Stack %(stack_name)s %(action)s failed for "
"tenant %(stack_owner)s"),
LOG.error("Stack %(stack_name)s %(action)s failed for "
"tenant %(stack_owner)s",
{'stack_name': stack.stack_name,
'stack_owner': stack.stack_owner,
'action': action})
@ -1484,8 +1481,8 @@ class HeatDriver(object):
time.sleep(STACK_ACTION_RETRY_WAIT)
time_waited = time_waited + STACK_ACTION_RETRY_WAIT
if time_waited >= wait_timeout:
LOG.error(_LE("Stack %(action)s not completed within "
"%(wait)s seconds"),
LOG.error("Stack %(action)s not completed within "
"%(wait)s seconds",
{'action': action,
'wait': wait_timeout,
'stack': stack_id})
@ -1499,10 +1496,10 @@ class HeatDriver(object):
pass
return
else:
LOG.error(_LE(
LOG.error(
"Stack %(stack_name)s %(action)s not "
"completed within %(time)s seconds where "
"stack owner is %(stack_owner)s"),
"stack owner is %(stack_owner)s",
{'stack_name': stack.stack_name,
'action': action,
'time': wait_timeout,
@ -1529,7 +1526,7 @@ class HeatDriver(object):
elif stack.stack_status == 'UPDATE_COMPLETE':
return success_status
elif stack.stack_status == 'DELETE_COMPLETE':
LOG.info(_LI("Stack %(stack)s is deleted"),
LOG.info("Stack %(stack)s is deleted",
{'stack': stack_id})
return failure_status
elif stack.stack_status == 'CREATE_FAILED':
@ -1562,7 +1559,7 @@ class HeatDriver(object):
elif stack.stack_status == 'UPDATE_COMPLETE':
return success_status
elif stack.stack_status == 'DELETE_COMPLETE':
LOG.info(_LI("Stack %(stack)s is deleted"),
LOG.info("Stack %(stack)s is deleted",
{'stack': stack_id})
return failure_status
elif stack.stack_status == 'CREATE_FAILED':
@ -1589,7 +1586,7 @@ class HeatDriver(object):
elif stack.stack_status == 'CREATE_COMPLETE':
return failure_status
elif stack.stack_status == 'DELETE_COMPLETE':
LOG.info(_LI("Stack %(stack)s is deleted"),
LOG.info("Stack %(stack)s is deleted",
{'stack': stack_id})
if network_function:
self._post_stack_cleanup(network_function)
@ -1683,8 +1680,8 @@ class HeatDriver(object):
stack = hcm.retry(heatclient.create, stack_name,
stack_template, stack_params)
stack_id = stack['stack']['id']
LOG.info(_LI("Created stack with ID %(stack_id)s and "
"name %(stack_name)s for provider PTG %(provider)s"),
LOG.info("Created stack with ID %(stack_id)s and "
"name %(stack_name)s for provider PTG %(provider)s",
{'stack_id': stack_id, 'stack_name': stack_name,
'provider': provider['id']})
@ -1735,8 +1732,8 @@ class HeatDriver(object):
stack_template, stack_params)
stack_id = stack['stack']['id']
LOG.info(_LI("Created stack with ID %(stack_id)s and "
"name %(stack_name)s for provider PTG %(provider)s"),
LOG.info("Created stack with ID %(stack_id)s and "
"name %(stack_name)s for provider PTG %(provider)s",
{'stack_id': stack_id, 'stack_name': stack_name,
'provider': provider['id']})
@ -1755,8 +1752,8 @@ class HeatDriver(object):
except Exception as err:
# Log the error and continue with VM delete in case of *aas
# cleanup failure
LOG.exception(_LE("Cleaning up the service chain stack failed "
"with Error: %(error)s"), {'error': err})
LOG.exception("Cleaning up the service chain stack failed "
"with Error: %(error)s", {'error': err})
return None
return stack_id
@ -1783,8 +1780,8 @@ class HeatDriver(object):
return None
if not base_mode_support and not mgmt_ip:
LOG.error(_LE("Service information is not available with Service "
"Orchestrator on node update"))
LOG.error("Service information is not available with Service "
"Orchestrator on node update")
return None
stack_template, stack_params = self._update_node_config(
@ -1866,8 +1863,8 @@ class HeatDriver(object):
pt_added_or_removed=True)
return stack_id
except Exception:
LOG.exception(_LE("Processing policy target %(operation)s "
" failed"), {'operation': operation})
LOG.exception("Processing policy target %(operation)s "
" failed", {'operation': operation})
return None
def notify_chain_parameters_updated(self, network_function_details):
@ -1902,7 +1899,7 @@ class HeatDriver(object):
return None
return stack_id
except Exception:
LOG.exception(_LE(
LOG.exception(
"Processing policy target group "
"%(operation)s failed"), {'operation': operation})
"%(operation)s failed", {'operation': operation})
return None

View File

@ -15,7 +15,6 @@ from oslo_serialization import jsonutils
from oslo_utils import uuidutils
from sqlalchemy.orm import exc
from gbpservice._i18n import _LW
from gbpservice.nfp.common import exceptions as nfp_exc
from gbpservice.nfp.orchestrator.db import common_db_mixin
from gbpservice.nfp.orchestrator.db import nfp_db_model
@ -686,8 +685,8 @@ class NFPDbBase(common_db_mixin.CommonDbMixin):
return self._get_gw_info_dict(session.query(svc_gw).filter(
svc_gw.network_function_id == nf_id).one())
except exc.NoResultFound:
LOG.warning(_LW("Gateway detail doesn't exist for Network Function"
" %s "), nf_id)
LOG.warning("Gateway detail doesn't exist for Network Function"
" %s ", nf_id)
raise
def _get_gw_info_dict(self, gw):

View File

@ -12,8 +12,6 @@
import ast
from collections import defaultdict
from gbpservice._i18n import _LE
from gbpservice._i18n import _LW
from gbpservice.nfp.common import constants as nfp_constants
from gbpservice.nfp.common import data_formatter as df
@ -131,8 +129,8 @@ class OrchestrationDriver(object):
network_handler)
device_data['interfaces'] = [mgmt_interface]
except Exception as e:
LOG.exception(_LE('Failed to get interfaces for device creation.'
'Error: %(error)s'), {'error': e})
LOG.exception('Failed to get interfaces for device creation.'
'Error: %(error)s', {'error': e})
def _delete_interfaces(self, device_data, interfaces,
network_handler=None):
@ -148,8 +146,8 @@ class OrchestrationDriver(object):
if attr in nfp_constants.METADATA_SUPPORTED_ATTRIBUTES:
provider_metadata[attr] = ast.literal_eval(metadata[attr])
except Exception as e:
LOG.error(_LE('Wrong metadata: %(metadata)s provided for '
'image name: %(image_name)s. Error: %(error)s'),
LOG.error('Wrong metadata: %(metadata)s provided for '
'image name: %(image_name)s. Error: %(error)s',
{'image_name': image_name, 'metadata': metadata,
'error': e})
return None
@ -211,8 +209,8 @@ class OrchestrationDriver(object):
LOG.debug("No provider metadata specified in image,"
" proceeding with default values")
except Exception:
LOG.error(_LE("Error while getting metadata for image name:"
"%(image_name)s, proceeding with default values"),
LOG.error("Error while getting metadata for image name:"
"%(image_name)s, proceeding with default values",
{'image_name': image_name})
return provider_metadata
@ -235,8 +233,8 @@ class OrchestrationDriver(object):
LOG.debug("No provider metadata specified in image,"
" proceeding with default values")
except Exception:
LOG.error(_LE("Error while getting metadata for image name: "
"%(image_name)s, proceeding with default values"),
LOG.error("Error while getting metadata for image name: "
"%(image_name)s, proceeding with default values",
{'image_name': image_name})
return provider_metadata
@ -274,8 +272,8 @@ class OrchestrationDriver(object):
image_id = nova.get_image_id(token, admin_tenant_id, image_name)
return image_id
except Exception as e:
LOG.error(_LE('Failed to get image id for device creation.'
' image name: %(image_name)s. Error: %(error)s'),
LOG.error('Failed to get image id for device creation.'
' image name: %(image_name)s. Error: %(error)s',
{'image_name': image_name, 'error': e})
def create_instance(self, nova, token, admin_tenant_id,
@ -291,8 +289,8 @@ class OrchestrationDriver(object):
server_grp_id=server_grp_id)
return instance_id
except Exception as e:
LOG.error(_LE('Failed to create instance.'
'Error: %(error)s'), {'error': e})
LOG.error('Failed to create instance.'
'Error: %(error)s', {'error': e})
def get_neutron_port_details(self, network_handler, token, port_id):
try:
@ -314,8 +312,8 @@ class OrchestrationDriver(object):
exc_type, exc_value, exc_traceback = sys.exc_info()
LOG.error(traceback.format_exception(exc_type, exc_value,
exc_traceback))
LOG.error(_LE('Failed to get management port details. '
'Error: %(error)s'), {'error': e})
LOG.error('Failed to get management port details. '
'Error: %(error)s', {'error': e})
@_set_network_handler
def create_network_function_device(self, device_data,
@ -383,8 +381,8 @@ class OrchestrationDriver(object):
interfaces_to_attach,
device_data)
except Exception as e:
LOG.error(_LE('Failed to fetch list of interfaces to attach'
' for device creation %(error)s'), {'error': e})
LOG.error('Failed to fetch list of interfaces to attach'
' for device creation %(error)s', {'error': e})
self._delete_interfaces(device_data, interfaces,
network_handler=network_handler)
return None
@ -479,25 +477,25 @@ class OrchestrationDriver(object):
server_grp_id_result=None):
interfaces = device_data.pop('interfaces', None)
if not interfaces:
LOG.exception(_LE('Failed to get interfaces for device creation.'))
LOG.exception('Failed to get interfaces for device creation.')
return None, _, _
image_id = image_id_result.get('result', None)
if not image_id:
LOG.error(_LE('Failed to get image id for device creation.'))
LOG.error('Failed to get image id for device creation.')
self._delete_interfaces(device_data, interfaces,
network_handler=network_handler)
return None, _, _
if server_grp_id_result and not server_grp_id_result.get('result'):
LOG.error(_LE('Validation failed for Nova anti-affinity '
'server group.'))
LOG.error('Validation failed for Nova anti-affinity '
'server group.')
return None, _, _
provider_metadata = provider_metadata_result.get('result', None)
if not provider_metadata:
LOG.warning(_LW('Failed to get provider metadata for'
' device creation.'))
LOG.warning('Failed to get provider metadata for'
' device creation.')
provider_metadata = {}
return interfaces, image_id, provider_metadata
@ -559,8 +557,8 @@ class OrchestrationDriver(object):
admin_tenant_id = device_data['admin_tenant_id']
instance_id = instance_id_result.get('result', None)
if not instance_id:
LOG.error(_LE('Failed to create instance with device data:'
'%(data)s.'),
LOG.error('Failed to create instance with device data:'
'%(data)s.',
{'data': device_data})
self._delete_interfaces(device_data, interfaces,
network_handler=network_handler)
@ -569,7 +567,7 @@ class OrchestrationDriver(object):
mgmt_neutron_port_info = port_details_result.get('result', None)
if not mgmt_neutron_port_info:
LOG.error(_LE('Failed to get management port details. '))
LOG.error('Failed to get management port details. ')
with nfp_ctx_mgr.NovaContextManager as ncm:
ncm.retry(self.compute_handler_nova.delete_instance,
token,
@ -646,8 +644,8 @@ class OrchestrationDriver(object):
interfaces,
network_handler=network_handler)
except Exception as e:
LOG.error(_LE('Failed to delete the management data port(s). '
'Error: %(error)s'), {'error': e})
LOG.error('Failed to delete the management data port(s). '
'Error: %(error)s', {'error': e})
def get_network_function_device_status(self, device_data,
ignore_failure=False):
@ -789,8 +787,8 @@ class OrchestrationDriver(object):
executor.fire()
except Exception as e:
LOG.error(_LE('Failed to plug interface(s) to the device.'
'Error: %(error)s'), {'error': e})
LOG.error('Failed to plug interface(s) to the device.'
'Error: %(error)s', {'error': e})
return None
else:
return True
@ -924,7 +922,7 @@ class OrchestrationDriver(object):
network_handler = self.network_handlers[nfp_constants.NEUTRON_MODE]
network_handler.delete_port(token, port_id)
except Exception as exc:
LOG.error(_LE("Failed to delete port %(port_id)s. Error: %(exc)s"),
LOG.error("Failed to delete port %(port_id)s. Error: %(exc)s",
{"port_id": port_id, 'exc': exc})
def _get_port_from_pt(self, device_data, pt_id):
@ -942,7 +940,7 @@ class OrchestrationDriver(object):
for pt in device_data['consumer']['pt']:
if pt['id'] == pt_id:
return pt['port_id']
LOG.error(_LE('Policy Target %(pt_id) not found in provided data'),
LOG.error('Policy Target %(pt_id) not found in provided data',
{'pt_id': pt_id})
return port_id
@ -1003,8 +1001,8 @@ class OrchestrationDriver(object):
'port_classification',
'port_model'])
):
LOG.error(_LE('Incomplete device data received for delete '
'network function device.'))
LOG.error('Incomplete device data received for delete '
'network function device.')
return None
token = self._get_token(device_data.get('token'))
@ -1027,8 +1025,8 @@ class OrchestrationDriver(object):
devices_data['provider'])
)
except Exception:
LOG.error(_LE('Failed to get provider port details'
' for get device config info operation'))
LOG.error('Failed to get provider port details'
' for get device config info operation')
return None
elif port['port_classification'] == nfp_constants.CONSUMER:
try:
@ -1038,8 +1036,8 @@ class OrchestrationDriver(object):
devices_data['consumer'])
)
except Exception:
LOG.error(_LE('Failed to get consumer port details'
' for get device config info operation'))
LOG.error('Failed to get consumer port details'
' for get device config info operation')
return None
device_data.update({

View File

@ -14,8 +14,6 @@
import oslo_messaging as messaging
from gbpservice._i18n import _
from gbpservice._i18n import _LE
from gbpservice._i18n import _LI
from gbpservice.nfp.common import constants as nfp_constants
from gbpservice.nfp.common import topics as nsf_topics
from gbpservice.nfp.common import utils as nfp_utils
@ -112,14 +110,14 @@ class RpcHandler(object):
NFI = event_data.get('network_function_instance_id')
if NFD and NF and NFI:
LOG.info(_LI("Created event %(event_name)s with"
" NF:%(nf)s ,NFI:%(nfi)s and NFD:%(nfd)s"),
LOG.info("Created event %(event_name)s with"
" NF:%(nf)s ,NFI:%(nfi)s and NFD:%(nfd)s",
{'event_name': event_id,
'nf': NF,
'nfi': NFI,
'nfd': NFD})
else:
LOG.info(_LI("Created event %(event_name)s "),
LOG.info("Created event %(event_name)s ",
{'event_name': event_id})
def _create_event(self, event_id, event_data=None, key=None,
@ -182,7 +180,7 @@ class RpcHandler(object):
event_id = self.rpc_event_mapping[resource][0]
if result.lower() != 'success':
LOG.info(_LI("RPC Handler response data:%(data)s"),
LOG.info("RPC Handler response data:%(data)s",
{'data': data})
if is_delete_request:
# Ignore any deletion errors, generate SUCCESS event
@ -323,20 +321,20 @@ class DeviceOrchestrator(nfp_api.NfpEventHandler):
NFI = event_data.get('network_function_instance_id')
if NFD and NF and NFI:
LOG.info(_LI("Received event %(event_name)s with "
"NF:%(nf)s ,NFI:%(nfi)s and NFD:%(nfd)s"),
LOG.info("Received event %(event_name)s with "
"NF:%(nf)s ,NFI:%(nfi)s and NFD:%(nfd)s",
{'event_name': event.id,
'nf': NF,
'nfi': NFI,
'nfd': NFD})
else:
LOG.info(_LI("Received event %(event_name)s "),
LOG.info("Received event %(event_name)s ",
{'event_name': event.id})
event_handler = self.event_method_mapping(event.id)
event_handler(event)
except Exception as e:
LOG.error(_LE("error in processing event: %(event_id)s for "
"event data %(event_data)s. error: %(error)s"),
LOG.error("error in processing event: %(event_id)s for "
"event data %(event_data)s. error: %(error)s",
{'event_id': event.id, 'event_data': event.data,
'error': e})
_, _, tb = sys.exc_info()
@ -356,13 +354,13 @@ class DeviceOrchestrator(nfp_api.NfpEventHandler):
nf = None
nfi = None
if nf and nfi:
LOG.info(_LI("Created event %(event_name)s with NF:%(nf)s and "
"NFI:%(nfi)s "),
LOG.info("Created event %(event_name)s with NF:%(nf)s and "
"NFI:%(nfi)s ",
{'event_name': event_id,
'nf': nf,
'nfi': nfi})
else:
LOG.info(_LI("Created event %(event_name)s "),
LOG.info("Created event %(event_name)s ",
{'event_name': event_id})
def _create_event(self, event_id, event_data=None,
@ -407,11 +405,11 @@ class DeviceOrchestrator(nfp_api.NfpEventHandler):
self._controller.event_complete(ev)
def event_cancelled(self, ev, reason):
LOG.info(_LI("Poll event %(event_id)s cancelled."),
LOG.info("Poll event %(event_id)s cancelled.",
{'event_id': ev.id})
if ev.id == 'DEVICE_SPAWNING':
LOG.info(_LI("Device is not up still after 10secs of launch"))
LOG.info("Device is not up still after 10secs of launch")
# create event DEVICE_NOT_UP
device = self._prepare_failure_case_device_data(ev.data)
self._create_event(event_id='DEVICE_NOT_UP',
@ -420,10 +418,10 @@ class DeviceOrchestrator(nfp_api.NfpEventHandler):
self._update_network_function_device_db(device,
'DEVICE_NOT_UP')
if ev.id == 'DEVICE_BEING_DELETED':
LOG.info(_LI("Device is not deleted completely."
" Continuing further cleanup of resources."
" Possibly there could be stale port resources"
" on Compute"))
LOG.info("Device is not deleted completely."
" Continuing further cleanup of resources."
" Possibly there could be stale port resources"
" on Compute")
device = ev.data
orchestration_driver = self._get_orchestration_driver(
device['service_details']['service_vendor'])
@ -745,15 +743,15 @@ class DeviceOrchestrator(nfp_api.NfpEventHandler):
nfd_request = self._prepare_failure_case_device_data(nfp_context)
service_details = nfp_context['service_details']
LOG.info(_LI("Received event CREATE NETWORK FUNCTION "
"DEVICE request."))
LOG.info("Received event CREATE NETWORK FUNCTION "
"DEVICE request.")
orchestration_driver = self._get_orchestration_driver(
service_details['service_vendor'])
device_data = self._prepare_device_data_from_nfp_context(nfp_context)
LOG.info(_LI("Creating new device:%(device)s"),
LOG.info("Creating new device:%(device)s",
{'device': nfd_request})
device_data['volume_support'] = (
self.config.device_orchestrator.volume_support)
@ -765,7 +763,7 @@ class DeviceOrchestrator(nfp_api.NfpEventHandler):
orchestration_driver.create_network_function_device(
device_data))
if not driver_device_info:
LOG.info(_LI("Device creation failed"))
LOG.info("Device creation failed")
self._create_event(event_id='DEVICE_ERROR',
event_data=nfd_request,
is_internal_event=True)
@ -824,8 +822,8 @@ class DeviceOrchestrator(nfp_api.NfpEventHandler):
orchestration_driver.get_network_function_device_status(device))
if is_device_up == nfp_constants.ACTIVE:
LOG.info(_LI("Device with NFD:%(id)s came up for "
"tenant:%(tenant)s "),
LOG.info("Device with NFD:%(id)s came up for "
"tenant:%(tenant)s ",
{'id': network_function_device['id'],
'tenant': tenant_id})
self._post_device_up_event_graph(nfp_context)
@ -916,9 +914,9 @@ class DeviceOrchestrator(nfp_api.NfpEventHandler):
self._update_network_function_device_db(
network_function_device, nfp_constants.ACTIVE)
LOG.info(_LI(
LOG.info(
"Configuration completed for device with NFD:%(device_id)s. "
"Updated DB status to ACTIVE."),
"Updated DB status to ACTIVE.",
{'device_id': network_function_device['id']})
LOG.debug("Device detail:%s",
network_function_device)
@ -1331,8 +1329,8 @@ class DeviceOrchestrator(nfp_api.NfpEventHandler):
self._controller.event_complete(event, result="SUCCESS")
return
device = self._prepare_device_data_fast(network_function_details)
LOG.info(_LI("Recieved DELETE NETWORK FUNCTION "
"DEVICE request "))
LOG.info("Recieved DELETE NETWORK FUNCTION "
"DEVICE request ")
device['event_desc'] = event.desc.to_dict()
self._create_event(event_id='DELETE_CONFIGURATION',
event_data=device,
@ -1474,7 +1472,7 @@ class DeviceOrchestrator(nfp_api.NfpEventHandler):
# Error Handling
def handle_device_create_error(self, event):
device = event.data
LOG.error(_LE("Device creation failed, for device %(device)s"),
LOG.error("Device creation failed, for device %(device)s",
{'device': device})
device['network_function_device_id'] = device.get('id')
self._create_event(event_id='DEVICE_CREATE_FAILED',
@ -1563,8 +1561,8 @@ class DeviceOrchestrator(nfp_api.NfpEventHandler):
def handle_driver_error(self, event):
device = event.data
LOG.error(_LE("Exception occured in driver, driver returned None "
" for device %(device)s"), {'device': device})
LOG.error("Exception occured in driver, driver returned None "
" for device %(device)s", {'device': device})
status = nfp_constants.ERROR
desc = 'Exception in driver, driver return None'
self._update_network_function_device_db(device, status, desc)
@ -1631,8 +1629,8 @@ class NDOConfiguratorRpcApi(object):
def create_network_function_device_config(self, device_data,
config_params):
self._update_params(device_data, config_params, operation='create')
LOG.info(_LI("Sending create NFD config request to configurator "
"for NF:%(nf_id)s "),
LOG.info("Sending create NFD config request to configurator "
"for NF:%(nf_id)s ",
{'nf_id': config_params['info']['context']['nf_id']})
transport.send_request_to_configurator(self.conf,
@ -1645,7 +1643,7 @@ class NDOConfiguratorRpcApi(object):
config_params):
self._update_params(device_data, config_params, operation='delete')
config_params['info']['context']['nfp_context'] = device_data
LOG.info(_LI("Sending delete NFD config request to configurator "))
LOG.info("Sending delete NFD config request to configurator ")
transport.send_request_to_configurator(self.conf,
self.context,

View File

@ -17,9 +17,6 @@ from oslo_log import helpers as log_helpers
import oslo_messaging
from gbpservice._i18n import _
from gbpservice._i18n import _LE
from gbpservice._i18n import _LI
from gbpservice._i18n import _LW
from gbpservice.nfp.common import constants as nfp_constants
from gbpservice.nfp.common import exceptions as nfp_exc
from gbpservice.nfp.common import topics as nfp_rpc_topics
@ -124,8 +121,8 @@ class RpcHandler(object):
Function Instance
'''
module_context.init(network_function)
LOG.info(_LI("Received RPC call for CREATE NETWORK FUNCTION for "
"tenant:%(tenant_id)s"),
LOG.info("Received RPC call for CREATE NETWORK FUNCTION for "
"tenant:%(tenant_id)s",
{'tenant_id': network_function[
'resource_owner_context']['tenant_id']})
@ -151,7 +148,7 @@ class RpcHandler(object):
Returns the Network functions from DB
'''
module_context.init()
LOG.info(_LI("Received RPC call for GET NETWORK FUNCTIONS "))
LOG.info("Received RPC call for GET NETWORK FUNCTIONS ")
service_orchestrator = ServiceOrchestrator(self._controller, self.conf)
return service_orchestrator.get_network_functions(
context, filters)
@ -166,8 +163,8 @@ class RpcHandler(object):
'''
module_context.init()
LOG.info(_LI("Received RPC call for UPDATE NETWORK FUNCTION for NF:"
"%(network_function_id)s"),
LOG.info("Received RPC call for UPDATE NETWORK FUNCTION for NF:"
"%(network_function_id)s",
{'network_function_id': network_function_id})
service_orchestrator = ServiceOrchestrator(self._controller, self.conf)
service_orchestrator.update_network_function(
@ -182,8 +179,8 @@ class RpcHandler(object):
Results in an Event for async processing of Network Function Instance.
'''
module_context.init()
LOG.info(_LI("Received RPC call for DELETE NETWORK FUNCTION for NF:"
"%(network_function_id)s"),
LOG.info("Received RPC call for DELETE NETWORK FUNCTION for NF:"
"%(network_function_id)s",
{'network_function_id': network_function_id})
service_orchestrator = ServiceOrchestrator(self._controller, self.conf)
@ -199,9 +196,9 @@ class RpcHandler(object):
Results in an Event for async processing of Network Function Instance.
'''
module_context.init()
LOG.info(_LI("Received RPC call for POLICY TARGET ADDED NOTIFICATION "
"for NF:"
" %(network_function_id)s"),
LOG.info("Received RPC call for POLICY TARGET ADDED NOTIFICATION "
"for NF:"
" %(network_function_id)s",
{'network_function_id': network_function_id})
service_orchestrator = ServiceOrchestrator(self._controller, self.conf)
service_orchestrator.handle_policy_target_added(
@ -216,8 +213,8 @@ class RpcHandler(object):
Results in an Event for async processing of Network Function Instance.
'''
module_context.init()
LOG.info(_LI("Received RPC call for POLICY TARGET REMOVED "
"NOTIFICATION for NF:%(network_function_id)s"),
LOG.info("Received RPC call for POLICY TARGET REMOVED "
"NOTIFICATION for NF:%(network_function_id)s",
{'network_function_id': network_function_id})
service_orchestrator = ServiceOrchestrator(self._controller, self.conf)
service_orchestrator.handle_policy_target_removed(
@ -232,8 +229,8 @@ class RpcHandler(object):
Results in an Event for async processing of Network Function Instance.
'''
module_context.init()
LOG.info(_LI("Received RPC call CONSUMER PTG ADDED NOTIFICATION "
"for NF:%(network_function_id)s"),
LOG.info("Received RPC call CONSUMER PTG ADDED NOTIFICATION "
"for NF:%(network_function_id)s",
{'network_function_id': network_function_id})
service_orchestrator = ServiceOrchestrator(self._controller, self.conf)
service_orchestrator.handle_consumer_ptg_added(
@ -248,8 +245,8 @@ class RpcHandler(object):
Results in an Event for async processing of Network Function Instance.
'''
module_context.init()
LOG.info(_LI("Received RPC call for CONSUMER PTG REMOVED NOTIFICATION "
"for NF:%(network_function_id)s"),
LOG.info("Received RPC call for CONSUMER PTG REMOVED NOTIFICATION "
"for NF:%(network_function_id)s",
{'network_function_id': network_function_id})
service_orchestrator = ServiceOrchestrator(self._controller, self.conf)
service_orchestrator.handle_consumer_ptg_removed(
@ -334,13 +331,13 @@ class RpcHandlerConfigurator(object):
NF = None
NFI = None
if NF and NFI:
LOG.info(_LI("Created event %(event_name)s with NF:%(nf)s "
"and NFI:%(nfi)s "),
LOG.info("Created event %(event_name)s with NF:%(nf)s "
"and NFI:%(nfi)s ",
{'event_name': event_id,
'nf': NF,
'nfi': NFI})
else:
LOG.info(_LI("Created event %(event_name)s "),
LOG.info("Created event %(event_name)s ",
{'event_name': event_id})
def _create_event(self, event_id, event_data=None,
@ -588,20 +585,20 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
NF = None
NFI = None
if NF and NFI:
LOG.info(_LI("Received event %(event_name)s with NF:%(nf)s and "
"NFI:%(nfi)s "),
LOG.info("Received event %(event_name)s with NF:%(nf)s and "
"NFI:%(nfi)s ",
{'event_name': event.id,
'nf': NF,
'nfi': NFI})
else:
LOG.info(_LI("Received event %(event_name)s "),
LOG.info("Received event %(event_name)s ",
{'event_name': event.id})
try:
event_handler = self.event_method_mapping(event.id)
event_handler(event)
except Exception as e:
LOG.exception(_LE("Error in processing event: %(event_id)s for "
"event data %(event_data)s. Error: %(error)s"),
LOG.exception("Error in processing event: %(event_id)s for "
"event data %(event_data)s. Error: %(error)s",
{'event_id': event.id, 'event_data': event.data,
'error': e})
_, _, tb = sys.exc_info()
@ -609,27 +606,26 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
raise e
def handle_poll_event(self, event):
LOG.info(_LI("Received poll event %(id)s"),
LOG.info("Received poll event %(id)s",
{'id': event.id})
try:
event_handler = self.event_method_mapping(event.id)
return event_handler(event)
except Exception:
LOG.exception(_LE("Error in processing poll event: "
"%(event_id)s"), {'event_id': event.id})
LOG.exception("Error in processing poll event: "
"%(event_id)s", {'event_id': event.id})
def event_cancelled(self, event, reason):
nfp_context = event.context
if event.id == 'CHECK_USER_CONFIG_COMPLETE':
network_function = nfp_context['network_function']
LOG.info(_LI("Applying user config failed for "
"NF:%(network_function_id)s "
"with reason %(reason)s"
""), {'network_function_id': network_function[
'id'],
'reason': str(reason)})
LOG.info("Applying user config failed for "
"NF:%(network_function_id)s "
"with reason %(reason)s"
" ", {'network_function_id': network_function[
'id'], 'reason': str(reason)})
operation = nfp_context['log_context'].get('path')
LOG.error(_LE("[Event:Service%(operation)sFailed]"),
LOG.error("[Event:Service%(operation)sFailed]",
{'operation': operation.capitalize()})
LOG.event('%s network function failed.' % operation.capitalize(),
stats_type=nfp_constants.error_event)
@ -649,13 +645,13 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
elif event.id == 'APPLY_USER_CONFIG_IN_PROGRESS' or (
event.id == 'UPDATE_USER_CONFIG_STILL_IN_PROGRESS'):
request_data = event.data
LOG.info(_LI("Applying user config failed for "
"NF: %(network_function_id)s data:"
"%(data)s with reason %(reason)s"
""), {'data': request_data,
'network_function_id': request_data[
'network_function_id'],
'reason': str(reason)})
LOG.info("Applying user config failed for "
"NF: %(network_function_id)s data:"
"%(data)s with reason %(reason)s"
"", {'data': request_data,
'network_function_id': request_data[
'network_function_id'],
'reason': str(reason)})
updated_network_function = {'status': nfp_constants.ERROR}
with nfp_ctx_mgr.DbContextManager as dcm:
@ -666,7 +662,7 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
updated_network_function)
operation = nfp_context['log_context'].get('path')
LOG.error(_LE("[Event:Service%(operation)sFailed]"),
LOG.error("[Event:Service%(operation)sFailed]",
{'operation': operation.capitalize()})
LOG.event('%s network function failed.' % operation.capitalize(),
stats_type=nfp_constants.error_event)
@ -702,13 +698,13 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
NF = None
NFI = None
if NF and NFI:
LOG.info(_LI("Created event %(event_name)s with NF:%(nf)s and "
"NFI:%(nfi)s "),
LOG.info("Created event %(event_name)s with NF:%(nf)s and "
"NFI:%(nfi)s ",
{'event_name': event_id,
'nf': NF,
'nfi': NFI})
else:
LOG.info(_LI("Created event %(event_name)s "),
LOG.info("Created event %(event_name)s ",
{'event_name': event_id})
# REVISIT(ashu): Merge this _create_event, and above one to have
# single function.
@ -788,9 +784,9 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
tag_str, config_str = self.config_driver.parse_template_config_string(
service_config_str)
if not config_str:
LOG.error(_LE('Exception while parsing config string, config '
'string: %(config_str)s is improper for '
'network_function id: %(network_function_id)s'),
LOG.error('Exception while parsing config string, config '
'string: %(config_str)s is improper for '
'network_function id: %(network_function_id)s',
{'config_str': service_config_str,
'network_function_id': network_function_id})
self.handle_driver_error(network_function_id)
@ -823,9 +819,9 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
network_function_id = network_function_data[
'network_function_details']['network_function']['id']
if not config_str:
LOG.error(_LE('Exception while parsing config string, config '
'string: %(config_str)s is improper for '
'network_function id: %(network_function_id)s'),
LOG.error('Exception while parsing config string, config '
'string: %(config_str)s is improper for '
'network_function id: %(network_function_id)s',
{'config_str': service_config_str,
'network_function_id': network_function_id})
self.handle_driver_error(network_function_id)
@ -877,11 +873,11 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
def _report_logging_info(self, nf, nfi, service_type,
service_vendor):
LOG.info(_LI("[TenantID:%(tenant_id)s, "
"ServiceChainID:%(service_chain_id)s, "
"ServiceInstanceID:%(service_instance_id)s, "
"ServiceType:%(service_type)s, "
"ServiceProvider:%(service_provider)s]"),
LOG.info("[TenantID:%(tenant_id)s, "
"ServiceChainID:%(service_chain_id)s, "
"ServiceInstanceID:%(service_instance_id)s, "
"ServiceType:%(service_type)s, "
"ServiceProvider:%(service_provider)s]",
{'tenant_id': nf['tenant_id'],
'service_chain_id': nf['service_chain_id'],
'service_instance_id': nfi['id'],
@ -961,7 +957,7 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
nfp_context['log_context']['meta_id'] = network_function['id']
nfp_context['log_context']['auth_token'] = context.auth_token
LOG.info(_LI("[Event:ServiceCreateInitiated]"))
LOG.info("[Event:ServiceCreateInitiated]")
LOG.event("Started create network function.",
stats_type=nfp_constants.request_event)
@ -973,8 +969,8 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
nfp_context['service_details'] = service_details
nfp_context['share_existing_device'] = False
nfp_context['base_mode'] = base_mode_support
LOG.info(_LI("Handling RPC call CREATE NETWORK FUNCTION for "
"%(service_type)s with tenant:%(tenant_id)s"),
LOG.info("Handling RPC call CREATE NETWORK FUNCTION for "
"%(service_type)s with tenant:%(tenant_id)s",
{'tenant_id': tenant_id,
'service_type': service_profile['service_type']})
if base_mode_support:
@ -1009,7 +1005,7 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
network_function_id,
{'service_config': user_config,
'status': nfp_constants.PENDING_UPDATE})
LOG.info(_LI("[Event:ServiceUpdateInitiated]"))
LOG.info("[Event:ServiceUpdateInitiated]")
LOG.event("Started update network function.",
stats_type=nfp_constants.request_event)
@ -1054,7 +1050,7 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
with nfp_ctx_mgr.DbContextManager:
self.db_handler.delete_network_function(
self.db_session, network_function_id)
LOG.info(_LI("[Event:ServiceDeleteCompleted]"))
LOG.info("[Event:ServiceDeleteCompleted]")
LOG.event("Completed delete network function.",
stats_type=nfp_constants.response_event)
@ -1087,7 +1083,7 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
network_function_id, network_function)
nfp_context.update(network_function_details)
LOG.info(_LI("[Event:ServiceDeleteInitiated]"))
LOG.info("[Event:ServiceDeleteInitiated]")
LOG.event("Started delete network function.",
stats_type=nfp_constants.request_event)
if not base_mode_support:
@ -1231,8 +1227,8 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
nfp_context['log_context']['nfi_id'] = nfi['id']
LOG.info(_LI("Creating event CREATE NETWORK FUNCTION DEVICE "
"for NF: %(network_function_id)s"),
LOG.info("Creating event CREATE NETWORK FUNCTION DEVICE "
"for NF: %(network_function_id)s",
{'network_function_id': network_function['id']})
ev = self._controller.new_event(
@ -1385,10 +1381,10 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
neutron_resource_desc = (
self.config_driver.get_neutron_resource_description(nfp_context))
if not neutron_resource_desc:
LOG.error(_LE(
LOG.error(
"Preparing neutron resource description failed in "
"config driver, marking user config as Failed for "
"network function: %(nf)s"), {'nf': network_function})
"network function: %(nf)s", {'nf': network_function})
nfp_context['network_function_id'] = network_function['id']
binding_key = nfp_context['service_details'][
'service_vendor'].lower() + network_function['id']
@ -1488,7 +1484,7 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
request_data['network_function_id'])
network_function = network_function_details['network_function']
LOG.info(_LI("[Event:ServiceUpdateInitiated]"))
LOG.info("[Event:ServiceUpdateInitiated]")
LOG.event("Started update network function.",
stats_type=nfp_constants.request_event)
nfi = network_function_details.get('network_function_instance', None)
@ -1570,7 +1566,7 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
nfp_context = event.context
operation = nfp_context['log_context'].get('path')
LOG.error(_LE("[Event:Service%(operation)sFailed]"),
LOG.error("[Event:Service%(operation)sFailed]",
{'operation': operation.capitalize()})
LOG.event('%s network function failed.' % operation.capitalize(),
stats_type=nfp_constants.error_event)
@ -1578,9 +1574,9 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
# Trigger RPC to notify the Create_Service caller with status
def handle_driver_error(self, network_function_id):
LOG.error(_LE("Error occurred while processing network function "
"CRUD operations, marking network function: %(nf_id)s "
"as ERROR to initiate cleanup."),
LOG.error("Error occurred while processing network function "
"CRUD operations, marking network function: %(nf_id)s "
"as ERROR to initiate cleanup.",
{'nf_id': network_function_id})
network_function_details = self.get_network_function_details(
network_function_id)
@ -1592,7 +1588,7 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
network_function_id, network_function)
nfp_context = module_context.get()
operation = nfp_context['log_context'].get('path')
LOG.error(_LE("[Event:Service%(operation)sFailed]"),
LOG.error("[Event:Service%(operation)sFailed]",
{'operation': operation.capitalize()})
LOG.event('%s network function failed.' % operation.capitalize(),
stats_type=nfp_constants.error_event)
@ -1664,10 +1660,10 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
request_data['config_policy_id'], request_data['tenant_id'],
request_data['network_function_details'])
if config_status == nfp_constants.ERROR:
LOG.info(_LI("Applying user config failed for "
"NF:%(network_function_id)s "), {
'network_function_id':
request_data['network_function_id']})
LOG.info("Applying user config failed for "
"NF:%(network_function_id)s ", {
'network_function_id':
request_data['network_function_id']})
updated_network_function = {'status': nfp_constants.ERROR}
with nfp_ctx_mgr.DbContextManager as dcm:
dcm.lock(
@ -1676,7 +1672,7 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
request_data['network_function_id'],
updated_network_function)
operation = event.context['log_context'].get('path')
LOG.error(_LE("[Event:Service%(operation)sFailed]"),
LOG.error("[Event:Service%(operation)sFailed]",
{'operation': operation.capitalize()})
LOG.event('%s network function failed.' % operation.capitalize(),
stats_type=nfp_constants.error_event)
@ -1710,8 +1706,8 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
max_times=self.UPDATE_USER_CONFIG_MAXRETRY)
return STOP_POLLING
updated_network_function = {'status': nfp_constants.ACTIVE}
LOG.info(_LI("Applying user config is successfull moving "
"NF:%(network_function_id)s to ACTIVE"),
LOG.info("Applying user config is successfull moving "
"NF:%(network_function_id)s to ACTIVE",
{'network_function_id':
request_data['network_function_id']})
with nfp_ctx_mgr.DbContextManager as dcm:
@ -1722,7 +1718,7 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
updated_network_function)
operation = event.context['log_context'].get('path')
LOG.info(_LI("[Event:Service%(operation)sCompleted]"),
LOG.info("[Event:Service%(operation)sCompleted]",
{'operation': operation.capitalize()})
LOG.event('Completed %s network function.' % operation,
stats_type=nfp_constants.response_event)
@ -1751,8 +1747,8 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
network_function = nfp_context['network_function']
updated_network_function = {'status': nfp_constants.ACTIVE}
LOG.info(_LI("Applying user config is successfull moving "
"NF: %(network_function_id)s to ACTIVE"),
LOG.info("Applying user config is successfull moving "
"NF: %(network_function_id)s to ACTIVE",
{'network_function_id': network_function['id']})
with nfp_ctx_mgr.DbContextManager as dcm:
dcm.lock(self.db_session, self.db_handler.update_network_function,
@ -1760,7 +1756,7 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
updated_network_function)
operation = nfp_context['log_context'].get('path')
LOG.info(_LI("[Event:Service%(operation)sCompleted]"),
LOG.info("[Event:Service%(operation)sCompleted]",
{'operation': operation.capitalize()})
LOG.event('Completed %s network function.' % operation,
stats_type=nfp_constants.response_event)
@ -1787,10 +1783,10 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
config_status = self.config_driver.check_config_complete(nfp_context)
if config_status == nfp_constants.ERROR:
LOG.info(_LI("Applying user config failed for "
"NF: %(network_function_id)s"), {
'network_function_id':
network_function['id']})
LOG.info("Applying user config failed for "
"NF: %(network_function_id)s", {
'network_function_id':
network_function['id']})
# Complete the original event APPLY_USER_CONFIG here
event_desc = nfp_context.pop('event_desc', None)
apply_config_event = self._controller.new_event(
@ -1833,8 +1829,8 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
network_function)
except Exception as err:
# REVISIT: May be we need a count before removing the poll event
LOG.error(_LE("Error: %(err)s while verifying configuration "
"delete completion."), {'err': err})
LOG.error("Error: %(err)s while verifying configuration "
"delete completion.", {'err': err})
self._create_event('USER_CONFIG_DELETE_FAILED',
event_data=request_data, is_internal_event=True)
self._controller.event_complete(event)
@ -1850,8 +1846,8 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
# Trigger RPC to notify the Create_Service caller with status
elif config_status == nfp_constants.COMPLETED:
updated_network_function = {'status': nfp_constants.ACTIVE}
LOG.info(_LI("Applying user config is successfull moving "
"NF:%(network_function_id)s to ACTIVE"),
LOG.info("Applying user config is successfull moving "
"NF:%(network_function_id)s to ACTIVE",
{'network_function_id':
request_data['network_function_id']})
with nfp_ctx_mgr.DbContextManager as dcm:
@ -1862,7 +1858,7 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
updated_network_function)
operation = event.context['log_context'].get('path')
LOG.info(_LI("[Event:Service%(operation)sCompleted]"),
LOG.info("[Event:Service%(operation)sCompleted]",
{'operation': operation.capitalize()})
LOG.event('Completed %s network function.' % operation,
stats_type=nfp_constants.response_event)
@ -1890,8 +1886,8 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
request_data['config_policy_id'], request_data['tenant_id'])
except Exception as err:
# REVISIT: May be we need a count before removing the poll event
LOG.error(_LE("Error: %(err)s while verifying configuration "
"delete completion."), {'err': err})
LOG.error("Error: %(err)s while verifying configuration "
"delete completion.", {'err': err})
# self._create_event('USER_CONFIG_DELETE_FAILED',
# event_data=event_data, is_internal_event=True)
self._controller.event_complete(event)
@ -1954,8 +1950,8 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
self.db_handler.update_network_function,
network_function_id,
network_function)
LOG.info(_LI("Applying user config is successfull moving "
"NF: %(network_function_id)s to ACTIVE"),
LOG.info("Applying user config is successfull moving "
"NF: %(network_function_id)s to ACTIVE",
{'network_function_id':
network_function_id})
else:
@ -1981,8 +1977,8 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
def handle_update_user_config_failed(self, event):
event_data = event.data
network_function_id = event_data['network_function_id']
LOG.error(_LE("NSO: updating user config failed, moving "
"network function %(network_function_id)s to ERROR"),
LOG.error("NSO: updating user config failed, moving "
"network function %(network_function_id)s to ERROR",
{'network_function_id': network_function_id})
self.handle_user_config_failed(event)
@ -1998,7 +1994,7 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
updated_network_function)
# Trigger RPC to notify the Create_Service caller with status
operation = event.context['log_context'].get('path')
LOG.error(_LE("[Event:Service%(operation)sFailed]"),
LOG.error("[Event:Service%(operation)sFailed]",
{'operation': operation.capitalize()})
LOG.event('%s network function failed.' % operation.capitalize(),
stats_type=nfp_constants.error_event)
@ -2024,9 +2020,9 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
self.config_driver.is_update_config_supported(
request_data['service_type'])):
updated_network_function.update({'status': nfp_constants.ACTIVE})
LOG.warning(_LW(
LOG.warning(
"Failed to delete old stack id: %(stack_id)s in"
"firewall update case, Need to manually delete it"),
"firewall update case, Need to manually delete it",
{"stack_id": request_data['config_policy_id']})
with nfp_ctx_mgr.DbContextManager as dcm:
@ -2042,7 +2038,7 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
results = event.result
for result in results:
if result.result.lower() != 'success':
LOG.error(_LE("Event: %(result_id)s failed"),
LOG.error("Event: %(result_id)s failed",
{'result_id': result.id})
network_function_details = event.context
@ -2065,7 +2061,7 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
with nfp_ctx_mgr.DbContextManager:
self.db_handler.delete_network_function(
self.db_session, nf['id'])
LOG.info(_LI("[Event:ServiceDeleteCompleted]"))
LOG.info("[Event:ServiceDeleteCompleted]")
LOG.event("Completed delete network function.",
stats_type=nfp_constants.response_event)
@ -2076,7 +2072,7 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
nf_id=nf_id,
service_type=service_type)
LOG.info(_LI("Deleted NF:%(nf_id)s "),
LOG.info("Deleted NF:%(nf_id)s ",
{'nf_id': nf['id']})
self._controller.event_complete(event)
@ -2095,7 +2091,7 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
with nfp_ctx_mgr.DbContextManager:
self.db_handler.delete_network_function(
self.db_session, nfi['network_function_id'])
LOG.info(_LI("[Event:ServiceDeleteCompleted]"))
LOG.info("[Event:ServiceDeleteCompleted]")
LOG.event("Completed delete network function.",
stats_type=nfp_constants.response_event)
@ -2105,7 +2101,7 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
nf_id=nf_id,
service_type=service_type)
LOG.info(_LI("Deleted NF:%(nf_id)s "),
LOG.info("Deleted NF:%(nf_id)s ",
{'nf_id': nf_id})
# Inform delete service caller with delete completed RPC
@ -2120,13 +2116,13 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
self.db_session, network_function_id)
return network_function
except nfp_exc.NetworkFunctionNotFound:
LOG.warning(_LW("Failed to retrieve Network Function details for"
" %(network_function)s"),
LOG.warning("Failed to retrieve Network Function details for"
" %(network_function)s",
{'network_function': network_function_id})
return None
except Exception:
LOG.exception(_LE("Failed to retrieve Network Function details for"
" %(network_function)s"),
LOG.exception("Failed to retrieve Network Function details for"
" %(network_function)s",
{'network_function': network_function_id})
return None
@ -2431,8 +2427,8 @@ class ServiceOrchestrator(nfp_api.NfpEventHandler):
self.db_session, port_id)
return port_info
except Exception:
LOG.exception(_LE("Failed to retrieve Port Info for"
" %(port_id)s"),
LOG.exception("Failed to retrieve Port Info for"
" %(port_id)s",
{'port_id': port_id})
return None
@ -2615,7 +2611,7 @@ class NSOConfiguratorRpcApi(object):
config_tag)
self._update_params(user_config_data,
config_params, operation='create')
LOG.info(_LI("Sending create heat config request to configurator "))
LOG.info("Sending create heat config request to configurator ")
LOG.debug("Sending create heat config request to configurator "
"with config_params = %s",
config_params)
@ -2632,7 +2628,7 @@ class NSOConfiguratorRpcApi(object):
config_tag)
self._update_params(user_config_data,
config_params, operation='delete')
LOG.info(_LI("Sending delete heat config request to configurator "))
LOG.info("Sending delete heat config request to configurator ")
LOG.debug("Sending delete heat config request to configurator "
" with config_params = %s",
config_params)
@ -2648,7 +2644,7 @@ class NSOConfiguratorRpcApi(object):
config_tag)
self._update_params(user_config_data,
config_params, operation='update')
LOG.info(_LI("Sending update heat config request to configurator. "))
LOG.info("Sending update heat config request to configurator. ")
transport.send_request_to_configurator(self.conf,
self.context,
@ -2662,8 +2658,8 @@ class NSOConfiguratorRpcApi(object):
config_tag)
self._update_params(user_config_data,
config_params, operation='pt_add')
LOG.info(_LI("Sending Policy Target and heat config request to "
"configurator ."))
LOG.info("Sending Policy Target and heat config request to "
"configurator .")
transport.send_request_to_configurator(self.conf,
self.context,
@ -2677,8 +2673,8 @@ class NSOConfiguratorRpcApi(object):
config_tag)
self._update_params(user_config_data,
config_params, operation='pt_remove')
LOG.info(_LI("Sending Policy Target remove heat config request to "
"configurator. "))
LOG.info("Sending Policy Target remove heat config request to "
"configurator. ")
transport.send_request_to_configurator(self.conf,
self.context,
@ -2692,8 +2688,8 @@ class NSOConfiguratorRpcApi(object):
config_tag)
self._update_params(user_config_data,
config_params, operation='consumer_add')
LOG.info(_LI("Sending consumer and heat config request to "
"configurator ."))
LOG.info("Sending consumer and heat config request to "
"configurator .")
transport.send_request_to_configurator(self.conf,
self.context,
@ -2707,8 +2703,8 @@ class NSOConfiguratorRpcApi(object):
config_tag)
self._update_params(user_config_data,
config_params, operation='consumer_remove')
LOG.info(_LI("Sending consumer remove heat config request to "
"configurator ."))
LOG.info("Sending consumer remove heat config request to "
"configurator .")
transport.send_request_to_configurator(self.conf,
self.context,