
Kill some of the variables marked as unused by flakes8. This should allow to enable F841 check in the future. Only trivial cases with no function calls and obviously pure functions (like datetime.now(), or len()) are cleaned up here. Part 1, split to reduce conflicts. Change-Id: I82854349574ec4bcb9336aba626eefdaed81a8c8
401 lines
15 KiB
Python
401 lines
15 KiB
Python
# vim: tabstop=4 shiftwidth=4 softtabstop=4
|
|
|
|
# Copyright 2010 United States Government as represented by the
|
|
# Administrator of the National Aeronautics and Space Administration.
|
|
# Copyright 2011 Justin Santa Barbara
|
|
# All Rights Reserved.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
|
|
"""Generic Node base class for all workers that run on hosts."""
|
|
|
|
import os
|
|
import random
|
|
import sys
|
|
|
|
from oslo.config import cfg
|
|
|
|
from nova import conductor
|
|
from nova import context
|
|
from nova import exception
|
|
from nova.openstack.common import importutils
|
|
from nova.openstack.common import log as logging
|
|
from nova.openstack.common import rpc
|
|
from nova.openstack.common import service
|
|
from nova import servicegroup
|
|
from nova import utils
|
|
from nova import version
|
|
from nova import wsgi
|
|
|
|
LOG = logging.getLogger(__name__)
|
|
|
|
service_opts = [
|
|
cfg.IntOpt('report_interval',
|
|
default=10,
|
|
help='seconds between nodes reporting state to datastore'),
|
|
cfg.BoolOpt('periodic_enable',
|
|
default=True,
|
|
help='enable periodic tasks'),
|
|
cfg.IntOpt('periodic_fuzzy_delay',
|
|
default=60,
|
|
help='range of seconds to randomly delay when starting the'
|
|
' periodic task scheduler to reduce stampeding.'
|
|
' (Disable by setting to 0)'),
|
|
cfg.ListOpt('enabled_apis',
|
|
default=['ec2', 'osapi_compute', 'metadata'],
|
|
help='a list of APIs to enable by default'),
|
|
cfg.ListOpt('enabled_ssl_apis',
|
|
default=[],
|
|
help='a list of APIs with enabled SSL'),
|
|
cfg.StrOpt('ec2_listen',
|
|
default="0.0.0.0",
|
|
help='IP address for EC2 API to listen'),
|
|
cfg.IntOpt('ec2_listen_port',
|
|
default=8773,
|
|
help='port for ec2 api to listen'),
|
|
cfg.IntOpt('ec2_workers',
|
|
default=None,
|
|
help='Number of workers for EC2 API service'),
|
|
cfg.StrOpt('osapi_compute_listen',
|
|
default="0.0.0.0",
|
|
help='IP address for OpenStack API to listen'),
|
|
cfg.IntOpt('osapi_compute_listen_port',
|
|
default=8774,
|
|
help='list port for osapi compute'),
|
|
cfg.IntOpt('osapi_compute_workers',
|
|
default=None,
|
|
help='Number of workers for OpenStack API service'),
|
|
cfg.StrOpt('metadata_manager',
|
|
default='nova.api.manager.MetadataManager',
|
|
help='OpenStack metadata service manager'),
|
|
cfg.StrOpt('metadata_listen',
|
|
default="0.0.0.0",
|
|
help='IP address for metadata api to listen'),
|
|
cfg.IntOpt('metadata_listen_port',
|
|
default=8775,
|
|
help='port for metadata api to listen'),
|
|
cfg.IntOpt('metadata_workers',
|
|
default=None,
|
|
help='Number of workers for metadata service'),
|
|
cfg.StrOpt('compute_manager',
|
|
default='nova.compute.manager.ComputeManager',
|
|
help='full class name for the Manager for compute'),
|
|
cfg.StrOpt('console_manager',
|
|
default='nova.console.manager.ConsoleProxyManager',
|
|
help='full class name for the Manager for console proxy'),
|
|
cfg.StrOpt('cert_manager',
|
|
default='nova.cert.manager.CertManager',
|
|
help='full class name for the Manager for cert'),
|
|
cfg.StrOpt('network_manager',
|
|
default='nova.network.manager.VlanManager',
|
|
help='full class name for the Manager for network'),
|
|
cfg.StrOpt('scheduler_manager',
|
|
default='nova.scheduler.manager.SchedulerManager',
|
|
help='full class name for the Manager for scheduler'),
|
|
cfg.IntOpt('service_down_time',
|
|
default=60,
|
|
help='maximum time since last check-in for up service'),
|
|
]
|
|
|
|
CONF = cfg.CONF
|
|
CONF.register_opts(service_opts)
|
|
CONF.import_opt('host', 'nova.netconf')
|
|
|
|
|
|
class Service(service.Service):
|
|
"""Service object for binaries running on hosts.
|
|
|
|
A service takes a manager and enables rpc by listening to queues based
|
|
on topic. It also periodically runs tasks on the manager and reports
|
|
it state to the database services table.
|
|
"""
|
|
|
|
def __init__(self, host, binary, topic, manager, report_interval=None,
|
|
periodic_enable=None, periodic_fuzzy_delay=None,
|
|
periodic_interval_max=None, db_allowed=True,
|
|
*args, **kwargs):
|
|
super(Service, self).__init__()
|
|
self.host = host
|
|
self.binary = binary
|
|
self.topic = topic
|
|
self.manager_class_name = manager
|
|
# NOTE(russellb) We want to make sure to create the servicegroup API
|
|
# instance early, before creating other things such as the manager,
|
|
# that will also create a servicegroup API instance. Internally, the
|
|
# servicegroup only allocates a single instance of the driver API and
|
|
# we want to make sure that our value of db_allowed is there when it
|
|
# gets created. For that to happen, this has to be the first instance
|
|
# of the servicegroup API.
|
|
self.servicegroup_api = servicegroup.API(db_allowed=db_allowed)
|
|
manager_class = importutils.import_class(self.manager_class_name)
|
|
self.manager = manager_class(host=self.host, *args, **kwargs)
|
|
self.report_interval = report_interval
|
|
self.periodic_enable = periodic_enable
|
|
self.periodic_fuzzy_delay = periodic_fuzzy_delay
|
|
self.periodic_interval_max = periodic_interval_max
|
|
self.saved_args, self.saved_kwargs = args, kwargs
|
|
self.backdoor_port = None
|
|
self.conductor_api = conductor.API(use_local=db_allowed)
|
|
self.conductor_api.wait_until_ready(context.get_admin_context())
|
|
|
|
def start(self):
|
|
verstr = version.version_string_with_package()
|
|
LOG.audit(_('Starting %(topic)s node (version %(version)s)'),
|
|
{'topic': self.topic, 'version': verstr})
|
|
self.basic_config_check()
|
|
self.manager.init_host()
|
|
self.model_disconnected = False
|
|
ctxt = context.get_admin_context()
|
|
try:
|
|
self.service_ref = self.conductor_api.service_get_by_args(ctxt,
|
|
self.host, self.binary)
|
|
self.service_id = self.service_ref['id']
|
|
except exception.NotFound:
|
|
self.service_ref = self._create_service_ref(ctxt)
|
|
|
|
if self.backdoor_port is not None:
|
|
self.manager.backdoor_port = self.backdoor_port
|
|
|
|
self.conn = rpc.create_connection(new=True)
|
|
LOG.debug(_("Creating Consumer connection for Service %s") %
|
|
self.topic)
|
|
|
|
self.manager.pre_start_hook(rpc_connection=self.conn)
|
|
|
|
rpc_dispatcher = self.manager.create_rpc_dispatcher(self.backdoor_port)
|
|
|
|
# Share this same connection for these Consumers
|
|
self.conn.create_consumer(self.topic, rpc_dispatcher, fanout=False)
|
|
|
|
node_topic = '%s.%s' % (self.topic, self.host)
|
|
self.conn.create_consumer(node_topic, rpc_dispatcher, fanout=False)
|
|
|
|
self.conn.create_consumer(self.topic, rpc_dispatcher, fanout=True)
|
|
|
|
# Consume from all consumers in a thread
|
|
self.conn.consume_in_thread()
|
|
|
|
self.manager.post_start_hook()
|
|
|
|
LOG.debug(_("Join ServiceGroup membership for this service %s")
|
|
% self.topic)
|
|
# Add service to the ServiceGroup membership group.
|
|
self.servicegroup_api.join(self.host, self.topic, self)
|
|
|
|
if self.periodic_enable:
|
|
if self.periodic_fuzzy_delay:
|
|
initial_delay = random.randint(0, self.periodic_fuzzy_delay)
|
|
else:
|
|
initial_delay = None
|
|
|
|
self.tg.add_dynamic_timer(self.periodic_tasks,
|
|
initial_delay=initial_delay,
|
|
periodic_interval_max=
|
|
self.periodic_interval_max)
|
|
|
|
def _create_service_ref(self, context):
|
|
svc_values = {
|
|
'host': self.host,
|
|
'binary': self.binary,
|
|
'topic': self.topic,
|
|
'report_count': 0
|
|
}
|
|
service = self.conductor_api.service_create(context, svc_values)
|
|
self.service_id = service['id']
|
|
return service
|
|
|
|
def __getattr__(self, key):
|
|
manager = self.__dict__.get('manager', None)
|
|
return getattr(manager, key)
|
|
|
|
@classmethod
|
|
def create(cls, host=None, binary=None, topic=None, manager=None,
|
|
report_interval=None, periodic_enable=None,
|
|
periodic_fuzzy_delay=None, periodic_interval_max=None,
|
|
db_allowed=True):
|
|
"""Instantiates class and passes back application object.
|
|
|
|
:param host: defaults to CONF.host
|
|
:param binary: defaults to basename of executable
|
|
:param topic: defaults to bin_name - 'nova-' part
|
|
:param manager: defaults to CONF.<topic>_manager
|
|
:param report_interval: defaults to CONF.report_interval
|
|
:param periodic_enable: defaults to CONF.periodic_enable
|
|
:param periodic_fuzzy_delay: defaults to CONF.periodic_fuzzy_delay
|
|
:param periodic_interval_max: if set, the max time to wait between runs
|
|
|
|
"""
|
|
if not host:
|
|
host = CONF.host
|
|
if not binary:
|
|
binary = os.path.basename(sys.argv[0])
|
|
if not topic:
|
|
topic = binary.rpartition('nova-')[2]
|
|
if not manager:
|
|
manager_cls = ('%s_manager' %
|
|
binary.rpartition('nova-')[2])
|
|
manager = CONF.get(manager_cls, None)
|
|
if report_interval is None:
|
|
report_interval = CONF.report_interval
|
|
if periodic_enable is None:
|
|
periodic_enable = CONF.periodic_enable
|
|
if periodic_fuzzy_delay is None:
|
|
periodic_fuzzy_delay = CONF.periodic_fuzzy_delay
|
|
service_obj = cls(host, binary, topic, manager,
|
|
report_interval=report_interval,
|
|
periodic_enable=periodic_enable,
|
|
periodic_fuzzy_delay=periodic_fuzzy_delay,
|
|
periodic_interval_max=periodic_interval_max,
|
|
db_allowed=db_allowed)
|
|
|
|
return service_obj
|
|
|
|
def kill(self):
|
|
"""Destroy the service object in the datastore."""
|
|
self.stop()
|
|
try:
|
|
self.conductor_api.service_destroy(context.get_admin_context(),
|
|
self.service_id)
|
|
except exception.NotFound:
|
|
LOG.warn(_('Service killed that has no database entry'))
|
|
|
|
def stop(self):
|
|
try:
|
|
self.conn.close()
|
|
except Exception:
|
|
pass
|
|
|
|
super(Service, self).stop()
|
|
|
|
def periodic_tasks(self, raise_on_error=False):
|
|
"""Tasks to be run at a periodic interval."""
|
|
ctxt = context.get_admin_context()
|
|
return self.manager.periodic_tasks(ctxt, raise_on_error=raise_on_error)
|
|
|
|
def basic_config_check(self):
|
|
"""Perform basic config checks before starting processing."""
|
|
# Make sure the tempdir exists and is writable
|
|
try:
|
|
with utils.tempdir():
|
|
pass
|
|
except Exception as e:
|
|
LOG.error(_('Temporary directory is invalid: %s'), e)
|
|
sys.exit(1)
|
|
|
|
|
|
class WSGIService(object):
|
|
"""Provides ability to launch API from a 'paste' configuration."""
|
|
|
|
def __init__(self, name, loader=None, use_ssl=False, max_url_len=None):
|
|
"""Initialize, but do not start the WSGI server.
|
|
|
|
:param name: The name of the WSGI server given to the loader.
|
|
:param loader: Loads the WSGI application using the given name.
|
|
:returns: None
|
|
|
|
"""
|
|
self.name = name
|
|
self.manager = self._get_manager()
|
|
self.loader = loader or wsgi.Loader()
|
|
self.app = self.loader.load_app(name)
|
|
self.host = getattr(CONF, '%s_listen' % name, "0.0.0.0")
|
|
self.port = getattr(CONF, '%s_listen_port' % name, 0)
|
|
self.workers = getattr(CONF, '%s_workers' % name, None)
|
|
self.use_ssl = use_ssl
|
|
self.server = wsgi.Server(name,
|
|
self.app,
|
|
host=self.host,
|
|
port=self.port,
|
|
use_ssl=self.use_ssl,
|
|
max_url_len=max_url_len)
|
|
# Pull back actual port used
|
|
self.port = self.server.port
|
|
self.backdoor_port = None
|
|
|
|
def _get_manager(self):
|
|
"""Initialize a Manager object appropriate for this service.
|
|
|
|
Use the service name to look up a Manager subclass from the
|
|
configuration and initialize an instance. If no class name
|
|
is configured, just return None.
|
|
|
|
:returns: a Manager instance, or None.
|
|
|
|
"""
|
|
fl = '%s_manager' % self.name
|
|
if fl not in CONF:
|
|
return None
|
|
|
|
manager_class_name = CONF.get(fl, None)
|
|
if not manager_class_name:
|
|
return None
|
|
|
|
manager_class = importutils.import_class(manager_class_name)
|
|
return manager_class()
|
|
|
|
def start(self):
|
|
"""Start serving this service using loaded configuration.
|
|
|
|
Also, retrieve updated port number in case '0' was passed in, which
|
|
indicates a random port should be used.
|
|
|
|
:returns: None
|
|
|
|
"""
|
|
if self.manager:
|
|
self.manager.init_host()
|
|
self.manager.pre_start_hook()
|
|
if self.backdoor_port is not None:
|
|
self.manager.backdoor_port = self.backdoor_port
|
|
self.server.start()
|
|
if self.manager:
|
|
self.manager.post_start_hook()
|
|
|
|
def stop(self):
|
|
"""Stop serving this API.
|
|
|
|
:returns: None
|
|
|
|
"""
|
|
self.server.stop()
|
|
|
|
def wait(self):
|
|
"""Wait for the service to stop serving this API.
|
|
|
|
:returns: None
|
|
|
|
"""
|
|
self.server.wait()
|
|
|
|
|
|
def process_launcher():
|
|
return service.ProcessLauncher()
|
|
|
|
|
|
# NOTE(vish): the global launcher is to maintain the existing
|
|
# functionality of calling service.serve +
|
|
# service.wait
|
|
_launcher = None
|
|
|
|
|
|
def serve(server, workers=None):
|
|
global _launcher
|
|
if _launcher:
|
|
raise RuntimeError(_('serve() can only be called once'))
|
|
|
|
_launcher = service.launch(server, workers=workers)
|
|
|
|
|
|
def wait():
|
|
_launcher.wait()
|