Merge "Revert "Remove ClusteredComputeManager""
This commit is contained in:
0
ironic/nova/__init__.py
Normal file
0
ironic/nova/__init__.py
Normal file
0
ironic/nova/compute/__init__.py
Normal file
0
ironic/nova/compute/__init__.py
Normal file
122
ironic/nova/compute/manager.py
Normal file
122
ironic/nova/compute/manager.py
Normal file
@@ -0,0 +1,122 @@
|
|||||||
|
# coding=utf-8
|
||||||
|
#
|
||||||
|
# Copyright 2014 Red Hat, Inc.
|
||||||
|
# Copyright 2013 Hewlett-Packard Development Company, L.P.
|
||||||
|
# All Rights Reserved.
|
||||||
|
#
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||||
|
# not use this file except in compliance with the License. You may obtain
|
||||||
|
# a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||||
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||||
|
# License for the specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
|
||||||
|
"""Short term workaround for friction in the Nova compute manager with Ironic.
|
||||||
|
|
||||||
|
https://etherpad.openstack.org/p/ironic-nova-friction contains current design
|
||||||
|
work. The goal here is to generalise the areas where n-c talking to a clustered
|
||||||
|
hypervisor has issues, and long term fold them into the main ComputeManager.
|
||||||
|
"""
|
||||||
|
|
||||||
|
from oslo_concurrency import lockutils
|
||||||
|
from oslo_log import log as logging
|
||||||
|
|
||||||
|
from nova.compute import manager
|
||||||
|
import nova.context
|
||||||
|
from nova.i18n import _LW
|
||||||
|
|
||||||
|
|
||||||
|
LOG = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
CCM_SEMAPHORE = 'clustered_compute_manager'
|
||||||
|
|
||||||
|
|
||||||
|
# TODO(jroll) delete this in Ocata
|
||||||
|
class ClusteredComputeManager(manager.ComputeManager):
|
||||||
|
|
||||||
|
def __init__(self, *args, **kwargs):
|
||||||
|
LOG.warning(_LW('ClusteredComputeManager is deprecated. As of the '
|
||||||
|
'Newton release of Nova, the ironic '
|
||||||
|
'virt driver directly supports using multiple '
|
||||||
|
'compute hosts without ClusteredComputeManager. '
|
||||||
|
'Users should unset the `compute_manager` '
|
||||||
|
'configuration option to use Nova\'s default '
|
||||||
|
'instead. ClusteredComputeManager will be removed '
|
||||||
|
'during the Ocata cycle.'))
|
||||||
|
super(ClusteredComputeManager, self).__init__(*args, **kwargs)
|
||||||
|
|
||||||
|
def init_host(self):
|
||||||
|
"""Initialization for a clustered compute service."""
|
||||||
|
self.driver.init_host(host=self.host)
|
||||||
|
# Not used currently.
|
||||||
|
# context = nova.context.get_admin_context()
|
||||||
|
# instances = instance_obj.InstanceList.get_by_host(
|
||||||
|
# context, self.host, expected_attrs=['info_cache'])
|
||||||
|
|
||||||
|
# defer_iptables_apply is moot for clusters - no local iptables
|
||||||
|
# if CONF.defer_iptables_apply:
|
||||||
|
# self.driver.filter_defer_apply_on()
|
||||||
|
|
||||||
|
self.init_virt_events()
|
||||||
|
|
||||||
|
# try:
|
||||||
|
# evacuation is moot for a clustered hypervisor
|
||||||
|
# # checking that instance was not already evacuated to other host
|
||||||
|
# self._destroy_evacuated_instances(context)
|
||||||
|
# Don't run _init_instance until we solve the partitioning problem
|
||||||
|
# - with N n-cpu's all claiming the same hostname, running
|
||||||
|
# _init_instance here would lead to race conditions where each runs
|
||||||
|
# _init_instance concurrently.
|
||||||
|
# for instance in instances:
|
||||||
|
# self._init_instance(context, instance)
|
||||||
|
# finally:
|
||||||
|
# defer_iptables_apply is moot for clusters - no local iptables
|
||||||
|
# if CONF.defer_iptables_apply:
|
||||||
|
# self.driver.filter_defer_apply_off()
|
||||||
|
|
||||||
|
def pre_start_hook(self):
|
||||||
|
"""Update our available resources
|
||||||
|
|
||||||
|
After the service is initialized, but before we fully bring
|
||||||
|
the service up by listening on RPC queues, make sure to update
|
||||||
|
our available resources (and indirectly our available nodes).
|
||||||
|
"""
|
||||||
|
# This is an optimisation to immediately advertise resources but
|
||||||
|
# the periodic task will update them eventually anyway, so ignore
|
||||||
|
# errors as they may be transient (e.g. the scheduler isn't
|
||||||
|
# available...). XXX(lifeless) this applies to all ComputeManagers
|
||||||
|
# and once I feature freeze is over we should push that to nova
|
||||||
|
# directly.
|
||||||
|
try:
|
||||||
|
self.update_available_resource(nova.context.get_admin_context())
|
||||||
|
except Exception:
|
||||||
|
pass
|
||||||
|
|
||||||
|
@lockutils.synchronized(CCM_SEMAPHORE, 'ironic-')
|
||||||
|
def _update_resources(self, node_uuid):
|
||||||
|
"""Update the specified node resource
|
||||||
|
|
||||||
|
Updates the resources for instance while protecting against a race on
|
||||||
|
self._resource_tracker_dict.
|
||||||
|
:param node_uuid: UUID of the Ironic node to update resources for.
|
||||||
|
"""
|
||||||
|
self.update_available_resource_for_node(
|
||||||
|
nova.context.get_admin_context(), node_uuid)
|
||||||
|
|
||||||
|
def terminate_instance(self, context, instance, bdms, reservations):
|
||||||
|
"""Terminate an instance on a node.
|
||||||
|
|
||||||
|
We override this method and force a post-termination update to Nova's
|
||||||
|
resources. This avoids having to wait for a Nova periodic task tick
|
||||||
|
before nodes can be reused.
|
||||||
|
"""
|
||||||
|
super(ClusteredComputeManager, self).terminate_instance(context,
|
||||||
|
instance,
|
||||||
|
bdms,
|
||||||
|
reservations)
|
||||||
|
self._update_resources(instance.node)
|
@@ -1,5 +0,0 @@
|
|||||||
---
|
|
||||||
upgrade:
|
|
||||||
- |
|
|
||||||
The deprecated ironic.nova.ClusteredComputerManager module is now removed.
|
|
||||||
This is not required with nova >= 14.0.0 (Newton).
|
|
Reference in New Issue
Block a user