674ef05b0a
Now that nothing is using publish_service_capabilities it can be disabled, but not removed in order to not break compute rpcapi backwards compatibility. Part of bp no-compute-fanout-to-scheduler. Change-Id: I80c49c46138fd6ee89cb08ffbbced72ada4de72e
300 lines
13 KiB
Python
300 lines
13 KiB
Python
# vim: tabstop=4 shiftwidth=4 softtabstop=4
|
|
|
|
# Copyright (c) 2010 OpenStack Foundation
|
|
# Copyright 2010 United States Government as represented by the
|
|
# Administrator of the National Aeronautics and Space Administration.
|
|
# All Rights Reserved.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
|
|
"""
|
|
Scheduler Service
|
|
"""
|
|
|
|
from oslo.config import cfg
|
|
|
|
from nova.compute import rpcapi as compute_rpcapi
|
|
from nova.compute import task_states
|
|
from nova.compute import utils as compute_utils
|
|
from nova.compute import vm_states
|
|
from nova.conductor import api as conductor_api
|
|
from nova.conductor.tasks import live_migrate
|
|
from nova import exception
|
|
from nova import manager
|
|
from nova.objects import instance as instance_obj
|
|
from nova.openstack.common import excutils
|
|
from nova.openstack.common import importutils
|
|
from nova.openstack.common import jsonutils
|
|
from nova.openstack.common import log as logging
|
|
from nova.openstack.common import periodic_task
|
|
from nova.openstack.common.rpc import common as rpc_common
|
|
from nova import quota
|
|
from nova.scheduler import utils as scheduler_utils
|
|
|
|
|
|
LOG = logging.getLogger(__name__)
|
|
|
|
scheduler_driver_opt = cfg.StrOpt('scheduler_driver',
|
|
default='nova.scheduler.filter_scheduler.FilterScheduler',
|
|
help='Default driver to use for the scheduler')
|
|
|
|
CONF = cfg.CONF
|
|
CONF.register_opt(scheduler_driver_opt)
|
|
|
|
QUOTAS = quota.QUOTAS
|
|
|
|
|
|
class SchedulerManager(manager.Manager):
|
|
"""Chooses a host to run instances on."""
|
|
|
|
RPC_API_VERSION = '2.9'
|
|
|
|
def __init__(self, scheduler_driver=None, *args, **kwargs):
|
|
if not scheduler_driver:
|
|
scheduler_driver = CONF.scheduler_driver
|
|
self.driver = importutils.import_object(scheduler_driver)
|
|
self.compute_rpcapi = compute_rpcapi.ComputeAPI()
|
|
super(SchedulerManager, self).__init__(service_name='scheduler',
|
|
*args, **kwargs)
|
|
|
|
def update_service_capabilities(self, context, service_name,
|
|
host, capabilities):
|
|
"""Process a capability update from a service node."""
|
|
#NOTE(jogo) This is deprecated, but is used by the deprecated
|
|
# publish_service_capabilities call. So this can begin its removal
|
|
# process once publish_service_capabilities is removed.
|
|
if not isinstance(capabilities, list):
|
|
capabilities = [capabilities]
|
|
for capability in capabilities:
|
|
if capability is None:
|
|
capability = {}
|
|
self.driver.update_service_capabilities(service_name, host,
|
|
capability)
|
|
|
|
def create_volume(self, context, volume_id, snapshot_id,
|
|
reservations=None, image_id=None):
|
|
#function removed in RPC API 2.3
|
|
pass
|
|
|
|
@rpc_common.client_exceptions(exception.NoValidHost,
|
|
exception.ComputeServiceUnavailable,
|
|
exception.InvalidHypervisorType,
|
|
exception.UnableToMigrateToSelf,
|
|
exception.DestinationHypervisorTooOld,
|
|
exception.InvalidLocalStorage,
|
|
exception.InvalidSharedStorage,
|
|
exception.MigrationPreCheckError)
|
|
def live_migration(self, context, instance, dest,
|
|
block_migration, disk_over_commit):
|
|
try:
|
|
self._schedule_live_migration(context, instance, dest,
|
|
block_migration, disk_over_commit)
|
|
except (exception.NoValidHost,
|
|
exception.ComputeServiceUnavailable,
|
|
exception.InvalidHypervisorType,
|
|
exception.UnableToMigrateToSelf,
|
|
exception.DestinationHypervisorTooOld,
|
|
exception.InvalidLocalStorage,
|
|
exception.InvalidSharedStorage,
|
|
exception.MigrationPreCheckError) as ex:
|
|
request_spec = {'instance_properties': {
|
|
'uuid': instance['uuid'], },
|
|
}
|
|
with excutils.save_and_reraise_exception():
|
|
self._set_vm_state_and_notify('live_migration',
|
|
dict(vm_state=instance['vm_state'],
|
|
task_state=None,
|
|
expected_task_state=task_states.MIGRATING,),
|
|
context, ex, request_spec)
|
|
except Exception as ex:
|
|
request_spec = {'instance_properties': {
|
|
'uuid': instance['uuid'], },
|
|
}
|
|
with excutils.save_and_reraise_exception():
|
|
self._set_vm_state_and_notify('live_migration',
|
|
{'vm_state': vm_states.ERROR},
|
|
context, ex, request_spec)
|
|
|
|
def _schedule_live_migration(self, context, instance, dest,
|
|
block_migration, disk_over_commit):
|
|
task = live_migrate.LiveMigrationTask(context, instance,
|
|
dest, block_migration, disk_over_commit)
|
|
return task.execute()
|
|
|
|
def run_instance(self, context, request_spec, admin_password,
|
|
injected_files, requested_networks, is_first_time,
|
|
filter_properties, legacy_bdm_in_spec=True):
|
|
"""Tries to call schedule_run_instance on the driver.
|
|
Sets instance vm_state to ERROR on exceptions
|
|
"""
|
|
instance_uuids = request_spec['instance_uuids']
|
|
with compute_utils.EventReporter(context, conductor_api.LocalAPI(),
|
|
'schedule', *instance_uuids):
|
|
try:
|
|
return self.driver.schedule_run_instance(context,
|
|
request_spec, admin_password, injected_files,
|
|
requested_networks, is_first_time, filter_properties,
|
|
legacy_bdm_in_spec)
|
|
|
|
except exception.NoValidHost as ex:
|
|
# don't re-raise
|
|
self._set_vm_state_and_notify('run_instance',
|
|
{'vm_state': vm_states.ERROR,
|
|
'task_state': None},
|
|
context, ex, request_spec)
|
|
except Exception as ex:
|
|
with excutils.save_and_reraise_exception():
|
|
self._set_vm_state_and_notify('run_instance',
|
|
{'vm_state': vm_states.ERROR,
|
|
'task_state': None},
|
|
context, ex, request_spec)
|
|
|
|
def prep_resize(self, context, image, request_spec, filter_properties,
|
|
instance, instance_type, reservations):
|
|
"""Tries to call schedule_prep_resize on the driver.
|
|
Sets instance vm_state to ACTIVE on NoHostFound
|
|
Sets vm_state to ERROR on other exceptions
|
|
"""
|
|
instance_uuid = instance['uuid']
|
|
with compute_utils.EventReporter(context, conductor_api.LocalAPI(),
|
|
'schedule', instance_uuid):
|
|
try:
|
|
request_spec['num_instances'] = len(
|
|
request_spec['instance_uuids'])
|
|
hosts = self.driver.select_destinations(
|
|
context, request_spec, filter_properties)
|
|
host_state = hosts[0]
|
|
|
|
scheduler_utils.populate_filter_properties(filter_properties,
|
|
host_state)
|
|
# context is not serializable
|
|
filter_properties.pop('context', None)
|
|
|
|
(host, node) = (host_state['host'], host_state['nodename'])
|
|
attrs = ['metadata', 'system_metadata', 'info_cache',
|
|
'security_groups']
|
|
inst_obj = instance_obj.Instance._from_db_object(
|
|
context, instance_obj.Instance(), instance,
|
|
expected_attrs=attrs)
|
|
self.compute_rpcapi.prep_resize(
|
|
context, image, inst_obj, instance_type, host,
|
|
reservations, request_spec=request_spec,
|
|
filter_properties=filter_properties, node=node)
|
|
|
|
except exception.NoValidHost as ex:
|
|
vm_state = instance.get('vm_state', vm_states.ACTIVE)
|
|
self._set_vm_state_and_notify('prep_resize',
|
|
{'vm_state': vm_state,
|
|
'task_state': None},
|
|
context, ex, request_spec)
|
|
if reservations:
|
|
QUOTAS.rollback(context, reservations)
|
|
except Exception as ex:
|
|
with excutils.save_and_reraise_exception():
|
|
self._set_vm_state_and_notify('prep_resize',
|
|
{'vm_state': vm_states.ERROR,
|
|
'task_state': None},
|
|
context, ex, request_spec)
|
|
if reservations:
|
|
QUOTAS.rollback(context, reservations)
|
|
|
|
def _set_vm_state_and_notify(self, method, updates, context, ex,
|
|
request_spec):
|
|
scheduler_utils.set_vm_state_and_notify(
|
|
context, 'scheduler', method, updates, ex, request_spec, self.db)
|
|
|
|
# NOTE(hanlind): This method can be removed in v3.0 of the RPC API.
|
|
def show_host_resources(self, context, host):
|
|
"""Shows the physical/usage resource given by hosts.
|
|
|
|
:param context: security context
|
|
:param host: hostname
|
|
:returns:
|
|
example format is below::
|
|
|
|
{'resource':D, 'usage':{proj_id1:D, proj_id2:D}}
|
|
D: {'vcpus': 3, 'memory_mb': 2048, 'local_gb': 2048,
|
|
'vcpus_used': 12, 'memory_mb_used': 10240,
|
|
'local_gb_used': 64}
|
|
|
|
"""
|
|
# Getting compute node info and related instances info
|
|
service_ref = self.db.service_get_by_compute_host(context, host)
|
|
instance_refs = self.db.instance_get_all_by_host(context,
|
|
service_ref['host'])
|
|
|
|
# Getting total available/used resource
|
|
compute_ref = service_ref['compute_node'][0]
|
|
resource = {'vcpus': compute_ref['vcpus'],
|
|
'memory_mb': compute_ref['memory_mb'],
|
|
'local_gb': compute_ref['local_gb'],
|
|
'vcpus_used': compute_ref['vcpus_used'],
|
|
'memory_mb_used': compute_ref['memory_mb_used'],
|
|
'local_gb_used': compute_ref['local_gb_used']}
|
|
usage = dict()
|
|
if not instance_refs:
|
|
return {'resource': resource, 'usage': usage}
|
|
|
|
# Getting usage resource per project
|
|
project_ids = [i['project_id'] for i in instance_refs]
|
|
project_ids = list(set(project_ids))
|
|
for project_id in project_ids:
|
|
vcpus = [i['vcpus'] for i in instance_refs
|
|
if i['project_id'] == project_id]
|
|
|
|
mem = [i['memory_mb'] for i in instance_refs
|
|
if i['project_id'] == project_id]
|
|
|
|
root = [i['root_gb'] for i in instance_refs
|
|
if i['project_id'] == project_id]
|
|
|
|
ephemeral = [i['ephemeral_gb'] for i in instance_refs
|
|
if i['project_id'] == project_id]
|
|
|
|
usage[project_id] = {'vcpus': sum(vcpus),
|
|
'memory_mb': sum(mem),
|
|
'root_gb': sum(root),
|
|
'ephemeral_gb': sum(ephemeral)}
|
|
|
|
return {'resource': resource, 'usage': usage}
|
|
|
|
@periodic_task.periodic_task
|
|
def _expire_reservations(self, context):
|
|
QUOTAS.expire(context)
|
|
|
|
# NOTE(russellb) This method can be removed in 3.0 of this API. It is
|
|
# deprecated in favor of the method in the base API.
|
|
def get_backdoor_port(self, context):
|
|
return self.backdoor_port
|
|
|
|
@rpc_common.client_exceptions(exception.NoValidHost)
|
|
def select_hosts(self, context, request_spec, filter_properties):
|
|
"""Returns host(s) best suited for this request_spec
|
|
and filter_properties.
|
|
"""
|
|
hosts = self.driver.select_hosts(context, request_spec,
|
|
filter_properties)
|
|
return jsonutils.to_primitive(hosts)
|
|
|
|
@rpc_common.client_exceptions(exception.NoValidHost)
|
|
def select_destinations(self, context, request_spec, filter_properties):
|
|
"""Returns destinations(s) best suited for this request_spec and
|
|
filter_properties.
|
|
|
|
The result should be a list of dicts with 'host', 'nodename' and
|
|
'limits' as keys.
|
|
"""
|
|
dests = self.driver.select_destinations(context, request_spec,
|
|
filter_properties)
|
|
return jsonutils.to_primitive(dests)
|