Merge "Shrink Simple Scheduler"
This commit is contained in:
@@ -18,7 +18,9 @@
|
|||||||
# under the License.
|
# under the License.
|
||||||
|
|
||||||
"""
|
"""
|
||||||
Simple Scheduler
|
Simple Scheduler - for Volumes
|
||||||
|
|
||||||
|
Note: Deprecated in Folsom. Will be removed along with nova-volumes
|
||||||
"""
|
"""
|
||||||
|
|
||||||
from nova import db
|
from nova import db
|
||||||
@@ -31,18 +33,9 @@ from nova import utils
|
|||||||
|
|
||||||
|
|
||||||
simple_scheduler_opts = [
|
simple_scheduler_opts = [
|
||||||
cfg.IntOpt("max_cores",
|
|
||||||
default=16,
|
|
||||||
help="maximum number of instance cores to allow per host"),
|
|
||||||
cfg.IntOpt("max_gigabytes",
|
cfg.IntOpt("max_gigabytes",
|
||||||
default=10000,
|
default=10000,
|
||||||
help="maximum number of volume gigabytes to allow per host"),
|
help="maximum number of volume gigabytes to allow per host"),
|
||||||
cfg.IntOpt("max_networks",
|
|
||||||
default=1000,
|
|
||||||
help="maximum number of networks to allow per host"),
|
|
||||||
cfg.BoolOpt('skip_isolated_core_check',
|
|
||||||
default=True,
|
|
||||||
help='Allow overcommitting vcpus on isolated hosts'),
|
|
||||||
]
|
]
|
||||||
|
|
||||||
FLAGS = flags.FLAGS
|
FLAGS = flags.FLAGS
|
||||||
@@ -52,64 +45,6 @@ FLAGS.register_opts(simple_scheduler_opts)
|
|||||||
class SimpleScheduler(chance.ChanceScheduler):
|
class SimpleScheduler(chance.ChanceScheduler):
|
||||||
"""Implements Naive Scheduler that tries to find least loaded host."""
|
"""Implements Naive Scheduler that tries to find least loaded host."""
|
||||||
|
|
||||||
def _schedule_instance(self, context, instance_opts, *_args, **_kwargs):
|
|
||||||
"""Picks a host that is up and has the fewest running instances."""
|
|
||||||
elevated = context.elevated()
|
|
||||||
|
|
||||||
availability_zone = instance_opts.get('availability_zone')
|
|
||||||
|
|
||||||
zone, host = FLAGS.default_schedule_zone, None
|
|
||||||
if availability_zone:
|
|
||||||
zone, _x, host = availability_zone.partition(':')
|
|
||||||
|
|
||||||
if host and context.is_admin:
|
|
||||||
service = db.service_get_by_args(elevated, host, 'nova-compute')
|
|
||||||
if not utils.service_is_up(service):
|
|
||||||
raise exception.WillNotSchedule(host=host)
|
|
||||||
return host
|
|
||||||
|
|
||||||
results = db.service_get_all_compute_sorted(elevated)
|
|
||||||
in_isolation = instance_opts['image_ref'] in FLAGS.isolated_images
|
|
||||||
check_cores = not in_isolation or not FLAGS.skip_isolated_core_check
|
|
||||||
if zone:
|
|
||||||
results = [(service, cores) for (service, cores) in results
|
|
||||||
if service['availability_zone'] == zone]
|
|
||||||
for result in results:
|
|
||||||
(service, instance_cores) = result
|
|
||||||
if in_isolation and service['host'] not in FLAGS.isolated_hosts:
|
|
||||||
# isloated images run on isolated hosts
|
|
||||||
continue
|
|
||||||
if service['host'] in FLAGS.isolated_hosts and not in_isolation:
|
|
||||||
# images that aren't isolated only run on general hosts
|
|
||||||
continue
|
|
||||||
if (check_cores and
|
|
||||||
instance_cores + instance_opts['vcpus'] > FLAGS.max_cores):
|
|
||||||
msg = _("Not enough allocatable CPU cores remaining")
|
|
||||||
raise exception.NoValidHost(reason=msg)
|
|
||||||
if utils.service_is_up(service) and not service['disabled']:
|
|
||||||
return service['host']
|
|
||||||
msg = _("Is the appropriate service running?")
|
|
||||||
raise exception.NoValidHost(reason=msg)
|
|
||||||
|
|
||||||
def schedule_run_instance(self, context, request_spec, reservations,
|
|
||||||
*_args, **_kwargs):
|
|
||||||
num_instances = request_spec.get('num_instances', 1)
|
|
||||||
instances = []
|
|
||||||
for num in xrange(num_instances):
|
|
||||||
host = self._schedule_instance(context,
|
|
||||||
request_spec['instance_properties'], *_args, **_kwargs)
|
|
||||||
request_spec['instance_properties']['launch_index'] = num
|
|
||||||
instance_ref = self.create_instance_db_entry(context,
|
|
||||||
request_spec, reservations)
|
|
||||||
driver.cast_to_compute_host(context, host, 'run_instance',
|
|
||||||
instance_uuid=instance_ref['uuid'], **_kwargs)
|
|
||||||
instances.append(driver.encode_instance(instance_ref))
|
|
||||||
# So if we loop around, create_instance_db_entry will actually
|
|
||||||
# create a new entry, instead of assume it's been created
|
|
||||||
# already
|
|
||||||
del request_spec['instance_properties']['uuid']
|
|
||||||
return instances
|
|
||||||
|
|
||||||
def schedule_create_volume(self, context, volume_id, *_args, **_kwargs):
|
def schedule_create_volume(self, context, volume_id, *_args, **_kwargs):
|
||||||
"""Picks a host that is up and has the fewest volumes."""
|
"""Picks a host that is up and has the fewest volumes."""
|
||||||
elevated = context.elevated()
|
elevated = context.elevated()
|
||||||
|
|||||||
@@ -143,3 +143,43 @@ class MultiDriverTestCase(test_scheduler.SchedulerTestCase):
|
|||||||
mgr.update_service_capabilities('foo_svc', 'foo_host', 'foo_caps')
|
mgr.update_service_capabilities('foo_svc', 'foo_host', 'foo_caps')
|
||||||
self.assertTrue(mgr.drivers['compute'].is_update_caps_called)
|
self.assertTrue(mgr.drivers['compute'].is_update_caps_called)
|
||||||
self.assertTrue(mgr.drivers['volume'].is_update_caps_called)
|
self.assertTrue(mgr.drivers['volume'].is_update_caps_called)
|
||||||
|
|
||||||
|
|
||||||
|
class SimpleSchedulerTestCase(MultiDriverTestCase):
|
||||||
|
"""Test case for simple driver."""
|
||||||
|
|
||||||
|
driver_cls = multi.MultiScheduler
|
||||||
|
|
||||||
|
def setUp(self):
|
||||||
|
super(SimpleSchedulerTestCase, self).setUp()
|
||||||
|
base_name = 'nova.tests.scheduler.test_multi_scheduler.%s'
|
||||||
|
compute_cls_name = base_name % 'FakeComputeScheduler'
|
||||||
|
volume_cls_name = 'nova.scheduler.simple.SimpleScheduler'
|
||||||
|
default_cls_name = base_name % 'FakeDefaultScheduler'
|
||||||
|
self.flags(compute_scheduler_driver=compute_cls_name,
|
||||||
|
volume_scheduler_driver=volume_cls_name,
|
||||||
|
default_scheduler_driver=default_cls_name)
|
||||||
|
self._manager = multi.MultiScheduler()
|
||||||
|
|
||||||
|
def test_update_service_capabilities(self):
|
||||||
|
def fake_update_service_capabilities(self, service, host, caps):
|
||||||
|
self.is_update_caps_called = True
|
||||||
|
|
||||||
|
mgr = self._manager
|
||||||
|
self.stubs.Set(driver.Scheduler,
|
||||||
|
'update_service_capabilities',
|
||||||
|
fake_update_service_capabilities)
|
||||||
|
self.assertFalse(mgr.drivers['compute'].is_update_caps_called)
|
||||||
|
mgr.update_service_capabilities('foo_svc', 'foo_host', 'foo_caps')
|
||||||
|
self.assertTrue(mgr.drivers['compute'].is_update_caps_called)
|
||||||
|
self.assertTrue(mgr.drivers['volume'].is_update_caps_called)
|
||||||
|
|
||||||
|
def test_drivers_inited(self):
|
||||||
|
mgr = self._manager
|
||||||
|
self.assertEqual(len(mgr.drivers), 3)
|
||||||
|
self.assertTrue(mgr.drivers['compute'].is_fake_compute)
|
||||||
|
self.assertTrue(mgr.drivers['volume'] is not None)
|
||||||
|
self.assertTrue(mgr.drivers['default'].is_fake_default)
|
||||||
|
|
||||||
|
def test_proxy_calls(self):
|
||||||
|
pass
|
||||||
|
|||||||
Reference in New Issue
Block a user