merge with trunk
This commit is contained in:
commit
3f2f5d8944
1
Authors
1
Authors
@ -30,6 +30,7 @@ Ilya Alekseyev <ialekseev@griddynamics.com>
|
||||
Jason Koelker <jason@koelker.net>
|
||||
Jay Pipes <jaypipes@gmail.com>
|
||||
Jesse Andrews <anotherjesse@gmail.com>
|
||||
Jimmy Bergman <jimmy@sigint.se>
|
||||
Joe Heck <heckj@mac.com>
|
||||
Joel Moore <joelbm24@gmail.com>
|
||||
Johannes Erdfelt <johannes.erdfelt@rackspace.com>
|
||||
|
@ -187,7 +187,7 @@ class CloudController(object):
|
||||
'mpi': mpi}}
|
||||
|
||||
for image_type in ['kernel', 'ramdisk']:
|
||||
if '%s_id' % image_type in instance_ref:
|
||||
if instance_ref.get('%s_id' % image_type):
|
||||
ec2_id = self._image_ec2_id(instance_ref['%s_id' % image_type],
|
||||
self._image_type(image_type))
|
||||
data['meta-data']['%s-id' % image_type] = ec2_id
|
||||
@ -613,7 +613,7 @@ class CloudController(object):
|
||||
# TODO(vish): Instance should be None at db layer instead of
|
||||
# trying to lazy load, but for now we turn it into
|
||||
# a dict to avoid an error.
|
||||
return {'volumeSet': [self._format_volume(context, dict(volume))]}
|
||||
return self._format_volume(context, dict(volume))
|
||||
|
||||
def delete_volume(self, context, volume_id, **kwargs):
|
||||
volume_id = ec2utils.ec2_id_to_id(volume_id)
|
||||
@ -726,7 +726,9 @@ class CloudController(object):
|
||||
instance['mac_address'])
|
||||
|
||||
i['privateDnsName'] = fixed_addr
|
||||
i['privateIpAddress'] = fixed_addr
|
||||
i['publicDnsName'] = floating_addr
|
||||
i['ipAddress'] = floating_addr or fixed_addr
|
||||
i['dnsName'] = i['publicDnsName'] or i['privateDnsName']
|
||||
i['keyName'] = instance['key_name']
|
||||
|
||||
|
@ -40,7 +40,7 @@ import nova.api.openstack
|
||||
from nova.scheduler import api as scheduler_api
|
||||
|
||||
|
||||
LOG = logging.getLogger('server')
|
||||
LOG = logging.getLogger('nova.api.openstack.servers')
|
||||
FLAGS = flags.FLAGS
|
||||
|
||||
|
||||
@ -321,6 +321,7 @@ class Controller(common.OpenstackController):
|
||||
return exc.HTTPAccepted()
|
||||
|
||||
def _action_rebuild(self, input_dict, req, id):
|
||||
LOG.debug(_("Rebuild server action is not implemented"))
|
||||
return faults.Fault(exc.HTTPNotImplemented())
|
||||
|
||||
def _action_resize(self, input_dict, req, id):
|
||||
@ -336,18 +337,20 @@ class Controller(common.OpenstackController):
|
||||
except Exception, e:
|
||||
LOG.exception(_("Error in resize %s"), e)
|
||||
return faults.Fault(exc.HTTPBadRequest())
|
||||
return faults.Fault(exc.HTTPAccepted())
|
||||
return exc.HTTPAccepted()
|
||||
|
||||
def _action_reboot(self, input_dict, req, id):
|
||||
try:
|
||||
if 'reboot' in input_dict and 'type' in input_dict['reboot']:
|
||||
reboot_type = input_dict['reboot']['type']
|
||||
except Exception:
|
||||
raise faults.Fault(exc.HTTPNotImplemented())
|
||||
else:
|
||||
LOG.exception(_("Missing argument 'type' for reboot"))
|
||||
return faults.Fault(exc.HTTPUnprocessableEntity())
|
||||
try:
|
||||
# TODO(gundlach): pass reboot_type, support soft reboot in
|
||||
# virt driver
|
||||
self.compute_api.reboot(req.environ['nova.context'], id)
|
||||
except:
|
||||
except Exception, e:
|
||||
LOG.exception(_("Error in reboot %s"), e)
|
||||
return faults.Fault(exc.HTTPUnprocessableEntity())
|
||||
return exc.HTTPAccepted()
|
||||
|
||||
|
@ -17,8 +17,7 @@
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""
|
||||
Handles all processes relating to instances (guest vms).
|
||||
"""Handles all processes relating to instances (guest vms).
|
||||
|
||||
The :py:class:`ComputeManager` class is a :py:class:`nova.manager.Manager` that
|
||||
handles RPC calls relating to creating instances. It is responsible for
|
||||
@ -33,6 +32,7 @@ terminating it.
|
||||
by :func:`nova.utils.import_object`
|
||||
:volume_manager: Name of class that handles persistent storage, loaded by
|
||||
:func:`nova.utils.import_object`
|
||||
|
||||
"""
|
||||
|
||||
import datetime
|
||||
@ -55,6 +55,7 @@ from nova import utils
|
||||
from nova.compute import power_state
|
||||
from nova.virt import driver
|
||||
|
||||
|
||||
FLAGS = flags.FLAGS
|
||||
flags.DEFINE_string('instances_path', '$state_path/instances',
|
||||
'where instances are stored on disk')
|
||||
@ -74,19 +75,14 @@ flags.DEFINE_integer("rescue_timeout", 0,
|
||||
"Automatically unrescue an instance after N seconds."
|
||||
" Set to 0 to disable.")
|
||||
|
||||
|
||||
LOG = logging.getLogger('nova.compute.manager')
|
||||
|
||||
|
||||
def checks_instance_lock(function):
|
||||
"""
|
||||
decorator used for preventing action against locked instances
|
||||
unless, of course, you happen to be admin
|
||||
|
||||
"""
|
||||
|
||||
"""Decorator to prevent action against locked instances for non-admins."""
|
||||
@functools.wraps(function)
|
||||
def decorated_function(self, context, instance_id, *args, **kwargs):
|
||||
|
||||
LOG.info(_("check_instance_lock: decorating: |%s|"), function,
|
||||
context=context)
|
||||
LOG.info(_("check_instance_lock: arguments: |%(self)s| |%(context)s|"
|
||||
@ -112,7 +108,6 @@ def checks_instance_lock(function):
|
||||
|
||||
|
||||
class ComputeManager(manager.SchedulerDependentManager):
|
||||
|
||||
"""Manages the running instances from creation to destruction."""
|
||||
|
||||
def __init__(self, compute_driver=None, *args, **kwargs):
|
||||
@ -136,9 +131,7 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
*args, **kwargs)
|
||||
|
||||
def init_host(self):
|
||||
"""Do any initialization that needs to be run if this is a
|
||||
standalone service.
|
||||
"""
|
||||
"""Initialization for a standalone compute service."""
|
||||
self.driver.init_host(host=self.host)
|
||||
|
||||
def _update_state(self, context, instance_id):
|
||||
@ -153,16 +146,18 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
self.db.instance_set_state(context, instance_id, state)
|
||||
|
||||
def get_console_topic(self, context, **kwargs):
|
||||
"""Retrieves the console host for a project on this host
|
||||
Currently this is just set in the flags for each compute
|
||||
host."""
|
||||
"""Retrieves the console host for a project on this host.
|
||||
|
||||
Currently this is just set in the flags for each compute host.
|
||||
|
||||
"""
|
||||
#TODO(mdragon): perhaps make this variable by console_type?
|
||||
return self.db.queue_get_for(context,
|
||||
FLAGS.console_topic,
|
||||
FLAGS.console_host)
|
||||
|
||||
def get_network_topic(self, context, **kwargs):
|
||||
"""Retrieves the network host for a project on this host"""
|
||||
"""Retrieves the network host for a project on this host."""
|
||||
# TODO(vish): This method should be memoized. This will make
|
||||
# the call to get_network_host cheaper, so that
|
||||
# it can pas messages instead of checking the db
|
||||
@ -179,15 +174,23 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
return self.driver.get_console_pool_info(console_type)
|
||||
|
||||
@exception.wrap_exception
|
||||
def refresh_security_group_rules(self, context,
|
||||
security_group_id, **kwargs):
|
||||
"""This call passes straight through to the virtualization driver."""
|
||||
def refresh_security_group_rules(self, context, security_group_id,
|
||||
**kwargs):
|
||||
"""Tell the virtualization driver to refresh security group rules.
|
||||
|
||||
Passes straight through to the virtualization driver.
|
||||
|
||||
"""
|
||||
return self.driver.refresh_security_group_rules(security_group_id)
|
||||
|
||||
@exception.wrap_exception
|
||||
def refresh_security_group_members(self, context,
|
||||
security_group_id, **kwargs):
|
||||
"""This call passes straight through to the virtualization driver."""
|
||||
"""Tell the virtualization driver to refresh security group members.
|
||||
|
||||
Passes straight through to the virtualization driver.
|
||||
|
||||
"""
|
||||
return self.driver.refresh_security_group_members(security_group_id)
|
||||
|
||||
@exception.wrap_exception
|
||||
@ -249,7 +252,7 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
@exception.wrap_exception
|
||||
@checks_instance_lock
|
||||
def terminate_instance(self, context, instance_id):
|
||||
"""Terminate an instance on this machine."""
|
||||
"""Terminate an instance on this host."""
|
||||
context = context.elevated()
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
LOG.audit(_("Terminating instance %s"), instance_id, context=context)
|
||||
@ -297,7 +300,7 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
@exception.wrap_exception
|
||||
@checks_instance_lock
|
||||
def reboot_instance(self, context, instance_id):
|
||||
"""Reboot an instance on this server."""
|
||||
"""Reboot an instance on this host."""
|
||||
context = context.elevated()
|
||||
self._update_state(context, instance_id)
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
@ -321,7 +324,7 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
|
||||
@exception.wrap_exception
|
||||
def snapshot_instance(self, context, instance_id, image_id):
|
||||
"""Snapshot an instance on this server."""
|
||||
"""Snapshot an instance on this host."""
|
||||
context = context.elevated()
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
|
||||
@ -344,7 +347,7 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
@exception.wrap_exception
|
||||
@checks_instance_lock
|
||||
def set_admin_password(self, context, instance_id, new_pass=None):
|
||||
"""Set the root/admin password for an instance on this server."""
|
||||
"""Set the root/admin password for an instance on this host."""
|
||||
context = context.elevated()
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
instance_id = instance_ref['id']
|
||||
@ -365,7 +368,7 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
@exception.wrap_exception
|
||||
@checks_instance_lock
|
||||
def inject_file(self, context, instance_id, path, file_contents):
|
||||
"""Write a file to the specified path on an instance on this server"""
|
||||
"""Write a file to the specified path in an instance on this host."""
|
||||
context = context.elevated()
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
instance_id = instance_ref['id']
|
||||
@ -383,44 +386,34 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
@exception.wrap_exception
|
||||
@checks_instance_lock
|
||||
def rescue_instance(self, context, instance_id):
|
||||
"""Rescue an instance on this server."""
|
||||
"""Rescue an instance on this host."""
|
||||
context = context.elevated()
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
LOG.audit(_('instance %s: rescuing'), instance_id, context=context)
|
||||
self.db.instance_set_state(
|
||||
context,
|
||||
instance_id,
|
||||
power_state.NOSTATE,
|
||||
'rescuing')
|
||||
self.db.instance_set_state(context,
|
||||
instance_id,
|
||||
power_state.NOSTATE,
|
||||
'rescuing')
|
||||
self.network_manager.setup_compute_network(context, instance_id)
|
||||
self.driver.rescue(
|
||||
instance_ref,
|
||||
lambda result: self._update_state_callback(
|
||||
self,
|
||||
context,
|
||||
instance_id,
|
||||
result))
|
||||
_update_state = lambda result: self._update_state_callback(
|
||||
self, context, instance_id, result)
|
||||
self.driver.rescue(instance_ref, _update_state)
|
||||
self._update_state(context, instance_id)
|
||||
|
||||
@exception.wrap_exception
|
||||
@checks_instance_lock
|
||||
def unrescue_instance(self, context, instance_id):
|
||||
"""Rescue an instance on this server."""
|
||||
"""Rescue an instance on this host."""
|
||||
context = context.elevated()
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
LOG.audit(_('instance %s: unrescuing'), instance_id, context=context)
|
||||
self.db.instance_set_state(
|
||||
context,
|
||||
instance_id,
|
||||
power_state.NOSTATE,
|
||||
'unrescuing')
|
||||
self.driver.unrescue(
|
||||
instance_ref,
|
||||
lambda result: self._update_state_callback(
|
||||
self,
|
||||
context,
|
||||
instance_id,
|
||||
result))
|
||||
self.db.instance_set_state(context,
|
||||
instance_id,
|
||||
power_state.NOSTATE,
|
||||
'unrescuing')
|
||||
_update_state = lambda result: self._update_state_callback(
|
||||
self, context, instance_id, result)
|
||||
self.driver.unrescue(instance_ref, _update_state)
|
||||
self._update_state(context, instance_id)
|
||||
|
||||
@staticmethod
|
||||
@ -431,7 +424,7 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
@exception.wrap_exception
|
||||
@checks_instance_lock
|
||||
def confirm_resize(self, context, instance_id, migration_id):
|
||||
"""Destroys the source instance"""
|
||||
"""Destroys the source instance."""
|
||||
context = context.elevated()
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
self.driver.destroy(instance_ref)
|
||||
@ -439,9 +432,12 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
@exception.wrap_exception
|
||||
@checks_instance_lock
|
||||
def revert_resize(self, context, instance_id, migration_id):
|
||||
"""Destroys the new instance on the destination machine,
|
||||
reverts the model changes, and powers on the old
|
||||
instance on the source machine"""
|
||||
"""Destroys the new instance on the destination machine.
|
||||
|
||||
Reverts the model changes, and powers on the old instance on the
|
||||
source machine.
|
||||
|
||||
"""
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
migration_ref = self.db.migration_get(context, migration_id)
|
||||
|
||||
@ -458,9 +454,12 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
@exception.wrap_exception
|
||||
@checks_instance_lock
|
||||
def finish_revert_resize(self, context, instance_id, migration_id):
|
||||
"""Finishes the second half of reverting a resize, powering back on
|
||||
the source instance and reverting the resized attributes in the
|
||||
database"""
|
||||
"""Finishes the second half of reverting a resize.
|
||||
|
||||
Power back on the source instance and revert the resized attributes
|
||||
in the database.
|
||||
|
||||
"""
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
migration_ref = self.db.migration_get(context, migration_id)
|
||||
instance_type = self.db.instance_type_get_by_flavor_id(context,
|
||||
@ -480,8 +479,11 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
@exception.wrap_exception
|
||||
@checks_instance_lock
|
||||
def prep_resize(self, context, instance_id, flavor_id):
|
||||
"""Initiates the process of moving a running instance to another
|
||||
host, possibly changing the RAM and disk size in the process"""
|
||||
"""Initiates the process of moving a running instance to another host.
|
||||
|
||||
Possibly changes the RAM and disk size in the process.
|
||||
|
||||
"""
|
||||
context = context.elevated()
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
if instance_ref['host'] == FLAGS.host:
|
||||
@ -513,35 +515,38 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
@exception.wrap_exception
|
||||
@checks_instance_lock
|
||||
def resize_instance(self, context, instance_id, migration_id):
|
||||
"""Starts the migration of a running instance to another host"""
|
||||
"""Starts the migration of a running instance to another host."""
|
||||
migration_ref = self.db.migration_get(context, migration_id)
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
self.db.migration_update(context, migration_id,
|
||||
{'status': 'migrating', })
|
||||
self.db.migration_update(context,
|
||||
migration_id,
|
||||
{'status': 'migrating'})
|
||||
|
||||
disk_info = self.driver.migrate_disk_and_power_off(instance_ref,
|
||||
migration_ref['dest_host'])
|
||||
self.db.migration_update(context, migration_id,
|
||||
{'status': 'post-migrating', })
|
||||
disk_info = self.driver.migrate_disk_and_power_off(
|
||||
instance_ref, migration_ref['dest_host'])
|
||||
self.db.migration_update(context,
|
||||
migration_id,
|
||||
{'status': 'post-migrating'})
|
||||
|
||||
# Make sure the service exists before sending a message.
|
||||
_service = self.db.service_get_by_host_and_topic(context,
|
||||
migration_ref['dest_compute'], FLAGS.compute_topic)
|
||||
topic = self.db.queue_get_for(context, FLAGS.compute_topic,
|
||||
migration_ref['dest_compute'])
|
||||
rpc.cast(context, topic,
|
||||
{'method': 'finish_resize',
|
||||
'args': {
|
||||
'migration_id': migration_id,
|
||||
'instance_id': instance_id,
|
||||
'disk_info': disk_info, },
|
||||
})
|
||||
service = self.db.service_get_by_host_and_topic(
|
||||
context, migration_ref['dest_compute'], FLAGS.compute_topic)
|
||||
topic = self.db.queue_get_for(context,
|
||||
FLAGS.compute_topic,
|
||||
migration_ref['dest_compute'])
|
||||
rpc.cast(context, topic, {'method': 'finish_resize',
|
||||
'args': {'migration_id': migration_id,
|
||||
'instance_id': instance_id,
|
||||
'disk_info': disk_info}})
|
||||
|
||||
@exception.wrap_exception
|
||||
@checks_instance_lock
|
||||
def finish_resize(self, context, instance_id, migration_id, disk_info):
|
||||
"""Completes the migration process by setting up the newly transferred
|
||||
disk and turning on the instance on its new host machine"""
|
||||
"""Completes the migration process.
|
||||
|
||||
Sets up the newly transferred disk and turns on the instance at its
|
||||
new host machine.
|
||||
|
||||
"""
|
||||
migration_ref = self.db.migration_get(context, migration_id)
|
||||
instance_ref = self.db.instance_get(context,
|
||||
migration_ref['instance_id'])
|
||||
@ -566,7 +571,7 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
@exception.wrap_exception
|
||||
@checks_instance_lock
|
||||
def pause_instance(self, context, instance_id):
|
||||
"""Pause an instance on this server."""
|
||||
"""Pause an instance on this host."""
|
||||
context = context.elevated()
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
LOG.audit(_('instance %s: pausing'), instance_id, context=context)
|
||||
@ -583,7 +588,7 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
@exception.wrap_exception
|
||||
@checks_instance_lock
|
||||
def unpause_instance(self, context, instance_id):
|
||||
"""Unpause a paused instance on this server."""
|
||||
"""Unpause a paused instance on this host."""
|
||||
context = context.elevated()
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
LOG.audit(_('instance %s: unpausing'), instance_id, context=context)
|
||||
@ -599,7 +604,7 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
|
||||
@exception.wrap_exception
|
||||
def get_diagnostics(self, context, instance_id):
|
||||
"""Retrieve diagnostics for an instance on this server."""
|
||||
"""Retrieve diagnostics for an instance on this host."""
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
|
||||
if instance_ref["state"] == power_state.RUNNING:
|
||||
@ -610,10 +615,7 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
@exception.wrap_exception
|
||||
@checks_instance_lock
|
||||
def suspend_instance(self, context, instance_id):
|
||||
"""
|
||||
suspend the instance with instance_id
|
||||
|
||||
"""
|
||||
"""Suspend the given instance."""
|
||||
context = context.elevated()
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
LOG.audit(_('instance %s: suspending'), instance_id, context=context)
|
||||
@ -629,10 +631,7 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
@exception.wrap_exception
|
||||
@checks_instance_lock
|
||||
def resume_instance(self, context, instance_id):
|
||||
"""
|
||||
resume the suspended instance with instance_id
|
||||
|
||||
"""
|
||||
"""Resume the given suspended instance."""
|
||||
context = context.elevated()
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
LOG.audit(_('instance %s: resuming'), instance_id, context=context)
|
||||
@ -647,10 +646,7 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
|
||||
@exception.wrap_exception
|
||||
def lock_instance(self, context, instance_id):
|
||||
"""
|
||||
lock the instance with instance_id
|
||||
|
||||
"""
|
||||
"""Lock the given instance."""
|
||||
context = context.elevated()
|
||||
|
||||
LOG.debug(_('instance %s: locking'), instance_id, context=context)
|
||||
@ -658,10 +654,7 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
|
||||
@exception.wrap_exception
|
||||
def unlock_instance(self, context, instance_id):
|
||||
"""
|
||||
unlock the instance with instance_id
|
||||
|
||||
"""
|
||||
"""Unlock the given instance."""
|
||||
context = context.elevated()
|
||||
|
||||
LOG.debug(_('instance %s: unlocking'), instance_id, context=context)
|
||||
@ -669,10 +662,7 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
|
||||
@exception.wrap_exception
|
||||
def get_lock(self, context, instance_id):
|
||||
"""
|
||||
return the boolean state of (instance with instance_id)'s lock
|
||||
|
||||
"""
|
||||
"""Return the boolean state of the given instance's lock."""
|
||||
context = context.elevated()
|
||||
LOG.debug(_('instance %s: getting locked state'), instance_id,
|
||||
context=context)
|
||||
@ -681,10 +671,7 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
|
||||
@checks_instance_lock
|
||||
def reset_network(self, context, instance_id):
|
||||
"""
|
||||
Reset networking on the instance.
|
||||
|
||||
"""
|
||||
"""Reset networking on the given instance."""
|
||||
context = context.elevated()
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
LOG.debug(_('instance %s: reset network'), instance_id,
|
||||
@ -693,10 +680,7 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
|
||||
@checks_instance_lock
|
||||
def inject_network_info(self, context, instance_id):
|
||||
"""
|
||||
Inject network info for the instance.
|
||||
|
||||
"""
|
||||
"""Inject network info for the given instance."""
|
||||
context = context.elevated()
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
LOG.debug(_('instance %s: inject network info'), instance_id,
|
||||
@ -705,7 +689,7 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
|
||||
@exception.wrap_exception
|
||||
def get_console_output(self, context, instance_id):
|
||||
"""Send the console output for an instance."""
|
||||
"""Send the console output for the given instance."""
|
||||
context = context.elevated()
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
LOG.audit(_("Get console output for instance %s"), instance_id,
|
||||
@ -714,20 +698,18 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
|
||||
@exception.wrap_exception
|
||||
def get_ajax_console(self, context, instance_id):
|
||||
"""Return connection information for an ajax console"""
|
||||
"""Return connection information for an ajax console."""
|
||||
context = context.elevated()
|
||||
LOG.debug(_("instance %s: getting ajax console"), instance_id)
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
|
||||
return self.driver.get_ajax_console(instance_ref)
|
||||
|
||||
@exception.wrap_exception
|
||||
def get_vnc_console(self, context, instance_id):
|
||||
"""Return connection information for an vnc console."""
|
||||
"""Return connection information for a vnc console."""
|
||||
context = context.elevated()
|
||||
LOG.debug(_("instance %s: getting vnc console"), instance_id)
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
|
||||
return self.driver.get_vnc_console(instance_ref)
|
||||
|
||||
@checks_instance_lock
|
||||
@ -781,7 +763,7 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
|
||||
@exception.wrap_exception
|
||||
def compare_cpu(self, context, cpu_info):
|
||||
"""Checks the host cpu is compatible to a cpu given by xml.
|
||||
"""Checks that the host cpu is compatible with a cpu given by xml.
|
||||
|
||||
:param context: security context
|
||||
:param cpu_info: json string obtained from virConnect.getCapabilities
|
||||
@ -802,7 +784,6 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
:returns: tmpfile name(basename)
|
||||
|
||||
"""
|
||||
|
||||
dirpath = FLAGS.instances_path
|
||||
fd, tmp_file = tempfile.mkstemp(dir=dirpath)
|
||||
LOG.debug(_("Creating tmpfile %s to notify to other "
|
||||
@ -819,7 +800,6 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
:param filename: confirm existence of FLAGS.instances_path/thisfile
|
||||
|
||||
"""
|
||||
|
||||
tmp_file = os.path.join(FLAGS.instances_path, filename)
|
||||
if not os.path.exists(tmp_file):
|
||||
raise exception.NotFound(_('%s not found') % tmp_file)
|
||||
@ -832,7 +812,6 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
:param filename: remove existence of FLAGS.instances_path/thisfile
|
||||
|
||||
"""
|
||||
|
||||
tmp_file = os.path.join(FLAGS.instances_path, filename)
|
||||
os.remove(tmp_file)
|
||||
|
||||
@ -844,7 +823,6 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
:returns: See driver.update_available_resource()
|
||||
|
||||
"""
|
||||
|
||||
return self.driver.update_available_resource(context, self.host)
|
||||
|
||||
def pre_live_migration(self, context, instance_id, time=None):
|
||||
@ -854,7 +832,6 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
:param instance_id: nova.db.sqlalchemy.models.Instance.Id
|
||||
|
||||
"""
|
||||
|
||||
if not time:
|
||||
time = greenthread
|
||||
|
||||
@ -913,7 +890,6 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
:param dest: destination host
|
||||
|
||||
"""
|
||||
|
||||
# Get instance for error handling.
|
||||
instance_ref = self.db.instance_get(context, instance_id)
|
||||
i_name = instance_ref.name
|
||||
@ -1009,12 +985,10 @@ class ComputeManager(manager.SchedulerDependentManager):
|
||||
|
||||
:param ctxt: security context
|
||||
:param instance_id: nova.db.sqlalchemy.models.Instance.Id
|
||||
:param host:
|
||||
DB column value is updated by this hostname.
|
||||
if none, the host instance currently running is selected.
|
||||
:param host: DB column value is updated by this hostname.
|
||||
If none, the host instance currently running is selected.
|
||||
|
||||
"""
|
||||
|
||||
if not host:
|
||||
host = instance_ref['host']
|
||||
|
||||
|
@ -15,23 +15,19 @@
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""
|
||||
Handles ConsoleProxy API requests
|
||||
"""
|
||||
"""Handles ConsoleProxy API requests."""
|
||||
|
||||
from nova import exception
|
||||
from nova.db import base
|
||||
|
||||
|
||||
from nova import flags
|
||||
from nova import rpc
|
||||
from nova.db import base
|
||||
|
||||
|
||||
FLAGS = flags.FLAGS
|
||||
|
||||
|
||||
class API(base.Base):
|
||||
"""API for spining up or down console proxy connections"""
|
||||
"""API for spinning up or down console proxy connections."""
|
||||
|
||||
def __init__(self, **kwargs):
|
||||
super(API, self).__init__(**kwargs)
|
||||
@ -51,8 +47,8 @@ class API(base.Base):
|
||||
self.db.queue_get_for(context,
|
||||
FLAGS.console_topic,
|
||||
pool['host']),
|
||||
{"method": "remove_console",
|
||||
"args": {"console_id": console['id']}})
|
||||
{'method': 'remove_console',
|
||||
'args': {'console_id': console['id']}})
|
||||
|
||||
def create_console(self, context, instance_id):
|
||||
instance = self.db.instance_get(context, instance_id)
|
||||
@ -63,13 +59,12 @@ class API(base.Base):
|
||||
# here.
|
||||
rpc.cast(context,
|
||||
self._get_console_topic(context, instance['host']),
|
||||
{"method": "add_console",
|
||||
"args": {"instance_id": instance_id}})
|
||||
{'method': 'add_console',
|
||||
'args': {'instance_id': instance_id}})
|
||||
|
||||
def _get_console_topic(self, context, instance_host):
|
||||
topic = self.db.queue_get_for(context,
|
||||
FLAGS.compute_topic,
|
||||
instance_host)
|
||||
return rpc.call(context,
|
||||
topic,
|
||||
{"method": "get_console_topic", "args": {'fake': 1}})
|
||||
return rpc.call(context, topic, {'method': 'get_console_topic',
|
||||
'args': {'fake': 1}})
|
||||
|
@ -15,9 +15,7 @@
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""
|
||||
Fake ConsoleProxy driver for tests.
|
||||
"""
|
||||
"""Fake ConsoleProxy driver for tests."""
|
||||
|
||||
from nova import exception
|
||||
|
||||
@ -27,32 +25,32 @@ class FakeConsoleProxy(object):
|
||||
|
||||
@property
|
||||
def console_type(self):
|
||||
return "fake"
|
||||
return 'fake'
|
||||
|
||||
def setup_console(self, context, console):
|
||||
"""Sets up actual proxies"""
|
||||
"""Sets up actual proxies."""
|
||||
pass
|
||||
|
||||
def teardown_console(self, context, console):
|
||||
"""Tears down actual proxies"""
|
||||
"""Tears down actual proxies."""
|
||||
pass
|
||||
|
||||
def init_host(self):
|
||||
"""Start up any config'ed consoles on start"""
|
||||
"""Start up any config'ed consoles on start."""
|
||||
pass
|
||||
|
||||
def generate_password(self, length=8):
|
||||
"""Returns random console password"""
|
||||
return "fakepass"
|
||||
"""Returns random console password."""
|
||||
return 'fakepass'
|
||||
|
||||
def get_port(self, context):
|
||||
"""get available port for consoles that need one"""
|
||||
"""Get available port for consoles that need one."""
|
||||
return 5999
|
||||
|
||||
def fix_pool_password(self, password):
|
||||
"""Trim password to length, and any other massaging"""
|
||||
"""Trim password to length, and any other massaging."""
|
||||
return password
|
||||
|
||||
def fix_console_password(self, password):
|
||||
"""Trim password to length, and any other massaging"""
|
||||
"""Trim password to length, and any other massaging."""
|
||||
return password
|
||||
|
@ -15,9 +15,7 @@
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""
|
||||
Console Proxy Service
|
||||
"""
|
||||
"""Console Proxy Service."""
|
||||
|
||||
import functools
|
||||
import socket
|
||||
@ -29,6 +27,7 @@ from nova import manager
|
||||
from nova import rpc
|
||||
from nova import utils
|
||||
|
||||
|
||||
FLAGS = flags.FLAGS
|
||||
flags.DEFINE_string('console_driver',
|
||||
'nova.console.xvp.XVPConsoleProxy',
|
||||
@ -41,9 +40,11 @@ flags.DEFINE_string('console_public_hostname',
|
||||
|
||||
|
||||
class ConsoleProxyManager(manager.Manager):
|
||||
"""Sets up and tears down any console proxy connections.
|
||||
|
||||
""" Sets up and tears down any proxy connections needed for accessing
|
||||
instance consoles securely"""
|
||||
Needed for accessing instance consoles securely.
|
||||
|
||||
"""
|
||||
|
||||
def __init__(self, console_driver=None, *args, **kwargs):
|
||||
if not console_driver:
|
||||
@ -67,7 +68,7 @@ class ConsoleProxyManager(manager.Manager):
|
||||
pool['id'],
|
||||
instance_id)
|
||||
except exception.NotFound:
|
||||
logging.debug(_("Adding console"))
|
||||
logging.debug(_('Adding console'))
|
||||
if not password:
|
||||
password = utils.generate_password(8)
|
||||
if not port:
|
||||
@ -115,8 +116,8 @@ class ConsoleProxyManager(manager.Manager):
|
||||
self.db.queue_get_for(context,
|
||||
FLAGS.compute_topic,
|
||||
instance_host),
|
||||
{"method": "get_console_pool_info",
|
||||
"args": {"console_type": console_type}})
|
||||
{'method': 'get_console_pool_info',
|
||||
'args': {'console_type': console_type}})
|
||||
pool_info['password'] = self.driver.fix_pool_password(
|
||||
pool_info['password'])
|
||||
pool_info['host'] = self.host
|
||||
|
@ -15,9 +15,7 @@
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""
|
||||
VMRC console drivers.
|
||||
"""
|
||||
"""VMRC console drivers."""
|
||||
|
||||
import base64
|
||||
import json
|
||||
@ -27,6 +25,8 @@ from nova import flags
|
||||
from nova import log as logging
|
||||
from nova.virt.vmwareapi import vim_util
|
||||
|
||||
|
||||
FLAGS = flags.FLAGS
|
||||
flags.DEFINE_integer('console_vmrc_port',
|
||||
443,
|
||||
"port for VMware VMRC connections")
|
||||
@ -34,8 +34,6 @@ flags.DEFINE_integer('console_vmrc_error_retries',
|
||||
10,
|
||||
"number of retries for retrieving VMRC information")
|
||||
|
||||
FLAGS = flags.FLAGS
|
||||
|
||||
|
||||
class VMRCConsole(object):
|
||||
"""VMRC console driver with ESX credentials."""
|
||||
@ -69,34 +67,34 @@ class VMRCConsole(object):
|
||||
return password
|
||||
|
||||
def generate_password(self, vim_session, pool, instance_name):
|
||||
"""
|
||||
Returns VMRC Connection credentials.
|
||||
"""Returns VMRC Connection credentials.
|
||||
|
||||
Return string is of the form '<VM PATH>:<ESX Username>@<ESX Password>'.
|
||||
|
||||
"""
|
||||
username, password = pool['username'], pool['password']
|
||||
vms = vim_session._call_method(vim_util, "get_objects",
|
||||
"VirtualMachine", ["name", "config.files.vmPathName"])
|
||||
vms = vim_session._call_method(vim_util, 'get_objects',
|
||||
'VirtualMachine', ['name', 'config.files.vmPathName'])
|
||||
vm_ds_path_name = None
|
||||
vm_ref = None
|
||||
for vm in vms:
|
||||
vm_name = None
|
||||
ds_path_name = None
|
||||
for prop in vm.propSet:
|
||||
if prop.name == "name":
|
||||
if prop.name == 'name':
|
||||
vm_name = prop.val
|
||||
elif prop.name == "config.files.vmPathName":
|
||||
elif prop.name == 'config.files.vmPathName':
|
||||
ds_path_name = prop.val
|
||||
if vm_name == instance_name:
|
||||
vm_ref = vm.obj
|
||||
vm_ds_path_name = ds_path_name
|
||||
break
|
||||
if vm_ref is None:
|
||||
raise exception.NotFound(_("instance - %s not present") %
|
||||
raise exception.NotFound(_('instance - %s not present') %
|
||||
instance_name)
|
||||
json_data = json.dumps({"vm_id": vm_ds_path_name,
|
||||
"username": username,
|
||||
"password": password})
|
||||
json_data = json.dumps({'vm_id': vm_ds_path_name,
|
||||
'username': username,
|
||||
'password': password})
|
||||
return base64.b64encode(json_data)
|
||||
|
||||
def is_otp(self):
|
||||
@ -115,28 +113,28 @@ class VMRCSessionConsole(VMRCConsole):
|
||||
return 'vmrc+session'
|
||||
|
||||
def generate_password(self, vim_session, pool, instance_name):
|
||||
"""
|
||||
Returns a VMRC Session.
|
||||
"""Returns a VMRC Session.
|
||||
|
||||
Return string is of the form '<VM MOID>:<VMRC Ticket>'.
|
||||
|
||||
"""
|
||||
vms = vim_session._call_method(vim_util, "get_objects",
|
||||
"VirtualMachine", ["name"])
|
||||
vm_ref = None
|
||||
vms = vim_session._call_method(vim_util, 'get_objects',
|
||||
'VirtualMachine', ['name'])
|
||||
vm_ref = NoneV
|
||||
for vm in vms:
|
||||
if vm.propSet[0].val == instance_name:
|
||||
vm_ref = vm.obj
|
||||
if vm_ref is None:
|
||||
raise exception.NotFound(_("instance - %s not present") %
|
||||
raise exception.NotFound(_('instance - %s not present') %
|
||||
instance_name)
|
||||
virtual_machine_ticket = \
|
||||
vim_session._call_method(
|
||||
vim_session._get_vim(),
|
||||
"AcquireCloneTicket",
|
||||
'AcquireCloneTicket',
|
||||
vim_session._get_vim().get_service_content().sessionManager)
|
||||
json_data = json.dumps({"vm_id": str(vm_ref.value),
|
||||
"username": virtual_machine_ticket,
|
||||
"password": virtual_machine_ticket})
|
||||
json_data = json.dumps({'vm_id': str(vm_ref.value),
|
||||
'username': virtual_machine_ticket,
|
||||
'password': virtual_machine_ticket})
|
||||
return base64.b64encode(json_data)
|
||||
|
||||
def is_otp(self):
|
||||
|
@ -15,9 +15,7 @@
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""
|
||||
VMRC Console Manager.
|
||||
"""
|
||||
"""VMRC Console Manager."""
|
||||
|
||||
from nova import exception
|
||||
from nova import flags
|
||||
@ -25,24 +23,21 @@ from nova import log as logging
|
||||
from nova import manager
|
||||
from nova import rpc
|
||||
from nova import utils
|
||||
from nova.virt.vmwareapi_conn import VMWareAPISession
|
||||
from nova.virt import vmwareapi_conn
|
||||
|
||||
|
||||
LOG = logging.getLogger("nova.console.vmrc_manager")
|
||||
|
||||
|
||||
FLAGS = flags.FLAGS
|
||||
flags.DEFINE_string('console_public_hostname',
|
||||
'',
|
||||
flags.DEFINE_string('console_public_hostname', '',
|
||||
'Publicly visible name for this console host')
|
||||
flags.DEFINE_string('console_driver',
|
||||
'nova.console.vmrc.VMRCConsole',
|
||||
flags.DEFINE_string('console_driver', 'nova.console.vmrc.VMRCConsole',
|
||||
'Driver to use for the console')
|
||||
|
||||
|
||||
class ConsoleVMRCManager(manager.Manager):
|
||||
|
||||
"""
|
||||
Manager to handle VMRC connections needed for accessing instance consoles.
|
||||
"""
|
||||
"""Manager to handle VMRC connections for accessing instance consoles."""
|
||||
|
||||
def __init__(self, console_driver=None, *args, **kwargs):
|
||||
self.driver = utils.import_object(FLAGS.console_driver)
|
||||
@ -56,16 +51,17 @@ class ConsoleVMRCManager(manager.Manager):
|
||||
"""Get VIM session for the pool specified."""
|
||||
vim_session = None
|
||||
if pool['id'] not in self.sessions.keys():
|
||||
vim_session = VMWareAPISession(pool['address'],
|
||||
pool['username'],
|
||||
pool['password'],
|
||||
FLAGS.console_vmrc_error_retries)
|
||||
vim_session = vmwareapi_conn.VMWareAPISession(
|
||||
pool['address'],
|
||||
pool['username'],
|
||||
pool['password'],
|
||||
FLAGS.console_vmrc_error_retries)
|
||||
self.sessions[pool['id']] = vim_session
|
||||
return self.sessions[pool['id']]
|
||||
|
||||
def _generate_console(self, context, pool, name, instance_id, instance):
|
||||
"""Sets up console for the instance."""
|
||||
LOG.debug(_("Adding console"))
|
||||
LOG.debug(_('Adding console'))
|
||||
|
||||
password = self.driver.generate_password(
|
||||
self._get_vim_session(pool),
|
||||
@ -84,9 +80,10 @@ class ConsoleVMRCManager(manager.Manager):
|
||||
@exception.wrap_exception
|
||||
def add_console(self, context, instance_id, password=None,
|
||||
port=None, **kwargs):
|
||||
"""
|
||||
Adds a console for the instance. If it is one time password, then we
|
||||
generate new console credentials.
|
||||
"""Adds a console for the instance.
|
||||
|
||||
If it is one time password, then we generate new console credentials.
|
||||
|
||||
"""
|
||||
instance = self.db.instance_get(context, instance_id)
|
||||
host = instance['host']
|
||||
@ -97,19 +94,17 @@ class ConsoleVMRCManager(manager.Manager):
|
||||
pool['id'],
|
||||
instance_id)
|
||||
if self.driver.is_otp():
|
||||
console = self._generate_console(
|
||||
context,
|
||||
pool,
|
||||
name,
|
||||
instance_id,
|
||||
instance)
|
||||
console = self._generate_console(context,
|
||||
pool,
|
||||
name,
|
||||
instance_id,
|
||||
instance)
|
||||
except exception.NotFound:
|
||||
console = self._generate_console(
|
||||
context,
|
||||
pool,
|
||||
name,
|
||||
instance_id,
|
||||
instance)
|
||||
console = self._generate_console(context,
|
||||
pool,
|
||||
name,
|
||||
instance_id,
|
||||
instance)
|
||||
return console['id']
|
||||
|
||||
@exception.wrap_exception
|
||||
@ -118,13 +113,11 @@ class ConsoleVMRCManager(manager.Manager):
|
||||
try:
|
||||
console = self.db.console_get(context, console_id)
|
||||
except exception.NotFound:
|
||||
LOG.debug(_("Tried to remove non-existent console "
|
||||
"%(console_id)s.") %
|
||||
{'console_id': console_id})
|
||||
LOG.debug(_('Tried to remove non-existent console '
|
||||
'%(console_id)s.') % {'console_id': console_id})
|
||||
return
|
||||
LOG.debug(_("Removing console "
|
||||
"%(console_id)s.") %
|
||||
{'console_id': console_id})
|
||||
LOG.debug(_('Removing console '
|
||||
'%(console_id)s.') % {'console_id': console_id})
|
||||
self.db.console_delete(context, console_id)
|
||||
self.driver.teardown_console(context, console)
|
||||
|
||||
@ -139,11 +132,11 @@ class ConsoleVMRCManager(manager.Manager):
|
||||
console_type)
|
||||
except exception.NotFound:
|
||||
pool_info = rpc.call(context,
|
||||
self.db.queue_get_for(context,
|
||||
FLAGS.compute_topic,
|
||||
instance_host),
|
||||
{"method": "get_console_pool_info",
|
||||
"args": {"console_type": console_type}})
|
||||
self.db.queue_get_for(context,
|
||||
FLAGS.compute_topic,
|
||||
instance_host),
|
||||
{'method': 'get_console_pool_info',
|
||||
'args': {'console_type': console_type}})
|
||||
pool_info['password'] = self.driver.fix_pool_password(
|
||||
pool_info['password'])
|
||||
pool_info['host'] = self.host
|
||||
|
@ -15,16 +15,14 @@
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""
|
||||
XVP (Xenserver VNC Proxy) driver.
|
||||
"""
|
||||
"""XVP (Xenserver VNC Proxy) driver."""
|
||||
|
||||
import fcntl
|
||||
import os
|
||||
import signal
|
||||
import subprocess
|
||||
|
||||
from Cheetah.Template import Template
|
||||
from Cheetah import Template
|
||||
|
||||
from nova import context
|
||||
from nova import db
|
||||
@ -33,6 +31,8 @@ from nova import flags
|
||||
from nova import log as logging
|
||||
from nova import utils
|
||||
|
||||
|
||||
FLAGS = flags.FLAGS
|
||||
flags.DEFINE_string('console_xvp_conf_template',
|
||||
utils.abspath('console/xvp.conf.template'),
|
||||
'XVP conf template')
|
||||
@ -47,12 +47,11 @@ flags.DEFINE_string('console_xvp_log',
|
||||
'XVP log file')
|
||||
flags.DEFINE_integer('console_xvp_multiplex_port',
|
||||
5900,
|
||||
"port for XVP to multiplex VNC connections on")
|
||||
FLAGS = flags.FLAGS
|
||||
'port for XVP to multiplex VNC connections on')
|
||||
|
||||
|
||||
class XVPConsoleProxy(object):
|
||||
"""Sets up XVP config, and manages xvp daemon"""
|
||||
"""Sets up XVP config, and manages XVP daemon."""
|
||||
|
||||
def __init__(self):
|
||||
self.xvpconf_template = open(FLAGS.console_xvp_conf_template).read()
|
||||
@ -61,50 +60,51 @@ class XVPConsoleProxy(object):
|
||||
|
||||
@property
|
||||
def console_type(self):
|
||||
return "vnc+xvp"
|
||||
return 'vnc+xvp'
|
||||
|
||||
def get_port(self, context):
|
||||
"""get available port for consoles that need one"""
|
||||
"""Get available port for consoles that need one."""
|
||||
#TODO(mdragon): implement port selection for non multiplex ports,
|
||||
# we are not using that, but someone else may want
|
||||
# it.
|
||||
return FLAGS.console_xvp_multiplex_port
|
||||
|
||||
def setup_console(self, context, console):
|
||||
"""Sets up actual proxies"""
|
||||
"""Sets up actual proxies."""
|
||||
self._rebuild_xvp_conf(context.elevated())
|
||||
|
||||
def teardown_console(self, context, console):
|
||||
"""Tears down actual proxies"""
|
||||
"""Tears down actual proxies."""
|
||||
self._rebuild_xvp_conf(context.elevated())
|
||||
|
||||
def init_host(self):
|
||||
"""Start up any config'ed consoles on start"""
|
||||
"""Start up any config'ed consoles on start."""
|
||||
ctxt = context.get_admin_context()
|
||||
self._rebuild_xvp_conf(ctxt)
|
||||
|
||||
def fix_pool_password(self, password):
|
||||
"""Trim password to length, and encode"""
|
||||
"""Trim password to length, and encode."""
|
||||
return self._xvp_encrypt(password, is_pool_password=True)
|
||||
|
||||
def fix_console_password(self, password):
|
||||
"""Trim password to length, and encode"""
|
||||
"""Trim password to length, and encode."""
|
||||
return self._xvp_encrypt(password)
|
||||
|
||||
def _rebuild_xvp_conf(self, context):
|
||||
logging.debug(_("Rebuilding xvp conf"))
|
||||
logging.debug(_('Rebuilding xvp conf'))
|
||||
pools = [pool for pool in
|
||||
db.console_pool_get_all_by_host_type(context, self.host,
|
||||
self.console_type)
|
||||
if pool['consoles']]
|
||||
if not pools:
|
||||
logging.debug("No console pools!")
|
||||
logging.debug('No console pools!')
|
||||
self._xvp_stop()
|
||||
return
|
||||
conf_data = {'multiplex_port': FLAGS.console_xvp_multiplex_port,
|
||||
'pools': pools,
|
||||
'pass_encode': self.fix_console_password}
|
||||
config = str(Template(self.xvpconf_template, searchList=[conf_data]))
|
||||
config = str(Template.Template(self.xvpconf_template,
|
||||
searchList=[conf_data]))
|
||||
self._write_conf(config)
|
||||
self._xvp_restart()
|
||||
|
||||
@ -114,7 +114,7 @@ class XVPConsoleProxy(object):
|
||||
cfile.write(config)
|
||||
|
||||
def _xvp_stop(self):
|
||||
logging.debug(_("Stopping xvp"))
|
||||
logging.debug(_('Stopping xvp'))
|
||||
pid = self._xvp_pid()
|
||||
if not pid:
|
||||
return
|
||||
@ -127,19 +127,19 @@ class XVPConsoleProxy(object):
|
||||
def _xvp_start(self):
|
||||
if self._xvp_check_running():
|
||||
return
|
||||
logging.debug(_("Starting xvp"))
|
||||
logging.debug(_('Starting xvp'))
|
||||
try:
|
||||
utils.execute('xvp',
|
||||
'-p', FLAGS.console_xvp_pid,
|
||||
'-c', FLAGS.console_xvp_conf,
|
||||
'-l', FLAGS.console_xvp_log)
|
||||
except exception.ProcessExecutionError, err:
|
||||
logging.error(_("Error starting xvp: %s") % err)
|
||||
logging.error(_('Error starting xvp: %s') % err)
|
||||
|
||||
def _xvp_restart(self):
|
||||
logging.debug(_("Restarting xvp"))
|
||||
logging.debug(_('Restarting xvp'))
|
||||
if not self._xvp_check_running():
|
||||
logging.debug(_("xvp not running..."))
|
||||
logging.debug(_('xvp not running...'))
|
||||
self._xvp_start()
|
||||
else:
|
||||
pid = self._xvp_pid()
|
||||
@ -178,7 +178,9 @@ class XVPConsoleProxy(object):
|
||||
|
||||
Note that xvp's obfuscation should not be considered 'real' encryption.
|
||||
It simply DES encrypts the passwords with static keys plainly viewable
|
||||
in the xvp source code."""
|
||||
in the xvp source code.
|
||||
|
||||
"""
|
||||
maxlen = 8
|
||||
flag = '-e'
|
||||
if is_pool_password:
|
||||
|
@ -16,9 +16,7 @@
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""
|
||||
RequestContext: context for requests that persist through all of nova.
|
||||
"""
|
||||
"""RequestContext: context for requests that persist through all of nova."""
|
||||
|
||||
import datetime
|
||||
import random
|
||||
@ -28,6 +26,12 @@ from nova import utils
|
||||
|
||||
|
||||
class RequestContext(object):
|
||||
"""Security context and request information.
|
||||
|
||||
Represents the user taking a given action within the system.
|
||||
|
||||
"""
|
||||
|
||||
def __init__(self, user, project, is_admin=None, read_deleted=False,
|
||||
remote_address=None, timestamp=None, request_id=None):
|
||||
if hasattr(user, 'id'):
|
||||
|
@ -15,10 +15,11 @@
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
"""
|
||||
Wrappers around standard crypto data elements.
|
||||
|
||||
"""Wrappers around standard crypto data elements.
|
||||
|
||||
Includes root and intermediate CAs, SSH key_pairs and x509 certificates.
|
||||
|
||||
"""
|
||||
|
||||
import base64
|
||||
@ -43,6 +44,8 @@ from nova import log as logging
|
||||
|
||||
|
||||
LOG = logging.getLogger("nova.crypto")
|
||||
|
||||
|
||||
FLAGS = flags.FLAGS
|
||||
flags.DEFINE_string('ca_file', 'cacert.pem', _('Filename of root CA'))
|
||||
flags.DEFINE_string('key_file',
|
||||
@ -90,13 +93,13 @@ def key_path(project_id=None):
|
||||
def fetch_ca(project_id=None, chain=True):
|
||||
if not FLAGS.use_project_ca:
|
||||
project_id = None
|
||||
buffer = ""
|
||||
buffer = ''
|
||||
if project_id:
|
||||
with open(ca_path(project_id), "r") as cafile:
|
||||
with open(ca_path(project_id), 'r') as cafile:
|
||||
buffer += cafile.read()
|
||||
if not chain:
|
||||
return buffer
|
||||
with open(ca_path(None), "r") as cafile:
|
||||
with open(ca_path(None), 'r') as cafile:
|
||||
buffer += cafile.read()
|
||||
return buffer
|
||||
|
||||
@ -143,7 +146,7 @@ def ssl_pub_to_ssh_pub(ssl_public_key, name='root', suffix='nova'):
|
||||
|
||||
|
||||
def revoke_cert(project_id, file_name):
|
||||
"""Revoke a cert by file name"""
|
||||
"""Revoke a cert by file name."""
|
||||
start = os.getcwd()
|
||||
os.chdir(ca_folder(project_id))
|
||||
# NOTE(vish): potential race condition here
|
||||
@ -155,14 +158,14 @@ def revoke_cert(project_id, file_name):
|
||||
|
||||
|
||||
def revoke_certs_by_user(user_id):
|
||||
"""Revoke all user certs"""
|
||||
"""Revoke all user certs."""
|
||||
admin = context.get_admin_context()
|
||||
for cert in db.certificate_get_all_by_user(admin, user_id):
|
||||
revoke_cert(cert['project_id'], cert['file_name'])
|
||||
|
||||
|
||||
def revoke_certs_by_project(project_id):
|
||||
"""Revoke all project certs"""
|
||||
"""Revoke all project certs."""
|
||||
# NOTE(vish): This is somewhat useless because we can just shut down
|
||||
# the vpn.
|
||||
admin = context.get_admin_context()
|
||||
@ -171,29 +174,29 @@ def revoke_certs_by_project(project_id):
|
||||
|
||||
|
||||
def revoke_certs_by_user_and_project(user_id, project_id):
|
||||
"""Revoke certs for user in project"""
|
||||
"""Revoke certs for user in project."""
|
||||
admin = context.get_admin_context()
|
||||
for cert in db.certificate_get_all_by_user(admin, user_id, project_id):
|
||||
revoke_cert(cert['project_id'], cert['file_name'])
|
||||
|
||||
|
||||
def _project_cert_subject(project_id):
|
||||
"""Helper to generate user cert subject"""
|
||||
"""Helper to generate user cert subject."""
|
||||
return FLAGS.project_cert_subject % (project_id, utils.isotime())
|
||||
|
||||
|
||||
def _vpn_cert_subject(project_id):
|
||||
"""Helper to generate user cert subject"""
|
||||
"""Helper to generate user cert subject."""
|
||||
return FLAGS.vpn_cert_subject % (project_id, utils.isotime())
|
||||
|
||||
|
||||
def _user_cert_subject(user_id, project_id):
|
||||
"""Helper to generate user cert subject"""
|
||||
"""Helper to generate user cert subject."""
|
||||
return FLAGS.user_cert_subject % (project_id, user_id, utils.isotime())
|
||||
|
||||
|
||||
def generate_x509_cert(user_id, project_id, bits=1024):
|
||||
"""Generate and sign a cert for user in project"""
|
||||
"""Generate and sign a cert for user in project."""
|
||||
subject = _user_cert_subject(user_id, project_id)
|
||||
tmpdir = tempfile.mkdtemp()
|
||||
keyfile = os.path.abspath(os.path.join(tmpdir, 'temp.key'))
|
||||
@ -205,7 +208,7 @@ def generate_x509_cert(user_id, project_id, bits=1024):
|
||||
csr = open(csrfile).read()
|
||||
shutil.rmtree(tmpdir)
|
||||
(serial, signed_csr) = sign_csr(csr, project_id)
|
||||
fname = os.path.join(ca_folder(project_id), "newcerts/%s.pem" % serial)
|
||||
fname = os.path.join(ca_folder(project_id), 'newcerts/%s.pem' % serial)
|
||||
cert = {'user_id': user_id,
|
||||
'project_id': project_id,
|
||||
'file_name': fname}
|
||||
@ -227,8 +230,8 @@ def _ensure_project_folder(project_id):
|
||||
|
||||
def generate_vpn_files(project_id):
|
||||
project_folder = ca_folder(project_id)
|
||||
csr_fn = os.path.join(project_folder, "server.csr")
|
||||
crt_fn = os.path.join(project_folder, "server.crt")
|
||||
csr_fn = os.path.join(project_folder, 'server.csr')
|
||||
crt_fn = os.path.join(project_folder, 'server.crt')
|
||||
|
||||
genvpn_sh_path = os.path.join(os.path.dirname(__file__),
|
||||
'CA',
|
||||
@ -241,10 +244,10 @@ def generate_vpn_files(project_id):
|
||||
# TODO(vish): the shell scripts could all be done in python
|
||||
utils.execute('sh', genvpn_sh_path,
|
||||
project_id, _vpn_cert_subject(project_id))
|
||||
with open(csr_fn, "r") as csrfile:
|
||||
with open(csr_fn, 'r') as csrfile:
|
||||
csr_text = csrfile.read()
|
||||
(serial, signed_csr) = sign_csr(csr_text, project_id)
|
||||
with open(crt_fn, "w") as crtfile:
|
||||
with open(crt_fn, 'w') as crtfile:
|
||||
crtfile.write(signed_csr)
|
||||
os.chdir(start)
|
||||
|
||||
@ -261,12 +264,12 @@ def sign_csr(csr_text, project_id=None):
|
||||
|
||||
def _sign_csr(csr_text, ca_folder):
|
||||
tmpfolder = tempfile.mkdtemp()
|
||||
inbound = os.path.join(tmpfolder, "inbound.csr")
|
||||
outbound = os.path.join(tmpfolder, "outbound.csr")
|
||||
csrfile = open(inbound, "w")
|
||||
inbound = os.path.join(tmpfolder, 'inbound.csr')
|
||||
outbound = os.path.join(tmpfolder, 'outbound.csr')
|
||||
csrfile = open(inbound, 'w')
|
||||
csrfile.write(csr_text)
|
||||
csrfile.close()
|
||||
LOG.debug(_("Flags path: %s"), ca_folder)
|
||||
LOG.debug(_('Flags path: %s'), ca_folder)
|
||||
start = os.getcwd()
|
||||
# Change working dir to CA
|
||||
if not os.path.exists(ca_folder):
|
||||
@ -276,13 +279,13 @@ def _sign_csr(csr_text, ca_folder):
|
||||
'./openssl.cnf', '-infiles', inbound)
|
||||
out, _err = utils.execute('openssl', 'x509', '-in', outbound,
|
||||
'-serial', '-noout')
|
||||
serial = string.strip(out.rpartition("=")[2])
|
||||
serial = string.strip(out.rpartition('=')[2])
|
||||
os.chdir(start)
|
||||
with open(outbound, "r") as crtfile:
|
||||
with open(outbound, 'r') as crtfile:
|
||||
return (serial, crtfile.read())
|
||||
|
||||
|
||||
def mkreq(bits, subject="foo", ca=0):
|
||||
def mkreq(bits, subject='foo', ca=0):
|
||||
pk = M2Crypto.EVP.PKey()
|
||||
req = M2Crypto.X509.Request()
|
||||
rsa = M2Crypto.RSA.gen_key(bits, 65537, callback=lambda: None)
|
||||
@ -314,7 +317,7 @@ def mkcacert(subject='nova', years=1):
|
||||
cert.set_not_before(now)
|
||||
cert.set_not_after(nowPlusYear)
|
||||
issuer = M2Crypto.X509.X509_Name()
|
||||
issuer.C = "US"
|
||||
issuer.C = 'US'
|
||||
issuer.CN = subject
|
||||
cert.set_issuer(issuer)
|
||||
cert.set_pubkey(pkey)
|
||||
@ -352,13 +355,15 @@ def mkcacert(subject='nova', years=1):
|
||||
# http://code.google.com/p/boto
|
||||
|
||||
def compute_md5(fp):
|
||||
"""
|
||||
"""Compute an md5 hash.
|
||||
|
||||
:type fp: file
|
||||
:param fp: File pointer to the file to MD5 hash. The file pointer will be
|
||||
reset to the beginning of the file before the method returns.
|
||||
|
||||
:rtype: tuple
|
||||
:return: the hex digest version of the MD5 hash
|
||||
:returns: the hex digest version of the MD5 hash
|
||||
|
||||
"""
|
||||
m = hashlib.md5()
|
||||
fp.seek(0)
|
||||
|
@ -15,8 +15,8 @@
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
"""
|
||||
Defines interface for DB access.
|
||||
|
||||
"""Defines interface for DB access.
|
||||
|
||||
The underlying driver is loaded as a :class:`LazyPluggable`.
|
||||
|
||||
@ -30,6 +30,7 @@ The underlying driver is loaded as a :class:`LazyPluggable`.
|
||||
|
||||
:enable_new_services: when adding a new service to the database, is it in the
|
||||
pool of available hardware (Default: True)
|
||||
|
||||
"""
|
||||
|
||||
from nova import exception
|
||||
@ -86,7 +87,7 @@ def service_get(context, service_id):
|
||||
|
||||
|
||||
def service_get_by_host_and_topic(context, host, topic):
|
||||
"""Get a service by host it's on and topic it listens to"""
|
||||
"""Get a service by host it's on and topic it listens to."""
|
||||
return IMPL.service_get_by_host_and_topic(context, host, topic)
|
||||
|
||||
|
||||
@ -113,7 +114,7 @@ def service_get_all_compute_by_host(context, host):
|
||||
def service_get_all_compute_sorted(context):
|
||||
"""Get all compute services sorted by instance count.
|
||||
|
||||
Returns a list of (Service, instance_count) tuples.
|
||||
:returns: a list of (Service, instance_count) tuples.
|
||||
|
||||
"""
|
||||
return IMPL.service_get_all_compute_sorted(context)
|
||||
@ -122,7 +123,7 @@ def service_get_all_compute_sorted(context):
|
||||
def service_get_all_network_sorted(context):
|
||||
"""Get all network services sorted by network count.
|
||||
|
||||
Returns a list of (Service, network_count) tuples.
|
||||
:returns: a list of (Service, network_count) tuples.
|
||||
|
||||
"""
|
||||
return IMPL.service_get_all_network_sorted(context)
|
||||
@ -131,7 +132,7 @@ def service_get_all_network_sorted(context):
|
||||
def service_get_all_volume_sorted(context):
|
||||
"""Get all volume services sorted by volume count.
|
||||
|
||||
Returns a list of (Service, volume_count) tuples.
|
||||
:returns: a list of (Service, volume_count) tuples.
|
||||
|
||||
"""
|
||||
return IMPL.service_get_all_volume_sorted(context)
|
||||
@ -241,7 +242,7 @@ def floating_ip_count_by_project(context, project_id):
|
||||
|
||||
|
||||
def floating_ip_deallocate(context, address):
|
||||
"""Deallocate an floating ip by address"""
|
||||
"""Deallocate an floating ip by address."""
|
||||
return IMPL.floating_ip_deallocate(context, address)
|
||||
|
||||
|
||||
@ -253,7 +254,7 @@ def floating_ip_destroy(context, address):
|
||||
def floating_ip_disassociate(context, address):
|
||||
"""Disassociate an floating ip from a fixed ip by address.
|
||||
|
||||
Returns the address of the existing fixed ip.
|
||||
:returns: the address of the existing fixed ip.
|
||||
|
||||
"""
|
||||
return IMPL.floating_ip_disassociate(context, address)
|
||||
@ -294,22 +295,22 @@ def floating_ip_update(context, address, values):
|
||||
####################
|
||||
|
||||
def migration_update(context, id, values):
|
||||
"""Update a migration instance"""
|
||||
"""Update a migration instance."""
|
||||
return IMPL.migration_update(context, id, values)
|
||||
|
||||
|
||||
def migration_create(context, values):
|
||||
"""Create a migration record"""
|
||||
"""Create a migration record."""
|
||||
return IMPL.migration_create(context, values)
|
||||
|
||||
|
||||
def migration_get(context, migration_id):
|
||||
"""Finds a migration by the id"""
|
||||
"""Finds a migration by the id."""
|
||||
return IMPL.migration_get(context, migration_id)
|
||||
|
||||
|
||||
def migration_get_by_instance_and_status(context, instance_id, status):
|
||||
"""Finds a migration by the instance id its migrating"""
|
||||
"""Finds a migration by the instance id its migrating."""
|
||||
return IMPL.migration_get_by_instance_and_status(context, instance_id,
|
||||
status)
|
||||
|
||||
@ -579,7 +580,9 @@ def network_create_safe(context, values):
|
||||
|
||||
def network_delete_safe(context, network_id):
|
||||
"""Delete network with key network_id.
|
||||
|
||||
This method assumes that the network is not associated with any project
|
||||
|
||||
"""
|
||||
return IMPL.network_delete_safe(context, network_id)
|
||||
|
||||
@ -674,7 +677,6 @@ def project_get_network(context, project_id, associate=True):
|
||||
network if one is not found, otherwise it returns None.
|
||||
|
||||
"""
|
||||
|
||||
return IMPL.project_get_network(context, project_id, associate)
|
||||
|
||||
|
||||
@ -722,7 +724,9 @@ def iscsi_target_create_safe(context, values):
|
||||
|
||||
The device is not returned. If the create violates the unique
|
||||
constraints because the iscsi_target and host already exist,
|
||||
no exception is raised."""
|
||||
no exception is raised.
|
||||
|
||||
"""
|
||||
return IMPL.iscsi_target_create_safe(context, values)
|
||||
|
||||
|
||||
@ -1050,10 +1054,7 @@ def project_delete(context, project_id):
|
||||
|
||||
|
||||
def host_get_networks(context, host):
|
||||
"""Return all networks for which the given host is the designated
|
||||
network host.
|
||||
|
||||
"""
|
||||
"""All networks for which the given host is the network host."""
|
||||
return IMPL.host_get_networks(context, host)
|
||||
|
||||
|
||||
@ -1115,38 +1116,40 @@ def console_get(context, console_id, instance_id=None):
|
||||
|
||||
|
||||
def instance_type_create(context, values):
|
||||
"""Create a new instance type"""
|
||||
"""Create a new instance type."""
|
||||
return IMPL.instance_type_create(context, values)
|
||||
|
||||
|
||||
def instance_type_get_all(context, inactive=False):
|
||||
"""Get all instance types"""
|
||||
"""Get all instance types."""
|
||||
return IMPL.instance_type_get_all(context, inactive)
|
||||
|
||||
|
||||
def instance_type_get_by_id(context, id):
|
||||
"""Get instance type by id"""
|
||||
"""Get instance type by id."""
|
||||
return IMPL.instance_type_get_by_id(context, id)
|
||||
|
||||
|
||||
def instance_type_get_by_name(context, name):
|
||||
"""Get instance type by name"""
|
||||
"""Get instance type by name."""
|
||||
return IMPL.instance_type_get_by_name(context, name)
|
||||
|
||||
|
||||
def instance_type_get_by_flavor_id(context, id):
|
||||
"""Get instance type by name"""
|
||||
"""Get instance type by name."""
|
||||
return IMPL.instance_type_get_by_flavor_id(context, id)
|
||||
|
||||
|
||||
def instance_type_destroy(context, name):
|
||||
"""Delete a instance type"""
|
||||
"""Delete a instance type."""
|
||||
return IMPL.instance_type_destroy(context, name)
|
||||
|
||||
|
||||
def instance_type_purge(context, name):
|
||||
"""Purges (removes) an instance type from DB
|
||||
Use instance_type_destroy for most cases
|
||||
"""Purges (removes) an instance type from DB.
|
||||
|
||||
Use instance_type_destroy for most cases
|
||||
|
||||
"""
|
||||
return IMPL.instance_type_purge(context, name)
|
||||
|
||||
@ -1183,15 +1186,15 @@ def zone_get_all(context):
|
||||
|
||||
|
||||
def instance_metadata_get(context, instance_id):
|
||||
"""Get all metadata for an instance"""
|
||||
"""Get all metadata for an instance."""
|
||||
return IMPL.instance_metadata_get(context, instance_id)
|
||||
|
||||
|
||||
def instance_metadata_delete(context, instance_id, key):
|
||||
"""Delete the given metadata item"""
|
||||
"""Delete the given metadata item."""
|
||||
IMPL.instance_metadata_delete(context, instance_id, key)
|
||||
|
||||
|
||||
def instance_metadata_update_or_create(context, instance_id, metadata):
|
||||
"""Create or update instance metadata"""
|
||||
"""Create or update instance metadata."""
|
||||
IMPL.instance_metadata_update_or_create(context, instance_id, metadata)
|
||||
|
@ -16,20 +16,20 @@
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""
|
||||
Base class for classes that need modular database access.
|
||||
"""
|
||||
"""Base class for classes that need modular database access."""
|
||||
|
||||
from nova import utils
|
||||
from nova import flags
|
||||
|
||||
|
||||
FLAGS = flags.FLAGS
|
||||
flags.DEFINE_string('db_driver', 'nova.db.api',
|
||||
'driver to use for database access')
|
||||
|
||||
|
||||
class Base(object):
|
||||
"""DB driver is injected in the init method"""
|
||||
"""DB driver is injected in the init method."""
|
||||
|
||||
def __init__(self, db_driver=None):
|
||||
if not db_driver:
|
||||
db_driver = FLAGS.db_driver
|
||||
|
@ -15,11 +15,13 @@
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""Database setup and migration commands."""
|
||||
|
||||
from nova import flags
|
||||
from nova import utils
|
||||
|
||||
|
||||
FLAGS = flags.FLAGS
|
||||
flags.DECLARE('db_backend', 'nova.db.api')
|
||||
|
||||
|
@ -54,10 +54,12 @@ def upgrade(migrate_engine):
|
||||
|
||||
instances.create_column(c_instance_type_id)
|
||||
|
||||
type_names = {}
|
||||
recs = migrate_engine.execute(instance_types.select())
|
||||
for row in recs:
|
||||
type_id = row[0]
|
||||
type_name = row[1]
|
||||
type_names[row[0]] = row[1]
|
||||
|
||||
for type_id, type_name in type_names.iteritems():
|
||||
migrate_engine.execute(instances.update()\
|
||||
.where(instances.c.instance_type == type_name)\
|
||||
.values(instance_type_id=type_id))
|
||||
|
@ -16,31 +16,34 @@
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""
|
||||
Nova base exception handling, including decorator for re-raising
|
||||
Nova-type exceptions. SHOULD include dedicated exception logging.
|
||||
"""Nova base exception handling.
|
||||
|
||||
Includes decorator for re-raising Nova-type exceptions.
|
||||
|
||||
SHOULD include dedicated exception logging.
|
||||
|
||||
"""
|
||||
|
||||
from nova import log as logging
|
||||
|
||||
|
||||
LOG = logging.getLogger('nova.exception')
|
||||
|
||||
|
||||
class ProcessExecutionError(IOError):
|
||||
|
||||
def __init__(self, stdout=None, stderr=None, exit_code=None, cmd=None,
|
||||
description=None):
|
||||
if description is None:
|
||||
description = _("Unexpected error while running command.")
|
||||
description = _('Unexpected error while running command.')
|
||||
if exit_code is None:
|
||||
exit_code = '-'
|
||||
message = _("%(description)s\nCommand: %(cmd)s\n"
|
||||
"Exit code: %(exit_code)s\nStdout: %(stdout)r\n"
|
||||
"Stderr: %(stderr)r") % locals()
|
||||
message = _('%(description)s\nCommand: %(cmd)s\n'
|
||||
'Exit code: %(exit_code)s\nStdout: %(stdout)r\n'
|
||||
'Stderr: %(stderr)r') % locals()
|
||||
IOError.__init__(self, message)
|
||||
|
||||
|
||||
class Error(Exception):
|
||||
|
||||
def __init__(self, message=None):
|
||||
super(Error, self).__init__(message)
|
||||
|
||||
@ -97,7 +100,7 @@ class TimeoutException(Error):
|
||||
|
||||
|
||||
class DBError(Error):
|
||||
"""Wraps an implementation specific exception"""
|
||||
"""Wraps an implementation specific exception."""
|
||||
def __init__(self, inner_exception):
|
||||
self.inner_exception = inner_exception
|
||||
super(DBError, self).__init__(str(inner_exception))
|
||||
@ -108,7 +111,7 @@ def wrap_db_error(f):
|
||||
try:
|
||||
return f(*args, **kwargs)
|
||||
except Exception, e:
|
||||
LOG.exception(_('DB exception wrapped'))
|
||||
LOG.exception(_('DB exception wrapped.'))
|
||||
raise DBError(e)
|
||||
return _wrap
|
||||
_wrap.func_name = f.func_name
|
||||
|
@ -18,14 +18,14 @@
|
||||
|
||||
"""Super simple fake memcache client."""
|
||||
|
||||
import utils
|
||||
from nova import utils
|
||||
|
||||
|
||||
class Client(object):
|
||||
"""Replicates a tiny subset of memcached client interface."""
|
||||
|
||||
def __init__(self, *args, **kwargs):
|
||||
"""Ignores the passed in args"""
|
||||
"""Ignores the passed in args."""
|
||||
self.cache = {}
|
||||
|
||||
def get(self, key):
|
||||
|
@ -16,9 +16,13 @@
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""
|
||||
"""Command-line flag library.
|
||||
|
||||
Wraps gflags.
|
||||
|
||||
Package-level global flags are defined here, the rest are defined
|
||||
where they're used.
|
||||
|
||||
"""
|
||||
|
||||
import getopt
|
||||
@ -145,10 +149,12 @@ class FlagValues(gflags.FlagValues):
|
||||
|
||||
|
||||
class StrWrapper(object):
|
||||
"""Wrapper around FlagValues objects
|
||||
"""Wrapper around FlagValues objects.
|
||||
|
||||
Wraps FlagValues objects for string.Template so that we're
|
||||
sure to return strings."""
|
||||
sure to return strings.
|
||||
|
||||
"""
|
||||
def __init__(self, context_objs):
|
||||
self.context_objs = context_objs
|
||||
|
||||
@ -169,6 +175,7 @@ def _GetCallingModule():
|
||||
|
||||
We generally use this function to get the name of the module calling a
|
||||
DEFINE_foo... function.
|
||||
|
||||
"""
|
||||
# Walk down the stack to find the first globals dict that's not ours.
|
||||
for depth in range(1, sys.getrecursionlimit()):
|
||||
@ -192,6 +199,7 @@ def __GetModuleName(globals_dict):
|
||||
Returns:
|
||||
A string (the name of the module) or None (if the module could not
|
||||
be identified.
|
||||
|
||||
"""
|
||||
for name, module in sys.modules.iteritems():
|
||||
if getattr(module, '__dict__', None) is globals_dict:
|
||||
@ -326,7 +334,7 @@ DEFINE_integer('auth_token_ttl', 3600, 'Seconds for auth tokens to linger')
|
||||
DEFINE_string('state_path', os.path.join(os.path.dirname(__file__), '../'),
|
||||
"Top-level directory for maintaining nova's state")
|
||||
DEFINE_string('lock_path', os.path.join(os.path.dirname(__file__), '../'),
|
||||
"Directory for lock files")
|
||||
'Directory for lock files')
|
||||
DEFINE_string('logdir', None, 'output to a per-service log file in named '
|
||||
'directory')
|
||||
|
||||
|
@ -14,6 +14,7 @@
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""Implementation of an fake image service"""
|
||||
|
||||
import copy
|
||||
@ -69,14 +70,14 @@ class FakeImageService(service.BaseImageService):
|
||||
image = self.images.get(image_id)
|
||||
if image:
|
||||
return copy.deepcopy(image)
|
||||
LOG.warn("Unable to find image id %s. Have images: %s",
|
||||
LOG.warn('Unable to find image id %s. Have images: %s',
|
||||
image_id, self.images)
|
||||
raise exception.NotFound
|
||||
|
||||
def create(self, context, data):
|
||||
"""Store the image data and return the new image id.
|
||||
|
||||
:raises Duplicate if the image already exist.
|
||||
:raises: Duplicate if the image already exist.
|
||||
|
||||
"""
|
||||
image_id = int(data['id'])
|
||||
@ -88,7 +89,7 @@ class FakeImageService(service.BaseImageService):
|
||||
def update(self, context, image_id, data):
|
||||
"""Replace the contents of the given image with the new data.
|
||||
|
||||
:raises NotFound if the image does not exist.
|
||||
:raises: NotFound if the image does not exist.
|
||||
|
||||
"""
|
||||
image_id = int(image_id)
|
||||
@ -99,7 +100,7 @@ class FakeImageService(service.BaseImageService):
|
||||
def delete(self, context, image_id):
|
||||
"""Delete the given image.
|
||||
|
||||
:raises NotFound if the image does not exist.
|
||||
:raises: NotFound if the image does not exist.
|
||||
|
||||
"""
|
||||
image_id = int(image_id)
|
||||
|
@ -14,6 +14,7 @@
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""Implementation of an image service that uses Glance as the backend"""
|
||||
|
||||
from __future__ import absolute_import
|
||||
@ -31,16 +32,18 @@ from nova.image import service
|
||||
|
||||
LOG = logging.getLogger('nova.image.glance')
|
||||
|
||||
|
||||
FLAGS = flags.FLAGS
|
||||
|
||||
|
||||
GlanceClient = utils.import_class('glance.client.Client')
|
||||
|
||||
|
||||
class GlanceImageService(service.BaseImageService):
|
||||
"""Provides storage and retrieval of disk image objects within Glance."""
|
||||
|
||||
GLANCE_ONLY_ATTRS = ["size", "location", "disk_format",
|
||||
"container_format"]
|
||||
GLANCE_ONLY_ATTRS = ['size', 'location', 'disk_format',
|
||||
'container_format']
|
||||
|
||||
# NOTE(sirp): Overriding to use _translate_to_service provided by
|
||||
# BaseImageService
|
||||
@ -56,9 +59,7 @@ class GlanceImageService(service.BaseImageService):
|
||||
self.client = client
|
||||
|
||||
def index(self, context):
|
||||
"""
|
||||
Calls out to Glance for a list of images available
|
||||
"""
|
||||
"""Calls out to Glance for a list of images available."""
|
||||
# NOTE(sirp): We need to use `get_images_detailed` and not
|
||||
# `get_images` here because we need `is_public` and `properties`
|
||||
# included so we can filter by user
|
||||
@ -71,9 +72,7 @@ class GlanceImageService(service.BaseImageService):
|
||||
return filtered
|
||||
|
||||
def detail(self, context):
|
||||
"""
|
||||
Calls out to Glance for a list of detailed image information
|
||||
"""
|
||||
"""Calls out to Glance for a list of detailed image information."""
|
||||
filtered = []
|
||||
image_metas = self.client.get_images_detailed()
|
||||
for image_meta in image_metas:
|
||||
@ -83,9 +82,7 @@ class GlanceImageService(service.BaseImageService):
|
||||
return filtered
|
||||
|
||||
def show(self, context, image_id):
|
||||
"""
|
||||
Returns a dict containing image data for the given opaque image id.
|
||||
"""
|
||||
"""Returns a dict with image data for the given opaque image id."""
|
||||
try:
|
||||
image_meta = self.client.get_image_meta(image_id)
|
||||
except glance_exception.NotFound:
|
||||
@ -98,9 +95,7 @@ class GlanceImageService(service.BaseImageService):
|
||||
return base_image_meta
|
||||
|
||||
def show_by_name(self, context, name):
|
||||
"""
|
||||
Returns a dict containing image data for the given name.
|
||||
"""
|
||||
"""Returns a dict containing image data for the given name."""
|
||||
# TODO(vish): replace this with more efficient call when glance
|
||||
# supports it.
|
||||
image_metas = self.detail(context)
|
||||
@ -110,9 +105,7 @@ class GlanceImageService(service.BaseImageService):
|
||||
raise exception.NotFound
|
||||
|
||||
def get(self, context, image_id, data):
|
||||
"""
|
||||
Calls out to Glance for metadata and data and writes data.
|
||||
"""
|
||||
"""Calls out to Glance for metadata and data and writes data."""
|
||||
try:
|
||||
image_meta, image_chunks = self.client.get_image(image_id)
|
||||
except glance_exception.NotFound:
|
||||
@ -125,16 +118,16 @@ class GlanceImageService(service.BaseImageService):
|
||||
return base_image_meta
|
||||
|
||||
def create(self, context, image_meta, data=None):
|
||||
"""
|
||||
Store the image data and return the new image id.
|
||||
"""Store the image data and return the new image id.
|
||||
|
||||
:raises: AlreadyExists if the image already exist.
|
||||
|
||||
:raises AlreadyExists if the image already exist.
|
||||
"""
|
||||
# Translate Base -> Service
|
||||
LOG.debug(_("Creating image in Glance. Metadata passed in %s"),
|
||||
LOG.debug(_('Creating image in Glance. Metadata passed in %s'),
|
||||
image_meta)
|
||||
sent_service_image_meta = self._translate_to_service(image_meta)
|
||||
LOG.debug(_("Metadata after formatting for Glance %s"),
|
||||
LOG.debug(_('Metadata after formatting for Glance %s'),
|
||||
sent_service_image_meta)
|
||||
|
||||
recv_service_image_meta = self.client.add_image(
|
||||
@ -142,14 +135,15 @@ class GlanceImageService(service.BaseImageService):
|
||||
|
||||
# Translate Service -> Base
|
||||
base_image_meta = self._translate_to_base(recv_service_image_meta)
|
||||
LOG.debug(_("Metadata returned from Glance formatted for Base %s"),
|
||||
LOG.debug(_('Metadata returned from Glance formatted for Base %s'),
|
||||
base_image_meta)
|
||||
return base_image_meta
|
||||
|
||||
def update(self, context, image_id, image_meta, data=None):
|
||||
"""Replace the contents of the given image with the new data.
|
||||
|
||||
:raises NotFound if the image does not exist.
|
||||
:raises: NotFound if the image does not exist.
|
||||
|
||||
"""
|
||||
# NOTE(vish): show is to check if image is available
|
||||
self.show(context, image_id)
|
||||
@ -162,10 +156,10 @@ class GlanceImageService(service.BaseImageService):
|
||||
return base_image_meta
|
||||
|
||||
def delete(self, context, image_id):
|
||||
"""
|
||||
Delete the given image.
|
||||
"""Delete the given image.
|
||||
|
||||
:raises: NotFound if the image does not exist.
|
||||
|
||||
:raises NotFound if the image does not exist.
|
||||
"""
|
||||
# NOTE(vish): show is to check if image is available
|
||||
self.show(context, image_id)
|
||||
@ -176,16 +170,12 @@ class GlanceImageService(service.BaseImageService):
|
||||
return result
|
||||
|
||||
def delete_all(self):
|
||||
"""
|
||||
Clears out all images
|
||||
"""
|
||||
"""Clears out all images."""
|
||||
pass
|
||||
|
||||
@classmethod
|
||||
def _translate_to_base(cls, image_meta):
|
||||
"""Overriding the base translation to handle conversion to datetime
|
||||
objects
|
||||
"""
|
||||
"""Override translation to handle conversion to datetime objects."""
|
||||
image_meta = service.BaseImageService._propertify_metadata(
|
||||
image_meta, cls.SERVICE_IMAGE_ATTRS)
|
||||
image_meta = _convert_timestamps_to_datetimes(image_meta)
|
||||
@ -194,9 +184,7 @@ class GlanceImageService(service.BaseImageService):
|
||||
|
||||
# utility functions
|
||||
def _convert_timestamps_to_datetimes(image_meta):
|
||||
"""
|
||||
Returns image with known timestamp fields converted to datetime objects
|
||||
"""
|
||||
"""Returns image with timestamp fields converted to datetime objects."""
|
||||
for attr in ['created_at', 'updated_at', 'deleted_at']:
|
||||
if image_meta.get(attr):
|
||||
image_meta[attr] = _parse_glance_iso8601_timestamp(
|
||||
@ -205,10 +193,8 @@ def _convert_timestamps_to_datetimes(image_meta):
|
||||
|
||||
|
||||
def _parse_glance_iso8601_timestamp(timestamp):
|
||||
"""
|
||||
Parse a subset of iso8601 timestamps into datetime objects
|
||||
"""
|
||||
iso_formats = ["%Y-%m-%dT%H:%M:%S.%f", "%Y-%m-%dT%H:%M:%S"]
|
||||
"""Parse a subset of iso8601 timestamps into datetime objects."""
|
||||
iso_formats = ['%Y-%m-%dT%H:%M:%S.%f', '%Y-%m-%dT%H:%M:%S']
|
||||
|
||||
for iso_format in iso_formats:
|
||||
try:
|
||||
@ -216,5 +202,5 @@ def _parse_glance_iso8601_timestamp(timestamp):
|
||||
except ValueError:
|
||||
pass
|
||||
|
||||
raise ValueError(_("%(timestamp)s does not follow any of the "
|
||||
"signatures: %(ISO_FORMATS)s") % locals())
|
||||
raise ValueError(_('%(timestamp)s does not follow any of the '
|
||||
'signatures: %(ISO_FORMATS)s') % locals())
|
||||
|
@ -23,14 +23,15 @@ import shutil
|
||||
from nova import exception
|
||||
from nova import flags
|
||||
from nova import log as logging
|
||||
from nova.image import service
|
||||
from nova import utils
|
||||
from nova.image import service
|
||||
|
||||
|
||||
FLAGS = flags.FLAGS
|
||||
flags.DEFINE_string('images_path', '$state_path/images',
|
||||
'path to decrypted images')
|
||||
|
||||
|
||||
LOG = logging.getLogger('nova.image.local')
|
||||
|
||||
|
||||
@ -56,9 +57,8 @@ class LocalImageService(service.BaseImageService):
|
||||
try:
|
||||
unhexed_image_id = int(image_dir, 16)
|
||||
except ValueError:
|
||||
LOG.error(
|
||||
_("%s is not in correct directory naming format"\
|
||||
% image_dir))
|
||||
LOG.error(_('%s is not in correct directory naming format')
|
||||
% image_dir)
|
||||
else:
|
||||
images.append(unhexed_image_id)
|
||||
return images
|
||||
@ -148,7 +148,8 @@ class LocalImageService(service.BaseImageService):
|
||||
|
||||
def delete(self, context, image_id):
|
||||
"""Delete the given image.
|
||||
Raises NotFound if the image does not exist.
|
||||
|
||||
:raises: NotFound if the image does not exist.
|
||||
|
||||
"""
|
||||
# NOTE(vish): show is to check if image is available
|
||||
|
@ -16,13 +16,9 @@
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""
|
||||
Proxy AMI-related calls from the cloud controller, to the running
|
||||
objectstore service.
|
||||
"""
|
||||
"""Proxy AMI-related calls from cloud controller to objectstore service."""
|
||||
|
||||
import binascii
|
||||
import eventlet
|
||||
import os
|
||||
import shutil
|
||||
import tarfile
|
||||
@ -30,6 +26,7 @@ import tempfile
|
||||
from xml.etree import ElementTree
|
||||
|
||||
import boto.s3.connection
|
||||
import eventlet
|
||||
|
||||
from nova import crypto
|
||||
from nova import exception
|
||||
@ -46,7 +43,8 @@ flags.DEFINE_string('image_decryption_dir', '/tmp',
|
||||
|
||||
|
||||
class S3ImageService(service.BaseImageService):
|
||||
"""Wraps an existing image service to support s3 based register"""
|
||||
"""Wraps an existing image service to support s3 based register."""
|
||||
|
||||
def __init__(self, service=None, *args, **kwargs):
|
||||
if service is None:
|
||||
service = utils.import_object(FLAGS.image_service)
|
||||
@ -54,7 +52,11 @@ class S3ImageService(service.BaseImageService):
|
||||
self.service.__init__(*args, **kwargs)
|
||||
|
||||
def create(self, context, metadata, data=None):
|
||||
"""metadata['properties'] should contain image_location"""
|
||||
"""Create an image.
|
||||
|
||||
metadata['properties'] should contain image_location.
|
||||
|
||||
"""
|
||||
image = self._s3_create(context, metadata)
|
||||
return image
|
||||
|
||||
@ -100,12 +102,12 @@ class S3ImageService(service.BaseImageService):
|
||||
return local_filename
|
||||
|
||||
def _s3_create(self, context, metadata):
|
||||
"""Gets a manifext from s3 and makes an image"""
|
||||
"""Gets a manifext from s3 and makes an image."""
|
||||
|
||||
image_path = tempfile.mkdtemp(dir=FLAGS.image_decryption_dir)
|
||||
|
||||
image_location = metadata['properties']['image_location']
|
||||
bucket_name = image_location.split("/")[0]
|
||||
bucket_name = image_location.split('/')[0]
|
||||
manifest_path = image_location[len(bucket_name) + 1:]
|
||||
bucket = self._conn(context).get_bucket(bucket_name)
|
||||
key = bucket.get_key(manifest_path)
|
||||
@ -116,7 +118,7 @@ class S3ImageService(service.BaseImageService):
|
||||
image_type = 'machine'
|
||||
|
||||
try:
|
||||
kernel_id = manifest.find("machine_configuration/kernel_id").text
|
||||
kernel_id = manifest.find('machine_configuration/kernel_id').text
|
||||
if kernel_id == 'true':
|
||||
image_format = 'aki'
|
||||
image_type = 'kernel'
|
||||
@ -125,7 +127,7 @@ class S3ImageService(service.BaseImageService):
|
||||
kernel_id = None
|
||||
|
||||
try:
|
||||
ramdisk_id = manifest.find("machine_configuration/ramdisk_id").text
|
||||
ramdisk_id = manifest.find('machine_configuration/ramdisk_id').text
|
||||
if ramdisk_id == 'true':
|
||||
image_format = 'ari'
|
||||
image_type = 'ramdisk'
|
||||
@ -134,7 +136,7 @@ class S3ImageService(service.BaseImageService):
|
||||
ramdisk_id = None
|
||||
|
||||
try:
|
||||
arch = manifest.find("machine_configuration/architecture").text
|
||||
arch = manifest.find('machine_configuration/architecture').text
|
||||
except Exception:
|
||||
arch = 'x86_64'
|
||||
|
||||
@ -160,7 +162,7 @@ class S3ImageService(service.BaseImageService):
|
||||
def delayed_create():
|
||||
"""This handles the fetching and decrypting of the part files."""
|
||||
parts = []
|
||||
for fn_element in manifest.find("image").getiterator("filename"):
|
||||
for fn_element in manifest.find('image').getiterator('filename'):
|
||||
part = self._download_file(bucket, fn_element.text, image_path)
|
||||
parts.append(part)
|
||||
|
||||
@ -174,9 +176,9 @@ class S3ImageService(service.BaseImageService):
|
||||
metadata['properties']['image_state'] = 'decrypting'
|
||||
self.service.update(context, image_id, metadata)
|
||||
|
||||
hex_key = manifest.find("image/ec2_encrypted_key").text
|
||||
hex_key = manifest.find('image/ec2_encrypted_key').text
|
||||
encrypted_key = binascii.a2b_hex(hex_key)
|
||||
hex_iv = manifest.find("image/ec2_encrypted_iv").text
|
||||
hex_iv = manifest.find('image/ec2_encrypted_iv').text
|
||||
encrypted_iv = binascii.a2b_hex(hex_iv)
|
||||
|
||||
# FIXME(vish): grab key from common service so this can run on
|
||||
@ -214,7 +216,7 @@ class S3ImageService(service.BaseImageService):
|
||||
process_input=encrypted_key,
|
||||
check_exit_code=False)
|
||||
if err:
|
||||
raise exception.Error(_("Failed to decrypt private key: %s")
|
||||
raise exception.Error(_('Failed to decrypt private key: %s')
|
||||
% err)
|
||||
iv, err = utils.execute('openssl',
|
||||
'rsautl',
|
||||
@ -223,8 +225,8 @@ class S3ImageService(service.BaseImageService):
|
||||
process_input=encrypted_iv,
|
||||
check_exit_code=False)
|
||||
if err:
|
||||
raise exception.Error(_("Failed to decrypt initialization "
|
||||
"vector: %s") % err)
|
||||
raise exception.Error(_('Failed to decrypt initialization '
|
||||
'vector: %s') % err)
|
||||
|
||||
_out, err = utils.execute('openssl', 'enc',
|
||||
'-d', '-aes-128-cbc',
|
||||
@ -234,14 +236,14 @@ class S3ImageService(service.BaseImageService):
|
||||
'-out', '%s' % (decrypted_filename,),
|
||||
check_exit_code=False)
|
||||
if err:
|
||||
raise exception.Error(_("Failed to decrypt image file "
|
||||
"%(image_file)s: %(err)s") %
|
||||
raise exception.Error(_('Failed to decrypt image file '
|
||||
'%(image_file)s: %(err)s') %
|
||||
{'image_file': encrypted_filename,
|
||||
'err': err})
|
||||
|
||||
@staticmethod
|
||||
def _untarzip_image(path, filename):
|
||||
tar_file = tarfile.open(filename, "r|gz")
|
||||
tar_file = tarfile.open(filename, 'r|gz')
|
||||
tar_file.extractall(path)
|
||||
image_file = tar_file.getnames()[0]
|
||||
tar_file.close()
|
||||
|
@ -20,7 +20,7 @@ from nova import utils
|
||||
|
||||
|
||||
class BaseImageService(object):
|
||||
"""Base class for providing image search and retrieval services
|
||||
"""Base class for providing image search and retrieval services.
|
||||
|
||||
ImageService exposes two concepts of metadata:
|
||||
|
||||
@ -35,7 +35,9 @@ class BaseImageService(object):
|
||||
This means that ImageServices will return BASE_IMAGE_ATTRS as keys in the
|
||||
metadata dict, all other attributes will be returned as keys in the nested
|
||||
'properties' dict.
|
||||
|
||||
"""
|
||||
|
||||
BASE_IMAGE_ATTRS = ['id', 'name', 'created_at', 'updated_at',
|
||||
'deleted_at', 'deleted', 'status', 'is_public']
|
||||
|
||||
@ -45,23 +47,18 @@ class BaseImageService(object):
|
||||
SERVICE_IMAGE_ATTRS = []
|
||||
|
||||
def index(self, context):
|
||||
"""
|
||||
Returns a sequence of mappings of id and name information about
|
||||
images.
|
||||
"""List images.
|
||||
|
||||
:rtype: array
|
||||
:retval: a sequence of mappings with the following signature
|
||||
{'id': opaque id of image, 'name': name of image}
|
||||
:returns: a sequence of mappings with the following signature
|
||||
{'id': opaque id of image, 'name': name of image}
|
||||
|
||||
"""
|
||||
raise NotImplementedError
|
||||
|
||||
def detail(self, context):
|
||||
"""
|
||||
Returns a sequence of mappings of detailed information about images.
|
||||
"""Detailed information about an images.
|
||||
|
||||
:rtype: array
|
||||
:retval: a sequence of mappings with the following signature
|
||||
:returns: a sequence of mappings with the following signature
|
||||
{'id': opaque id of image,
|
||||
'name': name of image,
|
||||
'created_at': creation datetime object,
|
||||
@ -77,15 +74,14 @@ class BaseImageService(object):
|
||||
NotImplementedError, in which case Nova will emulate this method
|
||||
with repeated calls to show() for each image received from the
|
||||
index() method.
|
||||
|
||||
"""
|
||||
raise NotImplementedError
|
||||
|
||||
def show(self, context, image_id):
|
||||
"""
|
||||
Returns a dict containing image metadata for the given opaque image id.
|
||||
|
||||
:retval a mapping with the following signature:
|
||||
"""Detailed information about an image.
|
||||
|
||||
:returns: a mapping with the following signature:
|
||||
{'id': opaque id of image,
|
||||
'name': name of image,
|
||||
'created_at': creation datetime object,
|
||||
@ -96,54 +92,56 @@ class BaseImageService(object):
|
||||
'is_public': boolean indicating if image is public
|
||||
}, ...
|
||||
|
||||
:raises NotFound if the image does not exist
|
||||
:raises: NotFound if the image does not exist
|
||||
|
||||
"""
|
||||
raise NotImplementedError
|
||||
|
||||
def get(self, context, data):
|
||||
"""
|
||||
Returns a dict containing image metadata and writes image data to data.
|
||||
"""Get an image.
|
||||
|
||||
:param data: a file-like object to hold binary image data
|
||||
:returns: a dict containing image metadata, writes image data to data.
|
||||
:raises: NotFound if the image does not exist
|
||||
|
||||
:raises NotFound if the image does not exist
|
||||
"""
|
||||
raise NotImplementedError
|
||||
|
||||
def create(self, context, metadata, data=None):
|
||||
"""
|
||||
Store the image metadata and data and return the new image metadata.
|
||||
"""Store the image metadata and data.
|
||||
|
||||
:raises AlreadyExists if the image already exist.
|
||||
:returns: the new image metadata.
|
||||
:raises: AlreadyExists if the image already exist.
|
||||
|
||||
"""
|
||||
raise NotImplementedError
|
||||
|
||||
def update(self, context, image_id, metadata, data=None):
|
||||
"""Update the given image metadata and data and return the metadata
|
||||
"""Update the given image metadata and data and return the metadata.
|
||||
|
||||
:raises NotFound if the image does not exist.
|
||||
:raises: NotFound if the image does not exist.
|
||||
|
||||
"""
|
||||
raise NotImplementedError
|
||||
|
||||
def delete(self, context, image_id):
|
||||
"""
|
||||
Delete the given image.
|
||||
"""Delete the given image.
|
||||
|
||||
:raises NotFound if the image does not exist.
|
||||
:raises: NotFound if the image does not exist.
|
||||
|
||||
"""
|
||||
raise NotImplementedError
|
||||
|
||||
@staticmethod
|
||||
def _is_image_available(context, image_meta):
|
||||
"""
|
||||
"""Check image availability.
|
||||
|
||||
Images are always available if they are public or if the user is an
|
||||
admin.
|
||||
|
||||
Otherwise, we filter by project_id (if present) and then fall-back to
|
||||
images owned by user.
|
||||
|
||||
"""
|
||||
# FIXME(sirp): We should be filtering by user_id on the Glance side
|
||||
# for security; however, we can't do that until we get authn/authz
|
||||
@ -169,29 +167,32 @@ class BaseImageService(object):
|
||||
|
||||
This is used by subclasses to expose only a metadata dictionary that
|
||||
is the same across ImageService implementations.
|
||||
|
||||
"""
|
||||
return cls._propertify_metadata(metadata, cls.BASE_IMAGE_ATTRS)
|
||||
|
||||
@classmethod
|
||||
def _translate_to_service(cls, metadata):
|
||||
"""Return a metadata dictionary that is usable by the ImageService
|
||||
subclass.
|
||||
"""Return a metadata dict that is usable by the ImageService subclass.
|
||||
|
||||
As an example, Glance has additional attributes (like 'location'); the
|
||||
BaseImageService considers these properties, but we need to translate
|
||||
these back to first-class attrs for sending to Glance. This method
|
||||
handles this by allowing you to specify the attributes an ImageService
|
||||
considers first-class.
|
||||
|
||||
"""
|
||||
if not cls.SERVICE_IMAGE_ATTRS:
|
||||
raise NotImplementedError(_("Cannot use this without specifying "
|
||||
"SERVICE_IMAGE_ATTRS for subclass"))
|
||||
raise NotImplementedError(_('Cannot use this without specifying '
|
||||
'SERVICE_IMAGE_ATTRS for subclass'))
|
||||
return cls._propertify_metadata(metadata, cls.SERVICE_IMAGE_ATTRS)
|
||||
|
||||
@staticmethod
|
||||
def _propertify_metadata(metadata, keys):
|
||||
"""Return a dict with any unrecognized keys placed in the nested
|
||||
'properties' dict.
|
||||
"""Move unknown keys to a nested 'properties' dict.
|
||||
|
||||
:returns: a new dict with the keys moved.
|
||||
|
||||
"""
|
||||
flattened = utils.flatten_dict(metadata)
|
||||
attributes, properties = utils.partition_dict(flattened, keys)
|
||||
|
49
nova/log.py
49
nova/log.py
@ -16,16 +16,15 @@
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""
|
||||
Nova logging handler.
|
||||
"""Nova logging handler.
|
||||
|
||||
This module adds to logging functionality by adding the option to specify
|
||||
a context object when calling the various log methods. If the context object
|
||||
is not specified, default formatting is used.
|
||||
|
||||
It also allows setting of formatting information through flags.
|
||||
"""
|
||||
|
||||
"""
|
||||
|
||||
import cStringIO
|
||||
import inspect
|
||||
@ -41,34 +40,28 @@ from nova import version
|
||||
|
||||
|
||||
FLAGS = flags.FLAGS
|
||||
|
||||
flags.DEFINE_string('logging_context_format_string',
|
||||
'%(asctime)s %(levelname)s %(name)s '
|
||||
'[%(request_id)s %(user)s '
|
||||
'%(project)s] %(message)s',
|
||||
'format string to use for log messages with context')
|
||||
|
||||
flags.DEFINE_string('logging_default_format_string',
|
||||
'%(asctime)s %(levelname)s %(name)s [-] '
|
||||
'%(message)s',
|
||||
'format string to use for log messages without context')
|
||||
|
||||
flags.DEFINE_string('logging_debug_format_suffix',
|
||||
'from (pid=%(process)d) %(funcName)s'
|
||||
' %(pathname)s:%(lineno)d',
|
||||
'data to append to log format when level is DEBUG')
|
||||
|
||||
flags.DEFINE_string('logging_exception_prefix',
|
||||
'(%(name)s): TRACE: ',
|
||||
'prefix each line of exception output with this format')
|
||||
|
||||
flags.DEFINE_list('default_log_levels',
|
||||
['amqplib=WARN',
|
||||
'sqlalchemy=WARN',
|
||||
'boto=WARN',
|
||||
'eventlet.wsgi.server=WARN'],
|
||||
'list of logger=LEVEL pairs')
|
||||
|
||||
flags.DEFINE_bool('use_syslog', False, 'output to syslog')
|
||||
flags.DEFINE_string('logfile', None, 'output to named file')
|
||||
|
||||
@ -83,6 +76,8 @@ WARN = logging.WARN
|
||||
INFO = logging.INFO
|
||||
DEBUG = logging.DEBUG
|
||||
NOTSET = logging.NOTSET
|
||||
|
||||
|
||||
# methods
|
||||
getLogger = logging.getLogger
|
||||
debug = logging.debug
|
||||
@ -93,6 +88,8 @@ error = logging.error
|
||||
exception = logging.exception
|
||||
critical = logging.critical
|
||||
log = logging.log
|
||||
|
||||
|
||||
# handlers
|
||||
StreamHandler = logging.StreamHandler
|
||||
WatchedFileHandler = logging.handlers.WatchedFileHandler
|
||||
@ -127,17 +124,18 @@ def _get_log_file_path(binary=None):
|
||||
|
||||
|
||||
class NovaLogger(logging.Logger):
|
||||
"""
|
||||
NovaLogger manages request context and formatting.
|
||||
"""NovaLogger manages request context and formatting.
|
||||
|
||||
This becomes the class that is instanciated by logging.getLogger.
|
||||
|
||||
"""
|
||||
|
||||
def __init__(self, name, level=NOTSET):
|
||||
logging.Logger.__init__(self, name, level)
|
||||
self.setup_from_flags()
|
||||
|
||||
def setup_from_flags(self):
|
||||
"""Setup logger from flags"""
|
||||
"""Setup logger from flags."""
|
||||
level = NOTSET
|
||||
for pair in FLAGS.default_log_levels:
|
||||
logger, _sep, level_name = pair.partition('=')
|
||||
@ -148,7 +146,7 @@ class NovaLogger(logging.Logger):
|
||||
self.setLevel(level)
|
||||
|
||||
def _log(self, level, msg, args, exc_info=None, extra=None, context=None):
|
||||
"""Extract context from any log call"""
|
||||
"""Extract context from any log call."""
|
||||
if not extra:
|
||||
extra = {}
|
||||
if context:
|
||||
@ -157,17 +155,17 @@ class NovaLogger(logging.Logger):
|
||||
return logging.Logger._log(self, level, msg, args, exc_info, extra)
|
||||
|
||||
def addHandler(self, handler):
|
||||
"""Each handler gets our custom formatter"""
|
||||
"""Each handler gets our custom formatter."""
|
||||
handler.setFormatter(_formatter)
|
||||
return logging.Logger.addHandler(self, handler)
|
||||
|
||||
def audit(self, msg, *args, **kwargs):
|
||||
"""Shortcut for our AUDIT level"""
|
||||
"""Shortcut for our AUDIT level."""
|
||||
if self.isEnabledFor(AUDIT):
|
||||
self._log(AUDIT, msg, args, **kwargs)
|
||||
|
||||
def exception(self, msg, *args, **kwargs):
|
||||
"""Logging.exception doesn't handle kwargs, so breaks context"""
|
||||
"""Logging.exception doesn't handle kwargs, so breaks context."""
|
||||
if not kwargs.get('exc_info'):
|
||||
kwargs['exc_info'] = 1
|
||||
self.error(msg, *args, **kwargs)
|
||||
@ -181,14 +179,13 @@ class NovaLogger(logging.Logger):
|
||||
for k in env.keys():
|
||||
if not isinstance(env[k], str):
|
||||
env.pop(k)
|
||||
message = "Environment: %s" % json.dumps(env)
|
||||
message = 'Environment: %s' % json.dumps(env)
|
||||
kwargs.pop('exc_info')
|
||||
self.error(message, **kwargs)
|
||||
|
||||
|
||||
class NovaFormatter(logging.Formatter):
|
||||
"""
|
||||
A nova.context.RequestContext aware formatter configured through flags.
|
||||
"""A nova.context.RequestContext aware formatter configured through flags.
|
||||
|
||||
The flags used to set format strings are: logging_context_foramt_string
|
||||
and logging_default_format_string. You can also specify
|
||||
@ -197,10 +194,11 @@ class NovaFormatter(logging.Formatter):
|
||||
|
||||
For information about what variables are available for the formatter see:
|
||||
http://docs.python.org/library/logging.html#formatter
|
||||
|
||||
"""
|
||||
|
||||
def format(self, record):
|
||||
"""Uses contextstring if request_id is set, otherwise default"""
|
||||
"""Uses contextstring if request_id is set, otherwise default."""
|
||||
if record.__dict__.get('request_id', None):
|
||||
self._fmt = FLAGS.logging_context_format_string
|
||||
else:
|
||||
@ -214,20 +212,21 @@ class NovaFormatter(logging.Formatter):
|
||||
return logging.Formatter.format(self, record)
|
||||
|
||||
def formatException(self, exc_info, record=None):
|
||||
"""Format exception output with FLAGS.logging_exception_prefix"""
|
||||
"""Format exception output with FLAGS.logging_exception_prefix."""
|
||||
if not record:
|
||||
return logging.Formatter.formatException(self, exc_info)
|
||||
stringbuffer = cStringIO.StringIO()
|
||||
traceback.print_exception(exc_info[0], exc_info[1], exc_info[2],
|
||||
None, stringbuffer)
|
||||
lines = stringbuffer.getvalue().split("\n")
|
||||
lines = stringbuffer.getvalue().split('\n')
|
||||
stringbuffer.close()
|
||||
formatted_lines = []
|
||||
for line in lines:
|
||||
pl = FLAGS.logging_exception_prefix % record.__dict__
|
||||
fl = "%s%s" % (pl, line)
|
||||
fl = '%s%s' % (pl, line)
|
||||
formatted_lines.append(fl)
|
||||
return "\n".join(formatted_lines)
|
||||
return '\n'.join(formatted_lines)
|
||||
|
||||
|
||||
_formatter = NovaFormatter()
|
||||
|
||||
@ -241,7 +240,7 @@ class NovaRootLogger(NovaLogger):
|
||||
NovaLogger.__init__(self, name, level)
|
||||
|
||||
def setup_from_flags(self):
|
||||
"""Setup logger from flags"""
|
||||
"""Setup logger from flags."""
|
||||
global _filelog
|
||||
if FLAGS.use_syslog:
|
||||
self.syslog = SysLogHandler(address='/dev/log')
|
||||
|
@ -16,7 +16,8 @@
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""
|
||||
"""Base Manager class.
|
||||
|
||||
Managers are responsible for a certain aspect of the sytem. It is a logical
|
||||
grouping of code relating to a portion of the system. In general other
|
||||
components should be using the manager to make changes to the components that
|
||||
@ -49,16 +50,19 @@ Managers will often provide methods for initial setup of a host or periodic
|
||||
tasksto a wrapping service.
|
||||
|
||||
This module provides Manager, a base class for managers.
|
||||
|
||||
"""
|
||||
|
||||
from nova import utils
|
||||
from nova import flags
|
||||
from nova import log as logging
|
||||
from nova import utils
|
||||
from nova.db import base
|
||||
from nova.scheduler import api
|
||||
|
||||
|
||||
FLAGS = flags.FLAGS
|
||||
|
||||
|
||||
LOG = logging.getLogger('nova.manager')
|
||||
|
||||
|
||||
@ -70,23 +74,29 @@ class Manager(base.Base):
|
||||
super(Manager, self).__init__(db_driver)
|
||||
|
||||
def periodic_tasks(self, context=None):
|
||||
"""Tasks to be run at a periodic interval"""
|
||||
"""Tasks to be run at a periodic interval."""
|
||||
pass
|
||||
|
||||
def init_host(self):
|
||||
"""Do any initialization that needs to be run if this is a standalone
|
||||
service. Child classes should override this method."""
|
||||
"""Handle initialization if this is a standalone service.
|
||||
|
||||
Child classes should override this method.
|
||||
|
||||
"""
|
||||
pass
|
||||
|
||||
|
||||
class SchedulerDependentManager(Manager):
|
||||
"""Periodically send capability updates to the Scheduler services.
|
||||
Services that need to update the Scheduler of their capabilities
|
||||
should derive from this class. Otherwise they can derive from
|
||||
manager.Manager directly. Updates are only sent after
|
||||
update_service_capabilities is called with non-None values."""
|
||||
|
||||
def __init__(self, host=None, db_driver=None, service_name="undefined"):
|
||||
Services that need to update the Scheduler of their capabilities
|
||||
should derive from this class. Otherwise they can derive from
|
||||
manager.Manager directly. Updates are only sent after
|
||||
update_service_capabilities is called with non-None values.
|
||||
|
||||
"""
|
||||
|
||||
def __init__(self, host=None, db_driver=None, service_name='undefined'):
|
||||
self.last_capabilities = None
|
||||
self.service_name = service_name
|
||||
super(SchedulerDependentManager, self).__init__(host, db_driver)
|
||||
@ -96,9 +106,9 @@ class SchedulerDependentManager(Manager):
|
||||
self.last_capabilities = capabilities
|
||||
|
||||
def periodic_tasks(self, context=None):
|
||||
"""Pass data back to the scheduler at a periodic interval"""
|
||||
"""Pass data back to the scheduler at a periodic interval."""
|
||||
if self.last_capabilities:
|
||||
LOG.debug(_("Notifying Schedulers of capabilities ..."))
|
||||
LOG.debug(_('Notifying Schedulers of capabilities ...'))
|
||||
api.update_service_capabilities(context, self.service_name,
|
||||
self.host, self.last_capabilities)
|
||||
|
||||
|
@ -15,16 +15,15 @@
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
"""
|
||||
Quotas for instances, volumes, and floating ips
|
||||
"""
|
||||
|
||||
"""Quotas for instances, volumes, and floating ips."""
|
||||
|
||||
from nova import db
|
||||
from nova import exception
|
||||
from nova import flags
|
||||
|
||||
FLAGS = flags.FLAGS
|
||||
|
||||
FLAGS = flags.FLAGS
|
||||
flags.DEFINE_integer('quota_instances', 10,
|
||||
'number of instances allowed per project')
|
||||
flags.DEFINE_integer('quota_cores', 20,
|
||||
@ -64,7 +63,7 @@ def get_quota(context, project_id):
|
||||
|
||||
|
||||
def allowed_instances(context, num_instances, instance_type):
|
||||
"""Check quota and return min(num_instances, allowed_instances)"""
|
||||
"""Check quota and return min(num_instances, allowed_instances)."""
|
||||
project_id = context.project_id
|
||||
context = context.elevated()
|
||||
used_instances, used_cores = db.instance_data_get_for_project(context,
|
||||
@ -79,7 +78,7 @@ def allowed_instances(context, num_instances, instance_type):
|
||||
|
||||
|
||||
def allowed_volumes(context, num_volumes, size):
|
||||
"""Check quota and return min(num_volumes, allowed_volumes)"""
|
||||
"""Check quota and return min(num_volumes, allowed_volumes)."""
|
||||
project_id = context.project_id
|
||||
context = context.elevated()
|
||||
used_volumes, used_gigabytes = db.volume_data_get_for_project(context,
|
||||
@ -95,7 +94,7 @@ def allowed_volumes(context, num_volumes, size):
|
||||
|
||||
|
||||
def allowed_floating_ips(context, num_floating_ips):
|
||||
"""Check quota and return min(num_floating_ips, allowed_floating_ips)"""
|
||||
"""Check quota and return min(num_floating_ips, allowed_floating_ips)."""
|
||||
project_id = context.project_id
|
||||
context = context.elevated()
|
||||
used_floating_ips = db.floating_ip_count_by_project(context, project_id)
|
||||
@ -105,7 +104,7 @@ def allowed_floating_ips(context, num_floating_ips):
|
||||
|
||||
|
||||
def allowed_metadata_items(context, num_metadata_items):
|
||||
"""Check quota; return min(num_metadata_items,allowed_metadata_items)"""
|
||||
"""Check quota; return min(num_metadata_items,allowed_metadata_items)."""
|
||||
project_id = context.project_id
|
||||
context = context.elevated()
|
||||
quota = get_quota(context, project_id)
|
||||
@ -114,20 +113,20 @@ def allowed_metadata_items(context, num_metadata_items):
|
||||
|
||||
|
||||
def allowed_injected_files(context):
|
||||
"""Return the number of injected files allowed"""
|
||||
"""Return the number of injected files allowed."""
|
||||
return FLAGS.quota_max_injected_files
|
||||
|
||||
|
||||
def allowed_injected_file_content_bytes(context):
|
||||
"""Return the number of bytes allowed per injected file content"""
|
||||
"""Return the number of bytes allowed per injected file content."""
|
||||
return FLAGS.quota_max_injected_file_content_bytes
|
||||
|
||||
|
||||
def allowed_injected_file_path_bytes(context):
|
||||
"""Return the number of bytes allowed in an injected file path"""
|
||||
"""Return the number of bytes allowed in an injected file path."""
|
||||
return FLAGS.quota_max_injected_file_path_bytes
|
||||
|
||||
|
||||
class QuotaError(exception.ApiError):
|
||||
"""Quota Exceeeded"""
|
||||
"""Quota Exceeeded."""
|
||||
pass
|
||||
|
152
nova/rpc.py
152
nova/rpc.py
@ -16,9 +16,12 @@
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""
|
||||
AMQP-based RPC. Queues have consumers and publishers.
|
||||
"""AMQP-based RPC.
|
||||
|
||||
Queues have consumers and publishers.
|
||||
|
||||
No fan-out support yet.
|
||||
|
||||
"""
|
||||
|
||||
import json
|
||||
@ -40,17 +43,19 @@ from nova import log as logging
|
||||
from nova import utils
|
||||
|
||||
|
||||
FLAGS = flags.FLAGS
|
||||
LOG = logging.getLogger('nova.rpc')
|
||||
|
||||
|
||||
FLAGS = flags.FLAGS
|
||||
flags.DEFINE_integer('rpc_thread_pool_size', 1024, 'Size of RPC thread pool')
|
||||
|
||||
|
||||
class Connection(carrot_connection.BrokerConnection):
|
||||
"""Connection instance object"""
|
||||
"""Connection instance object."""
|
||||
|
||||
@classmethod
|
||||
def instance(cls, new=True):
|
||||
"""Returns the instance"""
|
||||
"""Returns the instance."""
|
||||
if new or not hasattr(cls, '_instance'):
|
||||
params = dict(hostname=FLAGS.rabbit_host,
|
||||
port=FLAGS.rabbit_port,
|
||||
@ -71,9 +76,11 @@ class Connection(carrot_connection.BrokerConnection):
|
||||
|
||||
@classmethod
|
||||
def recreate(cls):
|
||||
"""Recreates the connection instance
|
||||
"""Recreates the connection instance.
|
||||
|
||||
This is necessary to recover from some network errors/disconnects"""
|
||||
This is necessary to recover from some network errors/disconnects.
|
||||
|
||||
"""
|
||||
try:
|
||||
del cls._instance
|
||||
except AttributeError, e:
|
||||
@ -84,10 +91,12 @@ class Connection(carrot_connection.BrokerConnection):
|
||||
|
||||
|
||||
class Consumer(messaging.Consumer):
|
||||
"""Consumer base class
|
||||
"""Consumer base class.
|
||||
|
||||
Contains methods for connecting the fetch method to async loops.
|
||||
|
||||
Contains methods for connecting the fetch method to async loops
|
||||
"""
|
||||
|
||||
def __init__(self, *args, **kwargs):
|
||||
for i in xrange(FLAGS.rabbit_max_retries):
|
||||
if i > 0:
|
||||
@ -100,19 +109,18 @@ class Consumer(messaging.Consumer):
|
||||
fl_host = FLAGS.rabbit_host
|
||||
fl_port = FLAGS.rabbit_port
|
||||
fl_intv = FLAGS.rabbit_retry_interval
|
||||
LOG.error(_("AMQP server on %(fl_host)s:%(fl_port)d is"
|
||||
" unreachable: %(e)s. Trying again in %(fl_intv)d"
|
||||
" seconds.")
|
||||
% locals())
|
||||
LOG.error(_('AMQP server on %(fl_host)s:%(fl_port)d is'
|
||||
' unreachable: %(e)s. Trying again in %(fl_intv)d'
|
||||
' seconds.') % locals())
|
||||
self.failed_connection = True
|
||||
if self.failed_connection:
|
||||
LOG.error(_("Unable to connect to AMQP server "
|
||||
"after %d tries. Shutting down."),
|
||||
LOG.error(_('Unable to connect to AMQP server '
|
||||
'after %d tries. Shutting down.'),
|
||||
FLAGS.rabbit_max_retries)
|
||||
sys.exit(1)
|
||||
|
||||
def fetch(self, no_ack=None, auto_ack=None, enable_callbacks=False):
|
||||
"""Wraps the parent fetch with some logic for failed connections"""
|
||||
"""Wraps the parent fetch with some logic for failed connection."""
|
||||
# TODO(vish): the logic for failed connections and logging should be
|
||||
# refactored into some sort of connection manager object
|
||||
try:
|
||||
@ -125,14 +133,14 @@ class Consumer(messaging.Consumer):
|
||||
self.declare()
|
||||
super(Consumer, self).fetch(no_ack, auto_ack, enable_callbacks)
|
||||
if self.failed_connection:
|
||||
LOG.error(_("Reconnected to queue"))
|
||||
LOG.error(_('Reconnected to queue'))
|
||||
self.failed_connection = False
|
||||
# NOTE(vish): This is catching all errors because we really don't
|
||||
# want exceptions to be logged 10 times a second if some
|
||||
# persistent failure occurs.
|
||||
except Exception, e: # pylint: disable=W0703
|
||||
if not self.failed_connection:
|
||||
LOG.exception(_("Failed to fetch message from queue: %s" % e))
|
||||
LOG.exception(_('Failed to fetch message from queue: %s' % e))
|
||||
self.failed_connection = True
|
||||
|
||||
def attach_to_eventlet(self):
|
||||
@ -143,8 +151,9 @@ class Consumer(messaging.Consumer):
|
||||
|
||||
|
||||
class AdapterConsumer(Consumer):
|
||||
"""Calls methods on a proxy object based on method and args"""
|
||||
def __init__(self, connection=None, topic="broadcast", proxy=None):
|
||||
"""Calls methods on a proxy object based on method and args."""
|
||||
|
||||
def __init__(self, connection=None, topic='broadcast', proxy=None):
|
||||
LOG.debug(_('Initing the Adapter Consumer for %s') % topic)
|
||||
self.proxy = proxy
|
||||
self.pool = greenpool.GreenPool(FLAGS.rpc_thread_pool_size)
|
||||
@ -156,13 +165,14 @@ class AdapterConsumer(Consumer):
|
||||
|
||||
@exception.wrap_exception
|
||||
def _receive(self, message_data, message):
|
||||
"""Magically looks for a method on the proxy object and calls it
|
||||
"""Magically looks for a method on the proxy object and calls it.
|
||||
|
||||
Message data should be a dictionary with two keys:
|
||||
method: string representing the method to call
|
||||
args: dictionary of arg: value
|
||||
|
||||
Example: {'method': 'echo', 'args': {'value': 42}}
|
||||
|
||||
"""
|
||||
LOG.debug(_('received %s') % message_data)
|
||||
msg_id = message_data.pop('_msg_id', None)
|
||||
@ -189,22 +199,23 @@ class AdapterConsumer(Consumer):
|
||||
if msg_id:
|
||||
msg_reply(msg_id, rval, None)
|
||||
except Exception as e:
|
||||
logging.exception("Exception during message handling")
|
||||
logging.exception('Exception during message handling')
|
||||
if msg_id:
|
||||
msg_reply(msg_id, None, sys.exc_info())
|
||||
return
|
||||
|
||||
|
||||
class Publisher(messaging.Publisher):
|
||||
"""Publisher base class"""
|
||||
"""Publisher base class."""
|
||||
pass
|
||||
|
||||
|
||||
class TopicAdapterConsumer(AdapterConsumer):
|
||||
"""Consumes messages on a specific topic"""
|
||||
exchange_type = "topic"
|
||||
"""Consumes messages on a specific topic."""
|
||||
|
||||
def __init__(self, connection=None, topic="broadcast", proxy=None):
|
||||
exchange_type = 'topic'
|
||||
|
||||
def __init__(self, connection=None, topic='broadcast', proxy=None):
|
||||
self.queue = topic
|
||||
self.routing_key = topic
|
||||
self.exchange = FLAGS.control_exchange
|
||||
@ -214,27 +225,29 @@ class TopicAdapterConsumer(AdapterConsumer):
|
||||
|
||||
|
||||
class FanoutAdapterConsumer(AdapterConsumer):
|
||||
"""Consumes messages from a fanout exchange"""
|
||||
exchange_type = "fanout"
|
||||
"""Consumes messages from a fanout exchange."""
|
||||
|
||||
def __init__(self, connection=None, topic="broadcast", proxy=None):
|
||||
self.exchange = "%s_fanout" % topic
|
||||
exchange_type = 'fanout'
|
||||
|
||||
def __init__(self, connection=None, topic='broadcast', proxy=None):
|
||||
self.exchange = '%s_fanout' % topic
|
||||
self.routing_key = topic
|
||||
unique = uuid.uuid4().hex
|
||||
self.queue = "%s_fanout_%s" % (topic, unique)
|
||||
self.queue = '%s_fanout_%s' % (topic, unique)
|
||||
self.durable = False
|
||||
LOG.info(_("Created '%(exchange)s' fanout exchange "
|
||||
"with '%(key)s' routing key"),
|
||||
dict(exchange=self.exchange, key=self.routing_key))
|
||||
LOG.info(_('Created "%(exchange)s" fanout exchange '
|
||||
'with "%(key)s" routing key'),
|
||||
dict(exchange=self.exchange, key=self.routing_key))
|
||||
super(FanoutAdapterConsumer, self).__init__(connection=connection,
|
||||
topic=topic, proxy=proxy)
|
||||
|
||||
|
||||
class TopicPublisher(Publisher):
|
||||
"""Publishes messages on a specific topic"""
|
||||
exchange_type = "topic"
|
||||
"""Publishes messages on a specific topic."""
|
||||
|
||||
def __init__(self, connection=None, topic="broadcast"):
|
||||
exchange_type = 'topic'
|
||||
|
||||
def __init__(self, connection=None, topic='broadcast'):
|
||||
self.routing_key = topic
|
||||
self.exchange = FLAGS.control_exchange
|
||||
self.durable = False
|
||||
@ -243,20 +256,22 @@ class TopicPublisher(Publisher):
|
||||
|
||||
class FanoutPublisher(Publisher):
|
||||
"""Publishes messages to a fanout exchange."""
|
||||
exchange_type = "fanout"
|
||||
|
||||
exchange_type = 'fanout'
|
||||
|
||||
def __init__(self, topic, connection=None):
|
||||
self.exchange = "%s_fanout" % topic
|
||||
self.queue = "%s_fanout" % topic
|
||||
self.exchange = '%s_fanout' % topic
|
||||
self.queue = '%s_fanout' % topic
|
||||
self.durable = False
|
||||
LOG.info(_("Creating '%(exchange)s' fanout exchange"),
|
||||
dict(exchange=self.exchange))
|
||||
LOG.info(_('Creating "%(exchange)s" fanout exchange'),
|
||||
dict(exchange=self.exchange))
|
||||
super(FanoutPublisher, self).__init__(connection=connection)
|
||||
|
||||
|
||||
class DirectConsumer(Consumer):
|
||||
"""Consumes messages directly on a channel specified by msg_id"""
|
||||
exchange_type = "direct"
|
||||
"""Consumes messages directly on a channel specified by msg_id."""
|
||||
|
||||
exchange_type = 'direct'
|
||||
|
||||
def __init__(self, connection=None, msg_id=None):
|
||||
self.queue = msg_id
|
||||
@ -268,8 +283,9 @@ class DirectConsumer(Consumer):
|
||||
|
||||
|
||||
class DirectPublisher(Publisher):
|
||||
"""Publishes messages directly on a channel specified by msg_id"""
|
||||
exchange_type = "direct"
|
||||
"""Publishes messages directly on a channel specified by msg_id."""
|
||||
|
||||
exchange_type = 'direct'
|
||||
|
||||
def __init__(self, connection=None, msg_id=None):
|
||||
self.routing_key = msg_id
|
||||
@ -279,9 +295,9 @@ class DirectPublisher(Publisher):
|
||||
|
||||
|
||||
def msg_reply(msg_id, reply=None, failure=None):
|
||||
"""Sends a reply or an error on the channel signified by msg_id
|
||||
"""Sends a reply or an error on the channel signified by msg_id.
|
||||
|
||||
failure should be a sys.exc_info() tuple.
|
||||
Failure should be a sys.exc_info() tuple.
|
||||
|
||||
"""
|
||||
if failure:
|
||||
@ -303,17 +319,20 @@ def msg_reply(msg_id, reply=None, failure=None):
|
||||
|
||||
|
||||
class RemoteError(exception.Error):
|
||||
"""Signifies that a remote class has raised an exception
|
||||
"""Signifies that a remote class has raised an exception.
|
||||
|
||||
Containes a string representation of the type of the original exception,
|
||||
the value of the original exception, and the traceback. These are
|
||||
sent to the parent as a joined string so printing the exception
|
||||
contains all of the relevent info."""
|
||||
contains all of the relevent info.
|
||||
|
||||
"""
|
||||
|
||||
def __init__(self, exc_type, value, traceback):
|
||||
self.exc_type = exc_type
|
||||
self.value = value
|
||||
self.traceback = traceback
|
||||
super(RemoteError, self).__init__("%s %s\n%s" % (exc_type,
|
||||
super(RemoteError, self).__init__('%s %s\n%s' % (exc_type,
|
||||
value,
|
||||
traceback))
|
||||
|
||||
@ -339,6 +358,7 @@ def _pack_context(msg, context):
|
||||
context out into a bunch of separate keys. If we want to support
|
||||
more arguments in rabbit messages, we may want to do the same
|
||||
for args at some point.
|
||||
|
||||
"""
|
||||
context = dict([('_context_%s' % key, value)
|
||||
for (key, value) in context.to_dict().iteritems()])
|
||||
@ -346,11 +366,11 @@ def _pack_context(msg, context):
|
||||
|
||||
|
||||
def call(context, topic, msg):
|
||||
"""Sends a message on a topic and wait for a response"""
|
||||
LOG.debug(_("Making asynchronous call on %s ..."), topic)
|
||||
"""Sends a message on a topic and wait for a response."""
|
||||
LOG.debug(_('Making asynchronous call on %s ...'), topic)
|
||||
msg_id = uuid.uuid4().hex
|
||||
msg.update({'_msg_id': msg_id})
|
||||
LOG.debug(_("MSG_ID is %s") % (msg_id))
|
||||
LOG.debug(_('MSG_ID is %s') % (msg_id))
|
||||
_pack_context(msg, context)
|
||||
|
||||
class WaitMessage(object):
|
||||
@ -387,8 +407,8 @@ def call(context, topic, msg):
|
||||
|
||||
|
||||
def cast(context, topic, msg):
|
||||
"""Sends a message on a topic without waiting for a response"""
|
||||
LOG.debug(_("Making asynchronous cast on %s..."), topic)
|
||||
"""Sends a message on a topic without waiting for a response."""
|
||||
LOG.debug(_('Making asynchronous cast on %s...'), topic)
|
||||
_pack_context(msg, context)
|
||||
conn = Connection.instance()
|
||||
publisher = TopicPublisher(connection=conn, topic=topic)
|
||||
@ -397,8 +417,8 @@ def cast(context, topic, msg):
|
||||
|
||||
|
||||
def fanout_cast(context, topic, msg):
|
||||
"""Sends a message on a fanout exchange without waiting for a response"""
|
||||
LOG.debug(_("Making asynchronous fanout cast..."))
|
||||
"""Sends a message on a fanout exchange without waiting for a response."""
|
||||
LOG.debug(_('Making asynchronous fanout cast...'))
|
||||
_pack_context(msg, context)
|
||||
conn = Connection.instance()
|
||||
publisher = FanoutPublisher(topic, connection=conn)
|
||||
@ -407,14 +427,14 @@ def fanout_cast(context, topic, msg):
|
||||
|
||||
|
||||
def generic_response(message_data, message):
|
||||
"""Logs a result and exits"""
|
||||
"""Logs a result and exits."""
|
||||
LOG.debug(_('response %s'), message_data)
|
||||
message.ack()
|
||||
sys.exit(0)
|
||||
|
||||
|
||||
def send_message(topic, message, wait=True):
|
||||
"""Sends a message for testing"""
|
||||
"""Sends a message for testing."""
|
||||
msg_id = uuid.uuid4().hex
|
||||
message.update({'_msg_id': msg_id})
|
||||
LOG.debug(_('topic is %s'), topic)
|
||||
@ -425,14 +445,14 @@ def send_message(topic, message, wait=True):
|
||||
queue=msg_id,
|
||||
exchange=msg_id,
|
||||
auto_delete=True,
|
||||
exchange_type="direct",
|
||||
exchange_type='direct',
|
||||
routing_key=msg_id)
|
||||
consumer.register_callback(generic_response)
|
||||
|
||||
publisher = messaging.Publisher(connection=Connection.instance(),
|
||||
exchange=FLAGS.control_exchange,
|
||||
durable=False,
|
||||
exchange_type="topic",
|
||||
exchange_type='topic',
|
||||
routing_key=topic)
|
||||
publisher.send(message)
|
||||
publisher.close()
|
||||
@ -441,8 +461,8 @@ def send_message(topic, message, wait=True):
|
||||
consumer.wait()
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
# NOTE(vish): you can send messages from the command line using
|
||||
# topic and a json sting representing a dictionary
|
||||
# for the method
|
||||
if __name__ == '__main__':
|
||||
# You can send messages from the command line using
|
||||
# topic and a json string representing a dictionary
|
||||
# for the method
|
||||
send_message(sys.argv[1], json.loads(sys.argv[2]))
|
||||
|
@ -17,9 +17,7 @@
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""
|
||||
Generic Node baseclass for all workers that run on hosts
|
||||
"""
|
||||
"""Generic Node baseclass for all workers that run on hosts."""
|
||||
|
||||
import inspect
|
||||
import os
|
||||
@ -30,13 +28,11 @@ from eventlet import event
|
||||
from eventlet import greenthread
|
||||
from eventlet import greenpool
|
||||
|
||||
from sqlalchemy.exc import OperationalError
|
||||
|
||||
from nova import context
|
||||
from nova import db
|
||||
from nova import exception
|
||||
from nova import log as logging
|
||||
from nova import flags
|
||||
from nova import log as logging
|
||||
from nova import rpc
|
||||
from nova import utils
|
||||
from nova import version
|
||||
@ -79,7 +75,7 @@ class Service(object):
|
||||
|
||||
def start(self):
|
||||
vcs_string = version.version_string_with_vcs()
|
||||
logging.audit(_("Starting %(topic)s node (version %(vcs_string)s)"),
|
||||
logging.audit(_('Starting %(topic)s node (version %(vcs_string)s)'),
|
||||
{'topic': self.topic, 'vcs_string': vcs_string})
|
||||
self.manager.init_host()
|
||||
self.model_disconnected = False
|
||||
@ -140,29 +136,24 @@ class Service(object):
|
||||
return getattr(manager, key)
|
||||
|
||||
@classmethod
|
||||
def create(cls,
|
||||
host=None,
|
||||
binary=None,
|
||||
topic=None,
|
||||
manager=None,
|
||||
report_interval=None,
|
||||
periodic_interval=None):
|
||||
def create(cls, host=None, binary=None, topic=None, manager=None,
|
||||
report_interval=None, periodic_interval=None):
|
||||
"""Instantiates class and passes back application object.
|
||||
|
||||
Args:
|
||||
host, defaults to FLAGS.host
|
||||
binary, defaults to basename of executable
|
||||
topic, defaults to bin_name - "nova-" part
|
||||
manager, defaults to FLAGS.<topic>_manager
|
||||
report_interval, defaults to FLAGS.report_interval
|
||||
periodic_interval, defaults to FLAGS.periodic_interval
|
||||
:param host: defaults to FLAGS.host
|
||||
:param binary: defaults to basename of executable
|
||||
:param topic: defaults to bin_name - 'nova-' part
|
||||
:param manager: defaults to FLAGS.<topic>_manager
|
||||
:param report_interval: defaults to FLAGS.report_interval
|
||||
:param periodic_interval: defaults to FLAGS.periodic_interval
|
||||
|
||||
"""
|
||||
if not host:
|
||||
host = FLAGS.host
|
||||
if not binary:
|
||||
binary = os.path.basename(inspect.stack()[-1][1])
|
||||
if not topic:
|
||||
topic = binary.rpartition("nova-")[2]
|
||||
topic = binary.rpartition('nova-')[2]
|
||||
if not manager:
|
||||
manager = FLAGS.get('%s_manager' % topic, None)
|
||||
if not report_interval:
|
||||
@ -175,12 +166,12 @@ class Service(object):
|
||||
return service_obj
|
||||
|
||||
def kill(self):
|
||||
"""Destroy the service object in the datastore"""
|
||||
"""Destroy the service object in the datastore."""
|
||||
self.stop()
|
||||
try:
|
||||
db.service_destroy(context.get_admin_context(), self.service_id)
|
||||
except exception.NotFound:
|
||||
logging.warn(_("Service killed that has no database entry"))
|
||||
logging.warn(_('Service killed that has no database entry'))
|
||||
|
||||
def stop(self):
|
||||
for x in self.timers:
|
||||
@ -198,7 +189,7 @@ class Service(object):
|
||||
pass
|
||||
|
||||
def periodic_tasks(self):
|
||||
"""Tasks to be run at a periodic interval"""
|
||||
"""Tasks to be run at a periodic interval."""
|
||||
self.manager.periodic_tasks(context.get_admin_context())
|
||||
|
||||
def report_state(self):
|
||||
@ -208,8 +199,8 @@ class Service(object):
|
||||
try:
|
||||
service_ref = db.service_get(ctxt, self.service_id)
|
||||
except exception.NotFound:
|
||||
logging.debug(_("The service database object disappeared, "
|
||||
"Recreating it."))
|
||||
logging.debug(_('The service database object disappeared, '
|
||||
'Recreating it.'))
|
||||
self._create_service_ref(ctxt)
|
||||
service_ref = db.service_get(ctxt, self.service_id)
|
||||
|
||||
@ -218,23 +209,24 @@ class Service(object):
|
||||
{'report_count': service_ref['report_count'] + 1})
|
||||
|
||||
# TODO(termie): make this pattern be more elegant.
|
||||
if getattr(self, "model_disconnected", False):
|
||||
if getattr(self, 'model_disconnected', False):
|
||||
self.model_disconnected = False
|
||||
logging.error(_("Recovered model server connection!"))
|
||||
logging.error(_('Recovered model server connection!'))
|
||||
|
||||
# TODO(vish): this should probably only catch connection errors
|
||||
except Exception: # pylint: disable=W0702
|
||||
if not getattr(self, "model_disconnected", False):
|
||||
if not getattr(self, 'model_disconnected', False):
|
||||
self.model_disconnected = True
|
||||
logging.exception(_("model server went away"))
|
||||
logging.exception(_('model server went away'))
|
||||
|
||||
|
||||
class WsgiService(object):
|
||||
"""Base class for WSGI based services.
|
||||
|
||||
For each api you define, you must also define these flags:
|
||||
:<api>_listen: The address on which to listen
|
||||
:<api>_listen_port: The port on which to listen
|
||||
:<api>_listen: The address on which to listen
|
||||
:<api>_listen_port: The port on which to listen
|
||||
|
||||
"""
|
||||
|
||||
def __init__(self, conf, apis):
|
||||
@ -250,13 +242,14 @@ class WsgiService(object):
|
||||
|
||||
|
||||
class ApiService(WsgiService):
|
||||
"""Class for our nova-api service"""
|
||||
"""Class for our nova-api service."""
|
||||
|
||||
@classmethod
|
||||
def create(cls, conf=None):
|
||||
if not conf:
|
||||
conf = wsgi.paste_config_file(FLAGS.api_paste_config)
|
||||
if not conf:
|
||||
message = (_("No paste configuration found for: %s"),
|
||||
message = (_('No paste configuration found for: %s'),
|
||||
FLAGS.api_paste_config)
|
||||
raise exception.Error(message)
|
||||
api_endpoints = ['ec2', 'osapi']
|
||||
@ -280,11 +273,11 @@ def serve(*services):
|
||||
FLAGS.ParseNewFlags()
|
||||
|
||||
name = '_'.join(x.binary for x in services)
|
||||
logging.debug(_("Serving %s"), name)
|
||||
logging.debug(_("Full set of FLAGS:"))
|
||||
logging.debug(_('Serving %s'), name)
|
||||
logging.debug(_('Full set of FLAGS:'))
|
||||
for flag in FLAGS:
|
||||
flag_get = FLAGS.get(flag, None)
|
||||
logging.debug("%(flag)s : %(flag_get)s" % locals())
|
||||
logging.debug('%(flag)s : %(flag_get)s' % locals())
|
||||
|
||||
for x in services:
|
||||
x.start()
|
||||
@ -315,20 +308,20 @@ def serve_wsgi(cls, conf=None):
|
||||
|
||||
|
||||
def _run_wsgi(paste_config_file, apis):
|
||||
logging.debug(_("Using paste.deploy config at: %s"), paste_config_file)
|
||||
logging.debug(_('Using paste.deploy config at: %s'), paste_config_file)
|
||||
apps = []
|
||||
for api in apis:
|
||||
config = wsgi.load_paste_configuration(paste_config_file, api)
|
||||
if config is None:
|
||||
logging.debug(_("No paste configuration for app: %s"), api)
|
||||
logging.debug(_('No paste configuration for app: %s'), api)
|
||||
continue
|
||||
logging.debug(_("App Config: %(api)s\n%(config)r") % locals())
|
||||
logging.info(_("Running %s API"), api)
|
||||
logging.debug(_('App Config: %(api)s\n%(config)r') % locals())
|
||||
logging.info(_('Running %s API'), api)
|
||||
app = wsgi.load_paste_app(paste_config_file, api)
|
||||
apps.append((app, getattr(FLAGS, "%s_listen_port" % api),
|
||||
getattr(FLAGS, "%s_listen" % api)))
|
||||
apps.append((app, getattr(FLAGS, '%s_listen_port' % api),
|
||||
getattr(FLAGS, '%s_listen' % api)))
|
||||
if len(apps) == 0:
|
||||
logging.error(_("No known API applications configured in %s."),
|
||||
logging.error(_('No known API applications configured in %s.'),
|
||||
paste_config_file)
|
||||
return
|
||||
|
||||
|
50
nova/test.py
50
nova/test.py
@ -16,12 +16,12 @@
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""
|
||||
Base classes for our unit tests.
|
||||
Allows overriding of flags for use of fakes,
|
||||
and some black magic for inline callbacks.
|
||||
"""
|
||||
"""Base classes for our unit tests.
|
||||
|
||||
Allows overriding of flags for use of fakes, and some black magic for
|
||||
inline callbacks.
|
||||
|
||||
"""
|
||||
|
||||
import datetime
|
||||
import functools
|
||||
@ -52,9 +52,9 @@ flags.DEFINE_bool('fake_tests', True,
|
||||
|
||||
|
||||
def skip_if_fake(func):
|
||||
"""Decorator that skips a test if running in fake mode"""
|
||||
"""Decorator that skips a test if running in fake mode."""
|
||||
def _skipper(*args, **kw):
|
||||
"""Wrapped skipper function"""
|
||||
"""Wrapped skipper function."""
|
||||
if FLAGS.fake_tests:
|
||||
raise unittest.SkipTest('Test cannot be run in fake mode')
|
||||
else:
|
||||
@ -63,9 +63,10 @@ def skip_if_fake(func):
|
||||
|
||||
|
||||
class TestCase(unittest.TestCase):
|
||||
"""Test case base class for all unit tests"""
|
||||
"""Test case base class for all unit tests."""
|
||||
|
||||
def setUp(self):
|
||||
"""Run before each test method to initialize test environment"""
|
||||
"""Run before each test method to initialize test environment."""
|
||||
super(TestCase, self).setUp()
|
||||
# NOTE(vish): We need a better method for creating fixtures for tests
|
||||
# now that we have some required db setup for the system
|
||||
@ -86,8 +87,7 @@ class TestCase(unittest.TestCase):
|
||||
self._original_flags = FLAGS.FlagValuesDict()
|
||||
|
||||
def tearDown(self):
|
||||
"""Runs after each test method to finalize/tear down test
|
||||
environment."""
|
||||
"""Runs after each test method to tear down test environment."""
|
||||
try:
|
||||
self.mox.UnsetStubs()
|
||||
self.stubs.UnsetAll()
|
||||
@ -121,7 +121,7 @@ class TestCase(unittest.TestCase):
|
||||
pass
|
||||
|
||||
def flags(self, **kw):
|
||||
"""Override flag variables for a test"""
|
||||
"""Override flag variables for a test."""
|
||||
for k, v in kw.iteritems():
|
||||
if k in self.flag_overrides:
|
||||
self.reset_flags()
|
||||
@ -131,7 +131,11 @@ class TestCase(unittest.TestCase):
|
||||
setattr(FLAGS, k, v)
|
||||
|
||||
def reset_flags(self):
|
||||
"""Resets all flag variables for the test. Runs after each test"""
|
||||
"""Resets all flag variables for the test.
|
||||
|
||||
Runs after each test.
|
||||
|
||||
"""
|
||||
FLAGS.Reset()
|
||||
for k, v in self._original_flags.iteritems():
|
||||
setattr(FLAGS, k, v)
|
||||
@ -158,7 +162,6 @@ class TestCase(unittest.TestCase):
|
||||
|
||||
def _monkey_patch_wsgi(self):
|
||||
"""Allow us to kill servers spawned by wsgi.Server."""
|
||||
# TODO(termie): change these patterns to use functools
|
||||
self.original_start = wsgi.Server.start
|
||||
|
||||
@functools.wraps(self.original_start)
|
||||
@ -189,12 +192,13 @@ class TestCase(unittest.TestCase):
|
||||
If you don't care (or don't know) a given value, you can specify
|
||||
the string DONTCARE as the value. This will cause that dict-item
|
||||
to be skipped.
|
||||
|
||||
"""
|
||||
def raise_assertion(msg):
|
||||
d1str = str(d1)
|
||||
d2str = str(d2)
|
||||
base_msg = ("Dictionaries do not match. %(msg)s d1: %(d1str)s "
|
||||
"d2: %(d2str)s" % locals())
|
||||
base_msg = ('Dictionaries do not match. %(msg)s d1: %(d1str)s '
|
||||
'd2: %(d2str)s' % locals())
|
||||
raise AssertionError(base_msg)
|
||||
|
||||
d1keys = set(d1.keys())
|
||||
@ -202,8 +206,8 @@ class TestCase(unittest.TestCase):
|
||||
if d1keys != d2keys:
|
||||
d1only = d1keys - d2keys
|
||||
d2only = d2keys - d1keys
|
||||
raise_assertion("Keys in d1 and not d2: %(d1only)s. "
|
||||
"Keys in d2 and not d1: %(d2only)s" % locals())
|
||||
raise_assertion('Keys in d1 and not d2: %(d1only)s. '
|
||||
'Keys in d2 and not d1: %(d2only)s' % locals())
|
||||
|
||||
for key in d1keys:
|
||||
d1value = d1[key]
|
||||
@ -217,19 +221,19 @@ class TestCase(unittest.TestCase):
|
||||
"d2['%(key)s']=%(d2value)s" % locals())
|
||||
|
||||
def assertDictListMatch(self, L1, L2):
|
||||
"""Assert a list of dicts are equivalent"""
|
||||
"""Assert a list of dicts are equivalent."""
|
||||
def raise_assertion(msg):
|
||||
L1str = str(L1)
|
||||
L2str = str(L2)
|
||||
base_msg = ("List of dictionaries do not match: %(msg)s "
|
||||
"L1: %(L1str)s L2: %(L2str)s" % locals())
|
||||
base_msg = ('List of dictionaries do not match: %(msg)s '
|
||||
'L1: %(L1str)s L2: %(L2str)s' % locals())
|
||||
raise AssertionError(base_msg)
|
||||
|
||||
L1count = len(L1)
|
||||
L2count = len(L2)
|
||||
if L1count != L2count:
|
||||
raise_assertion("Length mismatch: len(L1)=%(L1count)d != "
|
||||
"len(L2)=%(L2count)d" % locals())
|
||||
raise_assertion('Length mismatch: len(L1)=%(L1count)d != '
|
||||
'len(L2)=%(L2count)d' % locals())
|
||||
|
||||
for d1, d2 in zip(L1, L2):
|
||||
self.assertDictMatch(d1, d2)
|
||||
|
@ -618,7 +618,8 @@ class IptablesFirewallTestCase(test.TestCase):
|
||||
instance_ref = db.instance_create(self.context,
|
||||
{'user_id': 'fake',
|
||||
'project_id': 'fake',
|
||||
'mac_address': '56:12:12:12:12:12'})
|
||||
'mac_address': '56:12:12:12:12:12',
|
||||
'instance_type_id': 1})
|
||||
ip = '10.11.12.13'
|
||||
|
||||
network_ref = db.project_get_network(self.context,
|
||||
@ -841,7 +842,8 @@ class NWFilterTestCase(test.TestCase):
|
||||
instance_ref = db.instance_create(self.context,
|
||||
{'user_id': 'fake',
|
||||
'project_id': 'fake',
|
||||
'mac_address': '00:A0:C9:14:C8:29'})
|
||||
'mac_address': '00:A0:C9:14:C8:29',
|
||||
'instance_type_id': 1})
|
||||
inst_id = instance_ref['id']
|
||||
|
||||
ip = '10.11.12.13'
|
||||
|
147
nova/utils.py
147
nova/utils.py
@ -17,9 +17,7 @@
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""
|
||||
System-level utilities and helper functions.
|
||||
"""
|
||||
"""Utilities and helper functions."""
|
||||
|
||||
import base64
|
||||
import datetime
|
||||
@ -43,9 +41,8 @@ from eventlet import event
|
||||
from eventlet import greenthread
|
||||
from eventlet import semaphore
|
||||
from eventlet.green import subprocess
|
||||
None
|
||||
|
||||
from nova import exception
|
||||
from nova.exception import ProcessExecutionError
|
||||
from nova import flags
|
||||
from nova import log as logging
|
||||
|
||||
@ -56,7 +53,7 @@ FLAGS = flags.FLAGS
|
||||
|
||||
|
||||
def import_class(import_str):
|
||||
"""Returns a class from a string including module and class"""
|
||||
"""Returns a class from a string including module and class."""
|
||||
mod_str, _sep, class_str = import_str.rpartition('.')
|
||||
try:
|
||||
__import__(mod_str)
|
||||
@ -67,7 +64,7 @@ def import_class(import_str):
|
||||
|
||||
|
||||
def import_object(import_str):
|
||||
"""Returns an object including a module or module and class"""
|
||||
"""Returns an object including a module or module and class."""
|
||||
try:
|
||||
__import__(import_str)
|
||||
return sys.modules[import_str]
|
||||
@ -99,11 +96,12 @@ def vpn_ping(address, port, timeout=0.05, session_id=None):
|
||||
cli_id = 64 bit identifier
|
||||
? = unknown, probably flags/padding
|
||||
bit 9 was 1 and the rest were 0 in testing
|
||||
|
||||
"""
|
||||
if session_id is None:
|
||||
session_id = random.randint(0, 0xffffffffffffffff)
|
||||
sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
|
||||
data = struct.pack("!BQxxxxxx", 0x38, session_id)
|
||||
data = struct.pack('!BQxxxxxx', 0x38, session_id)
|
||||
sock.sendto(data, (address, port))
|
||||
sock.settimeout(timeout)
|
||||
try:
|
||||
@ -112,7 +110,7 @@ def vpn_ping(address, port, timeout=0.05, session_id=None):
|
||||
return False
|
||||
finally:
|
||||
sock.close()
|
||||
fmt = "!BQxxxxxQxxxx"
|
||||
fmt = '!BQxxxxxQxxxx'
|
||||
if len(received) != struct.calcsize(fmt):
|
||||
print struct.calcsize(fmt)
|
||||
return False
|
||||
@ -122,15 +120,8 @@ def vpn_ping(address, port, timeout=0.05, session_id=None):
|
||||
|
||||
|
||||
def fetchfile(url, target):
|
||||
LOG.debug(_("Fetching %s") % url)
|
||||
# c = pycurl.Curl()
|
||||
# fp = open(target, "wb")
|
||||
# c.setopt(c.URL, url)
|
||||
# c.setopt(c.WRITEDATA, fp)
|
||||
# c.perform()
|
||||
# c.close()
|
||||
# fp.close()
|
||||
execute("curl", "--fail", url, "-o", target)
|
||||
LOG.debug(_('Fetching %s') % url)
|
||||
execute('curl', '--fail', url, '-o', target)
|
||||
|
||||
|
||||
def execute(*cmd, **kwargs):
|
||||
@ -147,7 +138,7 @@ def execute(*cmd, **kwargs):
|
||||
while attempts > 0:
|
||||
attempts -= 1
|
||||
try:
|
||||
LOG.debug(_("Running cmd (subprocess): %s"), ' '.join(cmd))
|
||||
LOG.debug(_('Running cmd (subprocess): %s'), ' '.join(cmd))
|
||||
env = os.environ.copy()
|
||||
if addl_env:
|
||||
env.update(addl_env)
|
||||
@ -163,20 +154,21 @@ def execute(*cmd, **kwargs):
|
||||
result = obj.communicate()
|
||||
obj.stdin.close()
|
||||
if obj.returncode:
|
||||
LOG.debug(_("Result was %s") % obj.returncode)
|
||||
LOG.debug(_('Result was %s') % obj.returncode)
|
||||
if type(check_exit_code) == types.IntType \
|
||||
and obj.returncode != check_exit_code:
|
||||
(stdout, stderr) = result
|
||||
raise ProcessExecutionError(exit_code=obj.returncode,
|
||||
stdout=stdout,
|
||||
stderr=stderr,
|
||||
cmd=' '.join(cmd))
|
||||
raise exception.ProcessExecutionError(
|
||||
exit_code=obj.returncode,
|
||||
stdout=stdout,
|
||||
stderr=stderr,
|
||||
cmd=' '.join(cmd))
|
||||
return result
|
||||
except ProcessExecutionError:
|
||||
except exception.ProcessExecutionError:
|
||||
if not attempts:
|
||||
raise
|
||||
else:
|
||||
LOG.debug(_("%r failed. Retrying."), cmd)
|
||||
LOG.debug(_('%r failed. Retrying.'), cmd)
|
||||
if delay_on_retry:
|
||||
greenthread.sleep(random.randint(20, 200) / 100.0)
|
||||
finally:
|
||||
@ -188,13 +180,13 @@ def execute(*cmd, **kwargs):
|
||||
|
||||
def ssh_execute(ssh, cmd, process_input=None,
|
||||
addl_env=None, check_exit_code=True):
|
||||
LOG.debug(_("Running cmd (SSH): %s"), ' '.join(cmd))
|
||||
LOG.debug(_('Running cmd (SSH): %s'), ' '.join(cmd))
|
||||
if addl_env:
|
||||
raise exception.Error("Environment not supported over SSH")
|
||||
raise exception.Error(_('Environment not supported over SSH'))
|
||||
|
||||
if process_input:
|
||||
# This is (probably) fixable if we need it...
|
||||
raise exception.Error("process_input not supported over SSH")
|
||||
raise exception.Error(_('process_input not supported over SSH'))
|
||||
|
||||
stdin_stream, stdout_stream, stderr_stream = ssh.exec_command(cmd)
|
||||
channel = stdout_stream.channel
|
||||
@ -212,7 +204,7 @@ def ssh_execute(ssh, cmd, process_input=None,
|
||||
|
||||
# exit_status == -1 if no exit code was returned
|
||||
if exit_status != -1:
|
||||
LOG.debug(_("Result was %s") % exit_status)
|
||||
LOG.debug(_('Result was %s') % exit_status)
|
||||
if check_exit_code and exit_status != 0:
|
||||
raise exception.ProcessExecutionError(exit_code=exit_status,
|
||||
stdout=stdout,
|
||||
@ -251,7 +243,7 @@ def debug(arg):
|
||||
|
||||
|
||||
def runthis(prompt, *cmd, **kwargs):
|
||||
LOG.debug(_("Running %s"), (" ".join(cmd)))
|
||||
LOG.debug(_('Running %s'), (' '.join(cmd)))
|
||||
rv, err = execute(*cmd, **kwargs)
|
||||
|
||||
|
||||
@ -266,48 +258,49 @@ def generate_mac():
|
||||
random.randint(0x00, 0x7f),
|
||||
random.randint(0x00, 0xff),
|
||||
random.randint(0x00, 0xff)]
|
||||
return ':'.join(map(lambda x: "%02x" % x, mac))
|
||||
return ':'.join(map(lambda x: '%02x' % x, mac))
|
||||
|
||||
|
||||
# Default symbols to use for passwords. Avoids visually confusing characters.
|
||||
# ~6 bits per symbol
|
||||
DEFAULT_PASSWORD_SYMBOLS = ("23456789" # Removed: 0,1
|
||||
"ABCDEFGHJKLMNPQRSTUVWXYZ" # Removed: I, O
|
||||
"abcdefghijkmnopqrstuvwxyz") # Removed: l
|
||||
DEFAULT_PASSWORD_SYMBOLS = ('23456789' # Removed: 0,1
|
||||
'ABCDEFGHJKLMNPQRSTUVWXYZ' # Removed: I, O
|
||||
'abcdefghijkmnopqrstuvwxyz') # Removed: l
|
||||
|
||||
|
||||
# ~5 bits per symbol
|
||||
EASIER_PASSWORD_SYMBOLS = ("23456789" # Removed: 0, 1
|
||||
"ABCDEFGHJKLMNPQRSTUVWXYZ") # Removed: I, O
|
||||
EASIER_PASSWORD_SYMBOLS = ('23456789' # Removed: 0, 1
|
||||
'ABCDEFGHJKLMNPQRSTUVWXYZ') # Removed: I, O
|
||||
|
||||
|
||||
def generate_password(length=20, symbols=DEFAULT_PASSWORD_SYMBOLS):
|
||||
"""Generate a random password from the supplied symbols.
|
||||
|
||||
Believed to be reasonably secure (with a reasonable password length!)
|
||||
|
||||
"""
|
||||
r = random.SystemRandom()
|
||||
return "".join([r.choice(symbols) for _i in xrange(length)])
|
||||
return ''.join([r.choice(symbols) for _i in xrange(length)])
|
||||
|
||||
|
||||
def last_octet(address):
|
||||
return int(address.split(".")[-1])
|
||||
return int(address.split('.')[-1])
|
||||
|
||||
|
||||
def get_my_linklocal(interface):
|
||||
try:
|
||||
if_str = execute("ip", "-f", "inet6", "-o", "addr", "show", interface)
|
||||
condition = "\s+inet6\s+([0-9a-f:]+)/\d+\s+scope\s+link"
|
||||
if_str = execute('ip', '-f', 'inet6', '-o', 'addr', 'show', interface)
|
||||
condition = '\s+inet6\s+([0-9a-f:]+)/\d+\s+scope\s+link'
|
||||
links = [re.search(condition, x) for x in if_str[0].split('\n')]
|
||||
address = [w.group(1) for w in links if w is not None]
|
||||
if address[0] is not None:
|
||||
return address[0]
|
||||
else:
|
||||
raise exception.Error(_("Link Local address is not found.:%s")
|
||||
raise exception.Error(_('Link Local address is not found.:%s')
|
||||
% if_str)
|
||||
except Exception as ex:
|
||||
raise exception.Error(_("Couldn't get Link Local IP of %(interface)s"
|
||||
" :%(ex)s") % locals())
|
||||
" :%(ex)s") % locals())
|
||||
|
||||
|
||||
def to_global_ipv6(prefix, mac):
|
||||
@ -319,15 +312,15 @@ def to_global_ipv6(prefix, mac):
|
||||
return (mac64_addr ^ netaddr.IPAddress('::0200:0:0:0') | maskIP).\
|
||||
format()
|
||||
except TypeError:
|
||||
raise TypeError(_("Bad mac for to_global_ipv6: %s") % mac)
|
||||
raise TypeError(_('Bad mac for to_global_ipv6: %s') % mac)
|
||||
|
||||
|
||||
def to_mac(ipv6_address):
|
||||
address = netaddr.IPAddress(ipv6_address)
|
||||
mask1 = netaddr.IPAddress("::ffff:ffff:ffff:ffff")
|
||||
mask2 = netaddr.IPAddress("::0200:0:0:0")
|
||||
mask1 = netaddr.IPAddress('::ffff:ffff:ffff:ffff')
|
||||
mask2 = netaddr.IPAddress('::0200:0:0:0')
|
||||
mac64 = netaddr.EUI(int(address & mask1 ^ mask2)).words
|
||||
return ":".join(["%02x" % i for i in mac64[0:3] + mac64[5:8]])
|
||||
return ':'.join(['%02x' % i for i in mac64[0:3] + mac64[5:8]])
|
||||
|
||||
|
||||
def utcnow():
|
||||
@ -341,7 +334,7 @@ utcnow.override_time = None
|
||||
|
||||
|
||||
def is_older_than(before, seconds):
|
||||
"""Return True if before is older than seconds"""
|
||||
"""Return True if before is older than seconds."""
|
||||
return utcnow() - before > datetime.timedelta(seconds=seconds)
|
||||
|
||||
|
||||
@ -379,7 +372,7 @@ def isotime(at=None):
|
||||
|
||||
|
||||
def parse_isotime(timestr):
|
||||
"""Turn an iso formatted time back into a datetime"""
|
||||
"""Turn an iso formatted time back into a datetime."""
|
||||
return datetime.datetime.strptime(timestr, TIME_FORMAT)
|
||||
|
||||
|
||||
@ -433,16 +426,19 @@ class LazyPluggable(object):
|
||||
|
||||
|
||||
class LoopingCallDone(Exception):
|
||||
"""The poll-function passed to LoopingCall can raise this exception to
|
||||
"""Exception to break out and stop a LoopingCall.
|
||||
|
||||
The poll-function passed to LoopingCall can raise this exception to
|
||||
break out of the loop normally. This is somewhat analogous to
|
||||
StopIteration.
|
||||
|
||||
An optional return-value can be included as the argument to the exception;
|
||||
this return-value will be returned by LoopingCall.wait()
|
||||
|
||||
"""
|
||||
|
||||
def __init__(self, retvalue=True):
|
||||
""":param retvalue: Value that LoopingCall.wait() should return"""
|
||||
""":param retvalue: Value that LoopingCall.wait() should return."""
|
||||
self.retvalue = retvalue
|
||||
|
||||
|
||||
@ -493,7 +489,7 @@ def xhtml_escape(value):
|
||||
http://github.com/facebook/tornado/blob/master/tornado/escape.py
|
||||
|
||||
"""
|
||||
return saxutils.escape(value, {'"': """})
|
||||
return saxutils.escape(value, {'"': '"'})
|
||||
|
||||
|
||||
def utf8(value):
|
||||
@ -504,7 +500,7 @@ def utf8(value):
|
||||
|
||||
"""
|
||||
if isinstance(value, unicode):
|
||||
return value.encode("utf-8")
|
||||
return value.encode('utf-8')
|
||||
assert isinstance(value, str)
|
||||
return value
|
||||
|
||||
@ -554,7 +550,7 @@ class _NoopContextManager(object):
|
||||
|
||||
|
||||
def synchronized(name, external=False):
|
||||
"""Synchronization decorator
|
||||
"""Synchronization decorator.
|
||||
|
||||
Decorating a method like so:
|
||||
@synchronized('mylock')
|
||||
@ -578,6 +574,7 @@ def synchronized(name, external=False):
|
||||
multiple processes. This means that if two different workers both run a
|
||||
a method decorated with @synchronized('mylock', external=True), only one
|
||||
of them will execute at a time.
|
||||
|
||||
"""
|
||||
|
||||
def wrap(f):
|
||||
@ -590,13 +587,13 @@ def synchronized(name, external=False):
|
||||
_semaphores[name] = semaphore.Semaphore()
|
||||
sem = _semaphores[name]
|
||||
LOG.debug(_('Attempting to grab semaphore "%(lock)s" for method '
|
||||
'"%(method)s"...' % {"lock": name,
|
||||
"method": f.__name__}))
|
||||
'"%(method)s"...' % {'lock': name,
|
||||
'method': f.__name__}))
|
||||
with sem:
|
||||
if external:
|
||||
LOG.debug(_('Attempting to grab file lock "%(lock)s" for '
|
||||
'method "%(method)s"...' %
|
||||
{"lock": name, "method": f.__name__}))
|
||||
{'lock': name, 'method': f.__name__}))
|
||||
lock_file_path = os.path.join(FLAGS.lock_path,
|
||||
'nova-%s.lock' % name)
|
||||
lock = lockfile.FileLock(lock_file_path)
|
||||
@ -617,21 +614,23 @@ def synchronized(name, external=False):
|
||||
|
||||
|
||||
def get_from_path(items, path):
|
||||
""" Returns a list of items matching the specified path. Takes an
|
||||
XPath-like expression e.g. prop1/prop2/prop3, and for each item in items,
|
||||
looks up items[prop1][prop2][prop3]. Like XPath, if any of the
|
||||
"""Returns a list of items matching the specified path.
|
||||
|
||||
Takes an XPath-like expression e.g. prop1/prop2/prop3, and for each item
|
||||
in items, looks up items[prop1][prop2][prop3]. Like XPath, if any of the
|
||||
intermediate results are lists it will treat each list item individually.
|
||||
A 'None' in items or any child expressions will be ignored, this function
|
||||
will not throw because of None (anywhere) in items. The returned list
|
||||
will contain no None values."""
|
||||
will contain no None values.
|
||||
|
||||
"""
|
||||
if path is None:
|
||||
raise exception.Error("Invalid mini_xpath")
|
||||
raise exception.Error('Invalid mini_xpath')
|
||||
|
||||
(first_token, sep, remainder) = path.partition("/")
|
||||
(first_token, sep, remainder) = path.partition('/')
|
||||
|
||||
if first_token == "":
|
||||
raise exception.Error("Invalid mini_xpath")
|
||||
if first_token == '':
|
||||
raise exception.Error('Invalid mini_xpath')
|
||||
|
||||
results = []
|
||||
|
||||
@ -645,7 +644,7 @@ def get_from_path(items, path):
|
||||
for item in items:
|
||||
if item is None:
|
||||
continue
|
||||
get_method = getattr(item, "get", None)
|
||||
get_method = getattr(item, 'get', None)
|
||||
if get_method is None:
|
||||
continue
|
||||
child = get_method(first_token)
|
||||
@ -666,7 +665,7 @@ def get_from_path(items, path):
|
||||
|
||||
|
||||
def flatten_dict(dict_, flattened=None):
|
||||
"""Recursively flatten a nested dictionary"""
|
||||
"""Recursively flatten a nested dictionary."""
|
||||
flattened = flattened or {}
|
||||
for key, value in dict_.iteritems():
|
||||
if hasattr(value, 'iteritems'):
|
||||
@ -677,9 +676,7 @@ def flatten_dict(dict_, flattened=None):
|
||||
|
||||
|
||||
def partition_dict(dict_, keys):
|
||||
"""Return two dicts, one containing only `keys` the other containing
|
||||
everything but `keys`
|
||||
"""
|
||||
"""Return two dicts, one with `keys` the other with everything else."""
|
||||
intersection = {}
|
||||
difference = {}
|
||||
for key, value in dict_.iteritems():
|
||||
@ -691,9 +688,7 @@ def partition_dict(dict_, keys):
|
||||
|
||||
|
||||
def map_dict_keys(dict_, key_map):
|
||||
"""Return a dictionary in which the dictionaries keys are mapped to
|
||||
new keys.
|
||||
"""
|
||||
"""Return a dict in which the dictionaries keys are mapped to new keys."""
|
||||
mapped = {}
|
||||
for key, value in dict_.iteritems():
|
||||
mapped_key = key_map[key] if key in key_map else key
|
||||
@ -702,15 +697,15 @@ def map_dict_keys(dict_, key_map):
|
||||
|
||||
|
||||
def subset_dict(dict_, keys):
|
||||
"""Return a dict that only contains a subset of keys"""
|
||||
"""Return a dict that only contains a subset of keys."""
|
||||
subset = partition_dict(dict_, keys)[0]
|
||||
return subset
|
||||
|
||||
|
||||
def check_isinstance(obj, cls):
|
||||
"""Checks that obj is of type cls, and lets PyLint infer types"""
|
||||
"""Checks that obj is of type cls, and lets PyLint infer types."""
|
||||
if isinstance(obj, cls):
|
||||
return obj
|
||||
raise Exception(_("Expected object of type: %s") % (str(cls)))
|
||||
raise Exception(_('Expected object of type: %s') % (str(cls)))
|
||||
# TODO(justinsb): Can we make this better??
|
||||
return cls() # Ugly PyLint hack
|
||||
|
@ -21,9 +21,9 @@ except ImportError:
|
||||
'revision_id': 'LOCALREVISION',
|
||||
'revno': 0}
|
||||
|
||||
|
||||
NOVA_VERSION = ['2011', '3']
|
||||
YEAR, COUNT = NOVA_VERSION
|
||||
|
||||
FINAL = False # This becomes true at Release Candidate time
|
||||
|
||||
|
||||
@ -39,8 +39,8 @@ def version_string():
|
||||
|
||||
|
||||
def vcs_version_string():
|
||||
return "%s:%s" % (version_info['branch_nick'], version_info['revision_id'])
|
||||
return '%s:%s' % (version_info['branch_nick'], version_info['revision_id'])
|
||||
|
||||
|
||||
def version_string_with_vcs():
|
||||
return "%s-%s" % (canonical_version_string(), vcs_version_string())
|
||||
return '%s-%s' % (canonical_version_string(), vcs_version_string())
|
||||
|
@ -154,8 +154,8 @@ def _get_net_and_prefixlen(cidr):
|
||||
|
||||
|
||||
def _get_ip_version(cidr):
|
||||
net = IPy.IP(cidr)
|
||||
return int(net.version())
|
||||
net = IPy.IP(cidr)
|
||||
return int(net.version())
|
||||
|
||||
|
||||
def _get_network_info(instance):
|
||||
@ -165,9 +165,10 @@ def _get_network_info(instance):
|
||||
|
||||
ip_addresses = db.fixed_ip_get_all_by_instance(admin_context,
|
||||
instance['id'])
|
||||
|
||||
networks = db.network_get_all_by_instance(admin_context,
|
||||
instance['id'])
|
||||
flavor = db.instance_type_get_by_id(admin_context,
|
||||
instance['instance_type_id'])
|
||||
network_info = []
|
||||
|
||||
for network in networks:
|
||||
@ -191,7 +192,9 @@ def _get_network_info(instance):
|
||||
mapping = {
|
||||
'label': network['label'],
|
||||
'gateway': network['gateway'],
|
||||
'broadcast': network['broadcast'],
|
||||
'mac': instance['mac_address'],
|
||||
'rxtx_cap': flavor['rxtx_cap'],
|
||||
'dns': [network['dns']],
|
||||
'ips': [ip_dict(ip) for ip in network_ips]}
|
||||
|
||||
@ -359,28 +362,19 @@ class LibvirtConnection(driver.ComputeDriver):
|
||||
locals())
|
||||
raise
|
||||
|
||||
# We'll save this for when we do shutdown,
|
||||
# instead of destroy - but destroy returns immediately
|
||||
timer = utils.LoopingCall(f=None)
|
||||
def _wait_for_destroy():
|
||||
"""Called at an interval until the VM is gone."""
|
||||
instance_name = instance['name']
|
||||
|
||||
while True:
|
||||
try:
|
||||
state = self.get_info(instance['name'])['state']
|
||||
db.instance_set_state(context.get_admin_context(),
|
||||
instance['id'], state)
|
||||
if state == power_state.SHUTOFF:
|
||||
break
|
||||
state = self.get_info(instance_name)['state']
|
||||
except exception.NotFound:
|
||||
msg = _("Instance %s destroyed successfully.") % instance_name
|
||||
LOG.info(msg)
|
||||
raise utils.LoopingCallDone
|
||||
|
||||
# Let's not hammer on the DB
|
||||
time.sleep(1)
|
||||
except Exception as ex:
|
||||
msg = _("Error encountered when destroying instance '%(id)s': "
|
||||
"%(ex)s") % {"id": instance["id"], "ex": ex}
|
||||
LOG.debug(msg)
|
||||
db.instance_set_state(context.get_admin_context(),
|
||||
instance['id'],
|
||||
power_state.SHUTOFF)
|
||||
break
|
||||
timer = utils.LoopingCall(_wait_for_destroy)
|
||||
timer.start(interval=0.5, now=True)
|
||||
|
||||
self.firewall_driver.unfilter_instance(instance)
|
||||
|
||||
@ -522,6 +516,12 @@ class LibvirtConnection(driver.ComputeDriver):
|
||||
|
||||
@exception.wrap_exception
|
||||
def reboot(self, instance):
|
||||
"""Reboot a virtual machine, given an instance reference.
|
||||
|
||||
This method actually destroys and re-creates the domain to ensure the
|
||||
reboot happens, as the guest OS cannot ignore this action.
|
||||
|
||||
"""
|
||||
self.destroy(instance, False)
|
||||
xml = self.to_xml(instance)
|
||||
self.firewall_driver.setup_basic_filtering(instance)
|
||||
@ -529,24 +529,23 @@ class LibvirtConnection(driver.ComputeDriver):
|
||||
self._create_new_domain(xml)
|
||||
self.firewall_driver.apply_instance_filter(instance)
|
||||
|
||||
timer = utils.LoopingCall(f=None)
|
||||
|
||||
def _wait_for_reboot():
|
||||
try:
|
||||
state = self.get_info(instance['name'])['state']
|
||||
db.instance_set_state(context.get_admin_context(),
|
||||
instance['id'], state)
|
||||
if state == power_state.RUNNING:
|
||||
LOG.debug(_('instance %s: rebooted'), instance['name'])
|
||||
timer.stop()
|
||||
except Exception, exn:
|
||||
LOG.exception(_('_wait_for_reboot failed: %s'), exn)
|
||||
db.instance_set_state(context.get_admin_context(),
|
||||
instance['id'],
|
||||
power_state.SHUTDOWN)
|
||||
timer.stop()
|
||||
"""Called at an interval until the VM is running again."""
|
||||
instance_name = instance['name']
|
||||
|
||||
timer.f = _wait_for_reboot
|
||||
try:
|
||||
state = self.get_info(instance_name)['state']
|
||||
except exception.NotFound:
|
||||
msg = _("During reboot, %s disappeared.") % instance_name
|
||||
LOG.error(msg)
|
||||
raise utils.LoopingCallDone
|
||||
|
||||
if state == power_state.RUNNING:
|
||||
msg = _("Instance %s rebooted successfully.") % instance_name
|
||||
LOG.info(msg)
|
||||
raise utils.LoopingCallDone
|
||||
|
||||
timer = utils.LoopingCall(_wait_for_reboot)
|
||||
return timer.start(interval=0.5, now=True)
|
||||
|
||||
@exception.wrap_exception
|
||||
@ -566,7 +565,15 @@ class LibvirtConnection(driver.ComputeDriver):
|
||||
raise exception.ApiError("resume not supported for libvirt")
|
||||
|
||||
@exception.wrap_exception
|
||||
def rescue(self, instance, callback=None):
|
||||
def rescue(self, instance):
|
||||
"""Loads a VM using rescue images.
|
||||
|
||||
A rescue is normally performed when something goes wrong with the
|
||||
primary images and data needs to be corrected/recovered. Rescuing
|
||||
should not edit or over-ride the original image, only allow for
|
||||
data recovery.
|
||||
|
||||
"""
|
||||
self.destroy(instance, False)
|
||||
|
||||
xml = self.to_xml(instance, rescue=True)
|
||||
@ -576,29 +583,33 @@ class LibvirtConnection(driver.ComputeDriver):
|
||||
self._create_image(instance, xml, '.rescue', rescue_images)
|
||||
self._create_new_domain(xml)
|
||||
|
||||
timer = utils.LoopingCall(f=None)
|
||||
|
||||
def _wait_for_rescue():
|
||||
try:
|
||||
state = self.get_info(instance['name'])['state']
|
||||
db.instance_set_state(None, instance['id'], state)
|
||||
if state == power_state.RUNNING:
|
||||
LOG.debug(_('instance %s: rescued'), instance['name'])
|
||||
timer.stop()
|
||||
except Exception, exn:
|
||||
LOG.exception(_('_wait_for_rescue failed: %s'), exn)
|
||||
db.instance_set_state(None,
|
||||
instance['id'],
|
||||
power_state.SHUTDOWN)
|
||||
timer.stop()
|
||||
"""Called at an interval until the VM is running again."""
|
||||
instance_name = instance['name']
|
||||
|
||||
timer.f = _wait_for_rescue
|
||||
try:
|
||||
state = self.get_info(instance_name)['state']
|
||||
except exception.NotFound:
|
||||
msg = _("During reboot, %s disappeared.") % instance_name
|
||||
LOG.error(msg)
|
||||
raise utils.LoopingCallDone
|
||||
|
||||
if state == power_state.RUNNING:
|
||||
msg = _("Instance %s rescued successfully.") % instance_name
|
||||
LOG.info(msg)
|
||||
raise utils.LoopingCallDone
|
||||
|
||||
timer = utils.LoopingCall(_wait_for_rescue)
|
||||
return timer.start(interval=0.5, now=True)
|
||||
|
||||
@exception.wrap_exception
|
||||
def unrescue(self, instance, callback=None):
|
||||
# NOTE(vish): Because reboot destroys and recreates an instance using
|
||||
# the normal xml file, we can just call reboot here
|
||||
def unrescue(self, instance):
|
||||
"""Reboot the VM which is being rescued back into primary images.
|
||||
|
||||
Because reboot destroys and re-creates instances, unresue should
|
||||
simply call reboot.
|
||||
|
||||
"""
|
||||
self.reboot(instance)
|
||||
|
||||
@exception.wrap_exception
|
||||
@ -610,10 +621,6 @@ class LibvirtConnection(driver.ComputeDriver):
|
||||
@exception.wrap_exception
|
||||
def spawn(self, instance, network_info=None):
|
||||
xml = self.to_xml(instance, False, network_info)
|
||||
db.instance_set_state(context.get_admin_context(),
|
||||
instance['id'],
|
||||
power_state.NOSTATE,
|
||||
'launching')
|
||||
self.firewall_driver.setup_basic_filtering(instance, network_info)
|
||||
self.firewall_driver.prepare_instance_filter(instance, network_info)
|
||||
self._create_image(instance, xml, network_info)
|
||||
@ -626,25 +633,23 @@ class LibvirtConnection(driver.ComputeDriver):
|
||||
instance['name'])
|
||||
domain.setAutostart(1)
|
||||
|
||||
timer = utils.LoopingCall(f=None)
|
||||
|
||||
def _wait_for_boot():
|
||||
try:
|
||||
state = self.get_info(instance['name'])['state']
|
||||
db.instance_set_state(context.get_admin_context(),
|
||||
instance['id'], state)
|
||||
if state == power_state.RUNNING:
|
||||
LOG.debug(_('instance %s: booted'), instance['name'])
|
||||
timer.stop()
|
||||
except:
|
||||
LOG.exception(_('instance %s: failed to boot'),
|
||||
instance['name'])
|
||||
db.instance_set_state(context.get_admin_context(),
|
||||
instance['id'],
|
||||
power_state.SHUTDOWN)
|
||||
timer.stop()
|
||||
"""Called at an interval until the VM is running."""
|
||||
instance_name = instance['name']
|
||||
|
||||
timer.f = _wait_for_boot
|
||||
try:
|
||||
state = self.get_info(instance_name)['state']
|
||||
except exception.NotFound:
|
||||
msg = _("During reboot, %s disappeared.") % instance_name
|
||||
LOG.error(msg)
|
||||
raise utils.LoopingCallDone
|
||||
|
||||
if state == power_state.RUNNING:
|
||||
msg = _("Instance %s spawned successfully.") % instance_name
|
||||
LOG.info(msg)
|
||||
raise utils.LoopingCallDone
|
||||
|
||||
timer = utils.LoopingCall(_wait_for_boot)
|
||||
return timer.start(interval=0.5, now=True)
|
||||
|
||||
def _flush_xen_console(self, virsh_output):
|
||||
@ -1045,21 +1050,24 @@ class LibvirtConnection(driver.ComputeDriver):
|
||||
return xml
|
||||
|
||||
def get_info(self, instance_name):
|
||||
# NOTE(justinsb): When libvirt isn't running / can't connect, we get:
|
||||
# libvir: Remote error : unable to connect to
|
||||
# '/var/run/libvirt/libvirt-sock', libvirtd may need to be started:
|
||||
# No such file or directory
|
||||
"""Retrieve information from libvirt for a specific instance name.
|
||||
|
||||
If a libvirt error is encountered during lookup, we might raise a
|
||||
NotFound exception or Error exception depending on how severe the
|
||||
libvirt error is.
|
||||
|
||||
"""
|
||||
try:
|
||||
virt_dom = self._conn.lookupByName(instance_name)
|
||||
except libvirt.libvirtError as e:
|
||||
errcode = e.get_error_code()
|
||||
if errcode == libvirt.VIR_ERR_NO_DOMAIN:
|
||||
raise exception.NotFound(_("Instance %s not found")
|
||||
% instance_name)
|
||||
LOG.warning(_("Error from libvirt during lookup. "
|
||||
"Code=%(errcode)s Error=%(e)s") %
|
||||
locals())
|
||||
raise
|
||||
except libvirt.libvirtError as ex:
|
||||
error_code = ex.get_error_code()
|
||||
if error_code == libvirt.VIR_ERR_NO_DOMAIN:
|
||||
msg = _("Instance %s not found") % instance_name
|
||||
raise exception.NotFound(msg)
|
||||
|
||||
msg = _("Error from libvirt while looking up %(instance_name)s: "
|
||||
"[Error Code %(error_code)s] %(ex)s") % locals()
|
||||
raise exception.Error(msg)
|
||||
|
||||
(state, max_mem, mem, num_cpu, cpu_time) = virt_dom.info()
|
||||
return {'state': state,
|
||||
|
139
nova/wsgi.py
139
nova/wsgi.py
@ -17,9 +17,7 @@
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""
|
||||
Utility methods for working with WSGI servers
|
||||
"""
|
||||
"""Utility methods for working with WSGI servers."""
|
||||
|
||||
import os
|
||||
import sys
|
||||
@ -33,7 +31,6 @@ import routes.middleware
|
||||
import webob
|
||||
import webob.dec
|
||||
import webob.exc
|
||||
|
||||
from paste import deploy
|
||||
|
||||
from nova import exception
|
||||
@ -66,7 +63,7 @@ class Server(object):
|
||||
def start(self, application, port, host='0.0.0.0', backlog=128):
|
||||
"""Run a WSGI server with the given application."""
|
||||
arg0 = sys.argv[0]
|
||||
logging.audit(_("Starting %(arg0)s on %(host)s:%(port)s") % locals())
|
||||
logging.audit(_('Starting %(arg0)s on %(host)s:%(port)s') % locals())
|
||||
socket = eventlet.listen((host, port), backlog=backlog)
|
||||
self.pool.spawn_n(self._run, application, socket)
|
||||
|
||||
@ -87,22 +84,22 @@ class Server(object):
|
||||
class Request(webob.Request):
|
||||
|
||||
def best_match_content_type(self):
|
||||
"""
|
||||
Determine the most acceptable content-type based on the
|
||||
query extension then the Accept header
|
||||
"""
|
||||
"""Determine the most acceptable content-type.
|
||||
|
||||
parts = self.path.rsplit(".", 1)
|
||||
Based on the query extension then the Accept header.
|
||||
|
||||
"""
|
||||
parts = self.path.rsplit('.', 1)
|
||||
|
||||
if len(parts) > 1:
|
||||
format = parts[1]
|
||||
if format in ["json", "xml"]:
|
||||
return "application/{0}".format(parts[1])
|
||||
if format in ['json', 'xml']:
|
||||
return 'application/{0}'.format(parts[1])
|
||||
|
||||
ctypes = ["application/json", "application/xml"]
|
||||
ctypes = ['application/json', 'application/xml']
|
||||
bm = self.accept.best_match(ctypes)
|
||||
|
||||
return bm or "application/json"
|
||||
return bm or 'application/json'
|
||||
|
||||
def get_content_type(self):
|
||||
allowed_types = ("application/xml", "application/json")
|
||||
@ -122,7 +119,7 @@ class Application(object):
|
||||
|
||||
@classmethod
|
||||
def factory(cls, global_config, **local_config):
|
||||
"""Used for paste app factories in paste.deploy config fles.
|
||||
"""Used for paste app factories in paste.deploy config files.
|
||||
|
||||
Any local configuration (that is, values under the [app:APPNAME]
|
||||
section of the paste config) will be passed into the `__init__` method
|
||||
@ -177,8 +174,9 @@ class Application(object):
|
||||
|
||||
See the end of http://pythonpaste.org/webob/modules/dec.html
|
||||
for more info.
|
||||
|
||||
"""
|
||||
raise NotImplementedError(_("You must implement __call__"))
|
||||
raise NotImplementedError(_('You must implement __call__'))
|
||||
|
||||
|
||||
class Middleware(Application):
|
||||
@ -188,11 +186,12 @@ class Middleware(Application):
|
||||
initialized that will be called next. By default the middleware will
|
||||
simply call its wrapped app, or you can override __call__ to customize its
|
||||
behavior.
|
||||
|
||||
"""
|
||||
|
||||
@classmethod
|
||||
def factory(cls, global_config, **local_config):
|
||||
"""Used for paste app factories in paste.deploy config fles.
|
||||
"""Used for paste app factories in paste.deploy config files.
|
||||
|
||||
Any local configuration (that is, values under the [filter:APPNAME]
|
||||
section of the paste config) will be passed into the `__init__` method
|
||||
@ -244,20 +243,24 @@ class Middleware(Application):
|
||||
|
||||
|
||||
class Debug(Middleware):
|
||||
"""Helper class that can be inserted into any WSGI application chain
|
||||
to get information about the request and response."""
|
||||
"""Helper class for debugging a WSGI application.
|
||||
|
||||
Can be inserted into any WSGI application chain to get information
|
||||
about the request and response.
|
||||
|
||||
"""
|
||||
|
||||
@webob.dec.wsgify(RequestClass=Request)
|
||||
def __call__(self, req):
|
||||
print ("*" * 40) + " REQUEST ENVIRON"
|
||||
print ('*' * 40) + ' REQUEST ENVIRON'
|
||||
for key, value in req.environ.items():
|
||||
print key, "=", value
|
||||
print key, '=', value
|
||||
print
|
||||
resp = req.get_response(self.application)
|
||||
|
||||
print ("*" * 40) + " RESPONSE HEADERS"
|
||||
print ('*' * 40) + ' RESPONSE HEADERS'
|
||||
for (key, value) in resp.headers.iteritems():
|
||||
print key, "=", value
|
||||
print key, '=', value
|
||||
print
|
||||
|
||||
resp.app_iter = self.print_generator(resp.app_iter)
|
||||
@ -266,11 +269,8 @@ class Debug(Middleware):
|
||||
|
||||
@staticmethod
|
||||
def print_generator(app_iter):
|
||||
"""
|
||||
Iterator that prints the contents of a wrapper string iterator
|
||||
when iterated.
|
||||
"""
|
||||
print ("*" * 40) + " BODY"
|
||||
"""Iterator that prints the contents of a wrapper string."""
|
||||
print ('*' * 40) + ' BODY'
|
||||
for part in app_iter:
|
||||
sys.stdout.write(part)
|
||||
sys.stdout.flush()
|
||||
@ -279,13 +279,10 @@ class Debug(Middleware):
|
||||
|
||||
|
||||
class Router(object):
|
||||
"""
|
||||
WSGI middleware that maps incoming requests to WSGI apps.
|
||||
"""
|
||||
"""WSGI middleware that maps incoming requests to WSGI apps."""
|
||||
|
||||
def __init__(self, mapper):
|
||||
"""
|
||||
Create a router for the given routes.Mapper.
|
||||
"""Create a router for the given routes.Mapper.
|
||||
|
||||
Each route in `mapper` must specify a 'controller', which is a
|
||||
WSGI app to call. You'll probably want to specify an 'action' as
|
||||
@ -297,15 +294,16 @@ class Router(object):
|
||||
sc = ServerController()
|
||||
|
||||
# Explicit mapping of one route to a controller+action
|
||||
mapper.connect(None, "/svrlist", controller=sc, action="list")
|
||||
mapper.connect(None, '/svrlist', controller=sc, action='list')
|
||||
|
||||
# Actions are all implicitly defined
|
||||
mapper.resource("server", "servers", controller=sc)
|
||||
mapper.resource('server', 'servers', controller=sc)
|
||||
|
||||
# Pointing to an arbitrary WSGI app. You can specify the
|
||||
# {path_info:.*} parameter so the target app can be handed just that
|
||||
# section of the URL.
|
||||
mapper.connect(None, "/v1.0/{path_info:.*}", controller=BlogApp())
|
||||
mapper.connect(None, '/v1.0/{path_info:.*}', controller=BlogApp())
|
||||
|
||||
"""
|
||||
self.map = mapper
|
||||
self._router = routes.middleware.RoutesMiddleware(self._dispatch,
|
||||
@ -313,19 +311,22 @@ class Router(object):
|
||||
|
||||
@webob.dec.wsgify(RequestClass=Request)
|
||||
def __call__(self, req):
|
||||
"""
|
||||
Route the incoming request to a controller based on self.map.
|
||||
"""Route the incoming request to a controller based on self.map.
|
||||
|
||||
If no match, return a 404.
|
||||
|
||||
"""
|
||||
return self._router
|
||||
|
||||
@staticmethod
|
||||
@webob.dec.wsgify(RequestClass=Request)
|
||||
def _dispatch(req):
|
||||
"""
|
||||
"""Dispatch the request to the appropriate controller.
|
||||
|
||||
Called by self._router after matching the incoming request to a route
|
||||
and putting the information into req.environ. Either returns 404
|
||||
or the routed WSGI app's response.
|
||||
|
||||
"""
|
||||
match = req.environ['wsgiorg.routing_args'][1]
|
||||
if not match:
|
||||
@ -335,19 +336,19 @@ class Router(object):
|
||||
|
||||
|
||||
class Controller(object):
|
||||
"""
|
||||
"""WSGI app that dispatched to methods.
|
||||
|
||||
WSGI app that reads routing information supplied by RoutesMiddleware
|
||||
and calls the requested action method upon itself. All action methods
|
||||
must, in addition to their normal parameters, accept a 'req' argument
|
||||
which is the incoming wsgi.Request. They raise a webob.exc exception,
|
||||
or return a dict which will be serialized by requested content type.
|
||||
|
||||
"""
|
||||
|
||||
@webob.dec.wsgify(RequestClass=Request)
|
||||
def __call__(self, req):
|
||||
"""
|
||||
Call the method specified in req.environ by RoutesMiddleware.
|
||||
"""
|
||||
"""Call the method specified in req.environ by RoutesMiddleware."""
|
||||
arg_dict = req.environ['wsgiorg.routing_args'][1]
|
||||
action = arg_dict['action']
|
||||
method = getattr(self, action)
|
||||
@ -365,7 +366,7 @@ class Controller(object):
|
||||
body = self._serialize(result, content_type, default_xmlns)
|
||||
|
||||
response = webob.Response()
|
||||
response.headers["Content-Type"] = content_type
|
||||
response.headers['Content-Type'] = content_type
|
||||
response.body = body
|
||||
msg_dict = dict(url=req.url, status=response.status_int)
|
||||
msg = _("%(url)s returned with HTTP %(status)d") % msg_dict
|
||||
@ -375,12 +376,13 @@ class Controller(object):
|
||||
return result
|
||||
|
||||
def _serialize(self, data, content_type, default_xmlns):
|
||||
"""
|
||||
Serialize the given dict to the provided content_type.
|
||||
"""Serialize the given dict to the provided content_type.
|
||||
|
||||
Uses self._serialization_metadata if it exists, which is a dict mapping
|
||||
MIME types to information needed to serialize to that type.
|
||||
|
||||
"""
|
||||
_metadata = getattr(type(self), "_serialization_metadata", {})
|
||||
_metadata = getattr(type(self), '_serialization_metadata', {})
|
||||
|
||||
serializer = Serializer(_metadata, default_xmlns)
|
||||
try:
|
||||
@ -389,12 +391,13 @@ class Controller(object):
|
||||
raise webob.exc.HTTPNotAcceptable()
|
||||
|
||||
def _deserialize(self, data, content_type):
|
||||
"""
|
||||
Deserialize the request body to the specefied content type.
|
||||
"""Deserialize the request body to the specefied content type.
|
||||
|
||||
Uses self._serialization_metadata if it exists, which is a dict mapping
|
||||
MIME types to information needed to serialize to that type.
|
||||
|
||||
"""
|
||||
_metadata = getattr(type(self), "_serialization_metadata", {})
|
||||
_metadata = getattr(type(self), '_serialization_metadata', {})
|
||||
serializer = Serializer(_metadata)
|
||||
return serializer.deserialize(data, content_type)
|
||||
|
||||
@ -404,23 +407,22 @@ class Controller(object):
|
||||
|
||||
|
||||
class Serializer(object):
|
||||
"""
|
||||
Serializes and deserializes dictionaries to certain MIME types.
|
||||
"""
|
||||
"""Serializes and deserializes dictionaries to certain MIME types."""
|
||||
|
||||
def __init__(self, metadata=None, default_xmlns=None):
|
||||
"""
|
||||
Create a serializer based on the given WSGI environment.
|
||||
"""Create a serializer based on the given WSGI environment.
|
||||
|
||||
'metadata' is an optional dict mapping MIME types to information
|
||||
needed to serialize a dictionary to that type.
|
||||
|
||||
"""
|
||||
self.metadata = metadata or {}
|
||||
self.default_xmlns = default_xmlns
|
||||
|
||||
def _get_serialize_handler(self, content_type):
|
||||
handlers = {
|
||||
"application/json": self._to_json,
|
||||
"application/xml": self._to_xml,
|
||||
'application/json': self._to_json,
|
||||
'application/xml': self._to_xml,
|
||||
}
|
||||
|
||||
try:
|
||||
@ -429,29 +431,27 @@ class Serializer(object):
|
||||
raise exception.InvalidContentType()
|
||||
|
||||
def serialize(self, data, content_type):
|
||||
"""
|
||||
Serialize a dictionary into a string of the specified content type.
|
||||
"""
|
||||
"""Serialize a dictionary into the specified content type."""
|
||||
return self._get_serialize_handler(content_type)(data)
|
||||
|
||||
def deserialize(self, datastring, content_type):
|
||||
"""
|
||||
Deserialize a string to a dictionary.
|
||||
"""Deserialize a string to a dictionary.
|
||||
|
||||
The string must be in the format of a supported MIME type.
|
||||
|
||||
"""
|
||||
return self.get_deserialize_handler(content_type)(datastring)
|
||||
|
||||
def get_deserialize_handler(self, content_type):
|
||||
handlers = {
|
||||
"application/json": self._from_json,
|
||||
"application/xml": self._from_xml,
|
||||
'application/json': self._from_json,
|
||||
'application/xml': self._from_xml,
|
||||
}
|
||||
|
||||
try:
|
||||
return handlers[content_type]
|
||||
except Exception:
|
||||
raise exception.InvalidContentType(_("Invalid content type %s"
|
||||
raise exception.InvalidContentType(_('Invalid content type %s'
|
||||
% content_type))
|
||||
|
||||
def _from_json(self, datastring):
|
||||
@ -464,11 +464,11 @@ class Serializer(object):
|
||||
return {node.nodeName: self._from_xml_node(node, plurals)}
|
||||
|
||||
def _from_xml_node(self, node, listnames):
|
||||
"""
|
||||
Convert a minidom node to a simple Python type.
|
||||
"""Convert a minidom node to a simple Python type.
|
||||
|
||||
listnames is a collection of names of XML nodes whose subnodes should
|
||||
be considered list items.
|
||||
|
||||
"""
|
||||
if len(node.childNodes) == 1 and node.childNodes[0].nodeType == 3:
|
||||
return node.childNodes[0].nodeValue
|
||||
@ -575,7 +575,6 @@ def paste_config_file(basename):
|
||||
* /etc/nova, which may not be diffrerent from state_path on your distro
|
||||
|
||||
"""
|
||||
|
||||
configfiles = [basename,
|
||||
os.path.join(FLAGS.state_path, 'etc', 'nova', basename),
|
||||
os.path.join(FLAGS.state_path, 'etc', basename),
|
||||
@ -591,7 +590,7 @@ def load_paste_configuration(filename, appname):
|
||||
filename = os.path.abspath(filename)
|
||||
config = None
|
||||
try:
|
||||
config = deploy.appconfig("config:%s" % filename, name=appname)
|
||||
config = deploy.appconfig('config:%s' % filename, name=appname)
|
||||
except LookupError:
|
||||
pass
|
||||
return config
|
||||
@ -602,7 +601,7 @@ def load_paste_app(filename, appname):
|
||||
filename = os.path.abspath(filename)
|
||||
app = None
|
||||
try:
|
||||
app = deploy.loadapp("config:%s" % filename, name=appname)
|
||||
app = deploy.loadapp('config:%s' % filename, name=appname)
|
||||
except LookupError:
|
||||
pass
|
||||
return app
|
||||
|
Loading…
Reference in New Issue
Block a user