Merge "Use six.moves.range for Python 3"
This commit is contained in:
commit
2108a66572
doc/ext
nova
api/metadata
cells
compute
db/sqlalchemy
image
scheduler
tests
functional
unit
api
ec2
openstack
cells
compute
conductor
consoleauth
db
fake_ldap.pyfake_network.pynetwork
scheduler
test_quota.pyutils.pyvirt
virt
hyperv
libvirt
xenapi
@ -25,7 +25,7 @@ def process_todo_nodes(app, doctree, fromdocname):
|
|||||||
# remove the item that was added in the constructor, since I'm tired of
|
# remove the item that was added in the constructor, since I'm tired of
|
||||||
# reading through docutils for the proper way to construct an empty list
|
# reading through docutils for the proper way to construct an empty list
|
||||||
lists = []
|
lists = []
|
||||||
for i in xrange(5):
|
for i in range(5):
|
||||||
lists.append(nodes.bullet_list("", nodes.Text('', '')))
|
lists.append(nodes.bullet_list("", nodes.Text('', '')))
|
||||||
lists[i].remove(lists[i][0])
|
lists[i].remove(lists[i][0])
|
||||||
lists[i]['classes'].append('todo_list')
|
lists[i]['classes'].append('todo_list')
|
||||||
|
@ -13,6 +13,7 @@
|
|||||||
# License for the specific language governing permissions and limitations
|
# License for the specific language governing permissions and limitations
|
||||||
# under the License.
|
# under the License.
|
||||||
|
|
||||||
|
from six.moves import range
|
||||||
from webob import exc
|
from webob import exc
|
||||||
|
|
||||||
from nova import context
|
from nova import context
|
||||||
@ -42,7 +43,7 @@ def convert_password(context, password):
|
|||||||
"""
|
"""
|
||||||
password = password or ''
|
password = password or ''
|
||||||
meta = {}
|
meta = {}
|
||||||
for i in xrange(CHUNKS):
|
for i in range(CHUNKS):
|
||||||
meta['password_%d' % i] = password[:CHUNK_LENGTH]
|
meta['password_%d' % i] = password[:CHUNK_LENGTH]
|
||||||
password = password[CHUNK_LENGTH:]
|
password = password[CHUNK_LENGTH:]
|
||||||
return meta
|
return meta
|
||||||
|
@ -24,6 +24,7 @@ from oslo_log import log as logging
|
|||||||
import oslo_messaging
|
import oslo_messaging
|
||||||
from oslo_utils import importutils
|
from oslo_utils import importutils
|
||||||
from oslo_utils import timeutils
|
from oslo_utils import timeutils
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova.cells import messaging
|
from nova.cells import messaging
|
||||||
from nova.cells import state as cells_state
|
from nova.cells import state as cells_state
|
||||||
@ -169,7 +170,7 @@ class CellsManager(manager.Manager):
|
|||||||
|
|
||||||
rd_context = ctxt.elevated(read_deleted='yes')
|
rd_context = ctxt.elevated(read_deleted='yes')
|
||||||
|
|
||||||
for i in xrange(CONF.cells.instance_update_num_instances):
|
for i in range(CONF.cells.instance_update_num_instances):
|
||||||
while True:
|
while True:
|
||||||
# Yield to other greenthreads
|
# Yield to other greenthreads
|
||||||
time.sleep(0)
|
time.sleep(0)
|
||||||
|
@ -40,6 +40,7 @@ from oslo_utils import importutils
|
|||||||
from oslo_utils import timeutils
|
from oslo_utils import timeutils
|
||||||
from oslo_utils import uuidutils
|
from oslo_utils import uuidutils
|
||||||
import six
|
import six
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova.cells import state as cells_state
|
from nova.cells import state as cells_state
|
||||||
from nova.cells import utils as cells_utils
|
from nova.cells import utils as cells_utils
|
||||||
@ -228,7 +229,7 @@ class _BaseMessage(object):
|
|||||||
responses = []
|
responses = []
|
||||||
wait_time = CONF.cells.call_timeout
|
wait_time = CONF.cells.call_timeout
|
||||||
try:
|
try:
|
||||||
for x in xrange(num_responses):
|
for x in range(num_responses):
|
||||||
json_responses = self.resp_queue.get(timeout=wait_time)
|
json_responses = self.resp_queue.get(timeout=wait_time)
|
||||||
responses.extend(json_responses)
|
responses.extend(json_responses)
|
||||||
except queue.Empty:
|
except queue.Empty:
|
||||||
|
@ -21,6 +21,7 @@ import time
|
|||||||
|
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_log import log as logging
|
from oslo_log import log as logging
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova.cells import filters
|
from nova.cells import filters
|
||||||
from nova.cells import weights
|
from nova.cells import weights
|
||||||
@ -223,7 +224,7 @@ class CellsScheduler(base.Base):
|
|||||||
filter_properties, method, method_kwargs):
|
filter_properties, method, method_kwargs):
|
||||||
"""Pick a cell where we should create a new instance(s)."""
|
"""Pick a cell where we should create a new instance(s)."""
|
||||||
try:
|
try:
|
||||||
for i in xrange(max(0, CONF.cells.scheduler_retries) + 1):
|
for i in range(max(0, CONF.cells.scheduler_retries) + 1):
|
||||||
try:
|
try:
|
||||||
target_cells = self._grab_target_cells(filter_properties)
|
target_cells = self._grab_target_cells(filter_properties)
|
||||||
if target_cells is None:
|
if target_cells is None:
|
||||||
|
@ -35,6 +35,7 @@ from oslo_utils import timeutils
|
|||||||
from oslo_utils import units
|
from oslo_utils import units
|
||||||
from oslo_utils import uuidutils
|
from oslo_utils import uuidutils
|
||||||
import six
|
import six
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova import availability_zones
|
from nova import availability_zones
|
||||||
from nova import block_device
|
from nova import block_device
|
||||||
@ -910,7 +911,7 @@ class API(base.Base):
|
|||||||
LOG.debug("Going to run %s instances..." % num_instances)
|
LOG.debug("Going to run %s instances..." % num_instances)
|
||||||
instances = []
|
instances = []
|
||||||
try:
|
try:
|
||||||
for i in xrange(num_instances):
|
for i in range(num_instances):
|
||||||
instance = objects.Instance(context=context)
|
instance = objects.Instance(context=context)
|
||||||
instance.update(base_options)
|
instance.update(base_options)
|
||||||
instance = self.create_db_entry_for_new_instance(
|
instance = self.create_db_entry_for_new_instance(
|
||||||
|
@ -48,6 +48,7 @@ from oslo_utils import excutils
|
|||||||
from oslo_utils import strutils
|
from oslo_utils import strutils
|
||||||
from oslo_utils import timeutils
|
from oslo_utils import timeutils
|
||||||
import six
|
import six
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova import block_device
|
from nova import block_device
|
||||||
from nova.cells import rpcapi as cells_rpcapi
|
from nova.cells import rpcapi as cells_rpcapi
|
||||||
@ -3006,7 +3007,7 @@ class ComputeManager(manager.Manager):
|
|||||||
excess = len(images) - rotation
|
excess = len(images) - rotation
|
||||||
LOG.debug("Rotating out %d backups", excess,
|
LOG.debug("Rotating out %d backups", excess,
|
||||||
instance=instance)
|
instance=instance)
|
||||||
for i in xrange(excess):
|
for i in range(excess):
|
||||||
image = images.pop()
|
image = images.pop()
|
||||||
image_id = image['id']
|
image_id = image['id']
|
||||||
LOG.debug("Deleting image %s", image_id,
|
LOG.debug("Deleting image %s", image_id,
|
||||||
|
@ -36,6 +36,7 @@ from oslo_utils import excutils
|
|||||||
from oslo_utils import timeutils
|
from oslo_utils import timeutils
|
||||||
from oslo_utils import uuidutils
|
from oslo_utils import uuidutils
|
||||||
import six
|
import six
|
||||||
|
from six.moves import range
|
||||||
from sqlalchemy import and_
|
from sqlalchemy import and_
|
||||||
from sqlalchemy import Boolean
|
from sqlalchemy import Boolean
|
||||||
from sqlalchemy.exc import NoSuchTableError
|
from sqlalchemy.exc import NoSuchTableError
|
||||||
@ -4826,7 +4827,7 @@ def flavor_extra_specs_delete(context, flavor_id, key):
|
|||||||
@require_context
|
@require_context
|
||||||
def flavor_extra_specs_update_or_create(context, flavor_id, specs,
|
def flavor_extra_specs_update_or_create(context, flavor_id, specs,
|
||||||
max_retries=10):
|
max_retries=10):
|
||||||
for attempt in xrange(max_retries):
|
for attempt in range(max_retries):
|
||||||
try:
|
try:
|
||||||
session = get_session()
|
session = get_session()
|
||||||
with session.begin():
|
with session.begin():
|
||||||
@ -5522,7 +5523,7 @@ def aggregate_metadata_delete(context, aggregate_id, key):
|
|||||||
def aggregate_metadata_add(context, aggregate_id, metadata, set_delete=False,
|
def aggregate_metadata_add(context, aggregate_id, metadata, set_delete=False,
|
||||||
max_retries=10):
|
max_retries=10):
|
||||||
all_keys = metadata.keys()
|
all_keys = metadata.keys()
|
||||||
for attempt in xrange(max_retries):
|
for attempt in range(max_retries):
|
||||||
try:
|
try:
|
||||||
session = get_session()
|
session = get_session()
|
||||||
with session.begin():
|
with session.begin():
|
||||||
|
@ -31,6 +31,7 @@ from oslo_serialization import jsonutils
|
|||||||
from oslo_utils import netutils
|
from oslo_utils import netutils
|
||||||
from oslo_utils import timeutils
|
from oslo_utils import timeutils
|
||||||
import six
|
import six
|
||||||
|
from six.moves import range
|
||||||
import six.moves.urllib.parse as urlparse
|
import six.moves.urllib.parse as urlparse
|
||||||
|
|
||||||
from nova import exception
|
from nova import exception
|
||||||
@ -212,7 +213,7 @@ class GlanceClientWrapper(object):
|
|||||||
glanceclient.exc.CommunicationError)
|
glanceclient.exc.CommunicationError)
|
||||||
num_attempts = 1 + CONF.glance.num_retries
|
num_attempts = 1 + CONF.glance.num_retries
|
||||||
|
|
||||||
for attempt in xrange(1, num_attempts + 1):
|
for attempt in range(1, num_attempts + 1):
|
||||||
client = self.client or self._create_onetime_client(context,
|
client = self.client or self._create_onetime_client(context,
|
||||||
version)
|
version)
|
||||||
try:
|
try:
|
||||||
|
@ -23,6 +23,7 @@ import random
|
|||||||
|
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_log import log as logging
|
from oslo_log import log as logging
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova import exception
|
from nova import exception
|
||||||
from nova.i18n import _
|
from nova.i18n import _
|
||||||
@ -140,7 +141,7 @@ class FilterScheduler(driver.Scheduler):
|
|||||||
|
|
||||||
selected_hosts = []
|
selected_hosts = []
|
||||||
num_instances = request_spec.get('num_instances', 1)
|
num_instances = request_spec.get('num_instances', 1)
|
||||||
for num in xrange(num_instances):
|
for num in range(num_instances):
|
||||||
# Filter local hosts based on requirements ...
|
# Filter local hosts based on requirements ...
|
||||||
hosts = self.host_manager.get_filtered_hosts(hosts,
|
hosts = self.host_manager.get_filtered_hosts(hosts,
|
||||||
filter_properties, index=num)
|
filter_properties, index=num)
|
||||||
|
@ -47,7 +47,7 @@ class ServersTestBase(integrated_helpers._IntegratedTestBase):
|
|||||||
'conductor', manager='nova.conductor.manager.ConductorManager')
|
'conductor', manager='nova.conductor.manager.ConductorManager')
|
||||||
|
|
||||||
def _wait_for_state_change(self, server, from_status):
|
def _wait_for_state_change(self, server, from_status):
|
||||||
for i in xrange(0, 50):
|
for i in range(0, 50):
|
||||||
server = self.api.get_server(server['id'])
|
server = self.api.get_server(server['id'])
|
||||||
if server['status'] != from_status:
|
if server['status'] != from_status:
|
||||||
break
|
break
|
||||||
|
@ -14,6 +14,7 @@
|
|||||||
# under the License.
|
# under the License.
|
||||||
|
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova.cells import rpcapi as cells_rpcapi
|
from nova.cells import rpcapi as cells_rpcapi
|
||||||
from nova.cells import state
|
from nova.cells import state
|
||||||
@ -62,7 +63,7 @@ class CellsSampleJsonTest(api_sample_base.ApiSampleTestBaseV3):
|
|||||||
return cell
|
return cell
|
||||||
raise exception.CellNotFound(cell_name=cell_name)
|
raise exception.CellNotFound(cell_name=cell_name)
|
||||||
|
|
||||||
for x in xrange(num_cells):
|
for x in range(num_cells):
|
||||||
cell = models.Cell()
|
cell = models.Cell()
|
||||||
our_id = self.cells_next_id
|
our_id = self.cells_next_id
|
||||||
self.cells_next_id += 1
|
self.cells_next_id += 1
|
||||||
|
@ -19,6 +19,7 @@ import mock
|
|||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_utils import timeutils
|
from oslo_utils import timeutils
|
||||||
import requests
|
import requests
|
||||||
|
from six.moves import range
|
||||||
import webob
|
import webob
|
||||||
import webob.dec
|
import webob.dec
|
||||||
import webob.exc
|
import webob.exc
|
||||||
@ -53,7 +54,7 @@ class LockoutTestCase(test.NoDBTestCase):
|
|||||||
|
|
||||||
def _send_bad_attempts(self, access_key, num_attempts=1):
|
def _send_bad_attempts(self, access_key, num_attempts=1):
|
||||||
"""Fail x."""
|
"""Fail x."""
|
||||||
for i in xrange(num_attempts):
|
for i in range(num_attempts):
|
||||||
req = webob.Request.blank('/?AWSAccessKeyId=%s&die=1' % access_key)
|
req = webob.Request.blank('/?AWSAccessKeyId=%s&die=1' % access_key)
|
||||||
self.assertEqual(req.get_response(self.lockout).status_int, 403)
|
self.assertEqual(req.get_response(self.lockout).status_int, 403)
|
||||||
|
|
||||||
|
@ -17,6 +17,7 @@ import mock
|
|||||||
from mox3 import mox
|
from mox3 import mox
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_serialization import jsonutils
|
from oslo_serialization import jsonutils
|
||||||
|
from six.moves import range
|
||||||
from webob import exc
|
from webob import exc
|
||||||
|
|
||||||
from nova.api.openstack.compute import extensions
|
from nova.api.openstack.compute import extensions
|
||||||
@ -322,7 +323,7 @@ class BlockDeviceMappingTestV21(test.TestCase):
|
|||||||
self.stubs.Set(compute_api.API, '_validate_bdm', _validate_bdm)
|
self.stubs.Set(compute_api.API, '_validate_bdm', _validate_bdm)
|
||||||
self.stubs.Set(objects.Instance, 'destroy', _instance_destroy)
|
self.stubs.Set(objects.Instance, 'destroy', _instance_destroy)
|
||||||
|
|
||||||
for _unused in xrange(len(bdm_exceptions)):
|
for _unused in range(len(bdm_exceptions)):
|
||||||
params = {block_device_mapping.ATTRIBUTE_NAME:
|
params = {block_device_mapping.ATTRIBUTE_NAME:
|
||||||
[self.bdm[0].copy()]}
|
[self.bdm[0].copy()]}
|
||||||
self.assertRaises(exc.HTTPBadRequest,
|
self.assertRaises(exc.HTTPBadRequest,
|
||||||
|
@ -17,6 +17,7 @@ import datetime
|
|||||||
|
|
||||||
import mock
|
import mock
|
||||||
from oslo_utils import timeutils
|
from oslo_utils import timeutils
|
||||||
|
from six.moves import range
|
||||||
import webob
|
import webob
|
||||||
|
|
||||||
from nova.api.openstack.compute.contrib import simple_tenant_usage as \
|
from nova.api.openstack.compute.contrib import simple_tenant_usage as \
|
||||||
@ -99,7 +100,7 @@ def fake_instance_get_active_by_window_joined(context, begin, end,
|
|||||||
x,
|
x,
|
||||||
project_id if project_id else
|
project_id if project_id else
|
||||||
"faketenant_%s" % (x / SERVERS))
|
"faketenant_%s" % (x / SERVERS))
|
||||||
for x in xrange(TENANTS * SERVERS)]
|
for x in range(TENANTS * SERVERS)]
|
||||||
|
|
||||||
|
|
||||||
@mock.patch.object(db, 'instance_get_active_by_window_joined',
|
@mock.patch.object(db, 'instance_get_active_by_window_joined',
|
||||||
@ -126,7 +127,7 @@ class SimpleTenantUsageTestV21(test.TestCase):
|
|||||||
req.environ['nova.context'] = self.admin_context
|
req.environ['nova.context'] = self.admin_context
|
||||||
res_dict = self.controller.index(req)
|
res_dict = self.controller.index(req)
|
||||||
usages = res_dict['tenant_usages']
|
usages = res_dict['tenant_usages']
|
||||||
for i in xrange(TENANTS):
|
for i in range(TENANTS):
|
||||||
self.assertEqual(int(usages[i]['total_hours']),
|
self.assertEqual(int(usages[i]['total_hours']),
|
||||||
SERVERS * HOURS)
|
SERVERS * HOURS)
|
||||||
self.assertEqual(int(usages[i]['total_local_gb_usage']),
|
self.assertEqual(int(usages[i]['total_local_gb_usage']),
|
||||||
@ -178,20 +179,20 @@ class SimpleTenantUsageTestV21(test.TestCase):
|
|||||||
|
|
||||||
def test_verify_detailed_index(self):
|
def test_verify_detailed_index(self):
|
||||||
usages = self._get_tenant_usages('1')
|
usages = self._get_tenant_usages('1')
|
||||||
for i in xrange(TENANTS):
|
for i in range(TENANTS):
|
||||||
servers = usages[i]['server_usages']
|
servers = usages[i]['server_usages']
|
||||||
for j in xrange(SERVERS):
|
for j in range(SERVERS):
|
||||||
self.assertEqual(int(servers[j]['hours']), HOURS)
|
self.assertEqual(int(servers[j]['hours']), HOURS)
|
||||||
|
|
||||||
def test_verify_simple_index(self):
|
def test_verify_simple_index(self):
|
||||||
usages = self._get_tenant_usages(detailed='0')
|
usages = self._get_tenant_usages(detailed='0')
|
||||||
for i in xrange(TENANTS):
|
for i in range(TENANTS):
|
||||||
self.assertIsNone(usages[i].get('server_usages'))
|
self.assertIsNone(usages[i].get('server_usages'))
|
||||||
|
|
||||||
def test_verify_simple_index_empty_param(self):
|
def test_verify_simple_index_empty_param(self):
|
||||||
# NOTE(lzyeval): 'detailed=&start=..&end=..'
|
# NOTE(lzyeval): 'detailed=&start=..&end=..'
|
||||||
usages = self._get_tenant_usages()
|
usages = self._get_tenant_usages()
|
||||||
for i in xrange(TENANTS):
|
for i in range(TENANTS):
|
||||||
self.assertIsNone(usages[i].get('server_usages'))
|
self.assertIsNone(usages[i].get('server_usages'))
|
||||||
|
|
||||||
def _test_verify_show(self, start, stop):
|
def _test_verify_show(self, start, stop):
|
||||||
@ -206,8 +207,8 @@ class SimpleTenantUsageTestV21(test.TestCase):
|
|||||||
servers = usage['server_usages']
|
servers = usage['server_usages']
|
||||||
self.assertEqual(len(usage['server_usages']), TENANTS * SERVERS)
|
self.assertEqual(len(usage['server_usages']), TENANTS * SERVERS)
|
||||||
uuids = ['00000000-0000-0000-0000-00000000000000%02d' %
|
uuids = ['00000000-0000-0000-0000-00000000000000%02d' %
|
||||||
x for x in xrange(SERVERS)]
|
x for x in range(SERVERS)]
|
||||||
for j in xrange(SERVERS):
|
for j in range(SERVERS):
|
||||||
delta = STOP - START
|
delta = STOP - START
|
||||||
uptime = delta.days * 24 * 3600 + delta.seconds
|
uptime = delta.days * 24 * 3600 + delta.seconds
|
||||||
self.assertEqual(int(servers[j]['uptime']), uptime)
|
self.assertEqual(int(servers[j]['uptime']), uptime)
|
||||||
|
@ -27,6 +27,7 @@ from mox3 import mox
|
|||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_serialization import jsonutils
|
from oslo_serialization import jsonutils
|
||||||
from oslo_utils import timeutils
|
from oslo_utils import timeutils
|
||||||
|
from six.moves import range
|
||||||
import six.moves.urllib.parse as urlparse
|
import six.moves.urllib.parse as urlparse
|
||||||
import testtools
|
import testtools
|
||||||
import webob
|
import webob
|
||||||
@ -580,7 +581,7 @@ class ServersControllerTest(ControllerTest):
|
|||||||
|
|
||||||
servers = res_dict['servers']
|
servers = res_dict['servers']
|
||||||
self.assertEqual([s['id'] for s in servers],
|
self.assertEqual([s['id'] for s in servers],
|
||||||
[fakes.get_fake_uuid(i) for i in xrange(len(servers))])
|
[fakes.get_fake_uuid(i) for i in range(len(servers))])
|
||||||
|
|
||||||
servers_links = res_dict['servers_links']
|
servers_links = res_dict['servers_links']
|
||||||
self.assertEqual(servers_links[0]['rel'], 'next')
|
self.assertEqual(servers_links[0]['rel'], 'next')
|
||||||
@ -612,7 +613,7 @@ class ServersControllerTest(ControllerTest):
|
|||||||
|
|
||||||
servers = res['servers']
|
servers = res['servers']
|
||||||
self.assertEqual([s['id'] for s in servers],
|
self.assertEqual([s['id'] for s in servers],
|
||||||
[fakes.get_fake_uuid(i) for i in xrange(len(servers))])
|
[fakes.get_fake_uuid(i) for i in range(len(servers))])
|
||||||
|
|
||||||
servers_links = res['servers_links']
|
servers_links = res['servers_links']
|
||||||
self.assertEqual(servers_links[0]['rel'], 'next')
|
self.assertEqual(servers_links[0]['rel'], 'next')
|
||||||
@ -636,7 +637,7 @@ class ServersControllerTest(ControllerTest):
|
|||||||
|
|
||||||
servers = res['servers']
|
servers = res['servers']
|
||||||
self.assertEqual([s['id'] for s in servers],
|
self.assertEqual([s['id'] for s in servers],
|
||||||
[fakes.get_fake_uuid(i) for i in xrange(len(servers))])
|
[fakes.get_fake_uuid(i) for i in range(len(servers))])
|
||||||
|
|
||||||
servers_links = res['servers_links']
|
servers_links = res['servers_links']
|
||||||
self.assertEqual(servers_links[0]['rel'], 'next')
|
self.assertEqual(servers_links[0]['rel'], 'next')
|
||||||
@ -1283,7 +1284,7 @@ class ServersControllerTest(ControllerTest):
|
|||||||
def return_servers_with_host(context, *args, **kwargs):
|
def return_servers_with_host(context, *args, **kwargs):
|
||||||
return [fakes.stub_instance(i + 1, 'fake', 'fake', host=i % 2,
|
return [fakes.stub_instance(i + 1, 'fake', 'fake', host=i % 2,
|
||||||
uuid=fakes.get_fake_uuid(i))
|
uuid=fakes.get_fake_uuid(i))
|
||||||
for i in xrange(5)]
|
for i in range(5)]
|
||||||
|
|
||||||
self.stubs.Set(db, 'instance_get_all_by_filters_sort',
|
self.stubs.Set(db, 'instance_get_all_by_filters_sort',
|
||||||
return_servers_with_host)
|
return_servers_with_host)
|
||||||
|
@ -23,6 +23,7 @@ import StringIO
|
|||||||
import mock
|
import mock
|
||||||
from oslo_serialization import jsonutils
|
from oslo_serialization import jsonutils
|
||||||
import six
|
import six
|
||||||
|
from six.moves import range
|
||||||
import webob
|
import webob
|
||||||
|
|
||||||
from nova.api.openstack.compute import limits
|
from nova.api.openstack.compute import limits
|
||||||
@ -502,7 +503,7 @@ class LimiterTest(BaseLimitTestSuite):
|
|||||||
|
|
||||||
def _check(self, num, verb, url, username=None):
|
def _check(self, num, verb, url, username=None):
|
||||||
"""Check and yield results from checks."""
|
"""Check and yield results from checks."""
|
||||||
for x in xrange(num):
|
for x in range(num):
|
||||||
yield self.limiter.check_for_delay(verb, url, username)[0]
|
yield self.limiter.check_for_delay(verb, url, username)[0]
|
||||||
|
|
||||||
def _check_sum(self, num, verb, url, username=None):
|
def _check_sum(self, num, verb, url, username=None):
|
||||||
|
@ -27,6 +27,7 @@ import mock
|
|||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_serialization import jsonutils
|
from oslo_serialization import jsonutils
|
||||||
from oslo_utils import timeutils
|
from oslo_utils import timeutils
|
||||||
|
from six.moves import range
|
||||||
import six.moves.urllib.parse as urlparse
|
import six.moves.urllib.parse as urlparse
|
||||||
import testtools
|
import testtools
|
||||||
import webob
|
import webob
|
||||||
@ -542,7 +543,7 @@ class ServersControllerTest(ControllerTest):
|
|||||||
|
|
||||||
servers = res_dict['servers']
|
servers = res_dict['servers']
|
||||||
self.assertEqual([s['id'] for s in servers],
|
self.assertEqual([s['id'] for s in servers],
|
||||||
[fakes.get_fake_uuid(i) for i in xrange(len(servers))])
|
[fakes.get_fake_uuid(i) for i in range(len(servers))])
|
||||||
|
|
||||||
servers_links = res_dict['servers_links']
|
servers_links = res_dict['servers_links']
|
||||||
self.assertEqual(servers_links[0]['rel'], 'next')
|
self.assertEqual(servers_links[0]['rel'], 'next')
|
||||||
@ -574,7 +575,7 @@ class ServersControllerTest(ControllerTest):
|
|||||||
|
|
||||||
servers = res['servers']
|
servers = res['servers']
|
||||||
self.assertEqual([s['id'] for s in servers],
|
self.assertEqual([s['id'] for s in servers],
|
||||||
[fakes.get_fake_uuid(i) for i in xrange(len(servers))])
|
[fakes.get_fake_uuid(i) for i in range(len(servers))])
|
||||||
|
|
||||||
servers_links = res['servers_links']
|
servers_links = res['servers_links']
|
||||||
self.assertEqual(servers_links[0]['rel'], 'next')
|
self.assertEqual(servers_links[0]['rel'], 'next')
|
||||||
@ -598,7 +599,7 @@ class ServersControllerTest(ControllerTest):
|
|||||||
|
|
||||||
servers = res['servers']
|
servers = res['servers']
|
||||||
self.assertEqual([s['id'] for s in servers],
|
self.assertEqual([s['id'] for s in servers],
|
||||||
[fakes.get_fake_uuid(i) for i in xrange(len(servers))])
|
[fakes.get_fake_uuid(i) for i in range(len(servers))])
|
||||||
|
|
||||||
servers_links = res['servers_links']
|
servers_links = res['servers_links']
|
||||||
self.assertEqual(servers_links[0]['rel'], 'next')
|
self.assertEqual(servers_links[0]['rel'], 'next')
|
||||||
@ -1332,7 +1333,7 @@ class ServersControllerTest(ControllerTest):
|
|||||||
def return_servers_with_host(context, *args, **kwargs):
|
def return_servers_with_host(context, *args, **kwargs):
|
||||||
return [fakes.stub_instance(i + 1, 'fake', 'fake', host=i % 2,
|
return [fakes.stub_instance(i + 1, 'fake', 'fake', host=i % 2,
|
||||||
uuid=fakes.get_fake_uuid(i))
|
uuid=fakes.get_fake_uuid(i))
|
||||||
for i in xrange(5)]
|
for i in range(5)]
|
||||||
|
|
||||||
self.stubs.Set(db, 'instance_get_all_by_filters',
|
self.stubs.Set(db, 'instance_get_all_by_filters',
|
||||||
return_servers_with_host)
|
return_servers_with_host)
|
||||||
|
@ -21,6 +21,7 @@ from oslo_utils import netutils
|
|||||||
from oslo_utils import timeutils
|
from oslo_utils import timeutils
|
||||||
import routes
|
import routes
|
||||||
import six
|
import six
|
||||||
|
from six.moves import range
|
||||||
import webob
|
import webob
|
||||||
import webob.dec
|
import webob.dec
|
||||||
import webob.request
|
import webob.request
|
||||||
@ -410,7 +411,7 @@ def fake_instance_get_all_by_filters(num_servers=5, **kwargs):
|
|||||||
if 'sort_dirs' in kwargs:
|
if 'sort_dirs' in kwargs:
|
||||||
kwargs.pop('sort_dirs')
|
kwargs.pop('sort_dirs')
|
||||||
|
|
||||||
for i in xrange(num_servers):
|
for i in range(num_servers):
|
||||||
uuid = get_fake_uuid(i)
|
uuid = get_fake_uuid(i)
|
||||||
server = stub_instance(id=i + 1, uuid=uuid,
|
server = stub_instance(id=i + 1, uuid=uuid,
|
||||||
**kwargs)
|
**kwargs)
|
||||||
|
@ -56,7 +56,7 @@ class RequestTest(test.NoDBTestCase):
|
|||||||
def test_cache_and_retrieve_instances(self):
|
def test_cache_and_retrieve_instances(self):
|
||||||
request = wsgi.Request.blank('/foo')
|
request = wsgi.Request.blank('/foo')
|
||||||
instances = []
|
instances = []
|
||||||
for x in xrange(3):
|
for x in range(3):
|
||||||
instances.append({'uuid': 'uuid%s' % x})
|
instances.append({'uuid': 'uuid%s' % x})
|
||||||
# Store 2
|
# Store 2
|
||||||
request.cache_db_instances(instances[:2])
|
request.cache_db_instances(instances[:2])
|
||||||
@ -77,7 +77,7 @@ class RequestTest(test.NoDBTestCase):
|
|||||||
def test_cache_and_retrieve_compute_nodes(self):
|
def test_cache_and_retrieve_compute_nodes(self):
|
||||||
request = wsgi.Request.blank('/foo')
|
request = wsgi.Request.blank('/foo')
|
||||||
compute_nodes = []
|
compute_nodes = []
|
||||||
for x in xrange(3):
|
for x in range(3):
|
||||||
compute_nodes.append({'id': 'id%s' % x})
|
compute_nodes.append({'id': 'id%s' % x})
|
||||||
# Store 2
|
# Store 2
|
||||||
request.cache_db_compute_nodes(compute_nodes[:2])
|
request.cache_db_compute_nodes(compute_nodes[:2])
|
||||||
|
@ -21,6 +21,7 @@ import datetime
|
|||||||
import mock
|
import mock
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_utils import timeutils
|
from oslo_utils import timeutils
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova.cells import messaging
|
from nova.cells import messaging
|
||||||
from nova.cells import utils as cells_utils
|
from nova.cells import utils as cells_utils
|
||||||
@ -279,7 +280,7 @@ class CellsManagerClassTestCase(test.NoDBTestCase):
|
|||||||
expected_response = []
|
expected_response = []
|
||||||
# 3 cells... so 3 responses. Each response is a list of services.
|
# 3 cells... so 3 responses. Each response is a list of services.
|
||||||
# Manager should turn these into a single list of responses.
|
# Manager should turn these into a single list of responses.
|
||||||
for i in xrange(3):
|
for i in range(3):
|
||||||
cell_name = 'path!to!cell%i' % i
|
cell_name = 'path!to!cell%i' % i
|
||||||
services = []
|
services = []
|
||||||
for service in FAKE_SERVICES:
|
for service in FAKE_SERVICES:
|
||||||
@ -405,7 +406,7 @@ class CellsManagerClassTestCase(test.NoDBTestCase):
|
|||||||
# 3 cells... so 3 responses. Each response is a list of task log
|
# 3 cells... so 3 responses. Each response is a list of task log
|
||||||
# entries. Manager should turn these into a single list of
|
# entries. Manager should turn these into a single list of
|
||||||
# task log entries.
|
# task log entries.
|
||||||
for i in xrange(num):
|
for i in range(num):
|
||||||
cell_name = 'path!to!cell%i' % i
|
cell_name = 'path!to!cell%i' % i
|
||||||
task_logs = []
|
task_logs = []
|
||||||
for task_log in FAKE_TASK_LOGS:
|
for task_log in FAKE_TASK_LOGS:
|
||||||
@ -468,7 +469,7 @@ class CellsManagerClassTestCase(test.NoDBTestCase):
|
|||||||
expected_response = []
|
expected_response = []
|
||||||
# 3 cells... so 3 responses. Each response is a list of computes.
|
# 3 cells... so 3 responses. Each response is a list of computes.
|
||||||
# Manager should turn these into a single list of responses.
|
# Manager should turn these into a single list of responses.
|
||||||
for i in xrange(3):
|
for i in range(3):
|
||||||
cell_name = 'path!to!cell%i' % i
|
cell_name = 'path!to!cell%i' % i
|
||||||
compute_nodes = []
|
compute_nodes = []
|
||||||
for compute_node in FAKE_COMPUTE_NODES:
|
for compute_node in FAKE_COMPUTE_NODES:
|
||||||
|
@ -75,7 +75,7 @@ class CellsRPCDriverTestCase(test.NoDBTestCase):
|
|||||||
def stop(self):
|
def stop(self):
|
||||||
call_info['stopped'].append(self)
|
call_info['stopped'].append(self)
|
||||||
|
|
||||||
fake_servers = [FakeRPCServer() for x in xrange(5)]
|
fake_servers = [FakeRPCServer() for x in range(5)]
|
||||||
self.driver.rpc_servers = fake_servers
|
self.driver.rpc_servers = fake_servers
|
||||||
self.driver.stop_servers()
|
self.driver.stop_servers()
|
||||||
self.assertEqual(fake_servers, call_info['stopped'])
|
self.assertEqual(fake_servers, call_info['stopped'])
|
||||||
|
@ -79,7 +79,7 @@ class CellsSchedulerTestCase(test.TestCase):
|
|||||||
self.my_cell_state = self.state_manager.get_my_state()
|
self.my_cell_state = self.state_manager.get_my_state()
|
||||||
self.ctxt = context.RequestContext('fake', 'fake')
|
self.ctxt = context.RequestContext('fake', 'fake')
|
||||||
instance_uuids = []
|
instance_uuids = []
|
||||||
for x in xrange(3):
|
for x in range(3):
|
||||||
instance_uuids.append(uuidutils.generate_uuid())
|
instance_uuids.append(uuidutils.generate_uuid())
|
||||||
self.instance_uuids = instance_uuids
|
self.instance_uuids = instance_uuids
|
||||||
self.instances = [objects.Instance(uuid=uuid, id=id)
|
self.instances = [objects.Instance(uuid=uuid, id=id)
|
||||||
|
@ -6130,7 +6130,7 @@ class ComputeTestCase(BaseTestCase):
|
|||||||
|
|
||||||
instance_map = {}
|
instance_map = {}
|
||||||
instances = []
|
instances = []
|
||||||
for x in xrange(8):
|
for x in range(8):
|
||||||
inst_uuid = 'fake-uuid-%s' % x
|
inst_uuid = 'fake-uuid-%s' % x
|
||||||
instance_map[inst_uuid] = fake_instance.fake_db_instance(
|
instance_map[inst_uuid] = fake_instance.fake_db_instance(
|
||||||
uuid=inst_uuid, host=CONF.host, created_at=None)
|
uuid=inst_uuid, host=CONF.host, created_at=None)
|
||||||
@ -6408,7 +6408,7 @@ class ComputeTestCase(BaseTestCase):
|
|||||||
filters = {'vm_state': vm_states.BUILDING, 'host': CONF.host}
|
filters = {'vm_state': vm_states.BUILDING, 'host': CONF.host}
|
||||||
# these are the ones that are expired
|
# these are the ones that are expired
|
||||||
old_instances = []
|
old_instances = []
|
||||||
for x in xrange(4):
|
for x in range(4):
|
||||||
instance = {'uuid': str(uuid.uuid4()), 'created_at': created_at}
|
instance = {'uuid': str(uuid.uuid4()), 'created_at': created_at}
|
||||||
instance.update(filters)
|
instance.update(filters)
|
||||||
old_instances.append(fake_instance.fake_db_instance(**instance))
|
old_instances.append(fake_instance.fake_db_instance(**instance))
|
||||||
@ -7506,7 +7506,7 @@ class ComputeAPITestCase(BaseTestCase):
|
|||||||
self.assertEqual([], instance.security_groups.objects)
|
self.assertEqual([], instance.security_groups.objects)
|
||||||
|
|
||||||
def test_default_hostname_generator(self):
|
def test_default_hostname_generator(self):
|
||||||
fake_uuids = [str(uuid.uuid4()) for x in xrange(4)]
|
fake_uuids = [str(uuid.uuid4()) for x in range(4)]
|
||||||
|
|
||||||
orig_populate = self.compute_api._populate_instance_for_create
|
orig_populate = self.compute_api._populate_instance_for_create
|
||||||
|
|
||||||
|
@ -1083,7 +1083,7 @@ class ComputeManagerUnitTestCase(test.NoDBTestCase):
|
|||||||
|
|
||||||
def test_get_instances_on_driver(self):
|
def test_get_instances_on_driver(self):
|
||||||
driver_instances = []
|
driver_instances = []
|
||||||
for x in xrange(10):
|
for x in range(10):
|
||||||
driver_instances.append(fake_instance.fake_db_instance())
|
driver_instances.append(fake_instance.fake_db_instance())
|
||||||
|
|
||||||
self.mox.StubOutWithMock(self.compute.driver,
|
self.mox.StubOutWithMock(self.compute.driver,
|
||||||
@ -1128,7 +1128,7 @@ class ComputeManagerUnitTestCase(test.NoDBTestCase):
|
|||||||
|
|
||||||
all_instances = []
|
all_instances = []
|
||||||
driver_instances = []
|
driver_instances = []
|
||||||
for x in xrange(10):
|
for x in range(10):
|
||||||
instance = fake_instance.fake_db_instance(name='inst-%i' % x,
|
instance = fake_instance.fake_db_instance(name='inst-%i' % x,
|
||||||
id=x)
|
id=x)
|
||||||
if x % 2:
|
if x % 2:
|
||||||
|
@ -1013,7 +1013,7 @@ class _BaseTaskTestCase(object):
|
|||||||
instances = [objects.Instance(context=self.context,
|
instances = [objects.Instance(context=self.context,
|
||||||
id=i,
|
id=i,
|
||||||
uuid=uuid.uuid4(),
|
uuid=uuid.uuid4(),
|
||||||
flavor=instance_type) for i in xrange(2)]
|
flavor=instance_type) for i in range(2)]
|
||||||
instance_type_p = obj_base.obj_to_primitive(instance_type)
|
instance_type_p = obj_base.obj_to_primitive(instance_type)
|
||||||
instance_properties = instance_obj.compat_instance(instances[0])
|
instance_properties = instance_obj.compat_instance(instances[0])
|
||||||
|
|
||||||
@ -1095,7 +1095,7 @@ class _BaseTaskTestCase(object):
|
|||||||
|
|
||||||
def test_build_instances_scheduler_failure(self):
|
def test_build_instances_scheduler_failure(self):
|
||||||
instances = [fake_instance.fake_instance_obj(self.context)
|
instances = [fake_instance.fake_instance_obj(self.context)
|
||||||
for i in xrange(2)]
|
for i in range(2)]
|
||||||
image = {'fake-data': 'should_pass_silently'}
|
image = {'fake-data': 'should_pass_silently'}
|
||||||
spec = {'fake': 'specs',
|
spec = {'fake': 'specs',
|
||||||
'instance_properties': instances[0]}
|
'instance_properties': instances[0]}
|
||||||
@ -1928,7 +1928,7 @@ class ConductorTaskTestCase(_BaseTaskTestCase, test_compute.BaseTestCase):
|
|||||||
|
|
||||||
def test_build_instances_instance_not_found(self):
|
def test_build_instances_instance_not_found(self):
|
||||||
instances = [fake_instance.fake_instance_obj(self.context)
|
instances = [fake_instance.fake_instance_obj(self.context)
|
||||||
for i in xrange(2)]
|
for i in range(2)]
|
||||||
self.mox.StubOutWithMock(instances[0], 'refresh')
|
self.mox.StubOutWithMock(instances[0], 'refresh')
|
||||||
self.mox.StubOutWithMock(instances[1], 'refresh')
|
self.mox.StubOutWithMock(instances[1], 'refresh')
|
||||||
image = {'fake-data': 'should_pass_silently'}
|
image = {'fake-data': 'should_pass_silently'}
|
||||||
@ -1986,7 +1986,7 @@ class ConductorTaskTestCase(_BaseTaskTestCase, test_compute.BaseTestCase):
|
|||||||
def test_build_instances_info_cache_not_found(self, build_request_spec,
|
def test_build_instances_info_cache_not_found(self, build_request_spec,
|
||||||
setup_instance_group):
|
setup_instance_group):
|
||||||
instances = [fake_instance.fake_instance_obj(self.context)
|
instances = [fake_instance.fake_instance_obj(self.context)
|
||||||
for i in xrange(2)]
|
for i in range(2)]
|
||||||
image = {'fake-data': 'should_pass_silently'}
|
image = {'fake-data': 'should_pass_silently'}
|
||||||
destinations = [{'host': 'host1', 'nodename': 'node1', 'limits': []},
|
destinations = [{'host': 'host1', 'nodename': 'node1', 'limits': []},
|
||||||
{'host': 'host2', 'nodename': 'node2', 'limits': []}]
|
{'host': 'host2', 'nodename': 'node2', 'limits': []}]
|
||||||
|
@ -60,7 +60,7 @@ class ConsoleauthTestCase(test.TestCase):
|
|||||||
fake_validate_console_port)
|
fake_validate_console_port)
|
||||||
|
|
||||||
def test_multiple_tokens_for_instance(self):
|
def test_multiple_tokens_for_instance(self):
|
||||||
tokens = [u"token" + str(i) for i in xrange(10)]
|
tokens = [u"token" + str(i) for i in range(10)]
|
||||||
|
|
||||||
self._stub_validate_console_port(True)
|
self._stub_validate_console_port(True)
|
||||||
|
|
||||||
@ -73,7 +73,7 @@ class ConsoleauthTestCase(test.TestCase):
|
|||||||
self.assertTrue(self.manager_api.check_token(self.context, token))
|
self.assertTrue(self.manager_api.check_token(self.context, token))
|
||||||
|
|
||||||
def test_delete_tokens_for_instance(self):
|
def test_delete_tokens_for_instance(self):
|
||||||
tokens = [u"token" + str(i) for i in xrange(10)]
|
tokens = [u"token" + str(i) for i in range(10)]
|
||||||
for token in tokens:
|
for token in tokens:
|
||||||
self.manager_api.authorize_console(self.context, token, 'novnc',
|
self.manager_api.authorize_console(self.context, token, 'novnc',
|
||||||
'127.0.0.1', '8080', 'host',
|
'127.0.0.1', '8080', 'host',
|
||||||
|
@ -35,6 +35,7 @@ from oslo_serialization import jsonutils
|
|||||||
from oslo_utils import timeutils
|
from oslo_utils import timeutils
|
||||||
from oslo_utils import uuidutils
|
from oslo_utils import uuidutils
|
||||||
import six
|
import six
|
||||||
|
from six.moves import range
|
||||||
from sqlalchemy import Column
|
from sqlalchemy import Column
|
||||||
from sqlalchemy.dialects import sqlite
|
from sqlalchemy.dialects import sqlite
|
||||||
from sqlalchemy.exc import OperationalError
|
from sqlalchemy.exc import OperationalError
|
||||||
@ -6496,7 +6497,7 @@ class S3ImageTestCase(test.TestCase):
|
|||||||
def setUp(self):
|
def setUp(self):
|
||||||
super(S3ImageTestCase, self).setUp()
|
super(S3ImageTestCase, self).setUp()
|
||||||
self.ctxt = context.get_admin_context()
|
self.ctxt = context.get_admin_context()
|
||||||
self.values = [uuidutils.generate_uuid() for i in xrange(3)]
|
self.values = [uuidutils.generate_uuid() for i in range(3)]
|
||||||
self.images = [db.s3_image_create(self.ctxt, uuid)
|
self.images = [db.s3_image_create(self.ctxt, uuid)
|
||||||
for uuid in self.values]
|
for uuid in self.values]
|
||||||
|
|
||||||
@ -6755,7 +6756,7 @@ class ComputeNodeTestCase(test.TestCase, ModelsObjectComparatorMixin):
|
|||||||
def test_compute_node_search_by_hypervisor(self):
|
def test_compute_node_search_by_hypervisor(self):
|
||||||
nodes_created = []
|
nodes_created = []
|
||||||
new_service = copy.copy(self.service_dict)
|
new_service = copy.copy(self.service_dict)
|
||||||
for i in xrange(3):
|
for i in range(3):
|
||||||
new_service['binary'] += str(i)
|
new_service['binary'] += str(i)
|
||||||
new_service['topic'] += str(i)
|
new_service['topic'] += str(i)
|
||||||
service = db.service_create(self.ctxt, new_service)
|
service = db.service_create(self.ctxt, new_service)
|
||||||
@ -6882,7 +6883,7 @@ class ProviderFwRuleTestCase(test.TestCase, ModelsObjectComparatorMixin):
|
|||||||
'2001:4f8:3:ba::/64',
|
'2001:4f8:3:ba::/64',
|
||||||
'2001:4f8:3:ba:2e0:81ff:fe22:d1f1/128']
|
'2001:4f8:3:ba:2e0:81ff:fe22:d1f1/128']
|
||||||
values = []
|
values = []
|
||||||
for i in xrange(len(cidr_samples)):
|
for i in range(len(cidr_samples)):
|
||||||
rule = {}
|
rule = {}
|
||||||
rule['protocol'] = 'foo' + str(i)
|
rule['protocol'] = 'foo' + str(i)
|
||||||
rule['from_port'] = 9999 + i
|
rule['from_port'] = 9999 + i
|
||||||
@ -6922,7 +6923,7 @@ class CertificateTestCase(test.TestCase, ModelsObjectComparatorMixin):
|
|||||||
'file_name': 'filename'
|
'file_name': 'filename'
|
||||||
}
|
}
|
||||||
return [{k: v + str(x) for k, v in base_values.iteritems()}
|
return [{k: v + str(x) for k, v in base_values.iteritems()}
|
||||||
for x in xrange(1, 4)]
|
for x in range(1, 4)]
|
||||||
|
|
||||||
def _certificates_create(self):
|
def _certificates_create(self):
|
||||||
return [db.certificate_create(self.ctxt, cert)
|
return [db.certificate_create(self.ctxt, cert)
|
||||||
@ -6983,7 +6984,7 @@ class ConsoleTestCase(test.TestCase, ModelsObjectComparatorMixin):
|
|||||||
'password': 'pass' + str(x),
|
'password': 'pass' + str(x),
|
||||||
'port': 7878 + x,
|
'port': 7878 + x,
|
||||||
'pool_id': self.console_pools[x]['id']}
|
'pool_id': self.console_pools[x]['id']}
|
||||||
for x in xrange(len(pools_data))]
|
for x in range(len(pools_data))]
|
||||||
self.consoles = [db.console_create(self.ctxt, val)
|
self.consoles = [db.console_create(self.ctxt, val)
|
||||||
for val in self.console_data]
|
for val in self.console_data]
|
||||||
|
|
||||||
@ -7086,7 +7087,7 @@ class CellTestCase(test.TestCase, ModelsObjectComparatorMixin):
|
|||||||
|
|
||||||
def _create_cells(self):
|
def _create_cells(self):
|
||||||
test_values = []
|
test_values = []
|
||||||
for x in xrange(1, 4):
|
for x in range(1, 4):
|
||||||
modified_val = {k: self._cell_value_modify(v, x)
|
modified_val = {k: self._cell_value_modify(v, x)
|
||||||
for k, v in self._get_cell_base_values().iteritems()}
|
for k, v in self._get_cell_base_values().iteritems()}
|
||||||
db.cell_create(self.ctxt, modified_val)
|
db.cell_create(self.ctxt, modified_val)
|
||||||
@ -8592,7 +8593,7 @@ class TestDBInstanceTags(test.TestCase):
|
|||||||
uuid = self._create_instance()
|
uuid = self._create_instance()
|
||||||
tag = 'tag'
|
tag = 'tag'
|
||||||
|
|
||||||
for x in xrange(5):
|
for x in range(5):
|
||||||
db.instance_tag_add(self.context, uuid, tag)
|
db.instance_tag_add(self.context, uuid, tag)
|
||||||
|
|
||||||
tag_refs = db.instance_tag_get_by_instance_uuid(self.context, uuid)
|
tag_refs = db.instance_tag_get_by_instance_uuid(self.context, uuid)
|
||||||
|
@ -24,6 +24,7 @@ library to work with nova.
|
|||||||
import fnmatch
|
import fnmatch
|
||||||
|
|
||||||
from oslo_serialization import jsonutils
|
from oslo_serialization import jsonutils
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
|
|
||||||
class Store(object):
|
class Store(object):
|
||||||
@ -140,7 +141,7 @@ def _paren_groups(source):
|
|||||||
count = 0
|
count = 0
|
||||||
start = 0
|
start = 0
|
||||||
result = []
|
result = []
|
||||||
for pos in xrange(len(source)):
|
for pos in range(len(source)):
|
||||||
if source[pos] == '(':
|
if source[pos] == '(':
|
||||||
if count == 0:
|
if count == 0:
|
||||||
start = pos
|
start = pos
|
||||||
|
@ -15,6 +15,7 @@
|
|||||||
|
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_serialization import jsonutils
|
from oslo_serialization import jsonutils
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova.compute import api as compute_api
|
from nova.compute import api as compute_api
|
||||||
from nova.compute import manager as compute_manager
|
from nova.compute import manager as compute_manager
|
||||||
@ -221,12 +222,12 @@ def fake_vif(x):
|
|||||||
|
|
||||||
|
|
||||||
def floating_ip_ids():
|
def floating_ip_ids():
|
||||||
for i in xrange(1, 100):
|
for i in range(1, 100):
|
||||||
yield i
|
yield i
|
||||||
|
|
||||||
|
|
||||||
def fixed_ip_ids():
|
def fixed_ip_ids():
|
||||||
for i in xrange(1, 100):
|
for i in range(1, 100):
|
||||||
yield i
|
yield i
|
||||||
|
|
||||||
|
|
||||||
@ -237,7 +238,7 @@ fixed_ip_id = fixed_ip_ids()
|
|||||||
def next_fixed_ip(network_id, num_floating_ips=0):
|
def next_fixed_ip(network_id, num_floating_ips=0):
|
||||||
next_id = next(fixed_ip_id)
|
next_id = next(fixed_ip_id)
|
||||||
f_ips = [FakeModel(**next_floating_ip(next_id))
|
f_ips = [FakeModel(**next_floating_ip(next_id))
|
||||||
for i in xrange(num_floating_ips)]
|
for i in range(num_floating_ips)]
|
||||||
return {'id': next_id,
|
return {'id': next_id,
|
||||||
'network_id': network_id,
|
'network_id': network_id,
|
||||||
'address': '192.168.%d.%03d' % (network_id, (next_id + 99)),
|
'address': '192.168.%d.%03d' % (network_id, (next_id + 99)),
|
||||||
@ -295,8 +296,8 @@ def fake_get_instance_nw_info(stubs, num_networks=1, ips_per_vif=2,
|
|||||||
def fixed_ips_fake(*args, **kwargs):
|
def fixed_ips_fake(*args, **kwargs):
|
||||||
global fixed_ips
|
global fixed_ips
|
||||||
ips = [next_fixed_ip(i, floating_ips_per_fixed_ip)
|
ips = [next_fixed_ip(i, floating_ips_per_fixed_ip)
|
||||||
for i in xrange(1, num_networks + 1)
|
for i in range(1, num_networks + 1)
|
||||||
for j in xrange(ips_per_vif)]
|
for j in range(ips_per_vif)]
|
||||||
fixed_ips = ips
|
fixed_ips = ips
|
||||||
return ips
|
return ips
|
||||||
|
|
||||||
|
@ -17,6 +17,8 @@ from mox3 import mox
|
|||||||
from neutronclient.common import exceptions as n_exc
|
from neutronclient.common import exceptions as n_exc
|
||||||
from neutronclient.v2_0 import client
|
from neutronclient.v2_0 import client
|
||||||
|
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova import context
|
from nova import context
|
||||||
from nova import exception
|
from nova import exception
|
||||||
from nova.network.neutronv2 import api as neutronapi
|
from nova.network.neutronv2 import api as neutronapi
|
||||||
@ -222,7 +224,7 @@ class TestNeutronDriver(test.NoDBTestCase):
|
|||||||
device_ids = []
|
device_ids = []
|
||||||
ports = []
|
ports = []
|
||||||
sg_bindings = {}
|
sg_bindings = {}
|
||||||
for i in xrange(0, num_servers):
|
for i in range(0, num_servers):
|
||||||
server_id = "server-%d" % i
|
server_id = "server-%d" % i
|
||||||
port_id = "port-%d" % i
|
port_id = "port-%d" % i
|
||||||
servers.append({'id': server_id})
|
servers.append({'id': server_id})
|
||||||
@ -232,7 +234,7 @@ class TestNeutronDriver(test.NoDBTestCase):
|
|||||||
'security_groups': [sg1_id, sg2_id]})
|
'security_groups': [sg1_id, sg2_id]})
|
||||||
sg_bindings[server_id] = [{'name': 'wol'}, {'name': 'eor'}]
|
sg_bindings[server_id] = [{'name': 'wol'}, {'name': 'eor'}]
|
||||||
|
|
||||||
for x in xrange(0, num_servers, max_query):
|
for x in range(0, num_servers, max_query):
|
||||||
self.moxed_client.list_ports(
|
self.moxed_client.list_ports(
|
||||||
device_id=device_ids[x:x + max_query]).\
|
device_id=device_ids[x:x + max_query]).\
|
||||||
AndReturn({'ports': ports[x:x + max_query]})
|
AndReturn({'ports': ports[x:x + max_query]})
|
||||||
|
@ -134,7 +134,7 @@ class FixedIPTests(test.NoDBTestCase):
|
|||||||
|
|
||||||
def test_add_floating_ip_repeatedly_only_one_instance(self):
|
def test_add_floating_ip_repeatedly_only_one_instance(self):
|
||||||
fixed_ip = model.FixedIP(address='192.168.1.100')
|
fixed_ip = model.FixedIP(address='192.168.1.100')
|
||||||
for i in xrange(10):
|
for i in range(10):
|
||||||
fixed_ip.add_floating_ip('192.168.1.101')
|
fixed_ip.add_floating_ip('192.168.1.101')
|
||||||
self.assertEqual(fixed_ip['floating_ips'], ['192.168.1.101'])
|
self.assertEqual(fixed_ip['floating_ips'], ['192.168.1.101'])
|
||||||
|
|
||||||
@ -199,7 +199,7 @@ class SubnetTests(test.NoDBTestCase):
|
|||||||
subnet = fake_network_cache_model.new_subnet()
|
subnet = fake_network_cache_model.new_subnet()
|
||||||
route1 = fake_network_cache_model.new_route()
|
route1 = fake_network_cache_model.new_route()
|
||||||
route2 = fake_network_cache_model.new_route({'cidr': '1.1.1.1/24'})
|
route2 = fake_network_cache_model.new_route({'cidr': '1.1.1.1/24'})
|
||||||
for i in xrange(10):
|
for i in range(10):
|
||||||
subnet.add_route(route2)
|
subnet.add_route(route2)
|
||||||
self.assertEqual(subnet['routes'], [route1, route2])
|
self.assertEqual(subnet['routes'], [route1, route2])
|
||||||
|
|
||||||
@ -214,7 +214,7 @@ class SubnetTests(test.NoDBTestCase):
|
|||||||
|
|
||||||
def test_add_dns_a_lot(self):
|
def test_add_dns_a_lot(self):
|
||||||
subnet = fake_network_cache_model.new_subnet()
|
subnet = fake_network_cache_model.new_subnet()
|
||||||
for i in xrange(10):
|
for i in range(10):
|
||||||
subnet.add_dns(fake_network_cache_model.new_ip(
|
subnet.add_dns(fake_network_cache_model.new_ip(
|
||||||
dict(address='9.9.9.9')))
|
dict(address='9.9.9.9')))
|
||||||
self.assertEqual(subnet['dns'],
|
self.assertEqual(subnet['dns'],
|
||||||
@ -236,7 +236,7 @@ class SubnetTests(test.NoDBTestCase):
|
|||||||
|
|
||||||
def test_add_ip_a_lot(self):
|
def test_add_ip_a_lot(self):
|
||||||
subnet = fake_network_cache_model.new_subnet()
|
subnet = fake_network_cache_model.new_subnet()
|
||||||
for i in xrange(10):
|
for i in range(10):
|
||||||
subnet.add_ip(fake_network_cache_model.new_fixed_ip(
|
subnet.add_ip(fake_network_cache_model.new_fixed_ip(
|
||||||
dict(address='192.168.1.102')))
|
dict(address='192.168.1.102')))
|
||||||
self.assertEqual(subnet['ips'],
|
self.assertEqual(subnet['ips'],
|
||||||
@ -293,7 +293,7 @@ class NetworkTests(test.NoDBTestCase):
|
|||||||
|
|
||||||
def test_add_subnet_a_lot(self):
|
def test_add_subnet_a_lot(self):
|
||||||
network = fake_network_cache_model.new_network()
|
network = fake_network_cache_model.new_network()
|
||||||
for i in xrange(10):
|
for i in range(10):
|
||||||
network.add_subnet(fake_network_cache_model.new_subnet(
|
network.add_subnet(fake_network_cache_model.new_subnet(
|
||||||
dict(cidr='0.0.0.0')))
|
dict(cidr='0.0.0.0')))
|
||||||
self.assertEqual(network['subnets'],
|
self.assertEqual(network['subnets'],
|
||||||
|
@ -27,6 +27,7 @@ from oslo_config import cfg
|
|||||||
from oslo_serialization import jsonutils
|
from oslo_serialization import jsonutils
|
||||||
from oslo_utils import timeutils
|
from oslo_utils import timeutils
|
||||||
import six
|
import six
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova.compute import flavors
|
from nova.compute import flavors
|
||||||
from nova import context
|
from nova import context
|
||||||
@ -596,7 +597,7 @@ class TestNeutronv2Base(test.TestCase):
|
|||||||
nets = number == 1 and self.nets1 or self.nets2
|
nets = number == 1 and self.nets1 or self.nets2
|
||||||
self.moxed_client.list_networks(
|
self.moxed_client.list_networks(
|
||||||
id=net_ids).AndReturn({'networks': nets})
|
id=net_ids).AndReturn({'networks': nets})
|
||||||
for i in xrange(1, number + 1):
|
for i in range(1, number + 1):
|
||||||
float_data = number == 1 and self.float_data1 or self.float_data2
|
float_data = number == 1 and self.float_data1 or self.float_data2
|
||||||
for ip in port_data[i - 1]['fixed_ips']:
|
for ip in port_data[i - 1]['fixed_ips']:
|
||||||
float_data = [x for x in float_data
|
float_data = [x for x in float_data
|
||||||
@ -619,7 +620,7 @@ class TestNeutronv2Base(test.TestCase):
|
|||||||
net_info_cache)
|
net_info_cache)
|
||||||
instance = self._fake_instance_object_with_info_cache(self.instance)
|
instance = self._fake_instance_object_with_info_cache(self.instance)
|
||||||
nw_inf = api.get_instance_nw_info(self.context, instance)
|
nw_inf = api.get_instance_nw_info(self.context, instance)
|
||||||
for i in xrange(0, number):
|
for i in range(0, number):
|
||||||
self._verify_nw_info(nw_inf, i)
|
self._verify_nw_info(nw_inf, i)
|
||||||
|
|
||||||
def _allocate_for_instance(self, net_idx=1, **kwargs):
|
def _allocate_for_instance(self, net_idx=1, **kwargs):
|
||||||
|
@ -15,6 +15,7 @@
|
|||||||
|
|
||||||
import mock
|
import mock
|
||||||
from oslo_utils import timeutils
|
from oslo_utils import timeutils
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova import exception
|
from nova import exception
|
||||||
from nova.scheduler import caching_scheduler
|
from nova.scheduler import caching_scheduler
|
||||||
@ -144,7 +145,7 @@ class CachingSchedulerTestCase(test_scheduler.SchedulerTestCase):
|
|||||||
|
|
||||||
request_spec = self._get_fake_request_spec()
|
request_spec = self._get_fake_request_spec()
|
||||||
host_states = []
|
host_states = []
|
||||||
for x in xrange(hosts):
|
for x in range(hosts):
|
||||||
host_state = self._get_fake_host_state(x)
|
host_state = self._get_fake_host_state(x)
|
||||||
host_states.append(host_state)
|
host_states.append(host_state)
|
||||||
self.driver.all_host_states = host_states
|
self.driver.all_host_states = host_states
|
||||||
@ -152,7 +153,7 @@ class CachingSchedulerTestCase(test_scheduler.SchedulerTestCase):
|
|||||||
def run_test():
|
def run_test():
|
||||||
a = timeutils.utcnow()
|
a = timeutils.utcnow()
|
||||||
|
|
||||||
for x in xrange(requests):
|
for x in range(requests):
|
||||||
self.driver.select_destinations(
|
self.driver.select_destinations(
|
||||||
self.context, request_spec, {})
|
self.context, request_spec, {})
|
||||||
|
|
||||||
|
@ -18,6 +18,8 @@ Tests For Scheduler Host Filters.
|
|||||||
import inspect
|
import inspect
|
||||||
import sys
|
import sys
|
||||||
|
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova import filters
|
from nova import filters
|
||||||
from nova import loadables
|
from nova import loadables
|
||||||
from nova import test
|
from nova import test
|
||||||
@ -72,15 +74,15 @@ class FiltersTestCase(test.NoDBTestCase):
|
|||||||
# call gets to processing 'obj2'. We then return 'False' for it.
|
# call gets to processing 'obj2'. We then return 'False' for it.
|
||||||
# After that, 'obj3' gets yielded 'total_iterations' number of
|
# After that, 'obj3' gets yielded 'total_iterations' number of
|
||||||
# times.
|
# times.
|
||||||
for x in xrange(total_iterations):
|
for x in range(total_iterations):
|
||||||
base_filter._filter_one('obj1', filter_properties).AndReturn(True)
|
base_filter._filter_one('obj1', filter_properties).AndReturn(True)
|
||||||
base_filter._filter_one('obj2', filter_properties).AndReturn(False)
|
base_filter._filter_one('obj2', filter_properties).AndReturn(False)
|
||||||
for x in xrange(total_iterations):
|
for x in range(total_iterations):
|
||||||
base_filter._filter_one('obj3', filter_properties).AndReturn(True)
|
base_filter._filter_one('obj3', filter_properties).AndReturn(True)
|
||||||
self.mox.ReplayAll()
|
self.mox.ReplayAll()
|
||||||
|
|
||||||
objs = iter(filter_obj_list)
|
objs = iter(filter_obj_list)
|
||||||
for x in xrange(total_iterations):
|
for x in range(total_iterations):
|
||||||
# Pass in generators returned from previous calls.
|
# Pass in generators returned from previous calls.
|
||||||
objs = base_filter.filter_all(objs, filter_properties)
|
objs = base_filter.filter_all(objs, filter_properties)
|
||||||
self.assertTrue(inspect.isgenerator(objs))
|
self.assertTrue(inspect.isgenerator(objs))
|
||||||
|
@ -67,9 +67,9 @@ class HostManagerTestCase(test.NoDBTestCase):
|
|||||||
self.flags(scheduler_default_filters=['FakeFilterClass1'])
|
self.flags(scheduler_default_filters=['FakeFilterClass1'])
|
||||||
self.host_manager = host_manager.HostManager()
|
self.host_manager = host_manager.HostManager()
|
||||||
self.fake_hosts = [host_manager.HostState('fake_host%s' % x,
|
self.fake_hosts = [host_manager.HostState('fake_host%s' % x,
|
||||||
'fake-node') for x in xrange(1, 5)]
|
'fake-node') for x in range(1, 5)]
|
||||||
self.fake_hosts += [host_manager.HostState('fake_multihost',
|
self.fake_hosts += [host_manager.HostState('fake_multihost',
|
||||||
'fake-node%s' % x) for x in xrange(1, 5)]
|
'fake-node%s' % x) for x in range(1, 5)]
|
||||||
|
|
||||||
def test_load_filters(self):
|
def test_load_filters(self):
|
||||||
filters = self.host_manager._load_filters()
|
filters = self.host_manager._load_filters()
|
||||||
@ -388,7 +388,7 @@ class HostManagerTestCase(test.NoDBTestCase):
|
|||||||
|
|
||||||
self.assertEqual(len(host_states_map), 4)
|
self.assertEqual(len(host_states_map), 4)
|
||||||
# Check that .service is set properly
|
# Check that .service is set properly
|
||||||
for i in xrange(4):
|
for i in range(4):
|
||||||
compute_node = fakes.COMPUTE_NODES[i]
|
compute_node = fakes.COMPUTE_NODES[i]
|
||||||
host = compute_node['host']
|
host = compute_node['host']
|
||||||
node = compute_node['hypervisor_hostname']
|
node = compute_node['hypervisor_hostname']
|
||||||
|
@ -18,6 +18,7 @@ import datetime
|
|||||||
|
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_utils import timeutils
|
from oslo_utils import timeutils
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova import compute
|
from nova import compute
|
||||||
from nova.compute import flavors
|
from nova.compute import flavors
|
||||||
@ -174,39 +175,39 @@ class QuotaIntegrationTestCase(test.TestCase):
|
|||||||
|
|
||||||
def test_max_injected_files(self):
|
def test_max_injected_files(self):
|
||||||
files = []
|
files = []
|
||||||
for i in xrange(CONF.quota_injected_files):
|
for i in range(CONF.quota_injected_files):
|
||||||
files.append(('/my/path%d' % i, 'config = test\n'))
|
files.append(('/my/path%d' % i, 'config = test\n'))
|
||||||
self._create_with_injected_files(files) # no QuotaError
|
self._create_with_injected_files(files) # no QuotaError
|
||||||
|
|
||||||
def test_too_many_injected_files(self):
|
def test_too_many_injected_files(self):
|
||||||
files = []
|
files = []
|
||||||
for i in xrange(CONF.quota_injected_files + 1):
|
for i in range(CONF.quota_injected_files + 1):
|
||||||
files.append(('/my/path%d' % i, 'my\ncontent%d\n' % i))
|
files.append(('/my/path%d' % i, 'my\ncontent%d\n' % i))
|
||||||
self.assertRaises(exception.QuotaError,
|
self.assertRaises(exception.QuotaError,
|
||||||
self._create_with_injected_files, files)
|
self._create_with_injected_files, files)
|
||||||
|
|
||||||
def test_max_injected_file_content_bytes(self):
|
def test_max_injected_file_content_bytes(self):
|
||||||
max = CONF.quota_injected_file_content_bytes
|
max = CONF.quota_injected_file_content_bytes
|
||||||
content = ''.join(['a' for i in xrange(max)])
|
content = ''.join(['a' for i in range(max)])
|
||||||
files = [('/test/path', content)]
|
files = [('/test/path', content)]
|
||||||
self._create_with_injected_files(files) # no QuotaError
|
self._create_with_injected_files(files) # no QuotaError
|
||||||
|
|
||||||
def test_too_many_injected_file_content_bytes(self):
|
def test_too_many_injected_file_content_bytes(self):
|
||||||
max = CONF.quota_injected_file_content_bytes
|
max = CONF.quota_injected_file_content_bytes
|
||||||
content = ''.join(['a' for i in xrange(max + 1)])
|
content = ''.join(['a' for i in range(max + 1)])
|
||||||
files = [('/test/path', content)]
|
files = [('/test/path', content)]
|
||||||
self.assertRaises(exception.QuotaError,
|
self.assertRaises(exception.QuotaError,
|
||||||
self._create_with_injected_files, files)
|
self._create_with_injected_files, files)
|
||||||
|
|
||||||
def test_max_injected_file_path_bytes(self):
|
def test_max_injected_file_path_bytes(self):
|
||||||
max = CONF.quota_injected_file_path_length
|
max = CONF.quota_injected_file_path_length
|
||||||
path = ''.join(['a' for i in xrange(max)])
|
path = ''.join(['a' for i in range(max)])
|
||||||
files = [(path, 'config = quotatest')]
|
files = [(path, 'config = quotatest')]
|
||||||
self._create_with_injected_files(files) # no QuotaError
|
self._create_with_injected_files(files) # no QuotaError
|
||||||
|
|
||||||
def test_too_many_injected_file_path_bytes(self):
|
def test_too_many_injected_file_path_bytes(self):
|
||||||
max = CONF.quota_injected_file_path_length
|
max = CONF.quota_injected_file_path_length
|
||||||
path = ''.join(['a' for i in xrange(max + 1)])
|
path = ''.join(['a' for i in range(max + 1)])
|
||||||
files = [(path, 'config = quotatest')]
|
files = [(path, 'config = quotatest')]
|
||||||
self.assertRaises(exception.QuotaError,
|
self.assertRaises(exception.QuotaError,
|
||||||
self._create_with_injected_files, files)
|
self._create_with_injected_files, files)
|
||||||
|
@ -19,6 +19,7 @@ import sys
|
|||||||
|
|
||||||
import mock
|
import mock
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova.compute import flavors
|
from nova.compute import flavors
|
||||||
import nova.context
|
import nova.context
|
||||||
@ -145,7 +146,7 @@ def get_test_network_info(count=1):
|
|||||||
|
|
||||||
return vif
|
return vif
|
||||||
|
|
||||||
return network_model.NetworkInfo([current() for x in xrange(0, count)])
|
return network_model.NetworkInfo([current() for x in range(0, count)])
|
||||||
|
|
||||||
|
|
||||||
def is_osx():
|
def is_osx():
|
||||||
|
@ -15,6 +15,8 @@
|
|||||||
|
|
||||||
import mock
|
import mock
|
||||||
|
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova import exception
|
from nova import exception
|
||||||
from nova import test
|
from nova import test
|
||||||
from nova.virt.hyperv import constants
|
from nova.virt.hyperv import constants
|
||||||
@ -270,7 +272,7 @@ class VMUtilsTestCase(test.NoDBTestCase):
|
|||||||
def test_get_free_controller_slot_exception(self):
|
def test_get_free_controller_slot_exception(self):
|
||||||
fake_drive = mock.MagicMock()
|
fake_drive = mock.MagicMock()
|
||||||
type(fake_drive).AddressOnParent = mock.PropertyMock(
|
type(fake_drive).AddressOnParent = mock.PropertyMock(
|
||||||
side_effect=xrange(constants.SCSI_CONTROLLER_SLOTS_NUMBER))
|
side_effect=range(constants.SCSI_CONTROLLER_SLOTS_NUMBER))
|
||||||
|
|
||||||
with mock.patch.object(self._vmutils,
|
with mock.patch.object(self._vmutils,
|
||||||
'get_attached_disks') as fake_get_attached_disks:
|
'get_attached_disks') as fake_get_attached_disks:
|
||||||
|
@ -43,6 +43,7 @@ from oslo_utils import timeutils
|
|||||||
from oslo_utils import units
|
from oslo_utils import units
|
||||||
from oslo_utils import uuidutils
|
from oslo_utils import uuidutils
|
||||||
import six
|
import six
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova.api.metadata import base as instance_metadata
|
from nova.api.metadata import base as instance_metadata
|
||||||
from nova.compute import arch
|
from nova.compute import arch
|
||||||
@ -8258,7 +8259,7 @@ class LibvirtConnTestCase(test.NoDBTestCase):
|
|||||||
return_infos = [(libvirt_driver.VIR_DOMAIN_RUNNING,) + info_tuple]
|
return_infos = [(libvirt_driver.VIR_DOMAIN_RUNNING,) + info_tuple]
|
||||||
return_shutdowns = [shutdown_count.append("shutdown")]
|
return_shutdowns = [shutdown_count.append("shutdown")]
|
||||||
retry_countdown = retry_interval
|
retry_countdown = retry_interval
|
||||||
for x in xrange(min(seconds_to_shutdown, timeout)):
|
for x in range(min(seconds_to_shutdown, timeout)):
|
||||||
return_infos.append(
|
return_infos.append(
|
||||||
(libvirt_driver.VIR_DOMAIN_RUNNING,) + info_tuple)
|
(libvirt_driver.VIR_DOMAIN_RUNNING,) + info_tuple)
|
||||||
if retry_countdown == 0:
|
if retry_countdown == 0:
|
||||||
@ -8932,7 +8933,7 @@ class LibvirtConnTestCase(test.NoDBTestCase):
|
|||||||
|
|
||||||
# Generate mempages list per cell
|
# Generate mempages list per cell
|
||||||
cell_mempages = list()
|
cell_mempages = list()
|
||||||
for cellid in xrange(cells_per_host):
|
for cellid in range(cells_per_host):
|
||||||
mempages_0 = vconfig.LibvirtConfigCapsNUMAPages()
|
mempages_0 = vconfig.LibvirtConfigCapsNUMAPages()
|
||||||
mempages_0.size = 4
|
mempages_0.size = 4
|
||||||
mempages_0.total = 1024 * cellid
|
mempages_0.total = 1024 * cellid
|
||||||
|
@ -824,7 +824,7 @@ class ImageCacheManagerTestCase(test.NoDBTestCase):
|
|||||||
def fake_get_all_by_filters(context, *args, **kwargs):
|
def fake_get_all_by_filters(context, *args, **kwargs):
|
||||||
was['called'] = True
|
was['called'] = True
|
||||||
instances = []
|
instances = []
|
||||||
for x in xrange(2):
|
for x in range(2):
|
||||||
instances.append(fake_instance.fake_db_instance(
|
instances.append(fake_instance.fake_db_instance(
|
||||||
image_ref='1',
|
image_ref='1',
|
||||||
uuid=x,
|
uuid=x,
|
||||||
|
@ -685,9 +685,9 @@ class TestDriverBlockDevice(test.NoDBTestCase):
|
|||||||
|
|
||||||
# Test that the ephemerals work as expected
|
# Test that the ephemerals work as expected
|
||||||
test_ephemerals = [self.driver_classes['ephemeral'](
|
test_ephemerals = [self.driver_classes['ephemeral'](
|
||||||
self.ephemeral_bdm) for _ in xrange(2)]
|
self.ephemeral_bdm) for _ in range(2)]
|
||||||
expected = [self.ephemeral_legacy_driver_bdm.copy()
|
expected = [self.ephemeral_legacy_driver_bdm.copy()
|
||||||
for _ in xrange(2)]
|
for _ in range(2)]
|
||||||
expected[0]['virtual_name'] = 'ephemeral0'
|
expected[0]['virtual_name'] = 'ephemeral0'
|
||||||
expected[0]['num'] = 0
|
expected[0]['num'] = 0
|
||||||
expected[1]['virtual_name'] = 'ephemeral1'
|
expected[1]['virtual_name'] = 'ephemeral1'
|
||||||
|
@ -377,7 +377,7 @@ class XenAPIVMTestCase(stubs.XenAPITestBase):
|
|||||||
|
|
||||||
def test_list_instance_uuids(self):
|
def test_list_instance_uuids(self):
|
||||||
uuids = []
|
uuids = []
|
||||||
for x in xrange(1, 4):
|
for x in range(1, 4):
|
||||||
instance = self._create_instance()
|
instance = self._create_instance()
|
||||||
uuids.append(instance['uuid'])
|
uuids.append(instance['uuid'])
|
||||||
instance_uuids = self.conn.list_instance_uuids()
|
instance_uuids = self.conn.list_instance_uuids()
|
||||||
@ -1625,7 +1625,7 @@ class XenAPIDiffieHellmanTestCase(test.NoDBTestCase):
|
|||||||
self._test_encryption('\n\nMessage with leading newlines.')
|
self._test_encryption('\n\nMessage with leading newlines.')
|
||||||
|
|
||||||
def test_encrypt_really_long_message(self):
|
def test_encrypt_really_long_message(self):
|
||||||
self._test_encryption(''.join(['abcd' for i in xrange(1024)]))
|
self._test_encryption(''.join(['abcd' for i in range(1024)]))
|
||||||
|
|
||||||
|
|
||||||
# FIXME(sirp): convert this to use XenAPITestBaseNoDB
|
# FIXME(sirp): convert this to use XenAPITestBaseNoDB
|
||||||
|
@ -47,6 +47,7 @@ from oslo_utils import excutils
|
|||||||
from oslo_utils import importutils
|
from oslo_utils import importutils
|
||||||
from oslo_utils import timeutils
|
from oslo_utils import timeutils
|
||||||
import six
|
import six
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova import exception
|
from nova import exception
|
||||||
from nova.i18n import _, _LE, _LW
|
from nova.i18n import _, _LE, _LW
|
||||||
@ -249,7 +250,7 @@ def novadir():
|
|||||||
|
|
||||||
def generate_uid(topic, size=8):
|
def generate_uid(topic, size=8):
|
||||||
characters = '01234567890abcdefghijklmnopqrstuvwxyz'
|
characters = '01234567890abcdefghijklmnopqrstuvwxyz'
|
||||||
choices = [random.choice(characters) for _x in xrange(size)]
|
choices = [random.choice(characters) for _x in range(size)]
|
||||||
return '%s-%s' % (topic, ''.join(choices))
|
return '%s-%s' % (topic, ''.join(choices))
|
||||||
|
|
||||||
|
|
||||||
@ -384,7 +385,7 @@ def generate_password(length=None, symbolgroups=DEFAULT_PASSWORD_SYMBOLS):
|
|||||||
|
|
||||||
# then fill with random characters from all symbol groups
|
# then fill with random characters from all symbol groups
|
||||||
symbols = ''.join(symbolgroups)
|
symbols = ''.join(symbolgroups)
|
||||||
password.extend([r.choice(symbols) for _i in xrange(length)])
|
password.extend([r.choice(symbols) for _i in range(length)])
|
||||||
|
|
||||||
# finally shuffle to ensure first x characters aren't from a
|
# finally shuffle to ensure first x characters aren't from a
|
||||||
# predictable group
|
# predictable group
|
||||||
|
@ -27,6 +27,7 @@ if sys.platform == 'win32':
|
|||||||
|
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_log import log as logging
|
from oslo_log import log as logging
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova import exception
|
from nova import exception
|
||||||
from nova.i18n import _, _LW
|
from nova.i18n import _, _LW
|
||||||
@ -729,7 +730,7 @@ class VMUtils(object):
|
|||||||
attached_disks = self.get_attached_disks(scsi_controller_path)
|
attached_disks = self.get_attached_disks(scsi_controller_path)
|
||||||
used_slots = [int(disk.AddressOnParent) for disk in attached_disks]
|
used_slots = [int(disk.AddressOnParent) for disk in attached_disks]
|
||||||
|
|
||||||
for slot in xrange(constants.SCSI_CONTROLLER_SLOTS_NUMBER):
|
for slot in range(constants.SCSI_CONTROLLER_SLOTS_NUMBER):
|
||||||
if slot not in used_slots:
|
if slot not in used_slots:
|
||||||
return slot
|
return slot
|
||||||
raise HyperVException(_("Exceeded the maximum number of slots"))
|
raise HyperVException(_("Exceeded the maximum number of slots"))
|
||||||
|
@ -25,6 +25,7 @@ import time
|
|||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_log import log as logging
|
from oslo_log import log as logging
|
||||||
from oslo_utils import excutils
|
from oslo_utils import excutils
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova import exception
|
from nova import exception
|
||||||
from nova.i18n import _, _LE, _LW
|
from nova.i18n import _, _LE, _LW
|
||||||
@ -284,7 +285,7 @@ class ISCSIVolumeDriver(object):
|
|||||||
# The WMI query in get_device_number_for_target can incorrectly
|
# The WMI query in get_device_number_for_target can incorrectly
|
||||||
# return no data when the system is under load. This issue can
|
# return no data when the system is under load. This issue can
|
||||||
# be avoided by adding a retry.
|
# be avoided by adding a retry.
|
||||||
for i in xrange(CONF.hyperv.mounted_disk_query_retry_count):
|
for i in range(CONF.hyperv.mounted_disk_query_retry_count):
|
||||||
device_number = self._volutils.get_device_number_for_target(
|
device_number = self._volutils.get_device_number_for_target(
|
||||||
target_iqn, target_lun)
|
target_iqn, target_lun)
|
||||||
if device_number in (None, -1):
|
if device_number in (None, -1):
|
||||||
|
@ -28,6 +28,7 @@ import time
|
|||||||
|
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_log import log as logging
|
from oslo_log import log as logging
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova.i18n import _
|
from nova.i18n import _
|
||||||
from nova import utils
|
from nova import utils
|
||||||
@ -86,7 +87,7 @@ class VolumeUtils(basevolumeutils.BaseVolumeUtils):
|
|||||||
if retry_count < 2:
|
if retry_count < 2:
|
||||||
retry_count = 2
|
retry_count = 2
|
||||||
|
|
||||||
for attempt in xrange(retry_count):
|
for attempt in range(retry_count):
|
||||||
try:
|
try:
|
||||||
session_info = self.execute('iscsicli.exe', 'SessionList')
|
session_info = self.execute('iscsicli.exe', 'SessionList')
|
||||||
if session_info.find(target_iqn) == -1:
|
if session_info.find(target_iqn) == -1:
|
||||||
|
@ -26,6 +26,7 @@ if sys.platform == 'win32':
|
|||||||
|
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_log import log as logging
|
from oslo_log import log as logging
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova.i18n import _
|
from nova.i18n import _
|
||||||
from nova import utils
|
from nova import utils
|
||||||
@ -78,7 +79,7 @@ class VolumeUtilsV2(basevolumeutils.BaseVolumeUtils):
|
|||||||
if retry_count < 2:
|
if retry_count < 2:
|
||||||
retry_count = 2
|
retry_count = 2
|
||||||
|
|
||||||
for attempt in xrange(retry_count):
|
for attempt in range(retry_count):
|
||||||
target = self._conn_storage.query("SELECT * FROM MSFT_iSCSITarget "
|
target = self._conn_storage.query("SELECT * FROM MSFT_iSCSITarget "
|
||||||
"WHERE NodeAddress='%s' " %
|
"WHERE NodeAddress='%s' " %
|
||||||
target_iqn)
|
target_iqn)
|
||||||
|
@ -54,6 +54,7 @@ from oslo_utils import strutils
|
|||||||
from oslo_utils import timeutils
|
from oslo_utils import timeutils
|
||||||
from oslo_utils import units
|
from oslo_utils import units
|
||||||
import six
|
import six
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova.api.metadata import base as instance_metadata
|
from nova.api.metadata import base as instance_metadata
|
||||||
from nova import block_device
|
from nova import block_device
|
||||||
@ -2042,7 +2043,7 @@ class LibvirtDriver(driver.ComputeDriver):
|
|||||||
# call takes to return.
|
# call takes to return.
|
||||||
self._prepare_pci_devices_for_use(
|
self._prepare_pci_devices_for_use(
|
||||||
pci_manager.get_instance_pci_devs(instance, 'all'))
|
pci_manager.get_instance_pci_devs(instance, 'all'))
|
||||||
for x in xrange(CONF.libvirt.wait_soft_reboot_seconds):
|
for x in range(CONF.libvirt.wait_soft_reboot_seconds):
|
||||||
dom = self._host.get_domain(instance)
|
dom = self._host.get_domain(instance)
|
||||||
state = self._get_power_state(dom)
|
state = self._get_power_state(dom)
|
||||||
new_domid = dom.ID()
|
new_domid = dom.ID()
|
||||||
@ -5243,7 +5244,7 @@ class LibvirtDriver(driver.ComputeDriver):
|
|||||||
|
|
||||||
# nwfilters may be defined in a separate thread in the case
|
# nwfilters may be defined in a separate thread in the case
|
||||||
# of libvirt non-blocking mode, so we wait for completion
|
# of libvirt non-blocking mode, so we wait for completion
|
||||||
timeout_count = range(CONF.live_migration_retry_count)
|
timeout_count = list(range(CONF.live_migration_retry_count))
|
||||||
while timeout_count:
|
while timeout_count:
|
||||||
if self.firewall_driver.instance_filter_exists(instance,
|
if self.firewall_driver.instance_filter_exists(instance,
|
||||||
network_info):
|
network_info):
|
||||||
|
@ -23,6 +23,7 @@ from eventlet import queue
|
|||||||
from eventlet import timeout
|
from eventlet import timeout
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
from oslo_log import log as logging
|
from oslo_log import log as logging
|
||||||
|
from six.moves import range
|
||||||
|
|
||||||
from nova import context
|
from nova import context
|
||||||
from nova import exception
|
from nova import exception
|
||||||
@ -128,7 +129,7 @@ class XenAPISession(object):
|
|||||||
return url
|
return url
|
||||||
|
|
||||||
def _populate_session_pool(self, url, user, pw, exception):
|
def _populate_session_pool(self, url, user, pw, exception):
|
||||||
for i in xrange(CONF.xenserver.connection_concurrent - 1):
|
for i in range(CONF.xenserver.connection_concurrent - 1):
|
||||||
session = self._create_session(url)
|
session = self._create_session(url)
|
||||||
with timeout.Timeout(CONF.xenserver.login_timeout, exception):
|
with timeout.Timeout(CONF.xenserver.login_timeout, exception):
|
||||||
session.login_with_password(user, pw,
|
session.login_with_password(user, pw,
|
||||||
@ -217,7 +218,7 @@ class XenAPISession(object):
|
|||||||
"""Allows a plugin to raise RetryableError so we can try again."""
|
"""Allows a plugin to raise RetryableError so we can try again."""
|
||||||
attempts = num_retries + 1
|
attempts = num_retries + 1
|
||||||
sleep_time = 0.5
|
sleep_time = 0.5
|
||||||
for attempt in xrange(1, attempts + 1):
|
for attempt in range(1, attempts + 1):
|
||||||
try:
|
try:
|
||||||
if attempt > 1:
|
if attempt > 1:
|
||||||
time.sleep(sleep_time)
|
time.sleep(sleep_time)
|
||||||
|
@ -37,6 +37,7 @@ from oslo_utils import strutils
|
|||||||
from oslo_utils import timeutils
|
from oslo_utils import timeutils
|
||||||
from oslo_utils import units
|
from oslo_utils import units
|
||||||
import six
|
import six
|
||||||
|
from six.moves import range
|
||||||
import six.moves.urllib.parse as urlparse
|
import six.moves.urllib.parse as urlparse
|
||||||
|
|
||||||
from nova.api.metadata import base as instance_metadata
|
from nova.api.metadata import base as instance_metadata
|
||||||
@ -382,7 +383,7 @@ def _should_retry_unplug_vbd(err):
|
|||||||
def unplug_vbd(session, vbd_ref, this_vm_ref):
|
def unplug_vbd(session, vbd_ref, this_vm_ref):
|
||||||
# make sure that perform at least once
|
# make sure that perform at least once
|
||||||
max_attempts = max(0, CONF.xenserver.num_vbd_unplug_retries) + 1
|
max_attempts = max(0, CONF.xenserver.num_vbd_unplug_retries) + 1
|
||||||
for num_attempt in xrange(1, max_attempts + 1):
|
for num_attempt in range(1, max_attempts + 1):
|
||||||
try:
|
try:
|
||||||
if num_attempt > 1:
|
if num_attempt > 1:
|
||||||
greenthread.sleep(1)
|
greenthread.sleep(1)
|
||||||
@ -1366,7 +1367,7 @@ def _make_uuid_stack():
|
|||||||
# which does not have the `uuid` module. To work around this,
|
# which does not have the `uuid` module. To work around this,
|
||||||
# we generate the uuids here (under Python 2.6+) and
|
# we generate the uuids here (under Python 2.6+) and
|
||||||
# pass them as arguments
|
# pass them as arguments
|
||||||
return [str(uuid.uuid4()) for i in xrange(MAX_VDI_CHAIN_SIZE)]
|
return [str(uuid.uuid4()) for i in range(MAX_VDI_CHAIN_SIZE)]
|
||||||
|
|
||||||
|
|
||||||
def _image_uses_bittorrent(context, instance):
|
def _image_uses_bittorrent(context, instance):
|
||||||
@ -2099,7 +2100,7 @@ def _wait_for_vhd_coalesce(session, instance, sr_ref, vdi_ref,
|
|||||||
# Its possible that other coalesce operation happen, so we need
|
# Its possible that other coalesce operation happen, so we need
|
||||||
# to consider the full chain, rather than just the most recent parent.
|
# to consider the full chain, rather than just the most recent parent.
|
||||||
good_parent_uuids = vdi_uuid_list[1:]
|
good_parent_uuids = vdi_uuid_list[1:]
|
||||||
for i in xrange(max_attempts):
|
for i in range(max_attempts):
|
||||||
# NOTE(sirp): This rescan is necessary to ensure the VM's `sm_config`
|
# NOTE(sirp): This rescan is necessary to ensure the VM's `sm_config`
|
||||||
# matches the underlying VHDs.
|
# matches the underlying VHDs.
|
||||||
# This can also kick XenServer into performing a pending coalesce.
|
# This can also kick XenServer into performing a pending coalesce.
|
||||||
@ -2146,7 +2147,7 @@ def _remap_vbd_dev(dev):
|
|||||||
|
|
||||||
def _wait_for_device(dev):
|
def _wait_for_device(dev):
|
||||||
"""Wait for device node to appear."""
|
"""Wait for device node to appear."""
|
||||||
for i in xrange(0, CONF.xenserver.block_device_creation_timeout):
|
for i in range(0, CONF.xenserver.block_device_creation_timeout):
|
||||||
dev_path = utils.make_dev_path(dev)
|
dev_path = utils.make_dev_path(dev)
|
||||||
if os.path.exists(dev_path):
|
if os.path.exists(dev_path):
|
||||||
return
|
return
|
||||||
|
Loading…
x
Reference in New Issue
Block a user