4475a5ca00
Commit [1] added wrong decorator to test class located in "manila_tempest_tests/tests/api/test_quotas.py" module that made it invisible for tempest. It should be "ddt.ddt" instead of "ddt.data" as it is now. Also, fix negative quota tests that were testing wrong thing. It should have been testing "quota-sets" and "os-quota-sets", not "services" and "os-services" as it is now. [1] I82f00114db985b4b3bf4db0a64191559508ac600 Change-Id: Ie0eb7d32b7b032ffdb7f7dd47f68841211e7d7a6 Closes-Bug: #1635588
944 lines
38 KiB
Python
944 lines
38 KiB
Python
# Copyright 2014 Mirantis Inc.
|
|
# All Rights Reserved.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
|
|
import copy
|
|
import inspect
|
|
import re
|
|
import traceback
|
|
|
|
from oslo_concurrency import lockutils
|
|
from oslo_log import log
|
|
import six
|
|
from tempest import clients
|
|
from tempest.common import credentials_factory as common_creds
|
|
from tempest.common import dynamic_creds
|
|
from tempest import config
|
|
from tempest.lib.common.utils import data_utils
|
|
from tempest.lib import exceptions
|
|
from tempest import test
|
|
|
|
from manila_tempest_tests.common import constants
|
|
from manila_tempest_tests.services.share.json import shares_client
|
|
from manila_tempest_tests.services.share.v2.json import (
|
|
shares_client as shares_v2_client)
|
|
from manila_tempest_tests import share_exceptions
|
|
from manila_tempest_tests import utils
|
|
|
|
CONF = config.CONF
|
|
LOG = log.getLogger(__name__)
|
|
|
|
# Test tags related to test direction
|
|
TAG_POSITIVE = "positive"
|
|
TAG_NEGATIVE = "negative"
|
|
|
|
# Test tags related to service involvement
|
|
TAG_API = "api"
|
|
TAG_BACKEND = "backend"
|
|
TAG_API_WITH_BACKEND = "api_with_backend"
|
|
|
|
TAGS_MAPPER = {
|
|
"p": TAG_POSITIVE,
|
|
"n": TAG_NEGATIVE,
|
|
"a": TAG_API,
|
|
"b": TAG_BACKEND,
|
|
"ab": TAG_API_WITH_BACKEND,
|
|
}
|
|
TAGS_PATTERN = re.compile(
|
|
r"(?=.*\[.*\b(%(p)s|%(n)s)\b.*\])(?=.*\[.*\b(%(a)s|%(b)s|%(ab)s)\b.*\])" %
|
|
TAGS_MAPPER)
|
|
|
|
|
|
def verify_test_has_appropriate_tags(self):
|
|
if not TAGS_PATTERN.match(self.id()):
|
|
msg = (
|
|
"Required attributes either not set or set improperly. "
|
|
"Two test attributes are expected:\n"
|
|
" - one of '%(p)s' or '%(n)s' and \n"
|
|
" - one of '%(a)s', '%(b)s' or '%(ab)s'."
|
|
) % TAGS_MAPPER
|
|
raise self.failureException(msg)
|
|
|
|
|
|
class handle_cleanup_exceptions(object):
|
|
"""Handle exceptions raised with cleanup operations.
|
|
|
|
Always suppress errors when exceptions.NotFound or exceptions.Forbidden
|
|
are raised.
|
|
Suppress all other exceptions only in case config opt
|
|
'suppress_errors_in_cleanup' in config group 'share' is True.
|
|
"""
|
|
|
|
def __enter__(self):
|
|
return self
|
|
|
|
def __exit__(self, exc_type, exc_value, exc_traceback):
|
|
if not (isinstance(exc_value,
|
|
(exceptions.NotFound, exceptions.Forbidden)) or
|
|
CONF.share.suppress_errors_in_cleanup):
|
|
return False # Do not suppress error if any
|
|
if exc_traceback:
|
|
LOG.error("Suppressed cleanup error in Manila: "
|
|
"\n%s" % traceback.format_exc())
|
|
return True # Suppress error if any
|
|
|
|
|
|
def network_synchronized(f):
|
|
|
|
def wrapped_func(self, *args, **kwargs):
|
|
with_isolated_creds = True if len(args) > 2 else False
|
|
no_lock_required = kwargs.get(
|
|
"isolated_creds_client", with_isolated_creds)
|
|
if no_lock_required:
|
|
# Usage of not reusable network. No need in lock.
|
|
return f(self, *args, **kwargs)
|
|
|
|
# Use lock assuming reusage of common network.
|
|
@lockutils.synchronized("manila_network_lock", external=True)
|
|
def source_func(self, *args, **kwargs):
|
|
return f(self, *args, **kwargs)
|
|
|
|
return source_func(self, *args, **kwargs)
|
|
|
|
return wrapped_func
|
|
|
|
|
|
skip_if_microversion_not_supported = utils.skip_if_microversion_not_supported
|
|
skip_if_microversion_lt = utils.skip_if_microversion_lt
|
|
|
|
|
|
class BaseSharesTest(test.BaseTestCase):
|
|
"""Base test case class for all Manila API tests."""
|
|
|
|
credentials = ('primary', )
|
|
force_tenant_isolation = False
|
|
protocols = ["nfs", "cifs", "glusterfs", "hdfs", "cephfs"]
|
|
|
|
# Will be cleaned up in resource_cleanup
|
|
class_resources = []
|
|
|
|
# Will be cleaned up in tearDown method
|
|
method_resources = []
|
|
|
|
# Will be cleaned up in resource_cleanup
|
|
class_isolated_creds = []
|
|
|
|
# Will be cleaned up in tearDown method
|
|
method_isolated_creds = []
|
|
|
|
def skip_if_microversion_not_supported(self, microversion):
|
|
if not utils.is_microversion_supported(microversion):
|
|
raise self.skipException(
|
|
"Microversion '%s' is not supported." % microversion)
|
|
|
|
def skip_if_microversion_lt(self, microversion):
|
|
if utils.is_microversion_lt(CONF.share.max_api_microversion,
|
|
microversion):
|
|
raise self.skipException(
|
|
"Microversion must be greater than or equal to '%s'." %
|
|
microversion)
|
|
|
|
@classmethod
|
|
def get_client_with_isolated_creds(cls,
|
|
name=None,
|
|
type_of_creds="admin",
|
|
cleanup_in_class=False,
|
|
client_version='1'):
|
|
"""Creates isolated creds.
|
|
|
|
:param name: name, will be used for naming ic and related stuff
|
|
:param type_of_creds: admin, alt or primary
|
|
:param cleanup_in_class: defines place where to delete
|
|
:returns: SharesClient -- shares client with isolated creds.
|
|
:returns: To client added dict attr 'creds' with
|
|
:returns: key elements 'tenant' and 'user'.
|
|
"""
|
|
if name is None:
|
|
# Get name of test method
|
|
name = inspect.stack()[1][3]
|
|
if len(name) > 32:
|
|
name = name[0:32]
|
|
|
|
# Choose type of isolated creds
|
|
ic = dynamic_creds.DynamicCredentialProvider(
|
|
identity_version=CONF.identity.auth_version,
|
|
name=name,
|
|
admin_role=CONF.identity.admin_role,
|
|
admin_creds=common_creds.get_configured_admin_credentials())
|
|
if "admin" in type_of_creds:
|
|
creds = ic.get_admin_creds().credentials
|
|
elif "alt" in type_of_creds:
|
|
creds = ic.get_alt_creds().credentials
|
|
else:
|
|
creds = ic.get_credentials(type_of_creds).credentials
|
|
ic.type_of_creds = type_of_creds
|
|
|
|
# create client with isolated creds
|
|
os = clients.Manager(credentials=creds)
|
|
if client_version == '1':
|
|
client = shares_client.SharesClient(os.auth_provider)
|
|
elif client_version == '2':
|
|
client = shares_v2_client.SharesV2Client(os.auth_provider)
|
|
|
|
# Set place where will be deleted isolated creds
|
|
ic_res = {
|
|
"method": ic.clear_creds,
|
|
"deleted": False,
|
|
}
|
|
if cleanup_in_class:
|
|
cls.class_isolated_creds.insert(0, ic_res)
|
|
else:
|
|
cls.method_isolated_creds.insert(0, ic_res)
|
|
|
|
# Provide share network
|
|
if CONF.share.multitenancy_enabled:
|
|
if (not CONF.service_available.neutron and
|
|
CONF.share.create_networks_when_multitenancy_enabled):
|
|
raise cls.skipException("Neutron support is required")
|
|
nc = os.networks_client
|
|
share_network_id = cls.provide_share_network(client, nc, ic)
|
|
client.share_network_id = share_network_id
|
|
resource = {
|
|
"type": "share_network",
|
|
"id": client.share_network_id,
|
|
"client": client,
|
|
}
|
|
if cleanup_in_class:
|
|
cls.class_resources.insert(0, resource)
|
|
else:
|
|
cls.method_resources.insert(0, resource)
|
|
return client
|
|
|
|
@classmethod
|
|
def verify_nonempty(cls, *args):
|
|
if not all(args):
|
|
msg = "Missing API credentials in configuration."
|
|
raise cls.skipException(msg)
|
|
|
|
@classmethod
|
|
def setup_clients(cls):
|
|
super(BaseSharesTest, cls).setup_clients()
|
|
os = getattr(cls, 'os_%s' % cls.credentials[0])
|
|
os.shares_client = shares_client.SharesClient(os.auth_provider)
|
|
|
|
if CONF.identity.auth_version == 'v3':
|
|
project_id = os.auth_provider.auth_data[1]['project']['id']
|
|
else:
|
|
project_id = os.auth_provider.auth_data[1]['token']['tenant']['id']
|
|
cls.tenant_id = project_id
|
|
cls.user_id = os.auth_provider.auth_data[1]['user']['id']
|
|
|
|
cls.shares_client = os.shares_client
|
|
os.shares_v2_client = shares_v2_client.SharesV2Client(
|
|
os.auth_provider)
|
|
cls.shares_v2_client = os.shares_v2_client
|
|
if CONF.share.multitenancy_enabled:
|
|
if (not CONF.service_available.neutron and
|
|
CONF.share.create_networks_when_multitenancy_enabled):
|
|
raise cls.skipException("Neutron support is required")
|
|
share_network_id = cls.provide_share_network(
|
|
cls.shares_v2_client, os.networks_client)
|
|
cls.shares_client.share_network_id = share_network_id
|
|
cls.shares_v2_client.share_network_id = share_network_id
|
|
|
|
@classmethod
|
|
def resource_setup(cls):
|
|
if not (any(p in CONF.share.enable_protocols
|
|
for p in cls.protocols) and
|
|
CONF.service_available.manila):
|
|
skip_msg = "Manila is disabled"
|
|
raise cls.skipException(skip_msg)
|
|
super(BaseSharesTest, cls).resource_setup()
|
|
|
|
def setUp(self):
|
|
super(BaseSharesTest, self).setUp()
|
|
self.addCleanup(self.clear_isolated_creds)
|
|
self.addCleanup(self.clear_resources)
|
|
verify_test_has_appropriate_tags(self)
|
|
|
|
@classmethod
|
|
def resource_cleanup(cls):
|
|
cls.clear_resources(cls.class_resources)
|
|
cls.clear_isolated_creds(cls.class_isolated_creds)
|
|
super(BaseSharesTest, cls).resource_cleanup()
|
|
|
|
@classmethod
|
|
@network_synchronized
|
|
def provide_share_network(cls, shares_client, networks_client,
|
|
isolated_creds_client=None,
|
|
ignore_multitenancy_config=False):
|
|
"""Used for finding/creating share network for multitenant driver.
|
|
|
|
This method creates/gets entity share-network for one tenant. This
|
|
share-network will be used for creation of service vm.
|
|
|
|
:param shares_client: shares client, which requires share-network
|
|
:param networks_client: network client from same tenant as shares
|
|
:param isolated_creds_client: DynamicCredentialProvider instance
|
|
If provided, then its networking will be used if needed.
|
|
If not provided, then common network will be used if needed.
|
|
:param ignore_multitenancy_config: provide a share network regardless
|
|
of 'multitenancy_enabled' configuration value.
|
|
:returns: str -- share network id for shares_client tenant
|
|
:returns: None -- if single-tenant driver used
|
|
"""
|
|
|
|
sc = shares_client
|
|
search_word = "reusable"
|
|
sn_name = "autogenerated_by_tempest_%s" % search_word
|
|
|
|
if (not ignore_multitenancy_config and
|
|
not CONF.share.multitenancy_enabled):
|
|
# Assumed usage of a single-tenant driver
|
|
share_network_id = None
|
|
else:
|
|
if sc.share_network_id:
|
|
# Share-network already exists, use it
|
|
share_network_id = sc.share_network_id
|
|
elif not CONF.share.create_networks_when_multitenancy_enabled:
|
|
share_network_id = None
|
|
|
|
# Try get suitable share-network
|
|
share_networks = sc.list_share_networks_with_detail()
|
|
for sn in share_networks:
|
|
if (sn["neutron_net_id"] is None and
|
|
sn["neutron_subnet_id"] is None and
|
|
sn["name"] and search_word in sn["name"]):
|
|
share_network_id = sn["id"]
|
|
break
|
|
|
|
# Create new share-network if one was not found
|
|
if share_network_id is None:
|
|
sn_desc = "This share-network was created by tempest"
|
|
sn = sc.create_share_network(name=sn_name,
|
|
description=sn_desc)
|
|
share_network_id = sn["id"]
|
|
else:
|
|
net_id = subnet_id = share_network_id = None
|
|
|
|
if not isolated_creds_client:
|
|
# Search for networks, created in previous runs
|
|
service_net_name = "share-service"
|
|
networks = networks_client.list_networks()
|
|
if "networks" in networks.keys():
|
|
networks = networks["networks"]
|
|
for network in networks:
|
|
if (service_net_name in network["name"] and
|
|
sc.tenant_id == network['tenant_id']):
|
|
net_id = network["id"]
|
|
if len(network["subnets"]) > 0:
|
|
subnet_id = network["subnets"][0]
|
|
break
|
|
|
|
# Create suitable network
|
|
if net_id is None or subnet_id is None:
|
|
ic = dynamic_creds.DynamicCredentialProvider(
|
|
identity_version=CONF.identity.auth_version,
|
|
name=service_net_name,
|
|
admin_role=CONF.identity.admin_role,
|
|
admin_creds=(
|
|
common_creds.
|
|
get_configured_admin_credentials()))
|
|
net_data = ic._create_network_resources(sc.tenant_id)
|
|
network, subnet, router = net_data
|
|
net_id = network["id"]
|
|
subnet_id = subnet["id"]
|
|
|
|
# Try get suitable share-network
|
|
share_networks = sc.list_share_networks_with_detail()
|
|
for sn in share_networks:
|
|
if (net_id == sn["neutron_net_id"] and
|
|
subnet_id == sn["neutron_subnet_id"] and
|
|
sn["name"] and search_word in sn["name"]):
|
|
share_network_id = sn["id"]
|
|
break
|
|
else:
|
|
sn_name = "autogenerated_by_tempest_for_isolated_creds"
|
|
# Use precreated network and subnet from isolated creds
|
|
net_id = isolated_creds_client.get_credentials(
|
|
isolated_creds_client.type_of_creds).network['id']
|
|
subnet_id = isolated_creds_client.get_credentials(
|
|
isolated_creds_client.type_of_creds).subnet['id']
|
|
|
|
# Create suitable share-network
|
|
if share_network_id is None:
|
|
sn_desc = "This share-network was created by tempest"
|
|
sn = sc.create_share_network(name=sn_name,
|
|
description=sn_desc,
|
|
neutron_net_id=net_id,
|
|
neutron_subnet_id=subnet_id)
|
|
share_network_id = sn["id"]
|
|
|
|
return share_network_id
|
|
|
|
@classmethod
|
|
def _create_share(cls, share_protocol=None, size=None, name=None,
|
|
snapshot_id=None, description=None, metadata=None,
|
|
share_network_id=None, share_type_id=None,
|
|
consistency_group_id=None, client=None,
|
|
cleanup_in_class=True, is_public=False, **kwargs):
|
|
client = client or cls.shares_v2_client
|
|
description = description or "Tempest's share"
|
|
share_network_id = share_network_id or client.share_network_id or None
|
|
metadata = metadata or {}
|
|
size = size or CONF.share.share_size
|
|
kwargs.update({
|
|
'share_protocol': share_protocol,
|
|
'size': size,
|
|
'name': name,
|
|
'snapshot_id': snapshot_id,
|
|
'description': description,
|
|
'metadata': metadata,
|
|
'share_network_id': share_network_id,
|
|
'share_type_id': share_type_id,
|
|
'is_public': is_public,
|
|
})
|
|
if consistency_group_id:
|
|
kwargs['consistency_group_id'] = consistency_group_id
|
|
|
|
share = client.create_share(**kwargs)
|
|
resource = {"type": "share", "id": share["id"], "client": client,
|
|
"consistency_group_id": consistency_group_id}
|
|
cleanup_list = (cls.class_resources if cleanup_in_class else
|
|
cls.method_resources)
|
|
cleanup_list.insert(0, resource)
|
|
return share
|
|
|
|
@classmethod
|
|
def migrate_share(
|
|
cls, share_id, dest_host, wait_for_status, client=None,
|
|
force_host_assisted_migration=False, new_share_network_id=None,
|
|
new_share_type_id=None, **kwargs):
|
|
client = client or cls.shares_v2_client
|
|
client.migrate_share(
|
|
share_id, dest_host,
|
|
force_host_assisted_migration=force_host_assisted_migration,
|
|
new_share_network_id=new_share_network_id,
|
|
writable=False, preserve_metadata=False, nondisruptive=False,
|
|
new_share_type_id=new_share_type_id, **kwargs)
|
|
share = client.wait_for_migration_status(
|
|
share_id, dest_host, wait_for_status, **kwargs)
|
|
return share
|
|
|
|
@classmethod
|
|
def migration_complete(cls, share_id, dest_host, client=None, **kwargs):
|
|
client = client or cls.shares_v2_client
|
|
client.migration_complete(share_id, **kwargs)
|
|
share = client.wait_for_migration_status(
|
|
share_id, dest_host, 'migration_success', **kwargs)
|
|
return share
|
|
|
|
@classmethod
|
|
def migration_cancel(cls, share_id, dest_host, client=None, **kwargs):
|
|
client = client or cls.shares_v2_client
|
|
client.migration_cancel(share_id, **kwargs)
|
|
share = client.wait_for_migration_status(
|
|
share_id, dest_host, 'migration_cancelled', **kwargs)
|
|
return share
|
|
|
|
@classmethod
|
|
def create_share(cls, *args, **kwargs):
|
|
"""Create one share and wait for available state. Retry if allowed."""
|
|
result = cls.create_shares([{"args": args, "kwargs": kwargs}])
|
|
return result[0]
|
|
|
|
@classmethod
|
|
def create_shares(cls, share_data_list):
|
|
"""Creates several shares in parallel with retries.
|
|
|
|
Use this method when you want to create more than one share at same
|
|
time. Especially if config option 'share.share_creation_retry_number'
|
|
has value more than zero (0).
|
|
All shares will be expected to have 'available' status with or without
|
|
recreation else error will be raised.
|
|
|
|
:param share_data_list: list -- list of dictionaries with 'args' and
|
|
'kwargs' for '_create_share' method of this base class.
|
|
example of data:
|
|
share_data_list=[{'args': ['quuz'], 'kwargs': {'foo': 'bar'}}}]
|
|
:returns: list -- list of shares created using provided data.
|
|
"""
|
|
|
|
for d in share_data_list:
|
|
if not isinstance(d, dict):
|
|
raise exceptions.TempestException(
|
|
"Expected 'dict', got '%s'" % type(d))
|
|
if "args" not in d:
|
|
d["args"] = []
|
|
if "kwargs" not in d:
|
|
d["kwargs"] = {}
|
|
if len(d) > 2:
|
|
raise exceptions.TempestException(
|
|
"Expected only 'args' and 'kwargs' keys. "
|
|
"Provided %s" % list(d))
|
|
|
|
data = []
|
|
for d in share_data_list:
|
|
client = d["kwargs"].pop("client", cls.shares_v2_client)
|
|
local_d = {
|
|
"args": d["args"],
|
|
"kwargs": copy.deepcopy(d["kwargs"]),
|
|
}
|
|
local_d["kwargs"]["client"] = client
|
|
local_d["share"] = cls._create_share(
|
|
*local_d["args"], **local_d["kwargs"])
|
|
local_d["cnt"] = 0
|
|
local_d["available"] = False
|
|
data.append(local_d)
|
|
|
|
while not all(d["available"] for d in data):
|
|
for d in data:
|
|
if d["available"]:
|
|
continue
|
|
client = d["kwargs"]["client"]
|
|
share_id = d["share"]["id"]
|
|
try:
|
|
client.wait_for_share_status(share_id, "available")
|
|
d["available"] = True
|
|
except (share_exceptions.ShareBuildErrorException,
|
|
exceptions.TimeoutException) as e:
|
|
if CONF.share.share_creation_retry_number > d["cnt"]:
|
|
d["cnt"] += 1
|
|
msg = ("Share '%s' failed to be built. "
|
|
"Trying create another." % share_id)
|
|
LOG.error(msg)
|
|
LOG.error(e)
|
|
cg_id = d["kwargs"].get("consistency_group_id")
|
|
if cg_id:
|
|
# NOTE(vponomaryov): delete errored share
|
|
# immediately in case share is part of CG.
|
|
client.delete_share(
|
|
share_id,
|
|
params={"consistency_group_id": cg_id})
|
|
client.wait_for_resource_deletion(
|
|
share_id=share_id)
|
|
d["share"] = cls._create_share(
|
|
*d["args"], **d["kwargs"])
|
|
else:
|
|
raise
|
|
|
|
return [d["share"] for d in data]
|
|
|
|
@classmethod
|
|
def create_consistency_group(cls, client=None, cleanup_in_class=True,
|
|
share_network_id=None, **kwargs):
|
|
client = client or cls.shares_v2_client
|
|
if kwargs.get('source_cgsnapshot_id') is None:
|
|
kwargs['share_network_id'] = (share_network_id or
|
|
client.share_network_id or None)
|
|
consistency_group = client.create_consistency_group(**kwargs)
|
|
resource = {
|
|
"type": "consistency_group",
|
|
"id": consistency_group["id"],
|
|
"client": client}
|
|
if cleanup_in_class:
|
|
cls.class_resources.insert(0, resource)
|
|
else:
|
|
cls.method_resources.insert(0, resource)
|
|
|
|
if kwargs.get('source_cgsnapshot_id'):
|
|
new_cg_shares = client.list_shares(
|
|
detailed=True,
|
|
params={'consistency_group_id': consistency_group['id']})
|
|
|
|
for share in new_cg_shares:
|
|
resource = {"type": "share",
|
|
"id": share["id"],
|
|
"client": client,
|
|
"consistency_group_id": share.get(
|
|
'consistency_group_id')}
|
|
if cleanup_in_class:
|
|
cls.class_resources.insert(0, resource)
|
|
else:
|
|
cls.method_resources.insert(0, resource)
|
|
|
|
client.wait_for_consistency_group_status(consistency_group['id'],
|
|
'available')
|
|
return consistency_group
|
|
|
|
@classmethod
|
|
def create_snapshot_wait_for_active(cls, share_id, name=None,
|
|
description=None, force=False,
|
|
client=None, cleanup_in_class=True):
|
|
if client is None:
|
|
client = cls.shares_v2_client
|
|
if description is None:
|
|
description = "Tempest's snapshot"
|
|
snapshot = client.create_snapshot(share_id, name, description, force)
|
|
resource = {
|
|
"type": "snapshot",
|
|
"id": snapshot["id"],
|
|
"client": client,
|
|
}
|
|
if cleanup_in_class:
|
|
cls.class_resources.insert(0, resource)
|
|
else:
|
|
cls.method_resources.insert(0, resource)
|
|
client.wait_for_snapshot_status(snapshot["id"], "available")
|
|
return snapshot
|
|
|
|
@classmethod
|
|
def create_cgsnapshot_wait_for_active(cls, consistency_group_id,
|
|
name=None, description=None,
|
|
client=None, cleanup_in_class=True,
|
|
**kwargs):
|
|
client = client or cls.shares_v2_client
|
|
if description is None:
|
|
description = "Tempest's cgsnapshot"
|
|
cgsnapshot = client.create_cgsnapshot(consistency_group_id,
|
|
name=name,
|
|
description=description,
|
|
**kwargs)
|
|
resource = {
|
|
"type": "cgsnapshot",
|
|
"id": cgsnapshot["id"],
|
|
"client": client,
|
|
}
|
|
if cleanup_in_class:
|
|
cls.class_resources.insert(0, resource)
|
|
else:
|
|
cls.method_resources.insert(0, resource)
|
|
client.wait_for_cgsnapshot_status(cgsnapshot["id"], "available")
|
|
return cgsnapshot
|
|
|
|
@classmethod
|
|
def get_availability_zones(cls, client=None):
|
|
"""List the availability zones for "manila-share" services
|
|
|
|
that are currently in "up" state.
|
|
"""
|
|
client = client or cls.shares_v2_client
|
|
cls.services = client.list_services()
|
|
zones = [service['zone'] for service in cls.services if
|
|
service['binary'] == "manila-share" and
|
|
service['state'] == 'up']
|
|
return zones
|
|
|
|
def get_pools_for_replication_domain(self):
|
|
# Get the list of pools for the replication domain
|
|
pools = self.admin_client.list_pools(detail=True)['pools']
|
|
instance_host = self.shares[0]['host']
|
|
host_pool = [p for p in pools if p['name'] == instance_host][0]
|
|
rep_domain = host_pool['capabilities']['replication_domain']
|
|
pools_in_rep_domain = [p for p in pools if p['capabilities'][
|
|
'replication_domain'] == rep_domain]
|
|
return rep_domain, pools_in_rep_domain
|
|
|
|
@classmethod
|
|
def create_share_replica(cls, share_id, availability_zone, client=None,
|
|
cleanup_in_class=False, cleanup=True):
|
|
client = client or cls.shares_v2_client
|
|
replica = client.create_share_replica(share_id, availability_zone)
|
|
resource = {
|
|
"type": "share_replica",
|
|
"id": replica["id"],
|
|
"client": client,
|
|
"share_id": share_id,
|
|
}
|
|
# NOTE(Yogi1): Cleanup needs to be disabled during promotion tests.
|
|
if cleanup:
|
|
if cleanup_in_class:
|
|
cls.class_resources.insert(0, resource)
|
|
else:
|
|
cls.method_resources.insert(0, resource)
|
|
client.wait_for_share_replica_status(
|
|
replica["id"], constants.STATUS_AVAILABLE)
|
|
return replica
|
|
|
|
@classmethod
|
|
def delete_share_replica(cls, replica_id, client=None):
|
|
client = client or cls.shares_v2_client
|
|
try:
|
|
client.delete_share_replica(replica_id)
|
|
client.wait_for_resource_deletion(replica_id=replica_id)
|
|
except exceptions.NotFound:
|
|
pass
|
|
|
|
@classmethod
|
|
def promote_share_replica(cls, replica_id, client=None):
|
|
client = client or cls.shares_v2_client
|
|
replica = client.promote_share_replica(replica_id)
|
|
client.wait_for_share_replica_status(
|
|
replica["id"],
|
|
constants.REPLICATION_STATE_ACTIVE,
|
|
status_attr="replica_state")
|
|
return replica
|
|
|
|
@classmethod
|
|
def create_share_network(cls, client=None,
|
|
cleanup_in_class=False, **kwargs):
|
|
if client is None:
|
|
client = cls.shares_client
|
|
share_network = client.create_share_network(**kwargs)
|
|
resource = {
|
|
"type": "share_network",
|
|
"id": share_network["id"],
|
|
"client": client,
|
|
}
|
|
if cleanup_in_class:
|
|
cls.class_resources.insert(0, resource)
|
|
else:
|
|
cls.method_resources.insert(0, resource)
|
|
return share_network
|
|
|
|
@classmethod
|
|
def create_security_service(cls, ss_type="ldap", client=None,
|
|
cleanup_in_class=False, **kwargs):
|
|
if client is None:
|
|
client = cls.shares_client
|
|
security_service = client.create_security_service(ss_type, **kwargs)
|
|
resource = {
|
|
"type": "security_service",
|
|
"id": security_service["id"],
|
|
"client": client,
|
|
}
|
|
if cleanup_in_class:
|
|
cls.class_resources.insert(0, resource)
|
|
else:
|
|
cls.method_resources.insert(0, resource)
|
|
return security_service
|
|
|
|
@classmethod
|
|
def create_share_type(cls, name, is_public=True, client=None,
|
|
cleanup_in_class=True, **kwargs):
|
|
if client is None:
|
|
client = cls.shares_v2_client
|
|
share_type = client.create_share_type(name, is_public, **kwargs)
|
|
resource = {
|
|
"type": "share_type",
|
|
"id": share_type["share_type"]["id"],
|
|
"client": client,
|
|
}
|
|
if cleanup_in_class:
|
|
cls.class_resources.insert(0, resource)
|
|
else:
|
|
cls.method_resources.insert(0, resource)
|
|
return share_type
|
|
|
|
@staticmethod
|
|
def add_required_extra_specs_to_dict(extra_specs=None):
|
|
dhss = six.text_type(CONF.share.multitenancy_enabled)
|
|
snapshot_support = six.text_type(
|
|
CONF.share.capability_snapshot_support)
|
|
required = {
|
|
"driver_handles_share_servers": dhss,
|
|
"snapshot_support": snapshot_support,
|
|
}
|
|
if extra_specs:
|
|
required.update(extra_specs)
|
|
return required
|
|
|
|
@classmethod
|
|
def clear_isolated_creds(cls, creds=None):
|
|
if creds is None:
|
|
creds = cls.method_isolated_creds
|
|
for ic in creds:
|
|
if "deleted" not in ic.keys():
|
|
ic["deleted"] = False
|
|
if not ic["deleted"]:
|
|
with handle_cleanup_exceptions():
|
|
ic["method"]()
|
|
ic["deleted"] = True
|
|
|
|
@classmethod
|
|
def clear_share_replicas(cls, share_id, client=None):
|
|
client = client or cls.shares_v2_client
|
|
share_replicas = client.list_share_replicas(
|
|
share_id=share_id)
|
|
|
|
for replica in share_replicas:
|
|
try:
|
|
cls.delete_share_replica(replica['id'])
|
|
except exceptions.BadRequest:
|
|
# Ignore the exception due to deletion of last active replica
|
|
pass
|
|
|
|
@classmethod
|
|
def clear_resources(cls, resources=None):
|
|
"""Deletes resources, that were created in test suites.
|
|
|
|
This method tries to remove resources from resource list,
|
|
if it is not found, assumed it was deleted in test itself.
|
|
It is expected, that all resources were added as LIFO
|
|
due to restriction of deletion resources, that is in the chain.
|
|
|
|
:param resources: dict with keys 'type','id','client' and 'deleted'
|
|
"""
|
|
|
|
if resources is None:
|
|
resources = cls.method_resources
|
|
for res in resources:
|
|
if "deleted" not in res.keys():
|
|
res["deleted"] = False
|
|
if "client" not in res.keys():
|
|
res["client"] = cls.shares_client
|
|
if not(res["deleted"]):
|
|
res_id = res['id']
|
|
client = res["client"]
|
|
with handle_cleanup_exceptions():
|
|
if res["type"] is "share":
|
|
cls.clear_share_replicas(res_id)
|
|
cg_id = res.get('consistency_group_id')
|
|
if cg_id:
|
|
params = {'consistency_group_id': cg_id}
|
|
client.delete_share(res_id, params=params)
|
|
else:
|
|
client.delete_share(res_id)
|
|
client.wait_for_resource_deletion(share_id=res_id)
|
|
elif res["type"] is "snapshot":
|
|
client.delete_snapshot(res_id)
|
|
client.wait_for_resource_deletion(snapshot_id=res_id)
|
|
elif res["type"] is "share_network":
|
|
client.delete_share_network(res_id)
|
|
client.wait_for_resource_deletion(sn_id=res_id)
|
|
elif res["type"] is "security_service":
|
|
client.delete_security_service(res_id)
|
|
client.wait_for_resource_deletion(ss_id=res_id)
|
|
elif res["type"] is "share_type":
|
|
client.delete_share_type(res_id)
|
|
client.wait_for_resource_deletion(st_id=res_id)
|
|
elif res["type"] is "consistency_group":
|
|
client.delete_consistency_group(res_id)
|
|
client.wait_for_resource_deletion(cg_id=res_id)
|
|
elif res["type"] is "cgsnapshot":
|
|
client.delete_cgsnapshot(res_id)
|
|
client.wait_for_resource_deletion(cgsnapshot_id=res_id)
|
|
elif res["type"] is "share_replica":
|
|
client.delete_share_replica(res_id)
|
|
client.wait_for_resource_deletion(replica_id=res_id)
|
|
else:
|
|
LOG.warning("Provided unsupported resource type for "
|
|
"cleanup '%s'. Skipping." % res["type"])
|
|
res["deleted"] = True
|
|
|
|
@classmethod
|
|
def generate_share_network_data(self):
|
|
data = {
|
|
"name": data_utils.rand_name("sn-name"),
|
|
"description": data_utils.rand_name("sn-desc"),
|
|
"neutron_net_id": data_utils.rand_name("net-id"),
|
|
"neutron_subnet_id": data_utils.rand_name("subnet-id"),
|
|
}
|
|
return data
|
|
|
|
@classmethod
|
|
def generate_security_service_data(self):
|
|
data = {
|
|
"name": data_utils.rand_name("ss-name"),
|
|
"description": data_utils.rand_name("ss-desc"),
|
|
"dns_ip": utils.rand_ip(),
|
|
"server": utils.rand_ip(),
|
|
"domain": data_utils.rand_name("ss-domain"),
|
|
"user": data_utils.rand_name("ss-user"),
|
|
"password": data_utils.rand_name("ss-password"),
|
|
}
|
|
return data
|
|
|
|
# Useful assertions
|
|
def assertDictMatch(self, d1, d2, approx_equal=False, tolerance=0.001):
|
|
"""Assert two dicts are equivalent.
|
|
|
|
This is a 'deep' match in the sense that it handles nested
|
|
dictionaries appropriately.
|
|
|
|
NOTE:
|
|
|
|
If you don't care (or don't know) a given value, you can specify
|
|
the string DONTCARE as the value. This will cause that dict-item
|
|
to be skipped.
|
|
|
|
"""
|
|
def raise_assertion(msg):
|
|
d1str = str(d1)
|
|
d2str = str(d2)
|
|
base_msg = ('Dictionaries do not match. %(msg)s d1: %(d1str)s '
|
|
'd2: %(d2str)s' %
|
|
{"msg": msg, "d1str": d1str, "d2str": d2str})
|
|
raise AssertionError(base_msg)
|
|
|
|
d1keys = set(d1.keys())
|
|
d2keys = set(d2.keys())
|
|
if d1keys != d2keys:
|
|
d1only = d1keys - d2keys
|
|
d2only = d2keys - d1keys
|
|
raise_assertion('Keys in d1 and not d2: %(d1only)s. '
|
|
'Keys in d2 and not d1: %(d2only)s' %
|
|
{"d1only": d1only, "d2only": d2only})
|
|
|
|
for key in d1keys:
|
|
d1value = d1[key]
|
|
d2value = d2[key]
|
|
try:
|
|
error = abs(float(d1value) - float(d2value))
|
|
within_tolerance = error <= tolerance
|
|
except (ValueError, TypeError):
|
|
# If both values aren't convertible to float, just ignore
|
|
# ValueError if arg is a str, TypeError if it's something else
|
|
# (like None)
|
|
within_tolerance = False
|
|
|
|
if hasattr(d1value, 'keys') and hasattr(d2value, 'keys'):
|
|
self.assertDictMatch(d1value, d2value)
|
|
elif 'DONTCARE' in (d1value, d2value):
|
|
continue
|
|
elif approx_equal and within_tolerance:
|
|
continue
|
|
elif d1value != d2value:
|
|
raise_assertion("d1['%(key)s']=%(d1value)s != "
|
|
"d2['%(key)s']=%(d2value)s" %
|
|
{
|
|
"key": key,
|
|
"d1value": d1value,
|
|
"d2value": d2value
|
|
})
|
|
|
|
|
|
class BaseSharesAltTest(BaseSharesTest):
|
|
"""Base test case class for all Shares Alt API tests."""
|
|
credentials = ('alt', )
|
|
|
|
|
|
class BaseSharesAdminTest(BaseSharesTest):
|
|
"""Base test case class for all Shares Admin API tests."""
|
|
credentials = ('admin', )
|
|
|
|
|
|
class BaseSharesMixedTest(BaseSharesTest):
|
|
"""Base test case class for all Shares API tests with all user roles."""
|
|
credentials = ('primary', 'alt', 'admin')
|
|
|
|
@classmethod
|
|
def setup_clients(cls):
|
|
super(BaseSharesMixedTest, cls).setup_clients()
|
|
cls.admin_shares_client = shares_client.SharesClient(
|
|
cls.os_admin.auth_provider)
|
|
cls.admin_shares_v2_client = shares_v2_client.SharesV2Client(
|
|
cls.os_admin.auth_provider)
|
|
cls.alt_shares_client = shares_client.SharesClient(
|
|
cls.os_alt.auth_provider)
|
|
cls.alt_shares_v2_client = shares_v2_client.SharesV2Client(
|
|
cls.os_alt.auth_provider)
|
|
|
|
if CONF.share.multitenancy_enabled:
|
|
admin_share_network_id = cls.provide_share_network(
|
|
cls.admin_shares_v2_client, cls.os_admin.networks_client)
|
|
cls.admin_shares_client.share_network_id = admin_share_network_id
|
|
cls.admin_shares_v2_client.share_network_id = (
|
|
admin_share_network_id)
|
|
|
|
alt_share_network_id = cls.provide_share_network(
|
|
cls.alt_shares_v2_client, cls.os_alt.networks_client)
|
|
cls.alt_shares_client.share_network_id = alt_share_network_id
|
|
cls.alt_shares_v2_client.share_network_id = alt_share_network_id
|