Merge "Rename Bay to Cluster in api"
This commit is contained in:
commit
01fa411521
|
@ -20,6 +20,21 @@
|
|||
"baymodel:update": "rule:default",
|
||||
"baymodel:publish": "rule:admin_or_owner",
|
||||
|
||||
"cluster:create": "rule:default",
|
||||
"cluster:delete": "rule:default",
|
||||
"cluster:detail": "rule:default",
|
||||
"cluster:get": "rule:default",
|
||||
"cluster:get_all": "rule:default",
|
||||
"cluster:update": "rule:default",
|
||||
|
||||
"clustertemplate:create": "rule:default",
|
||||
"clustertemplate:delete": "rule:default",
|
||||
"clustertemplate:detail": "rule:default",
|
||||
"clustertemplate:get": "rule:default",
|
||||
"clustertemplate:get_all": "rule:default",
|
||||
"clustertemplate:update": "rule:default",
|
||||
"clustertemplate:publish": "rule:admin_or_owner",
|
||||
|
||||
"rc:create": "rule:default",
|
||||
"rc:delete": "rule:default",
|
||||
"rc:detail": "rule:default",
|
||||
|
|
|
@ -69,7 +69,7 @@ class Root(base.APIBase):
|
|||
root = Root()
|
||||
root.name = "OpenStack Magnum API"
|
||||
root.description = ("Magnum is an OpenStack project which aims to "
|
||||
"provide container management.")
|
||||
"provide container cluster management.")
|
||||
root.versions = [Version.convert('v1', "CURRENT",
|
||||
versions.CURRENT_MAX_VER,
|
||||
versions.BASE_VER)]
|
||||
|
|
|
@ -27,6 +27,8 @@ from magnum.api.controllers import link
|
|||
from magnum.api.controllers.v1 import bay
|
||||
from magnum.api.controllers.v1 import baymodel
|
||||
from magnum.api.controllers.v1 import certificate
|
||||
from magnum.api.controllers.v1 import cluster
|
||||
from magnum.api.controllers.v1 import cluster_template
|
||||
from magnum.api.controllers.v1 import magnum_services
|
||||
from magnum.api.controllers import versions as ver
|
||||
from magnum.api import expose
|
||||
|
@ -77,6 +79,12 @@ class V1(controllers_base.APIBase):
|
|||
bays = [link.Link]
|
||||
"""Links to the bays resource"""
|
||||
|
||||
clustertemplates = [link.Link]
|
||||
"""Links to the clustertemplates resource"""
|
||||
|
||||
clusters = [link.Link]
|
||||
"""Links to the clusters resource"""
|
||||
|
||||
certificates = [link.Link]
|
||||
"""Links to the certificates resource"""
|
||||
|
||||
|
@ -108,6 +116,19 @@ class V1(controllers_base.APIBase):
|
|||
pecan.request.host_url,
|
||||
'bays', '',
|
||||
bookmark=True)]
|
||||
v1.clustertemplates = [link.Link.make_link('self',
|
||||
pecan.request.host_url,
|
||||
'clustertemplates', ''),
|
||||
link.Link.make_link('bookmark',
|
||||
pecan.request.host_url,
|
||||
'clustertemplates', '',
|
||||
bookmark=True)]
|
||||
v1.clusters = [link.Link.make_link('self', pecan.request.host_url,
|
||||
'clusters', ''),
|
||||
link.Link.make_link('bookmark',
|
||||
pecan.request.host_url,
|
||||
'clusters', '',
|
||||
bookmark=True)]
|
||||
v1.certificates = [link.Link.make_link('self', pecan.request.host_url,
|
||||
'certificates', ''),
|
||||
link.Link.make_link('bookmark',
|
||||
|
@ -128,6 +149,8 @@ class Controller(controllers_base.Controller):
|
|||
|
||||
bays = bay.BaysController()
|
||||
baymodels = baymodel.BayModelsController()
|
||||
clusters = cluster.ClustersController()
|
||||
clustertemplates = cluster_template.ClusterTemplatesController()
|
||||
certificates = certificate.CertificateController()
|
||||
mservices = magnum_services.MagnumServiceController()
|
||||
|
||||
|
|
|
@ -80,7 +80,7 @@ class Bay(base.APIBase):
|
|||
try:
|
||||
baymodel = api_utils.get_resource('BayModel', value)
|
||||
self._baymodel_id = baymodel.uuid
|
||||
except exception.BayModelNotFound as e:
|
||||
except exception.ClusterTemplateNotFound as e:
|
||||
# Change error code because 404 (NotFound) is inappropriate
|
||||
# response for a POST request to create a Bay
|
||||
e.code = 400 # BadRequest
|
||||
|
|
|
@ -343,7 +343,7 @@ class BayModelsController(base.Controller):
|
|||
if baymodel_dict['public']:
|
||||
if not policy.enforce(context, "baymodel:publish", None,
|
||||
do_raise=False):
|
||||
raise exception.BaymodelPublishDenied()
|
||||
raise exception.ClusterTemplatePublishDenied()
|
||||
|
||||
# NOTE(yuywz): We will generate a random human-readable name for
|
||||
# baymodel if the name is not spcified by user.
|
||||
|
@ -386,7 +386,7 @@ class BayModelsController(base.Controller):
|
|||
if baymodel.public != new_baymodel.public:
|
||||
if not policy.enforce(context, "baymodel:publish", None,
|
||||
do_raise=False):
|
||||
raise exception.BaymodelPublishDenied()
|
||||
raise exception.ClusterTemplatePublishDenied()
|
||||
|
||||
# Update only the fields that have changed
|
||||
for field in objects.BayModel.fields:
|
||||
|
|
|
@ -48,7 +48,7 @@ class Certificate(base.APIBase):
|
|||
try:
|
||||
self._bay = api_utils.get_resource('Bay', value)
|
||||
self._bay_uuid = self._bay.uuid
|
||||
except exception.BayNotFound as e:
|
||||
except exception.ClusterNotFound as e:
|
||||
# Change error code because 404 (NotFound) is inappropriate
|
||||
# response for a POST request to create a Bay
|
||||
e.code = 400 # BadRequest
|
||||
|
|
|
@ -0,0 +1,468 @@
|
|||
# Copyright 2013 UnitedStack Inc.
|
||||
# All Rights Reserved.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
import uuid
|
||||
|
||||
from oslo_log import log as logging
|
||||
from oslo_utils import timeutils
|
||||
import pecan
|
||||
import wsme
|
||||
from wsme import types as wtypes
|
||||
|
||||
from magnum.api import attr_validator
|
||||
from magnum.api.controllers import base
|
||||
from magnum.api.controllers import link
|
||||
from magnum.api.controllers.v1 import collection
|
||||
from magnum.api.controllers.v1 import types
|
||||
from magnum.api import expose
|
||||
from magnum.api import utils as api_utils
|
||||
from magnum.api.validation import validate_bay_properties
|
||||
from magnum.common import clients
|
||||
from magnum.common import exception
|
||||
from magnum.common import name_generator
|
||||
from magnum.common import policy
|
||||
from magnum.i18n import _LW
|
||||
from magnum import objects
|
||||
from magnum.objects import fields
|
||||
|
||||
LOG = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class ClusterPatchType(types.JsonPatchType):
|
||||
@staticmethod
|
||||
def mandatory_attrs():
|
||||
return ['/cluster_template_id']
|
||||
|
||||
@staticmethod
|
||||
def internal_attrs():
|
||||
internal_attrs = ['/api_address', '/node_addresses',
|
||||
'/master_addresses', '/stack_id',
|
||||
'/ca_cert_ref', '/magnum_cert_ref',
|
||||
'/trust_id', '/trustee_user_name',
|
||||
'/trustee_password', '/trustee_user_id']
|
||||
return types.JsonPatchType.internal_attrs() + internal_attrs
|
||||
|
||||
|
||||
class ClusterID(wtypes.Base):
|
||||
"""API representation of a cluster ID
|
||||
|
||||
This class enforces type checking and value constraints, and converts
|
||||
between the internal object model and the API representation of a cluster
|
||||
ID.
|
||||
"""
|
||||
|
||||
uuid = types.uuid
|
||||
"""Unique UUID for this cluster"""
|
||||
|
||||
def __init__(self, uuid):
|
||||
self.uuid = uuid
|
||||
|
||||
|
||||
class Cluster(base.APIBase):
|
||||
"""API representation of a cluster.
|
||||
|
||||
This class enforces type checking and value constraints, and converts
|
||||
between the internal object model and the API representation of a bay.
|
||||
"""
|
||||
|
||||
_cluster_template_id = None
|
||||
|
||||
def _get_cluster_template_id(self):
|
||||
return self._cluster_template_id
|
||||
|
||||
def _set_cluster_template_id(self, value):
|
||||
if value and self._cluster_template_id != value:
|
||||
try:
|
||||
cluster_template = api_utils.get_resource('BayModel', value)
|
||||
self._cluster_template_id = cluster_template.uuid
|
||||
except exception.ClusterTemplateNotFound as e:
|
||||
# Change error code because 404 (NotFound) is inappropriate
|
||||
# response for a POST request to create a Bay
|
||||
e.code = 400 # BadRequest
|
||||
raise
|
||||
elif value == wtypes.Unset:
|
||||
self._cluster_template_id = wtypes.Unset
|
||||
|
||||
uuid = types.uuid
|
||||
"""Unique UUID for this cluster"""
|
||||
|
||||
name = wtypes.StringType(min_length=1, max_length=242,
|
||||
pattern='^[a-zA-Z][a-zA-Z0-9_.-]*$')
|
||||
"""Name of this cluster, max length is limited to 242 because of heat
|
||||
stack requires max length limit to 255, and Magnum amend a uuid length"""
|
||||
|
||||
cluster_template_id = wsme.wsproperty(wtypes.text,
|
||||
_get_cluster_template_id,
|
||||
_set_cluster_template_id,
|
||||
mandatory=True)
|
||||
"""The cluster_template UUID"""
|
||||
|
||||
node_count = wsme.wsattr(wtypes.IntegerType(minimum=1), default=1)
|
||||
"""The node count for this cluster. Default to 1 if not set"""
|
||||
|
||||
master_count = wsme.wsattr(wtypes.IntegerType(minimum=1), default=1)
|
||||
"""The number of master nodes for this cluster. Default to 1 if not set"""
|
||||
|
||||
create_timeout = wsme.wsattr(wtypes.IntegerType(minimum=0), default=60)
|
||||
"""Timeout for creating the cluster in minutes. Default to 60 if not set"""
|
||||
|
||||
links = wsme.wsattr([link.Link], readonly=True)
|
||||
"""A list containing a self link and associated cluster links"""
|
||||
|
||||
stack_id = wsme.wsattr(wtypes.text, readonly=True)
|
||||
"""Stack id of the heat stack"""
|
||||
|
||||
status = wtypes.Enum(str, *fields.BayStatus.ALL)
|
||||
"""Status of the cluster from the heat stack"""
|
||||
|
||||
status_reason = wtypes.text
|
||||
"""Status reason of the cluster from the heat stack"""
|
||||
|
||||
discovery_url = wtypes.text
|
||||
"""Url used for cluster node discovery"""
|
||||
|
||||
api_address = wsme.wsattr(wtypes.text, readonly=True)
|
||||
"""Api address of cluster master node"""
|
||||
|
||||
node_addresses = wsme.wsattr([wtypes.text], readonly=True)
|
||||
"""IP addresses of cluster agent nodes"""
|
||||
|
||||
master_addresses = wsme.wsattr([wtypes.text], readonly=True)
|
||||
"""IP addresses of cluster master nodes"""
|
||||
|
||||
faults = wsme.wsattr(wtypes.DictType(str, wtypes.text))
|
||||
"""Fault info collected from the heat resources of this cluster"""
|
||||
|
||||
def __init__(self, **kwargs):
|
||||
super(Cluster, self).__init__()
|
||||
self.fields = []
|
||||
for field in objects.Bay.fields:
|
||||
# Skip fields we do not expose.
|
||||
if not hasattr(self, field):
|
||||
continue
|
||||
self.fields.append(field)
|
||||
setattr(self, field, kwargs.get(field, wtypes.Unset))
|
||||
|
||||
# Set the renamed attributes for clusters
|
||||
self.fields.append('cluster_template_id')
|
||||
if 'cluster_template_id' in kwargs.keys():
|
||||
setattr(self, 'cluster_template_id',
|
||||
kwargs.get('cluster_template_id', wtypes.Unset))
|
||||
else:
|
||||
setattr(self, 'cluster_template_id', kwargs.get('baymodel_id',
|
||||
wtypes.Unset))
|
||||
|
||||
self.fields.append('create_timeout')
|
||||
if 'create_timeout' in kwargs.keys():
|
||||
setattr(self, 'create_timeout', kwargs.get('create_timeout',
|
||||
wtypes.Unset))
|
||||
else:
|
||||
setattr(self, 'create_timeout', kwargs.get('bay_create_timeout',
|
||||
wtypes.Unset))
|
||||
|
||||
self.fields.append('faults')
|
||||
if 'faults' in kwargs.keys():
|
||||
setattr(self, 'faults', kwargs.get('faults', wtypes.Unset))
|
||||
else:
|
||||
setattr(self, 'faults', kwargs.get('bay_faults', wtypes.Unset))
|
||||
|
||||
@staticmethod
|
||||
def _convert_with_links(cluster, url, expand=True):
|
||||
if not expand:
|
||||
cluster.unset_fields_except(['uuid', 'name', 'cluster_template_id',
|
||||
'node_count', 'status',
|
||||
'create_timeout', 'master_count',
|
||||
'stack_id'])
|
||||
|
||||
cluster.links = [link.Link.make_link('self', url,
|
||||
'bays', cluster.uuid),
|
||||
link.Link.make_link('bookmark', url,
|
||||
'bays', cluster.uuid,
|
||||
bookmark=True)]
|
||||
return cluster
|
||||
|
||||
@classmethod
|
||||
def convert_with_links(cls, rpc_bay, expand=True):
|
||||
cluster = Cluster(**rpc_bay.as_dict())
|
||||
return cls._convert_with_links(cluster, pecan.request.host_url, expand)
|
||||
|
||||
@classmethod
|
||||
def sample(cls, expand=True):
|
||||
temp_id = '4a96ac4b-2447-43f1-8ca6-9fd6f36d146d'
|
||||
sample = cls(uuid='27e3153e-d5bf-4b7e-b517-fb518e17f34c',
|
||||
name='example',
|
||||
cluster_template_id=temp_id,
|
||||
node_count=2,
|
||||
master_count=1,
|
||||
create_timeout=15,
|
||||
stack_id='49dc23f5-ffc9-40c3-9d34-7be7f9e34d63',
|
||||
status=fields.BayStatus.CREATE_COMPLETE,
|
||||
status_reason="CREATE completed successfully",
|
||||
api_address='172.24.4.3',
|
||||
node_addresses=['172.24.4.4', '172.24.4.5'],
|
||||
created_at=timeutils.utcnow(),
|
||||
updated_at=timeutils.utcnow())
|
||||
return cls._convert_with_links(sample, 'http://localhost:9511', expand)
|
||||
|
||||
def as_dict(self):
|
||||
"""Render this object as a dict of its fields."""
|
||||
|
||||
# Override this for updated cluster values
|
||||
d = super(Cluster, self).as_dict()
|
||||
|
||||
if 'cluster_template_id' in d.keys():
|
||||
d['baymodel_id'] = d['cluster_template_id']
|
||||
del d['cluster_template_id']
|
||||
|
||||
if 'create_timeout' in d.keys():
|
||||
d['bay_create_timeout'] = d['create_timeout']
|
||||
del d['create_timeout']
|
||||
|
||||
if 'faults' in d.keys():
|
||||
d['bay_faults'] = d['faults']
|
||||
del d['faults']
|
||||
|
||||
return d
|
||||
|
||||
|
||||
class ClusterCollection(collection.Collection):
|
||||
"""API representation of a collection of clusters."""
|
||||
|
||||
clusters = [Cluster]
|
||||
"""A list containing cluster objects"""
|
||||
|
||||
def __init__(self, **kwargs):
|
||||
self._type = 'clusters'
|
||||
|
||||
@staticmethod
|
||||
def convert_with_links(rpc_bays, limit, url=None, expand=False, **kwargs):
|
||||
collection = ClusterCollection()
|
||||
collection.clusters = [Cluster.convert_with_links(p, expand)
|
||||
for p in rpc_bays]
|
||||
collection.next = collection.get_next(limit, url=url, **kwargs)
|
||||
return collection
|
||||
|
||||
@classmethod
|
||||
def sample(cls):
|
||||
sample = cls()
|
||||
sample.clusters = [Cluster.sample(expand=False)]
|
||||
return sample
|
||||
|
||||
|
||||
class ClustersController(base.Controller):
|
||||
"""REST controller for Clusters."""
|
||||
|
||||
def __init__(self):
|
||||
super(ClustersController, self).__init__()
|
||||
|
||||
_custom_actions = {
|
||||
'detail': ['GET'],
|
||||
}
|
||||
|
||||
def _generate_name_for_cluster(self, context):
|
||||
"""Generate a random name like: zeta-22-bay."""
|
||||
name_gen = name_generator.NameGenerator()
|
||||
name = name_gen.generate()
|
||||
return name + '-cluster'
|
||||
|
||||
def _get_clusters_collection(self, marker, limit,
|
||||
sort_key, sort_dir, expand=False,
|
||||
resource_url=None):
|
||||
|
||||
limit = api_utils.validate_limit(limit)
|
||||
sort_dir = api_utils.validate_sort_dir(sort_dir)
|
||||
|
||||
marker_obj = None
|
||||
if marker:
|
||||
marker_obj = objects.Bay.get_by_uuid(pecan.request.context,
|
||||
marker)
|
||||
|
||||
clusters = objects.Bay.list(pecan.request.context, limit,
|
||||
marker_obj, sort_key=sort_key,
|
||||
sort_dir=sort_dir)
|
||||
|
||||
return ClusterCollection.convert_with_links(clusters, limit,
|
||||
url=resource_url,
|
||||
expand=expand,
|
||||
sort_key=sort_key,
|
||||
sort_dir=sort_dir)
|
||||
|
||||
@expose.expose(ClusterCollection, types.uuid, int, wtypes.text,
|
||||
wtypes.text)
|
||||
def get_all(self, marker=None, limit=None, sort_key='id',
|
||||
sort_dir='asc'):
|
||||
"""Retrieve a list of clusters.
|
||||
|
||||
:param marker: pagination marker for large data sets.
|
||||
:param limit: maximum number of resources to return in a single result.
|
||||
:param sort_key: column to sort results by. Default: id.
|
||||
:param sort_dir: direction to sort. "asc" or "desc". Default: asc.
|
||||
"""
|
||||
context = pecan.request.context
|
||||
policy.enforce(context, 'cluster:get_all',
|
||||
action='cluster:get_all')
|
||||
return self._get_clusters_collection(marker, limit, sort_key,
|
||||
sort_dir)
|
||||
|
||||
@expose.expose(ClusterCollection, types.uuid, int, wtypes.text,
|
||||
wtypes.text)
|
||||
def detail(self, marker=None, limit=None, sort_key='id',
|
||||
sort_dir='asc'):
|
||||
"""Retrieve a list of clusters with detail.
|
||||
|
||||
:param marker: pagination marker for large data sets.
|
||||
:param limit: maximum number of resources to return in a single result.
|
||||
:param sort_key: column to sort results by. Default: id.
|
||||
:param sort_dir: direction to sort. "asc" or "desc". Default: asc.
|
||||
"""
|
||||
context = pecan.request.context
|
||||
policy.enforce(context, 'cluster:detail',
|
||||
action='cluster:detail')
|
||||
|
||||
# NOTE(lucasagomes): /detail should only work against collections
|
||||
parent = pecan.request.path.split('/')[:-1][-1]
|
||||
if parent != "clusters":
|
||||
raise exception.HTTPNotFound
|
||||
|
||||
expand = True
|
||||
resource_url = '/'.join(['clusters', 'detail'])
|
||||
return self._get_clusters_collection(marker, limit,
|
||||
sort_key, sort_dir, expand,
|
||||
resource_url)
|
||||
|
||||
def _collect_fault_info(self, context, cluster):
|
||||
"""Collect fault info from heat resources of given cluster
|
||||
|
||||
and store them into cluster.faults.
|
||||
"""
|
||||
osc = clients.OpenStackClients(context)
|
||||
filters = {'status': 'FAILED'}
|
||||
try:
|
||||
failed_resources = osc.heat().resources.list(
|
||||
cluster.stack_id, nested_depth=2, filters=filters)
|
||||
except Exception as e:
|
||||
failed_resources = []
|
||||
LOG.warning(_LW("Failed to retrieve failed resources for "
|
||||
"cluster %(cluster)s from Heat stack "
|
||||
"%(stack)s due to error: %(e)s"),
|
||||
{'cluster': cluster.uuid,
|
||||
'stack': cluster.stack_id, 'e': e},
|
||||
exc_info=True)
|
||||
|
||||
return {res.resource_name: res.resource_status_reason
|
||||
for res in failed_resources}
|
||||
|
||||
@expose.expose(Cluster, types.uuid_or_name)
|
||||
def get_one(self, bay_ident):
|
||||
"""Retrieve information about the given bay.
|
||||
|
||||
:param bay_ident: UUID of a bay or logical name of the bay.
|
||||
"""
|
||||
context = pecan.request.context
|
||||
cluster = api_utils.get_resource('Bay', bay_ident)
|
||||
policy.enforce(context, 'cluster:get', cluster,
|
||||
action='cluster:get')
|
||||
|
||||
cluster = Cluster.convert_with_links(cluster)
|
||||
|
||||
if cluster.status in fields.BayStatus.STATUS_FAILED:
|
||||
cluster.faults = self._collect_fault_info(context, cluster)
|
||||
|
||||
return cluster
|
||||
|
||||
@expose.expose(ClusterID, body=Cluster, status_code=202)
|
||||
def post(self, cluster):
|
||||
"""Create a new cluster.
|
||||
|
||||
:param cluster: a cluster within the request body.
|
||||
"""
|
||||
context = pecan.request.context
|
||||
policy.enforce(context, 'cluster:create',
|
||||
action='cluster:create')
|
||||
temp_id = cluster.cluster_template_id
|
||||
cluster_template = objects.BayModel.get_by_uuid(context, temp_id)
|
||||
cluster_dict = cluster.as_dict()
|
||||
|
||||
attr_validator.validate_os_resources(context,
|
||||
cluster_template.as_dict())
|
||||
attr_validator.validate_master_count(cluster_dict,
|
||||
cluster_template.as_dict())
|
||||
|
||||
cluster_dict['project_id'] = context.project_id
|
||||
cluster_dict['user_id'] = context.user_id
|
||||
# NOTE(yuywz): We will generate a random human-readable name for
|
||||
# cluster if the name is not specified by user.
|
||||
name = cluster_dict.get('name') or \
|
||||
self._generate_name_for_cluster(context)
|
||||
cluster_dict['name'] = name
|
||||
|
||||
new_cluster = objects.Bay(context, **cluster_dict)
|
||||
new_cluster.uuid = uuid.uuid4()
|
||||
pecan.request.rpcapi.bay_create_async(new_cluster,
|
||||
cluster.create_timeout)
|
||||
|
||||
return ClusterID(new_cluster.uuid)
|
||||
|
||||
@wsme.validate(types.uuid, [ClusterPatchType])
|
||||
@expose.expose(ClusterID, types.uuid_or_name, body=[ClusterPatchType],
|
||||
status_code=202)
|
||||
def patch(self, cluster_ident, patch):
|
||||
"""Update an existing bay.
|
||||
|
||||
:param cluster_ident: UUID or logical name of a bay.
|
||||
:param patch: a json PATCH document to apply to this bay.
|
||||
"""
|
||||
context = pecan.request.context
|
||||
cluster = api_utils.get_resource('Bay', cluster_ident)
|
||||
policy.enforce(context, 'cluster:update', cluster,
|
||||
action='cluster:update')
|
||||
try:
|
||||
cluster_dict = cluster.as_dict()
|
||||
new_cluster = Cluster(**api_utils.apply_jsonpatch(cluster_dict,
|
||||
patch))
|
||||
except api_utils.JSONPATCH_EXCEPTIONS as e:
|
||||
raise exception.PatchError(patch=patch, reason=e)
|
||||
|
||||
# Update only the fields that have changed
|
||||
for field in objects.Bay.fields:
|
||||
try:
|
||||
patch_val = getattr(new_cluster, field)
|
||||
except AttributeError:
|
||||
# Ignore fields that aren't exposed in the API
|
||||
continue
|
||||
if patch_val == wtypes.Unset:
|
||||
patch_val = None
|
||||
if cluster[field] != patch_val:
|
||||
cluster[field] = patch_val
|
||||
|
||||
delta = cluster.obj_what_changed()
|
||||
|
||||
validate_bay_properties(delta)
|
||||
|
||||
pecan.request.rpcapi.bay_update_async(cluster)
|
||||
return ClusterID(cluster.uuid)
|
||||
|
||||
@expose.expose(None, types.uuid_or_name, status_code=204)
|
||||
def delete(self, cluster_ident):
|
||||
"""Delete a cluster.
|
||||
|
||||
:param cluster_ident: UUID of cluster or logical name of the cluster.
|
||||
"""
|
||||
context = pecan.request.context
|
||||
cluster = api_utils.get_resource('Bay', cluster_ident)
|
||||
policy.enforce(context, 'cluster:delete', cluster,
|
||||
action='cluster:delete')
|
||||
|
||||
pecan.request.rpcapi.bay_delete_async(cluster.uuid)
|
|
@ -0,0 +1,434 @@
|
|||
# All Rights Reserved.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
from oslo_utils import timeutils
|
||||
import pecan
|
||||
import wsme
|
||||
from wsme import types as wtypes
|
||||
|
||||
from magnum.api import attr_validator
|
||||
from magnum.api.controllers import base
|
||||
from magnum.api.controllers import link
|
||||
from magnum.api.controllers.v1 import collection
|
||||
from magnum.api.controllers.v1 import types
|
||||
from magnum.api import expose
|
||||
from magnum.api import utils as api_utils
|
||||
from magnum.api import validation
|
||||
from magnum.common import clients
|
||||
from magnum.common import exception
|
||||
from magnum.common import name_generator
|
||||
from magnum.common import policy
|
||||
from magnum import objects
|
||||
from magnum.objects import fields
|
||||
|
||||
|
||||
class ClusterTemplatePatchType(types.JsonPatchType):
|
||||
|
||||
@staticmethod
|
||||
def mandatory_attrs():
|
||||
return ['/image_id', '/keypair_id', '/external_network_id', '/coe',
|
||||
'/tls_disabled', '/public', '/registry_enabled',
|
||||
'/server_type', '/cluster_distro', '/network_driver']
|
||||
|
||||
|
||||
class ClusterTemplate(base.APIBase):
|
||||
"""API representation of a clustertemplate.
|
||||
|
||||
This class enforces type checking and value constraints, and converts
|
||||
between the internal object model and the API representation of
|
||||
a clustertemplate.
|
||||
"""
|
||||
|
||||
uuid = types.uuid
|
||||
"""Unique UUID for this clustertemplate"""
|
||||
|
||||
name = wtypes.StringType(min_length=1, max_length=255)
|
||||
"""The name of the clustertemplate"""
|
||||
|
||||
coe = wtypes.Enum(str, *fields.BayType.ALL, mandatory=True)
|
||||
"""The Container Orchestration Engine for this clustertemplate"""
|
||||
|
||||
image_id = wsme.wsattr(wtypes.StringType(min_length=1, max_length=255),
|
||||
mandatory=True)
|
||||
"""The image name or UUID to use as an image for this clustertemplate"""
|
||||
|
||||
flavor_id = wtypes.StringType(min_length=1, max_length=255)
|
||||
"""The flavor of this clustertemplate"""
|
||||
|
||||
master_flavor_id = wtypes.StringType(min_length=1, max_length=255)
|
||||
"""The flavor of the master node for this clustertemplate"""
|
||||
|
||||
dns_nameserver = wtypes.IPv4AddressType()
|
||||
"""The DNS nameserver address"""
|
||||
|
||||
keypair_id = wsme.wsattr(wtypes.StringType(min_length=1, max_length=255),
|
||||
mandatory=True)
|
||||
"""The name or id of the nova ssh keypair"""
|
||||
|
||||
external_network_id = wtypes.StringType(min_length=1, max_length=255)
|
||||
"""The external network to attach the cluster"""
|
||||
|
||||
fixed_network = wtypes.StringType(min_length=1, max_length=255)
|
||||
"""The fixed network name to attach the cluster"""
|
||||
|
||||
fixed_subnet = wtypes.StringType(min_length=1, max_length=255)
|
||||
"""The fixed subnet name to attach the cluster"""
|
||||
|
||||
network_driver = wtypes.StringType(min_length=1, max_length=255)
|
||||
"""The name of the driver used for instantiating container networks"""
|
||||
|
||||
apiserver_port = wtypes.IntegerType(minimum=1024, maximum=65535)
|
||||
"""The API server port for k8s"""
|
||||
|
||||
docker_volume_size = wtypes.IntegerType(minimum=1)
|
||||
"""The size in GB of the docker volume"""
|
||||
|
||||
cluster_distro = wtypes.StringType(min_length=1, max_length=255)
|
||||
"""The Cluster distro for the cluster, ex - coreos, fedora-atomic."""
|
||||
|
||||
links = wsme.wsattr([link.Link], readonly=True)
|
||||
"""A list containing a self link and associated clustertemplate links"""
|
||||
|
||||
http_proxy = wtypes.StringType(min_length=1, max_length=255)
|
||||
"""Address of a proxy that will receive all HTTP requests and relay them.
|
||||
The format is a URL including a port number.
|
||||
"""
|
||||
|
||||
https_proxy = wtypes.StringType(min_length=1, max_length=255)
|
||||
"""Address of a proxy that will receive all HTTPS requests and relay them.
|
||||
The format is a URL including a port number.
|
||||
"""
|
||||
|
||||
no_proxy = wtypes.StringType(min_length=1, max_length=255)
|
||||
"""A comma separated list of ips for which proxies should not
|
||||
used in the cluster
|
||||
"""
|
||||
|
||||
volume_driver = wtypes.StringType(min_length=1, max_length=255)
|
||||
"""The name of the driver used for instantiating container volume driver"""
|
||||
|
||||
registry_enabled = wsme.wsattr(types.boolean, default=False)
|
||||
"""Indicates whether the docker registry is enabled"""
|
||||
|
||||
labels = wtypes.DictType(str, str)
|
||||
"""One or more key/value pairs"""
|
||||
|
||||
tls_disabled = wsme.wsattr(types.boolean, default=False)
|
||||
"""Indicates whether the TLS should be disabled"""
|
||||
|
||||
public = wsme.wsattr(types.boolean, default=False)
|
||||
"""Indicates whether the clustertemplate is public or not."""
|
||||
|
||||
server_type = wsme.wsattr(wtypes.StringType(min_length=1,
|
||||
max_length=255),
|
||||
default='vm')
|
||||
"""Server type for this clustertemplate """
|
||||
|
||||
insecure_registry = wtypes.StringType(min_length=1, max_length=255)
|
||||
"""insecure registry url when create clustertemplate """
|
||||
|
||||
docker_storage_driver = wtypes.Enum(str, *fields.DockerStorageDriver.ALL)
|
||||
"""Docker storage driver"""
|
||||
|
||||
master_lb_enabled = wsme.wsattr(types.boolean, default=False)
|
||||
"""Indicates whether created bays should have a load balancer for master
|
||||
nodes or not.
|
||||
"""
|
||||
|
||||
floating_ip_enabled = wsme.wsattr(types.boolean, default=True)
|
||||
"""Indicates whether created bays should have a floating ip or not."""
|
||||
|
||||
def __init__(self, **kwargs):
|
||||
self.fields = []
|
||||
for field in objects.BayModel.fields:
|
||||
# Skip fields we do not expose.
|
||||
if not hasattr(self, field):
|
||||
continue
|
||||
self.fields.append(field)
|
||||
setattr(self, field, kwargs.get(field, wtypes.Unset))
|
||||
|
||||
@staticmethod
|
||||
def _convert_with_links(cluster_template, url):
|
||||
cluster_template.links = [link.Link.make_link('self', url,
|
||||
'clustertemplates',
|
||||
cluster_template.uuid),
|
||||
link.Link.make_link('bookmark', url,
|
||||
'clustertemplates',
|
||||
cluster_template.uuid,
|
||||
bookmark=True)]
|
||||
return cluster_template
|
||||
|
||||
@classmethod
|
||||
def convert_with_links(cls, rpc_baymodel):
|
||||
cluster_template = ClusterTemplate(**rpc_baymodel.as_dict())
|
||||
return cls._convert_with_links(cluster_template,
|
||||
pecan.request.host_url)
|
||||
|
||||
@classmethod
|
||||
def sample(cls):
|
||||
sample = cls(
|
||||
uuid='27e3153e-d5bf-4b7e-b517-fb518e17f34c',
|
||||
name='example',
|
||||
image_id='Fedora-k8s',
|
||||
flavor_id='m1.small',
|
||||
master_flavor_id='m1.small',
|
||||
dns_nameserver='8.8.1.1',
|
||||
keypair_id='keypair1',
|
||||
external_network_id='ffc44e4a-2319-4062-bce0-9ae1c38b05ba',
|
||||
fixed_network='private',
|
||||
fixed_subnet='private-subnet',
|
||||
network_driver='libnetwork',
|
||||
volume_driver='cinder',
|
||||
apiserver_port=8080,
|
||||
docker_volume_size=25,
|
||||
docker_storage_driver='devicemapper',
|
||||
cluster_distro='fedora-atomic',
|
||||
coe=fields.BayType.KUBERNETES,
|
||||
http_proxy='http://proxy.com:123',
|
||||
https_proxy='https://proxy.com:123',
|
||||
no_proxy='192.168.0.1,192.168.0.2,192.168.0.3',
|
||||
labels={'key1': 'val1', 'key2': 'val2'},
|
||||
server_type='vm',
|
||||
insecure_registry='10.238.100.100:5000',
|
||||
created_at=timeutils.utcnow(),
|
||||
updated_at=timeutils.utcnow(),
|
||||
public=False,
|
||||
master_lb_enabled=False,
|
||||
floating_ip_enabled=True)
|
||||
return cls._convert_with_links(sample, 'http://localhost:9511')
|
||||
|
||||
|
||||
class ClusterTemplateCollection(collection.Collection):
|
||||
"""API representation of a collection of clustertemplates."""
|
||||
|
||||
clustertemplates = [ClusterTemplate]
|
||||
"""A list containing clustertemplates objects"""
|
||||
|
||||
def __init__(self, **kwargs):
|
||||
self._type = 'clustertemplates'
|
||||
|
||||
@staticmethod
|
||||
def convert_with_links(rpc_baymodels, limit, url=None, **kwargs):
|
||||
collection = ClusterTemplateCollection()
|
||||
collection.clustertemplates = [ClusterTemplate.convert_with_links(p)
|
||||
for p in rpc_baymodels]
|
||||
collection.next = collection.get_next(limit, url=url, **kwargs)
|
||||
return collection
|
||||
|
||||
@classmethod
|
||||
def sample(cls):
|
||||
sample = cls()
|
||||
sample.clustertemplates = [ClusterTemplate.sample()]
|
||||
return sample
|
||||
|
||||
|
||||
class ClusterTemplatesController(base.Controller):
|
||||
"""REST controller for ClusterTemplates."""
|
||||
|
||||
_custom_actions = {
|
||||
'detail': ['GET'],
|
||||
}
|
||||
|
||||
def _generate_name_for_cluster_template(self, context):
|
||||
"""Generate a random name like: zeta-22-model."""
|
||||
|
||||
name_gen = name_generator.NameGenerator()
|
||||
name = name_gen.generate()
|
||||
return name + '-template'
|
||||
|
||||
def _get_cluster_templates_collection(self, marker, limit,
|
||||
sort_key, sort_dir,
|
||||
resource_url=None):
|
||||
|
||||
limit = api_utils.validate_limit(limit)
|
||||
sort_dir = api_utils.validate_sort_dir(sort_dir)
|
||||
|
||||
marker_obj = None
|
||||
if marker:
|
||||
marker_obj = objects.BayModel.get_by_uuid(pecan.request.context,
|
||||
marker)
|
||||
|
||||
cluster_templates = objects.BayModel.list(pecan.request.context, limit,
|
||||
marker_obj,
|
||||
sort_key=sort_key,
|
||||
sort_dir=sort_dir)
|
||||
|
||||
return ClusterTemplateCollection.convert_with_links(cluster_templates,
|
||||
limit,
|
||||
url=resource_url,
|
||||
sort_key=sort_key,
|
||||
sort_dir=sort_dir)
|
||||
|
||||
@expose.expose(ClusterTemplateCollection, types.uuid, int, wtypes.text,
|
||||
wtypes.text)
|
||||
def get_all(self, marker=None, limit=None, sort_key='id',
|
||||
sort_dir='asc'):
|
||||
"""Retrieve a list of baymodels.
|
||||
|
||||
:param marker: pagination marker for large data sets.
|
||||
:param limit: maximum number of resources to return in a single result.
|
||||
:param sort_key: column to sort results by. Default: id.
|
||||
:param sort_dir: direction to sort. "asc" or "desc". Default: asc.
|
||||
"""
|
||||
context = pecan.request.context
|
||||
policy.enforce(context, 'clustertemplate:get_all',
|
||||
action='clustertemplate:get_all')
|
||||
return self._get_cluster_templates_collection(marker, limit, sort_key,
|
||||
sort_dir)
|
||||
|
||||
@expose.expose(ClusterTemplateCollection, types.uuid, int, wtypes.text,
|
||||
wtypes.text)
|
||||
def detail(self, marker=None, limit=None, sort_key='id',
|
||||
sort_dir='asc'):
|
||||
"""Retrieve a list of clustertemplates with detail.
|
||||
|
||||
:param marker: pagination marker for large data sets.
|
||||
:param limit: maximum number of resources to return in a single result.
|
||||
:param sort_key: column to sort results by. Default: id.
|
||||
:param sort_dir: direction to sort. "asc" or "desc". Default: asc.
|
||||
"""
|
||||
context = pecan.request.context
|
||||
policy.enforce(context, 'clustertemplate:detail',
|
||||
action='clustertemplate:detail')
|
||||
|
||||
# NOTE(lucasagomes): /detail should only work against collections
|
||||
parent = pecan.request.path.split('/')[:-1][-1]
|
||||
if parent != "clustertemplates":
|
||||
raise exception.HTTPNotFound
|
||||
|
||||
resource_url = '/'.join(['clustertemplates', 'detail'])
|
||||
return self._get_cluster_templates_collection(marker, limit,
|
||||
sort_key, sort_dir,
|
||||
resource_url)
|
||||
|
||||
@expose.expose(ClusterTemplate, types.uuid_or_name)
|
||||
def get_one(self, cluster_template_ident):
|
||||
"""Retrieve information about the given clustertemplate.
|
||||
|
||||
:param cluster_template_ident: UUID or logical name of a
|
||||
clustertemplate.
|
||||
"""
|
||||
context = pecan.request.context
|
||||
cluster_template = api_utils.get_resource('BayModel',
|
||||
cluster_template_ident)
|
||||
if not cluster_template.public:
|
||||
policy.enforce(context, 'clustertemplate:get', cluster_template,
|
||||
action='clustertemplate:get')
|
||||
|
||||
return ClusterTemplate.convert_with_links(cluster_template)
|
||||
|
||||
@expose.expose(ClusterTemplate, body=ClusterTemplate, status_code=201)
|
||||
@validation.enforce_network_driver_types_create()
|
||||
@validation.enforce_volume_driver_types_create()
|
||||
@validation.enforce_volume_storage_size_create()
|
||||
def post(self, cluster_template):
|
||||
"""Create a new cluster_template.
|
||||
|
||||
:param cluster_template: a cluster_template within the request body.
|
||||
"""
|
||||
context = pecan.request.context
|
||||
policy.enforce(context, 'clustertemplate:create',
|
||||
action='clustertemplate:create')
|
||||
cluster_template_dict = cluster_template.as_dict()
|
||||
cli = clients.OpenStackClients(context)
|
||||
attr_validator.validate_os_resources(context, cluster_template_dict)
|
||||
image_data = attr_validator.validate_image(cli,
|
||||
cluster_template_dict[
|
||||
'image_id'])
|
||||
cluster_template_dict['cluster_distro'] = image_data['os_distro']
|
||||
cluster_template_dict['project_id'] = context.project_id
|
||||
cluster_template_dict['user_id'] = context.user_id
|
||||
# check permissions for making cluster_template public
|
||||
if cluster_template_dict['public']:
|
||||
if not policy.enforce(context, "clustertemplate:publish", None,
|
||||
do_raise=False):
|
||||
raise exception.ClusterTemplatePublishDenied()
|
||||
|
||||
# NOTE(yuywz): We will generate a random human-readable name for
|
||||
# cluster_template if the name is not specified by user.
|
||||
arg_name = cluster_template_dict.get('name')
|
||||
name = arg_name or self._generate_name_for_cluster_template(context)
|
||||
cluster_template_dict['name'] = name
|
||||
|
||||
new_cluster_template = objects.BayModel(context,
|
||||
**cluster_template_dict)
|
||||
new_cluster_template.create()
|
||||
# Set the HTTP Location Header
|
||||
pecan.response.location = link.build_url('clustertemplates',
|
||||
new_cluster_template.uuid)
|
||||
return ClusterTemplate.convert_with_links(new_cluster_template)
|
||||
|
||||
@wsme.validate(types.uuid_or_name, [ClusterTemplatePatchType])
|
||||
@expose.expose(ClusterTemplate, types.uuid_or_name,
|
||||
body=[ClusterTemplatePatchType])
|
||||
@validation.enforce_network_driver_types_update()
|
||||
@validation.enforce_volume_driver_types_update()
|
||||
def patch(self, cluster_template_ident, patch):
|
||||
"""Update an existing cluster_template.
|
||||
|
||||
:param cluster_template_ident: UUID or logic name of a
|
||||
cluster_template.
|
||||
:param patch: a json PATCH document to apply to this
|
||||
cluster_template.
|
||||
"""
|
||||
context = pecan.request.context
|
||||
cluster_template = api_utils.get_resource('BayModel',
|
||||
cluster_template_ident)
|
||||
policy.enforce(context, 'clustertemplate:update', cluster_template,
|
||||
action='clustertemplate:update')
|
||||
try:
|
||||
cluster_template_dict = cluster_template.as_dict()
|
||||
new_cluster_template = ClusterTemplate(**api_utils.apply_jsonpatch(
|
||||
cluster_template_dict,
|
||||
patch))
|
||||
except api_utils.JSONPATCH_EXCEPTIONS as e:
|
||||
raise exception.PatchError(patch=patch, reason=e)
|
||||
|
||||
new_cluster_template_dict = new_cluster_template.as_dict()
|
||||
attr_validator.validate_os_resources(context,
|
||||
new_cluster_template_dict)
|
||||
# check permissions when updating baymodel public flag
|
||||
if cluster_template.public != new_cluster_template.public:
|
||||
if not policy.enforce(context, "clustertemplate:publish", None,
|
||||
do_raise=False):
|
||||
raise exception.ClusterTemplatePublishDenied()
|
||||
|
||||
# Update only the fields that have changed
|
||||
for field in objects.BayModel.fields:
|
||||
try:
|
||||
patch_val = getattr(new_cluster_template, field)
|
||||
except AttributeError:
|
||||
# Ignore fields that aren't exposed in the API
|
||||
continue
|
||||
if patch_val == wtypes.Unset:
|
||||
patch_val = None
|
||||
if cluster_template[field] != patch_val:
|
||||
cluster_template[field] = patch_val
|
||||
|
||||
cluster_template.save()
|
||||
return ClusterTemplate.convert_with_links(cluster_template)
|
||||
|
||||
@expose.expose(None, types.uuid_or_name, status_code=204)
|
||||
def delete(self, cluster_template_ident):
|
||||
"""Delete a cluster_template.
|
||||
|
||||
:param cluster_template_ident: UUID or logical name of a
|
||||
cluster_template.
|
||||
"""
|
||||
context = pecan.request.context
|
||||
cluster_template = api_utils.get_resource('BayModel',
|
||||
cluster_template_ident)
|
||||
policy.enforce(context, 'clustertemplate:delete', cluster_template,
|
||||
action='clustertemplate:delete')
|
||||
cluster_template.destroy()
|
|
@ -229,28 +229,29 @@ class FileSystemNotSupported(MagnumException):
|
|||
"File system %(fs)s is not supported.")
|
||||
|
||||
|
||||
class BayModelNotFound(ResourceNotFound):
|
||||
message = _("Baymodel %(baymodel)s could not be found.")
|
||||
class ClusterTemplateNotFound(ResourceNotFound):
|
||||
message = _("ClusterTemplate %(clustertemplate)s could not be found.")
|
||||
|
||||
|
||||
class BayModelAlreadyExists(Conflict):
|
||||
message = _("A baymodel with UUID %(uuid)s already exists.")
|
||||
class ClusterTemplateAlreadyExists(Conflict):
|
||||
message = _("A ClusterTemplate with UUID %(uuid)s already exists.")
|
||||
|
||||
|
||||
class BayModelReferenced(Invalid):
|
||||
message = _("Baymodel %(baymodel)s is referenced by one or multiple bays.")
|
||||
class ClusterTemplateReferenced(Invalid):
|
||||
message = _("ClusterTemplate %(clustertemplate)s is referenced by one or"
|
||||
" multiple clusters.")
|
||||
|
||||
|
||||
class BaymodelPublishDenied(NotAuthorized):
|
||||
message = _("Not authorized to set public flag for baymodel.")
|
||||
class ClusterTemplatePublishDenied(NotAuthorized):
|
||||
message = _("Not authorized to set public flag for cluster template.")
|
||||
|
||||
|
||||
class BayNotFound(ResourceNotFound):
|
||||
message = _("Bay %(bay)s could not be found.")
|
||||
class ClusterNotFound(ResourceNotFound):
|
||||
message = _("Cluster %(cluster)s could not be found.")
|
||||
|
||||
|
||||
class BayAlreadyExists(Conflict):
|
||||
message = _("A bay with UUID %(uuid)s already exists.")
|
||||
class ClusterAlreadyExists(Conflict):
|
||||
message = _("A cluster with UUID %(uuid)s already exists.")
|
||||
|
||||
|
||||
class ContainerNotFound(ResourceNotFound):
|
||||
|
|
|
@ -234,7 +234,7 @@ class Handler(object):
|
|||
trust_manager.delete_trustee_and_trust(osc, context, bay)
|
||||
cert_manager.delete_certificates_from_bay(bay, context=context)
|
||||
bay.destroy()
|
||||
except exception.BayNotFound:
|
||||
except exception.ClusterNotFound:
|
||||
LOG.info(_LI('The bay %s has been deleted by others.'), uuid)
|
||||
conductor_utils.notify_about_bay_operation(
|
||||
context, taxonomy.ACTION_DELETE, taxonomy.OUTCOME_SUCCESS)
|
||||
|
@ -348,7 +348,7 @@ class HeatPoller(object):
|
|||
cert_manager.delete_certificates_from_bay(self.bay,
|
||||
context=self.context)
|
||||
self.bay.destroy()
|
||||
except exception.BayNotFound:
|
||||
except exception.ClusterNotFound:
|
||||
LOG.info(_LI('The bay %s has been deleted by others.')
|
||||
% self.bay.uuid)
|
||||
|
||||
|
|
|
@ -118,7 +118,7 @@ class Connection(object):
|
|||
|
||||
:param bay_id: The id or uuid of a bay.
|
||||
:returns: A bay.
|
||||
:raises: BayNotFound
|
||||
:raises: ClusterNotFound
|
||||
"""
|
||||
|
||||
@abc.abstractmethod
|
||||
|
@ -201,7 +201,7 @@ class Connection(object):
|
|||
|
||||
:param baymodel_id: The id or uuid of a baymodel.
|
||||
:returns: A baymodel.
|
||||
:raises: BayModelNotFound
|
||||
:raises: ClusterTemplateNotFound
|
||||
"""
|
||||
|
||||
@abc.abstractmethod
|
||||
|
|
|
@ -157,7 +157,7 @@ class Connection(api.Connection):
|
|||
try:
|
||||
bay.save()
|
||||
except db_exc.DBDuplicateEntry:
|
||||
raise exception.BayAlreadyExists(uuid=values['uuid'])
|
||||
raise exception.ClusterAlreadyExists(uuid=values['uuid'])
|
||||
return bay
|
||||
|
||||
def get_bay_by_id(self, context, bay_id):
|
||||
|
@ -167,7 +167,7 @@ class Connection(api.Connection):
|
|||
try:
|
||||
return query.one()
|
||||
except NoResultFound:
|
||||
raise exception.BayNotFound(bay=bay_id)
|
||||
raise exception.ClusterNotFound(cluster=bay_id)
|
||||
|
||||
def get_bay_by_name(self, context, bay_name):
|
||||
query = model_query(models.Bay)
|
||||
|
@ -179,7 +179,7 @@ class Connection(api.Connection):
|
|||
raise exception.Conflict('Multiple bays exist with same name.'
|
||||
' Please use the bay uuid instead.')
|
||||
except NoResultFound:
|
||||
raise exception.BayNotFound(bay=bay_name)
|
||||
raise exception.ClusterNotFound(cluster=bay_name)
|
||||
|
||||
def get_bay_by_uuid(self, context, bay_uuid):
|
||||
query = model_query(models.Bay)
|
||||
|
@ -188,7 +188,7 @@ class Connection(api.Connection):
|
|||
try:
|
||||
return query.one()
|
||||
except NoResultFound:
|
||||
raise exception.BayNotFound(bay=bay_uuid)
|
||||
raise exception.ClusterNotFound(cluster=bay_uuid)
|
||||
|
||||
def destroy_bay(self, bay_id):
|
||||
session = get_session()
|
||||
|
@ -199,7 +199,7 @@ class Connection(api.Connection):
|
|||
try:
|
||||
query.one()
|
||||
except NoResultFound:
|
||||
raise exception.BayNotFound(bay=bay_id)
|
||||
raise exception.ClusterNotFound(cluster=bay_id)
|
||||
|
||||
query.delete()
|
||||
|
||||
|
@ -219,7 +219,7 @@ class Connection(api.Connection):
|
|||
try:
|
||||
ref = query.with_lockmode('update').one()
|
||||
except NoResultFound:
|
||||
raise exception.BayNotFound(bay=bay_id)
|
||||
raise exception.ClusterNotFound(cluster=bay_id)
|
||||
|
||||
if 'provision_state' in values:
|
||||
values['provision_updated_at'] = timeutils.utcnow()
|
||||
|
@ -264,7 +264,7 @@ class Connection(api.Connection):
|
|||
try:
|
||||
baymodel.save()
|
||||
except db_exc.DBDuplicateEntry:
|
||||
raise exception.BayModelAlreadyExists(uuid=values['uuid'])
|
||||
raise exception.ClusterTemplateAlreadyExists(uuid=values['uuid'])
|
||||
return baymodel
|
||||
|
||||
def get_baymodel_by_id(self, context, baymodel_id):
|
||||
|
@ -276,7 +276,8 @@ class Connection(api.Connection):
|
|||
try:
|
||||
return query.one()
|
||||
except NoResultFound:
|
||||
raise exception.BayModelNotFound(baymodel=baymodel_id)
|
||||
raise exception.ClusterTemplateNotFound(
|
||||
clustertemplate=baymodel_id)
|
||||
|
||||
def get_baymodel_by_uuid(self, context, baymodel_uuid):
|
||||
query = model_query(models.BayModel)
|
||||
|
@ -287,7 +288,8 @@ class Connection(api.Connection):
|
|||
try:
|
||||
return query.one()
|
||||
except NoResultFound:
|
||||
raise exception.BayModelNotFound(baymodel=baymodel_uuid)
|
||||
raise exception.ClusterTemplateNotFound(
|
||||
clustertemplate=baymodel_uuid)
|
||||
|
||||
def get_baymodel_by_name(self, context, baymodel_name):
|
||||
query = model_query(models.BayModel)
|
||||
|
@ -301,7 +303,8 @@ class Connection(api.Connection):
|
|||
raise exception.Conflict('Multiple baymodels exist with same name.'
|
||||
' Please use the baymodel uuid instead.')
|
||||
except NoResultFound:
|
||||
raise exception.BayModelNotFound(baymodel=baymodel_name)
|
||||
raise exception.ClusterTemplateNotFound(
|
||||
clustertemplate=baymodel_name)
|
||||
|
||||
def _is_baymodel_referenced(self, session, baymodel_uuid):
|
||||
"""Checks whether the baymodel is referenced by bay(s)."""
|
||||
|
@ -324,10 +327,12 @@ class Connection(api.Connection):
|
|||
try:
|
||||
baymodel_ref = query.one()
|
||||
except NoResultFound:
|
||||
raise exception.BayModelNotFound(baymodel=baymodel_id)
|
||||
raise exception.ClusterTemplateNotFound(
|
||||
clustertemplate=baymodel_id)
|
||||
|
||||
if self._is_baymodel_referenced(session, baymodel_ref['uuid']):
|
||||
raise exception.BayModelReferenced(baymodel=baymodel_id)
|
||||
raise exception.ClusterTemplateReferenced(
|
||||
clustertemplate=baymodel_id)
|
||||
|
||||
query.delete()
|
||||
|
||||
|
@ -347,12 +352,14 @@ class Connection(api.Connection):
|
|||
try:
|
||||
ref = query.with_lockmode('update').one()
|
||||
except NoResultFound:
|
||||
raise exception.BayModelNotFound(baymodel=baymodel_id)
|
||||
raise exception.ClusterTemplateNotFound(
|
||||
clustertemplate=baymodel_id)
|
||||
|
||||
if self._is_baymodel_referenced(session, ref['uuid']):
|
||||
# we only allow to update baymodel to be public
|
||||
if not self._is_publishing_baymodel(values):
|
||||
raise exception.BayModelReferenced(baymodel=baymodel_id)
|
||||
raise exception.ClusterTemplateReferenced(
|
||||
clustertemplate=baymodel_id)
|
||||
|
||||
ref.update(values)
|
||||
return ref
|
||||
|
|
|
@ -165,7 +165,7 @@ class MagnumPeriodicTasks(periodic_task.PeriodicTasks):
|
|||
def _sync_deleted_stack(self, bay):
|
||||
try:
|
||||
bay.destroy()
|
||||
except exception.BayNotFound:
|
||||
except exception.ClusterNotFound:
|
||||
LOG.info(_LI('The bay %s has been deleted by others.'), bay.uuid)
|
||||
else:
|
||||
LOG.info(_LI("Bay with id %(id)s not found in heat "
|
||||
|
|
|
@ -0,0 +1,176 @@
|
|||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
from oslo_log import log as logging
|
||||
from tempest.lib import exceptions
|
||||
|
||||
from magnum.i18n import _LE
|
||||
from magnum.i18n import _LI
|
||||
from magnum.i18n import _LW
|
||||
from magnum.tests.functional.api.v1.models import cluster_id_model
|
||||
from magnum.tests.functional.api.v1.models import cluster_model
|
||||
from magnum.tests.functional.common import client
|
||||
from magnum.tests.functional.common import utils
|
||||
|
||||
|
||||
class ClusterClient(client.MagnumClient):
|
||||
"""Encapsulates REST calls and maps JSON to/from models"""
|
||||
|
||||
LOG = logging.getLogger(__name__)
|
||||
|
||||
@classmethod
|
||||
def clusters_uri(cls, filters=None):
|
||||
"""Construct clusters uri with optional filters
|
||||
|
||||
:param filters: Optional k:v dict that's converted to url query
|
||||
:returns: url string
|
||||
"""
|
||||
|
||||
url = "/clusters"
|
||||
if filters:
|
||||
url = cls.add_filters(url, filters)
|
||||
return url
|
||||
|
||||
@classmethod
|
||||
def cluster_uri(cls, cluster_id):
|
||||
"""Construct cluster uri
|
||||
|
||||
:param cluster_id: cluster uuid or name
|
||||
:returns: url string
|
||||
"""
|
||||
|
||||
return "{0}/{1}".format(cls.clusters_uri(), cluster_id)
|
||||
|
||||
def list_clusters(self, filters=None, **kwargs):
|
||||
"""Makes GET /clusters request and returns ClusterCollection
|
||||
|
||||
Abstracts REST call to return all clusters
|
||||
|
||||
:param filters: Optional k:v dict that's converted to url query
|
||||
:returns: response object and ClusterCollection object
|
||||
"""
|
||||
|
||||
resp, body = self.get(self.clusters_uri(filters), **kwargs)
|
||||
return self.deserialize(resp, body, cluster_model.ClusterCollection)
|
||||
|
||||
def get_cluster(self, cluster_id, **kwargs):
|
||||
"""Makes GET /cluster request and returns ClusterEntity
|
||||
|
||||
Abstracts REST call to return a single cluster based on uuid or name
|
||||
|
||||
:param cluster_id: cluster uuid or name
|
||||
:returns: response object and ClusterCollection object
|
||||
"""
|
||||
|
||||
resp, body = self.get(self.cluster_uri(cluster_id))
|
||||
return self.deserialize(resp, body, cluster_model.ClusterEntity)
|
||||
|
||||
def post_cluster(self, model, **kwargs):
|
||||
"""Makes POST /cluster request and returns ClusterIdEntity
|
||||
|
||||
Abstracts REST call to create new cluster
|
||||
|
||||
:param model: ClusterEntity
|
||||
:returns: response object and ClusterIdEntity object
|
||||
"""
|
||||
|
||||
resp, body = self.post(
|
||||
self.clusters_uri(),
|
||||
body=model.to_json(), **kwargs)
|
||||
return self.deserialize(resp, body, cluster_id_model.ClusterIdEntity)
|
||||
|
||||
def patch_cluster(self, cluster_id, clusterpatch_listmodel, **kwargs):
|
||||
"""Makes PATCH /cluster request and returns ClusterIdEntity
|
||||
|
||||
Abstracts REST call to update cluster attributes
|
||||
|
||||
:param cluster_id: UUID of cluster
|
||||
:param clusterpatch_listmodel: ClusterPatchCollection
|
||||
:returns: response object and ClusterIdEntity object
|
||||
"""
|
||||
|
||||
resp, body = self.patch(
|
||||
self.cluster_uri(cluster_id),
|
||||
body=clusterpatch_listmodel.to_json(), **kwargs)
|
||||
return self.deserialize(resp, body, cluster_id_model.ClusterIdEntity)
|
||||
|
||||
def delete_cluster(self, cluster_id, **kwargs):
|
||||
"""Makes DELETE /cluster request and returns response object
|
||||
|
||||
Abstracts REST call to delete cluster based on uuid or name
|
||||
|
||||
:param cluster_id: UUID or name of cluster
|
||||
:returns: response object
|
||||
"""
|
||||
|
||||
return self.delete(self.cluster_uri(cluster_id), **kwargs)
|
||||
|
||||
def wait_for_cluster_to_delete(self, cluster_id):
|
||||
utils.wait_for_condition(
|
||||
lambda: self.does_cluster_not_exist(cluster_id), 10, 600)
|
||||
|
||||
def wait_for_created_cluster(self, cluster_id, delete_on_error=True):
|
||||
try:
|
||||
utils.wait_for_condition(
|
||||
lambda: self.does_cluster_exist(cluster_id), 10, 1800)
|
||||
except Exception:
|
||||
# In error state. Clean up the cluster id if desired
|
||||
self.LOG.error(_LE('Cluster %s entered an exception state.') %
|
||||
cluster_id)
|
||||
if delete_on_error:
|
||||
self.LOG.error(_LE('We will attempt to delete clusters now.'))
|
||||
self.delete_cluster(cluster_id)
|
||||
self.wait_for_cluster_to_delete(cluster_id)
|
||||
raise
|
||||
|
||||
def wait_for_final_state(self, cluster_id):
|
||||
utils.wait_for_condition(
|
||||
lambda: self.is_cluster_in_final_state(cluster_id), 10, 1800)
|
||||
|
||||
def is_cluster_in_final_state(self, cluster_id):
|
||||
try:
|
||||
resp, model = self.get_cluster(cluster_id)
|
||||
if model.status in ['CREATED', 'CREATE_COMPLETE',
|
||||
'ERROR', 'CREATE_FAILED']:
|
||||
self.LOG.info(_LI('Cluster %s succeeded.') % cluster_id)
|
||||
return True
|
||||
else:
|
||||
return False
|
||||
except exceptions.NotFound:
|
||||
self.LOG.warning(_LW('Cluster %s is not found.') % cluster_id)
|
||||
return False
|
||||
|
||||
def does_cluster_exist(self, cluster_id):
|
||||
try:
|
||||
resp, model = self.get_cluster(cluster_id)
|
||||
if model.status in ['CREATED', 'CREATE_COMPLETE']:
|
||||
self.LOG.info(_LI('Cluster %s is created.') % cluster_id)
|
||||
return True
|
||||
elif model.status in ['ERROR', 'CREATE_FAILED']:
|
||||
self.LOG.error(_LE('Cluster %s is in fail state.') %
|
||||
cluster_id)
|
||||
raise exceptions.ServerFault(
|
||||
"Got into an error condition: %s for %s" %
|
||||
(model.status, cluster_id))
|
||||
else:
|
||||
return False
|
||||
except exceptions.NotFound:
|
||||
self.LOG.warning(_LW('Cluster %s is not found.') % cluster_id)
|
||||
return False
|
||||
|
||||
def does_cluster_not_exist(self, cluster_id):
|
||||
try:
|
||||
self.get_cluster(cluster_id)
|
||||
except exceptions.NotFound:
|
||||
self.LOG.warning(_LW('Cluster %s is not found.') % cluster_id)
|
||||
return True
|
||||
return False
|
|
@ -0,0 +1,113 @@
|
|||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
from magnum.tests.functional.api.v1.models import cluster_template_model
|
||||
from magnum.tests.functional.common import client
|
||||
|
||||
|
||||
class ClusterTemplateClient(client.MagnumClient):
|
||||
"""Encapsulates REST calls and maps JSON to/from models"""
|
||||
|
||||
@classmethod
|
||||
def cluster_templates_uri(cls, filters=None):
|
||||
"""Construct clustertemplates uri with optional filters
|
||||
|
||||
:param filters: Optional k:v dict that's converted to url query
|
||||
:returns: url string
|
||||
"""
|
||||
|
||||
url = "/clustertemplates"
|
||||
if filters:
|
||||
url = cls.add_filters(url, filters)
|
||||
return url
|
||||
|
||||
@classmethod
|
||||
def cluster_template_uri(cls, cluster_template_id):
|
||||
"""Construct cluster_template uri
|
||||
|
||||
:param cluster_template_id: cluster_template uuid or name
|
||||
:returns: url string
|
||||
"""
|
||||
|
||||
return "{0}/{1}".format(cls.cluster_templates_uri(),
|
||||
cluster_template_id)
|
||||
|
||||
def list_cluster_templates(self, filters=None, **kwargs):
|
||||
"""Makes GET /clustertemplates request
|
||||
|
||||
Abstracts REST call to return all clustertemplates
|
||||
|
||||
:param filters: Optional k:v dict that's converted to url query
|
||||
:returns: response object and ClusterTemplateCollection object
|
||||
"""
|
||||
|
||||
resp, body = self.get(self.cluster_templates_uri(filters), **kwargs)
|
||||
collection = cluster_template_model.ClusterTemplateCollection
|
||||
return self.deserialize(resp, body, collection)
|
||||
|
||||
def get_cluster_template(self, cluster_template_id, **kwargs):
|
||||
"""Makes GET /clustertemplate request and returns ClusterTemplateEntity
|
||||
|
||||
Abstracts REST call to return a single clustertempalte based on uuid
|
||||
or name
|
||||
|
||||
:param cluster_template_id: clustertempalte uuid or name
|
||||
:returns: response object and ClusterTemplateCollection object
|
||||
"""
|
||||
|
||||
resp, body = self.get(self.cluster_template_uri(cluster_template_id))
|
||||
return self.deserialize(resp, body,
|
||||
cluster_template_model.ClusterTemplateEntity)
|
||||
|
||||
def post_cluster_template(self, model, **kwargs):
|
||||
"""Makes POST /clustertemplate request
|
||||
|
||||
Abstracts REST call to create new clustertemplate
|
||||
|
||||
:param model: ClusterTemplateEntity
|
||||
:returns: response object and ClusterTemplateEntity object
|
||||
"""
|
||||
|
||||
resp, body = self.post(
|
||||
self.cluster_templates_uri(),
|
||||
body=model.to_json(), **kwargs)
|
||||
entity = cluster_template_model.ClusterTemplateEntity
|
||||
return self.deserialize(resp, body, entity)
|
||||
|
||||
def patch_cluster_template(self, cluster_template_id,
|
||||
cluster_templatepatch_listmodel, **kwargs):
|
||||
"""Makes PATCH /clustertemplate and returns ClusterTemplateEntity
|
||||
|
||||
Abstracts REST call to update clustertemplate attributes
|
||||
|
||||
:param cluster_template_id: UUID of clustertemplate
|
||||
:param cluster_templatepatch_listmodel: ClusterTemplatePatchCollection
|
||||
:returns: response object and ClusterTemplateEntity object
|
||||
"""
|
||||
|
||||
resp, body = self.patch(
|
||||
self.cluster_template_uri(cluster_template_id),
|
||||
body=cluster_templatepatch_listmodel.to_json(), **kwargs)
|
||||
return self.deserialize(resp, body,
|
||||
cluster_template_model.ClusterTemplateEntity)
|
||||
|
||||
def delete_cluster_template(self, cluster_template_id, **kwargs):
|
||||
"""Makes DELETE /clustertemplate request and returns response object
|
||||
|
||||
Abstracts REST call to delete clustertemplate based on uuid or name
|
||||
|
||||
:param cluster_template_id: UUID or name of clustertemplate
|
||||
:returns: response object
|
||||
"""
|
||||
|
||||
return self.delete(self.cluster_template_uri(cluster_template_id),
|
||||
**kwargs)
|
|
@ -0,0 +1,24 @@
|
|||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
from magnum.tests.functional.common import models
|
||||
|
||||
|
||||
class ClusterIdData(models.BaseModel):
|
||||
"""Data that encapsulates ClusterId attributes"""
|
||||
pass
|
||||
|
||||
|
||||
class ClusterIdEntity(models.EntityModel):
|
||||
"""Entity Model that represents a single instance of CertData"""
|
||||
ENTITY_NAME = 'clusterid'
|
||||
MODEL_TYPE = ClusterIdData
|
|
@ -0,0 +1,30 @@
|
|||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
from magnum.tests.functional.common import models
|
||||
|
||||
|
||||
class ClusterData(models.BaseModel):
|
||||
"""Data that encapsulates cluster attributes"""
|
||||
pass
|
||||
|
||||
|
||||
class ClusterEntity(models.EntityModel):
|
||||
"""Entity Model that represents a single instance of ClusterData"""
|
||||
ENTITY_NAME = 'cluster'
|
||||
MODEL_TYPE = ClusterData
|
||||
|
||||
|
||||
class ClusterCollection(models.CollectionModel):
|
||||
"""Collection Model that represents a list of ClusterData objects"""
|
||||
COLLECTION_NAME = 'clusterlists'
|
||||
MODEL_TYPE = ClusterData
|
|
@ -0,0 +1,30 @@
|
|||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
from magnum.tests.functional.common import models
|
||||
|
||||
|
||||
class ClusterTemplateData(models.BaseModel):
|
||||
"""Data that encapsulates clustertemplate attributes"""
|
||||
pass
|
||||
|
||||
|
||||
class ClusterTemplateEntity(models.EntityModel):
|
||||
"""Entity Model that represents a single instance of ClusterTemplateData"""
|
||||
ENTITY_NAME = 'clustertemplate'
|
||||
MODEL_TYPE = ClusterTemplateData
|
||||
|
||||
|
||||
class ClusterTemplateCollection(models.CollectionModel):
|
||||
"""Collection that represents a list of ClusterTemplateData objects"""
|
||||
COLLECTION_NAME = 'clustertemplatelists'
|
||||
MODEL_TYPE = ClusterTemplateData
|
|
@ -0,0 +1,77 @@
|
|||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
import json
|
||||
|
||||
from magnum.tests.functional.common import models
|
||||
|
||||
|
||||
class ClusterTemplatePatchData(models.BaseModel):
|
||||
"""Data that encapsulates clustertemplatepatch attributes"""
|
||||
pass
|
||||
|
||||
|
||||
class ClusterTemplatePatchEntity(models.EntityModel):
|
||||
"""Model that represents a single instance of ClusterTemplatePatchData"""
|
||||
ENTITY_NAME = 'clustertemplatepatch'
|
||||
MODEL_TYPE = ClusterTemplatePatchData
|
||||
|
||||
|
||||
class ClusterTemplatePatchCollection(models.CollectionModel):
|
||||
"""Model that represents a list of ClusterTemplatePatchData objects"""
|
||||
MODEL_TYPE = ClusterTemplatePatchData
|
||||
COLLECTION_NAME = 'clustertemplatepatchlist'
|
||||
|
||||
def to_json(self):
|
||||
"""Converts ClusterTemplatePatchCollection to json
|
||||
|
||||
Retrieves list from COLLECTION_NAME attribute and converts each object
|
||||
to dict, appending it to a list. Then converts the entire list to
|
||||
json
|
||||
|
||||
This is required due to COLLECTION_NAME holding a list of objects that
|
||||
needed to be converted to dict individually
|
||||
|
||||
:returns: json object
|
||||
"""
|
||||
|
||||
data = getattr(self, ClusterTemplatePatchCollection.COLLECTION_NAME)
|
||||
collection = []
|
||||
for d in data:
|
||||
collection.append(d.to_dict())
|
||||
return json.dumps(collection)
|
||||
|
||||
@classmethod
|
||||
def from_dict(cls, data):
|
||||
"""Converts dict to ClusterTemplatePatchData
|
||||
|
||||
Converts data dict to list of ClusterTemplatePatchData objects and
|
||||
stores it in COLLECTION_NAME
|
||||
|
||||
Example of dict data:
|
||||
|
||||
[{
|
||||
"path": "/name",
|
||||
"value": "myname",
|
||||
"op": "replace"
|
||||
}]
|
||||
|
||||
:param data: dict of patch data
|
||||
:returns: json object
|
||||
"""
|
||||
|
||||
model = cls()
|
||||
collection = []
|
||||
for d in data:
|
||||
collection.append(cls.MODEL_TYPE.from_dict(d))
|
||||
setattr(model, cls.COLLECTION_NAME, collection)
|
||||
return model
|
|
@ -0,0 +1,76 @@
|
|||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
import json
|
||||
|
||||
from magnum.tests.functional.common import models
|
||||
|
||||
|
||||
class ClusterPatchData(models.BaseModel):
|
||||
"""Data that encapsulates clusterpatch attributes"""
|
||||
pass
|
||||
|
||||
|
||||
class ClusterPatchEntity(models.EntityModel):
|
||||
"""Entity Model that represents a single instance of ClusterPatchData"""
|
||||
ENTITY_NAME = 'clusterpatch'
|
||||
MODEL_TYPE = ClusterPatchData
|
||||
|
||||
|
||||
class ClusterPatchCollection(models.CollectionModel):
|
||||
"""Collection Model that represents a list of ClusterPatchData objects"""
|
||||
MODEL_TYPE = ClusterPatchData
|
||||
COLLECTION_NAME = 'clusterpatchlist'
|
||||
|
||||
def to_json(self):
|
||||
"""Converts ClusterPatchCollection to json
|
||||
|
||||
Retrieves list from COLLECTION_NAME attribute and converts each object
|
||||
to dict, appending it to a list. Then converts the entire list to json
|
||||
|
||||
This is required due to COLLECTION_NAME holding a list of objects that
|
||||
needed to be converted to dict individually
|
||||
|
||||
:returns: json object
|
||||
"""
|
||||
|
||||
data = getattr(self, ClusterPatchCollection.COLLECTION_NAME)
|
||||
collection = []
|
||||
for d in data:
|
||||
collection.append(d.to_dict())
|
||||
return json.dumps(collection)
|
||||
|
||||
@classmethod
|
||||
def from_dict(cls, data):
|
||||
"""Converts dict to ClusterPatchData
|
||||
|
||||
Converts data dict to list of ClusterPatchData objects and stores it
|
||||
in COLLECTION_NAME
|
||||
|
||||
Example of dict data:
|
||||
|
||||
[{
|
||||
"path": "/name",
|
||||
"value": "myname",
|
||||
"op": "replace"
|
||||
}]
|
||||
|
||||
:param data: dict of patch data
|
||||
:returns: json object
|
||||
"""
|
||||
|
||||
model = cls()
|
||||
collection = []
|
||||
for d in data:
|
||||
collection.append(cls.MODEL_TYPE.from_dict(d))
|
||||
setattr(model, cls.COLLECTION_NAME, collection)
|
||||
return model
|
|
@ -18,7 +18,6 @@ from tempest.lib.common.utils import data_utils
|
|||
from tempest.lib import exceptions
|
||||
import testtools
|
||||
|
||||
from magnum.objects.fields import BayStatus
|
||||
from magnum.tests.functional.api import base
|
||||
from magnum.tests.functional.common import config
|
||||
from magnum.tests.functional.common import datagen
|
||||
|
@ -131,56 +130,6 @@ class BayTest(base.BaseTempestTest):
|
|||
resp, model = self.bay_client.get_bay(bay_id)
|
||||
return resp, model
|
||||
|
||||
# (dimtruck) Combining all these tests in one because
|
||||
# they time out on the gate (2 hours not enough)
|
||||
@testtools.testcase.attr('positive')
|
||||
def test_create_list_and_delete_bays(self):
|
||||
gen_model = datagen.valid_bay_data(
|
||||
baymodel_id=self.baymodel.uuid, node_count=1)
|
||||
|
||||
# test bay create
|
||||
_, temp_model = self._create_bay(gen_model)
|
||||
self.assertEqual(BayStatus.CREATE_IN_PROGRESS, temp_model.status)
|
||||
self.assertIsNone(temp_model.status_reason)
|
||||
|
||||
# test bay list
|
||||
resp, model = self.bay_client.list_bays()
|
||||
self.assertEqual(200, resp.status)
|
||||
self.assertGreater(len(model.bays), 0)
|
||||
self.assertIn(
|
||||
temp_model.uuid, list([x['uuid'] for x in model.bays]))
|
||||
|
||||
# test invalid bay update
|
||||
patch_model = datagen.bay_name_patch_data()
|
||||
self.assertRaises(
|
||||
exceptions.BadRequest,
|
||||
self.bay_client.patch_bay,
|
||||
temp_model.uuid, patch_model)
|
||||
|
||||
# test bay delete
|
||||
self._delete_bay(temp_model.uuid)
|
||||
self.bays.remove(temp_model.uuid)
|
||||
|
||||
@testtools.testcase.attr('positive')
|
||||
def test_create_delete_bays_async(self):
|
||||
gen_model = datagen.valid_bay_data(
|
||||
baymodel_id=self.baymodel.uuid, node_count=1)
|
||||
|
||||
# test bay create
|
||||
_, temp_model = self._create_bay(gen_model, is_async=True)
|
||||
self.assertNotIn('status', temp_model)
|
||||
|
||||
# test bay list
|
||||
resp, model = self.bay_client.list_bays()
|
||||
self.assertEqual(200, resp.status)
|
||||
self.assertGreater(len(model.bays), 0)
|
||||
self.assertIn(
|
||||
temp_model.uuid, list([x['uuid'] for x in model.bays]))
|
||||
|
||||
# test bay delete
|
||||
self._delete_bay(temp_model.uuid)
|
||||
self.bays.remove(temp_model.uuid)
|
||||
|
||||
@testtools.testcase.attr('negative')
|
||||
def test_create_bay_for_nonexisting_baymodel(self):
|
||||
gen_model = datagen.valid_bay_data(baymodel_id='this-does-not-exist')
|
||||
|
@ -265,37 +214,3 @@ class BayTest(base.BaseTempestTest):
|
|||
self.assertRaises(
|
||||
exceptions.NotFound,
|
||||
self.bay_client.delete_bay, data_utils.rand_uuid())
|
||||
|
||||
@testtools.testcase.attr('positive')
|
||||
def test_certificate_sign_and_show(self):
|
||||
first_model = datagen.valid_bay_data(baymodel_id=self.baymodel.uuid,
|
||||
name='test')
|
||||
_, bay_model = self._create_bay(first_model)
|
||||
|
||||
# test ca show
|
||||
resp, model = self.cert_client.get_cert(
|
||||
bay_model.uuid)
|
||||
self.LOG.debug("cert resp: %s" % resp)
|
||||
self.assertEqual(200, resp.status)
|
||||
self.assertEqual(model.bay_uuid, bay_model.uuid)
|
||||
self.assertIsNotNone(model.pem)
|
||||
self.assertIn('-----BEGIN CERTIFICATE-----', model.pem)
|
||||
self.assertIn('-----END CERTIFICATE-----', model.pem)
|
||||
|
||||
# test ca sign
|
||||
model = datagen.cert_data(bay_uuid=bay_model.uuid)
|
||||
resp, model = self.cert_client.post_cert(model)
|
||||
self.LOG.debug("cert resp: %s" % resp)
|
||||
self.assertEqual(201, resp.status)
|
||||
self.assertEqual(model.bay_uuid, bay_model.uuid)
|
||||
self.assertIsNotNone(model.pem)
|
||||
self.assertIn('-----BEGIN CERTIFICATE-----', model.pem)
|
||||
self.assertIn('-----END CERTIFICATE-----', model.pem)
|
||||
|
||||
# test ca sign invalid
|
||||
model = datagen.cert_data(bay_uuid=bay_model.uuid,
|
||||
csr_data="invalid_csr")
|
||||
self.assertRaises(
|
||||
exceptions.BadRequest,
|
||||
self.cert_client.post_cert,
|
||||
model)
|
||||
|
|
|
@ -0,0 +1,240 @@
|
|||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
import fixtures
|
||||
|
||||
from oslo_log import log as logging
|
||||
from oslo_utils import uuidutils
|
||||
from tempest.lib.common.utils import data_utils
|
||||
from tempest.lib import exceptions
|
||||
import testtools
|
||||
|
||||
from magnum.tests.functional.api import base
|
||||
from magnum.tests.functional.common import config
|
||||
from magnum.tests.functional.common import datagen
|
||||
|
||||
|
||||
class ClusterTest(base.BaseTempestTest):
|
||||
|
||||
"""Tests for cluster CRUD."""
|
||||
|
||||
LOG = logging.getLogger(__name__)
|
||||
|
||||
def __init__(self, *args, **kwargs):
|
||||
super(ClusterTest, self).__init__(*args, **kwargs)
|
||||
self.clusters = []
|
||||
self.creds = None
|
||||
self.keypair = None
|
||||
self.cluster_template = None
|
||||
self.cluster_template_client = None
|
||||
self.keypairs_client = None
|
||||
self.cluster_client = None
|
||||
self.cert_client = None
|
||||
|
||||
def setUp(self):
|
||||
try:
|
||||
super(ClusterTest, self).setUp()
|
||||
(self.creds, self.keypair) = self.get_credentials_with_keypair(
|
||||
type_of_creds='default')
|
||||
(self.cluster_template_client,
|
||||
self.keypairs_client) = self.get_clients_with_existing_creds(
|
||||
creds=self.creds,
|
||||
type_of_creds='default',
|
||||
request_type='cluster_template')
|
||||
(self.cluster_client, _) = self.get_clients_with_existing_creds(
|
||||
creds=self.creds,
|
||||
type_of_creds='default',
|
||||
request_type='cluster')
|
||||
(self.cert_client, _) = self.get_clients_with_existing_creds(
|
||||
creds=self.creds,
|
||||
type_of_creds='default',
|
||||
request_type='cert')
|
||||
model = datagen.valid_swarm_cluster_template()
|
||||
_, self.cluster_template = self._create_cluster_template(model)
|
||||
|
||||
# NOTE (dimtruck) by default tempest sets timeout to 20 mins.
|
||||
# We need more time.
|
||||
test_timeout = 1800
|
||||
self.useFixture(fixtures.Timeout(test_timeout, gentle=True))
|
||||
except Exception:
|
||||
self.tearDown()
|
||||
raise
|
||||
|
||||
def tearDown(self):
|
||||
try:
|
||||
cluster_list = self.clusters[:]
|
||||
for cluster_id in cluster_list:
|
||||
self._delete_cluster(cluster_id)
|
||||
self.clusters.remove(cluster_id)
|
||||
if self.cluster_template:
|
||||
self._delete_cluster_template(self.cluster_template.uuid)
|
||||
finally:
|
||||
super(ClusterTest, self).tearDown()
|
||||
|
||||
def _create_cluster_template(self, cm_model):
|
||||
self.LOG.debug('We will create a clustertemplate for %s' % cm_model)
|
||||
resp, model = self.cluster_template_client.post_cluster_template(
|
||||
cm_model)
|
||||
return resp, model
|
||||
|
||||
def _delete_cluster_template(self, cm_id):
|
||||
self.LOG.debug('We will delete a clustertemplate for %s' % cm_id)
|
||||
resp, model = self.cluster_template_client.delete_cluster_template(
|
||||
cm_id)
|
||||
return resp, model
|
||||
|
||||
def _create_cluster(self, cluster_model):
|
||||
self.LOG.debug('We will create cluster for %s' % cluster_model)
|
||||
resp, model = self.cluster_client.post_cluster(cluster_model)
|
||||
self.LOG.debug('Response: %s' % resp)
|
||||
self.assertEqual(202, resp.status)
|
||||
self.assertIsNotNone(model.uuid)
|
||||
self.assertTrue(uuidutils.is_uuid_like(model.uuid))
|
||||
self.clusters.append(model.uuid)
|
||||
self.cluster_uuid = model.uuid
|
||||
if config.Config.copy_logs:
|
||||
self.addOnException(self.copy_logs_handler(
|
||||
lambda: list(
|
||||
[self._get_cluster_by_id(model.uuid)[1].master_addresses,
|
||||
self._get_cluster_by_id(model.uuid)[1].node_addresses]),
|
||||
self.cluster_template.coe,
|
||||
self.keypair))
|
||||
self.cluster_client.wait_for_created_cluster(model.uuid,
|
||||
delete_on_error=False)
|
||||
return resp, model
|
||||
|
||||
def _delete_cluster(self, cluster_id):
|
||||
self.LOG.debug('We will delete a cluster for %s' % cluster_id)
|
||||
resp, model = self.cluster_client.delete_cluster(cluster_id)
|
||||
self.assertEqual(204, resp.status)
|
||||
self.cluster_client.wait_for_cluster_to_delete(cluster_id)
|
||||
self.assertRaises(exceptions.NotFound, self.cert_client.get_cert,
|
||||
cluster_id)
|
||||
return resp, model
|
||||
|
||||
def _get_cluster_by_id(self, cluster_id):
|
||||
resp, model = self.cluster_client.get_cluster(cluster_id)
|
||||
return resp, model
|
||||
|
||||
# (dimtruck) Combining all these tests in one because
|
||||
# they time out on the gate (2 hours not enough)
|
||||
@testtools.testcase.attr('positive')
|
||||
def test_create_list_sign_delete_clusters(self):
|
||||
gen_model = datagen.valid_cluster_data(
|
||||
cluster_template_id=self.cluster_template.uuid, node_count=1)
|
||||
|
||||
# test cluster create
|
||||
_, cluster_model = self._create_cluster(gen_model)
|
||||
self.assertNotIn('status', cluster_model)
|
||||
|
||||
# test cluster list
|
||||
resp, cluster_list_model = self.cluster_client.list_clusters()
|
||||
self.assertEqual(200, resp.status)
|
||||
self.assertGreater(len(cluster_list_model.clusters), 0)
|
||||
self.assertIn(
|
||||
cluster_model.uuid, list([x['uuid']
|
||||
for x in cluster_list_model.clusters]))
|
||||
|
||||
# test invalid cluster update
|
||||
patch_model = datagen.cluster_name_patch_data()
|
||||
self.assertRaises(
|
||||
exceptions.BadRequest,
|
||||
self.cluster_client.patch_cluster,
|
||||
cluster_model.uuid, patch_model)
|
||||
|
||||
# test ca show
|
||||
resp, cert_model = self.cert_client.get_cert(
|
||||
cluster_model.uuid)
|
||||
self.LOG.debug("cert resp: %s" % resp)
|
||||
self.assertEqual(200, resp.status)
|
||||
self.assertEqual(cert_model.bay_uuid, cluster_model.uuid)
|
||||
self.assertIsNotNone(cert_model.pem)
|
||||
self.assertIn('-----BEGIN CERTIFICATE-----', cert_model.pem)
|
||||
self.assertIn('-----END CERTIFICATE-----', cert_model.pem)
|
||||
|
||||
# test ca sign
|
||||
cert_data_model = datagen.cert_data(cluster_model.uuid)
|
||||
resp, cert_model = self.cert_client.post_cert(cert_data_model)
|
||||
self.LOG.debug("cert resp: %s" % resp)
|
||||
self.assertEqual(201, resp.status)
|
||||
self.assertEqual(cert_model.bay_uuid, cluster_model.uuid)
|
||||
self.assertIsNotNone(cert_model.pem)
|
||||
self.assertIn('-----BEGIN CERTIFICATE-----', cert_model.pem)
|
||||
self.assertIn('-----END CERTIFICATE-----', cert_model.pem)
|
||||
|
||||
# test ca sign invalid
|
||||
cert_data_model = datagen.cert_data(cluster_model.uuid,
|
||||
csr_data="invalid_csr")
|
||||
self.assertRaises(
|
||||
exceptions.BadRequest,
|
||||
self.cert_client.post_cert,
|
||||
cert_data_model)
|
||||
|
||||
# test cluster delete
|
||||
self._delete_cluster(cluster_model.uuid)
|
||||
self.clusters.remove(cluster_model.uuid)
|
||||
|
||||
@testtools.testcase.attr('negative')
|
||||
def test_create_cluster_for_nonexisting_cluster_template(self):
|
||||
cm_id = 'this-does-not-exist'
|
||||
gen_model = datagen.valid_cluster_data(cluster_template_id=cm_id)
|
||||
self.assertRaises(
|
||||
exceptions.BadRequest,
|
||||
self.cluster_client.post_cluster, gen_model)
|
||||
|
||||
@testtools.testcase.attr('negative')
|
||||
def test_create_cluster_with_node_count_0(self):
|
||||
gen_model = datagen.valid_cluster_data(
|
||||
cluster_template_id=self.cluster_template.uuid, node_count=0)
|
||||
self.assertRaises(
|
||||
exceptions.BadRequest,
|
||||
self.cluster_client.post_cluster, gen_model)
|
||||
|
||||
@testtools.testcase.attr('negative')
|
||||
def test_create_cluster_with_zero_masters(self):
|
||||
uuid = self.cluster_template.uuid
|
||||
gen_model = datagen.valid_cluster_data(cluster_template_id=uuid,
|
||||
master_count=0)
|
||||
self.assertRaises(
|
||||
exceptions.BadRequest,
|
||||
self.cluster_client.post_cluster, gen_model)
|
||||
|
||||
@testtools.testcase.attr('negative')
|
||||
def test_create_cluster_with_nonexisting_flavor(self):
|
||||
gen_model = \
|
||||
datagen.cluster_template_data_with_valid_keypair_image_flavor()
|
||||
resp, cluster_template = self._create_cluster_template(gen_model)
|
||||
self.assertEqual(201, resp.status)
|
||||
self.assertIsNotNone(cluster_template.uuid)
|
||||
|
||||
uuid = cluster_template.uuid
|
||||
gen_model = datagen.valid_cluster_data(cluster_template_id=uuid)
|
||||
gen_model.flavor_id = 'aaa'
|
||||
self.assertRaises(exceptions.BadRequest,
|
||||
self.cluster_client.post_cluster, gen_model)
|
||||
|
||||
resp, _ = self._delete_cluster_template(cluster_template.uuid)
|
||||
self.assertEqual(204, resp.status)
|
||||
|
||||
@testtools.testcase.attr('negative')
|
||||
def test_update_cluster_for_nonexisting_cluster(self):
|
||||
patch_model = datagen.cluster_name_patch_data()
|
||||
|
||||
self.assertRaises(
|
||||
exceptions.NotFound,
|
||||
self.cluster_client.patch_cluster, 'fooo', patch_model)
|
||||
|
||||
@testtools.testcase.attr('negative')
|
||||
def test_delete_cluster_for_nonexisting_cluster(self):
|
||||
self.assertRaises(
|
||||
exceptions.NotFound,
|
||||
self.cluster_client.delete_cluster, data_utils.rand_uuid())
|
|
@ -0,0 +1,239 @@
|
|||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
|
||||
from tempest.lib.common.utils import data_utils
|
||||
from tempest.lib import exceptions
|
||||
import testtools
|
||||
|
||||
from magnum.tests.functional.api import base
|
||||
from magnum.tests.functional.common import datagen
|
||||
|
||||
|
||||
class ClusterTemplateTest(base.BaseTempestTest):
|
||||
|
||||
"""Tests for clustertemplate CRUD."""
|
||||
|
||||
def __init__(self, *args, **kwargs):
|
||||
super(ClusterTemplateTest, self).__init__(*args, **kwargs)
|
||||
self.cluster_templates = []
|
||||
self.cluster_template_client = None
|
||||
self.keypairs_client = None
|
||||
|
||||
def setUp(self):
|
||||
try:
|
||||
super(ClusterTemplateTest, self).setUp()
|
||||
(self.cluster_template_client,
|
||||
self.keypairs_client) = self.get_clients_with_new_creds(
|
||||
type_of_creds='default',
|
||||
request_type='cluster_template')
|
||||
except Exception:
|
||||
self.tearDown()
|
||||
raise
|
||||
|
||||
def tearDown(self):
|
||||
for cluster_template_id in self.cluster_templates:
|
||||
self._delete_cluster_template(cluster_template_id)
|
||||
self.cluster_templates.remove(cluster_template_id)
|
||||
super(ClusterTemplateTest, self).tearDown()
|
||||
|
||||
def _create_cluster_template(self, cmodel_model):
|
||||
resp, model = \
|
||||
self.cluster_template_client.post_cluster_template(cmodel_model)
|
||||
self.assertEqual(201, resp.status)
|
||||
self.cluster_templates.append(model.uuid)
|
||||
return resp, model
|
||||
|
||||
def _delete_cluster_template(self, model_id):
|
||||
resp, model = \
|
||||
self.cluster_template_client.delete_cluster_template(model_id)
|
||||
self.assertEqual(204, resp.status)
|
||||
return resp, model
|
||||
|
||||
@testtools.testcase.attr('positive')
|
||||
def test_list_cluster_templates(self):
|
||||
gen_model = \
|
||||
datagen.cluster_template_data_with_valid_keypair_image_flavor()
|
||||
_, temp_model = self._create_cluster_template(gen_model)
|
||||
resp, model = self.cluster_template_client.list_cluster_templates()
|
||||
self.assertEqual(200, resp.status)
|
||||
self.assertGreater(len(model.clustertemplates), 0)
|
||||
self.assertIn(
|
||||
temp_model.uuid,
|
||||
list([x['uuid'] for x in model.clustertemplates]))
|
||||
|
||||
@testtools.testcase.attr('positive')
|
||||
def test_create_cluster_template(self):
|
||||
gen_model = \
|
||||
datagen.cluster_template_data_with_valid_keypair_image_flavor()
|
||||
resp, model = self._create_cluster_template(gen_model)
|
||||
|
||||
@testtools.testcase.attr('positive')
|
||||
def test_create_get_public_cluster_template(self):
|
||||
gen_model = datagen.valid_swarm_cluster_template(is_public=True)
|
||||
resp, model = self._create_cluster_template(gen_model)
|
||||
|
||||
resp, model = \
|
||||
self.cluster_template_client.get_cluster_template(model.uuid)
|
||||
self.assertEqual(200, resp.status)
|
||||
self.assertTrue(model.public)
|
||||
|
||||
@testtools.testcase.attr('positive')
|
||||
def test_update_cluster_template_public_by_uuid(self):
|
||||
path = "/public"
|
||||
gen_model = \
|
||||
datagen.cluster_template_data_with_valid_keypair_image_flavor()
|
||||
resp, old_model = self._create_cluster_template(gen_model)
|
||||
|
||||
patch_model = datagen.cluster_template_replace_patch_data(path,
|
||||
value=True)
|
||||
resp, new_model = self.cluster_template_client.patch_cluster_template(
|
||||
old_model.uuid, patch_model)
|
||||
self.assertEqual(200, resp.status)
|
||||
|
||||
resp, model = self.cluster_template_client.get_cluster_template(
|
||||
new_model.uuid)
|
||||
self.assertEqual(200, resp.status)
|
||||
self.assertTrue(model.public)
|
||||
|
||||
@testtools.testcase.attr('positive')
|
||||
def test_update_cluster_template_by_uuid(self):
|
||||
gen_model = \
|
||||
datagen.cluster_template_data_with_valid_keypair_image_flavor()
|
||||
resp, old_model = self._create_cluster_template(gen_model)
|
||||
|
||||
patch_model = datagen.cluster_template_name_patch_data()
|
||||
resp, new_model = self.cluster_template_client.patch_cluster_template(
|
||||
old_model.uuid, patch_model)
|
||||
self.assertEqual(200, resp.status)
|
||||
|
||||
resp, model = \
|
||||
self.cluster_template_client.get_cluster_template(new_model.uuid)
|
||||
self.assertEqual(200, resp.status)
|
||||
self.assertEqual(old_model.uuid, new_model.uuid)
|
||||
self.assertEqual(model.name, new_model.name)
|
||||
|
||||
@testtools.testcase.attr('positive')
|
||||
def test_delete_cluster_template_by_uuid(self):
|
||||
gen_model = \
|
||||
datagen.cluster_template_data_with_valid_keypair_image_flavor()
|
||||
resp, model = self._create_cluster_template(gen_model)
|
||||
resp, _ = self.cluster_template_client.delete_cluster_template(
|
||||
model.uuid)
|
||||
self.assertEqual(204, resp.status)
|
||||
self.cluster_templates.remove(model.uuid)
|
||||
|
||||
@testtools.testcase.attr('positive')
|
||||
def test_delete_cluster_template_by_name(self):
|
||||
gen_model = \
|
||||
datagen.cluster_template_data_with_valid_keypair_image_flavor()
|
||||
resp, model = self._create_cluster_template(gen_model)
|
||||
resp, _ = self.cluster_template_client.delete_cluster_template(
|
||||
model.name)
|
||||
self.assertEqual(204, resp.status)
|
||||
self.cluster_templates.remove(model.uuid)
|
||||
|
||||
@testtools.testcase.attr('negative')
|
||||
def test_get_cluster_template_by_uuid_404(self):
|
||||
self.assertRaises(
|
||||
exceptions.NotFound,
|
||||
self.cluster_template_client.get_cluster_template,
|
||||
data_utils.rand_uuid())
|
||||
|
||||
@testtools.testcase.attr('negative')
|
||||
def test_update_cluster_template_404(self):
|
||||
patch_model = datagen.cluster_template_name_patch_data()
|
||||
|
||||
self.assertRaises(
|
||||
exceptions.NotFound,
|
||||
self.cluster_template_client.patch_cluster_template,
|
||||
data_utils.rand_uuid(), patch_model)
|
||||
|
||||
@testtools.testcase.attr('negative')
|
||||
def test_delete_cluster_template_404(self):
|
||||
self.assertRaises(
|
||||
exceptions.NotFound,
|
||||
self.cluster_template_client.delete_cluster_template,
|
||||
data_utils.rand_uuid())
|
||||
|
||||
@testtools.testcase.attr('negative')
|
||||
def test_get_cluster_template_by_name_404(self):
|
||||
self.assertRaises(
|
||||
exceptions.NotFound,
|
||||
self.cluster_template_client.get_cluster_template, 'fooo')
|
||||
|
||||
@testtools.testcase.attr('negative')
|
||||
def test_update_cluster_template_name_not_found(self):
|
||||
patch_model = datagen.cluster_template_name_patch_data()
|
||||
|
||||
self.assertRaises(
|
||||
exceptions.NotFound,
|
||||
self.cluster_template_client.patch_cluster_template,
|
||||
'fooo', patch_model)
|
||||
|
||||
@testtools.testcase.attr('negative')
|
||||
def test_delete_cluster_template_by_name_404(self):
|
||||
self.assertRaises(
|
||||
exceptions.NotFound,
|
||||
self.cluster_template_client.get_cluster_template, 'fooo')
|
||||
|
||||
@testtools.testcase.attr('negative')
|
||||
def test_create_cluster_template_missing_image(self):
|
||||
gen_model = datagen.cluster_template_data_with_missing_image()
|
||||
self.assertRaises(
|
||||
exceptions.BadRequest,
|
||||
self.cluster_template_client.post_cluster_template, gen_model)
|
||||
|
||||
@testtools.testcase.attr('negative')
|
||||
def test_create_cluster_template_missing_flavor(self):
|
||||
gen_model = datagen.cluster_template_data_with_missing_flavor()
|
||||
self.assertRaises(
|
||||
exceptions.BadRequest,
|
||||
self.cluster_template_client.post_cluster_template, gen_model)
|
||||
|
||||
@testtools.testcase.attr('negative')
|
||||
def test_create_cluster_template_missing_keypair(self):
|
||||
gen_model = \
|
||||
datagen.cluster_template_data_with_missing_keypair()
|
||||
self.assertRaises(
|
||||
exceptions.NotFound,
|
||||
self.cluster_template_client.post_cluster_template, gen_model)
|
||||
|
||||
@testtools.testcase.attr('negative')
|
||||
def test_update_cluster_template_invalid_patch(self):
|
||||
# get json object
|
||||
gen_model = \
|
||||
datagen.cluster_template_data_with_valid_keypair_image_flavor()
|
||||
resp, old_model = self._create_cluster_template(gen_model)
|
||||
|
||||
self.assertRaises(
|
||||
exceptions.BadRequest,
|
||||
self.cluster_template_client.patch_cluster_template,
|
||||
data_utils.rand_uuid(), gen_model)
|
||||
|
||||
@testtools.testcase.attr('negative')
|
||||
def test_create_cluster_template_invalid_network_driver(self):
|
||||
gen_model = \
|
||||
datagen.cluster_template_data_with_valid_keypair_image_flavor()
|
||||
gen_model.network_driver = 'invalid_network_driver'
|
||||
self.assertRaises(
|
||||
exceptions.BadRequest,
|
||||
self.cluster_template_client.post_cluster_template, gen_model)
|
||||
|
||||
@testtools.testcase.attr('negative')
|
||||
def test_create_cluster_template_invalid_volume_driver(self):
|
||||
gen_model = \
|
||||
datagen.cluster_template_data_with_valid_keypair_image_flavor()
|
||||
gen_model.volume_driver = 'invalid_volume_driver'
|
||||
self.assertRaises(
|
||||
exceptions.BadRequest,
|
||||
self.cluster_template_client.post_cluster_template, gen_model)
|
|
@ -79,9 +79,9 @@ class BaseMagnumTest(base.BaseTestCase):
|
|||
])
|
||||
except Exception:
|
||||
cls.LOG.error(msg)
|
||||
msg = (_LE("failed to copy from %{node_address}s "
|
||||
"to %{base_path}s%{log_name}s-"
|
||||
"%{node_address}s") %
|
||||
msg = (_LE("failed to copy from %(node_address)s "
|
||||
"to %(base_path)s%(log_name)s-"
|
||||
"%(node_address)s") %
|
||||
{'node_address': node_address,
|
||||
'base_path': "/opt/stack/logs/bay-nodes/",
|
||||
'log_name': log_name})
|
||||
|
|
|
@ -23,6 +23,10 @@ from magnum.tests.functional.api.v1.models import baymodel_model
|
|||
from magnum.tests.functional.api.v1.models import baymodelpatch_model
|
||||
from magnum.tests.functional.api.v1.models import baypatch_model
|
||||
from magnum.tests.functional.api.v1.models import cert_model
|
||||
from magnum.tests.functional.api.v1.models import cluster_model
|
||||
from magnum.tests.functional.api.v1.models import cluster_template_model
|
||||
from magnum.tests.functional.api.v1.models import cluster_templatepatch_model
|
||||
from magnum.tests.functional.api.v1.models import clusterpatch_model
|
||||
from magnum.tests.functional.common import config
|
||||
|
||||
|
||||
|
@ -334,3 +338,277 @@ def cert_data(bay_uuid, csr_data=None):
|
|||
model = cert_model.CertEntity.from_dict(data)
|
||||
|
||||
return model
|
||||
|
||||
|
||||
def cluster_template_data(**kwargs):
|
||||
"""Generates random cluster_template data
|
||||
|
||||
Keypair and image id cannot be random for the cluster_template to be valid
|
||||
due to validations for the presence of keypair and image id prior to
|
||||
cluster_template creation.
|
||||
|
||||
:param keypair_id: keypair name
|
||||
:param image_id: image id or name
|
||||
:returns: ClusterTemplateEntity with generated data
|
||||
"""
|
||||
|
||||
data = {
|
||||
"name": data_utils.rand_name('cluster'),
|
||||
"coe": "swarm",
|
||||
"tls_disabled": False,
|
||||
"network_driver": None,
|
||||
"volume_driver": None,
|
||||
"docker_volume_size": 3,
|
||||
"labels": {},
|
||||
"public": False,
|
||||
"fixed_network": "192.168.0.0/24",
|
||||
"dns_nameserver": "8.8.8.8",
|
||||
"flavor_id": data_utils.rand_name('cluster'),
|
||||
"master_flavor_id": data_utils.rand_name('cluster'),
|
||||
"external_network_id": config.Config.nic_id,
|
||||
"keypair_id": data_utils.rand_name('cluster'),
|
||||
"image_id": data_utils.rand_name('cluster')
|
||||
}
|
||||
|
||||
data.update(kwargs)
|
||||
model = cluster_template_model.ClusterTemplateEntity.from_dict(data)
|
||||
|
||||
return model
|
||||
|
||||
|
||||
def cluster_template_replace_patch_data(path,
|
||||
value=data_utils.rand_name('cluster')):
|
||||
"""Generates random ClusterTemplate patch data
|
||||
|
||||
:param path: path to replace
|
||||
:param value: value to replace in patch
|
||||
:returns: ClusterTemplatePatchCollection with generated data
|
||||
"""
|
||||
|
||||
data = [{
|
||||
"path": path,
|
||||
"value": value,
|
||||
"op": "replace"
|
||||
}]
|
||||
collection = cluster_templatepatch_model.ClusterTemplatePatchCollection
|
||||
return collection.from_dict(data)
|
||||
|
||||
|
||||
def cluster_template_remove_patch_data(path):
|
||||
"""Generates ClusterTempalte patch data by removing value
|
||||
|
||||
:param path: path to remove
|
||||
:returns: BayModelPatchCollection with generated data
|
||||
"""
|
||||
|
||||
data = [{
|
||||
"path": path,
|
||||
"op": "remove"
|
||||
}]
|
||||
collection = cluster_templatepatch_model.ClusterTemplatePatchCollection
|
||||
return collection.from_dict(data)
|
||||
|
||||
|
||||
def cluster_template_name_patch_data(name=data_utils.rand_name('cluster')):
|
||||
"""Generates random cluster_template patch data
|
||||
|
||||
:param name: name to replace in patch
|
||||
:returns: ClusterTemplatePatchCollection with generated data
|
||||
"""
|
||||
|
||||
data = [{
|
||||
"path": "/name",
|
||||
"value": name,
|
||||
"op": "replace"
|
||||
}]
|
||||
collection = cluster_templatepatch_model.ClusterTemplatePatchCollection
|
||||
return collection.from_dict(data)
|
||||
|
||||
|
||||
def cluster_template_flavor_patch_data(flavor=data_utils.rand_name('cluster')):
|
||||
"""Generates random cluster_template patch data
|
||||
|
||||
:param flavor: flavor to replace in patch
|
||||
:returns: ClusterTemplatePatchCollection with generated data
|
||||
"""
|
||||
|
||||
data = [{
|
||||
"path": "/flavor_id",
|
||||
"value": flavor,
|
||||
"op": "replace"
|
||||
}]
|
||||
collection = cluster_templatepatch_model.ClusterTemplatePatchCollection
|
||||
return collection.from_dict(data)
|
||||
|
||||
|
||||
def cluster_template_data_with_valid_keypair_image_flavor():
|
||||
"""Generates random clustertemplate data with valid data
|
||||
|
||||
:returns: ClusterTemplateEntity with generated data
|
||||
"""
|
||||
master_flavor = config.Config.master_flavor_id
|
||||
return cluster_template_data(keypair_id=config.Config.keypair_id,
|
||||
image_id=config.Config.image_id,
|
||||
flavor_id=config.Config.flavor_id,
|
||||
master_flavor_id=master_flavor)
|
||||
|
||||
|
||||
def cluster_template_data_with_missing_image():
|
||||
"""Generates random cluster_template data with missing image
|
||||
|
||||
:returns: ClusterTemplateEntity with generated data
|
||||
"""
|
||||
|
||||
return cluster_template_data(
|
||||
keypair_id=config.Config.keypair_id,
|
||||
flavor_id=config.Config.flavor_id,
|
||||
master_flavor_id=config.Config.master_flavor_id)
|
||||
|
||||
|
||||
def cluster_template_data_with_missing_flavor():
|
||||
"""Generates random cluster_template data with missing flavor
|
||||
|
||||
:returns: ClusterTemplateEntity with generated data
|
||||
"""
|
||||
|
||||
return cluster_template_data(keypair_id=config.Config.keypair_id,
|
||||
image_id=config.Config.image_id)
|
||||
|
||||
|
||||
def cluster_template_data_with_missing_keypair():
|
||||
"""Generates random cluster_template data with missing keypair
|
||||
|
||||
:returns: ClusterTemplateEntity with generated data
|
||||
"""
|
||||
|
||||
return cluster_template_data(
|
||||
image_id=config.Config.image_id,
|
||||
flavor_id=config.Config.flavor_id,
|
||||
master_flavor_id=config.Config.master_flavor_id)
|
||||
|
||||
|
||||
def cluster_template_valid_data_with_specific_coe(coe):
|
||||
"""Generates random cluster_template data with valid keypair and image
|
||||
|
||||
:param coe: coe
|
||||
:returns: ClusterTemplateEntity with generated data
|
||||
"""
|
||||
|
||||
return cluster_template_data(keypair_id=config.Config.keypair_id,
|
||||
image_id=config.Config.image_id, coe=coe)
|
||||
|
||||
|
||||
def valid_swarm_cluster_template(is_public=False):
|
||||
"""Generates a valid swarm cluster_template with valid data
|
||||
|
||||
:returns: ClusterTemplateEntity with generated data
|
||||
"""
|
||||
master_flavor_id = config.Config.master_flavor_id
|
||||
return cluster_template_data(image_id=config.Config.image_id,
|
||||
fixed_network="192.168.0.0/24",
|
||||
flavor_id=config.Config.flavor_id,
|
||||
public=is_public,
|
||||
dns_nameserver=config.Config.dns_nameserver,
|
||||
master_flavor_id=master_flavor_id,
|
||||
keypair_id=config.Config.keypair_id,
|
||||
coe="swarm", docker_volume_size=3,
|
||||
cluster_distro=None,
|
||||
external_network_id=config.Config.nic_id,
|
||||
http_proxy=None, https_proxy=None,
|
||||
no_proxy=None, network_driver=None,
|
||||
volume_driver=None, labels={},
|
||||
tls_disabled=False)
|
||||
|
||||
|
||||
def cluster_data(name=data_utils.rand_name('cluster'),
|
||||
cluster_template_id=data_utils.rand_uuid(),
|
||||
node_count=random_int(1, 5), discovery_url=gen_random_ip(),
|
||||
create_timeout=random_int(1, 30),
|
||||
master_count=random_int(1, 5)):
|
||||
"""Generates random cluster data
|
||||
|
||||
cluster_template_id cannot be random for the cluster to be valid due to
|
||||
validations for the presence of clustertemplate prior to clustertemplate
|
||||
creation.
|
||||
|
||||
:param name: cluster name (must be unique)
|
||||
:param cluster_template_id: clustertemplate unique id (must already exist)
|
||||
:param node_count: number of agents for cluster
|
||||
:param discovery_url: url provided for node discovery
|
||||
:param create_timeout: timeout in minutes for cluster create
|
||||
:param master_count: number of master nodes for the cluster
|
||||
:returns: ClusterEntity with generated data
|
||||
"""
|
||||
|
||||
data = {
|
||||
"name": name,
|
||||
"cluster_template_id": cluster_template_id,
|
||||
"node_count": node_count,
|
||||
"discovery_url": None,
|
||||
"create_timeout": create_timeout,
|
||||
"master_count": master_count
|
||||
}
|
||||
model = cluster_model.ClusterEntity.from_dict(data)
|
||||
|
||||
return model
|
||||
|
||||
|
||||
def valid_cluster_data(cluster_template_id,
|
||||
name=data_utils.rand_name('cluster'),
|
||||
node_count=1, master_count=1, create_timeout=None):
|
||||
"""Generates random cluster data with valid
|
||||
|
||||
:param cluster_template_id: clustertemplate unique id that already exists
|
||||
:param name: cluster name (must be unique)
|
||||
:param node_count: number of agents for cluster
|
||||
:returns: ClusterEntity with generated data
|
||||
"""
|
||||
|
||||
return cluster_data(cluster_template_id=cluster_template_id, name=name,
|
||||
master_count=master_count, node_count=node_count,
|
||||
create_timeout=create_timeout)
|
||||
|
||||
|
||||
def cluster_name_patch_data(name=data_utils.rand_name('cluster')):
|
||||
"""Generates random clustertemplate patch data
|
||||
|
||||
:param name: name to replace in patch
|
||||
:returns: ClusterPatchCollection with generated data
|
||||
"""
|
||||
|
||||
data = [{
|
||||
"path": "/name",
|
||||
"value": name,
|
||||
"op": "replace"
|
||||
}]
|
||||
return clusterpatch_model.ClusterPatchCollection.from_dict(data)
|
||||
|
||||
|
||||
def cluster_api_addy_patch_data(address='0.0.0.0'):
|
||||
"""Generates random cluster patch data
|
||||
|
||||
:param name: name to replace in patch
|
||||
:returns: ClusterPatchCollection with generated data
|
||||
"""
|
||||
|
||||
data = [{
|
||||
"path": "/api_address",
|
||||
"value": address,
|
||||
"op": "replace"
|
||||
}]
|
||||
return clusterpatch_model.ClusterPatchCollection.from_dict(data)
|
||||
|
||||
|
||||
def cluster_node_count_patch_data(node_count=2):
|
||||
"""Generates random cluster patch data
|
||||
|
||||
:param name: name to replace in patch
|
||||
:returns: ClusterPatchCollection with generated data
|
||||
"""
|
||||
|
||||
data = [{
|
||||
"path": "/node_count",
|
||||
"value": node_count,
|
||||
"op": "replace"
|
||||
}]
|
||||
return clusterpatch_model.ClusterPatchCollection.from_dict(data)
|
||||
|
|
|
@ -16,6 +16,8 @@ from tempest.common import credentials_factory as common_creds
|
|||
from magnum.tests.functional.api.v1.clients import bay_client
|
||||
from magnum.tests.functional.api.v1.clients import baymodel_client
|
||||
from magnum.tests.functional.api.v1.clients import cert_client
|
||||
from magnum.tests.functional.api.v1.clients import cluster_client
|
||||
from magnum.tests.functional.api.v1.clients import cluster_template_client
|
||||
from magnum.tests.functional.api.v1.clients import magnum_service_client
|
||||
from magnum.tests.functional.common import client
|
||||
from magnum.tests.functional.common import config
|
||||
|
@ -29,17 +31,21 @@ class Manager(clients.Manager):
|
|||
super(Manager, self).__init__(credentials, 'container-infra')
|
||||
self.auth_provider.orig_base_url = self.auth_provider.base_url
|
||||
self.auth_provider.base_url = self.bypassed_base_url
|
||||
auth = self.auth_provider
|
||||
if request_type == 'baymodel':
|
||||
self.client = baymodel_client.BayModelClient(self.auth_provider)
|
||||
self.client = baymodel_client.BayModelClient(auth)
|
||||
elif request_type == 'bay':
|
||||
self.client = bay_client.BayClient(self.auth_provider)
|
||||
self.client = bay_client.BayClient(auth)
|
||||
elif request_type == 'cert':
|
||||
self.client = cert_client.CertClient(self.auth_provider)
|
||||
self.client = cert_client.CertClient(auth)
|
||||
elif request_type == 'cluster_template':
|
||||
self.client = cluster_template_client.ClusterTemplateClient(auth)
|
||||
elif request_type == 'cluster':
|
||||
self.client = cluster_client.ClusterClient(auth)
|
||||
elif request_type == 'service':
|
||||
self.client = magnum_service_client.MagnumServiceClient(
|
||||
self.auth_provider)
|
||||
self.client = magnum_service_client.MagnumServiceClient(auth)
|
||||
else:
|
||||
self.client = client.MagnumClient(self.auth_provider)
|
||||
self.client = client.MagnumClient(auth)
|
||||
|
||||
def bypassed_base_url(self, filters, auth_data=None):
|
||||
if (config.Config.magnum_url and
|
||||
|
|
|
@ -33,20 +33,20 @@ class TestRootController(api_base.FunctionalTest):
|
|||
super(TestRootController, self).setUp()
|
||||
self.root_expected = {
|
||||
u'description': u'Magnum is an OpenStack project which '
|
||||
'aims to provide container management.',
|
||||
'aims to provide container cluster management.',
|
||||
u'name': u'OpenStack Magnum API',
|
||||
u'versions': [{u'id': u'v1',
|
||||
u'links':
|
||||
[{u'href': u'http://localhost/v1/',
|
||||
u'rel': u'self'}],
|
||||
u'links':
|
||||
[{u'href': u'http://localhost/v1/',
|
||||
u'rel': u'self'}],
|
||||
u'status': u'CURRENT',
|
||||
u'max_version': u'1.3',
|
||||
u'min_version': u'1.1'}]}
|
||||
|
||||
self.v1_expected = {
|
||||
u'media_types':
|
||||
[{u'base': u'application/json',
|
||||
u'type': u'application/vnd.openstack.magnum.v1+json'}],
|
||||
[{u'base': u'application/json',
|
||||
u'type': u'application/vnd.openstack.magnum.v1+json'}],
|
||||
u'links': [{u'href': u'http://localhost/v1/',
|
||||
u'rel': u'self'},
|
||||
{u'href':
|
||||
|
@ -61,6 +61,15 @@ class TestRootController(api_base.FunctionalTest):
|
|||
u'rel': u'self'},
|
||||
{u'href': u'http://localhost/baymodels/',
|
||||
u'rel': u'bookmark'}],
|
||||
u'clusters': [{u'href': u'http://localhost/v1/clusters/',
|
||||
u'rel': u'self'},
|
||||
{u'href': u'http://localhost/clusters/',
|
||||
u'rel': u'bookmark'}],
|
||||
u'clustertemplates':
|
||||
[{u'href': u'http://localhost/v1/clustertemplates/',
|
||||
u'rel': u'self'},
|
||||
{u'href': u'http://localhost/clustertemplates/',
|
||||
u'rel': u'bookmark'}],
|
||||
u'id': u'v1',
|
||||
u'certificates': [{u'href': u'http://localhost/v1/certificates/',
|
||||
u'rel': u'self'},
|
||||
|
@ -199,7 +208,6 @@ class TestHeathcheck(api_base.FunctionalTest):
|
|||
|
||||
|
||||
class TestV1Routing(api_base.FunctionalTest):
|
||||
|
||||
def test_route_checks_version(self):
|
||||
self.get_json('/')
|
||||
self._check_version.assert_called_once_with(mock.ANY,
|
||||
|
@ -207,7 +215,6 @@ class TestV1Routing(api_base.FunctionalTest):
|
|||
|
||||
|
||||
class TestCheckVersions(test_base.TestCase):
|
||||
|
||||
def setUp(self):
|
||||
super(TestCheckVersions, self).setUp()
|
||||
|
||||
|
|
|
@ -0,0 +1,895 @@
|
|||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
# You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
import datetime
|
||||
|
||||
import mock
|
||||
from oslo_config import cfg
|
||||
from oslo_utils import timeutils
|
||||
from oslo_utils import uuidutils
|
||||
|
||||
from magnum.api import attr_validator
|
||||
from magnum.api.controllers.v1 import cluster as api_cluster
|
||||
from magnum.common import exception
|
||||
from magnum.conductor import api as rpcapi
|
||||
from magnum import objects
|
||||
from magnum.tests import base
|
||||
from magnum.tests.unit.api import base as api_base
|
||||
from magnum.tests.unit.api import utils as apiutils
|
||||
from magnum.tests.unit.objects import utils as obj_utils
|
||||
|
||||
|
||||
class TestClusterObject(base.TestCase):
|
||||
def test_cluster_init(self):
|
||||
cluster_dict = apiutils.cluster_post_data(cluster_template_id=None)
|
||||
del cluster_dict['node_count']
|
||||
del cluster_dict['master_count']
|
||||
del cluster_dict['create_timeout']
|
||||
cluster = api_cluster.Cluster(**cluster_dict)
|
||||
self.assertEqual(1, cluster.node_count)
|
||||
self.assertEqual(1, cluster.master_count)
|
||||
self.assertEqual(60, cluster.create_timeout)
|
||||
|
||||
|
||||
class TestListCluster(api_base.FunctionalTest):
|
||||
_cluster_attrs = ("name", "cluster_template_id", "node_count", "status",
|
||||
"master_count", "stack_id", "create_timeout")
|
||||
|
||||
_expand_cluster_attrs = ("name", "cluster_template_id", "node_count",
|
||||
"status", "api_address", "discovery_url",
|
||||
"node_addresses", "master_count",
|
||||
"master_addresses", "stack_id",
|
||||
"create_timeout", "status_reason")
|
||||
|
||||
def setUp(self):
|
||||
super(TestListCluster, self).setUp()
|
||||
obj_utils.create_test_cluster_template(self.context)
|
||||
|
||||
def test_empty(self):
|
||||
response = self.get_json('/clusters')
|
||||
self.assertEqual([], response['clusters'])
|
||||
|
||||
def test_one(self):
|
||||
cluster = obj_utils.create_test_cluster(self.context)
|
||||
response = self.get_json('/clusters')
|
||||
self.assertEqual(cluster.uuid, response['clusters'][0]["uuid"])
|
||||
self._verify_attrs(self._cluster_attrs, response['clusters'][0])
|
||||
|
||||
# Verify attrs do not appear from cluster's get_all response
|
||||
none_attrs = \
|
||||
set(self._expand_cluster_attrs) - set(self._cluster_attrs)
|
||||
self._verify_attrs(none_attrs, response['clusters'][0],
|
||||
positive=False)
|
||||
|
||||
def test_get_one(self):
|
||||
cluster = obj_utils.create_test_cluster(self.context)
|
||||
response = self.get_json('/clusters/%s' % cluster['uuid'])
|
||||
self.assertEqual(cluster.uuid, response['uuid'])
|
||||
self._verify_attrs(self._expand_cluster_attrs, response)
|
||||
|
||||
@mock.patch('magnum.common.clients.OpenStackClients.heat')
|
||||
def test_get_one_failed_cluster(self, mock_heat):
|
||||
fake_resources = mock.MagicMock()
|
||||
fake_resources.resource_name = 'fake_name'
|
||||
fake_resources.resource_status_reason = 'fake_reason'
|
||||
|
||||
ht = mock.MagicMock()
|
||||
ht.resources.list.return_value = [fake_resources]
|
||||
mock_heat.return_value = ht
|
||||
|
||||
cluster = obj_utils.create_test_cluster(self.context,
|
||||
status='CREATE_FAILED')
|
||||
response = self.get_json('/clusters/%s' % cluster['uuid'])
|
||||
self.assertEqual(cluster.uuid, response['uuid'])
|
||||
self.assertEqual({'fake_name': 'fake_reason'}, response['faults'])
|
||||
|
||||
@mock.patch('magnum.common.clients.OpenStackClients.heat')
|
||||
def test_get_one_failed_cluster_heatclient_exception(self, mock_heat):
|
||||
mock_heat.resources.list.side_effect = Exception('fake')
|
||||
cluster = obj_utils.create_test_cluster(self.context,
|
||||
status='CREATE_FAILED')
|
||||
response = self.get_json('/clusters/%s' % cluster['uuid'])
|
||||
self.assertEqual(cluster.uuid, response['uuid'])
|
||||
self.assertEqual({}, response['faults'])
|
||||
|
||||
def test_get_one_by_name(self):
|
||||
cluster = obj_utils.create_test_cluster(self.context)
|
||||
response = self.get_json('/clusters/%s' % cluster['name'])
|
||||
self.assertEqual(cluster.uuid, response['uuid'])
|
||||
self._verify_attrs(self._expand_cluster_attrs, response)
|
||||
|
||||
def test_get_one_by_name_not_found(self):
|
||||
response = self.get_json(
|
||||
'/clusters/not_found',
|
||||
expect_errors=True)
|
||||
self.assertEqual(404, response.status_int)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_get_one_by_name_multiple_cluster(self):
|
||||
obj_utils.create_test_cluster(self.context, name='test_cluster',
|
||||
uuid=uuidutils.generate_uuid())
|
||||
obj_utils.create_test_cluster(self.context, name='test_cluster',
|
||||
uuid=uuidutils.generate_uuid())
|
||||
response = self.get_json('/clusters/test_cluster',
|
||||
expect_errors=True)
|
||||
self.assertEqual(409, response.status_int)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_get_all_with_pagination_marker(self):
|
||||
cluster_list = []
|
||||
for id_ in range(4):
|
||||
temp_uuid = uuidutils.generate_uuid()
|
||||
cluster = obj_utils.create_test_cluster(self.context, id=id_,
|
||||
uuid=temp_uuid)
|
||||
cluster_list.append(cluster)
|
||||
|
||||
response = self.get_json('/clusters?limit=3&marker=%s'
|
||||
% cluster_list[2].uuid)
|
||||
self.assertEqual(1, len(response['clusters']))
|
||||
self.assertEqual(cluster_list[-1].uuid,
|
||||
response['clusters'][0]['uuid'])
|
||||
|
||||
def test_detail(self):
|
||||
cluster = obj_utils.create_test_cluster(self.context)
|
||||
response = self.get_json('/clusters/detail')
|
||||
self.assertEqual(cluster.uuid, response['clusters'][0]["uuid"])
|
||||
self._verify_attrs(self._expand_cluster_attrs,
|
||||
response['clusters'][0])
|
||||
|
||||
def test_detail_with_pagination_marker(self):
|
||||
cluster_list = []
|
||||
for id_ in range(4):
|
||||
temp_uuid = uuidutils.generate_uuid()
|
||||
cluster = obj_utils.create_test_cluster(self.context, id=id_,
|
||||
uuid=temp_uuid)
|
||||
cluster_list.append(cluster)
|
||||
|
||||
response = self.get_json('/clusters/detail?limit=3&marker=%s'
|
||||
% cluster_list[2].uuid)
|
||||
self.assertEqual(1, len(response['clusters']))
|
||||
self.assertEqual(cluster_list[-1].uuid,
|
||||
response['clusters'][0]['uuid'])
|
||||
self._verify_attrs(self._expand_cluster_attrs,
|
||||
response['clusters'][0])
|
||||
|
||||
def test_detail_against_single(self):
|
||||
cluster = obj_utils.create_test_cluster(self.context)
|
||||
response = self.get_json('/clusters/%s/detail' % cluster['uuid'],
|
||||
expect_errors=True)
|
||||
self.assertEqual(404, response.status_int)
|
||||
|
||||
def test_many(self):
|
||||
bm_list = []
|
||||
for id_ in range(5):
|
||||
temp_uuid = uuidutils.generate_uuid()
|
||||
cluster = obj_utils.create_test_cluster(self.context, id=id_,
|
||||
uuid=temp_uuid)
|
||||
bm_list.append(cluster.uuid)
|
||||
response = self.get_json('/clusters')
|
||||
self.assertEqual(len(bm_list), len(response['clusters']))
|
||||
uuids = [b['uuid'] for b in response['clusters']]
|
||||
self.assertEqual(sorted(bm_list), sorted(uuids))
|
||||
|
||||
def test_links(self):
|
||||
uuid = uuidutils.generate_uuid()
|
||||
obj_utils.create_test_cluster(self.context, id=1, uuid=uuid)
|
||||
response = self.get_json('/clusters/%s' % uuid)
|
||||
self.assertIn('links', response.keys())
|
||||
self.assertEqual(2, len(response['links']))
|
||||
self.assertIn(uuid, response['links'][0]['href'])
|
||||
for l in response['links']:
|
||||
bookmark = l['rel'] == 'bookmark'
|
||||
self.assertTrue(self.validate_link(l['href'],
|
||||
bookmark=bookmark))
|
||||
|
||||
def test_collection_links(self):
|
||||
for id_ in range(5):
|
||||
obj_utils.create_test_cluster(self.context, id=id_,
|
||||
uuid=uuidutils.generate_uuid())
|
||||
response = self.get_json('/clusters/?limit=3')
|
||||
self.assertEqual(3, len(response['clusters']))
|
||||
|
||||
next_marker = response['clusters'][-1]['uuid']
|
||||
self.assertIn(next_marker, response['next'])
|
||||
|
||||
def test_collection_links_default_limit(self):
|
||||
cfg.CONF.set_override('max_limit', 3, 'api')
|
||||
for id_ in range(5):
|
||||
obj_utils.create_test_cluster(self.context, id=id_,
|
||||
uuid=uuidutils.generate_uuid())
|
||||
response = self.get_json('/clusters')
|
||||
self.assertEqual(3, len(response['clusters']))
|
||||
|
||||
next_marker = response['clusters'][-1]['uuid']
|
||||
self.assertIn(next_marker, response['next'])
|
||||
|
||||
|
||||
class TestPatch(api_base.FunctionalTest):
|
||||
def setUp(self):
|
||||
super(TestPatch, self).setUp()
|
||||
self.cluster_template_obj = obj_utils.create_test_cluster_template(
|
||||
self.context)
|
||||
self.cluster_obj = obj_utils.create_test_cluster(
|
||||
self.context, name='cluster_example_A', node_count=3)
|
||||
p = mock.patch.object(rpcapi.API, 'bay_update_async')
|
||||
self.mock_bay_update = p.start()
|
||||
self.mock_bay_update.side_effect = self._simulate_rpc_bay_update
|
||||
self.addCleanup(p.stop)
|
||||
|
||||
def _simulate_rpc_bay_update(self, bay):
|
||||
bay.save()
|
||||
return bay
|
||||
|
||||
@mock.patch('oslo_utils.timeutils.utcnow')
|
||||
def test_replace_ok(self, mock_utcnow):
|
||||
new_node_count = 4
|
||||
test_time = datetime.datetime(2000, 1, 1, 0, 0)
|
||||
mock_utcnow.return_value = test_time
|
||||
|
||||
response = self.patch_json('/clusters/%s' % self.cluster_obj.uuid,
|
||||
[{'path': '/node_count',
|
||||
'value': new_node_count,
|
||||
'op': 'replace'}])
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(202, response.status_code)
|
||||
|
||||
response = self.get_json('/clusters/%s' % self.cluster_obj.uuid)
|
||||
self.assertEqual(new_node_count, response['node_count'])
|
||||
return_updated_at = timeutils.parse_isotime(
|
||||
response['updated_at']).replace(tzinfo=None)
|
||||
self.assertEqual(test_time, return_updated_at)
|
||||
# Assert nothing else was changed
|
||||
self.assertEqual(self.cluster_obj.uuid, response['uuid'])
|
||||
self.assertEqual(self.cluster_obj.baymodel_id,
|
||||
response['cluster_template_id'])
|
||||
|
||||
@mock.patch('oslo_utils.timeutils.utcnow')
|
||||
def test_replace_ok_by_name(self, mock_utcnow):
|
||||
new_node_count = 4
|
||||
test_time = datetime.datetime(2000, 1, 1, 0, 0)
|
||||
mock_utcnow.return_value = test_time
|
||||
|
||||
response = self.patch_json('/clusters/%s' % self.cluster_obj.name,
|
||||
[{'path': '/node_count',
|
||||
'value': new_node_count,
|
||||
'op': 'replace'}])
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(202, response.status_code)
|
||||
|
||||
response = self.get_json('/clusters/%s' % self.cluster_obj.uuid)
|
||||
self.assertEqual(new_node_count, response['node_count'])
|
||||
return_updated_at = timeutils.parse_isotime(
|
||||
response['updated_at']).replace(tzinfo=None)
|
||||
self.assertEqual(test_time, return_updated_at)
|
||||
# Assert nothing else was changed
|
||||
self.assertEqual(self.cluster_obj.uuid, response['uuid'])
|
||||
self.assertEqual(self.cluster_obj.baymodel_id,
|
||||
response['cluster_template_id'])
|
||||
|
||||
@mock.patch('oslo_utils.timeutils.utcnow')
|
||||
def test_replace_ok_by_name_not_found(self, mock_utcnow):
|
||||
name = 'not_found'
|
||||
test_time = datetime.datetime(2000, 1, 1, 0, 0)
|
||||
mock_utcnow.return_value = test_time
|
||||
|
||||
response = self.patch_json('/clusters/%s' % name,
|
||||
[{'path': '/name', 'value': name,
|
||||
'op': 'replace'}],
|
||||
expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(404, response.status_code)
|
||||
|
||||
def test_replace_cluster_template_id_failed(self):
|
||||
cluster_template = obj_utils.create_test_cluster_template(
|
||||
self.context,
|
||||
uuid=uuidutils.generate_uuid())
|
||||
response = self.patch_json('/clusters/%s' % self.cluster_obj.uuid,
|
||||
[{'path': '/cluster_template_id',
|
||||
'value': cluster_template.uuid,
|
||||
'op': 'replace'}],
|
||||
expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(400, response.status_code)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
@mock.patch('oslo_utils.timeutils.utcnow')
|
||||
def test_replace_ok_by_name_multiple_cluster(self, mock_utcnow):
|
||||
test_time = datetime.datetime(2000, 1, 1, 0, 0)
|
||||
mock_utcnow.return_value = test_time
|
||||
|
||||
obj_utils.create_test_cluster(self.context, name='test_cluster',
|
||||
uuid=uuidutils.generate_uuid())
|
||||
obj_utils.create_test_cluster(self.context, name='test_cluster',
|
||||
uuid=uuidutils.generate_uuid())
|
||||
|
||||
response = self.patch_json('/clusters/test_cluster',
|
||||
[{'path': '/name',
|
||||
'value': 'test_cluster',
|
||||
'op': 'replace'}],
|
||||
expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(409, response.status_code)
|
||||
|
||||
def test_replace_non_existent_cluster_template_id(self):
|
||||
response = self.patch_json('/clusters/%s' % self.cluster_obj.uuid,
|
||||
[{'path': '/cluster_template_id',
|
||||
'value': uuidutils.generate_uuid(),
|
||||
'op': 'replace'}],
|
||||
expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(400, response.status_code)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_replace_invalid_node_count(self):
|
||||
response = self.patch_json('/clusters/%s' % self.cluster_obj.uuid,
|
||||
[{'path': '/node_count', 'value': -1,
|
||||
'op': 'replace'}],
|
||||
expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(400, response.status_code)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_replace_non_existent_cluster(self):
|
||||
response = self.patch_json('/clusters/%s' %
|
||||
uuidutils.generate_uuid(),
|
||||
[{'path': '/name',
|
||||
'value': 'cluster_example_B',
|
||||
'op': 'replace'}],
|
||||
expect_errors=True)
|
||||
self.assertEqual(404, response.status_int)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_replace_cluster_name_failed(self):
|
||||
response = self.patch_json('/clusters/%s' % self.cluster_obj.uuid,
|
||||
[{'path': '/name',
|
||||
'value': 'cluster_example_B',
|
||||
'op': 'replace'}],
|
||||
expect_errors=True)
|
||||
self.assertEqual(400, response.status_int)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_add_non_existent_property(self):
|
||||
response = self.patch_json(
|
||||
'/clusters/%s' % self.cluster_obj.uuid,
|
||||
[{'path': '/foo', 'value': 'bar', 'op': 'add'}],
|
||||
expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(400, response.status_int)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_remove_ok(self):
|
||||
response = self.get_json('/clusters/%s' % self.cluster_obj.uuid)
|
||||
self.assertIsNotNone(response['name'])
|
||||
|
||||
response = self.patch_json('/clusters/%s' % self.cluster_obj.uuid,
|
||||
[{'path': '/node_count',
|
||||
'op': 'remove'}])
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(202, response.status_code)
|
||||
|
||||
response = self.get_json('/clusters/%s' % self.cluster_obj.uuid)
|
||||
# only allow node_count for cluster, and default value is 1
|
||||
self.assertEqual(1, response['node_count'])
|
||||
# Assert nothing else was changed
|
||||
self.assertEqual(self.cluster_obj.uuid, response['uuid'])
|
||||
self.assertEqual(self.cluster_obj.baymodel_id,
|
||||
response['cluster_template_id'])
|
||||
self.assertEqual(self.cluster_obj.name, response['name'])
|
||||
self.assertEqual(self.cluster_obj.master_count,
|
||||
response['master_count'])
|
||||
|
||||
def test_remove_uuid(self):
|
||||
response = self.patch_json('/clusters/%s' % self.cluster_obj.uuid,
|
||||
[{'path': '/uuid', 'op': 'remove'}],
|
||||
expect_errors=True)
|
||||
self.assertEqual(400, response.status_int)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_remove_cluster_template_id(self):
|
||||
response = self.patch_json('/clusters/%s' % self.cluster_obj.uuid,
|
||||
[{'path': '/cluster_template_id',
|
||||
'op': 'remove'}],
|
||||
expect_errors=True)
|
||||
self.assertEqual(400, response.status_int)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_remove_non_existent_property(self):
|
||||
response = self.patch_json(
|
||||
'/clusters/%s' % self.cluster_obj.uuid,
|
||||
[{'path': '/non-existent', 'op': 'remove'}],
|
||||
expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(400, response.status_code)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
|
||||
class TestPost(api_base.FunctionalTest):
|
||||
def setUp(self):
|
||||
super(TestPost, self).setUp()
|
||||
self.cluster_template = obj_utils.create_test_cluster_template(
|
||||
self.context)
|
||||
p = mock.patch.object(rpcapi.API, 'bay_create_async')
|
||||
self.mock_bay_create = p.start()
|
||||
self.mock_bay_create.side_effect = self._simulate_rpc_bay_create
|
||||
self.addCleanup(p.stop)
|
||||
p = mock.patch.object(attr_validator, 'validate_os_resources')
|
||||
self.mock_valid_os_res = p.start()
|
||||
self.addCleanup(p.stop)
|
||||
|
||||
def _simulate_rpc_bay_create(self, bay, bay_create_timeout):
|
||||
bay.create()
|
||||
return bay
|
||||
|
||||
@mock.patch('oslo_utils.timeutils.utcnow')
|
||||
def test_create_cluster(self, mock_utcnow):
|
||||
bdict = apiutils.cluster_post_data()
|
||||
test_time = datetime.datetime(2000, 1, 1, 0, 0)
|
||||
mock_utcnow.return_value = test_time
|
||||
|
||||
response = self.post_json('/clusters', bdict)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(202, response.status_int)
|
||||
self.assertTrue(uuidutils.is_uuid_like(response.json['uuid']))
|
||||
|
||||
def test_create_cluster_set_project_id_and_user_id(self):
|
||||
bdict = apiutils.cluster_post_data()
|
||||
|
||||
def _simulate_rpc_bay_create(bay, bay_create_timeout):
|
||||
self.assertEqual(self.context.project_id, bay.project_id)
|
||||
self.assertEqual(self.context.user_id, bay.user_id)
|
||||
bay.create()
|
||||
return bay
|
||||
|
||||
self.mock_bay_create.side_effect = _simulate_rpc_bay_create
|
||||
|
||||
self.post_json('/clusters', bdict)
|
||||
|
||||
def test_create_cluster_doesnt_contain_id(self):
|
||||
with mock.patch.object(self.dbapi, 'create_bay',
|
||||
wraps=self.dbapi.create_bay) as cc_mock:
|
||||
bdict = apiutils.cluster_post_data(name='cluster_example_A')
|
||||
response = self.post_json('/clusters', bdict)
|
||||
cc_mock.assert_called_once_with(mock.ANY)
|
||||
# Check that 'id' is not in first arg of positional args
|
||||
self.assertNotIn('id', cc_mock.call_args[0][0])
|
||||
self.assertTrue(uuidutils.is_uuid_like(response.json['uuid']))
|
||||
|
||||
def test_create_cluster_generate_uuid(self):
|
||||
bdict = apiutils.cluster_post_data()
|
||||
del bdict['uuid']
|
||||
|
||||
response = self.post_json('/clusters', bdict)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(202, response.status_int)
|
||||
self.assertTrue(uuidutils.is_uuid_like(response.json['uuid']))
|
||||
|
||||
def test_create_cluster_no_cluster_template_id(self):
|
||||
bdict = apiutils.cluster_post_data()
|
||||
del bdict['cluster_template_id']
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(400, response.status_int)
|
||||
|
||||
def test_create_cluster_with_non_existent_cluster_template_id(self):
|
||||
temp_uuid = uuidutils.generate_uuid()
|
||||
bdict = apiutils.cluster_post_data(cluster_template_id=temp_uuid)
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(400, response.status_int)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_create_cluster_with_cluster_template_name(self):
|
||||
modelname = self.cluster_template.name
|
||||
bdict = apiutils.cluster_post_data(cluster_template_id=modelname)
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(202, response.status_int)
|
||||
|
||||
def test_create_cluster_with_node_count_zero(self):
|
||||
bdict = apiutils.cluster_post_data()
|
||||
bdict['node_count'] = 0
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(400, response.status_int)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_create_cluster_with_node_count_negative(self):
|
||||
bdict = apiutils.cluster_post_data()
|
||||
bdict['node_count'] = -1
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(400, response.status_int)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_create_cluster_with_no_node_count(self):
|
||||
bdict = apiutils.cluster_post_data()
|
||||
del bdict['node_count']
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(202, response.status_int)
|
||||
|
||||
def test_create_cluster_with_master_count_zero(self):
|
||||
bdict = apiutils.cluster_post_data()
|
||||
bdict['master_count'] = 0
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(400, response.status_int)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_create_cluster_with_no_master_count(self):
|
||||
bdict = apiutils.cluster_post_data()
|
||||
del bdict['master_count']
|
||||
response = self.post_json('/clusters', bdict)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(202, response.status_int)
|
||||
|
||||
def test_create_cluster_with_invalid_long_name(self):
|
||||
bdict = apiutils.cluster_post_data(name='x' * 243)
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(400, response.status_int)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_create_cluster_with_invalid_integer_name(self):
|
||||
bdict = apiutils.cluster_post_data(name='123456')
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(400, response.status_int)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_create_cluster_with_invalid_integer_str_name(self):
|
||||
bdict = apiutils.cluster_post_data(name='123456test_cluster')
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(400, response.status_int)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_create_cluster_with_hyphen_invalid_at_start_name(self):
|
||||
bdict = apiutils.cluster_post_data(name='-test_cluster')
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(400, response.status_int)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_create_cluster_with_period_invalid_at_start_name(self):
|
||||
bdict = apiutils.cluster_post_data(name='.test_cluster')
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(400, response.status_int)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_create_cluster_with_underscore_invalid_at_start_name(self):
|
||||
bdict = apiutils.cluster_post_data(name='_test_cluster')
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(400, response.status_int)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_create_cluster_with_valid_str_int_name(self):
|
||||
bdict = apiutils.cluster_post_data(name='test_cluster123456')
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(202, response.status_int)
|
||||
|
||||
def test_create_cluster_with_hyphen_valid_name(self):
|
||||
bdict = apiutils.cluster_post_data(name='test-cluster')
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(202, response.status_int)
|
||||
|
||||
def test_create_cluster_with_period_valid_name(self):
|
||||
bdict = apiutils.cluster_post_data(name='test.cluster')
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(202, response.status_int)
|
||||
|
||||
def test_create_cluster_with_period_at_end_valid_name(self):
|
||||
bdict = apiutils.cluster_post_data(name='testcluster.')
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(202, response.status_int)
|
||||
|
||||
def test_create_cluster_with_hyphen_at_end_valid_name(self):
|
||||
bdict = apiutils.cluster_post_data(name='testcluster-')
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(202, response.status_int)
|
||||
|
||||
def test_create_cluster_with_underscore_at_end_valid_name(self):
|
||||
bdict = apiutils.cluster_post_data(name='testcluster_')
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(202, response.status_int)
|
||||
|
||||
def test_create_cluster_with_mix_special_char_valid_name(self):
|
||||
bdict = apiutils.cluster_post_data(name='test.-_cluster')
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(202, response.status_int)
|
||||
|
||||
def test_create_cluster_with_capital_letter_start_valid_name(self):
|
||||
bdict = apiutils.cluster_post_data(name='Testcluster')
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(202, response.status_int)
|
||||
|
||||
def test_create_cluster_with_invalid_empty_name(self):
|
||||
bdict = apiutils.cluster_post_data(name='')
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(400, response.status_int)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_create_cluster_without_name(self):
|
||||
bdict = apiutils.cluster_post_data()
|
||||
del bdict['name']
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(202, response.status_int)
|
||||
|
||||
def test_create_cluster_with_timeout_none(self):
|
||||
bdict = apiutils.cluster_post_data()
|
||||
bdict['create_timeout'] = None
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(202, response.status_int)
|
||||
|
||||
def test_create_cluster_with_no_timeout(self):
|
||||
def _simulate_rpc_bay_create(bay, bay_create_timeout):
|
||||
self.assertEqual(60, bay_create_timeout)
|
||||
bay.create()
|
||||
return bay
|
||||
|
||||
self.mock_bay_create.side_effect = _simulate_rpc_bay_create
|
||||
bdict = apiutils.cluster_post_data()
|
||||
del bdict['create_timeout']
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(202, response.status_int)
|
||||
|
||||
def test_create_cluster_with_timeout_negative(self):
|
||||
bdict = apiutils.cluster_post_data()
|
||||
bdict['create_timeout'] = -1
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(400, response.status_int)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_create_cluster_with_timeout_zero(self):
|
||||
bdict = apiutils.cluster_post_data()
|
||||
bdict['create_timeout'] = 0
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(202, response.status_int)
|
||||
|
||||
def test_create_cluster_with_invalid_flavor(self):
|
||||
bdict = apiutils.cluster_post_data()
|
||||
self.mock_valid_os_res.side_effect = exception.FlavorNotFound(
|
||||
'test-flavor')
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertTrue(self.mock_valid_os_res.called)
|
||||
self.assertEqual(400, response.status_int)
|
||||
|
||||
def test_create_cluster_with_invalid_ext_network(self):
|
||||
bdict = apiutils.cluster_post_data()
|
||||
self.mock_valid_os_res.side_effect = \
|
||||
exception.ExternalNetworkNotFound('test-net')
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertTrue(self.mock_valid_os_res.called)
|
||||
self.assertEqual(400, response.status_int)
|
||||
|
||||
def test_create_cluster_with_invalid_keypair(self):
|
||||
bdict = apiutils.cluster_post_data()
|
||||
self.mock_valid_os_res.side_effect = exception.KeyPairNotFound(
|
||||
'test-key')
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertTrue(self.mock_valid_os_res.called)
|
||||
self.assertEqual(404, response.status_int)
|
||||
|
||||
def test_create_cluster_with_nonexist_image(self):
|
||||
bdict = apiutils.cluster_post_data()
|
||||
self.mock_valid_os_res.side_effect = exception.ImageNotFound(
|
||||
'test-img')
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertTrue(self.mock_valid_os_res.called)
|
||||
self.assertEqual(400, response.status_int)
|
||||
|
||||
def test_create_cluster_with_multi_images_same_name(self):
|
||||
bdict = apiutils.cluster_post_data()
|
||||
self.mock_valid_os_res.side_effect = exception.Conflict('test-img')
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertTrue(self.mock_valid_os_res.called)
|
||||
self.assertEqual(409, response.status_int)
|
||||
|
||||
def test_create_cluster_with_on_os_distro_image(self):
|
||||
bdict = apiutils.cluster_post_data()
|
||||
self.mock_valid_os_res.side_effect = \
|
||||
exception.OSDistroFieldNotFound('img')
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertTrue(self.mock_valid_os_res.called)
|
||||
self.assertEqual(400, response.status_int)
|
||||
|
||||
def test_create_cluster_with_no_lb_one_node(self):
|
||||
cluster_template = obj_utils.create_test_cluster_template(
|
||||
self.context, name='foo', uuid='foo', master_lb_enabled=False)
|
||||
bdict = apiutils.cluster_post_data(
|
||||
cluster_template_id=cluster_template.name, master_count=1)
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(202, response.status_int)
|
||||
|
||||
def test_create_cluster_with_no_lb_multi_node(self):
|
||||
cluster_template = obj_utils.create_test_cluster_template(
|
||||
self.context, name='foo', uuid='foo', master_lb_enabled=False)
|
||||
bdict = apiutils.cluster_post_data(
|
||||
cluster_template_id=cluster_template.name, master_count=3)
|
||||
response = self.post_json('/clusters', bdict, expect_errors=True)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertEqual(400, response.status_int)
|
||||
|
||||
|
||||
class TestDelete(api_base.FunctionalTest):
|
||||
def setUp(self):
|
||||
super(TestDelete, self).setUp()
|
||||
self.cluster_template = obj_utils.create_test_cluster_template(
|
||||
self.context)
|
||||
self.cluster = obj_utils.create_test_cluster(self.context)
|
||||
p = mock.patch.object(rpcapi.API, 'bay_delete_async')
|
||||
self.mock_bay_delete = p.start()
|
||||
self.mock_bay_delete.side_effect = self._simulate_rpc_bay_delete
|
||||
self.addCleanup(p.stop)
|
||||
|
||||
def _simulate_rpc_bay_delete(self, bay_uuid):
|
||||
bay = objects.Bay.get_by_uuid(self.context, bay_uuid)
|
||||
bay.destroy()
|
||||
|
||||
def test_delete_cluster(self):
|
||||
self.delete('/clusters/%s' % self.cluster.uuid)
|
||||
response = self.get_json('/clusters/%s' % self.cluster.uuid,
|
||||
expect_errors=True)
|
||||
self.assertEqual(404, response.status_int)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_delete_cluster_not_found(self):
|
||||
uuid = uuidutils.generate_uuid()
|
||||
response = self.delete('/clusters/%s' % uuid, expect_errors=True)
|
||||
self.assertEqual(404, response.status_int)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_delete_cluster_with_name_not_found(self):
|
||||
response = self.delete('/clusters/not_found', expect_errors=True)
|
||||
self.assertEqual(404, response.status_int)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
def test_delete_cluster_with_name(self):
|
||||
response = self.delete('/clusters/%s' % self.cluster.name,
|
||||
expect_errors=True)
|
||||
self.assertEqual(204, response.status_int)
|
||||
|
||||
def test_delete_multiple_cluster_by_name(self):
|
||||
obj_utils.create_test_cluster(self.context, name='test_cluster',
|
||||
uuid=uuidutils.generate_uuid())
|
||||
obj_utils.create_test_cluster(self.context, name='test_cluster',
|
||||
uuid=uuidutils.generate_uuid())
|
||||
response = self.delete('/clusters/test_cluster', expect_errors=True)
|
||||
self.assertEqual(409, response.status_int)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertTrue(response.json['errors'])
|
||||
|
||||
|
||||
class TestClusterPolicyEnforcement(api_base.FunctionalTest):
|
||||
def setUp(self):
|
||||
super(TestClusterPolicyEnforcement, self).setUp()
|
||||
obj_utils.create_test_cluster_template(self.context)
|
||||
|
||||
def _common_policy_check(self, rule, func, *arg, **kwarg):
|
||||
self.policy.set_rules({rule: "project:non_fake"})
|
||||
response = func(*arg, **kwarg)
|
||||
self.assertEqual(403, response.status_int)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertTrue(
|
||||
"Policy doesn't allow %s to be performed." % rule,
|
||||
response.json['errors'][0]['detail'])
|
||||
|
||||
def test_policy_disallow_get_all(self):
|
||||
self._common_policy_check(
|
||||
"cluster:get_all", self.get_json, '/clusters', expect_errors=True)
|
||||
|
||||
def test_policy_disallow_get_one(self):
|
||||
self.cluster = obj_utils.create_test_cluster(self.context)
|
||||
self._common_policy_check(
|
||||
"cluster:get", self.get_json, '/clusters/%s' % self.cluster.uuid,
|
||||
expect_errors=True)
|
||||
|
||||
def test_policy_disallow_detail(self):
|
||||
self._common_policy_check(
|
||||
"cluster:detail", self.get_json,
|
||||
'/clusters/%s/detail' % uuidutils.generate_uuid(),
|
||||
expect_errors=True)
|
||||
|
||||
def test_policy_disallow_update(self):
|
||||
self.cluster = obj_utils.create_test_cluster(self.context,
|
||||
name='cluster_example_A',
|
||||
node_count=3)
|
||||
self._common_policy_check(
|
||||
"cluster:update", self.patch_json, '/clusters/%s' %
|
||||
self.cluster.name,
|
||||
[{'path': '/name', 'value': "new_name", 'op': 'replace'}],
|
||||
expect_errors=True)
|
||||
|
||||
def test_policy_disallow_create(self):
|
||||
bdict = apiutils.cluster_post_data(name='cluster_example_A')
|
||||
self._common_policy_check(
|
||||
"cluster:create", self.post_json, '/clusters', bdict,
|
||||
expect_errors=True)
|
||||
|
||||
def _simulate_rpc_bay_delete(self, bay_uuid):
|
||||
bay = objects.Bay.get_by_uuid(self.context, bay_uuid)
|
||||
bay.destroy()
|
||||
|
||||
def test_policy_disallow_delete(self):
|
||||
p = mock.patch.object(rpcapi.API, 'bay_delete')
|
||||
self.mock_bay_delete = p.start()
|
||||
self.mock_bay_delete.side_effect = self._simulate_rpc_bay_delete
|
||||
self.addCleanup(p.stop)
|
||||
self.cluster = obj_utils.create_test_cluster(self.context)
|
||||
self._common_policy_check(
|
||||
"cluster:delete", self.delete, '/clusters/%s' %
|
||||
self.cluster.uuid,
|
||||
expect_errors=True)
|
||||
|
||||
def _owner_check(self, rule, func, *args, **kwargs):
|
||||
self.policy.set_rules({rule: "user_id:%(user_id)s"})
|
||||
response = func(*args, **kwargs)
|
||||
self.assertEqual(403, response.status_int)
|
||||
self.assertEqual('application/json', response.content_type)
|
||||
self.assertTrue(
|
||||
"Policy doesn't allow %s to be performed." % rule,
|
||||
response.json['errors'][0]['detail'])
|
||||
|
||||
def test_policy_only_owner_get_one(self):
|
||||
cluster = obj_utils.create_test_cluster(self.context,
|
||||
user_id='another')
|
||||
self._owner_check("cluster:get", self.get_json,
|
||||
'/clusters/%s' % cluster.uuid,
|
||||
expect_errors=True)
|
||||
|
||||
def test_policy_only_owner_update(self):
|
||||
cluster = obj_utils.create_test_cluster(self.context,
|
||||
user_id='another')
|
||||
self._owner_check(
|
||||
"cluster:update", self.patch_json,
|
||||
'/clusters/%s' % cluster.uuid,
|
||||
[{'path': '/name', 'value': "new_name", 'op': 'replace'}],
|
||||
expect_errors=True)
|
||||
|
||||
def test_policy_only_owner_delete(self):
|
||||
cluster = obj_utils.create_test_cluster(self.context,
|
||||
user_id='another')
|
||||
self._owner_check("cluster:delete", self.delete,
|
||||
'/clusters/%s' % cluster.uuid,
|
||||
expect_errors=True)
|
File diff suppressed because it is too large
Load Diff
|
@ -18,6 +18,8 @@ import pytz
|
|||
|
||||
from magnum.api.controllers.v1 import bay as bay_controller
|
||||
from magnum.api.controllers.v1 import baymodel as baymodel_controller
|
||||
from magnum.api.controllers.v1 import cluster as cluster_controller
|
||||
from magnum.api.controllers.v1 import cluster_template as cluster_tmp_ctrl
|
||||
from magnum.tests.unit.db import utils
|
||||
|
||||
|
||||
|
@ -33,6 +35,12 @@ def baymodel_post_data(**kw):
|
|||
return remove_internal(baymodel, internal)
|
||||
|
||||
|
||||
def cluster_template_post_data(**kw):
|
||||
cluster_template = utils.get_test_baymodel(**kw)
|
||||
internal = cluster_tmp_ctrl.ClusterTemplatePatchType.internal_attrs()
|
||||
return remove_internal(cluster_template, internal)
|
||||
|
||||
|
||||
def bay_post_data(**kw):
|
||||
bay = utils.get_test_bay(**kw)
|
||||
bay['bay_create_timeout'] = kw.get('bay_create_timeout', 15)
|
||||
|
@ -40,6 +48,17 @@ def bay_post_data(**kw):
|
|||
return remove_internal(bay, internal)
|
||||
|
||||
|
||||
def cluster_post_data(**kw):
|
||||
cluster = utils.get_test_bay(**kw)
|
||||
cluster['create_timeout'] = kw.get('create_timeout', 15)
|
||||
cluster['cluster_template_id'] = kw.get('cluster_template_id',
|
||||
cluster['baymodel_id'])
|
||||
del cluster['bay_create_timeout']
|
||||
del cluster['baymodel_id']
|
||||
internal = cluster_controller.ClusterPatchType.internal_attrs()
|
||||
return remove_internal(cluster, internal)
|
||||
|
||||
|
||||
def cert_post_data(**kw):
|
||||
return {
|
||||
'bay_uuid': kw.get('bay_uuid', '5d12f6fd-a196-4bf0-ae4c-1f639a523a52'),
|
||||
|
|
|
@ -456,7 +456,7 @@ class TestHandler(db_base.DbTestCase):
|
|||
self.assertEqual(1,
|
||||
cert_manager.delete_certificates_from_bay.call_count)
|
||||
# The bay has been destroyed
|
||||
self.assertRaises(exception.BayNotFound,
|
||||
self.assertRaises(exception.ClusterNotFound,
|
||||
objects.Bay.get, self.context, self.bay.uuid)
|
||||
|
||||
@patch('magnum.conductor.handlers.bay_conductor.cert_manager')
|
||||
|
|
|
@ -34,7 +34,7 @@ class DbBayTestCase(base.DbTestCase):
|
|||
|
||||
def test_create_bay_already_exists(self):
|
||||
utils.create_test_bay()
|
||||
self.assertRaises(exception.BayAlreadyExists,
|
||||
self.assertRaises(exception.ClusterAlreadyExists,
|
||||
utils.create_test_bay)
|
||||
|
||||
def test_get_bay_by_id(self):
|
||||
|
@ -56,10 +56,10 @@ class DbBayTestCase(base.DbTestCase):
|
|||
self.assertEqual(bay.uuid, res.uuid)
|
||||
|
||||
def test_get_bay_that_does_not_exist(self):
|
||||
self.assertRaises(exception.BayNotFound,
|
||||
self.assertRaises(exception.ClusterNotFound,
|
||||
self.dbapi.get_bay_by_id,
|
||||
self.context, 999)
|
||||
self.assertRaises(exception.BayNotFound,
|
||||
self.assertRaises(exception.ClusterNotFound,
|
||||
self.dbapi.get_bay_by_uuid,
|
||||
self.context,
|
||||
'12345678-9999-0000-aaaa-123456789012')
|
||||
|
@ -174,7 +174,7 @@ class DbBayTestCase(base.DbTestCase):
|
|||
self.assertIsNotNone(self.dbapi.get_bay_by_id(self.context,
|
||||
bay.id))
|
||||
self.dbapi.destroy_bay(bay.id)
|
||||
self.assertRaises(exception.BayNotFound,
|
||||
self.assertRaises(exception.ClusterNotFound,
|
||||
self.dbapi.get_bay_by_id,
|
||||
self.context, bay.id)
|
||||
|
||||
|
@ -183,12 +183,12 @@ class DbBayTestCase(base.DbTestCase):
|
|||
self.assertIsNotNone(self.dbapi.get_bay_by_uuid(self.context,
|
||||
bay.uuid))
|
||||
self.dbapi.destroy_bay(bay.uuid)
|
||||
self.assertRaises(exception.BayNotFound,
|
||||
self.assertRaises(exception.ClusterNotFound,
|
||||
self.dbapi.get_bay_by_uuid, self.context,
|
||||
bay.uuid)
|
||||
|
||||
def test_destroy_bay_that_does_not_exist(self):
|
||||
self.assertRaises(exception.BayNotFound,
|
||||
self.assertRaises(exception.ClusterNotFound,
|
||||
self.dbapi.destroy_bay,
|
||||
'12345678-9999-0000-aaaa-123456789012')
|
||||
|
||||
|
@ -202,7 +202,7 @@ class DbBayTestCase(base.DbTestCase):
|
|||
|
||||
def test_update_bay_not_found(self):
|
||||
bay_uuid = uuidutils.generate_uuid()
|
||||
self.assertRaises(exception.BayNotFound, self.dbapi.update_bay,
|
||||
self.assertRaises(exception.ClusterNotFound, self.dbapi.update_bay,
|
||||
bay_uuid, {'node_count': 5})
|
||||
|
||||
def test_update_bay_uuid(self):
|
||||
|
|
|
@ -100,7 +100,7 @@ class DbBaymodelTestCase(base.DbTestCase):
|
|||
self.assertEqual(bm['id'], baymodel.id)
|
||||
|
||||
def test_get_baymodel_that_does_not_exist(self):
|
||||
self.assertRaises(exception.BayModelNotFound,
|
||||
self.assertRaises(exception.ClusterTemplateNotFound,
|
||||
self.dbapi.get_baymodel_by_id, self.context, 666)
|
||||
|
||||
def test_get_baymodel_by_name(self):
|
||||
|
@ -128,7 +128,7 @@ class DbBaymodelTestCase(base.DbTestCase):
|
|||
self.context, 'bm')
|
||||
|
||||
def test_get_baymodel_by_name_not_found(self):
|
||||
self.assertRaises(exception.BayModelNotFound,
|
||||
self.assertRaises(exception.ClusterTemplateNotFound,
|
||||
self.dbapi.get_baymodel_by_name,
|
||||
self.context, 'not_found')
|
||||
|
||||
|
@ -138,7 +138,7 @@ class DbBaymodelTestCase(base.DbTestCase):
|
|||
self.assertEqual('updated-model', res.name)
|
||||
|
||||
def test_update_baymodel_that_does_not_exist(self):
|
||||
self.assertRaises(exception.BayModelNotFound,
|
||||
self.assertRaises(exception.ClusterTemplateNotFound,
|
||||
self.dbapi.update_baymodel, 666, {'name': ''})
|
||||
|
||||
def test_update_baymodel_uuid(self):
|
||||
|
@ -150,7 +150,7 @@ class DbBaymodelTestCase(base.DbTestCase):
|
|||
def test_destroy_baymodel(self):
|
||||
bm = utils.create_test_baymodel()
|
||||
self.dbapi.destroy_baymodel(bm['id'])
|
||||
self.assertRaises(exception.BayModelNotFound,
|
||||
self.assertRaises(exception.ClusterTemplateNotFound,
|
||||
self.dbapi.get_baymodel_by_id,
|
||||
self.context, bm['id'])
|
||||
|
||||
|
@ -160,23 +160,23 @@ class DbBaymodelTestCase(base.DbTestCase):
|
|||
self.assertIsNotNone(self.dbapi.get_baymodel_by_uuid(self.context,
|
||||
uuid))
|
||||
self.dbapi.destroy_baymodel(uuid)
|
||||
self.assertRaises(exception.BayModelNotFound,
|
||||
self.assertRaises(exception.ClusterTemplateNotFound,
|
||||
self.dbapi.get_baymodel_by_uuid, self.context, uuid)
|
||||
|
||||
def test_destroy_baymodel_that_does_not_exist(self):
|
||||
self.assertRaises(exception.BayModelNotFound,
|
||||
self.assertRaises(exception.ClusterTemplateNotFound,
|
||||
self.dbapi.destroy_baymodel, 666)
|
||||
|
||||
def test_destroy_baymodel_that_referenced_by_bays(self):
|
||||
bm = utils.create_test_baymodel()
|
||||
bay = utils.create_test_bay(baymodel_id=bm['uuid'])
|
||||
self.assertEqual(bm['uuid'], bay.baymodel_id)
|
||||
self.assertRaises(exception.BayModelReferenced,
|
||||
self.assertRaises(exception.ClusterTemplateReferenced,
|
||||
self.dbapi.destroy_baymodel, bm['id'])
|
||||
|
||||
def test_create_baymodel_already_exists(self):
|
||||
uuid = uuidutils.generate_uuid()
|
||||
utils.create_test_baymodel(id=1, uuid=uuid)
|
||||
self.assertRaises(exception.BayModelAlreadyExists,
|
||||
self.assertRaises(exception.ClusterTemplateAlreadyExists,
|
||||
utils.create_test_baymodel,
|
||||
id=2, uuid=uuid)
|
||||
|
|
|
@ -53,7 +53,7 @@ def create_test_baymodel(context, **kw):
|
|||
baymodel = get_test_baymodel(context, **kw)
|
||||
try:
|
||||
baymodel.create()
|
||||
except exception.BayModelAlreadyExists:
|
||||
except exception.ClusterTemplateAlreadyExists:
|
||||
baymodel = objects.BayModel.get(context, baymodel.uuid)
|
||||
return baymodel
|
||||
|
||||
|
@ -87,6 +87,42 @@ def create_test_bay(context, **kw):
|
|||
return bay
|
||||
|
||||
|
||||
def get_test_cluster_template(context, **kw):
|
||||
"""Return a ClusterTemplate object with appropriate attributes.
|
||||
|
||||
NOTE: Object model is the same for ClusterTemplate and
|
||||
BayModel
|
||||
"""
|
||||
return get_test_baymodel(context, **kw)
|
||||
|
||||
|
||||
def create_test_cluster_template(context, **kw):
|
||||
"""Create and return a test ClusterTemplate object.
|
||||
|
||||
NOTE: Object model is the same for ClusterTemplate and
|
||||
BayModel
|
||||
"""
|
||||
return create_test_baymodel(context, **kw)
|
||||
|
||||
|
||||
def get_test_cluster(context, **kw):
|
||||
"""Return a Cluster object with appropriate attributes.
|
||||
|
||||
NOTE: Object model is the same for Cluster and
|
||||
Bay
|
||||
"""
|
||||
return get_test_bay(context, **kw)
|
||||
|
||||
|
||||
def create_test_cluster(context, **kw):
|
||||
"""Create and return a test cluster object.
|
||||
|
||||
NOTE: Object model is the same for Cluster and
|
||||
Bay
|
||||
"""
|
||||
return create_test_bay(context, **kw)
|
||||
|
||||
|
||||
def get_test_x509keypair(context, **kw):
|
||||
"""Return a X509KeyPair object with appropriate attributes.
|
||||
|
||||
|
|
Loading…
Reference in New Issue