da974e0208
Infbc435ce16
, we stopped passing 'parameters' to stack update call. This is backward incompatible as the 'parameters' set for the stack before an upgrade would take precedence over the ones set in 'parameter_defaults' of plan env. Change-Id: Ia1d823386da687154c433136bf22ae6a0291573d Closes-Bug: #1749636 (cherry picked from commit7213aa5669
)
153 lines
6.1 KiB
Python
153 lines
6.1 KiB
Python
# Copyright 2016 Red Hat, Inc.
|
|
# All Rights Reserved.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
import collections
|
|
import logging
|
|
|
|
from mistral.workflow import utils as mistral_workflow_utils
|
|
|
|
|
|
from tripleo_common.actions import parameters as parameters_actions
|
|
from tripleo_common.actions import templates
|
|
from tripleo_common import constants
|
|
from tripleo_common import update
|
|
|
|
LOG = logging.getLogger(__name__)
|
|
|
|
|
|
def get_group_resources_after_delete(groupname, res_to_delete, resources):
|
|
group = next(res for res in resources if
|
|
res.resource_name == groupname and
|
|
res.resource_type == constants.RESOURCE_GROUP_TYPE)
|
|
members = []
|
|
for res in resources:
|
|
stack_name, stack_id = next(
|
|
x['href'] for x in res.links if
|
|
x['rel'] == 'stack').rsplit('/', 2)[1:]
|
|
# desired new count of nodes after delete operation should be
|
|
# count of all existing nodes in ResourceGroup which are not
|
|
# in set of nodes being deleted. Also nodes in any delete state
|
|
# from a previous failed update operation are not included in
|
|
# overall count (if such nodes exist)
|
|
if (stack_id == group.physical_resource_id and
|
|
res not in res_to_delete and
|
|
not res.resource_status.startswith('DELETE')):
|
|
|
|
members.append(res)
|
|
|
|
return members
|
|
|
|
|
|
class ScaleDownAction(templates.ProcessTemplatesAction):
|
|
"""Deletes overcloud nodes
|
|
|
|
Before calling this method, ensure you have updated the plan
|
|
with any templates or environment files as needed.
|
|
"""
|
|
|
|
def __init__(self, timeout, nodes=[],
|
|
container=constants.DEFAULT_CONTAINER_NAME):
|
|
self.nodes = nodes
|
|
self.timeout_mins = timeout
|
|
super(ScaleDownAction, self).__init__(container)
|
|
|
|
def _update_stack(self, parameters={},
|
|
timeout_mins=constants.STACK_TIMEOUT_DEFAULT):
|
|
# TODO(rbrady): migrate _update_stack to it's own action and update
|
|
# the workflow for scale down
|
|
|
|
# update the plan parameters with the scaled down parameters
|
|
update_params_action = parameters_actions.UpdateParametersAction(
|
|
parameters, self.container)
|
|
updated_plan = update_params_action.run()
|
|
if isinstance(updated_plan, mistral_workflow_utils.Result):
|
|
return updated_plan
|
|
|
|
processed_data = super(ScaleDownAction, self).run()
|
|
if isinstance(processed_data, mistral_workflow_utils.Result):
|
|
return processed_data
|
|
|
|
update.add_breakpoints_cleanup_into_env(processed_data['environment'])
|
|
|
|
fields = processed_data.copy()
|
|
fields['timeout_mins'] = timeout_mins
|
|
fields['existing'] = True
|
|
# As we do a PATCH update when deleting nodes, parameters set for a
|
|
# stack before upgrade to newton (ex. ComputeRemovalPolicies),
|
|
# would still take precedence over the ones set in parameter_defaults
|
|
# after upgrade. Clear these parameters for backward compatibility.
|
|
fields['clear_parameters'] = list(parameters.keys())
|
|
|
|
LOG.debug('stack update params: %s', fields)
|
|
self.get_orchestration_client().stacks.update(self.container,
|
|
**fields)
|
|
|
|
def _get_removal_params_from_heat(self, resources_by_role, resources):
|
|
stack_params = {}
|
|
for role, role_resources in resources_by_role.items():
|
|
param_name = "{0}Count".format(role)
|
|
|
|
# get real count of nodes for each role. *Count stack parameters
|
|
# can not be used because stack parameters return parameters
|
|
# passed by user no matter if previous update operation succeeded
|
|
# or not
|
|
group_members = get_group_resources_after_delete(
|
|
role, role_resources, resources)
|
|
stack_params[param_name] = str(len(group_members))
|
|
|
|
# add instance resource names into removal_policies
|
|
# so heat knows which instances should be removed
|
|
removal_param = "{0}RemovalPolicies".format(role)
|
|
stack_params[removal_param] = [{
|
|
'resource_list': [r.resource_name for r in role_resources]
|
|
}]
|
|
|
|
return stack_params
|
|
|
|
def run(self):
|
|
heatclient = self.get_orchestration_client()
|
|
resources = heatclient.resources.list(self.container, nested_depth=5)
|
|
resources_by_role = collections.defaultdict(list)
|
|
instance_list = list(self.nodes)
|
|
for res in resources:
|
|
try:
|
|
instance_list.remove(res.physical_resource_id)
|
|
except ValueError:
|
|
continue
|
|
|
|
stack_name, stack_id = next(
|
|
x['href'] for x in res.links if
|
|
x['rel'] == 'stack').rsplit('/', 2)[1:]
|
|
# get resource to remove from resource group (it's parent resource
|
|
# of nova server)
|
|
role_resource = next(x for x in resources if
|
|
x.physical_resource_id == stack_id)
|
|
# get the role name which is parent resource name in Heat
|
|
role = role_resource.parent_resource
|
|
resources_by_role[role].append(role_resource)
|
|
|
|
resources_by_role = dict(resources_by_role)
|
|
|
|
if instance_list:
|
|
raise ValueError(
|
|
"Couldn't find following instances in stack %s: %s" %
|
|
(self.container, ','.join(instance_list)))
|
|
|
|
# decrease count for each role (or resource group) and set removal
|
|
# policy for each resource group
|
|
stack_params = self._get_removal_params_from_heat(
|
|
resources_by_role, resources)
|
|
|
|
self._update_stack(parameters=stack_params)
|