a6a6b84258
This change adds the capability to rename the subcloud after bootstrap or during subcloud rehome operation. Added a field in the database to separate the region name from the subcloud name. The region name determines the subcloud reference in the Openstack core, through which it is possible to access the endpoints of a given subcloud. Since the region name cannot be changed, this commit adds the ability to maintain a unique region name based on the UUID format, and allows subcloud renaming when necessary without any endpoint impact. The region is randomly generated to configure the subcloud when it is created and only applies to future subclouds. For those systems that have existing subclouds, the region will be the same as on day 0, that is, region will keep the same name as the subcloud, but subclouds can be renamed. This topic involves changes to dcmanager, dcmanager-client and GUI. To ensure the region name reference needed by the cert-monitor, a mechanism to determine if the request is coming from the cert-monitor has been created. Usage for subcloud rename: dcmanager subcloud update <subcloud-name> --name <new-name> Usage for subcloud rehoming: dcmanager subcloud add --name <subcloud-name> --migrate ... Note: Upgrade test from StarlingX 8 -> 9 for this commit is deferred until upgrade functionality in master is restored. Any issue found during upgrade test will be addressed in a separate commit Test Plan: PASS: Run dcmanager subcloud passing subcommands: - add/delete/migrate/list/show/show --detail - errors/manage/unmanage/reinstall/reconfig - update/deploy PASS: Run dcmanager subcloud add supplying --name parameter and validate the operation is not allowed PASS: Run dcmanager supplying subcommands: - kube/patch/prestage strategies PASS: Run dcmanager to apply patch and remove it PASS: Run dcmanager subcloud-backup: - create/delete/restore/show/upload PASS: Run subcloud-group: - add/delete/list/list-subclouds/show/update PASS: Run dcmanager subcloud strategy for: - patch/kubernetes/firmware PASS: Run dcmanager subcloud update command passing --name parameter supplying the following values: - current subcloud name (not changed) - different existing subcloud name PASS: Run dcmanager to migrate a subcloud passing --name parameter supplying a new subcloud name PASS: Run dcmanager to migrate a subcloud without --name parameter PASS: Run dcmanager to migrate a subcloud passing --name parameter supplying a new subcloud name and different subcloud name in bootstrap file PASS: Test dcmanager API response using cURL command line to validate new region name field PASS: Run full DC sanity and regression Story: 2010788 Task: 48217 Signed-off-by: Cristian Mondo <cristian.mondo@windriver.com> Change-Id: Id04f42504b8e325d9ec3880c240fe4a06e3a20b7
201 lines
7.9 KiB
Python
201 lines
7.9 KiB
Python
#
|
|
# Copyright (c) 2020-2022 Wind River Systems, Inc.
|
|
#
|
|
# SPDX-License-Identifier: Apache-2.0
|
|
#
|
|
import mock
|
|
from oslo_utils import uuidutils
|
|
|
|
from dccommon import consts as dccommon_consts
|
|
from dccommon.drivers.openstack import vim
|
|
from dcmanager.common import consts
|
|
from dcmanager.common import exceptions as exception
|
|
from dcmanager.common import scheduler
|
|
from dcmanager.db.sqlalchemy import api as db_api
|
|
from dcmanager.orchestrator.orch_thread import OrchThread
|
|
|
|
from dcmanager.tests.unit.common import fake_strategy
|
|
from dcmanager.tests.unit.fakes import FakeVimClient
|
|
from dcmanager.tests.unit.fakes import FakeVimStrategy
|
|
from dcmanager.tests.unit.orchestrator.test_base import TestSwUpdate
|
|
|
|
|
|
# rather than invoke a thread, we invoke the function immediately
|
|
def non_threaded_start(some_function, some_arguments):
|
|
some_function(some_arguments)
|
|
|
|
|
|
class TestFwOrchThread(TestSwUpdate):
|
|
@staticmethod
|
|
def create_subcloud(ctxt, name, group_id):
|
|
values = {
|
|
"name": name,
|
|
"description": "subcloud1 description",
|
|
"location": "subcloud1 location",
|
|
'software_version': "18.03",
|
|
"management_subnet": "192.168.101.0/24",
|
|
"management_gateway_ip": "192.168.101.1",
|
|
"management_start_ip": "192.168.101.3",
|
|
"management_end_ip": "192.168.101.4",
|
|
"systemcontroller_gateway_ip": "192.168.204.101",
|
|
'deploy_status': "not-deployed",
|
|
'error_description': 'No errors present',
|
|
'region_name': uuidutils.generate_uuid().replace("-", ""),
|
|
'openstack_installed': False,
|
|
'group_id': group_id,
|
|
'data_install': 'data from install',
|
|
}
|
|
subcloud = db_api.subcloud_create(ctxt, **values)
|
|
state = dccommon_consts.MANAGEMENT_MANAGED
|
|
subcloud = db_api.subcloud_update(ctxt, subcloud.id,
|
|
management_state=state)
|
|
return subcloud
|
|
|
|
# Setting DEFAULT_STRATEGY_TYPE to firmware will setup the firmware
|
|
# orchestration worker, and will mock away the other orch threads
|
|
DEFAULT_STRATEGY_TYPE = consts.SW_UPDATE_TYPE_FIRMWARE
|
|
|
|
def setUp(self):
|
|
super(TestFwOrchThread, self).setUp()
|
|
|
|
# Mock the vim client defined in the base state class
|
|
self.vim_client = FakeVimClient()
|
|
p = mock.patch.object(OrchThread, 'get_vim_client')
|
|
self.mock_vim_client = p.start()
|
|
self.mock_vim_client.return_value = self.vim_client
|
|
self.addCleanup(p.stop)
|
|
|
|
self.vim_client.create_strategy = mock.MagicMock()
|
|
self.vim_client.delete_strategy = mock.MagicMock()
|
|
self.vim_client.get_strategy = mock.MagicMock()
|
|
|
|
def setup_strategy(self, state):
|
|
return fake_strategy.create_fake_strategy(
|
|
self.ctx,
|
|
consts.SW_UPDATE_TYPE_FIRMWARE,
|
|
max_parallel_subclouds=2,
|
|
state=state)
|
|
|
|
def test_delete_strategy_no_steps(self):
|
|
# The 'strategy'should be 'deleting'
|
|
self.strategy = self.setup_strategy(
|
|
state=consts.SW_UPDATE_STATE_DELETING)
|
|
|
|
# invoke the strategy (not strategy step) operation on the orch thread
|
|
self.worker.delete(self.strategy)
|
|
|
|
# There are no strategy steps, so no vim api calls should be invoked
|
|
self.vim_client.get_strategy.assert_not_called()
|
|
|
|
# Verify the strategy was deleted
|
|
self.assertRaises(exception.NotFound,
|
|
db_api.sw_update_strategy_get,
|
|
self.ctx,
|
|
consts.SW_UPDATE_TYPE_FIRMWARE)
|
|
|
|
@mock.patch.object(scheduler.ThreadGroupManager, 'start')
|
|
@mock.patch.object(OrchThread, 'perform_state_action')
|
|
def test_apply_strategy(self, mock_perform_state_action,
|
|
mock_start):
|
|
mock_start.side_effect = non_threaded_start
|
|
self.strategy = self.setup_strategy(
|
|
state=consts.SW_UPDATE_STATE_APPLYING)
|
|
subcloud2 = self.create_subcloud(self.ctxt, 'subcloud2', 1)
|
|
subcloud3 = self.create_subcloud(self.ctxt, 'subcloud3', 1)
|
|
subcloud4 = self.create_subcloud(self.ctxt, 'subcloud4', 1)
|
|
|
|
self.setup_strategy_step(
|
|
subcloud2.id, consts.STRATEGY_STATE_INITIAL)
|
|
self.setup_strategy_step(
|
|
subcloud3.id, consts.STRATEGY_STATE_INITIAL)
|
|
self.setup_strategy_step(
|
|
subcloud4.id, consts.STRATEGY_STATE_INITIAL)
|
|
|
|
self.worker.apply(self.strategy)
|
|
|
|
steps = db_api.strategy_step_get_all(self.ctx)
|
|
|
|
# the orchestrator can orchestrate 2 subclouds at a time
|
|
self.assertEqual(steps[0].state, consts.STRATEGY_STATE_IMPORTING_FIRMWARE)
|
|
self.assertEqual(steps[1].state, consts.STRATEGY_STATE_IMPORTING_FIRMWARE)
|
|
self.assertEqual(steps[2].state, consts.STRATEGY_STATE_INITIAL)
|
|
|
|
# subcloud3 orchestration finished first
|
|
db_api.strategy_step_update(self.ctx,
|
|
subcloud3.id,
|
|
state=consts.STRATEGY_STATE_COMPLETE)
|
|
|
|
self.worker.apply(self.strategy)
|
|
|
|
steps = db_api.strategy_step_get_all(self.ctx)
|
|
|
|
# the subcloud3 finished thus the subcloud 4 should start
|
|
self.assertEqual(steps[0].state, consts.STRATEGY_STATE_IMPORTING_FIRMWARE)
|
|
self.assertEqual(steps[1].state, consts.STRATEGY_STATE_COMPLETE)
|
|
self.assertEqual(steps[2].state, consts.STRATEGY_STATE_IMPORTING_FIRMWARE)
|
|
|
|
@mock.patch.object(scheduler.ThreadGroupManager, 'start')
|
|
def test_delete_strategy_single_step_no_vim_strategy(self, mock_start):
|
|
# The 'strategy' needs to be in 'deleting'
|
|
self.strategy = self.setup_strategy(
|
|
state=consts.SW_UPDATE_STATE_DELETING)
|
|
|
|
self.subcloud = self.setup_subcloud()
|
|
self.setup_strategy_step(
|
|
self.subcloud.id, consts.STRATEGY_STATE_CREATING_FW_UPDATE_STRATEGY)
|
|
|
|
# If the subcloud does not have a vim strategy, it raises an exception
|
|
self.vim_client.get_strategy.side_effect = Exception
|
|
|
|
mock_start.side_effect = non_threaded_start
|
|
|
|
# invoke the strategy (not strategy step) operation on the orch thread
|
|
self.worker.delete(self.strategy)
|
|
|
|
# There is a step, so the vim strategy should be queried
|
|
self.vim_client.get_strategy.assert_called()
|
|
|
|
# Verify the strategy was deleted
|
|
self.assertRaises(exception.NotFound,
|
|
db_api.sw_update_strategy_get,
|
|
self.ctx,
|
|
consts.SW_UPDATE_TYPE_FIRMWARE)
|
|
|
|
# Verify the steps were deleted
|
|
steps = db_api.strategy_step_get_all(self.ctx)
|
|
self.assertEqual(steps, [])
|
|
|
|
@mock.patch.object(scheduler.ThreadGroupManager, 'start')
|
|
def test_delete_strategy_single_step_with_vim_strategy(self, mock_start):
|
|
|
|
mock_start.side_effect = non_threaded_start
|
|
|
|
# The 'strategy' needs to be in 'deleting'
|
|
self.strategy = self.setup_strategy(
|
|
state=consts.SW_UPDATE_STATE_DELETING)
|
|
|
|
self.subcloud = self.setup_subcloud()
|
|
self.setup_strategy_step(
|
|
self.subcloud.id, consts.STRATEGY_STATE_CREATING_FW_UPDATE_STRATEGY)
|
|
|
|
# the subcloud returns a vim strategy
|
|
vim_strategy = FakeVimStrategy(state=vim.STATE_APPLIED)
|
|
self.vim_client.get_strategy.return_value = vim_strategy
|
|
|
|
# invoke the strategy (not strategy step) operation on the orch thread
|
|
self.worker.delete(self.strategy)
|
|
|
|
# There is a step, so the vim strategy should be queried and deleted
|
|
self.vim_client.get_strategy.assert_called()
|
|
self.vim_client.delete_strategy.assert_called()
|
|
|
|
# Verify the strategy was deleted
|
|
self.assertRaises(exception.NotFound,
|
|
db_api.sw_update_strategy_get,
|
|
self.ctx,
|
|
consts.SW_UPDATE_TYPE_FIRMWARE)
|
|
|
|
# Verify the steps were deleted
|
|
steps = db_api.strategy_step_get_all(self.ctx)
|
|
self.assertEqual(steps, [])
|