
This commit optimizes subcloud state manager's queries used to perform a bulk update for a subcloud's availability and endpoint status, which were being used in [1]. Previously, each database query was made separately for every update for either the availability status and/or the endpoint(s) during the audit process, which resulted in duplicated state and database calls. In [1] the RPC state calls were significantly reduced by creating a single request for each subcloud and. Consequently, all queries that were made in the database in separate steps of the process started to be executed at once, resulting in approximately 23 [2] queries per subcloud for a complete audit. With this commit, the maximum number of database transactions is reduced from 23 to 3. Test plan: 1. PASS: Unmanage a subcloud and verify that all of its endpoints' sync status become unknown. 2. PASS: Manage a subcloud and verify that all of its endpoints' sync status become in-sync 3. PASS: Apply a patch in the system controller and verify that all of the subclouds' patching sync status becomes out-of-sync. 4. PASS: Apply the patch in the subclouds and verify that their patching sync status becomes in-sync. 5. PASS: Verify that the hourly unconditional update for the subcloud's availability status updates the database [1] https://review.opendev.org/c/starlingx/distcloud/+/922058 [2] Analysis of the number of requests considering dcmanager's audit Subcloud becoming online: - subcloud_get_by_region_name: 1 - fm's get_fault: 1 - fm's set_fault or clear_fault: 1 - subcloud_update: 1 - for each endpoint audited by dcmanager (7): - subcloud_get_by_region_name: 1 (removed) - subcloud_get_with_status: 1 (removed) - subcloud_endpoint_status_db_model_to_dict: 9 (removed) - subcloud_status_update: 1 (changed to one query for all endpoints) - fm's get_fault: 1 - fm's set_fault or clear_fault: 1 Total: - 39 queries that are now 19, considering fm's database. - 23 queries that are now 3 in dcmanager's database Note that the totals does not include the db_model_to_dict request because it does not query the database. Subcloud becoming offline: - subcloud_get_by_region_name: 2 reduced to 1 - fm's get_fault: 10 (one for each endpoint and the availability) - fm's set_fault or clear_fault: 10 - subcloud_get_with_status: 1 (removed) - subcloud_endpoint_status_db_model_to_dict: 9 (removed) - subcloud_status_update_endpoints: 1 - subcloud_update: 1 Story: 2011106 Task: 50433 Change-Id: I34b8604bf445cc0ebdc02c5959a919221e62de5a Signed-off-by: Raphael Lima <Raphael.Lima@windriver.com>
4437 lines
195 KiB
Python
4437 lines
195 KiB
Python
# Copyright (c) 2017-2024 Wind River Systems, Inc.
|
|
# All Rights Reserved.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
#
|
|
|
|
import base64
|
|
import collections
|
|
import copy
|
|
import datetime
|
|
import filecmp
|
|
import json
|
|
import os
|
|
from os import path as os_path
|
|
import sys
|
|
import threading
|
|
import time
|
|
from urllib import request
|
|
import uuid
|
|
import yaml
|
|
|
|
from eventlet.green import subprocess
|
|
import mock
|
|
import netaddr
|
|
from oslo_concurrency import lockutils
|
|
from oslo_utils import timeutils
|
|
from tsconfig.tsconfig import SW_VERSION
|
|
|
|
from dccommon import consts as dccommon_consts
|
|
from dccommon.drivers.openstack import dcmanager_v1
|
|
from dccommon.exceptions import PlaybookExecutionFailed
|
|
from dccommon import kubeoperator
|
|
from dccommon import ostree_mount
|
|
from dccommon import subcloud_enrollment
|
|
from dccommon import subcloud_install
|
|
from dccommon.utils import AnsiblePlaybook
|
|
from dcmanager.common import consts
|
|
from dcmanager.common import exceptions
|
|
from dcmanager.common import prestage
|
|
from dcmanager.common import utils as cutils
|
|
from dcmanager.db.sqlalchemy import api as db_api
|
|
from dcmanager.manager import subcloud_manager
|
|
from dcmanager.manager import system_peer_manager
|
|
from dcmanager.state import subcloud_state_manager
|
|
from dcmanager.tests import base
|
|
from dcmanager.tests.unit.common import fake_subcloud
|
|
from dcmanager.tests.unit.db import test_subcloud_alarms
|
|
from dcmanager.tests.unit.manager import test_system_peer_manager
|
|
from dcmanager.tests import utils
|
|
|
|
sys.modules['fm_core'] = mock.Mock()
|
|
|
|
ANS_PATH = dccommon_consts.ANSIBLE_OVERRIDES_PATH
|
|
FAKE_PREVIOUS_SW_VERSION = '21.12'
|
|
FAKE_SW_VERSION = '22.12'
|
|
FAKE_ADMIN_USER_ID = 1
|
|
FAKE_SYSINV_USER_ID = 2
|
|
FAKE_DCMANAGER_USER_ID = 3
|
|
FAKE_ADMIN_PROJECT_ID = 1
|
|
FAKE_SERVICE_PROJECT_ID = 2
|
|
|
|
|
|
class FakeEndpoint(object):
|
|
def __init__(self, endpoint_name, region, service_id):
|
|
self.endpoint_name = endpoint_name
|
|
self.region = region
|
|
self.service_id = service_id
|
|
|
|
|
|
class FakeUser(object):
|
|
def __init__(self, name, id):
|
|
self.name = name
|
|
self.id = id
|
|
|
|
|
|
FAKE_USERS = [
|
|
FakeUser(dccommon_consts.ADMIN_USER_NAME, FAKE_ADMIN_USER_ID),
|
|
FakeUser(dccommon_consts.SYSINV_USER_NAME, FAKE_SYSINV_USER_ID),
|
|
FakeUser(dccommon_consts.DCMANAGER_USER_NAME, FAKE_DCMANAGER_USER_ID)
|
|
]
|
|
|
|
|
|
class FakeProject(object):
|
|
def __init__(self, name, id):
|
|
self.name = name
|
|
self.id = id
|
|
|
|
|
|
FAKE_PROJECTS = [
|
|
FakeProject(dccommon_consts.ADMIN_PROJECT_NAME, FAKE_ADMIN_PROJECT_ID),
|
|
FakeProject(dccommon_consts.SERVICES_USER_NAME, FAKE_SERVICE_PROJECT_ID)
|
|
]
|
|
|
|
|
|
class FakeService(object):
|
|
def __init__(self, type, id):
|
|
self.type = type
|
|
self.id = id
|
|
|
|
|
|
FAKE_SERVICES = [
|
|
FakeService(dccommon_consts.ENDPOINT_TYPE_PLATFORM, 1),
|
|
FakeService(dccommon_consts.ENDPOINT_TYPE_IDENTITY, 2),
|
|
FakeService(dccommon_consts.ENDPOINT_TYPE_PATCHING, 3),
|
|
FakeService(dccommon_consts.ENDPOINT_TYPE_FM, 4),
|
|
FakeService(dccommon_consts.ENDPOINT_TYPE_NFV, 5),
|
|
FakeService(dccommon_consts.ENDPOINT_TYPE_DC_CERT, 6),
|
|
FakeService(dccommon_consts.ENDPOINT_TYPE_SOFTWARE, 7)
|
|
]
|
|
|
|
|
|
class FakeKeystoneClient(object):
|
|
def __init__(self):
|
|
self.user_list = FAKE_USERS
|
|
self.project_list = FAKE_PROJECTS
|
|
self.services_list = FAKE_SERVICES
|
|
self.keystone_client = mock.MagicMock()
|
|
self.session = mock.MagicMock()
|
|
self.endpoint_cache = mock.MagicMock()
|
|
|
|
def get_enabled_users(self, id_only):
|
|
if not id_only:
|
|
return self.user_list
|
|
else:
|
|
return None
|
|
|
|
def get_enabled_projects(self, id_only):
|
|
if not id_only:
|
|
return self.project_list
|
|
else:
|
|
return None
|
|
|
|
def delete_endpoints(self, region_name):
|
|
pass
|
|
|
|
def delete_region(self, region_name):
|
|
pass
|
|
|
|
|
|
class FakeSystem(object):
|
|
def __init__(self, uuid):
|
|
self.uuid = uuid
|
|
self.name = 'fake_name'
|
|
|
|
|
|
class FakeController(object):
|
|
def __init__(self, hostname):
|
|
self.hostname = hostname
|
|
|
|
|
|
FAKE_CONTROLLERS = [
|
|
FakeController(
|
|
'controller-0'
|
|
),
|
|
FakeController(
|
|
'controller-1'
|
|
),
|
|
]
|
|
|
|
|
|
class FakeManagementInterface(object):
|
|
def __init__(self, uuid):
|
|
self.uuid = uuid
|
|
|
|
|
|
FAKE_MGMT_INTERFACES = [
|
|
FakeManagementInterface(
|
|
'47cb1222-21a9-4ee0-b1f9-0b37de345f65'
|
|
),
|
|
FakeManagementInterface(
|
|
'0106bdf0-1662-48cc-b6b3-664c91147843'
|
|
),
|
|
]
|
|
|
|
|
|
class FakeManagementPool(object):
|
|
def __init__(self):
|
|
self.floating_address = 'fdff:719a:bf60:233::2'
|
|
self.network = 'fdff:719a:bf60:233::'
|
|
self.prefix = 64
|
|
|
|
|
|
class FakeOamAddresses(object):
|
|
def __init__(self):
|
|
self.oam_floating_ip = '2620:10a:a001:d41::260'
|
|
self.oam_subnet = '2620:10a:a001:d41::/64'
|
|
|
|
|
|
class FakeSysinvClient(object):
|
|
def __init__(self):
|
|
self.hosts = FAKE_CONTROLLERS
|
|
self.interfaces = FAKE_MGMT_INTERFACES
|
|
self.mgmt_pool = FakeManagementPool()
|
|
self.oam_addresses = FakeOamAddresses()
|
|
|
|
def get_controller_hosts(self):
|
|
return self.hosts
|
|
|
|
def get_management_interface(self, hostname):
|
|
if hostname == 'controller-0':
|
|
return self.interfaces[0]
|
|
else:
|
|
return self.interfaces[1]
|
|
|
|
def get_management_address_pool(self):
|
|
return self.mgmt_pool
|
|
|
|
def get_oam_addresses(self):
|
|
return self.oam_addresses
|
|
|
|
def get_system(self):
|
|
return FakeSystem(str(uuid.uuid4()))
|
|
|
|
|
|
FAKE_SUBCLOUD_PRESTAGE_PAYLOAD = {
|
|
"install_values": fake_subcloud.FAKE_SUBCLOUD_INSTALL_VALUES,
|
|
"subcloud_name": 'subcloud1',
|
|
"force": False,
|
|
"oam_floating_ip": '10.10.10.12',
|
|
"software_version": "20.12",
|
|
"sysadmin_password": "testpasswd",
|
|
}
|
|
|
|
FAKE_PRESTAGE_RELEASE = '22.12'
|
|
FAKE_SUBCLOUD_SW_VERSION = '21.12'
|
|
FAKE_PRESTAGE_PAYLOAD = {
|
|
"subcloud_name": "subcloud1",
|
|
"oam_floating_ip": "10.10.10.12",
|
|
"sysadmin_password": (base64.b64encode
|
|
('testpass'.encode("utf-8"))).decode('ascii'),
|
|
"force": False,
|
|
"release": FAKE_PRESTAGE_RELEASE
|
|
}
|
|
|
|
FAKE_MGMT_IF_UUIDS = [
|
|
'47cb1222-21a9-4ee0-b1f9-0b37de345f65',
|
|
'0106bdf0-1662-48cc-b6b3-664c91147843'
|
|
]
|
|
|
|
FAKE_CACHED_REGIONONE_DATA = {
|
|
"admin_user_id": FAKE_USERS[0].id,
|
|
"sysinv_user_id": FAKE_USERS[1].id,
|
|
"dcmanager_user_id": FAKE_USERS[2].id,
|
|
"admin_project_id": FAKE_PROJECTS[0].id,
|
|
"services_project_id": FAKE_PROJECTS[1].id,
|
|
"mgmt_interface_uuids": FAKE_MGMT_IF_UUIDS,
|
|
"expiry": datetime.datetime.utcnow() + datetime.timedelta(seconds=3600),
|
|
'mgmt_pool': FakeManagementPool(),
|
|
'oam_addresses': FakeOamAddresses()
|
|
}
|
|
|
|
FAKE_BACKUP_DELETE_LOAD = {
|
|
"sysadmin_password": "testpasswd",
|
|
"subcloud": 1,
|
|
"software_version": "22.12"
|
|
}
|
|
|
|
FAKE_BACKUP_DELETE_LOCAL_LOAD = {
|
|
"sysadmin_password": "testpasswd",
|
|
"subcloud": 1,
|
|
"software_version": "22.12",
|
|
"local_only": True
|
|
}
|
|
|
|
FAKE_BACKUP_DELETE_LOAD_1 = {
|
|
"sysadmin_password": "testpasswd",
|
|
"subcloud": 1,
|
|
"software_version": "22.12",
|
|
"local_only": False
|
|
}
|
|
|
|
FAKE_BACKUP_CREATE_LOAD = {
|
|
"sysadmin_password": "testpasswd",
|
|
"subcloud": 1,
|
|
}
|
|
|
|
FAKE_BACKUP_CREATE_LOAD_1 = {
|
|
"sysadmin_password": "testpasswd",
|
|
"subcloud": 1,
|
|
"local_only": False,
|
|
"registry_images": False,
|
|
}
|
|
|
|
FAKE_BACKUP_RESTORE_LOAD = {
|
|
"sysadmin_password": "testpasswd",
|
|
"subcloud": 1
|
|
}
|
|
|
|
FAKE_BACKUP_RESTORE_LOAD_WITH_INSTALL = {
|
|
"sysadmin_password": "testpasswd",
|
|
"subcloud": 1,
|
|
"install_values": fake_subcloud.FAKE_SUBCLOUD_INSTALL_VALUES
|
|
}
|
|
|
|
SERVICE_ENDPOINTS = [dccommon_consts.ENDPOINT_TYPE_PLATFORM,
|
|
dccommon_consts.ENDPOINT_TYPE_IDENTITY,
|
|
dccommon_consts.ENDPOINT_TYPE_PATCHING,
|
|
dccommon_consts.ENDPOINT_TYPE_FM,
|
|
dccommon_consts.ENDPOINT_TYPE_NFV,
|
|
dccommon_consts.ENDPOINT_TYPE_SOFTWARE]
|
|
|
|
|
|
class Subcloud(object):
|
|
def __init__(self, data, is_online):
|
|
self.id = data['id']
|
|
self.name = data['name']
|
|
self.description = data['description']
|
|
self.location = data['location']
|
|
self.software_version = data['software-version']
|
|
self.management_state = dccommon_consts.MANAGEMENT_UNMANAGED
|
|
if is_online:
|
|
self.availability_status = dccommon_consts.AVAILABILITY_ONLINE
|
|
else:
|
|
self.availability_status = dccommon_consts.AVAILABILITY_OFFLINE
|
|
self.deploy_status = data['deploy_status']
|
|
self.error_description = data['error_description']
|
|
self.management_subnet = data['management_subnet']
|
|
self.management_gateway_ip = data['management_gateway_address']
|
|
self.management_start_ip = data['management_start_address']
|
|
self.management_end_ip = data['management_end_address']
|
|
self.external_oam_subnet = data['external_oam_subnet']
|
|
self.external_oam_gateway_address = \
|
|
data['external_oam_gateway_address']
|
|
self.external_oam_floating_address = \
|
|
data['external_oam_floating_address']
|
|
self.systemcontroller_gateway_ip = \
|
|
data['systemcontroller_gateway_address']
|
|
self.data_install = data['data_install']
|
|
self.created_at = timeutils.utcnow()
|
|
self.updated_at = timeutils.utcnow()
|
|
|
|
|
|
class BaseTestSubcloudManager(base.DCManagerTestCase):
|
|
"""Base class for testing subcloud Manager"""
|
|
|
|
def setUp(self):
|
|
super().setUp()
|
|
|
|
self._mock_audit_rpc_client()
|
|
self._mock_rpc_subcloud_state_client()
|
|
# TODO(sgorre): Need to improve same mock method
|
|
# two times for different targets
|
|
self._mock_openstack_driver(subcloud_install)
|
|
self.mock_install_openstack_driver = self.mock_openstack_driver
|
|
self._mock_sysinv_client(subcloud_install)
|
|
self.mock_install_sysinv_client = self.mock_sysinv_client
|
|
self._mock_openstack_driver(subcloud_manager)
|
|
self._mock_sysinv_client(subcloud_manager)
|
|
self._mock_dcorch_api()
|
|
self._mock_dcmanager_api()
|
|
self._mock_context()
|
|
self._mock_log(subcloud_manager)
|
|
self._mock_subcloud_manager_keyring()
|
|
self._mock_utils_create_subcloud_inventory()
|
|
self._mock_utils_delete_subcloud_inventory()
|
|
self._mock_utils_get_playbook_for_software_version()
|
|
self._mock_subprocess_run()
|
|
self._mock_ansible_run_playbook()
|
|
self._mock_os_mkdir()
|
|
self._mock_os_listdir()
|
|
self._mock_os_path_isdir()
|
|
self._mock_os_path_exists()
|
|
self._mock_os_remove()
|
|
self._mock_ostree_mount_validate_ostree_iso_mount()
|
|
self._mock_get_local_system()
|
|
self.sm = subcloud_manager.SubcloudManager()
|
|
|
|
self.subcloud = self.create_subcloud_static(self.ctx)
|
|
self.test_system_peer_manager = \
|
|
test_system_peer_manager.TestSystemPeerManager
|
|
self.system_peer = self.test_system_peer_manager.create_system_peer_static(
|
|
self.ctx, peer_name='SystemPeer1'
|
|
)
|
|
self.peer_group = self.create_subcloud_peer_group_static(self.ctx)
|
|
|
|
self.mock_keyring.get_password.return_value = "testpassword"
|
|
self.fake_install_values = copy.copy(
|
|
fake_subcloud.FAKE_SUBCLOUD_INSTALL_VALUES)
|
|
self.fake_bootstrap_values = copy.copy(
|
|
fake_subcloud.FAKE_BOOTSTRAP_FILE_DATA)
|
|
|
|
self.fake_payload = {"sysadmin_password": "testpass",
|
|
"deploy_playbook": "test_playbook.yaml",
|
|
"deploy_overrides": "test_overrides.yaml",
|
|
"deploy_chart": "test_chart.yaml",
|
|
"deploy_config": "subcloud1.yaml"}
|
|
self.fake_payload_install = {'bmc_password': 'bmc_pass',
|
|
'install_values': self.fake_install_values,
|
|
'software_version': SW_VERSION,
|
|
'sysadmin_password': 'sys_pass'}
|
|
self.fake_payload_enroll = {'bmc_password': 'bmc_pass',
|
|
'install_values': self.fake_install_values,
|
|
'software_version': SW_VERSION,
|
|
'sysadmin_password': 'sys_pass',
|
|
'admin_password': 'sys_pass'
|
|
}
|
|
self.fake_payload_enroll = dict(self.fake_payload_enroll,
|
|
**self.fake_bootstrap_values,
|
|
**self.fake_install_values)
|
|
|
|
rel_version = self.fake_payload_enroll.get('software_version')
|
|
|
|
self.iso_dir = (f'/opt/platform/iso/{rel_version}/'
|
|
f'nodes/{self.subcloud.name}')
|
|
self.iso_file = f'{self.iso_dir}/seed.iso'
|
|
|
|
# Reset the regionone_data cache between tests
|
|
subcloud_manager.SubcloudManager.regionone_data = \
|
|
collections.defaultdict(dict)
|
|
|
|
def patched_isdir(self, path):
|
|
return path != self.iso_dir
|
|
|
|
def _mock_dcorch_api(self):
|
|
"""Mock the DCOrch API"""
|
|
|
|
p = mock.patch('dcorch.rpc.client.EngineWorkerClient')
|
|
self.mock_dcorch_api = p.start()
|
|
self.addCleanup(p.stop)
|
|
|
|
def _mock_dcmanager_api(self):
|
|
"""Mock the DCManager notifications"""
|
|
|
|
p = mock.patch('dcmanager.rpc.client.DCManagerNotifications')
|
|
self.mock_dcmanager_api = p.start()
|
|
self.addCleanup(p.stop)
|
|
|
|
def _mock_utils_create_subcloud_inventory(self):
|
|
"""Mock utils create_subcloud_inventory"""
|
|
|
|
mock_patch = mock.patch.object(cutils, 'create_subcloud_inventory')
|
|
self.mock_create_subcloud_inventory = mock_patch.start()
|
|
self.addCleanup(mock_patch.stop)
|
|
|
|
def _mock_utils_delete_subcloud_inventory(self):
|
|
"""Mock utils's delete_subcloud_inventory"""
|
|
|
|
mock_patch = mock.patch.object(cutils,
|
|
'delete_subcloud_inventory')
|
|
self.mock_delete_subcloud_inventory = mock_patch.start()
|
|
self.addCleanup(mock_patch.stop)
|
|
|
|
def _mock_utils_get_playbook_for_software_version(self):
|
|
"""Mock utils's get_playbook_for_software_version"""
|
|
|
|
mock_patch = mock.patch.object(cutils,
|
|
'get_playbook_for_software_version')
|
|
self.mock_get_playbook_for_software_version = mock_patch.start()
|
|
self.addCleanup(mock_patch.stop)
|
|
|
|
def _mock_subcloud_manager_keyring(self):
|
|
"""Mock subcloud manager's keyring"""
|
|
|
|
mock_patch = mock.patch.object(subcloud_manager, 'keyring')
|
|
self.mock_keyring = mock_patch.start()
|
|
self.addCleanup(mock_patch.stop)
|
|
|
|
def _mock_subcloud_manager_get_cached_regionone_data(self):
|
|
"""Mock subcloud manager's _get_cached_regionone_data"""
|
|
|
|
mock_patch = mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_get_cached_regionone_data')
|
|
self.mock_get_cached_regionone_data = mock_patch.start()
|
|
self.addCleanup(mock_patch.stop)
|
|
|
|
def _mock_subcloud_manager_create_addn_hosts_dc(self):
|
|
"""Mock subcloud manager's _create_addn_hosts_dc"""
|
|
|
|
mock_patch = mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_addn_hosts_dc')
|
|
self.mock_create_addn_hosts = mock_patch.start()
|
|
self.addCleanup(mock_patch.stop)
|
|
|
|
def _mock_subcloud_manager_write_subcloud_ansible_config(self):
|
|
"""Mock subcloud manager's _write_subcloud_ansible_config"""
|
|
|
|
mock_patch = mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_write_subcloud_ansible_config')
|
|
self.mock_write_subcloud_ansible_config = mock_patch.start()
|
|
self.addCleanup(mock_patch.stop)
|
|
|
|
def _mock_subcloud_manager_run_subcloud_install(self):
|
|
"""Mock subcloud manager's _run_subcloud_install"""
|
|
|
|
mock_patch = mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_run_subcloud_install')
|
|
self.mock_run_subcloud_install = mock_patch.start()
|
|
self.addCleanup(mock_patch.stop)
|
|
|
|
def _mock_ostree_mount_validate_ostree_iso_mount(self):
|
|
"""Mock ostree_mount validate_ostree_iso_mount"""
|
|
|
|
mock_patch = mock.patch.object(ostree_mount,
|
|
'validate_ostree_iso_mount')
|
|
self.mock_validate_ostree_iso_mount = mock_patch.start()
|
|
|
|
def _mock_subcloud_manager_run_subcloud_enroll(self):
|
|
"""Mock subcloud manager's _run_subcloud_enroll"""
|
|
|
|
mock_patch = mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_run_subcloud_enroll')
|
|
self.mock_run_subcloud_enroll = mock_patch.start()
|
|
self.addCleanup(mock_patch.stop)
|
|
|
|
def _mock_subcloud_manager_create_intermediate_ca_cert(self):
|
|
"""Mock subcloud manager's _create_intermediate_ca_cert"""
|
|
|
|
mock_patch = mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_intermediate_ca_cert')
|
|
self.mock_create_intermediate_ca_cert = mock_patch.start()
|
|
self.addCleanup(mock_patch.stop)
|
|
|
|
def _mock_subcloud_manager_compose_install_command(self):
|
|
"""Mock subcloud manager's compose_install_command"""
|
|
|
|
mock_patch = mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'compose_install_command')
|
|
self.mock_compose_install_command = mock_patch.start()
|
|
self.addCleanup(mock_patch.stop)
|
|
|
|
def _mock_subcloud_manager_compose_enroll_command(self):
|
|
"""Mock subcloud manager compose_enroll_command"""
|
|
|
|
mock_patch = mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'compose_enroll_command')
|
|
self.mock_compose_enroll_command = mock_patch.start()
|
|
self.addCleanup(mock_patch.stop)
|
|
|
|
def _mock_netaddr_ipaddress(self):
|
|
"""Mock netaddr's IPAddress"""
|
|
|
|
mock_patch = mock.patch.object(netaddr, 'IPAddress')
|
|
self.mock_netaddr_ipaddress = mock_patch.start()
|
|
self.addCleanup(mock_patch.stop)
|
|
|
|
def _mock_subprocess_run(self):
|
|
"""Mock subprocess' run"""
|
|
|
|
mock_patch = mock.patch.object(subprocess, 'run')
|
|
self.mock_subprocess_run = mock_patch.start()
|
|
self.addCleanup(mock_patch.stop)
|
|
|
|
def _mock_context(self):
|
|
"""Mock the context"""
|
|
|
|
p = mock.patch.object(subcloud_manager, 'dcmanager_context')
|
|
self.mock_context = p.start()
|
|
self.mock_context.get_admin_context.return_value = self.ctx
|
|
self.addCleanup(p.stop)
|
|
|
|
@staticmethod
|
|
def create_subcloud_static(ctxt, **kwargs):
|
|
values = {
|
|
"name": "subcloud1",
|
|
"description": "subcloud1 description",
|
|
"location": "subcloud1 location",
|
|
'software_version': "18.03",
|
|
"management_subnet": "192.168.101.0/24",
|
|
"management_gateway_ip": "192.168.101.1",
|
|
"management_start_ip": "192.168.101.2",
|
|
"management_end_ip": "192.168.101.50",
|
|
"systemcontroller_gateway_ip": "192.168.204.101",
|
|
'deploy_status': "not-deployed",
|
|
'error_description': "No errors present",
|
|
'region_name': base.SUBCLOUD_1['region_name'],
|
|
'openstack_installed': False,
|
|
'group_id': 1,
|
|
'data_install': 'data from install',
|
|
}
|
|
values.update(kwargs)
|
|
return db_api.subcloud_create(ctxt, **values)
|
|
|
|
@staticmethod
|
|
def create_subcloud_peer_group_static(ctxt, **kwargs):
|
|
values = {
|
|
"peer_group_name": "pgname",
|
|
"system_leader_id": "12e0cb13-2c5c-480e-b0ea-9161fc03f3ef",
|
|
"system_leader_name": "DC0",
|
|
"group_priority": 0,
|
|
"group_state": "enabled",
|
|
"max_subcloud_rehoming": 50,
|
|
"migration_status": None
|
|
}
|
|
values.update(kwargs)
|
|
return db_api.subcloud_peer_group_create(ctxt, **values)
|
|
|
|
def test_init(self):
|
|
self.assertIsNotNone(self.sm)
|
|
self.assertEqual('subcloud_manager', self.sm.service_name)
|
|
self.assertEqual('localhost', self.sm.host)
|
|
self.assertEqual(self.ctx, self.sm.context)
|
|
|
|
|
|
class TestSubcloudManager(BaseTestSubcloudManager):
|
|
"""Test class for testing subcloud Manager"""
|
|
|
|
def setUp(self):
|
|
super().setUp()
|
|
self._mock_netaddr_ipaddress()
|
|
self.values = utils.create_subcloud_dict(base.SUBCLOUD_SAMPLE_DATA_0)
|
|
self.test_system_peer_manager.create_peer_group_association_static(
|
|
self.ctx,
|
|
system_peer_id=self.system_peer.id,
|
|
peer_group_id=self.peer_group.id)
|
|
self.payload = {'name': self.subcloud.name,
|
|
'description': "subcloud description",
|
|
'location': "subcloud location",
|
|
'management_subnet': "192.168.102.0/24",
|
|
'management_start_ip': "192.168.102.5",
|
|
'management_end_ip': "192.168.102.49",
|
|
'management_gateway_ip': "192.168.102.1"}
|
|
|
|
def test_network_reconf_same_subnet_endpoint_type_platform(self):
|
|
self.payload.update({
|
|
'management_subnet': "192.168.101.0/24",
|
|
'management_start_address': "192.168.101.3",
|
|
'management_end_ip': "192.168.101.49",
|
|
'management_gateway_ip': "192.168.101.1"})
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_DONE
|
|
endpoint_1 = FakeEndpoint('endpoint1', 'regionOne', '1')
|
|
self.mock_openstack_driver().keystone_client.endpoints.list.return_value = \
|
|
[endpoint_1]
|
|
self.mks_client = self.mock_openstack_driver()
|
|
db_api.subcloud_update(
|
|
self.ctx, self.subcloud.id,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE)
|
|
for endpoint_type in SERVICE_ENDPOINTS:
|
|
self.mock_openstack_driver().keystone_client.services.\
|
|
get.return_value.type = endpoint_type
|
|
self.sm._update_services_endpoint(
|
|
self.ctx, self.payload, self.subcloud.region_name, self.mks_client
|
|
)
|
|
self.mock_log.info.assert_called_with(
|
|
'Update services endpoint to 192.168.101.3 in subcloud region '
|
|
f'{base.SUBCLOUD_1["region_name"]}')
|
|
self.assertEqual(
|
|
self.mock_dcmanager_api().subcloud_sysinv_endpoint_update.call_count, 6)
|
|
|
|
@mock.patch.object(kubeoperator, 'KubeOperator')
|
|
@mock.patch.object(json, 'dumps')
|
|
def test_generate_subcloud_ansible_config_initial_deployment(
|
|
self, mock_dumps, mock_kubeoperator
|
|
):
|
|
self._mock_builtins_open()
|
|
self.subcloud['region_name'] = self.values['region_name']
|
|
self.sm.generate_subcloud_ansible_config(self.subcloud,
|
|
self.values,
|
|
True)
|
|
self.mock_keyring.get_password.assert_called()
|
|
self.mock_create_subcloud_inventory.assert_called_once()
|
|
mock_kubeoperator.assert_called_once()
|
|
|
|
def test_generate_subcloud_ansible_config(self):
|
|
self._mock_subcloud_manager_write_subcloud_ansible_config()
|
|
self._mock_subcloud_manager_create_intermediate_ca_cert()
|
|
self._mock_builtins_open()
|
|
self.subcloud['region_name'] = self.values['region_name']
|
|
initial_deployment = False
|
|
self.sm.generate_subcloud_ansible_config(self.subcloud,
|
|
self.values,
|
|
initial_deployment)
|
|
self.mock_keyring.get_password.assert_called()
|
|
self.mock_create_subcloud_inventory.assert_called_once()
|
|
self.mock_create_intermediate_ca_cert.assert_called_once()
|
|
self.mock_write_subcloud_ansible_config.assert_called_once()
|
|
|
|
def test_generate_subcloud_ansible_config_failed(self):
|
|
self.mock_create_subcloud_inventory.side_effect = base.FakeException('boom')
|
|
self.subcloud['region_name'] = self.values['region_name']
|
|
self.assertRaises(
|
|
base.FakeException, self.sm.generate_subcloud_ansible_config,
|
|
self.subcloud, self.values)
|
|
|
|
def test_get_subcloud_name_by_region_name(self):
|
|
self.subcloud.update(
|
|
{'name': fake_subcloud.FAKE_SUBCLOUD_DATA['name'],
|
|
'region_name': fake_subcloud.FAKE_SUBCLOUD_DATA['region_name']}
|
|
)
|
|
self.sm.get_subcloud_name_by_region_name(self.ctx, self.subcloud.region_name)
|
|
|
|
ret = db_api.subcloud_get_by_region_name(self.ctx, self.subcloud.region_name)
|
|
self.assertEqual(self.subcloud.name, ret.name)
|
|
|
|
def test_get_peer_system_list_with_unavailable_peer(self):
|
|
db_api.system_peer_update(
|
|
self.ctx, self.system_peer.id,
|
|
availability_state=consts.SYSTEM_PEER_AVAILABILITY_STATE_UNAVAILABLE
|
|
)
|
|
self.sm._get_peer_system_list(self.peer_group)
|
|
self.mock_log.warning.assert_called_once_with(
|
|
"Peer system %s offline, skip checking" % self.system_peer.peer_name)
|
|
|
|
def test_get_peer_system_list_availability_state(self):
|
|
db_api.system_peer_update(
|
|
self.ctx, self.system_peer.id,
|
|
availability_state=consts.SYSTEM_PEER_AVAILABILITY_STATE_AVAILABLE
|
|
)
|
|
return_peer = self.sm._get_peer_system_list(self.peer_group)
|
|
self.assertEqual(len(return_peer), 1)
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'subcloud_deploy_install', return_value=True)
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'subcloud_deploy_bootstrap', return_value=True)
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'subcloud_deploy_config', return_value=True)
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'subcloud_deploy_complete', return_value=True)
|
|
def test_run_deploy_phases(
|
|
self, mock_subcloud_deploy_complete, mock_subcloud_deploy_config,
|
|
mock_subcloud_deploy_bootstrap, mock_subcloud_deploy_install,
|
|
):
|
|
deploy_phases_to_run = [consts.DEPLOY_PHASE_COMPLETE,
|
|
consts.DEPLOY_PHASE_CONFIG,
|
|
consts.DEPLOY_PHASE_INSTALL,
|
|
consts.DEPLOY_PHASE_BOOTSTRAP
|
|
]
|
|
manager = mock.Mock()
|
|
manager.attach_mock(
|
|
mock_subcloud_deploy_install, 'mock_subcloud_deploy_install')
|
|
manager.attach_mock(
|
|
mock_subcloud_deploy_bootstrap, 'mock_subcloud_deploy_bootstrap')
|
|
manager.attach_mock(
|
|
mock_subcloud_deploy_config, 'mock_subcloud_deploy_config')
|
|
manager.attach_mock(
|
|
mock_subcloud_deploy_complete, 'mock_subcloud_deploy_complete')
|
|
values = utils.create_subcloud_dict(base.SUBCLOUD_SAMPLE_DATA_0)
|
|
self.sm.run_deploy_phases(
|
|
self.ctx, self.subcloud.id, values, deploy_phases_to_run
|
|
)
|
|
expected_calls = [
|
|
mock.call.mock_subcloud_deploy_install(
|
|
self.ctx, self.subcloud.id, values, False),
|
|
mock.call.mock_subcloud_deploy_bootstrap(
|
|
self.ctx, self.subcloud.id, values, False),
|
|
mock.call.mock_subcloud_deploy_config(
|
|
self.ctx, self.subcloud.id, values, False),
|
|
mock.call.mock_subcloud_deploy_complete(self.ctx, self.subcloud.id)]
|
|
self.assertEqual(expected_calls, manager.mock_calls)
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'subcloud_deploy_install')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'subcloud_deploy_bootstrap')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'subcloud_deploy_config')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'subcloud_deploy_complete')
|
|
def test_run_deploy_phases_failed(
|
|
self, mock_subcloud_deploy_complete, mock_subcloud_deploy_config,
|
|
mock_subcloud_deploy_bootstrap, mock_subcloud_deploy_install,
|
|
):
|
|
values = utils.create_subcloud_dict(base.SUBCLOUD_SAMPLE_DATA_0)
|
|
mock_subcloud_deploy_config.return_value = False
|
|
deploy_phases_to_run = [consts.DEPLOY_PHASE_COMPLETE,
|
|
consts.DEPLOY_PHASE_CONFIG,
|
|
consts.DEPLOY_PHASE_INSTALL,
|
|
consts.DEPLOY_PHASE_BOOTSTRAP
|
|
]
|
|
ret = self.sm.run_deploy_phases(
|
|
self.ctx, self.subcloud.id, values, deploy_phases_to_run
|
|
)
|
|
self.assertEqual(ret, False)
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_update_backup_status_by_ids')
|
|
def test_mark_invalid_subclouds_for_backup_validation_failed(
|
|
self, mock_update_backup_status_by_ids
|
|
):
|
|
mock_update_backup_status_by_ids.side_effect = \
|
|
exceptions.DCManagerException
|
|
self.assertRaises(
|
|
exceptions.DCManagerException,
|
|
self.sm._mark_invalid_subclouds_for_backup,
|
|
self.ctx, [self.subcloud]
|
|
)
|
|
self.mock_log.exception.assert_called_once_with(
|
|
'Subcloud backup validation failed'
|
|
)
|
|
|
|
def test_mark_invalid_subclouds_for_backup_validation_success(self):
|
|
self.sm._mark_invalid_subclouds_for_backup(
|
|
self.ctx, [self.subcloud])
|
|
self.mock_log.warn.assert_called_once_with(
|
|
'The following subclouds are not online and/or managed '
|
|
'and/or in a valid deploy state, and will not be backed up: %s',
|
|
self.subcloud.name
|
|
)
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.BACKUP_STATE_VALIDATE_FAILED,
|
|
updated_subcloud.backup_status)
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_delete_subcloud_routes')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_update_services_endpoint')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_subcloud_route')
|
|
def test_update_subcloud_network_reconfiguration(
|
|
self, mock_create_route,
|
|
mock_update_endpoints, mock_delete_route):
|
|
self._mock_subcloud_manager_create_addn_hosts_dc()
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_RECONFIGURING_NETWORK
|
|
db_api.subcloud_update(
|
|
self.ctx, self.subcloud.id,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE)
|
|
|
|
self.sm._run_network_reconfiguration(
|
|
self.subcloud.name, mock.ANY, None,
|
|
self.payload, self.ctx, self.subcloud)
|
|
|
|
self.mock_ansible_run_playbook.assert_called_once()
|
|
self.mock_openstack_driver.assert_called_once()
|
|
mock_create_route.assert_called_once()
|
|
mock_update_endpoints.assert_called_once()
|
|
mock_delete_route.assert_called_once()
|
|
self.mock_create_addn_hosts.assert_called_once()
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
for key, value in self.payload.items():
|
|
self.assertEqual(value, updated_subcloud[key])
|
|
self.assertEqual(consts.DEPLOY_STATE_DONE,
|
|
updated_subcloud.deploy_status)
|
|
|
|
@mock.patch.object(kubeoperator, 'KubeOperator')
|
|
def test_run_network_reconfiguration_fail(self, mock_kube_operator):
|
|
self.mock_ansible_run_playbook.side_effect = PlaybookExecutionFailed()
|
|
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_RECONFIGURING_NETWORK
|
|
|
|
self.sm._run_network_reconfiguration(self.subcloud.name,
|
|
mock.ANY, None, self.payload,
|
|
self.ctx, self.subcloud)
|
|
self.mock_log.error.assert_called_once_with(
|
|
'FAILED reconfiguring-network playbook of (subcloud1).'
|
|
'\ncheck individual log at /var/log/dcmanager/ansible'
|
|
'/subcloud1_playbook_output.log for detailed output ')
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_RECONFIGURING_NETWORK_FAILED,
|
|
updated_subcloud.deploy_status)
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_delete_subcloud_routes')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_update_services_endpoint')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_subcloud_route')
|
|
def test_configure_system_controller_network(
|
|
self, mock_create_route,
|
|
mock_update_endpoints, mock_delete_route):
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_RECONFIGURING_NETWORK
|
|
db_api.subcloud_update(
|
|
self.ctx, self.subcloud.id,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE)
|
|
|
|
self.payload.update({
|
|
'management_subnet': "192.168.101.0/24",
|
|
'management_start_ip': "192.168.101.3",
|
|
'management_end_ip': "192.168.101.49",
|
|
'management_gateway_ip': "192.168.101.1"})
|
|
|
|
self.sm._configure_system_controller_network(
|
|
self.ctx, self.payload, self.subcloud)
|
|
|
|
self.mock_openstack_driver.assert_called_once()
|
|
mock_create_route.assert_called_once()
|
|
mock_update_endpoints.assert_called_once()
|
|
mock_delete_route.assert_not_called()
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_delete_subcloud_routes')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_update_services_endpoint')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_subcloud_route')
|
|
def test_configure_system_controller_network_failed_to_update_endpoints(
|
|
self, mock_create_route,
|
|
mock_update_endpoints, mock_delete_route):
|
|
mock_update_endpoints.side_effect = Exception('boom')
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_RECONFIGURING_NETWORK
|
|
db_api.subcloud_update(
|
|
self.ctx, self.subcloud.id,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE)
|
|
|
|
self.sm._configure_system_controller_network(
|
|
self.ctx, self.payload, self.subcloud)
|
|
|
|
self.mock_openstack_driver.assert_called_once()
|
|
mock_create_route.assert_called_once()
|
|
mock_update_endpoints.assert_called_once()
|
|
self.assertFalse(mock_delete_route.called)
|
|
self.mock_log.exception.assert_called_once_with(
|
|
f'Failed to update subcloud {self.subcloud.name} endpoints')
|
|
mock_delete_route.assert_not_called()
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_RECONFIGURING_NETWORK_FAILED,
|
|
updated_subcloud.deploy_status)
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_delete_subcloud_routes')
|
|
def test_configure_system_controller_network_failed(
|
|
self, mock_delete_route
|
|
):
|
|
self.mock_sysinv_client.side_effect = Exception('boom')
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_RECONFIGURING_NETWORK
|
|
db_api.subcloud_update(
|
|
self.ctx, self.subcloud.id,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE)
|
|
|
|
self.sm._configure_system_controller_network(
|
|
self.ctx, self.payload, self.subcloud)
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(
|
|
self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_RECONFIGURING_NETWORK_FAILED,
|
|
updated_subcloud.deploy_status)
|
|
self.mock_openstack_driver.assert_called_once()
|
|
self.assertFalse(mock_delete_route.called)
|
|
self.mock_log.exception.assert_called_once_with(
|
|
f'Failed to create route to subcloud {self.subcloud.name}.')
|
|
self.assertEqual(consts.DEPLOY_STATE_RECONFIGURING_NETWORK_FAILED,
|
|
updated_subcloud.deploy_status)
|
|
|
|
|
|
class TestSubcloudDeploy(BaseTestSubcloudManager):
|
|
"""Test class for testing subcloud deploy"""
|
|
|
|
def setUp(self):
|
|
super().setUp()
|
|
|
|
self._mock_builtins_open()
|
|
self._mock_subcloud_manager_create_addn_hosts_dc()
|
|
self._mock_subcloud_manager_run_subcloud_install()
|
|
self._mock_subcloud_manager_create_intermediate_ca_cert()
|
|
self.mock_openstack_driver().keystone_client = FakeKeystoneClient()
|
|
self._mock_subcloud_manager_get_cached_regionone_data()
|
|
self.mock_get_cached_regionone_data.return_value = \
|
|
FAKE_CACHED_REGIONONE_DATA
|
|
|
|
self.payload = {**fake_subcloud.FAKE_BOOTSTRAP_VALUE,
|
|
**fake_subcloud.FAKE_BOOTSTRAP_FILE_DATA,
|
|
'sysadmin_password': 'testpass'}
|
|
|
|
def test_subcloud_deploy_install(self):
|
|
self._mock_subcloud_manager_compose_install_command()
|
|
self.mock_run_subcloud_install.return_value = True
|
|
self.fake_payload_install['software_version'] = FAKE_PREVIOUS_SW_VERSION
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_PRE_INSTALL
|
|
self.fake_install_values['software_version'] = SW_VERSION
|
|
|
|
self.sm.subcloud_deploy_install(
|
|
self.ctx, self.subcloud.id, payload=self.fake_payload_install)
|
|
|
|
self.mock_compose_install_command.assert_called_once_with(
|
|
self.subcloud.name,
|
|
cutils.get_ansible_filename(
|
|
self.subcloud.name, consts.INVENTORY_FILE_POSTFIX),
|
|
FAKE_PREVIOUS_SW_VERSION)
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx,
|
|
self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_INSTALLED,
|
|
updated_subcloud.deploy_status)
|
|
|
|
def test_subcloud_deploy_install_failed(self):
|
|
self.mock_run_subcloud_install.side_effect = Exception('boom')
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_PRE_INSTALL
|
|
self.fake_install_values['software_version'] = SW_VERSION
|
|
|
|
self.sm.subcloud_deploy_install(self.ctx, self.subcloud.id,
|
|
payload=self.fake_payload_install)
|
|
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx,
|
|
self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_PRE_INSTALL_FAILED,
|
|
updated_subcloud.deploy_status)
|
|
|
|
def test_subcloud_deploy_create(self):
|
|
self._mock_subcloud_manager_write_subcloud_ansible_config()
|
|
values = utils.create_subcloud_dict(base.SUBCLOUD_SAMPLE_DATA_0)
|
|
values['deploy_status'] = consts.DEPLOY_STATE_NONE
|
|
|
|
# dcmanager add_subcloud queries the data from the db
|
|
subcloud = self.create_subcloud_static(self.ctx, name=values['name'],
|
|
region_name=values['region_name'])
|
|
values['id'] = subcloud.id
|
|
|
|
subcloud_dict = self.sm.subcloud_deploy_create(
|
|
self.ctx, subcloud.id, payload=values)
|
|
|
|
self.mock_get_cached_regionone_data.assert_called_once()
|
|
self.mock_sysinv_client().create_route.assert_called()
|
|
self.mock_dcorch_api().add_subcloud.assert_called_once()
|
|
self.mock_create_addn_hosts.assert_called_once()
|
|
self.mock_create_subcloud_inventory.assert_called_once()
|
|
self.mock_write_subcloud_ansible_config.assert_called_once()
|
|
self.mock_keyring.get_password.assert_called()
|
|
self.mock_create_intermediate_ca_cert.assert_called_once()
|
|
|
|
# Verify subcloud was updated with correct values
|
|
self.assertEqual(consts.DEPLOY_STATE_CREATED,
|
|
subcloud_dict['deploy-status'])
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, values['name'])
|
|
self.assertEqual(consts.DEPLOY_STATE_CREATED,
|
|
updated_subcloud.deploy_status)
|
|
|
|
def test_subcloud_deploy_create_failed(self):
|
|
values = utils.create_subcloud_dict(base.SUBCLOUD_SAMPLE_DATA_0)
|
|
values['deploy_status'] = consts.DEPLOY_STATE_NONE
|
|
|
|
# dcmanager add_subcloud queries the data from the db
|
|
subcloud = self.create_subcloud_static(self.ctx, name=values['name'],
|
|
region_name=values['region_name'])
|
|
values['id'] = subcloud.id
|
|
|
|
self.mock_openstack_driver.side_effect = Exception('boom')
|
|
|
|
subcloud_dict = self.sm.subcloud_deploy_create(
|
|
self.ctx, subcloud.id, payload=values)
|
|
|
|
# Verify subcloud was updated with correct values
|
|
self.assertEqual(consts.DEPLOY_STATE_CREATE_FAILED,
|
|
subcloud_dict['deploy-status'])
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, values['name'])
|
|
self.assertEqual(consts.DEPLOY_STATE_CREATE_FAILED,
|
|
updated_subcloud.deploy_status)
|
|
|
|
def test_deploy_create_secondary(self):
|
|
values = utils.create_subcloud_dict(base.SUBCLOUD_SAMPLE_DATA_0)
|
|
values['deploy_status'] = consts.DEPLOY_STATE_NONE
|
|
sysadmin_password = values['sysadmin_password']
|
|
subcloud = self.create_subcloud_static(
|
|
self.ctx, name=values['name'], region_name=values['region_name']
|
|
)
|
|
values.update({'id': subcloud.id, 'secondary': 'true',
|
|
'ansible_ssh_pass': sysadmin_password,
|
|
'ansible_become_pass': sysadmin_password,
|
|
'admin_password': sysadmin_password})
|
|
|
|
subcloud_dict = self.sm.subcloud_deploy_create(
|
|
self.ctx, subcloud.id, payload=values)
|
|
|
|
self.assertEqual(
|
|
consts.DEPLOY_STATE_SECONDARY, subcloud_dict['deploy-status']
|
|
)
|
|
|
|
def test_deploy_create_secondary_without_additional_values(self):
|
|
values = utils.create_subcloud_dict(base.SUBCLOUD_SAMPLE_DATA_0)
|
|
values['deploy_status'] = consts.DEPLOY_STATE_NONE
|
|
subcloud = self.create_subcloud_static(
|
|
self.ctx, name=values['name'], region_name=values['region_name']
|
|
)
|
|
values['id'] = subcloud.id
|
|
values['secondary'] = 'true'
|
|
|
|
subcloud_dict = self.sm.subcloud_deploy_create(
|
|
self.ctx, subcloud.id, payload=values)
|
|
|
|
self.assertEqual(
|
|
consts.DEPLOY_STATE_SECONDARY, subcloud_dict['deploy-status']
|
|
)
|
|
|
|
def test_deploy_create_secondary_failed(self):
|
|
self.mock_openstack_driver.side_effect = Exception('boom')
|
|
values = utils.create_subcloud_dict(base.SUBCLOUD_SAMPLE_DATA_0)
|
|
values['deploy_status'] = consts.DEPLOY_STATE_NONE
|
|
subcloud = self.create_subcloud_static(
|
|
self.ctx, name=values['name'], region_name=values['region_name']
|
|
)
|
|
values['id'] = subcloud.id
|
|
values['secondary'] = 'true'
|
|
subcloud_dict = self.sm.subcloud_deploy_create(
|
|
self.ctx, subcloud.id, payload=values)
|
|
self.assertEqual(
|
|
consts.DEPLOY_STATE_SECONDARY_FAILED, subcloud_dict['deploy-status']
|
|
)
|
|
|
|
@mock.patch.object(cutils, 'update_values_on_yaml_file')
|
|
def test_subcloud_deploy_bootstrap(self, mock_update_yml):
|
|
self.mock_get_playbook_for_software_version.return_value = SW_VERSION
|
|
self.mock_keyring.get_password.return_value = "testpass"
|
|
self.mock_ansible_run_playbook.return_value = False
|
|
self.fake_install_values['software_version'] = SW_VERSION
|
|
|
|
subcloud = fake_subcloud.create_fake_subcloud(
|
|
self.ctx,
|
|
name=fake_subcloud.FAKE_BOOTSTRAP_FILE_DATA["name"],
|
|
deploy_status=consts.DEPLOY_STATE_INSTALLED,
|
|
data_install=json.dumps(self.fake_install_values)
|
|
)
|
|
self.payload["bootstrap-address"] = "10.10.10.12"
|
|
|
|
self.sm.subcloud_deploy_bootstrap(self.ctx, subcloud.id, self.payload)
|
|
|
|
self.mock_ansible_run_playbook.assert_called_once()
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx,
|
|
self.payload['name'])
|
|
self.assertEqual(consts.DEPLOY_STATE_BOOTSTRAPPED,
|
|
updated_subcloud.deploy_status)
|
|
# Verify the subcloud rehomed flag is False after bootstrapped
|
|
self.assertFalse(updated_subcloud.rehomed)
|
|
|
|
def test_subcloud_deploy_bootstrap_run_playbook_failed(self):
|
|
self.mock_ansible_run_playbook.side_effect = PlaybookExecutionFailed()
|
|
|
|
subcloud = fake_subcloud.create_fake_subcloud(
|
|
self.ctx,
|
|
name=fake_subcloud.FAKE_BOOTSTRAP_FILE_DATA["name"],
|
|
deploy_status=consts.DEPLOY_STATE_INSTALLED,
|
|
data_install=json.dumps(self.fake_install_values)
|
|
)
|
|
|
|
self.sm.subcloud_deploy_bootstrap(self.ctx, subcloud.id, self.payload)
|
|
|
|
self.mock_ansible_run_playbook.assert_called_once()
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx,
|
|
self.payload['name'])
|
|
self.assertEqual(consts.DEPLOY_STATE_BOOTSTRAP_FAILED,
|
|
updated_subcloud.deploy_status)
|
|
# Verify the subcloud rehomed flag is False after bootstrapped
|
|
self.assertFalse(updated_subcloud.rehomed)
|
|
self.mock_log.error.assert_called_once_with(
|
|
'FAILED bootstrapping playbook of (fake subcloud1).\ncheck'
|
|
' individual log at /var/log/dcmanager/ansible/fake'
|
|
' subcloud1_playbook_output.log for detailed output ')
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_deploy_bootstrap_prep')
|
|
def test_subcloud_deploy_bootstrap_failed(self, mock_bootstrap_prep):
|
|
mock_bootstrap_prep.side_effect = Exception('boom')
|
|
|
|
subcloud = fake_subcloud.create_fake_subcloud(
|
|
self.ctx,
|
|
name=fake_subcloud.FAKE_BOOTSTRAP_FILE_DATA["name"],
|
|
deploy_status=consts.DEPLOY_STATE_INSTALLED)
|
|
|
|
self.sm.subcloud_deploy_bootstrap(self.ctx, subcloud.id, self.payload)
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx,
|
|
self.payload['name'])
|
|
self.assertEqual(consts.DEPLOY_STATE_PRE_BOOTSTRAP_FAILED,
|
|
updated_subcloud.deploy_status)
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_prepare_for_deployment')
|
|
def test_configure_subcloud(self, mock_prepare_for_deployment):
|
|
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_PRE_CONFIG
|
|
|
|
self.fake_payload[consts.BOOTSTRAP_ADDRESS] = \
|
|
fake_subcloud.FAKE_BOOTSTRAP_VALUE[consts.BOOTSTRAP_ADDRESS]
|
|
|
|
self.sm.subcloud_deploy_config(
|
|
self.ctx, self.subcloud.id, payload=self.fake_payload)
|
|
mock_prepare_for_deployment.assert_called_once()
|
|
self.mock_create_subcloud_inventory.assert_called_once()
|
|
|
|
def test_configure_subcloud_failed(self):
|
|
self.mock_ansible_run_playbook.side_effect = PlaybookExecutionFailed()
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_PRE_CONFIG
|
|
self.fake_payload.update({
|
|
"user_uploaded_artifacts": "fake_files",
|
|
consts.BOOTSTRAP_ADDRESS:
|
|
fake_subcloud.FAKE_BOOTSTRAP_VALUE[consts.BOOTSTRAP_ADDRESS]
|
|
})
|
|
self.sm.subcloud_deploy_config(self.ctx,
|
|
self.subcloud.id,
|
|
payload=self.fake_payload)
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx,
|
|
self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_CONFIG_FAILED,
|
|
updated_subcloud.deploy_status)
|
|
self.mock_create_subcloud_inventory.assert_called_once()
|
|
|
|
def test_configure_subcloud_pre_config_failed(self):
|
|
|
|
ret = self.sm.subcloud_deploy_config(self.ctx,
|
|
self.subcloud.id,
|
|
payload=self.fake_payload)
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx,
|
|
self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_PRE_CONFIG_FAILED,
|
|
updated_subcloud.deploy_status)
|
|
self.mock_log.exception.assert_called_once_with(
|
|
f'Failed to configure {self.subcloud.name}')
|
|
self.assertFalse(ret)
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_prepare_for_deployment')
|
|
@mock.patch.object(cutils, 'update_values_on_yaml_file')
|
|
def test_subcloud_deploy_resume(
|
|
self, mock_update_yml,
|
|
mock_prepare_for_deployment
|
|
):
|
|
self.mock_get_playbook_for_software_version.return_value = FAKE_SW_VERSION
|
|
self.mock_ansible_run_playbook.return_value = False
|
|
self.mock_run_subcloud_install.return_value = True
|
|
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_CREATED
|
|
|
|
deploy_states_to_run = [consts.DEPLOY_PHASE_INSTALL,
|
|
consts.DEPLOY_PHASE_BOOTSTRAP,
|
|
consts.DEPLOY_PHASE_CONFIG]
|
|
|
|
fake_payload_bootstrap = {**fake_subcloud.FAKE_BOOTSTRAP_VALUE,
|
|
**fake_subcloud.FAKE_BOOTSTRAP_FILE_DATA}
|
|
fake_payload_bootstrap["sysadmin_password"] = "testpass"
|
|
fake_payload = {**self.fake_payload_install,
|
|
**fake_payload_bootstrap,
|
|
**self.fake_payload}
|
|
|
|
self.sm.subcloud_deploy_resume(self.ctx, self.subcloud.id,
|
|
self.subcloud.name,
|
|
fake_payload,
|
|
deploy_states_to_run)
|
|
mock_prepare_for_deployment.assert_called_once()
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx,
|
|
self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_DONE,
|
|
updated_subcloud.deploy_status)
|
|
|
|
@mock.patch.object(AnsiblePlaybook, 'run_abort')
|
|
def test_subcloud_deploy_abort_failed(self, mock_run_abort):
|
|
self.subcloud.update(
|
|
{'name': fake_subcloud.FAKE_SUBCLOUD_DATA['name'],
|
|
'deploy_status': consts.DEPLOY_STATE_INSTALLING}
|
|
)
|
|
mock_run_abort.side_effect = base.FakeException('boom')
|
|
self.assertRaises(base.FakeException, self.sm.subcloud_deploy_abort,
|
|
self.ctx, self.subcloud.id, self.subcloud.deploy_status)
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx,
|
|
self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_INSTALL_FAILED,
|
|
updated_subcloud.deploy_status)
|
|
|
|
@mock.patch.object(AnsiblePlaybook, 'run_abort')
|
|
def test_subprocessors_terminated_before_abort(self,
|
|
mock_run_abort):
|
|
self.subcloud.update(
|
|
{'name': fake_subcloud.FAKE_SUBCLOUD_DATA['name'],
|
|
'deploy_status': consts.DEPLOY_STATE_INSTALLING}
|
|
)
|
|
mock_run_abort.return_value = False
|
|
|
|
self.sm.subcloud_deploy_abort(
|
|
self.ctx, self.subcloud.id, self.subcloud.deploy_status)
|
|
ret = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_ABORTING_INSTALL,
|
|
ret.deploy_status)
|
|
mock_run_abort.assert_called_once()
|
|
|
|
def test_subcloud_deploy_complete(self):
|
|
self.subcloud.update(
|
|
{'name': fake_subcloud.FAKE_SUBCLOUD_DATA['name'],
|
|
'deploy_status': consts.DEPLOY_STATE_INSTALLING}
|
|
)
|
|
self.sm.subcloud_deploy_complete(self.ctx, self.subcloud.id)
|
|
ret = db_api.subcloud_get_by_region_name(
|
|
self.ctx, self.subcloud.region_name
|
|
)
|
|
self.assertEqual(consts.DEPLOY_STATE_DONE, ret.deploy_status)
|
|
|
|
@mock.patch.object(cutils, 'get_region_name')
|
|
@mock.patch.object(subcloud_enrollment.SubcloudEnrollmentInit, 'prep')
|
|
def test_deploy_subcloud_enroll(
|
|
self, mock_subcloud_enrollment_prep, mock_get_region_name):
|
|
|
|
mock_run_patch_patch = mock.patch('eventlet.green.subprocess.run')
|
|
mock_mkdtemp_patch = mock.patch('tempfile.mkdtemp')
|
|
mock_makedirs_patch = mock.patch('os.makedirs')
|
|
mock_rmtree_patch = mock.patch('shutil.rmtree')
|
|
|
|
self.seed_data_dir = '/temp/seed_data'
|
|
mock_get_region_name.return_value = '11111'
|
|
|
|
self.mock_run = mock_run_patch_patch.start()
|
|
self.mock_mkdtemp = mock_mkdtemp_patch.start()
|
|
self.mock_makedirs = mock_makedirs_patch.start()
|
|
self.mock_rmtree = mock_rmtree_patch.start()
|
|
|
|
self.addCleanup(mock_run_patch_patch.stop)
|
|
self.addCleanup(mock_mkdtemp_patch.stop)
|
|
self.addCleanup(mock_makedirs_patch.stop)
|
|
self.addCleanup(mock_rmtree_patch.stop)
|
|
|
|
self._mock_builtins_open()
|
|
|
|
self.mock_builtins_open.side_effect = mock.mock_open()
|
|
self.mock_os_path_exists.return_value = True
|
|
self.mock_mkdtemp.return_value = self.seed_data_dir
|
|
self.mock_os_path_isdir.return_value = True
|
|
self.mock_run.return_value = mock.MagicMock(returncode=0,
|
|
stdout=b'Success')
|
|
|
|
self._mock_subcloud_manager_compose_enroll_command()
|
|
self.fake_payload_enroll['software_version'] = FAKE_PREVIOUS_SW_VERSION
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_PRE_INIT_ENROLL
|
|
self.fake_payload_enroll['software_version'] = SW_VERSION
|
|
|
|
with mock.patch('os.path.isdir', side_effect=self.patched_isdir):
|
|
self.sm.subcloud_deploy_enroll(
|
|
self.ctx, self.subcloud.id, payload=self.fake_payload_enroll)
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx,
|
|
self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_ENROLLED,
|
|
updated_subcloud.deploy_status)
|
|
|
|
@mock.patch.object(subcloud_enrollment.SubcloudEnrollmentInit,
|
|
'prep')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_deploy_install_prep')
|
|
def test_subcloud_deploy_pre_init_enroll_failed(
|
|
self, mock_deploy_install_prep, mock_subcloud_enrollment_prep):
|
|
|
|
mock_deploy_install_prep.side_effect = base.FakeException('boom')
|
|
|
|
subcloud = fake_subcloud.create_fake_subcloud(
|
|
self.ctx,
|
|
name=fake_subcloud.FAKE_BOOTSTRAP_FILE_DATA["name"],
|
|
deploy_status=consts.DEPLOY_STATE_CREATED,
|
|
data_install=json.dumps(self.fake_payload_enroll['install_values'])
|
|
)
|
|
|
|
self.sm.subcloud_init_enroll(
|
|
self.ctx, subcloud.id, self.fake_payload_enroll)
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx,
|
|
self.payload['name'])
|
|
|
|
self.assertEqual(consts.DEPLOY_STATE_PRE_INIT_ENROLL_FAILED,
|
|
updated_subcloud.deploy_status)
|
|
|
|
@mock.patch.object(subcloud_enrollment.SubcloudEnrollmentInit,
|
|
'prep')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_subcloud_endpoints')
|
|
def test_subcloud_deploy_enroll_failed(
|
|
self, mock_create_subcloud_endpoints, mock_subcloud_enrollment_prep):
|
|
mock_create_subcloud_endpoints.side_effect = base.FakeException('boom')
|
|
|
|
subcloud = fake_subcloud.create_fake_subcloud(
|
|
self.ctx,
|
|
name=fake_subcloud.FAKE_BOOTSTRAP_FILE_DATA["name"],
|
|
deploy_status=consts.DEPLOY_STATE_INIT_ENROLL_COMPLETE,
|
|
data_install=json.dumps(self.fake_payload_enroll['install_values'])
|
|
)
|
|
|
|
self.sm.subcloud_deploy_enroll(
|
|
self.ctx, subcloud.id, self.fake_payload_enroll)
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx,
|
|
self.payload['name'])
|
|
|
|
self.assertEqual(consts.DEPLOY_STATE_PRE_ENROLL_FAILED,
|
|
updated_subcloud.deploy_status)
|
|
|
|
@mock.patch.object(subcloud_enrollment.SubcloudEnrollmentInit,
|
|
'prep')
|
|
@mock.patch.object(subcloud_enrollment.SubcloudEnrollmentInit,
|
|
'enroll_init')
|
|
@mock.patch.object(cutils, 'get_region_name')
|
|
def test_subcloud_deploy_enroll_run_playbook_failed(
|
|
self, mock_get_region_name,
|
|
mock_subcloud_enrollment_prep,
|
|
mock_subcloud_enrollment_enroll_init):
|
|
|
|
self.mock_ansible_run_playbook.side_effect = PlaybookExecutionFailed()
|
|
mock_get_region_name.return_value = "11111"
|
|
|
|
subcloud = fake_subcloud.create_fake_subcloud(
|
|
self.ctx,
|
|
name=fake_subcloud.FAKE_BOOTSTRAP_FILE_DATA["name"],
|
|
deploy_status=consts.DEPLOY_STATE_PRE_ENROLL_COMPLETE,
|
|
data_install=json.dumps(self.fake_payload_enroll['install_values'])
|
|
)
|
|
|
|
self.sm.subcloud_deploy_enroll(
|
|
self.ctx, subcloud.id, self.fake_payload_enroll)
|
|
|
|
self.mock_ansible_run_playbook.assert_called_once()
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx,
|
|
self.payload['name'])
|
|
self.assertEqual(consts.DEPLOY_STATE_ENROLL_FAILED,
|
|
updated_subcloud.deploy_status)
|
|
# Verify the subcloud rehomed flag is False after bootstrapped
|
|
self.assertFalse(updated_subcloud.rehomed)
|
|
self.mock_log.error.assert_called_once_with(
|
|
'Enroll failed for subcloud fake subcloud1: '
|
|
'FAILED enrolling playbook of (fake subcloud1).'
|
|
'\ncheck individual log at '
|
|
'/var/log/dcmanager/ansible/fake '
|
|
'subcloud1_playbook_output.log for detailed output ')
|
|
|
|
|
|
class TestSubcloudAdd(BaseTestSubcloudManager):
|
|
"""Test class for testing subcloud add"""
|
|
|
|
def setUp(self):
|
|
super().setUp()
|
|
self._mock_subcloud_manager_write_subcloud_ansible_config()
|
|
self._mock_subcloud_manager_create_addn_hosts_dc()
|
|
self._mock_subcloud_manager_create_intermediate_ca_cert()
|
|
self.mock_openstack_driver().keystone_client = FakeKeystoneClient()
|
|
self._mock_subcloud_manager_get_cached_regionone_data()
|
|
self.mock_get_cached_regionone_data.return_value = \
|
|
FAKE_CACHED_REGIONONE_DATA
|
|
|
|
@mock.patch.object(cutils, 'get_oam_addresses')
|
|
@mock.patch.object(subcloud_install.SubcloudInstall, 'prep')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_write_deploy_files')
|
|
@mock.patch.object(cutils, 'update_values_on_yaml_file')
|
|
def test_add_subcloud(
|
|
self, mock_update_yml, mock_write_deploy_files,
|
|
mock_install_prep, mock_oam_address
|
|
):
|
|
# Prepare the payload
|
|
install_values = copy.copy(fake_subcloud.FAKE_SUBCLOUD_INSTALL_VALUES)
|
|
install_values['software_version'] = SW_VERSION
|
|
payload = {**fake_subcloud.FAKE_BOOTSTRAP_VALUE,
|
|
**fake_subcloud.FAKE_BOOTSTRAP_FILE_DATA,
|
|
"sysadmin_password": "testpass",
|
|
'bmc_password': 'bmc_pass',
|
|
'install_values': install_values,
|
|
'software_version': FAKE_PREVIOUS_SW_VERSION,
|
|
"deploy_playbook": "test_playbook.yaml",
|
|
"deploy_overrides": "test_overrides.yaml",
|
|
"deploy_chart": "test_chart.yaml",
|
|
"deploy_config": "subcloud1.yaml",
|
|
"user_uploaded_artifacts": True}
|
|
|
|
# Create subcloud in DB
|
|
subcloud = self.create_subcloud_static(self.ctx, name=payload['name'])
|
|
payload['region_name'] = subcloud.region_name
|
|
|
|
# Mock return values
|
|
self.mock_get_playbook_for_software_version.return_value = SW_VERSION
|
|
self.mock_keyring.get_password.return_value = payload['sysadmin_password']
|
|
self.mock_ansible_run_playbook.return_value = False
|
|
|
|
# Call the add method
|
|
self.sm.add_subcloud(self.ctx, subcloud.id, payload)
|
|
|
|
# Verify results
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_DONE,
|
|
updated_subcloud.deploy_status)
|
|
|
|
mock_write_deploy_files.assert_called()
|
|
self.mock_keyring.get_password.assert_called()
|
|
mock_update_yml.assert_called()
|
|
self.mock_create_subcloud_inventory.assert_called()
|
|
self.mock_get_playbook_for_software_version.assert_called_once()
|
|
self.assertEqual(self.mock_ansible_run_playbook.call_count, 3)
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'compose_rehome_command')
|
|
def test_add_subcloud_with_migration_option(
|
|
self,
|
|
mock_compose_rehome_command
|
|
):
|
|
values = utils.create_subcloud_dict(base.SUBCLOUD_SAMPLE_DATA_0)
|
|
values['deploy_status'] = consts.DEPLOY_STATE_NONE
|
|
values['migrate'] = 'true'
|
|
sysadmin_password = values['sysadmin_password']
|
|
|
|
# dcmanager add_subcloud queries the data from the db
|
|
subcloud = self.create_subcloud_static(self.ctx, name=values['name'],
|
|
region_name=values['region_name'])
|
|
|
|
self.mock_keyring.get_password.return_value = sysadmin_password
|
|
|
|
self.sm.add_subcloud(self.ctx, subcloud.id, payload=values)
|
|
|
|
self.mock_get_cached_regionone_data.assert_called_once()
|
|
self.mock_sysinv_client().create_route.assert_called()
|
|
self.mock_dcorch_api().add_subcloud.assert_called_once()
|
|
self.mock_create_addn_hosts.assert_called_once()
|
|
self.mock_create_subcloud_inventory.assert_called_once()
|
|
self.mock_write_subcloud_ansible_config.assert_called_once()
|
|
self.mock_create_intermediate_ca_cert.assert_called_once()
|
|
mock_compose_rehome_command.assert_called_once_with(
|
|
values['name'],
|
|
values['region_name'],
|
|
self.sm._get_ansible_filename(values['name'],
|
|
consts.INVENTORY_FILE_POSTFIX),
|
|
subcloud['software_version'])
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, values['name'])
|
|
self.assertEqual(consts.DEPLOY_STATE_DONE,
|
|
updated_subcloud.deploy_status)
|
|
# Verify subcloud rehomed flag is true
|
|
self.assertTrue(updated_subcloud.rehomed)
|
|
|
|
# Verify that the password fields are present
|
|
written_payload = self.mock_write_subcloud_ansible_config.call_args.args[1]
|
|
expected_subset = {'ansible_become_pass': sysadmin_password,
|
|
'ansible_ssh_pass': sysadmin_password,
|
|
'admin_password': sysadmin_password}
|
|
# Check that expected_subset is a subset of written_payload
|
|
self.assertTrue(expected_subset.items() <= written_payload.items())
|
|
|
|
def test_add_subcloud_create_failed(self):
|
|
values = utils.create_subcloud_dict(base.SUBCLOUD_SAMPLE_DATA_0)
|
|
services = FAKE_SERVICES
|
|
|
|
# dcmanager add_subcloud queries the data from the db
|
|
subcloud = self.create_subcloud_static(self.ctx, name=values['name'],
|
|
region_name=values['region_name'])
|
|
|
|
self.mock_dcorch_api().add_subcloud.side_effect = Exception('boom')
|
|
self.mock_openstack_driver().services_list = services
|
|
|
|
self.sm.add_subcloud(self.ctx, subcloud.id, payload=values)
|
|
self.mock_get_cached_regionone_data.assert_called_once()
|
|
self.mock_sysinv_client().create_route.assert_called()
|
|
|
|
# Verify subcloud was updated with correct values
|
|
subcloud = db_api.subcloud_get_by_name(self.ctx, values['name'])
|
|
self.assertEqual(consts.DEPLOY_STATE_CREATE_FAILED,
|
|
subcloud.deploy_status)
|
|
|
|
def test_add_subcloud_with_migrate_option_prep_failed(self):
|
|
values = utils.create_subcloud_dict(base.SUBCLOUD_SAMPLE_DATA_0)
|
|
values['migrate'] = 'true'
|
|
services = FAKE_SERVICES
|
|
|
|
# dcmanager add_subcloud queries the data from the db
|
|
subcloud = self.create_subcloud_static(self.ctx, name=values['name'],
|
|
region_name=values['region_name'])
|
|
|
|
self.mock_dcorch_api().add_subcloud.side_effect = Exception('boom')
|
|
self.mock_openstack_driver().services_list = services
|
|
self.mock_keyring.get_password.return_vaue = "testpass"
|
|
|
|
self.sm.add_subcloud(self.ctx, subcloud.id, payload=values)
|
|
self.mock_get_cached_regionone_data.assert_called_once()
|
|
self.mock_sysinv_client().create_route.assert_called()
|
|
|
|
# Verify subcloud was updated with correct values
|
|
subcloud = db_api.subcloud_get_by_name(self.ctx, values['name'])
|
|
self.assertEqual(consts.DEPLOY_STATE_REHOME_PREP_FAILED,
|
|
subcloud.deploy_status)
|
|
|
|
|
|
class TestSubcloudDelete(BaseTestSubcloudManager):
|
|
"""Test class for testing subcloud delete"""
|
|
|
|
def setUp(self):
|
|
super().setUp()
|
|
self.mock_openstack_driver().keystone_client = FakeKeystoneClient()
|
|
self._mock_subcloud_manager_get_cached_regionone_data()
|
|
self.mock_get_cached_regionone_data.return_value = \
|
|
FAKE_CACHED_REGIONONE_DATA
|
|
|
|
@mock.patch.object(kubeoperator, 'KubeOperator')
|
|
def test_delete_subcloud(self, mock_kubeoperator):
|
|
self._mock_builtins_open()
|
|
self._mock_subcloud_manager_create_addn_hosts_dc()
|
|
self.sm.delete_subcloud(self.ctx, subcloud_id=self.subcloud.id)
|
|
self.mock_get_cached_regionone_data.assert_called_once()
|
|
self.mock_sysinv_client().delete_route.assert_called()
|
|
self.mock_create_addn_hosts.assert_called_once()
|
|
mock_kubeoperator().delete_cert_manager_certificate.assert_called_once()
|
|
|
|
# Verify subcloud was deleted
|
|
self.assertRaises(exceptions.SubcloudNameNotFound,
|
|
db_api.subcloud_get_by_name,
|
|
self.ctx,
|
|
self.subcloud.name)
|
|
|
|
def test_delete_subcloud_failed(self):
|
|
# Semantic checking
|
|
db_api.subcloud_update(
|
|
self.ctx, self.subcloud.id, management_state='managed'
|
|
)
|
|
self.assertRaises(exceptions.SubcloudNotUnmanaged,
|
|
self.sm.delete_subcloud,
|
|
self.ctx,
|
|
self.subcloud.id)
|
|
subcloud2 = self.create_subcloud_static(self.ctx, name='subcloud2')
|
|
db_api.subcloud_update(self.ctx, subcloud2.id, availability_status='online')
|
|
self.assertRaises(exceptions.SubcloudNotOffline,
|
|
self.sm.delete_subcloud,
|
|
self.ctx,
|
|
subcloud2.id)
|
|
|
|
@mock.patch('shutil.rmtree')
|
|
def test_cleanup_ansible_files(self, mock_rmtree):
|
|
self.sm._cleanup_ansible_files('subcloud1')
|
|
|
|
files = ('subcloud1.yml',
|
|
'subcloud1_deploy_values.yml',
|
|
'subcloud1_deploy_config.yml')
|
|
|
|
calls = []
|
|
for f in files:
|
|
filepath = os.path.join(ANS_PATH, f)
|
|
calls.append(mock.call(filepath))
|
|
|
|
install_dir = os.path.join(ANS_PATH,
|
|
'subcloud1')
|
|
|
|
self.mock_os_remove.assert_has_calls(calls, any_order=True)
|
|
mock_rmtree.assert_called_with(install_dir)
|
|
|
|
@mock.patch('shutil.rmtree')
|
|
def test_cleanup_ansible_files_exception(self, mock_rmtree):
|
|
self.mock_os_remove.side_effect = FileNotFoundError()
|
|
self.sm._cleanup_ansible_files('subcloud1')
|
|
self.mock_log.exception.assert_called_once_with(
|
|
"Unable to cleanup subcloud ansible files for subcloud: subcloud1")
|
|
|
|
|
|
class TestSubcloudUpdate(BaseTestSubcloudManager):
|
|
"""Test class for testing subcloud update"""
|
|
|
|
def setUp(self):
|
|
super().setUp()
|
|
self.fake_bootstrap_values = \
|
|
'{"name": "TestSubcloud", "system_mode": "simplex"}'
|
|
self._test_values()
|
|
|
|
def _test_values(self):
|
|
values = utils.create_subcloud_dict(base.SUBCLOUD_SAMPLE_DATA_0)
|
|
values['management_gateway_ip'] = "192.168.101.1"
|
|
values['management_start_ip'] = "192.168.101.2"
|
|
values['system_controller_network'] = 'fdff:719a:bf60:233::'
|
|
values['system_controller_network_prefix'] = 64
|
|
values['dc_root_ca_cert'] = 'fake_dc_root_ca_cert'
|
|
values['sc_ca_cert'] = 'fake_sc_ca_cert'
|
|
values['sc_ca_key'] = 'fake_sc_ca_key'
|
|
return values
|
|
|
|
def test_update_subcloud(self):
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_DONE,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE
|
|
)
|
|
self.sm.update_subcloud(self.ctx,
|
|
self.subcloud.id,
|
|
management_state=dccommon_consts.
|
|
MANAGEMENT_MANAGED,
|
|
description="subcloud new description",
|
|
location="subcloud new location")
|
|
|
|
self.mock_dcmanager_api().subcloud_managed.assert_called_once_with(
|
|
self.ctx, self.subcloud.region_name)
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(
|
|
self.ctx, self.subcloud.name)
|
|
self.assertEqual(dccommon_consts.MANAGEMENT_MANAGED,
|
|
updated_subcloud.management_state)
|
|
self.assertEqual("subcloud new description",
|
|
updated_subcloud.description)
|
|
self.assertEqual("subcloud new location",
|
|
updated_subcloud.location)
|
|
|
|
def test_update_subcloud_managed_rehome_pending(self):
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_DONE
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE,
|
|
deploy_status=consts.DEPLOY_STATE_REHOME_PENDING
|
|
)
|
|
|
|
self.sm.update_subcloud(self.ctx,
|
|
self.subcloud.id,
|
|
management_state=dccommon_consts.
|
|
MANAGEMENT_MANAGED)
|
|
|
|
self.mock_dcmanager_api().subcloud_managed.assert_called_once_with(
|
|
self.ctx, self.subcloud.region_name)
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_DONE,
|
|
updated_subcloud.deploy_status)
|
|
|
|
def test_update_subcloud_bootstrap_address_failed(self):
|
|
"""Failed updating bootstrap-address because bootstrap_values are missing."""
|
|
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_DONE
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE)
|
|
fake_bootstrap_address = '10.10.20.12'
|
|
self.assertRaisesRegex(
|
|
exceptions.BadRequest,
|
|
"Cannot update bootstrap_address"
|
|
" into rehome data, need to import bootstrap_values first",
|
|
self.sm.update_subcloud,
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
bootstrap_address=fake_bootstrap_address)
|
|
|
|
def test_update_subcloud_deploy_state_secondary(self):
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_DONE
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE,
|
|
management_state=dccommon_consts.MANAGEMENT_UNMANAGED
|
|
)
|
|
self.sm.update_subcloud(self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_SECONDARY)
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(
|
|
self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_SECONDARY,
|
|
updated_subcloud.deploy_status)
|
|
|
|
def test_update_subcloud_bootstrap_values(self):
|
|
fake_result = f'{{"saved_payload": {self.fake_bootstrap_values}}}'
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_DONE
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE)
|
|
self.sm.update_subcloud(self.ctx,
|
|
self.subcloud.id,
|
|
bootstrap_values=self.fake_bootstrap_values)
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(fake_result,
|
|
updated_subcloud.rehome_data)
|
|
|
|
def test_update_subcloud_bootstrap_address(self):
|
|
fake_result = ('{"saved_payload": {"name": "TestSubcloud", '
|
|
'"system_mode": "simplex", '
|
|
'"bootstrap-address": "123.123.123.123"}}')
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_DONE
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE)
|
|
self.sm.update_subcloud(self.ctx,
|
|
self.subcloud.id,
|
|
bootstrap_values=self.fake_bootstrap_values,
|
|
bootstrap_address="123.123.123.123")
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(fake_result,
|
|
updated_subcloud.rehome_data)
|
|
|
|
def test_update_subcloud_management_state_failed(self):
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_REHOME_PENDING,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE
|
|
)
|
|
self.mock_dcorch_api().update_subcloud_states.\
|
|
side_effect = Exception('boom')
|
|
|
|
self.sm.update_subcloud(self.ctx,
|
|
self.subcloud.id,
|
|
management_state=dccommon_consts.MANAGEMENT_MANAGED)
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(dccommon_consts.MANAGEMENT_UNMANAGED,
|
|
updated_subcloud.management_state)
|
|
self.assertEqual(consts.DEPLOY_STATE_REHOME_PENDING,
|
|
updated_subcloud.deploy_status)
|
|
|
|
def test_update_subcloud_deploy_state_rehome_pending(self):
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_DONE
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE,
|
|
management_state=dccommon_consts.MANAGEMENT_MANAGED
|
|
)
|
|
for state in (
|
|
consts.DEPLOY_STATE_REHOME_PENDING, consts.DEPLOY_STATE_SECONDARY
|
|
):
|
|
self.assertRaisesRegex(
|
|
exceptions.BadRequest, f"Unable to update deploy_status "
|
|
f"of subcloud {self.subcloud.name} to {state}",
|
|
self.sm.update_subcloud, self.ctx,
|
|
self.subcloud.id, deploy_status=state)
|
|
|
|
def test_update_subcloud_manage_with_invalid_deploy_status_failed(self):
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_DONE,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE
|
|
)
|
|
for state in (
|
|
consts.DEPLOY_STATE_REHOME_PENDING, consts.DEPLOY_STATE_SECONDARY
|
|
):
|
|
self.assertRaisesRegex(
|
|
exceptions.BadRequest,
|
|
f"Unable to manage {self.subcloud.name} while also updating "
|
|
f"its deploy_status to {state}: not allowed",
|
|
self.sm.update_subcloud,
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
management_state=dccommon_consts.MANAGEMENT_MANAGED,
|
|
deploy_status=state)
|
|
|
|
def test_update_subcloud_rehome_data(self):
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_DONE
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE
|
|
)
|
|
fake_result = ('{"saved_payload": {"name": "TestSubcloud", '
|
|
'"system_mode": "simplex", '
|
|
'"admin_password": "dGVzdHBhc3M=", '
|
|
'"bootstrap-address": "123.123.123.123"}}')
|
|
fake_bootstrap_values = '{"name": "TestSubcloud", \
|
|
"system_mode": "simplex", "sysadmin_password": "testpass", \
|
|
"ansible_ssh_pass": "fakepass", "ansible_become_pass": "fakepass",\
|
|
"admin_password": "testpass"}'
|
|
fake_bootstrap_address = '123.123.123.123'
|
|
|
|
self.sm.update_subcloud(self.ctx,
|
|
self.subcloud.id,
|
|
bootstrap_values=fake_bootstrap_values,
|
|
bootstrap_address=fake_bootstrap_address)
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(fake_result,
|
|
updated_subcloud.rehome_data)
|
|
|
|
def test_update_subcloud_with_install_values(self):
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_DONE,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE
|
|
)
|
|
self.sm.update_subcloud(self.ctx,
|
|
self.subcloud.id,
|
|
management_state=dccommon_consts.
|
|
MANAGEMENT_MANAGED,
|
|
description="subcloud new description",
|
|
location="subcloud new location",
|
|
data_install="install values")
|
|
|
|
self.mock_dcmanager_api().subcloud_managed.assert_called_once_with(
|
|
self.ctx, self.subcloud.region_name)
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(
|
|
self.ctx, self.subcloud.name)
|
|
self.assertEqual(dccommon_consts.MANAGEMENT_MANAGED,
|
|
updated_subcloud.management_state)
|
|
self.assertEqual("subcloud new description",
|
|
updated_subcloud.description)
|
|
self.assertEqual("subcloud new location",
|
|
updated_subcloud.location)
|
|
self.assertEqual("install values",
|
|
updated_subcloud.data_install)
|
|
|
|
def test_update_already_managed_subcloud(self):
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_DONE
|
|
db_api.subcloud_update(self.ctx,
|
|
self.subcloud.id,
|
|
management_state=dccommon_consts.MANAGEMENT_MANAGED)
|
|
|
|
self.assertRaisesRegex(
|
|
exceptions.BadRequest,
|
|
f"Subcloud {self.subcloud.name} already managed",
|
|
self.sm.update_subcloud, self.ctx,
|
|
self.subcloud.id,
|
|
management_state=dccommon_consts.MANAGEMENT_MANAGED)
|
|
self.mock_audit_rpc_client().trigger_subcloud_audits.assert_not_called()
|
|
|
|
def test_update_already_unmanaged_subcloud(self):
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_DONE
|
|
self.assertRaisesRegex(
|
|
exceptions.BadRequest,
|
|
f"Subcloud {self.subcloud.name} already unmanaged",
|
|
self.sm.update_subcloud, self.ctx,
|
|
self.subcloud.id,
|
|
management_state=dccommon_consts.MANAGEMENT_UNMANAGED)
|
|
self.mock_audit_rpc_client().trigger_subcloud_audits.assert_not_called()
|
|
|
|
def test_update_management_unmanaged(self):
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_DONE
|
|
db_api.subcloud_update(self.ctx, self.subcloud.id,
|
|
management_state=dccommon_consts.
|
|
MANAGEMENT_MANAGED)
|
|
self.sm.update_subcloud(self.ctx,
|
|
self.subcloud.id,
|
|
management_state=dccommon_consts.
|
|
MANAGEMENT_UNMANAGED)
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(dccommon_consts.MANAGEMENT_UNMANAGED,
|
|
updated_subcloud.management_state)
|
|
|
|
def test_manage_when_deploy_status_failed(self):
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_DEPLOY_FAILED
|
|
self.assertRaisesRegex(
|
|
exceptions.BadRequest,
|
|
f"Unable to manage {self.subcloud.name}: "
|
|
"its deploy_status must be either"
|
|
f" '{consts.DEPLOY_STATE_DONE}' or "
|
|
f"'{consts.DEPLOY_STATE_REHOME_PENDING}'",
|
|
self.sm.update_subcloud, self.ctx,
|
|
self.subcloud.id,
|
|
management_state=dccommon_consts.MANAGEMENT_MANAGED)
|
|
|
|
def test_manage_when_offline_without_force(self):
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_DONE,
|
|
availability_status=dccommon_consts.AVAILABILITY_OFFLINE)
|
|
self.assertRaises(exceptions.SubcloudNotOnline,
|
|
self.sm.update_subcloud, self.ctx,
|
|
self.subcloud.id,
|
|
management_state=dccommon_consts.MANAGEMENT_MANAGED)
|
|
|
|
def test_manage_when_offline_with_force(self):
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_DONE,
|
|
availability_status=dccommon_consts.AVAILABILITY_OFFLINE)
|
|
|
|
self.sm.update_subcloud(self.ctx,
|
|
self.subcloud.id,
|
|
management_state=dccommon_consts.
|
|
MANAGEMENT_MANAGED,
|
|
description="subcloud new description",
|
|
location="subcloud new location",
|
|
data_install="install values",
|
|
force=True)
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(dccommon_consts.MANAGEMENT_MANAGED,
|
|
updated_subcloud.management_state)
|
|
self.assertEqual("subcloud new description",
|
|
updated_subcloud.description)
|
|
self.assertEqual("subcloud new location",
|
|
updated_subcloud.location)
|
|
self.assertEqual("install values",
|
|
updated_subcloud.data_install)
|
|
|
|
def test_update_subcloud_group_id(self):
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_DONE,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE)
|
|
|
|
self.sm.update_subcloud(self.ctx,
|
|
self.subcloud.id,
|
|
management_state=dccommon_consts.
|
|
MANAGEMENT_MANAGED,
|
|
description="subcloud new description",
|
|
location="subcloud new location",
|
|
group_id=2)
|
|
self.mock_dcmanager_api().subcloud_managed.assert_called_once_with(
|
|
self.ctx, self.subcloud.region_name)
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(dccommon_consts.MANAGEMENT_MANAGED,
|
|
updated_subcloud.management_state)
|
|
self.assertEqual("subcloud new description",
|
|
updated_subcloud.description)
|
|
self.assertEqual("subcloud new location",
|
|
updated_subcloud.location)
|
|
self.assertEqual(2,
|
|
updated_subcloud.group_id)
|
|
|
|
def test_update_subcloud_endpoint_status(self):
|
|
self.assertIsNotNone(self.subcloud)
|
|
self.assertEqual(self.subcloud.management_state,
|
|
dccommon_consts.MANAGEMENT_UNMANAGED)
|
|
self.assertEqual(self.subcloud.availability_status,
|
|
dccommon_consts.AVAILABILITY_OFFLINE)
|
|
|
|
# create sync statuses for endpoints
|
|
for endpoint in [dccommon_consts.ENDPOINT_TYPE_PLATFORM,
|
|
dccommon_consts.ENDPOINT_TYPE_IDENTITY,
|
|
dccommon_consts.ENDPOINT_TYPE_PATCHING,
|
|
dccommon_consts.ENDPOINT_TYPE_FM,
|
|
dccommon_consts.ENDPOINT_TYPE_NFV,
|
|
dccommon_consts.ENDPOINT_TYPE_DC_CERT]:
|
|
status = db_api.subcloud_status_create(
|
|
self.ctx, self.subcloud.id, endpoint)
|
|
self.assertIsNotNone(status)
|
|
self.assertEqual(status.sync_status, dccommon_consts.SYNC_STATUS_UNKNOWN)
|
|
|
|
# Update/verify each status with the default sync state: out-of-sync
|
|
ssm = subcloud_state_manager.SubcloudStateManager()
|
|
for endpoint in [dccommon_consts.ENDPOINT_TYPE_PLATFORM,
|
|
dccommon_consts.ENDPOINT_TYPE_IDENTITY,
|
|
dccommon_consts.ENDPOINT_TYPE_PATCHING,
|
|
dccommon_consts.ENDPOINT_TYPE_FM,
|
|
dccommon_consts.ENDPOINT_TYPE_NFV,
|
|
dccommon_consts.ENDPOINT_TYPE_DC_CERT]:
|
|
# Update
|
|
ssm.update_subcloud_endpoint_status(
|
|
self.ctx, subcloud_region=self.subcloud.region_name,
|
|
endpoint_type=endpoint)
|
|
|
|
# Verify
|
|
updated_subcloud_status = db_api.subcloud_status_get(
|
|
self.ctx, self.subcloud.id, endpoint)
|
|
self.assertIsNotNone(updated_subcloud_status)
|
|
self.assertEqual(updated_subcloud_status.sync_status,
|
|
dccommon_consts.SYNC_STATUS_OUT_OF_SYNC)
|
|
|
|
# Attempt to update each status to be in-sync for an offline/unmanaged
|
|
# subcloud. This is not allowed. Verify no change.
|
|
for endpoint in [dccommon_consts.ENDPOINT_TYPE_PLATFORM,
|
|
dccommon_consts.ENDPOINT_TYPE_IDENTITY,
|
|
dccommon_consts.ENDPOINT_TYPE_PATCHING,
|
|
dccommon_consts.ENDPOINT_TYPE_FM,
|
|
dccommon_consts.ENDPOINT_TYPE_NFV,
|
|
dccommon_consts.ENDPOINT_TYPE_DC_CERT]:
|
|
ssm.update_subcloud_endpoint_status(
|
|
self.ctx, subcloud_region=self.subcloud.region_name,
|
|
endpoint_type=endpoint,
|
|
sync_status=dccommon_consts.SYNC_STATUS_IN_SYNC)
|
|
|
|
updated_subcloud_status = db_api.subcloud_status_get(
|
|
self.ctx, self.subcloud.id, endpoint)
|
|
self.assertIsNotNone(updated_subcloud_status)
|
|
# No change in status: Only online/managed clouds are updated
|
|
self.assertEqual(updated_subcloud_status.sync_status,
|
|
dccommon_consts.SYNC_STATUS_OUT_OF_SYNC)
|
|
|
|
# Attempt to update each status to be unknown for an offline/unmanaged
|
|
# subcloud. This is allowed.
|
|
ssm.update_subcloud_endpoint_status(
|
|
self.ctx, subcloud_region=self.subcloud.region_name,
|
|
endpoint_type=None,
|
|
sync_status=dccommon_consts.SYNC_STATUS_UNKNOWN)
|
|
|
|
for endpoint in [dccommon_consts.ENDPOINT_TYPE_PLATFORM,
|
|
dccommon_consts.ENDPOINT_TYPE_IDENTITY,
|
|
dccommon_consts.ENDPOINT_TYPE_PATCHING,
|
|
dccommon_consts.ENDPOINT_TYPE_FM,
|
|
dccommon_consts.ENDPOINT_TYPE_NFV,
|
|
dccommon_consts.ENDPOINT_TYPE_DC_CERT]:
|
|
updated_subcloud_status = db_api.subcloud_status_get(
|
|
self.ctx, self.subcloud.id, endpoint)
|
|
self.assertIsNotNone(updated_subcloud_status)
|
|
self.assertEqual(updated_subcloud_status.sync_status,
|
|
dccommon_consts.SYNC_STATUS_UNKNOWN)
|
|
|
|
# Attempt to update each status to be out-of-sync for an
|
|
# offline/unmanaged subcloud. Exclude one endpoint. This is allowed.
|
|
ssm.update_subcloud_endpoint_status(
|
|
self.ctx, subcloud_region=self.subcloud.region_name,
|
|
endpoint_type=None,
|
|
sync_status=dccommon_consts.SYNC_STATUS_OUT_OF_SYNC,
|
|
ignore_endpoints=[dccommon_consts.ENDPOINT_TYPE_DC_CERT])
|
|
|
|
for endpoint in [dccommon_consts.ENDPOINT_TYPE_PLATFORM,
|
|
dccommon_consts.ENDPOINT_TYPE_IDENTITY,
|
|
dccommon_consts.ENDPOINT_TYPE_PATCHING,
|
|
dccommon_consts.ENDPOINT_TYPE_FM,
|
|
dccommon_consts.ENDPOINT_TYPE_NFV]:
|
|
updated_subcloud_status = db_api.subcloud_status_get(
|
|
self.ctx, self.subcloud.id, endpoint)
|
|
self.assertIsNotNone(updated_subcloud_status)
|
|
self.assertEqual(updated_subcloud_status.sync_status,
|
|
dccommon_consts.SYNC_STATUS_OUT_OF_SYNC)
|
|
# Verify the dc-sync endpoint did not change
|
|
endpoint = dccommon_consts.ENDPOINT_TYPE_DC_CERT
|
|
updated_subcloud_status = db_api.subcloud_status_get(
|
|
self.ctx, self.subcloud.id, endpoint)
|
|
self.assertIsNotNone(updated_subcloud_status)
|
|
self.assertEqual(updated_subcloud_status.sync_status,
|
|
dccommon_consts.SYNC_STATUS_UNKNOWN)
|
|
|
|
# Set/verify the subcloud is online/unmanaged
|
|
db_api.subcloud_update(
|
|
self.ctx, self.subcloud.id,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE)
|
|
subcloud = db_api.subcloud_get(self.ctx, self.subcloud.id)
|
|
self.assertIsNotNone(subcloud)
|
|
self.assertEqual(subcloud.management_state,
|
|
dccommon_consts.MANAGEMENT_UNMANAGED)
|
|
self.assertEqual(subcloud.availability_status,
|
|
dccommon_consts.AVAILABILITY_ONLINE)
|
|
|
|
# Attempt to update each status to be in-sync for an online/unmanaged
|
|
# subcloud. This is not allowed. Verify no change.
|
|
for endpoint in [dccommon_consts.ENDPOINT_TYPE_PLATFORM,
|
|
dccommon_consts.ENDPOINT_TYPE_IDENTITY,
|
|
dccommon_consts.ENDPOINT_TYPE_PATCHING,
|
|
dccommon_consts.ENDPOINT_TYPE_FM,
|
|
dccommon_consts.ENDPOINT_TYPE_NFV]:
|
|
ssm.update_subcloud_endpoint_status(
|
|
self.ctx, subcloud_region=subcloud.region_name,
|
|
endpoint_type=endpoint,
|
|
sync_status=dccommon_consts.SYNC_STATUS_IN_SYNC)
|
|
|
|
updated_subcloud_status = db_api.subcloud_status_get(
|
|
self.ctx, subcloud.id, endpoint)
|
|
self.assertIsNotNone(updated_subcloud_status)
|
|
# No change in status: Only online/managed clouds are updated
|
|
self.assertEqual(updated_subcloud_status.sync_status,
|
|
dccommon_consts.SYNC_STATUS_OUT_OF_SYNC)
|
|
|
|
# Attempt to update dc-cert status to be in-sync for an
|
|
# online/unmanaged subcloud. This is allowed. Verify the change.
|
|
endpoint = dccommon_consts.ENDPOINT_TYPE_DC_CERT
|
|
ssm.update_subcloud_endpoint_status(
|
|
self.ctx, subcloud_region=subcloud.region_name,
|
|
endpoint_type=endpoint,
|
|
sync_status=dccommon_consts.SYNC_STATUS_IN_SYNC)
|
|
|
|
updated_subcloud_status = db_api.subcloud_status_get(
|
|
self.ctx, subcloud.id, endpoint)
|
|
self.assertIsNotNone(updated_subcloud_status)
|
|
self.assertEqual(updated_subcloud_status.sync_status,
|
|
dccommon_consts.SYNC_STATUS_IN_SYNC)
|
|
|
|
# Set/verify the subcloud is online/managed
|
|
db_api.subcloud_update(
|
|
self.ctx, subcloud.id,
|
|
management_state=dccommon_consts.MANAGEMENT_MANAGED)
|
|
subcloud = db_api.subcloud_get(self.ctx, subcloud.id)
|
|
self.assertIsNotNone(subcloud)
|
|
self.assertEqual(subcloud.management_state,
|
|
dccommon_consts.MANAGEMENT_MANAGED)
|
|
self.assertEqual(subcloud.availability_status,
|
|
dccommon_consts.AVAILABILITY_ONLINE)
|
|
|
|
# Attempt to update each status to be in-sync for an online/managed
|
|
# subcloud
|
|
for endpoint in [dccommon_consts.ENDPOINT_TYPE_PLATFORM,
|
|
dccommon_consts.ENDPOINT_TYPE_IDENTITY,
|
|
dccommon_consts.ENDPOINT_TYPE_PATCHING,
|
|
dccommon_consts.ENDPOINT_TYPE_FM,
|
|
dccommon_consts.ENDPOINT_TYPE_NFV,
|
|
dccommon_consts.ENDPOINT_TYPE_DC_CERT]:
|
|
ssm.update_subcloud_endpoint_status(
|
|
self.ctx, subcloud_region=subcloud.region_name,
|
|
endpoint_type=endpoint,
|
|
sync_status=dccommon_consts.SYNC_STATUS_IN_SYNC)
|
|
|
|
updated_subcloud_status = db_api.subcloud_status_get(
|
|
self.ctx, subcloud.id, endpoint)
|
|
self.assertIsNotNone(updated_subcloud_status)
|
|
self.assertEqual(updated_subcloud_status.sync_status,
|
|
dccommon_consts.SYNC_STATUS_IN_SYNC)
|
|
|
|
# Change the sync status to 'out-of-sync' and verify fair lock access
|
|
# based on subcloud name for each update
|
|
with mock.patch.object(lockutils, 'internal_fair_lock') as mock_lock:
|
|
for endpoint in [dccommon_consts.ENDPOINT_TYPE_PLATFORM,
|
|
dccommon_consts.ENDPOINT_TYPE_IDENTITY,
|
|
dccommon_consts.ENDPOINT_TYPE_PATCHING,
|
|
dccommon_consts.ENDPOINT_TYPE_FM,
|
|
dccommon_consts.ENDPOINT_TYPE_NFV,
|
|
dccommon_consts.ENDPOINT_TYPE_DC_CERT]:
|
|
ssm.update_subcloud_endpoint_status(
|
|
self.ctx, subcloud_region=subcloud.region_name,
|
|
endpoint_type=endpoint,
|
|
sync_status=dccommon_consts.SYNC_STATUS_OUT_OF_SYNC)
|
|
# Verify lock was called
|
|
mock_lock.assert_called_with(subcloud.region_name)
|
|
|
|
# Verify status was updated
|
|
updated_subcloud_status = db_api.subcloud_status_get(
|
|
self.ctx, subcloud.id, endpoint)
|
|
self.assertIsNotNone(updated_subcloud_status)
|
|
self.assertEqual(updated_subcloud_status.sync_status,
|
|
dccommon_consts.SYNC_STATUS_OUT_OF_SYNC)
|
|
|
|
def test_update_subcloud_availability_go_online(self):
|
|
self.assertIsNotNone(self.subcloud)
|
|
self.assertEqual(self.subcloud.availability_status,
|
|
dccommon_consts.AVAILABILITY_OFFLINE)
|
|
|
|
ssm = subcloud_state_manager.SubcloudStateManager()
|
|
db_api.subcloud_update(self.ctx, self.subcloud.id,
|
|
management_state=dccommon_consts.MANAGEMENT_MANAGED)
|
|
|
|
# create sync statuses for endpoints
|
|
for endpoint in [dccommon_consts.ENDPOINT_TYPE_PLATFORM,
|
|
dccommon_consts.ENDPOINT_TYPE_IDENTITY,
|
|
dccommon_consts.ENDPOINT_TYPE_PATCHING,
|
|
dccommon_consts.ENDPOINT_TYPE_FM,
|
|
dccommon_consts.ENDPOINT_TYPE_NFV,
|
|
dccommon_consts.ENDPOINT_TYPE_DC_CERT]:
|
|
status = db_api.subcloud_status_create(
|
|
self.ctx, self.subcloud.id, endpoint)
|
|
self.assertIsNotNone(status)
|
|
self.assertEqual(status.sync_status, dccommon_consts.SYNC_STATUS_UNKNOWN)
|
|
|
|
ssm.update_subcloud_availability(self.ctx, self.subcloud.region_name,
|
|
dccommon_consts.AVAILABILITY_ONLINE)
|
|
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, 'subcloud1')
|
|
# Verify the subcloud was set to online
|
|
self.assertEqual(updated_subcloud.availability_status,
|
|
dccommon_consts.AVAILABILITY_ONLINE)
|
|
# Verify notifying dcorch
|
|
self.mock_dcorch_api().update_subcloud_states.assert_called_once_with(
|
|
self.ctx, self.subcloud.region_name, updated_subcloud.management_state,
|
|
dccommon_consts.AVAILABILITY_ONLINE)
|
|
# Verify triggering audits
|
|
self.mock_audit_rpc_client().trigger_subcloud_audits.\
|
|
assert_called_once_with(self.ctx, self.subcloud.id)
|
|
|
|
self.mock_dcmanager_api().subcloud_online.\
|
|
assert_called_once_with(self.ctx, self.subcloud.region_name)
|
|
|
|
def test_bulk_update_subcloud_availability_and_endpoint_status(self):
|
|
availability_data = {
|
|
"availability_status": dccommon_consts.AVAILABILITY_OFFLINE,
|
|
"update_state_only": False,
|
|
"audit_fail_count": 1
|
|
}
|
|
endpoint_data = {
|
|
dccommon_consts.ENDPOINT_TYPE_LOAD: dccommon_consts.SYNC_STATUS_IN_SYNC,
|
|
dccommon_consts.ENDPOINT_TYPE_FIRMWARE:
|
|
dccommon_consts.SYNC_STATUS_OUT_OF_SYNC
|
|
}
|
|
endpoints = db_api.subcloud_status_get_all(self.ctx, self.subcloud.id)
|
|
|
|
db_api.subcloud_update(
|
|
self.ctx, self.subcloud.id,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE,
|
|
management_state=dccommon_consts.MANAGEMENT_MANAGED
|
|
)
|
|
|
|
ssm = subcloud_state_manager.SubcloudStateManager()
|
|
ssm.bulk_update_subcloud_availability_and_endpoint_status(
|
|
self.ctx, self.subcloud.name, self.subcloud.region_name,
|
|
availability_data, endpoint_data
|
|
)
|
|
|
|
updated_subcloud = db_api.subcloud_get(self.ctx, self.subcloud.id)
|
|
self.assertEqual(
|
|
updated_subcloud.availability_status,
|
|
availability_data["availability_status"]
|
|
)
|
|
|
|
new_endpoints = db_api.subcloud_status_get_all(self.ctx, self.subcloud.id)
|
|
for index, endpoint in enumerate(endpoints):
|
|
self.assertEqual(
|
|
endpoint.endpoint_type, new_endpoints[index].endpoint_type
|
|
)
|
|
if endpoint.endpoint_type in endpoint_data:
|
|
self.assertEqual(
|
|
new_endpoints[index].sync_status,
|
|
endpoint_data[endpoint.endpoint_type]
|
|
)
|
|
else:
|
|
self.assertEqual(
|
|
endpoint.sync_status, new_endpoints[index].sync_status
|
|
)
|
|
|
|
@mock.patch.object(
|
|
db_api, "subcloud_status_bulk_update_endpoints",
|
|
wraps=db_api.subcloud_status_bulk_update_endpoints
|
|
)
|
|
def test_bulk_update_endpoint_status_when_endpoint_status_is_the_same(
|
|
self, mock_db
|
|
):
|
|
"""Test bulk_update_endpoint_status updates the endpoint with same status
|
|
|
|
When the endpoint's status in the database is the same as the one it'll be
|
|
updated to, ensure that, instead of validating, bulk_update_endpoint_status
|
|
sets the same value in the database
|
|
"""
|
|
|
|
db_api.subcloud_update(
|
|
self.ctx, self.subcloud.id,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE,
|
|
management_state=dccommon_consts.MANAGEMENT_MANAGED
|
|
)
|
|
|
|
endpoint_data = {
|
|
dccommon_consts.ENDPOINT_TYPE_LOAD: dccommon_consts.SYNC_STATUS_IN_SYNC,
|
|
dccommon_consts.ENDPOINT_TYPE_FIRMWARE:
|
|
dccommon_consts.SYNC_STATUS_OUT_OF_SYNC
|
|
}
|
|
|
|
ssm = subcloud_state_manager.SubcloudStateManager()
|
|
ssm.bulk_update_subcloud_availability_and_endpoint_status(
|
|
self.ctx, self.subcloud.name, self.subcloud.region_name,
|
|
None, endpoint_data
|
|
)
|
|
|
|
self.assertEqual(mock_db.call_count, 1)
|
|
|
|
# Re-executing the method should result in the same amount of call counts
|
|
# for the database query since there are no updates
|
|
ssm.bulk_update_subcloud_availability_and_endpoint_status(
|
|
self.ctx, self.subcloud.name, self.subcloud.region_name,
|
|
None, endpoint_data
|
|
)
|
|
|
|
self.assertEqual(mock_db.call_count, 2)
|
|
|
|
def test_bulk_update_fails_with_invalid_region(self):
|
|
ssm = subcloud_state_manager.SubcloudStateManager()
|
|
|
|
self.assertRaises(
|
|
exceptions.SubcloudRegionNameNotFound,
|
|
ssm.bulk_update_subcloud_availability_and_endpoint_status,
|
|
self.ctx, self.subcloud.name, "fake", None, None
|
|
)
|
|
|
|
@mock.patch.object(subcloud_state_manager.SubcloudStateManager,
|
|
'_raise_or_clear_subcloud_status_alarm')
|
|
def test_update_state_only(self, mock_update_status_alarm):
|
|
self.assertIsNotNone(self.subcloud)
|
|
|
|
# Set the subcloud to online/managed
|
|
db_api.subcloud_update(
|
|
self.ctx, self.subcloud.id,
|
|
management_state=dccommon_consts.MANAGEMENT_UNMANAGED,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE
|
|
)
|
|
|
|
ssm = subcloud_state_manager.SubcloudStateManager()
|
|
|
|
with mock.patch.object(db_api, "subcloud_update") as subcloud_update_mock:
|
|
ssm.update_subcloud_availability(
|
|
self.ctx, self.subcloud.region_name,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE,
|
|
update_state_only=True
|
|
)
|
|
# Verify that the subcloud was not updated
|
|
subcloud_update_mock.assert_not_called()
|
|
|
|
# Verify alarm status update was attempted
|
|
mock_update_status_alarm.assert_called_once()
|
|
|
|
# Verify dcorch was notified
|
|
self.mock_dcorch_api().update_subcloud_states.assert_called_once_with(
|
|
self.ctx, self.subcloud.region_name, self.subcloud.management_state,
|
|
dccommon_consts.AVAILABILITY_ONLINE)
|
|
|
|
# Verify audits were not triggered
|
|
self.mock_audit_rpc_client.trigger_subcloud_audits.assert_not_called()
|
|
|
|
def test_update_subcloud_availability_go_online_unmanaged(self):
|
|
self.assertIsNotNone(self.subcloud)
|
|
self.assertEqual(self.subcloud.availability_status,
|
|
dccommon_consts.AVAILABILITY_OFFLINE)
|
|
|
|
ssm = subcloud_state_manager.SubcloudStateManager()
|
|
|
|
# Note that we have intentionally left the subcloud as "unmanaged"
|
|
|
|
# create sync statuses for endpoints
|
|
for endpoint in [dccommon_consts.ENDPOINT_TYPE_PLATFORM,
|
|
dccommon_consts.ENDPOINT_TYPE_IDENTITY,
|
|
dccommon_consts.ENDPOINT_TYPE_PATCHING,
|
|
dccommon_consts.ENDPOINT_TYPE_FM,
|
|
dccommon_consts.ENDPOINT_TYPE_NFV,
|
|
dccommon_consts.ENDPOINT_TYPE_DC_CERT]:
|
|
status = db_api.subcloud_status_create(
|
|
self.ctx, self.subcloud.id, endpoint)
|
|
self.assertIsNotNone(status)
|
|
self.assertEqual(status.sync_status, dccommon_consts.SYNC_STATUS_UNKNOWN)
|
|
|
|
ssm.update_subcloud_availability(self.ctx, self.subcloud.region_name,
|
|
dccommon_consts.AVAILABILITY_ONLINE)
|
|
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, 'subcloud1')
|
|
# Verify the subcloud was set to online
|
|
self.assertEqual(updated_subcloud.availability_status,
|
|
dccommon_consts.AVAILABILITY_ONLINE)
|
|
# Verify notifying dcorch
|
|
self.mock_dcorch_api().update_subcloud_states.assert_called_once_with(
|
|
self.ctx, self.subcloud.region_name, updated_subcloud.management_state,
|
|
dccommon_consts.AVAILABILITY_ONLINE)
|
|
# Verify triggering audits
|
|
self.mock_audit_rpc_client().trigger_subcloud_audits.\
|
|
assert_called_once_with(self.ctx, self.subcloud.id)
|
|
|
|
self.mock_dcmanager_api().subcloud_online.\
|
|
assert_called_once_with(self.ctx, self.subcloud.region_name)
|
|
|
|
def test_update_subcloud_availability_go_offline(self):
|
|
self.assertIsNotNone(self.subcloud)
|
|
|
|
# Set the subcloud to online/managed
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
management_state=dccommon_consts.MANAGEMENT_MANAGED,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE)
|
|
|
|
ssm = subcloud_state_manager.SubcloudStateManager()
|
|
|
|
# create sync statuses for endpoints and set them to in-sync
|
|
for endpoint in dccommon_consts.ENDPOINT_TYPES_LIST:
|
|
db_api.subcloud_status_create(
|
|
self.ctx, self.subcloud.id, endpoint)
|
|
ssm.update_subcloud_endpoint_status(
|
|
self.ctx, subcloud_region=self.subcloud.region_name,
|
|
endpoint_type=endpoint,
|
|
sync_status=dccommon_consts.SYNC_STATUS_IN_SYNC)
|
|
|
|
# We trigger a subcloud audits after updating the identity from unknown
|
|
# to in-sync
|
|
self.mock_audit_rpc_client().trigger_subcloud_audits.\
|
|
assert_called_once_with(self.ctx, self.subcloud.id)
|
|
|
|
# Audit fails once
|
|
audit_fail_count = 1
|
|
ssm.update_subcloud_availability(self.ctx, self.subcloud.region_name,
|
|
availability_status=None,
|
|
audit_fail_count=audit_fail_count)
|
|
# Verify the subcloud availability was not updated
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, 'subcloud1')
|
|
self.assertEqual(updated_subcloud.availability_status,
|
|
dccommon_consts.AVAILABILITY_ONLINE)
|
|
# Verify dcorch was not notified
|
|
self.mock_dcorch_api().update_subcloud_states.assert_not_called()
|
|
# Verify the audit_fail_count was updated
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, 'subcloud1')
|
|
self.assertEqual(updated_subcloud.audit_fail_count, audit_fail_count)
|
|
|
|
# Audit fails again
|
|
audit_fail_count = audit_fail_count + 1
|
|
ssm.update_subcloud_availability(self.ctx, self.subcloud.region_name,
|
|
dccommon_consts.AVAILABILITY_OFFLINE,
|
|
audit_fail_count=audit_fail_count)
|
|
|
|
# Verify the subcloud availability was updated
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, 'subcloud1')
|
|
self.assertEqual(updated_subcloud.availability_status,
|
|
dccommon_consts.AVAILABILITY_OFFLINE)
|
|
|
|
# Verify notifying dcorch
|
|
self.mock_dcorch_api().update_subcloud_states.assert_called_once_with(
|
|
self.ctx, self.subcloud.region_name, updated_subcloud.management_state,
|
|
dccommon_consts.AVAILABILITY_OFFLINE)
|
|
|
|
# Verify all endpoint statuses set to unknown
|
|
for self.subcloud, subcloud_status in db_api. \
|
|
subcloud_get_with_status(self.ctx, self.subcloud.id):
|
|
self.assertIsNotNone(subcloud_status)
|
|
self.assertEqual(subcloud_status.sync_status,
|
|
dccommon_consts.SYNC_STATUS_UNKNOWN)
|
|
|
|
def test_update_subcloud_identity_endpoint(self):
|
|
self.assertIsNotNone(self.subcloud)
|
|
for endpoint_type in dccommon_consts.ENDPOINT_TYPES_LIST:
|
|
subcloud_status = db_api.subcloud_status_get(
|
|
self.ctx, self.subcloud.id, endpoint_type)
|
|
self.assertIsNotNone(subcloud_status)
|
|
self.assertEqual(subcloud_status.sync_status,
|
|
dccommon_consts.SYNC_STATUS_UNKNOWN)
|
|
|
|
ssm = subcloud_state_manager.SubcloudStateManager()
|
|
|
|
# Set the subcloud to online/managed
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
management_state=dccommon_consts.MANAGEMENT_MANAGED,
|
|
first_identity_sync_complete=True,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE)
|
|
|
|
# Update identity endpoints statuses
|
|
endpoint = dccommon_consts.ENDPOINT_TYPE_IDENTITY
|
|
for original_sync_status in [dccommon_consts.SYNC_STATUS_IN_SYNC,
|
|
dccommon_consts.SYNC_STATUS_OUT_OF_SYNC,
|
|
dccommon_consts.SYNC_STATUS_UNKNOWN]:
|
|
|
|
for new_sync_status in [dccommon_consts.SYNC_STATUS_IN_SYNC,
|
|
dccommon_consts.SYNC_STATUS_OUT_OF_SYNC,
|
|
dccommon_consts.SYNC_STATUS_UNKNOWN]:
|
|
|
|
# Update identity to the original status
|
|
ssm.update_subcloud_endpoint_status(
|
|
self.ctx, subcloud_region=self.subcloud.region_name,
|
|
endpoint_type=endpoint,
|
|
sync_status=original_sync_status)
|
|
|
|
# Get the count of the trigger already called
|
|
trigger_subcloud_audits = \
|
|
self.mock_audit_rpc_client().trigger_subcloud_audits.call_count
|
|
|
|
# Update identity to new status and get the count of the trigger
|
|
# again
|
|
ssm.update_subcloud_endpoint_status(
|
|
self.ctx, subcloud_region=self.subcloud.region_name,
|
|
endpoint_type=endpoint,
|
|
sync_status=new_sync_status)
|
|
new_trigger_subcloud_audits = \
|
|
self.mock_audit_rpc_client().trigger_subcloud_audits.call_count
|
|
|
|
trigger_count = new_trigger_subcloud_audits - \
|
|
trigger_subcloud_audits
|
|
|
|
if original_sync_status == dccommon_consts.SYNC_STATUS_UNKNOWN and \
|
|
new_sync_status != dccommon_consts.SYNC_STATUS_UNKNOWN:
|
|
# Verify the subcloud patch and load audit is triggered once
|
|
self.assertEqual(trigger_count, 1)
|
|
else:
|
|
# Verify the subcloud patch and load audit is not triggered
|
|
self.assertEqual(trigger_count, 0)
|
|
|
|
def test_update_subcloud_sync_endpoint_type(self):
|
|
self.assertIsNotNone(self.subcloud)
|
|
|
|
endpoint_type_list = dccommon_consts.ENDPOINT_TYPES_LIST_OS
|
|
|
|
# Test openstack app installed
|
|
openstack_installed = True
|
|
self.sm.update_subcloud_sync_endpoint_type(
|
|
self.ctx, self.subcloud.region_name, endpoint_type_list,
|
|
openstack_installed
|
|
)
|
|
|
|
# Verify notifying dcorch to add subcloud sync endpoint type
|
|
self.mock_dcorch_api().add_subcloud_sync_endpoint_type.\
|
|
assert_called_once_with(self.ctx, self.subcloud.region_name,
|
|
endpoint_type_list)
|
|
|
|
# Verify the subcloud status created for os endpoints
|
|
for endpoint in endpoint_type_list:
|
|
subcloud_status = db_api.subcloud_status_get(
|
|
self.ctx, self.subcloud.id, endpoint)
|
|
self.assertIsNotNone(subcloud_status)
|
|
self.assertEqual(subcloud_status.sync_status,
|
|
dccommon_consts.SYNC_STATUS_UNKNOWN)
|
|
|
|
# Verify the subcloud openstack_installed was updated
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(updated_subcloud.openstack_installed, True)
|
|
|
|
# Test openstack app removed
|
|
openstack_installed = False
|
|
self.sm.update_subcloud_sync_endpoint_type(
|
|
self.ctx, self.subcloud.region_name, endpoint_type_list,
|
|
openstack_installed
|
|
)
|
|
# Verify notifying dcorch to remove subcloud sync endpoint type
|
|
self.mock_dcorch_api().remove_subcloud_sync_endpoint_type.\
|
|
assert_called_once_with(self.ctx, self.subcloud.region_name,
|
|
endpoint_type_list)
|
|
|
|
# Verify the subcloud status is deleted for os endpoints
|
|
for endpoint in endpoint_type_list:
|
|
self.assertRaises(exceptions.SubcloudStatusNotFound,
|
|
db_api.subcloud_status_get, self.ctx,
|
|
self.subcloud.id, endpoint)
|
|
|
|
# Verify the subcloud openstack_installed was updated
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(updated_subcloud.openstack_installed, False)
|
|
|
|
def test_get_ansible_filename(self):
|
|
filename = cutils.get_ansible_filename('subcloud1',
|
|
consts.INVENTORY_FILE_POSTFIX)
|
|
self.assertEqual(
|
|
filename,
|
|
f'{ANS_PATH}/subcloud1_inventory.yml'
|
|
)
|
|
|
|
def test_update_subcloud_peer_group_id(self):
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_DONE
|
|
fake_peer_group_id = 123
|
|
|
|
self.sm.update_subcloud(self.ctx,
|
|
self.subcloud.id,
|
|
peer_group_id=fake_peer_group_id)
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(fake_peer_group_id,
|
|
updated_subcloud.peer_group_id)
|
|
|
|
def test_update_subcloud_peer_group_id_to_none(self):
|
|
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_DONE
|
|
fake_peer_group_id = 123
|
|
|
|
self.sm.update_subcloud(self.ctx,
|
|
self.subcloud.id,
|
|
peer_group_id=fake_peer_group_id)
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(fake_peer_group_id,
|
|
updated_subcloud.peer_group_id)
|
|
self.sm.update_subcloud(self.ctx,
|
|
self.subcloud.id,
|
|
peer_group_id='NoNe')
|
|
# Verify subcloud was updated to None
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(None,
|
|
updated_subcloud.peer_group_id)
|
|
|
|
@mock.patch.object(threading.Thread, 'start')
|
|
def test_update_subcloud_with_network_reconfig_failed(self, mock_start):
|
|
self._mock_builtins_open()
|
|
self._mock_subcloud_manager_create_intermediate_ca_cert()
|
|
self.mock_os_listdir.return_value = ['testfile1', 'testfile2']
|
|
mock_start.side_effect = Exception('boom')
|
|
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_PRE_INSTALL
|
|
|
|
self.sm.update_subcloud_with_network_reconfig(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
self._test_values())
|
|
|
|
self.assertEqual(self.mock_builtins_open.call_count, 1)
|
|
self.mock_log.exception.assert_called_once_with(
|
|
'Failed to update subcloud subcloud-4')
|
|
|
|
@mock.patch.object(time, 'sleep')
|
|
@mock.patch.object(kubeoperator, 'KubeOperator')
|
|
def test_update_subcloud_with_network_reconfig(self,
|
|
mock_kube_operator,
|
|
mock_time_sleep):
|
|
self.mock_os_listdir.return_value = ['testfile1', 'testfile2']
|
|
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_PRE_INSTALL
|
|
self.sm.update_subcloud_with_network_reconfig(
|
|
self.ctx,
|
|
self.subcloud.id, self._test_values()
|
|
)
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_RECONFIGURING_NETWORK,
|
|
updated_subcloud.deploy_status)
|
|
|
|
def test_compose_update_command(self):
|
|
subcloud_update_command = self.sm.compose_update_command(
|
|
'subcloud1',
|
|
f'{ANS_PATH}/subcloud1_inventory.yml')
|
|
self.assertEqual(
|
|
subcloud_update_command,
|
|
['ansible-playbook', subcloud_manager.ANSIBLE_SUBCLOUD_UPDATE_PLAYBOOK,
|
|
'-i', f'{ANS_PATH}/subcloud1_inventory.yml', '--limit', 'subcloud1',
|
|
'--timeout', '180',
|
|
'-e', 'TEST.SW.VERSION',
|
|
'-e', 'subcloud_update_overrides={}/subcloud1_update_values.yml'
|
|
.format(ANS_PATH)]
|
|
)
|
|
|
|
def test_update_subcloud_sync_endpoint_type_region_name_not_found(self):
|
|
# Test openstack app installed
|
|
openstack_installed = True
|
|
self.assertRaises(
|
|
exceptions.SubcloudRegionNameNotFound,
|
|
self.sm.update_subcloud_sync_endpoint_type, self.ctx, 'test_region',
|
|
dccommon_consts.ENDPOINT_TYPES_LIST_OS, openstack_installed
|
|
)
|
|
|
|
def test_update_subcloud_sync_endpoint_type_failed(self):
|
|
endpoint_type_list = None
|
|
# Test openstack app installed
|
|
openstack_installed = True
|
|
self.sm.update_subcloud_sync_endpoint_type(self.ctx,
|
|
self.subcloud.region_name,
|
|
endpoint_type_list,
|
|
openstack_installed)
|
|
self.mock_log.exception.assert_called_once_with(
|
|
'Problem informing dcorch of subcloud sync endpoint type change,'
|
|
f' subcloud region: {base.SUBCLOUD_1["region_name"]}'
|
|
)
|
|
|
|
|
|
class TestSubcloudCompose(BaseTestSubcloudManager):
|
|
"""Test class for testing subcloud compose command"""
|
|
|
|
def setUp(self):
|
|
super().setUp()
|
|
|
|
def test_compose_install_command(self):
|
|
install_command = self.sm.compose_install_command(
|
|
'subcloud1',
|
|
f'{ANS_PATH}/subcloud1_inventory.yml',
|
|
FAKE_PREVIOUS_SW_VERSION)
|
|
self.assertEqual(
|
|
install_command,
|
|
['ansible-playbook', dccommon_consts.ANSIBLE_SUBCLOUD_INSTALL_PLAYBOOK,
|
|
'-i', f'{ANS_PATH}/subcloud1_inventory.yml', '--limit', 'subcloud1',
|
|
'-e', f"@{ANS_PATH}/subcloud1/"
|
|
'install_values.yml', '-e', "install_release_version=%s"
|
|
% FAKE_PREVIOUS_SW_VERSION, '-e', "rvmc_config_file=%s" % os.path.join(
|
|
ANS_PATH, 'subcloud1',
|
|
dccommon_consts.RVMC_CONFIG_FILE_NAME)]
|
|
)
|
|
|
|
@mock.patch('os.path.isfile')
|
|
def test_compose_bootstrap_command(self, mock_isfile):
|
|
mock_isfile.return_value = True
|
|
subcloud_name = base.SUBCLOUD_1['name']
|
|
subcloud_region = base.SUBCLOUD_1['region_name']
|
|
bootstrap_command = self.sm.compose_bootstrap_command(
|
|
subcloud_name,
|
|
subcloud_region,
|
|
f'{ANS_PATH}/subcloud1_inventory.yml',
|
|
FAKE_PREVIOUS_SW_VERSION)
|
|
self.assertEqual(
|
|
bootstrap_command,
|
|
[
|
|
'ansible-playbook',
|
|
cutils.get_playbook_for_software_version(
|
|
subcloud_manager.ANSIBLE_SUBCLOUD_PLAYBOOK,
|
|
FAKE_PREVIOUS_SW_VERSION),
|
|
'-i',
|
|
f'{ANS_PATH}/subcloud1_inventory.yml',
|
|
'--limit', '%s' % subcloud_name, '-e',
|
|
str("override_files_dir='%s' region_name=%s") %
|
|
(ANS_PATH, subcloud_region),
|
|
'-e', "install_release_version=%s" % FAKE_PREVIOUS_SW_VERSION
|
|
]
|
|
)
|
|
|
|
def test_compose_config_command(self):
|
|
|
|
config_command = self.sm.compose_config_command(
|
|
'subcloud1',
|
|
f'{ANS_PATH}/subcloud1_inventory.yml',
|
|
self.fake_payload)
|
|
self.assertEqual(
|
|
config_command,
|
|
[
|
|
'ansible-playbook', 'test_playbook.yaml', '-e',
|
|
f'@{ANS_PATH}/subcloud1_deploy_values.yml', '-i',
|
|
f'{ANS_PATH}/subcloud1_inventory.yml',
|
|
'--limit', 'subcloud1'
|
|
]
|
|
)
|
|
|
|
@mock.patch('os.path.isfile')
|
|
def test_compose_rehome_command_with_previous_sw_version(self, mock_isfile):
|
|
mock_isfile.return_value = True
|
|
subcloud_name = base.SUBCLOUD_1['name']
|
|
subcloud_region = base.SUBCLOUD_1['region_name']
|
|
|
|
rehome_command = self.sm.compose_rehome_command(
|
|
subcloud_name,
|
|
subcloud_region,
|
|
f'{ANS_PATH}/subcloud1_inventory.yml',
|
|
FAKE_PREVIOUS_SW_VERSION)
|
|
|
|
extra_vars = "override_files_dir='%s' region_name=%s" % (
|
|
ANS_PATH, subcloud_region)
|
|
extra_vars += (" validate_keystone_passwords_script='%s'" %
|
|
subcloud_manager.ANSIBLE_VALIDATE_KEYSTONE_PASSWORD_SCRIPT)
|
|
|
|
self.assertEqual(
|
|
rehome_command,
|
|
[
|
|
'ansible-playbook',
|
|
cutils.get_playbook_for_software_version(
|
|
subcloud_manager.ANSIBLE_SUBCLOUD_REHOME_PLAYBOOK,
|
|
FAKE_PREVIOUS_SW_VERSION),
|
|
'-i',
|
|
f'{ANS_PATH}/subcloud1_inventory.yml',
|
|
'--limit', subcloud_name, '--timeout',
|
|
subcloud_manager.REHOME_PLAYBOOK_TIMEOUT, '-e', extra_vars
|
|
]
|
|
)
|
|
|
|
@mock.patch('os.path.isfile')
|
|
def test_compose_rehome_command(self, mock_isfile):
|
|
mock_isfile.return_value = True
|
|
subcloud_name = base.SUBCLOUD_1['name']
|
|
subcloud_region = base.SUBCLOUD_1['region_name']
|
|
|
|
rehome_command = self.sm.compose_rehome_command(
|
|
subcloud_name,
|
|
subcloud_region,
|
|
f'{ANS_PATH}/subcloud1_inventory.yml',
|
|
SW_VERSION)
|
|
|
|
self.assertEqual(
|
|
rehome_command,
|
|
[
|
|
'ansible-playbook',
|
|
cutils.get_playbook_for_software_version(
|
|
subcloud_manager.ANSIBLE_SUBCLOUD_REHOME_PLAYBOOK,
|
|
SW_VERSION),
|
|
'-i',
|
|
f'{ANS_PATH}/subcloud1_inventory.yml',
|
|
'--limit', subcloud_name,
|
|
'--timeout', subcloud_manager.REHOME_PLAYBOOK_TIMEOUT,
|
|
'-e', str("override_files_dir='%s' region_name=%s") %
|
|
(ANS_PATH, subcloud_region)
|
|
]
|
|
)
|
|
|
|
def test_compose_backup_command(self):
|
|
backup_command = self.sm.compose_backup_command(
|
|
'subcloud1',
|
|
f'{ANS_PATH}/subcloud1_inventory.yml')
|
|
self.assertEqual(
|
|
backup_command,
|
|
['ansible-playbook',
|
|
subcloud_manager.ANSIBLE_SUBCLOUD_BACKUP_CREATE_PLAYBOOK,
|
|
'-i',
|
|
f'{ANS_PATH}/subcloud1_inventory.yml', '--limit', 'subcloud1', '-e',
|
|
'subcloud_bnr_overrides={}/subcloud1_backup_create_values.yml'
|
|
.format(ANS_PATH)]
|
|
)
|
|
|
|
def test_compose_backup_delete_command(self):
|
|
backup_delete_command = self.sm.compose_backup_delete_command(
|
|
'subcloud1',
|
|
f'{ANS_PATH}/subcloud1_inventory.yml')
|
|
self.assertEqual(
|
|
backup_delete_command,
|
|
['ansible-playbook',
|
|
'/usr/share/ansible/stx-ansible/playbooks/delete_subcloud_backup.yml',
|
|
'-e',
|
|
'subcloud_bnr_overrides={}/subcloud1_backup_delete_values.yml'
|
|
.format(ANS_PATH),
|
|
'-i', '/opt/dc-vault/ansible/subcloud1_inventory.yml',
|
|
'--limit', 'subcloud1']
|
|
)
|
|
|
|
|
|
class TestSubcloudRedeploy(BaseTestSubcloudManager):
|
|
"""Test class for testing subcloud redeploy"""
|
|
|
|
def setUp(self):
|
|
super().setUp()
|
|
|
|
@mock.patch.object(cutils, 'get_oam_addresses')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_prepare_for_deployment')
|
|
@mock.patch.object(cutils, 'update_values_on_yaml_file')
|
|
def test_subcloud_redeploy(self, mock_update_yml,
|
|
mock_prepare_for_deployment,
|
|
mock_oam_address):
|
|
self._mock_subcloud_manager_run_subcloud_install()
|
|
self.mock_get_playbook_for_software_version.return_value = FAKE_SW_VERSION
|
|
self.mock_ansible_run_playbook.return_value = False
|
|
self.mock_run_subcloud_install.return_value = True
|
|
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_CREATED
|
|
self.fake_install_values['software_version'] = SW_VERSION
|
|
fake_payload_bootstrap = {**fake_subcloud.FAKE_BOOTSTRAP_VALUE,
|
|
**fake_subcloud.FAKE_BOOTSTRAP_FILE_DATA}
|
|
fake_payload_bootstrap["sysadmin_password"] = "testpass"
|
|
fake_payload = {**self.fake_payload_install,
|
|
**fake_payload_bootstrap,
|
|
**self.fake_payload}
|
|
|
|
self.sm.redeploy_subcloud(self.ctx, self.subcloud.id, fake_payload)
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx,
|
|
self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_DONE,
|
|
updated_subcloud.deploy_status)
|
|
# Verify subcloud rehomed flag is False after re-deploy
|
|
self.assertFalse(updated_subcloud.rehomed)
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager, 'subcloud_deploy_config')
|
|
@mock.patch.object(subcloud_install.SubcloudInstall, 'prep')
|
|
@mock.patch.object(cutils, 'get_oam_addresses')
|
|
@mock.patch.object(cutils, 'update_values_on_yaml_file')
|
|
@mock.patch.object(subprocess, 'check_call')
|
|
def test_subcloud_redeploy_skip_deploy_config(self, mock_check_call,
|
|
mock_update_yml,
|
|
mock_oam_address, mock_prep,
|
|
mock_subcloud_deploy_config):
|
|
self.mock_get_playbook_for_software_version.return_value = FAKE_SW_VERSION
|
|
self.mock_ansible_run_playbook.return_value = False
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_CREATED
|
|
self.fake_install_values['software_version'] = SW_VERSION
|
|
|
|
fake_payload_bootstrap = {**fake_subcloud.FAKE_BOOTSTRAP_VALUE,
|
|
**fake_subcloud.FAKE_BOOTSTRAP_FILE_DATA}
|
|
fake_payload_bootstrap["sysadmin_password"] = "testpass"
|
|
fake_payload = {**self.fake_payload_install,
|
|
**fake_payload_bootstrap}
|
|
|
|
self.sm.redeploy_subcloud(self.ctx, self.subcloud.id, fake_payload)
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx,
|
|
self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_DONE,
|
|
updated_subcloud.deploy_status)
|
|
# Verify subcloud rehomed flag is False after re-deploy
|
|
self.assertFalse(updated_subcloud.rehomed)
|
|
mock_subcloud_deploy_config.assert_not_called()
|
|
|
|
def test_handle_subcloud_operations_in_progress(self):
|
|
# There are three types of transitory states
|
|
state_map = {
|
|
'deploy_status': subcloud_manager.TRANSITORY_STATES.copy(),
|
|
'backup_status': subcloud_manager.TRANSITORY_BACKUP_STATES.copy(),
|
|
'prestage_status': subcloud_manager.TRANSITORY_PRESTAGE_STATES.copy()}
|
|
|
|
# Any state not defined in the transitory states should not be modified
|
|
# NOTE: this should cover the test_handle_completed_subcloud_operations test
|
|
for states in state_map.values():
|
|
states["undefined state"] = "undefined state"
|
|
|
|
# Create subclouds
|
|
expected_states = {}
|
|
for state_type, states in state_map.items():
|
|
for initial_state, expected_state in states.items():
|
|
uid = str(uuid.uuid4())
|
|
self.subcloud['name'] = uid
|
|
self.subcloud['region_name'] = uid
|
|
db_api.subcloud_update(
|
|
self.ctx, self.subcloud.id, **{state_type: initial_state})
|
|
expected_states[uid] = (state_type, expected_state)
|
|
|
|
# Run the handle_subcloud_operations_in_progress function
|
|
self.sm.handle_subcloud_operations_in_progress()
|
|
|
|
# Verify the expected states
|
|
for subcloud in db_api.subcloud_get_all(self.ctx):
|
|
state_type, expected_state = expected_states[self.subcloud.name]
|
|
self.assertEqual(expected_state, subcloud.get(state_type))
|
|
|
|
|
|
class TestSubcloudBackup(BaseTestSubcloudManager):
|
|
"""Test class for testing subcloud backup"""
|
|
|
|
def setUp(self):
|
|
super().setUp()
|
|
self.mock_openstack_driver().keystone_client = FakeKeystoneClient()
|
|
self.values = copy.copy(FAKE_BACKUP_DELETE_LOAD_1)
|
|
self.backup_values = copy.copy(FAKE_BACKUP_CREATE_LOAD)
|
|
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_DONE,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE,
|
|
management_state=dccommon_consts.MANAGEMENT_MANAGED,
|
|
backup_datetime=None,
|
|
backup_status=consts.BACKUP_STATE_UNKNOWN)
|
|
|
|
@mock.patch.object(cutils, 'is_subcloud_healthy', return_value=True)
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_run_subcloud_backup_create_playbook')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_set_subcloud_backup_failure_alarm')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_clear_subcloud_backup_failure_alarm_if_exists')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_overrides_for_backup_or_restore')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_subcloud_inventory_file')
|
|
def test_backup_create_managed_online(
|
|
self, mock_create_inventory_file, mock_create_overrides,
|
|
mock_clear_alarm, mock_set_alarm, mock_run_playbook, mock_is_healthy,
|
|
):
|
|
mock_create_inventory_file.return_value = 'inventory_file.yml'
|
|
mock_create_overrides.return_value = 'overrides_file.yml'
|
|
|
|
self.sm.create_subcloud_backups(self.ctx, payload=self.backup_values)
|
|
|
|
mock_create_inventory_file.assert_called_once()
|
|
mock_create_overrides.assert_called_once()
|
|
mock_clear_alarm.assert_called_once()
|
|
mock_set_alarm.assert_not_called()
|
|
mock_run_playbook.assert_called_once()
|
|
mock_is_healthy.assert_called_once()
|
|
|
|
# Verify that subcloud has the correct backup status
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.BACKUP_STATE_PRE_BACKUP,
|
|
updated_subcloud.backup_status)
|
|
|
|
@mock.patch.object(cutils, 'is_subcloud_healthy', return_value=False)
|
|
def test_backup_create_managed_online_not_healthy(self,
|
|
mock_is_healthy):
|
|
self.backup_values['local_only'] = False
|
|
self.backup_values['registry_images'] = True
|
|
|
|
self.sm.create_subcloud_backups(self.ctx, payload=self.backup_values)
|
|
mock_is_healthy.assert_called_once()
|
|
# Verify that subcloud has the correct backup status
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.BACKUP_STATE_VALIDATE_FAILED,
|
|
updated_subcloud.backup_status)
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_backup_overrides_file')
|
|
def test_delete_subcloud_backup_failed(self,
|
|
mock_create_backup_overrides_file):
|
|
mock_create_backup_overrides_file.side_effect = Exception('boom')
|
|
self.subcloud['deploy_status'] = consts.BACKUP_STATE_COMPLETE_CENTRAL
|
|
self.sm._delete_subcloud_backup(
|
|
self.ctx, payload=self.values,
|
|
release_version=FAKE_SW_VERSION, subcloud=self.subcloud)
|
|
self.mock_log.exception.assert_called_once_with(
|
|
f'Failed to prepare subcloud {self.subcloud.name} for backup delete')
|
|
|
|
@mock.patch.object(cutils, 'is_subcloud_healthy',
|
|
return_value=True)
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_subcloud_inventory_file')
|
|
@mock.patch.object(os.path, 'join')
|
|
def test_backup_create_failed(self, mock_os_path_join,
|
|
mock_create_inventory_file,
|
|
mock_is_healthy):
|
|
mock_os_path_join.return_value = 'subcloud1_fake_file.yml'
|
|
mock_create_inventory_file.return_value = 'inventory_file.yml'
|
|
self.mock_ansible_run_playbook.side_effect = PlaybookExecutionFailed()
|
|
self.backup_values['local_only'] = True
|
|
self.backup_values['registry_images'] = True
|
|
|
|
self.sm.create_subcloud_backups(self.ctx, payload=self.backup_values)
|
|
|
|
mock_create_inventory_file.assert_called_once()
|
|
mock_is_healthy.assert_called_once()
|
|
|
|
# Verify that subcloud has the correct backup status
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.BACKUP_STATE_FAILED,
|
|
updated_subcloud.backup_status)
|
|
self.mock_log.error.assert_called_once_with(
|
|
f'FAILED backing-up playbook of ({self.subcloud.name}).'
|
|
'\ncheck individual log at subcloud1_fake_file.yml_playbook_output.log'
|
|
' for detailed output ')
|
|
|
|
def test_backup_create_managed_online_backup_state_in_progess(self):
|
|
self.backup_values['local_only'] = False
|
|
self.backup_values['registry_images'] = True
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
backup_status=consts.BACKUP_STATE_IN_PROGRESS)
|
|
self.sm.create_subcloud_backups(self.ctx, payload=self.backup_values)
|
|
# Verify that subcloud has the correct backup status
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.BACKUP_STATE_IN_PROGRESS,
|
|
updated_subcloud.backup_status)
|
|
Calls = [mock.call.debug('SubcloudManager initialization...'),
|
|
mock.call.info('Subcloud subcloud1 already'
|
|
' has a backup operation in progress'),
|
|
mock.call.info('Subcloud backup operation finished')]
|
|
self.mock_log.assert_has_calls(Calls)
|
|
|
|
@mock.patch.object(cutils, 'get_oam_addresses')
|
|
@mock.patch.object(cutils, 'is_subcloud_healthy', return_value=True)
|
|
def backup_create_managed_online_local_only(
|
|
self, mock_is_healthy, mock_get_oam_addresses
|
|
):
|
|
self.sm.create_subcloud_backups(self.ctx, payload=self.backup_values)
|
|
|
|
mock_is_healthy.assert_called_once()
|
|
|
|
# Verify that subcloud has the correct backup status
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.BACKUP_STATE_COMPLETE_LOCAL,
|
|
updated_subcloud.backup_status)
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_run_parallel_group_operation')
|
|
def test_backup_create_unmanaged_online(self, mock_parallel_group_operation):
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_NONE,
|
|
management_state=dccommon_consts.MANAGEMENT_UNMANAGED
|
|
)
|
|
|
|
self.sm.create_subcloud_backups(self.ctx, payload=self.backup_values)
|
|
|
|
mock_parallel_group_operation.assert_called_once()
|
|
|
|
# Verify that subcloud has the correct backup status
|
|
# consts.BACKUP_STATE_VALIDATE_FAILED
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.BACKUP_STATE_VALIDATE_FAILED,
|
|
updated_subcloud.backup_status)
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_run_parallel_group_operation')
|
|
def test_backup_create_unmanaged_offline(self, mock_parallel_group_operation):
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_NONE,
|
|
availability_status=dccommon_consts.AVAILABILITY_OFFLINE,
|
|
management_state=dccommon_consts.MANAGEMENT_UNMANAGED)
|
|
|
|
self.sm.create_subcloud_backups(self.ctx, payload=self.backup_values)
|
|
mock_parallel_group_operation.assert_called_once()
|
|
|
|
# Verify that subcloud has the correct backup status
|
|
# consts.BACKUP_STATE_VALIDATE_FAILED
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.BACKUP_STATE_VALIDATE_FAILED,
|
|
updated_subcloud.backup_status)
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_run_parallel_group_operation')
|
|
def test_backup_create_managed_offline(self, mock_parallel_group_operation):
|
|
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_NONE,
|
|
availability_status=dccommon_consts.AVAILABILITY_OFFLINE)
|
|
|
|
self.sm.create_subcloud_backups(self.ctx, payload=self.backup_values)
|
|
mock_parallel_group_operation.assert_called_once()
|
|
|
|
# Verify that subcloud has the correct backup status
|
|
# consts.BACKUP_STATE_VALIDATE_FAILED
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.BACKUP_STATE_VALIDATE_FAILED,
|
|
updated_subcloud.backup_status)
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_run_parallel_group_operation')
|
|
def test_backup_delete_managed_online(self, mock_parallel_group_operation):
|
|
values = copy.copy(FAKE_BACKUP_DELETE_LOAD)
|
|
values['local_only'] = False
|
|
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_NONE)
|
|
|
|
self.sm.delete_subcloud_backups(self.ctx, payload=values,
|
|
release_version=FAKE_SW_VERSION)
|
|
|
|
mock_parallel_group_operation.assert_called_once()
|
|
|
|
# Verify that subcloud has the correct backup status
|
|
# consts.BACKUP_STATE_UNKNOWN
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.BACKUP_STATE_UNKNOWN,
|
|
updated_subcloud.backup_status)
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_run_parallel_group_operation')
|
|
def test_backup_delete_managed_local_online(self, mock_parallel_group_operation):
|
|
|
|
values = copy.copy(FAKE_BACKUP_DELETE_LOCAL_LOAD)
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_NONE)
|
|
|
|
self.sm.delete_subcloud_backups(self.ctx, payload=values,
|
|
release_version=FAKE_SW_VERSION)
|
|
|
|
mock_parallel_group_operation.assert_called_once()
|
|
|
|
# Verify that subcloud has the correct backup status
|
|
# consts.BACKUP_STATE_UNKNOWN
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.BACKUP_STATE_UNKNOWN,
|
|
updated_subcloud.backup_status)
|
|
|
|
@mock.patch.object(cutils, 'is_subcloud_healthy', return_value=True)
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_backup_overrides_file')
|
|
@mock.patch.object(cutils, 'get_oam_addresses')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_clear_subcloud_backup_failure_alarm_if_exists')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'compose_backup_command')
|
|
def test_backup_create_subcloud(self,
|
|
mock_compose_backup_command,
|
|
mock_clear_subcloud_failure_alarm,
|
|
mock_oam_address,
|
|
mock_create_backup_file,
|
|
mock_is_healthy):
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
backup_status=consts.BACKUP_STATE_UNKNOWN)
|
|
values = copy.copy(FAKE_BACKUP_CREATE_LOAD_1)
|
|
override_file = os_path.join(
|
|
ANS_PATH, self.subcloud.name + "_backup_create_values.yml"
|
|
)
|
|
mock_create_backup_file.return_value = override_file
|
|
|
|
self.sm._backup_subcloud(self.ctx, payload=values, subcloud=self.subcloud)
|
|
self.mock_create_subcloud_inventory.assert_called_once()
|
|
mock_oam_address.return_value = '2620:10a:a001:d41::260'
|
|
self.mock_keyring.get_password.assert_called()
|
|
|
|
mock_create_backup_file.assert_called_once()
|
|
self.mock_ansible_run_playbook.assert_called_once()
|
|
mock_is_healthy.assert_called_once()
|
|
|
|
mock_compose_backup_command.assert_called_once()
|
|
mock_clear_subcloud_failure_alarm.assert_called_once()
|
|
|
|
self.mock_delete_subcloud_inventory.assert_called_once_with(override_file)
|
|
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.BACKUP_STATE_COMPLETE_CENTRAL,
|
|
updated_subcloud.backup_status)
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_subcloud_inventory_file')
|
|
def test_backup_create_subcloud_fail_to_create(
|
|
self, mock_create_subcloud_inventory_file):
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
backup_status=consts.BACKUP_STATE_UNKNOWN)
|
|
|
|
values = copy.copy(FAKE_BACKUP_CREATE_LOAD_1)
|
|
self.sm._backup_subcloud(self.ctx, payload=values, subcloud=self.subcloud)
|
|
|
|
mock_create_subcloud_inventory_file.side_effect = Exception(
|
|
'FakeFailure')
|
|
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.BACKUP_STATE_PREP_FAILED,
|
|
updated_subcloud.backup_status)
|
|
|
|
@mock.patch.object(cutils, 'get_oam_addresses')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'compose_backup_delete_command')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_backup_overrides_file')
|
|
def test_delete_subcloud_backup(
|
|
self, mock_create_backup_overrides_file,
|
|
mock_compose_backup_delete_command,
|
|
mock_oam_address
|
|
):
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
backup_status=consts.BACKUP_STATE_COMPLETE_CENTRAL)
|
|
|
|
override_file = os_path.join(
|
|
ANS_PATH, self.subcloud.name + "_backup_delete_values.yml"
|
|
)
|
|
mock_create_backup_overrides_file.return_value = override_file
|
|
|
|
self.sm._delete_subcloud_backup(
|
|
self.ctx, payload=self.values,
|
|
release_version=FAKE_SW_VERSION, subcloud=self.subcloud
|
|
)
|
|
|
|
mock_create_backup_overrides_file.assert_called_once()
|
|
mock_compose_backup_delete_command.assert_called_once()
|
|
self.mock_ansible_run_playbook.assert_called_once()
|
|
mock_oam_address.return_value = '2620:10a:a001:d41::260'
|
|
self.mock_create_subcloud_inventory.assert_not_called()
|
|
self.mock_delete_subcloud_inventory.assert_called_once_with(override_file)
|
|
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.BACKUP_STATE_UNKNOWN,
|
|
updated_subcloud.backup_status)
|
|
|
|
@mock.patch.object(cutils, 'get_oam_addresses')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'compose_backup_delete_command')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_backup_overrides_file')
|
|
def test_delete_subcloud_backup_local_only(
|
|
self, mock_create_subcloud_inventory_file,
|
|
mock_compose_backup_delete_command,
|
|
mock_oam_address):
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
backup_status=consts.BACKUP_STATE_COMPLETE_LOCAL)
|
|
|
|
self.values['local_only'] = True
|
|
|
|
override_file = os_path.join(
|
|
ANS_PATH,
|
|
self.subcloud.name + "_backup_delete_values.yml"
|
|
)
|
|
mock_create_subcloud_inventory_file.return_value = override_file
|
|
|
|
self.sm._delete_subcloud_backup(
|
|
self.ctx, payload=self.values,
|
|
release_version=FAKE_SW_VERSION, subcloud=self.subcloud)
|
|
|
|
mock_create_subcloud_inventory_file.assert_called_once()
|
|
mock_compose_backup_delete_command.assert_called_once()
|
|
self.mock_ansible_run_playbook.assert_called_once()
|
|
mock_oam_address.return_value = '2620:10a:a001:d41::260'
|
|
self.mock_create_subcloud_inventory.assert_called_once()
|
|
self.mock_delete_subcloud_inventory.assert_called_once_with(override_file)
|
|
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.BACKUP_STATE_UNKNOWN,
|
|
updated_subcloud.backup_status)
|
|
|
|
def test_compose_backup_delete_command_backup_stored_in_central_storage(self):
|
|
backup_delete_command = self.sm.compose_backup_delete_command(
|
|
'subcloud1')
|
|
self.assertEqual(
|
|
backup_delete_command,
|
|
['ansible-playbook',
|
|
'/usr/share/ansible/stx-ansible/playbooks/delete_subcloud_backup.yml',
|
|
'-e',
|
|
'subcloud_bnr_overrides={}/subcloud1_backup_delete_values.yml'
|
|
.format(ANS_PATH), '-e',
|
|
'inventory_hostname=subcloud1']
|
|
)
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'compose_backup_delete_command')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_backup_overrides_file')
|
|
def test_delete_subcloud_backup_playbook_execution_failed(
|
|
self, mock_create_backup_overrides_file,
|
|
mock_compose_backup_delete_command
|
|
):
|
|
self.mock_ansible_run_playbook.side_effect = PlaybookExecutionFailed()
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
backup_status=consts.BACKUP_STATE_COMPLETE_CENTRAL)
|
|
|
|
self.sm._delete_subcloud_backup(
|
|
self.ctx, payload=self.values,
|
|
release_version=FAKE_SW_VERSION, subcloud=self.subcloud
|
|
)
|
|
Calls = [
|
|
mock.call(
|
|
'Failed to delete backup for subcloud subcloud1, check individual'
|
|
' log at /var/log/dcmanager/ansible/subcloud1_playbook_output.log'
|
|
' for detailed output.'),
|
|
mock.call(
|
|
'FAILED failed playbook of (subcloud1).\ncheck individual'
|
|
' log at /var/log/dcmanager/ansible/subcloud1_playbook_output.log'
|
|
' for detailed output ')]
|
|
self.mock_log.error.assert_has_calls(Calls)
|
|
mock_create_backup_overrides_file.assert_called_once()
|
|
mock_compose_backup_delete_command.assert_called_once()
|
|
self.mock_ansible_run_playbook.assert_called_once()
|
|
self.mock_create_subcloud_inventory.assert_not_called()
|
|
|
|
|
|
class TestSubcloudPrestage(BaseTestSubcloudManager):
|
|
"""Test class for testing subcloud prestage"""
|
|
|
|
def setUp(self):
|
|
super().setUp()
|
|
|
|
@mock.patch.object(threading.Thread, 'start')
|
|
def test_prestage_no_subcloud(self, mock_thread_start):
|
|
values = copy.copy(FAKE_PRESTAGE_PAYLOAD)
|
|
values['subcloud_name'] = 'randomname'
|
|
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_NONE
|
|
e = self.assertRaises(exceptions.PrestagePreCheckFailedException,
|
|
self.sm.prestage_subcloud, self.ctx, values)
|
|
|
|
self.assertTrue('Subcloud does not exist'
|
|
in str(e))
|
|
|
|
@mock.patch.object(cutils, 'get_filename_by_prefix')
|
|
@mock.patch.object(prestage, '_run_ansible')
|
|
def test_prestage_remote_pass_with_img_list(self, mock_run_ansible,
|
|
mock_get_filename_by_prefix):
|
|
|
|
values = copy.copy(FAKE_PRESTAGE_PAYLOAD)
|
|
self.subcloud.update(
|
|
{'deploy_status': consts.DEPLOY_STATE_NONE,
|
|
'software_version': FAKE_SUBCLOUD_SW_VERSION}
|
|
)
|
|
|
|
mock_run_ansible.return_value = None
|
|
mock_get_filename_by_prefix.return_value = 'prestage_images_list.txt'
|
|
self.mock_os_path_isdir.return_value = True
|
|
prestage._prestage_standalone_thread(self.ctx, self.subcloud, payload=values)
|
|
|
|
# Verify that subcloud has the correct deploy status
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.PRESTAGE_STATE_COMPLETE,
|
|
updated_subcloud.prestage_status)
|
|
|
|
# Verify both of prestage package and image ansible playbooks were called
|
|
self.assertEqual(mock_run_ansible.call_count, 2)
|
|
# Verify the "image_list_file" was passed to the prestage image playbook
|
|
# for the remote prestage
|
|
self.assertIn('image_list_file',
|
|
mock_run_ansible.call_args_list[1].args[1][5])
|
|
# Verify the prestage request release was passed to the playbooks
|
|
self.assertIn(FAKE_PRESTAGE_RELEASE,
|
|
mock_run_ansible.call_args_list[0].args[1][5])
|
|
self.assertIn(FAKE_PRESTAGE_RELEASE,
|
|
mock_run_ansible.call_args_list[1].args[1][5])
|
|
|
|
@mock.patch.object(cutils, 'get_filename_by_prefix')
|
|
@mock.patch.object(prestage, '_run_ansible')
|
|
def test_prestage_remote_pass_without_img_list(self, mock_run_ansible,
|
|
mock_get_filename_by_prefix):
|
|
|
|
values = copy.copy(FAKE_PRESTAGE_PAYLOAD)
|
|
self.subcloud.update(
|
|
{'deploy_status': consts.DEPLOY_STATE_NONE,
|
|
'software_version': FAKE_SUBCLOUD_SW_VERSION}
|
|
)
|
|
|
|
mock_run_ansible.return_value = None
|
|
mock_get_filename_by_prefix.return_value = None
|
|
self.mock_os_path_isdir.return_value = True
|
|
prestage._prestage_standalone_thread(self.ctx, self.subcloud, payload=values)
|
|
|
|
# Verify that subcloud has the correct deploy status
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.PRESTAGE_STATE_COMPLETE,
|
|
updated_subcloud.prestage_status)
|
|
|
|
# Verify that only prestage package playbook is called
|
|
self.assertEqual(mock_run_ansible.call_count, 1)
|
|
|
|
# Verify the prestage request release was passed to the playbooks
|
|
self.assertTrue(
|
|
FAKE_PRESTAGE_RELEASE in mock_run_ansible.call_args_list[0].args[1][5])
|
|
|
|
@mock.patch.object(cutils, 'get_filename_by_prefix')
|
|
@mock.patch.object(prestage, '_run_ansible')
|
|
def test_prestage_local_pass_with_img_list(self, mock_run_ansible,
|
|
mock_get_filename_by_prefix):
|
|
|
|
values = copy.copy(FAKE_PRESTAGE_PAYLOAD)
|
|
self.subcloud.update(
|
|
{'deploy_status': consts.DEPLOY_STATE_NONE,
|
|
'software_version': FAKE_PRESTAGE_RELEASE}
|
|
)
|
|
|
|
mock_run_ansible.return_value = None
|
|
mock_get_filename_by_prefix.return_value = 'prestage_images_list.txt'
|
|
self.mock_os_path_isdir.return_value = True
|
|
prestage._prestage_standalone_thread(self.ctx, self.subcloud, payload=values)
|
|
|
|
# Verify that subcloud has the correct deploy status
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.PRESTAGE_STATE_COMPLETE,
|
|
updated_subcloud.prestage_status)
|
|
|
|
# Verify both of prestage package and image ansible playbooks were called
|
|
self.assertEqual(mock_run_ansible.call_count, 2)
|
|
# Verify the "image_list_file" was passed to the prestage image playbook
|
|
# for the local prestage
|
|
self.assertTrue(
|
|
'image_list_file' in mock_run_ansible.call_args_list[1].args[1][5])
|
|
# Verify the prestage request release was passed to the playbooks
|
|
self.assertTrue(
|
|
FAKE_PRESTAGE_RELEASE in mock_run_ansible.call_args_list[0].args[1][5])
|
|
self.assertTrue(
|
|
FAKE_PRESTAGE_RELEASE in mock_run_ansible.call_args_list[1].args[1][5])
|
|
|
|
@mock.patch.object(cutils, 'get_filename_by_prefix')
|
|
@mock.patch.object(prestage, '_run_ansible')
|
|
def test_prestage_local_pass_without_img_list(self, mock_run_ansible,
|
|
mock_get_filename_by_prefix):
|
|
|
|
values = copy.copy(FAKE_PRESTAGE_PAYLOAD)
|
|
self.subcloud.update(
|
|
{'deploy_status': consts.DEPLOY_STATE_NONE,
|
|
'software_version': FAKE_PRESTAGE_RELEASE}
|
|
)
|
|
|
|
mock_run_ansible.return_value = None
|
|
mock_get_filename_by_prefix.return_value = None
|
|
self.mock_os_path_isdir.return_value = True
|
|
prestage._prestage_standalone_thread(self.ctx, self.subcloud, payload=values)
|
|
|
|
# Verify that subcloud has the correct deploy status
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.PRESTAGE_STATE_COMPLETE,
|
|
updated_subcloud.prestage_status)
|
|
|
|
# Verify both of prestage package and image ansible playbooks were called
|
|
self.assertEqual(mock_run_ansible.call_count, 2)
|
|
# Verify the "image_list_file" was not passed to the prestage image playbook
|
|
# for the local prestage
|
|
self.assertTrue(
|
|
'image_list_file' not in mock_run_ansible.call_args_list[1].args[1][5])
|
|
# Verify the prestage request release was passed to the playbooks
|
|
self.assertIn(FAKE_PRESTAGE_RELEASE,
|
|
mock_run_ansible.call_args_list[0].args[1][5])
|
|
self.assertIn(FAKE_PRESTAGE_RELEASE,
|
|
mock_run_ansible.call_args_list[1].args[1][5])
|
|
|
|
@mock.patch.object(prestage, 'prestage_images')
|
|
@mock.patch.object(prestage, 'prestage_packages')
|
|
@mock.patch.object(prestage, '_run_ansible')
|
|
def test_prestage_subcloud_complete(self,
|
|
mock_run_ansible,
|
|
mock_prestage_packages,
|
|
mock_prestage_images):
|
|
|
|
values = copy.copy(FAKE_PRESTAGE_PAYLOAD)
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_NONE
|
|
prestage._prestage_standalone_thread(self.ctx, self.subcloud, payload=values)
|
|
mock_run_ansible.return_value = None
|
|
mock_prestage_packages.assert_called_once_with(
|
|
self.ctx, self.subcloud, values
|
|
)
|
|
mock_prestage_images.assert_called_once_with(self.ctx, self.subcloud, values)
|
|
self.mock_delete_subcloud_inventory.return_value = None
|
|
|
|
# Verify that subcloud has the "prestage-complete" deploy status
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.PRESTAGE_STATE_COMPLETE,
|
|
updated_subcloud.prestage_status)
|
|
|
|
def test_get_cached_regionone_data(self):
|
|
mock_keystone_client = FakeKeystoneClient()
|
|
mock_sysinv_client = FakeSysinvClient()
|
|
cached_regionone_data = self.sm._get_cached_regionone_data(
|
|
mock_keystone_client, mock_sysinv_client)
|
|
expiry1 = cached_regionone_data['expiry']
|
|
self.assertEqual(cached_regionone_data['dcmanager_user_id'],
|
|
FAKE_DCMANAGER_USER_ID)
|
|
self.assertEqual(cached_regionone_data['admin_project_id'],
|
|
FAKE_ADMIN_PROJECT_ID)
|
|
self.assertEqual(cached_regionone_data['mgmt_interface_uuids'],
|
|
FAKE_MGMT_IF_UUIDS)
|
|
self.assertEqual(cached_regionone_data['mgmt_pool'].floating_address,
|
|
'fdff:719a:bf60:233::2')
|
|
self.assertEqual(cached_regionone_data['oam_addresses'].oam_floating_ip,
|
|
'2620:10a:a001:d41::260')
|
|
# The expiry timestamp is likely a couple of seconds less than the time
|
|
# the cache is set when it gets here so check if the expiry is greater than
|
|
# 59m55s from now.
|
|
self.assertGreater(
|
|
cached_regionone_data['expiry'],
|
|
datetime.datetime.utcnow() + datetime.timedelta(seconds=3595))
|
|
cached_regionone_data = self.sm._get_cached_regionone_data(
|
|
mock_keystone_client, mock_sysinv_client)
|
|
expiry2 = cached_regionone_data['expiry']
|
|
self.assertEqual(expiry1, expiry2)
|
|
|
|
|
|
class TestSubcloudBackupRestore(BaseTestSubcloudManager):
|
|
"""Test class for testing subcloud backup restore"""
|
|
|
|
def setUp(self):
|
|
super().setUp()
|
|
self._mock_builtins_open()
|
|
self.values = copy.copy(FAKE_BACKUP_RESTORE_LOAD)
|
|
self.data_install = str(fake_subcloud.FAKE_SUBCLOUD_INSTALL_VALUES).replace(
|
|
'\'', '"')
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_run_subcloud_backup_restore_playbook')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_overrides_for_backup_or_restore')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_subcloud_inventory_file')
|
|
def test_backup_restore_unmanaged_online(self,
|
|
mock_create_inventory_file,
|
|
mock_create_overrides,
|
|
mock_run_playbook):
|
|
mock_create_inventory_file.return_value = 'inventory_file.yml'
|
|
mock_create_overrides.return_value = 'overrides_file.yml'
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_DONE,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE,
|
|
management_state=dccommon_consts.MANAGEMENT_UNMANAGED,
|
|
data_install=self.data_install)
|
|
|
|
self.sm.restore_subcloud_backups(self.ctx, payload=self.values)
|
|
|
|
mock_create_inventory_file.assert_called_once()
|
|
mock_create_overrides.assert_called_once()
|
|
mock_run_playbook.assert_called_once()
|
|
|
|
# Verify that subcloud has the correct deploy status
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_PRE_RESTORE,
|
|
updated_subcloud.deploy_status)
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_run_subcloud_backup_restore_playbook')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_overrides_for_backup_or_restore')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_subcloud_inventory_file')
|
|
def test_backup_restore_unmanaged_offline(self,
|
|
mock_create_inventory_file,
|
|
mock_create_overrides,
|
|
mock_run_playbook
|
|
):
|
|
|
|
values = copy.copy(FAKE_BACKUP_RESTORE_LOAD)
|
|
values['local_only'] = False
|
|
values['registry_images'] = False
|
|
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_NONE,
|
|
availability_status=dccommon_consts.AVAILABILITY_OFFLINE,
|
|
management_state=dccommon_consts.MANAGEMENT_UNMANAGED,
|
|
data_install=self.data_install)
|
|
|
|
self.sm.restore_subcloud_backups(self.ctx, payload=values)
|
|
|
|
mock_create_inventory_file.assert_called_once()
|
|
mock_create_overrides.assert_called_once()
|
|
mock_run_playbook.assert_called_once()
|
|
|
|
# Verify that subcloud has the correct deploy status
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_PRE_RESTORE,
|
|
updated_subcloud.deploy_status)
|
|
|
|
def test_backup_restore_managed_offline(self):
|
|
"""Subcloud must be unmanaged for backup restore operation."""
|
|
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_DONE,
|
|
availability_status=dccommon_consts.AVAILABILITY_OFFLINE,
|
|
management_state=dccommon_consts.MANAGEMENT_MANAGED)
|
|
|
|
return_log = self.sm.restore_subcloud_backups(self.ctx, payload=self.values)
|
|
expected_log = 'skipped for local backup restore operation'
|
|
|
|
self.assertIn(expected_log, return_log)
|
|
self.mock_log.info.assert_called_with(
|
|
'Subcloud restore backup operation finished.\nRestored subclouds: 0.'
|
|
' Invalid subclouds: 1. Failed subclouds: 0.')
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_run_subcloud_backup_restore_playbook')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_overrides_for_backup_or_restore')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_subcloud_inventory_file')
|
|
def test_backup_restore_with_install(self,
|
|
mock_create_inventory_file,
|
|
mock_create_overrides,
|
|
mock_run_restore_playbook
|
|
):
|
|
self._mock_subcloud_manager_run_subcloud_install()
|
|
self.mock_os_path_isdir.return_value = True
|
|
self.mock_os_listdir.return_value = ['test.iso', 'test.sig']
|
|
mock_create_inventory_file.return_value = 'inventory_file.yml'
|
|
mock_create_overrides.return_value = 'overrides_file.yml'
|
|
self.mock_run_subcloud_install.return_value = True
|
|
mock_run_restore_playbook.return_value = True
|
|
|
|
values = copy.copy(FAKE_BACKUP_RESTORE_LOAD_WITH_INSTALL)
|
|
values['with_install'] = True
|
|
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
data_install=self.data_install,
|
|
deploy_status=consts.DEPLOY_STATE_DONE,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE,
|
|
management_state=dccommon_consts.MANAGEMENT_UNMANAGED)
|
|
|
|
self.sm.restore_subcloud_backups(self.ctx, payload=values)
|
|
|
|
mock_create_inventory_file.assert_called_once()
|
|
mock_create_overrides.assert_called_once()
|
|
|
|
# Verify that subcloud has the correct deploy status
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_PRE_RESTORE,
|
|
updated_subcloud.deploy_status)
|
|
|
|
@mock.patch.object(subcloud_install.SubcloudInstall, 'prep')
|
|
def test_backup_restore_with_install_failed(self, mock_prep):
|
|
# FAILED installing playbook of (subcloud1).
|
|
# Backup restore failed for all applied subclouds.
|
|
self.mock_os_path_isdir.return_value = True
|
|
self.mock_os_listdir.return_value = ['test.iso', 'test.sig']
|
|
self.mock_ansible_run_playbook.side_effect = PlaybookExecutionFailed()
|
|
values = copy.copy(FAKE_BACKUP_RESTORE_LOAD_WITH_INSTALL)
|
|
values['with_install'] = True
|
|
values['local_only'] = False
|
|
values['registry_images'] = True
|
|
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
data_install=self.data_install,
|
|
deploy_status=consts.DEPLOY_STATE_DONE,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE,
|
|
management_state=dccommon_consts.MANAGEMENT_UNMANAGED)
|
|
|
|
self.assertRaises(exceptions.SubcloudBackupOperationFailed,
|
|
self.sm.restore_subcloud_backups, self.ctx, payload=values)
|
|
# Verify that subcloud has the correct deploy status
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_INSTALL_FAILED,
|
|
updated_subcloud.deploy_status)
|
|
|
|
def test_backup_restore_unmanage_online_complete_restore_val(self):
|
|
self.values['local_only'] = True
|
|
self.values['registry_images'] = True
|
|
self.values['restore_values'] = {
|
|
'bootstrap_address': {
|
|
'subcloud1': '10.10.20.12'
|
|
}
|
|
}
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_DONE,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE,
|
|
management_state=dccommon_consts.MANAGEMENT_UNMANAGED,
|
|
data_install=self.data_install)
|
|
|
|
self.sm.restore_subcloud_backups(self.ctx, payload=self.values)
|
|
# Verify that subcloud has the correct deploy status
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_DONE,
|
|
updated_subcloud.deploy_status)
|
|
Calls = [mock.call('Backup restore: Received restore_values for'
|
|
' subcloud subcloud1'),
|
|
mock.call('Successfully restore subcloud subcloud1'),
|
|
mock.call('Processed subcloud subcloud1 for backup restore '
|
|
'(operation 100% complete, 0 subcloud(s) remaining)'),
|
|
mock.call('Subcloud restore backup operation finished.\nRestored'
|
|
' subclouds: 1. Invalid subclouds: 0. '
|
|
'Failed subclouds: 0.')]
|
|
self.mock_log.info.assert_has_calls(Calls)
|
|
|
|
def test_backup_restore_unmanage_online_complete_backup_val(self):
|
|
self.values['local_only'] = True
|
|
self.values['registry_images'] = True
|
|
self.values['backup_values'] = copy.copy(FAKE_BACKUP_CREATE_LOAD)
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_DONE,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE,
|
|
management_state=dccommon_consts.MANAGEMENT_UNMANAGED,
|
|
data_install=self.data_install)
|
|
|
|
self.sm.restore_subcloud_backups(self.ctx, payload=self.values)
|
|
|
|
# Verify that subcloud has the correct deploy status
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_DONE,
|
|
updated_subcloud.deploy_status)
|
|
Calls = [mock.call('Successfully restore subcloud subcloud1'),
|
|
mock.call('Processed subcloud subcloud1 for backup restore'
|
|
' (operation 100% complete, 0 subcloud(s) remaining)'),
|
|
mock.call('Subcloud restore backup operation finished.'
|
|
'\nRestored subclouds: 1. Invalid subclouds:'
|
|
' 0. Failed subclouds: 0.')]
|
|
self.mock_log.info.assert_has_calls(Calls)
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_create_subcloud_inventory_file')
|
|
def test_backup_restore_failed(self, mock_create_inventory_file):
|
|
mock_create_inventory_file.side_effect = Exception('boom')
|
|
self.values['local_only'] = True
|
|
self.values['registry_images'] = True
|
|
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_DONE,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE,
|
|
management_state=dccommon_consts.MANAGEMENT_UNMANAGED,
|
|
data_install=self.data_install)
|
|
self.assertRaises(exceptions.SubcloudBackupOperationFailed,
|
|
self.sm.restore_subcloud_backups,
|
|
self.ctx, payload=self.values)
|
|
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_RESTORE_PREP_FAILED,
|
|
updated_subcloud.deploy_status)
|
|
self.mock_log.exception.assert_called_once_with(
|
|
f"Failed to prepare subcloud {self.subcloud.name} for backup restore")
|
|
|
|
def test_backup_restore_playbook_execution_failed(self):
|
|
self.mock_ansible_run_playbook.side_effect = PlaybookExecutionFailed()
|
|
self.values['local_only'] = True
|
|
self.values['registry_images'] = True
|
|
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_DONE,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE,
|
|
management_state=dccommon_consts.MANAGEMENT_UNMANAGED,
|
|
data_install=self.data_install)
|
|
self.assertRaises(exceptions.SubcloudBackupOperationFailed,
|
|
self.sm.restore_subcloud_backups,
|
|
self.ctx, payload=self.values)
|
|
# Verify that subcloud has the correct deploy status
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_RESTORE_FAILED,
|
|
updated_subcloud.deploy_status)
|
|
|
|
|
|
class TestSubcloudMigrate(BaseTestSubcloudManager):
|
|
"""Test class for testing subcloud migrate"""
|
|
|
|
def setUp(self):
|
|
super().setUp()
|
|
|
|
self.fake_bootstrap_values = \
|
|
"{'name': 'TestSubcloud', 'system_mode': 'simplex'}"
|
|
self.saved_payload = {
|
|
"deploy_status": consts.DEPLOY_STATE_DONE,
|
|
"rehome_data": '{"saved_payload": {"system_mode": "simplex",\
|
|
"name": "testsub", "bootstrap-address": "128.224.119.56",\
|
|
"admin_password": "TGk2OW51eA=="}}',
|
|
}
|
|
|
|
@mock.patch.object(subcloud_manager, 'db_api', side_effect=db_api)
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'generate_subcloud_ansible_config')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager, 'rehome_subcloud')
|
|
def test_migrate_subcloud(self, mock_rehome_subcloud,
|
|
mock_generate_subcloud_ansible_config,
|
|
mock_db_api):
|
|
# Prepare the test data
|
|
payload = {
|
|
"sysadmin_password": "TGk2OW51eA=="
|
|
}
|
|
payload_result = {
|
|
"name": self.subcloud.name,
|
|
"deploy_status": "secondary",
|
|
"rehome_data": {
|
|
"saved_payload": {
|
|
"system_mode": "simplex",
|
|
"name": "testsub",
|
|
"bootstrap-address": "128.224.119.56",
|
|
"sysadmin_password": "Li69nux",
|
|
"admin_password": "Li69nux"
|
|
}
|
|
},
|
|
}
|
|
db_api.subcloud_update(self.ctx, self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_SECONDARY,
|
|
rehome_data=self.saved_payload['rehome_data'])
|
|
self.sm.migrate_subcloud(self.ctx, self.subcloud.id, payload)
|
|
|
|
mock_generate_subcloud_ansible_config.assert_called_once_with(
|
|
mock.ANY, payload_result['rehome_data']['saved_payload'])
|
|
mock_rehome_subcloud.assert_called_once_with(mock.ANY, mock.ANY)
|
|
mock_db_api.subcloud_update.assert_called_once_with(
|
|
mock.ANY, mock.ANY, deploy_status=consts.DEPLOY_STATE_PRE_REHOME)
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_run_parallel_group_operation')
|
|
def test_batch_migrate_subcloud(self, mock_run_parallel_group_operation):
|
|
# Prepare the test data
|
|
rehome_data = '{"saved_payload": {"system_mode": "simplex",\
|
|
"name": "test_sub_migrate", "bootstrap-address": "128.224.119.56"}}'
|
|
payload = {
|
|
"sysadmin_password": self._create_password(),
|
|
"peer_group": self.peer_group.peer_group_name
|
|
}
|
|
subcloud = self.create_subcloud_static(
|
|
self.ctx,
|
|
name="sub_migrateable",
|
|
deploy_status=consts.DEPLOY_STATE_SECONDARY
|
|
)
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
subcloud.id,
|
|
peer_group_id=self.peer_group.id,
|
|
rehome_data=rehome_data,
|
|
management_state=dccommon_consts.MANAGEMENT_UNMANAGED)
|
|
subcloud = self.create_subcloud_static(
|
|
self.ctx,
|
|
name="sub_no_rehome_data",
|
|
deploy_status=consts.DEPLOY_STATE_SECONDARY
|
|
)
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
subcloud.id,
|
|
peer_group_id=self.peer_group.id,
|
|
management_state=dccommon_consts.MANAGEMENT_UNMANAGED)
|
|
subcloud = self.create_subcloud_static(
|
|
self.ctx,
|
|
name="sub_no_secondary"
|
|
)
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
subcloud.id,
|
|
peer_group_id=self.peer_group.id,
|
|
rehome_data=rehome_data,
|
|
management_state=dccommon_consts.MANAGEMENT_UNMANAGED)
|
|
subcloud = self.create_subcloud_static(
|
|
self.ctx,
|
|
name="sub_no_saved_payload"
|
|
)
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
subcloud.id,
|
|
peer_group_id=self.peer_group.id,
|
|
rehome_data="{}",
|
|
management_state=dccommon_consts.MANAGEMENT_UNMANAGED)
|
|
|
|
self.sm.batch_migrate_subcloud(self.ctx, payload)
|
|
mock_run_parallel_group_operation.assert_called_with(
|
|
"migrate", mock.ANY, mock.ANY, mock.ANY)
|
|
actual_args, _ = mock_run_parallel_group_operation.call_args
|
|
expect_subclouds = actual_args[3]
|
|
self.assertEqual(1, len(expect_subclouds))
|
|
self.assertEqual("sub_migrateable", expect_subclouds[0].name)
|
|
|
|
def test_batch_migrate_subcloud_failed(self):
|
|
# Prepare the test data
|
|
payload = {}
|
|
self.sm.batch_migrate_subcloud(self.ctx, payload)
|
|
self.mock_log.error.assert_called_once_with(
|
|
"Failed to migrate subcloud peer group, missing peer_group")
|
|
|
|
def test_batch_migrate_subcloud_failed_password(self):
|
|
# Prepare the test data
|
|
payload = {
|
|
"peer_group": self.peer_group.peer_group_name
|
|
}
|
|
self.sm.batch_migrate_subcloud(self.ctx, payload)
|
|
self.mock_log.error.assert_called_once_with(
|
|
"Failed to migrate subcloud peer group, missing sysadmin_password")
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_unmanage_system_peer_subcloud')
|
|
def test_migrate_manage_subcloud_called_unmanage_peer_subcloud(
|
|
self, mock_unmanage_system_peer_subcloud
|
|
):
|
|
self.sm._migrate_manage_subcloud(self.ctx, mock.ANY, [self.system_peer],
|
|
self.subcloud)
|
|
mock_unmanage_system_peer_subcloud.assert_called()
|
|
|
|
@mock.patch.object(system_peer_manager.SystemPeerManager,
|
|
'get_peer_dc_client')
|
|
@mock.patch.object(dcmanager_v1.DcmanagerClient, 'update_subcloud')
|
|
def test_unmanage_system_peer_subcloud_failed(self, mock_update_subcloud,
|
|
mock_peer_dc_client):
|
|
mock_peer_dc_client.return_value = mock.MagicMock
|
|
db_api.subcloud_update(
|
|
self.ctx, self.subcloud.id,
|
|
management_state=dccommon_consts.MANAGEMENT_UNMANAGED)
|
|
ret = self.sm._unmanage_system_peer_subcloud(
|
|
[self.system_peer], self.subcloud)
|
|
self.assertEqual(ret, False)
|
|
self.mock_log.exception.assert_called_with(
|
|
f'Failed to set unmanged for subcloud: {self.subcloud.region_name}'
|
|
f' on system {self.system_peer.peer_name} attempt: 2')
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_unmanage_system_peer_subcloud')
|
|
def test_migrate_manage_subcloud_not_called_unmanage_peer_subcloud(
|
|
self, mock_unmanage_system_peer_subcloud
|
|
):
|
|
# Give empty system peers
|
|
system_peers = []
|
|
self.sm._migrate_manage_subcloud(self.ctx,
|
|
mock.ANY, system_peers,
|
|
self.subcloud)
|
|
mock_unmanage_system_peer_subcloud.assert_not_called()
|
|
|
|
@mock.patch.object(system_peer_manager.SystemPeerManager,
|
|
'get_peer_dc_client')
|
|
def test_unmanage_system_peer_subcloud_ret_false(self,
|
|
mock_get_peer_dc_client):
|
|
mock_get_peer_dc_client.update_subcloud.return_value = None
|
|
ret = self.sm._unmanage_system_peer_subcloud(
|
|
[self.system_peer], self.subcloud)
|
|
self.assertEqual(ret, False)
|
|
mock_get_peer_dc_client().update_subcloud.assert_called()
|
|
|
|
@mock.patch.object(system_peer_manager.SystemPeerManager,
|
|
'get_peer_dc_client')
|
|
@mock.patch.object(dcmanager_v1.DcmanagerClient, 'update_subcloud')
|
|
def test_unmanage_system_peer_subcloud_ret_true(self,
|
|
mock_update_subcloud,
|
|
mock_get_peer_dc_client):
|
|
mock_update_subcloud.return_value = mock.MagicMock()
|
|
mock_get_peer_dc_client().get_subcloud.side_effect = \
|
|
exceptions.SubcloudNotUnmanaged()
|
|
ret = self.sm._unmanage_system_peer_subcloud(
|
|
[self.system_peer], self.subcloud)
|
|
self.assertEqual(ret, True)
|
|
mock_get_peer_dc_client().update_subcloud.assert_not_called()
|
|
|
|
@mock.patch.object(system_peer_manager.SystemPeerManager,
|
|
'get_peer_dc_client')
|
|
@mock.patch.object(dcmanager_v1.DcmanagerClient, 'update_subcloud')
|
|
def test_unmanage_system_peer_subcloud_update_subcloud(
|
|
self, mock_update_subcloud, mock_get_peer_dc_client
|
|
):
|
|
mock_get_peer_dc_client.return_value = mock.MagicMock()
|
|
ret = self.sm._unmanage_system_peer_subcloud(
|
|
[self.system_peer], self.subcloud)
|
|
self.assertEqual(ret, False)
|
|
mock_get_peer_dc_client().update_subcloud.assert_called_once()
|
|
|
|
@mock.patch.object(subcloud_manager.SubcloudManager, 'subcloud_deploy_create')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager, 'rehome_subcloud')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager, 'run_deploy_phases')
|
|
@mock.patch.object(subcloud_manager, 'db_api')
|
|
def test_add_subcloud_with_secondary_option(self, mock_db_api,
|
|
mock_run_deploy_phases,
|
|
mock_rehome_subcloud,
|
|
mock_subcloud_deploy_create):
|
|
# Prepare the test data
|
|
values = {
|
|
'name': 'TestSubcloud',
|
|
'sysadmin_password': '123',
|
|
'secondary': 'true',
|
|
'region_name': '2ec93dfb654846909efe61d1b39dd2ce'
|
|
}
|
|
|
|
# Call add_subcloud method with the test data
|
|
self.sm.add_subcloud(mock.MagicMock(), 1, values)
|
|
|
|
# Assert that the rehome_subcloud and run_deploy_phases methods were not
|
|
# called
|
|
mock_rehome_subcloud.assert_not_called()
|
|
mock_run_deploy_phases.assert_not_called()
|
|
|
|
mock_subcloud_deploy_create.assert_called_once()
|
|
|
|
# Assert that db_api.subcloud_update was not called for secondary subcloud
|
|
self.assertFalse(mock_db_api.subcloud_update.called)
|
|
|
|
def test_update_subcloud_bootstrap_address(self):
|
|
fake_result = ('{"saved_payload": {"name": "TestSubcloud", '
|
|
'"system_mode": "simplex", '
|
|
'"bootstrap-address": "123.123.123.123"}}')
|
|
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_DONE,
|
|
availability_status=dccommon_consts.AVAILABILITY_ONLINE)
|
|
|
|
self.sm.update_subcloud(self.ctx,
|
|
self.subcloud.id,
|
|
bootstrap_values=self.fake_bootstrap_values)
|
|
self.sm.update_subcloud(self.ctx,
|
|
self.subcloud.id,
|
|
bootstrap_address="123.123.123.123")
|
|
|
|
# Verify subcloud was updated with correct values
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(fake_result,
|
|
updated_subcloud.rehome_data)
|
|
|
|
@mock.patch.object(db_api, 'subcloud_peer_group_update')
|
|
@mock.patch.object(subcloud_manager.SubcloudManager,
|
|
'_run_parallel_group_operation')
|
|
def test_run_batch_migrate(self, mock_run_parallel_group_operation,
|
|
mock_subcloud_peer_group_update):
|
|
self._mock_builtins_open()
|
|
self.mock_get_local_system.return_value = FakeSysinvClient.get_system(self)
|
|
self.mock_openstack_driver.keystone_client = FakeKeystoneClient()
|
|
db_api.subcloud_peer_group_update(
|
|
self.ctx,
|
|
group_id=1,
|
|
peer_group_name="dc1-pg",
|
|
group_priority=0,
|
|
group_state="enabled",
|
|
system_leader_id="ac62f555-9386-42f1-b3a1-51ecb709409d",
|
|
system_leader_name="dc1-name",
|
|
migration_status=None)
|
|
self.saved_payload.update({
|
|
"name": self.subcloud.name, "deploy_status": "secondary",
|
|
"admin_password": self._create_password()
|
|
})
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_SECONDARY,
|
|
peer_group_id=self.peer_group.id,
|
|
rehome_data=self.saved_payload['rehome_data'],
|
|
management_state=dccommon_consts.MANAGEMENT_UNMANAGED)
|
|
|
|
self.sm.run_batch_migrate(self.ctx, self.peer_group, "TGk2OW51eA==")
|
|
Calls = [mock.call('No association found for peer group pgname'),
|
|
mock.call('Batch migrate operation finished')]
|
|
self.mock_log.info.assert_has_calls(Calls)
|
|
|
|
@mock.patch.object(db_api, 'subcloud_peer_group_update')
|
|
def test_run_batch_migrate_no_secondary_subclouds(
|
|
self, mock_subcloud_peer_group_update
|
|
):
|
|
self.mock_sysinv_client().return_value = FakeSysinvClient()
|
|
db_api.subcloud_peer_group_update(
|
|
self.ctx,
|
|
group_id=1,
|
|
peer_group_name="dc1-pg",
|
|
group_priority=0,
|
|
group_state="enabled",
|
|
system_leader_id="ac62f555-9386-42f1-b3a1-51ecb709409d",
|
|
system_leader_name="dc1-name",
|
|
migration_status=None)
|
|
self.saved_payload.update({
|
|
"name": self.subcloud.name, "deploy_status": "secondary",
|
|
"admin_password": self._create_password()
|
|
})
|
|
db_api.subcloud_update(
|
|
self.ctx,
|
|
self.subcloud.id,
|
|
peer_group_id=self.peer_group.id,
|
|
rehome_data=self.saved_payload['rehome_data'],
|
|
management_state=dccommon_consts.MANAGEMENT_UNMANAGED)
|
|
|
|
sysadmin_password = self._create_password()
|
|
|
|
self.sm.run_batch_migrate(self.ctx, self.peer_group, sysadmin_password)
|
|
Calls = [mock.call('Skipping subcloud subcloud1 from batch migration:'
|
|
' subcloud deploy_status is not in secondary,'
|
|
' rehome-failed or rehome-prep-failed'),
|
|
mock.call('No subclouds to be migrated in peer group:'
|
|
' pgname ending migration attempt')]
|
|
self.mock_log.info.assert_has_calls(Calls)
|
|
|
|
def test_migrate_subcloud_failed_non_existent_subcloud(self):
|
|
payload = {}
|
|
self.sm.migrate_subcloud(self.ctx, 'subcloud2', payload)
|
|
self.mock_log.error.assert_called_once_with(
|
|
"Failed to migrate, non-existent subcloud subcloud2")
|
|
|
|
def test_migrate_subcloud_failed_sysadmin_pswd_not_provided(self):
|
|
self.saved_payload['name'] = self.subcloud.name
|
|
payload = {}
|
|
db_api.subcloud_update(self.ctx, self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_DONE,
|
|
rehome_data=self.saved_payload['rehome_data'])
|
|
self.sm.migrate_subcloud(self.ctx, self.subcloud.id, payload)
|
|
self.mock_log.error.assert_called_once_with(
|
|
"Failed to migrate subcloud: subcloud1, must provide sysadmin_password")
|
|
|
|
def test_migrate_subcloud_failed_due_to_deploy_state(self):
|
|
self.saved_payload['name'] = self.subcloud.name
|
|
payload = {"sysadmin_password": self._create_password()}
|
|
db_api.subcloud_update(self.ctx, self.subcloud.id,
|
|
deploy_status=consts.DEPLOY_STATE_DONE,
|
|
rehome_data=self.saved_payload['rehome_data'])
|
|
self.sm.migrate_subcloud(self.ctx, self.subcloud.id, payload)
|
|
self.mock_log.error.assert_called_once_with(
|
|
f'Failed to migrate subcloud: {self.subcloud.name}, must be in '
|
|
'secondary or rehome failure state')
|
|
|
|
def test_rehome_subcloud_failed(self):
|
|
self._mock_utils_get_playbook_for_software_version()
|
|
self.mock_get_playbook_for_software_version.return_value = SW_VERSION
|
|
self.mock_subprocess_run.return_value.returncode = 0
|
|
self.mock_ansible_run_playbook.side_effect = PlaybookExecutionFailed()
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_NONE
|
|
|
|
self.sm.rehome_subcloud(self.ctx, self.subcloud)
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx,
|
|
self.subcloud.name)
|
|
self.assertEqual(consts.DEPLOY_STATE_REHOME_FAILED,
|
|
updated_subcloud.deploy_status)
|
|
|
|
|
|
class TestSubcloudInstall(BaseTestSubcloudManager):
|
|
"""Test class for testing subcloud install"""
|
|
|
|
def setUp(self):
|
|
super().setUp()
|
|
self._mock_builtins_open()
|
|
self._mock_openstack_driver(subcloud_install)
|
|
self._mock_sysinv_client(subcloud_install)
|
|
self._mock_subcloud_manager_compose_install_command()
|
|
self.fake_install_values = \
|
|
copy.copy(fake_subcloud.FAKE_SUBCLOUD_INSTALL_VALUES)
|
|
self.fake_install_values['software_version'] = SW_VERSION
|
|
self.fake_log_file = (
|
|
os.path.join(
|
|
consts.DC_ANSIBLE_LOG_DIR,
|
|
fake_subcloud.FAKE_SUBCLOUD_DATA['name']) + "_playbook_output.log"
|
|
)
|
|
self.mock_sysinv_client().get_oam_addresses.return_value = \
|
|
FakeOamAddresses()
|
|
self.mock_subprocess_run.return_value.returncode = 0
|
|
self.subcloud.update(
|
|
{'name': fake_subcloud.FAKE_SUBCLOUD_DATA['name'],
|
|
'deploy_status': consts.DEPLOY_STATE_PRE_INSTALL}
|
|
)
|
|
|
|
@mock.patch.object(request, 'urlretrieve')
|
|
def test_subcloud_install(self,
|
|
mock_request_urlretrieve):
|
|
mock_request_urlretrieve.return_value = "fake_path", "empty"
|
|
self.mock_ansible_run_playbook.return_value = False
|
|
|
|
install_success = self.sm._run_subcloud_install(
|
|
self.ctx, self.subcloud, self.mock_compose_install_command,
|
|
self.fake_log_file, self.fake_install_values
|
|
)
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx,
|
|
self.subcloud.name)
|
|
self.assertEqual(
|
|
consts.DEPLOY_STATE_INSTALLING, updated_subcloud.deploy_status
|
|
)
|
|
self.mock_ansible_run_playbook.assert_called_once()
|
|
self.assertTrue(install_success)
|
|
Calls = [mock.call('Preparing remote install of %s, version: %s',
|
|
self.subcloud.name, SW_VERSION),
|
|
mock.call(f'Starting remote install of {self.subcloud.name}'),
|
|
mock.call(f'Successfully installed {self.subcloud.name}')]
|
|
self.mock_log.info.assert_has_calls(Calls)
|
|
|
|
def test_subcloud_install_prep_failed(self):
|
|
install_success = self.sm._run_subcloud_install(
|
|
self.ctx, self.subcloud, self.mock_compose_install_command,
|
|
self.fake_log_file, self.fake_install_values
|
|
)
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(
|
|
consts.DEPLOY_STATE_PRE_INSTALL_FAILED, updated_subcloud.deploy_status
|
|
)
|
|
self.assertFalse(install_success)
|
|
|
|
@mock.patch.object(request, 'urlretrieve')
|
|
def test_subcloud_install_with_aborted_install(
|
|
self, mock_request_urlretrieve
|
|
):
|
|
mock_request_urlretrieve.return_value = "fake_path", "empty"
|
|
self.mock_ansible_run_playbook.return_value = True
|
|
self.subcloud['deploy_status'] = consts.DEPLOY_STATE_PRE_INSTALL_FAILED
|
|
|
|
install_success = self.sm._run_subcloud_install(
|
|
self.ctx, self.subcloud, self.mock_compose_install_command,
|
|
self.fake_log_file, self.fake_install_values
|
|
)
|
|
updated_subcloud = db_api.subcloud_get_by_name(self.ctx, self.subcloud.name)
|
|
self.assertEqual(
|
|
consts.DEPLOY_STATE_INSTALLING, updated_subcloud.deploy_status
|
|
)
|
|
self.mock_ansible_run_playbook.assert_called_once()
|
|
self.assertFalse(install_success)
|
|
|
|
|
|
class TestSubcloudRename(BaseTestSubcloudManager):
|
|
"""Test class for testing rename subcloud"""
|
|
|
|
def setUp(self):
|
|
super().setUp()
|
|
self._mock_builtins_open()
|
|
self.new_subcloud_name = 'testsubcloud'
|
|
self.alarms = test_subcloud_alarms.DBAPISubcloudAlarm
|
|
self.alarms.create_subcloud_alarms(self.ctx, self.subcloud.name)
|
|
|
|
@mock.patch.object(yaml, 'safe_load')
|
|
@mock.patch.object(filecmp, 'cmp')
|
|
@mock.patch.object(os, 'rename')
|
|
def test_rename_subcloud_ansible_files(
|
|
self, mock_rename, mock_filecmp, mock_load
|
|
):
|
|
cur_file = 'subcloud1.yml'
|
|
self.mock_os_listdir.return_value = [cur_file]
|
|
data = {'key1': 'value1', 'key2': 'value2'}
|
|
mock_load.return_value = data
|
|
self.mock_builtins_open().return_value = \
|
|
f'{ANS_PATH}/subcloud1_inventory.yml'
|
|
rehome_data = '{"saved_payload": {"system_mode": "simplex",\
|
|
"name": "test_sub_migrate", "bootstrap-address": "128.224.119.56"}}'
|
|
db_api.subcloud_update(self.ctx, self.subcloud.id,
|
|
rehome_data=rehome_data)
|
|
self.sm.rename_subcloud(
|
|
self.ctx, self.subcloud.id, self.subcloud.name, self.new_subcloud_name
|
|
)
|
|
ret = db_api.subcloud_get_by_name(self.ctx, self.new_subcloud_name)
|
|
self.assertEqual(self.new_subcloud_name, ret.name)
|
|
Calls = [
|
|
mock.call(f'{dccommon_consts.ANSIBLE_OVERRIDES_PATH}/subcloud1.yml',
|
|
f'{dccommon_consts.ANSIBLE_OVERRIDES_PATH}/testsubcloud.yml'),
|
|
mock.call(f'{consts.DC_ANSIBLE_LOG_DIR}/subcloud1.yml',
|
|
f'{consts.DC_ANSIBLE_LOG_DIR}/testsubcloud.yml')]
|
|
mock_rename.assert_has_calls(Calls)
|
|
|
|
@mock.patch.object(filecmp, 'cmp')
|
|
def test_rename_subcloud_inventory_file_failed(self, mock_filecmp):
|
|
self.mock_os_path_exists.return_value = False
|
|
self.mock_os_listdir.return_value = ['testfile1', 'testfile2']
|
|
|
|
db_api.subcloud_update(
|
|
self.ctx, self.subcloud.id, self.subcloud.name
|
|
)
|
|
self.sm.rename_subcloud(
|
|
self.ctx, self.subcloud.id, self.subcloud.name, self.new_subcloud_name
|
|
)
|
|
ret = db_api.subcloud_get_by_name(self.ctx, self.new_subcloud_name)
|
|
self.mock_log.warn.assert_called_once_with(
|
|
'Could not rename inventory file '
|
|
f'{dccommon_consts.ANSIBLE_OVERRIDES_PATH}/'
|
|
'testsubcloud_inventory.yml because it does not exist.'
|
|
)
|
|
self.assertEqual(self.new_subcloud_name, ret.name)
|
|
self.assertEqual(self.mock_os_listdir.call_count, 2)
|
|
|
|
|
|
class TestSubcloudEnrollment(BaseTestSubcloudManager):
|
|
"""Test class for testing Subcloud Enrollment"""
|
|
|
|
# TODO(srana): Add unit tests for SubcloudEnrollmentInit
|
|
# prep and enroll_init methods
|
|
|
|
def setUp(self):
|
|
super().setUp()
|
|
self.rel_version = '24.09'
|
|
self.subcloud_name = 'test_subcloud'
|
|
self.iso_dir = (f'/opt/platform/iso/{self.rel_version}/'
|
|
f'nodes/{self.subcloud_name}')
|
|
self.iso_file = f'{self.iso_dir}/seed.iso'
|
|
self.seed_data_dir = '/temp/seed_data'
|
|
self.enroll_init = subcloud_enrollment.\
|
|
SubcloudEnrollmentInit(self.subcloud_name)
|
|
self.fake_install_values = \
|
|
copy.copy(fake_subcloud.FAKE_SUBCLOUD_INSTALL_VALUES)
|
|
|
|
self.iso_values = {
|
|
'software_version': self.rel_version,
|
|
'admin_password': 'St8rlingX*',
|
|
'bootstrap_interface': 'enp2s1',
|
|
'external_oam_floating_address': '10.10.10.2',
|
|
'network_mask': '255.255.255.0',
|
|
'external_oam_gateway_address': '10.10.10.1',
|
|
'external_oam_subnet': '10.10.10.0/24',
|
|
'install_values': self.fake_install_values
|
|
}
|
|
|
|
mock_run_patch_patch = mock.patch('eventlet.green.subprocess.run')
|
|
mock_TemporaryDirectory_patch = mock.patch('tempfile.TemporaryDirectory')
|
|
mock_makedirs_patch = mock.patch('os.makedirs')
|
|
mock_rmtree_patch = mock.patch('shutil.rmtree')
|
|
|
|
self.mock_run = mock_run_patch_patch.start()
|
|
self.mock_TemporaryDirectory = mock_TemporaryDirectory_patch.start()
|
|
self.mock_makedirs = mock_makedirs_patch.start()
|
|
self.mock_rmtree = mock_rmtree_patch.start()
|
|
|
|
self.addCleanup(mock_run_patch_patch.stop)
|
|
self.addCleanup(mock_TemporaryDirectory_patch.stop)
|
|
self.addCleanup(mock_makedirs_patch.stop)
|
|
self.addCleanup(mock_rmtree_patch.stop)
|
|
|
|
self._mock_builtins_open()
|
|
|
|
self.mock_builtins_open.side_effect = mock.mock_open()
|
|
self.mock_os_path_exists.return_value = True
|
|
self.mock_TemporaryDirectory.return_value.__enter__.return_value = \
|
|
self.seed_data_dir
|
|
self.mock_os_path_isdir.return_value = True
|
|
self.mock_run.return_value = mock.MagicMock(returncode=0,
|
|
stdout=b'Success')
|
|
|
|
def patched_isdir(self, path):
|
|
return path != self.iso_dir
|
|
|
|
def test_build_seed_meta_config(self):
|
|
result = self.enroll_init._build_seed_meta_data(self.seed_data_dir,
|
|
self.iso_values)
|
|
|
|
self.assertTrue(result)
|
|
self.mock_builtins_open.assert_called_once_with(
|
|
f'{self.seed_data_dir}/meta-data',
|
|
'w')
|
|
|
|
def test_build_seed_user_config(self):
|
|
result = self.enroll_init._build_seed_user_config(self.seed_data_dir,
|
|
self.iso_values)
|
|
|
|
self.assertTrue(result)
|
|
self.mock_builtins_open.assert_called_once_with(
|
|
f'{self.seed_data_dir}/user-data',
|
|
'w')
|
|
|
|
# Test with incomplete iso_values, expect KeyError
|
|
copied_dict = self.iso_values.copy()
|
|
copied_dict.pop('admin_password')
|
|
|
|
test_func = lambda: self.enroll_init._build_seed_user_config(
|
|
self.seed_data_dir,
|
|
copied_dict)
|
|
|
|
self.assertRaises(KeyError, test_func)
|
|
|
|
def test_generate_seed_iso(self):
|
|
with mock.patch('os.path.isdir', side_effect=self.patched_isdir):
|
|
self.assertTrue(self.enroll_init._generate_seed_iso(self.iso_values))
|
|
# Iso command must be invoked (subprocess.run)
|
|
self.mock_run.assert_called_once()
|
|
# Temp seed data dir must be created
|
|
self.mock_TemporaryDirectory.assert_called_once_with(prefix='seed_')
|
|
# Seed files must be generted in temp seed dir
|
|
self.mock_builtins_open.assert_any_call(
|
|
f'{self.seed_data_dir}/meta-data',
|
|
'w')
|
|
self.mock_builtins_open.assert_any_call(
|
|
f'{self.seed_data_dir}/user-data',
|
|
'w')
|