Add tests to ensure snapshots across replicas

Related-Bug: #1546303
Depends-On: Ia4cd2a36e31418e7a3d1c218080caa632755fe16
Depends-On: Id318a4adc0faf64a4bef57252aa2f0d9083b82b1
Change-Id: I269225b976efe13b2cbb9e0648d541a063df70e5
This commit is contained in:
Yogesh 2015-09-29 23:41:02 -04:00 committed by Goutham Pacha Ravi
parent 547807b464
commit 7bf4f0262e
6 changed files with 249 additions and 20 deletions

View File

@ -77,6 +77,10 @@ elif [[ "$DRIVER" == "lvm" ]]; then
elif [[ "$DRIVER" == "zfsonlinux" ]]; then elif [[ "$DRIVER" == "zfsonlinux" ]]; then
echo "SHARE_DRIVER=manila.share.drivers.zfsonlinux.driver.ZFSonLinuxShareDriver" >> $localrc_path echo "SHARE_DRIVER=manila.share.drivers.zfsonlinux.driver.ZFSonLinuxShareDriver" >> $localrc_path
echo "RUN_MANILA_REPLICATION_TESTS=True" >> $localrc_path echo "RUN_MANILA_REPLICATION_TESTS=True" >> $localrc_path
# Set the replica_state_update_interval to 60 seconds to make
# replication tests run faster. The default is 300, which is greater than
# the build timeout for ZFS on the gate.
echo "MANILA_REPLICA_STATE_UPDATE_INTERVAL=60" >> $localrc_path
echo "MANILA_ZFSONLINUX_USE_SSH=True" >> $localrc_path echo "MANILA_ZFSONLINUX_USE_SSH=True" >> $localrc_path
fi fi

View File

@ -187,6 +187,9 @@ function configure_manila {
iniset $MANILA_CONF DEFAULT lvm_share_volume_group $SHARE_GROUP iniset $MANILA_CONF DEFAULT lvm_share_volume_group $SHARE_GROUP
# Set the replica_state_update_interval
iniset $MANILA_CONF DEFAULT replica_state_update_interval $MANILA_REPLICA_STATE_UPDATE_INTERVAL
if is_service_enabled neutron; then if is_service_enabled neutron; then
configure_auth_token_middleware $MANILA_CONF neutron $MANILA_AUTH_CACHE_DIR neutron configure_auth_token_middleware $MANILA_CONF neutron $MANILA_AUTH_CACHE_DIR neutron
fi fi

View File

@ -141,6 +141,9 @@ SMB_CONF=${SMB_CONF:-/etc/samba/smb.conf}
SMB_PRIVATE_DIR=${SMB_PRIVATE_DIR:-/var/lib/samba/private} SMB_PRIVATE_DIR=${SMB_PRIVATE_DIR:-/var/lib/samba/private}
CONFIGURE_BACKING_FILE=${CONFIGURE_BACKING_FILE:-"True"} CONFIGURE_BACKING_FILE=${CONFIGURE_BACKING_FILE:-"True"}
# Options for replication
MANILA_REPLICA_STATE_UPDATE_INTERVAL=${MANILA_REPLICA_STATE_UPDATE_INTERVAL:-300}
# Options for configuration of ZFSonLinux driver # Options for configuration of ZFSonLinux driver
# 'MANILA_ZFSONLINUX_ZPOOL_SIZE' defines size of each zpool. That value # 'MANILA_ZFSONLINUX_ZPOOL_SIZE' defines size of each zpool. That value
# will be used for creation of sparse files. # will be used for creation of sparse files.

View File

@ -468,7 +468,7 @@ class BaseSharesTest(test.BaseTestCase):
description=None, force=False, description=None, force=False,
client=None, cleanup_in_class=True): client=None, cleanup_in_class=True):
if client is None: if client is None:
client = cls.shares_client client = cls.shares_v2_client
if description is None: if description is None:
description = "Tempest's snapshot" description = "Tempest's snapshot"
snapshot = client.create_snapshot(share_id, name, description, force) snapshot = client.create_snapshot(share_id, name, description, force)
@ -521,6 +521,16 @@ class BaseSharesTest(test.BaseTestCase):
service['state'] == 'up'] service['state'] == 'up']
return zones return zones
def get_pools_for_replication_domain(self):
# Get the list of pools for the replication domain
pools = self.admin_client.list_pools(detail=True)['pools']
instance_host = self.shares[0]['host']
host_pool = [p for p in pools if p['name'] == instance_host][0]
rep_domain = host_pool['capabilities']['replication_domain']
pools_in_rep_domain = [p for p in pools if p['capabilities'][
'replication_domain'] == rep_domain]
return rep_domain, pools_in_rep_domain
@classmethod @classmethod
def create_share_replica(cls, share_id, availability_zone, client=None, def create_share_replica(cls, share_id, availability_zone, client=None,
cleanup_in_class=False, cleanup=True): cleanup_in_class=False, cleanup=True):
@ -545,8 +555,11 @@ class BaseSharesTest(test.BaseTestCase):
@classmethod @classmethod
def delete_share_replica(cls, replica_id, client=None): def delete_share_replica(cls, replica_id, client=None):
client = client or cls.shares_v2_client client = client or cls.shares_v2_client
client.delete_share_replica(replica_id) try:
client.wait_for_resource_deletion(replica_id=replica_id) client.delete_share_replica(replica_id)
client.wait_for_resource_deletion(replica_id=replica_id)
except exceptions.NotFound:
pass
@classmethod @classmethod
def promote_share_replica(cls, replica_id, client=None): def promote_share_replica(cls, replica_id, client=None):
@ -634,6 +647,19 @@ class BaseSharesTest(test.BaseTestCase):
ic["method"]() ic["method"]()
ic["deleted"] = True ic["deleted"] = True
@classmethod
def clear_share_replicas(cls, share_id, client=None):
client = client or cls.shares_v2_client
share_replicas = client.list_share_replicas(
share_id=share_id)
for replica in share_replicas:
try:
cls.delete_share_replica(replica['id'])
except exceptions.BadRequest:
# Ignore the exception due to deletion of last active replica
pass
@classmethod @classmethod
def clear_resources(cls, resources=None): def clear_resources(cls, resources=None):
"""Deletes resources, that were created in test suites. """Deletes resources, that were created in test suites.
@ -658,6 +684,7 @@ class BaseSharesTest(test.BaseTestCase):
client = res["client"] client = res["client"]
with handle_cleanup_exceptions(): with handle_cleanup_exceptions():
if res["type"] is "share": if res["type"] is "share":
cls.clear_share_replicas(res_id)
cg_id = res.get('consistency_group_id') cg_id = res.get('consistency_group_id')
if cg_id: if cg_id:
params = {'consistency_group_id': cg_id} params = {'consistency_group_id': cg_id}

View File

@ -108,16 +108,6 @@ class ReplicationTest(base.BaseSharesTest):
return [replica for replica in replica_list return [replica for replica in replica_list
if replica['replica_state'] == r_state] if replica['replica_state'] == r_state]
def _get_pools_for_replication_domain(self):
# Get the list of pools for the replication domain
pools = self.admin_client.list_pools(detail=True)['pools']
instance_host = self.shares[0]['host']
host_pool = [p for p in pools if p['name'] == instance_host][0]
rep_domain = host_pool['capabilities']['replication_domain']
pools_in_rep_domain = [p for p in pools if p['capabilities'][
'replication_domain'] == rep_domain]
return rep_domain, pools_in_rep_domain
def _verify_config_and_set_access_rule_data(self): def _verify_config_and_set_access_rule_data(self):
"""Verify the access rule configuration is enabled for NFS. """Verify the access rule configuration is enabled for NFS.
@ -157,9 +147,10 @@ class ReplicationTest(base.BaseSharesTest):
# Create the replica # Create the replica
self._verify_create_replica() self._verify_create_replica()
# Verify access rule transitions to 'active' state. # Verify access_rules_status transitions to 'active' state.
self.shares_v2_client.wait_for_access_rule_status( self.shares_v2_client.wait_for_share_status(
self.shares[0]["id"], rule["id"], constants.RULE_STATE_ACTIVE) self.shares[0]["id"], constants.RULE_STATE_ACTIVE,
status_attr='access_rules_status')
# Delete rule and wait for deletion # Delete rule and wait for deletion
self.shares_v2_client.delete_access_rule(self.shares[0]["id"], self.shares_v2_client.delete_access_rule(self.shares[0]["id"],
@ -174,17 +165,19 @@ class ReplicationTest(base.BaseSharesTest):
share_replica = self._verify_create_replica() share_replica = self._verify_create_replica()
# Add access rule # Add access rule
rule = self.shares_v2_client.create_access_rule( self.shares_v2_client.create_access_rule(
self.shares[0]["id"], access_type, access_to, 'ro') self.shares[0]["id"], access_type, access_to, 'ro')
self.shares_v2_client.wait_for_access_rule_status(
self.shares[0]["id"], rule["id"], constants.RULE_STATE_ACTIVE) self.shares_v2_client.wait_for_share_status(
self.shares[0]["id"], constants.RULE_STATE_ACTIVE,
status_attr='access_rules_status')
# Delete the replica # Delete the replica
self.delete_share_replica(share_replica["id"]) self.delete_share_replica(share_replica["id"])
@test.attr(type=["gate", ]) @test.attr(type=["gate", ])
def test_add_multiple_share_replicas(self): def test_add_multiple_share_replicas(self):
rep_domain, pools = self._get_pools_for_replication_domain() rep_domain, pools = self.get_pools_for_replication_domain()
if len(pools) < 3: if len(pools) < 3:
msg = ("Replication domain %(domain)s has only %(count)s pools. " msg = ("Replication domain %(domain)s has only %(count)s pools. "
"Need at least 3 pools to run this test." % "Need at least 3 pools to run this test." %

View File

@ -0,0 +1,199 @@
# Copyright 2016 Yogesh Kshirsagar
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from tempest import config
from tempest.lib.common.utils import data_utils
from tempest import test
import testtools
from manila_tempest_tests import clients_share as clients
from manila_tempest_tests.common import constants
from manila_tempest_tests import share_exceptions
from manila_tempest_tests.tests.api import base
CONF = config.CONF
_MIN_SUPPORTED_MICROVERSION = '2.11'
@testtools.skipUnless(CONF.share.run_replication_tests,
'Replication tests are disabled.')
@testtools.skipUnless(CONF.share.run_snapshot_tests,
'Snapshot tests disabled.')
@base.skip_if_microversion_lt(_MIN_SUPPORTED_MICROVERSION)
class ReplicationSnapshotTest(base.BaseSharesTest):
@classmethod
def resource_setup(cls):
super(ReplicationSnapshotTest, cls).resource_setup()
# Create share_type
name = data_utils.rand_name(constants.TEMPEST_MANILA_PREFIX)
cls.admin_client = clients.AdminManager().shares_v2_client
cls.replication_type = CONF.share.backend_replication_type
if cls.replication_type not in constants.REPLICATION_TYPE_CHOICES:
raise share_exceptions.ShareReplicationTypeException(
replication_type=cls.replication_type
)
cls.zones = cls.get_availability_zones(client=cls.admin_client)
cls.share_zone = cls.zones[0]
cls.replica_zone = cls.zones[-1]
cls.extra_specs = cls.add_required_extra_specs_to_dict(
{"replication_type": cls.replication_type})
share_type = cls.create_share_type(
name,
extra_specs=cls.extra_specs,
client=cls.admin_client)
cls.share_type = share_type["share_type"]
# Create share with above share_type
cls.creation_data = {'kwargs': {
'share_type_id': cls.share_type['id'],
'availability_zone': cls.share_zone,
}}
@test.attr(type=["gate", ])
def test_snapshot_after_share_replica(self):
"""Test the snapshot for replicated share.
Create replica first and then create a snapshot.
Verify that the snapshot is properly created under replica by
creating a share from that snapshot.
"""
share = self.create_share(share_type_id=self.share_type['id'],
availability_zone=self.share_zone)
original_replica = self.shares_v2_client.list_share_replicas(
share["id"])[0]
share_replica = self.create_share_replica(share["id"],
self.replica_zone,
cleanup=False)
self.addCleanup(self.delete_share_replica, original_replica['id'])
self.shares_v2_client.wait_for_share_replica_status(
share_replica['id'], constants.REPLICATION_STATE_IN_SYNC,
status_attr='replica_state')
snapshot = self.create_snapshot_wait_for_active(share["id"])
self.promote_share_replica(share_replica['id'])
self.delete_share_replica(original_replica['id'])
self.create_share(snapshot_id=snapshot['id'])
@test.attr(type=["gate", ])
def test_snapshot_before_share_replica(self):
"""Test the snapshot for replicated share.
Create snapshot before creating share replica for the same
share.
Verify snapshot by creating share from the snapshot.
"""
share = self.create_share(share_type_id=self.share_type['id'],
availability_zone=self.share_zone)
snapshot = self.create_snapshot_wait_for_active(share["id"])
original_replica = self.shares_v2_client.list_share_replicas(
share["id"])[0]
share_replica = self.create_share_replica(share["id"],
self.replica_zone,
cleanup=False)
self.addCleanup(self.delete_share_replica, original_replica['id'])
self.shares_v2_client.wait_for_share_replica_status(
share_replica['id'], constants.REPLICATION_STATE_IN_SYNC,
status_attr='replica_state')
# Wait for snapshot1 to become available
self.shares_v2_client.wait_for_snapshot_status(
snapshot['id'], "available")
self.promote_share_replica(share_replica['id'])
self.delete_share_replica(original_replica['id'])
self.create_share(snapshot_id=snapshot['id'])
@test.attr(type=["gate", ])
def test_snapshot_before_and_after_share_replica(self):
"""Test the snapshot for replicated share.
Verify that snapshot can be created before and after share replica
being created.
Verify snapshots by creating share from the snapshots.
"""
share = self.create_share(share_type_id=self.share_type['id'],
availability_zone=self.share_zone)
snapshot1 = self.create_snapshot_wait_for_active(share["id"])
original_replica = self.shares_v2_client.list_share_replicas(
share["id"])[0]
share_replica = self.create_share_replica(share["id"],
self.replica_zone,
cleanup=False)
self.addCleanup(self.delete_share_replica, original_replica['id'])
self.shares_v2_client.wait_for_share_replica_status(
share_replica['id'], constants.REPLICATION_STATE_IN_SYNC,
status_attr='replica_state')
snapshot2 = self.create_snapshot_wait_for_active(share["id"])
# Wait for snapshot1 to become available
self.shares_v2_client.wait_for_snapshot_status(
snapshot1['id'], "available")
self.promote_share_replica(share_replica['id'])
# Remove the original active replica to ensure that snapshot is
# still being created successfully.
self.delete_share_replica(original_replica['id'])
self.create_share(snapshot_id=snapshot1['id'])
self.create_share(snapshot_id=snapshot2['id'])
@test.attr(type=["gate", ])
def test_delete_snapshot_after_adding_replica(self):
"""Verify the snapshot delete.
Ensure that deleting the original snapshot also deletes the
snapshot from replica.
"""
share = self.create_share(share_type_id=self.share_type['id'],
availability_zone=self.share_zone)
share_replica = self.create_share_replica(share["id"],
self.replica_zone)
self.shares_v2_client.wait_for_share_replica_status(
share_replica['id'], constants.REPLICATION_STATE_IN_SYNC,
status_attr='replica_state')
snapshot = self.create_snapshot_wait_for_active(share["id"])
self.shares_v2_client.delete_snapshot(snapshot['id'])
self.shares_v2_client.wait_for_resource_deletion(
snapshot_id=snapshot["id"])
@test.attr(type=["gate", ])
def test_create_replica_from_snapshot_share(self):
"""Test replica for a share that was created from snapshot."""
share = self.create_share(share_type_id=self.share_type['id'],
availability_zone=self.share_zone)
orig_snapshot = self.create_snapshot_wait_for_active(share["id"])
snap_share = self.create_share(snapshot_id=orig_snapshot['id'])
original_replica = self.shares_v2_client.list_share_replicas(
snap_share["id"])[0]
share_replica = self.create_share_replica(snap_share["id"],
self.replica_zone,
cleanup=False)
self.addCleanup(self.delete_share_replica, original_replica['id'])
self.shares_v2_client.wait_for_share_replica_status(
share_replica['id'], constants.REPLICATION_STATE_IN_SYNC,
status_attr='replica_state')
self.promote_share_replica(share_replica['id'])
# Delete the demoted replica so promoted replica can be cleaned
# during the cleanup
self.delete_share_replica(original_replica['id'])