Support for consistency groups in ScaleIO driver
Add support for all consistency groups functionalities in ScaleIO driver. Also fixed a small mistake in a test regarding deleting a snapshot. DocImpact Implements: blueprint scaleio-consistency-groups Change-Id: Id8b52aeb546f9f5fa68b98a4e59bd3f12e78bbef
This commit is contained in:
parent
eac894c09f
commit
494a55973d
@ -49,10 +49,6 @@ class ScaleIODriver(scaleio.ScaleIODriver):
|
||||
*args,
|
||||
**kwargs)
|
||||
|
||||
def update_consistencygroup(self, context, group, add_volumes=None,
|
||||
remove_volumes=None):
|
||||
pass
|
||||
|
||||
def local_path(self, volume):
|
||||
pass
|
||||
|
||||
@ -62,28 +58,12 @@ class ScaleIODriver(scaleio.ScaleIODriver):
|
||||
def promote_replica(self, context, volume):
|
||||
pass
|
||||
|
||||
def delete_consistencygroup(self, context, group, volumes):
|
||||
pass
|
||||
|
||||
def create_consistencygroup_from_src(self, context, group, volumes,
|
||||
cgsnapshot=None, snapshots=None):
|
||||
pass
|
||||
|
||||
def create_replica_test_volume(self, volume, src_vref):
|
||||
pass
|
||||
|
||||
def create_consistencygroup(self, context, group):
|
||||
pass
|
||||
|
||||
def unmanage(self, volume):
|
||||
pass
|
||||
|
||||
def create_cgsnapshot(self, context, cgsnapshot, snapshots):
|
||||
pass
|
||||
|
||||
def delete_cgsnapshot(self, context, cgsnapshot, snapshots):
|
||||
pass
|
||||
|
||||
|
||||
class MockHTTPSResponse(requests.Response):
|
||||
"""Mock HTTP Response
|
||||
|
@ -0,0 +1,209 @@
|
||||
# Copyright (c) 2013 - 2016 EMC Corporation.
|
||||
# All Rights Reserved.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
import json
|
||||
|
||||
import mock
|
||||
|
||||
from cinder import context
|
||||
from cinder.tests.unit import fake_consistencygroup
|
||||
from cinder.tests.unit import fake_snapshot
|
||||
from cinder.tests.unit import fake_volume
|
||||
from cinder.tests.unit.volume.drivers.emc import scaleio
|
||||
from cinder.tests.unit.volume.drivers.emc.scaleio import mocks
|
||||
|
||||
|
||||
class TestConsistencyGroups(scaleio.TestScaleIODriver):
|
||||
"""Test cases for ``ScaleIODriver consistency groups support``"""
|
||||
|
||||
def setUp(self):
|
||||
"""Setup a test case environment.
|
||||
|
||||
Creates a fake volume object and sets up the required API responses.
|
||||
"""
|
||||
super(TestConsistencyGroups, self).setUp()
|
||||
self.ctx = context.RequestContext('fake', 'fake', auth_token=True)
|
||||
self.consistency_group = (
|
||||
fake_consistencygroup.fake_consistencyobject_obj(self.ctx,
|
||||
**{'id': 'cgid'}))
|
||||
fake_volume1 = fake_volume.fake_volume_obj(
|
||||
self.ctx,
|
||||
**{'id': 'volid1', 'provider_id': 'pid_1'})
|
||||
fake_volume2 = fake_volume.fake_volume_obj(
|
||||
self.ctx,
|
||||
**{'id': 'volid2', 'provider_id': 'pid_2'})
|
||||
fake_volume3 = fake_volume.fake_volume_obj(
|
||||
self.ctx,
|
||||
**{'id': 'volid3', 'provider_id': 'pid_3'})
|
||||
fake_volume4 = fake_volume.fake_volume_obj(
|
||||
self.ctx,
|
||||
**{'id': 'volid4', 'provider_id': 'pid_4'})
|
||||
self.volumes = [fake_volume1, fake_volume2]
|
||||
self.volumes2 = [fake_volume3, fake_volume4]
|
||||
fake_snapshot1 = fake_snapshot.fake_snapshot_obj(
|
||||
self.ctx,
|
||||
**{'id': 'snapid1', 'volume_id': 'volid1',
|
||||
'volume': fake_volume1})
|
||||
fake_snapshot2 = fake_snapshot.fake_snapshot_obj(
|
||||
self.ctx,
|
||||
**{'id': 'snapid2', 'volume_id': 'volid2', 'volume':
|
||||
fake_volume2})
|
||||
self.snapshots = [fake_snapshot1, fake_snapshot2]
|
||||
self.snapshot_reply = json.dumps({
|
||||
'volumeIdList': ['sid1', 'sid2'],
|
||||
'snapshotGroupId': 'sgid1'})
|
||||
self.HTTPS_MOCK_RESPONSES = {
|
||||
self.RESPONSE_MODE.Valid: {
|
||||
'instances/Volume::{}/action/removeVolume'.format(
|
||||
fake_volume1['provider_id']
|
||||
): fake_volume1['provider_id'],
|
||||
'instances/Volume::{}/action/removeVolume'.format(
|
||||
fake_volume2['provider_id']
|
||||
): fake_volume2['provider_id'],
|
||||
'instances/Volume::{}/action/removeMappedSdc'.format(
|
||||
fake_volume1['provider_id']
|
||||
): fake_volume1['provider_id'],
|
||||
'instances/Volume::{}/action/removeMappedSdc'.format(
|
||||
fake_volume2['provider_id']
|
||||
): fake_volume2['provider_id'],
|
||||
'instances/System/action/snapshotVolumes':
|
||||
self.snapshot_reply,
|
||||
},
|
||||
self.RESPONSE_MODE.BadStatus: {
|
||||
'instances/Volume::{}/action/removeVolume'.format(
|
||||
fake_volume1['provider_id']
|
||||
): mocks.MockHTTPSResponse(
|
||||
{
|
||||
'errorCode': 401,
|
||||
'message': 'BadStatus Volume Test',
|
||||
}, 401
|
||||
),
|
||||
'instances/Volume::{}/action/removeVolume'.format(
|
||||
fake_volume2['provider_id']
|
||||
): mocks.MockHTTPSResponse(
|
||||
{
|
||||
'errorCode': 401,
|
||||
'message': 'BadStatus Volume Test',
|
||||
}, 401
|
||||
),
|
||||
'instances/System/action/snapshotVolumes':
|
||||
self.BAD_STATUS_RESPONSE
|
||||
},
|
||||
}
|
||||
|
||||
def _fake_cgsnapshot(self):
|
||||
cgsnap = {'id': 'cgsid', 'name': 'testsnap',
|
||||
'consistencygroup_id': 'cgid', 'status': 'available'}
|
||||
return cgsnap
|
||||
|
||||
def test_create_consistencygroup(self):
|
||||
result = self.driver.create_consistencygroup(self.ctx,
|
||||
self.consistency_group)
|
||||
self.assertEqual('available', result['status'])
|
||||
|
||||
def test_delete_consistencygroup_valid(self):
|
||||
self.set_https_response_mode(self.RESPONSE_MODE.Valid)
|
||||
self.driver.configuration.set_override(
|
||||
'sio_unmap_volume_before_deletion',
|
||||
override=True)
|
||||
result_model_update, result_volumes_update = (
|
||||
self.driver.delete_consistencygroup(self.ctx,
|
||||
self.consistency_group,
|
||||
self.volumes))
|
||||
self.assertTrue(all(volume['status'] == 'deleted' for volume in
|
||||
result_volumes_update))
|
||||
self.assertEqual('deleted', result_model_update['status'])
|
||||
|
||||
def test_delete_consistency_group_fail(self):
|
||||
self.set_https_response_mode(self.RESPONSE_MODE.BadStatus)
|
||||
result_model_update, result_volumes_update = (
|
||||
self.driver.delete_consistencygroup(self.ctx,
|
||||
self.consistency_group,
|
||||
self.volumes))
|
||||
self.assertTrue(any(volume['status'] == 'error_deleting' for volume in
|
||||
result_volumes_update))
|
||||
self.assertTrue(result_model_update['status'] in ['error_deleting',
|
||||
'error'])
|
||||
|
||||
def test_create_consistencygroup_from_cg(self):
|
||||
self.set_https_response_mode(self.RESPONSE_MODE.Valid)
|
||||
result_model_update, result_volumes_model_update = (
|
||||
self.driver.create_consistencygroup_from_src(
|
||||
self.ctx, self.consistency_group, self.volumes2,
|
||||
source_cg=self.consistency_group, source_vols=self.volumes))
|
||||
self.assertEqual('available', result_model_update['status'])
|
||||
get_pid = lambda snapshot: snapshot['provider_id']
|
||||
volume_provider_list = list(map(get_pid, result_volumes_model_update))
|
||||
self.assertListEqual(volume_provider_list, ['sid1', 'sid2'])
|
||||
|
||||
def test_create_consistencygroup_from_cgs(self):
|
||||
self.snapshots[0]['provider_id'] = 'pid_1'
|
||||
self.snapshots[1]['provider_id'] = 'pid_2'
|
||||
self.set_https_response_mode(self.RESPONSE_MODE.Valid)
|
||||
result_model_update, result_volumes_model_update = (
|
||||
self.driver.create_consistencygroup_from_src(
|
||||
self.ctx, self.consistency_group, self.volumes2,
|
||||
cgsnapshot=self._fake_cgsnapshot(),
|
||||
snapshots=self.snapshots))
|
||||
self.assertEqual('available', result_model_update['status'])
|
||||
get_pid = lambda snapshot: snapshot['provider_id']
|
||||
volume_provider_list = list(map(get_pid, result_volumes_model_update))
|
||||
self.assertListEqual(['sid1', 'sid2'], volume_provider_list)
|
||||
|
||||
@mock.patch('cinder.objects.snapshot')
|
||||
@mock.patch('cinder.objects.snapshot')
|
||||
def test_create_cgsnapshots(self, snapshot1, snapshot2):
|
||||
type(snapshot1).volume = mock.PropertyMock(
|
||||
return_value=self.volumes[0])
|
||||
type(snapshot2).volume = mock.PropertyMock(
|
||||
return_value=self.volumes[1])
|
||||
snapshots = [snapshot1, snapshot2]
|
||||
self.set_https_response_mode(self.RESPONSE_MODE.Valid)
|
||||
result_model_update, result_snapshot_model_update = (
|
||||
self.driver.create_cgsnapshot(
|
||||
self.ctx,
|
||||
self._fake_cgsnapshot(),
|
||||
snapshots
|
||||
))
|
||||
self.assertEqual('available', result_model_update['status'])
|
||||
self.assertTrue(all(snapshot['status'] == 'available' for snapshot in
|
||||
result_snapshot_model_update))
|
||||
get_pid = lambda snapshot: snapshot['provider_id']
|
||||
snapshot_provider_list = list(map(get_pid,
|
||||
result_snapshot_model_update))
|
||||
self.assertListEqual(['sid1', 'sid2'], snapshot_provider_list)
|
||||
|
||||
@mock.patch('cinder.objects.snapshot')
|
||||
@mock.patch('cinder.objects.snapshot')
|
||||
def test_delete_cgsnapshots(self, snapshot1, snapshot2):
|
||||
type(snapshot1).volume = mock.PropertyMock(
|
||||
return_value=self.volumes[0])
|
||||
type(snapshot2).volume = mock.PropertyMock(
|
||||
return_value=self.volumes[1])
|
||||
type(snapshot1).provider_id = mock.PropertyMock(
|
||||
return_value='pid_1')
|
||||
type(snapshot2).provider_id = mock.PropertyMock(
|
||||
return_value='pid_2')
|
||||
snapshots = [snapshot1, snapshot2]
|
||||
self.set_https_response_mode(self.RESPONSE_MODE.Valid)
|
||||
result_model_update, result_snapshot_model_update = (
|
||||
self.driver.delete_cgsnapshot(
|
||||
self.ctx,
|
||||
self._fake_cgsnapshot(),
|
||||
snapshots
|
||||
))
|
||||
self.assertEqual('deleted', result_model_update['status'])
|
||||
self.assertTrue(all(snapshot['status'] == 'deleted' for snapshot in
|
||||
result_snapshot_model_update))
|
@ -88,9 +88,8 @@ class TestDeleteSnapShot(scaleio.TestScaleIODriver):
|
||||
self.driver.delete_snapshot(self.snapshot)
|
||||
|
||||
def test_delete_invalid_snapshot(self):
|
||||
self.set_https_response_mode(self.RESPONSE_MODE.Invalid)
|
||||
self.assertRaises(exception.VolumeBackendAPIException,
|
||||
self.driver.delete_snapshot, self.snapshot)
|
||||
self.set_https_response_mode(self.RESPONSE_MODE.Valid)
|
||||
self.driver.delete_snapshot(self.snapshot)
|
||||
|
||||
def test_delete_snapshot(self):
|
||||
"""Setting the unmap volume before delete flag for tests """
|
||||
|
@ -30,7 +30,7 @@ from six.moves import urllib
|
||||
|
||||
from cinder import context
|
||||
from cinder import exception
|
||||
from cinder.i18n import _, _LI, _LW
|
||||
from cinder.i18n import _, _LI, _LW, _LE
|
||||
from cinder.image import image_utils
|
||||
from cinder import utils
|
||||
from cinder.volume import driver
|
||||
@ -239,11 +239,11 @@ class ScaleIODriver(driver.VolumeDriver):
|
||||
extraspecs_limit = storage_type.get(extraspecs_key)
|
||||
if extraspecs_limit is not None:
|
||||
if qos_limit is not None:
|
||||
LOG.warning(_LW("QoS specs are overriding extraspecs"))
|
||||
LOG.warning(_LW("QoS specs are overriding extra_specs."))
|
||||
else:
|
||||
LOG.info(_LI("Using extraspecs for defining QoS specs "
|
||||
"will be deprecated in the next "
|
||||
"version of OpenStack, please use QoS specs"))
|
||||
LOG.info(_LI("Using extra_specs for defining QoS specs "
|
||||
"will be deprecated in the N release "
|
||||
"of OpenStack. Please use QoS specs."))
|
||||
return qos_limit if qos_limit is not None else extraspecs_limit
|
||||
|
||||
def _id_to_base64(self, id):
|
||||
@ -447,16 +447,7 @@ class ScaleIODriver(driver.VolumeDriver):
|
||||
'server_port': self.server_port}
|
||||
request = ("https://%(server_ip)s:%(server_port)s"
|
||||
"/api/instances/System/action/snapshotVolumes") % req_vars
|
||||
r = requests.post(
|
||||
request,
|
||||
data=json.dumps(params),
|
||||
headers=self._get_headers(),
|
||||
auth=(
|
||||
self.server_username,
|
||||
self.server_token),
|
||||
verify=self._get_verify_cert())
|
||||
r = self._check_response(r, request, False, params)
|
||||
response = r.json()
|
||||
r, response = self._execute_scaleio_post_request(params, request)
|
||||
LOG.info(_LI("Snapshot volume response: %s."), response)
|
||||
if r.status_code != OK_STATUS_CODE and "errorCode" in response:
|
||||
msg = (_("Failed creating snapshot for volume %(volname)s: "
|
||||
@ -468,6 +459,19 @@ class ScaleIODriver(driver.VolumeDriver):
|
||||
|
||||
return {'provider_id': response['volumeIdList'][0]}
|
||||
|
||||
def _execute_scaleio_post_request(self, params, request):
|
||||
r = requests.post(
|
||||
request,
|
||||
data=json.dumps(params),
|
||||
headers=self._get_headers(),
|
||||
auth=(
|
||||
self.server_username,
|
||||
self.server_token),
|
||||
verify=self._get_verify_cert())
|
||||
r = self._check_response(r, request, False, params)
|
||||
response = r.json()
|
||||
return r, response
|
||||
|
||||
def _check_response(self, response, request, is_get_request=True,
|
||||
params=None):
|
||||
if response.status_code == 401 or response.status_code == 403:
|
||||
@ -713,6 +717,7 @@ class ScaleIODriver(driver.VolumeDriver):
|
||||
stats['free_capacity_gb'] = 'unknown'
|
||||
stats['reserved_percentage'] = 0
|
||||
stats['QoS_support'] = True
|
||||
stats['consistencygroup_support'] = True
|
||||
|
||||
pools = []
|
||||
|
||||
@ -829,6 +834,7 @@ class ScaleIODriver(driver.VolumeDriver):
|
||||
'total_capacity_gb': total_capacity_gb,
|
||||
'free_capacity_gb': free_capacity_gb,
|
||||
'QoS_support': True,
|
||||
'consistencygroup_support': True,
|
||||
'reserved_percentage': 0
|
||||
}
|
||||
|
||||
@ -1024,7 +1030,7 @@ class ScaleIODriver(driver.VolumeDriver):
|
||||
LOG.info(_LI("Get Volume response: %s"), response)
|
||||
self._manage_existing_check_legal_response(r, existing_ref)
|
||||
if response['mappedSdcInfo'] is not None:
|
||||
reason = ("manage_existing cannot manage a volume "
|
||||
reason = _("manage_existing cannot manage a volume "
|
||||
"connected to hosts. Please disconnect this volume "
|
||||
"from existing hosts before importing")
|
||||
raise exception.ManageExistingInvalidReference(
|
||||
@ -1087,6 +1093,138 @@ class ScaleIODriver(driver.VolumeDriver):
|
||||
reason=reason
|
||||
)
|
||||
|
||||
def create_consistencygroup(self, context, group):
|
||||
"""Creates a consistency group.
|
||||
|
||||
ScaleIO won't create CG until cg-snapshot creation,
|
||||
db will maintain the volumes and CG relationship.
|
||||
"""
|
||||
LOG.info(_LI("Creating Consistency Group"))
|
||||
model_update = {'status': 'available'}
|
||||
return model_update
|
||||
|
||||
def delete_consistencygroup(self, context, group, volumes):
|
||||
"""Deletes a consistency group.
|
||||
|
||||
ScaleIO will delete the volumes of the CG.
|
||||
"""
|
||||
LOG.info(_LI("Deleting Consistency Group"))
|
||||
model_update = {'status': 'deleted'}
|
||||
error_statuses = ['error', 'error_deleting']
|
||||
volumes_model_update = []
|
||||
for volume in volumes:
|
||||
try:
|
||||
self._delete_volume(volume['provider_id'])
|
||||
update_item = {'id': volume['id'],
|
||||
'status': 'deleted'}
|
||||
volumes_model_update.append(update_item)
|
||||
except exception.VolumeBackendAPIException as err:
|
||||
update_item = {'id': volume['id'],
|
||||
'status': 'error_deleting'}
|
||||
volumes_model_update.append(update_item)
|
||||
if model_update['status'] not in error_statuses:
|
||||
model_update['status'] = 'error_deleting'
|
||||
LOG.error(_LE("Failed to delete the volume %(vol)s of CG. "
|
||||
"Exception: %(exception)s."),
|
||||
{'vol': volume['name'], 'exception': err})
|
||||
return model_update, volumes_model_update
|
||||
|
||||
def create_cgsnapshot(self, context, cgsnapshot, snapshots):
|
||||
"""Creates a cgsnapshot."""
|
||||
get_scaleio_snapshot_params = lambda snapshot: {
|
||||
'volumeId': snapshot.volume['provider_id'],
|
||||
'snapshotName': self._id_to_base64(snapshot['id'])}
|
||||
snapshotDefs = list(map(get_scaleio_snapshot_params, snapshots))
|
||||
r, response = self._snapshot_volume_group(snapshotDefs)
|
||||
LOG.info(_LI("Snapshot volume response: %s."), response)
|
||||
if r.status_code != OK_STATUS_CODE and "errorCode" in response:
|
||||
msg = (_("Failed creating snapshot for group: "
|
||||
"%(response)s.") %
|
||||
{'response': response['message']})
|
||||
LOG.error(msg)
|
||||
raise exception.VolumeBackendAPIException(data=msg)
|
||||
snapshot_model_update = []
|
||||
for snapshot, scaleio_id in zip(snapshots, response['volumeIdList']):
|
||||
update_item = {'id': snapshot['id'],
|
||||
'status': 'available',
|
||||
'provider_id': scaleio_id}
|
||||
snapshot_model_update.append(update_item)
|
||||
model_update = {'status': 'available'}
|
||||
return model_update, snapshot_model_update
|
||||
|
||||
def delete_cgsnapshot(self, context, cgsnapshot, snapshots):
|
||||
"""Deletes a cgsnapshot."""
|
||||
error_statuses = ['error', 'error_deleting']
|
||||
model_update = {'status': cgsnapshot['status']}
|
||||
snapshot_model_update = []
|
||||
for snapshot in snapshots:
|
||||
try:
|
||||
self._delete_volume(snapshot.provider_id)
|
||||
update_item = {'id': snapshot['id'],
|
||||
'status': 'deleted'}
|
||||
snapshot_model_update.append(update_item)
|
||||
except exception.VolumeBackendAPIException as err:
|
||||
update_item = {'id': snapshot['id'],
|
||||
'status': 'error_deleting'}
|
||||
snapshot_model_update.append(update_item)
|
||||
if model_update['status'] not in error_statuses:
|
||||
model_update['status'] = 'error_deleting'
|
||||
LOG.error(_LE("Failed to delete the snapshot %(snap)s "
|
||||
"of cgsnapshot: %(cgsnapshot_id)s. "
|
||||
"Exception: %(exception)s."),
|
||||
{'snap': snapshot['name'],
|
||||
'exception': err,
|
||||
'cgsnapshot_id': cgsnapshot.id})
|
||||
model_update['status'] = 'deleted'
|
||||
return model_update, snapshot_model_update
|
||||
|
||||
def create_consistencygroup_from_src(self, context, group, volumes,
|
||||
cgsnapshot=None, snapshots=None,
|
||||
source_cg=None, source_vols=None):
|
||||
"""Creates a consistency group from a source."""
|
||||
get_scaleio_snapshot_params = lambda src_volume, trg_volume: {
|
||||
'volumeId': src_volume['provider_id'],
|
||||
'snapshotName': self._id_to_base64(trg_volume['id'])}
|
||||
if cgsnapshot and snapshots:
|
||||
snapshotDefs = map(get_scaleio_snapshot_params, snapshots, volumes)
|
||||
else:
|
||||
snapshotDefs = map(get_scaleio_snapshot_params, source_vols,
|
||||
volumes)
|
||||
r, response = self._snapshot_volume_group(list(snapshotDefs))
|
||||
LOG.info(_LI("Snapshot volume response: %s."), response)
|
||||
if r.status_code != OK_STATUS_CODE and "errorCode" in response:
|
||||
msg = (_("Failed creating snapshot for group: "
|
||||
"%(response)s.") %
|
||||
{'response': response['message']})
|
||||
LOG.error(msg)
|
||||
raise exception.VolumeBackendAPIException(data=msg)
|
||||
volumes_model_update = []
|
||||
for volume, scaleio_id in zip(volumes, response['volumeIdList']):
|
||||
update_item = {'id': volume['id'],
|
||||
'status': 'available',
|
||||
'provider_id': scaleio_id}
|
||||
volumes_model_update.append(update_item)
|
||||
model_update = {'status': 'available'}
|
||||
return model_update, volumes_model_update
|
||||
|
||||
def update_consistencygroup(self, context, group,
|
||||
add_volumes=None, remove_volumes=None):
|
||||
"""Update a consistency group.
|
||||
|
||||
ScaleIO does not handle volume grouping.
|
||||
Cinder maintains volumes and CG relationship.
|
||||
"""
|
||||
return None, None, None
|
||||
|
||||
def _snapshot_volume_group(self, snapshotDefs):
|
||||
LOG.info(_LI("ScaleIO snapshot group of volumes"))
|
||||
params = {'snapshotDefs': snapshotDefs}
|
||||
req_vars = {'server_ip': self.server_ip,
|
||||
'server_port': self.server_port}
|
||||
request = ("https://%(server_ip)s:%(server_port)s"
|
||||
"/api/instances/System/action/snapshotVolumes") % req_vars
|
||||
return self._execute_scaleio_post_request(params, request)
|
||||
|
||||
def ensure_export(self, context, volume):
|
||||
"""Driver entry point to get the export info for an existing volume."""
|
||||
pass
|
||||
|
@ -0,0 +1,3 @@
|
||||
---
|
||||
features:
|
||||
- Add Consistency Group support in ScaleIO driver.
|
Loading…
Reference in New Issue
Block a user