From bf4ca3bcf099760ef145571b975badc80d42b2d6 Mon Sep 17 00:00:00 2001 From: Gaurang Tapase Date: Thu, 2 Jul 2015 04:41:58 -0400 Subject: [PATCH] Adding NFS support to the GPFS Driver This change helps create cinder volumes on a GPFS cluster, when the cinder services are not running directly on one of the GPFS nodes. The NFS export created on GPFS cluster will be locally mounted on a node running the cinder service. The cinder volume service will communicate with GPFS nodes over SSH to execute CLIs remotely. This change also ports the GPFSDriver to ABC driver model. Implements : blueprint ibm-gpfs-nfs-support DocImpact Needs an update in gpfs driver documentation, as this patch adds an NFS based access for GPFS cinder driver. Change-Id: I40e5098482d48552170c34e74be0c9e2f544fef5 --- cinder/tests/unit/test_gpfs.py | 210 +++++++++++++++++ cinder/volume/drivers/ibm/gpfs.py | 365 +++++++++++++++++++++++------- 2 files changed, 498 insertions(+), 77 deletions(-) diff --git a/cinder/tests/unit/test_gpfs.py b/cinder/tests/unit/test_gpfs.py index bb13ec0e0..6217eb870 100644 --- a/cinder/tests/unit/test_gpfs.py +++ b/cinder/tests/unit/test_gpfs.py @@ -67,6 +67,7 @@ class GPFSDriverTestCase(test.TestCase): self.image_id = '70a599e0-31e7-49b7-b260-868f441e862b' self.driver = gpfs.GPFSDriver(configuration=conf.Configuration(None)) + self.driver.gpfs_execute = self._execute_wrapper self.driver.set_execute(self._execute_wrapper) self.driver._cluster_id = '123456' self.driver._gpfs_device = '/dev/gpfs' @@ -95,6 +96,13 @@ class GPFSDriverTestCase(test.TestCase): def test_sizestr(self): self.assertEqual(gpfs._sizestr('10'), '10G') + @mock.patch('cinder.utils.execute') + def test_gpfs_local_execute(self, mock_exec): + mock_exec.return_value = 'test' + self.driver._gpfs_local_execute('test') + expected = [mock.call('test', run_as_root=True)] + self.assertEqual(expected, mock_exec.mock_calls) + @mock.patch('cinder.utils.execute') def test_get_gpfs_state_ok(self, mock_exec): mock_exec.return_value = ('mmgetstate::HEADER:version:reserved:' @@ -1815,3 +1823,205 @@ class GPFSDriverTestCase(test.TestCase): volume['host'] = host return (volume, new_type, diff, host) + + +class GPFSNFSDriverTestCase(test.TestCase): + driver_name = "cinder.volume.drivers.gpfs.GPFSNFSDriver" + TEST_NFS_EXPORT = 'nfs-host1:/export' + TEST_SIZE_IN_GB = 1 + TEST_EXTEND_SIZE_IN_GB = 2 + TEST_MNT_POINT = '/mnt/nfs' + TEST_MNT_POINT_BASE = '/mnt' + TEST_GPFS_MNT_POINT_BASE = '/export' + TEST_LOCAL_PATH = '/mnt/nfs/volume-123' + TEST_VOLUME_PATH = '/export/volume-123' + TEST_SNAP_PATH = '/export/snapshot-123' + + def _execute_wrapper(self, cmd, *args, **kwargs): + try: + kwargs.pop('run_as_root') + except KeyError: + pass + + return utils.execute(cmd, *args, **kwargs) + + def _fake_volume(self): + volume = {} + volume['id'] = '123456' + volume['name'] = 'test' + volume['size'] = 1000 + volume['consistencygroup_id'] = 'cg-1234' + return volume + + def _fake_snapshot(self): + snapshot = {} + snapshot['id'] = '12345' + snapshot['name'] = 'test-snap' + snapshot['size'] = 1000 + snapshot['volume_id'] = '123456' + snapshot['status'] = 'available' + return snapshot + + def setUp(self): + super(GPFSNFSDriverTestCase, self).setUp() + self.driver = gpfs.GPFSNFSDriver(configuration=conf. + Configuration(None)) + self.driver.gpfs_execute = self._execute_wrapper + self.driver.set_execute(self._execute_wrapper) + self.context = context.get_admin_context() + self.context.user_id = 'fake' + self.context.project_id = 'fake' + + @mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSNFSDriver.' + '_run_ssh') + def test_gpfs_remote_execute(self, mock_run_ssh): + mock_run_ssh.return_value = 'test' + self.driver._gpfs_remote_execute('test', check_exit_code=True) + expected = [mock.call(('test',), True)] + self.assertEqual(expected, mock_run_ssh.mock_calls) + + @mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSNFSDriver.' + '_ensure_shares_mounted') + def test_update_volume_stats(self, mock_ensure): + """Check update volume stats.""" + + mock_ensure.return_value = True + fake_avail = 80 * units.Gi + fake_size = 2 * fake_avail + fake_used = 10 * units.Gi + + with mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSNFSDriver.' + '_get_capacity_info', + return_value=(fake_avail, fake_size, fake_used)): + stats = self.driver.get_volume_stats() + self.assertEqual('GPFSNFS', stats['volume_backend_name']) + self.assertEqual('file', stats['storage_protocol']) + stats = self.driver.get_volume_stats(True) + self.assertEqual('GPFSNFS', stats['volume_backend_name']) + self.assertEqual('file', stats['storage_protocol']) + + def test_get_volume_path(self): + self.driver.configuration.gpfs_mount_point_base = ( + self.TEST_GPFS_MNT_POINT_BASE) + volume = self._fake_volume() + self.assertEqual('/export/consisgroup-cg-1234/test', + self.driver._get_volume_path(volume)) + volume['consistencygroup_id'] = None + self.assertEqual('/export/test', + self.driver._get_volume_path(volume)) + + @mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSNFSDriver.' + '_get_mount_point_for_share') + def test_local_path(self, mock_mount_point): + mock_mount_point.return_value = self.TEST_MNT_POINT + volume = self._fake_volume() + volume['provider_location'] = self.TEST_GPFS_MNT_POINT_BASE + self.assertEqual('/mnt/nfs/consisgroup-cg-1234/test', + self.driver.local_path(volume)) + volume['consistencygroup_id'] = None + self.assertEqual('/mnt/nfs/test', + self.driver.local_path(volume)) + + @mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSNFSDriver.' + '_get_volume_path') + def test_get_snapshot_path(self, mock_volume_path): + volume = self._fake_volume() + self.driver.db = mock.Mock() + self.driver.db.volume_get = mock.Mock() + self.driver.db.volume_get.return_value = volume + mock_volume_path.return_value = os.path.join(self. + TEST_GPFS_MNT_POINT_BASE, + volume['name']) + snapshot = self._fake_snapshot() + self.assertEqual('/export/test-snap', + self.driver._get_snapshot_path(snapshot)) + + @mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSNFSDriver.' + '_find_share') + @mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSDriver.' + 'create_volume') + def test_create_volume(self, + mock_create_volume, + mock_find_share): + volume = self._fake_volume() + mock_find_share.return_value = self.TEST_VOLUME_PATH + self.assertEqual({'provider_location': self.TEST_VOLUME_PATH}, + self.driver.create_volume(volume)) + + @mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSDriver.' + '_delete_gpfs_file') + @mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSNFSDriver.' + 'local_path') + @mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSNFSDriver.' + '_get_volume_path') + @mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSDriver.' + '_verify_gpfs_path_state') + def test_delete_volume(self, + mock_verify_gpfs_path_state, + mock_volume_path, + mock_local_path, + mock_delete_gpfs_file): + self.driver.delete_volume('') + + @mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSDriver.' + 'delete_snapshot') + def test_delete_snapshot(self, + mock_delete_snapshot): + self.driver.delete_snapshot('') + + @mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSDriver.' + '_resize_volume_file') + @mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSNFSDriver.' + '_find_share') + @mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSDriver.' + '_create_volume_from_snapshot') + def test_create_volume_from_snapshot(self, + mock_create_volume_from_snapshot, + mock_find_share, + mock_resize_volume_file): + volume = self._fake_volume() + snapshot = self._fake_snapshot() + mock_find_share.return_value = self.TEST_VOLUME_PATH + self.assertEqual({'provider_location': self.TEST_VOLUME_PATH}, + self.driver.create_volume_from_snapshot(volume, + snapshot)) + + @mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSDriver.' + '_resize_volume_file') + @mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSNFSDriver.' + '_find_share') + @mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSDriver.' + '_create_cloned_volume') + def test_create_cloned_volume(self, + mock_create_cloned_volume, + mock_find_share, + mock_resize_volume_file): + volume = self._fake_volume() + src_vref = self._fake_volume() + mock_find_share.return_value = self.TEST_VOLUME_PATH + self.assertEqual({'provider_location': self.TEST_VOLUME_PATH}, + self.driver.create_cloned_volume(volume, src_vref)) + + @mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSDriver.' + '_delete_gpfs_file') + @mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSDriver.' + '_do_backup') + @mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSDriver.' + '_create_backup_source') + @mock.patch('cinder.volume.drivers.ibm.gpfs.GPFSNFSDriver.' + 'local_path') + def test_backup_volume(self, + mock_local_path, + mock_create_backup_source, + mock_do_backup, + mock_delete_gpfs_file): + volume = self._fake_volume() + self.driver.db = mock.Mock() + self.driver.db.volume_get = mock.Mock() + self.driver.db.volume_get.return_value = volume + backup = {} + backup['volume_id'] = 'test' + backup['id'] = '123456' + backup_service = mock.Mock() + mock_local_path.return_value = self.TEST_VOLUME_PATH + self.driver.backup_volume('', backup, backup_service) diff --git a/cinder/volume/drivers/ibm/gpfs.py b/cinder/volume/drivers/ibm/gpfs.py index 02b1870ba..ae3853a39 100644 --- a/cinder/volume/drivers/ibm/gpfs.py +++ b/cinder/volume/drivers/ibm/gpfs.py @@ -34,8 +34,12 @@ from cinder.image import image_utils from cinder.openstack.common import fileutils from cinder import utils from cinder.volume import driver +from cinder.volume.drivers import nfs +from cinder.volume.drivers import remotefs +from cinder.volume.drivers.san import san GPFS_CLONE_MIN_RELEASE = 1200 +MIGRATION_ALLOWED_DEST_TYPE = ['GPFSDriver', 'GPFSNFSDriver'] LOG = logging.getLogger(__name__) @@ -103,25 +107,38 @@ def _sizestr(size_in_g): return '%sG' % size_in_g -class GPFSDriver(driver.VolumeDriver): +class GPFSDriver(driver.ConsistencyGroupVD, driver.ExtendVD, + driver.LocalVD, driver.TransferVD, driver.CloneableVD, + driver.CloneableImageVD, driver.SnapshotVD, + driver.RetypeVD, driver.MigrateVD, + driver.BaseVD): """Implements volume functions using GPFS primitives. Version history: 1.0.0 - Initial driver 1.1.0 - Add volume retype, refactor volume migration 1.2.0 - Add consistency group support + 1.3.0 - Add NFS based GPFS storage backend support """ - VERSION = "1.2.0" + VERSION = "1.3.0" def __init__(self, *args, **kwargs): super(GPFSDriver, self).__init__(*args, **kwargs) self.configuration.append_config_values(gpfs_opts) + self.gpfs_execute = self._gpfs_local_execute + self._execute = utils.execute + + def _gpfs_local_execute(self, *cmd, **kwargs): + if 'run_as_root' not in kwargs: + kwargs.update({'run_as_root': True}) + + return utils.execute(*cmd, **kwargs) def _get_gpfs_state(self): """Return GPFS state information.""" try: - (out, err) = self._execute('mmgetstate', '-Y', run_as_root=True) + (out, err) = self.gpfs_execute('mmgetstate', '-Y') return out except processutils.ProcessExecutionError as exc: LOG.error(_LE('Failed to issue mmgetstate command, error: %s.'), @@ -142,7 +159,7 @@ class GPFSDriver(driver.VolumeDriver): def _get_filesystem_from_path(self, path): """Return filesystem for specified path.""" try: - (out, err) = self._execute('df', path, run_as_root=True) + (out, err) = self.gpfs_execute('df', path) lines = out.splitlines() filesystem = lines[1].split()[0] return filesystem @@ -156,8 +173,7 @@ class GPFSDriver(driver.VolumeDriver): def _get_gpfs_cluster_id(self): """Return the id for GPFS cluster being used.""" try: - (out, err) = self._execute('mmlsconfig', 'clusterId', '-Y', - run_as_root=True) + (out, err) = self.gpfs_execute('mmlsconfig', 'clusterId', '-Y') lines = out.splitlines() value_token = lines[0].split(':').index('value') cluster_id = lines[1].split(':')[value_token] @@ -171,8 +187,7 @@ class GPFSDriver(driver.VolumeDriver): """Return the GPFS fileset for specified path.""" fs_regex = re.compile(r'.*fileset.name:\s+(?P\w+)', re.S) try: - (out, err) = self._execute('mmlsattr', '-L', path, - run_as_root=True) + (out, err) = self.gpfs_execute('mmlsattr', '-L', path) except processutils.ProcessExecutionError as exc: LOG.error(_LE('Failed to issue mmlsattr command on path %(path)s, ' 'error: %(error)s'), @@ -193,8 +208,7 @@ class GPFSDriver(driver.VolumeDriver): def _verify_gpfs_pool(self, storage_pool): """Return true if the specified pool is a valid GPFS storage pool.""" try: - self._execute('mmlspool', self._gpfs_device, storage_pool, - run_as_root=True) + self.gpfs_execute('mmlspool', self._gpfs_device, storage_pool) return True except processutils.ProcessExecutionError: return False @@ -211,8 +225,7 @@ class GPFSDriver(driver.VolumeDriver): raise exception.VolumeBackendAPIException(data=msg) try: - self._execute('mmchattr', '-P', new_pool, local_path, - run_as_root=True) + self.gpfs_execute('mmchattr', '-P', new_pool, local_path) LOG.debug('Updated storage pool with mmchattr to %s.', new_pool) return True except processutils.ProcessExecutionError as exc: @@ -229,8 +242,7 @@ class GPFSDriver(driver.VolumeDriver): """ filesystem = self._get_filesystem_from_path(path) try: - (out, err) = self._execute('mmlsfs', filesystem, '-V', '-Y', - run_as_root=True) + (out, err) = self.gpfs_execute('mmlsfs', filesystem, '-V', '-Y') except processutils.ProcessExecutionError as exc: LOG.error(_LE('Failed to issue mmlsfs command for path %(path)s, ' 'error: %(error)s.'), @@ -249,8 +261,9 @@ class GPFSDriver(driver.VolumeDriver): def _get_gpfs_cluster_release_level(self): """Return the GPFS version of current cluster.""" try: - (out, err) = self._execute('mmlsconfig', 'minreleaseLeveldaemon', - '-Y', run_as_root=True) + (out, err) = self.gpfs_execute('mmlsconfig', + 'minreleaseLeveldaemon', + '-Y') except processutils.ProcessExecutionError as exc: LOG.error(_LE('Failed to issue mmlsconfig command, error: %s.'), exc.stderr) @@ -267,7 +280,7 @@ class GPFSDriver(driver.VolumeDriver): If not part of a gpfs file system, raise ProcessExecutionError. """ try: - self._execute('mmlsattr', directory, run_as_root=True) + self.gpfs_execute('mmlsattr', directory) except processutils.ProcessExecutionError as exc: LOG.error(_LE('Failed to issue mmlsattr command ' 'for path %(path)s, ' @@ -292,7 +305,7 @@ class GPFSDriver(driver.VolumeDriver): def _set_rw_permission(self, path, modebits='660'): """Set permission bits for the path.""" - self._execute('chmod', modebits, path, run_as_root=True) + self.gpfs_execute('chmod', modebits, path) def _can_migrate_locally(self, host): """Return true if the host can migrate a volume locally.""" @@ -307,7 +320,8 @@ class GPFSDriver(driver.VolumeDriver): LOG.debug('Evaluate migration: unexpected location info, ' 'cannot migrate locally: %s.', info) return None - if dest_type != 'GPFSDriver' or dest_id != self._cluster_id: + if (dest_id != self._cluster_id or + dest_type not in MIGRATION_ALLOWED_DEST_TYPE): LOG.debug('Evaluate migration: different destination driver or ' 'cluster id in location info: %s.', info) return None @@ -428,7 +442,7 @@ class GPFSDriver(driver.VolumeDriver): """Creates file with 0 disk usage.""" sizestr = _sizestr(size) - self._execute('truncate', '-s', sizestr, path, run_as_root=True) + self.gpfs_execute('truncate', '-s', sizestr, path) def _allocate_file_blocks(self, path, size): """Preallocate file blocks by writing zeros.""" @@ -436,10 +450,9 @@ class GPFSDriver(driver.VolumeDriver): block_size_mb = 1 block_count = size * units.Gi / (block_size_mb * units.Mi) - self._execute('dd', 'if=/dev/zero', 'of=%s' % path, - 'bs=%dM' % block_size_mb, - 'count=%d' % block_count, - run_as_root=True) + self.gpfs_execute('dd', 'if=/dev/zero', 'of=%s' % path, + 'bs=%dM' % block_size_mb, + 'count=%d' % block_count) def _gpfs_change_attributes(self, options, path): """Update GPFS attributes on the specified file.""" @@ -448,7 +461,7 @@ class GPFSDriver(driver.VolumeDriver): cmd.extend(options) cmd.append(path) LOG.debug('Update volume attributes with mmchattr to %s.', options) - self._execute(*cmd, run_as_root=True) + self.gpfs_execute(*cmd) def _set_volume_attributes(self, volume, path, metadata): """Set various GPFS attributes for this volume.""" @@ -493,7 +506,7 @@ class GPFSDriver(driver.VolumeDriver): # Check if GPFS is mounted self._verify_gpfs_path_state(self.configuration.gpfs_mount_point_base) - volume_path = self.local_path(volume) + volume_path = self._get_volume_path(volume) volume_size = volume['size'] # Create a sparse file first; allocate blocks later if requested @@ -507,9 +520,7 @@ class GPFSDriver(driver.VolumeDriver): if not self.configuration.gpfs_sparse_volumes: self._allocate_file_blocks(volume_path, volume_size) - def create_volume_from_snapshot(self, volume, snapshot): - """Creates a GPFS volume from a snapshot.""" - + def _create_volume_from_snapshot(self, volume, snapshot): snapshot_path = self._get_snapshot_path(snapshot) # check if the snapshot lies in the same CG as the volume to be created # if yes, clone the volume from the snapshot, else perform full copy @@ -520,8 +531,7 @@ class GPFSDriver(driver.VolumeDriver): if (volume['consistencygroup_id'] == snap_parent_vol['consistencygroup_id']): clone = True - - volume_path = self.local_path(volume) + volume_path = self._get_volume_path(volume) if clone: self._create_gpfs_copy(src=snapshot_path, dest=volume_path) self._gpfs_redirect(volume_path) @@ -531,44 +541,70 @@ class GPFSDriver(driver.VolumeDriver): self._set_rw_permission(volume_path) v_metadata = volume.get('volume_metadata') self._set_volume_attributes(volume, volume_path, v_metadata) + + def create_volume_from_snapshot(self, volume, snapshot): + """Creates a GPFS volume from a snapshot.""" + self._create_volume_from_snapshot(volume, snapshot) virt_size = self._resize_volume_file(volume, volume['size']) return {'size': math.ceil(virt_size / units.Gi)} - def create_cloned_volume(self, volume, src_vref): - """Create a GPFS volume from another volume.""" + def _get_volume_path(self, volume): + return self.local_path(volume) - src = self.local_path(src_vref) - dest = self.local_path(volume) + def _create_cloned_volume(self, volume, src_vref): + src = self._get_volume_path(src_vref) + dest = self._get_volume_path(volume) if (volume['consistencygroup_id'] == src_vref['consistencygroup_id']): self._create_gpfs_clone(src, dest) else: self._gpfs_full_copy(src, dest) - self._set_rw_permission(dest) v_metadata = volume.get('volume_metadata') self._set_volume_attributes(volume, dest, v_metadata) + + def create_cloned_volume(self, volume, src_vref): + """Create a GPFS volume from another volume.""" + self._create_cloned_volume(volume, src_vref) virt_size = self._resize_volume_file(volume, volume['size']) return {'size': math.ceil(virt_size / units.Gi)} - def _delete_gpfs_file(self, fchild): + def _delete_gpfs_file(self, fchild, mount_point=None): """Delete a GPFS file and cleanup clone children.""" - if not os.path.exists(fchild): - return - (out, err) = self._execute('mmclone', 'show', fchild, run_as_root=True) + if mount_point is None: + if not os.path.exists(fchild): + return + else: + fchild_local_path = os.path.join(mount_point, + os.path.basename(fchild)) + if not os.path.exists(fchild_local_path): + return + + (out, err) = self.gpfs_execute('mmclone', 'show', fchild) fparent = None + delete_parent = False inode_regex = re.compile( r'.*\s+(?:yes|no)\s+\d+\s+(?P\d+)', re.M | re.S) match = inode_regex.match(out) if match: inode = match.group('inode') - path = os.path.dirname(fchild) + if mount_point is None: + path = os.path.dirname(fchild) + else: + path = mount_point + (out, err) = self._execute('find', path, '-maxdepth', '1', '-inum', inode, run_as_root=True) if out: fparent = out.split('\n', 1)[0] - self._execute( - 'rm', '-f', fchild, check_exit_code=False, run_as_root=True) + + if mount_point is None: + self._execute( + 'rm', '-f', fchild, check_exit_code=False, run_as_root=True) + else: + self._execute( + 'rm', '-f', fchild_local_path, check_exit_code=False, + run_as_root=True) # There is no need to check for volume references on this snapshot # because 'rm -f' itself serves as a simple and implicit check. If the @@ -578,10 +614,24 @@ class GPFSDriver(driver.VolumeDriver): # If there are such volumes, we quit recursion and let the other # volumes delete the snapshot later. If there are no references, rm # would succeed and the snapshot is deleted. - if not os.path.exists(fchild) and fparent: + if mount_point is None: + if not os.path.exists(fchild) and fparent: + delete_parent = True + else: + if not os.path.exists(fchild_local_path) and fparent: + delete_parent = True + + if delete_parent: fpbase = os.path.basename(fparent) if fpbase.endswith('.snap') or fpbase.endswith('.ts'): - self._delete_gpfs_file(fparent) + if mount_point is None: + self._delete_gpfs_file(fparent) + else: + fparent_remote_path = os.path.join(os.path.dirname(fchild), + fpbase) + fparent_mount_path = os.path.dirname(fparent) + self._delete_gpfs_file(fparent_remote_path, + fparent_mount_path) def delete_volume(self, volume): """Deletes a logical volume.""" @@ -601,13 +651,13 @@ class GPFSDriver(driver.VolumeDriver): max_depth = self.configuration.gpfs_max_clone_depth if max_depth == 0: return False - (out, err) = self._execute('mmclone', 'show', src, run_as_root=True) + (out, err) = self.gpfs_execute('mmclone', 'show', src) depth_regex = re.compile(r'.*\s+no\s+(?P\d+)', re.M | re.S) match = depth_regex.match(out) if match: depth = int(match.group('depth')) if depth > max_depth: - self._execute('mmclone', 'redirect', src, run_as_root=True) + self.gpfs_execute('mmclone', 'redirect', src) return True return False @@ -621,24 +671,22 @@ class GPFSDriver(driver.VolumeDriver): def _create_gpfs_copy(self, src, dest): """Create a GPFS file clone copy for the specified file.""" - self._execute('mmclone', 'copy', src, dest, run_as_root=True) + self.gpfs_execute('mmclone', 'copy', src, dest) def _gpfs_full_copy(self, src, dest): """Create a full copy from src to dest.""" - self._execute('cp', src, dest, - check_exit_code=True, run_as_root=True) + self.gpfs_execute('cp', src, dest, check_exit_code=True) def _create_gpfs_snap(self, src, dest=None): """Create a GPFS file clone snapshot for the specified file.""" if dest is None: - self._execute('mmclone', 'snap', src, run_as_root=True) + self.gpfs_execute('mmclone', 'snap', src) else: - self._execute('mmclone', 'snap', src, dest, run_as_root=True) + self.gpfs_execute('mmclone', 'snap', src, dest) def _is_gpfs_parent_file(self, gpfs_file): """Return true if the specified file is a gpfs clone parent.""" - out, err = self._execute('mmclone', 'show', gpfs_file, - run_as_root=True) + out, err = self.gpfs_execute('mmclone', 'show', gpfs_file) ptoken = out.splitlines().pop().split()[0] return ptoken == 'yes' @@ -660,9 +708,9 @@ class GPFSDriver(driver.VolumeDriver): # snapshots will also be deleted. snapshot_path = self._get_snapshot_path(snapshot) snapshot_ts_path = '%s.ts' % snapshot_path - self._execute('mv', snapshot_path, snapshot_ts_path, run_as_root=True) - self._execute('rm', '-f', snapshot_ts_path, - check_exit_code=False, run_as_root=True) + self.gpfs_execute('mv', snapshot_path, snapshot_ts_path) + self.gpfs_execute('rm', '-f', snapshot_ts_path, + check_exit_code=False) def _get_snapshot_path(self, snapshot): ctxt = context.get_admin_context() @@ -867,21 +915,28 @@ class GPFSDriver(driver.VolumeDriver): image_meta, self.local_path(volume)) + def _create_backup_source(self, volume, backup): + src_path = self._get_volume_path(volume) + dest_path = '%s_%s' % (src_path, backup['id']) + self._create_gpfs_clone(src_path, dest_path) + self._gpfs_redirect(src_path) + return dest_path + + def _do_backup(self, backup_path, backup, backup_service): + with utils.temporary_chown(backup_path): + with fileutils.file_open(backup_path) as backup_file: + backup_service.backup(backup, backup_file) + def backup_volume(self, context, backup, backup_service): """Create a new backup from an existing volume.""" volume = self.db.volume_get(context, backup['volume_id']) volume_path = self.local_path(volume) - LOG.debug('Begin backup of volume %s.', volume['name']) - - # create a snapshot that will be used as the backup source backup_path = '%s_%s' % (volume_path, backup['id']) - self._create_gpfs_clone(volume_path, backup_path) - self._gpfs_redirect(volume_path) - + # create a snapshot that will be used as the backup source + self._create_backup_source(volume, backup) try: - with utils.temporary_chown(backup_path): - with fileutils.file_open(backup_path) as backup_file: - backup_service.backup(backup, backup_file) + LOG.debug('Begin backup of volume %s.', volume['name']) + self._do_backup(backup_path, backup, backup_service) finally: # clean up snapshot file. If it is a clone parent, delete # will fail silently, but be cleaned up when volume is @@ -915,10 +970,10 @@ class GPFSDriver(driver.VolumeDriver): LOG.debug('Migration target is same cluster but different path, ' 'move the volume file.') - local_path = self.local_path(volume) + local_path = self._get_volume_path(volume) new_path = os.path.join(dest_path, volume['name']) try: - self._execute('mv', local_path, new_path, run_as_root=True) + self.gpfs_execute('mv', local_path, new_path) return (True, None) except processutils.ProcessExecutionError as exc: LOG.error(_LE('Driver-based migration of volume %(vol)s failed. ' @@ -1044,8 +1099,8 @@ class GPFSDriver(driver.VolumeDriver): cgpath = os.path.join(self.configuration.gpfs_mount_point_base, cgname) try: - self._execute('mmcrfileset', fsdev, cgname, - '--inode-space', 'new', run_as_root=True) + self.gpfs_execute('mmcrfileset', fsdev, cgname, + '--inode-space', 'new') except processutils.ProcessExecutionError as e: msg = (_('Failed to create consistency group: %(cgid)s. ' 'Error: %(excmsg)s.') % @@ -1054,8 +1109,8 @@ class GPFSDriver(driver.VolumeDriver): raise exception.VolumeBackendAPIException(data=msg) try: - self._execute('mmlinkfileset', fsdev, cgname, - '-J', cgpath, run_as_root=True) + self.gpfs_execute('mmlinkfileset', fsdev, cgname, + '-J', cgpath) except processutils.ProcessExecutionError as e: msg = (_('Failed to link fileset for the share %(cgname)s. ' 'Error: %(excmsg)s.') % @@ -1064,7 +1119,7 @@ class GPFSDriver(driver.VolumeDriver): raise exception.VolumeBackendAPIException(data=msg) try: - self._execute('chmod', '770', cgpath, run_as_root=True) + self.gpfs_execute('chmod', '770', cgpath) except processutils.ProcessExecutionError as e: msg = (_('Failed to set permissions for the consistency group ' '%(cgname)s. ' @@ -1088,8 +1143,7 @@ class GPFSDriver(driver.VolumeDriver): # Unlink and delete the fileset associated with the consistency group. # All of the volumes and volume snapshot data will also be deleted. try: - self._execute('mmunlinkfileset', fsdev, cgname, '-f', - run_as_root=True) + self.gpfs_execute('mmunlinkfileset', fsdev, cgname, '-f') except processutils.ProcessExecutionError as e: msg = (_('Failed to unlink fileset for consistency group ' '%(cgname)s. Error: %(excmsg)s.') % @@ -1098,8 +1152,7 @@ class GPFSDriver(driver.VolumeDriver): raise exception.VolumeBackendAPIException(data=msg) try: - self._execute('mmdelfileset', fsdev, cgname, '-f', - run_as_root=True) + self.gpfs_execute('mmdelfileset', fsdev, cgname, '-f') except processutils.ProcessExecutionError as e: msg = (_('Failed to delete fileset for consistency group ' '%(cgname)s. Error: %(excmsg)s.') % @@ -1139,3 +1192,161 @@ class GPFSDriver(driver.VolumeDriver): model_update = {'status': cgsnapshot['status']} return model_update, snapshots + + +class GPFSNFSDriver(GPFSDriver, nfs.NfsDriver, san.SanDriver): + """GPFS cinder driver extension. + + This extends the capability of existing GPFS cinder driver + to be able to create cinder volumes when cinder volume service + is not running on GPFS node. + """ + + def __init__(self, *args, **kwargs): + self._context = None + self._storage_pool = None + self._cluster_id = None + super(GPFSNFSDriver, self).__init__(*args, **kwargs) + self.gpfs_execute = self._gpfs_remote_execute + self.configuration.append_config_values(remotefs.nas_opts) + self.configuration.san_ip = self.configuration.nas_ip + self.configuration.san_login = self.configuration.nas_login + self.configuration.san_password = self.configuration.nas_password + self.configuration.san_private_key = ( + self.configuration.nas_private_key) + self.configuration.san_ssh_port = self.configuration.nas_ssh_port + + def _gpfs_remote_execute(self, *cmd, **kwargs): + check_exit_code = kwargs.pop('check_exit_code', None) + return self._run_ssh(cmd, check_exit_code) + + def do_setup(self, context): + super(GPFSNFSDriver, self).do_setup(context) + self._context = context + + def get_volume_stats(self, refresh=False): + """Get volume stats. + + If 'refresh' is True, or stats have never been updated, run update + the stats first. + """ + if not self._stats or refresh: + self._update_volume_stats() + + return self._stats + + def _update_volume_stats(self): + """Retrieve stats info from volume group.""" + + LOG.debug("Enter _update_volume_stats.") + gpfs_base = self.configuration.gpfs_mount_point_base + data = {} + backend_name = self.configuration.safe_get('volume_backend_name') + data['volume_backend_name'] = backend_name or 'GPFSNFS' + data['vendor_name'] = 'IBM' + data['driver_version'] = self.get_version() + data['storage_protocol'] = 'file' + + self._ensure_shares_mounted() + + global_capacity = 0 + global_free = 0 + for share in self._mounted_shares: + capacity, free, _used = self._get_capacity_info(share) + global_capacity += capacity + global_free += free + + data['total_capacity_gb'] = global_capacity / float(units.Gi) + data['free_capacity_gb'] = global_free / float(units.Gi) + data['reserved_percentage'] = 0 + data['QoS_support'] = False + data['storage_pool'] = self._storage_pool + data['location_info'] = ('GPFSNFSDriver:%(cluster_id)s:%(root_path)s' % + {'cluster_id': self._cluster_id, + 'root_path': gpfs_base}) + + data['consistencygroup_support'] = 'True' + self._stats = data + LOG.debug("Exit _update_volume_stats.") + + def _get_volume_path(self, volume): + """Returns remote GPFS path for the given volume.""" + export_path = self.configuration.gpfs_mount_point_base + if volume['consistencygroup_id'] is not None: + cgname = "consisgroup-%s" % volume['consistencygroup_id'] + volume_path = os.path.join(export_path, cgname, volume['name']) + else: + volume_path = os.path.join(export_path, volume['name']) + return volume_path + + def local_path(self, volume): + """Returns the local path for the specified volume.""" + remotefs_share = volume['provider_location'] + base_local_path = self._get_mount_point_for_share(remotefs_share) + + # Check if the volume is part of a consistency group and return + # the local_path accordingly. + if volume['consistencygroup_id'] is not None: + cgname = "consisgroup-%s" % volume['consistencygroup_id'] + volume_path = os.path.join(base_local_path, cgname, volume['name']) + else: + volume_path = os.path.join(base_local_path, volume['name']) + return volume_path + + def _get_snapshot_path(self, snapshot): + """Returns remote GPFS path for the given snapshot.""" + snap_parent_vol = self.db.volume_get(self._context, + snapshot['volume_id']) + snap_parent_vol_path = self._get_volume_path(snap_parent_vol) + snapshot_path = os.path.join(os.path.dirname(snap_parent_vol_path), + snapshot['name']) + return snapshot_path + + def create_volume(self, volume): + """Creates a GPFS volume.""" + super(GPFSNFSDriver, self).create_volume(volume) + volume['provider_location'] = self._find_share(volume['size']) + return {'provider_location': volume['provider_location']} + + def delete_volume(self, volume): + """Deletes a logical volume.""" + # Check if GPFS is mounted + self._verify_gpfs_path_state(self.configuration.gpfs_mount_point_base) + + volume_path = self._get_volume_path(volume) + mount_point = os.path.dirname(self.local_path(volume)) + # Delete all dependent snapshots, the snapshot will get deleted + # if the link count goes to zero, else rm will fail silently + self._delete_gpfs_file(volume_path, mount_point) + + def create_volume_from_snapshot(self, volume, snapshot): + """Creates a GPFS volume from a snapshot.""" + self._create_volume_from_snapshot(volume, snapshot) + volume['provider_location'] = self._find_share(volume['size']) + self._resize_volume_file(volume, volume['size']) + return {'provider_location': volume['provider_location']} + + def create_cloned_volume(self, volume, src_vref): + """Create a GPFS volume from another volume.""" + self._create_cloned_volume(volume, src_vref) + volume['provider_location'] = self._find_share(volume['size']) + self._resize_volume_file(volume, volume['size']) + return {'provider_location': volume['provider_location']} + + def backup_volume(self, context, backup, backup_service): + """Create a new backup from an existing volume.""" + volume = self.db.volume_get(context, backup['volume_id']) + volume_path = self.local_path(volume) + backup_path = '%s_%s' % (volume_path, backup['id']) + # create a snapshot that will be used as the backup source + backup_remote_path = self._create_backup_source(volume, backup) + try: + LOG.debug('Begin backup of volume %s.', volume['name']) + self._do_backup(backup_path, backup, backup_service) + finally: + # clean up snapshot file. If it is a clone parent, delete + # will fail silently, but be cleaned up when volume is + # eventually removed. This ensures we do not accumulate + # more than gpfs_max_clone_depth snap files. + backup_mount_path = os.path.dirname(backup_path) + self._delete_gpfs_file(backup_remote_path, backup_mount_path)