cinder/cinder/volume/drivers/netapp.py

1334 lines
55 KiB
Python

# vim: tabstop=4 shiftwidth=4 softtabstop=4
# Copyright (c) 2012 NetApp, Inc.
# Copyright (c) 2012 OpenStack LLC.
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
"""
Volume driver for NetApp storage systems.
This driver requires NetApp OnCommand 5.0 and one or more Data
ONTAP 7-mode storage systems with installed iSCSI licenses.
"""
import time
import suds
from suds import client
from suds.sax import text
from cinder import exception
from cinder import flags
from cinder.openstack.common import cfg
from cinder.openstack.common import log as logging
from cinder.volume import driver
from cinder.volume import volume_types
LOG = logging.getLogger(__name__)
netapp_opts = [
cfg.StrOpt('netapp_wsdl_url',
default=None,
help='URL of the WSDL file for the DFM server'),
cfg.StrOpt('netapp_login',
default=None,
help='User name for the DFM server'),
cfg.StrOpt('netapp_password',
default=None,
help='Password for the DFM server'),
cfg.StrOpt('netapp_server_hostname',
default=None,
help='Hostname for the DFM server'),
cfg.IntOpt('netapp_server_port',
default=8088,
help='Port number for the DFM server'),
cfg.StrOpt('netapp_storage_service',
default=None,
help=('Storage service to use for provisioning '
'(when volume_type=None)')),
cfg.StrOpt('netapp_storage_service_prefix',
default=None,
help=('Prefix of storage service name to use for '
'provisioning (volume_type name will be appended)')),
cfg.StrOpt('netapp_vfiler',
default=None,
help='Vfiler to use for provisioning'), ]
FLAGS = flags.FLAGS
FLAGS.register_opts(netapp_opts)
class DfmDataset(object):
def __init__(self, id, name, project, type):
self.id = id
self.name = name
self.project = project
self.type = type
class DfmLun(object):
def __init__(self, dataset, lunpath, id):
self.dataset = dataset
self.lunpath = lunpath
self.id = id
class NetAppISCSIDriver(driver.ISCSIDriver):
"""NetApp iSCSI volume driver."""
IGROUP_PREFIX = 'openstack-'
DATASET_PREFIX = 'OpenStack_'
DATASET_METADATA_PROJECT_KEY = 'OpenStackProject'
DATASET_METADATA_VOL_TYPE_KEY = 'OpenStackVolType'
def __init__(self, *args, **kwargs):
super(NetAppISCSIDriver, self).__init__(*args, **kwargs)
self.discovered_luns = []
self.discovered_datasets = []
self.lun_table = {}
def _check_fail(self, request, response):
"""Utility routine to handle checking ZAPI failures."""
if 'failed' == response.Status:
name = request.Name
reason = response.Reason
msg = _('API %(name)s failed: %(reason)s')
raise exception.VolumeBackendAPIException(data=msg % locals())
def _create_client(self, **kwargs):
"""Instantiate a web services client.
This method creates a "suds" client to make web services calls to the
DFM server. Note that the WSDL file is quite large and may take
a few seconds to parse.
"""
wsdl_url = kwargs['wsdl_url']
LOG.debug(_('Using WSDL: %s') % wsdl_url)
if kwargs['cache']:
self.client = client.Client(wsdl_url, username=kwargs['login'],
password=kwargs['password'])
else:
self.client = client.Client(wsdl_url, username=kwargs['login'],
password=kwargs['password'],
cache=None)
soap_url = 'http://%s:%s/apis/soap/v1' % (kwargs['hostname'],
kwargs['port'])
LOG.debug(_('Using DFM server: %s') % soap_url)
self.client.set_options(location=soap_url)
def _set_storage_service(self, storage_service):
"""Set the storage service to use for provisioning."""
LOG.debug(_('Using storage service: %s') % storage_service)
self.storage_service = storage_service
def _set_storage_service_prefix(self, storage_service_prefix):
"""Set the storage service prefix to use for provisioning."""
LOG.debug(_('Using storage service prefix: %s') %
storage_service_prefix)
self.storage_service_prefix = storage_service_prefix
def _set_vfiler(self, vfiler):
"""Set the vfiler to use for provisioning."""
LOG.debug(_('Using vfiler: %s') % vfiler)
self.vfiler = vfiler
def _check_flags(self):
"""Ensure that the flags we care about are set."""
required_flags = ['netapp_wsdl_url', 'netapp_login', 'netapp_password',
'netapp_server_hostname', 'netapp_server_port']
for flag in required_flags:
if not getattr(FLAGS, flag, None):
raise exception.InvalidInput(reason=_('%s is not set') % flag)
if not (FLAGS.netapp_storage_service or
FLAGS.netapp_storage_service_prefix):
raise exception.InvalidInput(
reason=_('Either '
'netapp_storage_service or '
'netapp_storage_service_prefix must '
'be set'))
def do_setup(self, context):
"""Setup the NetApp Volume driver.
Called one time by the manager after the driver is loaded.
Validate the flags we care about and setup the suds (web services)
client.
"""
self._check_flags()
self._create_client(
wsdl_url=FLAGS.netapp_wsdl_url,
login=FLAGS.netapp_login, password=FLAGS.netapp_password,
hostname=FLAGS.netapp_server_hostname,
port=FLAGS.netapp_server_port, cache=True)
self._set_storage_service(FLAGS.netapp_storage_service)
self._set_storage_service_prefix(FLAGS.netapp_storage_service_prefix)
self._set_vfiler(FLAGS.netapp_vfiler)
def check_for_setup_error(self):
"""Check that the driver is working and can communicate.
Invoke a web services API to make sure we can talk to the server.
Also perform the discovery of datasets and LUNs from DFM.
"""
self.client.service.DfmAbout()
LOG.debug(_("Connected to DFM server"))
self._discover_luns()
def _get_datasets(self):
"""Get the list of datasets from DFM."""
server = self.client.service
res = server.DatasetListInfoIterStart(IncludeMetadata=True)
tag = res.Tag
datasets = []
try:
while True:
res = server.DatasetListInfoIterNext(Tag=tag, Maximum=100)
if not res.Datasets:
break
datasets.extend(res.Datasets.DatasetInfo)
finally:
server.DatasetListInfoIterEnd(Tag=tag)
return datasets
def _discover_dataset_luns(self, dataset, volume):
"""Discover all of the LUNs in a dataset."""
server = self.client.service
res = server.DatasetMemberListInfoIterStart(
DatasetNameOrId=dataset.id,
IncludeExportsInfo=True,
IncludeIndirect=True,
MemberType='lun_path')
tag = res.Tag
suffix = None
if volume:
suffix = '/' + volume
try:
while True:
res = server.DatasetMemberListInfoIterNext(Tag=tag,
Maximum=100)
if (not hasattr(res, 'DatasetMembers') or
not res.DatasetMembers):
break
for member in res.DatasetMembers.DatasetMemberInfo:
if suffix and not member.MemberName.endswith(suffix):
continue
# MemberName is the full LUN path in this format:
# host:/volume/qtree/lun
lun = DfmLun(dataset, member.MemberName, member.MemberId)
self.discovered_luns.append(lun)
finally:
server.DatasetMemberListInfoIterEnd(Tag=tag)
def _discover_luns(self):
"""Discover the LUNs from DFM.
Discover all of the OpenStack-created datasets and LUNs in the DFM
database.
"""
datasets = self._get_datasets()
self.discovered_datasets = []
self.discovered_luns = []
for dataset in datasets:
if not dataset.DatasetName.startswith(self.DATASET_PREFIX):
continue
if (not hasattr(dataset, 'DatasetMetadata') or
not dataset.DatasetMetadata):
continue
project = None
type = None
for field in dataset.DatasetMetadata.DfmMetadataField:
if field.FieldName == self.DATASET_METADATA_PROJECT_KEY:
project = field.FieldValue
elif field.FieldName == self.DATASET_METADATA_VOL_TYPE_KEY:
type = field.FieldValue
if not project:
continue
ds = DfmDataset(dataset.DatasetId, dataset.DatasetName,
project, type)
self.discovered_datasets.append(ds)
self._discover_dataset_luns(ds, None)
dataset_count = len(self.discovered_datasets)
lun_count = len(self.discovered_luns)
msg = _("Discovered %(dataset_count)s datasets and %(lun_count)s LUNs")
LOG.debug(msg % locals())
self.lun_table = {}
def _get_job_progress(self, job_id):
"""Get progress of one running DFM job.
Obtain the latest progress report for the job and return the
list of progress events.
"""
server = self.client.service
res = server.DpJobProgressEventListIterStart(JobId=job_id)
tag = res.Tag
event_list = []
try:
while True:
res = server.DpJobProgressEventListIterNext(Tag=tag,
Maximum=100)
if not hasattr(res, 'ProgressEvents'):
break
event_list += res.ProgressEvents.DpJobProgressEventInfo
finally:
server.DpJobProgressEventListIterEnd(Tag=tag)
return event_list
def _wait_for_job(self, job_id):
"""Wait until a job terminates.
Poll the job until it completes or an error is detected. Return the
final list of progress events if it completes successfully.
"""
while True:
events = self._get_job_progress(job_id)
for event in events:
if event.EventStatus == 'error':
msg = _('Job failed: %s') % (event.ErrorMessage)
raise exception.VolumeBackendAPIException(data=msg)
if event.EventType == 'job-end':
return events
time.sleep(5)
def _dataset_name(self, project, ss_type):
"""Return the dataset name for a given project and volume type."""
_project = project.replace(' ', '_').replace('-', '_')
dataset_name = self.DATASET_PREFIX + _project
if not ss_type:
return dataset_name
_type = ss_type.replace(' ', '_').replace('-', '_')
return dataset_name + '_' + _type
def _get_dataset(self, dataset_name):
"""Lookup a dataset by name in the list of discovered datasets."""
for dataset in self.discovered_datasets:
if dataset.name == dataset_name:
return dataset
return None
def _create_dataset(self, dataset_name, project, ss_type):
"""Create a new dataset using the storage service.
The export settings are set to create iSCSI LUNs aligned for Linux.
Returns the ID of the new dataset.
"""
if ss_type and not self.storage_service_prefix:
msg = _('Attempt to use volume_type without specifying '
'netapp_storage_service_prefix flag.')
raise exception.VolumeBackendAPIException(data=msg)
if not (ss_type or self.storage_service):
msg = _('You must set the netapp_storage_service flag in order to '
'create volumes with no volume_type.')
raise exception.VolumeBackendAPIException(data=msg)
storage_service = self.storage_service
if ss_type:
storage_service = self.storage_service_prefix + ss_type
factory = self.client.factory
lunmap = factory.create('DatasetLunMappingInfo')
lunmap.IgroupOsType = 'linux'
export = factory.create('DatasetExportInfo')
export.DatasetExportProtocol = 'iscsi'
export.DatasetLunMappingInfo = lunmap
detail = factory.create('StorageSetInfo')
detail.DpNodeName = 'Primary data'
detail.DatasetExportInfo = export
if hasattr(self, 'vfiler') and self.vfiler:
detail.ServerNameOrId = self.vfiler
details = factory.create('ArrayOfStorageSetInfo')
details.StorageSetInfo = [detail]
field1 = factory.create('DfmMetadataField')
field1.FieldName = self.DATASET_METADATA_PROJECT_KEY
field1.FieldValue = project
field2 = factory.create('DfmMetadataField')
field2.FieldName = self.DATASET_METADATA_VOL_TYPE_KEY
field2.FieldValue = ss_type
metadata = factory.create('ArrayOfDfmMetadataField')
metadata.DfmMetadataField = [field1, field2]
res = self.client.service.StorageServiceDatasetProvision(
StorageServiceNameOrId=storage_service,
DatasetName=dataset_name,
AssumeConfirmation=True,
StorageSetDetails=details,
DatasetMetadata=metadata)
ds = DfmDataset(res.DatasetId, dataset_name, project, ss_type)
self.discovered_datasets.append(ds)
return ds
def _provision(self, name, description, project, ss_type, size):
"""Provision a LUN through provisioning manager.
The LUN will be created inside a dataset associated with the project.
If the dataset doesn't already exist, we create it using the storage
service specified in the cinder conf.
"""
dataset_name = self._dataset_name(project, ss_type)
dataset = self._get_dataset(dataset_name)
if not dataset:
dataset = self._create_dataset(dataset_name, project, ss_type)
info = self.client.factory.create('ProvisionMemberRequestInfo')
info.Name = name
if description:
info.Description = description
info.Size = size
info.MaximumSnapshotSpace = 2 * long(size)
server = self.client.service
lock_id = server.DatasetEditBegin(DatasetNameOrId=dataset.id)
try:
server.DatasetProvisionMember(EditLockId=lock_id,
ProvisionMemberRequestInfo=info)
res = server.DatasetEditCommit(EditLockId=lock_id,
AssumeConfirmation=True)
except (suds.WebFault, Exception):
server.DatasetEditRollback(EditLockId=lock_id)
msg = _('Failed to provision dataset member')
raise exception.VolumeBackendAPIException(data=msg)
lun_id = None
lunpath = None
for info in res.JobIds.JobInfo:
events = self._wait_for_job(info.JobId)
for event in events:
if event.EventType != 'lun-create':
continue
lunpath = event.ProgressLunInfo.LunName
lun_id = event.ProgressLunInfo.LunPathId
if not lun_id:
msg = _('No LUN was created by the provision job')
raise exception.VolumeBackendAPIException(data=msg)
lun = DfmLun(dataset, lunpath, lun_id)
self.discovered_luns.append(lun)
self.lun_table[name] = lun
def _get_ss_type(self, volume):
"""Get the storage service type for a volume."""
id = volume['volume_type_id']
if not id:
return None
volume_type = volume_types.get_volume_type(None, id)
if not volume_type:
return None
return volume_type['name']
def _remove_destroy(self, name, project):
"""Remove the LUN from the dataset, also destroying it.
Remove the LUN from the dataset and destroy the actual LUN and Qtree
on the storage system.
"""
lun = self._lookup_lun_for_volume(name, project)
lun_details = self._get_lun_details(lun.id)
member = self.client.factory.create('DatasetMemberParameter')
member.ObjectNameOrId = lun.id
members = self.client.factory.create('ArrayOfDatasetMemberParameter')
members.DatasetMemberParameter = [member]
server = self.client.service
lock_id = server.DatasetEditBegin(DatasetNameOrId=lun.dataset.id)
try:
server.DatasetRemoveMember(EditLockId=lock_id, Destroy=True,
DatasetMemberParameters=members)
res = server.DatasetEditCommit(EditLockId=lock_id,
AssumeConfirmation=True)
except (suds.WebFault, Exception):
server.DatasetEditRollback(EditLockId=lock_id)
msg = _('Failed to remove and delete dataset LUN member')
raise exception.VolumeBackendAPIException(data=msg)
for info in res.JobIds.JobInfo:
self._wait_for_job(info.JobId)
# Note: it's not possible to delete Qtree & his LUN in one transaction
member.ObjectNameOrId = lun_details.QtreeId
lock_id = server.DatasetEditBegin(DatasetNameOrId=lun.dataset.id)
try:
server.DatasetRemoveMember(EditLockId=lock_id, Destroy=True,
DatasetMemberParameters=members)
server.DatasetEditCommit(EditLockId=lock_id,
AssumeConfirmation=True)
except (suds.WebFault, Exception):
server.DatasetEditRollback(EditLockId=lock_id)
msg = _('Failed to remove and delete dataset Qtree member')
raise exception.VolumeBackendAPIException(data=msg)
def create_volume(self, volume):
"""Driver entry point for creating a new volume."""
default_size = '104857600' # 100 MB
gigabytes = 1073741824L # 2^30
name = volume['name']
project = volume['project_id']
display_name = volume['display_name']
display_description = volume['display_description']
description = None
if display_name:
if display_description:
description = display_name + "\n" + display_description
else:
description = display_name
elif display_description:
description = display_description
if int(volume['size']) == 0:
size = default_size
else:
size = str(int(volume['size']) * gigabytes)
ss_type = self._get_ss_type(volume)
self._provision(name, description, project, ss_type, size)
def _lookup_lun_for_volume(self, name, project):
"""Lookup the LUN that corresponds to the give volume.
Initial lookups involve a table scan of all of the discovered LUNs,
but later lookups are done instantly from the hashtable.
"""
if name in self.lun_table:
return self.lun_table[name]
lunpath_suffix = '/' + name
for lun in self.discovered_luns:
if lun.dataset.project != project:
continue
if lun.lunpath.endswith(lunpath_suffix):
self.lun_table[name] = lun
return lun
msg = _("No entry in LUN table for volume %s") % (name)
raise exception.VolumeBackendAPIException(data=msg)
def delete_volume(self, volume):
"""Driver entry point for destroying existing volumes."""
name = volume['name']
project = volume['project_id']
self._remove_destroy(name, project)
def _get_lun_details(self, lun_id):
"""Given the ID of a LUN, get the details about that LUN."""
server = self.client.service
res = server.LunListInfoIterStart(ObjectNameOrId=lun_id)
tag = res.Tag
try:
res = server.LunListInfoIterNext(Tag=tag, Maximum=1)
if hasattr(res, 'Luns') and res.Luns.LunInfo:
return res.Luns.LunInfo[0]
finally:
server.LunListInfoIterEnd(Tag=tag)
msg = _('Failed to get LUN details for LUN ID %s')
raise exception.VolumeBackendAPIException(data=msg % lun_id)
def _get_host_details(self, host_id):
"""Given the ID of a host, get the details about it.
A "host" is a storage system here.
"""
server = self.client.service
res = server.HostListInfoIterStart(ObjectNameOrId=host_id)
tag = res.Tag
try:
res = server.HostListInfoIterNext(Tag=tag, Maximum=1)
if hasattr(res, 'Hosts') and res.Hosts.HostInfo:
return res.Hosts.HostInfo[0]
finally:
server.HostListInfoIterEnd(Tag=tag)
msg = _('Failed to get host details for host ID %s')
raise exception.VolumeBackendAPIException(data=msg % host_id)
def _get_iqn_for_host(self, host_id):
"""Get the iSCSI Target Name for a storage system."""
request = self.client.factory.create('Request')
request.Name = 'iscsi-node-get-name'
response = self.client.service.ApiProxy(Target=host_id,
Request=request)
self._check_fail(request, response)
return response.Results['node-name'][0]
def _api_elem_is_empty(self, elem):
"""Return true if the API element should be considered empty.
Helper routine to figure out if a list returned from a proxy API
is empty. This is necessary because the API proxy produces nasty
looking XML.
"""
if type(elem) is not list:
return True
if 0 == len(elem):
return True
child = elem[0]
if isinstance(child, text.Text):
return True
if type(child) is str:
return True
return False
def _get_target_portal_for_host(self, host_id, host_address):
"""Get iSCSI target portal for a storage system.
Get the iSCSI Target Portal details for a particular IP address
on a storage system.
"""
request = self.client.factory.create('Request')
request.Name = 'iscsi-portal-list-info'
response = self.client.service.ApiProxy(Target=host_id,
Request=request)
self._check_fail(request, response)
portal = {}
portals = response.Results['iscsi-portal-list-entries']
if self._api_elem_is_empty(portals):
return portal
portal_infos = portals[0]['iscsi-portal-list-entry-info']
for portal_info in portal_infos:
portal['address'] = portal_info['ip-address'][0]
portal['port'] = portal_info['ip-port'][0]
portal['portal'] = portal_info['tpgroup-tag'][0]
if host_address == portal['address']:
break
return portal
def _get_export(self, volume):
"""Get the iSCSI export details for a volume.
Looks up the LUN in DFM based on the volume and project name, then get
the LUN's ID. We store that value in the database instead of the iSCSI
details because we will not have the true iSCSI details until masking
time (when initialize_connection() is called).
"""
name = volume['name']
project = volume['project_id']
lun = self._lookup_lun_for_volume(name, project)
return {'provider_location': lun.id}
def ensure_export(self, context, volume):
"""Driver entry point to get the export info for an existing volume."""
return self._get_export(volume)
def create_export(self, context, volume):
"""Driver entry point to get the export info for a new volume."""
return self._get_export(volume)
def remove_export(self, context, volume):
"""Driver exntry point to remove an export for a volume.
Since exporting is idempotent in this driver, we have nothing
to do for unexporting.
"""
pass
def _find_igroup_for_initiator(self, host_id, initiator_name):
"""Get the igroup for an initiator.
Look for an existing igroup (initiator group) on the storage system
containing a given iSCSI initiator and return the name of the igroup.
"""
request = self.client.factory.create('Request')
request.Name = 'igroup-list-info'
response = self.client.service.ApiProxy(Target=host_id,
Request=request)
self._check_fail(request, response)
igroups = response.Results['initiator-groups']
if self._api_elem_is_empty(igroups):
return None
igroup_infos = igroups[0]['initiator-group-info']
for igroup_info in igroup_infos:
if ('iscsi' != igroup_info['initiator-group-type'][0] or
'linux' != igroup_info['initiator-group-os-type'][0]):
continue
igroup_name = igroup_info['initiator-group-name'][0]
if not igroup_name.startswith(self.IGROUP_PREFIX):
continue
initiators = igroup_info['initiators'][0]['initiator-info']
for initiator in initiators:
if initiator_name == initiator['initiator-name'][0]:
return igroup_name
return None
def _create_igroup(self, host_id, initiator_name):
"""Create a new igroup.
Create a new igroup (initiator group) on the storage system to hold
the given iSCSI initiator. The group will only have 1 member and will
be named "openstack-${initiator_name}".
"""
igroup_name = self.IGROUP_PREFIX + initiator_name
request = self.client.factory.create('Request')
request.Name = 'igroup-create'
igroup_create_xml = (
'<initiator-group-name>%s</initiator-group-name>'
'<initiator-group-type>iscsi</initiator-group-type>'
'<os-type>linux</os-type><ostype>linux</ostype>')
request.Args = text.Raw(igroup_create_xml % igroup_name)
response = self.client.service.ApiProxy(Target=host_id,
Request=request)
self._check_fail(request, response)
request = self.client.factory.create('Request')
request.Name = 'igroup-add'
igroup_add_xml = (
'<initiator-group-name>%s</initiator-group-name>'
'<initiator>%s</initiator>')
request.Args = text.Raw(igroup_add_xml % (igroup_name, initiator_name))
response = self.client.service.ApiProxy(Target=host_id,
Request=request)
self._check_fail(request, response)
return igroup_name
def _get_lun_mappping(self, host_id, lunpath, igroup_name):
"""Get the mapping between a LUN and an igroup.
Check if a given LUN is already mapped to the given igroup (initiator
group). If the LUN is mapped, also return the LUN number for the
mapping.
"""
request = self.client.factory.create('Request')
request.Name = 'lun-map-list-info'
request.Args = text.Raw('<path>%s</path>' % (lunpath))
response = self.client.service.ApiProxy(Target=host_id,
Request=request)
self._check_fail(request, response)
igroups = response.Results['initiator-groups']
if self._api_elem_is_empty(igroups):
return {'mapped': False}
igroup_infos = igroups[0]['initiator-group-info']
for igroup_info in igroup_infos:
if igroup_name == igroup_info['initiator-group-name'][0]:
return {'mapped': True, 'lun_num': igroup_info['lun-id'][0]}
return {'mapped': False}
def _map_initiator(self, host_id, lunpath, igroup_name):
"""Map a LUN to an igroup.
Map the given LUN to the given igroup (initiator group). Return the LUN
number that the LUN was mapped to (the filer will choose the lowest
available number).
"""
request = self.client.factory.create('Request')
request.Name = 'lun-map'
lun_map_xml = ('<initiator-group>%s</initiator-group>'
'<path>%s</path>')
request.Args = text.Raw(lun_map_xml % (igroup_name, lunpath))
response = self.client.service.ApiProxy(Target=host_id,
Request=request)
self._check_fail(request, response)
return response.Results['lun-id-assigned'][0]
def _unmap_initiator(self, host_id, lunpath, igroup_name):
"""Unmap the given LUN from the given igroup (initiator group)."""
request = self.client.factory.create('Request')
request.Name = 'lun-unmap'
lun_unmap_xml = ('<initiator-group>%s</initiator-group>'
'<path>%s</path>')
request.Args = text.Raw(lun_unmap_xml % (igroup_name, lunpath))
response = self.client.service.ApiProxy(Target=host_id,
Request=request)
self._check_fail(request, response)
def _ensure_initiator_mapped(self, host_id, lunpath, initiator_name):
"""Ensure that a LUN is mapped to a particular initiator.
Check if a LUN is mapped to a given initiator already and create
the mapping if it is not. A new igroup will be created if needed.
Returns the LUN number for the mapping between the LUN and initiator
in both cases.
"""
lunpath = '/vol/' + lunpath
igroup_name = self._find_igroup_for_initiator(host_id, initiator_name)
if not igroup_name:
igroup_name = self._create_igroup(host_id, initiator_name)
mapping = self._get_lun_mappping(host_id, lunpath, igroup_name)
if mapping['mapped']:
return mapping['lun_num']
return self._map_initiator(host_id, lunpath, igroup_name)
def _ensure_initiator_unmapped(self, host_id, lunpath, initiator_name):
"""Ensure that a LUN is not mapped to a particular initiator.
Check if a LUN is mapped to a given initiator and remove the
mapping if it is. This does not destroy the igroup.
"""
lunpath = '/vol/' + lunpath
igroup_name = self._find_igroup_for_initiator(host_id, initiator_name)
if not igroup_name:
return
mapping = self._get_lun_mappping(host_id, lunpath, igroup_name)
if mapping['mapped']:
self._unmap_initiator(host_id, lunpath, igroup_name)
def initialize_connection(self, volume, connector):
"""Driver entry point to attach a volume to an instance.
Do the LUN masking on the storage system so the initiator can access
the LUN on the target. Also return the iSCSI properties so the
initiator can find the LUN. This implementation does not call
_get_iscsi_properties() to get the properties because cannot store the
LUN number in the database. We only find out what the LUN number will
be during this method call so we construct the properties dictionary
ourselves.
"""
initiator_name = connector['initiator']
lun_id = volume['provider_location']
if not lun_id:
msg = _("No LUN ID for volume %s") % volume['name']
raise exception.VolumeBackendAPIException(data=msg)
lun = self._get_lun_details(lun_id)
lun_num = self._ensure_initiator_mapped(lun.HostId, lun.LunPath,
initiator_name)
host = self._get_host_details(lun.HostId)
portal = self._get_target_portal_for_host(host.HostId,
host.HostAddress)
if not portal:
msg = _('Failed to get target portal for filer: %s')
raise exception.VolumeBackendAPIException(data=msg % host.HostName)
iqn = self._get_iqn_for_host(host.HostId)
if not iqn:
msg = _('Failed to get target IQN for filer: %s')
raise exception.VolumeBackendAPIException(data=msg % host.HostName)
properties = {}
properties['target_discovered'] = False
(address, port) = (portal['address'], portal['port'])
properties['target_portal'] = '%s:%s' % (address, port)
properties['target_iqn'] = iqn
properties['target_lun'] = lun_num
properties['volume_id'] = volume['id']
auth = volume['provider_auth']
if auth:
(auth_method, auth_username, auth_secret) = auth.split()
properties['auth_method'] = auth_method
properties['auth_username'] = auth_username
properties['auth_password'] = auth_secret
return {
'driver_volume_type': 'iscsi',
'data': properties,
}
def terminate_connection(self, volume, connector, **kwargs):
"""Driver entry point to unattach a volume from an instance.
Unmask the LUN on the storage system so the given intiator can no
longer access it.
"""
initiator_name = connector['initiator']
lun_id = volume['provider_location']
if not lun_id:
msg = _('No LUN ID for volume %s') % volume['name']
raise exception.VolumeBackendAPIException(data=msg)
lun = self._get_lun_details(lun_id)
self._ensure_initiator_unmapped(lun.HostId, lun.LunPath,
initiator_name)
def _is_clone_done(self, host_id, clone_op_id, volume_uuid):
"""Check the status of a clone operation.
Return True if done, False otherwise.
"""
request = self.client.factory.create('Request')
request.Name = 'clone-list-status'
clone_list_status_xml = (
'<clone-id><clone-id-info>'
'<clone-op-id>%s</clone-op-id>'
'<volume-uuid>%s</volume-uuid>'
'</clone-id-info></clone-id>')
request.Args = text.Raw(clone_list_status_xml % (clone_op_id,
volume_uuid))
response = self.client.service.ApiProxy(Target=host_id,
Request=request)
self._check_fail(request, response)
if isinstance(response.Results, text.Text):
return False
status = response.Results['status']
if self._api_elem_is_empty(status):
return False
ops_info = status[0]['ops-info'][0]
state = ops_info['clone-state'][0]
return 'completed' == state
def _clone_lun(self, host_id, src_path, dest_path, snap):
"""Create a clone of a NetApp LUN.
The clone initially consumes no space and is not space reserved.
"""
request = self.client.factory.create('Request')
request.Name = 'clone-start'
clone_start_xml = (
'<source-path>%s</source-path><no-snap>%s</no-snap>'
'<destination-path>%s</destination-path>')
if snap:
no_snap = 'false'
else:
no_snap = 'true'
request.Args = text.Raw(clone_start_xml % (src_path, no_snap,
dest_path))
response = self.client.service.ApiProxy(Target=host_id,
Request=request)
self._check_fail(request, response)
clone_id = response.Results['clone-id'][0]
clone_id_info = clone_id['clone-id-info'][0]
clone_op_id = clone_id_info['clone-op-id'][0]
volume_uuid = clone_id_info['volume-uuid'][0]
while not self._is_clone_done(host_id, clone_op_id, volume_uuid):
time.sleep(5)
def _refresh_dfm_luns(self, host_id):
"""Refresh the LUN list for one filer in DFM."""
server = self.client.service
server.DfmObjectRefresh(ObjectNameOrId=host_id, ChildType='lun_path')
while True:
time.sleep(15)
res = server.DfmMonitorTimestampList(HostNameOrId=host_id)
for timestamp in res.DfmMonitoringTimestamp:
if 'lun' != timestamp.MonitorName:
continue
if timestamp.LastMonitoringTimestamp:
return
def _destroy_lun(self, host_id, lun_path):
"""Destroy a LUN on the filer."""
request = self.client.factory.create('Request')
request.Name = 'lun-offline'
path_xml = '<path>%s</path>'
request.Args = text.Raw(path_xml % lun_path)
response = self.client.service.ApiProxy(Target=host_id,
Request=request)
self._check_fail(request, response)
request = self.client.factory.create('Request')
request.Name = 'lun-destroy'
request.Args = text.Raw(path_xml % lun_path)
response = self.client.service.ApiProxy(Target=host_id,
Request=request)
self._check_fail(request, response)
def _resize_volume(self, host_id, vol_name, new_size):
"""Resize the volume by the amount requested."""
request = self.client.factory.create('Request')
request.Name = 'volume-size'
volume_size_xml = (
'<volume>%s</volume><new-size>%s</new-size>')
request.Args = text.Raw(volume_size_xml % (vol_name, new_size))
response = self.client.service.ApiProxy(Target=host_id,
Request=request)
self._check_fail(request, response)
def _create_qtree(self, host_id, vol_name, qtree_name):
"""Create a qtree the filer."""
request = self.client.factory.create('Request')
request.Name = 'qtree-create'
qtree_create_xml = (
'<mode>0755</mode><volume>%s</volume><qtree>%s</qtree>')
request.Args = text.Raw(qtree_create_xml % (vol_name, qtree_name))
response = self.client.service.ApiProxy(Target=host_id,
Request=request)
self._check_fail(request, response)
def create_snapshot(self, snapshot):
"""Driver entry point for creating a snapshot.
This driver implements snapshots by using efficient single-file
(LUN) cloning.
"""
vol_name = snapshot['volume_name']
snapshot_name = snapshot['name']
project = snapshot['project_id']
lun = self._lookup_lun_for_volume(vol_name, project)
lun_id = lun.id
lun = self._get_lun_details(lun_id)
extra_gb = snapshot['volume_size']
new_size = '+%dg' % extra_gb
self._resize_volume(lun.HostId, lun.VolumeName, new_size)
# LunPath is the partial LUN path in this format: volume/qtree/lun
lun_path = str(lun.LunPath)
lun_name = lun_path[lun_path.rfind('/') + 1:]
qtree_path = '/vol/%s/%s' % (lun.VolumeName, lun.QtreeName)
src_path = '%s/%s' % (qtree_path, lun_name)
dest_path = '%s/%s' % (qtree_path, snapshot_name)
self._clone_lun(lun.HostId, src_path, dest_path, True)
def delete_snapshot(self, snapshot):
"""Driver entry point for deleting a snapshot."""
vol_name = snapshot['volume_name']
snapshot_name = snapshot['name']
project = snapshot['project_id']
lun = self._lookup_lun_for_volume(vol_name, project)
lun_id = lun.id
lun = self._get_lun_details(lun_id)
lun_path = '/vol/%s/%s/%s' % (lun.VolumeName, lun.QtreeName,
snapshot_name)
self._destroy_lun(lun.HostId, lun_path)
extra_gb = snapshot['volume_size']
new_size = '-%dg' % extra_gb
self._resize_volume(lun.HostId, lun.VolumeName, new_size)
def create_volume_from_snapshot(self, volume, snapshot):
"""Driver entry point for creating a new volume from a snapshot.
Many would call this "cloning" and in fact we use cloning to implement
this feature.
"""
vol_size = volume['size']
snap_size = snapshot['volume_size']
if vol_size != snap_size:
msg = _('Cannot create volume of size %(vol_size)s from '
'snapshot of size %(snap_size)s')
raise exception.VolumeBackendAPIException(data=msg % locals())
vol_name = snapshot['volume_name']
snapshot_name = snapshot['name']
project = snapshot['project_id']
lun = self._lookup_lun_for_volume(vol_name, project)
lun_id = lun.id
dataset = lun.dataset
old_type = dataset.type
new_type = self._get_ss_type(volume)
if new_type != old_type:
msg = _('Cannot create volume of type %(new_type)s from '
'snapshot of type %(old_type)s')
raise exception.VolumeBackendAPIException(data=msg % locals())
lun = self._get_lun_details(lun_id)
extra_gb = vol_size
new_size = '+%dg' % extra_gb
self._resize_volume(lun.HostId, lun.VolumeName, new_size)
clone_name = volume['name']
self._create_qtree(lun.HostId, lun.VolumeName, clone_name)
src_path = '/vol/%s/%s/%s' % (lun.VolumeName, lun.QtreeName,
snapshot_name)
dest_path = '/vol/%s/%s/%s' % (lun.VolumeName, clone_name, clone_name)
self._clone_lun(lun.HostId, src_path, dest_path, False)
self._refresh_dfm_luns(lun.HostId)
self._discover_dataset_luns(dataset, clone_name)
def create_cloned_volume(self, volume, src_vref):
"""Creates a clone of the specified volume."""
raise NotImplementedError()
class NetAppLun(object):
"""Represents a LUN on NetApp storage."""
def __init__(self, handle, name, size, metadata_dict):
self.handle = handle
self.name = name
self.size = size
self.metadata = metadata_dict
def get_metadata_property(self, prop):
"""Get the metadata property of a LUN."""
if prop in self.metadata:
return self.metadata[prop]
name = self.name
msg = _("No metadata property %(prop)s defined for the LUN %(name)s")
LOG.debug(msg % locals())
class NetAppCmodeISCSIDriver(driver.ISCSIDriver):
"""NetApp C-mode iSCSI volume driver."""
def __init__(self, *args, **kwargs):
super(NetAppCmodeISCSIDriver, self).__init__(*args, **kwargs)
self.lun_table = {}
def _create_client(self, **kwargs):
"""Instantiate a web services client.
This method creates a "suds" client to make web services calls to the
DFM server. Note that the WSDL file is quite large and may take
a few seconds to parse.
"""
wsdl_url = kwargs['wsdl_url']
LOG.debug(_('Using WSDL: %s') % wsdl_url)
if kwargs['cache']:
self.client = client.Client(wsdl_url, username=kwargs['login'],
password=kwargs['password'])
else:
self.client = client.Client(wsdl_url, username=kwargs['login'],
password=kwargs['password'],
cache=None)
def _check_flags(self):
"""Ensure that the flags we care about are set."""
required_flags = ['netapp_wsdl_url', 'netapp_login', 'netapp_password',
'netapp_server_hostname', 'netapp_server_port']
for flag in required_flags:
if not getattr(FLAGS, flag, None):
msg = _('%s is not set') % flag
raise exception.InvalidInput(data=msg)
def do_setup(self, context):
"""Setup the NetApp Volume driver.
Called one time by the manager after the driver is loaded.
Validate the flags we care about and setup the suds (web services)
client.
"""
self._check_flags()
self._create_client(
wsdl_url=FLAGS.netapp_wsdl_url,
login=FLAGS.netapp_login, password=FLAGS.netapp_password,
hostname=FLAGS.netapp_server_hostname,
port=FLAGS.netapp_server_port, cache=True)
def check_for_setup_error(self):
"""Check that the driver is working and can communicate.
Discovers the LUNs on the NetApp server.
"""
self.lun_table = {}
luns = self.client.service.ListLuns()
for lun in luns:
meta_dict = {}
if hasattr(lun, 'Metadata'):
meta_dict = self._create_dict_from_meta(lun.Metadata)
discovered_lun = NetAppLun(lun.Handle,
lun.Name,
lun.Size,
meta_dict)
self._add_lun_to_table(discovered_lun)
LOG.debug(_("Success getting LUN list from server"))
def create_volume(self, volume):
"""Driver entry point for creating a new volume."""
default_size = '104857600' # 100 MB
gigabytes = 1073741824L # 2^30
name = volume['name']
if int(volume['size']) == 0:
size = default_size
else:
size = str(int(volume['size']) * gigabytes)
extra_args = {}
extra_args['OsType'] = 'linux'
extra_args['QosType'] = self._get_qos_type(volume)
extra_args['Container'] = volume['project_id']
extra_args['Display'] = volume['display_name']
extra_args['Description'] = volume['display_description']
extra_args['SpaceReserved'] = True
server = self.client.service
metadata = self._create_metadata_list(extra_args)
lun = server.ProvisionLun(Name=name, Size=size,
Metadata=metadata)
LOG.debug(_("Created LUN with name %s") % name)
self._add_lun_to_table(
NetAppLun(lun.Handle,
lun.Name,
lun.Size,
self._create_dict_from_meta(lun.Metadata)))
def delete_volume(self, volume):
"""Driver entry point for destroying existing volumes."""
name = volume['name']
handle = self._get_lun_handle(name)
self.client.service.DestroyLun(Handle=handle)
LOG.debug(_("Destroyed LUN %s") % handle)
self.lun_table.pop(name)
def ensure_export(self, context, volume):
"""Driver entry point to get the export info for an existing volume."""
handle = self._get_lun_handle(volume['name'])
return {'provider_location': handle}
def create_export(self, context, volume):
"""Driver entry point to get the export info for a new volume."""
handle = self._get_lun_handle(volume['name'])
return {'provider_location': handle}
def remove_export(self, context, volume):
"""Driver exntry point to remove an export for a volume.
Since exporting is idempotent in this driver, we have nothing
to do for unexporting.
"""
pass
def initialize_connection(self, volume, connector):
"""Driver entry point to attach a volume to an instance.
Do the LUN masking on the storage system so the initiator can access
the LUN on the target. Also return the iSCSI properties so the
initiator can find the LUN. This implementation does not call
_get_iscsi_properties() to get the properties because cannot store the
LUN number in the database. We only find out what the LUN number will
be during this method call so we construct the properties dictionary
ourselves.
"""
initiator_name = connector['initiator']
handle = volume['provider_location']
server = self.client.service
server.MapLun(Handle=handle, InitiatorType="iscsi",
InitiatorName=initiator_name)
msg = _("Mapped LUN %(handle)s to the initiator %(initiator_name)s")
LOG.debug(msg % locals())
target_details_list = server.GetLunTargetDetails(
Handle=handle,
InitiatorType="iscsi",
InitiatorName=initiator_name)
msg = _("Succesfully fetched target details for LUN %(handle)s and "
"initiator %(initiator_name)s")
LOG.debug(msg % locals())
if not target_details_list:
msg = _('Failed to get LUN target details for the LUN %s')
raise exception.VolumeBackendAPIException(data=msg % handle)
target_details = target_details_list[0]
if not target_details.Address and target_details.Port:
msg = _('Failed to get target portal for the LUN %s')
raise exception.VolumeBackendAPIException(data=msg % handle)
iqn = target_details.Iqn
if not iqn:
msg = _('Failed to get target IQN for the LUN %s')
raise exception.VolumeBackendAPIException(data=msg % handle)
properties = {}
properties['target_discovered'] = False
(address, port) = (target_details.Address, target_details.Port)
properties['target_portal'] = '%s:%s' % (address, port)
properties['target_iqn'] = iqn
properties['target_lun'] = target_details.LunNumber
properties['volume_id'] = volume['id']
auth = volume['provider_auth']
if auth:
(auth_method, auth_username, auth_secret) = auth.split()
properties['auth_method'] = auth_method
properties['auth_username'] = auth_username
properties['auth_password'] = auth_secret
return {
'driver_volume_type': 'iscsi',
'data': properties,
}
def terminate_connection(self, volume, connector, **kwargs):
"""Driver entry point to unattach a volume from an instance.
Unmask the LUN on the storage system so the given intiator can no
longer access it.
"""
initiator_name = connector['initiator']
handle = volume['provider_location']
self.client.service.UnmapLun(Handle=handle, InitiatorType="iscsi",
InitiatorName=initiator_name)
msg = _("Unmapped LUN %(handle)s from the initiator "
"%(initiator_name)s")
LOG.debug(msg % locals())
def create_snapshot(self, snapshot):
"""Driver entry point for creating a snapshot.
This driver implements snapshots by using efficient single-file
(LUN) cloning.
"""
vol_name = snapshot['volume_name']
snapshot_name = snapshot['name']
lun = self.lun_table[vol_name]
extra_args = {'SpaceReserved': False}
self._clone_lun(lun.handle, snapshot_name, extra_args)
def delete_snapshot(self, snapshot):
"""Driver entry point for deleting a snapshot."""
handle = self._get_lun_handle(snapshot['name'])
self.client.service.DestroyLun(Handle=handle)
LOG.debug(_("Destroyed LUN %s") % handle)
def create_volume_from_snapshot(self, volume, snapshot):
"""Driver entry point for creating a new volume from a snapshot.
Many would call this "cloning" and in fact we use cloning to implement
this feature.
"""
snapshot_name = snapshot['name']
lun = self.lun_table[snapshot_name]
new_name = volume['name']
extra_args = {}
extra_args['OsType'] = 'linux'
extra_args['QosType'] = self._get_qos_type(volume)
extra_args['Container'] = volume['project_id']
extra_args['Display'] = volume['display_name']
extra_args['Description'] = volume['display_description']
extra_args['SpaceReserved'] = True
self._clone_lun(lun.handle, new_name, extra_args)
def _get_qos_type(self, volume):
"""Get the storage service type for a volume."""
type_id = volume['volume_type_id']
if not type_id:
return None
volume_type = volume_types.get_volume_type(None, type_id)
if not volume_type:
return None
return volume_type['name']
def _add_lun_to_table(self, lun):
"""Adds LUN to cache table."""
if not isinstance(lun, NetAppLun):
msg = _("Object is not a NetApp LUN.")
raise exception.VolumeBackendAPIException(data=msg)
self.lun_table[lun.name] = lun
def _clone_lun(self, handle, new_name, extra_args):
"""Clone LUN with the given handle to the new name."""
server = self.client.service
metadata = self._create_metadata_list(extra_args)
lun = server.CloneLun(Handle=handle, NewName=new_name,
Metadata=metadata)
LOG.debug(_("Cloned LUN with new name %s") % new_name)
self._add_lun_to_table(
NetAppLun(lun.Handle,
lun.Name,
lun.Size,
self._create_dict_from_meta(lun.Metadata)))
def _create_metadata_list(self, extra_args):
"""Creates metadata from kwargs."""
metadata = []
for key in extra_args.keys():
meta = self.client.factory.create("Metadata")
meta.Key = key
meta.Value = extra_args[key]
metadata.append(meta)
return metadata
def _get_lun_handle(self, name):
"""Get the details for a LUN from our cache table."""
if name not in self.lun_table:
LOG.warn(_("Could not find handle for LUN named %s") % name)
return None
return self.lun_table[name].handle
def _create_dict_from_meta(self, metadata):
"""Creates dictionary from metadata array."""
meta_dict = {}
if not metadata:
return meta_dict
for meta in metadata:
meta_dict[meta.Key] = meta.Value
return meta_dict
def copy_image_to_volume(self, context, volume, image_service, image_id):
"""Fetch the image from image_service and write it to the volume."""
raise NotImplementedError()
def copy_volume_to_image(self, context, volume, image_service, image_meta):
"""Copy the volume to the specified image."""
raise NotImplementedError()
def create_cloned_volume(self, volume, src_vref):
"""Creates a clone of the specified volume."""
raise NotImplementedError()