ae9c41727a
This patch allow to use nfs as cinder driver. To use it, in the localrc we can now set CINDER_DRIVER=nfs CINDER_NFS_SERVERPATH=172.16.0.50:/export_cinder The nfs-server is not setup by devstack. Change-Id: I8e240d00b58f272d04ab2c0922c551b1f7266260
557 lines
22 KiB
Plaintext
557 lines
22 KiB
Plaintext
# lib/cinder
|
|
# Install and start **Cinder** volume service
|
|
|
|
# Dependencies:
|
|
# - functions
|
|
# - DEST, DATA_DIR, STACK_USER must be defined
|
|
# SERVICE_{TENANT_NAME|PASSWORD} must be defined
|
|
# ``KEYSTONE_TOKEN_FORMAT`` must be defined
|
|
|
|
# stack.sh
|
|
# ---------
|
|
# install_cinder
|
|
# configure_cinder
|
|
# init_cinder
|
|
# start_cinder
|
|
# stop_cinder
|
|
# cleanup_cinder
|
|
|
|
# Save trace setting
|
|
XTRACE=$(set +o | grep xtrace)
|
|
set +o xtrace
|
|
|
|
|
|
# Defaults
|
|
# --------
|
|
|
|
# set up default driver
|
|
CINDER_DRIVER=${CINDER_DRIVER:-default}
|
|
|
|
# set up default directories
|
|
CINDER_DIR=$DEST/cinder
|
|
CINDERCLIENT_DIR=$DEST/python-cinderclient
|
|
CINDER_STATE_PATH=${CINDER_STATE_PATH:=$DATA_DIR/cinder}
|
|
CINDER_AUTH_CACHE_DIR=${CINDER_AUTH_CACHE_DIR:-/var/cache/cinder}
|
|
|
|
CINDER_CONF_DIR=/etc/cinder
|
|
CINDER_CONF=$CINDER_CONF_DIR/cinder.conf
|
|
CINDER_API_PASTE_INI=$CINDER_CONF_DIR/api-paste.ini
|
|
|
|
# Public facing bits
|
|
CINDER_SERVICE_HOST=${CINDER_SERVICE_HOST:-$SERVICE_HOST}
|
|
CINDER_SERVICE_PORT=${CINDER_SERVICE_PORT:-8776}
|
|
CINDER_SERVICE_PORT_INT=${CINDER_SERVICE_PORT_INT:-18776}
|
|
CINDER_SERVICE_PROTOCOL=${CINDER_SERVICE_PROTOCOL:-$SERVICE_PROTOCOL}
|
|
|
|
# Support entry points installation of console scripts
|
|
if [[ -d $CINDER_DIR/bin ]]; then
|
|
CINDER_BIN_DIR=$CINDER_DIR/bin
|
|
else
|
|
CINDER_BIN_DIR=$(get_python_exec_prefix)
|
|
fi
|
|
|
|
# Support for multi lvm backend configuration (default is no support)
|
|
CINDER_MULTI_LVM_BACKEND=$(trueorfalse False $CINDER_MULTI_LVM_BACKEND)
|
|
|
|
# Should cinder perform secure deletion of volumes?
|
|
# Defaults to true, can be set to False to avoid this bug when testing:
|
|
# https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1023755
|
|
CINDER_SECURE_DELETE=`trueorfalse True $CINDER_SECURE_DELETE`
|
|
|
|
# Cinder reports allocations back to the scheduler on periodic intervals
|
|
# it turns out we can get an "out of space" issue when we run tests too
|
|
# quickly just because cinder didn't realize we'd freed up resources.
|
|
# Make this configurable so that devstack-gate/tempest can set it to
|
|
# less than the 60 second default
|
|
# https://bugs.launchpad.net/cinder/+bug/1180976
|
|
CINDER_PERIODIC_INTERVAL=${CINDER_PERIODIC_INTERVAL:-60}
|
|
|
|
# Name of the lvm volume groups to use/create for iscsi volumes
|
|
VOLUME_GROUP=${VOLUME_GROUP:-stack-volumes}
|
|
VOLUME_BACKING_FILE=${VOLUME_BACKING_FILE:-$DATA_DIR/${VOLUME_GROUP}-backing-file}
|
|
VOLUME_BACKING_DEVICE=${VOLUME_BACKING_DEVICE:-}
|
|
|
|
# VOLUME_GROUP2 is used only if CINDER_MULTI_LVM_BACKEND = True
|
|
VOLUME_GROUP2=${VOLUME_GROUP2:-stack-volumes2}
|
|
VOLUME_BACKING_FILE2=${VOLUME_BACKING_FILE2:-$DATA_DIR/${VOLUME_GROUP2}-backing-file}
|
|
VOLUME_BACKING_DEVICE2=${VOLUME_BACKING_DEVICE2:-}
|
|
|
|
VOLUME_NAME_PREFIX=${VOLUME_NAME_PREFIX:-volume-}
|
|
|
|
|
|
# Functions
|
|
# ---------
|
|
# _clean_lvm_lv removes all cinder LVM volumes
|
|
# _clean_lvm_lv $VOLUME_GROUP $VOLUME_NAME_PREFIX
|
|
function _clean_lvm_lv() {
|
|
local vg=$1
|
|
local lv_prefix=$2
|
|
|
|
# Clean out existing volumes
|
|
for lv in `sudo lvs --noheadings -o lv_name $vg`; do
|
|
# lv_prefix prefixes the LVs we want
|
|
if [[ "${lv#$lv_prefix}" != "$lv" ]]; then
|
|
sudo lvremove -f $vg/$lv
|
|
fi
|
|
done
|
|
}
|
|
|
|
# _clean_lvm_backing_file() removes the backing file of the
|
|
# volume group used by cinder
|
|
# _clean_lvm_backing_file() $VOLUME_GROUP
|
|
function _clean_lvm_backing_file() {
|
|
local vg=$1
|
|
|
|
# if there is no logical volume left, it's safe to attempt a cleanup
|
|
# of the backing file
|
|
if [ -z "`sudo lvs --noheadings -o lv_name $vg`" ]; then
|
|
# if the backing physical device is a loop device, it was probably setup by devstack
|
|
VG_DEV=$(sudo losetup -j $DATA_DIR/${vg}-backing-file | awk -F':' '/backing-file/ { print $1}')
|
|
if [[ -n "$VG_DEV" ]]; then
|
|
sudo losetup -d $VG_DEV
|
|
rm -f $DATA_DIR/${vg}-backing-file
|
|
fi
|
|
fi
|
|
}
|
|
|
|
# cleanup_cinder() - Remove residual data files, anything left over from previous
|
|
# runs that a clean run would need to clean up
|
|
function cleanup_cinder() {
|
|
# ensure the volume group is cleared up because fails might
|
|
# leave dead volumes in the group
|
|
TARGETS=$(sudo tgtadm --op show --mode target)
|
|
if [ $? -ne 0 ]; then
|
|
# If tgt driver isn't running this won't work obviously
|
|
# So check the response and restart if need be
|
|
echo "tgtd seems to be in a bad state, restarting..."
|
|
if is_ubuntu; then
|
|
restart_service tgt
|
|
else
|
|
restart_service tgtd
|
|
fi
|
|
TARGETS=$(sudo tgtadm --op show --mode target)
|
|
fi
|
|
|
|
if [[ -n "$TARGETS" ]]; then
|
|
iqn_list=( $(grep --no-filename -r iqn $SCSI_PERSIST_DIR | sed 's/<target //' | sed 's/>//') )
|
|
for i in "${iqn_list[@]}"; do
|
|
echo removing iSCSI target: $i
|
|
sudo tgt-admin --delete $i
|
|
done
|
|
fi
|
|
|
|
if is_service_enabled cinder; then
|
|
sudo rm -rf $CINDER_STATE_PATH/volumes/*
|
|
fi
|
|
|
|
if is_ubuntu; then
|
|
stop_service tgt
|
|
else
|
|
stop_service tgtd
|
|
fi
|
|
|
|
# Campsite rule: leave behind a volume group at least as clean as we found it
|
|
_clean_lvm_lv $VOLUME_GROUP $VOLUME_NAME_PREFIX
|
|
_clean_lvm_backing_file $VOLUME_GROUP
|
|
|
|
if [ "$CINDER_MULTI_LVM_BACKEND" = "True" ]; then
|
|
_clean_lvm_lv $VOLUME_GROUP2 $VOLUME_NAME_PREFIX
|
|
_clean_lvm_backing_file $VOLUME_GROUP2
|
|
fi
|
|
}
|
|
|
|
# configure_cinder() - Set config files, create data dirs, etc
|
|
function configure_cinder() {
|
|
if [[ ! -d $CINDER_CONF_DIR ]]; then
|
|
sudo mkdir -p $CINDER_CONF_DIR
|
|
fi
|
|
sudo chown $STACK_USER $CINDER_CONF_DIR
|
|
|
|
cp -p $CINDER_DIR/etc/cinder/policy.json $CINDER_CONF_DIR
|
|
|
|
# Set the paths of certain binaries
|
|
CINDER_ROOTWRAP=$(get_rootwrap_location cinder)
|
|
|
|
# If Cinder ships the new rootwrap filters files, deploy them
|
|
# (owned by root) and add a parameter to $CINDER_ROOTWRAP
|
|
ROOTWRAP_CINDER_SUDOER_CMD="$CINDER_ROOTWRAP"
|
|
if [[ -d $CINDER_DIR/etc/cinder/rootwrap.d ]]; then
|
|
# Wipe any existing rootwrap.d files first
|
|
if [[ -d $CINDER_CONF_DIR/rootwrap.d ]]; then
|
|
sudo rm -rf $CINDER_CONF_DIR/rootwrap.d
|
|
fi
|
|
# Deploy filters to /etc/cinder/rootwrap.d
|
|
sudo mkdir -m 755 $CINDER_CONF_DIR/rootwrap.d
|
|
sudo cp $CINDER_DIR/etc/cinder/rootwrap.d/*.filters $CINDER_CONF_DIR/rootwrap.d
|
|
sudo chown -R root:root $CINDER_CONF_DIR/rootwrap.d
|
|
sudo chmod 644 $CINDER_CONF_DIR/rootwrap.d/*
|
|
# Set up rootwrap.conf, pointing to /etc/cinder/rootwrap.d
|
|
sudo cp $CINDER_DIR/etc/cinder/rootwrap.conf $CINDER_CONF_DIR/
|
|
sudo sed -e "s:^filters_path=.*$:filters_path=$CINDER_CONF_DIR/rootwrap.d:" -i $CINDER_CONF_DIR/rootwrap.conf
|
|
sudo chown root:root $CINDER_CONF_DIR/rootwrap.conf
|
|
sudo chmod 0644 $CINDER_CONF_DIR/rootwrap.conf
|
|
# Specify rootwrap.conf as first parameter to cinder-rootwrap
|
|
CINDER_ROOTWRAP="$CINDER_ROOTWRAP $CINDER_CONF_DIR/rootwrap.conf"
|
|
ROOTWRAP_CINDER_SUDOER_CMD="$CINDER_ROOTWRAP *"
|
|
fi
|
|
|
|
TEMPFILE=`mktemp`
|
|
echo "$USER ALL=(root) NOPASSWD: $ROOTWRAP_CINDER_SUDOER_CMD" >$TEMPFILE
|
|
chmod 0440 $TEMPFILE
|
|
sudo chown root:root $TEMPFILE
|
|
sudo mv $TEMPFILE /etc/sudoers.d/cinder-rootwrap
|
|
|
|
cp $CINDER_DIR/etc/cinder/api-paste.ini $CINDER_API_PASTE_INI
|
|
|
|
inicomment $CINDER_API_PASTE_INI filter:authtoken auth_host
|
|
inicomment $CINDER_API_PASTE_INI filter:authtoken auth_port
|
|
inicomment $CINDER_API_PASTE_INI filter:authtoken auth_protocol
|
|
inicomment $CINDER_API_PASTE_INI filter:authtoken admin_tenant_name
|
|
inicomment $CINDER_API_PASTE_INI filter:authtoken admin_user
|
|
inicomment $CINDER_API_PASTE_INI filter:authtoken admin_password
|
|
inicomment $CINDER_API_PASTE_INI filter:authtoken signing_dir
|
|
|
|
cp $CINDER_DIR/etc/cinder/cinder.conf.sample $CINDER_CONF
|
|
|
|
iniset $CINDER_CONF keystone_authtoken auth_host $KEYSTONE_AUTH_HOST
|
|
iniset $CINDER_CONF keystone_authtoken auth_port $KEYSTONE_AUTH_PORT
|
|
iniset $CINDER_CONF keystone_authtoken auth_protocol $KEYSTONE_AUTH_PROTOCOL
|
|
iniset $CINDER_CONF keystone_authtoken admin_tenant_name $SERVICE_TENANT_NAME
|
|
iniset $CINDER_CONF keystone_authtoken admin_user cinder
|
|
iniset $CINDER_CONF keystone_authtoken admin_password $SERVICE_PASSWORD
|
|
iniset $CINDER_CONF keystone_authtoken signing_dir $CINDER_AUTH_CACHE_DIR
|
|
|
|
iniset $CINDER_CONF DEFAULT auth_strategy keystone
|
|
iniset $CINDER_CONF DEFAULT debug $ENABLE_DEBUG_LOG_LEVEL
|
|
iniset $CINDER_CONF DEFAULT verbose True
|
|
if [ "$CINDER_MULTI_LVM_BACKEND" = "True" ]; then
|
|
iniset $CINDER_CONF DEFAULT enabled_backends lvmdriver-1,lvmdriver-2
|
|
iniset $CINDER_CONF lvmdriver-1 volume_group $VOLUME_GROUP
|
|
iniset $CINDER_CONF lvmdriver-1 volume_driver cinder.volume.drivers.lvm.LVMISCSIDriver
|
|
iniset $CINDER_CONF lvmdriver-1 volume_backend_name LVM_iSCSI
|
|
iniset $CINDER_CONF lvmdriver-2 volume_group $VOLUME_GROUP2
|
|
iniset $CINDER_CONF lvmdriver-2 volume_driver cinder.volume.drivers.lvm.LVMISCSIDriver
|
|
iniset $CINDER_CONF lvmdriver-2 volume_backend_name LVM_iSCSI_2
|
|
else
|
|
iniset $CINDER_CONF DEFAULT volume_group $VOLUME_GROUP
|
|
iniset $CINDER_CONF DEFAULT volume_name_template ${VOLUME_NAME_PREFIX}%s
|
|
fi
|
|
iniset $CINDER_CONF DEFAULT my_ip "$CINDER_SERVICE_HOST"
|
|
iniset $CINDER_CONF DEFAULT iscsi_helper tgtadm
|
|
iniset $CINDER_CONF DEFAULT sql_connection `database_connection_url cinder`
|
|
iniset $CINDER_CONF DEFAULT api_paste_config $CINDER_API_PASTE_INI
|
|
iniset $CINDER_CONF DEFAULT rootwrap_config "$CINDER_CONF_DIR/rootwrap.conf"
|
|
iniset $CINDER_CONF DEFAULT osapi_volume_extension cinder.api.contrib.standard_extensions
|
|
iniset $CINDER_CONF DEFAULT state_path $CINDER_STATE_PATH
|
|
iniset $CINDER_CONF DEFAULT lock_path $CINDER_STATE_PATH
|
|
iniset $CINDER_CONF DEFAULT periodic_interval $CINDER_PERIODIC_INTERVAL
|
|
|
|
if is_service_enabled ceilometer; then
|
|
iniset $CINDER_CONF DEFAULT notification_driver "cinder.openstack.common.notifier.rpc_notifier"
|
|
fi
|
|
|
|
if is_service_enabled tls-proxy; then
|
|
# Set the service port for a proxy to take the original
|
|
iniset $CINDER_CONF DEFAULT osapi_volume_listen_port $CINDER_SERVICE_PORT_INT
|
|
fi
|
|
|
|
if [ "$SYSLOG" != "False" ]; then
|
|
iniset $CINDER_CONF DEFAULT use_syslog True
|
|
fi
|
|
|
|
iniset_rpc_backend cinder $CINDER_CONF DEFAULT
|
|
|
|
if [[ "$CINDER_SECURE_DELETE" == "False" ]]; then
|
|
iniset $CINDER_CONF DEFAULT secure_delete False
|
|
iniset $CINDER_CONF DEFAULT volume_clear none
|
|
fi
|
|
|
|
# Format logging
|
|
if [ "$LOG_COLOR" == "True" ] && [ "$SYSLOG" == "False" ]; then
|
|
setup_colorized_logging $CINDER_CONF DEFAULT "project_id" "user_id"
|
|
fi
|
|
|
|
if [ "$CINDER_DRIVER" == "XenAPINFS" ]; then
|
|
(
|
|
set -u
|
|
iniset $CINDER_CONF DEFAULT volume_driver "cinder.volume.drivers.xenapi.sm.XenAPINFSDriver"
|
|
iniset $CINDER_CONF DEFAULT xenapi_connection_url "$CINDER_XENAPI_CONNECTION_URL"
|
|
iniset $CINDER_CONF DEFAULT xenapi_connection_username "$CINDER_XENAPI_CONNECTION_USERNAME"
|
|
iniset $CINDER_CONF DEFAULT xenapi_connection_password "$CINDER_XENAPI_CONNECTION_PASSWORD"
|
|
iniset $CINDER_CONF DEFAULT xenapi_nfs_server "$CINDER_XENAPI_NFS_SERVER"
|
|
iniset $CINDER_CONF DEFAULT xenapi_nfs_serverpath "$CINDER_XENAPI_NFS_SERVERPATH"
|
|
)
|
|
elif [ "$CINDER_DRIVER" == "nfs" ]; then
|
|
iniset $CINDER_CONF DEFAULT volume_driver "cinder.volume.drivers.nfs.NfsDriver"
|
|
iniset $CINDER_CONF DEFAULT nfs_shares_config "$CINDER_CONF_DIR/nfs_shares.conf"
|
|
echo "$CINDER_NFS_SERVERPATH" | sudo tee "$CINDER_CONF_DIR/nfs_shares.conf"
|
|
sudo chmod 666 $CINDER_CONF_DIR/nfs_shares.conf
|
|
elif [ "$CINDER_DRIVER" == "sheepdog" ]; then
|
|
iniset $CINDER_CONF DEFAULT volume_driver "cinder.volume.drivers.sheepdog.SheepdogDriver"
|
|
elif [ "$CINDER_DRIVER" == "glusterfs" ]; then
|
|
# To use glusterfs, set the following in localrc:
|
|
# CINDER_DRIVER=glusterfs
|
|
# CINDER_GLUSTERFS_SHARES="127.0.0.1:/vol1;127.0.0.1:/vol2"
|
|
# Shares are <host>:<volume> and separated by semicolons.
|
|
|
|
iniset $CINDER_CONF DEFAULT volume_driver "cinder.volume.drivers.glusterfs.GlusterfsDriver"
|
|
iniset $CINDER_CONF DEFAULT glusterfs_shares_config "$CINDER_CONF_DIR/glusterfs_shares"
|
|
touch $CINDER_CONF_DIR/glusterfs_shares
|
|
if [ ! -z "$CINDER_GLUSTERFS_SHARES" ]; then
|
|
CINDER_GLUSTERFS_SHARES=$(echo $CINDER_GLUSTERFS_SHARES | tr ";" "\n")
|
|
echo "$CINDER_GLUSTERFS_SHARES" > $CINDER_CONF_DIR/glusterfs_shares
|
|
fi
|
|
elif [ "$CINDER_DRIVER" == "vsphere" ]; then
|
|
echo_summary "Using VMware vCenter driver"
|
|
iniset $CINDER_CONF DEFAULT vmware_host_ip "$VMWAREAPI_IP"
|
|
iniset $CINDER_CONF DEFAULT vmware_host_username "$VMWAREAPI_USER"
|
|
iniset $CINDER_CONF DEFAULT vmware_host_password "$VMWAREAPI_PASSWORD"
|
|
iniset $CINDER_CONF DEFAULT volume_driver "cinder.volume.drivers.vmware.vmdk.VMwareVcVmdkDriver"
|
|
fi
|
|
|
|
if [[ is_fedora && $DISTRO =~ (rhel6) ]]; then
|
|
# Cinder clones are slightly larger due to some extra
|
|
# metadata. RHEL6 will not allow auto-extending of LV's
|
|
# without this, leading to clones giving hard-to-track disk
|
|
# I/O errors.
|
|
# see https://bugzilla.redhat.com/show_bug.cgi?id=975052
|
|
sudo sed -i~ \
|
|
-e 's/snapshot_autoextend_threshold =.*/snapshot_autoextend_threshold = 80/' \
|
|
-e 's/snapshot_autoextend_percent =.*/snapshot_autoextend_percent = 20/' \
|
|
/etc/lvm/lvm.conf
|
|
fi
|
|
iniset $CINDER_CONF keystone_authtoken auth_uri $KEYSTONE_SERVICE_PROTOCOL://$KEYSTONE_SERVICE_HOST:$KEYSTONE_SERVICE_PORT
|
|
iniset $CINDER_CONF keystone_authtoken admin_user cinder
|
|
iniset $CINDER_CONF keystone_authtoken admin_tenant_name $SERVICE_TENANT_NAME
|
|
iniset $CINDER_CONF keystone_authtoken admin_password $SERVICE_PASSWORD
|
|
|
|
}
|
|
|
|
# create_cinder_accounts() - Set up common required cinder accounts
|
|
|
|
# Tenant User Roles
|
|
# ------------------------------------------------------------------
|
|
# service cinder admin # if enabled
|
|
|
|
# Migrated from keystone_data.sh
|
|
create_cinder_accounts() {
|
|
|
|
SERVICE_TENANT=$(keystone tenant-list | awk "/ $SERVICE_TENANT_NAME / { print \$2 }")
|
|
ADMIN_ROLE=$(keystone role-list | awk "/ admin / { print \$2 }")
|
|
|
|
# Cinder
|
|
if [[ "$ENABLED_SERVICES" =~ "c-api" ]]; then
|
|
CINDER_USER=$(keystone user-create \
|
|
--name=cinder \
|
|
--pass="$SERVICE_PASSWORD" \
|
|
--tenant_id $SERVICE_TENANT \
|
|
--email=cinder@example.com \
|
|
| grep " id " | get_field 2)
|
|
keystone user-role-add \
|
|
--tenant-id $SERVICE_TENANT \
|
|
--user-id $CINDER_USER \
|
|
--role-id $ADMIN_ROLE
|
|
if [[ "$KEYSTONE_CATALOG_BACKEND" = 'sql' ]]; then
|
|
CINDER_SERVICE=$(keystone service-create \
|
|
--name=cinder \
|
|
--type=volume \
|
|
--description="Cinder Volume Service" \
|
|
| grep " id " | get_field 2)
|
|
keystone endpoint-create \
|
|
--region RegionOne \
|
|
--service_id $CINDER_SERVICE \
|
|
--publicurl "$CINDER_SERVICE_PROTOCOL://$CINDER_SERVICE_HOST:$CINDER_SERVICE_PORT/v1/\$(tenant_id)s" \
|
|
--adminurl "$CINDER_SERVICE_PROTOCOL://$CINDER_SERVICE_HOST:$CINDER_SERVICE_PORT/v1/\$(tenant_id)s" \
|
|
--internalurl "$CINDER_SERVICE_PROTOCOL://$CINDER_SERVICE_HOST:$CINDER_SERVICE_PORT/v1/\$(tenant_id)s"
|
|
CINDER_V2_SERVICE=$(keystone service-create \
|
|
--name=cinder \
|
|
--type=volumev2 \
|
|
--description="Cinder Volume Service V2" \
|
|
| grep " id " | get_field 2)
|
|
keystone endpoint-create \
|
|
--region RegionOne \
|
|
--service_id $CINDER_V2_SERVICE \
|
|
--publicurl "$CINDER_SERVICE_PROTOCOL://$CINDER_SERVICE_HOST:$CINDER_SERVICE_PORT/v2/\$(tenant_id)s" \
|
|
--adminurl "$CINDER_SERVICE_PROTOCOL://$CINDER_SERVICE_HOST:$CINDER_SERVICE_PORT/v2/\$(tenant_id)s" \
|
|
--internalurl "$CINDER_SERVICE_PROTOCOL://$CINDER_SERVICE_HOST:$CINDER_SERVICE_PORT/v2/\$(tenant_id)s"
|
|
|
|
fi
|
|
fi
|
|
}
|
|
|
|
# create_cinder_cache_dir() - Part of the init_cinder() process
|
|
function create_cinder_cache_dir() {
|
|
# Create cache dir
|
|
sudo mkdir -p $CINDER_AUTH_CACHE_DIR
|
|
sudo chown $STACK_USER $CINDER_AUTH_CACHE_DIR
|
|
rm -f $CINDER_AUTH_CACHE_DIR/*
|
|
}
|
|
|
|
create_cinder_volume_group() {
|
|
# According to the ``CINDER_MULTI_LVM_BACKEND`` value, configure one or two default volumes
|
|
# group called ``stack-volumes`` (and ``stack-volumes2``) for the volume
|
|
# service if it (they) does (do) not yet exist. If you don't wish to use a
|
|
# file backed volume group, create your own volume group called ``stack-volumes``
|
|
# and ``stack-volumes2`` before invoking ``stack.sh``.
|
|
#
|
|
# The two backing files are ``VOLUME_BACKING_FILE_SIZE`` in size, and they are stored in
|
|
# the ``DATA_DIR``.
|
|
|
|
if ! sudo vgs $VOLUME_GROUP; then
|
|
if [ -z "$VOLUME_BACKING_DEVICE" ]; then
|
|
# Only create if the file doesn't already exists
|
|
[[ -f $VOLUME_BACKING_FILE ]] || truncate -s $VOLUME_BACKING_FILE_SIZE $VOLUME_BACKING_FILE
|
|
DEV=`sudo losetup -f --show $VOLUME_BACKING_FILE`
|
|
|
|
# Only create if the loopback device doesn't contain $VOLUME_GROUP
|
|
if ! sudo vgs $VOLUME_GROUP; then
|
|
sudo vgcreate $VOLUME_GROUP $DEV
|
|
fi
|
|
else
|
|
sudo vgcreate $VOLUME_GROUP $VOLUME_BACKING_DEVICE
|
|
fi
|
|
fi
|
|
if [ "$CINDER_MULTI_LVM_BACKEND" = "True" ]; then
|
|
#set up the second volume if CINDER_MULTI_LVM_BACKEND is enabled
|
|
|
|
if ! sudo vgs $VOLUME_GROUP2; then
|
|
if [ -z "$VOLUME_BACKING_DEVICE2" ]; then
|
|
# Only create if the file doesn't already exists
|
|
[[ -f $VOLUME_BACKING_FILE2 ]] || truncate -s $VOLUME_BACKING_FILE_SIZE $VOLUME_BACKING_FILE2
|
|
|
|
DEV=`sudo losetup -f --show $VOLUME_BACKING_FILE2`
|
|
|
|
# Only create if the loopback device doesn't contain $VOLUME_GROUP
|
|
if ! sudo vgs $VOLUME_GROUP2; then
|
|
sudo vgcreate $VOLUME_GROUP2 $DEV
|
|
fi
|
|
else
|
|
sudo vgcreate $VOLUME_GROUP2 $VOLUME_BACKING_DEVICE2
|
|
fi
|
|
fi
|
|
fi
|
|
|
|
mkdir -p $CINDER_STATE_PATH/volumes
|
|
}
|
|
|
|
# init_cinder() - Initialize database and volume group
|
|
function init_cinder() {
|
|
# Force nova volumes off
|
|
NOVA_ENABLED_APIS=$(echo $NOVA_ENABLED_APIS | sed "s/osapi_volume,//")
|
|
|
|
if is_service_enabled $DATABASE_BACKENDS; then
|
|
# (Re)create cinder database
|
|
recreate_database cinder utf8
|
|
|
|
# Migrate cinder database
|
|
$CINDER_BIN_DIR/cinder-manage db sync
|
|
fi
|
|
|
|
if is_service_enabled c-vol; then
|
|
|
|
create_cinder_volume_group
|
|
|
|
if sudo vgs $VOLUME_GROUP; then
|
|
if is_fedora || is_suse; then
|
|
# service is not started by default
|
|
start_service tgtd
|
|
fi
|
|
|
|
# Remove iscsi targets
|
|
sudo tgtadm --op show --mode target | grep $VOLUME_NAME_PREFIX | grep Target | cut -f3 -d ' ' | sudo xargs -n1 tgt-admin --delete || true
|
|
# Start with a clean volume group
|
|
_clean_lvm_lv $VOLUME_GROUP $VOLUME_NAME_PREFIX
|
|
if [ "$CINDER_MULTI_LVM_BACKEND" = "True" ]; then
|
|
_clean_lvm_lv $VOLUME_GROUP2 $VOLUME_NAME_PREFIX
|
|
fi
|
|
fi
|
|
fi
|
|
|
|
create_cinder_cache_dir
|
|
}
|
|
|
|
# install_cinder() - Collect source and prepare
|
|
function install_cinder() {
|
|
git_clone $CINDER_REPO $CINDER_DIR $CINDER_BRANCH
|
|
setup_develop $CINDER_DIR
|
|
}
|
|
|
|
# install_cinderclient() - Collect source and prepare
|
|
function install_cinderclient() {
|
|
git_clone $CINDERCLIENT_REPO $CINDERCLIENT_DIR $CINDERCLIENT_BRANCH
|
|
setup_develop $CINDERCLIENT_DIR
|
|
sudo install -D -m 0644 -o $STACK_USER {$CINDERCLIENT_DIR/tools/,/etc/bash_completion.d/}cinder.bash_completion
|
|
}
|
|
|
|
# apply config.d approach for cinder volumes directory
|
|
function _configure_tgt_for_config_d() {
|
|
if [[ ! -d /etc/tgt/stack.d/ ]]; then
|
|
sudo ln -sf $CINDER_STATE_PATH/volumes /etc/tgt/stack.d
|
|
echo "include /etc/tgt/stack.d/*" | sudo tee -a /etc/tgt/targets.conf
|
|
fi
|
|
}
|
|
|
|
# start_cinder() - Start running processes, including screen
|
|
function start_cinder() {
|
|
if is_service_enabled c-vol; then
|
|
# Delete any old stack.conf
|
|
sudo rm -f /etc/tgt/conf.d/stack.conf
|
|
_configure_tgt_for_config_d
|
|
if is_ubuntu; then
|
|
# tgt in oneiric doesn't restart properly if tgtd isn't running
|
|
# do it in two steps
|
|
sudo stop tgt || true
|
|
sudo start tgt
|
|
elif is_fedora; then
|
|
# bypass redirection to systemctl during restart
|
|
sudo /sbin/service --skip-redirect tgtd restart
|
|
elif is_suse; then
|
|
restart_service tgtd
|
|
else
|
|
# note for other distros: unstack.sh also uses the tgt/tgtd service
|
|
# name, and would need to be adjusted too
|
|
exit_distro_not_supported "restarting tgt"
|
|
fi
|
|
# NOTE(gfidente): ensure tgtd is running in debug mode
|
|
sudo tgtadm --mode system --op update --name debug --value on
|
|
fi
|
|
|
|
screen_it c-api "cd $CINDER_DIR && $CINDER_BIN_DIR/cinder-api --config-file $CINDER_CONF"
|
|
screen_it c-sch "cd $CINDER_DIR && $CINDER_BIN_DIR/cinder-scheduler --config-file $CINDER_CONF"
|
|
screen_it c-bak "cd $CINDER_DIR && $CINDER_BIN_DIR/cinder-backup --config-file $CINDER_CONF"
|
|
screen_it c-vol "cd $CINDER_DIR && $CINDER_BIN_DIR/cinder-volume --config-file $CINDER_CONF"
|
|
|
|
# NOTE(jdg): For cinder, startup order matters. To ensure that repor_capabilities is received
|
|
# by the scheduler start the cinder-volume service last (or restart it) after the scheduler
|
|
# has started. This is a quick fix for lp bug/1189595
|
|
|
|
# Start proxies if enabled
|
|
if is_service_enabled c-api && is_service_enabled tls-proxy; then
|
|
start_tls_proxy '*' $CINDER_SERVICE_PORT $CINDER_SERVICE_HOST $CINDER_SERVICE_PORT_INT &
|
|
fi
|
|
}
|
|
|
|
# stop_cinder() - Stop running processes
|
|
function stop_cinder() {
|
|
# Kill the cinder screen windows
|
|
for serv in c-api c-bak c-sch c-vol; do
|
|
screen -S $SCREEN_NAME -p $serv -X kill
|
|
done
|
|
|
|
if is_service_enabled c-vol; then
|
|
if is_ubuntu; then
|
|
stop_service tgt
|
|
else
|
|
stop_service tgtd
|
|
fi
|
|
fi
|
|
}
|
|
|
|
|
|
# Restore xtrace
|
|
$XTRACE
|
|
|
|
# Local variables:
|
|
# mode: shell-script
|
|
# End:
|