#!/usr/bin/env bash set -o errexit -o nounset TOP_DIR=$(cd "$(dirname "$0")/.." && pwd) source "$TOP_DIR/config/paths" source "$CONFIG_DIR/credentials" source "$CONFIG_DIR/openstack" source "$LIB_DIR/functions.guest.sh" source "$CONFIG_DIR/admin-openstackrc.sh" exec_logfile indicate_current_auto #------------------------------------------------------------------------------ # Set up Block Storage service (cinder). # http://docs.openstack.org/kilo/install-guide/install/apt/content/cinder-install-storage-node.html #------------------------------------------------------------------------------ # Get FOURTH_OCTET for this node source "$CONFIG_DIR/config.$(hostname)" MY_MGMT_IP=$(get_ip_from_net_and_fourth "MGMT_NET" "$FOURTH_OCTET") echo "IP address of this node's interface in management network: $MY_MGMT_IP." echo "Installing qemu support package for non-raw image types." sudo apt-get install -y qemu echo "Installing the Logical Volume Manager (LVM)." sudo apt-get install -y lvm2 echo "Configuring LVM physical and logical volumes." # We don't have a dedicated physical partition for cinder to use; instead, # we configure a file-backed loopback device. cinder_loop_path=/var/lib/cinder-volumes cinder_loop_dev=/dev/loop2 sudo dd if=/dev/zero of=$cinder_loop_path bs=1 count=0 seek=4G sudo losetup $cinder_loop_dev $cinder_loop_path # Tell upstart to run losetup again when the system is rebooted cat << UPSTART | sudo tee "/etc/init/cinder-losetup.conf" description "Set up loop device for cinder." start on mounted MOUNTPOINT=/ task exec /sbin/losetup $cinder_loop_dev $cinder_loop_path UPSTART sudo pvcreate $cinder_loop_dev sudo vgcreate cinder-volumes $cinder_loop_dev # We could configure LVM to only use loopback devices or only the device # we just set up, but scanning our block devices to find our volume group # is fast enough. echo "Installing cinder." sudo apt-get install -y cinder-volume python-mysqldb conf=/etc/cinder/cinder.conf echo "Configuring $conf." function get_database_url { local db_user=$(service_to_db_user cinder) local db_password=$(service_to_db_password cinder) local database_host=controller-mgmt echo "mysql://$db_user:$db_password@$database_host/cinder" } database_url=$(get_database_url) echo "Setting database connection: $database_url." iniset_sudo $conf database connection "$database_url" # Configure [DEFAULT] section. iniset_sudo $conf DEFAULT rpc_backend rabbit iniset_sudo $conf oslo_messaging_rabbit rabbit_host controller-mgmt iniset_sudo $conf oslo_messaging_rabbit rabbit_userid openstack iniset_sudo $conf oslo_messaging_rabbit rabbit_password "$RABBIT_PASSWORD" iniset_sudo $conf DEFAULT auth_strategy keystone # Configure [keystone_authtoken] section. cinder_admin_user=$(service_to_user_name cinder) cinder_admin_password=$(service_to_user_password cinder) iniset_sudo $conf keystone_authtoken auth_uri http://controller-mgmt:5000 iniset_sudo $conf keystone_authtoken auth_url http://controller-mgmt:35357 iniset_sudo $conf keystone_authtoken auth_plugin password iniset_sudo $conf keystone_authtoken project_domain_id default iniset_sudo $conf keystone_authtoken user_domain_id default iniset_sudo $conf keystone_authtoken project_name "$SERVICE_PROJECT_NAME" iniset_sudo $conf keystone_authtoken username "$cinder_admin_user" iniset_sudo $conf keystone_authtoken password "$cinder_admin_password" iniset_sudo $conf DEFAULT my_ip "$MY_MGMT_IP" iniset_sudo $conf lvm volume_driver cinder.volume.drivers.lvm.LVMVolumeDriver iniset_sudo $conf lvm volume_group cinder-volumes iniset_sudo $conf lvm iscsi_protocol iscsi iniset_sudo $conf lvm iscsi_helper tgtadm iniset_sudo $conf DEFAULT enabled_backends lvm iniset_sudo $conf DEFAULT glance_host controller-mgmt iniset_sudo $conf oslo_concurrency lock_path /var/lock/cinder iniset_sudo $conf DEFAULT verbose True echo "Restarting cinder service." sudo service tgt restart sudo service cinder-volume restart sudo rm -f /var/lib/cinder/cinder.sqlite #------------------------------------------------------------------------------ # Verify the Block Storage installation # http://docs.openstack.org/kilo/install-guide/install/apt/content/cinder-verify.html #------------------------------------------------------------------------------ echo "Verifying Block Storage installation on controller node." echo "Sourcing the admin credentials." AUTH="source $CONFIG_DIR/admin-openstackrc.sh" # It takes time for Cinder to be aware of its services status. # Force restart cinder API and wait for 20 seconds. echo "Restarting Cinder API." node_ssh controller-mgmt "sudo service cinder-api restart" sleep 20 echo "Waiting for cinder to start." until node_ssh controller-mgmt "$AUTH; cinder service-list" >/dev/null 2>&1; do sleep 1 done echo "cinder service-list" node_ssh controller-mgmt "$AUTH; cinder service-list" function check_cinder_services { # It takes some time for cinder to detect its services and update its # status. This method will wait for 20 seconds to get the status of the # Cinder services. local i=1 while [ : ]; do if [[ -z $(node_ssh controller-mgmt "$AUTH; cinder service-list" | grep down) ]] > /dev/null 2>&1; then echo "Cinder services seem to be up and running!" return 0 fi i=$((i + 1)) if [[ "$i" -gt "20" ]]; then echo "Error, cinder services are not working as expected." exit 0 fi echo -n . sleep 1 done } # To avoid race conditions which were causing Cinder Volumes script to fail, # check the status of the cinder services. Cinder takes a few seconds before it # is aware of the exact status of its services. echo "Waiting for all cinder services to start." check_cinder_services echo "Sourcing the demo credentials." AUTH="source $CONFIG_DIR/demo-openstackrc.sh" echo "cinder create --display-name demo-volume1 1" node_ssh controller-mgmt "$AUTH; cinder create --display-name demo-volume1 1;sleep 20" echo -n "Waiting for cinder to list the new volume." until node_ssh controller-mgmt "$AUTH; cinder list | grep demo-volume1" > /dev/null 2>&1; do echo -n . sleep 1 done function wait_for_cinder_volume { # Wait for cinder volume to be created echo -n 'Waiting for cinder volume to be created.' local i=1 while [ : ]; do if [[ -z $(node_ssh controller-mgmt "$AUTH;cinder list" | grep creating) ]] > /dev/null 2>&1; then # Proceed if the state of cinder-volumes is error or created. # Cinder volumes cannot be deleted when it is in creating state. # Throw an error and stop this script. # The purpose of this method is to resolve cinder-volumes race. return 0 fi i=$((i + 1)) if [[ "$i" -gt "20" ]]; then echo "Error creating cinder volume." echo "[Warning]: Debug cinder volumes service on the compute node. Delete the cinder-volume demo-volume1. Script could not delete this volume." exit 0 fi echo -n . sleep 1 done echo } echo "Checking if volume is created." wait_for_cinder_volume echo "cinder delete demo-volume1" node_ssh controller-mgmt "$AUTH; cinder delete demo-volume1" echo "cinder list" node_ssh controller-mgmt "$AUTH; cinder list"