90bcd2ff4d
VMDK formats such as monolithicFlat and vmfs require two files to be fully consumable by the Nova drivers (a descriptor-data pair: *.vmdk and *-flat.vmdk). On the upload of the descriptor (*.vmdk), upload_image.sh should attempt to retrieve the *-flat.vmdk. The same way, the descriptor should be retrieved when a flat disk is uploaded. On success, the upload script will be able to use the flat disk as the image content and the relevant descriptor settings as the image metadata. Change-Id: I9214754029c46dd60b9e7d606d84d8819a498a8d Closes-Bug: #1252443
1948 lines
63 KiB
Plaintext
1948 lines
63 KiB
Plaintext
# functions - Common functions used by DevStack components
|
||
#
|
||
# The following variables are assumed to be defined by certain functions:
|
||
#
|
||
# - ``ENABLED_SERVICES``
|
||
# - ``ERROR_ON_CLONE``
|
||
# - ``FILES``
|
||
# - ``GLANCE_HOSTPORT``
|
||
# - ``OFFLINE``
|
||
# - ``PIP_DOWNLOAD_CACHE``
|
||
# - ``PIP_USE_MIRRORS``
|
||
# - ``RECLONE``
|
||
# - ``TRACK_DEPENDS``
|
||
# - ``http_proxy``, ``https_proxy``, ``no_proxy``
|
||
|
||
|
||
# Save trace setting
|
||
XTRACE=$(set +o | grep xtrace)
|
||
set +o xtrace
|
||
|
||
|
||
# Convert CIDR notation to a IPv4 netmask
|
||
# cidr2netmask cidr-bits
|
||
function cidr2netmask() {
|
||
local maskpat="255 255 255 255"
|
||
local maskdgt="254 252 248 240 224 192 128"
|
||
set -- ${maskpat:0:$(( ($1 / 8) * 4 ))}${maskdgt:$(( (7 - ($1 % 8)) * 4 )):3}
|
||
echo ${1-0}.${2-0}.${3-0}.${4-0}
|
||
}
|
||
|
||
|
||
# Return the network portion of the given IP address using netmask
|
||
# netmask is in the traditional dotted-quad format
|
||
# maskip ip-address netmask
|
||
function maskip() {
|
||
local ip=$1
|
||
local mask=$2
|
||
local l="${ip%.*}"; local r="${ip#*.}"; local n="${mask%.*}"; local m="${mask#*.}"
|
||
local subnet=$((${ip%%.*}&${mask%%.*})).$((${r%%.*}&${m%%.*})).$((${l##*.}&${n##*.})).$((${ip##*.}&${mask##*.}))
|
||
echo $subnet
|
||
}
|
||
|
||
|
||
# Exit 0 if address is in network or 1 if address is not in network
|
||
# ip-range is in CIDR notation: 1.2.3.4/20
|
||
# address_in_net ip-address ip-range
|
||
function address_in_net() {
|
||
local ip=$1
|
||
local range=$2
|
||
local masklen=${range#*/}
|
||
local network=$(maskip ${range%/*} $(cidr2netmask $masklen))
|
||
local subnet=$(maskip $ip $(cidr2netmask $masklen))
|
||
[[ $network == $subnet ]]
|
||
}
|
||
|
||
|
||
# Wrapper for ``apt-get`` to set cache and proxy environment variables
|
||
# Uses globals ``OFFLINE``, ``*_proxy``
|
||
# apt_get operation package [package ...]
|
||
function apt_get() {
|
||
[[ "$OFFLINE" = "True" || -z "$@" ]] && return
|
||
local sudo="sudo"
|
||
[[ "$(id -u)" = "0" ]] && sudo="env"
|
||
$sudo DEBIAN_FRONTEND=noninteractive \
|
||
http_proxy=$http_proxy https_proxy=$https_proxy \
|
||
no_proxy=$no_proxy \
|
||
apt-get --option "Dpkg::Options::=--force-confold" --assume-yes "$@"
|
||
}
|
||
|
||
|
||
# Gracefully cp only if source file/dir exists
|
||
# cp_it source destination
|
||
function cp_it {
|
||
if [ -e $1 ] || [ -d $1 ]; then
|
||
cp -pRL $1 $2
|
||
fi
|
||
}
|
||
|
||
|
||
# Prints backtrace info
|
||
# filename:lineno:function
|
||
function backtrace {
|
||
local level=$1
|
||
local deep=$((${#BASH_SOURCE[@]} - 1))
|
||
echo "[Call Trace]"
|
||
while [ $level -le $deep ]; do
|
||
echo "${BASH_SOURCE[$deep]}:${BASH_LINENO[$deep-1]}:${FUNCNAME[$deep-1]}"
|
||
deep=$((deep - 1))
|
||
done
|
||
}
|
||
|
||
|
||
# Prints line number and "message" then exits
|
||
# die $LINENO "message"
|
||
function die() {
|
||
local exitcode=$?
|
||
set +o xtrace
|
||
local line=$1; shift
|
||
if [ $exitcode == 0 ]; then
|
||
exitcode=1
|
||
fi
|
||
backtrace 2
|
||
err $line "$*"
|
||
exit $exitcode
|
||
}
|
||
|
||
|
||
# Checks an environment variable is not set or has length 0 OR if the
|
||
# exit code is non-zero and prints "message" and exits
|
||
# NOTE: env-var is the variable name without a '$'
|
||
# die_if_not_set $LINENO env-var "message"
|
||
function die_if_not_set() {
|
||
local exitcode=$?
|
||
FXTRACE=$(set +o | grep xtrace)
|
||
set +o xtrace
|
||
local line=$1; shift
|
||
local evar=$1; shift
|
||
if ! is_set $evar || [ $exitcode != 0 ]; then
|
||
die $line "$*"
|
||
fi
|
||
$FXTRACE
|
||
}
|
||
|
||
|
||
# Prints line number and "message" in error format
|
||
# err $LINENO "message"
|
||
function err() {
|
||
local exitcode=$?
|
||
errXTRACE=$(set +o | grep xtrace)
|
||
set +o xtrace
|
||
local msg="[ERROR] ${BASH_SOURCE[2]}:$1 $2"
|
||
echo $msg 1>&2;
|
||
if [[ -n ${SCREEN_LOGDIR} ]]; then
|
||
echo $msg >> "${SCREEN_LOGDIR}/error.log"
|
||
fi
|
||
$errXTRACE
|
||
return $exitcode
|
||
}
|
||
|
||
|
||
# Checks an environment variable is not set or has length 0 OR if the
|
||
# exit code is non-zero and prints "message"
|
||
# NOTE: env-var is the variable name without a '$'
|
||
# err_if_not_set $LINENO env-var "message"
|
||
function err_if_not_set() {
|
||
local exitcode=$?
|
||
errinsXTRACE=$(set +o | grep xtrace)
|
||
set +o xtrace
|
||
local line=$1; shift
|
||
local evar=$1; shift
|
||
if ! is_set $evar || [ $exitcode != 0 ]; then
|
||
err $line "$*"
|
||
fi
|
||
$errinsXTRACE
|
||
return $exitcode
|
||
}
|
||
|
||
|
||
# Prints line number and "message" in warning format
|
||
# warn $LINENO "message"
|
||
function warn() {
|
||
local exitcode=$?
|
||
errXTRACE=$(set +o | grep xtrace)
|
||
set +o xtrace
|
||
local msg="[WARNING] ${BASH_SOURCE[2]}:$1 $2"
|
||
echo $msg 1>&2;
|
||
if [[ -n ${SCREEN_LOGDIR} ]]; then
|
||
echo $msg >> "${SCREEN_LOGDIR}/error.log"
|
||
fi
|
||
$errXTRACE
|
||
return $exitcode
|
||
}
|
||
|
||
|
||
# HTTP and HTTPS proxy servers are supported via the usual environment variables [1]
|
||
# ``http_proxy``, ``https_proxy`` and ``no_proxy``. They can be set in
|
||
# ``localrc`` or on the command line if necessary::
|
||
#
|
||
# [1] http://www.w3.org/Daemon/User/Proxies/ProxyClients.html
|
||
#
|
||
# http_proxy=http://proxy.example.com:3128/ no_proxy=repo.example.net ./stack.sh
|
||
|
||
function export_proxy_variables() {
|
||
if [[ -n "$http_proxy" ]]; then
|
||
export http_proxy=$http_proxy
|
||
fi
|
||
if [[ -n "$https_proxy" ]]; then
|
||
export https_proxy=$https_proxy
|
||
fi
|
||
if [[ -n "$no_proxy" ]]; then
|
||
export no_proxy=$no_proxy
|
||
fi
|
||
}
|
||
|
||
|
||
# Grab a numbered field from python prettytable output
|
||
# Fields are numbered starting with 1
|
||
# Reverse syntax is supported: -1 is the last field, -2 is second to last, etc.
|
||
# get_field field-number
|
||
function get_field() {
|
||
while read data; do
|
||
if [ "$1" -lt 0 ]; then
|
||
field="(\$(NF$1))"
|
||
else
|
||
field="\$$(($1 + 1))"
|
||
fi
|
||
echo "$data" | awk -F'[ \t]*\\|[ \t]*' "{print $field}"
|
||
done
|
||
}
|
||
|
||
|
||
# Get the default value for HOST_IP
|
||
# get_default_host_ip fixed_range floating_range host_ip_iface host_ip
|
||
function get_default_host_ip() {
|
||
local fixed_range=$1
|
||
local floating_range=$2
|
||
local host_ip_iface=$3
|
||
local host_ip=$4
|
||
|
||
# Find the interface used for the default route
|
||
host_ip_iface=${host_ip_iface:-$(ip route | sed -n '/^default/{ s/.*dev \(\w\+\)\s\+.*/\1/; p; }' | head -1)}
|
||
# Search for an IP unless an explicit is set by ``HOST_IP`` environment variable
|
||
if [ -z "$host_ip" -o "$host_ip" == "dhcp" ]; then
|
||
host_ip=""
|
||
host_ips=`LC_ALL=C ip -f inet addr show ${host_ip_iface} | awk '/inet/ {split($2,parts,"/"); print parts[1]}'`
|
||
for IP in $host_ips; do
|
||
# Attempt to filter out IP addresses that are part of the fixed and
|
||
# floating range. Note that this method only works if the ``netaddr``
|
||
# python library is installed. If it is not installed, an error
|
||
# will be printed and the first IP from the interface will be used.
|
||
# If that is not correct set ``HOST_IP`` in ``localrc`` to the correct
|
||
# address.
|
||
if ! (address_in_net $IP $fixed_range || address_in_net $IP $floating_range); then
|
||
host_ip=$IP
|
||
break;
|
||
fi
|
||
done
|
||
fi
|
||
echo $host_ip
|
||
}
|
||
|
||
|
||
function _get_package_dir() {
|
||
local pkg_dir
|
||
if is_ubuntu; then
|
||
pkg_dir=$FILES/apts
|
||
elif is_fedora; then
|
||
pkg_dir=$FILES/rpms
|
||
elif is_suse; then
|
||
pkg_dir=$FILES/rpms-suse
|
||
else
|
||
exit_distro_not_supported "list of packages"
|
||
fi
|
||
echo "$pkg_dir"
|
||
}
|
||
|
||
|
||
# get_packages() collects a list of package names of any type from the
|
||
# prerequisite files in ``files/{apts|rpms}``. The list is intended
|
||
# to be passed to a package installer such as apt or yum.
|
||
#
|
||
# Only packages required for the services in 1st argument will be
|
||
# included. Two bits of metadata are recognized in the prerequisite files:
|
||
#
|
||
# - ``# NOPRIME`` defers installation to be performed later in `stack.sh`
|
||
# - ``# dist:DISTRO`` or ``dist:DISTRO1,DISTRO2`` limits the selection
|
||
# of the package to the distros listed. The distro names are case insensitive.
|
||
function get_packages() {
|
||
local services=$@
|
||
local package_dir=$(_get_package_dir)
|
||
local file_to_parse
|
||
local service
|
||
|
||
if [[ -z "$package_dir" ]]; then
|
||
echo "No package directory supplied"
|
||
return 1
|
||
fi
|
||
if [[ -z "$DISTRO" ]]; then
|
||
GetDistro
|
||
fi
|
||
for service in ${services//,/ }; do
|
||
# Allow individual services to specify dependencies
|
||
if [[ -e ${package_dir}/${service} ]]; then
|
||
file_to_parse="${file_to_parse} $service"
|
||
fi
|
||
# NOTE(sdague) n-api needs glance for now because that's where
|
||
# glance client is
|
||
if [[ $service == n-api ]]; then
|
||
if [[ ! $file_to_parse =~ nova ]]; then
|
||
file_to_parse="${file_to_parse} nova"
|
||
fi
|
||
if [[ ! $file_to_parse =~ glance ]]; then
|
||
file_to_parse="${file_to_parse} glance"
|
||
fi
|
||
elif [[ $service == c-* ]]; then
|
||
if [[ ! $file_to_parse =~ cinder ]]; then
|
||
file_to_parse="${file_to_parse} cinder"
|
||
fi
|
||
elif [[ $service == ceilometer-* ]]; then
|
||
if [[ ! $file_to_parse =~ ceilometer ]]; then
|
||
file_to_parse="${file_to_parse} ceilometer"
|
||
fi
|
||
elif [[ $service == s-* ]]; then
|
||
if [[ ! $file_to_parse =~ swift ]]; then
|
||
file_to_parse="${file_to_parse} swift"
|
||
fi
|
||
elif [[ $service == n-* ]]; then
|
||
if [[ ! $file_to_parse =~ nova ]]; then
|
||
file_to_parse="${file_to_parse} nova"
|
||
fi
|
||
elif [[ $service == g-* ]]; then
|
||
if [[ ! $file_to_parse =~ glance ]]; then
|
||
file_to_parse="${file_to_parse} glance"
|
||
fi
|
||
elif [[ $service == key* ]]; then
|
||
if [[ ! $file_to_parse =~ keystone ]]; then
|
||
file_to_parse="${file_to_parse} keystone"
|
||
fi
|
||
elif [[ $service == q-* ]]; then
|
||
if [[ ! $file_to_parse =~ neutron ]]; then
|
||
file_to_parse="${file_to_parse} neutron"
|
||
fi
|
||
fi
|
||
done
|
||
|
||
for file in ${file_to_parse}; do
|
||
local fname=${package_dir}/${file}
|
||
local OIFS line package distros distro
|
||
[[ -e $fname ]] || continue
|
||
|
||
OIFS=$IFS
|
||
IFS=$'\n'
|
||
for line in $(<${fname}); do
|
||
if [[ $line =~ "NOPRIME" ]]; then
|
||
continue
|
||
fi
|
||
|
||
# Assume we want this package
|
||
package=${line%#*}
|
||
inst_pkg=1
|
||
|
||
# Look for # dist:xxx in comment
|
||
if [[ $line =~ (.*)#.*dist:([^ ]*) ]]; then
|
||
# We are using BASH regexp matching feature.
|
||
package=${BASH_REMATCH[1]}
|
||
distros=${BASH_REMATCH[2]}
|
||
# In bash ${VAR,,} will lowecase VAR
|
||
# Look for a match in the distro list
|
||
if [[ ! ${distros,,} =~ ${DISTRO,,} ]]; then
|
||
# If no match then skip this package
|
||
inst_pkg=0
|
||
fi
|
||
fi
|
||
|
||
# Look for # testonly in comment
|
||
if [[ $line =~ (.*)#.*testonly.* ]]; then
|
||
package=${BASH_REMATCH[1]}
|
||
# Are we installing test packages? (test for the default value)
|
||
if [[ $INSTALL_TESTONLY_PACKAGES = "False" ]]; then
|
||
# If not installing test packages the skip this package
|
||
inst_pkg=0
|
||
fi
|
||
fi
|
||
|
||
if [[ $inst_pkg = 1 ]]; then
|
||
echo $package
|
||
fi
|
||
done
|
||
IFS=$OIFS
|
||
done
|
||
}
|
||
|
||
|
||
# Determine OS Vendor, Release and Update
|
||
# Tested with OS/X, Ubuntu, RedHat, CentOS, Fedora
|
||
# Returns results in global variables:
|
||
# os_VENDOR - vendor name
|
||
# os_RELEASE - release
|
||
# os_UPDATE - update
|
||
# os_PACKAGE - package type
|
||
# os_CODENAME - vendor's codename for release
|
||
# GetOSVersion
|
||
GetOSVersion() {
|
||
# Figure out which vendor we are
|
||
if [[ -x "`which sw_vers 2>/dev/null`" ]]; then
|
||
# OS/X
|
||
os_VENDOR=`sw_vers -productName`
|
||
os_RELEASE=`sw_vers -productVersion`
|
||
os_UPDATE=${os_RELEASE##*.}
|
||
os_RELEASE=${os_RELEASE%.*}
|
||
os_PACKAGE=""
|
||
if [[ "$os_RELEASE" =~ "10.7" ]]; then
|
||
os_CODENAME="lion"
|
||
elif [[ "$os_RELEASE" =~ "10.6" ]]; then
|
||
os_CODENAME="snow leopard"
|
||
elif [[ "$os_RELEASE" =~ "10.5" ]]; then
|
||
os_CODENAME="leopard"
|
||
elif [[ "$os_RELEASE" =~ "10.4" ]]; then
|
||
os_CODENAME="tiger"
|
||
elif [[ "$os_RELEASE" =~ "10.3" ]]; then
|
||
os_CODENAME="panther"
|
||
else
|
||
os_CODENAME=""
|
||
fi
|
||
elif [[ -x $(which lsb_release 2>/dev/null) ]]; then
|
||
os_VENDOR=$(lsb_release -i -s)
|
||
os_RELEASE=$(lsb_release -r -s)
|
||
os_UPDATE=""
|
||
os_PACKAGE="rpm"
|
||
if [[ "Debian,Ubuntu,LinuxMint" =~ $os_VENDOR ]]; then
|
||
os_PACKAGE="deb"
|
||
elif [[ "SUSE LINUX" =~ $os_VENDOR ]]; then
|
||
lsb_release -d -s | grep -q openSUSE
|
||
if [[ $? -eq 0 ]]; then
|
||
os_VENDOR="openSUSE"
|
||
fi
|
||
elif [[ $os_VENDOR == "openSUSE project" ]]; then
|
||
os_VENDOR="openSUSE"
|
||
elif [[ $os_VENDOR =~ Red.*Hat ]]; then
|
||
os_VENDOR="Red Hat"
|
||
fi
|
||
os_CODENAME=$(lsb_release -c -s)
|
||
elif [[ -r /etc/redhat-release ]]; then
|
||
# Red Hat Enterprise Linux Server release 5.5 (Tikanga)
|
||
# CentOS release 5.5 (Final)
|
||
# CentOS Linux release 6.0 (Final)
|
||
# Fedora release 16 (Verne)
|
||
# XenServer release 6.2.0-70446c (xenenterprise)
|
||
os_CODENAME=""
|
||
for r in "Red Hat" CentOS Fedora XenServer; do
|
||
os_VENDOR=$r
|
||
if [[ -n "`grep \"$r\" /etc/redhat-release`" ]]; then
|
||
ver=`sed -e 's/^.* \(.*\) (\(.*\)).*$/\1\|\2/' /etc/redhat-release`
|
||
os_CODENAME=${ver#*|}
|
||
os_RELEASE=${ver%|*}
|
||
os_UPDATE=${os_RELEASE##*.}
|
||
os_RELEASE=${os_RELEASE%.*}
|
||
break
|
||
fi
|
||
os_VENDOR=""
|
||
done
|
||
os_PACKAGE="rpm"
|
||
elif [[ -r /etc/SuSE-release ]]; then
|
||
for r in openSUSE "SUSE Linux"; do
|
||
if [[ "$r" = "SUSE Linux" ]]; then
|
||
os_VENDOR="SUSE LINUX"
|
||
else
|
||
os_VENDOR=$r
|
||
fi
|
||
|
||
if [[ -n "`grep \"$r\" /etc/SuSE-release`" ]]; then
|
||
os_CODENAME=`grep "CODENAME = " /etc/SuSE-release | sed 's:.* = ::g'`
|
||
os_RELEASE=`grep "VERSION = " /etc/SuSE-release | sed 's:.* = ::g'`
|
||
os_UPDATE=`grep "PATCHLEVEL = " /etc/SuSE-release | sed 's:.* = ::g'`
|
||
break
|
||
fi
|
||
os_VENDOR=""
|
||
done
|
||
os_PACKAGE="rpm"
|
||
# If lsb_release is not installed, we should be able to detect Debian OS
|
||
elif [[ -f /etc/debian_version ]] && [[ $(cat /proc/version) =~ "Debian" ]]; then
|
||
os_VENDOR="Debian"
|
||
os_PACKAGE="deb"
|
||
os_CODENAME=$(awk '/VERSION=/' /etc/os-release | sed 's/VERSION=//' | sed -r 's/\"|\(|\)//g' | awk '{print $2}')
|
||
os_RELEASE=$(awk '/VERSION_ID=/' /etc/os-release | sed 's/VERSION_ID=//' | sed 's/\"//g')
|
||
fi
|
||
export os_VENDOR os_RELEASE os_UPDATE os_PACKAGE os_CODENAME
|
||
}
|
||
|
||
|
||
# Translate the OS version values into common nomenclature
|
||
# Sets ``DISTRO`` from the ``os_*`` values
|
||
function GetDistro() {
|
||
GetOSVersion
|
||
if [[ "$os_VENDOR" =~ (Ubuntu) || "$os_VENDOR" =~ (Debian) ]]; then
|
||
# 'Everyone' refers to Ubuntu / Debian releases by the code name adjective
|
||
DISTRO=$os_CODENAME
|
||
elif [[ "$os_VENDOR" =~ (Fedora) ]]; then
|
||
# For Fedora, just use 'f' and the release
|
||
DISTRO="f$os_RELEASE"
|
||
elif [[ "$os_VENDOR" =~ (openSUSE) ]]; then
|
||
DISTRO="opensuse-$os_RELEASE"
|
||
elif [[ "$os_VENDOR" =~ (SUSE LINUX) ]]; then
|
||
# For SLE, also use the service pack
|
||
if [[ -z "$os_UPDATE" ]]; then
|
||
DISTRO="sle${os_RELEASE}"
|
||
else
|
||
DISTRO="sle${os_RELEASE}sp${os_UPDATE}"
|
||
fi
|
||
elif [[ "$os_VENDOR" =~ (Red Hat) || "$os_VENDOR" =~ (CentOS) ]]; then
|
||
# Drop the . release as we assume it's compatible
|
||
DISTRO="rhel${os_RELEASE::1}"
|
||
elif [[ "$os_VENDOR" =~ (XenServer) ]]; then
|
||
DISTRO="xs$os_RELEASE"
|
||
else
|
||
# Catch-all for now is Vendor + Release + Update
|
||
DISTRO="$os_VENDOR-$os_RELEASE.$os_UPDATE"
|
||
fi
|
||
export DISTRO
|
||
}
|
||
|
||
|
||
# Determine if current distribution is a Fedora-based distribution
|
||
# (Fedora, RHEL, CentOS, etc).
|
||
# is_fedora
|
||
function is_fedora {
|
||
if [[ -z "$os_VENDOR" ]]; then
|
||
GetOSVersion
|
||
fi
|
||
|
||
[ "$os_VENDOR" = "Fedora" ] || [ "$os_VENDOR" = "Red Hat" ] || [ "$os_VENDOR" = "CentOS" ]
|
||
}
|
||
|
||
|
||
# Determine if current distribution is a SUSE-based distribution
|
||
# (openSUSE, SLE).
|
||
# is_suse
|
||
function is_suse {
|
||
if [[ -z "$os_VENDOR" ]]; then
|
||
GetOSVersion
|
||
fi
|
||
|
||
[ "$os_VENDOR" = "openSUSE" ] || [ "$os_VENDOR" = "SUSE LINUX" ]
|
||
}
|
||
|
||
|
||
# Determine if current distribution is an Ubuntu-based distribution
|
||
# It will also detect non-Ubuntu but Debian-based distros
|
||
# is_ubuntu
|
||
function is_ubuntu {
|
||
if [[ -z "$os_PACKAGE" ]]; then
|
||
GetOSVersion
|
||
fi
|
||
[ "$os_PACKAGE" = "deb" ]
|
||
}
|
||
|
||
|
||
# Exit after outputting a message about the distribution not being supported.
|
||
# exit_distro_not_supported [optional-string-telling-what-is-missing]
|
||
function exit_distro_not_supported {
|
||
if [[ -z "$DISTRO" ]]; then
|
||
GetDistro
|
||
fi
|
||
|
||
if [ $# -gt 0 ]; then
|
||
die $LINENO "Support for $DISTRO is incomplete: no support for $@"
|
||
else
|
||
die $LINENO "Support for $DISTRO is incomplete."
|
||
fi
|
||
}
|
||
|
||
# Utility function for checking machine architecture
|
||
# is_arch arch-type
|
||
function is_arch {
|
||
ARCH_TYPE=$1
|
||
|
||
[ "($uname -m)" = "$ARCH_TYPE" ]
|
||
}
|
||
|
||
# Checks if installed Apache is <= given version
|
||
# $1 = x.y.z (version string of Apache)
|
||
function check_apache_version {
|
||
local cmd="apachectl"
|
||
if ! [[ -x $(which apachectl 2>/dev/null) ]]; then
|
||
cmd="/usr/sbin/apachectl"
|
||
fi
|
||
|
||
local version=$($cmd -v | grep version | grep -Po 'Apache/\K[^ ]*')
|
||
expr "$version" '>=' $1 > /dev/null
|
||
}
|
||
|
||
# git clone only if directory doesn't exist already. Since ``DEST`` might not
|
||
# be owned by the installation user, we create the directory and change the
|
||
# ownership to the proper user.
|
||
# Set global RECLONE=yes to simulate a clone when dest-dir exists
|
||
# Set global ERROR_ON_CLONE=True to abort execution with an error if the git repo
|
||
# does not exist (default is False, meaning the repo will be cloned).
|
||
# Uses global ``OFFLINE``
|
||
# git_clone remote dest-dir branch
|
||
function git_clone {
|
||
GIT_REMOTE=$1
|
||
GIT_DEST=$2
|
||
GIT_REF=$3
|
||
RECLONE=$(trueorfalse False $RECLONE)
|
||
|
||
if [[ "$OFFLINE" = "True" ]]; then
|
||
echo "Running in offline mode, clones already exist"
|
||
# print out the results so we know what change was used in the logs
|
||
cd $GIT_DEST
|
||
git show --oneline | head -1
|
||
return
|
||
fi
|
||
|
||
if echo $GIT_REF | egrep -q "^refs"; then
|
||
# If our branch name is a gerrit style refs/changes/...
|
||
if [[ ! -d $GIT_DEST ]]; then
|
||
[[ "$ERROR_ON_CLONE" = "True" ]] && \
|
||
die $LINENO "Cloning not allowed in this configuration"
|
||
git clone $GIT_REMOTE $GIT_DEST
|
||
fi
|
||
cd $GIT_DEST
|
||
git fetch $GIT_REMOTE $GIT_REF && git checkout FETCH_HEAD
|
||
else
|
||
# do a full clone only if the directory doesn't exist
|
||
if [[ ! -d $GIT_DEST ]]; then
|
||
[[ "$ERROR_ON_CLONE" = "True" ]] && \
|
||
die $LINENO "Cloning not allowed in this configuration"
|
||
git clone $GIT_REMOTE $GIT_DEST
|
||
cd $GIT_DEST
|
||
# This checkout syntax works for both branches and tags
|
||
git checkout $GIT_REF
|
||
elif [[ "$RECLONE" = "True" ]]; then
|
||
# if it does exist then simulate what clone does if asked to RECLONE
|
||
cd $GIT_DEST
|
||
# set the url to pull from and fetch
|
||
git remote set-url origin $GIT_REMOTE
|
||
git fetch origin
|
||
# remove the existing ignored files (like pyc) as they cause breakage
|
||
# (due to the py files having older timestamps than our pyc, so python
|
||
# thinks the pyc files are correct using them)
|
||
find $GIT_DEST -name '*.pyc' -delete
|
||
|
||
# handle GIT_REF accordingly to type (tag, branch)
|
||
if [[ -n "`git show-ref refs/tags/$GIT_REF`" ]]; then
|
||
git_update_tag $GIT_REF
|
||
elif [[ -n "`git show-ref refs/heads/$GIT_REF`" ]]; then
|
||
git_update_branch $GIT_REF
|
||
elif [[ -n "`git show-ref refs/remotes/origin/$GIT_REF`" ]]; then
|
||
git_update_remote_branch $GIT_REF
|
||
else
|
||
die $LINENO "$GIT_REF is neither branch nor tag"
|
||
fi
|
||
|
||
fi
|
||
fi
|
||
|
||
# print out the results so we know what change was used in the logs
|
||
cd $GIT_DEST
|
||
git show --oneline | head -1
|
||
}
|
||
|
||
|
||
# git update using reference as a branch.
|
||
# git_update_branch ref
|
||
function git_update_branch() {
|
||
|
||
GIT_BRANCH=$1
|
||
|
||
git checkout -f origin/$GIT_BRANCH
|
||
# a local branch might not exist
|
||
git branch -D $GIT_BRANCH || true
|
||
git checkout -b $GIT_BRANCH
|
||
}
|
||
|
||
|
||
# git update using reference as a branch.
|
||
# git_update_remote_branch ref
|
||
function git_update_remote_branch() {
|
||
|
||
GIT_BRANCH=$1
|
||
|
||
git checkout -b $GIT_BRANCH -t origin/$GIT_BRANCH
|
||
}
|
||
|
||
|
||
# git update using reference as a tag. Be careful editing source at that repo
|
||
# as working copy will be in a detached mode
|
||
# git_update_tag ref
|
||
function git_update_tag() {
|
||
|
||
GIT_TAG=$1
|
||
|
||
git tag -d $GIT_TAG
|
||
# fetching given tag only
|
||
git fetch origin tag $GIT_TAG
|
||
git checkout -f $GIT_TAG
|
||
}
|
||
|
||
|
||
# Comment an option in an INI file
|
||
# inicomment config-file section option
|
||
function inicomment() {
|
||
local file=$1
|
||
local section=$2
|
||
local option=$3
|
||
sed -i -e "/^\[$section\]/,/^\[.*\]/ s|^\($option[ \t]*=.*$\)|#\1|" "$file"
|
||
}
|
||
|
||
|
||
# Uncomment an option in an INI file
|
||
# iniuncomment config-file section option
|
||
function iniuncomment() {
|
||
local file=$1
|
||
local section=$2
|
||
local option=$3
|
||
sed -i -e "/^\[$section\]/,/^\[.*\]/ s|[^ \t]*#[ \t]*\($option[ \t]*=.*$\)|\1|" "$file"
|
||
}
|
||
|
||
|
||
# Get an option from an INI file
|
||
# iniget config-file section option
|
||
function iniget() {
|
||
local file=$1
|
||
local section=$2
|
||
local option=$3
|
||
local line
|
||
line=$(sed -ne "/^\[$section\]/,/^\[.*\]/ { /^$option[ \t]*=/ p; }" "$file")
|
||
echo ${line#*=}
|
||
}
|
||
|
||
|
||
# Determinate is the given option present in the INI file
|
||
# ini_has_option config-file section option
|
||
function ini_has_option() {
|
||
local file=$1
|
||
local section=$2
|
||
local option=$3
|
||
local line
|
||
line=$(sed -ne "/^\[$section\]/,/^\[.*\]/ { /^$option[ \t]*=/ p; }" "$file")
|
||
[ -n "$line" ]
|
||
}
|
||
|
||
|
||
# Set an option in an INI file
|
||
# iniset config-file section option value
|
||
function iniset() {
|
||
local file=$1
|
||
local section=$2
|
||
local option=$3
|
||
local value=$4
|
||
|
||
if ! grep -q "^\[$section\]" "$file" 2>/dev/null; then
|
||
# Add section at the end
|
||
echo -e "\n[$section]" >>"$file"
|
||
fi
|
||
if ! ini_has_option "$file" "$section" "$option"; then
|
||
# Add it
|
||
sed -i -e "/^\[$section\]/ a\\
|
||
$option = $value
|
||
" "$file"
|
||
else
|
||
# Replace it
|
||
sed -i -e "/^\[$section\]/,/^\[.*\]/ s|^\($option[ \t]*=[ \t]*\).*$|\1$value|" "$file"
|
||
fi
|
||
}
|
||
|
||
|
||
# Get a multiple line option from an INI file
|
||
# iniget_multiline config-file section option
|
||
function iniget_multiline() {
|
||
local file=$1
|
||
local section=$2
|
||
local option=$3
|
||
local values
|
||
values=$(sed -ne "/^\[$section\]/,/^\[.*\]/ { s/^$option[ \t]*=[ \t]*//gp; }" "$file")
|
||
echo ${values}
|
||
}
|
||
|
||
|
||
# Set a multiple line option in an INI file
|
||
# iniset_multiline config-file section option value1 value2 valu3 ...
|
||
function iniset_multiline() {
|
||
local file=$1
|
||
local section=$2
|
||
local option=$3
|
||
shift 3
|
||
local values
|
||
for v in $@; do
|
||
# The later sed command inserts each new value in the line next to
|
||
# the section identifier, which causes the values to be inserted in
|
||
# the reverse order. Do a reverse here to keep the original order.
|
||
values="$v ${values}"
|
||
done
|
||
if ! grep -q "^\[$section\]" "$file"; then
|
||
# Add section at the end
|
||
echo -e "\n[$section]" >>"$file"
|
||
else
|
||
# Remove old values
|
||
sed -i -e "/^\[$section\]/,/^\[.*\]/ { /^$option[ \t]*=/ d; }" "$file"
|
||
fi
|
||
# Add new ones
|
||
for v in $values; do
|
||
sed -i -e "/^\[$section\]/ a\\
|
||
$option = $v
|
||
" "$file"
|
||
done
|
||
}
|
||
|
||
|
||
# Append a new option in an ini file without replacing the old value
|
||
# iniadd config-file section option value1 value2 value3 ...
|
||
function iniadd() {
|
||
local file=$1
|
||
local section=$2
|
||
local option=$3
|
||
shift 3
|
||
local values="$(iniget_multiline $file $section $option) $@"
|
||
iniset_multiline $file $section $option $values
|
||
}
|
||
|
||
# Find out if a process exists by partial name.
|
||
# is_running name
|
||
function is_running() {
|
||
local name=$1
|
||
ps auxw | grep -v grep | grep ${name} > /dev/null
|
||
RC=$?
|
||
# some times I really hate bash reverse binary logic
|
||
return $RC
|
||
}
|
||
|
||
|
||
# is_service_enabled() checks if the service(s) specified as arguments are
|
||
# enabled by the user in ``ENABLED_SERVICES``.
|
||
#
|
||
# Multiple services specified as arguments are ``OR``'ed together; the test
|
||
# is a short-circuit boolean, i.e it returns on the first match.
|
||
#
|
||
# There are special cases for some 'catch-all' services::
|
||
# **nova** returns true if any service enabled start with **n-**
|
||
# **cinder** returns true if any service enabled start with **c-**
|
||
# **ceilometer** returns true if any service enabled start with **ceilometer**
|
||
# **glance** returns true if any service enabled start with **g-**
|
||
# **neutron** returns true if any service enabled start with **q-**
|
||
# **swift** returns true if any service enabled start with **s-**
|
||
# **trove** returns true if any service enabled start with **tr-**
|
||
# For backward compatibility if we have **swift** in ENABLED_SERVICES all the
|
||
# **s-** services will be enabled. This will be deprecated in the future.
|
||
#
|
||
# Cells within nova is enabled if **n-cell** is in ``ENABLED_SERVICES``.
|
||
# We also need to make sure to treat **n-cell-region** and **n-cell-child**
|
||
# as enabled in this case.
|
||
#
|
||
# Uses global ``ENABLED_SERVICES``
|
||
# is_service_enabled service [service ...]
|
||
function is_service_enabled() {
|
||
services=$@
|
||
for service in ${services}; do
|
||
[[ ,${ENABLED_SERVICES}, =~ ,${service}, ]] && return 0
|
||
[[ ${service} == n-cell-* && ${ENABLED_SERVICES} =~ "n-cell" ]] && return 0
|
||
[[ ${service} == "nova" && ${ENABLED_SERVICES} =~ "n-" ]] && return 0
|
||
[[ ${service} == "cinder" && ${ENABLED_SERVICES} =~ "c-" ]] && return 0
|
||
[[ ${service} == "ceilometer" && ${ENABLED_SERVICES} =~ "ceilometer-" ]] && return 0
|
||
[[ ${service} == "glance" && ${ENABLED_SERVICES} =~ "g-" ]] && return 0
|
||
[[ ${service} == "ironic" && ${ENABLED_SERVICES} =~ "ir-" ]] && return 0
|
||
[[ ${service} == "neutron" && ${ENABLED_SERVICES} =~ "q-" ]] && return 0
|
||
[[ ${service} == "trove" && ${ENABLED_SERVICES} =~ "tr-" ]] && return 0
|
||
[[ ${service} == "swift" && ${ENABLED_SERVICES} =~ "s-" ]] && return 0
|
||
[[ ${service} == s-* && ${ENABLED_SERVICES} =~ "swift" ]] && return 0
|
||
done
|
||
return 1
|
||
}
|
||
|
||
|
||
# remove extra commas from the input string (i.e. ``ENABLED_SERVICES``)
|
||
# _cleanup_service_list service-list
|
||
function _cleanup_service_list () {
|
||
echo "$1" | sed -e '
|
||
s/,,/,/g;
|
||
s/^,//;
|
||
s/,$//
|
||
'
|
||
}
|
||
|
||
|
||
# enable_service() adds the services passed as argument to the
|
||
# ``ENABLED_SERVICES`` list, if they are not already present.
|
||
#
|
||
# For example:
|
||
# enable_service qpid
|
||
#
|
||
# This function does not know about the special cases
|
||
# for nova, glance, and neutron built into is_service_enabled().
|
||
# Uses global ``ENABLED_SERVICES``
|
||
# enable_service service [service ...]
|
||
function enable_service() {
|
||
local tmpsvcs="${ENABLED_SERVICES}"
|
||
for service in $@; do
|
||
if ! is_service_enabled $service; then
|
||
tmpsvcs+=",$service"
|
||
fi
|
||
done
|
||
ENABLED_SERVICES=$(_cleanup_service_list "$tmpsvcs")
|
||
disable_negated_services
|
||
}
|
||
|
||
|
||
# disable_service() removes the services passed as argument to the
|
||
# ``ENABLED_SERVICES`` list, if they are present.
|
||
#
|
||
# For example:
|
||
# disable_service rabbit
|
||
#
|
||
# This function does not know about the special cases
|
||
# for nova, glance, and neutron built into is_service_enabled().
|
||
# Uses global ``ENABLED_SERVICES``
|
||
# disable_service service [service ...]
|
||
function disable_service() {
|
||
local tmpsvcs=",${ENABLED_SERVICES},"
|
||
local service
|
||
for service in $@; do
|
||
if is_service_enabled $service; then
|
||
tmpsvcs=${tmpsvcs//,$service,/,}
|
||
fi
|
||
done
|
||
ENABLED_SERVICES=$(_cleanup_service_list "$tmpsvcs")
|
||
}
|
||
|
||
|
||
# disable_all_services() removes all current services
|
||
# from ``ENABLED_SERVICES`` to reset the configuration
|
||
# before a minimal installation
|
||
# Uses global ``ENABLED_SERVICES``
|
||
# disable_all_services
|
||
function disable_all_services() {
|
||
ENABLED_SERVICES=""
|
||
}
|
||
|
||
|
||
# Remove all services starting with '-'. For example, to install all default
|
||
# services except rabbit (rabbit) set in ``localrc``:
|
||
# ENABLED_SERVICES+=",-rabbit"
|
||
# Uses global ``ENABLED_SERVICES``
|
||
# disable_negated_services
|
||
function disable_negated_services() {
|
||
local tmpsvcs="${ENABLED_SERVICES}"
|
||
local service
|
||
for service in ${tmpsvcs//,/ }; do
|
||
if [[ ${service} == -* ]]; then
|
||
tmpsvcs=$(echo ${tmpsvcs}|sed -r "s/(,)?(-)?${service#-}(,)?/,/g")
|
||
fi
|
||
done
|
||
ENABLED_SERVICES=$(_cleanup_service_list "$tmpsvcs")
|
||
}
|
||
|
||
|
||
# Distro-agnostic package installer
|
||
# install_package package [package ...]
|
||
function install_package() {
|
||
if is_ubuntu; then
|
||
[[ "$NO_UPDATE_REPOS" = "True" ]] || apt_get update
|
||
NO_UPDATE_REPOS=True
|
||
|
||
apt_get install "$@"
|
||
elif is_fedora; then
|
||
yum_install "$@"
|
||
elif is_suse; then
|
||
zypper_install "$@"
|
||
else
|
||
exit_distro_not_supported "installing packages"
|
||
fi
|
||
}
|
||
|
||
|
||
# Distro-agnostic package uninstaller
|
||
# uninstall_package package [package ...]
|
||
function uninstall_package() {
|
||
if is_ubuntu; then
|
||
apt_get purge "$@"
|
||
elif is_fedora; then
|
||
sudo yum remove -y "$@"
|
||
elif is_suse; then
|
||
sudo zypper rm "$@"
|
||
else
|
||
exit_distro_not_supported "uninstalling packages"
|
||
fi
|
||
}
|
||
|
||
|
||
# Distro-agnostic function to tell if a package is installed
|
||
# is_package_installed package [package ...]
|
||
function is_package_installed() {
|
||
if [[ -z "$@" ]]; then
|
||
return 1
|
||
fi
|
||
|
||
if [[ -z "$os_PACKAGE" ]]; then
|
||
GetOSVersion
|
||
fi
|
||
|
||
if [[ "$os_PACKAGE" = "deb" ]]; then
|
||
dpkg -s "$@" > /dev/null 2> /dev/null
|
||
elif [[ "$os_PACKAGE" = "rpm" ]]; then
|
||
rpm --quiet -q "$@"
|
||
else
|
||
exit_distro_not_supported "finding if a package is installed"
|
||
fi
|
||
}
|
||
|
||
|
||
# Test if the named environment variable is set and not zero length
|
||
# is_set env-var
|
||
function is_set() {
|
||
local var=\$"$1"
|
||
eval "[ -n \"$var\" ]" # For ex.: sh -c "[ -n \"$var\" ]" would be better, but several exercises depends on this
|
||
}
|
||
|
||
|
||
# Wrapper for ``pip install`` to set cache and proxy environment variables
|
||
# Uses globals ``OFFLINE``, ``PIP_DOWNLOAD_CACHE``, ``PIP_USE_MIRRORS``,
|
||
# ``TRACK_DEPENDS``, ``*_proxy``
|
||
# pip_install package [package ...]
|
||
function pip_install {
|
||
[[ "$OFFLINE" = "True" || -z "$@" ]] && return
|
||
if [[ -z "$os_PACKAGE" ]]; then
|
||
GetOSVersion
|
||
fi
|
||
if [[ $TRACK_DEPENDS = True ]]; then
|
||
source $DEST/.venv/bin/activate
|
||
CMD_PIP=$DEST/.venv/bin/pip
|
||
SUDO_PIP="env"
|
||
else
|
||
SUDO_PIP="sudo"
|
||
CMD_PIP=$(get_pip_command)
|
||
fi
|
||
|
||
# Mirror option not needed anymore because pypi has CDN available,
|
||
# but it's useful in certain circumstances
|
||
PIP_USE_MIRRORS=${PIP_USE_MIRRORS:-False}
|
||
if [[ "$PIP_USE_MIRRORS" != "False" ]]; then
|
||
PIP_MIRROR_OPT="--use-mirrors"
|
||
fi
|
||
|
||
# pip < 1.4 has a bug where it will use an already existing build
|
||
# directory unconditionally. Say an earlier component installs
|
||
# foo v1.1; pip will have built foo's source in
|
||
# /tmp/$USER-pip-build. Even if a later component specifies foo <
|
||
# 1.1, the existing extracted build will be used and cause
|
||
# confusing errors. By creating unique build directories we avoid
|
||
# this problem. See https://github.com/pypa/pip/issues/709
|
||
local pip_build_tmp=$(mktemp --tmpdir -d pip-build.XXXXX)
|
||
|
||
$SUDO_PIP PIP_DOWNLOAD_CACHE=${PIP_DOWNLOAD_CACHE:-/var/cache/pip} \
|
||
HTTP_PROXY=$http_proxy \
|
||
HTTPS_PROXY=$https_proxy \
|
||
NO_PROXY=$no_proxy \
|
||
$CMD_PIP install --build=${pip_build_tmp} \
|
||
$PIP_MIRROR_OPT $@ \
|
||
&& $SUDO_PIP rm -rf ${pip_build_tmp}
|
||
}
|
||
|
||
|
||
# Cleanup anything from /tmp on unstack
|
||
# clean_tmp
|
||
function cleanup_tmp {
|
||
local tmp_dir=${TMPDIR:-/tmp}
|
||
|
||
# see comments in pip_install
|
||
sudo rm -rf ${tmp_dir}/pip-build.*
|
||
}
|
||
|
||
# Service wrapper to restart services
|
||
# restart_service service-name
|
||
function restart_service() {
|
||
if is_ubuntu; then
|
||
sudo /usr/sbin/service $1 restart
|
||
else
|
||
sudo /sbin/service $1 restart
|
||
fi
|
||
}
|
||
|
||
|
||
# _run_process() is designed to be backgrounded by run_process() to simulate a
|
||
# fork. It includes the dirty work of closing extra filehandles and preparing log
|
||
# files to produce the same logs as screen_it(). The log filename is derived
|
||
# from the service name and global-and-now-misnamed SCREEN_LOGDIR
|
||
# _run_process service "command-line"
|
||
function _run_process() {
|
||
local service=$1
|
||
local command="$2"
|
||
|
||
# Undo logging redirections and close the extra descriptors
|
||
exec 1>&3
|
||
exec 2>&3
|
||
exec 3>&-
|
||
exec 6>&-
|
||
|
||
if [[ -n ${SCREEN_LOGDIR} ]]; then
|
||
exec 1>&${SCREEN_LOGDIR}/screen-${1}.${CURRENT_LOG_TIME}.log 2>&1
|
||
ln -sf ${SCREEN_LOGDIR}/screen-${1}.${CURRENT_LOG_TIME}.log ${SCREEN_LOGDIR}/screen-${1}.log
|
||
|
||
# TODO(dtroyer): Hack to get stdout from the Python interpreter for the logs.
|
||
export PYTHONUNBUFFERED=1
|
||
fi
|
||
|
||
exec /bin/bash -c "$command"
|
||
die "$service exec failure: $command"
|
||
}
|
||
|
||
|
||
# run_process() launches a child process that closes all file descriptors and
|
||
# then exec's the passed in command. This is meant to duplicate the semantics
|
||
# of screen_it() without screen. PIDs are written to
|
||
# $SERVICE_DIR/$SCREEN_NAME/$service.pid
|
||
# run_process service "command-line"
|
||
function run_process() {
|
||
local service=$1
|
||
local command="$2"
|
||
|
||
# Spawn the child process
|
||
_run_process "$service" "$command" &
|
||
echo $!
|
||
}
|
||
|
||
|
||
# Helper to launch a service in a named screen
|
||
# screen_it service "command-line"
|
||
function screen_it {
|
||
SCREEN_NAME=${SCREEN_NAME:-stack}
|
||
SERVICE_DIR=${SERVICE_DIR:-${DEST}/status}
|
||
USE_SCREEN=$(trueorfalse True $USE_SCREEN)
|
||
|
||
if is_service_enabled $1; then
|
||
# Append the service to the screen rc file
|
||
screen_rc "$1" "$2"
|
||
|
||
if [[ "$USE_SCREEN" = "True" ]]; then
|
||
screen -S $SCREEN_NAME -X screen -t $1
|
||
|
||
if [[ -n ${SCREEN_LOGDIR} ]]; then
|
||
screen -S $SCREEN_NAME -p $1 -X logfile ${SCREEN_LOGDIR}/screen-${1}.${CURRENT_LOG_TIME}.log
|
||
screen -S $SCREEN_NAME -p $1 -X log on
|
||
ln -sf ${SCREEN_LOGDIR}/screen-${1}.${CURRENT_LOG_TIME}.log ${SCREEN_LOGDIR}/screen-${1}.log
|
||
fi
|
||
|
||
# sleep to allow bash to be ready to be send the command - we are
|
||
# creating a new window in screen and then sends characters, so if
|
||
# bash isn't running by the time we send the command, nothing happens
|
||
sleep 1.5
|
||
|
||
NL=`echo -ne '\015'`
|
||
screen -S $SCREEN_NAME -p $1 -X stuff "$2 || echo \"$1 failed to start\" | tee \"$SERVICE_DIR/$SCREEN_NAME/$1.failure\"$NL"
|
||
else
|
||
# Spawn directly without screen
|
||
run_process "$1" "$2" >$SERVICE_DIR/$SCREEN_NAME/$service.pid
|
||
fi
|
||
fi
|
||
}
|
||
|
||
|
||
# Screen rc file builder
|
||
# screen_rc service "command-line"
|
||
function screen_rc {
|
||
SCREEN_NAME=${SCREEN_NAME:-stack}
|
||
SCREENRC=$TOP_DIR/$SCREEN_NAME-screenrc
|
||
if [[ ! -e $SCREENRC ]]; then
|
||
# Name the screen session
|
||
echo "sessionname $SCREEN_NAME" > $SCREENRC
|
||
# Set a reasonable statusbar
|
||
echo "hardstatus alwayslastline '$SCREEN_HARDSTATUS'" >> $SCREENRC
|
||
# Some distributions override PROMPT_COMMAND for the screen terminal type - turn that off
|
||
echo "setenv PROMPT_COMMAND /bin/true" >> $SCREENRC
|
||
echo "screen -t shell bash" >> $SCREENRC
|
||
fi
|
||
# If this service doesn't already exist in the screenrc file
|
||
if ! grep $1 $SCREENRC 2>&1 > /dev/null; then
|
||
NL=`echo -ne '\015'`
|
||
echo "screen -t $1 bash" >> $SCREENRC
|
||
echo "stuff \"$2$NL\"" >> $SCREENRC
|
||
fi
|
||
}
|
||
|
||
|
||
# Helper to remove the ``*.failure`` files under ``$SERVICE_DIR/$SCREEN_NAME``.
|
||
# This is used for ``service_check`` when all the ``screen_it`` are called finished
|
||
# init_service_check
|
||
function init_service_check() {
|
||
SCREEN_NAME=${SCREEN_NAME:-stack}
|
||
SERVICE_DIR=${SERVICE_DIR:-${DEST}/status}
|
||
|
||
if [[ ! -d "$SERVICE_DIR/$SCREEN_NAME" ]]; then
|
||
mkdir -p "$SERVICE_DIR/$SCREEN_NAME"
|
||
fi
|
||
|
||
rm -f "$SERVICE_DIR/$SCREEN_NAME"/*.failure
|
||
}
|
||
|
||
|
||
# Helper to get the status of each running service
|
||
# service_check
|
||
function service_check() {
|
||
local service
|
||
local failures
|
||
SCREEN_NAME=${SCREEN_NAME:-stack}
|
||
SERVICE_DIR=${SERVICE_DIR:-${DEST}/status}
|
||
|
||
|
||
if [[ ! -d "$SERVICE_DIR/$SCREEN_NAME" ]]; then
|
||
echo "No service status directory found"
|
||
return
|
||
fi
|
||
|
||
# Check if there is any falure flag file under $SERVICE_DIR/$SCREEN_NAME
|
||
failures=`ls "$SERVICE_DIR/$SCREEN_NAME"/*.failure 2>/dev/null`
|
||
|
||
for service in $failures; do
|
||
service=`basename $service`
|
||
service=${service%.failure}
|
||
echo "Error: Service $service is not running"
|
||
done
|
||
|
||
if [ -n "$failures" ]; then
|
||
echo "More details about the above errors can be found with screen, with ./rejoin-stack.sh"
|
||
fi
|
||
}
|
||
|
||
# Returns true if the directory is on a filesystem mounted via NFS.
|
||
function is_nfs_directory() {
|
||
local mount_type=`stat -f -L -c %T $1`
|
||
test "$mount_type" == "nfs"
|
||
}
|
||
|
||
# Only run the command if the target file (the last arg) is not on an
|
||
# NFS filesystem.
|
||
function _safe_permission_operation() {
|
||
local args=( $@ )
|
||
local last
|
||
local sudo_cmd
|
||
local dir_to_check
|
||
|
||
let last="${#args[*]} - 1"
|
||
|
||
dir_to_check=${args[$last]}
|
||
if [ ! -d "$dir_to_check" ]; then
|
||
dir_to_check=`dirname "$dir_to_check"`
|
||
fi
|
||
|
||
if is_nfs_directory "$dir_to_check" ; then
|
||
return 0
|
||
fi
|
||
|
||
if [[ $TRACK_DEPENDS = True ]]; then
|
||
sudo_cmd="env"
|
||
else
|
||
sudo_cmd="sudo"
|
||
fi
|
||
|
||
$sudo_cmd $@
|
||
}
|
||
|
||
# Only change ownership of a file or directory if it is not on an NFS
|
||
# filesystem.
|
||
function safe_chown() {
|
||
_safe_permission_operation chown $@
|
||
}
|
||
|
||
# Only change permissions of a file or directory if it is not on an
|
||
# NFS filesystem.
|
||
function safe_chmod() {
|
||
_safe_permission_operation chmod $@
|
||
}
|
||
|
||
# ``pip install -e`` the package, which processes the dependencies
|
||
# using pip before running `setup.py develop`
|
||
#
|
||
# Updates the dependencies in project_dir from the
|
||
# openstack/requirements global list before installing anything.
|
||
#
|
||
# Uses globals ``TRACK_DEPENDS``, ``REQUIREMENTS_DIR``
|
||
# setup_develop directory
|
||
function setup_develop() {
|
||
local project_dir=$1
|
||
|
||
echo "cd $REQUIREMENTS_DIR; $SUDO_CMD python update.py $project_dir"
|
||
|
||
# Don't update repo if local changes exist
|
||
(cd $project_dir && git diff --quiet)
|
||
local update_requirements=$?
|
||
|
||
if [ $update_requirements -eq 0 ]; then
|
||
(cd $REQUIREMENTS_DIR; \
|
||
$SUDO_CMD python update.py $project_dir)
|
||
fi
|
||
|
||
setup_develop_no_requirements_update $project_dir
|
||
|
||
# We've just gone and possibly modified the user's source tree in an
|
||
# automated way, which is considered bad form if it's a development
|
||
# tree because we've screwed up their next git checkin. So undo it.
|
||
#
|
||
# However... there are some circumstances, like running in the gate
|
||
# where we really really want the overridden version to stick. So provide
|
||
# a variable that tells us whether or not we should UNDO the requirements
|
||
# changes (this will be set to False in the OpenStack ci gate)
|
||
if [ $UNDO_REQUIREMENTS = "True" ]; then
|
||
if [ $update_requirements -eq 0 ]; then
|
||
(cd $project_dir && git reset --hard)
|
||
fi
|
||
fi
|
||
}
|
||
|
||
# ``pip install -e`` the package, which processes the dependencies
|
||
# using pip before running `setup.py develop`
|
||
# Uses globals ``STACK_USER``
|
||
# setup_develop_no_requirements_update directory
|
||
function setup_develop_no_requirements_update() {
|
||
local project_dir=$1
|
||
|
||
pip_install -e $project_dir
|
||
# ensure that further actions can do things like setup.py sdist
|
||
safe_chown -R $STACK_USER $1/*.egg-info
|
||
}
|
||
|
||
|
||
# Service wrapper to start services
|
||
# start_service service-name
|
||
function start_service() {
|
||
if is_ubuntu; then
|
||
sudo /usr/sbin/service $1 start
|
||
else
|
||
sudo /sbin/service $1 start
|
||
fi
|
||
}
|
||
|
||
|
||
# Service wrapper to stop services
|
||
# stop_service service-name
|
||
function stop_service() {
|
||
if is_ubuntu; then
|
||
sudo /usr/sbin/service $1 stop
|
||
else
|
||
sudo /sbin/service $1 stop
|
||
fi
|
||
}
|
||
|
||
|
||
# Normalize config values to True or False
|
||
# Accepts as False: 0 no No NO false False FALSE
|
||
# Accepts as True: 1 yes Yes YES true True TRUE
|
||
# VAR=$(trueorfalse default-value test-value)
|
||
function trueorfalse() {
|
||
local default=$1
|
||
local testval=$2
|
||
|
||
[[ -z "$testval" ]] && { echo "$default"; return; }
|
||
[[ "0 no No NO false False FALSE" =~ "$testval" ]] && { echo "False"; return; }
|
||
[[ "1 yes Yes YES true True TRUE" =~ "$testval" ]] && { echo "True"; return; }
|
||
echo "$default"
|
||
}
|
||
|
||
|
||
# Retrieve an image from a URL and upload into Glance.
|
||
# Uses the following variables:
|
||
#
|
||
# - ``FILES`` must be set to the cache dir
|
||
# - ``GLANCE_HOSTPORT``
|
||
#
|
||
# upload_image image-url glance-token
|
||
function upload_image() {
|
||
local image_url=$1
|
||
local token=$2
|
||
|
||
# Create a directory for the downloaded image tarballs.
|
||
mkdir -p $FILES/images
|
||
IMAGE_FNAME=`basename "$image_url"`
|
||
if [[ $image_url != file* ]]; then
|
||
# Downloads the image (uec ami+aki style), then extracts it.
|
||
if [[ ! -f $FILES/$IMAGE_FNAME || "$(stat -c "%s" $FILES/$IMAGE_FNAME)" = "0" ]]; then
|
||
wget -c $image_url -O $FILES/$IMAGE_FNAME
|
||
if [[ $? -ne 0 ]]; then
|
||
echo "Not found: $image_url"
|
||
return
|
||
fi
|
||
fi
|
||
IMAGE="$FILES/${IMAGE_FNAME}"
|
||
else
|
||
# File based URL (RFC 1738): file://host/path
|
||
# Remote files are not considered here.
|
||
# *nix: file:///home/user/path/file
|
||
# windows: file:///C:/Documents%20and%20Settings/user/path/file
|
||
IMAGE=$(echo $image_url | sed "s/^file:\/\///g")
|
||
if [[ ! -f $IMAGE || "$(stat -c "%s" $IMAGE)" == "0" ]]; then
|
||
echo "Not found: $image_url"
|
||
return
|
||
fi
|
||
fi
|
||
|
||
# OpenVZ-format images are provided as .tar.gz, but not decompressed prior to loading
|
||
if [[ "$image_url" =~ 'openvz' ]]; then
|
||
IMAGE_NAME="${IMAGE_FNAME%.tar.gz}"
|
||
glance --os-auth-token $token --os-image-url http://$GLANCE_HOSTPORT image-create --name "$IMAGE_NAME" --is-public=True --container-format ami --disk-format ami < "${IMAGE}"
|
||
return
|
||
fi
|
||
|
||
# vmdk format images
|
||
if [[ "$image_url" =~ '.vmdk' ]]; then
|
||
IMAGE_NAME="${IMAGE_FNAME%.vmdk}"
|
||
|
||
# Before we can upload vmdk type images to glance, we need to know it's
|
||
# disk type, storage adapter, and networking adapter. These values are
|
||
# passed to glance as custom properties.
|
||
# We take these values from the vmdk file if populated. Otherwise, we use
|
||
# vmdk filename, which is expected in the following format:
|
||
#
|
||
# <name>-<disk type>;<storage adapter>;<network adapter>
|
||
#
|
||
# If the filename does not follow the above format then the vsphere
|
||
# driver will supply default values.
|
||
|
||
vmdk_adapter_type=""
|
||
vmdk_disktype=""
|
||
vmdk_net_adapter=""
|
||
|
||
# vmdk adapter type
|
||
vmdk_adapter_type="$(head -25 $IMAGE | grep -a -F -m 1 'ddb.adapterType =' $IMAGE)"
|
||
vmdk_adapter_type="${vmdk_adapter_type#*\"}"
|
||
vmdk_adapter_type="${vmdk_adapter_type%?}"
|
||
|
||
# vmdk disk type
|
||
vmdk_create_type="$(head -25 $IMAGE | grep -a -F -m 1 'createType=' $IMAGE)"
|
||
vmdk_create_type="${vmdk_create_type#*\"}"
|
||
vmdk_create_type="${vmdk_create_type%?}"
|
||
|
||
descriptor_data_pair_msg="Monolithic flat and VMFS disks "`
|
||
`"should use a descriptor-data pair."
|
||
if [[ "$vmdk_create_type" = "monolithicSparse" ]]; then
|
||
vmdk_disktype="sparse"
|
||
elif [[ "$vmdk_create_type" = "monolithicFlat" || \
|
||
"$vmdk_create_type" = "vmfs" ]]; then
|
||
# Attempt to retrieve the *-flat.vmdk
|
||
flat_fname="$(head -25 $IMAGE | grep -G 'RW\|RDONLY [0-9]+ FLAT\|VMFS' $IMAGE)"
|
||
flat_fname="${flat_fname#*\"}"
|
||
flat_fname="${flat_fname%?}"
|
||
if [[ -z "$flat_name" ]]; then
|
||
flat_fname="$IMAGE_NAME-flat.vmdk"
|
||
fi
|
||
path_len=`expr ${#image_url} - ${#IMAGE_FNAME}`
|
||
flat_url="${image_url:0:$path_len}$flat_fname"
|
||
warn $LINENO "$descriptor_data_pair_msg"`
|
||
`" Attempt to retrieve the *-flat.vmdk: $flat_url"
|
||
if [[ $flat_url != file* ]]; then
|
||
if [[ ! -f $FILES/$flat_fname || \
|
||
"$(stat -c "%s" $FILES/$flat_fname)" = "0" ]]; then
|
||
wget -c $flat_url -O $FILES/$flat_fname
|
||
if [[ $? -ne 0 ]]; then
|
||
echo "Flat disk not found: $flat_url"
|
||
flat_found=false
|
||
fi
|
||
fi
|
||
if $flat_found; then
|
||
IMAGE="$FILES/${flat_fname}"
|
||
fi
|
||
else
|
||
IMAGE=$(echo $flat_url | sed "s/^file:\/\///g")
|
||
if [[ ! -f $IMAGE || "$(stat -c "%s" $IMAGE)" == "0" ]]; then
|
||
echo "Flat disk not found: $flat_url"
|
||
flat_found=false
|
||
fi
|
||
if ! $flat_found; then
|
||
IMAGE=$(echo $image_url | sed "s/^file:\/\///g")
|
||
fi
|
||
fi
|
||
if $flat_found; then
|
||
IMAGE_NAME="${flat_fname}"
|
||
fi
|
||
vmdk_disktype="preallocated"
|
||
elif [[ -z "$vmdk_create_type" ]]; then
|
||
# *-flat.vmdk provided: attempt to retrieve the descriptor (*.vmdk)
|
||
# to retrieve appropriate metadata
|
||
if [[ ${IMAGE_NAME: -5} != "-flat" ]]; then
|
||
warn $LINENO "Expected filename suffix: '-flat'."`
|
||
`" Filename provided: ${IMAGE_NAME}"
|
||
else
|
||
descriptor_fname="${IMAGE_NAME:0:${#IMAGE_NAME} - 5}.vmdk"
|
||
path_len=`expr ${#image_url} - ${#IMAGE_FNAME}`
|
||
flat_path="${image_url:0:$path_len}"
|
||
descriptor_url=$flat_path$descriptor_fname
|
||
warn $LINENO "$descriptor_data_pair_msg"`
|
||
`" Attempt to retrieve the descriptor *.vmdk: $descriptor_url"
|
||
if [[ $flat_path != file* ]]; then
|
||
if [[ ! -f $FILES/$descriptor_fname || \
|
||
"$(stat -c "%s" $FILES/$descriptor_fname)" = "0" ]]; then
|
||
wget -c $descriptor_url -O $FILES/$descriptor_fname
|
||
if [[ $? -ne 0 ]]; then
|
||
warn $LINENO "Descriptor not found $descriptor_url"
|
||
descriptor_found=false
|
||
fi
|
||
fi
|
||
descriptor_url="$FILES/$descriptor_fname"
|
||
else
|
||
descriptor_url=$(echo $descriptor_url | sed "s/^file:\/\///g")
|
||
if [[ ! -f $descriptor_url || \
|
||
"$(stat -c "%s" $descriptor_url)" == "0" ]]; then
|
||
warn $LINENO "Descriptor not found $descriptor_url"
|
||
descriptor_found=false
|
||
fi
|
||
fi
|
||
if $descriptor_found; then
|
||
vmdk_adapter_type="$(head -25 $descriptor_url |"`
|
||
`"grep -a -F -m 1 'ddb.adapterType =' $descriptor_url)"
|
||
vmdk_adapter_type="${vmdk_adapter_type#*\"}"
|
||
vmdk_adapter_type="${vmdk_adapter_type%?}"
|
||
fi
|
||
fi
|
||
#TODO(alegendre): handle streamOptimized once supported by the VMware driver.
|
||
vmdk_disktype="preallocated"
|
||
else
|
||
#TODO(alegendre): handle streamOptimized once supported by the VMware driver.
|
||
vmdk_disktype="preallocated"
|
||
fi
|
||
|
||
# NOTE: For backwards compatibility reasons, colons may be used in place
|
||
# of semi-colons for property delimiters but they are not permitted
|
||
# characters in NTFS filesystems.
|
||
property_string=`echo "$IMAGE_NAME" | grep -oP '(?<=-)(?!.*-).+[:;].+[:;].+$'`
|
||
IFS=':;' read -a props <<< "$property_string"
|
||
vmdk_disktype="${props[0]:-$vmdk_disktype}"
|
||
vmdk_adapter_type="${props[1]:-$vmdk_adapter_type}"
|
||
vmdk_net_adapter="${props[2]:-$vmdk_net_adapter}"
|
||
|
||
glance --os-auth-token $token --os-image-url http://$GLANCE_HOSTPORT image-create --name "$IMAGE_NAME" --is-public=True --container-format bare --disk-format vmdk --property vmware_disktype="$vmdk_disktype" --property vmware_adaptertype="$vmdk_adapter_type" --property hw_vif_model="$vmdk_net_adapter" < "${IMAGE}"
|
||
return
|
||
fi
|
||
|
||
# XenServer-vhd-ovf-format images are provided as .vhd.tgz
|
||
# and should not be decompressed prior to loading
|
||
if [[ "$image_url" =~ '.vhd.tgz' ]]; then
|
||
IMAGE_NAME="${IMAGE_FNAME%.vhd.tgz}"
|
||
glance --os-auth-token $token --os-image-url http://$GLANCE_HOSTPORT image-create --name "$IMAGE_NAME" --is-public=True --container-format=ovf --disk-format=vhd < "${IMAGE}"
|
||
return
|
||
fi
|
||
|
||
# .xen-raw.tgz suggests a Xen capable raw image inside a tgz.
|
||
# and should not be decompressed prior to loading.
|
||
# Setting metadata, so PV mode is used.
|
||
if [[ "$image_url" =~ '.xen-raw.tgz' ]]; then
|
||
IMAGE_NAME="${IMAGE_FNAME%.xen-raw.tgz}"
|
||
glance \
|
||
--os-auth-token $token \
|
||
--os-image-url http://$GLANCE_HOSTPORT \
|
||
image-create \
|
||
--name "$IMAGE_NAME" --is-public=True \
|
||
--container-format=tgz --disk-format=raw \
|
||
--property vm_mode=xen < "${IMAGE}"
|
||
return
|
||
fi
|
||
|
||
KERNEL=""
|
||
RAMDISK=""
|
||
DISK_FORMAT=""
|
||
CONTAINER_FORMAT=""
|
||
UNPACK=""
|
||
case "$IMAGE_FNAME" in
|
||
*.tar.gz|*.tgz)
|
||
# Extract ami and aki files
|
||
[ "${IMAGE_FNAME%.tar.gz}" != "$IMAGE_FNAME" ] &&
|
||
IMAGE_NAME="${IMAGE_FNAME%.tar.gz}" ||
|
||
IMAGE_NAME="${IMAGE_FNAME%.tgz}"
|
||
xdir="$FILES/images/$IMAGE_NAME"
|
||
rm -Rf "$xdir";
|
||
mkdir "$xdir"
|
||
tar -zxf $FILES/$IMAGE_FNAME -C "$xdir"
|
||
KERNEL=$(for f in "$xdir/"*-vmlinuz* "$xdir/"aki-*/image; do
|
||
[ -f "$f" ] && echo "$f" && break; done; true)
|
||
RAMDISK=$(for f in "$xdir/"*-initrd* "$xdir/"ari-*/image; do
|
||
[ -f "$f" ] && echo "$f" && break; done; true)
|
||
IMAGE=$(for f in "$xdir/"*.img "$xdir/"ami-*/image; do
|
||
[ -f "$f" ] && echo "$f" && break; done; true)
|
||
if [[ -z "$IMAGE_NAME" ]]; then
|
||
IMAGE_NAME=$(basename "$IMAGE" ".img")
|
||
fi
|
||
;;
|
||
*.img)
|
||
IMAGE_NAME=$(basename "$IMAGE" ".img")
|
||
format=$(qemu-img info ${IMAGE} | awk '/^file format/ { print $3; exit }')
|
||
if [[ ",qcow2,raw,vdi,vmdk,vpc," =~ ",$format," ]]; then
|
||
DISK_FORMAT=$format
|
||
else
|
||
DISK_FORMAT=raw
|
||
fi
|
||
CONTAINER_FORMAT=bare
|
||
;;
|
||
*.img.gz)
|
||
IMAGE_NAME=$(basename "$IMAGE" ".img.gz")
|
||
DISK_FORMAT=raw
|
||
CONTAINER_FORMAT=bare
|
||
UNPACK=zcat
|
||
;;
|
||
*.qcow2)
|
||
IMAGE_NAME=$(basename "$IMAGE" ".qcow2")
|
||
DISK_FORMAT=qcow2
|
||
CONTAINER_FORMAT=bare
|
||
;;
|
||
*.iso)
|
||
IMAGE_NAME=$(basename "$IMAGE" ".iso")
|
||
DISK_FORMAT=iso
|
||
CONTAINER_FORMAT=bare
|
||
;;
|
||
*) echo "Do not know what to do with $IMAGE_FNAME"; false;;
|
||
esac
|
||
|
||
if [ "$CONTAINER_FORMAT" = "bare" ]; then
|
||
if [ "$UNPACK" = "zcat" ]; then
|
||
glance --os-auth-token $token --os-image-url http://$GLANCE_HOSTPORT image-create --name "$IMAGE_NAME" --is-public True --container-format=$CONTAINER_FORMAT --disk-format $DISK_FORMAT < <(zcat --force "${IMAGE}")
|
||
else
|
||
glance --os-auth-token $token --os-image-url http://$GLANCE_HOSTPORT image-create --name "$IMAGE_NAME" --is-public True --container-format=$CONTAINER_FORMAT --disk-format $DISK_FORMAT < "${IMAGE}"
|
||
fi
|
||
else
|
||
# Use glance client to add the kernel the root filesystem.
|
||
# We parse the results of the first upload to get the glance ID of the
|
||
# kernel for use when uploading the root filesystem.
|
||
KERNEL_ID=""; RAMDISK_ID="";
|
||
if [ -n "$KERNEL" ]; then
|
||
KERNEL_ID=$(glance --os-auth-token $token --os-image-url http://$GLANCE_HOSTPORT image-create --name "$IMAGE_NAME-kernel" --is-public True --container-format aki --disk-format aki < "$KERNEL" | grep ' id ' | get_field 2)
|
||
fi
|
||
if [ -n "$RAMDISK" ]; then
|
||
RAMDISK_ID=$(glance --os-auth-token $token --os-image-url http://$GLANCE_HOSTPORT image-create --name "$IMAGE_NAME-ramdisk" --is-public True --container-format ari --disk-format ari < "$RAMDISK" | grep ' id ' | get_field 2)
|
||
fi
|
||
glance --os-auth-token $token --os-image-url http://$GLANCE_HOSTPORT image-create --name "${IMAGE_NAME%.img}" --is-public True --container-format ami --disk-format ami ${KERNEL_ID:+--property kernel_id=$KERNEL_ID} ${RAMDISK_ID:+--property ramdisk_id=$RAMDISK_ID} < "${IMAGE}"
|
||
fi
|
||
}
|
||
|
||
|
||
# Set the database backend to use
|
||
# When called from stackrc/localrc DATABASE_BACKENDS has not been
|
||
# initialized yet, just save the configuration selection and call back later
|
||
# to validate it.
|
||
#
|
||
# ``$1`` - the name of the database backend to use (mysql, postgresql, ...)
|
||
function use_database {
|
||
if [[ -z "$DATABASE_BACKENDS" ]]; then
|
||
# No backends registered means this is likely called from ``localrc``
|
||
# This is now deprecated usage
|
||
DATABASE_TYPE=$1
|
||
DEPRECATED_TEXT="$DEPRECATED_TEXT\nThe database backend needs to be properly set in ENABLED_SERVICES; use_database is deprecated localrc\n"
|
||
else
|
||
# This should no longer get called...here for posterity
|
||
use_exclusive_service DATABASE_BACKENDS DATABASE_TYPE $1
|
||
fi
|
||
}
|
||
|
||
|
||
# Toggle enable/disable_service for services that must run exclusive of each other
|
||
# $1 The name of a variable containing a space-separated list of services
|
||
# $2 The name of a variable in which to store the enabled service's name
|
||
# $3 The name of the service to enable
|
||
function use_exclusive_service {
|
||
local options=${!1}
|
||
local selection=$3
|
||
out=$2
|
||
[ -z $selection ] || [[ ! "$options" =~ "$selection" ]] && return 1
|
||
for opt in $options;do
|
||
[[ "$opt" = "$selection" ]] && enable_service $opt || disable_service $opt
|
||
done
|
||
eval "$out=$selection"
|
||
return 0
|
||
}
|
||
|
||
|
||
# Wait for an HTTP server to start answering requests
|
||
# wait_for_service timeout url
|
||
function wait_for_service() {
|
||
local timeout=$1
|
||
local url=$2
|
||
timeout $timeout sh -c "while ! curl --noproxy '*' -s $url >/dev/null; do sleep 1; done"
|
||
}
|
||
|
||
|
||
# Wrapper for ``yum`` to set proxy environment variables
|
||
# Uses globals ``OFFLINE``, ``*_proxy``
|
||
# yum_install package [package ...]
|
||
function yum_install() {
|
||
[[ "$OFFLINE" = "True" ]] && return
|
||
local sudo="sudo"
|
||
[[ "$(id -u)" = "0" ]] && sudo="env"
|
||
$sudo http_proxy=$http_proxy https_proxy=$https_proxy \
|
||
no_proxy=$no_proxy \
|
||
yum install -y "$@"
|
||
}
|
||
|
||
|
||
# zypper wrapper to set arguments correctly
|
||
# zypper_install package [package ...]
|
||
function zypper_install() {
|
||
[[ "$OFFLINE" = "True" ]] && return
|
||
local sudo="sudo"
|
||
[[ "$(id -u)" = "0" ]] && sudo="env"
|
||
$sudo http_proxy=$http_proxy https_proxy=$https_proxy \
|
||
zypper --non-interactive install --auto-agree-with-licenses "$@"
|
||
}
|
||
|
||
|
||
# ping check
|
||
# Uses globals ``ENABLED_SERVICES``
|
||
# ping_check from-net ip boot-timeout expected
|
||
function ping_check() {
|
||
if is_service_enabled neutron; then
|
||
_ping_check_neutron "$1" $2 $3 $4
|
||
return
|
||
fi
|
||
_ping_check_novanet "$1" $2 $3 $4
|
||
}
|
||
|
||
# ping check for nova
|
||
# Uses globals ``MULTI_HOST``, ``PRIVATE_NETWORK``
|
||
function _ping_check_novanet() {
|
||
local from_net=$1
|
||
local ip=$2
|
||
local boot_timeout=$3
|
||
local expected=${4:-"True"}
|
||
local check_command=""
|
||
MULTI_HOST=`trueorfalse False $MULTI_HOST`
|
||
if [[ "$MULTI_HOST" = "True" && "$from_net" = "$PRIVATE_NETWORK_NAME" ]]; then
|
||
return
|
||
fi
|
||
if [[ "$expected" = "True" ]]; then
|
||
check_command="while ! ping -c1 -w1 $ip; do sleep 1; done"
|
||
else
|
||
check_command="while ping -c1 -w1 $ip; do sleep 1; done"
|
||
fi
|
||
if ! timeout $boot_timeout sh -c "$check_command"; then
|
||
if [[ "$expected" = "True" ]]; then
|
||
die $LINENO "[Fail] Couldn't ping server"
|
||
else
|
||
die $LINENO "[Fail] Could ping server"
|
||
fi
|
||
fi
|
||
}
|
||
|
||
# Get ip of instance
|
||
function get_instance_ip(){
|
||
local vm_id=$1
|
||
local network_name=$2
|
||
local nova_result="$(nova show $vm_id)"
|
||
local ip=$(echo "$nova_result" | grep "$network_name" | get_field 2)
|
||
if [[ $ip = "" ]];then
|
||
echo "$nova_result"
|
||
die $LINENO "[Fail] Coudn't get ipaddress of VM"
|
||
fi
|
||
echo $ip
|
||
}
|
||
|
||
# ssh check
|
||
|
||
# ssh_check net-name key-file floating-ip default-user active-timeout
|
||
function ssh_check() {
|
||
if is_service_enabled neutron; then
|
||
_ssh_check_neutron "$1" $2 $3 $4 $5
|
||
return
|
||
fi
|
||
_ssh_check_novanet "$1" $2 $3 $4 $5
|
||
}
|
||
|
||
function _ssh_check_novanet() {
|
||
local NET_NAME=$1
|
||
local KEY_FILE=$2
|
||
local FLOATING_IP=$3
|
||
local DEFAULT_INSTANCE_USER=$4
|
||
local ACTIVE_TIMEOUT=$5
|
||
local probe_cmd=""
|
||
if ! timeout $ACTIVE_TIMEOUT sh -c "while ! ssh -o StrictHostKeyChecking=no -i $KEY_FILE ${DEFAULT_INSTANCE_USER}@$FLOATING_IP echo success; do sleep 1; done"; then
|
||
die $LINENO "server didn't become ssh-able!"
|
||
fi
|
||
}
|
||
|
||
|
||
# Add a user to a group.
|
||
# add_user_to_group user group
|
||
function add_user_to_group() {
|
||
local user=$1
|
||
local group=$2
|
||
|
||
if [[ -z "$os_VENDOR" ]]; then
|
||
GetOSVersion
|
||
fi
|
||
|
||
# SLE11 and openSUSE 12.2 don't have the usual usermod
|
||
if ! is_suse || [[ "$os_VENDOR" = "openSUSE" && "$os_RELEASE" != "12.2" ]]; then
|
||
sudo usermod -a -G "$group" "$user"
|
||
else
|
||
sudo usermod -A "$group" "$user"
|
||
fi
|
||
}
|
||
|
||
|
||
# Get the path to the direcotry where python executables are installed.
|
||
# get_python_exec_prefix
|
||
function get_python_exec_prefix() {
|
||
if is_fedora || is_suse; then
|
||
echo "/usr/bin"
|
||
else
|
||
echo "/usr/local/bin"
|
||
fi
|
||
}
|
||
|
||
|
||
# Get the location of the $module-rootwrap executables, where module is cinder
|
||
# or nova.
|
||
# get_rootwrap_location module
|
||
function get_rootwrap_location() {
|
||
local module=$1
|
||
|
||
echo "$(get_python_exec_prefix)/$module-rootwrap"
|
||
}
|
||
|
||
|
||
# Get the path to the pip command.
|
||
# get_pip_command
|
||
function get_pip_command() {
|
||
which pip || which pip-python
|
||
|
||
if [ $? -ne 0 ]; then
|
||
die $LINENO "Unable to find pip; cannot continue"
|
||
fi
|
||
}
|
||
|
||
|
||
# Path permissions sanity check
|
||
# check_path_perm_sanity path
|
||
function check_path_perm_sanity() {
|
||
# Ensure no element of the path has 0700 permissions, which is very
|
||
# likely to cause issues for daemons. Inspired by default 0700
|
||
# homedir permissions on RHEL and common practice of making DEST in
|
||
# the stack user's homedir.
|
||
|
||
local real_path=$(readlink -f $1)
|
||
local rebuilt_path=""
|
||
for i in $(echo ${real_path} | tr "/" " "); do
|
||
rebuilt_path=$rebuilt_path"/"$i
|
||
|
||
if [[ $(stat -c '%a' ${rebuilt_path}) = 700 ]]; then
|
||
echo "*** DEST path element"
|
||
echo "*** ${rebuilt_path}"
|
||
echo "*** appears to have 0700 permissions."
|
||
echo "*** This is very likely to cause fatal issues for devstack daemons."
|
||
|
||
if [[ -n "$SKIP_PATH_SANITY" ]]; then
|
||
return
|
||
else
|
||
echo "*** Set SKIP_PATH_SANITY to skip this check"
|
||
die $LINENO "Invalid path permissions"
|
||
fi
|
||
fi
|
||
done
|
||
}
|
||
|
||
|
||
# This function recursively compares versions, and is not meant to be
|
||
# called by anything other than vercmp_numbers below. This function does
|
||
# not work with alphabetic versions.
|
||
#
|
||
# _vercmp_r sep ver1 ver2
|
||
function _vercmp_r {
|
||
typeset sep
|
||
typeset -a ver1=() ver2=()
|
||
sep=$1; shift
|
||
ver1=("${@:1:sep}")
|
||
ver2=("${@:sep+1}")
|
||
|
||
if ((ver1 > ver2)); then
|
||
echo 1; return 0
|
||
elif ((ver2 > ver1)); then
|
||
echo -1; return 0
|
||
fi
|
||
|
||
if ((sep <= 1)); then
|
||
echo 0; return 0
|
||
fi
|
||
|
||
_vercmp_r $((sep-1)) "${ver1[@]:1}" "${ver2[@]:1}"
|
||
}
|
||
|
||
|
||
# This function compares two versions and is meant to be called by
|
||
# external callers. Please note the function assumes non-alphabetic
|
||
# versions. For example, this will work:
|
||
#
|
||
# vercmp_numbers 1.10 1.4
|
||
#
|
||
# The above will return "1", as 1.10 is greater than 1.4.
|
||
#
|
||
# vercmp_numbers 5.2 6.4
|
||
#
|
||
# The above will return "-1", as 5.2 is less than 6.4.
|
||
#
|
||
# vercmp_numbers 4.0 4.0
|
||
#
|
||
# The above will return "0", as the versions are equal.
|
||
#
|
||
# vercmp_numbers ver1 ver2
|
||
vercmp_numbers() {
|
||
typeset v1=$1 v2=$2 sep
|
||
typeset -a ver1 ver2
|
||
|
||
IFS=. read -ra ver1 <<< "$v1"
|
||
IFS=. read -ra ver2 <<< "$v2"
|
||
|
||
_vercmp_r "${#ver1[@]}" "${ver1[@]}" "${ver2[@]}"
|
||
}
|
||
|
||
|
||
# ``policy_add policy_file policy_name policy_permissions``
|
||
#
|
||
# Add a policy to a policy.json file
|
||
# Do nothing if the policy already exists
|
||
|
||
function policy_add() {
|
||
local policy_file=$1
|
||
local policy_name=$2
|
||
local policy_perm=$3
|
||
|
||
if grep -q ${policy_name} ${policy_file}; then
|
||
echo "Policy ${policy_name} already exists in ${policy_file}"
|
||
return
|
||
fi
|
||
|
||
# Add a terminating comma to policy lines without one
|
||
# Remove the closing '}' and all lines following to the end-of-file
|
||
local tmpfile=$(mktemp)
|
||
uniq ${policy_file} | sed -e '
|
||
s/]$/],/
|
||
/^[}]/,$d
|
||
' > ${tmpfile}
|
||
|
||
# Append policy and closing brace
|
||
echo " \"${policy_name}\": ${policy_perm}" >>${tmpfile}
|
||
echo "}" >>${tmpfile}
|
||
|
||
mv ${tmpfile} ${policy_file}
|
||
}
|
||
|
||
|
||
# This function sets log formatting options for colorizing log
|
||
# output to stdout. It is meant to be called by lib modules.
|
||
# The last two parameters are optional and can be used to specify
|
||
# non-default value for project and user format variables.
|
||
# Defaults are respectively 'project_name' and 'user_name'
|
||
#
|
||
# setup_colorized_logging something.conf SOMESECTION
|
||
function setup_colorized_logging() {
|
||
local conf_file=$1
|
||
local conf_section=$2
|
||
local project_var=${3:-"project_name"}
|
||
local user_var=${4:-"user_name"}
|
||
# Add color to logging output
|
||
iniset $conf_file $conf_section logging_context_format_string "%(asctime)s.%(msecs)03d %(color)s%(levelname)s %(name)s [[01;36m%(request_id)s [00;36m%("$user_var")s %("$project_var")s%(color)s] [01;35m%(instance)s%(color)s%(message)s[00m"
|
||
iniset $conf_file $conf_section logging_default_format_string "%(asctime)s.%(msecs)03d %(color)s%(levelname)s %(name)s [[00;36m-%(color)s] [01;35m%(instance)s%(color)s%(message)s[00m"
|
||
iniset $conf_file $conf_section logging_debug_format_suffix "[00;33mfrom (pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d[00m"
|
||
iniset $conf_file $conf_section logging_exception_prefix "%(color)s%(asctime)s.%(msecs)03d TRACE %(name)s [01;35m%(instance)s[00m"
|
||
}
|
||
|
||
# Restore xtrace
|
||
$XTRACE
|
||
|
||
|
||
# Local variables:
|
||
# mode: shell-script
|
||
# End:
|