 ea23e68664
			
		
	
	ea23e68664
	
	
	
		
			
			Added quantum_v2.sh for quantum v2 api exercise. Quantum exercise script requires the following settings: - Q_USE_NAMESPACE=False - LIBVIRT_FIREWALL_DRIVER=nova.virt.firewall.NoopFirewallDriver And the following services enabled: - quantum, q-svc, q-agt, q-dhcp Change-Id: I5767f94c94187a4ca0fd189244fa6f5781519ab4
		
			
				
	
	
		
			2224 lines
		
	
	
		
			85 KiB
		
	
	
	
		
			Bash
		
	
	
		
			Executable File
		
	
	
	
	
			
		
		
	
	
			2224 lines
		
	
	
		
			85 KiB
		
	
	
	
		
			Bash
		
	
	
		
			Executable File
		
	
	
	
	
| #!/usr/bin/env bash
 | ||
| 
 | ||
| # ``stack.sh`` is an opinionated OpenStack developer installation.  It
 | ||
| # installs and configures various combinations of **Glance**, **Horizon**,
 | ||
| # **Keystone**, **Nova**, **Quantum** and **Swift**
 | ||
| 
 | ||
| # This script allows you to specify configuration options of what git
 | ||
| # repositories to use, enabled services, network configuration and various
 | ||
| # passwords.  If you are crafty you can run the script on multiple nodes using
 | ||
| # shared settings for common resources (mysql, rabbitmq) and build a multi-node
 | ||
| # developer install.
 | ||
| 
 | ||
| # To keep this script simple we assume you are running on an **Ubuntu 11.10
 | ||
| # Oneiric** or **Ubuntu 12.04 Precise** machine.  It should work in a VM or
 | ||
| # physical server.  Additionally we put the list of ``apt`` and ``pip``
 | ||
| # dependencies and other configuration files in this repo.  So start by
 | ||
| # grabbing this script and the dependencies.
 | ||
| 
 | ||
| # Learn more and get the most recent version at http://devstack.org
 | ||
| 
 | ||
| # Keep track of the devstack directory
 | ||
| TOP_DIR=$(cd $(dirname "$0") && pwd)
 | ||
| 
 | ||
| # Import common functions
 | ||
| source $TOP_DIR/functions
 | ||
| 
 | ||
| # Determine what system we are running on.  This provides ``os_VENDOR``,
 | ||
| # ``os_RELEASE``, ``os_UPDATE``, ``os_PACKAGE``, ``os_CODENAME``
 | ||
| # and ``DISTRO``
 | ||
| GetDistro
 | ||
| 
 | ||
| 
 | ||
| # Settings
 | ||
| # ========
 | ||
| 
 | ||
| # ``stack.sh`` is customizable through setting environment variables.  If you
 | ||
| # want to override a setting you can set and export it::
 | ||
| #
 | ||
| #     export MYSQL_PASSWORD=anothersecret
 | ||
| #     ./stack.sh
 | ||
| #
 | ||
| # You can also pass options on a single line ``MYSQL_PASSWORD=simple ./stack.sh``
 | ||
| #
 | ||
| # Additionally, you can put any local variables into a ``localrc`` file::
 | ||
| #
 | ||
| #     MYSQL_PASSWORD=anothersecret
 | ||
| #     MYSQL_USER=hellaroot
 | ||
| #
 | ||
| # We try to have sensible defaults, so you should be able to run ``./stack.sh``
 | ||
| # in most cases.
 | ||
| #
 | ||
| # DevStack distributes ``stackrc`` which contains locations for the OpenStack
 | ||
| # repositories and branches to configure.  ``stackrc`` sources ``localrc`` to
 | ||
| # allow you to safely override those settings without being overwritten
 | ||
| # when updating DevStack.
 | ||
| if [[ ! -r $TOP_DIR/stackrc ]]; then
 | ||
|     echo "ERROR: missing $TOP_DIR/stackrc - did you grab more than just stack.sh?"
 | ||
|     exit 1
 | ||
| fi
 | ||
| source $TOP_DIR/stackrc
 | ||
| 
 | ||
| # HTTP and HTTPS proxy servers are supported via the usual environment variables
 | ||
| # ``http_proxy`` and ``https_proxy``.  Additionally if you would like to access
 | ||
| # to specific server directly and not through the proxy server, you can use
 | ||
| # ``no_proxy`` environment variable.  They can be set in ``localrc`` if necessary
 | ||
| # or on the command line::
 | ||
| #
 | ||
| #     http_proxy=http://proxy.example.com:3128/ no_proxy=repo.example.net ./stack.sh
 | ||
| if [[ -n "$http_proxy" ]]; then
 | ||
|     export http_proxy=$http_proxy
 | ||
| fi
 | ||
| if [[ -n "$https_proxy" ]]; then
 | ||
|     export https_proxy=$https_proxy
 | ||
| fi
 | ||
| if [[ -n "$no_proxy" ]]; then
 | ||
|     export no_proxy=$no_proxy
 | ||
| fi
 | ||
| 
 | ||
| # Destination path for installation ``DEST``
 | ||
| DEST=${DEST:-/opt/stack}
 | ||
| 
 | ||
| 
 | ||
| # Sanity Check
 | ||
| # ============
 | ||
| 
 | ||
| # Remove services which were negated in ENABLED_SERVICES
 | ||
| # using the "-" prefix (e.g., "-n-vol") instead of
 | ||
| # calling disable_service().
 | ||
| disable_negated_services
 | ||
| 
 | ||
| # Warn users who aren't on an explicitly supported distro, but allow them to
 | ||
| # override check and attempt installation with ``FORCE=yes ./stack``
 | ||
| if [[ ! ${DISTRO} =~ (oneiric|precise|quantal|f16|f17) ]]; then
 | ||
|     echo "WARNING: this script has not been tested on $DISTRO"
 | ||
|     if [[ "$FORCE" != "yes" ]]; then
 | ||
|         echo "If you wish to run this script anyway run with FORCE=yes"
 | ||
|         exit 1
 | ||
|     fi
 | ||
| fi
 | ||
| 
 | ||
| if [ "${DISTRO}" = "oneiric" ] && is_service_enabled qpid ; then
 | ||
|     # Qpid was introduced in precise
 | ||
|     echo "You must use Ubuntu Precise or newer for Qpid support."
 | ||
|     exit 1
 | ||
| fi
 | ||
| 
 | ||
| # Set the paths of certain binaries
 | ||
| if [[ "$os_PACKAGE" = "deb" ]]; then
 | ||
|     NOVA_ROOTWRAP=/usr/local/bin/nova-rootwrap
 | ||
| else
 | ||
|     NOVA_ROOTWRAP=/usr/bin/nova-rootwrap
 | ||
| fi
 | ||
| 
 | ||
| # ``stack.sh`` keeps function libraries here
 | ||
| # Make sure ``$TOP_DIR/lib`` directory is present
 | ||
| if [ ! -d $TOP_DIR/lib ]; then
 | ||
|     echo "ERROR: missing devstack/lib - did you grab more than just stack.sh?"
 | ||
|     exit 1
 | ||
| fi
 | ||
| 
 | ||
| # stack.sh keeps the list of ``apt`` and ``pip`` dependencies in external
 | ||
| # files, along with config templates and other useful files.  You can find these
 | ||
| # in the ``files`` directory (next to this script).  We will reference this
 | ||
| # directory using the ``FILES`` variable in this script.
 | ||
| FILES=$TOP_DIR/files
 | ||
| if [ ! -d $FILES ]; then
 | ||
|     echo "ERROR: missing devstack/files - did you grab more than just stack.sh?"
 | ||
|     exit 1
 | ||
| fi
 | ||
| 
 | ||
| # Check to see if we are already running DevStack
 | ||
| if type -p screen >/dev/null && screen -ls | egrep -q "[0-9].stack"; then
 | ||
|     echo "You are already running a stack.sh session."
 | ||
|     echo "To rejoin this session type 'screen -x stack'."
 | ||
|     echo "To destroy this session, kill the running screen."
 | ||
|     exit 1
 | ||
| fi
 | ||
| 
 | ||
| # Make sure we only have one volume service enabled.
 | ||
| if is_service_enabled cinder && is_service_enabled n-vol; then
 | ||
|     echo "ERROR: n-vol and cinder must not be enabled at the same time"
 | ||
|     exit 1
 | ||
| fi
 | ||
| 
 | ||
| # OpenStack is designed to be run as a regular user (Horizon will fail to run
 | ||
| # as root, since apache refused to startup serve content from root user).  If
 | ||
| # ``stack.sh`` is run as **root**, it automatically creates a **stack** user with
 | ||
| # sudo privileges and runs as that user.
 | ||
| 
 | ||
| if [[ $EUID -eq 0 ]]; then
 | ||
|     ROOTSLEEP=${ROOTSLEEP:-10}
 | ||
|     echo "You are running this script as root."
 | ||
|     echo "In $ROOTSLEEP seconds, we will create a user 'stack' and run as that user"
 | ||
|     sleep $ROOTSLEEP
 | ||
| 
 | ||
|     # since this script runs as a normal user, we need to give that user
 | ||
|     # ability to run sudo
 | ||
|     if [[ "$os_PACKAGE" = "deb" ]]; then
 | ||
|         dpkg -l sudo || apt_get update && install_package sudo
 | ||
|     else
 | ||
|         rpm -qa | grep sudo || install_package sudo
 | ||
|     fi
 | ||
|     if ! getent group stack >/dev/null; then
 | ||
|         echo "Creating a group called stack"
 | ||
|         groupadd stack
 | ||
|     fi
 | ||
|     if ! getent passwd stack >/dev/null; then
 | ||
|         echo "Creating a user called stack"
 | ||
|         useradd -g stack -s /bin/bash -d $DEST -m stack
 | ||
|     fi
 | ||
| 
 | ||
|     echo "Giving stack user passwordless sudo priviledges"
 | ||
|     # some uec images sudoers does not have a '#includedir'. add one.
 | ||
|     grep -q "^#includedir.*/etc/sudoers.d" /etc/sudoers ||
 | ||
|         echo "#includedir /etc/sudoers.d" >> /etc/sudoers
 | ||
|     ( umask 226 && echo "stack ALL=(ALL) NOPASSWD:ALL" \
 | ||
|         > /etc/sudoers.d/50_stack_sh )
 | ||
| 
 | ||
|     echo "Copying files to stack user"
 | ||
|     STACK_DIR="$DEST/${PWD##*/}"
 | ||
|     cp -r -f -T "$PWD" "$STACK_DIR"
 | ||
|     chown -R stack "$STACK_DIR"
 | ||
|     if [[ "$SHELL_AFTER_RUN" != "no" ]]; then
 | ||
|         exec su -c "set -e; cd $STACK_DIR; bash stack.sh; bash" stack
 | ||
|     else
 | ||
|         exec su -c "set -e; cd $STACK_DIR; bash stack.sh" stack
 | ||
|     fi
 | ||
|     exit 1
 | ||
| else
 | ||
|     # We're not root, make sure sudo is available
 | ||
|     if [[ "$os_PACKAGE" = "deb" ]]; then
 | ||
|         CHECK_SUDO_CMD="dpkg -l sudo"
 | ||
|     else
 | ||
|         CHECK_SUDO_CMD="rpm -q sudo"
 | ||
|     fi
 | ||
|     $CHECK_SUDO_CMD || die "Sudo is required.  Re-run stack.sh as root ONE TIME ONLY to set up sudo."
 | ||
| 
 | ||
|     # UEC images /etc/sudoers does not have a '#includedir'. add one.
 | ||
|     sudo grep -q "^#includedir.*/etc/sudoers.d" /etc/sudoers ||
 | ||
|         echo "#includedir /etc/sudoers.d" | sudo tee -a /etc/sudoers
 | ||
| 
 | ||
|     # Set up devstack sudoers
 | ||
|     TEMPFILE=`mktemp`
 | ||
|     echo "`whoami` ALL=(root) NOPASSWD:ALL" >$TEMPFILE
 | ||
|     # Some binaries might be under /sbin or /usr/sbin, so make sure sudo will
 | ||
|     # see them by forcing PATH
 | ||
|     echo "Defaults:`whoami` secure_path=/sbin:/usr/sbin:/usr/bin:/bin:/usr/local/sbin:/usr/local/bin" >> $TEMPFILE
 | ||
|     chmod 0440 $TEMPFILE
 | ||
|     sudo chown root:root $TEMPFILE
 | ||
|     sudo mv $TEMPFILE /etc/sudoers.d/50_stack_sh
 | ||
| 
 | ||
|     # Remove old file
 | ||
|     sudo rm -f /etc/sudoers.d/stack_sh_nova
 | ||
| fi
 | ||
| 
 | ||
| # Create the destination directory and ensure it is writable by the user
 | ||
| sudo mkdir -p $DEST
 | ||
| if [ ! -w $DEST ]; then
 | ||
|     sudo chown `whoami` $DEST
 | ||
| fi
 | ||
| 
 | ||
| # Set True to configure ``stack.sh`` to run cleanly without Internet access.
 | ||
| # ``stack.sh`` must have been previously run with Internet access to install
 | ||
| # prerequisites and initialize ``$DEST``.
 | ||
| OFFLINE=`trueorfalse False $OFFLINE`
 | ||
| 
 | ||
| # Set True to configure ``stack.sh`` to exit with an error code if it is asked
 | ||
| # to clone any git repositories.  If devstack is used in a testing environment,
 | ||
| # this may be used to ensure that the correct code is being tested.
 | ||
| ERROR_ON_CLONE=`trueorfalse False $ERROR_ON_CLONE`
 | ||
| 
 | ||
| # Destination path for service data
 | ||
| DATA_DIR=${DATA_DIR:-${DEST}/data}
 | ||
| sudo mkdir -p $DATA_DIR
 | ||
| sudo chown `whoami` $DATA_DIR
 | ||
| 
 | ||
| 
 | ||
| # Projects
 | ||
| # --------
 | ||
| 
 | ||
| # Get project function libraries
 | ||
| source $TOP_DIR/lib/cinder
 | ||
| source $TOP_DIR/lib/ceilometer
 | ||
| 
 | ||
| # Set the destination directories for openstack projects
 | ||
| NOVA_DIR=$DEST/nova
 | ||
| HORIZON_DIR=$DEST/horizon
 | ||
| GLANCE_DIR=$DEST/glance
 | ||
| GLANCECLIENT_DIR=$DEST/python-glanceclient
 | ||
| KEYSTONE_DIR=$DEST/keystone
 | ||
| NOVACLIENT_DIR=$DEST/python-novaclient
 | ||
| KEYSTONECLIENT_DIR=$DEST/python-keystoneclient
 | ||
| OPENSTACKCLIENT_DIR=$DEST/python-openstackclient
 | ||
| NOVNC_DIR=$DEST/noVNC
 | ||
| SWIFT_DIR=$DEST/swift
 | ||
| SWIFT3_DIR=$DEST/swift3
 | ||
| SWIFTCLIENT_DIR=$DEST/python-swiftclient
 | ||
| QUANTUM_DIR=$DEST/quantum
 | ||
| QUANTUM_CLIENT_DIR=$DEST/python-quantumclient
 | ||
| 
 | ||
| # Default Quantum Plugin
 | ||
| Q_PLUGIN=${Q_PLUGIN:-openvswitch}
 | ||
| # Default Quantum Port
 | ||
| Q_PORT=${Q_PORT:-9696}
 | ||
| # Default Quantum Host
 | ||
| Q_HOST=${Q_HOST:-localhost}
 | ||
| # Which Quantum API nova should use
 | ||
| # Default admin username
 | ||
| Q_ADMIN_USERNAME=${Q_ADMIN_USERNAME:-quantum}
 | ||
| # Default auth strategy
 | ||
| Q_AUTH_STRATEGY=${Q_AUTH_STRATEGY:-keystone}
 | ||
| # Use namespace or not
 | ||
| Q_USE_NAMESPACE=${Q_USE_NAMESPACE:-True}
 | ||
| 
 | ||
| # Name of the lvm volume group to use/create for iscsi volumes
 | ||
| VOLUME_GROUP=${VOLUME_GROUP:-stack-volumes}
 | ||
| VOLUME_NAME_PREFIX=${VOLUME_NAME_PREFIX:-volume-}
 | ||
| INSTANCE_NAME_PREFIX=${INSTANCE_NAME_PREFIX:-instance-}
 | ||
| 
 | ||
| # Nova supports pluggable schedulers.  ``FilterScheduler`` should work in most
 | ||
| # cases.
 | ||
| SCHEDULER=${SCHEDULER:-nova.scheduler.filter_scheduler.FilterScheduler}
 | ||
| 
 | ||
| # Set fixed and floating range here so we can make sure not to use addresses
 | ||
| # from either range when attempting to guess the ip to use for the host
 | ||
| FIXED_RANGE=${FIXED_RANGE:-10.0.0.0/24}
 | ||
| FLOATING_RANGE=${FLOATING_RANGE:-172.24.4.224/28}
 | ||
| 
 | ||
| # Find the interface used for the default route
 | ||
| HOST_IP_IFACE=${HOST_IP_IFACE:-$(ip route | sed -n '/^default/{ s/.*dev \(\w\+\)\s\+.*/\1/; p; }')}
 | ||
| # Search for an IP unless an explicit is set by ``HOST_IP`` environment variable
 | ||
| if [ -z "$HOST_IP" -o "$HOST_IP" == "dhcp" ]; then
 | ||
|     HOST_IP=""
 | ||
|     HOST_IPS=`LC_ALL=C ip -f inet addr show ${HOST_IP_IFACE} | awk '/inet/ {split($2,parts,"/");  print parts[1]}'`
 | ||
|     for IP in $HOST_IPS; do
 | ||
|         # Attempt to filter out ip addresses that are part of the fixed and
 | ||
|         # floating range. Note that this method only works if the 'netaddr'
 | ||
|         # python library is installed. If it is not installed, an error
 | ||
|         # will be printed and the first ip from the interface will be used.
 | ||
|         if ! (address_in_net $IP $FIXED_RANGE || address_in_net $IP $FLOATING_RANGE); then
 | ||
|             HOST_IP=$IP
 | ||
|             break;
 | ||
|         fi
 | ||
|     done
 | ||
|     if [ "$HOST_IP" == "" ]; then
 | ||
|         echo "Could not determine host ip address."
 | ||
|         echo "Either localrc specified dhcp on ${HOST_IP_IFACE} or defaulted"
 | ||
|         exit 1
 | ||
|     fi
 | ||
| fi
 | ||
| 
 | ||
| # Allow the use of an alternate hostname (such as localhost/127.0.0.1) for service endpoints.
 | ||
| SERVICE_HOST=${SERVICE_HOST:-$HOST_IP}
 | ||
| 
 | ||
| # Configure services to use syslog instead of writing to individual log files
 | ||
| SYSLOG=`trueorfalse False $SYSLOG`
 | ||
| SYSLOG_HOST=${SYSLOG_HOST:-$HOST_IP}
 | ||
| SYSLOG_PORT=${SYSLOG_PORT:-516}
 | ||
| 
 | ||
| # Use color for logging output
 | ||
| LOG_COLOR=`trueorfalse True $LOG_COLOR`
 | ||
| 
 | ||
| # Service startup timeout
 | ||
| SERVICE_TIMEOUT=${SERVICE_TIMEOUT:-60}
 | ||
| 
 | ||
| # Generic helper to configure passwords
 | ||
| function read_password {
 | ||
|     set +o xtrace
 | ||
|     var=$1; msg=$2
 | ||
|     pw=${!var}
 | ||
| 
 | ||
|     localrc=$TOP_DIR/localrc
 | ||
| 
 | ||
|     # If the password is not defined yet, proceed to prompt user for a password.
 | ||
|     if [ ! $pw ]; then
 | ||
|         # If there is no localrc file, create one
 | ||
|         if [ ! -e $localrc ]; then
 | ||
|             touch $localrc
 | ||
|         fi
 | ||
| 
 | ||
|         # Presumably if we got this far it can only be that our localrc is missing
 | ||
|         # the required password.  Prompt user for a password and write to localrc.
 | ||
|         echo ''
 | ||
|         echo '################################################################################'
 | ||
|         echo $msg
 | ||
|         echo '################################################################################'
 | ||
|         echo "This value will be written to your localrc file so you don't have to enter it "
 | ||
|         echo "again.  Use only alphanumeric characters."
 | ||
|         echo "If you leave this blank, a random default value will be used."
 | ||
|         pw=" "
 | ||
|         while true; do
 | ||
|             echo "Enter a password now:"
 | ||
|             read -e $var
 | ||
|             pw=${!var}
 | ||
|             [[ "$pw" = "`echo $pw | tr -cd [:alnum:]`" ]] && break
 | ||
|             echo "Invalid chars in password.  Try again:"
 | ||
|         done
 | ||
|         if [ ! $pw ]; then
 | ||
|             pw=`openssl rand -hex 10`
 | ||
|         fi
 | ||
|         eval "$var=$pw"
 | ||
|         echo "$var=$pw" >> $localrc
 | ||
|     fi
 | ||
|     set -o xtrace
 | ||
| }
 | ||
| 
 | ||
| 
 | ||
| # Nova Network Configuration
 | ||
| # --------------------------
 | ||
| 
 | ||
| # FIXME: more documentation about why these are important options.  Also
 | ||
| # we should make sure we use the same variable names as the option names.
 | ||
| 
 | ||
| if [ "$VIRT_DRIVER" = 'xenserver' ]; then
 | ||
|     PUBLIC_INTERFACE_DEFAULT=eth3
 | ||
|     # allow build_domU.sh to specify the flat network bridge via kernel args
 | ||
|     FLAT_NETWORK_BRIDGE_DEFAULT=$(grep -o 'flat_network_bridge=[[:alnum:]]*' /proc/cmdline | cut -d= -f 2 | sort -u)
 | ||
|     GUEST_INTERFACE_DEFAULT=eth1
 | ||
| else
 | ||
|     PUBLIC_INTERFACE_DEFAULT=br100
 | ||
|     FLAT_NETWORK_BRIDGE_DEFAULT=br100
 | ||
|     GUEST_INTERFACE_DEFAULT=eth0
 | ||
| fi
 | ||
| 
 | ||
| PUBLIC_INTERFACE=${PUBLIC_INTERFACE:-$PUBLIC_INTERFACE_DEFAULT}
 | ||
| FIXED_NETWORK_SIZE=${FIXED_NETWORK_SIZE:-256}
 | ||
| NETWORK_GATEWAY=${NETWORK_GATEWAY:-10.0.0.1}
 | ||
| NET_MAN=${NET_MAN:-FlatDHCPManager}
 | ||
| EC2_DMZ_HOST=${EC2_DMZ_HOST:-$SERVICE_HOST}
 | ||
| FLAT_NETWORK_BRIDGE=${FLAT_NETWORK_BRIDGE:-$FLAT_NETWORK_BRIDGE_DEFAULT}
 | ||
| VLAN_INTERFACE=${VLAN_INTERFACE:-$GUEST_INTERFACE_DEFAULT}
 | ||
| 
 | ||
| # Test floating pool and range are used for testing.  They are defined
 | ||
| # here until the admin APIs can replace nova-manage
 | ||
| TEST_FLOATING_POOL=${TEST_FLOATING_POOL:-test}
 | ||
| TEST_FLOATING_RANGE=${TEST_FLOATING_RANGE:-192.168.253.0/29}
 | ||
| 
 | ||
| # **MULTI_HOST** is a mode where each compute node runs its own network node.  This
 | ||
| # allows network operations and routing for a VM to occur on the server that is
 | ||
| # running the VM - removing a SPOF and bandwidth bottleneck.
 | ||
| MULTI_HOST=`trueorfalse False $MULTI_HOST`
 | ||
| 
 | ||
| # If you are using FlatDHCP on multiple hosts, set the ``FLAT_INTERFACE``
 | ||
| # variable but make sure that the interface doesn't already have an
 | ||
| # ip or you risk breaking things.
 | ||
| #
 | ||
| # **DHCP Warning**:  If your flat interface device uses DHCP, there will be a
 | ||
| # hiccup while the network is moved from the flat interface to the flat network
 | ||
| # bridge.  This will happen when you launch your first instance.  Upon launch
 | ||
| # you will lose all connectivity to the node, and the vm launch will probably
 | ||
| # fail.
 | ||
| #
 | ||
| # If you are running on a single node and don't need to access the VMs from
 | ||
| # devices other than that node, you can set the flat interface to the same
 | ||
| # value as ``FLAT_NETWORK_BRIDGE``.  This will stop the network hiccup from
 | ||
| # occurring.
 | ||
| FLAT_INTERFACE=${FLAT_INTERFACE:-$GUEST_INTERFACE_DEFAULT}
 | ||
| 
 | ||
| ## FIXME(ja): should/can we check that FLAT_INTERFACE is sane?
 | ||
| 
 | ||
| # Using Quantum networking:
 | ||
| #
 | ||
| # Make sure that quantum is enabled in ENABLED_SERVICES.  If it is the network
 | ||
| # manager will be set to the QuantumManager.  If you want to run Quantum on
 | ||
| # this host, make sure that q-svc is also in ENABLED_SERVICES.
 | ||
| #
 | ||
| # If you're planning to use the Quantum openvswitch plugin, set Q_PLUGIN to
 | ||
| # "openvswitch" and make sure the q-agt service is enabled in
 | ||
| # ENABLED_SERVICES.
 | ||
| #
 | ||
| # With Quantum networking the NET_MAN variable is ignored.
 | ||
| 
 | ||
| # MySQL & (RabbitMQ or Qpid)
 | ||
| # --------------------------
 | ||
| 
 | ||
| # We configure Nova, Horizon, Glance and Keystone to use MySQL as their
 | ||
| # database server.  While they share a single server, each has their own
 | ||
| # database and tables.
 | ||
| 
 | ||
| # By default this script will install and configure MySQL.  If you want to
 | ||
| # use an existing server, you can pass in the user/password/host parameters.
 | ||
| # You will need to send the same ``MYSQL_PASSWORD`` to every host if you are doing
 | ||
| # a multi-node DevStack installation.
 | ||
| MYSQL_HOST=${MYSQL_HOST:-localhost}
 | ||
| MYSQL_USER=${MYSQL_USER:-root}
 | ||
| read_password MYSQL_PASSWORD "ENTER A PASSWORD TO USE FOR MYSQL."
 | ||
| 
 | ||
| # NOTE: Don't specify /db in this string so we can use it for multiple services
 | ||
| BASE_SQL_CONN=${BASE_SQL_CONN:-mysql://$MYSQL_USER:$MYSQL_PASSWORD@$MYSQL_HOST}
 | ||
| 
 | ||
| # Rabbit connection info
 | ||
| if is_service_enabled rabbit; then
 | ||
|     RABBIT_HOST=${RABBIT_HOST:-localhost}
 | ||
|     read_password RABBIT_PASSWORD "ENTER A PASSWORD TO USE FOR RABBIT."
 | ||
| fi
 | ||
| 
 | ||
| # Glance connection info.  Note the port must be specified.
 | ||
| GLANCE_HOSTPORT=${GLANCE_HOSTPORT:-$SERVICE_HOST:9292}
 | ||
| 
 | ||
| 
 | ||
| # Swift
 | ||
| # -----
 | ||
| 
 | ||
| # TODO: add logging to different location.
 | ||
| 
 | ||
| # By default the location of swift drives and objects is located inside
 | ||
| # the swift source directory. SWIFT_DATA_DIR variable allow you to redefine
 | ||
| # this.
 | ||
| SWIFT_DATA_DIR=${SWIFT_DATA_DIR:-${DEST}/data/swift}
 | ||
| 
 | ||
| # We are going to have the configuration files inside the source
 | ||
| # directory, change SWIFT_CONFIG_DIR if you want to adjust that.
 | ||
| SWIFT_CONFIG_DIR=${SWIFT_CONFIG_DIR:-/etc/swift}
 | ||
| 
 | ||
| # DevStack will create a loop-back disk formatted as XFS to store the
 | ||
| # swift data. By default the disk size is 1 gigabyte. The variable
 | ||
| # SWIFT_LOOPBACK_DISK_SIZE specified in bytes allow you to change
 | ||
| # that.
 | ||
| SWIFT_LOOPBACK_DISK_SIZE=${SWIFT_LOOPBACK_DISK_SIZE:-1000000}
 | ||
| 
 | ||
| # The ring uses a configurable number of bits from a path’s MD5 hash as
 | ||
| # a partition index that designates a device. The number of bits kept
 | ||
| # from the hash is known as the partition power, and 2 to the partition
 | ||
| # power indicates the partition count. Partitioning the full MD5 hash
 | ||
| # ring allows other parts of the cluster to work in batches of items at
 | ||
| # once which ends up either more efficient or at least less complex than
 | ||
| # working with each item separately or the entire cluster all at once.
 | ||
| # By default we define 9 for the partition count (which mean 512).
 | ||
| SWIFT_PARTITION_POWER_SIZE=${SWIFT_PARTITION_POWER_SIZE:-9}
 | ||
| 
 | ||
| # This variable allows you to configure how many replicas you want to be
 | ||
| # configured for your Swift cluster.  By default the three replicas would need a
 | ||
| # bit of IO and Memory on a VM you may want to lower that to 1 if you want to do
 | ||
| # only some quick testing.
 | ||
| SWIFT_REPLICAS=${SWIFT_REPLICAS:-3}
 | ||
| 
 | ||
| if is_service_enabled swift; then
 | ||
|     # If we are using swift3, we can default the s3 port to swift instead
 | ||
|     # of nova-objectstore
 | ||
|     if is_service_enabled swift3;then
 | ||
|         S3_SERVICE_PORT=${S3_SERVICE_PORT:-8080}
 | ||
|     fi
 | ||
|     # We only ask for Swift Hash if we have enabled swift service.
 | ||
|     # SWIFT_HASH is a random unique string for a swift cluster that
 | ||
|     # can never change.
 | ||
|     read_password SWIFT_HASH "ENTER A RANDOM SWIFT HASH."
 | ||
| fi
 | ||
| 
 | ||
| # Set default port for nova-objectstore
 | ||
| S3_SERVICE_PORT=${S3_SERVICE_PORT:-3333}
 | ||
| 
 | ||
| 
 | ||
| # Keystone
 | ||
| # --------
 | ||
| 
 | ||
| # Service Token - Openstack components need to have an admin token
 | ||
| # to validate user tokens.
 | ||
| read_password SERVICE_TOKEN "ENTER A SERVICE_TOKEN TO USE FOR THE SERVICE ADMIN TOKEN."
 | ||
| # Services authenticate to Identity with servicename/SERVICE_PASSWORD
 | ||
| read_password SERVICE_PASSWORD "ENTER A SERVICE_PASSWORD TO USE FOR THE SERVICE AUTHENTICATION."
 | ||
| # Horizon currently truncates usernames and passwords at 20 characters
 | ||
| read_password ADMIN_PASSWORD "ENTER A PASSWORD TO USE FOR HORIZON AND KEYSTONE (20 CHARS OR LESS)."
 | ||
| 
 | ||
| # Set the tenant for service accounts in Keystone
 | ||
| SERVICE_TENANT_NAME=${SERVICE_TENANT_NAME:-service}
 | ||
| 
 | ||
| # Set Keystone interface configuration
 | ||
| KEYSTONE_API_PORT=${KEYSTONE_API_PORT:-5000}
 | ||
| KEYSTONE_AUTH_HOST=${KEYSTONE_AUTH_HOST:-$SERVICE_HOST}
 | ||
| KEYSTONE_AUTH_PORT=${KEYSTONE_AUTH_PORT:-35357}
 | ||
| KEYSTONE_AUTH_PROTOCOL=${KEYSTONE_AUTH_PROTOCOL:-http}
 | ||
| KEYSTONE_SERVICE_HOST=${KEYSTONE_SERVICE_HOST:-$SERVICE_HOST}
 | ||
| KEYSTONE_SERVICE_PORT=${KEYSTONE_SERVICE_PORT:-5000}
 | ||
| KEYSTONE_SERVICE_PROTOCOL=${KEYSTONE_SERVICE_PROTOCOL:-http}
 | ||
| 
 | ||
| 
 | ||
| # Horizon
 | ||
| # -------
 | ||
| 
 | ||
| # Allow overriding the default Apache user and group, default both to
 | ||
| # current user.
 | ||
| APACHE_USER=${APACHE_USER:-$USER}
 | ||
| APACHE_GROUP=${APACHE_GROUP:-$APACHE_USER}
 | ||
| 
 | ||
| 
 | ||
| # Log files
 | ||
| # ---------
 | ||
| 
 | ||
| # Set up logging for stack.sh
 | ||
| # Set LOGFILE to turn on logging
 | ||
| # We append '.xxxxxxxx' to the given name to maintain history
 | ||
| # where xxxxxxxx is a representation of the date the file was created
 | ||
| if [[ -n "$LOGFILE" || -n "$SCREEN_LOGDIR" ]]; then
 | ||
|     LOGDAYS=${LOGDAYS:-7}
 | ||
|     TIMESTAMP_FORMAT=${TIMESTAMP_FORMAT:-"%F-%H%M%S"}
 | ||
|     CURRENT_LOG_TIME=$(date "+$TIMESTAMP_FORMAT")
 | ||
| fi
 | ||
| 
 | ||
| if [[ -n "$LOGFILE" ]]; then
 | ||
|     # First clean up old log files.  Use the user-specified LOGFILE
 | ||
|     # as the template to search for, appending '.*' to match the date
 | ||
|     # we added on earlier runs.
 | ||
|     LOGDIR=$(dirname "$LOGFILE")
 | ||
|     LOGNAME=$(basename "$LOGFILE")
 | ||
|     mkdir -p $LOGDIR
 | ||
|     find $LOGDIR -maxdepth 1 -name $LOGNAME.\* -mtime +$LOGDAYS -exec rm {} \;
 | ||
| 
 | ||
|     LOGFILE=$LOGFILE.${CURRENT_LOG_TIME}
 | ||
|     # Redirect stdout/stderr to tee to write the log file
 | ||
|     exec 1> >( tee "${LOGFILE}" ) 2>&1
 | ||
|     echo "stack.sh log $LOGFILE"
 | ||
|     # Specified logfile name always links to the most recent log
 | ||
|     ln -sf $LOGFILE $LOGDIR/$LOGNAME
 | ||
| fi
 | ||
| 
 | ||
| # Set up logging of screen windows
 | ||
| # Set SCREEN_LOGDIR to turn on logging of screen windows to the
 | ||
| # directory specified in SCREEN_LOGDIR, we will log to the the file
 | ||
| # screen-$SERVICE_NAME-$TIMESTAMP.log in that dir and have a link
 | ||
| # screen-$SERVICE_NAME.log to the latest log file.
 | ||
| # Logs are kept for as long specified in LOGDAYS.
 | ||
| if [[ -n "$SCREEN_LOGDIR" ]]; then
 | ||
| 
 | ||
|     # We make sure the directory is created.
 | ||
|     if [[ -d "$SCREEN_LOGDIR" ]]; then
 | ||
|         # We cleanup the old logs
 | ||
|         find $SCREEN_LOGDIR -maxdepth 1 -name screen-\*.log -mtime +$LOGDAYS -exec rm {} \;
 | ||
|     else
 | ||
|         mkdir -p $SCREEN_LOGDIR
 | ||
|     fi
 | ||
| fi
 | ||
| 
 | ||
| # So that errors don't compound we exit on any errors so you see only the
 | ||
| # first error that occurred.
 | ||
| trap failed ERR
 | ||
| failed() {
 | ||
|     local r=$?
 | ||
|     set +o xtrace
 | ||
|     [ -n "$LOGFILE" ] && echo "${0##*/} failed: full log in $LOGFILE"
 | ||
|     exit $r
 | ||
| }
 | ||
| 
 | ||
| # Print the commands being run so that we can see the command that triggers
 | ||
| # an error.  It is also useful for following along as the install occurs.
 | ||
| set -o xtrace
 | ||
| 
 | ||
| 
 | ||
| # Install Packages
 | ||
| # ================
 | ||
| 
 | ||
| # Openstack uses a fair number of other projects.
 | ||
| 
 | ||
| # Install package requirements
 | ||
| if [[ "$os_PACKAGE" = "deb" ]]; then
 | ||
|     apt_get update
 | ||
|     install_package $(get_packages $FILES/apts)
 | ||
| else
 | ||
|     install_package $(get_packages $FILES/rpms)
 | ||
| fi
 | ||
| 
 | ||
| if [[ $SYSLOG != "False" ]]; then
 | ||
|     install_package rsyslog-relp
 | ||
| fi
 | ||
| 
 | ||
| if is_service_enabled rabbit; then
 | ||
|     # Install rabbitmq-server
 | ||
|     # the temp file is necessary due to LP: #878600
 | ||
|     tfile=$(mktemp)
 | ||
|     install_package rabbitmq-server > "$tfile" 2>&1
 | ||
|     cat "$tfile"
 | ||
|     rm -f "$tfile"
 | ||
| elif is_service_enabled qpid; then
 | ||
|     if [[ "$os_PACKAGE" = "rpm" ]]; then
 | ||
|         install_package qpid-cpp-server
 | ||
|     else
 | ||
|         install_package qpidd
 | ||
|     fi
 | ||
| fi
 | ||
| 
 | ||
| if is_service_enabled mysql; then
 | ||
| 
 | ||
|     if [[ "$os_PACKAGE" = "deb" ]]; then
 | ||
|         # Seed configuration with mysql password so that apt-get install doesn't
 | ||
|         # prompt us for a password upon install.
 | ||
|         cat <<MYSQL_PRESEED | sudo debconf-set-selections
 | ||
| mysql-server-5.1 mysql-server/root_password password $MYSQL_PASSWORD
 | ||
| mysql-server-5.1 mysql-server/root_password_again password $MYSQL_PASSWORD
 | ||
| mysql-server-5.1 mysql-server/start_on_boot boolean true
 | ||
| MYSQL_PRESEED
 | ||
|     fi
 | ||
| 
 | ||
|     # while ``.my.cnf`` is not needed for openstack to function, it is useful
 | ||
|     # as it allows you to access the mysql databases via ``mysql nova`` instead
 | ||
|     # of having to specify the username/password each time.
 | ||
|     if [[ ! -e $HOME/.my.cnf ]]; then
 | ||
|         cat <<EOF >$HOME/.my.cnf
 | ||
| [client]
 | ||
| user=$MYSQL_USER
 | ||
| password=$MYSQL_PASSWORD
 | ||
| host=$MYSQL_HOST
 | ||
| EOF
 | ||
|         chmod 0600 $HOME/.my.cnf
 | ||
|     fi
 | ||
|     # Install mysql-server
 | ||
|     install_package mysql-server
 | ||
| fi
 | ||
| 
 | ||
| if is_service_enabled quantum; then
 | ||
|     if [[ "$Q_PLUGIN" = "linuxbridge" ]]; then
 | ||
|         # Install deps
 | ||
|         # FIXME add to files/apts/quantum, but don't install if not needed!
 | ||
|         install_package python-configobj
 | ||
|     fi
 | ||
| fi
 | ||
| 
 | ||
| if is_service_enabled horizon; then
 | ||
|     if [[ "$os_PACKAGE" = "deb" ]]; then
 | ||
|         # Install apache2, which is NOPRIME'd
 | ||
|         install_package apache2 libapache2-mod-wsgi
 | ||
|     else
 | ||
|         sudo rm -f /etc/httpd/conf.d/000-*
 | ||
|         install_package httpd mod_wsgi
 | ||
|     fi
 | ||
| fi
 | ||
| 
 | ||
| if is_service_enabled q-agt; then
 | ||
|     if [[ "$Q_PLUGIN" = "openvswitch" ]]; then
 | ||
|         # Install deps
 | ||
|         # FIXME add to files/apts/quantum, but don't install if not needed!
 | ||
|         if [[ "$os_PACKAGE" = "deb" ]]; then
 | ||
|             kernel_version=`cat /proc/version | cut -d " " -f3`
 | ||
|             install_package make fakeroot dkms openvswitch-switch openvswitch-datapath-dkms linux-headers-$kernel_version
 | ||
|         else
 | ||
|             ### FIXME(dtroyer): Find RPMs for OpenVSwitch
 | ||
|             echo "OpenVSwitch packages need to be located"
 | ||
|         fi
 | ||
|     elif [[ "$Q_PLUGIN" = "linuxbridge" ]]; then
 | ||
|        install_package bridge-utils
 | ||
|     fi
 | ||
| fi
 | ||
| 
 | ||
| if is_service_enabled n-cpu; then
 | ||
| 
 | ||
|     # Virtualization Configuration
 | ||
|     # ~~~~~~~~~~~~~~~~~~~~~~~~~~~~
 | ||
|     if [[ "$os_PACKAGE" = "deb" ]]; then
 | ||
|         LIBVIRT_PKG_NAME=libvirt-bin
 | ||
|     else
 | ||
|         LIBVIRT_PKG_NAME=libvirt
 | ||
|     fi
 | ||
|     install_package $LIBVIRT_PKG_NAME
 | ||
|     # Install and configure **LXC** if specified.  LXC is another approach to
 | ||
|     # splitting a system into many smaller parts.  LXC uses cgroups and chroot
 | ||
|     # to simulate multiple systems.
 | ||
|     if [[ "$LIBVIRT_TYPE" == "lxc" ]]; then
 | ||
|         if [[ "$os_PACKAGE" = "deb" ]]; then
 | ||
|             if [[ "$DISTRO" > natty ]]; then
 | ||
|                 install_package cgroup-lite
 | ||
|             fi
 | ||
|         else
 | ||
|             ### FIXME(dtroyer): figure this out
 | ||
|             echo "RPM-based cgroup not implemented yet"
 | ||
|             yum_install libcgroup-tools
 | ||
|         fi
 | ||
|     fi
 | ||
| fi
 | ||
| 
 | ||
| if is_service_enabled swift; then
 | ||
|     # Install memcached for swift.
 | ||
|     install_package memcached
 | ||
| fi
 | ||
| 
 | ||
| TRACK_DEPENDS=${TRACK_DEPENDS:-False}
 | ||
| 
 | ||
| # Install python packages into a virtualenv so that we can track them
 | ||
| if [[ $TRACK_DEPENDS = True ]] ; then
 | ||
|     install_package python-virtualenv
 | ||
| 
 | ||
|     rm -rf $DEST/.venv
 | ||
|     virtualenv --system-site-packages $DEST/.venv
 | ||
|     source $DEST/.venv/bin/activate
 | ||
|     $DEST/.venv/bin/pip freeze > $DEST/requires-pre-pip
 | ||
| fi
 | ||
| 
 | ||
| # Install python requirements
 | ||
| pip_install $(get_packages $FILES/pips | sort -u)
 | ||
| 
 | ||
| # Check out OpenStack sources
 | ||
| git_clone $NOVA_REPO $NOVA_DIR $NOVA_BRANCH
 | ||
| 
 | ||
| # Check out the client libs that are used most
 | ||
| git_clone $KEYSTONECLIENT_REPO $KEYSTONECLIENT_DIR $KEYSTONECLIENT_BRANCH
 | ||
| git_clone $NOVACLIENT_REPO $NOVACLIENT_DIR $NOVACLIENT_BRANCH
 | ||
| git_clone $OPENSTACKCLIENT_REPO $OPENSTACKCLIENT_DIR $OPENSTACKCLIENT_BRANCH
 | ||
| git_clone $GLANCECLIENT_REPO $GLANCECLIENT_DIR $GLANCECLIENT_BRANCH
 | ||
| 
 | ||
| # glance, swift middleware and nova api needs keystone middleware
 | ||
| if is_service_enabled key g-api n-api swift; then
 | ||
|     # unified auth system (manages accounts/tokens)
 | ||
|     git_clone $KEYSTONE_REPO $KEYSTONE_DIR $KEYSTONE_BRANCH
 | ||
| fi
 | ||
| if is_service_enabled swift; then
 | ||
|     # storage service
 | ||
|     git_clone $SWIFT_REPO $SWIFT_DIR $SWIFT_BRANCH
 | ||
|     # storage service client and and Library
 | ||
|     git_clone $SWIFTCLIENT_REPO $SWIFTCLIENT_DIR $SWIFTCLIENT_BRANCH
 | ||
|     if is_service_enabled swift3; then
 | ||
|         # swift3 middleware to provide S3 emulation to Swift
 | ||
|         git_clone $SWIFT3_REPO $SWIFT3_DIR $SWIFT3_BRANCH
 | ||
|     fi
 | ||
| fi
 | ||
| if is_service_enabled g-api n-api; then
 | ||
|     # image catalog service
 | ||
|     git_clone $GLANCE_REPO $GLANCE_DIR $GLANCE_BRANCH
 | ||
| fi
 | ||
| if is_service_enabled n-novnc; then
 | ||
|     # a websockets/html5 or flash powered VNC console for vm instances
 | ||
|     git_clone $NOVNC_REPO $NOVNC_DIR $NOVNC_BRANCH
 | ||
| fi
 | ||
| if is_service_enabled horizon; then
 | ||
|     # django powered web control panel for openstack
 | ||
|     git_clone $HORIZON_REPO $HORIZON_DIR $HORIZON_BRANCH $HORIZON_TAG
 | ||
| fi
 | ||
| if is_service_enabled quantum; then
 | ||
|     git_clone $QUANTUM_CLIENT_REPO $QUANTUM_CLIENT_DIR $QUANTUM_CLIENT_BRANCH
 | ||
| fi
 | ||
| if is_service_enabled quantum; then
 | ||
|     # quantum
 | ||
|     git_clone $QUANTUM_REPO $QUANTUM_DIR $QUANTUM_BRANCH
 | ||
| fi
 | ||
| if is_service_enabled cinder; then
 | ||
|     install_cinder
 | ||
| fi
 | ||
| if is_service_enabled ceilometer; then
 | ||
|     install_ceilometer
 | ||
| fi
 | ||
| 
 | ||
| # Initialization
 | ||
| # ==============
 | ||
| 
 | ||
| # Set up our checkouts so they are installed into python path
 | ||
| # allowing ``import nova`` or ``import glance.client``
 | ||
| setup_develop $KEYSTONECLIENT_DIR
 | ||
| setup_develop $NOVACLIENT_DIR
 | ||
| setup_develop $OPENSTACKCLIENT_DIR
 | ||
| if is_service_enabled key g-api n-api swift; then
 | ||
|     setup_develop $KEYSTONE_DIR
 | ||
| fi
 | ||
| if is_service_enabled swift; then
 | ||
|     setup_develop $SWIFT_DIR
 | ||
|     setup_develop $SWIFTCLIENT_DIR
 | ||
| fi
 | ||
| if is_service_enabled swift3; then
 | ||
|     setup_develop $SWIFT3_DIR
 | ||
| fi
 | ||
| if is_service_enabled g-api n-api; then
 | ||
|     setup_develop $GLANCE_DIR
 | ||
| fi
 | ||
| 
 | ||
| # Do this _after_ glance is installed to override the old binary
 | ||
| setup_develop $GLANCECLIENT_DIR
 | ||
| 
 | ||
| setup_develop $NOVA_DIR
 | ||
| if is_service_enabled horizon; then
 | ||
|     setup_develop $HORIZON_DIR
 | ||
| fi
 | ||
| if is_service_enabled quantum; then
 | ||
|     setup_develop $QUANTUM_CLIENT_DIR
 | ||
|     setup_develop $QUANTUM_DIR
 | ||
| fi
 | ||
| if is_service_enabled cinder; then
 | ||
|     configure_cinder
 | ||
| fi
 | ||
| 
 | ||
| if [[ $TRACK_DEPENDS = True ]] ; then
 | ||
|     $DEST/.venv/bin/pip freeze > $DEST/requires-post-pip
 | ||
|     if ! diff -Nru $DEST/requires-pre-pip $DEST/requires-post-pip > $DEST/requires.diff ; then
 | ||
|         cat $DEST/requires.diff
 | ||
|     fi
 | ||
|     echo "Ran stack.sh in depend tracking mode, bailing out now"
 | ||
|     exit 0
 | ||
| fi
 | ||
| 
 | ||
| # Syslog
 | ||
| # ------
 | ||
| 
 | ||
| if [[ $SYSLOG != "False" ]]; then
 | ||
|     if [[ "$SYSLOG_HOST" = "$HOST_IP" ]]; then
 | ||
|         # Configure the master host to receive
 | ||
|         cat <<EOF >/tmp/90-stack-m.conf
 | ||
| \$ModLoad imrelp
 | ||
| \$InputRELPServerRun $SYSLOG_PORT
 | ||
| EOF
 | ||
|         sudo mv /tmp/90-stack-m.conf /etc/rsyslog.d
 | ||
|     else
 | ||
|         # Set rsyslog to send to remote host
 | ||
|         cat <<EOF >/tmp/90-stack-s.conf
 | ||
| *.*		:omrelp:$SYSLOG_HOST:$SYSLOG_PORT
 | ||
| EOF
 | ||
|         sudo mv /tmp/90-stack-s.conf /etc/rsyslog.d
 | ||
|     fi
 | ||
|     restart_service rsyslog
 | ||
| fi
 | ||
| 
 | ||
| 
 | ||
| # Rabbit or Qpid
 | ||
| # --------------
 | ||
| 
 | ||
| if is_service_enabled rabbit; then
 | ||
|     # Start rabbitmq-server
 | ||
|     if [[ "$os_PACKAGE" = "rpm" ]]; then
 | ||
|         # RPM doesn't start the service
 | ||
|         restart_service rabbitmq-server
 | ||
|     fi
 | ||
|     # change the rabbit password since the default is "guest"
 | ||
|     sudo rabbitmqctl change_password guest $RABBIT_PASSWORD
 | ||
| elif is_service_enabled qpid; then
 | ||
|     restart_service qpidd
 | ||
| fi
 | ||
| 
 | ||
| 
 | ||
| # Mysql
 | ||
| # -----
 | ||
| 
 | ||
| 
 | ||
| if is_service_enabled mysql; then
 | ||
| 
 | ||
|     #start mysql-server
 | ||
|     if [[ "$os_PACKAGE" = "rpm" ]]; then
 | ||
|         # RPM doesn't start the service
 | ||
|         start_service mysqld
 | ||
|         # Set the root password - only works the first time
 | ||
|         sudo mysqladmin -u root password $MYSQL_PASSWORD || true
 | ||
|     fi
 | ||
|     # Update the DB to give user ‘$MYSQL_USER’@’%’ full control of the all databases:
 | ||
|     sudo mysql -uroot -p$MYSQL_PASSWORD -h127.0.0.1 -e "GRANT ALL PRIVILEGES ON *.* TO '$MYSQL_USER'@'%' identified by '$MYSQL_PASSWORD';"
 | ||
| 
 | ||
|     # Update ``my.cnf`` for some local needs and restart the mysql service
 | ||
|     if [[ "$os_PACKAGE" = "deb" ]]; then
 | ||
|         MY_CONF=/etc/mysql/my.cnf
 | ||
|         MYSQL=mysql
 | ||
|     else
 | ||
|         MY_CONF=/etc/my.cnf
 | ||
|         MYSQL=mysqld
 | ||
|     fi
 | ||
| 
 | ||
|     # Change ‘bind-address’ from localhost (127.0.0.1) to any (0.0.0.0)
 | ||
|     sudo sed -i '/^bind-address/s/127.0.0.1/0.0.0.0/g' $MY_CONF
 | ||
| 
 | ||
|     # Set default db type to InnoDB
 | ||
|     if sudo grep -q "default-storage-engine" $MY_CONF; then
 | ||
|         # Change it
 | ||
|         sudo bash -c "source $TOP_DIR/functions; iniset $MY_CONF mysqld default-storage-engine InnoDB"
 | ||
|     else
 | ||
|         # Add it
 | ||
|         sudo sed -i -e "/^\[mysqld\]/ a \
 | ||
| default-storage-engine = InnoDB" $MY_CONF
 | ||
|     fi
 | ||
| 
 | ||
|     restart_service $MYSQL
 | ||
| fi
 | ||
| 
 | ||
| if [ -z "$SCREEN_HARDSTATUS" ]; then
 | ||
|     SCREEN_HARDSTATUS='%{= .} %-Lw%{= .}%> %n%f %t*%{= .}%+Lw%< %-=%{g}(%{d}%H/%l%{g})'
 | ||
| fi
 | ||
| 
 | ||
| # Our screenrc file builder
 | ||
| function screen_rc {
 | ||
|     SCREENRC=$TOP_DIR/stack-screenrc
 | ||
|     if [[ ! -e $SCREENRC ]]; then
 | ||
|         # Name the screen session
 | ||
|         echo "sessionname stack" > $SCREENRC
 | ||
|         # Set a reasonable statusbar
 | ||
|         echo "hardstatus alwayslastline '$SCREEN_HARDSTATUS'" >> $SCREENRC
 | ||
|         echo "screen -t stack bash" >> $SCREENRC
 | ||
|     fi
 | ||
|     # If this service doesn't already exist in the screenrc file
 | ||
|     if ! grep $1 $SCREENRC 2>&1 > /dev/null; then
 | ||
|         NL=`echo -ne '\015'`
 | ||
|         echo "screen -t $1 bash" >> $SCREENRC
 | ||
|         echo "stuff \"$2$NL\"" >> $SCREENRC
 | ||
|     fi
 | ||
| }
 | ||
| 
 | ||
| # Our screen helper to launch a service in a hidden named screen
 | ||
| function screen_it {
 | ||
|     NL=`echo -ne '\015'`
 | ||
|     if is_service_enabled $1; then
 | ||
|         # Append the service to the screen rc file
 | ||
|         screen_rc "$1" "$2"
 | ||
| 
 | ||
|         screen -S stack -X screen -t $1
 | ||
|         # sleep to allow bash to be ready to be send the command - we are
 | ||
|         # creating a new window in screen and then sends characters, so if
 | ||
|         # bash isn't running by the time we send the command, nothing happens
 | ||
|         sleep 1.5
 | ||
| 
 | ||
|         if [[ -n ${SCREEN_LOGDIR} ]]; then
 | ||
|             screen -S stack -p $1 -X logfile ${SCREEN_LOGDIR}/screen-${1}.${CURRENT_LOG_TIME}.log
 | ||
|             screen -S stack -p $1 -X log on
 | ||
|             ln -sf ${SCREEN_LOGDIR}/screen-${1}.${CURRENT_LOG_TIME}.log ${SCREEN_LOGDIR}/screen-${1}.log
 | ||
|         fi
 | ||
|         screen -S stack -p $1 -X stuff "$2$NL"
 | ||
|     fi
 | ||
| }
 | ||
| 
 | ||
| # Create a new named screen to run processes in
 | ||
| screen -d -m -S stack -t stack -s /bin/bash
 | ||
| sleep 1
 | ||
| # Set a reasonable statusbar
 | ||
| screen -r stack -X hardstatus alwayslastline "$SCREEN_HARDSTATUS"
 | ||
| 
 | ||
| 
 | ||
| # Horizon
 | ||
| # -------
 | ||
| 
 | ||
| # Set up the django horizon application to serve via apache/wsgi
 | ||
| 
 | ||
| if is_service_enabled horizon; then
 | ||
| 
 | ||
|     # Remove stale session database.
 | ||
|     rm -f $HORIZON_DIR/openstack_dashboard/local/dashboard_openstack.sqlite3
 | ||
| 
 | ||
|     # ``local_settings.py`` is used to override horizon default settings.
 | ||
|     local_settings=$HORIZON_DIR/openstack_dashboard/local/local_settings.py
 | ||
|     cp $FILES/horizon_settings.py $local_settings
 | ||
| 
 | ||
|     # Initialize the horizon database (it stores sessions and notices shown to
 | ||
|     # users).  The user system is external (keystone).
 | ||
|     cd $HORIZON_DIR
 | ||
|     python manage.py syncdb --noinput
 | ||
|     cd $TOP_DIR
 | ||
| 
 | ||
|     # Create an empty directory that apache uses as docroot
 | ||
|     sudo mkdir -p $HORIZON_DIR/.blackhole
 | ||
| 
 | ||
|     if [[ "$os_PACKAGE" = "deb" ]]; then
 | ||
|         APACHE_NAME=apache2
 | ||
|         APACHE_CONF=sites-available/horizon
 | ||
|         # Clean up the old config name
 | ||
|         sudo rm -f /etc/apache2/sites-enabled/000-default
 | ||
|         # Be a good citizen and use the distro tools here
 | ||
|         sudo touch /etc/$APACHE_NAME/$APACHE_CONF
 | ||
|         sudo a2ensite horizon
 | ||
|     else
 | ||
|         # Install httpd, which is NOPRIME'd
 | ||
|         APACHE_NAME=httpd
 | ||
|         APACHE_CONF=conf.d/horizon.conf
 | ||
|         sudo sed '/^Listen/s/^.*$/Listen 0.0.0.0:80/' -i /etc/httpd/conf/httpd.conf
 | ||
|     fi
 | ||
|     ## Configure apache to run horizon
 | ||
|     sudo sh -c "sed -e \"
 | ||
|         s,%USER%,$APACHE_USER,g;
 | ||
|         s,%GROUP%,$APACHE_GROUP,g;
 | ||
|         s,%HORIZON_DIR%,$HORIZON_DIR,g;
 | ||
|         s,%APACHE_NAME%,$APACHE_NAME,g;
 | ||
|         s,%DEST%,$DEST,g;
 | ||
|     \" $FILES/apache-horizon.template >/etc/$APACHE_NAME/$APACHE_CONF"
 | ||
|     restart_service $APACHE_NAME
 | ||
| fi
 | ||
| 
 | ||
| 
 | ||
| # Glance
 | ||
| # ------
 | ||
| 
 | ||
| if is_service_enabled g-reg; then
 | ||
|     GLANCE_CONF_DIR=/etc/glance
 | ||
|     if [[ ! -d $GLANCE_CONF_DIR ]]; then
 | ||
|         sudo mkdir -p $GLANCE_CONF_DIR
 | ||
|     fi
 | ||
|     sudo chown `whoami` $GLANCE_CONF_DIR
 | ||
|     GLANCE_IMAGE_DIR=$DEST/glance/images
 | ||
|     # Delete existing images
 | ||
|     rm -rf $GLANCE_IMAGE_DIR
 | ||
| 
 | ||
|     # Use local glance directories
 | ||
|     mkdir -p $GLANCE_IMAGE_DIR
 | ||
| 
 | ||
|     # (re)create glance database
 | ||
|     mysql -u$MYSQL_USER -p$MYSQL_PASSWORD -e 'DROP DATABASE IF EXISTS glance;'
 | ||
|     mysql -u$MYSQL_USER -p$MYSQL_PASSWORD -e 'CREATE DATABASE glance CHARACTER SET utf8;'
 | ||
| 
 | ||
|     # Copy over our glance configurations and update them
 | ||
|     GLANCE_REGISTRY_CONF=$GLANCE_CONF_DIR/glance-registry.conf
 | ||
|     cp $GLANCE_DIR/etc/glance-registry.conf $GLANCE_REGISTRY_CONF
 | ||
|     iniset $GLANCE_REGISTRY_CONF DEFAULT debug True
 | ||
|     inicomment $GLANCE_REGISTRY_CONF DEFAULT log_file
 | ||
|     iniset $GLANCE_REGISTRY_CONF DEFAULT sql_connection $BASE_SQL_CONN/glance?charset=utf8
 | ||
|     iniset $GLANCE_REGISTRY_CONF DEFAULT use_syslog $SYSLOG
 | ||
|     iniset $GLANCE_REGISTRY_CONF paste_deploy flavor keystone
 | ||
| 
 | ||
|     GLANCE_REGISTRY_PASTE_INI=$GLANCE_CONF_DIR/glance-registry-paste.ini
 | ||
|     cp $GLANCE_DIR/etc/glance-registry-paste.ini $GLANCE_REGISTRY_PASTE_INI
 | ||
|     iniset $GLANCE_REGISTRY_PASTE_INI filter:authtoken auth_host $KEYSTONE_AUTH_HOST
 | ||
|     iniset $GLANCE_REGISTRY_PASTE_INI filter:authtoken auth_port $KEYSTONE_AUTH_PORT
 | ||
|     iniset $GLANCE_REGISTRY_PASTE_INI filter:authtoken auth_protocol $KEYSTONE_AUTH_PROTOCOL
 | ||
|     iniset $GLANCE_REGISTRY_PASTE_INI filter:authtoken auth_uri $KEYSTONE_SERVICE_PROTOCOL://$KEYSTONE_SERVICE_HOST:$KEYSTONE_SERVICE_PORT/
 | ||
|     iniset $GLANCE_REGISTRY_PASTE_INI filter:authtoken admin_tenant_name $SERVICE_TENANT_NAME
 | ||
|     iniset $GLANCE_REGISTRY_PASTE_INI filter:authtoken admin_user glance
 | ||
|     iniset $GLANCE_REGISTRY_PASTE_INI filter:authtoken admin_password $SERVICE_PASSWORD
 | ||
| 
 | ||
|     GLANCE_API_CONF=$GLANCE_CONF_DIR/glance-api.conf
 | ||
|     cp $GLANCE_DIR/etc/glance-api.conf $GLANCE_API_CONF
 | ||
|     iniset $GLANCE_API_CONF DEFAULT debug True
 | ||
|     inicomment $GLANCE_API_CONF DEFAULT log_file
 | ||
|     iniset $GLANCE_API_CONF DEFAULT sql_connection $BASE_SQL_CONN/glance?charset=utf8
 | ||
|     iniset $GLANCE_API_CONF DEFAULT use_syslog $SYSLOG
 | ||
|     iniset $GLANCE_API_CONF DEFAULT filesystem_store_datadir $GLANCE_IMAGE_DIR/
 | ||
|     iniset $GLANCE_API_CONF paste_deploy flavor keystone
 | ||
| 
 | ||
|     # Store the images in swift if enabled.
 | ||
|     if is_service_enabled swift; then
 | ||
|         iniset $GLANCE_API_CONF DEFAULT default_store swift
 | ||
|         iniset $GLANCE_API_CONF DEFAULT swift_store_auth_address $KEYSTONE_SERVICE_PROTOCOL://$KEYSTONE_SERVICE_HOST:$KEYSTONE_SERVICE_PORT/v2.0/
 | ||
|         iniset $GLANCE_API_CONF DEFAULT swift_store_user $SERVICE_TENANT_NAME:glance
 | ||
|         iniset $GLANCE_API_CONF DEFAULT swift_store_key $SERVICE_PASSWORD
 | ||
|         iniset $GLANCE_API_CONF DEFAULT swift_store_create_container_on_put True
 | ||
|     fi
 | ||
| 
 | ||
|     GLANCE_API_PASTE_INI=$GLANCE_CONF_DIR/glance-api-paste.ini
 | ||
|     cp $GLANCE_DIR/etc/glance-api-paste.ini $GLANCE_API_PASTE_INI
 | ||
|     iniset $GLANCE_API_PASTE_INI filter:authtoken auth_host $KEYSTONE_AUTH_HOST
 | ||
|     iniset $GLANCE_API_PASTE_INI filter:authtoken auth_port $KEYSTONE_AUTH_PORT
 | ||
|     iniset $GLANCE_API_PASTE_INI filter:authtoken auth_protocol $KEYSTONE_AUTH_PROTOCOL
 | ||
|     iniset $GLANCE_API_PASTE_INI filter:authtoken auth_uri $KEYSTONE_SERVICE_PROTOCOL://$KEYSTONE_SERVICE_HOST:$KEYSTONE_SERVICE_PORT/
 | ||
|     iniset $GLANCE_API_PASTE_INI filter:authtoken admin_tenant_name $SERVICE_TENANT_NAME
 | ||
|     iniset $GLANCE_API_PASTE_INI filter:authtoken admin_user glance
 | ||
|     iniset $GLANCE_API_PASTE_INI filter:authtoken admin_password $SERVICE_PASSWORD
 | ||
| 
 | ||
|     GLANCE_POLICY_JSON=$GLANCE_CONF_DIR/policy.json
 | ||
|     cp $GLANCE_DIR/etc/policy.json $GLANCE_POLICY_JSON
 | ||
| 
 | ||
|     $GLANCE_DIR/bin/glance-manage db_sync
 | ||
| 
 | ||
| fi
 | ||
| 
 | ||
| 
 | ||
| # Quantum
 | ||
| # -------
 | ||
| 
 | ||
| if is_service_enabled quantum; then
 | ||
|     # Put config files in /etc/quantum for everyone to find
 | ||
|     if [[ ! -d /etc/quantum ]]; then
 | ||
|         sudo mkdir -p /etc/quantum
 | ||
|     fi
 | ||
|     sudo chown `whoami` /etc/quantum
 | ||
| 
 | ||
|     if [[ "$Q_PLUGIN" = "openvswitch" ]]; then
 | ||
|         Q_PLUGIN_CONF_PATH=etc/quantum/plugins/openvswitch
 | ||
|         Q_PLUGIN_CONF_FILENAME=ovs_quantum_plugin.ini
 | ||
|         Q_DB_NAME="ovs_quantum"
 | ||
|         Q_PLUGIN_CLASS="quantum.plugins.openvswitch.ovs_quantum_plugin.OVSQuantumPluginV2"
 | ||
|     elif [[ "$Q_PLUGIN" = "linuxbridge" ]]; then
 | ||
|         Q_PLUGIN_CONF_PATH=etc/quantum/plugins/linuxbridge
 | ||
|         Q_PLUGIN_CONF_FILENAME=linuxbridge_conf.ini
 | ||
|         Q_DB_NAME="quantum_linux_bridge"
 | ||
|         Q_PLUGIN_CLASS="quantum.plugins.linuxbridge.lb_quantum_plugin.LinuxBridgePluginV2"
 | ||
|     else
 | ||
|         echo "Unknown Quantum plugin '$Q_PLUGIN'.. exiting"
 | ||
|         exit 1
 | ||
|     fi
 | ||
| 
 | ||
|     # If needed, move config file from $QUANTUM_DIR/etc/quantum to /etc/quantum
 | ||
|     mkdir -p /$Q_PLUGIN_CONF_PATH
 | ||
|     Q_PLUGIN_CONF_FILE=$Q_PLUGIN_CONF_PATH/$Q_PLUGIN_CONF_FILENAME
 | ||
|     cp $QUANTUM_DIR/$Q_PLUGIN_CONF_FILE /$Q_PLUGIN_CONF_FILE
 | ||
| 
 | ||
|     sudo sed -i -e "s/^sql_connection =.*$/sql_connection = mysql:\/\/$MYSQL_USER:$MYSQL_PASSWORD@$MYSQL_HOST\/$Q_DB_NAME?charset=utf8/g" /$Q_PLUGIN_CONF_FILE
 | ||
| 
 | ||
|     OVS_ENABLE_TUNNELING=${OVS_ENABLE_TUNNELING:-True}
 | ||
|     if [[ "$Q_PLUGIN" = "openvswitch" && $OVS_ENABLE_TUNNELING = "True" ]]; then
 | ||
|         OVS_VERSION=`ovs-vsctl --version | head -n 1 | awk '{print $4;}'`
 | ||
|         if [ $OVS_VERSION \< "1.4" ] && ! is_service_enabled q-svc ; then
 | ||
|             echo "You are running OVS version $OVS_VERSION."
 | ||
|             echo "OVS 1.4+ is required for tunneling between multiple hosts."
 | ||
|             exit 1
 | ||
|         fi
 | ||
|         sudo sed -i -e "s/.*enable_tunneling = .*$/enable_tunneling = $OVS_ENABLE_TUNNELING/g" /$Q_PLUGIN_CONF_FILE
 | ||
|     fi
 | ||
| 
 | ||
|     Q_CONF_FILE=/etc/quantum/quantum.conf
 | ||
|     cp $QUANTUM_DIR/etc/quantum.conf $Q_CONF_FILE
 | ||
| fi
 | ||
| 
 | ||
| # Quantum service (for controller node)
 | ||
| if is_service_enabled q-svc; then
 | ||
|     Q_API_PASTE_FILE=/etc/quantum/api-paste.ini
 | ||
|     Q_POLICY_FILE=/etc/quantum/policy.json
 | ||
| 
 | ||
|     cp $QUANTUM_DIR/etc/api-paste.ini $Q_API_PASTE_FILE
 | ||
|     cp $QUANTUM_DIR/etc/policy.json $Q_POLICY_FILE
 | ||
| 
 | ||
|     if is_service_enabled mysql; then
 | ||
|             mysql -u$MYSQL_USER -p$MYSQL_PASSWORD -e "DROP DATABASE IF EXISTS $Q_DB_NAME;"
 | ||
|             mysql -u$MYSQL_USER -p$MYSQL_PASSWORD -e "CREATE DATABASE IF NOT EXISTS $Q_DB_NAME CHARACTER SET utf8;"
 | ||
|         else
 | ||
|             echo "mysql must be enabled in order to use the $Q_PLUGIN Quantum plugin."
 | ||
|             exit 1
 | ||
|     fi
 | ||
| 
 | ||
|     # Update either configuration file with plugin
 | ||
|     iniset $Q_CONF_FILE DEFAULT core_plugin $Q_PLUGIN_CLASS
 | ||
| 
 | ||
|     iniset $Q_CONF_FILE DEFAULT auth_strategy $Q_AUTH_STRATEGY
 | ||
|     iniset $Q_API_PASTE_FILE filter:authtoken auth_host $KEYSTONE_SERVICE_HOST
 | ||
|     iniset $Q_API_PASTE_FILE filter:authtoken auth_port $KEYSTONE_AUTH_PORT
 | ||
|     iniset $Q_API_PASTE_FILE filter:authtoken auth_protocol $KEYSTONE_SERVICE_PROTOCOL
 | ||
|     iniset $Q_API_PASTE_FILE filter:authtoken admin_tenant_name $SERVICE_TENANT_NAME
 | ||
|     iniset $Q_API_PASTE_FILE filter:authtoken admin_user $Q_ADMIN_USERNAME
 | ||
|     iniset $Q_API_PASTE_FILE filter:authtoken admin_password $SERVICE_PASSWORD
 | ||
| fi
 | ||
| 
 | ||
| # Quantum agent (for compute nodes)
 | ||
| if is_service_enabled q-agt; then
 | ||
|     if [[ "$Q_PLUGIN" = "openvswitch" ]]; then
 | ||
|         # Set up integration bridge
 | ||
|         OVS_BRIDGE=${OVS_BRIDGE:-br-int}
 | ||
|         for PORT in `sudo ovs-vsctl --no-wait list-ports $OVS_BRIDGE`; do
 | ||
|             if [[ "$PORT" =~ tap* ]]; then echo `sudo ip link delete $PORT` > /dev/null; fi
 | ||
|             sudo ovs-vsctl --no-wait del-port $OVS_BRIDGE $PORT
 | ||
|         done
 | ||
|         sudo ovs-vsctl --no-wait -- --if-exists del-br $OVS_BRIDGE
 | ||
|         sudo ovs-vsctl --no-wait add-br $OVS_BRIDGE
 | ||
|         sudo ovs-vsctl --no-wait br-set-external-id $OVS_BRIDGE bridge-id br-int
 | ||
|         sudo sed -i -e "s/.*local_ip = .*/local_ip = $HOST_IP/g" /$Q_PLUGIN_CONF_FILE
 | ||
|         AGENT_BINARY="$QUANTUM_DIR/quantum/plugins/openvswitch/agent/ovs_quantum_agent.py"
 | ||
|     elif [[ "$Q_PLUGIN" = "linuxbridge" ]]; then
 | ||
|        # Start up the quantum <-> linuxbridge agent
 | ||
|        # set the default network interface
 | ||
|        QUANTUM_LB_PRIVATE_INTERFACE=${QUANTUM_LB_PRIVATE_INTERFACE:-$GUEST_INTERFACE_DEFAULT}
 | ||
|        iniset /$Q_PLUGIN_CONF_FILE LINUX_BRIDGE physical_interface_mappings default:$QUANTUM_LB_PRIVATE_INTERFACE
 | ||
|        AGENT_BINARY="$QUANTUM_DIR/quantum/plugins/linuxbridge/agent/linuxbridge_quantum_agent.py"
 | ||
|     fi
 | ||
| fi
 | ||
| 
 | ||
| # Quantum DHCP
 | ||
| if is_service_enabled q-dhcp; then
 | ||
|     AGENT_DHCP_BINARY="$QUANTUM_DIR/bin/quantum-dhcp-agent"
 | ||
| 
 | ||
|     Q_DHCP_CONF_FILE=/etc/quantum/dhcp_agent.ini
 | ||
| 
 | ||
|     cp $QUANTUM_DIR/etc/dhcp_agent.ini $Q_DHCP_CONF_FILE
 | ||
| 
 | ||
|     # Set verbose
 | ||
|     iniset $Q_DHCP_CONF_FILE DEFAULT verbose True
 | ||
|     # Set debug
 | ||
|     iniset $Q_DHCP_CONF_FILE DEFAULT debug True
 | ||
|     iniset $Q_DHCP_CONF_FILE DEFAULT use_namespaces $Q_USE_NAMESPACE
 | ||
| 
 | ||
|     # Update database
 | ||
|     iniset $Q_DHCP_CONF_FILE DEFAULT db_connection "mysql:\/\/$MYSQL_USER:$MYSQL_PASSWORD@$MYSQL_HOST\/$Q_DB_NAME?charset=utf8"
 | ||
|     iniset $Q_DHCP_CONF_FILE DEFAULT auth_url "$KEYSTONE_SERVICE_PROTOCOL://$KEYSTONE_AUTH_HOST:$KEYSTONE_AUTH_PORT/v2.0"
 | ||
|     iniset $Q_DHCP_CONF_FILE DEFAULT admin_tenant_name $SERVICE_TENANT_NAME
 | ||
|     iniset $Q_DHCP_CONF_FILE DEFAULT admin_user $Q_ADMIN_USERNAME
 | ||
|     iniset $Q_DHCP_CONF_FILE DEFAULT admin_password $SERVICE_PASSWORD
 | ||
| 
 | ||
|     if [[ "$Q_PLUGIN" = "openvswitch" ]]; then
 | ||
|         iniset $Q_DHCP_CONF_FILE DEFAULT interface_driver quantum.agent.linux.interface.OVSInterfaceDriver
 | ||
|     elif [[ "$Q_PLUGIN" = "linuxbridge" ]]; then
 | ||
|         iniset $Q_DHCP_CONF_FILE DEFAULT interface_driver quantum.agent.linux.interface.BridgeInterfaceDriver
 | ||
|     fi
 | ||
| fi
 | ||
| 
 | ||
| # Quantum RPC support - must be updated prior to starting any of the services
 | ||
| if is_service_enabled quantum; then
 | ||
|     iniset $Q_CONF_FILE DEFAULT control_exchange quantum
 | ||
|     if is_service_enabled qpid ; then
 | ||
|         iniset $Q_CONF_FILE DEFAULT rpc_backend quantum.openstack.common.rpc.impl_qpid
 | ||
|     elif [ -n "$RABBIT_HOST" ] &&  [ -n "$RABBIT_PASSWORD" ]; then
 | ||
|         iniset $Q_CONF_FILE DEFAULT rabbit_host $RABBIT_HOST
 | ||
|         iniset $Q_CONF_FILE DEFAULT rabbit_password $RABBIT_PASSWORD
 | ||
|     fi
 | ||
| fi
 | ||
| 
 | ||
| # Start the Quantum services
 | ||
| screen_it q-svc "cd $QUANTUM_DIR && python $QUANTUM_DIR/bin/quantum-server --config-file $Q_CONF_FILE --config-file /$Q_PLUGIN_CONF_FILE"
 | ||
| 
 | ||
| # Start up the quantum agent
 | ||
| screen_it q-agt "sudo python $AGENT_BINARY --config-file $Q_CONF_FILE --config-file /$Q_PLUGIN_CONF_FILE"
 | ||
| 
 | ||
| # Start up the quantum agent
 | ||
| screen_it q-dhcp "sudo python $AGENT_DHCP_BINARY --config-file $Q_CONF_FILE --config-file=$Q_DHCP_CONF_FILE"
 | ||
| 
 | ||
| # Nova
 | ||
| # ----
 | ||
| 
 | ||
| # Put config files in /etc/nova for everyone to find
 | ||
| NOVA_CONF_DIR=/etc/nova
 | ||
| if [[ ! -d $NOVA_CONF_DIR ]]; then
 | ||
|     sudo mkdir -p $NOVA_CONF_DIR
 | ||
| fi
 | ||
| sudo chown `whoami` $NOVA_CONF_DIR
 | ||
| 
 | ||
| cp -p $NOVA_DIR/etc/nova/policy.json $NOVA_CONF_DIR
 | ||
| 
 | ||
| # If Nova ships the new rootwrap filters files, deploy them
 | ||
| # (owned by root) and add a parameter to $NOVA_ROOTWRAP
 | ||
| ROOTWRAP_SUDOER_CMD="$NOVA_ROOTWRAP"
 | ||
| if [[ -d $NOVA_DIR/etc/nova/rootwrap.d ]]; then
 | ||
|     # Wipe any existing rootwrap.d files first
 | ||
|     if [[ -d $NOVA_CONF_DIR/rootwrap.d ]]; then
 | ||
|         sudo rm -rf $NOVA_CONF_DIR/rootwrap.d
 | ||
|     fi
 | ||
|     # Deploy filters to /etc/nova/rootwrap.d
 | ||
|     sudo mkdir -m 755 $NOVA_CONF_DIR/rootwrap.d
 | ||
|     sudo cp $NOVA_DIR/etc/nova/rootwrap.d/*.filters $NOVA_CONF_DIR/rootwrap.d
 | ||
|     sudo chown -R root:root $NOVA_CONF_DIR/rootwrap.d
 | ||
|     sudo chmod 644 $NOVA_CONF_DIR/rootwrap.d/*
 | ||
|     # Set up rootwrap.conf, pointing to /etc/nova/rootwrap.d
 | ||
|     sudo cp $NOVA_DIR/etc/nova/rootwrap.conf $NOVA_CONF_DIR/
 | ||
|     sudo sed -e "s:^filters_path=.*$:filters_path=$NOVA_CONF_DIR/rootwrap.d:" -i $NOVA_CONF_DIR/rootwrap.conf
 | ||
|     sudo chown root:root $NOVA_CONF_DIR/rootwrap.conf
 | ||
|     sudo chmod 0644 $NOVA_CONF_DIR/rootwrap.conf
 | ||
|     # Specify rootwrap.conf as first parameter to nova-rootwrap
 | ||
|     NOVA_ROOTWRAP="$NOVA_ROOTWRAP $NOVA_CONF_DIR/rootwrap.conf"
 | ||
|     ROOTWRAP_SUDOER_CMD="$NOVA_ROOTWRAP *"
 | ||
| fi
 | ||
| 
 | ||
| # Set up the rootwrap sudoers for nova
 | ||
| TEMPFILE=`mktemp`
 | ||
| echo "$USER ALL=(root) NOPASSWD: $ROOTWRAP_SUDOER_CMD" >$TEMPFILE
 | ||
| chmod 0440 $TEMPFILE
 | ||
| sudo chown root:root $TEMPFILE
 | ||
| sudo mv $TEMPFILE /etc/sudoers.d/nova-rootwrap
 | ||
| 
 | ||
| if is_service_enabled n-api; then
 | ||
|     # Use the sample http middleware configuration supplied in the
 | ||
|     # Nova sources.  This paste config adds the configuration required
 | ||
|     # for Nova to validate Keystone tokens.
 | ||
| 
 | ||
|     # Allow rate limiting to be turned off for testing, like for Tempest
 | ||
|     # NOTE: Set API_RATE_LIMIT="False" to turn OFF rate limiting
 | ||
|     API_RATE_LIMIT=${API_RATE_LIMIT:-"True"}
 | ||
| 
 | ||
|     # Remove legacy paste config if present
 | ||
|     rm -f $NOVA_DIR/bin/nova-api-paste.ini
 | ||
| 
 | ||
|     # Get the sample configuration file in place
 | ||
|     cp $NOVA_DIR/etc/nova/api-paste.ini $NOVA_CONF_DIR
 | ||
| 
 | ||
|     # Rewrite the authtoken configration for our Keystone service.
 | ||
|     # This is a bit defensive to allow the sample file some varaince.
 | ||
|     sed -e "
 | ||
|         /^admin_token/i admin_tenant_name = $SERVICE_TENANT_NAME
 | ||
|         /admin_tenant_name/s/^.*$/admin_tenant_name = $SERVICE_TENANT_NAME/;
 | ||
|         /admin_user/s/^.*$/admin_user = nova/;
 | ||
|         /admin_password/s/^.*$/admin_password = $SERVICE_PASSWORD/;
 | ||
|         s,%SERVICE_TENANT_NAME%,$SERVICE_TENANT_NAME,g;
 | ||
|         s,%SERVICE_TOKEN%,$SERVICE_TOKEN,g;
 | ||
|     " -i $NOVA_CONF_DIR/api-paste.ini
 | ||
| fi
 | ||
| 
 | ||
| # Helper to clean iptables rules
 | ||
| function clean_iptables() {
 | ||
|     # Delete rules
 | ||
|     sudo iptables -S -v | sed "s/-c [0-9]* [0-9]* //g" | grep "nova" | grep "\-A" |  sed "s/-A/-D/g" | awk '{print "sudo iptables",$0}' | bash
 | ||
|     # Delete nat rules
 | ||
|     sudo iptables -S -v -t nat | sed "s/-c [0-9]* [0-9]* //g" | grep "nova" |  grep "\-A" | sed "s/-A/-D/g" | awk '{print "sudo iptables -t nat",$0}' | bash
 | ||
|     # Delete chains
 | ||
|     sudo iptables -S -v | sed "s/-c [0-9]* [0-9]* //g" | grep "nova" | grep "\-N" |  sed "s/-N/-X/g" | awk '{print "sudo iptables",$0}' | bash
 | ||
|     # Delete nat chains
 | ||
|     sudo iptables -S -v -t nat | sed "s/-c [0-9]* [0-9]* //g" | grep "nova" |  grep "\-N" | sed "s/-N/-X/g" | awk '{print "sudo iptables -t nat",$0}' | bash
 | ||
| }
 | ||
| 
 | ||
| if is_service_enabled n-cpu; then
 | ||
| 
 | ||
|     # Force IP forwarding on, just on case
 | ||
|     sudo sysctl -w net.ipv4.ip_forward=1
 | ||
| 
 | ||
|     # attempt to load modules: network block device - used to manage qcow images
 | ||
|     sudo modprobe nbd || true
 | ||
| 
 | ||
|     # Check for kvm (hardware based virtualization).  If unable to initialize
 | ||
|     # kvm, we drop back to the slower emulation mode (qemu).  Note: many systems
 | ||
|     # come with hardware virtualization disabled in BIOS.
 | ||
|     if [[ "$LIBVIRT_TYPE" == "kvm" ]]; then
 | ||
|         sudo modprobe kvm || true
 | ||
|         if [ ! -e /dev/kvm ]; then
 | ||
|             echo "WARNING: Switching to QEMU"
 | ||
|             LIBVIRT_TYPE=qemu
 | ||
|         fi
 | ||
|     fi
 | ||
| 
 | ||
|     # Install and configure **LXC** if specified.  LXC is another approach to
 | ||
|     # splitting a system into many smaller parts.  LXC uses cgroups and chroot
 | ||
|     # to simulate multiple systems.
 | ||
|     if [[ "$LIBVIRT_TYPE" == "lxc" ]]; then
 | ||
|         if [[ "$os_PACKAGE" = "deb" ]]; then
 | ||
|             if [[ ! "$DISTRO" > natty ]]; then
 | ||
|                 cgline="none /cgroup cgroup cpuacct,memory,devices,cpu,freezer,blkio 0 0"
 | ||
|                 sudo mkdir -p /cgroup
 | ||
|                 if ! grep -q cgroup /etc/fstab; then
 | ||
|                     echo "$cgline" | sudo tee -a /etc/fstab
 | ||
|                 fi
 | ||
|                 if ! mount -n | grep -q cgroup; then
 | ||
|                     sudo mount /cgroup
 | ||
|                 fi
 | ||
|             fi
 | ||
|         fi
 | ||
|     fi
 | ||
| 
 | ||
|     QEMU_CONF=/etc/libvirt/qemu.conf
 | ||
|     if is_service_enabled quantum && [[ $Q_PLUGIN = "openvswitch" ]] && ! sudo grep -q '^cgroup_device_acl' $QEMU_CONF ; then
 | ||
|         # Add /dev/net/tun to cgroup_device_acls, needed for type=ethernet interfaces
 | ||
|         sudo chmod 666 $QEMU_CONF
 | ||
|         sudo cat <<EOF >> /etc/libvirt/qemu.conf
 | ||
| cgroup_device_acl = [
 | ||
|     "/dev/null", "/dev/full", "/dev/zero",
 | ||
|     "/dev/random", "/dev/urandom",
 | ||
|     "/dev/ptmx", "/dev/kvm", "/dev/kqemu",
 | ||
|     "/dev/rtc", "/dev/hpet","/dev/net/tun",
 | ||
| ]
 | ||
| EOF
 | ||
|         sudo chmod 644 $QEMU_CONF
 | ||
|     fi
 | ||
| 
 | ||
|     if [[ "$os_PACKAGE" = "deb" ]]; then
 | ||
|         LIBVIRT_DAEMON=libvirt-bin
 | ||
|     else
 | ||
|         # http://wiki.libvirt.org/page/SSHPolicyKitSetup
 | ||
|         if ! grep ^libvirtd: /etc/group >/dev/null; then
 | ||
|             sudo groupadd libvirtd
 | ||
|         fi
 | ||
|         sudo bash -c 'cat <<EOF >/etc/polkit-1/localauthority/50-local.d/50-libvirt-remote-access.pkla
 | ||
| [libvirt Management Access]
 | ||
| Identity=unix-group:libvirtd
 | ||
| Action=org.libvirt.unix.manage
 | ||
| ResultAny=yes
 | ||
| ResultInactive=yes
 | ||
| ResultActive=yes
 | ||
| EOF'
 | ||
|         LIBVIRT_DAEMON=libvirtd
 | ||
|     fi
 | ||
|     # The user that nova runs as needs to be member of libvirtd group otherwise
 | ||
|     # nova-compute will be unable to use libvirt.
 | ||
|     sudo usermod -a -G libvirtd `whoami`
 | ||
|     # libvirt detects various settings on startup, as we potentially changed
 | ||
|     # the system configuration (modules, filesystems), we need to restart
 | ||
|     # libvirt to detect those changes.
 | ||
|     restart_service $LIBVIRT_DAEMON
 | ||
| 
 | ||
| 
 | ||
|     # Instance Storage
 | ||
|     # ~~~~~~~~~~~~~~~~
 | ||
| 
 | ||
|     # Nova stores each instance in its own directory.
 | ||
|     mkdir -p $NOVA_DIR/instances
 | ||
| 
 | ||
|     # You can specify a different disk to be mounted and used for backing the
 | ||
|     # virtual machines.  If there is a partition labeled nova-instances we
 | ||
|     # mount it (ext filesystems can be labeled via e2label).
 | ||
|     if [ -L /dev/disk/by-label/nova-instances ]; then
 | ||
|         if ! mount -n | grep -q $NOVA_DIR/instances; then
 | ||
|             sudo mount -L nova-instances $NOVA_DIR/instances
 | ||
|             sudo chown -R `whoami` $NOVA_DIR/instances
 | ||
|         fi
 | ||
|     fi
 | ||
| 
 | ||
|     # Clean iptables from previous runs
 | ||
|     clean_iptables
 | ||
| 
 | ||
|     # Destroy old instances
 | ||
|     instances=`sudo virsh list --all | grep $INSTANCE_NAME_PREFIX | sed "s/.*\($INSTANCE_NAME_PREFIX[0-9a-fA-F]*\).*/\1/g"`
 | ||
|     if [ ! "$instances" = "" ]; then
 | ||
|         echo $instances | xargs -n1 sudo virsh destroy || true
 | ||
|         echo $instances | xargs -n1 sudo virsh undefine || true
 | ||
|     fi
 | ||
| 
 | ||
|     # Logout and delete iscsi sessions
 | ||
|     sudo iscsiadm --mode node | grep $VOLUME_NAME_PREFIX | cut -d " " -f2 | xargs sudo iscsiadm --mode node --logout || true
 | ||
|     sudo iscsiadm --mode node | grep $VOLUME_NAME_PREFIX | cut -d " " -f2 | sudo iscsiadm --mode node --op delete || true
 | ||
| 
 | ||
|     # Clean out the instances directory.
 | ||
|     sudo rm -rf $NOVA_DIR/instances/*
 | ||
| fi
 | ||
| 
 | ||
| if is_service_enabled n-net q-dhcp; then
 | ||
|     # Delete traces of nova networks from prior runs
 | ||
|     sudo killall dnsmasq || true
 | ||
|     clean_iptables
 | ||
|     rm -rf $NOVA_DIR/networks
 | ||
|     mkdir -p $NOVA_DIR/networks
 | ||
| 
 | ||
|     # Force IP forwarding on, just on case
 | ||
|     sudo sysctl -w net.ipv4.ip_forward=1
 | ||
| fi
 | ||
| 
 | ||
| 
 | ||
| # Storage Service
 | ||
| # ---------------
 | ||
| 
 | ||
| if is_service_enabled swift; then
 | ||
| 
 | ||
|     # We make sure to kill all swift processes first
 | ||
|     swift-init all stop || true
 | ||
| 
 | ||
|     # We first do a bit of setup by creating the directories and
 | ||
|     # changing the permissions so we can run it as our user.
 | ||
| 
 | ||
|     USER_GROUP=$(id -g)
 | ||
|     sudo mkdir -p ${SWIFT_DATA_DIR}/drives
 | ||
|     sudo chown -R $USER:${USER_GROUP} ${SWIFT_DATA_DIR}
 | ||
| 
 | ||
|     # We then create a loopback disk and format it to XFS.
 | ||
|     if [[ -e ${SWIFT_DATA_DIR}/drives/images/swift.img ]]; then
 | ||
|         if egrep -q ${SWIFT_DATA_DIR}/drives/sdb1 /proc/mounts; then
 | ||
|             sudo umount ${SWIFT_DATA_DIR}/drives/sdb1
 | ||
|         fi
 | ||
|     else
 | ||
|         mkdir -p  ${SWIFT_DATA_DIR}/drives/images
 | ||
|         sudo touch  ${SWIFT_DATA_DIR}/drives/images/swift.img
 | ||
|         sudo chown $USER: ${SWIFT_DATA_DIR}/drives/images/swift.img
 | ||
| 
 | ||
|         dd if=/dev/zero of=${SWIFT_DATA_DIR}/drives/images/swift.img \
 | ||
|             bs=1024 count=0 seek=${SWIFT_LOOPBACK_DISK_SIZE}
 | ||
|     fi
 | ||
|     # Make a fresh XFS filesystem
 | ||
|     mkfs.xfs -f -i size=1024  ${SWIFT_DATA_DIR}/drives/images/swift.img
 | ||
| 
 | ||
|     # After the drive being created we mount the disk with a few mount
 | ||
|     # options to make it most efficient as possible for swift.
 | ||
|     mkdir -p ${SWIFT_DATA_DIR}/drives/sdb1
 | ||
|     if ! egrep -q ${SWIFT_DATA_DIR}/drives/sdb1 /proc/mounts; then
 | ||
|         sudo mount -t xfs -o loop,noatime,nodiratime,nobarrier,logbufs=8  \
 | ||
|             ${SWIFT_DATA_DIR}/drives/images/swift.img ${SWIFT_DATA_DIR}/drives/sdb1
 | ||
|     fi
 | ||
| 
 | ||
|     # We then create link to that mounted location so swift would know
 | ||
|     # where to go.
 | ||
|     for x in $(seq ${SWIFT_REPLICAS}); do
 | ||
|         sudo ln -sf ${SWIFT_DATA_DIR}/drives/sdb1/$x ${SWIFT_DATA_DIR}/$x; done
 | ||
| 
 | ||
|     # We now have to emulate a few different servers into one we
 | ||
|     # create all the directories needed for swift
 | ||
|     for x in $(seq ${SWIFT_REPLICAS}); do
 | ||
|             drive=${SWIFT_DATA_DIR}/drives/sdb1/${x}
 | ||
|             node=${SWIFT_DATA_DIR}/${x}/node
 | ||
|             node_device=${node}/sdb1
 | ||
|             [[ -d $node ]] && continue
 | ||
|             [[ -d $drive ]] && continue
 | ||
|             sudo install -o ${USER} -g $USER_GROUP -d $drive
 | ||
|             sudo install -o ${USER} -g $USER_GROUP -d $node_device
 | ||
|             sudo chown -R $USER: ${node}
 | ||
|     done
 | ||
| 
 | ||
|    sudo mkdir -p ${SWIFT_CONFIG_DIR}/{object,container,account}-server /var/run/swift
 | ||
|    sudo chown -R $USER: ${SWIFT_CONFIG_DIR} /var/run/swift
 | ||
| 
 | ||
|     if [[ "$SWIFT_CONFIG_DIR" != "/etc/swift" ]]; then
 | ||
|         # Some swift tools are hard-coded to use /etc/swift and are apparenty not going to be fixed.
 | ||
|         # Create a symlink if the config dir is moved
 | ||
|         sudo ln -sf ${SWIFT_CONFIG_DIR} /etc/swift
 | ||
|     fi
 | ||
| 
 | ||
|     # Swift use rsync to syncronize between all the different
 | ||
|     # partitions (which make more sense when you have a multi-node
 | ||
|     # setup) we configure it with our version of rsync.
 | ||
|     sed -e "
 | ||
|         s/%GROUP%/${USER_GROUP}/;
 | ||
|         s/%USER%/$USER/;
 | ||
|         s,%SWIFT_DATA_DIR%,$SWIFT_DATA_DIR,;
 | ||
|     " $FILES/swift/rsyncd.conf | sudo tee /etc/rsyncd.conf
 | ||
|     if [[ "$os_PACKAGE" = "deb" ]]; then
 | ||
|         sudo sed -i '/^RSYNC_ENABLE=false/ { s/false/true/ }' /etc/default/rsync
 | ||
|     else
 | ||
|         sudo sed -i '/disable *= *yes/ { s/yes/no/ }' /etc/xinetd.d/rsync
 | ||
|     fi
 | ||
| 
 | ||
|     if is_service_enabled swift3;then
 | ||
|         swift_auth_server="s3token "
 | ||
|     fi
 | ||
| 
 | ||
|     # By default Swift will be installed with the tempauth middleware
 | ||
|     # which has some default username and password if you have
 | ||
|     # configured keystone it will checkout the directory.
 | ||
|     if is_service_enabled key; then
 | ||
|         swift_auth_server+="authtoken keystoneauth"
 | ||
|     else
 | ||
|         swift_auth_server=tempauth
 | ||
|     fi
 | ||
| 
 | ||
|     SWIFT_CONFIG_PROXY_SERVER=${SWIFT_CONFIG_DIR}/proxy-server.conf
 | ||
|     cp ${SWIFT_DIR}/etc/proxy-server.conf-sample ${SWIFT_CONFIG_PROXY_SERVER}
 | ||
| 
 | ||
|     iniuncomment ${SWIFT_CONFIG_PROXY_SERVER} DEFAULT user
 | ||
|     iniset ${SWIFT_CONFIG_PROXY_SERVER} DEFAULT user ${USER}
 | ||
| 
 | ||
|     iniuncomment ${SWIFT_CONFIG_PROXY_SERVER} DEFAULT swift_dir
 | ||
|     iniset ${SWIFT_CONFIG_PROXY_SERVER} DEFAULT swift_dir ${SWIFT_CONFIG_DIR}
 | ||
| 
 | ||
|     iniuncomment ${SWIFT_CONFIG_PROXY_SERVER} DEFAULT workers
 | ||
|     iniset ${SWIFT_CONFIG_PROXY_SERVER} DEFAULT workers 1
 | ||
| 
 | ||
|     iniuncomment ${SWIFT_CONFIG_PROXY_SERVER} DEFAULT log_level
 | ||
|     iniset ${SWIFT_CONFIG_PROXY_SERVER} DEFAULT log_level DEBUG
 | ||
| 
 | ||
|     iniuncomment ${SWIFT_CONFIG_PROXY_SERVER} DEFAULT bind_port
 | ||
|     iniset ${SWIFT_CONFIG_PROXY_SERVER} DEFAULT bind_port ${SWIFT_DEFAULT_BIND_PORT:-8080}
 | ||
| 
 | ||
|     # Only enable Swift3 if we have it enabled in ENABLED_SERVICES
 | ||
|     is_service_enabled swift3 && swift3=swift3 || swift3=""
 | ||
| 
 | ||
|     iniset ${SWIFT_CONFIG_PROXY_SERVER} pipeline:main pipeline "catch_errors healthcheck cache ratelimit ${swift3} ${swift_auth_server} proxy-logging proxy-server"
 | ||
| 
 | ||
|     iniset ${SWIFT_CONFIG_PROXY_SERVER} app:proxy-server account_autocreate true
 | ||
| 
 | ||
|     # Configure Keystone
 | ||
|     sed -i '/^# \[filter:authtoken\]/,/^# \[filter:keystoneauth\]$/ s/^#[ \t]*//' ${SWIFT_CONFIG_PROXY_SERVER}
 | ||
|     iniset ${SWIFT_CONFIG_PROXY_SERVER} filter:authtoken auth_host $KEYSTONE_AUTH_HOST
 | ||
|     iniset ${SWIFT_CONFIG_PROXY_SERVER} filter:authtoken auth_port $KEYSTONE_AUTH_PORT
 | ||
|     iniset ${SWIFT_CONFIG_PROXY_SERVER} filter:authtoken auth_protocol $KEYSTONE_AUTH_PROTOCOL
 | ||
|     iniset ${SWIFT_CONFIG_PROXY_SERVER} filter:authtoken auth_uri $KEYSTONE_SERVICE_PROTOCOL://$KEYSTONE_SERVICE_HOST:$KEYSTONE_SERVICE_PORT/
 | ||
|     iniset ${SWIFT_CONFIG_PROXY_SERVER} filter:authtoken admin_tenant_name $SERVICE_TENANT_NAME
 | ||
|     iniset ${SWIFT_CONFIG_PROXY_SERVER} filter:authtoken admin_user swift
 | ||
|     iniset ${SWIFT_CONFIG_PROXY_SERVER} filter:authtoken admin_password $SERVICE_PASSWORD
 | ||
| 
 | ||
|     iniuncomment ${SWIFT_CONFIG_PROXY_SERVER} filter:keystoneauth use
 | ||
|     iniuncomment ${SWIFT_CONFIG_PROXY_SERVER} filter:keystoneauth operator_roles
 | ||
|     iniset ${SWIFT_CONFIG_PROXY_SERVER} filter:keystoneauth operator_roles "Member, admin"
 | ||
| 
 | ||
|     if is_service_enabled swift3;then
 | ||
|         cat <<EOF>>${SWIFT_CONFIG_PROXY_SERVER}
 | ||
| # NOTE(chmou): s3token middleware is not updated yet to use only
 | ||
| # username and password.
 | ||
| [filter:s3token]
 | ||
| paste.filter_factory = keystone.middleware.s3_token:filter_factory
 | ||
| auth_port = ${KEYSTONE_AUTH_PORT}
 | ||
| auth_host = ${KEYSTONE_AUTH_HOST}
 | ||
| auth_protocol = ${KEYSTONE_AUTH_PROTOCOL}
 | ||
| auth_token = ${SERVICE_TOKEN}
 | ||
| admin_token = ${SERVICE_TOKEN}
 | ||
| 
 | ||
| [filter:swift3]
 | ||
| use = egg:swift3#swift3
 | ||
| EOF
 | ||
|     fi
 | ||
| 
 | ||
|     cp ${SWIFT_DIR}/etc/swift.conf-sample ${SWIFT_CONFIG_DIR}/swift.conf
 | ||
|     iniset ${SWIFT_CONFIG_DIR}/swift.conf swift-hash swift_hash_path_suffix ${SWIFT_HASH}
 | ||
| 
 | ||
|     # We need to generate a object/account/proxy configuration
 | ||
|     # emulating 4 nodes on different ports we have a little function
 | ||
|     # that help us doing that.
 | ||
|     function generate_swift_configuration() {
 | ||
|         local server_type=$1
 | ||
|         local bind_port=$2
 | ||
|         local log_facility=$3
 | ||
|         local node_number
 | ||
|         local swift_node_config
 | ||
| 
 | ||
|         for node_number in $(seq ${SWIFT_REPLICAS}); do
 | ||
|             node_path=${SWIFT_DATA_DIR}/${node_number}
 | ||
|             swift_node_config=${SWIFT_CONFIG_DIR}/${server_type}-server/${node_number}.conf
 | ||
| 
 | ||
|             cp ${SWIFT_DIR}/etc/${server_type}-server.conf-sample ${swift_node_config}
 | ||
| 
 | ||
|             iniuncomment ${swift_node_config} DEFAULT user
 | ||
|             iniset ${swift_node_config} DEFAULT user ${USER}
 | ||
| 
 | ||
|             iniuncomment ${swift_node_config} DEFAULT bind_port
 | ||
|             iniset ${swift_node_config} DEFAULT bind_port ${bind_port}
 | ||
| 
 | ||
|             iniuncomment ${swift_node_config} DEFAULT swift_dir
 | ||
|             iniset ${swift_node_config} DEFAULT swift_dir ${SWIFT_CONFIG_DIR}
 | ||
| 
 | ||
|             iniuncomment ${swift_node_config} DEFAULT devices
 | ||
|             iniset ${swift_node_config} DEFAULT devices ${node_path}
 | ||
| 
 | ||
|             iniuncomment ${swift_node_config} DEFAULT log_facility
 | ||
|             iniset ${swift_node_config} DEFAULT log_facility LOG_LOCAL${log_facility}
 | ||
| 
 | ||
|             iniuncomment ${swift_node_config} DEFAULT mount_check
 | ||
|             iniset ${swift_node_config} DEFAULT mount_check false
 | ||
| 
 | ||
|             iniuncomment ${swift_node_config} ${server_type}-replicator vm_test_mode
 | ||
|             iniset ${swift_node_config} ${server_type}-replicator vm_test_mode yes
 | ||
| 
 | ||
|             bind_port=$(( ${bind_port} + 10 ))
 | ||
|             log_facility=$(( ${log_facility} + 1 ))
 | ||
|         done
 | ||
|     }
 | ||
|     generate_swift_configuration object 6010 2
 | ||
|     generate_swift_configuration container 6011 2
 | ||
|     generate_swift_configuration account 6012 2
 | ||
| 
 | ||
|     # We have some specific configuration for swift for rsyslog. See
 | ||
|     # the file /etc/rsyslog.d/10-swift.conf for more info.
 | ||
|     swift_log_dir=${SWIFT_DATA_DIR}/logs
 | ||
|     rm -rf ${swift_log_dir}
 | ||
|     mkdir -p ${swift_log_dir}/hourly
 | ||
|     sudo chown -R $USER:adm ${swift_log_dir}
 | ||
|     sed "s,%SWIFT_LOGDIR%,${swift_log_dir}," $FILES/swift/rsyslog.conf | sudo \
 | ||
|         tee /etc/rsyslog.d/10-swift.conf
 | ||
|     restart_service rsyslog
 | ||
| 
 | ||
|     # This is where we create three different rings for swift with
 | ||
|     # different object servers binding on different ports.
 | ||
|     pushd ${SWIFT_CONFIG_DIR} >/dev/null && {
 | ||
| 
 | ||
|         rm -f *.builder *.ring.gz backups/*.builder backups/*.ring.gz
 | ||
| 
 | ||
|         port_number=6010
 | ||
|         swift-ring-builder object.builder create ${SWIFT_PARTITION_POWER_SIZE} ${SWIFT_REPLICAS} 1
 | ||
|         for x in $(seq ${SWIFT_REPLICAS}); do
 | ||
|             swift-ring-builder object.builder add z${x}-127.0.0.1:${port_number}/sdb1 1
 | ||
|             port_number=$[port_number + 10]
 | ||
|         done
 | ||
|         swift-ring-builder object.builder rebalance
 | ||
| 
 | ||
|         port_number=6011
 | ||
|         swift-ring-builder container.builder create ${SWIFT_PARTITION_POWER_SIZE} ${SWIFT_REPLICAS} 1
 | ||
|         for x in $(seq ${SWIFT_REPLICAS}); do
 | ||
|             swift-ring-builder container.builder add z${x}-127.0.0.1:${port_number}/sdb1 1
 | ||
|             port_number=$[port_number + 10]
 | ||
|         done
 | ||
|         swift-ring-builder container.builder rebalance
 | ||
| 
 | ||
|         port_number=6012
 | ||
|         swift-ring-builder account.builder create ${SWIFT_PARTITION_POWER_SIZE} ${SWIFT_REPLICAS} 1
 | ||
|         for x in $(seq ${SWIFT_REPLICAS}); do
 | ||
|             swift-ring-builder account.builder add z${x}-127.0.0.1:${port_number}/sdb1 1
 | ||
|             port_number=$[port_number + 10]
 | ||
|         done
 | ||
|         swift-ring-builder account.builder rebalance
 | ||
| 
 | ||
|     } && popd >/dev/null
 | ||
| 
 | ||
|    # We then can start rsync.
 | ||
|     if [[ "$os_PACKAGE" = "deb" ]]; then
 | ||
|         sudo /etc/init.d/rsync restart || :
 | ||
|     else
 | ||
|         sudo systemctl start xinetd.service
 | ||
|     fi
 | ||
| 
 | ||
|    # First spawn all the swift services then kill the
 | ||
|    # proxy service so we can run it in foreground in screen.
 | ||
|    # ``swift-init ... {stop|restart}`` exits with '1' if no servers are running,
 | ||
|    # ignore it just in case
 | ||
|    swift-init all restart || true
 | ||
|    swift-init proxy stop || true
 | ||
| 
 | ||
|    unset s swift_hash swift_auth_server
 | ||
| fi
 | ||
| 
 | ||
| 
 | ||
| # Volume Service
 | ||
| # --------------
 | ||
| 
 | ||
| if is_service_enabled cinder; then
 | ||
|     init_cinder
 | ||
| elif is_service_enabled n-vol; then
 | ||
|     # Configure a default volume group called '`stack-volumes`' for the volume
 | ||
|     # service if it does not yet exist.  If you don't wish to use a file backed
 | ||
|     # volume group, create your own volume group called ``stack-volumes`` before
 | ||
|     # invoking ``stack.sh``.
 | ||
|     #
 | ||
|     # By default, the backing file is 5G in size, and is stored in ``/opt/stack/data``.
 | ||
| 
 | ||
|     if ! sudo vgs $VOLUME_GROUP; then
 | ||
|         VOLUME_BACKING_FILE=${VOLUME_BACKING_FILE:-$DATA_DIR/${VOLUME_GROUP}-backing-file}
 | ||
|         # Only create if the file doesn't already exists
 | ||
|         [[ -f $VOLUME_BACKING_FILE ]] || truncate -s $VOLUME_BACKING_FILE_SIZE $VOLUME_BACKING_FILE
 | ||
|         DEV=`sudo losetup -f --show $VOLUME_BACKING_FILE`
 | ||
|         # Only create if the loopback device doesn't contain $VOLUME_GROUP
 | ||
|         if ! sudo vgs $VOLUME_GROUP; then sudo vgcreate $VOLUME_GROUP $DEV; fi
 | ||
|     fi
 | ||
| 
 | ||
|     if sudo vgs $VOLUME_GROUP; then
 | ||
|         if [[ "$os_PACKAGE" = "rpm" ]]; then
 | ||
|             # RPM doesn't start the service
 | ||
|             start_service tgtd
 | ||
|         fi
 | ||
| 
 | ||
|         # Setup tgtd configuration files
 | ||
|         mkdir -p $NOVA_DIR/volumes
 | ||
| 
 | ||
|         # Remove nova iscsi targets
 | ||
|         sudo tgtadm --op show --mode target | grep $VOLUME_NAME_PREFIX | grep Target | cut -f3 -d ' ' | sudo xargs -n1 tgt-admin --delete || true
 | ||
|         # Clean out existing volumes
 | ||
|         for lv in `sudo lvs --noheadings -o lv_name $VOLUME_GROUP`; do
 | ||
|             # VOLUME_NAME_PREFIX prefixes the LVs we want
 | ||
|             if [[ "${lv#$VOLUME_NAME_PREFIX}" != "$lv" ]]; then
 | ||
|                 sudo lvremove -f $VOLUME_GROUP/$lv
 | ||
|             fi
 | ||
|         done
 | ||
|     fi
 | ||
| 
 | ||
|     if [[ "$os_PACKAGE" = "deb" ]]; then
 | ||
| 
 | ||
|         # Setup the tgt configuration file
 | ||
|         if [[ ! -f /etc/tgt/conf.d/nova.conf ]]; then
 | ||
|            sudo mkdir -p /etc/tgt/conf.d
 | ||
|            echo "include $NOVA_DIR/volumes/*" | sudo tee /etc/tgt/conf.d/nova.conf
 | ||
|         fi
 | ||
| 
 | ||
|         # tgt in oneiric doesn't restart properly if tgtd isn't running
 | ||
|         # do it in two steps
 | ||
|         sudo stop tgt || true
 | ||
|         sudo start tgt
 | ||
|     else
 | ||
|         restart_service tgtd
 | ||
|     fi
 | ||
| fi
 | ||
| 
 | ||
| # Support entry points installation of console scripts
 | ||
| if [ -d $NOVA_DIR/bin ] ; then
 | ||
|     NOVA_BIN_DIR=$NOVA_DIR/bin
 | ||
| else
 | ||
|     NOVA_BIN_DIR=/usr/local/bin
 | ||
| fi
 | ||
| 
 | ||
| NOVA_CONF=nova.conf
 | ||
| function add_nova_opt {
 | ||
|     echo "$1" >> $NOVA_CONF_DIR/$NOVA_CONF
 | ||
| }
 | ||
| 
 | ||
| # Remove legacy nova.conf
 | ||
| rm -f $NOVA_DIR/bin/nova.conf
 | ||
| 
 | ||
| # (re)create nova.conf
 | ||
| rm -f $NOVA_CONF_DIR/$NOVA_CONF
 | ||
| add_nova_opt "[DEFAULT]"
 | ||
| add_nova_opt "verbose=True"
 | ||
| add_nova_opt "auth_strategy=keystone"
 | ||
| add_nova_opt "allow_resize_to_same_host=True"
 | ||
| add_nova_opt "root_helper=sudo $NOVA_ROOTWRAP"
 | ||
| add_nova_opt "compute_scheduler_driver=$SCHEDULER"
 | ||
| add_nova_opt "dhcpbridge_flagfile=$NOVA_CONF_DIR/$NOVA_CONF"
 | ||
| add_nova_opt "fixed_range=$FIXED_RANGE"
 | ||
| add_nova_opt "s3_host=$SERVICE_HOST"
 | ||
| add_nova_opt "s3_port=$S3_SERVICE_PORT"
 | ||
| if is_service_enabled quantum; then
 | ||
|     add_nova_opt "network_api_class=nova.network.quantumv2.api.API"
 | ||
|     add_nova_opt "quantum_admin_username=$Q_ADMIN_USERNAME"
 | ||
|     add_nova_opt "quantum_admin_password=$SERVICE_PASSWORD"
 | ||
|     add_nova_opt "quantum_admin_auth_url=$KEYSTONE_SERVICE_PROTOCOL://$KEYSTONE_SERVICE_HOST:$KEYSTONE_AUTH_PORT/v2.0"
 | ||
|     add_nova_opt "quantum_auth_strategy=$Q_AUTH_STRATEGY"
 | ||
|     add_nova_opt "quantum_admin_tenant_name=$SERVICE_TENANT_NAME"
 | ||
|     add_nova_opt "quantum_url=http://$Q_HOST:$Q_PORT"
 | ||
| 
 | ||
|     if [[ "$Q_PLUGIN" = "openvswitch" ]]; then
 | ||
|         NOVA_VIF_DRIVER="nova.virt.libvirt.vif.LibvirtOpenVswitchDriver"
 | ||
|         LINUXNET_VIF_DRIVER="nova.network.linux_net.LinuxOVSInterfaceDriver"
 | ||
|     elif [[ "$Q_PLUGIN" = "linuxbridge" ]]; then
 | ||
|         NOVA_VIF_DRIVER="nova.virt.libvirt.vif.QuantumLinuxBridgeVIFDriver"
 | ||
|         LINUXNET_VIF_DRIVER="nova.network.linux_net.QuantumLinuxBridgeInterfaceDriver"
 | ||
|     fi
 | ||
|     add_nova_opt "libvirt_vif_type=ethernet"
 | ||
|     add_nova_opt "libvirt_vif_driver=$NOVA_VIF_DRIVER"
 | ||
|     add_nova_opt "linuxnet_interface_driver=$LINUXNET_VIF_DRIVER"
 | ||
| else
 | ||
|     add_nova_opt "network_manager=nova.network.manager.$NET_MAN"
 | ||
| fi
 | ||
| if is_service_enabled n-vol; then
 | ||
|     add_nova_opt "volume_group=$VOLUME_GROUP"
 | ||
|     add_nova_opt "volume_name_template=${VOLUME_NAME_PREFIX}%s"
 | ||
|     # oneiric no longer supports ietadm
 | ||
|     add_nova_opt "iscsi_helper=tgtadm"
 | ||
| fi
 | ||
| add_nova_opt "osapi_compute_extension=nova.api.openstack.compute.contrib.standard_extensions"
 | ||
| add_nova_opt "my_ip=$HOST_IP"
 | ||
| add_nova_opt "public_interface=$PUBLIC_INTERFACE"
 | ||
| add_nova_opt "vlan_interface=$VLAN_INTERFACE"
 | ||
| add_nova_opt "flat_network_bridge=$FLAT_NETWORK_BRIDGE"
 | ||
| if [ -n "$FLAT_INTERFACE" ]; then
 | ||
|     add_nova_opt "flat_interface=$FLAT_INTERFACE"
 | ||
| fi
 | ||
| add_nova_opt "sql_connection=$BASE_SQL_CONN/nova?charset=utf8"
 | ||
| add_nova_opt "libvirt_type=$LIBVIRT_TYPE"
 | ||
| add_nova_opt "libvirt_cpu_mode=none"
 | ||
| add_nova_opt "instance_name_template=${INSTANCE_NAME_PREFIX}%08x"
 | ||
| # All nova-compute workers need to know the vnc configuration options
 | ||
| # These settings don't hurt anything if n-xvnc and n-novnc are disabled
 | ||
| if is_service_enabled n-cpu; then
 | ||
|     NOVNCPROXY_URL=${NOVNCPROXY_URL:-"http://$SERVICE_HOST:6080/vnc_auto.html"}
 | ||
|     add_nova_opt "novncproxy_base_url=$NOVNCPROXY_URL"
 | ||
|     XVPVNCPROXY_URL=${XVPVNCPROXY_URL:-"http://$SERVICE_HOST:6081/console"}
 | ||
|     add_nova_opt "xvpvncproxy_base_url=$XVPVNCPROXY_URL"
 | ||
| fi
 | ||
| if [ "$VIRT_DRIVER" = 'xenserver' ]; then
 | ||
|     VNCSERVER_PROXYCLIENT_ADDRESS=${VNCSERVER_PROXYCLIENT_ADDRESS=169.254.0.1}
 | ||
| else
 | ||
|     VNCSERVER_PROXYCLIENT_ADDRESS=${VNCSERVER_PROXYCLIENT_ADDRESS=127.0.0.1}
 | ||
| fi
 | ||
| # Address on which instance vncservers will listen on compute hosts.
 | ||
| # For multi-host, this should be the management ip of the compute host.
 | ||
| VNCSERVER_LISTEN=${VNCSERVER_LISTEN=127.0.0.1}
 | ||
| add_nova_opt "vncserver_listen=$VNCSERVER_LISTEN"
 | ||
| add_nova_opt "vncserver_proxyclient_address=$VNCSERVER_PROXYCLIENT_ADDRESS"
 | ||
| add_nova_opt "api_paste_config=$NOVA_CONF_DIR/api-paste.ini"
 | ||
| add_nova_opt "image_service=nova.image.glance.GlanceImageService"
 | ||
| add_nova_opt "ec2_dmz_host=$EC2_DMZ_HOST"
 | ||
| if is_service_enabled qpid ; then
 | ||
|     add_nova_opt "rpc_backend=nova.rpc.impl_qpid"
 | ||
| elif [ -n "$RABBIT_HOST" ] &&  [ -n "$RABBIT_PASSWORD" ]; then
 | ||
|     add_nova_opt "rabbit_host=$RABBIT_HOST"
 | ||
|     add_nova_opt "rabbit_password=$RABBIT_PASSWORD"
 | ||
| fi
 | ||
| add_nova_opt "glance_api_servers=$GLANCE_HOSTPORT"
 | ||
| add_nova_opt "force_dhcp_release=True"
 | ||
| if [ -n "$INSTANCES_PATH" ]; then
 | ||
|     add_nova_opt "instances_path=$INSTANCES_PATH"
 | ||
| fi
 | ||
| if [ "$MULTI_HOST" != "False" ]; then
 | ||
|     add_nova_opt "multi_host=True"
 | ||
|     add_nova_opt "send_arp_for_ha=True"
 | ||
| fi
 | ||
| if [ "$SYSLOG" != "False" ]; then
 | ||
|     add_nova_opt "use_syslog=True"
 | ||
| fi
 | ||
| if [ "$API_RATE_LIMIT" != "True" ]; then
 | ||
|     add_nova_opt "api_rate_limit=False"
 | ||
| fi
 | ||
| if [ "$LOG_COLOR" == "True" ] && [ "$SYSLOG" == "False" ]; then
 | ||
|     # Add color to logging output
 | ||
|     add_nova_opt "logging_context_format_string=%(asctime)s %(color)s%(levelname)s %(name)s [[01;36m%(request_id)s [00;36m%(user_name)s %(project_name)s%(color)s] [01;35m%(instance)s%(color)s%(message)s[00m"
 | ||
|     add_nova_opt "logging_default_format_string=%(asctime)s %(color)s%(levelname)s %(name)s [[00;36m-%(color)s] [01;35m%(instance)s%(color)s%(message)s[00m"
 | ||
|     add_nova_opt "logging_debug_format_suffix=[00;33mfrom (pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d[00m"
 | ||
|     add_nova_opt "logging_exception_prefix=%(color)s%(asctime)s TRACE %(name)s [01;35m%(instance)s[00m"
 | ||
| else
 | ||
|     # Show user_name and project_name instead of user_id and project_id
 | ||
|     add_nova_opt "logging_context_format_string=%(asctime)s %(levelname)s %(name)s [%(request_id)s %(user_name)s %(project_name)s] %(instance)s%(message)s"
 | ||
| fi
 | ||
| 
 | ||
| # If cinder is enabled, use the cinder volume driver
 | ||
| if is_service_enabled cinder; then
 | ||
|     add_nova_opt "volume_api_class=nova.volume.cinder.API"
 | ||
| fi
 | ||
| 
 | ||
| # Provide some transition from EXTRA_FLAGS to EXTRA_OPTS
 | ||
| if [[ -z "$EXTRA_OPTS" && -n "$EXTRA_FLAGS" ]]; then
 | ||
|     EXTRA_OPTS=$EXTRA_FLAGS
 | ||
| fi
 | ||
| 
 | ||
| # You can define extra nova conf flags by defining the array EXTRA_OPTS,
 | ||
| # For Example: EXTRA_OPTS=(foo=true bar=2)
 | ||
| for I in "${EXTRA_OPTS[@]}"; do
 | ||
|     # Attempt to convert flags to options
 | ||
|     add_nova_opt ${I//--}
 | ||
| done
 | ||
| 
 | ||
| 
 | ||
| # XenServer
 | ||
| # ---------
 | ||
| 
 | ||
| if [ "$VIRT_DRIVER" = 'xenserver' ]; then
 | ||
|     read_password XENAPI_PASSWORD "ENTER A PASSWORD TO USE FOR XEN."
 | ||
|     add_nova_opt "compute_driver=xenapi.XenAPIDriver"
 | ||
|     XENAPI_CONNECTION_URL=${XENAPI_CONNECTION_URL:-"http://169.254.0.1"}
 | ||
|     XENAPI_USER=${XENAPI_USER:-"root"}
 | ||
|     add_nova_opt "xenapi_connection_url=$XENAPI_CONNECTION_URL"
 | ||
|     add_nova_opt "xenapi_connection_username=$XENAPI_USER"
 | ||
|     add_nova_opt "xenapi_connection_password=$XENAPI_PASSWORD"
 | ||
|     add_nova_opt "flat_injected=False"
 | ||
|     # Need to avoid crash due to new firewall support
 | ||
|     XEN_FIREWALL_DRIVER=${XEN_FIREWALL_DRIVER:-"nova.virt.firewall.IptablesFirewallDriver"}
 | ||
|     add_nova_opt "firewall_driver=$XEN_FIREWALL_DRIVER"
 | ||
| elif [ "$VIRT_DRIVER" = 'openvz' ]; then
 | ||
|     # TODO(deva): OpenVZ driver does not yet work if compute_driver is set here.
 | ||
|     #             Replace connection_type when this is fixed.
 | ||
|     #             add_nova_opt "compute_driver=openvz.connection.OpenVzConnection"
 | ||
|     add_nova_opt "connection_type=openvz"
 | ||
|     LIBVIRT_FIREWALL_DRIVER=${LIBVIRT_FIREWALL_DRIVER:-"nova.virt.libvirt.firewall.IptablesFirewallDriver"}
 | ||
|     add_nova_opt "firewall_driver=$LIBVIRT_FIREWALL_DRIVER"
 | ||
| else
 | ||
|     add_nova_opt "compute_driver=libvirt.LibvirtDriver"
 | ||
|     LIBVIRT_FIREWALL_DRIVER=${LIBVIRT_FIREWALL_DRIVER:-"nova.virt.libvirt.firewall.IptablesFirewallDriver"}
 | ||
|     add_nova_opt "firewall_driver=$LIBVIRT_FIREWALL_DRIVER"
 | ||
| fi
 | ||
| 
 | ||
| 
 | ||
| # Nova Database
 | ||
| # ~~~~~~~~~~~~~
 | ||
| 
 | ||
| # All nova components talk to a central database.  We will need to do this step
 | ||
| # only once for an entire cluster.
 | ||
| 
 | ||
| if is_service_enabled mysql && is_service_enabled nova; then
 | ||
|     # (re)create nova database
 | ||
|     mysql -u$MYSQL_USER -p$MYSQL_PASSWORD -e 'DROP DATABASE IF EXISTS nova;'
 | ||
|     # Explicitly use latin1: to avoid lp#829209, nova expects the database to
 | ||
|     # use latin1 by default, and then upgrades the database to utf8 (see the
 | ||
|     # 082_essex.py in nova)
 | ||
|     mysql -u$MYSQL_USER -p$MYSQL_PASSWORD -e 'CREATE DATABASE nova CHARACTER SET latin1;'
 | ||
| 
 | ||
|     # (re)create nova database
 | ||
|     $NOVA_BIN_DIR/nova-manage db sync
 | ||
| fi
 | ||
| 
 | ||
| 
 | ||
| # Launch Services
 | ||
| # ===============
 | ||
| 
 | ||
| # nova api crashes if we start it with a regular screen command,
 | ||
| # so send the start command by forcing text into the window.
 | ||
| # Only run the services specified in ``ENABLED_SERVICES``
 | ||
| 
 | ||
| # launch the glance registry service
 | ||
| if is_service_enabled g-reg; then
 | ||
|     screen_it g-reg "cd $GLANCE_DIR; bin/glance-registry --config-file=$GLANCE_CONF_DIR/glance-registry.conf"
 | ||
| fi
 | ||
| 
 | ||
| # launch the glance api and wait for it to answer before continuing
 | ||
| if is_service_enabled g-api; then
 | ||
|     screen_it g-api "cd $GLANCE_DIR; bin/glance-api --config-file=$GLANCE_CONF_DIR/glance-api.conf"
 | ||
|     echo "Waiting for g-api ($GLANCE_HOSTPORT) to start..."
 | ||
|     if ! timeout $SERVICE_TIMEOUT sh -c "while ! http_proxy= wget -q -O- http://$GLANCE_HOSTPORT; do sleep 1; done"; then
 | ||
|       echo "g-api did not start"
 | ||
|       exit 1
 | ||
|     fi
 | ||
| fi
 | ||
| 
 | ||
| if is_service_enabled key; then
 | ||
|     # (re)create keystone database
 | ||
|     mysql -u$MYSQL_USER -p$MYSQL_PASSWORD -e 'DROP DATABASE IF EXISTS keystone;'
 | ||
|     mysql -u$MYSQL_USER -p$MYSQL_PASSWORD -e 'CREATE DATABASE keystone CHARACTER SET utf8;'
 | ||
| 
 | ||
|     KEYSTONE_CONF_DIR=${KEYSTONE_CONF_DIR:-/etc/keystone}
 | ||
|     KEYSTONE_CONF=$KEYSTONE_CONF_DIR/keystone.conf
 | ||
|     KEYSTONE_CATALOG_BACKEND=${KEYSTONE_CATALOG_BACKEND:-template}
 | ||
| 
 | ||
|     if [[ ! -d $KEYSTONE_CONF_DIR ]]; then
 | ||
|         sudo mkdir -p $KEYSTONE_CONF_DIR
 | ||
|         sudo chown `whoami` $KEYSTONE_CONF_DIR
 | ||
|     fi
 | ||
| 
 | ||
|     if [[ "$KEYSTONE_CONF_DIR" != "$KEYSTONE_DIR/etc" ]]; then
 | ||
|         cp -p $KEYSTONE_DIR/etc/keystone.conf.sample $KEYSTONE_CONF
 | ||
|         cp -p $KEYSTONE_DIR/etc/policy.json $KEYSTONE_CONF_DIR
 | ||
|     fi
 | ||
| 
 | ||
|     # Rewrite stock keystone.conf:
 | ||
|     iniset $KEYSTONE_CONF DEFAULT admin_token "$SERVICE_TOKEN"
 | ||
|     iniset $KEYSTONE_CONF sql connection "$BASE_SQL_CONN/keystone?charset=utf8"
 | ||
|     iniset $KEYSTONE_CONF ec2 driver "keystone.contrib.ec2.backends.sql.Ec2"
 | ||
|     sed -e "
 | ||
|         /^pipeline.*ec2_extension crud_/s|ec2_extension crud_extension|ec2_extension s3_extension crud_extension|;
 | ||
|     " -i $KEYSTONE_CONF
 | ||
|     # Append the S3 bits
 | ||
|     iniset $KEYSTONE_CONF filter:s3_extension paste.filter_factory "keystone.contrib.s3:S3Extension.factory"
 | ||
| 
 | ||
|     if [[ "$KEYSTONE_CATALOG_BACKEND" = "sql" ]]; then
 | ||
|         # Configure keystone.conf to use sql
 | ||
|         iniset $KEYSTONE_CONF catalog driver keystone.catalog.backends.sql.Catalog
 | ||
|         inicomment $KEYSTONE_CONF catalog template_file
 | ||
|     else
 | ||
|         KEYSTONE_CATALOG=$KEYSTONE_CONF_DIR/default_catalog.templates
 | ||
|         cp -p $FILES/default_catalog.templates $KEYSTONE_CATALOG
 | ||
|         # Add swift endpoints to service catalog if swift is enabled
 | ||
|         if is_service_enabled swift; then
 | ||
|             echo "catalog.RegionOne.object_store.publicURL = http://%SERVICE_HOST%:8080/v1/AUTH_\$(tenant_id)s" >> $KEYSTONE_CATALOG
 | ||
|             echo "catalog.RegionOne.object_store.adminURL = http://%SERVICE_HOST%:8080/" >> $KEYSTONE_CATALOG
 | ||
|             echo "catalog.RegionOne.object_store.internalURL = http://%SERVICE_HOST%:8080/v1/AUTH_\$(tenant_id)s" >> $KEYSTONE_CATALOG
 | ||
|             echo "catalog.RegionOne.object_store.name = Swift Service" >> $KEYSTONE_CATALOG
 | ||
|         fi
 | ||
| 
 | ||
|         # Add quantum endpoints to service catalog if quantum is enabled
 | ||
|         if is_service_enabled quantum; then
 | ||
|             echo "catalog.RegionOne.network.publicURL = http://%SERVICE_HOST%:$Q_PORT/" >> $KEYSTONE_CATALOG
 | ||
|             echo "catalog.RegionOne.network.adminURL = http://%SERVICE_HOST%:$Q_PORT/" >> $KEYSTONE_CATALOG
 | ||
|             echo "catalog.RegionOne.network.internalURL = http://%SERVICE_HOST%:$Q_PORT/" >> $KEYSTONE_CATALOG
 | ||
|             echo "catalog.RegionOne.network.name = Quantum Service" >> $KEYSTONE_CATALOG
 | ||
|         fi
 | ||
| 
 | ||
|         sudo sed -e "
 | ||
|             s,%SERVICE_HOST%,$SERVICE_HOST,g;
 | ||
|             s,%S3_SERVICE_PORT%,$S3_SERVICE_PORT,g;
 | ||
|         " -i $KEYSTONE_CATALOG
 | ||
| 
 | ||
|         # Configure keystone.conf to use templates
 | ||
|         iniset $KEYSTONE_CONF catalog driver "keystone.catalog.backends.templated.TemplatedCatalog"
 | ||
|         iniset $KEYSTONE_CONF catalog template_file "$KEYSTONE_CATALOG"
 | ||
|     fi
 | ||
| 
 | ||
|     # Set up logging
 | ||
|     LOGGING_ROOT="devel"
 | ||
|     if [ "$SYSLOG" != "False" ]; then
 | ||
|         LOGGING_ROOT="$LOGGING_ROOT,production"
 | ||
|     fi
 | ||
|     KEYSTONE_LOG_CONFIG="--log-config $KEYSTONE_CONF_DIR/logging.conf"
 | ||
|     cp $KEYSTONE_DIR/etc/logging.conf.sample $KEYSTONE_CONF_DIR/logging.conf
 | ||
|     iniset $KEYSTONE_CONF_DIR/logging.conf logger_root level "DEBUG"
 | ||
|     iniset $KEYSTONE_CONF_DIR/logging.conf logger_root handlers "devel,production"
 | ||
| 
 | ||
|     # Initialize keystone database
 | ||
|     $KEYSTONE_DIR/bin/keystone-manage db_sync
 | ||
|     # set up certificates
 | ||
|     $KEYSTONE_DIR/bin/keystone-manage pki_setup
 | ||
| 
 | ||
|     # launch keystone and wait for it to answer before continuing
 | ||
|     screen_it key "cd $KEYSTONE_DIR && $KEYSTONE_DIR/bin/keystone-all --config-file $KEYSTONE_CONF $KEYSTONE_LOG_CONFIG -d --debug"
 | ||
|     echo "Waiting for keystone to start..."
 | ||
|     if ! timeout $SERVICE_TIMEOUT sh -c "while ! http_proxy= curl -s $KEYSTONE_AUTH_PROTOCOL://$SERVICE_HOST:$KEYSTONE_API_PORT/v2.0/ >/dev/null; do sleep 1; done"; then
 | ||
|       echo "keystone did not start"
 | ||
|       exit 1
 | ||
|     fi
 | ||
| 
 | ||
|     # keystone_data.sh creates services, admin and demo users, and roles.
 | ||
|     SERVICE_ENDPOINT=$KEYSTONE_AUTH_PROTOCOL://$KEYSTONE_AUTH_HOST:$KEYSTONE_AUTH_PORT/v2.0
 | ||
| 
 | ||
|     ADMIN_PASSWORD=$ADMIN_PASSWORD SERVICE_TENANT_NAME=$SERVICE_TENANT_NAME SERVICE_PASSWORD=$SERVICE_PASSWORD \
 | ||
|     SERVICE_TOKEN=$SERVICE_TOKEN SERVICE_ENDPOINT=$SERVICE_ENDPOINT SERVICE_HOST=$SERVICE_HOST \
 | ||
|     S3_SERVICE_PORT=$S3_SERVICE_PORT KEYSTONE_CATALOG_BACKEND=$KEYSTONE_CATALOG_BACKEND \
 | ||
|     DEVSTACK_DIR=$TOP_DIR ENABLED_SERVICES=$ENABLED_SERVICES \
 | ||
|         bash -x $FILES/keystone_data.sh
 | ||
| 
 | ||
|     # Set up auth creds now that keystone is bootstrapped
 | ||
|     export OS_AUTH_URL=$SERVICE_ENDPOINT
 | ||
|     export OS_TENANT_NAME=admin
 | ||
|     export OS_USERNAME=admin
 | ||
|     export OS_PASSWORD=$ADMIN_PASSWORD
 | ||
| 
 | ||
|     # Create an access key and secret key for nova ec2 register image
 | ||
|     if is_service_enabled swift3 && is_service_enabled nova; then
 | ||
|         NOVA_USER_ID=$(keystone user-list | grep ' nova ' | get_field 1)
 | ||
|         NOVA_TENANT_ID=$(keystone tenant-list | grep " $SERVICE_TENANT_NAME " | get_field 1)
 | ||
|         CREDS=$(keystone ec2-credentials-create --user_id $NOVA_USER_ID --tenant_id $NOVA_TENANT_ID)
 | ||
|         ACCESS_KEY=$(echo "$CREDS" | awk '/ access / { print $4 }')
 | ||
|         SECRET_KEY=$(echo "$CREDS" | awk '/ secret / { print $4 }')
 | ||
|         add_nova_opt "s3_access_key=$ACCESS_KEY"
 | ||
|         add_nova_opt "s3_secret_key=$SECRET_KEY"
 | ||
|         add_nova_opt "s3_affix_tenant=True"
 | ||
|     fi
 | ||
| fi
 | ||
| 
 | ||
| # Launch the nova-api and wait for it to answer before continuing
 | ||
| if is_service_enabled n-api; then
 | ||
|     add_nova_opt "enabled_apis=$NOVA_ENABLED_APIS"
 | ||
|     screen_it n-api "cd $NOVA_DIR && $NOVA_BIN_DIR/nova-api"
 | ||
|     echo "Waiting for nova-api to start..."
 | ||
|     if ! timeout $SERVICE_TIMEOUT sh -c "while ! http_proxy= wget -q -O- http://127.0.0.1:8774; do sleep 1; done"; then
 | ||
|       echo "nova-api did not start"
 | ||
|       exit 1
 | ||
|     fi
 | ||
| fi
 | ||
| 
 | ||
| # If we're using Quantum (i.e. q-svc is enabled), network creation has to
 | ||
| # happen after we've started the Quantum service.
 | ||
| if is_service_enabled q-svc; then
 | ||
|     TENANT_ID=$(keystone tenant-list | grep " demo " | get_field 1)
 | ||
| 
 | ||
|     # Create a small network
 | ||
|     # Since quantum command is executed in admin context at this point,
 | ||
|     # --tenant_id needs to be specified.
 | ||
|     NET_ID=$(quantum net-create --tenant_id $TENANT_ID net1 | grep ' id ' | get_field 2)
 | ||
|     quantum subnet-create --tenant_id $TENANT_ID --ip_version 4 --gateway $NETWORK_GATEWAY $NET_ID $FIXED_RANGE
 | ||
| elif is_service_enabled mysql && is_service_enabled nova; then
 | ||
|     # Create a small network
 | ||
|     $NOVA_BIN_DIR/nova-manage network create private $FIXED_RANGE 1 $FIXED_NETWORK_SIZE $NETWORK_CREATE_ARGS
 | ||
| 
 | ||
|     # Create some floating ips
 | ||
|     $NOVA_BIN_DIR/nova-manage floating create $FLOATING_RANGE
 | ||
| 
 | ||
|     # Create a second pool
 | ||
|     $NOVA_BIN_DIR/nova-manage floating create --ip_range=$TEST_FLOATING_RANGE --pool=$TEST_FLOATING_POOL
 | ||
| fi
 | ||
| 
 | ||
| # Launching nova-compute should be as simple as running ``nova-compute`` but
 | ||
| # have to do a little more than that in our script.  Since we add the group
 | ||
| # ``libvirtd`` to our user in this script, when nova-compute is run it is
 | ||
| # within the context of our original shell (so our groups won't be updated).
 | ||
| # Use 'sg' to execute nova-compute as a member of the libvirtd group.
 | ||
| # We don't check for is_service_enable as screen_it does it for us
 | ||
| screen_it n-cpu "cd $NOVA_DIR && sg libvirtd $NOVA_BIN_DIR/nova-compute"
 | ||
| screen_it n-crt "cd $NOVA_DIR && $NOVA_BIN_DIR/nova-cert"
 | ||
| screen_it n-vol "cd $NOVA_DIR && $NOVA_BIN_DIR/nova-volume"
 | ||
| screen_it n-net "cd $NOVA_DIR && $NOVA_BIN_DIR/nova-network"
 | ||
| screen_it n-sch "cd $NOVA_DIR && $NOVA_BIN_DIR/nova-scheduler"
 | ||
| screen_it n-novnc "cd $NOVNC_DIR && ./utils/nova-novncproxy --config-file $NOVA_CONF_DIR/$NOVA_CONF --web ."
 | ||
| screen_it n-xvnc "cd $NOVA_DIR && ./bin/nova-xvpvncproxy --config-file $NOVA_CONF_DIR/$NOVA_CONF"
 | ||
| screen_it n-cauth "cd $NOVA_DIR && ./bin/nova-consoleauth"
 | ||
| if is_service_enabled cinder; then
 | ||
|     start_cinder
 | ||
| fi
 | ||
| if is_service_enabled ceilometer; then
 | ||
|     configure_ceilometer
 | ||
|     start_ceilometer
 | ||
| fi
 | ||
| screen_it horizon "cd $HORIZON_DIR && sudo tail -f /var/log/$APACHE_NAME/horizon_error.log"
 | ||
| screen_it swift "cd $SWIFT_DIR && $SWIFT_DIR/bin/swift-proxy-server ${SWIFT_CONFIG_DIR}/proxy-server.conf -v"
 | ||
| 
 | ||
| # Starting the nova-objectstore only if swift3 service is not enabled.
 | ||
| # Swift will act as s3 objectstore.
 | ||
| is_service_enabled swift3 || \
 | ||
|     screen_it n-obj "cd $NOVA_DIR && $NOVA_BIN_DIR/nova-objectstore"
 | ||
| 
 | ||
| 
 | ||
| # Install Images
 | ||
| # ==============
 | ||
| 
 | ||
| # Upload an image to glance.
 | ||
| #
 | ||
| # The default image is cirros, a small testing image, which lets you login as root
 | ||
| #
 | ||
| # cirros also uses ``cloud-init``, supporting login via keypair and sending scripts as
 | ||
| # userdata.  See https://help.ubuntu.com/community/CloudInit for more on cloud-init
 | ||
| #
 | ||
| # Override ``IMAGE_URLS`` with a comma-separated list of uec images.
 | ||
| #
 | ||
| #  * **oneiric**: http://uec-images.ubuntu.com/oneiric/current/oneiric-server-cloudimg-amd64.tar.gz
 | ||
| #  * **precise**: http://uec-images.ubuntu.com/precise/current/precise-server-cloudimg-amd64.tar.gz
 | ||
| 
 | ||
| if is_service_enabled g-reg; then
 | ||
|     TOKEN=$(keystone  token-get | grep ' id ' | get_field 2)
 | ||
| 
 | ||
|     # Option to upload legacy ami-tty, which works with xenserver
 | ||
|     if [[ -n "$UPLOAD_LEGACY_TTY" ]]; then
 | ||
|         IMAGE_URLS="${IMAGE_URLS:+${IMAGE_URLS},}http://images.ansolabs.com/tty.tgz"
 | ||
|     fi
 | ||
| 
 | ||
|     for image_url in ${IMAGE_URLS//,/ }; do
 | ||
|         upload_image $image_url $TOKEN
 | ||
|     done
 | ||
| fi
 | ||
| 
 | ||
| 
 | ||
| # Run local script
 | ||
| # ================
 | ||
| 
 | ||
| # Run ``local.sh`` if it exists to perform user-managed tasks
 | ||
| if [[ -x $TOP_DIR/local.sh ]]; then
 | ||
|     echo "Running user script $TOP_DIR/local.sh"
 | ||
|     $TOP_DIR/local.sh
 | ||
| fi
 | ||
| 
 | ||
| 
 | ||
| # Fin
 | ||
| # ===
 | ||
| 
 | ||
| set +o xtrace
 | ||
| 
 | ||
| 
 | ||
| # Using the cloud
 | ||
| # ===============
 | ||
| 
 | ||
| echo ""
 | ||
| echo ""
 | ||
| echo ""
 | ||
| 
 | ||
| # If you installed Horizon on this server you should be able
 | ||
| # to access the site using your browser.
 | ||
| if is_service_enabled horizon; then
 | ||
|     echo "Horizon is now available at http://$SERVICE_HOST/"
 | ||
| fi
 | ||
| 
 | ||
| # If Keystone is present you can point ``nova`` cli to this server
 | ||
| if is_service_enabled key; then
 | ||
|     echo "Keystone is serving at $KEYSTONE_AUTH_PROTOCOL://$SERVICE_HOST:$KEYSTONE_API_PORT/v2.0/"
 | ||
|     echo "Examples on using novaclient command line is in exercise.sh"
 | ||
|     echo "The default users are: admin and demo"
 | ||
|     echo "The password: $ADMIN_PASSWORD"
 | ||
| fi
 | ||
| 
 | ||
| # Echo HOST_IP - useful for build_uec.sh, which uses dhcp to give the instance an address
 | ||
| echo "This is your host ip: $HOST_IP"
 | ||
| 
 | ||
| # Warn that ``EXTRA_FLAGS`` needs to be converted to ``EXTRA_OPTS``
 | ||
| if [[ -n "$EXTRA_FLAGS" ]]; then
 | ||
|     echo "WARNING: EXTRA_FLAGS is defined and may need to be converted to EXTRA_OPTS"
 | ||
| fi
 | ||
| 
 | ||
| # Indicate how long this took to run (bash maintained variable 'SECONDS')
 | ||
| echo "stack.sh completed in $SECONDS seconds."
 |