2011-09-11 03:22:13 -07:00
#!/usr/bin/env bash
2011-10-02 16:36:54 -04:00
# **stack.sh** is an opinionated openstack developer installation.
2011-10-28 14:00:21 -07:00
# This script installs and configures *nova*, *glance*, *horizon* and *keystone*
2011-09-11 03:22:13 -07:00
2011-10-20 10:07:10 -07:00
# This script allows you to specify configuration options of what git
2011-10-03 01:08:24 -04:00
# repositories to use, enabled services, network configuration and various
# passwords. If you are crafty you can run the script on multiple nodes using
# shared settings for common resources (mysql, rabbitmq) and build a multi-node
# developer install.
2011-10-02 16:53:21 -04:00
2011-11-20 09:55:44 -08:00
# To keep this script simple we assume you are running on an **Ubuntu 11.10
# Oneiric** machine. It should work in a VM or physical server. Additionally
# we put the list of *apt* and *pip* dependencies and other configuration files
# in this repo. So start by grabbing this script and the dependencies.
2011-09-15 21:28:23 -07:00
2011-10-02 16:36:54 -04:00
# Learn more and get the most recent version at http://devstack.org
2011-09-15 22:19:42 -07:00
# Sanity Check
# ============
2011-11-20 09:55:44 -08:00
# Warn users who aren't on oneiric, but allow them to override check and attempt
2011-09-15 22:19:42 -07:00
# installation with ``FORCE=yes ./stack``
2011-11-14 14:24:30 +01:00
DISTRO = $( lsb_release -c -s)
2011-11-20 09:55:44 -08:00
if [ [ ! ${ DISTRO } = ~ ( oneiric) ] ] ; then
echo "WARNING: this script has only been tested on oneiric"
2011-09-15 22:19:42 -07:00
if [ [ " $FORCE " != "yes" ] ] ; then
echo "If you wish to run this script anyway run with FORCE=yes"
exit 1
fi
fi
2011-10-19 09:24:17 -07:00
# Keep track of the current devstack directory.
TOP_DIR = $( cd $( dirname " $0 " ) && pwd )
2012-01-31 12:11:56 -06:00
# Import common functions
. $TOP_DIR /functions
2011-09-20 18:06:14 +00:00
# stack.sh keeps the list of **apt** and **pip** dependencies in external
2011-09-16 11:27:43 -07:00
# files, along with config templates and other useful files. You can find these
2011-09-20 18:06:14 +00:00
# in the ``files`` directory (next to this script). We will reference this
2011-09-16 11:31:16 -07:00
# directory using the ``FILES`` variable in this script.
2011-10-19 09:24:17 -07:00
FILES = $TOP_DIR /files
2011-09-16 11:31:16 -07:00
if [ ! -d $FILES ] ; then
echo "ERROR: missing devstack/files - did you grab more than just stack.sh?"
2011-09-15 22:19:42 -07:00
exit 1
fi
2011-10-12 07:17:11 +00:00
2011-10-17 14:07:11 -05:00
# Settings
# ========
# This script is customizable through setting environment variables. If you
# want to override a setting you can either::
#
# export MYSQL_PASSWORD=anothersecret
# ./stack.sh
#
# You can also pass options on a single line ``MYSQL_PASSWORD=simple ./stack.sh``
#
# Additionally, you can put any local variables into a ``localrc`` file, like::
#
# MYSQL_PASSWORD=anothersecret
# MYSQL_USER=hellaroot
#
# We try to have sensible defaults, so you should be able to run ``./stack.sh``
# in most cases.
#
2012-01-13 12:13:59 -06:00
# We support HTTP and HTTPS proxy servers via the usual environment variables
# http_proxy and https_proxy. They can be set in localrc if necessary or
# on the command line::
#
# http_proxy=http://proxy.example.com:3128/ ./stack.sh
#
2011-11-20 09:55:44 -08:00
# We source our settings from ``stackrc``. This file is distributed with devstack
# and contains locations for what repositories to use. If you want to use other
# repositories and branches, you can add your own settings with another file called
# ``localrc``
2011-10-17 14:07:11 -05:00
#
2011-10-20 10:07:10 -07:00
# If ``localrc`` exists, then ``stackrc`` will load those settings. This is
2011-11-02 17:57:11 +01:00
# useful for changing a branch or repository to test other versions. Also you
2011-10-17 14:07:11 -05:00
# can store your other settings like **MYSQL_PASSWORD** or **ADMIN_PASSWORD** instead
# of letting devstack generate random ones for you.
source ./stackrc
# Destination path for installation ``DEST``
DEST = ${ DEST :- /opt/stack }
2011-12-17 00:21:49 +00:00
# Check to see if we are already running a stack.sh
2012-02-07 18:13:44 +01:00
if type -p screen >/dev/null && screen -ls | egrep -q "[0-9].stack" ; then
2011-12-17 00:21:49 +00:00
echo "You are already running a stack.sh session."
echo "To rejoin this session type 'screen -x stack'."
echo "To destroy this session, kill the running screen."
exit 1
fi
2011-10-28 02:34:19 -04:00
2011-10-28 14:00:21 -07:00
# OpenStack is designed to be run as a regular user (Horizon will fail to run
2011-10-02 16:53:21 -04:00
# as root, since apache refused to startup serve content from root user). If
# stack.sh is run as root, it automatically creates a stack user with
2011-09-29 10:48:49 -07:00
# sudo privileges and runs as that user.
2011-10-02 17:47:32 -04:00
2011-09-28 14:08:26 -07:00
if [ [ $EUID -eq 0 ] ] ; then
2011-10-11 09:26:29 -05:00
ROOTSLEEP = ${ ROOTSLEEP :- 10 }
2011-10-03 23:10:55 -04:00
echo "You are running this script as root."
2011-10-11 09:26:29 -05:00
echo " In $ROOTSLEEP seconds, we will create a user 'stack' and run as that user "
sleep $ROOTSLEEP
2011-10-02 16:53:21 -04:00
2011-10-03 23:10:55 -04:00
# since this script runs as a normal user, we need to give that user
# ability to run sudo
2011-11-06 08:00:28 -08:00
dpkg -l sudo || apt_get update && apt_get install sudo
2011-10-02 16:53:21 -04:00
2011-10-07 15:18:10 +00:00
if ! getent passwd stack >/dev/null; then
2011-10-03 23:10:55 -04:00
echo "Creating a user called stack"
2011-10-17 14:07:11 -05:00
useradd -U -G sudo -s /bin/bash -d $DEST -m stack
2011-09-29 10:48:49 -07:00
fi
2011-10-07 15:18:10 +00:00
2011-10-02 16:53:21 -04:00
echo "Giving stack user passwordless sudo priviledges"
2011-11-20 09:55:44 -08:00
# some uec images sudoers does not have a '#includedir'. add one.
2011-10-12 20:32:16 -04:00
grep -q "^#includedir.*/etc/sudoers.d" /etc/sudoers ||
echo "#includedir /etc/sudoers.d" >> /etc/sudoers
2011-10-12 20:19:46 -04:00
( umask 226 && echo "stack ALL=(ALL) NOPASSWD:ALL" \
> /etc/sudoers.d/50_stack_sh )
2011-10-02 16:53:21 -04:00
2011-09-29 10:48:49 -07:00
echo "Copying files to stack user"
2011-10-17 14:07:11 -05:00
STACK_DIR = " $DEST / ${ PWD ##*/ } "
2011-12-15 12:00:31 -06:00
cp -r -f -T " $PWD " " $STACK_DIR "
2011-11-06 07:47:09 -08:00
chown -R stack " $STACK_DIR "
2011-10-10 08:06:14 -05:00
if [ [ " $SHELL_AFTER_RUN " != "no" ] ] ; then
2011-10-12 20:00:34 -04:00
exec su -c " set -e; cd $STACK_DIR ; bash stack.sh; bash " stack
2011-10-10 08:06:14 -05:00
else
2011-10-12 20:00:34 -04:00
exec su -c " set -e; cd $STACK_DIR ; bash stack.sh " stack
2011-10-10 08:06:14 -05:00
fi
2011-10-07 21:28:00 -04:00
exit 1
2011-10-27 11:18:09 -07:00
else
2011-10-28 12:20:07 -07:00
# Our user needs passwordless priviledges for certain commands which nova
2011-10-27 11:18:09 -07:00
# uses internally.
# Natty uec images sudoers does not have a '#includedir'. add one.
2011-10-27 11:20:38 -07:00
sudo grep -q "^#includedir.*/etc/sudoers.d" /etc/sudoers ||
echo "#includedir /etc/sudoers.d" | sudo tee -a /etc/sudoers
2011-10-28 02:34:19 -04:00
TEMPFILE = ` mktemp`
cat $FILES /sudo/nova > $TEMPFILE
sed -e " s,%USER%, $USER ,g " -i $TEMPFILE
chmod 0440 $TEMPFILE
sudo chown root:root $TEMPFILE
2011-10-28 12:27:20 -04:00
sudo mv $TEMPFILE /etc/sudoers.d/stack_sh_nova
2011-09-28 14:08:26 -07:00
fi
2011-12-16 22:40:46 -06:00
# Set True to configure stack.sh to run cleanly without Internet access.
# stack.sh must have been previously run with Internet access to install
# prerequisites and initialize $DEST.
OFFLINE = ` trueorfalse False $OFFLINE `
2011-09-12 11:59:38 -07:00
# Set the destination directories for openstack projects
2011-09-11 03:22:13 -07:00
NOVA_DIR = $DEST /nova
2011-10-28 14:00:21 -07:00
HORIZON_DIR = $DEST /horizon
2011-09-11 03:22:13 -07:00
GLANCE_DIR = $DEST /glance
KEYSTONE_DIR = $DEST /keystone
NOVACLIENT_DIR = $DEST /python-novaclient
2011-12-27 22:22:14 -08:00
KEYSTONECLIENT_DIR = $DEST /python-keystoneclient
2011-09-11 03:22:13 -07:00
NOVNC_DIR = $DEST /noVNC
2011-11-01 12:30:55 +01:00
SWIFT_DIR = $DEST /swift
2011-11-01 19:32:23 +01:00
SWIFT_KEYSTONE_DIR = $DEST /swift-keystone2
2011-10-27 18:18:20 -07:00
QUANTUM_DIR = $DEST /quantum
2012-01-25 17:22:15 -05:00
QUANTUM_CLIENT_DIR = $DEST /python-quantumclient
2012-01-23 11:17:27 -06:00
MELANGE_DIR = $DEST /melange
MELANGECLIENT_DIR = $DEST /python-melangeclient
2011-10-27 18:18:20 -07:00
# Default Quantum Plugin
Q_PLUGIN = ${ Q_PLUGIN :- openvswitch }
2011-12-12 23:04:58 +00:00
# Default Quantum Port
Q_PORT = ${ Q_PORT :- 9696 }
# Default Quantum Host
Q_HOST = ${ Q_HOST :- localhost }
2011-09-13 20:07:44 -07:00
2012-01-23 11:17:27 -06:00
# Default Melange Port
M_PORT = ${ M_PORT :- 9898 }
# Default Melange Host
M_HOST = ${ M_HOST :- localhost }
# Melange MAC Address Range
M_MAC_RANGE = ${ M_MAC_RANGE :- 404040 /24 }
2011-09-13 20:07:44 -07:00
# Specify which services to launch. These generally correspond to screen tabs
2012-01-26 12:29:51 -08:00
ENABLED_SERVICES = ${ ENABLED_SERVICES :- g -api,g-reg,key,n-api,n-crt,n-obj,n-cpu,n-net,n-sch,n-novnc,n-xvnc,n-cauth,horizon,mysql,rabbit }
2011-09-11 03:22:13 -07:00
2011-11-14 08:59:05 -08:00
# Name of the lvm volume group to use/create for iscsi volumes
VOLUME_GROUP = ${ VOLUME_GROUP :- nova -volumes }
2011-12-01 17:02:07 -06:00
VOLUME_NAME_PREFIX = ${ VOLUME_NAME_PREFIX :- volume - }
2011-12-17 00:21:49 +00:00
INSTANCE_NAME_PREFIX = ${ INSTANCE_NAME_PREFIX :- instance - }
2011-11-14 08:59:05 -08:00
2011-10-26 22:29:08 -07:00
# Nova hypervisor configuration. We default to libvirt whth **kvm** but will
# drop back to **qemu** if we are unable to load the kvm module. Stack.sh can
# also install an **LXC** based system.
VIRT_DRIVER = ${ VIRT_DRIVER :- libvirt }
2011-10-02 16:53:21 -04:00
LIBVIRT_TYPE = ${ LIBVIRT_TYPE :- kvm }
2011-10-02 17:47:32 -04:00
# nova supports pluggable schedulers. ``SimpleScheduler`` should work in most
# cases unless you are working on multi-zone mode.
2011-10-02 16:53:21 -04:00
SCHEDULER = ${ SCHEDULER :- nova .scheduler.simple.SimpleScheduler }
2012-01-19 13:28:21 -08:00
HOST_IP_IFACE = ${ HOST_IP_IFACE :- eth0 }
2011-11-07 13:18:28 -06:00
# Use the eth0 IP unless an explicit is set by ``HOST_IP`` environment variable
2012-01-19 13:28:21 -08:00
if [ -z " $HOST_IP " -o " $HOST_IP " = = "dhcp" ] ; then
HOST_IP = ` LC_ALL = C /sbin/ifconfig ${ HOST_IP_IFACE } | grep -m 1 'inet addr:' | cut -d: -f2 | awk '{print $1}' `
2011-11-07 13:24:00 -06:00
if [ " $HOST_IP " = "" ] ; then
2011-11-07 14:02:13 -06:00
echo "Could not determine host ip address."
2012-01-19 13:28:21 -08:00
echo " Either localrc specified dhcp on ${ HOST_IP_IFACE } or defaulted to eth0 "
2011-11-07 13:18:28 -06:00
exit 1
fi
2011-09-11 03:22:13 -07:00
fi
2011-12-27 23:22:14 -08:00
# Allow the use of an alternate hostname (such as localhost/127.0.0.1) for service endpoints.
SERVICE_HOST = ${ SERVICE_HOST :- $HOST_IP }
2011-11-22 17:48:10 -06:00
# Configure services to syslog instead of writing to individual log files
SYSLOG = ` trueorfalse False $SYSLOG `
SYSLOG_HOST = ${ SYSLOG_HOST :- $HOST_IP }
SYSLOG_PORT = ${ SYSLOG_PORT :- 516 }
2011-11-05 16:19:03 -05:00
# Service startup timeout
SERVICE_TIMEOUT = ${ SERVICE_TIMEOUT :- 60 }
2011-10-12 07:13:13 +00:00
# Generic helper to configure passwords
function read_password {
set +o xtrace
var = $1 ; msg = $2
pw = ${ !var }
2011-10-12 14:08:08 -07:00
localrc = $TOP_DIR /localrc
2011-10-12 07:17:11 +00:00
2011-10-12 07:13:13 +00:00
# If the password is not defined yet, proceed to prompt user for a password.
if [ ! $pw ] ; then
# If there is no localrc file, create one
2011-10-12 14:08:08 -07:00
if [ ! -e $localrc ] ; then
touch $localrc
2011-10-12 07:13:13 +00:00
fi
2011-10-20 10:07:10 -07:00
# Presumably if we got this far it can only be that our localrc is missing
2011-10-12 07:13:13 +00:00
# the required password. Prompt user for a password and write to localrc.
2011-10-12 14:08:08 -07:00
echo ''
echo '################################################################################'
echo $msg
echo '################################################################################'
2011-12-29 17:27:45 -06:00
echo "This value will be written to your localrc file so you don't have to enter it "
echo "again. Use only alphanumeric characters."
2011-10-12 14:08:08 -07:00
echo "If you leave this blank, a random default value will be used."
2011-12-29 17:27:45 -06:00
pw = " "
while true; do
echo "Enter a password now:"
read -e $var
pw = ${ !var }
[ [ " $pw " = " `echo $pw | tr -cd [:alnum:]` " ] ] && break
echo "Invalid chars in password. Try again:"
done
2011-10-12 14:08:08 -07:00
if [ ! $pw ] ; then
pw = ` openssl rand -hex 10`
2011-10-12 07:13:13 +00:00
fi
2011-10-12 14:08:08 -07:00
eval " $var = $pw "
echo " $var = $pw " >> $localrc
2011-10-12 07:13:13 +00:00
fi
set -o xtrace
}
2012-02-16 10:16:52 +00:00
# This function will check if the service(s) specified in argument is
# enabled by the user in ENABLED_SERVICES.
#
# If there is multiple services specified as argument it will act as a
# boolean OR or if any of the services specified on the command line
# return true.
#
# There is a special cases for some 'catch-all' services :
# nova would catch if any service enabled start by n-
# glance would catch if any service enabled start by g-
# quantum would catch if any service enabled start by q-
function is_service_enabled( ) {
services = $@
for service in ${ services } ; do
[ [ ,${ ENABLED_SERVICES } , = ~ ,${ service } , ] ] && return 0
[ [ ${ service } = = "nova" && ${ ENABLED_SERVICES } = ~ "n-" ] ] && return 0
[ [ ${ service } = = "glance" && ${ ENABLED_SERVICES } = ~ "g-" ] ] && return 0
[ [ ${ service } = = "quantum" && ${ ENABLED_SERVICES } = ~ "q-" ] ] && return 0
done
return 1
}
2011-10-12 07:13:13 +00:00
2011-10-02 16:53:21 -04:00
# Nova Network Configuration
# --------------------------
2011-10-20 10:07:10 -07:00
# FIXME: more documentation about why these are important flags. Also
2011-10-03 01:08:24 -04:00
# we should make sure we use the same variable names as the flag names.
2011-12-16 20:23:07 +00:00
PUBLIC_INTERFACE = ${ PUBLIC_INTERFACE :- br100 }
2011-09-26 12:48:31 -07:00
FIXED_RANGE = ${ FIXED_RANGE :- 10 .0.0.0/24 }
FIXED_NETWORK_SIZE = ${ FIXED_NETWORK_SIZE :- 256 }
2011-10-27 13:21:52 -07:00
FLOATING_RANGE = ${ FLOATING_RANGE :- 172 .24.4.224/28 }
2011-09-25 13:41:22 -07:00
NET_MAN = ${ NET_MAN :- FlatDHCPManager }
2011-12-27 23:22:14 -08:00
EC2_DMZ_HOST = ${ EC2_DMZ_HOST :- $SERVICE_HOST }
2011-09-20 09:39:50 -07:00
FLAT_NETWORK_BRIDGE = ${ FLAT_NETWORK_BRIDGE :- br100 }
2011-12-16 20:23:07 +00:00
VLAN_INTERFACE = ${ VLAN_INTERFACE :- eth0 }
2011-10-02 17:47:32 -04:00
2012-01-10 15:34:34 -06:00
# Test floating pool and range are used for testing. They are defined
# here until the admin APIs can replace nova-manage
TEST_FLOATING_POOL = ${ TEST_FLOATING_POOL :- test }
TEST_FLOATING_RANGE = ${ TEST_FLOATING_RANGE :- 192 .168.253.0/29 }
2011-10-02 17:47:32 -04:00
# Multi-host is a mode where each compute node runs its own network node. This
# allows network operations and routing for a VM to occur on the server that is
# running the VM - removing a SPOF and bandwidth bottleneck.
2011-11-05 16:12:20 -07:00
MULTI_HOST = ${ MULTI_HOST :- False }
2011-09-13 00:59:54 -07:00
2011-09-13 01:24:50 -07:00
# If you are using FlatDHCP on multiple hosts, set the ``FLAT_INTERFACE``
# variable but make sure that the interface doesn't already have an
# ip or you risk breaking things.
2011-10-03 01:08:24 -04:00
#
2011-10-20 10:07:10 -07:00
# **DHCP Warning**: If your flat interface device uses DHCP, there will be a
# hiccup while the network is moved from the flat interface to the flat network
# bridge. This will happen when you launch your first instance. Upon launch
# you will lose all connectivity to the node, and the vm launch will probably
2011-10-03 01:08:24 -04:00
# fail.
2011-10-20 10:07:10 -07:00
#
# If you are running on a single node and don't need to access the VMs from
2011-10-03 01:08:24 -04:00
# devices other than that node, you can set the flat interface to the same
2011-10-20 10:07:10 -07:00
# value as ``FLAT_NETWORK_BRIDGE``. This will stop the network hiccup from
2011-11-02 17:57:11 +01:00
# occurring.
2011-09-25 22:28:08 -07:00
FLAT_INTERFACE = ${ FLAT_INTERFACE :- eth0 }
2011-09-11 03:22:13 -07:00
2011-10-02 17:47:32 -04:00
## FIXME(ja): should/can we check that FLAT_INTERFACE is sane?
2011-10-27 18:18:20 -07:00
# Using Quantum networking:
#
2011-12-12 23:04:58 +00:00
# Make sure that quantum is enabled in ENABLED_SERVICES. If it is the network
# manager will be set to the QuantumManager. If you want to run Quantum on
# this host, make sure that q-svc is also in ENABLED_SERVICES.
2011-10-27 18:18:20 -07:00
#
# If you're planning to use the Quantum openvswitch plugin, set Q_PLUGIN to
# "openvswitch" and make sure the q-agt service is enabled in
# ENABLED_SERVICES.
#
# With Quantum networking the NET_MAN variable is ignored.
2012-01-23 11:17:27 -06:00
# Using Melange IPAM:
#
# Make sure that quantum and melange are enabled in ENABLED_SERVICES.
# If they are then the melange IPAM lib will be set in the QuantumManager.
# Adding m-svc to ENABLED_SERVICES will start the melange service on this
# host.
2011-09-13 01:24:50 -07:00
2011-10-02 16:53:21 -04:00
# MySQL & RabbitMQ
# ----------------
2011-10-28 14:00:21 -07:00
# We configure Nova, Horizon, Glance and Keystone to use MySQL as their
2011-10-02 16:53:21 -04:00
# database server. While they share a single server, each has their own
# database and tables.
2011-10-20 10:07:10 -07:00
# By default this script will install and configure MySQL. If you want to
2011-10-02 16:53:21 -04:00
# use an existing server, you can pass in the user/password/host parameters.
2011-10-12 07:13:13 +00:00
# You will need to send the same ``MYSQL_PASSWORD`` to every host if you are doing
2011-10-02 16:53:21 -04:00
# a multi-node devstack installation.
2011-10-19 15:38:10 -07:00
MYSQL_HOST = ${ MYSQL_HOST :- localhost }
2011-09-14 02:39:10 -07:00
MYSQL_USER = ${ MYSQL_USER :- root }
2011-10-12 07:13:13 +00:00
read_password MYSQL_PASSWORD "ENTER A PASSWORD TO USE FOR MYSQL."
2011-10-02 16:53:21 -04:00
2011-09-13 20:07:44 -07:00
# don't specify /db in this string, so we can use it for multiple services
2011-10-12 07:13:13 +00:00
BASE_SQL_CONN = ${ BASE_SQL_CONN :- mysql : // $MYSQL_USER : $MYSQL_PASSWORD @ $MYSQL_HOST }
2011-09-13 20:07:44 -07:00
# Rabbit connection info
RABBIT_HOST = ${ RABBIT_HOST :- localhost }
2011-10-12 07:13:13 +00:00
read_password RABBIT_PASSWORD "ENTER A PASSWORD TO USE FOR RABBIT."
2011-09-11 03:22:13 -07:00
2011-09-14 09:55:31 -07:00
# Glance connection info. Note the port must be specified.
2011-12-27 23:22:14 -08:00
GLANCE_HOSTPORT = ${ GLANCE_HOSTPORT :- $SERVICE_HOST : 9292 }
2011-09-14 09:55:31 -07:00
2011-11-01 12:30:55 +01:00
# SWIFT
# -----
2011-11-02 17:57:11 +01:00
# TODO: implement glance support
# TODO: add logging to different location.
2011-11-02 16:49:56 +01:00
2011-11-02 17:57:11 +01:00
# By default the location of swift drives and objects is located inside
2011-11-03 09:17:06 +01:00
# the swift source directory. SWIFT_DATA_LOCATION variable allow you to redefine
2011-11-02 17:57:11 +01:00
# this.
2011-11-03 09:17:06 +01:00
SWIFT_DATA_LOCATION = ${ SWIFT_DATA_LOCATION :- ${ SWIFT_DIR } /data }
2011-11-01 12:30:55 +01:00
2011-11-03 10:43:46 +01:00
# We are going to have the configuration files inside the source
# directory, change SWIFT_CONFIG_LOCATION if you want to adjust that.
SWIFT_CONFIG_LOCATION = ${ SWIFT_CONFIG_LOCATION :- ${ SWIFT_DIR } /config }
2011-11-01 12:30:55 +01:00
2011-11-02 17:57:11 +01:00
# devstack will create a loop-back disk formatted as XFS to store the
# swift data. By default the disk size is 1 gigabyte. The variable
# SWIFT_LOOPBACK_DISK_SIZE specified in bytes allow you to change
# that.
2011-11-01 12:30:55 +01:00
SWIFT_LOOPBACK_DISK_SIZE = ${ SWIFT_LOOPBACK_DISK_SIZE :- 1000000 }
2011-10-12 07:13:13 +00:00
2011-11-02 17:57:11 +01:00
# The ring uses a configurable number of bits from a path’ s MD5 hash as
# a partition index that designates a device. The number of bits kept
# from the hash is known as the partition power, and 2 to the partition
# power indicates the partition count. Partitioning the full MD5 hash
# ring allows other parts of the cluster to work in batches of items at
# once which ends up either more efficient or at least less complex than
# working with each item separately or the entire cluster all at once.
# By default we define 9 for the partition count (which mean 512).
2011-11-01 15:36:00 +01:00
SWIFT_PARTITION_POWER_SIZE = ${ SWIFT_PARTITION_POWER_SIZE :- 9 }
2011-11-03 16:19:14 +01:00
# We only ask for Swift Hash if we have enabled swift service.
2012-02-16 10:16:52 +00:00
if is_service_enabled swift; then
2011-11-03 16:19:14 +01:00
# SWIFT_HASH is a random unique string for a swift cluster that
# can never change.
read_password SWIFT_HASH "ENTER A RANDOM SWIFT HASH."
fi
2011-11-05 16:12:20 -07:00
2011-10-02 16:53:21 -04:00
# Keystone
# --------
2011-10-02 09:02:46 -07:00
# Service Token - Openstack components need to have an admin token
# to validate user tokens.
2011-10-12 07:13:13 +00:00
read_password SERVICE_TOKEN "ENTER A SERVICE_TOKEN TO USE FOR THE SERVICE ADMIN TOKEN."
2011-10-28 14:00:21 -07:00
# Horizon currently truncates usernames and passwords at 20 characters
read_password ADMIN_PASSWORD "ENTER A PASSWORD TO USE FOR HORIZON AND KEYSTONE (20 CHARS OR LESS)."
2011-10-02 09:02:46 -07:00
2012-01-16 22:27:20 -06:00
# Set Keystone interface configuration
KEYSTONE_AUTH_HOST = ${ KEYSTONE_AUTH_HOST :- $SERVICE_HOST }
KEYSTONE_AUTH_PORT = ${ KEYSTONE_AUTH_PORT :- 35357 }
KEYSTONE_AUTH_PROTOCOL = ${ KEYSTONE_AUTH_PROTOCOL :- http }
KEYSTONE_SERVICE_HOST = ${ KEYSTONE_SERVICE_HOST :- $SERVICE_HOST }
KEYSTONE_SERVICE_PORT = ${ KEYSTONE_SERVICE_PORT :- 5000 }
KEYSTONE_SERVICE_PROTOCOL = ${ KEYSTONE_SERVICE_PROTOCOL :- http }
2012-01-24 11:45:52 -06:00
# Horizon
# -------
# Allow overriding the default Apache user and group, default both to
# current user.
APACHE_USER = ${ APACHE_USER :- $USER }
APACHE_GROUP = ${ APACHE_GROUP :- $APACHE_USER }
2011-12-27 11:45:55 -06:00
# Log files
# ---------
# Set up logging for stack.sh
# Set LOGFILE to turn on logging
# We append '.xxxxxxxx' to the given name to maintain history
# where xxxxxxxx is a representation of the date the file was created
if [ [ -n " $LOGFILE " ] ] ; then
# First clean up old log files. Use the user-specified LOGFILE
# as the template to search for, appending '.*' to match the date
# we added on earlier runs.
LOGDAYS = ${ LOGDAYS :- 7 }
LOGDIR = $( dirname " $LOGFILE " )
LOGNAME = $( basename " $LOGFILE " )
find $LOGDIR -maxdepth 1 -name $LOGNAME .\* -mtime +$LOGDAYS -exec rm { } \;
TIMESTAMP_FORMAT = ${ TIMESTAMP_FORMAT :- "%F-%H%M%S" }
LOGFILE = $LOGFILE .$( date " + $TIMESTAMP_FORMAT " )
# Redirect stdout/stderr to tee to write the log file
exec 1> >( tee " ${ LOGFILE } " ) 2>& 1
echo " stack.sh log $LOGFILE "
# Specified logfile name always links to the most recent log
ln -sf $LOGFILE $LOGDIR /$LOGNAME
fi
2011-10-07 21:28:00 -04:00
# So that errors don't compound we exit on any errors so you see only the
2011-11-02 17:57:11 +01:00
# first error that occurred.
2011-10-07 21:28:00 -04:00
trap failed ERR
failed( ) {
local r = $?
set +o xtrace
[ -n " $LOGFILE " ] && echo " ${ 0 ##*/ } failed: full log in $LOGFILE "
exit $r
}
# Print the commands being run so that we can see the command that triggers
# an error. It is also useful for following along as the install occurs.
set -o xtrace
2011-10-19 14:30:37 -07:00
# create the destination directory and ensure it is writable by the user
2011-10-07 21:28:00 -04:00
sudo mkdir -p $DEST
2011-10-19 14:30:37 -07:00
if [ ! -w $DEST ] ; then
sudo chown ` whoami` $DEST
fi
2011-10-02 14:28:17 -04:00
2011-09-13 00:59:54 -07:00
# Install Packages
2011-09-13 01:24:50 -07:00
# ================
2011-09-13 00:59:54 -07:00
#
# Openstack uses a fair number of other projects.
2011-11-14 15:20:39 +01:00
# - We are going to install packages only for the services needed.
# - We are parsing the packages files and detecting metadatas.
2011-11-16 00:43:34 -06:00
# - If there is a NOPRIME as comment mean we are not doing the install
# just yet.
2011-12-15 12:00:31 -06:00
# - If we have the meta-keyword dist:DISTRO or
# dist:DISTRO1,DISTRO2 it will be installed only for those
2011-11-14 15:20:39 +01:00
# distros (case insensitive).
function get_packages( ) {
local file_to_parse = "general"
local service
2011-11-20 09:55:44 -08:00
2011-11-14 15:20:39 +01:00
for service in ${ ENABLED_SERVICES //,/ } ; do
2011-12-01 13:44:51 -08:00
# Allow individual services to specify dependencies
if [ [ -e $FILES /apts/${ service } ] ] ; then
file_to_parse = " ${ file_to_parse } $service "
fi
2011-11-14 15:20:39 +01:00
if [ [ $service = = n-* ] ] ; then
2011-11-14 15:26:13 +01:00
if [ [ ! $file_to_parse = ~ nova ] ] ; then
2011-11-14 15:20:39 +01:00
file_to_parse = " ${ file_to_parse } nova "
fi
2011-11-14 15:26:13 +01:00
elif [ [ $service = = g-* ] ] ; then
if [ [ ! $file_to_parse = ~ glance ] ] ; then
2011-11-14 15:20:39 +01:00
file_to_parse = " ${ file_to_parse } glance "
fi
2011-11-14 15:26:13 +01:00
elif [ [ $service = = key* ] ] ; then
if [ [ ! $file_to_parse = ~ keystone ] ] ; then
2011-11-14 15:20:39 +01:00
file_to_parse = " ${ file_to_parse } keystone "
fi
fi
done
2011-11-14 15:26:13 +01:00
for file in ${ file_to_parse } ; do
2011-11-14 15:20:39 +01:00
local fname = ${ FILES } /apts/${ file }
local OIFS line package distros distro
[ [ -e $fname ] ] || { echo " missing: $fname " ; exit 1 ; }
OIFS = $IFS
IFS = $'\n'
2011-11-14 22:16:11 +01:00
for line in $( <${ fname } ) ; do
2011-11-16 00:43:34 -06:00
if [ [ $line = ~ "NOPRIME" ] ] ; then
continue
fi
2011-11-14 15:26:13 +01:00
if [ [ $line = ~ ( .*) #.*dist:([^ ]*) ]]; then # We are using BASH regexp matching feature.
2011-11-14 15:20:39 +01:00
package = ${ BASH_REMATCH [1] }
distros = ${ BASH_REMATCH [2] }
2011-11-14 15:26:13 +01:00
for distro in ${ distros //,/ } ; do #In bash ${VAR,,} will lowecase VAR
2011-11-14 15:20:39 +01:00
[ [ ${ distro ,, } = = ${ DISTRO ,, } ] ] && echo $package
done
continue
fi
echo ${ line %#* }
done
IFS = $OIFS
done
}
2011-09-12 17:09:08 -07:00
# install apt requirements
2011-10-20 13:09:11 -04:00
apt_get update
2011-11-14 15:20:39 +01:00
apt_get install $( get_packages)
2011-09-12 17:09:08 -07:00
# install python requirements
2011-12-16 22:40:46 -06:00
pip_install ` cat $FILES /pips/* | uniq`
2011-09-12 17:09:08 -07:00
# compute service
2011-09-26 13:02:40 -07:00
git_clone $NOVA_REPO $NOVA_DIR $NOVA_BRANCH
2011-10-28 14:00:21 -07:00
# python client library to nova that horizon (and others) use
2012-02-03 21:40:32 -08:00
git_clone $KEYSTONECLIENT_REPO $KEYSTONECLIENT_DIR $KEYSTONECLIENT_BRANCH
2011-09-26 13:02:40 -07:00
git_clone $NOVACLIENT_REPO $NOVACLIENT_DIR $NOVACLIENT_BRANCH
2011-11-20 09:55:44 -08:00
# glance, swift middleware and nova api needs keystone middleware
2012-02-16 10:16:52 +00:00
if is_service_enabled key g-api n-api swift; then
2011-11-20 09:55:44 -08:00
# unified auth system (manages accounts/tokens)
git_clone $KEYSTONE_REPO $KEYSTONE_DIR $KEYSTONE_BRANCH
fi
2012-02-16 10:16:52 +00:00
if is_service_enabled swift; then
2011-11-10 13:05:13 -08:00
# storage service
git_clone $SWIFT_REPO $SWIFT_DIR $SWIFT_BRANCH
# swift + keystone middleware
git_clone $SWIFT_KEYSTONE_REPO $SWIFT_KEYSTONE_DIR $SWIFT_KEYSTONE_BRANCH
fi
2012-02-16 10:16:52 +00:00
if is_service_enabled g-api n-api; then
2011-11-10 13:05:13 -08:00
# image catalog service
git_clone $GLANCE_REPO $GLANCE_DIR $GLANCE_BRANCH
fi
2012-02-16 10:16:52 +00:00
if is_service_enabled n-novnc; then
2011-11-10 13:05:13 -08:00
# a websockets/html5 or flash powered VNC console for vm instances
git_clone $NOVNC_REPO $NOVNC_DIR $NOVNC_BRANCH
fi
2012-02-16 10:16:52 +00:00
if is_service_enabled horizon; then
2011-11-10 13:05:13 -08:00
# django powered web control panel for openstack
git_clone $HORIZON_REPO $HORIZON_DIR $HORIZON_BRANCH $HORIZON_TAG
fi
2012-02-16 10:16:52 +00:00
if is_service_enabled q-svc; then
2011-11-10 13:05:13 -08:00
# quantum
git_clone $QUANTUM_REPO $QUANTUM_DIR $QUANTUM_BRANCH
2012-01-25 17:22:15 -05:00
git_clone $QUANTUM_CLIENT_REPO $QUANTUM_CLIENT_DIR $QUANTUM_CLIENT_BRANCH
2011-11-10 13:05:13 -08:00
fi
2011-09-12 17:09:08 -07:00
2012-02-16 10:16:52 +00:00
if is_service_enabled m-svc; then
2012-01-23 11:17:27 -06:00
# melange
git_clone $MELANGE_REPO $MELANGE_DIR $MELANGE_BRANCH
fi
2012-02-16 10:16:52 +00:00
if is_service_enabled melange; then
2012-01-23 11:17:27 -06:00
git_clone $MELANGECLIENT_REPO $MELANGECLIENT_DIR $MELANGECLIENT_BRANCH
fi
2011-09-13 00:59:54 -07:00
# Initialization
2011-09-13 01:24:50 -07:00
# ==============
2011-09-13 00:59:54 -07:00
2011-09-16 16:28:13 -07:00
2011-09-12 17:09:08 -07:00
# setup our checkouts so they are installed into python path
2011-09-13 01:24:50 -07:00
# allowing ``import nova`` or ``import glance.client``
2012-02-06 21:21:52 +00:00
cd $KEYSTONECLIENT_DIR ; sudo python setup.py develop
cd $NOVACLIENT_DIR ; sudo python setup.py develop
2012-02-16 10:16:52 +00:00
if is_service_enabled key g-api n-api swift; then
2011-11-10 13:05:13 -08:00
cd $KEYSTONE_DIR ; sudo python setup.py develop
fi
2012-02-16 10:16:52 +00:00
if is_service_enabled swift; then
2011-11-10 13:05:13 -08:00
cd $SWIFT_DIR ; sudo python setup.py develop
cd $SWIFT_KEYSTONE_DIR ; sudo python setup.py develop
fi
2012-02-16 10:16:52 +00:00
if is_service_enabled g-api n-api; then
2011-11-10 13:05:13 -08:00
cd $GLANCE_DIR ; sudo python setup.py develop
fi
2011-10-16 12:24:11 -07:00
cd $NOVA_DIR ; sudo python setup.py develop
2012-02-16 10:16:52 +00:00
if is_service_enabled horizon; then
2011-11-23 10:10:53 -08:00
cd $HORIZON_DIR /horizon; sudo python setup.py develop
2011-11-10 13:05:13 -08:00
cd $HORIZON_DIR /openstack-dashboard; sudo python setup.py develop
fi
2012-02-16 10:16:52 +00:00
if is_service_enabled q-svc; then
2011-11-10 13:05:13 -08:00
cd $QUANTUM_DIR ; sudo python setup.py develop
fi
2012-02-16 10:16:52 +00:00
if is_service_enabled m-svc; then
2012-01-23 11:17:27 -06:00
cd $MELANGE_DIR ; sudo python setup.py develop
fi
2012-02-16 10:16:52 +00:00
if is_service_enabled melange; then
2012-01-23 11:17:27 -06:00
cd $MELANGECLIENT_DIR ; sudo python setup.py develop
fi
2011-09-12 17:09:08 -07:00
2011-11-22 17:48:10 -06:00
# Syslog
# ---------
if [ [ $SYSLOG != "False" ] ] ; then
apt_get install -y rsyslog-relp
if [ [ " $SYSLOG_HOST " = " $HOST_IP " ] ] ; then
# Configure the master host to receive
cat <<EOF >/tmp/90-stack-m.conf
\$ ModLoad imrelp
\$ InputRELPServerRun $SYSLOG_PORT
EOF
sudo mv /tmp/90-stack-m.conf /etc/rsyslog.d
else
# Set rsyslog to send to remote host
cat <<EOF >/tmp/90-stack-s.conf
*.* :omrelp:$SYSLOG_HOST :$SYSLOG_PORT
EOF
sudo mv /tmp/90-stack-s.conf /etc/rsyslog.d
fi
sudo /usr/sbin/service rsyslog restart
fi
2011-09-15 23:11:29 -07:00
# Rabbit
# ---------
2011-10-02 17:47:32 -04:00
2012-02-16 10:16:52 +00:00
if is_service_enabled rabbit; then
2011-09-15 23:11:29 -07:00
# Install and start rabbitmq-server
2011-10-20 12:41:40 -04:00
# the temp file is necessary due to LP: #878600
tfile = $( mktemp)
2011-10-20 13:09:11 -04:00
apt_get install rabbitmq-server > " $tfile " 2>& 1
2011-10-20 12:41:40 -04:00
cat " $tfile "
rm -f " $tfile "
2011-10-02 14:28:17 -04:00
# change the rabbit password since the default is "guest"
sudo rabbitmqctl change_password guest $RABBIT_PASSWORD
2011-09-15 23:11:29 -07:00
fi
2011-09-11 03:22:13 -07:00
2011-09-15 21:28:23 -07:00
# Mysql
# ---------
2011-10-02 17:47:32 -04:00
2012-02-16 10:16:52 +00:00
if is_service_enabled mysql; then
2011-10-02 17:47:32 -04:00
# Seed configuration with mysql password so that apt-get install doesn't
# prompt us for a password upon install.
cat <<MYSQL_PRESEED | sudo debconf-set-selections
2011-10-12 07:13:13 +00:00
mysql-server-5.1 mysql-server/root_password password $MYSQL_PASSWORD
mysql-server-5.1 mysql-server/root_password_again password $MYSQL_PASSWORD
2011-10-02 17:47:32 -04:00
mysql-server-5.1 mysql-server/start_on_boot boolean true
MYSQL_PRESEED
2011-10-13 11:40:16 -07:00
# while ``.my.cnf`` is not needed for openstack to function, it is useful
# as it allows you to access the mysql databases via ``mysql nova`` instead
# of having to specify the username/password each time.
2011-10-13 18:45:42 +01:00
if [ [ ! -e $HOME /.my.cnf ] ] ; then
cat <<EOF >$HOME /.my.cnf
[ client]
user = $MYSQL_USER
2011-10-13 15:07:36 -07:00
password = $MYSQL_PASSWORD
2011-10-13 18:45:42 +01:00
host = $MYSQL_HOST
EOF
chmod 0600 $HOME /.my.cnf
fi
2011-09-15 23:11:29 -07:00
# Install and start mysql-server
2011-10-20 13:09:11 -04:00
apt_get install mysql-server
2011-09-15 21:28:23 -07:00
# Update the DB to give user ‘ $MYSQL_USER’ @’ %’ full control of the all databases:
2011-10-11 09:26:29 -05:00
sudo mysql -uroot -p$MYSQL_PASSWORD -h127.0.0.1 -e " GRANT ALL PRIVILEGES ON *.* TO ' $MYSQL_USER '@'%' identified by ' $MYSQL_PASSWORD '; "
2011-09-15 21:28:23 -07:00
# Edit /etc/mysql/my.cnf to change ‘ bind-address’ from localhost (127.0.0.1) to any (0.0.0.0) and restart the mysql service:
sudo sed -i 's/127.0.0.1/0.0.0.0/g' /etc/mysql/my.cnf
sudo service mysql restart
fi
2011-10-28 14:00:21 -07:00
# Horizon
2011-09-13 01:24:50 -07:00
# ---------
2011-10-02 17:47:32 -04:00
2011-10-28 14:00:21 -07:00
# Setup the django horizon application to serve via apache/wsgi
2011-09-12 17:09:08 -07:00
2012-02-16 10:16:52 +00:00
if is_service_enabled horizon; then
2011-09-15 21:28:23 -07:00
2011-11-20 09:55:44 -08:00
# Install apache2, which is NOPRIME'd
apt_get install apache2 libapache2-mod-wsgi
2012-01-25 17:22:15 -05:00
# Link to quantum client directory.
rm -fr ${ HORIZON_DIR } /openstack-dashboard/quantum
ln -s ${ QUANTUM_CLIENT_DIR } /quantum ${ HORIZON_DIR } /openstack-dashboard/quantum
2011-09-15 16:52:43 -07:00
2012-01-26 12:59:26 -08:00
# Remove stale session database.
rm -f $HORIZON_DIR /openstack-dashboard/local/dashboard_openstack.sqlite3
2011-09-20 18:06:14 +00:00
2011-10-28 14:00:21 -07:00
# ``local_settings.py`` is used to override horizon default settings.
2011-11-23 10:10:53 -08:00
local_settings = $HORIZON_DIR /openstack-dashboard/local/local_settings.py
cp $FILES /horizon_settings.py $local_settings
# Enable quantum in dashboard, if requested
2012-02-16 10:16:52 +00:00
if is_service_enabled quantum; then
2011-11-23 10:10:53 -08:00
sudo sed -e "s,QUANTUM_ENABLED = False,QUANTUM_ENABLED = True,g" -i $local_settings
fi
2011-09-15 21:46:20 -07:00
2011-10-28 14:00:21 -07:00
# Initialize the horizon database (it stores sessions and notices shown to
2011-10-19 09:24:17 -07:00
# users). The user system is external (keystone).
2011-10-28 14:00:21 -07:00
cd $HORIZON_DIR /openstack-dashboard
2012-01-05 22:21:08 +00:00
python manage.py syncdb
2011-09-15 16:52:43 -07:00
# create an empty directory that apache uses as docroot
2011-10-28 14:00:21 -07:00
sudo mkdir -p $HORIZON_DIR /.blackhole
2011-09-15 16:52:43 -07:00
2011-10-28 14:00:21 -07:00
## Configure apache's 000-default to run horizon
2011-09-16 11:31:16 -07:00
sudo cp $FILES /000-default.template /etc/apache2/sites-enabled/000-default
2012-01-24 11:45:52 -06:00
sudo sed -e "
s,%USER%,$APACHE_USER ,g;
s,%GROUP%,$APACHE_GROUP ,g;
s,%HORIZON_DIR%,$HORIZON_DIR ,g;
" -i /etc/apache2/sites-enabled/000-default
2011-10-28 10:34:26 -07:00
sudo service apache2 restart
2011-09-15 16:52:43 -07:00
fi
2011-09-12 19:29:56 -07:00
2011-09-12 21:46:12 -07:00
2011-09-13 01:24:50 -07:00
# Glance
# ------
2012-02-16 10:16:52 +00:00
if is_service_enabled g-reg; then
2011-09-20 10:38:06 -07:00
GLANCE_IMAGE_DIR = $DEST /glance/images
2011-09-20 09:59:54 -07:00
# Delete existing images
rm -rf $GLANCE_IMAGE_DIR
# Use local glance directories
mkdir -p $GLANCE_IMAGE_DIR
2011-09-15 16:52:43 -07:00
# (re)create glance database
2011-10-12 07:13:13 +00:00
mysql -u$MYSQL_USER -p$MYSQL_PASSWORD -e 'DROP DATABASE IF EXISTS glance;'
mysql -u$MYSQL_USER -p$MYSQL_PASSWORD -e 'CREATE DATABASE glance;'
2011-10-19 09:24:17 -07:00
2012-01-16 22:27:20 -06:00
function glance_config {
sudo sed -e "
s,%KEYSTONE_AUTH_HOST%,$KEYSTONE_AUTH_HOST ,g;
s,%KEYSTONE_AUTH_PORT%,$KEYSTONE_AUTH_PORT ,g;
s,%KEYSTONE_AUTH_PROTOCOL%,$KEYSTONE_AUTH_PROTOCOL ,g;
s,%KEYSTONE_SERVICE_HOST%,$KEYSTONE_SERVICE_HOST ,g;
s,%KEYSTONE_SERVICE_PORT%,$KEYSTONE_SERVICE_PORT ,g;
s,%KEYSTONE_SERVICE_PROTOCOL%,$KEYSTONE_SERVICE_PROTOCOL ,g;
s,%SQL_CONN%,$BASE_SQL_CONN /glance,g;
s,%SERVICE_TOKEN%,$SERVICE_TOKEN ,g;
s,%DEST%,$DEST ,g;
s,%SYSLOG%,$SYSLOG ,g;
" -i $1
}
2011-10-19 09:24:17 -07:00
# Copy over our glance configurations and update them
2012-01-16 22:27:20 -06:00
GLANCE_REGISTRY_CONF = $GLANCE_DIR /etc/glance-registry.conf
cp $FILES /glance-registry.conf $GLANCE_REGISTRY_CONF
glance_config $GLANCE_REGISTRY_CONF
if [ [ -e $FILES /glance-registry-paste.ini ] ] ; then
GLANCE_REGISTRY_PASTE_INI = $GLANCE_DIR /etc/glance-registry-paste.ini
cp $FILES /glance-registry-paste.ini $GLANCE_REGISTRY_PASTE_INI
glance_config $GLANCE_REGISTRY_PASTE_INI
fi
2011-09-20 00:33:51 -07:00
GLANCE_API_CONF = $GLANCE_DIR /etc/glance-api.conf
cp $FILES /glance-api.conf $GLANCE_API_CONF
2012-01-16 22:27:20 -06:00
glance_config $GLANCE_API_CONF
if [ [ -e $FILES /glance-api-paste.ini ] ] ; then
GLANCE_API_PASTE_INI = $GLANCE_DIR /etc/glance-api-paste.ini
cp $FILES /glance-api-paste.ini $GLANCE_API_PASTE_INI
glance_config $GLANCE_API_PASTE_INI
fi
2011-09-15 16:52:43 -07:00
fi
2011-09-12 17:09:08 -07:00
2011-09-13 01:24:50 -07:00
# Nova
# ----
2012-02-13 11:22:36 -06:00
# Put config files in /etc/nova for everyone to find
NOVA_CONF = /etc/nova
if [ [ ! -d $NOVA_CONF ] ] ; then
sudo mkdir -p $NOVA_CONF
fi
sudo chown ` whoami` $NOVA_CONF
2012-02-16 10:16:52 +00:00
if is_service_enabled n-api; then
2011-11-06 11:18:26 -08:00
# We are going to use a sample http middleware configuration based on the
# one from the keystone project to launch nova. This paste config adds
2012-01-11 11:34:13 -08:00
# the configuration required for nova to validate keystone tokens.
# First we add a some extra data to the default paste config from nova
2012-02-13 11:22:36 -06:00
cp $NOVA_DIR /etc/nova/api-paste.ini $NOVA_CONF
2012-01-11 11:34:13 -08:00
# Then we add our own service token to the configuration
2012-02-13 11:22:36 -06:00
sed -e " s,%SERVICE_TOKEN%, $SERVICE_TOKEN ,g " -i $NOVA_CONF /api-paste.ini
2012-01-11 11:34:13 -08:00
# Finally, we change the pipelines in nova to use keystone
function replace_pipeline( ) {
2012-02-13 11:22:36 -06:00
sed " /\[pipeline: $1 \]/,/\[/s/^pipeline = .*/pipeline = $2 / " -i $NOVA_CONF /api-paste.ini
2012-01-11 11:34:13 -08:00
}
2012-01-23 17:36:42 -08:00
replace_pipeline "ec2cloud" "ec2faultwrap logrequest totoken authtoken keystonecontext cloudrequest authorizer validator ec2executor"
2012-01-11 11:34:13 -08:00
replace_pipeline "ec2admin" "ec2faultwrap logrequest totoken authtoken keystonecontext adminrequest authorizer ec2executor"
2012-01-23 14:41:50 -06:00
replace_pipeline "openstack_compute_api_v2" "faultwrap authtoken keystonecontext ratelimit osapi_compute_app_v2"
replace_pipeline "openstack_volume_api_v1" "faultwrap authtoken keystonecontext ratelimit osapi_volume_app_v1"
2011-10-24 21:56:25 -07:00
fi
2011-09-12 17:09:08 -07:00
2011-12-17 00:21:49 +00:00
# Helper to clean iptables rules
function clean_iptables( ) {
# Delete rules
sudo iptables -S -v | sed "s/-c [0-9]* [0-9]* //g" | grep "nova" | grep "\-A" | sed "s/-A/-D/g" | awk '{print "sudo iptables",$0}' | bash
# Delete nat rules
sudo iptables -S -v -t nat | sed "s/-c [0-9]* [0-9]* //g" | grep "nova" | grep "\-A" | sed "s/-A/-D/g" | awk '{print "sudo iptables -t nat",$0}' | bash
# Delete chains
sudo iptables -S -v | sed "s/-c [0-9]* [0-9]* //g" | grep "nova" | grep "\-N" | sed "s/-N/-X/g" | awk '{print "sudo iptables",$0}' | bash
# Delete nat chains
sudo iptables -S -v -t nat | sed "s/-c [0-9]* [0-9]* //g" | grep "nova" | grep "\-N" | sed "s/-N/-X/g" | awk '{print "sudo iptables -t nat",$0}' | bash
}
2012-02-16 10:16:52 +00:00
if is_service_enabled n-cpu; then
2011-09-13 13:17:22 -07:00
2011-10-02 17:47:32 -04:00
# Virtualization Configuration
# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~
2011-12-08 16:22:27 -05:00
apt_get install libvirt-bin
2011-10-02 17:47:32 -04:00
# attempt to load modules: network block device - used to manage qcow images
2011-09-15 16:52:43 -07:00
sudo modprobe nbd || true
2011-10-02 13:11:28 -04:00
2011-10-20 10:07:10 -07:00
# Check for kvm (hardware based virtualization). If unable to initialize
# kvm, we drop back to the slower emulation mode (qemu). Note: many systems
2011-10-19 09:24:17 -07:00
# come with hardware virtualization disabled in BIOS.
2011-10-03 22:48:30 -04:00
if [ [ " $LIBVIRT_TYPE " = = "kvm" ] ] ; then
2011-10-02 17:47:32 -04:00
sudo modprobe kvm || true
2011-10-02 13:11:28 -04:00
if [ ! -e /dev/kvm ] ; then
2011-10-02 17:47:32 -04:00
echo "WARNING: Switching to QEMU"
2011-10-02 13:11:28 -04:00
LIBVIRT_TYPE = qemu
fi
fi
2011-10-02 17:47:32 -04:00
# Install and configure **LXC** if specified. LXC is another approach to
# splitting a system into many smaller parts. LXC uses cgroups and chroot
# to simulate multiple systems.
2011-10-03 22:48:30 -04:00
if [ [ " $LIBVIRT_TYPE " = = "lxc" ] ] ; then
2011-12-08 16:21:52 -05:00
if [ [ " $DISTRO " > natty ] ] ; then
apt_get install cgroup-lite
else
cgline = "none /cgroup cgroup cpuacct,memory,devices,cpu,freezer,blkio 0 0"
sudo mkdir -p /cgroup
if ! grep -q cgroup /etc/fstab; then
echo " $cgline " | sudo tee -a /etc/fstab
fi
if ! mount -n | grep -q cgroup; then
sudo mount /cgroup
fi
2011-10-07 10:34:32 -04:00
fi
2011-09-16 16:28:13 -07:00
fi
2011-10-02 13:11:28 -04:00
2011-10-19 09:24:17 -07:00
# The user that nova runs as needs to be member of libvirtd group otherwise
# nova-compute will be unable to use libvirt.
2011-09-15 16:52:43 -07:00
sudo usermod -a -G libvirtd ` whoami`
2011-10-20 10:07:10 -07:00
# libvirt detects various settings on startup, as we potentially changed
2011-10-19 09:24:17 -07:00
# the system configuration (modules, filesystems), we need to restart
# libvirt to detect those changes.
2011-09-15 16:52:43 -07:00
sudo /etc/init.d/libvirt-bin restart
2011-09-13 13:17:22 -07:00
2011-10-02 17:47:32 -04:00
# Instance Storage
# ~~~~~~~~~~~~~~~~
# Nova stores each instance in its own directory.
2011-09-15 16:52:43 -07:00
mkdir -p $NOVA_DIR /instances
2011-09-12 17:09:08 -07:00
2011-10-19 09:24:17 -07:00
# You can specify a different disk to be mounted and used for backing the
2011-10-20 10:07:10 -07:00
# virtual machines. If there is a partition labeled nova-instances we
2011-10-19 09:24:17 -07:00
# mount it (ext filesystems can be labeled via e2label).
2011-09-15 16:52:43 -07:00
if [ -L /dev/disk/by-label/nova-instances ] ; then
2011-11-20 09:55:44 -08:00
if ! mount -n | grep -q $NOVA_DIR /instances; then
2011-10-19 09:24:17 -07:00
sudo mount -L nova-instances $NOVA_DIR /instances
sudo chown -R ` whoami` $NOVA_DIR /instances
fi
2011-09-15 16:52:43 -07:00
fi
2011-12-17 00:21:49 +00:00
# Clean iptables from previous runs
clean_iptables
# Destroy old instances
2012-01-04 09:32:48 -08:00
instances = ` virsh list --all | grep $INSTANCE_NAME_PREFIX | sed " s/.*\( $INSTANCE_NAME_PREFIX [0-9a-fA-F]*\).*/\1/g " `
2012-01-20 12:45:32 -08:00
if [ ! " $instances " = "" ] ; then
2012-01-04 09:32:48 -08:00
echo $instances | xargs -n1 virsh destroy || true
echo $instances | xargs -n1 virsh undefine || true
2011-12-17 00:21:49 +00:00
fi
2012-02-11 00:17:31 -08:00
# Logout and delete iscsi sessions
sudo iscsiadm --mode node | grep $VOLUME_NAME_PREFIX | cut -d " " -f2 | xargs sudo iscsiadm --mode node --logout || true
sudo iscsiadm --mode node | grep $VOLUME_NAME_PREFIX | cut -d " " -f2 | sudo iscsiadm --mode node --op delete || true
2011-10-02 17:47:32 -04:00
# Clean out the instances directory.
2011-10-09 17:50:38 -07:00
sudo rm -rf $NOVA_DIR /instances/*
2011-09-15 16:52:43 -07:00
fi
2012-02-16 10:16:52 +00:00
if is_service_enabled n-net; then
2011-12-17 00:21:49 +00:00
# Delete traces of nova networks from prior runs
2011-09-20 02:23:54 -07:00
sudo killall dnsmasq || true
2011-12-17 00:21:49 +00:00
clean_iptables
2011-09-15 16:52:43 -07:00
rm -rf $NOVA_DIR /networks
mkdir -p $NOVA_DIR /networks
fi
2011-11-01 12:30:55 +01:00
# Storage Service
2012-02-16 10:16:52 +00:00
if is_service_enabled swift; then
2011-11-02 17:57:11 +01:00
# We first do a bit of setup by creating the directories and
# changing the permissions so we can run it as our user.
2011-11-02 14:25:06 +01:00
USER_GROUP = $( id -g)
2011-11-03 09:17:06 +01:00
sudo mkdir -p ${ SWIFT_DATA_LOCATION } /drives
2011-11-22 13:04:40 +01:00
sudo chown -R $USER :${ USER_GROUP } ${ SWIFT_DATA_LOCATION }
2011-11-05 16:12:20 -07:00
2011-11-02 17:57:11 +01:00
# We then create a loopback disk and format it to XFS.
2011-11-22 13:04:40 +01:00
# TODO: Reset disks on new pass.
2011-11-14 15:26:13 +01:00
if [ [ ! -e ${ SWIFT_DATA_LOCATION } /drives/images/swift.img ] ] ; then
2011-11-03 09:17:06 +01:00
mkdir -p ${ SWIFT_DATA_LOCATION } /drives/images
sudo touch ${ SWIFT_DATA_LOCATION } /drives/images/swift.img
sudo chown $USER : ${ SWIFT_DATA_LOCATION } /drives/images/swift.img
2011-11-05 16:12:20 -07:00
2011-11-03 09:17:06 +01:00
dd if = /dev/zero of = ${ SWIFT_DATA_LOCATION } /drives/images/swift.img \
2011-11-02 16:49:56 +01:00
bs = 1024 count = 0 seek = ${ SWIFT_LOOPBACK_DISK_SIZE }
2011-11-03 09:17:06 +01:00
mkfs.xfs -f -i size = 1024 ${ SWIFT_DATA_LOCATION } /drives/images/swift.img
2011-11-01 12:30:55 +01:00
fi
2011-11-02 17:57:11 +01:00
# After the drive being created we mount the disk with a few mount
# options to make it most efficient as possible for swift.
2011-11-03 09:17:06 +01:00
mkdir -p ${ SWIFT_DATA_LOCATION } /drives/sdb1
2011-11-14 15:26:13 +01:00
if ! egrep -q ${ SWIFT_DATA_LOCATION } /drives/sdb1 /proc/mounts; then
2011-11-02 16:49:56 +01:00
sudo mount -t xfs -o loop,noatime,nodiratime,nobarrier,logbufs= 8 \
2011-11-03 09:17:06 +01:00
${ SWIFT_DATA_LOCATION } /drives/images/swift.img ${ SWIFT_DATA_LOCATION } /drives/sdb1
2011-11-01 13:08:29 +00:00
fi
2011-11-01 12:30:55 +01:00
2011-11-02 17:57:11 +01:00
# We then create link to that mounted location so swift would know
# where to go.
2011-11-03 09:17:06 +01:00
for x in { 1..4} ; do sudo ln -sf ${ SWIFT_DATA_LOCATION } /drives/sdb1/$x ${ SWIFT_DATA_LOCATION } /$x ; done
2011-11-05 16:12:20 -07:00
2011-11-02 17:57:11 +01:00
# We now have to emulate a few different servers into one we
2011-11-05 16:12:20 -07:00
# create all the directories needed for swift
2011-11-01 17:32:11 +01:00
tmpd = ""
2011-11-03 10:43:46 +01:00
for d in ${ SWIFT_DATA_LOCATION } /drives/sdb1/{ 1..4} \
${ SWIFT_CONFIG_LOCATION } /{ object,container,account} -server \
2011-11-14 15:26:13 +01:00
${ SWIFT_DATA_LOCATION } /{ 1..4} /node/sdb1 /var/run/swift; do
2011-11-01 17:32:11 +01:00
[ [ -d $d ] ] && continue
2011-11-02 14:25:06 +01:00
sudo install -o ${ USER } -g $USER_GROUP -d $d
2011-11-01 17:32:11 +01:00
done
2011-11-03 10:43:46 +01:00
# We do want to make sure this is all owned by our user.
sudo chown -R $USER : ${ SWIFT_DATA_LOCATION } /{ 1..4} /node
sudo chown -R $USER : ${ SWIFT_CONFIG_LOCATION }
2011-11-01 12:30:55 +01:00
2011-11-03 10:43:46 +01:00
# swift-init has a bug using /etc/swift until bug #885595 is fixed
# we have to create a link
2011-11-10 23:46:08 +01:00
sudo ln -sf ${ SWIFT_CONFIG_LOCATION } /etc/swift
2011-11-05 16:12:20 -07:00
2011-11-02 17:57:11 +01:00
# Swift use rsync to syncronize between all the different
# partitions (which make more sense when you have a multi-node
# setup) we configure it with our version of rsync.
2011-11-03 09:17:06 +01:00
sed -e " s/%GROUP%/ ${ USER_GROUP } /;s/%USER%/ $USER /;s,%SWIFT_DATA_LOCATION%, $SWIFT_DATA_LOCATION , " $FILES /swift/rsyncd.conf | sudo tee /etc/rsyncd.conf
2011-11-01 17:32:11 +01:00
sudo sed -i '/^RSYNC_ENABLE=false/ { s/false/true/ }' /etc/default/rsync
2011-11-01 19:32:23 +01:00
2011-11-02 17:57:11 +01:00
# By default Swift will be installed with the tempauth middleware
# which has some default username and password if you have
# configured keystone it will checkout the directory.
2012-02-16 10:16:52 +00:00
if is_service_enabled key; then
2011-11-01 19:32:23 +01:00
swift_auth_server = keystone
2011-11-10 19:44:58 +01:00
# We install the memcache server as this is will be used by the
# middleware to cache the tokens auths for a long this is needed.
apt_get install memcached
2011-11-01 19:32:23 +01:00
else
swift_auth_server = tempauth
fi
2011-11-02 17:57:11 +01:00
# We do the install of the proxy-server and swift configuration
# replacing a few directives to match our configuration.
2011-11-03 10:43:46 +01:00
sed " s,%SWIFT_CONFIG_LOCATION%, ${ SWIFT_CONFIG_LOCATION } ,;s/%USER%/ $USER /;s/%SERVICE_TOKEN%/ ${ SERVICE_TOKEN } /;s/%AUTH_SERVER%/ ${ swift_auth_server } / " \
$FILES /swift/proxy-server.conf| sudo tee ${ SWIFT_CONFIG_LOCATION } /proxy-server.conf
2011-11-01 12:30:55 +01:00
2011-11-03 10:43:46 +01:00
sed -e " s/%SWIFT_HASH%/ $SWIFT_HASH / " $FILES /swift/swift.conf > ${ SWIFT_CONFIG_LOCATION } /swift.conf
2011-11-01 12:30:55 +01:00
# We need to generate a object/account/proxy configuration
2011-11-02 17:57:11 +01:00
# emulating 4 nodes on different ports we have a little function
2011-11-01 12:30:55 +01:00
# that help us doing that.
function generate_swift_configuration( ) {
local server_type = $1
local bind_port = $2
local log_facility = $3
2011-11-01 19:32:23 +01:00
local node_number
2011-11-05 16:12:20 -07:00
2011-11-14 15:26:13 +01:00
for node_number in { 1..4} ; do
2011-11-03 09:17:06 +01:00
node_path = ${ SWIFT_DATA_LOCATION } /${ node_number }
2011-11-03 10:43:46 +01:00
sed -e " s,%SWIFT_CONFIG_LOCATION%, ${ SWIFT_CONFIG_LOCATION } ,;s,%USER%, $USER ,;s,%NODE_PATH%, ${ node_path } ,;s,%BIND_PORT%, ${ bind_port } ,;s,%LOG_FACILITY%, ${ log_facility } , " \
$FILES /swift/${ server_type } -server.conf > ${ SWIFT_CONFIG_LOCATION } /${ server_type } -server/${ node_number } .conf
2011-11-01 12:30:55 +01:00
bind_port = $(( ${ bind_port } + 10 ))
log_facility = $(( ${ log_facility } + 1 ))
done
}
generate_swift_configuration object 6010 2
generate_swift_configuration container 6011 2
generate_swift_configuration account 6012 2
2011-11-01 13:08:29 +00:00
2011-11-22 13:04:40 +01:00
# We have some specific configuration for swift for rsyslog. See
# the file /etc/rsyslog.d/10-swift.conf for more info.
swift_log_dir = ${ SWIFT_DATA_LOCATION } /logs
rm -rf ${ swift_log_dir }
mkdir -p ${ swift_log_dir } /hourly
sudo chown -R syslog:adm ${ swift_log_dir }
sed " s,%SWIFT_LOGDIR%, ${ swift_log_dir } , " $FILES /swift/rsyslog.conf | sudo \
tee /etc/rsyslog.d/10-swift.conf
sudo restart rsyslog
2012-01-11 11:34:13 -08:00
2011-11-02 17:57:11 +01:00
# We create two helper scripts :
#
# - swift-remakerings
# Allow to recreate rings from scratch.
# - swift-startmain
# Restart your full cluster.
#
2011-11-03 10:43:46 +01:00
sed -e " s,%SWIFT_CONFIG_LOCATION%, ${ SWIFT_CONFIG_LOCATION } ,;s/%SWIFT_PARTITION_POWER_SIZE%/ $SWIFT_PARTITION_POWER_SIZE / " $FILES /swift/swift-remakerings | \
2011-11-01 15:36:00 +01:00
sudo tee /usr/local/bin/swift-remakerings
2011-11-02 19:09:30 +01:00
sudo install -m755 $FILES /swift/swift-startmain /usr/local/bin/
2011-11-01 15:36:00 +01:00
sudo chmod +x /usr/local/bin/swift-*
2011-11-02 17:57:11 +01:00
# We then can start rsync.
2011-11-01 17:32:11 +01:00
sudo /etc/init.d/rsync restart || :
2011-11-05 16:12:20 -07:00
2011-11-02 17:57:11 +01:00
# Create our ring for the object/container/account.
2011-11-01 16:22:08 +01:00
/usr/local/bin/swift-remakerings
2011-11-02 17:57:11 +01:00
# And now we launch swift-startmain to get our cluster running
# ready to be tested.
2011-11-01 16:22:08 +01:00
/usr/local/bin/swift-startmain || :
2011-11-05 16:12:20 -07:00
2011-11-01 19:32:23 +01:00
unset s swift_hash swift_auth_server tmpd
2011-11-01 12:30:55 +01:00
fi
2011-10-20 10:12:58 -07:00
# Volume Service
# --------------
2012-02-16 10:16:52 +00:00
if is_service_enabled n-vol; then
2011-10-20 10:12:58 -07:00
#
# Configure a default volume group called 'nova-volumes' for the nova-volume
# service if it does not yet exist. If you don't wish to use a file backed
# volume group, create your own volume group called 'nova-volumes' before
# invoking stack.sh.
#
# By default, the backing file is 2G in size, and is stored in /opt/stack.
2011-11-20 09:55:44 -08:00
2012-01-14 01:08:34 +00:00
# install the package
apt_get install tgt
2011-11-20 09:55:44 -08:00
2011-12-01 17:02:07 -06:00
if ! sudo vgs $VOLUME_GROUP ; then
2011-11-03 18:19:21 -04:00
VOLUME_BACKING_FILE = ${ VOLUME_BACKING_FILE :- $DEST /nova-volumes-backing-file }
2011-10-20 10:26:30 -07:00
VOLUME_BACKING_FILE_SIZE = ${ VOLUME_BACKING_FILE_SIZE :- 2052M }
2011-12-01 17:02:07 -06:00
# Only create if the file doesn't already exists
[ [ -f $VOLUME_BACKING_FILE ] ] || truncate -s $VOLUME_BACKING_FILE_SIZE $VOLUME_BACKING_FILE
2011-10-20 10:12:58 -07:00
DEV = ` sudo losetup -f --show $VOLUME_BACKING_FILE `
2011-12-01 17:02:07 -06:00
# Only create if the loopback device doesn't contain $VOLUME_GROUP
if ! sudo vgs $VOLUME_GROUP ; then sudo vgcreate $VOLUME_GROUP $DEV ; fi
fi
if sudo vgs $VOLUME_GROUP ; then
2012-02-08 00:54:20 +00:00
# Remove nova iscsi targets
sudo tgtadm --op show --mode target | grep $VOLUME_NAME_PREFIX | grep Target | cut -f3 -d ' ' | sudo xargs -n1 tgt-admin --delete || true
2011-12-01 17:02:07 -06:00
# Clean out existing volumes
for lv in ` sudo lvs --noheadings -o lv_name $VOLUME_GROUP ` ; do
# VOLUME_NAME_PREFIX prefixes the LVs we want
if [ [ " ${ lv # $VOLUME_NAME_PREFIX } " != " $lv " ] ] ; then
sudo lvremove -f $VOLUME_GROUP /$lv
fi
done
2011-10-20 10:12:58 -07:00
fi
2012-01-14 01:08:34 +00:00
# tgt in oneiric doesn't restart properly if tgtd isn't running
# do it in two steps
sudo stop tgt || true
sudo start tgt
2011-10-20 10:12:58 -07:00
fi
2011-09-16 16:28:13 -07:00
function add_nova_flag {
2012-02-13 11:22:36 -06:00
echo " $1 " >> $NOVA_CONF /nova.conf
2011-09-16 16:28:13 -07:00
}
# (re)create nova.conf
2012-02-13 11:22:36 -06:00
rm -f $NOVA_CONF /nova.conf
2011-09-16 16:28:13 -07:00
add_nova_flag "--verbose"
2011-10-27 12:55:29 -07:00
add_nova_flag "--allow_admin_api"
2011-09-25 22:28:08 -07:00
add_nova_flag " --scheduler_driver= $SCHEDULER "
2012-02-22 10:18:31 -06:00
add_nova_flag " --dhcpbridge_flagfile= $NOVA_CONF /nova.conf "
2011-10-29 23:30:10 -04:00
add_nova_flag " --fixed_range= $FIXED_RANGE "
2012-02-16 10:16:52 +00:00
if is_service_enabled n-obj; then
2012-01-19 23:28:46 +00:00
add_nova_flag " --s3_host= $SERVICE_HOST "
fi
2012-02-16 10:16:52 +00:00
if is_service_enabled quantum; then
2011-10-27 18:18:20 -07:00
add_nova_flag "--network_manager=nova.network.quantum.manager.QuantumManager"
2011-12-12 23:04:58 +00:00
add_nova_flag " --quantum_connection_host= $Q_HOST "
add_nova_flag " --quantum_connection_port= $Q_PORT "
2012-01-23 11:17:27 -06:00
2012-02-16 10:16:52 +00:00
if is_service_enabled melange; then
2012-01-23 11:17:27 -06:00
add_nova_flag "--quantum_ipam_lib=nova.network.quantum.melange_ipam_lib"
add_nova_flag "--use_melange_mac_generation"
add_nova_flag " --melange_host= $M_HOST "
add_nova_flag " --melange_port= $M_PORT "
fi
2012-02-16 10:16:52 +00:00
if is_service_enabled q-svc && [ [ " $Q_PLUGIN " = "openvswitch" ] ] ; then
2011-10-27 18:18:20 -07:00
add_nova_flag "--libvirt_vif_type=ethernet"
add_nova_flag "--libvirt_vif_driver=nova.virt.libvirt.vif.LibvirtOpenVswitchDriver"
2011-11-29 06:36:03 +00:00
add_nova_flag "--linuxnet_interface_driver=nova.network.linux_net.LinuxOVSInterfaceDriver"
2012-01-16 14:46:01 -08:00
add_nova_flag "--quantum_use_dhcp"
2011-10-27 18:18:20 -07:00
fi
else
add_nova_flag " --network_manager=nova.network.manager. $NET_MAN "
fi
2012-02-16 10:16:52 +00:00
if is_service_enabled n-vol; then
2011-11-14 08:59:05 -08:00
add_nova_flag " --volume_group= $VOLUME_GROUP "
2011-12-01 17:02:07 -06:00
add_nova_flag " --volume_name_template= ${ VOLUME_NAME_PREFIX } %08x "
2012-01-14 01:08:34 +00:00
# oneiric no longer supports ietadm
add_nova_flag "--iscsi_helper=tgtadm"
2011-11-14 08:59:05 -08:00
fi
2012-01-26 12:29:51 -08:00
add_nova_flag "--osapi_compute_extension=nova.api.openstack.compute.contrib.standard_extensions"
2011-09-16 16:28:13 -07:00
add_nova_flag " --my_ip= $HOST_IP "
2011-09-20 09:39:50 -07:00
add_nova_flag " --public_interface= $PUBLIC_INTERFACE "
add_nova_flag " --vlan_interface= $VLAN_INTERFACE "
2011-09-16 16:28:13 -07:00
add_nova_flag " --sql_connection= $BASE_SQL_CONN /nova "
add_nova_flag " --libvirt_type= $LIBVIRT_TYPE "
2011-12-17 00:21:49 +00:00
add_nova_flag " --instance_name_template= ${ INSTANCE_NAME_PREFIX } %08x "
2012-02-03 20:17:22 +00:00
# All nova-compute workers need to know the vnc configuration options
# These settings don't hurt anything if n-xvnc and n-novnc are disabled
2012-02-16 10:16:52 +00:00
if is_service_enabled n-cpu; then
2012-01-17 15:46:53 -08:00
NOVNCPROXY_URL = ${ NOVNCPROXY_URL :- " http:// $SERVICE_HOST :6080/vnc_auto.html " }
add_nova_flag " --novncproxy_base_url= $NOVNCPROXY_URL "
XVPVNCPROXY_URL = ${ XVPVNCPROXY_URL :- " http:// $SERVICE_HOST :6081/console " }
add_nova_flag " --xvpvncproxy_base_url= $XVPVNCPROXY_URL "
fi
if [ " $VIRT_DRIVER " = 'xenserver' ] ; then
VNCSERVER_PROXYCLIENT_ADDRESS = ${ VNCSERVER_PROXYCLIENT_ADDRESS =169.254.0.1 }
else
VNCSERVER_PROXYCLIENT_ADDRESS = ${ VNCSERVER_PROXYCLIENT_ADDRESS =127.0.0.1 }
fi
2012-01-26 09:38:33 -08:00
# Address on which instance vncservers will listen on compute hosts.
# For multi-host, this should be the management ip of the compute host.
VNCSERVER_LISTEN = ${ VNCSERVER_LISTEN =127.0.0.1 }
add_nova_flag " --vncserver_listen= $VNCSERVER_LISTEN "
2012-01-17 15:46:53 -08:00
add_nova_flag " --vncserver_proxyclient_address= $VNCSERVER_PROXYCLIENT_ADDRESS "
2012-02-13 11:22:36 -06:00
add_nova_flag " --api_paste_config= $NOVA_CONF /api-paste.ini "
2011-09-16 16:28:13 -07:00
add_nova_flag "--image_service=nova.image.glance.GlanceImageService"
add_nova_flag " --ec2_dmz_host= $EC2_DMZ_HOST "
add_nova_flag " --rabbit_host= $RABBIT_HOST "
2011-10-02 14:28:17 -04:00
add_nova_flag " --rabbit_password= $RABBIT_PASSWORD "
2011-09-16 16:28:13 -07:00
add_nova_flag " --glance_api_servers= $GLANCE_HOSTPORT "
2011-11-09 00:12:00 -08:00
add_nova_flag "--force_dhcp_release"
2011-10-28 12:20:07 -07:00
if [ -n " $INSTANCES_PATH " ] ; then
add_nova_flag " --instances_path= $INSTANCES_PATH "
2011-10-28 13:34:38 -07:00
fi
2011-11-05 16:12:20 -07:00
if [ " $MULTI_HOST " != "False" ] ; then
add_nova_flag "--multi_host"
add_nova_flag "--send_arp_for_ha"
2011-09-16 16:28:13 -07:00
fi
2011-10-26 15:44:27 -04:00
if [ " $SYSLOG " != "False" ] ; then
2011-11-05 16:12:20 -07:00
add_nova_flag "--use_syslog"
2011-10-26 15:44:27 -04:00
fi
2011-09-16 16:28:13 -07:00
2011-11-20 09:55:44 -08:00
# You can define extra nova conf flags by defining the array EXTRA_FLAGS,
# For Example: EXTRA_FLAGS=(--foo --bar=2)
for I in " ${ EXTRA_FLAGS [@] } " ; do
2011-11-21 15:15:43 -08:00
add_nova_flag $I
2011-11-20 09:55:44 -08:00
done
2011-10-26 22:29:08 -07:00
# XenServer
# ---------
if [ " $VIRT_DRIVER " = 'xenserver' ] ; then
read_password XENAPI_PASSWORD "ENTER A PASSWORD TO USE FOR XEN."
add_nova_flag "--connection_type=xenapi"
add_nova_flag "--xenapi_connection_url=http://169.254.0.1"
add_nova_flag "--xenapi_connection_username=root"
add_nova_flag " --xenapi_connection_password= $XENAPI_PASSWORD "
2012-01-11 17:35:40 -08:00
add_nova_flag "--noflat_injected"
2011-10-26 22:29:08 -07:00
add_nova_flag "--flat_interface=eth1"
2012-01-11 17:35:40 -08:00
add_nova_flag "--flat_network_bridge=xapi1"
2011-10-26 22:29:08 -07:00
add_nova_flag "--public_interface=eth3"
2012-01-17 15:46:53 -08:00
# Need to avoid crash due to new firewall support
XEN_FIREWALL_DRIVER = ${ XEN_FIREWALL_DRIVER :- "nova.virt.firewall.IptablesFirewallDriver" }
add_nova_flag " --firewall_driver= $XEN_FIREWALL_DRIVER "
2011-10-26 22:29:08 -07:00
else
2012-01-23 14:45:21 -08:00
add_nova_flag "--connection_type=libvirt"
2012-01-31 14:33:19 +00:00
LIBVIRT_FIREWALL_DRIVER = ${ LIBVIRT_FIREWALL_DRIVER :- "nova.virt.libvirt.firewall.IptablesFirewallDriver" }
add_nova_flag " --firewall_driver= $LIBVIRT_FIREWALL_DRIVER "
2011-10-26 22:29:08 -07:00
add_nova_flag " --flat_network_bridge= $FLAT_NETWORK_BRIDGE "
if [ -n " $FLAT_INTERFACE " ] ; then
add_nova_flag " --flat_interface= $FLAT_INTERFACE "
fi
fi
2011-10-02 17:47:32 -04:00
# Nova Database
# ~~~~~~~~~~~~~
# All nova components talk to a central database. We will need to do this step
# only once for an entire cluster.
2012-02-16 10:16:52 +00:00
if is_service_enabled mysql; then
2011-09-16 21:37:36 -07:00
# (re)create nova database
2011-10-12 07:13:13 +00:00
mysql -u$MYSQL_USER -p$MYSQL_PASSWORD -e 'DROP DATABASE IF EXISTS nova;'
mysql -u$MYSQL_USER -p$MYSQL_PASSWORD -e 'CREATE DATABASE nova;'
2011-10-02 17:47:32 -04:00
# (re)create nova database
2011-09-16 21:37:36 -07:00
$NOVA_DIR /bin/nova-manage db sync
fi
2011-09-13 01:24:50 -07:00
# Launch Services
# ===============
2011-09-13 00:59:54 -07:00
2011-09-12 19:29:56 -07:00
# nova api crashes if we start it with a regular screen command,
# so send the start command by forcing text into the window.
2011-09-13 13:17:22 -07:00
# Only run the services specified in ``ENABLED_SERVICES``
2011-09-16 15:18:53 -07:00
# our screen helper to launch a service in a hidden named screen
2011-09-12 19:29:56 -07:00
function screen_it {
2011-09-16 15:18:53 -07:00
NL = ` echo -ne '\015' `
2012-02-16 10:16:52 +00:00
if is_service_enabled $1 ; then
2012-02-09 16:27:58 +01:00
screen -S stack -X screen -t $1
# sleep to allow bash to be ready to be send the command - we are
# creating a new window in screen and then sends characters, so if
# bash isn't running by the time we send the command, nothing happens
sleep 1.5
screen -S stack -p $1 -X stuff " $2 $NL "
2011-09-13 11:28:56 -07:00
fi
2011-09-12 19:29:56 -07:00
}
2011-09-16 16:30:55 -07:00
# create a new named screen to run processes in
2012-02-09 16:36:15 +01:00
screen -d -m -S stack -t stack -s /bin/bash
2011-09-16 16:30:55 -07:00
sleep 1
2011-12-10 13:55:44 -08:00
# set a reasonable statusbar
screen -r stack -X hardstatus alwayslastline "%-Lw%{= BW}%50>%n%f* %t%{-}%+Lw%< %= %H"
2011-09-16 16:30:55 -07:00
2011-11-02 17:57:11 +01:00
# launch the glance registry service
2012-02-16 10:16:52 +00:00
if is_service_enabled g-reg; then
2011-09-19 14:46:53 -07:00
screen_it g-reg " cd $GLANCE_DIR ; bin/glance-registry --config-file=etc/glance-registry.conf "
fi
2011-10-02 17:50:41 -04:00
# launch the glance api and wait for it to answer before continuing
2012-02-16 10:16:52 +00:00
if is_service_enabled g-api; then
2011-09-19 14:46:53 -07:00
screen_it g-api " cd $GLANCE_DIR ; bin/glance-api --config-file=etc/glance-api.conf "
2011-10-11 09:26:29 -05:00
echo " Waiting for g-api ( $GLANCE_HOSTPORT ) to start... "
2012-01-13 12:13:59 -06:00
if ! timeout $SERVICE_TIMEOUT sh -c " while ! http_proxy= wget -q -O- http:// $GLANCE_HOSTPORT ; do sleep 1; done " ; then
2011-10-11 09:26:29 -05:00
echo "g-api did not start"
exit 1
fi
2011-09-19 14:46:53 -07:00
fi
2012-02-16 10:16:52 +00:00
if is_service_enabled key; then
2012-01-11 01:59:00 +00:00
# (re)create keystone database
mysql -u$MYSQL_USER -p$MYSQL_PASSWORD -e 'DROP DATABASE IF EXISTS keystone;'
mysql -u$MYSQL_USER -p$MYSQL_PASSWORD -e 'CREATE DATABASE keystone;'
# Configure keystone.conf
KEYSTONE_CONF = $KEYSTONE_DIR /etc/keystone.conf
cp $FILES /keystone.conf $KEYSTONE_CONF
sudo sed -e " s,%SQL_CONN%, $BASE_SQL_CONN /keystone,g " -i $KEYSTONE_CONF
sudo sed -e " s,%DEST%, $DEST ,g " -i $KEYSTONE_CONF
sudo sed -e " s,%SERVICE_TOKEN%, $SERVICE_TOKEN ,g " -i $KEYSTONE_CONF
2012-01-11 02:04:39 +00:00
sudo sed -e " s,%KEYSTONE_DIR%, $KEYSTONE_DIR ,g " -i $KEYSTONE_CONF
2012-01-11 01:59:00 +00:00
KEYSTONE_CATALOG = $KEYSTONE_DIR /etc/default_catalog.templates
cp $FILES /default_catalog.templates $KEYSTONE_CATALOG
sudo sed -e " s,%SERVICE_HOST%, $SERVICE_HOST ,g " -i $KEYSTONE_CATALOG
if [ " $SYSLOG " != "False" ] ; then
2012-01-18 10:19:15 -08:00
cp $KEYSTONE_DIR /etc/logging.conf.sample $KEYSTONE_DIR /etc/logging.conf
2012-01-11 01:59:00 +00:00
sed -i -e '/^handlers=devel$/s/=devel/=production/' \
2012-01-18 10:19:15 -08:00
$KEYSTONE_DIR /etc/logging.conf
2012-01-11 01:59:00 +00:00
sed -i -e "/^log_file/s/log_file/\#log_file/" \
$KEYSTONE_DIR /etc/keystone.conf
2012-01-18 10:19:15 -08:00
KEYSTONE_LOG_CONFIG = " --log-config $KEYSTONE_DIR /etc/logging.conf "
2012-01-11 01:59:00 +00:00
fi
fi
2011-10-02 17:50:41 -04:00
# launch the keystone and wait for it to answer before continuing
2012-02-16 10:16:52 +00:00
if is_service_enabled key; then
2012-02-02 17:31:19 -08:00
screen_it key " cd $KEYSTONE_DIR && $KEYSTONE_DIR /bin/keystone-all --config-file $KEYSTONE_CONF $KEYSTONE_LOG_CONFIG -d --debug "
2011-10-11 09:26:29 -05:00
echo "Waiting for keystone to start..."
2012-01-11 01:59:00 +00:00
if ! timeout $SERVICE_TIMEOUT sh -c " while ! http_proxy= wget -q -O- $KEYSTONE_SERVICE_PROTOCOL :// $KEYSTONE_SERVICE_HOST : $KEYSTONE_SERVICE_PORT /v2.0/; do sleep 1; done " ; then
2011-10-11 09:26:29 -05:00
echo "keystone did not start"
exit 1
fi
2012-01-11 01:59:00 +00:00
# initialize keystone with default users/endpoints
2012-01-18 10:19:15 -08:00
pushd $KEYSTONE_DIR
2012-01-11 01:59:00 +00:00
$KEYSTONE_DIR /bin/keystone-manage db_sync
2012-01-18 10:19:15 -08:00
popd
2012-02-09 03:50:57 +00:00
# keystone_data.sh creates services, admin and demo users, and roles.
SERVICE_ENDPOINT = $KEYSTONE_AUTH_PROTOCOL ://$KEYSTONE_AUTH_HOST :$KEYSTONE_AUTH_PORT /v2.0
ADMIN_PASSWORD = $ADMIN_PASSWORD SERVICE_TOKEN = $SERVICE_TOKEN SERVICE_ENDPOINT = $SERVICE_ENDPOINT DEVSTACK_DIR = $TOP_DIR ENABLED_SERVICES = $ENABLED_SERVICES bash $FILES /keystone_data.sh
2011-09-19 14:46:53 -07:00
fi
2012-01-11 01:59:00 +00:00
2011-10-02 17:50:41 -04:00
# launch the nova-api and wait for it to answer before continuing
2012-02-16 10:16:52 +00:00
if is_service_enabled n-api; then
2011-09-20 09:51:16 -07:00
screen_it n-api " cd $NOVA_DIR && $NOVA_DIR /bin/nova-api "
2011-10-11 09:26:29 -05:00
echo "Waiting for nova-api to start..."
2012-01-13 12:13:59 -06:00
if ! timeout $SERVICE_TIMEOUT sh -c "while ! http_proxy= wget -q -O- http://127.0.0.1:8774; do sleep 1; done" ; then
2011-10-11 09:26:29 -05:00
echo "nova-api did not start"
exit 1
fi
2011-09-19 14:46:53 -07:00
fi
2011-10-27 18:18:20 -07:00
2011-12-12 23:04:58 +00:00
# Quantum service
2012-02-16 10:16:52 +00:00
if is_service_enabled q-svc; then
2011-10-27 18:18:20 -07:00
if [ [ " $Q_PLUGIN " = "openvswitch" ] ] ; then
2011-12-12 23:04:58 +00:00
# Install deps
# FIXME add to files/apts/quantum, but don't install if not needed!
apt_get install openvswitch-switch openvswitch-datapath-dkms
# Create database for the plugin/agent
2012-02-16 10:16:52 +00:00
if is_service_enabled mysql; then
2012-01-23 01:56:22 -08:00
mysql -u$MYSQL_USER -p$MYSQL_PASSWORD -e 'DROP DATABASE IF EXISTS ovs_quantum;'
2011-10-27 18:18:20 -07:00
mysql -u$MYSQL_USER -p$MYSQL_PASSWORD -e 'CREATE DATABASE IF NOT EXISTS ovs_quantum;'
else
2011-10-28 08:28:26 -07:00
echo " mysql must be enabled in order to use the $Q_PLUGIN Quantum plugin. "
exit 1
2011-10-27 18:18:20 -07:00
fi
2011-12-12 23:04:58 +00:00
QUANTUM_PLUGIN_INI_FILE = $QUANTUM_DIR /etc/plugins.ini
# Make sure we're using the openvswitch plugin
sed -i -e " s/^provider =.* $/provider = quantum.plugins.openvswitch.ovs_quantum_plugin.OVSQuantumPlugin/g " $QUANTUM_PLUGIN_INI_FILE
2011-10-27 18:18:20 -07:00
fi
2012-01-25 17:22:15 -05:00
screen_it q-svc " cd $QUANTUM_DIR && PYTHONPATH=.: $QUANTUM_CLIENT_DIR : $PYTHONPATH python $QUANTUM_DIR /bin/quantum-server $QUANTUM_DIR /etc/quantum.conf "
2011-10-27 18:18:20 -07:00
fi
# Quantum agent (for compute nodes)
2012-02-16 10:16:52 +00:00
if is_service_enabled q-agt; then
2011-10-27 18:18:20 -07:00
if [ [ " $Q_PLUGIN " = "openvswitch" ] ] ; then
# Set up integration bridge
OVS_BRIDGE = ${ OVS_BRIDGE :- br -int }
sudo ovs-vsctl --no-wait -- --if-exists del-br $OVS_BRIDGE
sudo ovs-vsctl --no-wait add-br $OVS_BRIDGE
sudo ovs-vsctl --no-wait br-set-external-id $OVS_BRIDGE bridge-id br-int
2012-01-16 14:46:01 -08:00
# Start up the quantum <-> openvswitch agent
QUANTUM_OVS_CONFIG_FILE = $QUANTUM_DIR /etc/quantum/plugins/openvswitch/ovs_quantum_plugin.ini
sed -i -e " s/^sql_connection =.* $/sql_connection = mysql:\/\/ $MYSQL_USER : $MYSQL_PASSWORD @ $MYSQL_HOST \/ovs_quantum/g " $QUANTUM_OVS_CONFIG_FILE
screen_it q-agt " sleep 4; sudo python $QUANTUM_DIR /quantum/plugins/openvswitch/agent/ovs_quantum_agent.py $QUANTUM_OVS_CONFIG_FILE -v "
2011-10-27 18:18:20 -07:00
fi
fi
2012-01-23 11:17:27 -06:00
# Melange service
2012-02-16 10:16:52 +00:00
if is_service_enabled m-svc; then
if is_service_enabled mysql; then
2012-01-23 11:17:27 -06:00
mysql -u$MYSQL_USER -p$MYSQL_PASSWORD -e 'DROP DATABASE IF EXISTS melange;'
mysql -u$MYSQL_USER -p$MYSQL_PASSWORD -e 'CREATE DATABASE melange;'
else
echo " mysql must be enabled in order to use the $Q_PLUGIN Quantum plugin. "
exit 1
fi
MELANGE_CONFIG_FILE = $MELANGE_DIR /etc/melange/melange.conf
cp $MELANGE_CONFIG_FILE .sample $MELANGE_CONFIG_FILE
sed -i -e " s/^sql_connection =.* $/sql_connection = mysql:\/\/ $MYSQL_USER : $MYSQL_PASSWORD @ $MYSQL_HOST \/melange/g " $MELANGE_CONFIG_FILE
cd $MELANGE_DIR && PYTHONPATH = .:$PYTHONPATH python $MELANGE_DIR /bin/melange-manage --config-file= $MELANGE_CONFIG_FILE db_sync
screen_it m-svc " cd $MELANGE_DIR && PYTHONPATH=.: $PYTHONPATH python $MELANGE_DIR /bin/melange-server --config-file= $MELANGE_CONFIG_FILE "
echo "Waiting for melange to start..."
if ! timeout $SERVICE_TIMEOUT sh -c "while ! http_proxy= wget -q -O- http://127.0.0.1:9898; do sleep 1; done" ; then
echo "melange-server did not start"
exit 1
fi
melange mac_address_range create cidr = $M_MAC_RANGE
fi
2011-10-28 08:28:26 -07:00
# If we're using Quantum (i.e. q-svc is enabled), network creation has to
# happen after we've started the Quantum service.
2012-02-16 10:16:52 +00:00
if is_service_enabled mysql; then
2011-10-27 18:18:20 -07:00
# create a small network
$NOVA_DIR /bin/nova-manage network create private $FIXED_RANGE 1 $FIXED_NETWORK_SIZE
2012-02-16 10:16:52 +00:00
if is_service_enabled q-svc; then
2011-10-27 18:18:20 -07:00
echo "Not creating floating IPs (not supported by QuantumManager)"
else
# create some floating ips
$NOVA_DIR /bin/nova-manage floating create $FLOATING_RANGE
2012-01-10 15:34:34 -06:00
# create a second pool
$NOVA_DIR /bin/nova-manage floating create --ip_range= $TEST_FLOATING_RANGE --pool= $TEST_FLOATING_POOL
2011-10-27 18:18:20 -07:00
fi
fi
2012-01-23 11:17:27 -06:00
2011-09-20 18:06:14 +00:00
# Launching nova-compute should be as simple as running ``nova-compute`` but
# have to do a little more than that in our script. Since we add the group
2011-09-16 15:18:53 -07:00
# ``libvirtd`` to our user in this script, when nova-compute is run it is
2011-09-20 18:06:14 +00:00
# within the context of our original shell (so our groups won't be updated).
2011-10-07 21:03:16 -04:00
# Use 'sg' to execute nova-compute as a member of the libvirtd group.
screen_it n-cpu " cd $NOVA_DIR && sg libvirtd $NOVA_DIR /bin/nova-compute "
2012-01-19 23:28:46 +00:00
screen_it n-crt " cd $NOVA_DIR && $NOVA_DIR /bin/nova-cert "
screen_it n-obj " cd $NOVA_DIR && $NOVA_DIR /bin/nova-objectstore "
2011-10-20 10:12:58 -07:00
screen_it n-vol " cd $NOVA_DIR && $NOVA_DIR /bin/nova-volume "
2011-09-20 09:51:16 -07:00
screen_it n-net " cd $NOVA_DIR && $NOVA_DIR /bin/nova-network "
screen_it n-sch " cd $NOVA_DIR && $NOVA_DIR /bin/nova-scheduler "
2012-02-16 10:16:52 +00:00
if is_service_enabled n-novnc; then
2012-02-22 10:18:31 -06:00
screen_it n-novnc " cd $NOVNC_DIR && ./utils/nova-novncproxy --flagfile $NOVA_CONF /nova.conf --web . "
2012-01-17 15:46:53 -08:00
fi
2012-02-16 10:16:52 +00:00
if is_service_enabled n-xvnc; then
2012-02-22 10:18:31 -06:00
screen_it n-xvnc " cd $NOVA_DIR && ./bin/nova-xvpvncproxy --flagfile $NOVA_CONF /nova.conf "
2012-01-17 15:46:53 -08:00
fi
2012-02-16 10:16:52 +00:00
if is_service_enabled n-cauth; then
2012-01-17 15:46:53 -08:00
screen_it n-cauth " cd $NOVA_DIR && ./bin/nova-consoleauth "
2011-11-10 13:05:13 -08:00
fi
2012-02-16 10:16:52 +00:00
if is_service_enabled horizon; then
2011-11-10 13:05:13 -08:00
screen_it horizon " cd $HORIZON_DIR && sudo tail -f /var/log/apache2/error.log "
fi
2011-09-12 17:09:08 -07:00
2011-09-13 01:24:50 -07:00
# Install Images
# ==============
2011-09-12 18:08:04 -07:00
2011-10-13 23:03:23 -07:00
# Upload an image to glance.
2011-10-03 01:08:24 -04:00
#
2011-10-13 23:03:23 -07:00
# The default image is a small ***TTY*** testing image, which lets you login
# the username/password of root/password.
#
# TTY also uses cloud-init, supporting login via keypair and sending scripts as
# userdata. See https://help.ubuntu.com/community/CloudInit for more on cloud-init
#
2011-11-02 17:57:11 +01:00
# Override ``IMAGE_URLS`` with a comma-separated list of uec images.
2011-10-19 10:30:19 -07:00
#
# * **natty**: http://uec-images.ubuntu.com/natty/current/natty-server-cloudimg-amd64.tar.gz
# * **oneiric**: http://uec-images.ubuntu.com/oneiric/current/oneiric-server-cloudimg-amd64.tar.gz
2011-10-02 23:42:56 -04:00
2012-02-16 10:16:52 +00:00
if is_service_enabled g-reg; then
2011-10-14 10:20:30 -07:00
# Create a directory for the downloaded image tarballs.
2011-10-02 23:42:56 -04:00
mkdir -p $FILES /images
2012-01-11 22:31:59 +00:00
ADMIN_USER = admin
ADMIN_TENANT = admin
TOKEN = ` curl -s -d " {\"auth\":{\"passwordCredentials\": {\"username\": \" $ADMIN_USER \", \"password\": \" $ADMIN_PASSWORD \"}, \"tenantName\": \" $ADMIN_TENANT \"}} " -H "Content-type: application/json" http://$HOST_IP :5000/v2.0/tokens | python -c "import sys; import json; tok = json.loads(sys.stdin.read()); print tok['access']['token']['id'];" `
2011-10-26 22:29:08 -07:00
# Option to upload legacy ami-tty, which works with xenserver
if [ $UPLOAD_LEGACY_TTY ] ; then
if [ ! -f $FILES /tty.tgz ] ; then
wget -c http://images.ansolabs.com/tty.tgz -O $FILES /tty.tgz
fi
tar -zxf $FILES /tty.tgz -C $FILES /images
2012-01-11 22:31:59 +00:00
RVAL = ` glance add -A $TOKEN name = "tty-kernel" is_public = true container_format = aki disk_format = aki < $FILES /images/aki-tty/image`
2011-10-26 22:29:08 -07:00
KERNEL_ID = ` echo $RVAL | cut -d":" -f2 | tr -d " " `
2012-01-11 22:31:59 +00:00
RVAL = ` glance add -A $TOKEN name = "tty-ramdisk" is_public = true container_format = ari disk_format = ari < $FILES /images/ari-tty/image`
2011-10-26 22:29:08 -07:00
RAMDISK_ID = ` echo $RVAL | cut -d":" -f2 | tr -d " " `
2012-01-11 22:31:59 +00:00
glance add -A $TOKEN name = "tty" is_public = true container_format = ami disk_format = ami kernel_id = $KERNEL_ID ramdisk_id = $RAMDISK_ID < $FILES /images/ami-tty/image
2011-10-26 22:29:08 -07:00
fi
2011-10-14 09:31:09 -07:00
for image_url in ${ IMAGE_URLS //,/ } ; do
# Downloads the image (uec ami+aki style), then extracts it.
2011-10-24 21:37:00 -07:00
IMAGE_FNAME = ` basename " $image_url " `
2011-10-14 09:31:09 -07:00
if [ ! -f $FILES /$IMAGE_FNAME ] ; then
wget -c $image_url -O $FILES /$IMAGE_FNAME
fi
2011-09-12 18:08:04 -07:00
2011-12-08 16:22:51 -05:00
KERNEL = ""
RAMDISK = ""
case " $IMAGE_FNAME " in
*.tar.gz| *.tgz)
# Extract ami and aki files
[ " ${ IMAGE_FNAME %.tar.gz } " != " $IMAGE_FNAME " ] &&
IMAGE_NAME = " ${ IMAGE_FNAME %.tar.gz } " ||
IMAGE_NAME = " ${ IMAGE_FNAME %.tgz } "
xdir = " $FILES /images/ $IMAGE_NAME "
rm -Rf " $xdir " ;
mkdir " $xdir "
tar -zxf $FILES /$IMAGE_FNAME -C " $xdir "
KERNEL = $( for f in " $xdir / " *-vmlinuz*; do
[ -f " $f " ] && echo " $f " && break; done ; true )
RAMDISK = $( for f in " $xdir / " *-initrd*; do
[ -f " $f " ] && echo " $f " && break; done ; true )
IMAGE = $( for f in " $xdir / " *.img; do
[ -f " $f " ] && echo " $f " && break; done ; true )
[ -n " $IMAGE_NAME " ]
IMAGE_NAME = $( basename " $IMAGE " ".img" )
; ;
*.img)
IMAGE = " $FILES / $IMAGE_FNAME " ;
IMAGE_NAME = $( basename " $IMAGE " ".img" )
; ;
*.img.gz)
IMAGE = " $FILES / ${ IMAGE_FNAME } "
IMAGE_NAME = $( basename " $IMAGE " ".img.gz" )
; ;
*) echo " Do not know what to do with $IMAGE_FNAME " ; false; ;
esac
2011-09-12 18:08:04 -07:00
2011-10-14 09:31:09 -07:00
# Use glance client to add the kernel the root filesystem.
# We parse the results of the first upload to get the glance ID of the
# kernel for use when uploading the root filesystem.
2011-12-08 16:22:51 -05:00
KERNEL_ID = "" ; RAMDISK_ID = "" ;
if [ -n " $KERNEL " ] ; then
2012-01-11 22:31:59 +00:00
RVAL = ` glance add -A $TOKEN name = " $IMAGE_NAME -kernel " is_public = true container_format = aki disk_format = aki < " $KERNEL " `
2011-12-08 16:22:51 -05:00
KERNEL_ID = ` echo $RVAL | cut -d":" -f2 | tr -d " " `
fi
if [ -n " $RAMDISK " ] ; then
2012-01-11 22:31:59 +00:00
RVAL = ` glance add -A $TOKEN name = " $IMAGE_NAME -ramdisk " is_public = true container_format = ari disk_format = ari < " $RAMDISK " `
2011-12-08 16:22:51 -05:00
RAMDISK_ID = ` echo $RVAL | cut -d":" -f2 | tr -d " " `
fi
2012-01-11 22:31:59 +00:00
glance add -A $TOKEN name = " ${ IMAGE_NAME %.img } " is_public = true container_format = ami disk_format = ami ${ KERNEL_ID : +kernel_id= $KERNEL_ID } ${ RAMDISK_ID : +ramdisk_id= $RAMDISK_ID } < <( zcat --force " ${ IMAGE } " )
2011-10-14 09:31:09 -07:00
done
2011-09-15 16:52:43 -07:00
fi
2011-09-15 21:28:23 -07:00
2011-10-07 14:51:07 +00:00
# Fin
# ===
2011-12-27 11:45:55 -06:00
set +o xtrace
2011-10-07 14:51:07 +00:00
2011-09-15 21:28:23 -07:00
# Using the cloud
# ===============
2011-11-01 20:06:55 -07:00
echo ""
echo ""
echo ""
2011-10-28 14:00:21 -07:00
# If you installed the horizon on this server, then you should be able
2011-09-20 18:06:14 +00:00
# to access the site using your browser.
2012-02-16 10:16:52 +00:00
if is_service_enabled horizon; then
2011-12-27 23:22:14 -08:00
echo " horizon is now available at http:// $SERVICE_HOST / "
2011-09-15 21:28:23 -07:00
fi
# If keystone is present, you can point nova cli to this server
2012-02-16 10:16:52 +00:00
if is_service_enabled key; then
2012-01-16 22:27:20 -06:00
echo " keystone is serving at $KEYSTONE_SERVICE_PROTOCOL :// $KEYSTONE_SERVICE_HOST : $KEYSTONE_SERVICE_PORT /v2.0/ "
2011-09-15 21:28:23 -07:00
echo "examples on using novaclient command line is in exercise.sh"
2011-10-02 14:11:17 -04:00
echo "the default users are: admin and demo"
echo " the password: $ADMIN_PASSWORD "
2011-09-15 21:28:23 -07:00
fi
2011-09-28 19:49:40 -05:00
2011-12-27 23:22:14 -08:00
# Echo HOST_IP - useful for build_uec.sh, which uses dhcp to give the instance an address
echo " This is your host ip: $HOST_IP "
# Indicate how long this took to run (bash maintained variable 'SECONDS')
2011-10-07 14:51:07 +00:00
echo " stack.sh completed in $SECONDS seconds. "