Ansible playbooks for deploying OpenStack.
Go to file
Jesse Pretorius b707e13333 Update apt cache after adding repos or pinning
This patch implements a forced apt cache update after a new apt
repository is added, or any apt pinning is implemented.

This ensures that apt is sure about what packages are available to it,
and which packages/repositories should be preferred above others.

Change-Id: I2f6e9fa33c7765e8ed8aeaf4ea60be696125c186
Closes-Bug: #1421679
2015-02-13 16:17:25 +00:00
etc Implement local scripts for all gate tests 2015-02-12 15:59:11 +00:00
rpc_deployment Update apt cache after adding repos or pinning 2015-02-13 16:17:25 +00:00
scripts Implement local scripts for all gate tests 2015-02-12 15:59:11 +00:00
ssh Initial Commit 2014-08-26 18:08:15 -05:00
.gitignore Initial Commit 2014-08-26 18:08:15 -05:00
.gitreview Adds .gitreview file to repo 2014-12-02 17:22:12 -06:00
CONTRIBUTING.md Workflow documentation is now in infra-manual 2014-12-05 03:30:45 +00:00
LICENSE Initial Commit 2014-08-26 18:08:15 -05:00
README.rst add links to launchpad and gerrit to README 2015-01-15 14:32:05 +00:00
requirements.txt Use HTTPS when installing from mirror.rackspace.com 2015-01-15 11:32:18 -06:00

README.rst

Openstack Deployment with Ansible

date

2014-09-25 09:00

tags

lxc, openstack, cloud, ansible

category

*nix

Official Documentation

Comprehensive installation guides, including FAQs and release notes, can be found at http://docs.rackspace.com

Bug tracking and release management can be found in Launchpad

Code reviews will be managed in Gerrit

Playbook Support

OpenStack:
  • keystone
  • glance-api
  • glance-registry
  • cinder-api
  • cinder-scheduler
  • cinder-volume
  • nova-api
  • nova-api-ec2
  • nova-api-metadata
  • nova-api-os-compute
  • nova-compute
  • nova-conductor
  • nova-scheduler
  • heat-api
  • heat-api-cfn
  • heat-api-cloudwatch
  • heat-engine
  • horizon
  • neutron-server
  • neutron-dhcp-agent
  • neutron-metadata-agent
  • neutron-linuxbridge-agent
Infrastructure:
  • galera
  • rabbitmq
  • logstash
  • elastic-search
  • kibana

Assumptions

This repo assumes that you have setup the host servers that will be running the OpenStack infrastructure with three bridged network devices named: br-mgmt, br-vxlan, br-vlan. These bridges will be used throughout the OpenStack infrastructure.

The repo also relies on configuration files found in the /etc directory of this repo. If you are running Ansible from an "unprivileged" host, you can place the contents of the /etc/ directory in your home folder; this would be in a directory similar to /home/<myusername>/rpc_deploy/. Once you have the file in place, you will have to enter the details of your environment in the rpc_user_config.yml file; please see the file for how this should look. After you have a bridged network and the files/directory in place, continue on to _Base Usage.

Base Usage

All commands must be executed from the rpc_deployment directory. From this directory you will have access to all of the playbooks, roles, and variables. It is recommended that you create an override file to contain any and all variables that you wish to override for the deployment. While the override file is is not required it will make life a bit easier. The default override file for the RPC environment is the user_variables.yml file.

All of the variables that you may wish to update are in the vars/ directory, however you should also be aware that services will pull in base group variables as found in inventory/group_vars.

All playbooks exist in the playbooks/ directory and are grouped in different sub-directories.

All of the keys, tokens, and passwords are in the user_variables.yml file. This file contains no preset passwords. To setup your keys, passwords, and tokens you will need to either edit this file manually or use the script pw-token-gen.py. Example:

# Generate the tokens
scripts/pw-token-gen.py --file /etc/rpc_deploy/user_variables.yml

Example usage from the rpc_deployment directory in the ansible-rpc-lxc repository

# Run setup on all hosts:
ansible-playbook -e @vars/user_variables.yml playbooks/setup/host-setup.yml

# Run infrastructure on all hosts
ansible-playbook -e @vars/user_variables.yml playbooks/infrastructure/infrastructure-setup.yml

# Setup and configure openstack within your spec'd containers
ansible-playbook -e @vars/user_variables.yml playbooks/openstack/openstack-setup.yml

About Inventory

All things that Ansible cares about are located in inventory. In the Rackspace Private Cloud all inventory is dynamically generated using the previously mentioned configuration files. While this is a dynamically generated inventory, it is not 100% generated on every run. The inventory is saved in a file named rpc_inventory.json and is located in the directory where you've located your user configuration files. On every run a backup of the inventory json file is created in both the current working directory as well as the location where the user configuration files exist. The inventory json file is a living document and is intended to grow as the environment scales in infrastructure. This means that the inventory file will be appended to as you add more nodes and or change the container affinity from within the rpc_user_config.yml file. It is recommended that the base inventory file be backed up to a safe location upon the completion of a deployment operation. While the dynamic inventory processor has guards in it to ensure that the built inventory is not adversely effected by programmatic operations this does not guard against user error and/or catastrophic failure.

Scaling

If you are scaling the environment using the dynamically generated inventory you should know that the inventory was designed to generate new entries in inventory and not remove entries from inventory. These playbooks will build an environment to spec so if container affinity is changed and or a node is added or removed from an environment the user configuration file will need to be modified as well as the inventory json. For this reason it is recommended that should a physical node need replacing it should be renamed the same as the previous one. This will make things easier when rebuilding the environment. Additionally if a container is needing to be replaced it is better to simply remove the misbehaving container and rebuild it using the existing inventory.

Notes

  • Library has an experimental keystone module which adds keystone: support to Ansible.
  • Library has an experimental swift module which adds swift: support to Ansible.
  • Library has an experimental neutron module which adds keystone: support to Ansible.
  • Library has an experimental glance module which adds keystone: support to Ansible.
  • Library has an experimental lxc module which adds lxc: support to Ansible.
  • Library has an experimental memcached module which adds lxc: support to Ansible.
  • Library has an experimental name2int module which adds lxc: support to Ansible.