11 KiB
Home OpenStack-Salt Development Documentation
OpenStack-Salt Heat deployment
All-in-one (AIO) deployments are a great way to setup an OpenStack-Salt cloud for:
- a service development environment
- an overview of how all of the OpenStack services and roles play together
- a simple lab deployment for testing
It is possible to run full size proof-of-concept deployment on OpenStack with Heat template, the stack has following requirements for cluster deployment:
- At least 200GB disk space
- 70GB RAM
The single-node deployment has following requirements:
- At least 80GB disk space
- 16GB RAM
Available Heat templates
We have prepared two generic OpenStack Salt lab templates, OpenStack in single and OpenStack in cluster configuration. Both are deployed by custom parametrized bootstrap script, which sets up Salt master with OpenStack Salt formula ecosystem and example metadata.
Openstack-salt single setup
The openstack_single environment consists of three
nodes.
| FQDN | Role | IP |
| config.openstack.local | Salt master node | 10.10.10.200 |
| control.openstack.local | OpenStack control node | 10.10.10.201 |
| compute.openstack.local | OpenStack compute node | 10.10.10.202 |
Openstack-salt cluster setup
The openstack_cluster environment consists of six
nodes.
| FQDN | Role | IP |
| config.openstack.local | Salt master node | 10.10.10.200 |
| control01.openstack.local | OpenStack control node | 10.10.10.201 |
| control02.openstack.local | OpenStack control node | 10.10.10.202 |
| control03.openstack.local | OpenStack control node | 10.10.10.203 |
| compute01.openstack.local | OpenStack compute node | 10.10.10.211 |
| compute02.openstack.local | OpenStack compute node | 10.10.10.212 |
Heat client setup
The preffered way of installing OpenStack clients is isolated Python environment. To creat Python environment and install compatible OpenStack clients, you need to install build tools first.
Ubuntu installation
Install required packages:
$ apt-get install python-dev python-pip python-virtualenv build-essentialNow create and activate virtualenv venv-heat so you can install specific versions of OpenStack clients.
$ virtualenv venv-heat
$ source ./venv-heat/bin/activateUse requirements.txt from the OpenStack-Salt heat templates repository to install tested versions of clients into activated environment.
$ pip install -r requirements.txtThe summary of clients for OpenStack. Following clients were tested with Juno and Kilo Openstack versions.
../../../scripts/requirements/heat.txt
If everything goes right, you should be able to use openstack clients, heat, nova, etc.
Connecting to OpenStack cloud
Setup OpenStack credentials so you can use openstack clients. You can
download openrc file from Openstack dashboard and source it
or execute following commands with filled credentials:
$ vim ~/openrc
export OS_AUTH_URL=https://<openstack_endpoint>:5000/v2.0
export OS_USERNAME=<username>
export OS_PASSWORD=<password>
export OS_TENANT_NAME=<tenant>Now source the OpenStack credentials:
$ source openrcTo test your sourced variables:
$ env | grep OSSome resources required for heat environment deployment.
Get network ID
The public network is needed for setting up both testing heat stacks. The network ID can be found in Openstack Dashboard or by running following command:
$ neutron net-listGet image ID
Image ID is required to run OpenStack Salt lab templates, Ubuntu 14.04 LTS is required as config_image and image for one of the supported platforms is required as instance_image, used for OpenStack instances. To lookup for actual installed images run:
$ glance image-listLaunching the Heat stack
Download heat templates from this repository.
$ git clone git@github.com:openstack/openstack-salt.git
$ cd doc/source/_static/scripts/Now you need to customize env files for stacks, see examples in envs
directory doc/source/_static/scripts/envs and set required
parameters.
Full examples of env files for the two respective stacks:
OpenStack templates generic parameters
- public_net_id
-
name of external network
- instance_image
-
image for OpenStack instances, must correspond with os_distribution
- config_image
-
image for Salt master node, currently only Ubuntu 14.04 supported
- key_value
-
paste your SSH key here
- salt_source
-
salt-master installation source
- options:
-
- pkg
- pip
- default:
-
- pkg
- salt_version
-
salt-master version
- options:
-
- latest
- 2015.8.11
- 2015.8.10
- 2015.8.9
- ...
- default:
-
- latest
- formula_source
-
salt formulas source
- options:
-
- git
- pkg
- default:
-
- git
- formula_path
-
path to formulas
- default:
-
- /usr/share/salt-formulas
- formula_branch
-
formulas git branch
- default:
-
- master
- reclass_address
-
reclass git repository
- reclass_branch
-
reclass git branch
- default:
-
- master
- os_version
-
OpenStack release version
- options:
-
- kilo
- default:
-
- kilo
- os_distribution
-
OpenStack nodes distribution
- options:
-
- ubuntu
- redhat
- default:
-
- ubuntu
- os_networking
-
OpenStack networking engine
- options:
-
- opencontrail
- neutron
- default:
-
- opencontrail
- os_deployment
-
OpenStack architecture
- options:
-
- single
- cluster
- default:
-
- single
- config_hostname
-
salt-master hostname
- default:
-
- config
- config_address
-
salt-master internal IP address
- default:
-
- 10.10.10.200
OpenStack single specific parameters
- config_domain
-
salt-master domain
- default:
-
- openstack.local
- ctl01_name
-
OS controller hostname
- default:
-
- control
- cmp01_name
-
OS compute hostname
- default:
-
- compute
- prx01_name
-
OS proxy hostname
- default:
-
- proxy
OpenStack cluster specific parameters
- config_domain
-
salt-master domain
- default:
-
- openstack-ha.local
- ctl01_name
-
OS controller 1 hostname
- default:
-
- control01
- ctl02_name
-
OS controller 2 hostname
- default:
-
- control02
- ctl03_name
-
OS controller 3 hostname
- default:
-
- control03
- cmp01_name
-
OS compute 1 hostname
- default:
-
- compute01
- cmp02_name
-
OS compute 2 hostname
- default:
-
- compute02
- prx01_name
-
OS proxy hostname
- default:
-
- proxy
openstack_single.hot environment examples
OpenStack Kilo on Ubuntu with OpenContrail networking
/_static/scripts/envs/openstack_single_kilo_ubuntu_opencontrail.env.example
/_static/scripts/envs/openstack_single_kilo_ubuntu_opencontrail.env.example
OpenStack Kilo on Ubuntu with Neutron DVR networking
/_static/scripts/envs/openstack_single_kilo_ubuntu_neutron.env.example
/_static/scripts/envs/openstack_single_kilo_ubuntu_neutron.env.example
OpenStack Kilo on CentOS/RHEL with OpenContrail networking
/_static/scripts/envs/openstack_single_kilo_redhat_opencontrail.env.example
/_static/scripts/envs/openstack_single_kilo_redhat_opencontrail.env.example
openstack_cluster.hot environment examples
OpenStack Kilo on Ubuntu with OpenContrail networking
/_static/scripts/envs/openstack_cluster_kilo_ubuntu_opencontrail.env.example
/_static/scripts/envs/openstack_cluster_kilo_ubuntu_opencontrail.env.example
OpenStack Kilo on CentOS/RHEL with OpenContrail networking
/_static/scripts/envs/openstack_cluster_kilo_ubuntu_opencontrail.env.example
/_static/scripts/envs/openstack_cluster_kilo_ubuntu_opencontrail.env.example
$ heat stack-create -e envs/ENV_FILE -f openstack_single.hot
$ heat stack-create -e envs/ENV_FILE -f openstack_cluster.hotIf everything goes right, stack should be ready in a few minutes. You can verify by running following commands:
$ heat stack-list
$ nova listYou should be also able to log in as root to public IP provided by
nova list command. When this cluster is deployed, you
canlog in to the instances through the Salt master node.
Current state of supported env configurations
| ENV configuration | Single Status | Cluster Status |
| OS Kilo Ubuntu with OpenContrail | Stable | Stable |
| OS Kilo RedHat single with OpenContrail | Experimental | Experimental |
| OS Kilo Ubuntu single with Neutron DVR | Experimental | NONE |
OpenStack Heat templates
OpenStack single Heat template
/_static/scripts/openstack_single.hot
/_static/scripts/openstack_single.hot
OpenStack cluster Heat template
/_static/scripts/openstack_cluster.hot
/_static/scripts/openstack_cluster.hot
Openstack-salt testing labs
You can use publicly available labs offered by technology partners.
Testing lab at tcp cloud
Company tcp cloud has provided 100 cores and 400 GB of RAM divided in 5 separate projects, each with quotas set to 20 cores and 80 GB of RAM. Each project is capable of running both single and cluster deployments.
- Endpoint URL:
| User | Project | Domain |
| openstack_salt_user01 | openstack_salt_lab01 | default |
| openstack_salt_user02 | openstack_salt_lab02 | default |
| openstack_salt_user03 | openstack_salt_lab03 | default |
| openstack_salt_user04 | openstack_salt_lab04 | default |
| openstack_salt_user05 | openstack_salt_lab05 | default |
To get the access credentials and full support, visit
#openstack-salt IRC channel.