tripleo-heat-templates/deployment
John Eckersberg e80356e454 rabbitmq: Open ports 25673-25683 for CLI tools
Since RabbitMQ 3.7.4, the CLI tools (rabbitmqctl and friends)
parallelize the querying of information from cluster members.  In
order to receive stream data back, the cli instance binds and
registers itself on an available port (default between 35672 and
35682, inclusive).  If these ports are firewalled off, then
rabbitmqctl commands such as list_queues will hang waiting for data
from remote cluster members.

This patch does two things:

1) Reconfigures rabbitmqctl to bind to 25673-25683 instead of the
default range of 35672-35682.  This ensures the ports are not in the
ephemeral port range and avoids unintended collisions.

2) Opens the firewall on 25673-25683 to enable communication.

Resolves: rhbz#1811680

Change-Id: If5caa51cd9a3aef97d06d491dde1d5129cc1a569
(cherry picked from commit a2bc2e10b0)
(cherry picked from commit 40a1e5ba18)
2020-04-01 10:34:43 +00:00
..
aide Move Aide to deployments 2019-03-28 08:24:40 -04:00
aodh Remove stray conditional from aodh-evaluator tasks 2020-03-10 16:52:10 +00:00
apache Disable a directory listing of /icons in httpd. 2019-04-01 14:14:47 +09:00
barbican Fix default network in barbican deployment 2019-08-12 13:43:02 -04:00
cavium Move containers-common.yaml into deployment 2019-04-14 18:15:12 -04:00
ceilometer Stop services for unupgraded controllers 2019-07-16 15:24:04 +02:00
ceph-ansible Add swiftoperator role on ceph-rgw template 2020-03-11 11:44:14 +01:00
cinder HA: minor update of arbitrary container image name 2020-03-17 12:26:11 +01:00
clients Include python-panko client. 2019-03-18 14:49:28 -04:00
container-image-prepare Honor Debug for container image prepare 2019-11-01 13:09:01 -06:00
database HA: minor update of arbitrary container image name 2020-03-17 12:26:11 +01:00
deprecated [Stein-Only] Create a dedicated log file for healthchecks for collectd 2020-01-09 13:28:36 +01:00
ec2 Support TLS deployments with KernelDisableIPv6 enabled 2019-07-10 07:05:26 +00:00
etcd Set setype on etcd's service directory 2019-03-07 14:36:08 -05:00
experimental Add bind mount for config setup 2019-07-17 08:11:40 -04:00
glance Make sure glance_api_tls_proxy logs are persisted on the host 2019-11-29 13:53:23 +00:00
gnocchi Stop services for unupgraded controllers 2019-07-16 15:24:04 +02:00
haproxy HA: minor update of arbitrary container image name 2020-03-17 12:26:11 +01:00
heat Fix debug hiera keys 2019-11-07 23:16:29 +00:00
horizon Use /var/tmp on host to store temporal files for image upload via Horizon 2019-08-29 08:01:03 +09:00
image-serve We have to allow httpd to listen on those ports in some cases. 2019-03-21 08:57:59 +01:00
ironic Remove pre-upgrade best-effort online data migrations 2019-08-27 16:19:06 +02:00
iscsid Avoid checking rc if we're in dry-run mode 2019-06-28 05:21:48 +00:00
keepalived Fix keepalived logging on disk 2020-01-30 03:05:40 +00:00
kernel Use package instead yum. 2020-02-10 13:50:08 +01:00
keystone Update environment var for keystone bootstrap 2019-12-17 09:47:39 -07:00
logging Install and start Rsyslog on the Undercloud & Standalone 2019-11-04 22:43:07 +01:00
login-defs Move login-defs to deployment directory 2019-01-22 13:48:57 -07:00
logrotate Install tmpwatch on the overcloud 2019-12-19 13:32:31 +01:00
manila HA: minor update of arbitrary container image name 2020-03-17 12:26:11 +01:00
memcached Stop services for unupgraded controllers 2019-07-16 15:24:04 +02:00
messaging Move containers-common.yaml into deployment 2019-04-14 18:15:12 -04:00
metrics Mount /var/run rw 2019-07-23 09:31:51 +02:00
mistral Fix debug hiera keys 2019-11-07 23:16:29 +00:00
multipathd Move containers-common.yaml into deployment 2019-04-14 18:15:12 -04:00
neutron Add setfacl statements for neutron metadata proxy 2020-02-25 17:32:08 +00:00
nova [TRAIN and before] Introduce ContainerCpusetCpus 2020-03-14 18:57:54 -04:00
octavia Ensure /var/run/octavia is present upon reboot 2020-02-28 14:55:57 +00:00
ovn HA: minor update of arbitrary container image name 2020-03-17 12:26:11 +01:00
pacemaker clustercheck: use fqdn instead of ip for bind address 2020-01-20 13:56:46 +00:00
podman [Stein-only] add no_log to the podman login 2020-02-20 15:27:52 -07:00
qdr Move containers-common.yaml into deployment 2019-04-14 18:15:12 -04:00
rabbitmq rabbitmq: Open ports 25673-25683 for CLI tools 2020-04-01 10:34:43 +00:00
sahara Fix debug hiera keys 2019-11-07 23:16:29 +00:00
securetty Move securetty to deployment dir 2019-01-22 13:45:40 -07:00
selinux Move selinux to deployment folder 2019-01-22 13:21:13 -07:00
snmp Snmp - Use net_cidr_map for firewall rules 2019-01-06 18:21:54 +01:00
sshd Avoid dangling firewall rule for ssh access 2019-01-22 14:49:46 +01:00
swift Enable additional Swift healtchecks 2019-12-13 12:56:20 +00:00
tacker Move containers-common.yaml into deployment 2019-04-14 18:15:12 -04:00
time Revert "Add ntpdate to ensure hwclock sync" 2019-12-13 22:17:42 +00:00
timesync [stable only] Use service_name in heira for firewall rules 2020-02-18 07:32:24 +00:00
tripleo-firewall Merge "Add a daemon-reload to the tripleo-iptables services" into stable/stein 2019-08-16 01:40:13 +00:00
tripleo-packages Force facts cache refreshing after OS upgrade. 2020-02-20 12:58:30 +00:00
tuned flatten tuned service configuration 2018-12-14 13:13:40 -05:00
undercloud Put on hold the Ansible package update until is fixed 2019-06-17 12:19:22 +02:00
zaqar Resolve broken zaqar container caused by logging issues 2019-09-13 01:38:29 +00:00
containers-common.yaml HA: fix <service>_restart_bundle with minor update workflow 2019-09-06 13:45:11 +02:00
README.rst Update/combine docker/services/README.rst 2019-04-14 18:15:14 -04:00

TripleO Deployments

This directory contains files that represent individual service deployments, orchestration tools, and the configuration tools used to deploy them.

Directory Structure

Each logical grouping of services will have a directory. Example: 'timesync'. Within this directory related timesync services would exist to for example configure timesync services on baremetal or via containers.

Filenaming conventions

As a convention each deployments service filename will reflect both the deployment engine (baremetal, or containers) along with the config tool used to deploy that service.

The convention is <service-name>-<engine>-<config management tool>.

Examples:

deployment/aodh/aodh-api-container-puppet.yaml (containerized Aodh service configured with Puppet)

deployment/aodh/aodh-api-container-ansible.yaml (containerized Aodh service configured with Ansible)

deployment/timesync/chrony-baremetal-ansible.yaml (baremetal Chrony service configured with Ansible)

deployment/timesync/chrony-baremetal-puppet.yaml (baremetal Chrony service configured with Puppet)

Building Kolla Images

TripleO currently relies on Kolla(Dockerfile) containers. Kolla supports container customization and we are making use of this feature within TripleO to inject puppet (our configuration tool of choice) into the Kolla base images. A variety of other customizations are being made via the tripleo-common/container-images/tripleo_kolla_template_overrides.j2 file.

To build Kolla images for TripleO adjust your kolla config1 to build your centos base image with puppet using the example below:

$ cat template-overrides.j2 {% extends parent_template %} {% set base_centos_binary_packages_append = ['puppet'] %} {% set nova_scheduler_packages_append = ['openstack-tripleo-common'] %}

kolla-build --base centos --template-override template-overrides.j2

Containerized Deployment Template Structure

Each deployment template may define a set of output values control the underlying service deployment in a variety of ways. These output sections are specific to the TripleO deployment architecture. The following sections are available for containerized services.

  • config_settings: This section contains service specific hiera data can be used to generate config files for each service. This data is ultimately processed via the container-puppet.py tool which generates config files for each service according to the settings here.

  • kolla_config: Contains YAML that represents how to map config files into the kolla container. This config file is typically mapped into the container itself at the /var/lib/kolla/config_files/config.json location and drives how kolla's external config mechanisms work.

  • docker_config: Data that is passed to paunch tool to configure a container, or step of containers at each step. See the available steps documented below which are implemented by TripleO's cluster deployment architecture.

  • puppet_config: This section is a nested set of key value pairs that drive the creation of config files using puppet. Required parameters include:

    • puppet_tags: Puppet resource tag names that are used to generate config files with puppet. Only the named config resources are used to generate a config file. Any service that specifies tags will have the default tags of 'file,concat,file_line,augeas,cron' appended to the setting. Example: keystone_config
    • config_volume: The name of the volume (directory) where config files will be generated for this service. Use this as the location to bind mount into the running Kolla container for configuration.
    • config_image: The name of the container image that will be used for generating configuration files. This is often the same container that the runtime service uses. Some services share a common set of config files which are generated in a common base container.
    • step_config: This setting controls the manifest that is used to create container config files via puppet. The puppet tags below are used along with this manifest to generate a config directory for this container.
  • container_puppet_tasks: This section provides data to drive the container-puppet.py tool directly. The task is executed only once within the cluster (not on each node) and is useful for several puppet snippets we require for initialization of things like keystone endpoints, database users, etc. See container-puppet.py for formatting.

Deployment steps

Similar to baremetal containers are brought up in a stepwise manner. The current architecture supports bringing up baremetal services alongside of containers. For each step the baremetal puppet manifests are executed first and then any containers are brought up afterwards.

Steps correlate to the following:

Pre) Containers config files generated per hiera settings. 1) Load Balancer configuration baremetal a) step 1 baremetal b) step 1 containers 2) Core Services (Database/Rabbit/NTP/etc.) a) step 2 baremetal b) step 2 containers 3) Early Openstack Service setup (Ringbuilder, etc.) a) step 3 baremetal b) step 3 containers 4) General OpenStack Services a) step 4 baremetal b) step 4 containers c) Keystone containers post initialization (tenant,service,endpoint creation) 5) Service activation (Pacemaker), online data migration a) step 5 baremetal b) step 5 containers

Update steps:

All services have an associated update_tasks output that is an ansible snippet that will be run during update in an rolling update that is expected to run in a rolling update fashion (one node at a time)

For Controller (where pacemaker is running) we have the following states:
  1. Step=1: stop the cluster on the updated node;
  2. Step=2: Pull the latest image and retag the it pcmklatest
  3. Step=3: yum upgrade happens on the host.
  4. Step=4: Restart the cluster on the node
  5. Step=5: Verification: Currently we test that the pacemaker services are running.

Then the usual deploy steps are run which pull in the latest image for all containerized services and the updated configuration if any.

Note: as pacemaker is not containerized, the points 1 and 4 happen in puppet/services/pacemaker.yaml.

Fast-forward Upgrade Steps

Each service template may optionally define a fast_forward_upgrade_tasks key, which is a list of Ansible tasks to be performed during the fast-forward upgrade process. As with Upgrade steps each task is associated to a particular step provided as a variable and used along with a release variable by a basic conditional that determines when the task should run.

Steps are broken down into two categories, prep tasks executed across all hosts and bootstrap tasks executed on a single host for a given role.

The individual steps then correspond to the following tasks during the upgrade:

Prep steps:

  • Step=0: Check running services
  • Step=1: Stop the service
  • Step=2: Stop the cluster
  • Step=3: Update repos

Bootstrap steps:

  • Step=4: DB backups
  • Step=5: Pre package update commands
  • Step=6: Package updates
  • Step=7: Post package update commands
  • Step=8: DB syncs
  • Step=9: Verification

  1. See the override file which can be used to build Kolla packages that work with TripleO.↩︎