Ceilometer |
Log in to a controller node CLI.
Restart the Ceilometer services:
# service ceilometer-agent-central restart
# service ceilometer-api restart
# service ceilometer-agent-notification restart
# service ceilometer-collector status restart
Verify the status of the Ceilometer services. See service-status .
Repeat step 1 - 3 on all controller nodes.
|
Cinder |
Log in to a controller node CLI.
Restart the Cinder services:
# service cinder-api restart
# service cinder-scheduler restart
Verify the status of the Cinder services. See service-status .
Repeat step 1 - 3 on all controller nodes.
On every node with Cinder role, run:
# service cinder-volume restart
# service cinder-backup restart
Verify the status of the cinder-volume and
cinder-backup services.
|
Corosync/Pacemaker |
Log in to a controller node CLI.
Restart the Corosync and Pacemaker services:
# service corosync restart
# service pacemaker restart
Verify the status of the Corosync and Pacemaker services. See
service-status .
Repeat step 1 - 3 on all controller nodes.
|
Glance |
Log in to a controller node CLI.
Restart the Glance services:
# service glance-api restart
# service glance-registry restart
Verify the status of the Glance services. See service-status .
Repeat step 1 - 3 on all controller nodes.
|
Horizon |
Since the Horizon service is available through the Apache server,
you should restart the Apache service on all controller nodes:
Log in to a controller node CLI.
Restart the Apache server:
# service apache2 restart
Verify whether the Apache service is successfully running after
restart:
# service apache2 status
Verify whether the Apache ports are opened and listening:
# netstat -nltp | egrep apache2
Repeat step 1 - 3 on all controller nodes.
|
Ironic |
Log in to a controller node CLI.
Restart the Ironic services:
# service ironic-api restart
# service ironic-conductor restart
Verify the status of the Ironic services. See service-status .
Repeat step 1 - 3 on all controller nodes.
On any controller node, run the following command for the
nova-compute service configured to work with Ironic:
# crm resource restart p_nova_compute_ironic
Verify the status of the p_nova_compute_ironic
service.
|
Keystone |
Since the Keystone service is available through the Apache server,
complete the following steps on all controller nodes:
Log in to a controller node CLI.
Restart the Apache server:
# service apache2 restart
Verify whether the Apache service is successfully running after
restart:
# service apache2 status
Verify whether the Apache ports are opened and listening:
# netstat -nltp | egrep apache2
Repeat step 1 - 3 on all controller nodes.
|
MySQL |
Log in to any controller node CLI.
Run the following command:
# pcs status | grep -A1 mysql
In the output, the resource clone_p_mysql should be in
the Started status.
Disable the clone_p_mysql resource:
# pcs resource disable clone_p_mysqld
Verify that the resource clone_p_mysqld is in the
Stopped status:
# pcs status | grep -A2 mysql
It may take some time for this resource to be stopped on all
controller nodes.
Disable the clone_p_mysql resource:
# pcs resource enable clone_p_mysqld
Verify that the resource clone_p_mysqld is in the
Started status again on all controller nodes:
# pcs status | grep -A2 mysql
Use the pcs
commands instead of crm for restarting the service. The pcs tool
correctly stops the service according to the quorum policy preventing
MySQL failures.
|
Neutron |
Use the following restart steps for the DHCP Neutron agent as an
example for all Neutron agents.
Log in to any controller node CLI.
Verify the DHCP agent status:
# pcs resource show | grep -A1 neutron-dhcp-agent
The output should contain the list of all controllers in the
Started status.
Stop the DHCP agent:
# pcs resource disable clone_neutron-dhcp-agent
Verify the Corosync status of the DHCP agent:
# pcs resource show | grep -A1 neutron-dhcp-agent
The output should contain the list of all controllers in the
Stopped status.
Verify the neutron-dhcp-agent status on the
OpenStack side:
# neutron agent-list
The output table should contain the DHCP agents for every controller
node with xxx in the alive column.
Start the DHCP agent on every controller node:
# pcs resource enable clone_neutron-dhcp-agent
Verify the DHCP agent status:
# pcs resource show | grep -A1 neutron-dhcp-agent
The output should contain the list of all controllers in the
Started status.
Verify the neutron-dhcp-agent status on the
OpenStack side:
# neutron agent-list
The output table should contain the DHCP agents for every controller
node with :-) in the alive column and
True in the admin_state_up column.
|
Nova |
Log in to a controller node CLI.
Restart the Nova services:
# service nova-api restart
# service nova-cert restart
# service nova-compute restart
# service nova-conductor restart
# service nova-consoleauth restart
# service nova-novncproxy restart
# service nova-scheduler restart
# service nova-spicehtml5proxy restart
# service nova-xenvncproxy restart
Verify the status of the Nova services. See service-status .
Repeat step 1 - 3 on all controller nodes.
On every compute node, run:
# service nova-compute restart
Verify the status of the nova-compute
service.
|
RabbitMQ |
Log in to any controller node CLI.
Disable the RabbitMQ service:
# pcs resource disable master_p_rabbitmq-server
Verify whether the service is stopped:
# pcs status | grep -A2 rabbitmq
Enable the service:
# pcs resource enable master_p_rabbitmq-server
During the startup process, the output of the pcs status command can
show all existing RabbitMQ services in the Slaves
mode.
Verify the service status:
# rabbitmqctl cluster_status
In the output, the running_nodes field should contain
all controllers’ host names in the rabbit@<HOSTNAME>
format. The partitions field should be empty.
|
Swift |
Log in to a controller node CLI.
Restart the Swift services:
# service swift-account-auditor restart
# service swift-account restart
# service swift-account-reaper restart
# service swift-account-replicator restart
# service swift-container-auditor restart
# service swift-container restart
# service swift-container-reconciler restart
# service swift-container-replicator restart
# service swift-container-sync restart
# service swift-container-updater restart
# service swift-object-auditor restart
# service swift-object restart
# service swift-object-reconstructor restart
# service swift-object-replicator restart
# service swift-object-updater restart
# service swift-proxy restart
Verify the status of the Swift services. See service-status .
Repeat step 1 - 3 on all controller nodes.
|