2018-10-17 15:14:32 -07:00
|
|
|
|
.. _controller-storage:
|
|
|
|
|
|
|
|
|
|
===================================================================
|
|
|
|
|
StarlingX/Installation Guide Virtual Environment/Controller Storage
|
|
|
|
|
===================================================================
|
|
|
|
|
|
|
|
|
|
-----------------
|
|
|
|
|
Preparing Servers
|
|
|
|
|
-----------------
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
**********
|
2018-10-17 15:14:32 -07:00
|
|
|
|
Bare Metal
|
|
|
|
|
**********
|
|
|
|
|
|
|
|
|
|
Required Servers:
|
|
|
|
|
|
|
|
|
|
- Controllers: 2
|
|
|
|
|
- Computes: 2 - 100
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
^^^^^^^^^^^^^^^^^^^^^
|
2018-10-17 15:14:32 -07:00
|
|
|
|
Hardware Requirements
|
|
|
|
|
^^^^^^^^^^^^^^^^^^^^^
|
|
|
|
|
|
|
|
|
|
The recommended minimum requirements for the physical servers where
|
|
|
|
|
StarlingX Controller Storage will be deployed, include:
|
|
|
|
|
|
|
|
|
|
- ‘Minimum’ Processor:
|
|
|
|
|
|
|
|
|
|
- Dual-CPU Intel® Xeon® E5 26xx Family (SandyBridge) 8 cores/socket
|
|
|
|
|
|
|
|
|
|
- Memory:
|
|
|
|
|
|
|
|
|
|
- 64 GB Controller
|
|
|
|
|
- 32 GB Compute
|
|
|
|
|
|
|
|
|
|
- BIOS:
|
|
|
|
|
|
|
|
|
|
- Hyper-Threading Tech: Enabled
|
|
|
|
|
- Virtualization Technology: Enabled
|
|
|
|
|
- VT for Directed I/O: Enabled
|
|
|
|
|
- CPU Power and Performance Policy: Performance
|
|
|
|
|
- CPU C State Control: Disabled
|
|
|
|
|
- Plug & Play BMC Detection: Disabled
|
|
|
|
|
|
|
|
|
|
- Primary Disk:
|
|
|
|
|
|
|
|
|
|
- 500 GB SDD or NVMe Controller
|
|
|
|
|
- 120 GB (min. 10K RPM) Compute
|
|
|
|
|
|
|
|
|
|
- Additional Disks:
|
|
|
|
|
|
|
|
|
|
- 1 or more 500 GB disks (min. 10K RPM) Compute
|
|
|
|
|
|
|
|
|
|
- Network Ports\*
|
|
|
|
|
|
|
|
|
|
- Management: 10GE Controller, Compute
|
|
|
|
|
- OAM: 10GE Controller
|
|
|
|
|
- Data: n x 10GE Compute
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
*******************
|
2018-10-17 15:14:32 -07:00
|
|
|
|
Virtual Environment
|
|
|
|
|
*******************
|
|
|
|
|
|
|
|
|
|
Run the libvirt qemu setup scripts. Setting up virtualized OAM and
|
|
|
|
|
Management networks:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
$ bash setup_network.sh
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Building xmls for definition of virtual servers:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
2018-10-22 07:28:21 -05:00
|
|
|
|
$ bash setup_controller_storage.sh -i <starlingx iso image>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
The xml server definitions that are created by the previous script are:
|
|
|
|
|
|
|
|
|
|
- controller-0
|
|
|
|
|
- controller-1
|
|
|
|
|
- compute-0
|
|
|
|
|
- compute-1
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
2018-10-22 07:28:21 -05:00
|
|
|
|
Powering Up a Virtual Server
|
|
|
|
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
|
|
|
|
|
|
|
|
|
To power up a virtual server, run the following command:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
$ sudo virsh start <server-xml-name>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
e.g.
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
$ sudo virsh start controller-0
|
2018-10-17 15:14:32 -07:00
|
|
|
|
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
2018-10-17 15:14:32 -07:00
|
|
|
|
Accessing Virtual Server Consoles
|
|
|
|
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
|
|
|
|
|
|
|
|
|
The xml for virtual servers in stx-tools repo, deployment/libvirt,
|
|
|
|
|
provides both graphical and text consoles.
|
|
|
|
|
|
|
|
|
|
Access the graphical console in virt-manager by right-click on the
|
|
|
|
|
domain (the server) and selecting "Open".
|
|
|
|
|
|
|
|
|
|
Access the textual console with the command "virsh console $DOMAIN",
|
|
|
|
|
where DOMAIN is the name of the server shown in virsh.
|
|
|
|
|
|
|
|
|
|
When booting the controller-0 for the first time, both the serial and
|
|
|
|
|
graphical consoles will present the initial configuration menu for the
|
|
|
|
|
cluster. One can select serial or graphical console for controller-0.
|
|
|
|
|
For the other nodes however only serial is used, regardless of which
|
|
|
|
|
option is selected.
|
|
|
|
|
|
|
|
|
|
Open the graphic console on all servers before powering them on to
|
|
|
|
|
observe the boot device selection and PXI boot progress. Run "virsh
|
|
|
|
|
console $DOMAIN" command promptly after power on to see the initial boot
|
|
|
|
|
sequence which follows the boot device selection. One has a few seconds
|
|
|
|
|
to do this.
|
|
|
|
|
|
|
|
|
|
------------------------------
|
|
|
|
|
Controller-0 Host Installation
|
|
|
|
|
------------------------------
|
|
|
|
|
|
|
|
|
|
Installing controller-0 involves initializing a host with software and
|
|
|
|
|
then applying a bootstrap configuration from the command line. The
|
|
|
|
|
configured bootstrapped host becomes Controller-0.
|
|
|
|
|
|
|
|
|
|
Procedure:
|
|
|
|
|
|
|
|
|
|
#. Power on the server that will be controller-0 with the StarlingX ISO
|
|
|
|
|
on a USB in a bootable USB slot.
|
|
|
|
|
#. Configure the controller using the config_controller script.
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
*************************
|
2018-10-17 15:14:32 -07:00
|
|
|
|
Initializing Controller-0
|
|
|
|
|
*************************
|
|
|
|
|
|
|
|
|
|
This section describes how to initialize StarlingX in host Controller-0.
|
|
|
|
|
Except where noted, all the commands must be executed from a console of
|
|
|
|
|
the host.
|
|
|
|
|
|
|
|
|
|
Power on the host to be configured as Controller-0, with the StarlingX
|
|
|
|
|
ISO on a USB in a bootable USB slot. Wait for the console to show the
|
|
|
|
|
StarlingX ISO booting options:
|
|
|
|
|
|
|
|
|
|
- **Standard Controller Configuration**
|
|
|
|
|
|
|
|
|
|
- When the installer is loaded and the installer welcome screen
|
|
|
|
|
appears in the Controller-0 host, select the type of installation
|
|
|
|
|
"Standard Controller Configuration".
|
|
|
|
|
|
|
|
|
|
- **Graphical Console**
|
|
|
|
|
|
|
|
|
|
- Select the "Graphical Console" as the console to use during
|
|
|
|
|
installation.
|
|
|
|
|
|
|
|
|
|
- **Standard Security Boot Profile**
|
|
|
|
|
|
|
|
|
|
- Select "Standard Security Boot Profile" as the Security Profile.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Monitor the initialization. When it is complete, a reboot is initiated
|
|
|
|
|
on the Controller-0 host, briefly displays a GNU GRUB screen, and then
|
|
|
|
|
boots automatically into the StarlingX image.
|
|
|
|
|
|
|
|
|
|
Log into Controller-0 as user wrsroot, with password wrsroot. The
|
|
|
|
|
first time you log in as wrsroot, you are required to change your
|
|
|
|
|
password. Enter the current password (wrsroot):
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
Changing password for wrsroot.
|
|
|
|
|
(current) UNIX Password:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Enter a new password for the wrsroot account:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
New password:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Enter the new password again to confirm it:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
Retype new password:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Controller-0 is initialized with StarlingX, and is ready for
|
|
|
|
|
configuration.
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
************************
|
2018-10-17 15:14:32 -07:00
|
|
|
|
Configuring Controller-0
|
|
|
|
|
************************
|
|
|
|
|
|
|
|
|
|
This section describes how to perform the Controller-0 configuration
|
|
|
|
|
interactively just to bootstrap system with minimum critical data.
|
|
|
|
|
Except where noted, all the commands must be executed from the console
|
|
|
|
|
of the active controller (here assumed to be controller-0).
|
|
|
|
|
|
|
|
|
|
When run interactively, the config_controller script presents a series
|
|
|
|
|
of prompts for initial configuration of StarlingX:
|
|
|
|
|
|
|
|
|
|
- For the Virtual Environment, you can accept all the default values
|
|
|
|
|
immediately after ‘system date and time’.
|
|
|
|
|
- For a Physical Deployment, answer the bootstrap configuration
|
|
|
|
|
questions with answers applicable to your particular physical setup.
|
|
|
|
|
|
|
|
|
|
The script is used to configure the first controller in the StarlingX
|
|
|
|
|
cluster as controller-0. The prompts are grouped by configuration
|
|
|
|
|
area. To start the script interactively, use the following command
|
|
|
|
|
with no parameters:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
controller-0:~$ sudo config_controller
|
|
|
|
|
System Configuration
|
|
|
|
|
================
|
|
|
|
|
Enter ! at any prompt to abort...
|
|
|
|
|
...
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Accept all the default values immediately after ‘system date and time’.
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
...
|
|
|
|
|
Applying configuration (this will take several minutes):
|
|
|
|
|
|
|
|
|
|
01/08: Creating bootstrap configuration ... DONE
|
|
|
|
|
02/08: Applying bootstrap manifest ... DONE
|
|
|
|
|
03/08: Persisting local configuration ... DONE
|
|
|
|
|
04/08: Populating initial system inventory ... DONE
|
|
|
|
|
05:08: Creating system configuration ... DONE
|
|
|
|
|
06:08: Applying controller manifest ... DONE
|
|
|
|
|
07:08: Finalize controller configuration ... DONE
|
|
|
|
|
08:08: Waiting for service activation ... DONE
|
|
|
|
|
|
|
|
|
|
Configuration was applied
|
|
|
|
|
|
|
|
|
|
Please complete any out of service commissioning steps with system commands and unlock controller to proceed.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
After config_controller bootstrap configuration, REST API, CLI and
|
|
|
|
|
Horizon interfaces are enabled on the controller-0 OAM IP Address. The
|
|
|
|
|
remaining installation instructions will use the CLI.
|
|
|
|
|
|
|
|
|
|
---------------------------------
|
|
|
|
|
Controller-0 and System Provision
|
|
|
|
|
---------------------------------
|
|
|
|
|
|
|
|
|
|
On Controller-0, acquire Keystone administrative privileges:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
controller-0:~$ source /etc/nova/openrc
|
|
|
|
|
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
*********************************************
|
2018-10-17 15:14:32 -07:00
|
|
|
|
Configuring Provider Networks at Installation
|
|
|
|
|
*********************************************
|
|
|
|
|
|
|
|
|
|
You must set up provider networks at installation so that you can attach
|
|
|
|
|
data interfaces and unlock the compute nodes.
|
|
|
|
|
|
|
|
|
|
Set up one provider network of the vlan type, named providernet-a:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ neutron providernet-create providernet-a --type=vlan
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ neutron providernet-range-create --name providernet-a-range1 --range 100-400 providernet-a
|
|
|
|
|
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
*************************************
|
2018-10-17 15:14:32 -07:00
|
|
|
|
Configuring Cinder on Controller Disk
|
|
|
|
|
*************************************
|
|
|
|
|
|
|
|
|
|
Review the available disk space and capacity and obtain the uuid of the
|
|
|
|
|
physical disk
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-disk-list controller-0
|
|
|
|
|
+--------------------------------------+-----------+---------+---------+-------+------------+--------------+...
|
|
|
|
|
| uuid | device_no | device_ | device_ | size_ | available_ | rpm |...
|
|
|
|
|
| | de | num | type | gib | gib | |...
|
|
|
|
|
+--------------------------------------+-----------+---------+---------+-------+------------+--------------+...
|
|
|
|
|
| 004f4c09-2f61-46c5-8def-99b2bdeed83c | /dev/sda | 2048 | HDD | 200.0 | 0.0 | |...
|
|
|
|
|
| 89694799-0dd8-4532-8636-c0d8aabfe215 | /dev/sdb | 2064 | HDD | 200.0 | 199.997 | |...
|
|
|
|
|
+--------------------------------------+-----------+---------+---------+-------+------------+--------------+...
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Create the 'cinder-volumes' local volume group
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-lvg-add controller-0 cinder-volumes
|
|
|
|
|
+-----------------+--------------------------------------+
|
|
|
|
|
| Property | Value |
|
|
|
|
|
+-----------------+--------------------------------------+
|
|
|
|
|
| lvm_vg_name | cinder-volumes |
|
|
|
|
|
| vg_state | adding |
|
|
|
|
|
| uuid | ece4c755-241c-4363-958e-85e9e3d12917 |
|
|
|
|
|
| ihost_uuid | 150284e2-fb60-4169-ae75-7f444b8ca9bf |
|
|
|
|
|
| lvm_vg_access | None |
|
|
|
|
|
| lvm_max_lv | 0 |
|
|
|
|
|
| lvm_cur_lv | 0 |
|
|
|
|
|
| lvm_max_pv | 0 |
|
|
|
|
|
| lvm_cur_pv | 0 |
|
|
|
|
|
| lvm_vg_size_gib | 0.00 |
|
|
|
|
|
| lvm_vg_total_pe | 0 |
|
|
|
|
|
| lvm_vg_free_pe | 0 |
|
|
|
|
|
| created_at | 2018-08-22T03:59:30.685718+00:00 |
|
|
|
|
|
| updated_at | None |
|
|
|
|
|
| parameters | {u'lvm_type': u'thin'} |
|
|
|
|
|
+-----------------+--------------------------------------+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Create a disk partition to add to the volume group
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-disk-partition-add controller-0 89694799-0dd8-4532-8636-c0d8aabfe215 199 -t lvm_phys_vol
|
|
|
|
|
+-------------+--------------------------------------------------+
|
|
|
|
|
| Property | Value |
|
|
|
|
|
+-------------+--------------------------------------------------+
|
|
|
|
|
| device_path | /dev/disk/by-path/pci-0000:00:03.0-ata-2.0-part1 |
|
|
|
|
|
| device_node | /dev/sdb1 |
|
|
|
|
|
| type_guid | ba5eba11-0000-1111-2222-000000000001 |
|
|
|
|
|
| type_name | None |
|
|
|
|
|
| start_mib | None |
|
|
|
|
|
| end_mib | None |
|
|
|
|
|
| size_mib | 203776 |
|
|
|
|
|
| uuid | 9ba2d76a-6ae2-4bfa-ad48-57b62d102e80 |
|
|
|
|
|
| ihost_uuid | 150284e2-fb60-4169-ae75-7f444b8ca9bf |
|
|
|
|
|
| idisk_uuid | 89694799-0dd8-4532-8636-c0d8aabfe215 |
|
|
|
|
|
| ipv_uuid | None |
|
|
|
|
|
| status | Creating |
|
|
|
|
|
| created_at | 2018-08-22T04:03:40.761221+00:00 |
|
|
|
|
|
| updated_at | None |
|
|
|
|
|
+-------------+--------------------------------------------------+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Wait for the new partition to be created (i.e. status=Ready)
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-disk-partition-list controller-0 --disk 89694799-0dd8-4532-8636-c0d8aabfe215
|
|
|
|
|
+--------------------------------------+...+------------+...+---------------------+----------+--------+
|
|
|
|
|
| uuid |...| device_nod |...| type_name | size_mib | status |
|
|
|
|
|
| |...| e |...| | | |
|
|
|
|
|
+--------------------------------------+...+------------+...+---------------------+----------+--------+
|
|
|
|
|
| 9ba2d76a-6ae2-4bfa-ad48-57b62d102e80 |...| /dev/sdb1 |...| LVM Physical Volume | 199.0 | Ready |
|
|
|
|
|
| |...| |...| | | |
|
|
|
|
|
| |...| |...| | | |
|
|
|
|
|
+--------------------------------------+...+------------+...+---------------------+----------+--------+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Add the partition to the volume group
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-pv-add controller-0 cinder-volumes 9ba2d76a-6ae2-4bfa-ad48-57b62d102e80
|
|
|
|
|
+--------------------------+--------------------------------------------------+
|
|
|
|
|
| Property | Value |
|
|
|
|
|
+--------------------------+--------------------------------------------------+
|
|
|
|
|
| uuid | 060dc47e-bc17-40f4-8f09-5326ef0e86a5 |
|
|
|
|
|
| pv_state | adding |
|
|
|
|
|
| pv_type | partition |
|
|
|
|
|
| disk_or_part_uuid | 9ba2d76a-6ae2-4bfa-ad48-57b62d102e80 |
|
|
|
|
|
| disk_or_part_device_node | /dev/sdb1 |
|
|
|
|
|
| disk_or_part_device_path | /dev/disk/by-path/pci-0000:00:03.0-ata-2.0-part1 |
|
|
|
|
|
| lvm_pv_name | /dev/sdb1 |
|
|
|
|
|
| lvm_vg_name | cinder-volumes |
|
|
|
|
|
| lvm_pv_uuid | None |
|
|
|
|
|
| lvm_pv_size_gib | 0.0 |
|
|
|
|
|
| lvm_pe_total | 0 |
|
|
|
|
|
| lvm_pe_alloced | 0 |
|
|
|
|
|
| ihost_uuid | 150284e2-fb60-4169-ae75-7f444b8ca9bf |
|
|
|
|
|
| created_at | 2018-08-22T04:06:54.008632+00:00 |
|
|
|
|
|
| updated_at | None |
|
|
|
|
|
+--------------------------+--------------------------------------------------+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Enable LVM Backend.
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system storage-backend-add lvm -s cinder --confirmed
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Wait for the storage backend to leave "configuring" state. Confirm LVM
|
|
|
|
|
Backend storage is configured:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system storage-backend-list
|
|
|
|
|
+--------------------------------------+------------+---------+------------+------+----------+...
|
|
|
|
|
| uuid | name | backend | state | task | services |...
|
|
|
|
|
+--------------------------------------+------------+---------+------------+------+----------+...
|
|
|
|
|
| 1daf3e5b-4122-459f-9dba-d2e92896e718 | file-store | file | configured | None | glance |...
|
|
|
|
|
| a4607355-be7e-4c5c-bf87-c71a0e2ad380 | lvm-store | lvm | configured | None | cinder |...
|
|
|
|
|
+--------------------------------------+------------+---------+------------+------+----------+...
|
|
|
|
|
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
**********************
|
2018-10-17 15:14:32 -07:00
|
|
|
|
Unlocking Controller-0
|
|
|
|
|
**********************
|
|
|
|
|
|
|
|
|
|
You must unlock controller-0 so that you can use it to install the
|
|
|
|
|
remaining hosts. On Controller-0, acquire Keystone administrative
|
|
|
|
|
privileges. Use the system host-unlock command:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-unlock controller-0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
The host is rebooted. During the reboot, the command line is
|
|
|
|
|
unavailable, and any ssh connections are dropped. To monitor the
|
|
|
|
|
progress of the reboot, use the controller-0 console.
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
****************************************
|
2018-10-17 15:14:32 -07:00
|
|
|
|
Verifying the Controller-0 Configuration
|
|
|
|
|
****************************************
|
|
|
|
|
|
|
|
|
|
On Controller-0, acquire Keystone administrative privileges:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
controller-0:~$ source /etc/nova/openrc
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Verify that the StarlingX controller services are running:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system service-list
|
|
|
|
|
+-----+-------------------------------+--------------+----------------+
|
|
|
|
|
| id | service_name | hostname | state |
|
|
|
|
|
+-----+-------------------------------+--------------+----------------+
|
|
|
|
|
...
|
|
|
|
|
| 1 | oam-ip | controller-0 | enabled-active |
|
|
|
|
|
| 2 | management-ip | controller-0 | enabled-active |
|
|
|
|
|
...
|
|
|
|
|
+-----+-------------------------------+--------------+----------------+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Verify that controller-0 is unlocked, enabled, and available:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-list
|
|
|
|
|
+----+--------------+-------------+----------------+-------------+--------------+
|
|
|
|
|
| id | hostname | personality | administrative | operational | availability |
|
|
|
|
|
+----+--------------+-------------+----------------+-------------+--------------+
|
|
|
|
|
| 1 | controller-0 | controller | unlocked | enabled | available |
|
|
|
|
|
+----+--------------+-------------+----------------+-------------+--------------+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
-----------------------------------------
|
|
|
|
|
Controller-1 / Compute Hosts Installation
|
|
|
|
|
-----------------------------------------
|
|
|
|
|
|
|
|
|
|
After initializing and configuring an active controller, you can add and
|
|
|
|
|
configure a backup controller and additional compute hosts. For each
|
|
|
|
|
host do the following:
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
*****************
|
2018-10-17 15:14:32 -07:00
|
|
|
|
Initializing Host
|
|
|
|
|
*****************
|
|
|
|
|
|
|
|
|
|
Power on Host. In host console you will see:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
Waiting for this node to be configured.
|
|
|
|
|
|
|
|
|
|
Please configure the personality for this node from the
|
|
|
|
|
controller node in order to proceed.
|
|
|
|
|
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
***************************************
|
2018-10-17 15:14:32 -07:00
|
|
|
|
Updating Host Host Name and Personality
|
|
|
|
|
***************************************
|
|
|
|
|
|
|
|
|
|
On Controller-0, acquire Keystone administrative privileges:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
controller-0:~$ source /etc/nova/openrc
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Wait for Controller-0 to discover new host, list the host until new
|
|
|
|
|
UNKNOWN host shows up in table:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-list
|
|
|
|
|
+----+--------------+-------------+----------------+-------------+--------------+
|
|
|
|
|
| id | hostname | personality | administrative | operational | availability |
|
|
|
|
|
+----+--------------+-------------+----------------+-------------+--------------+
|
|
|
|
|
| 1 | controller-0 | controller | unlocked | enabled | available |
|
|
|
|
|
| 2 | None | None | locked | disabled | offline |
|
|
|
|
|
+----+--------------+-------------+----------------+-------------+--------------+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Use the system host-update to update host personality attribute:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-update 2 personality=controller hostname=controller-1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Or for compute-0:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-update 3 personality=compute hostname=compute-0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
See also: 'system help host-update'
|
|
|
|
|
|
|
|
|
|
Unless it is known that the host's configuration can support the
|
|
|
|
|
installation of more than one node, it is recommended that the
|
|
|
|
|
installation and configuration of each node be serialized. For example,
|
|
|
|
|
if the entire cluster has its virtual disks hosted on the host's root
|
|
|
|
|
disk which happens to be a single rotational type hard disk, then the
|
|
|
|
|
host cannot (reliably) support parallel node installation.
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
***************
|
2018-10-17 15:14:32 -07:00
|
|
|
|
Monitoring Host
|
|
|
|
|
***************
|
|
|
|
|
|
|
|
|
|
On Controller-0, you can monitor the installation progress by running
|
|
|
|
|
the system host-show command for the host periodically. Progress is
|
|
|
|
|
shown in the install_state field.
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-show <host> | grep install
|
|
|
|
|
| install_output | text |
|
|
|
|
|
| install_state | booting |
|
|
|
|
|
| install_state_info | None |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Wait while the host is configured and rebooted. Up to 20 minutes may be
|
|
|
|
|
required for a reboot, depending on hardware. When the reboot is
|
|
|
|
|
complete, the host is reported as Locked, Disabled, and Online.
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
*************
|
2018-10-17 15:14:32 -07:00
|
|
|
|
Listing Hosts
|
|
|
|
|
*************
|
|
|
|
|
|
|
|
|
|
Once all Nodes have been installed, configured and rebooted, on
|
|
|
|
|
Controller-0 list the hosts:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-list
|
|
|
|
|
+----+--------------+-------------+----------------+-------------+--------------+
|
|
|
|
|
| id | hostname | personality | administrative | operational | availability |
|
|
|
|
|
+----+--------------+-------------+----------------+-------------+--------------+
|
|
|
|
|
| 1 | controller-0 | controller | unlocked | enabled | available |
|
|
|
|
|
| 2 | controller-1 | controller | locked | disabled | online |
|
|
|
|
|
| 3 | compute-0 | compute | locked | disabled | online |
|
|
|
|
|
| 4 | compute-1 | compute | locked | disabled | online |
|
|
|
|
|
+----+--------------+-------------+----------------+-------------+--------------+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
-------------------------
|
|
|
|
|
Controller-1 Provisioning
|
|
|
|
|
-------------------------
|
|
|
|
|
|
|
|
|
|
On Controller-0, list hosts
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-list
|
|
|
|
|
+----+--------------+-------------+----------------+-------------+--------------+
|
|
|
|
|
| id | hostname | personality | administrative | operational | availability |
|
|
|
|
|
+----+--------------+-------------+----------------+-------------+--------------+
|
|
|
|
|
...
|
|
|
|
|
| 2 | controller-1 | controller | locked | disabled | online |
|
|
|
|
|
...
|
|
|
|
|
+----+--------------+-------------+----------------+-------------+--------------+
|
|
|
|
|
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
***********************************************
|
2018-10-17 15:14:32 -07:00
|
|
|
|
Provisioning Network Interfaces on Controller-1
|
|
|
|
|
***********************************************
|
|
|
|
|
|
|
|
|
|
In order to list out hardware port names, types, pci-addresses that have
|
|
|
|
|
been discovered:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-port-list controller-1
|
|
|
|
|
|
|
|
|
|
|
2018-10-22 07:28:21 -05:00
|
|
|
|
Provision the oam interface for Controller-1:
|
2018-10-17 15:14:32 -07:00
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-if-modify -n <oam interface> -c platform --networks oam controller-1 <oam interface>
|
|
|
|
|
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
************************************
|
2018-10-17 15:14:32 -07:00
|
|
|
|
Provisioning Storage on Controller-1
|
|
|
|
|
************************************
|
|
|
|
|
|
|
|
|
|
Review the available disk space and capacity and obtain the uuid of the
|
|
|
|
|
physical disk
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-disk-list controller-1
|
|
|
|
|
+--------------------------------------+-----------+---------+---------+-------+------------+
|
|
|
|
|
| uuid | device_no | device_ | device_ | size_ | available_ |
|
|
|
|
|
| | de | num | type | gib | gib |
|
|
|
|
|
+--------------------------------------+-----------+---------+---------+-------+------------+
|
|
|
|
|
| f7ce53db-7843-457e-8422-3c8f9970b4f2 | /dev/sda | 2048 | HDD | 200.0 | 0.0 |
|
|
|
|
|
| 70b83394-968e-4f0d-8a99-7985cd282a21 | /dev/sdb | 2064 | HDD | 200.0 | 199.997 |
|
|
|
|
|
+--------------------------------------+-----------+---------+---------+-------+------------+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Assign Cinder storage to the physical disk
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-lvg-add controller-1 cinder-volumes
|
|
|
|
|
+-----------------+--------------------------------------+
|
|
|
|
|
| Property | Value |
|
|
|
|
|
+-----------------+--------------------------------------+
|
|
|
|
|
| lvm_vg_name | cinder-volumes |
|
|
|
|
|
| vg_state | adding |
|
|
|
|
|
| uuid | 22d8b94a-200a-4fd5-b1f5-7015ddf10d0b |
|
|
|
|
|
| ihost_uuid | 06827025-eacb-45e6-bb88-1a649f7404ec |
|
|
|
|
|
| lvm_vg_access | None |
|
|
|
|
|
| lvm_max_lv | 0 |
|
|
|
|
|
| lvm_cur_lv | 0 |
|
|
|
|
|
| lvm_max_pv | 0 |
|
|
|
|
|
| lvm_cur_pv | 0 |
|
|
|
|
|
| lvm_vg_size_gib | 0.00 |
|
|
|
|
|
| lvm_vg_total_pe | 0 |
|
|
|
|
|
| lvm_vg_free_pe | 0 |
|
|
|
|
|
| created_at | 2018-08-22T05:33:44.608913+00:00 |
|
|
|
|
|
| updated_at | None |
|
|
|
|
|
| parameters | {u'lvm_type': u'thin'} |
|
|
|
|
|
+-----------------+--------------------------------------+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Create a disk partition to add to the volume group based on uuid of the
|
|
|
|
|
physical disk
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-disk-partition-add controller-1 70b83394-968e-4f0d-8a99-7985cd282a21 199 -t lvm_phys_vol
|
|
|
|
|
+-------------+--------------------------------------------------+
|
|
|
|
|
| Property | Value |
|
|
|
|
|
+-------------+--------------------------------------------------+
|
|
|
|
|
| device_path | /dev/disk/by-path/pci-0000:00:03.0-ata-2.0-part1 |
|
|
|
|
|
| device_node | /dev/sdb1 |
|
|
|
|
|
| type_guid | ba5eba11-0000-1111-2222-000000000001 |
|
|
|
|
|
| type_name | None |
|
|
|
|
|
| start_mib | None |
|
|
|
|
|
| end_mib | None |
|
|
|
|
|
| size_mib | 203776 |
|
|
|
|
|
| uuid | 16a1c5cb-620c-47a3-be4b-022eafd122ee |
|
|
|
|
|
| ihost_uuid | 06827025-eacb-45e6-bb88-1a649f7404ec |
|
|
|
|
|
| idisk_uuid | 70b83394-968e-4f0d-8a99-7985cd282a21 |
|
|
|
|
|
| ipv_uuid | None |
|
|
|
|
|
| status | Creating (on unlock) |
|
|
|
|
|
| created_at | 2018-08-22T05:36:42.123770+00:00 |
|
|
|
|
|
| updated_at | None |
|
|
|
|
|
+-------------+--------------------------------------------------+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Wait for the new partition to be created (i.e. status=Ready)
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-disk-partition-list controller-1 --disk 70b83394-968e-4f0d-8a99-7985cd282a21
|
|
|
|
|
+--------------------------------------+...+------------+...+-------+--------+----------------------+
|
|
|
|
|
| uuid |...| device_nod | ... | size_g | status |
|
|
|
|
|
| |...| e | ... | ib | |
|
|
|
|
|
+--------------------------------------+...+------------+ ... +--------+----------------------+
|
|
|
|
|
| 16a1c5cb-620c-47a3-be4b-022eafd122ee |...| /dev/sdb1 | ... | 199.0 | Creating (on unlock) |
|
|
|
|
|
| |...| | ... | | |
|
|
|
|
|
| |...| | ... | | |
|
|
|
|
|
+--------------------------------------+...+------------+...+--------+----------------------+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Add the partition to the volume group
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-pv-add controller-1 cinder-volumes 16a1c5cb-620c-47a3-be4b-022eafd122ee
|
|
|
|
|
+--------------------------+--------------------------------------------------+
|
|
|
|
|
| Property | Value |
|
|
|
|
|
+--------------------------+--------------------------------------------------+
|
|
|
|
|
| uuid | 01d79ed2-717f-428e-b9bc-23894203b35b |
|
|
|
|
|
| pv_state | adding |
|
|
|
|
|
| pv_type | partition |
|
|
|
|
|
| disk_or_part_uuid | 16a1c5cb-620c-47a3-be4b-022eafd122ee |
|
|
|
|
|
| disk_or_part_device_node | /dev/sdb1 |
|
|
|
|
|
| disk_or_part_device_path | /dev/disk/by-path/pci-0000:00:03.0-ata-2.0-part1 |
|
|
|
|
|
| lvm_pv_name | /dev/sdb1 |
|
|
|
|
|
| lvm_vg_name | cinder-volumes |
|
|
|
|
|
| lvm_pv_uuid | None |
|
|
|
|
|
| lvm_pv_size_gib | 0.0 |
|
|
|
|
|
| lvm_pe_total | 0 |
|
|
|
|
|
| lvm_pe_alloced | 0 |
|
|
|
|
|
| ihost_uuid | 06827025-eacb-45e6-bb88-1a649f7404ec |
|
|
|
|
|
| created_at | 2018-08-22T05:44:34.715289+00:00 |
|
|
|
|
|
| updated_at | None |
|
|
|
|
|
+--------------------------+--------------------------------------------------+
|
|
|
|
|
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
**********************
|
2018-10-17 15:14:32 -07:00
|
|
|
|
Unlocking Controller-1
|
|
|
|
|
**********************
|
|
|
|
|
|
|
|
|
|
Unlock Controller-1
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-unlock controller-1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Wait while the Controller-1 is rebooted. Up to 10 minutes may be
|
|
|
|
|
required for a reboot, depending on hardware.
|
|
|
|
|
|
|
|
|
|
**REMARK:** Controller-1 will remain in 'degraded' state until
|
|
|
|
|
data-syncing is complete. The duration is dependant on the
|
|
|
|
|
virtualization host's configuration - i.e., the number and configuration
|
|
|
|
|
of physical disks used to host the nodes' virtual disks. Also, the
|
|
|
|
|
management network is expected to have link capacity of 10000 (1000 is
|
|
|
|
|
not supported due to excessive data-sync time). Use 'fm alarm-list' to
|
|
|
|
|
confirm status.
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-list
|
|
|
|
|
+----+--------------+-------------+----------------+-------------+--------------+
|
|
|
|
|
| id | hostname | personality | administrative | operational | availability |
|
|
|
|
|
+----+--------------+-------------+----------------+-------------+--------------+
|
|
|
|
|
| 1 | controller-0 | controller | unlocked | enabled | available |
|
|
|
|
|
| 2 | controller-1 | controller | unlocked | enabled | available |
|
|
|
|
|
...
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
----------------------
|
|
|
|
|
Compute Host Provision
|
|
|
|
|
----------------------
|
|
|
|
|
|
|
|
|
|
You must configure the network interfaces and the storage disks on a
|
|
|
|
|
host before you can unlock it. For each Compute Host do the following:
|
|
|
|
|
|
|
|
|
|
On Controller-0, acquire Keystone administrative privileges:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
controller-0:~$ source /etc/nova/openrc
|
|
|
|
|
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
*************************************************
|
2018-10-17 15:14:32 -07:00
|
|
|
|
Provisioning Network Interfaces on a Compute Host
|
|
|
|
|
*************************************************
|
|
|
|
|
|
|
|
|
|
On Controller-0, in order to list out hardware port names, types,
|
|
|
|
|
pci-addresses that have been discovered:
|
|
|
|
|
|
|
|
|
|
- **Only in Virtual Environment**: Ensure that the interface used is
|
|
|
|
|
one of those attached to host bridge with model type "virtio" (i.e.,
|
|
|
|
|
eth1000 and eth1001). The model type "e1000" emulated devices will
|
|
|
|
|
not work for provider networks:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-port-list compute-0
|
|
|
|
|
|
|
|
|
|
|
2018-10-22 07:28:21 -05:00
|
|
|
|
Provision the data interface for Compute:
|
2018-10-17 15:14:32 -07:00
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-if-modify -p providernet-a -c data compute-0 eth1000
|
|
|
|
|
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
***************************
|
2018-10-17 15:14:32 -07:00
|
|
|
|
VSwitch Virtual Environment
|
|
|
|
|
***************************
|
|
|
|
|
|
|
|
|
|
**Only in Virtual Environment**. If the compute has more than 4 cpus,
|
|
|
|
|
the system will auto-configure the vswitch to use 2 cores. However some
|
|
|
|
|
virtual environments do not properly support multi-queue required in a
|
|
|
|
|
multi-cpu environment. Therefore run the following command to reduce the
|
|
|
|
|
vswitch cores to 1:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-cpu-modify compute-0 -f vswitch -p0 1
|
|
|
|
|
+--------------------------------------+-------+-----------+-------+--------+...
|
|
|
|
|
| uuid | log_c | processor | phy_c | thread |...
|
|
|
|
|
| | ore | | ore | |...
|
|
|
|
|
+--------------------------------------+-------+-----------+-------+--------+...
|
|
|
|
|
| a3b5620c-28b1-4fe0-9e97-82950d8582c2 | 0 | 0 | 0 | 0 |...
|
|
|
|
|
| f2e91c2b-bfc5-4f2a-9434-bceb7e5722c3 | 1 | 0 | 1 | 0 |...
|
|
|
|
|
| 18a98743-fdc4-4c0c-990f-3c1cb2df8cb3 | 2 | 0 | 2 | 0 |...
|
|
|
|
|
| 690d25d2-4f99-4ba1-a9ba-0484eec21cc7 | 3 | 0 | 3 | 0 |...
|
|
|
|
|
+--------------------------------------+-------+-----------+-------+--------+...
|
|
|
|
|
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
**************************************
|
2018-10-17 15:14:32 -07:00
|
|
|
|
Provisioning Storage on a Compute Host
|
|
|
|
|
**************************************
|
|
|
|
|
|
|
|
|
|
Review the available disk space and capacity and obtain the uuid(s) of
|
|
|
|
|
the physical disk(s) to be used for nova local:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-disk-list compute-0
|
|
|
|
|
+--------------------------------------+-----------+---------+---------+-------+------------+...
|
|
|
|
|
| uuid | device_no | device_ | device_ | size_ | available_ |...
|
|
|
|
|
| | de | num | type | gib | gib |...
|
|
|
|
|
+--------------------------------------+-----------+---------+---------+-------+------------+...
|
|
|
|
|
| 8a9d2c09-d3a7-4781-bd06-f7abf603713a | /dev/sda | 2048 | HDD | 200.0 | 172.164 |...
|
|
|
|
|
| 5ad61bd1-795a-4a76-96ce-39433ef55ca5 | /dev/sdb | 2064 | HDD | 200.0 | 199.997 |...
|
|
|
|
|
+--------------------------------------+-----------+---------+---------+-------+------------+...
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Create the 'nova-local' local volume group:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-lvg-add compute-0 nova-local
|
|
|
|
|
+-----------------+-------------------------------------------------------------------+
|
|
|
|
|
| Property | Value |
|
|
|
|
|
+-----------------+-------------------------------------------------------------------+
|
|
|
|
|
| lvm_vg_name | nova-local |
|
|
|
|
|
| vg_state | adding |
|
|
|
|
|
| uuid | 18898640-c8b7-4bbd-a323-4bf3e35fee4d |
|
|
|
|
|
| ihost_uuid | da1cbe93-cec5-4f64-b211-b277e4860ab3 |
|
|
|
|
|
| lvm_vg_access | None |
|
|
|
|
|
| lvm_max_lv | 0 |
|
|
|
|
|
| lvm_cur_lv | 0 |
|
|
|
|
|
| lvm_max_pv | 0 |
|
|
|
|
|
| lvm_cur_pv | 0 |
|
|
|
|
|
| lvm_vg_size_gib | 0.00 |
|
|
|
|
|
| lvm_vg_total_pe | 0 |
|
|
|
|
|
| lvm_vg_free_pe | 0 |
|
|
|
|
|
| created_at | 2018-08-22T08:00:51.945160+00:00 |
|
|
|
|
|
| updated_at | None |
|
|
|
|
|
| parameters | {u'concurrent_disk_operations': 2, u'instance_backing': u'image'} |
|
|
|
|
|
+-----------------+-------------------------------------------------------------------+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Create a disk partition to add to the volume group based on uuid of the
|
|
|
|
|
physical disk:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-pv-add compute-0 nova-local 5ad61bd1-795a-4a76-96ce-39433ef55ca5
|
|
|
|
|
+--------------------------+--------------------------------------------+
|
|
|
|
|
| Property | Value |
|
|
|
|
|
+--------------------------+--------------------------------------------+
|
|
|
|
|
| uuid | 4c81745b-286a-4850-ba10-305e19cee78c |
|
|
|
|
|
| pv_state | adding |
|
|
|
|
|
| pv_type | disk |
|
|
|
|
|
| disk_or_part_uuid | 5ad61bd1-795a-4a76-96ce-39433ef55ca5 |
|
|
|
|
|
| disk_or_part_device_node | /dev/sdb |
|
|
|
|
|
| disk_or_part_device_path | /dev/disk/by-path/pci-0000:00:03.0-ata-2.0 |
|
|
|
|
|
| lvm_pv_name | /dev/sdb |
|
|
|
|
|
| lvm_vg_name | nova-local |
|
|
|
|
|
| lvm_pv_uuid | None |
|
|
|
|
|
| lvm_pv_size_gib | 0.0 |
|
|
|
|
|
| lvm_pe_total | 0 |
|
|
|
|
|
| lvm_pe_alloced | 0 |
|
|
|
|
|
| ihost_uuid | da1cbe93-cec5-4f64-b211-b277e4860ab3 |
|
|
|
|
|
| created_at | 2018-08-22T08:07:14.205690+00:00 |
|
|
|
|
|
| updated_at | None |
|
|
|
|
|
+--------------------------+--------------------------------------------+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Specify the local storage space as local copy-on-write image volumes in
|
|
|
|
|
nova-local:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-lvg-modify -b image -s 10240 compute-0 nova-local
|
|
|
|
|
+-----------------+-------------------------------------------------------------------+
|
|
|
|
|
| Property | Value |
|
|
|
|
|
+-----------------+-------------------------------------------------------------------+
|
|
|
|
|
| lvm_vg_name | nova-local |
|
|
|
|
|
| vg_state | adding |
|
|
|
|
|
| uuid | 18898640-c8b7-4bbd-a323-4bf3e35fee4d |
|
|
|
|
|
| ihost_uuid | da1cbe93-cec5-4f64-b211-b277e4860ab3 |
|
|
|
|
|
| lvm_vg_access | None |
|
|
|
|
|
| lvm_max_lv | 0 |
|
|
|
|
|
| lvm_cur_lv | 0 |
|
|
|
|
|
| lvm_max_pv | 0 |
|
|
|
|
|
| lvm_cur_pv | 0 |
|
|
|
|
|
| lvm_vg_size_gib | 0.00 |
|
|
|
|
|
| lvm_vg_total_pe | 0 |
|
|
|
|
|
| lvm_vg_free_pe | 0 |
|
|
|
|
|
| created_at | 2018-08-22T08:00:51.945160+00:00 |
|
|
|
|
|
| updated_at | None |
|
|
|
|
|
| parameters | {u'concurrent_disk_operations': 2, u'instance_backing': u'image'} |
|
|
|
|
|
+-----------------+-------------------------------------------------------------------+
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
************************
|
2018-10-17 15:14:32 -07:00
|
|
|
|
Unlocking a Compute Host
|
|
|
|
|
************************
|
|
|
|
|
|
|
|
|
|
On Controller-0, use the system host-unlock command to unlock the
|
|
|
|
|
Compute node:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-unlock compute-0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Wait while the Compute node is rebooted. Up to 10 minutes may be
|
|
|
|
|
required for a reboot, depending on hardware. The host is rebooted, and
|
|
|
|
|
its Availability State is reported as In-Test, followed by
|
|
|
|
|
unlocked/enabled.
|
|
|
|
|
|
|
|
|
|
-------------------
|
|
|
|
|
System Health Check
|
|
|
|
|
-------------------
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
***********************
|
2018-10-17 15:14:32 -07:00
|
|
|
|
Listing StarlingX Nodes
|
|
|
|
|
***********************
|
|
|
|
|
|
|
|
|
|
On Controller-0, after a few minutes, all nodes shall be reported as
|
|
|
|
|
Unlocked, Enabled, and Available:
|
|
|
|
|
|
|
|
|
|
::
|
|
|
|
|
|
|
|
|
|
[wrsroot@controller-0 ~(keystone_admin)]$ system host-list
|
|
|
|
|
+----+--------------+-------------+----------------+-------------+--------------+
|
|
|
|
|
| id | hostname | personality | administrative | operational | availability |
|
|
|
|
|
+----+--------------+-------------+----------------+-------------+--------------+
|
|
|
|
|
| 1 | controller-0 | controller | unlocked | enabled | available |
|
|
|
|
|
| 2 | controller-1 | controller | unlocked | enabled | available |
|
|
|
|
|
| 3 | compute-0 | compute | unlocked | enabled | available |
|
|
|
|
|
| 4 | compute-1 | compute | unlocked | enabled | available |
|
|
|
|
|
+----+--------------+-------------+----------------+-------------+--------------+
|
|
|
|
|
|
2018-11-12 15:37:24 -08:00
|
|
|
|
*****************
|
2018-10-17 15:14:32 -07:00
|
|
|
|
System Alarm List
|
|
|
|
|
*****************
|
|
|
|
|
|
|
|
|
|
When all nodes are Unlocked, Enabled and Available: check 'fm alarm-list' for issues.
|
|
|
|
|
|
|
|
|
|
Your StarlingX deployment is now up and running with 2x HA Controllers with Cinder
|
|
|
|
|
Storage, 2x Computes and all OpenStack services up and running. You can now proceed
|
|
|
|
|
with standard OpenStack APIs, CLIs and/or Horizon to load Glance Images, configure
|
|
|
|
|
Nova Flavors, configure Neutron networks and launch Nova Virtual Machines.
|