Fixed \_ as the output was not rendering correctly (pick r5 updates only)

Fixed Patchset 4 comments
Fixed Patchset 3 comments and added additional updates
Signed-off-by: Juanita-Balaraj <juanita.balaraj@windriver.com>
Change-Id: I7482afc3a90bbdc94b6ecd8b6ac39d831b8a45db
Signed-off-by: Juanita-Balaraj <juanita.balaraj@windriver.com>
(cherry picked from commit 265d96bed1)
This commit is contained in:
Juanita-Balaraj
2021-10-26 17:31:52 -04:00
committed by Ron Stone
parent 803d73eeef
commit 10f84f422e
20 changed files with 116 additions and 115 deletions

View File

@@ -51,7 +51,7 @@ commands to manage containerized applications provided as part of |prod|.
where:
**<app\_name>**
**<app_name>**
The name of the application to show details.
For example:
@@ -83,7 +83,7 @@ commands to manage containerized applications provided as part of |prod|.
where the following are optional arguments:
**<app\_name>**
**<app_name>**
Assigns a custom name for application. You can use this name to
interact with the application in the future.
@@ -92,7 +92,7 @@ commands to manage containerized applications provided as part of |prod|.
and the following is a positional argument:
**<tar\_file>**
**<tar_file>**
The path to the tar file containing the application to be uploaded.
For example:
@@ -126,7 +126,7 @@ commands to manage containerized applications provided as part of |prod|.
where the following is a positional argument:
**<app\_name>**
**<app_name>**
The name of the application.
and the following is an optional argument:
@@ -175,7 +175,7 @@ commands to manage containerized applications provided as part of |prod|.
+---------------------+--------------------------------+---------------+
- To show the overrides for a particular chart, use the following command.
System overrides are displayed in the **system\_overrides** section of
System overrides are displayed in the **system_overrides** section of
the **Property** column.
.. code-block:: none
@@ -185,10 +185,10 @@ commands to manage containerized applications provided as part of |prod|.
where the following are positional arguments:
**<app\_name>**
**<app_name>**
The name of the application.
**< chart\_name>**
**< chart_name>**
The name of the chart.
**<namespace>**
@@ -212,10 +212,10 @@ commands to manage containerized applications provided as part of |prod|.
where the following are positional arguments:
**<app\_name>**
**<app_name>**
The name of the application.
**<chart\_name>**
**<chart_name>**
The name of the chart.
**<namespace>**
@@ -257,7 +257,7 @@ commands to manage containerized applications provided as part of |prod|.
| | DEBUG: true |
+----------------+-------------------+
The user overrides are shown in the **user\_overrides** section of the
The user overrides are shown in the **user_overrides** section of the
**Property** column.
.. note::
@@ -280,10 +280,10 @@ commands to manage containerized applications provided as part of |prod|.
and the following are positional arguments:
**<app\_name>**
**<app_name>**
The name of the application.
**<chart\_name>**
**<chart_name>**
The name of the chart.
**<namespace>**
@@ -302,10 +302,10 @@ commands to manage containerized applications provided as part of |prod|.
where the following are positional arguments:
**<app\_name>**
**<app_name>**
The name of the application.
**<chart\_name>**
**<chart_name>**
The name of the chart.
**<namespace>**
@@ -334,7 +334,7 @@ commands to manage containerized applications provided as part of |prod|.
and the following is a positional argument:
**<app\_name>**
**<app_name>**
The name of the application to apply.
For example:
@@ -366,7 +366,7 @@ commands to manage containerized applications provided as part of |prod|.
where:
**<app\_name>**
**<app_name>**
The name of the application to abort.
For example:
@@ -389,7 +389,7 @@ commands to manage containerized applications provided as part of |prod|.
where the following are optional arguments:
**<app\_name>**
**<app_name>**
The name of the application to update.
You can look up the name of an application using the
@@ -417,7 +417,7 @@ commands to manage containerized applications provided as part of |prod|.
and the following is a positional argument which must come last:
**<tar\_file>**
**<tar_file>**
The tar file containing the application manifest, Helm charts and
configuration file.
@@ -431,7 +431,7 @@ commands to manage containerized applications provided as part of |prod|.
where:
**<app\_name>**
**<app_name>**
The name of the application to remove.
For example:
@@ -466,7 +466,7 @@ commands to manage containerized applications provided as part of |prod|.
where:
**<app\_name>**
**<app_name>**
The name of the application to delete.
You must run :command:`application-remove` before deleting an application.

View File

@@ -52,7 +52,7 @@ example, the user **testuser** is correct in the following URL, while
cloud/systemController, access to the local registry can only be done using
registry.local:9001. registry.central:9001 will be inaccessible. Installing
a |CA|-signed certificate for the registry and the certificate of the |CA| as
an 'ssl\_ca' certificate will remove this restriction.
an 'ssl_ca' certificate will remove this restriction.
For more information about Docker commands, see
`https://docs.docker.com/engine/reference/commandline/docker/ <https://docs.docker.com/engine/reference/commandline/docker/>`__.

View File

@@ -132,7 +132,7 @@ conditions are in place:
#. Install network connectivity required for the subcloud.
#. Ensure that the backup file are available on the controller. Run both
Ansible Restore playbooks, restore\_platform.yml and restore\_user\_images.yml.
Ansible Restore playbooks, restore_platform.yml and restore_user_images.yml.
For more information on restoring the back up file, see :ref:`Run Restore
Playbook Locally on the Controller
<running-restore-playbook-locally-on-the-controller>`, and :ref:`Run
@@ -143,7 +143,7 @@ conditions are in place:
The backup files contain the system data and updates.
#. If the backup file contains patches, Ansible Restore playbook
restore\_platform.yml will apply the patches and prompt you to reboot the
restore_platform.yml will apply the patches and prompt you to reboot the
system, you will need to re-run Ansible Restore playbook.
The current software version on the controller is compared against the
@@ -166,7 +166,7 @@ conditions are in place:
Rerun the Ansible Playbook if there were patches applied and you were
prompted to reboot the system.
#. Restore the local registry using the file restore\_user\_images.yml.
#. Restore the local registry using the file restore_user_images.yml.
This must be done before unlocking controller-0.
@@ -180,7 +180,7 @@ conditions are in place:
becomes operational.
#. If the system is a Distributed Cloud system controller, restore the **dc-vault**
using the restore\_dc\_vault.yml playbook. Perform this step after unlocking
using the restore_dc_vault.yml playbook. Perform this step after unlocking
controller-0:
.. code-block:: none
@@ -265,7 +265,7 @@ conditions are in place:
| 6 | compute-1 | worker | locked |disabled |offline |
+----+-------------+------------+---------------+-----------+------------+
#. Restore storage configuration. If :command:`wipe\_ceph\_osds` is set to
#. Restore storage configuration. If :command:`wipe_ceph_osds` is set to
**True**, follow the same procedure used to restore **controller-1**,
beginning with host **storage-0** and proceeding in sequence.
@@ -279,12 +279,12 @@ conditions are in place:
the restore procedure without interruption.
Standard with Controller Storage install or reinstall depends on the
:command:`wipe\_ceph\_osds` configuration:
:command:`wipe_ceph_osds` configuration:
#. If :command:`wipe\_ceph\_osds` is set to **true**, reinstall the
#. If :command:`wipe_ceph_osds` is set to **true**, reinstall the
storage hosts.
#. If :command:`wipe\_ceph\_osds` is set to **false** \(default
#. If :command:`wipe_ceph_osds` is set to **false** \(default
option\), do not reinstall the storage hosts.
.. caution::
@@ -319,9 +319,9 @@ conditions are in place:
.. caution::
Do not proceed until the Ceph cluster is healthy and the message
HEALTH\_OK appears.
HEALTH_OK appears.
If the message HEALTH\_WARN appears, wait a few minutes and then try
If the message HEALTH_WARN appears, wait a few minutes and then try
again. If the warning condition persists, consult the public
documentation for troubleshooting Ceph monitors \(for example,
`http://docs.ceph.com/docs/master/rados/troubleshooting/troubleshootin

View File

@@ -15,7 +15,7 @@ Use the following command to run the Ansible Backup playbook and back up the
~(keystone_admin)]$ ansible-playbook /usr/share/ansible/stx-ansible/playbooks/backup.yml -e "ansible_become_pass=<sysadmin password> admin_password=<sysadmin password>" -e "backup_user_local_registry=true"
The <admin\_password> and <ansible\_become\_pass\> need to be set correctly
The <admin_password> and <ansible_become_pass> need to be set correctly
using the ``-e`` option on the command line, or an override file, or in the
Ansible secret file.
@@ -23,44 +23,44 @@ The output files will be named:
.. _running-ansible-backup-playbook-locally-on-the-controller-ul-wj1-vxh-pmb:
- inventory\_hostname\_platform\_backup\_timestamp.tgz
- inventory_hostname_platform_backup_timestamp.tgz
- inventory\_hostname\_openstack\_backup\_timestamp.tgz
- inventory_hostname_openstack_backup_timestamp.tgz
- inventory\_hostname\_docker\_local\_registry\_backup\_timestamp.tgz
- inventory_hostname_docker_local_registry_backup_timestamp.tgz
- inventory\_hostname\_dc\_vault\_backup\_timestamp.tgz
- inventory_hostname_dc_vault_backup_timestamp.tgz
The variables prefix can be overridden using the ``-e`` option on the command
line or by using an override file.
.. _running-ansible-backup-playbook-locally-on-the-controller-ul-rdp-gyh-pmb:
- platform\_backup\_filename\_prefix
- platform_backup_filename_prefix
- openstack\_backup\_filename\_prefix
- openstack_backup_filename_prefix
- docker\_local\_registry\_backup\_filename\_prefix
- docker_local_registry_backup_filename_prefix
- dc\_vault\_backup\_filename\_prefix
- dc_vault_backup_filename_prefix
The generated backup tar files will be displayed in the following format,
for example:
.. _running-ansible-backup-playbook-locally-on-the-controller-ul-p3b-f13-pmb:
- localhost\_docker\_local\_registry\_backup\_2020\_07\_15\_21\_24\_22.tgz
- localhost_docker_local_registry_backup_2020_07_15_21_24_22.tgz
- localhost\_platform\_backup\_2020\_07\_15\_21\_24\_22.tgz
- localhost_platform_backup_2020_07_15_21_24_22.tgz
- localhost\_openstack\_backup\_2020\_07\_15\_21\_24\_22.tgz
- localhost_openstack_backup_2020_07_15_21_24_22.tgz
- localhost\_dc\_vault\_backup\_2020\_07\_15\_21\_24\_22.tgz
- localhost_dc_vault_backup_2020_07_15_21_24_22.tgz
These files are located by default in the /opt/backups directory on
controller-0, and contains the complete system backup.
If the default location needs to be modified, the variable backup\_dir can
If the default location needs to be modified, the variable backup_dir can
be overridden using the ``-e`` option on the command line or by using an
override file.

View File

@@ -73,24 +73,24 @@ In this method you can run Ansible Restore playbook and point to controller-0.
expects both the **initial_backup_dir** and **backup_filename**
to be specified.
- The backup\_filename is the platform backup tar file. It must be
- The backup_filename is the platform backup tar file. It must be
provided using the ``-e`` option on the command line, for example:
.. code-block:: none
-e backup\_filename= localhost_platform_backup_2019_07_15_14_46_37.tgz
- The initial\_backup\_dir is the location on the Ansible control
- The initial_backup_dir is the location on the Ansible control
machine where the platform backup tar file is placed to restore the
platform. It must be provided using ``-e`` option on the command line.
- The :command:`admin\_password`, :command:`ansible\_become\_pass`,
and :command:`ansible\_ssh\_pass` need to be set correctly using
- The :command:`admin_password`, :command:`ansible_become_pass`,
and :command:`ansible_ssh_pass` need to be set correctly using
the ``-e`` option on the command line or in the Ansible secret file.
:command:`ansible\_ssh\_pass` is the password to the sysadmin user
:command:`ansible_ssh_pass` is the password to the sysadmin user
on controller-0.
- The :command:`ansible\_remote\_tmp` should be set to a new
- The :command:`ansible_remote_tmp` should be set to a new
directory \(not required to create it ahead of time\) under
/home/sysadmin on controller-0 using the ``-e`` option on the command
line.
@@ -106,7 +106,7 @@ In this method you can run Ansible Restore playbook and point to controller-0.
the patches and prompt you to reboot the system. Then you will need to
re-run Ansible Restore playbook.
#. After running the restore\_platform.yml playbook, you can restore the local
#. After running the restore_platform.yml playbook, you can restore the local
registry images.
.. note::
@@ -119,33 +119,33 @@ In this method you can run Ansible Restore playbook and point to controller-0.
where optional-extra-vars can be:
- The backup\_filename is the local registry backup tar file. It
- The backup_filename is the local registry backup tar file. It
must be provided using the ``-e`` option on the command line, for
example:
.. code-block:: none
-e backup\_filename= localhost_docker_local_registry_backup_2020_07_15_21_24_22.tgz
-e backup_filename= localhost_docker_local_registry_backup_2020_07_15_21_24_22.tgz
- The initial\_backup\_dir is the location on the Ansible control
- The initial_backup_dir is the location on the Ansible control
machine where the platform backup tar file is located. It must be
provided using ``-e`` option on the command line.
- The :command:`ansible\_become\_pass`, and
:command:`ansible\_ssh\_pass` need to be set correctly using the
- The :command:`ansible_become_pass`, and
:command:`ansible_ssh_pass` need to be set correctly using the
``-e`` option on the command line or in the Ansible secret file.
:command:`ansible\_ssh\_pass` is the password to the sysadmin user
:command:`ansible_ssh_pass` is the password to the sysadmin user
on controller-0.
- The backup\_dir should be set to a directory on controller-0.
- The backup_dir should be set to a directory on controller-0.
The directory must have sufficient space for local registry backup
to be copied. The backup\_dir is set using the ``-e`` option on the
to be copied. The backup_dir is set using the ``-e`` option on the
command line.
- The :command:`ansible\_remote\_tmp` should be set to a new
- The :command:`ansible_remote_tmp` should be set to a new
directory on controller-0. Ansible will use this directory to copy
files, and the directory must have sufficient space for local
registry backup to be copied. The :command:`ansible\_remote\_tmp`
registry backup to be copied. The :command:`ansible_remote_tmp`
is set using the ``-e`` option on the command line.
For example, run the local registry restore playbook, where

View File

@@ -33,13 +33,13 @@ your server is isolated from the public Internet.
username: <your_my-registry.io_username>
password: <your_my-registry.io_password>
Where ``<your\_my-registry.io\_username>`` and
``<your\_my-registry.io\_password>`` are your login credentials for the
Where ``<your_my-registry.io_username>`` and
``<your_my-registry.io_password>`` are your login credentials for the
``<my-registry.io>`` private Docker registry.
.. note::
``<my-registry.io>`` must be a DNS name resolvable by the dns servers
configured in the ``dns\_servers:`` structure of the ansible bootstrap
configured in the ``dns_servers:`` structure of the ansible bootstrap
override file /home/sysadmin/localhost.yml.
#. For any additional local registry images required, use the full image name

View File

@@ -138,13 +138,13 @@ controller for access by subclouds. For example:
**--subcloud-apply-type**
Determines whether the subclouds are upgraded in parallel, or serially. If
this is not specified using the CLI, the values for subcloud\_update\_type
this is not specified using the CLI, the values for subcloud_update_type
defined for each subcloud group will be used by default.
**--max-parallel-subclouds**
Sets the maximum number of subclouds that can be upgraded in parallel
\(default 20\). If this is not specified using the CLI, the values for
max\_parallel\_subclouds defined for each subcloud group will be used by
max_parallel_subclouds defined for each subcloud group will be used by
default.
**--stop-on-failure**

View File

@@ -35,8 +35,8 @@ You must be in **SystemController** mode. To change the mode, see
tab.
.. image:: figures/vhy1525122582274.png
#. On the Cloud Patching Orchestration tab, click **Create Strategy**.
@@ -49,7 +49,7 @@ You must be in **SystemController** mode. To change the mode, see
parallel or serially.
If this is not specified using the |CLI|, the values for
:command:`subcloud\_update\_type` defined for each subcloud group will
:command:`subcloud_update_type` defined for each subcloud group will
be used by default.
**max-parallel-subclouds**
@@ -57,7 +57,7 @@ You must be in **SystemController** mode. To change the mode, see
\(default 20\).
If this is not specified using the |CLI|, the values for
:command:`max\_parallel\_subclouds` defined for each subcloud group
:command:`max_parallel_subclouds` defined for each subcloud group
will be used by default.
**stop-on-failure**
@@ -82,7 +82,7 @@ You must be in **SystemController** mode. To change the mode, see
To change the update strategy settings, you must delete the update
strategy and create a new one.
.. seealso::
.. seealso::
:ref:`Customizing the Update Configuration for Distributed Cloud Update
Orchestration <customizing-the-update-configuration-for-distributed-cloud-update-orchestration>`

View File

@@ -54,7 +54,7 @@ device image updates, including |FPGA| updates.
.. code-block:: none
~(keystone_admin)]$ system host-device-list <hostname_or_id>
~(keystone_admin)]$ system host-device-list <hostname_or_id>
To list devices from the central cloud, run:
@@ -143,7 +143,7 @@ device image updates, including |FPGA| updates.
parallel, or serially.
If this is not specified using the |CLI|, the values for
:command:`subcloud\_update\_type` defined for each subcloud group will
:command:`subcloud_update_type` defined for each subcloud group will
be used by default.
**max-parallel-subclouds**
@@ -151,7 +151,7 @@ device image updates, including |FPGA| updates.
\(default 20\).
If this is not specified using the |CLI|, the values for
:command:`max\_parallel\_subclouds` defined for each subcloud group
:command:`max_parallel_subclouds` defined for each subcloud group
will be used by default.
**stop-on-failure**
@@ -198,7 +198,7 @@ device image updates, including |FPGA| updates.
| created_at | 2020-08-11T18:13:40.576659 |
| updated_at | 2020-08-11T18:13:56.525459 |
+------------------------+----------------------------+
#. Monitor progress as the strategy is applied.
@@ -217,7 +217,7 @@ device image updates, including |FPGA| updates.
+-----------+-------+----------+------------------------------+----------------------------+----------------------------+
| subcloud3 | 2 | applying | apply phase is 18% complete | 2020-08-13 14:12:13.457588 | None |
+-----------+-------+----------+------------------------------+----------------------------+----------------------------+
- To monitor the step currently being performed on a specific subcloud, do the following:

View File

@@ -12,7 +12,7 @@ connected to the Central Cloud over L3 networks.
The Central Cloud has two regions: RegionOne, used to manage the nodes in the
Central Cloud, and System Controller, used to manage the subclouds in the
|prod-dc| system. You can select RegionOne or System Controller regions from the
Horizon Web interface or by setting the <OS\_REGION\_NAME> environment variable
Horizon Web interface or by setting the <OS_REGION_NAME> environment variable
if using the CLI.
**Central Cloud**

View File

@@ -42,7 +42,7 @@ following conditions:
require remote install using Redfish.
- Redfish |BMC| is required for orchestrated subcloud upgrades. The install
values, and :command:`bmc\_password` for each |AIO-SX| subcloud controller
values, and :command:`bmc_password` for each |AIO-SX| subcloud controller
must be provided using the following |CLI| command on the System Controller.
.. note::
@@ -84,7 +84,7 @@ following conditions:
each subcloud, use the following command to remove the previous upgrade
data:
:command:`sudo rm /opt/platform-backup/upgrade\_data\*`
:command:`sudo rm /opt/platform-backup/upgrade_data\*`
.. rubric:: |proc|
@@ -93,7 +93,7 @@ following conditions:
#. Review the upgrade status for the subclouds.
After the System Controller upgrade is completed, wait for 10 minutes for
the **load\_sync\_status** of all subclouds to be updated.
the **load_sync_status** of all subclouds to be updated.
To identify which subclouds are upgrade-current \(in-sync\), use the
:command:`subcloud list` command. For example:
@@ -169,7 +169,7 @@ following conditions:
upgraded in parallel, or serially.
If this is not specified using the CLI, the values for
:command:`subcloud\_update\_type` defined for each subcloud group will
:command:`subcloud_update_type` defined for each subcloud group will
be used by default.
**max-parallel-subclouds**
@@ -177,7 +177,7 @@ following conditions:
\(default 20\).
If this is not specified using the CLI, the values for
:command:`max\_parallel\_subclouds` defined for each subcloud group
:command:`max_parallel_subclouds` defined for each subcloud group
will be used by default.
**stop-on-failure**

View File

@@ -31,8 +31,8 @@ subcloud, the subcloud installation has these phases:
After a successful remote installation of a subcloud in a Distributed Cloud
system, a subsequent remote reinstallation fails because of an existing ssh
key entry in the /root/.ssh/known\_hosts on the System Controller. In this
case, delete the host key entry, if present, from /root/.ssh/known\_hosts
key entry in the /root/.ssh/known_hosts on the System Controller. In this
case, delete the host key entry, if present, from /root/.ssh/known_hosts
on the System Controller before doing reinstallations.
.. rubric:: |prereq|
@@ -203,7 +203,7 @@ subcloud, the subcloud installation has these phases:
password: <sysinv_password>
type: docker
Where <sysinv\_password> can be found by running the following command as
Where <sysinv_password> can be found by running the following command as
'sysadmin' on the Central Cloud:
.. code-block:: none
@@ -230,7 +230,7 @@ subcloud, the subcloud installation has these phases:
If you prefer to install container images from the default WRS |AWS| ECR
external registries, make the following substitutions for the
**docker\_registries** sections of the file.
**docker_registries** sections of the file.
.. code-block:: none
@@ -332,9 +332,9 @@ subcloud, the subcloud installation has these phases:
bootstrapping and deployment by monitoring the following log files on the
active controller in the Central Cloud.
/var/log/dcmanager/<subcloud\_name>\_install\_<date\_stamp>.log.
/var/log/dcmanager/<subcloud_name>_install_<date_stamp>.log.
/var/log/dcmanager/<subcloud\_name>\_bootstrap\_<date\_stamp>.log.
/var/log/dcmanager/<subcloud_name>_bootstrap_<date_stamp>.log.
For example:

View File

@@ -29,7 +29,7 @@ configuration file.
To send user data when calling nova boot, use the ``--user-data
/path/to/filename`` option, or use the Heat service and set the
``user\_data`` property and ``user\_data\_format`` to RAW.
``user_data`` property and ``user_data_format`` to RAW.
On initialization, the |VM| queries the metadata service through either
the EC2 compatibility API. For example:

View File

@@ -35,12 +35,12 @@ The procedure for adding an IP address depends on the interface type.
~(keystone_admin)$ systemhost-if-add <hostname> -V --vlan_id -c --ifclass <interfacename> <ethname>
**Virtual Function interfaces**
You can create an SROIV VF interface on top of an existing SROIV VF
You can create an |SRIOV| |VF| interface on top of an existing |SRIOV| |VF|
interface in order to configure a subset of virtual functions with
different drivers. For example, if the ethernet SR-IOV interface is
configured with the kernel VF driver, you can create a VF interface to
different drivers. For example, if the ethernet |SRIOV| interface is
configured with the kernel VF driver, you can create a |VF| interface to
configure a subset of virtual functions with the vfio driver that can be
used with userspace libraries such as DPDK.
used with userspace libraries such as |DPDK|.
.. code-block:: none
@@ -57,6 +57,7 @@ They can only be modified to use different physical ports when required.
On the second worker and storage nodes, the Ethernet interface for the
internal management network is attached automatically to support
installation using |PXE| booting.
On the initial controller node, the interface for the internal management
network is attached according to the settings specified during the
Ansible bootstrapping of the system.
@@ -77,12 +78,12 @@ They can only be modified to use different physical ports when required.
where the following options are available:
``node``
The name or UUID of the worker node.
The name or |UUID| of the worker node.
``ifname``
The name of the interface.
``ip\_address``
``ip_address``
An IPv4 or IPv6 address.
``prefix``

View File

@@ -49,7 +49,7 @@ Remarks
:command:`system modify --description="A TEST"` is logged to sysinv-api.log because it issues a REST POST call
:command:`system snmp-comm-delete "TEST\_COMMUNITY1"` - is logged to sysinv-api.log because it issues a REST DELETE call
:command:`system snmp-comm-delete "TEST_COMMUNITY1"` - is logged to sysinv-api.log because it issues a REST DELETE call
- If the :command:`sysinv` command only issues a REST GET call, it is not logged.

View File

@@ -86,12 +86,12 @@ configuration is required in order to use :command:`helm`.
.. note::
In order for your remote host to trust the certificate used by
the |prod-long| K8S API, you must ensure that the
``k8s\_root\_ca\_cert`` specified at install time is a trusted
``k8s_root_ca_cert`` specified at install time is a trusted
|CA| certificate by your host. Follow the instructions for adding
a trusted |CA| certificate for the operating system distribution
of your particular host.
If you did not specify a ``k8s\_root\_ca\_cert`` at install
If you did not specify a ``k8s_root_ca_cert`` at install
time, then specify ``--insecure-skip-tls-verify``, as shown below.
The following example configures the default ~/.kube/config. See the

View File

@@ -85,7 +85,7 @@ procedure.
.. _configuring-an-external-netapp-deployment-as-the-storage-backend-mod-localhost:
#. Configure Netapps configurable parameters and run the provided
install\_netapp\_backend.yml ansible playbook to enable connectivity to
install_netapp_backend.yml ansible playbook to enable connectivity to
Netapp as a storage backend for |prod|.
#. Provide Netapp backend configurable parameters in an overrides yaml
@@ -98,11 +98,11 @@ procedure.
The following parameters are mandatory:
``ansible\_become\_pass``
``ansible_become_pass``
Provide the admin password.
``netapp\_backends``
``name``
``netapp_backends``
**name**
A name for the storage class.
``provisioner``
@@ -136,18 +136,18 @@ procedure.
The following parameters are optional:
``trident\_setup\_dir``
``trident_setup_dir``
Set a staging directory for generated configuration files. The
default is /tmp/trident.
``trident\_namespace``
``trident_namespace``
Set this option to use an alternate Kubernetes namespace.
``trident\_rest\_api\_port``
``trident_rest_api_port``
Use an alternate port for the Trident REST API. The default is
8000.
``trident\_install\_extra\_params``
``trident_install_extra_params``
Add extra space-separated parameters when installing trident.
For complete listings of available parameters, see
@@ -190,8 +190,8 @@ procedure.
username: "admin"
password: "secret"
This file is sectioned into ``netapp\_k8s\_storageclass``,
``netapp\_k8s\_snapshotstorageclasses``, and ``netapp\_backends``.
This file is sectioned into ``netapp_k8s_storageclass``,
``netapp_k8s_snapshotstorageclasses``, and ``netapp_backends``.
You can add multiple backends and/or storage classes.
.. note::

View File

@@ -18,7 +18,7 @@ complement on a host.
In the Horizon Web interface and in |CLI| output, both identifications are
shown. For example, the output of the :command:`system host-disk-show`
command includes both the **device\_node** and the **device\_path**.
command includes both the **device_node** and the **device_path**.
.. code-block:: none

View File

@@ -24,7 +24,7 @@ This procedure uses standard Helm mechanisms to install a second
#. Capture a list of monitors.
This will be stored in the environment variable ``<MON\_LIST>`` and
This will be stored in the environment variable ``<MON_LIST>`` and
used in the following step.
.. code-block:: none

View File

@@ -24,7 +24,7 @@ You can change the space allotted for the Ceph monitor, if required.
~(keystone_admin)]$ system ceph-mon-modify <controller> ceph_mon_gib=<size>
where ``<partition\_size>`` is the size in GiB to use for the Ceph monitor.
where ``<partition_size>`` is the size in GiB to use for the Ceph monitor.
The value must be between 21 and 40 GiB.
.. code-block:: none
@@ -140,10 +140,10 @@ The following are optional arguments:
For a Ceph backend, this is a user-assigned name for the backend. The
default is **ceph-store** for a Ceph backend.
**-t,** ``--tier\_uuid``
**-t,** ``--tier_uuid``
For a Ceph backend, is the UUID of a storage tier to back.
**-c,** ``--ceph\_conf``
**-c,** ``--ceph_conf``
Location of the Ceph configuration file used for provisioning an
external backend.
@@ -152,7 +152,7 @@ The following are optional arguments:
reversible.
``--ceph-mon-gib``
For a Ceph backend, this is the space in gibibytes allotted for the
For a Ceph backend, this is the space in GB allotted for the
Ceph monitor.
.. note::