Add ceph commands in the 800 series alarm (dsR6)

Following updates made in alarms in https://review.opendev.org/c/starlingx/fault/+/872600

Closes-bug: 2011718

Signed-off-by: Elisamara Aoki Goncalves <elisamaraaoki.goncalves@windriver.com>
Change-Id: I71c037b28657fa00b385e000b6be9a422115e6d0
This commit is contained in:
Elisamara Aoki Goncalves
2023-02-14 12:12:41 -03:00
parent 6a6d82226b
commit ce7b81b32a

View File

@@ -21,7 +21,8 @@ health of the system.
* - **Alarm ID: 800.001** * - **Alarm ID: 800.001**
- Storage Alarm Condition: - Storage Alarm Condition:
1 mons down, quorum 1,2 controller-1,storage-0 Possible data loss. Any mds, mon or osd is unavailable in storage
replication group.
* - Entity Instance * - Entity Instance
- cluster=<dist-fs-uuid> - cluster=<dist-fs-uuid>
* - Degrade Affecting Severity: * - Degrade Affecting Severity:
@@ -29,14 +30,8 @@ health of the system.
* - Severity: * - Severity:
- C/M\* - C/M\*
* - Proposed Repair Action * - Proposed Repair Action
- If problem persists, contact next level of support and provide the - Check the state of the Ceph cluster with :command:`ceph -s`. If problem
output of the following commands: persists, contact next level of support.
- ceph status
- ceph fs status
- system host-list
- system cluster-list
- system cluster-show <cluster-name>
----- -----
@@ -54,13 +49,7 @@ health of the system.
* - Severity: * - Severity:
- m - m
* - Proposed Repair Action * - Proposed Repair Action
- Update ceph storage pool quotas to use all available tier space and provide the output of the - Update Ceph storage pool quotas to use all available tier space.
following commands:
- ceph status
- ceph fs status
- system host-fs-list <hostname>
- system controllerfs-list
----- -----
@@ -78,15 +67,10 @@ health of the system.
- C\* - C\*
* - Proposed Repair Action * - Proposed Repair Action
- Ensure storage hosts from replication group are unlocked and available. - Ensure storage hosts from replication group are unlocked and available.
Check if OSDs of each storage host are up and running. If problem Check replication group state with :command:`system host-list`. Check if
persists contact next level of support and provide the output of the OSDs of each storage host are up and running. Check the state of the
following commands: Ceph OSDs with :command:`ceph osd stat` OR :command:`ceph osd tree`. If
problem persists, contact next level of support.
- ceph status
- ceph fs status
- system host-list
- system cluster-list
- system cluster-show <cluster-name>
----- -----
@@ -104,15 +88,10 @@ health of the system.
- M\* - M\*
* - Proposed Repair Action * - Proposed Repair Action
- Ensure storage hosts from replication group are unlocked and available. - Ensure storage hosts from replication group are unlocked and available.
Check if OSDs of each storage host are up and running. If problem Check replication group state with :command:`system host-list`. Check if
persists contact next level of support and provide the output of the OSDs of each storage host are up and running. Check the state of the
following commands: Ceph OSDs with :command:`ceph osd stat` AND/OR :command:`ceph osd tree`.
If problem persists, contact next level of support.
- ceph status
- ceph fs status
- system host-list
- system cluster-list
- system cluster-show <cluster-name>
----- -----
@@ -132,16 +111,7 @@ health of the system.
* - Severity: * - Severity:
- C/M\* - C/M\*
* - Proposed Repair Action * - Proposed Repair Action
- Remove failed PV and associated Storage Device then recreate them and - Remove failed PV and associated Storage Device then recreate them.
provide the output of the following commands:
- ceph status
- ceph fs status
- system helm-override-show platform-integ-apps rbd-provisioner kube-system
AND/OR
- system helm-override-show platform-integ-apps cephfs-provisioner kube-system
----- -----
@@ -164,15 +134,11 @@ health of the system.
* - Severity: * - Severity:
- C\* - C\*
* - Proposed Repair Action * - Proposed Repair Action
- Increase Storage Space Allotment for Cinder on the 'lvm' backend. - Increase Storage Space Allotment for Cinder on the 'lvm' backend. Try
Consult the user documentation for more details. If problem persists, the following commands: :command:`vgextend <VG name> <PV name>` or
contact next level of support and provide the output of the :command:`vgextend -L +<size extension> <PV name>`. Check status with
following commands: :command:`vgdisplay`. Consult the System Administration Manual for more
details. If problem persists, contact next level of support.
- ceph status
- ceph fs status
- system host-fs-list <hostname>
- system controllerfs-list
----- -----
@@ -189,12 +155,9 @@ health of the system.
* - Severity: * - Severity:
- C\* - C\*
* - Proposed Repair Action * - Proposed Repair Action
- Update backend setting to reapply configuration. Consult the user - Update backend setting to reapply configuration. Use the following
documentation for more details. If problem persists, contact next level commands to try again:
of support and provide the output of the :command:`system storage-backend-delete <storage-backend-name>` AND
following commands: :command:`system storage-backend-add <storage-backend-name>`.
Consult the user documentation for more details. If problem persists,
- ceph status contact next level of support.
- ceph fs status
- system storage-backend-list
- system storage-backend-show <storage-backend name>