Merge "Add ceph commands in the 800 series alarm (dsR6)" into r/stx.6.0
This commit is contained in:
@@ -21,7 +21,8 @@ health of the system.
|
||||
* - **Alarm ID: 800.001**
|
||||
- Storage Alarm Condition:
|
||||
|
||||
1 mons down, quorum 1,2 controller-1,storage-0
|
||||
Possible data loss. Any mds, mon or osd is unavailable in storage
|
||||
replication group.
|
||||
* - Entity Instance
|
||||
- cluster=<dist-fs-uuid>
|
||||
* - Degrade Affecting Severity:
|
||||
@@ -29,14 +30,8 @@ health of the system.
|
||||
* - Severity:
|
||||
- C/M\*
|
||||
* - Proposed Repair Action
|
||||
- If problem persists, contact next level of support and provide the
|
||||
output of the following commands:
|
||||
|
||||
- ceph status
|
||||
- ceph fs status
|
||||
- system host-list
|
||||
- system cluster-list
|
||||
- system cluster-show <cluster-name>
|
||||
- Check the state of the Ceph cluster with :command:`ceph -s`. If problem
|
||||
persists, contact next level of support.
|
||||
|
||||
-----
|
||||
|
||||
@@ -54,13 +49,7 @@ health of the system.
|
||||
* - Severity:
|
||||
- m
|
||||
* - Proposed Repair Action
|
||||
- Update ceph storage pool quotas to use all available tier space and provide the output of the
|
||||
following commands:
|
||||
|
||||
- ceph status
|
||||
- ceph fs status
|
||||
- system host-fs-list <hostname>
|
||||
- system controllerfs-list
|
||||
- Update Ceph storage pool quotas to use all available tier space.
|
||||
|
||||
-----
|
||||
|
||||
@@ -78,15 +67,10 @@ health of the system.
|
||||
- C\*
|
||||
* - Proposed Repair Action
|
||||
- Ensure storage hosts from replication group are unlocked and available.
|
||||
Check if OSDs of each storage host are up and running. If problem
|
||||
persists contact next level of support and provide the output of the
|
||||
following commands:
|
||||
|
||||
- ceph status
|
||||
- ceph fs status
|
||||
- system host-list
|
||||
- system cluster-list
|
||||
- system cluster-show <cluster-name>
|
||||
Check replication group state with :command:`system host-list`. Check if
|
||||
OSDs of each storage host are up and running. Check the state of the
|
||||
Ceph OSDs with :command:`ceph osd stat` OR :command:`ceph osd tree`. If
|
||||
problem persists, contact next level of support.
|
||||
|
||||
-----
|
||||
|
||||
@@ -104,15 +88,10 @@ health of the system.
|
||||
- M\*
|
||||
* - Proposed Repair Action
|
||||
- Ensure storage hosts from replication group are unlocked and available.
|
||||
Check if OSDs of each storage host are up and running. If problem
|
||||
persists contact next level of support and provide the output of the
|
||||
following commands:
|
||||
|
||||
- ceph status
|
||||
- ceph fs status
|
||||
- system host-list
|
||||
- system cluster-list
|
||||
- system cluster-show <cluster-name>
|
||||
Check replication group state with :command:`system host-list`. Check if
|
||||
OSDs of each storage host are up and running. Check the state of the
|
||||
Ceph OSDs with :command:`ceph osd stat` AND/OR :command:`ceph osd tree`.
|
||||
If problem persists, contact next level of support.
|
||||
|
||||
-----
|
||||
|
||||
@@ -132,16 +111,7 @@ health of the system.
|
||||
* - Severity:
|
||||
- C/M\*
|
||||
* - Proposed Repair Action
|
||||
- Remove failed PV and associated Storage Device then recreate them and
|
||||
provide the output of the following commands:
|
||||
|
||||
- ceph status
|
||||
- ceph fs status
|
||||
- system helm-override-show platform-integ-apps rbd-provisioner kube-system
|
||||
|
||||
AND/OR
|
||||
|
||||
- system helm-override-show platform-integ-apps cephfs-provisioner kube-system
|
||||
- Remove failed PV and associated Storage Device then recreate them.
|
||||
|
||||
-----
|
||||
|
||||
@@ -164,15 +134,11 @@ health of the system.
|
||||
* - Severity:
|
||||
- C\*
|
||||
* - Proposed Repair Action
|
||||
- Increase Storage Space Allotment for Cinder on the 'lvm' backend.
|
||||
Consult the user documentation for more details. If problem persists,
|
||||
contact next level of support and provide the output of the
|
||||
following commands:
|
||||
|
||||
- ceph status
|
||||
- ceph fs status
|
||||
- system host-fs-list <hostname>
|
||||
- system controllerfs-list
|
||||
- Increase Storage Space Allotment for Cinder on the 'lvm' backend. Try
|
||||
the following commands: :command:`vgextend <VG name> <PV name>` or
|
||||
:command:`vgextend -L +<size extension> <PV name>`. Check status with
|
||||
:command:`vgdisplay`. Consult the System Administration Manual for more
|
||||
details. If problem persists, contact next level of support.
|
||||
|
||||
-----
|
||||
|
||||
@@ -189,12 +155,9 @@ health of the system.
|
||||
* - Severity:
|
||||
- C\*
|
||||
* - Proposed Repair Action
|
||||
- Update backend setting to reapply configuration. Consult the user
|
||||
documentation for more details. If problem persists, contact next level
|
||||
of support and provide the output of the
|
||||
following commands:
|
||||
|
||||
- ceph status
|
||||
- ceph fs status
|
||||
- system storage-backend-list
|
||||
- system storage-backend-show <storage-backend name>
|
||||
- Update backend setting to reapply configuration. Use the following
|
||||
commands to try again:
|
||||
:command:`system storage-backend-delete <storage-backend-name>` AND
|
||||
:command:`system storage-backend-add <storage-backend-name>`.
|
||||
Consult the user documentation for more details. If problem persists,
|
||||
contact next level of support.
|
||||
|
||||
Reference in New Issue
Block a user