Add ceph commands in the 800 series alarm document

When a 800-Series alarm occurs, users refer to the documentation to
know what kind of error is shown. But sometimes that is not enough
information.
The output of some commands can be useful information and could
save time when solving issues related to the storage alarms.

Closes-bug: 2004601

Test plan
PASS: * Build fm packages and deploy an ISO containing new fm
        packages.
      * Trigger alarms that were modified by this commit,
        (e.g. shutdown a controller).
      * Run fm alarm-list --uuid and copy the uuid of a 800-series
        alarm.
      * Run fm alarm-show <uuid> and check that the field
        has changed.

Signed-off-by: Agustin Carranza <agustin.carranza@windriver.com>
Change-Id: I94e2719b55b4fc14b692439526b5b47204460ac7
This commit is contained in:
Agustin Carranza 2023-02-02 17:15:07 -03:00
parent ca8be6b866
commit 0e1321913b
1 changed files with 18 additions and 4 deletions

View File

@ -1,7 +1,7 @@
--- ---
# #
# Copyright (c) 2013-2022 Wind River Systems, Inc. # Copyright (c) 2013-2023 Wind River Systems, Inc.
# #
# SPDX-License-Identifier: Apache-2.0 # SPDX-License-Identifier: Apache-2.0
# #
@ -3125,11 +3125,12 @@
800.001: 800.001:
Type: Alarm Type: Alarm
Description: |- Description: |-
Storage Alarm Condition: Possible data loss. Any mds, mon or osd is unavailable in storage replication group.
1 mons down, quorum 1,2 controller-1,storage-0
Entity_Instance_ID: cluster=<dist-fs-uuid> Entity_Instance_ID: cluster=<dist-fs-uuid>
Severity: [critical, major] Severity: [critical, major]
Proposed_Repair_Action: "If problem persists, contact next level of support." Proposed_Repair_Action: "Manually restart Ceph processes and check the state of the Ceph cluster with
'ceph -s'
If problem persists, contact next level of support."
Maintenance_Action: Maintenance_Action:
Inhibit_Alarms: Inhibit_Alarms:
Alarm_Type: equipment Alarm_Type: equipment
@ -3149,7 +3150,10 @@
Entity_Instance_ID: cluster=<dist-fs-uuid>.peergroup=<group-x> Entity_Instance_ID: cluster=<dist-fs-uuid>.peergroup=<group-x>
Severity: [critical] Severity: [critical]
Proposed_Repair_Action: "Ensure storage hosts from replication group are unlocked and available. Proposed_Repair_Action: "Ensure storage hosts from replication group are unlocked and available.
Check replication group state with 'system host-list'
Check if OSDs of each storage host are up and running. Check if OSDs of each storage host are up and running.
Manually restart Ceph processes and check the state of the Ceph OSDs with
'ceph osd stat' OR 'ceph osd tree'
If problem persists, contact next level of support." If problem persists, contact next level of support."
Maintenance_Action: Maintenance_Action:
Inhibit_Alarms: Inhibit_Alarms:
@ -3169,7 +3173,10 @@
Entity_Instance_ID: cluster=<dist-fs-uuid>.peergroup=<group-x> Entity_Instance_ID: cluster=<dist-fs-uuid>.peergroup=<group-x>
Severity: [major] Severity: [major]
Proposed_Repair_Action: "Ensure storage hosts from replication group are unlocked and available. Proposed_Repair_Action: "Ensure storage hosts from replication group are unlocked and available.
Check replication group state with 'system host-list'
Check if OSDs of each storage host are up and running. Check if OSDs of each storage host are up and running.
Manually restart Ceph processes and check the state of the Ceph OSDs with
'ceph osd stat' AND/OR 'ceph osd tree'
If problem persists, contact next level of support." If problem persists, contact next level of support."
Maintenance_Action: Maintenance_Action:
Inhibit_Alarms: Inhibit_Alarms:
@ -3298,6 +3305,9 @@
Entity_Instance_ID: <hostname>.lvmthinpool=<VG name>/<Pool name> Entity_Instance_ID: <hostname>.lvmthinpool=<VG name>/<Pool name>
Severity: critical Severity: critical
Proposed_Repair_Action: "Increase Storage Space Allotment for Cinder on the 'lvm' backend. Proposed_Repair_Action: "Increase Storage Space Allotment for Cinder on the 'lvm' backend.
Try the following commands:
'vgextend <VG name> <PV name>' or 'vgextend -L +<size extension> <PV name>
Check status with 'vgdisplay'
Consult the System Administration Manual for more details. Consult the System Administration Manual for more details.
If problem persists, contact next level of support." If problem persists, contact next level of support."
Maintenance_Action: Maintenance_Action:
@ -3318,6 +3328,10 @@
Entity_Instance_ID: storage_backend=<storage-backend-name> Entity_Instance_ID: storage_backend=<storage-backend-name>
Severity: critical Severity: critical
Proposed_Repair_Action: "Update backend setting to reapply configuration. Proposed_Repair_Action: "Update backend setting to reapply configuration.
Use the following commands to try again:
'system storage-backend-delete <storage-backend-name>'
AND
'system storage-backend-add <storage-backend-name>'
Consult the System Administration Manual for more details. Consult the System Administration Manual for more details.
If problem persists, contact next level of support." If problem persists, contact next level of support."
Maintenance_Action: Maintenance_Action: