The template to monitor SAN NetApp FAS3220 cluster by Zabbix SNMP agent.
Zabbix version: 6.0 and higher.
This template has been tested on:
Zabbix should be configured according to the instructions in the Templates out of the box section.
Create a host for FAS3220 with cluster management IP as SNMPv2 interface.
Link the template to the host.
Customize macro values if needed.
Name | Description | Default |
---|---|---|
{$IF.UTIL.MAX} | 95 |
|
{$IF.ERRORS.WARN} | ||
{$CPU.UTIL.CRIT} | The critical threshold of the CPU utilization expressed in %. |
90 |
{$FAS3220.FS.NAME.MATCHES} | This macro is used in filesystems discovery. Can be overridden on the host or linked template level. |
.* |
{$FAS3220.FS.NAME.NOT_MATCHES} | This macro is used in filesystems discovery. Can be overridden on the host or linked template level. |
snapshot |
{$FAS3220.FS.TYPE.MATCHES} | This macro is used in filesystems discovery. Can be overridden on the host or linked template level. Value should be integer: 2 - flexibleVolume, 3 - aggregate, 4 - stripedAggregate, 5 - stripedVolume. |
.* |
{$FAS3220.FS.TYPE.NOT_MATCHES} | This macro is used in filesystems discovery. Can be overridden on the host or linked template level. Value should be integer: 2 - flexibleVolume, 3 - aggregate, 4 - stripedAggregate, 5 - stripedVolume. |
CHANGE_IF_NEEDED |
{$FAS3220.NET.PORT.TYPE.MATCHES} | This macro is used in net ports discovery. Can be overridden on the host or linked template level. {#TYPE} is integer. Possible values: physical, if-group, vlan, undef. |
.* |
{$FAS3220.NET.PORT.TYPE.NOT_MATCHES} | This macro is used in net ports discovery. Can be overridden on the host or linked template level. {#TYPE} is integer. Possible values: physical, if-group, vlan, undef. |
CHANGE_IF_NEEDED |
{$FAS3220.NET.PORT.ROLE.MATCHES} | This macro is used in net ports discovery. Can be overridden on the host or linked template level. {#ROLE} is integer. Possible values: 0 - undef 1 - cluster 2 - data 3 - node-mgmt 4 - intercluster 5 - cluster-mgmt |
.* |
{$FAS3220.NET.PORT.ROLE.NOT_MATCHES} | This macro is used in net ports discovery. Can be overridden on the host or linked template level. {#ROLE} is integer. Possible values: 0 - undef 1 - cluster 2 - data 3 - node-mgmt 4 - intercluster 5 - cluster-mgmt |
CHANGE_IF_NEEDED |
{$FAS3220.NET.PORT.NAME.MATCHES} | This macro is used in net ports discovery. Can be overridden on the host or linked template level. |
.* |
{$FAS3220.NET.PORT.NAME.NOT_MATCHES} | This macro is used in net ports discovery. Can be overridden on the host or linked template level. |
CHANGE_IF_NEEDED |
{$FAS3220.FS.PUSED.MAX.CRIT} | Maximum percentage of disk used. Can be used with {#FSNAME} as context. |
90 |
{$FAS3220.FS.AVAIL.MIN.CRIT} | Minimum available space on the disk. Can be used with {#FSNAME} as context. |
10G |
{$FAS3220.FS.TIME} | The time during which disk usage may exceed the threshold. Can be used with {#FSNAME} as context. |
10m |
{$FAS3220.FS.USE.PCT} | Macro define what threshold will be used for disk space trigger: 0 - use Bytes ({$FAS3220.FS.AVAIL.MIN.CRIT}) 1 - use percents ({$FAS3220.FS.PUSED.MAX.CRIT}) Can be used with {#FSNAME} as context. |
1 |
{$SNMP.TIMEOUT} | 5m |
|
{$ICMPLOSSWARN} | 20 |
|
{$ICMPRESPONSETIME_WARN} | 0.15 |
Name | Description | Type | Key and additional info |
---|---|---|---|
NetApp FAS3220: Product version | MIB: NETAPP-MIB Version string for the software running on this platform. |
SNMP agent | fas3220.inventory[productVersion] Preprocessing
|
NetApp FAS3220: Product firmware version | Version string for the firmware running on this platform. |
SNMP agent | fas3220.inventory[productFirmwareVersion] Preprocessing
|
NetApp FAS3220: Failed disks count | The number of disks that are currently broken. |
SNMP agent | fas3220.disk[diskFailedCount] Preprocessing
|
NetApp FAS3220: Failed disks message | If diskFailedCount is non-zero, this is a string describing the failed disk or disks. Each failed disk is described. |
SNMP agent | fas3220.disk[diskFailedMessage] Preprocessing
|
NetApp FAS3220: Uptime (network) | MIB: SNMPv2-MIB The time (in hundredths of a second) since the network management portion of the system was last re-initialized. |
SNMP agent | system.net.uptime[sysUpTime.0] Preprocessing
|
NetApp FAS3220: Uptime (hardware) | MIB: HOST-RESOURCES-MIB The amount of time since this host was last initialized. Note that this is different from sysUpTime in the SNMPv2-MIB [RFC1907] because sysUpTime is the uptime of the network management portion of the system. |
SNMP agent | system.hw.uptime[hrSystemUptime.0] Preprocessing
|
NetApp FAS3220: SNMP traps (fallback) | The item is used to collect all SNMP traps unmatched by other snmptrap items |
SNMP trap | snmptrap.fallback |
NetApp FAS3220: System location | MIB: SNMPv2-MIB The physical location of this node (e.g., `telephone closet, 3rd floor'). If the location is unknown, the value is the zero-length string. |
SNMP agent | system.location[sysLocation.0] Preprocessing
|
NetApp FAS3220: System contact details | MIB: SNMPv2-MIB The textual identification of the contact person for this managed node, together with information on how to contact this person. If no contact information is known, the value is the zero-length string. |
SNMP agent | system.contact[sysContact.0] Preprocessing
|
NetApp FAS3220: System object ID | MIB: SNMPv2-MIB The vendor's authoritative identification of the network management subsystem contained in the entity. This value is allocated within the SMI enterprises subtree (1.3.6.1.4.1) and provides an easy and unambiguous means for determining |
SNMP agent | system.objectid[sysObjectID.0] Preprocessing
|
NetApp FAS3220: System name | MIB: SNMPv2-MIB An administratively-assigned name for this managed node.By convention, this is the node's fully-qualified domain name. If the name is unknown, the value is the zero-length string. |
SNMP agent | system.name Preprocessing
|
NetApp FAS3220: System description | MIB: SNMPv2-MIB A textual description of the entity. This value should include the full name and version identification of the system's hardware type, software operating-system, and networking software. |
SNMP agent | system.descr[sysDescr.0] Preprocessing
|
NetApp FAS3220: SNMP agent availability | Availability of SNMP checks on the host. The value of this item corresponds to availability icons in the host list. Possible value: 0 - not available 1 - available 2 - unknown |
Zabbix internal | zabbix[host,snmp,available] |
NetApp FAS3220: ICMP ping | Simple check | icmpping | |
NetApp FAS3220: ICMP loss | Simple check | icmppingloss | |
NetApp FAS3220: ICMP response time | Simple check | icmppingsec |
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
NetApp FAS3220: Number of failed disks has changed | {{ITEM.LASTVALUE2}.regsub("(.*)", \1)} |
last(/NetApp FAS3220 by SNMP/fas3220.disk[diskFailedCount])>0 and last(/NetApp FAS3220 by SNMP/fas3220.disk[diskFailedMessage],#1)<>last(/NetApp FAS3220 by SNMP/fas3220.disk[diskFailedMessage],#2) |Warning |
||
NetApp FAS3220: Host has been restarted | Uptime is less than 10 minutes. |
(last(/NetApp FAS3220 by SNMP/system.hw.uptime[hrSystemUptime.0])>0 and last(/NetApp FAS3220 by SNMP/system.hw.uptime[hrSystemUptime.0])<10m) or (last(/NetApp FAS3220 by SNMP/system.hw.uptime[hrSystemUptime.0])=0 and last(/NetApp FAS3220 by SNMP/system.net.uptime[sysUpTime.0])<10m) |Warning |
Manual close: Yes Depends on:
|
|
NetApp FAS3220: System name has changed | The name of the system has changed. Acknowledge to close the problem manually. |
last(/NetApp FAS3220 by SNMP/system.name,#1)<>last(/NetApp FAS3220 by SNMP/system.name,#2) and length(last(/NetApp FAS3220 by SNMP/system.name))>0 |Info |
Manual close: Yes | |
NetApp FAS3220: No SNMP data collection | SNMP is not available for polling. Please check device connectivity and SNMP settings. |
max(/NetApp FAS3220 by SNMP/zabbix[host,snmp,available],{$SNMP.TIMEOUT})=0 |Warning |
Depends on:
|
|
NetApp FAS3220: Unavailable by ICMP ping | Last three attempts returned timeout. Please check device connectivity. |
max(/NetApp FAS3220 by SNMP/icmpping,#3)=0 |High |
||
NetApp FAS3220: High ICMP ping loss | min(/NetApp FAS3220 by SNMP/icmppingloss,5m)>{$ICMP_LOSS_WARN} and min(/NetApp FAS3220 by SNMP/icmppingloss,5m)<100 |Warning |
Depends on:
|
||
NetApp FAS3220: High ICMP ping response time | avg(/NetApp FAS3220 by SNMP/icmppingsec,5m)>{$ICMP_RESPONSE_TIME_WARN} |Warning |
Depends on:
|
Name | Description | Type | Key and additional info |
---|---|---|---|
CPU discovery | Discovery of CPU metrics per node |
SNMP agent | fas3220.cpu.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Node {#NODE.NAME}: CPU utilization | The average, over the last minute, of the percentage of time that this processor was not idle. |
SNMP agent | fas3220.cpu[cDOTCpuBusyTimePerCent, "{#NODE.NAME}"] |
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Node {#NODE.NAME}: High CPU utilization | The CPU utilization is too high. The system might be slow to respond. |
min(/NetApp FAS3220 by SNMP/fas3220.cpu[cDOTCpuBusyTimePerCent, "{#NODE.NAME}"],5m)>{$CPU.UTIL.CRIT} |Warning |
Name | Description | Type | Key and additional info |
---|---|---|---|
Cluster metrics discovery | Discovery of Cluster metrics per node |
SNMP agent | fas3220.cluster.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Node {#NODE.NAME}: Location | Node Location. Same as sysLocation for a specific node. |
SNMP agent | fas3220.cluster[nodeLocation, "{#NODE.NAME}"] Preprocessing
|
Node {#NODE.NAME}: Model | Node Model. Same as productModel for a specific node. |
SNMP agent | fas3220.cluster[nodeModel, "{#NODE.NAME}"] Preprocessing
|
Node {#NODE.NAME}: Serial number | Node Serial Number. Same as productSerialNum for a specific node. |
SNMP agent | fas3220.cluster[nodeSerialNumber, "{#NODE.NAME}"] Preprocessing
|
Node {#NODE.NAME}: Uptime | Node uptime. Same as sysUpTime for a specific node. |
SNMP agent | fas3220.cluster[nodeUptime, "{#NODE.NAME}"] Preprocessing
|
Node {#NODE.NAME}: Health | Whether or not the node can communicate with the cluster. |
SNMP agent | fas3220.cluster[nodeHealth, "{#NODE.NAME}"] Preprocessing
|
Node {#NODE.NAME}: NVRAM battery status | An indication of the current status of the NVRAM battery or batteries. Batteries which are fully or partially discharged may not fully protect the system during a crash. The end-of-life status values are based on the manufacturer's recommended life for the batteries. Possible values: ok(1), partiallyDischarged(2), fullyDischarged(3), notPresent(4), nearEndOfLife(5), atEndOfLife(6), unknown(7), overCharged(8), fullyCharged(9). |
SNMP agent | fas3220.cluster[nodeNvramBatteryStatus, "{#NODE.NAME}"] Preprocessing
|
Node {#NODE.NAME}: Over-temperature | An indication of whether the hardware is currently operating outside of its recommended temperature range. The hardware will shutdown if the temperature exceeds critical thresholds. |
SNMP agent | fas3220.cluster[nodeEnvOverTemperature, "{#NODE.NAME}"] Preprocessing
|
Node {#NODE.NAME}: Failed FAN count | Count of the number of chassis fans that are not operating within the recommended RPM range. |
SNMP agent | fas3220.cluster[nodeEnvFailedFanCount, "{#NODE.NAME}"] Preprocessing
|
Node {#NODE.NAME}: Failed FAN message | Text message describing current condition of chassis fans. This is useful only if envFailedFanCount is not zero. |
SNMP agent | fas3220.cluster[nodeEnvFailedFanMessage, "{#NODE.NAME}"] Preprocessing
|
Node {#NODE.NAME}: Degraded power supplies count | Count of the number of power supplies that are in degraded mode. |
SNMP agent | fas3220.cluster[nodeEnvFailedPowerSupplyCount, "{#NODE.NAME}"] Preprocessing
|
Node {#NODE.NAME}: Degraded power supplies message | Text message describing the state of any power supplies that are currently degraded. This is useful only if envFailedPowerSupplyCount is not zero. |
SNMP agent | fas3220.cluster[nodeEnvFailedPowerSupplyMessage, "{#NODE.NAME}"] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Node {#NODE.NAME}: Host has been restarted | Uptime is less than 10 minutes. |
last(/NetApp FAS3220 by SNMP/fas3220.cluster[nodeUptime, "{#NODE.NAME}"])<10m |Info |
Manual close: Yes | |
Node {#NODE.NAME}: Node can not communicate with the cluster | last(/NetApp FAS3220 by SNMP/fas3220.cluster[nodeHealth, "{#NODE.NAME}"])=0 |High |
Manual close: Yes | ||
Node {#NODE.NAME}: NVRAM battery status is not OK | last(/NetApp FAS3220 by SNMP/fas3220.cluster[nodeNvramBatteryStatus, "{#NODE.NAME}"])<>1 |Average |
Manual close: Yes | ||
Node {#NODE.NAME}: Temperature is over than recommended | The hardware will shutdown if the temperature exceeds critical thresholds. |
last(/NetApp FAS3220 by SNMP/fas3220.cluster[nodeEnvOverTemperature, "{#NODE.NAME}"])=2 |High |
||
Node {#NODE.NAME}: Failed FAN count is over than zero | {{ITEM.VALUE2}.regsub("(.*)", \1)} |
last(/NetApp FAS3220 by SNMP/fas3220.cluster[nodeEnvFailedFanCount, "{#NODE.NAME}"])>0 and last(/NetApp FAS3220 by SNMP/fas3220.cluster[nodeEnvFailedFanMessage, "{#NODE.NAME}"])=last(/NetApp FAS3220 by SNMP/fas3220.cluster[nodeEnvFailedFanMessage, "{#NODE.NAME}"]) |High |
||
Node {#NODE.NAME}: Degraded power supplies count is more than zero | {{ITEM.VALUE2}.regsub("(.*)", \1)} |
last(/NetApp FAS3220 by SNMP/fas3220.cluster[nodeEnvFailedPowerSupplyCount, "{#NODE.NAME}"])>0 and last(/NetApp FAS3220 by SNMP/fas3220.cluster[nodeEnvFailedPowerSupplyMessage, "{#NODE.NAME}"])=last(/NetApp FAS3220 by SNMP/fas3220.cluster[nodeEnvFailedPowerSupplyMessage, "{#NODE.NAME}"]) |Average |
Name | Description | Type | Key and additional info |
---|---|---|---|
HA discovery | Discovery of high availability metrics per node |
SNMP agent | fas3220.ha.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Node {#NODE.NAME}: Cannot takeover cause | The reason node cannot take over it's HA partner {#PARTNER.NAME}. Possible states: ok(1), unknownReason(2), disabledByOperator(3), interconnectOffline(4), disabledByPartner(5), takeoverFailed(6), mailboxIsInDegradedState(7), partnermailboxIsInUninitialisedState(8), mailboxVersionMismatch(9), nvramSizeMismatch(10), kernelVersionMismatch(11), partnerIsInBootingStage(12), diskshelfIsTooHot(13), partnerIsPerformingRevert(14), nodeIsPerformingRevert(15), sametimePartnerIsAlsoTryingToTakeUsOver(16), alreadyInTakenoverMode(17), nvramLogUnsynchronized(18), stateofBackupMailboxIsDoubtful(19). |
SNMP agent | fas3220.ha[haCannotTakeoverCause, "{#NODE.NAME}"] Preprocessing
|
Node {#NODE.NAME}: HA settings | High Availability configuration settings. The value notConfigured(1) indicates that the HA is not licensed. The thisNodeDead(5) setting indicates that this node has been takenover. |
SNMP agent | fas3220.ha[haSettings, "{#NODE.NAME}"] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Node {#NODE.NAME}: Node cannot takeover it's HA partner {#PARTNER.NAME}. Reason: {ITEM.VALUE} | Possible reasons: |
last(/NetApp FAS3220 by SNMP/fas3220.ha[haCannotTakeoverCause, "{#NODE.NAME}"])<>1 |High |
||
Node {#NODE.NAME}: Node has been taken over | The thisNodeDead(5) setting indicates that this node has been takenover. |
last(/NetApp FAS3220 by SNMP/fas3220.ha[haSettings, "{#NODE.NAME}"])=5 |High |
||
Node {#NODE.NAME}: HA is not licensed | The value notConfigured(1) indicates that the HA is not licensed. |
last(/NetApp FAS3220 by SNMP/fas3220.ha[haSettings, "{#NODE.NAME}"])=1 |Average |
Name | Description | Type | Key and additional info |
---|---|---|---|
Filesystems discovery | Filesystems discovery with filter. |
SNMP agent | fas3220.fs.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
{#VSERVER}{#FSNAME}: Total space used | The total disk space that is in use on {#FSNAME}. |
SNMP agent | fas3220.fs[df64UsedKBytes, "{#VSERVER}{#FSNAME}"] Preprocessing
|
{#VSERVER}{#FSNAME}: Total space available | The total disk space that is free for use on {#FSNAME}. |
SNMP agent | fas3220.fs[df64AvailKBytes, "{#VSERVER}{#FSNAME}"] Preprocessing
|
{#VSERVER}{#FSNAME}: Total space | The total capacity in bytes for {#FSNAME}. |
SNMP agent | fas3220.fs[df64TotalKBytes, "{#VSERVER}{#FSNAME}"] Preprocessing
|
{#VSERVER}{#FSNAME}: Used space percents | The percentage of disk space currently in use on {#FSNAME}. |
SNMP agent | fas3220.fs[dfPerCentKBytesCapacity, "{#VSERVER}{#FSNAME}"] |
{#VSERVER}{#FSNAME}: Saved by compression percents | Provides the percentage of compression savings in a volume, which is ((comprsaved/used)) * 10(comprsaved + 0). This is only returned for volumes. |
SNMP agent | fas3220.fs[dfCompressSavedPercent, "{#VSERVER}{#FSNAME}"] |
{#VSERVER}{#FSNAME}: Saved by deduplication percents | Provides the percentage of deduplication savings in a volume, which is ((dedupsaved/(dedupsaved + used)) * 100). This is only returned for volumes. |
SNMP agent | fas3220.fs[dfDedupeSavedPercent, "{#VSERVER}{#FSNAME}"] |
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
{#VSERVER}{#FSNAME}: Disk space is too low | min(/NetApp FAS3220 by SNMP/fas3220.fs[df64AvailKBytes, "{#VSERVER}{#FSNAME}"],{$FAS3220.FS.TIME:"{#FSNAME}"})<{$FAS3220.FS.AVAIL.MIN.CRIT:"{#FSNAME}"} and {$FAS3220.FS.USE.PCT:"{#FSNAME}"}=0 |High |
|||
{#VSERVER}{#FSNAME}: Disk space is too low | max(/NetApp FAS3220 by SNMP/fas3220.fs[dfPerCentKBytesCapacity, "{#VSERVER}{#FSNAME}"],{$FAS3220.FS.TIME:"{#FSNAME}"})>{$FAS3220.FS.PUSED.MAX.CRIT:"{#FSNAME}"} and {$FAS3220.FS.USE.PCT:"{#FSNAME}"}=1 |High |
Name | Description | Type | Key and additional info |
---|---|---|---|
Network ports discovery | Network interfaces discovery with filter. |
SNMP agent | fas3220.net.discovery Preprocessing
|
Name | Description | Type | Key and additional info |
---|---|---|---|
Node {#NODE}: port {#IFNAME} ({#TYPE}): Up by an administrator | Indicates whether the port status is set 'UP' by an administrator. |
SNMP agent | fas3220.net.port[netportUpAdmin, "{#NODE}", "{#IFNAME}"] Preprocessing
|
Node {#NODE}: port {#IFNAME} ({#TYPE}): Role | Role of the port. A port must have one of the following roles: cluster(1), data(2), mgmt(3), intercluster(4), cluster-mgmt(5) or undef(0). The cluster port is used to communicate to other node(s) in the cluster. The data port services clients' requests. It is where all the file requests come in. The management port is used by administrator to manage resources within a node. The intercluster port is used to communicate to other cluster. The cluster-mgmt port is used to manage resources within the cluster. The undef role is for the port that has not yet been assigned a role. |
SNMP agent | fas3220.net.port[netportRole, "{#NODE}", "{#IFNAME}"] Preprocessing
|
Node {#NODE}: port {#IFNAME} ({#TYPE}): Speed | The speed appears on the port. It can be either undef(0), auto(1), ten Mb/s(2), hundred Mb/s(3), one Gb/s(4), or ten Gb/s(5). |
SNMP agent | fas3220.net.port[netportSpeedOper, "{#NODE}", "{#IFNAME}"] Preprocessing
|
Node {#NODE}: port {#IFNAME} ({#TYPE}): Bits received | The total number of octets received on the interface, including framing characters. |
SNMP agent | fas3220.net.if[if64InOctets, "{#NODE}", "{#IFNAME}"] Preprocessing
|
Node {#NODE}: port {#IFNAME} ({#TYPE}): Bits sent | The total number of octets transmitted out of the interface, including framing characters. |
SNMP agent | fas3220.net.if[if64OutOctets, "{#NODE}", "{#IFNAME}"] Preprocessing
|
Node {#NODE}: port {#IFNAME} ({#TYPE}): Inbound packets with errors | MIB: IF-MIB The number of inbound packets that contained errors preventing them from being deliverable to a higher-layer protocol. |
SNMP agent | fas3220.net.if[if64InErrors, "{#NODE}", "{#IFNAME}"] Preprocessing
|
Node {#NODE}: port {#IFNAME} ({#TYPE}): Outbound packets with errors | MIB: IF-MIB The number of outbound packets that could not be transmitted because of errors. |
SNMP agent | fas3220.net.if[if64OutErrors, "{#NODE}", "{#IFNAME}"] Preprocessing
|
Node {#NODE}: port {#IFNAME} ({#TYPE}): Inbound packets discarded | MIB: IF-MIB The number of inbound packets that were chosen to be discarded even though no errors had been detected to prevent their being deliverable to a higher-layer protocol. One possible reason for discarding such a packet could be to free up buffer space. |
SNMP agent | fas3220.net.if[if64InDiscards, "{#NODE}", "{#IFNAME}"] Preprocessing
|
Node {#NODE}: port {#IFNAME} ({#TYPE}): Outbound packets discarded | MIB: IF-MIB The number of outbound packets that were chosen to be discarded even though no errors had been detected to prevent their being transmitted. One possible reason for discarding such a packet could be to free up buffer space. |
SNMP agent | fas3220.net.if[if64OutDiscards, "{#NODE}", "{#IFNAME}"] Preprocessing
|
Node {#NODE}: port {#IFNAME} ({#TYPE}): State | The link-state of the port. Normally it is either UP(2) or DOWN(3). |
SNMP agent | fas3220.net.port[netportLinkState, "{#NODE}", "{#IFNAME}"] Preprocessing
|
Node {#NODE}: port {#IFNAME} ({#TYPE}): Health | The health status of the port. |
SNMP agent | fas3220.net.port[netportHealthStatus, "{#NODE}", "{#IFNAME}"] Preprocessing
|
Node {#NODE}: port {#IFNAME} ({#TYPE}): Health degraded reason | The list of reasons why the port is marked as degraded. |
SNMP agent | fas3220.net.port[netportDegradedReason, "{#NODE}", "{#IFNAME}"] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Node {#NODE}: port {#IFNAME} ({#TYPE}): High error rate | It recovers when it is below 80% of the |
min(/NetApp FAS3220 by SNMP/fas3220.net.if[if64InErrors, "{#NODE}", "{#IFNAME}"],5m)>{$IF.ERRORS.WARN:"{#IFNAME}"} or min(/NetApp FAS3220 by SNMP/fas3220.net.if[if64OutErrors, "{#NODE}", "{#IFNAME}"],5m)>{$IF.ERRORS.WARN:"{#IFNAME}"} |Warning |
Manual close: Yes | |
Node {#NODE}: port {#IFNAME} ({#TYPE}): Link down | Link state is not UP and the port status is set 'UP' by an administrator. |
last(/NetApp FAS3220 by SNMP/fas3220.net.port[netportLinkState, "{#NODE}", "{#IFNAME}"])<>2 and last(/NetApp FAS3220 by SNMP/fas3220.net.port[netportUpAdmin, "{#NODE}", "{#IFNAME}"])=1 |Average |
Manual close: Yes | |
Node {#NODE}: port {#IFNAME} ({#TYPE}): Port is not healthy | {{ITEM.LASTVALUE2}.regsub("(.*)", \1)} |
last(/NetApp FAS3220 by SNMP/fas3220.net.port[netportHealthStatus, "{#NODE}", "{#IFNAME}"])<>0 and length(last(/NetApp FAS3220 by SNMP/fas3220.net.port[netportDegradedReason, "{#NODE}", "{#IFNAME}"]))>0 |Info |
Please report any issues with the template at https://support.zabbix.com
You can also provide feedback, discuss the template, or ask for help at ZABBIX forums
The template to monitor SAN NetApp AFF A700 cluster by Zabbix HTTP agent.
Zabbix version: 6.0 and higher.
This template has been tested on:
Zabbix should be configured according to the instructions in the Templates out of the box section.
1. Create a host for AFF A700 with cluster management IP as the Zabbix agent interface.
2. Link the template to the host.
3. Customize macro values if needed.
Name | Description | Default |
---|---|---|
{$URL} | AFF700 cluster URL address. |
|
{$USERNAME} | AFF700 user name. |
|
{$PASSWORD} | AFF700 user password. |
|
{$HTTP.AGENT.TIMEOUT} | The HTTP agent timeout to wait for a response from AFF700. |
3s |
Name | Description | Type | Key and additional info |
---|---|---|---|
NetApp AFF A700: Get cluster | HTTP agent | netapp.cluster.get | |
NetApp AFF A700: Get nodes | HTTP agent | netapp.nodes.get | |
NetApp AFF A700: Get disks | HTTP agent | netapp.disks.get | |
NetApp AFF A700: Get volumes | HTTP agent | netapp.volumes.get | |
NetApp AFF A700: Get ethernet ports | HTTP agent | netapp.ports.eth.get | |
NetApp AFF A700: Get FC ports | HTTP agent | netapp.ports.fc.get | |
NetApp AFF A700: Get SVMs | HTTP agent | netapp.svms.get | |
NetApp AFF A700: Get LUNs | HTTP agent | netapp.luns.get | |
NetApp AFF A700: Get chassis | HTTP agent | netapp.chassis.get | |
NetApp AFF A700: Get FRUs | HTTP agent | netapp.frus.get Preprocessing
|
|
NetApp AFF A700: Cluster software version | This returns the cluster version information. When the cluster has more than one node, the cluster version is equivalent to the lowest of generation, major, and minor versions on all nodes. |
Dependent item | netapp.cluster.version Preprocessing
|
NetApp AFF A700: Cluster name | The name of the cluster. |
Dependent item | netapp.cluster.name Preprocessing
|
NetApp AFF A700: Cluster location | The location of the cluster. |
Dependent item | netapp.cluster.location Preprocessing
|
NetApp AFF A700: Cluster status | The status of the cluster: ok, error, partialnodata, partialnoresponse, partialothererror, negativedelta, backfilleddata, inconsistentdeltatime, inconsistentolddata. |
Dependent item | netapp.cluster.status Preprocessing
|
NetApp AFF A700: Cluster throughput, other rate | Throughput bytes observed at the storage object. Performance metric for other I/O operations. Other I/O operations can be metadata operations, such as directory lookups and so on. |
Dependent item | netapp.cluster.statistics.throughput.other.rate Preprocessing
|
NetApp AFF A700: Cluster throughput, read rate | Throughput bytes observed at the storage object. Performance metric for read I/O operations. |
Dependent item | netapp.cluster.statistics.throughput.read.rate Preprocessing
|
NetApp AFF A700: Cluster throughput, write rate | Throughput bytes observed at the storage object. Performance metric for write I/O operations. |
Dependent item | netapp.cluster.statistics.throughput.write.rate Preprocessing
|
NetApp AFF A700: Cluster throughput, total rate | Throughput bytes observed at the storage object. Performance metric aggregated over all types of I/O operations. |
Dependent item | netapp.cluster.statistics.throughput.total.rate Preprocessing
|
NetApp AFF A700: Cluster IOPS, other rate | The number of I/O operations observed at the storage object. Performance metric for other I/O operations. Other I/O operations can be metadata operations, such as directory lookups and so on. |
Dependent item | netapp.cluster.statistics.iops.other.rate Preprocessing
|
NetApp AFF A700: Cluster IOPS, read rate | The number of I/O operations observed at the storage object. Performance metric for read I/O operations. |
Dependent item | netapp.cluster.statistics.iops.read.rate Preprocessing
|
NetApp AFF A700: Cluster IOPS, write rate | The number of I/O operations observed at the storage object. Performance metric for write I/O operations. |
Dependent item | netapp.cluster.statistics.iops.write.rate Preprocessing
|
NetApp AFF A700: Cluster IOPS, total rate | The number of I/O operations observed at the storage object. Performance metric aggregated over all types of I/O operations. |
Dependent item | netapp.cluster.statistics.iops.total.rate Preprocessing
|
NetApp AFF A700: Cluster latency, other | The average latency per I/O operation in milliseconds observed at the storage object. Performance metric for other I/O operations. Other I/O operations can be metadata operations, such as directory lookups and so on. |
Calculated | netapp.cluster.statistics.latency.other |
NetApp AFF A700: Cluster latency, read | The average latency per I/O operation in milliseconds observed at the storage object. Performance metric for read I/O operations. |
Calculated | netapp.cluster.statistics.latency.read |
NetApp AFF A700: Cluster latency, write | The average latency per I/O operation in milliseconds observed at the storage object. Performance metric for write I/O operations. |
Calculated | netapp.cluster.statistics.latency.write |
NetApp AFF A700: Cluster latency, total | The average latency per I/O operation in milliseconds observed at the storage object. Performance metric aggregated over all types of I/O operations. |
Calculated | netapp.cluster.statistics.latency.total |
NetApp AFF A700: Cluster latency raw, other | The raw latency in microseconds observed at the storage object. This can be divided by the raw IOPS value to calculate the average latency per I/O operation. Performance metric for other I/O operations. Other I/O operations can be metadata operations, such as directory lookups and so on. |
Dependent item | netapp.cluster.statistics.latency_raw.other Preprocessing
|
NetApp AFF A700: Cluster latency raw, read | The raw latency in microseconds observed at the storage object. This can be divided by the raw IOPS value to calculate the average latency per I/O operation. Performance metric for read I/O operations. |
Dependent item | netapp.cluster.statistics.latency_raw.read Preprocessing
|
NetApp AFF A700: Cluster latency raw, write | The raw latency in microseconds observed at the storage object. This can be divided by the raw IOPS value to calculate the average latency per I/O operation. Performance metric for write I/O operations. |
Dependent item | netapp.cluster.statistics.latency_raw.write Preprocessing
|
NetApp AFF A700: Cluster latency raw, total | The raw latency in microseconds observed at the storage object. This can be divided by the raw IOPS value to calculate the average latency per I/O operation. Performance metric aggregated over all types of I/O operations. |
Dependent item | netapp.cluster.statistics.latency_raw.total Preprocessing
|
NetApp AFF A700: Cluster IOPS raw, other | The number of I/O operations observed at the storage object. Performance metric for other I/O operations. Other I/O operations can be metadata operations, such as directory lookups and so on. |
Dependent item | netapp.cluster.statistics.iops_raw.other Preprocessing
|
NetApp AFF A700: Cluster IOPS raw, read | The number of I/O operations observed at the storage object. Performance metric for read I/O operations. |
Dependent item | netapp.cluster.statistics.iops_raw.read Preprocessing
|
NetApp AFF A700: Cluster IOPS raw, write | The number of I/O operations observed at the storage object. Performance metric for write I/O operations. |
Dependent item | netapp.cluster.statistics.iops_raw.write Preprocessing
|
NetApp AFF A700: Cluster IOPS raw, total | The number of I/O operations observed at the storage object. Performance metric aggregated over all types of I/O operations. |
Dependent item | netapp.cluster.statistics.iops_raw.total Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
NetApp AFF A700: Version has changed | The NetApp AFF A700 version has changed. Acknowledge to close the problem manually. |
last(/NetApp AFF A700 by HTTP/netapp.cluster.version,#1)<>last(/NetApp AFF A700 by HTTP/netapp.cluster.version,#2) and length(last(/NetApp AFF A700 by HTTP/netapp.cluster.version))>0 |Info |
Manual close: Yes | |
NetApp AFF A700: Cluster status is abnormal | Any errors associated with the sample. For example, if the aggregation of data over multiple nodes fails then any of the partial errors might be returned, “ok” on success, or “error” on any internal uncategorized failure. Whenever a sample collection is missed but done at a later time, it is back filled to the previous 15 second timestamp and tagged with "backfilleddata". “Inconsistent deltatime” is encountered when the time between two collections is not the same for all nodes. Therefore, the aggregated value might be over or under inflated. “Negativedelta” is returned when an expected monotonically increasing value has decreased in value. “Inconsistentolddata” is returned when one or more nodes do not have the latest data. |
(last(/NetApp AFF A700 by HTTP/netapp.cluster.status)<>"ok") |Average |
Name | Description | Type | Key and additional info |
---|---|---|---|
Nodes discovery | HTTP agent | netapp.nodes.discovery Preprocessing
|
Name | Description | Type | Key and additional info |
---|---|---|---|
{#NODENAME}: Software version | This returns the cluster version information. When the cluster has more than one node, the cluster version is equivalent to the lowest of generation, major, and minor versions on all nodes. |
Dependent item | netapp.node.version[{#NODENAME}] Preprocessing
|
{#NODENAME}: Location | The location of the node. |
Dependent item | netapp.nodes.location[{#NODENAME}] Preprocessing
|
{#NODENAME}: State | State of the node: up - Node is up and operational. booting - Node is booting up. down - Node has stopped or is dumping core. takenover - Node has been taken over by its HA partner and is not yet waiting for giveback. waitingfor_giveback - Node has been taken over by its HA partner and is waiting for the HA partner to giveback disks. degraded - Node has one or more critical services offline. unknown - Node or its HA partner cannot be contacted and there is no information on the node's state. |
Dependent item | netapp.nodes.state[{#NODENAME}] Preprocessing
|
{#NODENAME}: Membership | Possible values: available - If a node is available, this means it is detected on the internal cluster network and can be added to the cluster. Nodes that have a membership of “available” are not returned when a GET request is called when the cluster exists. A query on the “membership” property for available must be provided to scan for nodes on the cluster network. Nodes that have a membership of “available” are returned automatically before a cluster is created. joining - Joining nodes are in the process of being added to the cluster. The node may be progressing through the steps to become a member or might have failed. The job to add the node or create the cluster provides details on the current progress of the node. member - Nodes that are members have successfully joined the cluster. |
Dependent item | netapp.nodes.membership[{#NODENAME}] Preprocessing
|
{#NODENAME}: Uptime | The total time, in seconds, that the node has been up. |
Dependent item | netapp.nodes.uptime[{#NODENAME}] Preprocessing
|
{#NODENAME}: Controller over temperature | Specifies whether the hardware is currently operating outside of its recommended temperature range. The hardware shuts down if the temperature exceeds critical thresholds. Possible values: over, normal |
Dependent item | netapp.nodes.controller.over_temperature[{#NODENAME}] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
{#NODENAME}: Version has changed | {#NODENAME} version has changed. Acknowledge to close the problem manually. |
last(/NetApp AFF A700 by HTTP/netapp.node.version[{#NODENAME}],#1)<>last(/NetApp AFF A700 by HTTP/netapp.node.version[{#NODENAME}],#2) and length(last(/NetApp AFF A700 by HTTP/netapp.node.version[{#NODENAME}]))>0 |Info |
Manual close: Yes | |
{#NODENAME}: Node state is abnormal | The state of the node is different from up: |
(last(/NetApp AFF A700 by HTTP/netapp.nodes.state[{#NODENAME}])<>"up") |Average |
||
{#NODENAME}: Node has been restarted | Uptime is less than 10 minutes. |
last(/NetApp AFF A700 by HTTP/netapp.nodes.uptime[{#NODENAME}])<10m |Info |
Manual close: Yes | |
{#NODENAME}: Node has over temperature | The hardware shuts down if the temperature exceeds critical thresholds(item's value is "over"). |
(last(/NetApp AFF A700 by HTTP/netapp.nodes.controller.over_temperature[{#NODENAME}])<>"normal") |Average |
Name | Description | Type | Key and additional info |
---|---|---|---|
Ethernet ports discovery | HTTP agent | netapp.ports.ether.discovery Preprocessing
|
Name | Description | Type | Key and additional info |
---|---|---|---|
{#ETHPORTNAME}: State | The operational state of the port. Possible values: up, down. |
Dependent item | netapp.port.eth.state[{#NODENAME},{#ETHPORTNAME}] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
{#ETHPORTNAME}: Ethernet port of the Node "{#NODENAME}" is down | Something is wrong with the ethernet port. |
(last(/NetApp AFF A700 by HTTP/netapp.port.eth.state[{#NODENAME},{#ETHPORTNAME}],#1)<>last(/NetApp AFF A700 by HTTP/netapp.port.eth.state[{#NODENAME},{#ETHPORTNAME}],#2) and last(/NetApp AFF A700 by HTTP/netapp.port.eth.state[{#NODENAME},{#ETHPORTNAME}])="down") |Average |
Manual close: Yes |
Name | Description | Type | Key and additional info |
---|---|---|---|
FC ports discovery | HTTP agent | netapp.ports.fc.discovery Preprocessing
|
Name | Description | Type | Key and additional info |
---|---|---|---|
{#FCPORTNAME}: Description | A description of the FC port. |
Dependent item | netapp.port.fc.description[{#NODENAME},{#FCPORTNAME}] Preprocessing
|
{#FCPORTNAME}: State | The operational state of the FC port. Possible values: startup - The port is booting up. linknotconnected - The port has finished initialization, but a link with the fabric is not established. online - The port is initialized and a link with the fabric has been established. linkdisconnected - The link was present at one point on this port but is currently not established. offlinedbyuser - The port is administratively disabled. offlinedbysystem - The port is set to offline by the system. This happens when the port encounters too many errors. nodeoffline - The state information for the port cannot be retrieved. The node is offline or inaccessible. |
Dependent item | netapp.port.fc.state[{#NODENAME},{#FCPORTNAME}] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
{#FCPORTNAME}: FC port of the Node "{#NODENAME}" has state different from "online" | Something is wrong with the FC port. |
(last(/NetApp AFF A700 by HTTP/netapp.port.fc.state[{#NODENAME},{#FCPORTNAME}],#1)<>last(/NetApp AFF A700 by HTTP/netapp.port.fc.state[{#NODENAME},{#FCPORTNAME}],#2) and last(/NetApp AFF A700 by HTTP/netapp.port.fc.state[{#NODENAME},{#FCPORTNAME}])<>"online") |Average |
Manual close: Yes |
Name | Description | Type | Key and additional info |
---|---|---|---|
Disks discovery | HTTP agent | netapp.disks.discovery Preprocessing
|
Name | Description | Type | Key and additional info |
---|---|---|---|
{#DISKNAME}: State | The state of the disk. Possible values: broken, copy, maintenance, partner, pending, present, reconstructing, removed, spare, unfail, zeroing |
Dependent item | netapp.disk.state[{#NODENAME},{#DISKNAME}] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
{#DISKNAME}: Disk of the Node "{#NODENAME}" has state different from "present" | Something is wrong with the disk. |
(last(/NetApp AFF A700 by HTTP/netapp.disk.state[{#NODENAME},{#DISKNAME}],#1)<>last(/NetApp AFF A700 by HTTP/netapp.disk.state[{#NODENAME},{#DISKNAME}],#2) and last(/NetApp AFF A700 by HTTP/netapp.disk.state[{#NODENAME},{#DISKNAME}])<>"present") |Average |
Manual close: Yes |
Name | Description | Type | Key and additional info |
---|---|---|---|
Chassis discovery | HTTP agent | netapp.chassis.discovery Preprocessing
|
Name | Description | Type | Key and additional info |
---|---|---|---|
{#ID}: State | The chassis state: ok, error. |
Dependent item | netapp.chassis.state[{#ID}] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
{#ID}: Chassis has something errors | Something is wrong with the chassis. |
(last(/NetApp AFF A700 by HTTP/netapp.chassis.state[{#ID}],#1)<>last(/NetApp AFF A700 by HTTP/netapp.chassis.state[{#ID}],#2) and last(/NetApp AFF A700 by HTTP/netapp.chassis.state[{#ID}])="error") |Average |
Manual close: Yes |
Name | Description | Type | Key and additional info |
---|---|---|---|
FRUs discovery | Dependent item | netapp.frus.discovery Preprocessing
|
Name | Description | Type | Key and additional info |
---|---|---|---|
{#FRUID}: State | The FRU state: ok, error. |
Dependent item | netapp.chassis.fru.state[{#CHASSISID},{#FRUID}] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
{#FRUID}: FRU of the chassis "{#ID}" state is error | Something is wrong with the FRU. |
(last(/NetApp AFF A700 by HTTP/netapp.chassis.fru.state[{#CHASSISID},{#FRUID}],#1)<>last(/NetApp AFF A700 by HTTP/netapp.chassis.fru.state[{#CHASSISID},{#FRUID}],#2) and last(/NetApp AFF A700 by HTTP/netapp.chassis.fru.state[{#CHASSISID},{#FRUID}])="error") |Average |
Manual close: Yes |
Name | Description | Type | Key and additional info |
---|---|---|---|
SVMs discovery | HTTP agent | netapp.svms.discovery Preprocessing
|
Name | Description | Type | Key and additional info |
---|---|---|---|
{#SVMNAME}: State | SVM state: starting, running, stopping, stopped, deleting. |
Dependent item | netapp.svm.state[{#SVMNAME}] Preprocessing
|
{#SVMNAME}: Comment | The comment for the SVM. |
Dependent item | netapp.svm.comment[{#SVMNAME}] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
{#SVMNAME}: SVM state is abnormal | Something is wrong with the SVM. |
(last(/NetApp AFF A700 by HTTP/netapp.svm.state[{#SVMNAME}],#1)<>last(/NetApp AFF A700 by HTTP/netapp.svm.state[{#SVMNAME}],#2) and last(/NetApp AFF A700 by HTTP/netapp.svm.state[{#SVMNAME}])<>"running") |Average |
Manual close: Yes |
Name | Description | Type | Key and additional info |
---|---|---|---|
LUNs discovery | HTTP agent | netapp.luns.discovery Preprocessing
|
Name | Description | Type | Key and additional info |
---|---|---|---|
{#LUNNAME}: State | The state of the LUN. Normal states for a LUN are online and offline. Other states indicate errors. Possible values: foreignlunerror, nvfail, offline, online, space_error. |
Dependent item | netapp.lun.status.state[{#SVMNAME},{#LUNNAME}] Preprocessing
|
{#LUNNAME}: Container state | The state of the volume and aggregate that contain the LUN: online, aggregateoffline, volumeoffline. LUNs are only available when their containers are available. |
Dependent item | netapp.lun.status.container_state[{#SVMNAME},{#LUNNAME}] Preprocessing
|
{#LUNNAME}: Space size | The total provisioned size of the LUN. |
Dependent item | netapp.lun.space.size[{#SVMNAME},{#LUNNAME}] Preprocessing
|
{#LUNNAME}: Space used | The amount of space consumed by the main data stream of the LUN. |
Dependent item | netapp.lun.space.used[{#SVMNAME},{#LUNNAME}] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
{#LUNNAME}: LUN of the SVM "{#SVMNAME}" has abnormal state | Normal states for a LUN are online and offline. Other states indicate errors. |
(last(/NetApp AFF A700 by HTTP/netapp.lun.status.state[{#SVMNAME},{#LUNNAME}],#1)<>last(/NetApp AFF A700 by HTTP/netapp.lun.status.state[{#SVMNAME},{#LUNNAME}],#2) and last(/NetApp AFF A700 by HTTP/netapp.lun.status.state[{#SVMNAME},{#LUNNAME}])<>"online") |Average |
Manual close: Yes | |
{#LUNNAME}: LUN of the SVM "{#SVMNAME}" has abnormal container state | LUNs are only available when their containers are available. |
(last(/NetApp AFF A700 by HTTP/netapp.lun.status.container_state[{#SVMNAME},{#LUNNAME}],#1)<>last(/NetApp AFF A700 by HTTP/netapp.lun.status.container_state[{#SVMNAME},{#LUNNAME}],#2) and last(/NetApp AFF A700 by HTTP/netapp.lun.status.container_state[{#SVMNAME},{#LUNNAME}])<>"online") |Average |
Manual close: Yes |
Name | Description | Type | Key and additional info |
---|---|---|---|
Volumes discovery | HTTP agent | netapp.volumes.discovery Preprocessing
|
Name | Description | Type | Key and additional info |
---|---|---|---|
{#VOLUMENAME}: Comment | A comment for the volume. |
Dependent item | netapp.volume.comment[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: State | Volume state. A volume can only be brought online if it is offline. Taking a volume offline removes its junction path. The 'mixed' state applies to FlexGroup volumes only and cannot be specified as a target state. An 'error' state implies that the volume is not in a state to serve data. |
Dependent item | netapp.volume.state[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: Type | Type of the volume. rw - read-write volume. dp - data-protection volume. ls - load-sharing dp volume. |
Dependent item | netapp.volume.type[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: SVM name | The volume belongs this SVM. |
Dependent item | netapp.volume.svm_name[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: Space size | Total provisioned size. The default size is equal to the minimum size of 20MB, in bytes. |
Dependent item | netapp.volume.space_size[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: Available size | The available space, in bytes. |
Dependent item | netapp.volume.space_available[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: Used size | The virtual space used (includes volume reserves) before storage efficiency, in bytes. |
Dependent item | netapp.volume.space_used[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: Volume throughput, other rate | Throughput bytes observed at the storage object. Performance metric for other I/O operations. Other I/O operations can be metadata operations, such as directory lookups and so on. |
Dependent item | netapp.volume.statistics.throughput.other.rate[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: Volume throughput, read rate | Throughput bytes observed at the storage object. Performance metric for read I/O operations. |
Dependent item | netapp.volume.statistics.throughput.read.rate[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: Volume throughput, write rate | Throughput bytes observed at the storage object. Performance metric for write I/O operations. |
Dependent item | netapp.volume.statistics.throughput.write.rate[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: Volume throughput, total rate | Throughput bytes observed at the storage object. Performance metric aggregated over all types of I/O operations. |
Dependent item | netapp.volume.statistics.throughput.total.rate[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: Volume IOPS, other rate | The number of I/O operations observed at the storage object. Performance metric for other I/O operations. Other I/O operations can be metadata operations, such as directory lookups and so on. |
Dependent item | netapp.volume.statistics.iops.other.rate[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: Volume IOPS, read rate | The number of I/O operations observed at the storage object. Performance metric for read I/O operations. |
Dependent item | netapp.volume.statistics.iops.read.rate[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: Volume IOPS, write rate | The number of I/O operations observed at the storage object. Performance metric for write I/O operations. |
Dependent item | netapp.volume.statistics.iops.write.rate[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: Volume IOPS, total rate | The number of I/O operations observed at the storage object. Performance metric aggregated over all types of I/O operations. |
Dependent item | netapp.volume.statistics.iops.total.rate[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: Volume latency, other | The average latency per I/O operation in milliseconds observed at the storage object. Performance metric for other I/O operations. Other I/O operations can be metadata operations, such as directory lookups and so on. |
Calculated | netapp.volume.statistics.latency.other[{#VOLUMENAME}] |
{#VOLUMENAME}: Volume latency, read | The average latency per I/O operation in milliseconds observed at the storage object. Performance metric for read I/O operations. |
Calculated | netapp.volume.statistics.latency.read[{#VOLUMENAME}] |
{#VOLUMENAME}: Volume latency, write | The average latency per I/O operation in milliseconds observed at the storage object. Performance metric for write I/O operations. |
Calculated | netapp.volume.statistics.latency.write[{#VOLUMENAME}] |
{#VOLUMENAME}: Volume latency, total | The average latency per I/O operation in milliseconds observed at the storage object. Performance metric aggregated over all types of I/O operations. |
Calculated | netapp.volume.statistics.latency.total[{#VOLUMENAME}] |
{#VOLUMENAME}: Volume latency raw, other | The raw latency in microseconds observed at the storage object. Performance metric for other I/O operations. Other I/O operations can be metadata operations, such as directory lookups and so on. |
Dependent item | netapp.volume.statistics.latency_raw.other[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: Volume latency raw, read | The raw latency in microseconds observed at the storage object. Performance metric for read I/O operations. |
Dependent item | netapp.volume.statistics.latency_raw.read[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: Volume latency raw, write | The raw latency in microseconds observed at the storage object. Performance metric for write I/O operations. |
Dependent item | netapp.volume.statistics.latency_raw.write[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: Volume latency raw, total | The raw latency in microseconds observed at the storage object. Performance metric aggregated over all types of I/O operations. |
Dependent item | netapp.volume.statistics.latency_raw.total[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: Volume IOPS raw, other | The number of I/O operations observed at the storage object. Performance metric for other I/O operations. Other I/O operations can be metadata operations, such as directory lookups and so on. |
Dependent item | netapp.volume.statistics.iops_raw.other[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: Volume IOPS raw, read | The number of I/O operations observed at the storage object. Performance metric for read I/O operations. |
Dependent item | netapp.volume.statistics.iops_raw.read[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: Volume IOPS raw, write | The number of I/O operations observed at the storage object. Performance metric for write I/O operations. |
Dependent item | netapp.volume.statistics.iops_raw.write[{#VOLUMENAME}] Preprocessing
|
{#VOLUMENAME}: Volume IOPS raw, total | The number of I/O operations observed at the storage object. Performance metric aggregated over all types of I/O operations. |
Dependent item | netapp.volume.statistics.iops_raw.total[{#VOLUMENAME}] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
{#VOLUMENAME}: Volume state is abnormal | A volume can only be brought online if it is offline. Taking a volume offline removes its junction path. The 'mixed' state applies to FlexGroup volumes only and cannot be specified as a target state. An 'error' state implies that the volume is not in a state to serve data. |
(last(/NetApp AFF A700 by HTTP/netapp.volume.state[{#VOLUMENAME}],#1)<>last(/NetApp AFF A700 by HTTP/netapp.volume.state[{#VOLUMENAME}],#2) and last(/NetApp AFF A700 by HTTP/netapp.volume.state[{#VOLUMENAME}])<>"online") |Average |
Manual close: Yes |
Please report any issues with the template at https://support.zabbix.com
You can also provide feedback, discuss the template, or ask for help at ZABBIX forums
This template is developed to monitor SAN Huawei OceanStor V6 via the Zabbix SNMP agent.
Zabbix version: 6.0 and higher.
This template has been tested on:
Zabbix should be configured according to the instructions in the Templates out of the box section.
1. Create a host for Huawei OceanStor V6 with the controller management IP as the SNMP interface.
2. Link the template to the host.
3. Customize macro values if needed.
Name | Description | Default |
---|---|---|
{$CPU.UTIL.CRIT} | The critical threshold of CPU utilization expressed in %. |
90 |
{$HUAWEI.OCEANSTOR_V6.MEM.MAX.WARN} | Maximum percentage of memory used. |
90 |
{$HUAWEI.OCEANSTOR_V6.MEM.MAX.TIME} | The time during which memory usage may exceed the threshold. |
5m |
{$HUAWEI.OCEANSTOR_V6.TEMP.MAX.WARN} | Maximum enclosure temperature. |
35 |
{$HUAWEI.OCEANSTOR_V6.TEMP.MAX.TIME} | The time during which the enclosure temperature may exceed the threshold. |
3m |
{$HUAWEI.OCEANSTOR_V6.DISK.TEMP.MAX.WARN} | Maximum disk temperature. Can be used with |
45 |
{$HUAWEI.OCEANSTOR_V6.DISK.TEMP.MAX.TIME} | The time during which the disk temperature may exceed the threshold. |
5m |
{$HUAWEI.OCEANSTOR_V6.LUN.IO.TIME.MAX.WARN} | Maximum average I/O response time of a LUN in seconds. |
0.0001 |
{$HUAWEI.OCEANSTOR_V6.LUN.IO.TIME.MAX.TIME} | The time during which the average I/O response time of a LUN may exceed the threshold. |
5m |
{$SNMP.TIMEOUT} | 5m |
|
{$ICMPLOSSWARN} | 20 |
|
{$ICMPRESPONSETIME_WARN} | 0.15 |
Name | Description | Type | Key and additional info |
---|---|---|---|
Huawei OceanStor V6: Status | System status. |
SNMP agent | huawei.oceanstor.v6.status Preprocessing
|
Huawei OceanStor V6: Version | The device version. |
SNMP agent | huawei.oceanstor.v6.version Preprocessing
|
Huawei OceanStor V6: Capacity total | Total capacity of a device. |
SNMP agent | huawei.oceanstor.v6.capacity.total Preprocessing
|
Huawei OceanStor V6: Capacity used | Used capacity of a device. |
SNMP agent | huawei.oceanstor.v6.capacity.used Preprocessing
|
Huawei OceanStor V6: Uptime (network) | MIB: SNMPv2-MIB The time (in hundredths of a second) since the network management portion of the system was last re-initialized. |
SNMP agent | system.net.uptime[sysUpTime.0] Preprocessing
|
Huawei OceanStor V6: Uptime (hardware) | MIB: HOST-RESOURCES-MIB The amount of time since this host was last initialized. Note that this is different from sysUpTime in the SNMPv2-MIB [RFC1907] because sysUpTime is the uptime of the network management portion of the system. |
SNMP agent | system.hw.uptime[hrSystemUptime.0] Preprocessing
|
Huawei OceanStor V6: SNMP traps (fallback) | The item is used to collect all SNMP traps unmatched by other snmptrap items |
SNMP trap | snmptrap.fallback |
Huawei OceanStor V6: System location | MIB: SNMPv2-MIB The physical location of this node (e.g., `telephone closet, 3rd floor'). If the location is unknown, the value is the zero-length string. |
SNMP agent | system.location[sysLocation.0] Preprocessing
|
Huawei OceanStor V6: System contact details | MIB: SNMPv2-MIB The textual identification of the contact person for this managed node, together with information on how to contact this person. If no contact information is known, the value is the zero-length string. |
SNMP agent | system.contact[sysContact.0] Preprocessing
|
Huawei OceanStor V6: System object ID | MIB: SNMPv2-MIB The vendor's authoritative identification of the network management subsystem contained in the entity. This value is allocated within the SMI enterprises subtree (1.3.6.1.4.1) and provides an easy and unambiguous means for determining |
SNMP agent | system.objectid[sysObjectID.0] Preprocessing
|
Huawei OceanStor V6: System name | MIB: SNMPv2-MIB An administratively-assigned name for this managed node.By convention, this is the node's fully-qualified domain name. If the name is unknown, the value is the zero-length string. |
SNMP agent | system.name Preprocessing
|
Huawei OceanStor V6: System description | MIB: SNMPv2-MIB A textual description of the entity. This value should include the full name and version identification of the system's hardware type, software operating-system, and networking software. |
SNMP agent | system.descr[sysDescr.0] Preprocessing
|
Huawei OceanStor V6: SNMP agent availability | Availability of SNMP checks on the host. The value of this item corresponds to availability icons in the host list. Possible value: 0 - not available 1 - available 2 - unknown |
Zabbix internal | zabbix[host,snmp,available] |
Huawei OceanStor V6: ICMP ping | Simple check | icmpping | |
Huawei OceanStor V6: ICMP loss | Simple check | icmppingloss | |
Huawei OceanStor V6: ICMP response time | Simple check | icmppingsec |
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Huawei OceanStor V6: Storage version has changed | last(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.version,#1)<>last(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.version,#2) and length(last(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.version))>0 |Info |
Manual close: Yes | ||
Huawei OceanStor V6: Host has been restarted | Uptime is less than 10 minutes. |
(last(/Huawei OceanStor V6 by SNMP/system.hw.uptime[hrSystemUptime.0])>0 and last(/Huawei OceanStor V6 by SNMP/system.hw.uptime[hrSystemUptime.0])<10m) or (last(/Huawei OceanStor V6 by SNMP/system.hw.uptime[hrSystemUptime.0])=0 and last(/Huawei OceanStor V6 by SNMP/system.net.uptime[sysUpTime.0])<10m) |Warning |
Manual close: Yes Depends on:
|
|
Huawei OceanStor V6: System name has changed | The name of the system has changed. Acknowledge to close the problem manually. |
last(/Huawei OceanStor V6 by SNMP/system.name,#1)<>last(/Huawei OceanStor V6 by SNMP/system.name,#2) and length(last(/Huawei OceanStor V6 by SNMP/system.name))>0 |Info |
Manual close: Yes | |
Huawei OceanStor V6: No SNMP data collection | SNMP is not available for polling. Please check device connectivity and SNMP settings. |
max(/Huawei OceanStor V6 by SNMP/zabbix[host,snmp,available],{$SNMP.TIMEOUT})=0 |Warning |
Depends on:
|
|
Huawei OceanStor V6: Unavailable by ICMP ping | Last three attempts returned timeout. Please check device connectivity. |
max(/Huawei OceanStor V6 by SNMP/icmpping,#3)=0 |High |
||
Huawei OceanStor V6: High ICMP ping loss | min(/Huawei OceanStor V6 by SNMP/icmppingloss,5m)>{$ICMP_LOSS_WARN} and min(/Huawei OceanStor V6 by SNMP/icmppingloss,5m)<100 |Warning |
Depends on:
|
||
Huawei OceanStor V6: High ICMP ping response time | avg(/Huawei OceanStor V6 by SNMP/icmppingsec,5m)>{$ICMP_RESPONSE_TIME_WARN} |Warning |
Depends on:
|
Name | Description | Type | Key and additional info |
---|---|---|---|
Controller discovery | Discovery of controllers. |
SNMP agent | huawei.oceanstor.v6.controller.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Huawei OceanStor V6: Controller [{#ID}]: CPU utilization | CPU utilization of the controller. |
SNMP agent | huawei.oceanstor.v6.controller.cpu["{#ID}"] |
Huawei OceanStor V6: Controller [{#ID}]: Memory utilization | Memory utilization of the controller. |
SNMP agent | huawei.oceanstor.v6.controller.memory["{#ID}"] |
Huawei OceanStor V6: Controller [{#ID}]: Health status | Controller health status. |
SNMP agent | huawei.oceanstor.v6.controller.health_status["{#ID}"] Preprocessing
|
Huawei OceanStor V6: Controller [{#ID}]: Running status | Controller running status. |
SNMP agent | huawei.oceanstor.v6.controller.running_status["{#ID}"] Preprocessing
|
Huawei OceanStor V6: Controller [{#ID}]: Role | Controller role. |
SNMP agent | huawei.oceanstor.v6.controller.role["{#ID}"] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Huawei OceanStor V6: Controller [{#ID}]: High CPU utilization | The CPU utilization is too high. The system might be slow to respond. |
min(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.controller.cpu["{#ID}"],5m)>{$CPU.UTIL.CRIT} |Warning |
||
Huawei OceanStor V6: Controller [{#ID}]: Memory usage is too high | min(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.controller.memory["{#ID}"],{$HUAWEI.OCEANSTOR_V6.MEM.MAX.TIME})>{$HUAWEI.OCEANSTOR_V6.MEM.MAX.WARN} |Average |
|||
Huawei OceanStor V6: Controller [{#ID}]: Health status is not Normal | last(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.controller.health_status["{#ID}"])<>1 |High |
|||
Huawei OceanStor V6: Controller [{#ID}]: Running status is not Online | last(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.controller.running_status["{#ID}"])<>27 |Average |
|||
Huawei OceanStor V6: Controller [{#ID}]: Role has been changed | last(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.controller.role["{#ID}"],#1)<>last(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.controller.role["{#ID}"],#2) |Warning |
Manual close: Yes |
Name | Description | Type | Key and additional info |
---|---|---|---|
Enclosure discovery | Discovery of enclosures. |
SNMP agent | huawei.oceanstor.v6.enclosure.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Huawei OceanStor V6: Enclosure [{#NAME}]: Health status | Enclosure health status. |
SNMP agent | huawei.oceanstor.v6.enclosure.health_status["{#NAME}"] Preprocessing
|
Huawei OceanStor V6: Enclosure [{#NAME}]: Running status | Enclosure running status. |
SNMP agent | huawei.oceanstor.v6.enclosure.running_status["{#NAME}"] Preprocessing
|
Huawei OceanStor V6: Enclosure [{#NAME}]: Temperature | Enclosure temperature. |
SNMP agent | huawei.oceanstor.v6.enclosure.temperature["{#NAME}"] |
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Huawei OceanStor V6: Enclosure [{#NAME}]: Health status is not Normal | last(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.enclosure.health_status["{#NAME}"])<>1 |High |
|||
Huawei OceanStor V6: Enclosure [{#NAME}]: Running status is not Online | last(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.enclosure.running_status["{#NAME}"])<>27 |Average |
|||
Huawei OceanStor V6: Enclosure [{#NAME}]: Temperature is too high | min(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.enclosure.temperature["{#NAME}"],{$HUAWEI.OCEANSTOR_V6.TEMP.MAX.TIME})>{$HUAWEI.OCEANSTOR_V6.TEMP.MAX.WARN} |High |
Name | Description | Type | Key and additional info |
---|---|---|---|
Fan discovery | Discovery of fans. |
SNMP agent | huawei.oceanstor.v6.fan.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Huawei OceanStor V6: Fan [{#ID}] on [{#LOCATION}]: Health status | Health status of a fan. |
SNMP agent | huawei.oceanstor.v6.fan.health_status["{#ID}:{#LOCATION}"] Preprocessing
|
Huawei OceanStor V6: Fan [{#ID}] on [{#LOCATION}]: Running status | Operating status of a fan. |
SNMP agent | huawei.oceanstor.v6.fan.running_status["{#ID}:{#LOCATION}"] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Huawei OceanStor V6: Fan [{#ID}] on [{#LOCATION}]: Health status is not Normal | last(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.fan.health_status["{#ID}:{#LOCATION}"])<>1 |High |
|||
Huawei OceanStor V6: Fan [{#ID}] on [{#LOCATION}]: Running status is not Running | last(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.fan.running_status["{#ID}:{#LOCATION}"])<>2 |Average |
Name | Description | Type | Key and additional info |
---|---|---|---|
BBU discovery | Discovery of BBUs. |
SNMP agent | huawei.oceanstor.v6.bbu.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Huawei OceanStor V6: BBU [{#ID}] on [{#LOCATION}]: Health status | Health status of a BBU. |
SNMP agent | huawei.oceanstor.v6.bbu.health_status["{#ID}:{#LOCATION}"] Preprocessing
|
Huawei OceanStor V6: BBU [{#ID}] on [{#LOCATION}]: Running status | Running status of a BBU. |
SNMP agent | huawei.oceanstor.v6.bbu.running_status["{#ID}:{#LOCATION}"] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Huawei OceanStor V6: BBU [{#ID}] on [{#LOCATION}]: Health status is not Normal | last(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.bbu.health_status["{#ID}:{#LOCATION}"])<>1 |High |
|||
Huawei OceanStor V6: BBU [{#ID}] on [{#LOCATION}]: Running status is not Online | last(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.bbu.running_status["{#ID}:{#LOCATION}"])<>27 |Average |
Name | Description | Type | Key and additional info |
---|---|---|---|
Disk discovery | Discovery of disks. |
SNMP agent | huawei.oceanstor.v6.disk.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Huawei OceanStor V6: Disk [{#MODEL}] on [{#LOCATION}]: Health status | Disk health status. |
SNMP agent | huawei.oceanstor.v6.disk.health_status["{#ID}"] Preprocessing
|
Huawei OceanStor V6: Disk [{#MODEL}] on [{#LOCATION}]: Running status | Disk running status. |
SNMP agent | huawei.oceanstor.v6.disk.running_status["{#ID}"] Preprocessing
|
Huawei OceanStor V6: Disk [{#MODEL}] on [{#LOCATION}]: Temperature | Disk temperature. |
SNMP agent | huawei.oceanstor.v6.disk.temperature["{#ID}"] |
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Huawei OceanStor V6: Disk [{#MODEL}] on [{#LOCATION}]: Health status is not Normal | last(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.disk.health_status["{#ID}"])<>1 |High |
|||
Huawei OceanStor V6: Disk [{#MODEL}] on [{#LOCATION}]: Running status is not Online | last(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.disk.running_status["{#ID}"])<>27 |Average |
|||
Huawei OceanStor V6: Disk [{#MODEL}] on [{#LOCATION}]: Temperature is too high | min(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.disk.temperature["{#ID}"],{$HUAWEI.OCEANSTOR_V6.DISK.TEMP.MAX.TIME})>{$HUAWEI.OCEANSTOR_V6.DISK.TEMP.MAX.WARN:"{#MODEL}"} |High |
Name | Description | Type | Key and additional info |
---|---|---|---|
Node performance discovery | Discovery of node performance counters. |
SNMP agent | huawei.oceanstor.v6.node.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Huawei OceanStor V6: Node [{#NODE}]: CPU utilization | CPU utilization of the node. |
SNMP agent | huawei.oceanstor.v6.node.cpu["{#NODE}"] |
Huawei OceanStor V6: Node [{#NODE}]: Total I/O per second | Total IOPS of the node. |
SNMP agent | huawei.oceanstor.v6.node.iops.total["{#NODE}"] |
Huawei OceanStor V6: Node [{#NODE}]: Read operations per second | Read IOPS of the node. |
SNMP agent | huawei.oceanstor.v6.node.iops.read["{#NODE}"] |
Huawei OceanStor V6: Node [{#NODE}]: Write operations per second | Write IOPS of the node. |
SNMP agent | huawei.oceanstor.v6.node.iops.write["{#NODE}"] |
Huawei OceanStor V6: Node [{#NODE}]: Total traffic per second | Total bandwidth for the node. |
SNMP agent | huawei.oceanstor.v6.node.bps.total["{#NODE}"] Preprocessing
|
Huawei OceanStor V6: Node [{#NODE}]: Read traffic per second | Read bandwidth for the node. |
SNMP agent | huawei.oceanstor.v6.node.bps.read["{#NODE}"] Preprocessing
|
Huawei OceanStor V6: Node [{#NODE}]: Write traffic per second | Write bandwidth for the node. |
SNMP agent | huawei.oceanstor.v6.node.bps.write["{#NODE}"] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Huawei OceanStor V6: Node [{#NODE}]: High CPU utilization | The CPU utilization is too high. The system might be slow to respond. |
min(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.node.cpu["{#NODE}"],5m)>{$CPU.UTIL.CRIT} |Warning |
Name | Description | Type | Key and additional info |
---|---|---|---|
LUN discovery | Discovery of LUNs. |
SNMP agent | huawei.oceanstor.v6.lun.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Huawei OceanStor V6: LUN [{#NAME}]: Status | Status of the LUN. |
SNMP agent | huawei.oceanstor.v6.lun.status["{#NAME}"] Preprocessing
|
Huawei OceanStor V6: LUN [{#NAME}]: Average total I/O latency | Average I/O latency of the node. |
SNMP agent | huawei.oceanstor.v6.lun.latency.total["{#NAME}"] Preprocessing
|
Huawei OceanStor V6: LUN [{#NAME}]: Average read I/O latency | Average read I/O response time. |
SNMP agent | huawei.oceanstor.v6.lun.latency.read["{#NAME}"] Preprocessing
|
Huawei OceanStor V6: LUN [{#NAME}]: Average write I/O latency | Average write I/O response time. |
SNMP agent | huawei.oceanstor.v6.lun.latency.write["{#NAME}"] Preprocessing
|
Huawei OceanStor V6: LUN [{#NAME}]: Total I/O per second | Current IOPS of the LUN. |
SNMP agent | huawei.oceanstor.v6.lun.iops.total["{#NAME}"] |
Huawei OceanStor V6: LUN [{#NAME}]: Read operations per second | Read IOPS of the node. |
SNMP agent | huawei.oceanstor.v6.lun.iops.read["{#NAME}"] |
Huawei OceanStor V6: LUN [{#NAME}]: Write operations per second | Write IOPS of the node. |
SNMP agent | huawei.oceanstor.v6.lun.iops.write["{#NAME}"] |
Huawei OceanStor V6: LUN [{#NAME}]: Total traffic per second | Current total bandwidth for the LUN. |
SNMP agent | huawei.oceanstor.v6.lun.bps.total["{#NAME}"] Preprocessing
|
Huawei OceanStor V6: LUN [{#NAME}]: Read traffic per second | Current read bandwidth for the LUN. |
SNMP agent | huawei.oceanstor.v6.lun.bps.read["{#NAME}"] Preprocessing
|
Huawei OceanStor V6: LUN [{#NAME}]: Write traffic per second | Current write bandwidth for the LUN. |
SNMP agent | huawei.oceanstor.v6.lun.bps.write["{#NAME}"] Preprocessing
|
Huawei OceanStor V6: LUN [{#NAME}]: Capacity | Capacity of the LUN. |
SNMP agent | huawei.oceanstor.v6.lun.capacity["{#NAME}"] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Huawei OceanStor V6: LUN [{#NAME}]: Status is not Normal | last(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.lun.status["{#NAME}"])<>1 |Average |
|||
Huawei OceanStor V6: LUN [{#NAME}]: Average I/O response time is too high | min(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.lun.latency.total["{#NAME}"],{$HUAWEI.OCEANSTOR_V6.LUN.IO.TIME.MAX.TIME})>{$HUAWEI.OCEANSTOR_V6.LUN.IO.TIME.MAX.WARN} |Warning |
Name | Description | Type | Key and additional info |
---|---|---|---|
Storage pool discovery | Discovery of storage pools. |
SNMP agent | huawei.oceanstor.v6.pool.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Huawei OceanStor V6: Storage pool [{#NAME}]: Health status | Health status of a storage pool. |
SNMP agent | huawei.oceanstor.v6.pool.health_status["{#NAME}"] Preprocessing
|
Huawei OceanStor V6: Storage pool [{#NAME}]: Running status | Operating status of a storage pool. |
SNMP agent | huawei.oceanstor.v6.pool.running_status["{#NAME}"] Preprocessing
|
Huawei OceanStor V6: Storage pool [{#NAME}]: Capacity total | Total capacity of a storage pool. |
SNMP agent | huawei.oceanstor.v6.pool.capacity.total["{#NAME}"] Preprocessing
|
Huawei OceanStor V6: Storage pool [{#NAME}]: Capacity free | Available capacity of a storage pool. |
SNMP agent | huawei.oceanstor.v6.pool.capacity.free["{#NAME}"] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Huawei OceanStor V6: Storage pool [{#NAME}]: Health status is not Normal | last(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.pool.health_status["{#NAME}"])<>1 |High |
|||
Huawei OceanStor V6: Storage pool [{#NAME}]: Running status is not Online | last(/Huawei OceanStor V6 by SNMP/huawei.oceanstor.v6.pool.running_status["{#NAME}"])<>27 |Average |
Please report any issues with the template at https://support.zabbix.com
You can also provide feedback, discuss the template, or ask for help at ZABBIX forums
The template to monitor SAN Huawei OceanStor 5300 V5 by Zabbix SNMP agent.
Zabbix version: 6.0 and higher.
This template has been tested on:
Zabbix should be configured according to the instructions in the Templates out of the box section.
1. Create a host for Huawei OceanStor 5300 V5 with controller management IP as SNMPv2 interface.
2. Link the template to the host.
3. Customize macro values if needed.
Name | Description | Default |
---|---|---|
{$CPU.UTIL.CRIT} | The critical threshold of the CPU utilization expressed in %. |
90 |
{$HUAWEI.5300.MEM.MAX.WARN} | Maximum percentage of memory used |
90 |
{$HUAWEI.5300.MEM.MAX.TIME} | The time during which memory usage may exceed the threshold. |
5m |
{$HUAWEI.5300.TEMP.MAX.WARN} | Maximum temperature of enclosure |
35 |
{$HUAWEI.5300.TEMP.MAX.TIME} | The time during which temperature of enclosure may exceed the threshold. |
3m |
{$HUAWEI.5300.DISK.TEMP.MAX.WARN} | Maximum temperature of disk. Can be used with {#MODEL} as context. |
45 |
{$HUAWEI.5300.DISK.TEMP.MAX.TIME} | The time during which temperature of disk may exceed the threshold. |
5m |
{$HUAWEI.5300.NODE.IO.DELAY.MAX.WARN} | Maximum average I/O latency of node in milliseconds. |
20 |
{$HUAWEI.5300.NODE.IO.DELAY.MAX.TIME} | The time during which average I/O latency of node may exceed the threshold. |
5m |
{$HUAWEI.5300.LUN.IO.TIME.MAX.WARN} | Maximum average I/O response time of LUN in milliseconds. |
100 |
{$HUAWEI.5300.LUN.IO.TIME.MAX.TIME} | The time during which average I/O response time of LUN may exceed the threshold. |
5m |
{$HUAWEI.5300.POOL.CAPACITY.THRESH.TIME} | The time during which free capacity may exceed the {#THRESHOLD} from hwInfoStoragePoolFullThreshold. |
5m |
{$SNMP.TIMEOUT} | 5m |
|
{$ICMPLOSSWARN} | 20 |
|
{$ICMPRESPONSETIME_WARN} | 0.15 |
Name | Description | Type | Key and additional info |
---|---|---|---|
OceanStor 5300 V5: Status | System running status. |
SNMP agent | huawei.5300.v5[status] Preprocessing
|
OceanStor 5300 V5: Version | The device version. |
SNMP agent | huawei.5300.v5[version] Preprocessing
|
OceanStor 5300 V5: Capacity total | Total capacity of a device. |
SNMP agent | huawei.5300.v5[totalCapacity] Preprocessing
|
OceanStor 5300 V5: Capacity used | Used capacity of a device. |
SNMP agent | huawei.5300.v5[usedCapacity] Preprocessing
|
OceanStor 5300 V5: Uptime (network) | MIB: SNMPv2-MIB The time (in hundredths of a second) since the network management portion of the system was last re-initialized. |
SNMP agent | system.net.uptime[sysUpTime.0] Preprocessing
|
OceanStor 5300 V5: Uptime (hardware) | MIB: HOST-RESOURCES-MIB The amount of time since this host was last initialized. Note that this is different from sysUpTime in the SNMPv2-MIB [RFC1907] because sysUpTime is the uptime of the network management portion of the system. |
SNMP agent | system.hw.uptime[hrSystemUptime.0] Preprocessing
|
OceanStor 5300 V5: SNMP traps (fallback) | The item is used to collect all SNMP traps unmatched by other snmptrap items |
SNMP trap | snmptrap.fallback |
OceanStor 5300 V5: System location | MIB: SNMPv2-MIB The physical location of this node (e.g., `telephone closet, 3rd floor'). If the location is unknown, the value is the zero-length string. |
SNMP agent | system.location[sysLocation.0] Preprocessing
|
OceanStor 5300 V5: System contact details | MIB: SNMPv2-MIB The textual identification of the contact person for this managed node, together with information on how to contact this person. If no contact information is known, the value is the zero-length string. |
SNMP agent | system.contact[sysContact.0] Preprocessing
|
OceanStor 5300 V5: System object ID | MIB: SNMPv2-MIB The vendor's authoritative identification of the network management subsystem contained in the entity. This value is allocated within the SMI enterprises subtree (1.3.6.1.4.1) and provides an easy and unambiguous means for determining |
SNMP agent | system.objectid[sysObjectID.0] Preprocessing
|
OceanStor 5300 V5: System name | MIB: SNMPv2-MIB An administratively-assigned name for this managed node.By convention, this is the node's fully-qualified domain name. If the name is unknown, the value is the zero-length string. |
SNMP agent | system.name Preprocessing
|
OceanStor 5300 V5: System description | MIB: SNMPv2-MIB A textual description of the entity. This value should include the full name and version identification of the system's hardware type, software operating-system, and networking software. |
SNMP agent | system.descr[sysDescr.0] Preprocessing
|
OceanStor 5300 V5: SNMP agent availability | Availability of SNMP checks on the host. The value of this item corresponds to availability icons in the host list. Possible value: 0 - not available 1 - available 2 - unknown |
Zabbix internal | zabbix[host,snmp,available] |
OceanStor 5300 V5: ICMP ping | Simple check | icmpping | |
OceanStor 5300 V5: ICMP loss | Simple check | icmppingloss | |
OceanStor 5300 V5: ICMP response time | Simple check | icmppingsec |
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
OceanStor 5300 V5: Storage version has been changed | OceanStor 5300 V5 version has changed. Acknowledge to close the problem manually. |
last(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[version],#1)<>last(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[version],#2) and length(last(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[version]))>0 |Info |
Manual close: Yes | |
OceanStor 5300 V5: Host has been restarted | Uptime is less than 10 minutes. |
(last(/Huawei OceanStor 5300 V5 by SNMP/system.hw.uptime[hrSystemUptime.0])>0 and last(/Huawei OceanStor 5300 V5 by SNMP/system.hw.uptime[hrSystemUptime.0])<10m) or (last(/Huawei OceanStor 5300 V5 by SNMP/system.hw.uptime[hrSystemUptime.0])=0 and last(/Huawei OceanStor 5300 V5 by SNMP/system.net.uptime[sysUpTime.0])<10m) |Warning |
Manual close: Yes Depends on:
|
|
OceanStor 5300 V5: System name has changed | The name of the system has changed. Acknowledge to close the problem manually. |
last(/Huawei OceanStor 5300 V5 by SNMP/system.name,#1)<>last(/Huawei OceanStor 5300 V5 by SNMP/system.name,#2) and length(last(/Huawei OceanStor 5300 V5 by SNMP/system.name))>0 |Info |
Manual close: Yes | |
OceanStor 5300 V5: No SNMP data collection | SNMP is not available for polling. Please check device connectivity and SNMP settings. |
max(/Huawei OceanStor 5300 V5 by SNMP/zabbix[host,snmp,available],{$SNMP.TIMEOUT})=0 |Warning |
Depends on:
|
|
OceanStor 5300 V5: Unavailable by ICMP ping | Last three attempts returned timeout. Please check device connectivity. |
max(/Huawei OceanStor 5300 V5 by SNMP/icmpping,#3)=0 |High |
||
OceanStor 5300 V5: High ICMP ping loss | min(/Huawei OceanStor 5300 V5 by SNMP/icmppingloss,5m)>{$ICMP_LOSS_WARN} and min(/Huawei OceanStor 5300 V5 by SNMP/icmppingloss,5m)<100 |Warning |
Depends on:
|
||
OceanStor 5300 V5: High ICMP ping response time | avg(/Huawei OceanStor 5300 V5 by SNMP/icmppingsec,5m)>{$ICMP_RESPONSE_TIME_WARN} |Warning |
Depends on:
|
Name | Description | Type | Key and additional info |
---|---|---|---|
Controllers discovery | Discovery of controllers |
SNMP agent | huawei.5300.controllers.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Controller {#ID}: CPU utilization | CPU usage of a controller {#ID}. |
SNMP agent | huawei.5300.v5[hwInfoControllerCPUUsage, "{#ID}"] |
Controller {#ID}: Memory utilization | Memory usage of a controller {#ID}. |
SNMP agent | huawei.5300.v5[hwInfoControllerMemoryUsage, "{#ID}"] |
Controller {#ID}: Health status | Controller health status. For details, see definition of Enum Values (HEALTHSTATUSE). https://support.huawei.com/enterprise/en/centralized-storage/oceanstor-5300-v5-pid-22462029?category=reference-guides&subcategory=mib-reference |
SNMP agent | huawei.5300.v5[hwInfoControllerHealthStatus, "{#ID}"] Preprocessing
|
Controller {#ID}: Running status | Controller running status. For details, see definition of Enum Values (RUNNINGSTATUSE). https://support.huawei.com/enterprise/en/centralized-storage/oceanstor-5300-v5-pid-22462029?category=reference-guides&subcategory=mib-reference |
SNMP agent | huawei.5300.v5[hwInfoControllerRunningStatus, "{#ID}"] Preprocessing
|
Controller {#ID}: Role | Controller role. |
SNMP agent | huawei.5300.v5[hwInfoControllerRole, "{#ID}"] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Controller {#ID}: High CPU utilization | The CPU utilization is too high. The system might be slow to respond. |
min(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwInfoControllerCPUUsage, "{#ID}"],5m)>{$CPU.UTIL.CRIT} |Warning |
||
Controller {#ID}: Memory usage is too high | min(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwInfoControllerMemoryUsage, "{#ID}"],{$HUAWEI.5300.MEM.MAX.TIME})>{$HUAWEI.5300.MEM.MAX.WARN} |Average |
|||
Controller {#ID}: Health status is not Normal | last(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwInfoControllerHealthStatus, "{#ID}"])<>1 |High |
|||
Controller {#ID}: Running status is not Online | last(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwInfoControllerRunningStatus, "{#ID}"])<>27 |Average |
|||
Controller {#ID}: Role has been changed | last(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwInfoControllerRole, "{#ID}"],#1)<>last(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwInfoControllerRole, "{#ID}"],#2) |Warning |
Manual close: Yes |
Name | Description | Type | Key and additional info |
---|---|---|---|
Enclosure discovery | Discovery of enclosures |
SNMP agent | huawei.5300.enclosure.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Enclosure {#NAME}: Health status | Enclosure health status. For details, see definition of Enum Values (HEALTHSTATUSE). https://support.huawei.com/enterprise/en/centralized-storage/oceanstor-5300-v5-pid-22462029?category=reference-guides&subcategory=mib-reference |
SNMP agent | huawei.5300.v5[hwInfoEnclosureHealthStatus, "{#NAME}"] Preprocessing
|
Enclosure {#NAME}: Running status | Enclosure running status. For details, see definition of Enum Values (RUNNINGSTATUSE). https://support.huawei.com/enterprise/en/centralized-storage/oceanstor-5300-v5-pid-22462029?category=reference-guides&subcategory=mib-reference |
SNMP agent | huawei.5300.v5[hwInfoEnclosureRunningStatus, "{#NAME}"] Preprocessing
|
Enclosure {#NAME}: Temperature | Enclosure temperature. |
SNMP agent | huawei.5300.v5[hwInfoEnclosureTemperature, "{#NAME}"] |
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Enclosure {#NAME}: Health status is not Normal | last(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwInfoEnclosureHealthStatus, "{#NAME}"])<>1 |High |
|||
Enclosure {#NAME}: Running status is not Online | last(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwInfoEnclosureRunningStatus, "{#NAME}"])<>27 |Average |
|||
Enclosure {#NAME}: Temperature is too high | min(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwInfoEnclosureTemperature, "{#NAME}"],{$HUAWEI.5300.TEMP.MAX.TIME})>{$HUAWEI.5300.TEMP.MAX.WARN} |High |
Name | Description | Type | Key and additional info |
---|---|---|---|
FANs discovery | Discovery of FANs |
SNMP agent | huawei.5300.fan.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
FAN {#ID} on {#LOCATION}: Health status | Health status of a fan. For details, see definition of Enum Values (HEALTHSTATUSE). https://support.huawei.com/enterprise/en/centralized-storage/oceanstor-5300-v5-pid-22462029?category=reference-guides&subcategory=mib-reference |
SNMP agent | huawei.5300.v5[hwInfoFanHealthStatus, "{#ID}:{#LOCATION}"] Preprocessing
|
FAN {#ID} on {#LOCATION}: Running status | Operating status of a fan. For details, see definition of Enum Values (RUNNINGSTATUSE). https://support.huawei.com/enterprise/en/centralized-storage/oceanstor-5300-v5-pid-22462029?category=reference-guides&subcategory=mib-reference |
SNMP agent | huawei.5300.v5[hwInfoFanRunningStatus, "{#ID}:{#LOCATION}"] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
FAN {#ID} on {#LOCATION}: Health status is not Normal | last(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwInfoFanHealthStatus, "{#ID}:{#LOCATION}"])<>1 |High |
|||
FAN {#ID} on {#LOCATION}: Running status is not Running | last(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwInfoFanRunningStatus, "{#ID}:{#LOCATION}"])<>2 |Average |
Name | Description | Type | Key and additional info |
---|---|---|---|
BBU discovery | Discovery of BBU |
SNMP agent | huawei.5300.bbu.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
BBU {#ID} on {#LOCATION}: Health status | Health status of a BBU. For details, see definition of Enum Values (HEALTHSTATUSE). https://support.huawei.com/enterprise/en/centralized-storage/oceanstor-5300-v5-pid-22462029?category=reference-guides&subcategory=mib-reference |
SNMP agent | huawei.5300.v5[hwInfoBBUHealthStatus, "{#ID}:{#LOCATION}"] Preprocessing
|
BBU {#ID} on {#LOCATION}: Running status | Running status of a BBU. For details, see definition of Enum Values (RUNNINGSTATUSE). https://support.huawei.com/enterprise/en/centralized-storage/oceanstor-5300-v5-pid-22462029?category=reference-guides&subcategory=mib-reference |
SNMP agent | huawei.5300.v5[hwInfoBBURunningStatus, "{#ID}:{#LOCATION}"] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
BBU {#ID} on {#LOCATION}: Health status is not Normal | last(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwInfoBBUHealthStatus, "{#ID}:{#LOCATION}"])<>1 |High |
|||
BBU {#ID} on {#LOCATION}: Running status is not Online | last(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwInfoBBURunningStatus, "{#ID}:{#LOCATION}"])<>2 |Average |
Name | Description | Type | Key and additional info |
---|---|---|---|
Disks discovery | Discovery of disks |
SNMP agent | huawei.5300.disks.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Disk {#MODEL} on {#LOCATION}: Health status | Disk health status. For details, see definition of Enum Values (HEALTHSTATUSE). https://support.huawei.com/enterprise/en/centralized-storage/oceanstor-5300-v5-pid-22462029?category=reference-guides&subcategory=mib-reference |
SNMP agent | huawei.5300.v5[hwInfoDiskHealthStatus, "{#ID}"] Preprocessing
|
Disk {#MODEL} on {#LOCATION}: Running status | Disk running status. For details, see definition of Enum Values (RUNNINGSTATUSE). https://support.huawei.com/enterprise/en/centralized-storage/oceanstor-5300-v5-pid-22462029?category=reference-guides&subcategory=mib-reference |
SNMP agent | huawei.5300.v5[hwInfoDiskRunningStatus, "{#ID}"] Preprocessing
|
Disk {#MODEL} on {#LOCATION}: Temperature | Disk temperature. |
SNMP agent | huawei.5300.v5[hwInfoDiskTemperature, "{#ID}"] |
Disk {#MODEL} on {#LOCATION}: Health score | Health score of a disk. If the value is 255, indicating invalid. |
SNMP agent | huawei.5300.v5[hwInfoDiskHealthMark, "{#ID}"] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Disk {#MODEL} on {#LOCATION}: Health status is not Normal | last(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwInfoDiskHealthStatus, "{#ID}"])<>1 |High |
|||
Disk {#MODEL} on {#LOCATION}: Running status is not Online | last(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwInfoDiskRunningStatus, "{#ID}"])<>27 |Average |
|||
Disk {#MODEL} on {#LOCATION}: Temperature is too high | min(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwInfoDiskTemperature, "{#ID}"],{$HUAWEI.5300.DISK.TEMP.MAX.TIME})>{$HUAWEI.5300.DISK.TEMP.MAX.WARN:"{#MODEL}"} |High |
Name | Description | Type | Key and additional info |
---|---|---|---|
Nodes performance discovery | Discovery of nodes performance counters |
SNMP agent | huawei.5300.nodes.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Node {#NODE}: CPU utilization | CPU usage of the node {#NODE}. |
SNMP agent | huawei.5300.v5[hwPerfNodeCPUUsage, "{#NODE}"] |
Node {#NODE}: Average I/O latency | Average I/O latency of the node. |
SNMP agent | huawei.5300.v5[hwPerfNodeDelay, "{#NODE}"] |
Node {#NODE}: Total I/O per second | Total IOPS of the node. |
SNMP agent | huawei.5300.v5[hwPerfNodeTotalIOPS, "{#NODE}"] |
Node {#NODE}: Read operations per second | Read IOPS of the node. |
SNMP agent | huawei.5300.v5[hwPerfNodeReadIOPS, "{#NODE}"] |
Node {#NODE}: Write operations per second | Write IOPS of the node. |
SNMP agent | huawei.5300.v5[hwPerfNodeWriteIOPS, "{#NODE}"] |
Node {#NODE}: Total traffic per second | Total bandwidth for the node. |
SNMP agent | huawei.5300.v5[hwPerfNodeTotalTraffic, "{#NODE}"] Preprocessing
|
Node {#NODE}: Read traffic per second | Read bandwidth for the node. |
SNMP agent | huawei.5300.v5[hwPerfNodeReadTraffic, "{#NODE}"] Preprocessing
|
Node {#NODE}: Write traffic per second | Write bandwidth for the node. |
SNMP agent | huawei.5300.v5[hwPerfNodeWriteTraffic, "{#NODE}"] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Node {#NODE}: High CPU utilization | The CPU utilization is too high. The system might be slow to respond. |
min(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwPerfNodeCPUUsage, "{#NODE}"],5m)>{$CPU.UTIL.CRIT} |Warning |
||
Node {#NODE}: Average I/O latency is too high | min(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwPerfNodeDelay, "{#NODE}"],{$HUAWEI.5300.NODE.IO.DELAY.MAX.TIME})>{$HUAWEI.5300.NODE.IO.DELAY.MAX.WARN} |Warning |
Name | Description | Type | Key and additional info |
---|---|---|---|
LUNs discovery | Discovery of LUNs |
SNMP agent | huawei.5300.lun.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
LUN {#NAME}: Status | Status of the LUN. |
SNMP agent | huawei.5300.v5[hwStorageLunStatus, "{#NAME}"] Preprocessing
|
LUN {#NAME}: Average total I/O latency | Average I/O latency of the node in milliseconds. |
SNMP agent | huawei.5300.v5[hwPerfLunAverageIOResponseTime, "{#NAME}"] Preprocessing
|
LUN {#NAME}: Average read I/O latency | Average read I/O response time in milliseconds. |
SNMP agent | huawei.5300.v5[hwPerfLunAverageReadIOLatency, "{#NAME}"] Preprocessing
|
LUN {#NAME}: Average write I/O latency | Average write I/O response time in milliseconds. |
SNMP agent | huawei.5300.v5[hwPerfLunAverageWriteIOLatency, "{#NAME}"] Preprocessing
|
LUN {#NAME}: Total I/O per second | Current IOPS of the LUN. |
SNMP agent | huawei.5300.v5[hwPerfLunTotalIOPS, "{#NAME}"] |
LUN {#NAME}: Read operations per second | Read IOPS of the node. |
SNMP agent | huawei.5300.v5[hwPerfLunReadIOPS, "{#NAME}"] |
LUN {#NAME}: Write operations per second | Write IOPS of the node. |
SNMP agent | huawei.5300.v5[hwPerfLunWriteIOPS, "{#NAME}"] |
LUN {#NAME}: Total traffic per second | Current total bandwidth for the LUN. |
SNMP agent | huawei.5300.v5[hwPerfLunTotalTraffic, "{#NAME}"] Preprocessing
|
LUN {#NAME}: Read traffic per second | Current read bandwidth for the LUN. |
SNMP agent | huawei.5300.v5[hwPerfLunReadTraffic, "{#NAME}"] Preprocessing
|
LUN {#NAME}: Write traffic per second | Current write bandwidth for the LUN. |
SNMP agent | huawei.5300.v5[hwPerfLunWriteTraffic, "{#NAME}"] Preprocessing
|
LUN {#NAME}: Capacity | Capacity of the LUN. |
SNMP agent | huawei.5300.v5[hwStorageLunCapacity, "{#NAME}"] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
LUN {#NAME}: Status is not Normal | last(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwStorageLunStatus, "{#NAME}"])<>1 |Average |
|||
LUN {#NAME}: Average I/O response time is too high | min(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwPerfLunAverageIOResponseTime, "{#NAME}"],{$HUAWEI.5300.LUN.IO.TIME.MAX.TIME})>{$HUAWEI.5300.LUN.IO.TIME.MAX.WARN} |Warning |
Name | Description | Type | Key and additional info |
---|---|---|---|
Storage pools discovery | Discovery of storage pools |
SNMP agent | huawei.5300.pool.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Pool {#NAME}: Health status | Health status of a storage pool. For details, see definition of Enum Values (HEALTHSTATUSE). https://support.huawei.com/enterprise/en/centralized-storage/oceanstor-5300-v5-pid-22462029?category=reference-guides&subcategory=mib-reference |
SNMP agent | huawei.5300.v5[hwInfoStoragePoolHealthStatus, "{#NAME}"] Preprocessing
|
Pool {#NAME}: Running status | Operating status of a storage pool. For details, see definition of Enum Values (RUNNINGSTATUSE). https://support.huawei.com/enterprise/en/centralized-storage/oceanstor-5300-v5-pid-22462029?category=reference-guides&subcategory=mib-reference |
SNMP agent | huawei.5300.v5[hwInfoStoragePoolRunningStatus, "{#NAME}"] Preprocessing
|
Pool {#NAME}: Capacity total | Total capacity of a storage pool. |
SNMP agent | huawei.5300.v5[hwInfoStoragePoolTotalCapacity, "{#NAME}"] Preprocessing
|
Pool {#NAME}: Capacity free | Available capacity of a storage pool. |
SNMP agent | huawei.5300.v5[hwInfoStoragePoolFreeCapacity, "{#NAME}"] Preprocessing
|
Pool {#NAME}: Capacity used | Used capacity of a storage pool. |
SNMP agent | huawei.5300.v5[hwInfoStoragePoolSubscribedCapacity, "{#NAME}"] Preprocessing
|
Pool {#NAME}: Capacity used percentage | Used capacity of a storage pool in percents. |
Calculated | huawei.5300.v5[hwInfoStoragePoolFreeCapacityPct, "{#NAME}"] |
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Pool {#NAME}: Health status is not Normal | last(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwInfoStoragePoolHealthStatus, "{#NAME}"])<>1 |High |
|||
Pool {#NAME}: Running status is not Online | last(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwInfoStoragePoolRunningStatus, "{#NAME}"])<>27 |Average |
|||
Pool {#NAME}: Used capacity is too high | min(/Huawei OceanStor 5300 V5 by SNMP/huawei.5300.v5[hwInfoStoragePoolFreeCapacityPct, "{#NAME}"],{$HUAWEI.5300.POOL.CAPACITY.THRESH.TIME})>{#THRESHOLD} |Average |
Please report any issues with the template at https://support.zabbix.com
You can also provide feedback, discuss the template, or ask for help at ZABBIX forums
The template to monitor HPE Primera by HTTP. It works without any external scripts and uses the script item.
Zabbix version: 6.0 and higher.
This template has been tested on:
Zabbix should be configured according to the instructions in the Templates out of the box section.
startwsapi
.
To check WSAPI state use command: showwsapi
.Name | Description | Default |
---|---|---|
{$HPE.PRIMERA.API.PASSWORD} | Specify password for WSAPI. |
|
{$HPE.PRIMERA.API.USERNAME} | Specify user name for WSAPI. |
zabbix |
{$HPE.PRIMERA.LLD.FILTER.TASK.NAME.MATCHES} | Filter of discoverable tasks by name. |
CHANGE_IF_NEEDED |
{$HPE.PRIMERA.LLD.FILTER.TASK.NAME.NOT_MATCHES} | Filter to exclude discovered tasks by name. |
.* |
{$HPE.PRIMERA.LLD.FILTER.TASK.TYPE.MATCHES} | Filter of discoverable tasks by type. |
.* |
{$HPE.PRIMERA.LLD.FILTER.TASK.TYPE.NOT_MATCHES} | Filter to exclude discovered tasks by type. |
CHANGE_IF_NEEDED |
{$HPE.PRIMERA.DATA.TIMEOUT} | Response timeout for WSAPI. |
15s |
{$HPE.PRIMERA.API.SCHEME} | The WSAPI scheme (http/https). |
https |
{$HPE.PRIMERA.API.PORT} | The WSAPI port. |
443 |
{$HPE.PRIMERA.VOLUME.NAME.MATCHES} | This macro is used in filters of volume discovery rule. |
.* |
{$HPE.PRIMERA.VOLUME.NAME.NOT_MATCHES} | This macro is used in filters of volume discovery rule. |
^(admin|.srdata|.mgmtdata)$ |
{$HPE.PRIMERA.CPG.NAME.MATCHES} | This macro is used in filters of CPGs discovery rule. |
.* |
{$HPE.PRIMERA.CPG.NAME.NOT_MATCHES} | This macro is used in filters of CPGs discovery rule. |
CHANGE_IF_NEEDED |
Name | Description | Type | Key and additional info |
---|---|---|---|
HPE Primera: Get data | The JSON with result of WSAPI requests. |
Script | hpe.primera.get.data |
HPE Primera: Get errors | A list of errors from WSAPI requests. |
Dependent item | hpe.primera.get.errors Preprocessing
|
HPE Primera: Get disks data | Disks data. |
Dependent item | hpe.primera.get.disks Preprocessing
|
HPE Primera: Get CPGs data | Common provisioning groups data. |
Dependent item | hpe.primera.get.cpgs Preprocessing
|
HPE Primera: Get hosts data | Hosts data. |
Dependent item | hpe.primera.get.hosts Preprocessing
|
HPE Primera: Get ports data | Ports data. |
Dependent item | hpe.primera.get.ports Preprocessing
|
HPE Primera: Get system data | System data. |
Dependent item | hpe.primera.get.system Preprocessing
|
HPE Primera: Get tasks data | Tasks data. |
Dependent item | hpe.primera.get.tasks Preprocessing
|
HPE Primera: Get volumes data | Volumes data. |
Dependent item | hpe.primera.get.volumes Preprocessing
|
HPE Primera: Capacity allocated | Allocated capacity in the system. |
Dependent item | hpe.primera.system.capacity.allocated Preprocessing
|
HPE Primera: Chunklet size | Chunklet size. |
Dependent item | hpe.primera.system.chunklet.size Preprocessing
|
HPE Primera: System contact | Contact of the system. |
Dependent item | hpe.primera.system.contact Preprocessing
|
HPE Primera: Capacity failed | Failed capacity in the system. |
Dependent item | hpe.primera.system.capacity.failed Preprocessing
|
HPE Primera: Capacity free | Free capacity in the system. |
Dependent item | hpe.primera.system.capacity.free Preprocessing
|
HPE Primera: System location | Location of the system. |
Dependent item | hpe.primera.system.location Preprocessing
|
HPE Primera: Model | System model. |
Dependent item | hpe.primera.system.model Preprocessing
|
HPE Primera: System name | System name. |
Dependent item | hpe.primera.system.name Preprocessing
|
HPE Primera: Serial number | System serial number. |
Dependent item | hpe.primera.system.serial_number Preprocessing
|
HPE Primera: Software version number | Storage system software version number. |
Dependent item | hpe.primera.system.sw_version Preprocessing
|
HPE Primera: Capacity total | Total capacity in the system. |
Dependent item | hpe.primera.system.capacity.total Preprocessing
|
HPE Primera: Nodes total | Total number of nodes in the system. |
Dependent item | hpe.primera.system.nodes.total Preprocessing
|
HPE Primera: Nodes online | Number of online nodes in the system. |
Dependent item | hpe.primera.system.nodes.online Preprocessing
|
HPE Primera: Disks total | Number of physical disks. |
Dependent item | hpe.primera.disks.total Preprocessing
|
HPE Primera: Service ping | Checks if the service is running and accepting TCP connections. |
Simple check | net.tcp.service["{$HPE.PRIMERA.API.SCHEME}","{HOST.CONN}","{$HPE.PRIMERA.API.PORT}"] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
HPE Primera: There are errors in requests to WSAPI | Zabbix has received errors in requests to WSAPI. |
length(last(/HPE Primera by HTTP/hpe.primera.get.errors))>0 |Average |
Depends on:
|
|
HPE Primera: Service is unavailable | max(/HPE Primera by HTTP/net.tcp.service["{$HPE.PRIMERA.API.SCHEME}","{HOST.CONN}","{$HPE.PRIMERA.API.PORT}"],5m)=0 |High |
Manual close: Yes |
Name | Description | Type | Key and additional info |
---|---|---|---|
Common provisioning groups discovery | List of CPGs resources. |
Dependent item | hpe.primera.cpg.discovery Preprocessing
|
Name | Description | Type | Key and additional info |
---|---|---|---|
CPG [{#NAME}]: Get CPG data | CPG {#NAME} data |
Dependent item | hpe.primera.cpg["{#ID}",data] Preprocessing
|
CPG [{#NAME}]: Degraded state | Detailed state of the CPG: LDSNOTSTARTED (1) - LDs not started. NOTSTARTED (2) - VV not started. NEEDSCHECK (3) - check for consistency. NEEDSMAINTCHECK (4) - maintenance check is required. INTERNALCONSISTENCYERROR (5) - internal consistency error. SNAPDATAINVALID (6) - invalid snapshot data. PRESERVED (7) - unavailable LD sets due to missing chunklets. Preserved remaining VV data. STALE (8) - parts of the VV contain old data because of a copy-on-write operation. COPYFAILED (9) - a promote or copy operation to this volume failed. DEGRADEDAVAIL (10) - degraded due to availability. DEGRADEDPERF (11) - degraded due to performance. PROMOTING (12) - volume is the current target of a promote operation. COPYTARGET (13) - volume is the current target of a physical copy operation. RESYNCTARGET (14) - volume is the current target of a resynchronized copy operation. TUNING (15) - volume tuning is in progress. CLOSING (16) - volume is closing. REMOVING (17) - removing the volume. REMOVINGRETRY (18) - retrying a volume removal operation. CREATING (19) - creating a volume. COPYSOURCE (20) - copy source. IMPORTING (21) - importing a volume. CONVERTING (22) - converting a volume. INVALID (23) - invalid. EXCLUSIVE (24) - local storage system has exclusive access to the volume. CONSISTENT (25) - volume is being imported consistently along with other volumes in the VV set. STANDBY (26) - volume in standby mode. SDMETAINCONSISTENT (27) - SD Meta Inconsistent. SDNEEDSFIX (28) - SD needs fix. SDMETAFIXING (29) - SD meta fix. UNKNOWN (999) - unknown state. NOTSUPPORTEDBY_WSAPI (1000) - state not supported by WSAPI. |
Dependent item | hpe.primera.cpg.state["{#ID}",degraded] Preprocessing
|
CPG [{#NAME}]: Failed state | Detailed state of the CPG: LDSNOTSTARTED (1) - LDs not started. NOTSTARTED (2) - VV not started. NEEDSCHECK (3) - check for consistency. NEEDSMAINTCHECK (4) - maintenance check is required. INTERNALCONSISTENCYERROR (5) - internal consistency error. SNAPDATAINVALID (6) - invalid snapshot data. PRESERVED (7) - unavailable LD sets due to missing chunklets. Preserved remaining VV data. STALE (8) - parts of the VV contain old data because of a copy-on-write operation. COPYFAILED (9) - a promote or copy operation to this volume failed. DEGRADEDAVAIL (10) - degraded due to availability. DEGRADEDPERF (11) - degraded due to performance. PROMOTING (12) - volume is the current target of a promote operation. COPYTARGET (13) - volume is the current target of a physical copy operation. RESYNCTARGET (14) - volume is the current target of a resynchronized copy operation. TUNING (15) - volume tuning is in progress. CLOSING (16) - volume is closing. REMOVING (17) - removing the volume. REMOVINGRETRY (18) - retrying a volume removal operation. CREATING (19) - creating a volume. COPYSOURCE (20) - copy source. IMPORTING (21) - importing a volume. CONVERTING (22) - converting a volume. INVALID (23) - invalid. EXCLUSIVE (24) - local storage system has exclusive access to the volume. CONSISTENT (25) - volume is being imported consistently along with other volumes in the VV set. STANDBY (26) - volume in standby mode. SDMETAINCONSISTENT (27) - SD Meta Inconsistent. SDNEEDSFIX (28) - SD needs fix. SDMETAFIXING (29) - SD meta fix. UNKNOWN (999) - unknown state. NOTSUPPORTEDBY_WSAPI (1000) - state not supported by WSAPI. |
Dependent item | hpe.primera.cpg.state["{#ID}",failed] Preprocessing
|
CPG [{#NAME}]: CPG space: Free | Free CPG space. |
Dependent item | hpe.primera.cpg.space["{#ID}",free] Preprocessing
|
CPG [{#NAME}]: Number of FPVVs | Number of FPVVs (Fully Provisioned Virtual Volumes) allocated in the CPG. |
Dependent item | hpe.primera.cpg.fpvv["{#ID}",count] Preprocessing
|
CPG [{#NAME}]: Number of TPVVs | Number of TPVVs (Thinly Provisioned Virtual Volumes) allocated in the CPG. |
Dependent item | hpe.primera.cpg.tpvv["{#ID}",count] Preprocessing
|
CPG [{#NAME}]: Number of TDVVs | Number of TDVVs (Thinly Deduplicated Virtual Volume) created in the CPG. |
Dependent item | hpe.primera.cpg.tdvv["{#ID}",count] Preprocessing
|
CPG [{#NAME}]: Raw space: Free | Raw free space. |
Dependent item | hpe.primera.cpg.space.raw["{#ID}",free] Preprocessing
|
CPG [{#NAME}]: Raw space: Shared | Raw shared space. |
Dependent item | hpe.primera.cpg.space.raw["{#ID}",shared] Preprocessing
|
CPG [{#NAME}]: Raw space: Total | Raw total space. |
Dependent item | hpe.primera.cpg.space.raw["{#ID}",total] Preprocessing
|
CPG [{#NAME}]: CPG space: Shared | Shared CPG space. |
Dependent item | hpe.primera.cpg.space["{#ID}",shared] Preprocessing
|
CPG [{#NAME}]: State | Overall state of the CPG: NORMAL (1) - normal operation; DEGRADED (2) - degraded state; FAILED (3) - abnormal operation; UNKNOWN (99) - unknown state. |
Dependent item | hpe.primera.cpg.state["{#ID}"] Preprocessing
|
CPG [{#NAME}]: Logical disk space: Snapshot administration: Total (raw) | Total physical (raw) logical disk space in snapshot administration. |
Dependent item | hpe.primera.cpg.space.sa["{#ID}",raw_total] Preprocessing
|
CPG [{#NAME}]: Logical disk space: Snapshot data: Total (raw) | Total physical (raw) logical disk space in snapshot data space. |
Dependent item | hpe.primera.cpg.space.sd["{#ID}",raw_total] Preprocessing
|
CPG [{#NAME}]: Logical disk space: User space: Total (raw) | Total physical (raw) logical disk space in user data space. |
Dependent item | hpe.primera.cpg.space.usr["{#ID}",raw_total] Preprocessing
|
CPG [{#NAME}]: Logical disk space: Snapshot administration: Total | Total logical disk space in snapshot administration. |
Dependent item | hpe.primera.cpg.space.sa["{#ID}",total] Preprocessing
|
CPG [{#NAME}]: Logical disk space: Snapshot data: Total | Total logical disk space in snapshot data space. |
Dependent item | hpe.primera.cpg.space.sd["{#ID}",total] Preprocessing
|
CPG [{#NAME}]: Logical disk space: User space: Total | Total logical disk space in user data space. |
Dependent item | hpe.primera.cpg.space.usr["{#ID}",total] Preprocessing
|
CPG [{#NAME}]: CPG space: Total | Total CPG space. |
Dependent item | hpe.primera.cpg.space["{#ID}",total] Preprocessing
|
CPG [{#NAME}]: Logical disk space: Snapshot administration: Used (raw) | Amount of physical (raw) logical disk used in snapshot administration. |
Dependent item | hpe.primera.cpg.space.sa["{#ID}",raw_used] Preprocessing
|
CPG [{#NAME}]: Logical disk space: Snapshot data: Used (raw) | Amount of physical (raw) logical disk used in snapshot data space. |
Dependent item | hpe.primera.cpg.space.sd["{#ID}",raw_used] Preprocessing
|
CPG [{#NAME}]: Logical disk space: User space: Used (raw) | Amount of physical (raw) logical disk used in user data space. |
Dependent item | hpe.primera.cpg.space.usr["{#ID}",raw_used] Preprocessing
|
CPG [{#NAME}]: Logical disk space: Snapshot administration: Used | Amount of logical disk used in snapshot administration. |
Dependent item | hpe.primera.cpg.space.sa["{#ID}",used] Preprocessing
|
CPG [{#NAME}]: Logical disk space: Snapshot data: Used | Amount of logical disk used in snapshot data space. |
Dependent item | hpe.primera.cpg.space.sd["{#ID}",used] Preprocessing
|
CPG [{#NAME}]: Logical disk space: User space: Used | Amount of logical disk used in user data space. |
Dependent item | hpe.primera.cpg.space.usr["{#ID}",used] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
CPG [{#NAME}]: Degraded | CPG [{#NAME}] is in degraded state. |
last(/HPE Primera by HTTP/hpe.primera.cpg.state["{#ID}"])=2 |Average |
||
CPG [{#NAME}]: Failed | CPG [{#NAME}] is in failed state. |
last(/HPE Primera by HTTP/hpe.primera.cpg.state["{#ID}"])=3 |High |
Name | Description | Type | Key and additional info |
---|---|---|---|
Disks discovery | List of physical disk resources. |
Dependent item | hpe.primera.disks.discovery Preprocessing
|
Name | Description | Type | Key and additional info |
---|---|---|---|
Disk [{#POSITION}]: Get disk data | Disk [{#POSITION}] data |
Dependent item | hpe.primera.disk["{#ID}",data] Preprocessing
|
Disk [{#POSITION}]: Firmware version | Physical disk firmware version. |
Dependent item | hpe.primera.disk["{#ID}",fw_version] Preprocessing
|
Disk [{#POSITION}]: Free size | Physical disk free size. |
Dependent item | hpe.primera.disk["{#ID}",free_size] Preprocessing
|
Disk [{#POSITION}]: Manufacturer | Physical disk manufacturer. |
Dependent item | hpe.primera.disk["{#ID}",manufacturer] Preprocessing
|
Disk [{#POSITION}]: Model | Manufacturer's device ID for disk. |
Dependent item | hpe.primera.disk["{#ID}",model] Preprocessing
|
Disk [{#POSITION}]: Path A0 degraded | Indicates if this is a degraded path for the disk. |
Dependent item | hpe.primera.disk["{#ID}",loopa0degraded] Preprocessing
|
Disk [{#POSITION}]: Path A1 degraded | Indicates if this is a degraded path for the disk. |
Dependent item | hpe.primera.disk["{#ID}",loopa1degraded] Preprocessing
|
Disk [{#POSITION}]: Path B0 degraded | Indicates if this is a degraded path for the disk. |
Dependent item | hpe.primera.disk["{#ID}",loopb0degraded] Preprocessing
|
Disk [{#POSITION}]: Path B1 degraded | Indicates if this is a degraded path for the disk. |
Dependent item | hpe.primera.disk["{#ID}",loopb1degraded] Preprocessing
|
Disk [{#POSITION}]: RPM | RPM of the physical disk. |
Dependent item | hpe.primera.disk["{#ID}",rpm] Preprocessing
|
Disk [{#POSITION}]: Serial number | Disk drive serial number. |
Dependent item | hpe.primera.disk["{#ID}",serial_number] Preprocessing
|
Disk [{#POSITION}]: State | State of the physical disk: Normal (1) - physical disk is in Normal state; Degraded (2) - physical disk is not operating normally; New (3) - physical disk is new, needs to be admitted; Failed (4) - physical disk has failed; Unknown (99) - physical disk state is unknown. |
Dependent item | hpe.primera.disk["{#ID}",state] Preprocessing
|
Disk [{#POSITION}]: Total size | Physical disk total size. |
Dependent item | hpe.primera.disk["{#ID}",total_size] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Disk [{#POSITION}]: Path A0 degraded | Disk [{#POSITION}] path A0 in degraded state. |
last(/HPE Primera by HTTP/hpe.primera.disk["{#ID}",loop_a0_degraded])=1 |Average |
||
Disk [{#POSITION}]: Path A1 degraded | Disk [{#POSITION}] path A1 in degraded state. |
last(/HPE Primera by HTTP/hpe.primera.disk["{#ID}",loop_a1_degraded])=1 |Average |
||
Disk [{#POSITION}]: Path B0 degraded | Disk [{#POSITION}] path B0 in degraded state. |
last(/HPE Primera by HTTP/hpe.primera.disk["{#ID}",loop_b0_degraded])=1 |Average |
||
Disk [{#POSITION}]: Path B1 degraded | Disk [{#POSITION}] path B1 in degraded state. |
last(/HPE Primera by HTTP/hpe.primera.disk["{#ID}",loop_b1_degraded])=1 |Average |
||
Disk [{#POSITION}]: Degraded | Disk [{#POSITION}] in degraded state. |
last(/HPE Primera by HTTP/hpe.primera.disk["{#ID}",state])=2 |Average |
||
Disk [{#POSITION}]: Failed | Disk [{#POSITION}] in failed state. |
last(/HPE Primera by HTTP/hpe.primera.disk["{#ID}",state])=3 |High |
||
Disk [{#POSITION}]: Unknown issue | Disk [{#POSITION}] in unknown state. |
last(/HPE Primera by HTTP/hpe.primera.disk["{#ID}",state])=99 |Info |
Name | Description | Type | Key and additional info |
---|---|---|---|
Hosts discovery | List of host properties. |
Dependent item | hpe.primera.hosts.discovery Preprocessing
|
Name | Description | Type | Key and additional info |
---|---|---|---|
Host [{#NAME}]: Get host data | Host [{#NAME}] data |
Dependent item | hpe.primera.host["{#ID}",data] Preprocessing
|
Host [{#NAME}]: Comment | Additional information for the host. |
Dependent item | hpe.primera.host["{#ID}",comment] Preprocessing
|
Host [{#NAME}]: Contact | The host's owner and contact. |
Dependent item | hpe.primera.host["{#ID}",contact] Preprocessing
|
Host [{#NAME}]: IP address | The host's IP address. |
Dependent item | hpe.primera.host["{#ID}",ipaddress] Preprocessing
|
Host [{#NAME}]: Location | The host's location. |
Dependent item | hpe.primera.host["{#ID}",location] Preprocessing
|
Host [{#NAME}]: Model | The host's model. |
Dependent item | hpe.primera.host["{#ID}",model] Preprocessing
|
Host [{#NAME}]: OS | The operating system running on the host. |
Dependent item | hpe.primera.host["{#ID}",os] Preprocessing
|
Name | Description | Type | Key and additional info |
---|---|---|---|
Ports discovery | List of ports. |
Dependent item | hpe.primera.ports.discovery Preprocessing
|
Name | Description | Type | Key and additional info |
---|---|---|---|
Port [{#NODE}:{#SLOT}:{#CARD.PORT}]: Get port data | Port [{#NODE}:{#SLOT}:{#CARD.PORT}] data |
Dependent item | hpe.primera.port["{#NODE}:{#SLOT}:{#CARD.PORT}",data] Preprocessing
|
Port [{#NODE}:{#SLOT}:{#CARD.PORT}]: Failover state | The state of the failover operation, shown for the two ports indicated in the N:S:P and Partner columns. The value can be one of the following: none (1) - no failover in operation; failoverpending (2) - in the process of failing over to partner; failedover (3) - failed over to partner; active (4) - the partner port is failed over to this port; activedown (5) - the partner port is failed over to this port, but this port is down; activefailed (6) - the partner port is failed over to this port, but this port is down; failback_pending (7) - in the process of failing back from partner. |
Dependent item | hpe.primera.port["{#NODE}:{#SLOT}:{#CARD.PORT}",failover_state] Preprocessing
|
Port [{#NODE}:{#SLOT}:{#CARD.PORT}]: Link state | Port link state: CONFIGWAIT (1) - configuration wait; ALPAWAIT (2) - ALPA wait; LOGINWAIT (3) - login wait; READY (4) - link is ready; LOSSSYNC (5) - link is loss sync; ERRORSTATE (6) - in error state; XXX (7) - xxx; NONPARTICIPATE (8) - link did not participate; COREDUMP (9) - taking coredump; OFFLINE (10) - link is offline; FWDEAD (11) - firmware is dead; IDLEFORRESET (12) - link is idle for reset; DHCPINPROGRESS (13) - DHCP is in progress; PENDINGRESET (14) - link reset is pending; NEW (15) - link in new. This value is applicable for only virtual ports; DISABLED (16) - link in disabled. This value is applicable for only virtual ports; DOWN (17) - link in down. This value is applicable for only virtual ports; FAILED (18) - link in failed. This value is applicable for only virtual ports; PURGING (19) - link in purging. This value is applicable for only virtual ports. |
Dependent item | hpe.primera.port["{#NODE}:{#SLOT}:{#CARD.PORT}",link_state] Preprocessing
|
Port [{#NODE}:{#SLOT}:{#CARD.PORT}]: Type | Port connection type: HOST (1) - FC port connected to hosts or fabric; DISK (2) - FC port connected to disks; FREE (3) - port is not connected to hosts or disks; IPORT (4) - port is in iport mode; RCFC (5) - FC port used for remote copy; PEER (6) - FC port used for data migration; RCIP (7) - IP (Ethernet) port used for remote copy; ISCSI (8) - iSCSI (Ethernet) port connected to hosts; CNA (9) - CNA port, which can be FCoE or iSCSI; FS (10) - Ethernet File Persona ports. |
Dependent item | hpe.primera.port["{#NODE}:{#SLOT}:{#CARD.PORT}",type] Preprocessing
|
Port [{#NODE}:{#SLOT}:{#CARD.PORT}]: Hardware type | Hardware type: FC (1) - Fibre channel HBA; ETH (2) - Ethernet NIC; iSCSI (3) - iSCSI HBA; CNA (4) - Converged network adapter; SAS (5) - SAS HBA; COMBO (6) - Combo card; NVME (7) - NVMe drive; UNKNOWN (99) - unknown hardware type. |
Dependent item | hpe.primera.port["{#NODE}:{#SLOT}:{#CARD.PORT}",hw_type] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Port [{#NODE}:{#SLOT}:{#CARD.PORT}]: Failover state is {ITEM.VALUE1} | Port [{#NODE}:{#SLOT}:{#CARD.PORT}] has failover error. |
last(/HPE Primera by HTTP/hpe.primera.port["{#NODE}:{#SLOT}:{#CARD.PORT}",failover_state])<>1 and last(/HPE Primera by HTTP/hpe.primera.port["{#NODE}:{#SLOT}:{#CARD.PORT}",failover_state])<>4 |Average |
||
Port [{#NODE}:{#SLOT}:{#CARD.PORT}]: Link state is {ITEM.VALUE1} | Port [{#NODE}:{#SLOT}:{#CARD.PORT}] not in ready state. |
last(/HPE Primera by HTTP/hpe.primera.port["{#NODE}:{#SLOT}:{#CARD.PORT}",link_state])<>4 and last(/HPE Primera by HTTP/hpe.primera.port["{#NODE}:{#SLOT}:{#CARD.PORT}",link_state])<>1 and last(/HPE Primera by HTTP/hpe.primera.port["{#NODE}:{#SLOT}:{#CARD.PORT}",link_state])<>3 and last(/HPE Primera by HTTP/hpe.primera.port["{#NODE}:{#SLOT}:{#CARD.PORT}",link_state])<>13 and last(/HPE Primera by HTTP/hpe.primera.port["{#NODE}:{#SLOT}:{#CARD.PORT}",link_state])<>15 and last(/HPE Primera by HTTP/hpe.primera.port["{#NODE}:{#SLOT}:{#CARD.PORT}",link_state])<>16 |High |
||
Port [{#NODE}:{#SLOT}:{#CARD.PORT}]: Link state is {ITEM.VALUE1} | Port [{#NODE}:{#SLOT}:{#CARD.PORT}] not in ready state. |
last(/HPE Primera by HTTP/hpe.primera.port["{#NODE}:{#SLOT}:{#CARD.PORT}",link_state])=1 or last(/HPE Primera by HTTP/hpe.primera.port["{#NODE}:{#SLOT}:{#CARD.PORT}",link_state])=3 or last(/HPE Primera by HTTP/hpe.primera.port["{#NODE}:{#SLOT}:{#CARD.PORT}",link_state])=13 or last(/HPE Primera by HTTP/hpe.primera.port["{#NODE}:{#SLOT}:{#CARD.PORT}",link_state])=15 or last(/HPE Primera by HTTP/hpe.primera.port["{#NODE}:{#SLOT}:{#CARD.PORT}",link_state])=16 |Average |
Name | Description | Type | Key and additional info |
---|---|---|---|
Tasks discovery | List of tasks started within last 24 hours. |
Dependent item | hpe.primera.tasks.discovery Preprocessing
|
Name | Description | Type | Key and additional info |
---|---|---|---|
Task [{#NAME}]: Get task data | Task [{#NAME}] data |
Dependent item | hpe.primera.task["{#ID}",data] Preprocessing
|
Task [{#NAME}]: Finish time | Task finish time. |
Dependent item | hpe.primera.task["{#ID}",finish_time] Preprocessing
|
Task [{#NAME}]: Start time | Task start time. |
Dependent item | hpe.primera.task["{#ID}",start_time] Preprocessing
|
Task [{#NAME}]: Status | Task status: DONE (1) - task is finished; ACTIVE (2) - task is in progress; CANCELLED (3) - task is canceled; FAILED (4) - task failed. |
Dependent item | hpe.primera.task["{#ID}",status] Preprocessing
|
Task [{#NAME}]: Type | Task type: VVCOPY (1) - track the physical copy operations; PHYSCOPYRESYNC (2) - track physical copy resynchronization operations; MOVEREGIONS (3) - track region move operations; PROMOTESV (4) - track virtual-copy promotions; REMOTECOPYSYNC (5) - track remote copy group synchronizations; REMOTECOPYREVERSE (6) - track the reversal of a remote copy group; REMOTECOPYFAILOVER (7) - track the change-over of a secondary volume group to a primaryvolume group;REMOTECOPYRECOVER (8) - track synchronization start after a failover operation from originalsecondary cluster to original primary cluster; REMOTECOPYRESTORE (9) - tracks the restoration process for groups that have already been recovered; COMPACTCPG (10) - track space consolidation in CPGs; COMPACTIDS (11) - track space consolidation in logical disks; SNAPSHOTACCOUNTING (12) - track progress of snapshot space usage accounting; CHECKVV (13) - track the progress of the check-volume operation; SCHEDULEDTASK (14) - track tasks that have been executed by the system scheduler; SYSTEMTASK (15) - track tasks that are periodically run by the storage system; BACKGROUNDTASK (16) - track commands started using the starttask command; IMPORTVV (17) - track tasks that migrate data to the local storage system; ONLINECOPY (18) - track physical copy of the volume while online (createvvcopy-online command); CONVERTVV (19) - track tasks that convert a volume from an FPVV to a TPVV, and the reverse; BACKGROUNDCOMMAND (20) - track background command tasks; CLXSYNC (21) - track CLX synchronization tasks; CLXRECOVERY (22) - track CLX recovery tasks; TUNESD (23) - tune copy space; TUNEVV (24) - tune virtual volume; TUNEVVROLLBACK (25) - tune virtual volume rollback; TUNEVVRESTART (26) - tune virtual volume restart; SYSTEMTUNING (27) - system tuning; NODERESCUE (28) - node rescue; REPAIRSYNC (29) - remote copy repair sync; REMOTECOPYSWOVER (30) - remote copy switchover; DEFRAGMENTATION (31) - defragmentation; ENCRYPTIONCHANGE (32) - encryption change; REMOTECOPYFAILSAFE (33) - remote copy failsafe; TUNETPVV (34) - tune thin virtual volume; REMOTECOPYCHGMODE (35) - remote copy change mode; ONLINEPROMOTE (37) - online promote snap; RELOCATEPD (38) - relocate PD; PERIODICCSS (39) - remote copy periodic CSS; TUNEVVLARGE (40) - tune large virtual volume; SDMETAFIXER (41) - compression SD meta fixer; DEDUPDRYRUN (42) - preview dedup ratio; COMPRDRYRUN (43) - compression estimation; DEDUPCOMPRDRYRUN (44) - compression and dedup estimation; UNKNOWN (99) - unknown task type. |
Dependent item | hpe.primera.task["{#ID}",type] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Task [{#NAME}]: Cancelled | Task [{#NAME}] is cancelled. |
last(/HPE Primera by HTTP/hpe.primera.task["{#ID}",status])=3 |Info |
||
Task [{#NAME}]: Failed | Task [{#NAME}] is failed. |
last(/HPE Primera by HTTP/hpe.primera.task["{#ID}",status])=4 |Average |
Name | Description | Type | Key and additional info |
---|---|---|---|
Volumes discovery | List of storage volume resources. |
Dependent item | hpe.primera.volumes.discovery Preprocessing
|
Name | Description | Type | Key and additional info |
---|---|---|---|
Volume [{#NAME}]: Get volume data | Volume [{#NAME}] data |
Dependent item | hpe.primera.volume["{#ID}",data] Preprocessing
|
Volume [{#NAME}]: Administrative space: Free | Free administrative space. |
Dependent item | hpe.primera.volume.space.admin["{#ID}",free] Preprocessing
|
Volume [{#NAME}]: Administrative space: Raw reserved | Raw reserved administrative space. |
Dependent item | hpe.primera.volume.space.admin["{#ID}",raw_reserved] Preprocessing
|
Volume [{#NAME}]: Administrative space: Reserved | Reserved administrative space. |
Dependent item | hpe.primera.volume.space.admin["{#ID}",reserved] Preprocessing
|
Volume [{#NAME}]: Administrative space: Used | Used administrative space. |
Dependent item | hpe.primera.volume.space.admin["{#ID}",used] Preprocessing
|
Volume [{#NAME}]: Compaction ratio | The compaction ratio indicates the overall amount of storage space saved with thin technology. |
Dependent item | hpe.primera.volume.capacity.efficiency["{#ID}",compaction] Preprocessing
|
Volume [{#NAME}]: Compression state | Volume compression state: YES (1) - compression is enabled on the volume; NO (2) - compression is disabled on the volume; OFF (3) - compression is turned off; NA (4) - compression is not available on the volume. |
Dependent item | hpe.primera.volume.state["{#ID}",compression] Preprocessing
|
Volume [{#NAME}]: Deduplication state | Volume deduplication state: YES (1) - enables deduplication on the volume; NO (2) - disables deduplication on the volume; NA (3) - deduplication is not available; OFF (4) - deduplication is turned off. |
Dependent item | hpe.primera.volume.state["{#ID}",deduplication] Preprocessing
|
Volume [{#NAME}]: Degraded state | Volume detailed state: LDSNOTSTARTED (1) - LDs not started. NOTSTARTED (2) - VV not started. NEEDSCHECK (3) - check for consistency. NEEDSMAINTCHECK (4) - maintenance check is required. INTERNALCONSISTENCYERROR (5) - internal consistency error. SNAPDATAINVALID (6) - invalid snapshot data. PRESERVED (7) - unavailable LD sets due to missing chunklets. Preserved remaining VV data. STALE (8) - parts of the VV contain old data because of a copy-on-write operation. COPYFAILED (9) - a promote or copy operation to this volume failed. DEGRADEDAVAIL (10) - degraded due to availability. DEGRADEDPERF (11) - degraded due to performance. PROMOTING (12) - volume is the current target of a promote operation. COPYTARGET (13) - volume is the current target of a physical copy operation. RESYNCTARGET (14) - volume is the current target of a resynchronized copy operation. TUNING (15) - volume tuning is in progress. CLOSING (16) - volume is closing. REMOVING (17) - removing the volume. REMOVINGRETRY (18) - retrying a volume removal operation. CREATING (19) - creating a volume. COPYSOURCE (20) - copy source. IMPORTING (21) - importing a volume. CONVERTING (22) - converting a volume. INVALID (23) - invalid. EXCLUSIVE (24) - local storage system has exclusive access to the volume. CONSISTENT (25) - volume is being imported consistently along with other volumes in the VV set. STANDBY (26) - volume in standby mode. SDMETAINCONSISTENT (27) - SD Meta Inconsistent. SDNEEDSFIX (28) - SD needs fix. SDMETAFIXING (29) - SD meta fix. UNKNOWN (999) - unknown state. NOTSUPPORTEDBY_WSAPI (1000) - state not supported by WSAPI. |
Dependent item | hpe.primera.volume.state["{#ID}",degraded] Preprocessing
|
Volume [{#NAME}]: Failed state | Volume detailed state: LDSNOTSTARTED (1) - LDs not started. NOTSTARTED (2) - VV not started. NEEDSCHECK (3) - check for consistency. NEEDSMAINTCHECK (4) - maintenance check is required. INTERNALCONSISTENCYERROR (5) - internal consistency error. SNAPDATAINVALID (6) - invalid snapshot data. PRESERVED (7) - unavailable LD sets due to missing chunklets. Preserved remaining VV data. STALE (8) - parts of the VV contain old data because of a copy-on-write operation. COPYFAILED (9) - a promote or copy operation to this volume failed. DEGRADEDAVAIL (10) - degraded due to availability. DEGRADEDPERF (11) - degraded due to performance. PROMOTING (12) - volume is the current target of a promote operation. COPYTARGET (13) - volume is the current target of a physical copy operation. RESYNCTARGET (14) - volume is the current target of a resynchronized copy operation. TUNING (15) - volume tuning is in progress. CLOSING (16) - volume is closing. REMOVING (17) - removing the volume. REMOVINGRETRY (18) - retrying a volume removal operation. CREATING (19) - creating a volume. COPYSOURCE (20) - copy source. IMPORTING (21) - importing a volume. CONVERTING (22) - converting a volume. INVALID (23) - invalid. EXCLUSIVE (24) - local storage system has exclusive access to the volume. CONSISTENT (25) - volume is being imported consistently along with other volumes in the VV set. STANDBY (26) - volume in standby mode. SDMETAINCONSISTENT (27) - SD Meta Inconsistent. SDNEEDSFIX (28) - SD needs fix. SDMETAFIXING (29) - SD meta fix. UNKNOWN (999) - unknown state. NOTSUPPORTEDBY_WSAPI (1000) - state not supported by WSAPI. |
Dependent item | hpe.primera.volume.state["{#ID}",failed] Preprocessing
|
Volume [{#NAME}]: Overprovisioning ratio | Overprovisioning capacity efficiency ratio. |
Dependent item | hpe.primera.volume.capacity.efficiency["{#ID}",overprovisioning] Preprocessing
|
Volume [{#NAME}]: Remote copy status | Remote copy status of the volume: NONE (1) - volume is not associated with remote copy; PRIMARY (2) - volume is the primary copy; SECONDARY (3) - volume is the secondary copy; SNAP (4) - volume is the remote copy snapshot; SYNC (5) - volume is a remote copy snapshot being used for synchronization; DELETE (6) - volume is a remote copy snapshot that is marked for deletion; UNKNOWN (99) - remote copy status is unknown for this volume. |
Dependent item | hpe.primera.volume.status["{#ID}",rcopy] Preprocessing
|
Volume [{#NAME}]: Snapshot space: Free | Free snapshot space. |
Dependent item | hpe.primera.volume.space.snapshot["{#ID}",free] Preprocessing
|
Volume [{#NAME}]: Snapshot space: Raw reserved | Raw reserved snapshot space. |
Dependent item | hpe.primera.volume.space.snapshot["{#ID}",raw_reserved] Preprocessing
|
Volume [{#NAME}]: Snapshot space: Reserved | Reserved snapshot space. |
Dependent item | hpe.primera.volume.space.snapshot["{#ID}",reserved] Preprocessing
|
Volume [{#NAME}]: Snapshot space: Used | Used snapshot space. |
Dependent item | hpe.primera.volume.space.snapshot["{#ID}",used] Preprocessing
|
Volume [{#NAME}]: State | State of the volume: NORMAL (1) - normal operation; DEGRADED (2) - degraded state; FAILED (3) - abnormal operation; UNKNOWN (99) - unknown state. |
Dependent item | hpe.primera.volume.state["{#ID}"] Preprocessing
|
Volume [{#NAME}]: Storage space saved using compression | Indicates the amount of storage space saved using compression. |
Dependent item | hpe.primera.volume.capacity.efficiency["{#ID}",compression] Preprocessing
|
Volume [{#NAME}]: Storage space saved using deduplication | Indicates the amount of storage space saved using deduplication. |
Dependent item | hpe.primera.volume.capacity.efficiency["{#ID}",deduplication] Preprocessing
|
Volume [{#NAME}]: Storage space saved using deduplication and compression | Indicates the amount of storage space saved using deduplication and compression together. |
Dependent item | hpe.primera.volume.capacity.efficiency["{#ID}",reduction] Preprocessing
|
Volume [{#NAME}]: Total reserved space | Total reserved space. |
Dependent item | hpe.primera.volume.space.total["{#ID}",reserved] Preprocessing
|
Volume [{#NAME}]: Total space | Virtual size of volume. |
Dependent item | hpe.primera.volume.space.total["{#ID}",size] Preprocessing
|
Volume [{#NAME}]: Total used space | Total used space. Sum of used user space and used snapshot space. |
Dependent item | hpe.primera.volume.space.total["{#ID}",used] Preprocessing
|
Volume [{#NAME}]: User space: Free | Free user space. |
Dependent item | hpe.primera.volume.space.user["{#ID}",free] Preprocessing
|
Volume [{#NAME}]: User space: Raw reserved | Raw reserved user space. |
Dependent item | hpe.primera.volume.space.user["{#ID}",raw_reserved] Preprocessing
|
Volume [{#NAME}]: User space: Reserved | Reserved user space. |
Dependent item | hpe.primera.volume.space.user["{#ID}",reserved] Preprocessing
|
Volume [{#NAME}]: User space: Used | Used user space. |
Dependent item | hpe.primera.volume.space.user["{#ID}",used] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Volume [{#NAME}]: Degraded | Volume [{#NAME}] is in degraded state. |
last(/HPE Primera by HTTP/hpe.primera.volume.state["{#ID}"])=2 |Average |
||
Volume [{#NAME}]: Failed | Volume [{#NAME}] is in failed state. |
last(/HPE Primera by HTTP/hpe.primera.volume.state["{#ID}"])=3 |High |
Please report any issues with the template at https://support.zabbix.com
You can also provide feedback, discuss the template, or ask for help at ZABBIX forums
The template to monitor HPE MSA 2060 by HTTP. It works without any external scripts and uses the script item.
Zabbix version: 6.0 and higher.
This template has been tested on:
Zabbix should be configured according to the instructions in the Templates out of the box section.
Name | Description | Default |
---|---|---|
{$HPE.MSA.API.SCHEME} | Connection scheme for API. |
https |
{$HPE.MSA.API.PORT} | Connection port for API. |
443 |
{$HPE.MSA.DATA.TIMEOUT} | Response timeout for API. |
30s |
{$HPE.MSA.API.USERNAME} | Specify user name for API. |
zabbix |
{$HPE.MSA.API.PASSWORD} | Specify password for API. |
|
{$HPE.MSA.DISKS.GROUP.PUSED.MAX.WARN} | The warning threshold of the disk group space utilization in %. |
80 |
{$HPE.MSA.DISKS.GROUP.PUSED.MAX.CRIT} | The critical threshold of the disk group space utilization in %. |
90 |
{$HPE.MSA.POOL.PUSED.MAX.WARN} | The warning threshold of the pool space utilization in %. |
80 |
{$HPE.MSA.POOL.PUSED.MAX.CRIT} | The critical threshold of the pool space utilization in %. |
90 |
{$HPE.MSA.CONTROLLER.CPU.UTIL.CRIT} | The critical threshold of the CPU utilization expressed in %. |
90 |
Name | Description | Type | Key and additional info |
---|---|---|---|
HPE MSA: Get data | The JSON with result of API requests. |
Script | hpe.msa.get.data |
Get system | The system data. |
Dependent item | hpe.msa.get.system Preprocessing
|
Get FRU | FRU data. |
Dependent item | hpe.msa.get.fru Preprocessing
|
Get fans | Fans data. |
Dependent item | hpe.msa.get.fans Preprocessing
|
Get disks | Disks data. |
Dependent item | hpe.msa.get.disks Preprocessing
|
Get enclosures | Enclosures data. |
Dependent item | hpe.msa.get.enclosures Preprocessing
|
Get ports | Ports data. |
Dependent item | hpe.msa.get.ports Preprocessing
|
Get power supplies | Power supplies data. |
Dependent item | hpe.msa.get.power_supplies Preprocessing
|
Get pools | Pools data. |
Dependent item | hpe.msa.get.pools Preprocessing
|
Get controllers | Controllers data. |
Dependent item | hpe.msa.get.controllers Preprocessing
|
Get controller statistics | Controllers statistics data. |
Dependent item | hpe.msa.get.controller_statistics Preprocessing
|
Get disk groups | Disk groups data. |
Dependent item | hpe.msa.get.disks.groups Preprocessing
|
Get disk group statistics | Disk groups statistics data. |
Dependent item | hpe.msa.disks.get.groups.statistics Preprocessing
|
Get volumes | Volumes data. |
Dependent item | hpe.msa.get.volumes Preprocessing
|
Get volume statistics | Volumes statistics data. |
Dependent item | hpe.msa.get.volumes.statistics Preprocessing
|
Get method errors | A list of method errors from API requests. |
Dependent item | hpe.msa.get.errors Preprocessing
|
Product ID | The product model identifier. |
Dependent item | hpe.msa.system.product_id Preprocessing
|
System contact | The name of the person who administers the system. |
Dependent item | hpe.msa.system.contact Preprocessing
|
System information | A brief description of what the system is used for or how it is configured. |
Dependent item | hpe.msa.system.info Preprocessing
|
System location | The location of the system. |
Dependent item | hpe.msa.system.location Preprocessing
|
System name | The name of the storage system. |
Dependent item | hpe.msa.system.name Preprocessing
|
Vendor name | The vendor name. |
Dependent item | hpe.msa.system.vendor_name Preprocessing
|
System health | System health status. |
Dependent item | hpe.msa.system.health Preprocessing
|
HPE MSA: Service ping | Check if HTTP/HTTPS service accepts TCP connections. |
Simple check | net.tcp.service["{$HPE.MSA.API.SCHEME}","{HOST.CONN}","{$HPE.MSA.API.PORT}"] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
There are errors in method requests to API | There are errors in method requests to API. |
length(last(/HPE MSA 2060 Storage by HTTP/hpe.msa.get.errors))>0 |Average |
Depends on:
|
|
System health is in degraded state | System health is in degraded state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.system.health)=1 |Warning |
||
System health is in fault state | System health is in fault state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.system.health)=2 |Average |
||
System health is in unknown state | System health is in unknown state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.system.health)=3 |Info |
||
Service is down or unavailable | HTTP/HTTPS service is down or unable to establish TCP connection. |
max(/HPE MSA 2060 Storage by HTTP/net.tcp.service["{$HPE.MSA.API.SCHEME}","{HOST.CONN}","{$HPE.MSA.API.PORT}"],5m)=0 |High |
Name | Description | Type | Key and additional info |
---|---|---|---|
Controllers discovery | Discover controllers. |
Dependent item | hpe.msa.controllers.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Controller [{#CONTROLLER.ID}]: Get data | The discovered controller data. |
Dependent item | hpe.msa.get.controllers["{#CONTROLLER.ID}",data] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Get statistics data | The discovered controller statistics data. |
Dependent item | hpe.msa.get.controller_statistics["{#CONTROLLER.ID}",data] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Firmware version | Storage controller firmware version. |
Dependent item | hpe.msa.controllers["{#CONTROLLER.ID}",firmware] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Part number | Part number of the controller. |
Dependent item | hpe.msa.controllers["{#CONTROLLER.ID}",part_number] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Serial number | Storage controller serial number. |
Dependent item | hpe.msa.controllers["{#CONTROLLER.ID}",serial_number] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Health | Controller health status. |
Dependent item | hpe.msa.controllers["{#CONTROLLER.ID}",health] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Status | Storage controller status. |
Dependent item | hpe.msa.controllers["{#CONTROLLER.ID}",status] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Disks | Number of disks in the storage system. |
Dependent item | hpe.msa.controllers["{#CONTROLLER.ID}",disks] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Pools | Number of pools in the storage system. |
Dependent item | hpe.msa.controllers["{#CONTROLLER.ID}",pools] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Disk groups | Number of disk groups in the storage system. |
Dependent item | hpe.msa.controllers["{#CONTROLLER.ID}",disk_groups] Preprocessing
|
Controller [{#CONTROLLER.ID}]: IP address | Controller network port IP address. |
Dependent item | hpe.msa.controllers["{#CONTROLLER.ID}",ip_address] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Cache memory size | Controller cache memory size. |
Dependent item | hpe.msa.controllers.cache["{#CONTROLLER.ID}",total] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Cache: Write utilization | Percentage of write cache in use, from 0 to 100. |
Dependent item | hpe.msa.controllers.cache.write["{#CONTROLLER.ID}",util] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Cache: Read hits, rate | For the controller that owns the volume, the number of times the block to be read is found in cache per second. |
Dependent item | hpe.msa.controllers.cache.read.hits["{#CONTROLLER.ID}",rate] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Cache: Read misses, rate | For the controller that owns the volume, the number of times the block to be read is not found in cache per second. |
Dependent item | hpe.msa.controllers.cache.read.misses["{#CONTROLLER.ID}",rate] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Cache: Write hits, rate | For the controller that owns the volume, the number of times the block written to is found in cache per second. |
Dependent item | hpe.msa.controllers.cache.write.hits["{#CONTROLLER.ID}",rate] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Cache: Write misses, rate | For the controller that owns the volume, the number of times the block written to is not found in cache per second. |
Dependent item | hpe.msa.controllers.cache.write.misses["{#CONTROLLER.ID}",rate] Preprocessing
|
Controller [{#CONTROLLER.ID}]: CPU utilization | Percentage of time the CPU is busy, from 0 to 100. |
Dependent item | hpe.msa.controllers.cpu["{#CONTROLLER.ID}",util] Preprocessing
|
Controller [{#CONTROLLER.ID}]: IOPS, total rate | Input/output operations per second, calculated over the interval since these statistics were last requested or reset. This value will be zero if it has not been requested or reset since a controller restart. |
Dependent item | hpe.msa.controllers.iops.total["{#CONTROLLER.ID}",rate] Preprocessing
|
Controller [{#CONTROLLER.ID}]: IOPS, read rate | Number of read operations per second. |
Dependent item | hpe.msa.controllers.iops.read["{#CONTROLLER.ID}",rate] Preprocessing
|
Controller [{#CONTROLLER.ID}]: IOPS, write rate | Number of write operations per second. |
Dependent item | hpe.msa.controllers.iops.write["{#CONTROLLER.ID}",rate] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Data transfer rate: Total | The data transfer rate, in bytes per second, calculated over the interval since these statistics were last requested or reset. This value will be zero if it has not been requested or reset since a controller restart. |
Dependent item | hpe.msa.controllers.data_transfer.total["{#CONTROLLER.ID}",rate] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Data transfer rate: Reads | The data read rate, in bytes per second. |
Dependent item | hpe.msa.controllers.data_transfer.reads["{#CONTROLLER.ID}",rate] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Data transfer rate: Writes | The data write rate, in bytes per second. |
Dependent item | hpe.msa.controllers.data_transfer.writes["{#CONTROLLER.ID}",rate] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Uptime | Number of seconds since the controller was restarted. |
Dependent item | hpe.msa.controllers["{#CONTROLLER.ID}",uptime] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Controller [{#CONTROLLER.ID}]: Controller health is in degraded state | Controller health is in degraded state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.controllers["{#CONTROLLER.ID}",health])=1 |Warning |
Depends on:
|
|
Controller [{#CONTROLLER.ID}]: Controller health is in fault state | Controller health is in fault state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.controllers["{#CONTROLLER.ID}",health])=2 |Average |
Depends on:
|
|
Controller [{#CONTROLLER.ID}]: Controller health is in unknown state | Controller health is in unknown state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.controllers["{#CONTROLLER.ID}",health])=3 |Info |
Depends on:
|
|
Controller [{#CONTROLLER.ID}]: Controller is down | The controller is down. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.controllers["{#CONTROLLER.ID}",status])=1 |High |
||
Controller [{#CONTROLLER.ID}]: High CPU utilization | Controller CPU utilization is too high. The system might be slow to respond. |
min(/HPE MSA 2060 Storage by HTTP/hpe.msa.controllers.cpu["{#CONTROLLER.ID}",util],5m)>{$HPE.MSA.CONTROLLER.CPU.UTIL.CRIT} |Warning |
||
Controller [{#CONTROLLER.ID}]: Controller has been restarted | The controller uptime is less than 10 minutes. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.controllers["{#CONTROLLER.ID}",uptime])<10m |Warning |
Name | Description | Type | Key and additional info |
---|---|---|---|
Disk groups discovery | Discover disk groups. |
Dependent item | hpe.msa.disks.groups.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Disk group [{#NAME}]: Get data | The discovered disk group data. |
Dependent item | hpe.msa.get.disks.groups["{#NAME}",data] Preprocessing
|
Disk group [{#NAME}]: Get statistics data | The discovered disk group statistics data. |
Dependent item | hpe.msa.get.disks.groups.statistics["{#NAME}",data] Preprocessing
|
Disk group [{#NAME}]: Disks count | Number of disks in the disk group. |
Dependent item | hpe.msa.disks.groups["{#NAME}",disk_count] Preprocessing
|
Disk group [{#NAME}]: Pool space used | The percentage of pool capacity that the disk group occupies. |
Dependent item | hpe.msa.disks.groups.space["{#NAME}",pool_util] Preprocessing
|
Disk group [{#NAME}]: Health | Disk group health. |
Dependent item | hpe.msa.disks.groups["{#NAME}",health] Preprocessing
|
Disk group [{#NAME}]: Blocks size | The size of a block, in bytes. |
Dependent item | hpe.msa.disks.groups.blocks["{#NAME}",size] Preprocessing
|
Disk group [{#NAME}]: Blocks free | Free space in blocks. |
Dependent item | hpe.msa.disks.groups.blocks["{#NAME}",free] Preprocessing
|
Disk group [{#NAME}]: Blocks total | Total space in blocks. |
Dependent item | hpe.msa.disks.groups.blocks["{#NAME}",total] Preprocessing
|
Disk group [{#NAME}]: Space free | The free space in the disk group. |
Calculated | hpe.msa.disks.groups.space["{#NAME}",free] Preprocessing
|
Disk group [{#NAME}]: Space total | The capacity of the disk group. |
Calculated | hpe.msa.disks.groups.space["{#NAME}",total] Preprocessing
|
Disk group [{#NAME}]: Space utilization | The space utilization percentage in the disk group. |
Calculated | hpe.msa.disks.groups.space["{#NAME}",util] Preprocessing
|
Disk group [{#NAME}]: RAID type | The RAID level of the disk group. |
Dependent item | hpe.msa.disks.groups.raid["{#NAME}",type] Preprocessing
|
Disk group [{#NAME}]: Status | The status of the disk group: - CRIT: Critical. The disk group is online but isn't fault tolerant because some of it's disks are down. - DMGD: Damaged. The disk group is online and fault tolerant, but some of it's disks are damaged. - FTDN: Fault tolerant with a down disk.The disk group is online and fault tolerant, but some of it's disks are down. - FTOL: Fault tolerant. - MSNG: Missing. The disk group is online and fault tolerant, but some of it's disks are missing. - OFFL: Offline. Either the disk group is using offline initialization, or it's disks are down and data may be lost. - QTCR: Quarantined critical. The disk group is critical with at least one inaccessible disk. For example, two disks are inaccessible in a RAID 6 disk group or one disk is inaccessible for other fault-tolerant RAID levels. If the inaccessible disks come online or if after 60 seconds from being quarantined the disk group is QTCRor QTDN, the disk group is automatically dequarantined. - QTDN: Quarantined with a down disk. The RAID6 disk group has one inaccessible disk. The disk group is fault tolerant but degraded. If the inaccessible disks come online or if after 60 seconds from being quarantined the disk group is QTCRor QTDN, the disk group is automatically dequarantined. - QTOF: Quarantined offline. The disk group is offline with multiple inaccessible disks causing user data to be incomplete, or is an NRAID or RAID 0 disk group. - QTUN: Quarantined unsupported. The disk group contains data in a format that is not supported by this system. For example, this system does not support linear disk groups. - STOP: The disk group is stopped. - UNKN: Unknown. - UP: Up. The disk group is online and does not have fault-tolerant attributes. |
Dependent item | hpe.msa.disks.groups["{#NAME}",status] Preprocessing
|
Disk group [{#NAME}]: IOPS, total rate | Input/output operations per second, calculated over the interval since these statistics were last requested or reset. This value will be zero if it has not been requested or reset since a controller restart. |
Dependent item | hpe.msa.disks.groups.iops.total["{#NAME}",rate] Preprocessing
|
Disk group [{#NAME}]: Average response time: Total | Average response time for read and write operations, calculated over the interval since these statistics were last requested or reset. |
Dependent item | hpe.msa.disks.groups.avgrsptime["{#NAME}",total] Preprocessing
|
Disk group [{#NAME}]: Average response time: Read | Average response time for all read operations, calculated over the interval since these statistics were last requested or reset. |
Dependent item | hpe.msa.disks.groups.avgrsptime["{#NAME}",read] Preprocessing
|
Disk group [{#NAME}]: Average response time: Write | Average response time for all write operations, calculated over the interval since these statistics were last requested or reset. |
Dependent item | hpe.msa.disks.groups.avgrsptime["{#NAME}",write] Preprocessing
|
Disk group [{#NAME}]: IOPS, read rate | Number of read operations per second. |
Dependent item | hpe.msa.disks.groups.iops.read["{#NAME}",rate] Preprocessing
|
Disk group [{#NAME}]: IOPS, write rate | Number of write operations per second. |
Dependent item | hpe.msa.disks.groups.iops.write["{#NAME}",rate] Preprocessing
|
Disk group [{#NAME}]: Data transfer rate: Total | The data transfer rate, in bytes per second, calculated over the interval since these statistics were last requested or reset. This value will be zero if it has not been requested or reset since a controller restart. |
Dependent item | hpe.msa.disks.groups.data_transfer.total["{#NAME}",rate] Preprocessing
|
Disk group [{#NAME}]: Data transfer rate: Reads | The data read rate, in bytes per second. |
Dependent item | hpe.msa.disks.groups.data_transfer.reads["{#NAME}",rate] Preprocessing
|
Disk group [{#NAME}]: Data transfer rate: Writes | The data write rate, in bytes per second. |
Dependent item | hpe.msa.disks.groups.data_transfer.writes["{#NAME}",rate] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Disk group [{#NAME}]: Disk group health is in degraded state | Disk group health is in degraded state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",health])=1 |Warning |
||
Disk group [{#NAME}]: Disk group health is in fault state | Disk group health is in fault state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",health])=2 |Average |
||
Disk group [{#NAME}]: Disk group health is in unknown state | Disk group health is in unknown state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",health])=3 |Info |
||
Disk group [{#NAME}]: Disk group space is low | Disk group is running low on free space (less than {$HPE.MSA.DISKS.GROUP.PUSED.MAX.WARN:"{#NAME}"}% available). |
min(/HPE MSA 2060 Storage by HTTP/hpe.msa.disks.groups.space["{#NAME}",util],5m)>{$HPE.MSA.DISKS.GROUP.PUSED.MAX.WARN:"{#NAME}"} |Warning |
Depends on:
|
|
Disk group [{#NAME}]: Disk group space is critically low | Disk group is running low on free space (less than {$HPE.MSA.DISKS.GROUP.PUSED.MAX.CRIT:"{#NAME}"}% available). |
min(/HPE MSA 2060 Storage by HTTP/hpe.msa.disks.groups.space["{#NAME}",util],5m)>{$HPE.MSA.DISKS.GROUP.PUSED.MAX.CRIT:"{#NAME}"} |Average |
||
Disk group [{#NAME}]: Disk group is fault tolerant with a down disk | The disk group is online and fault tolerant, but some of it's disks are down. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",status])=1 |Average |
||
Disk group [{#NAME}]: Disk group has damaged disks | The disk group is online and fault tolerant, but some of it's disks are damaged. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",status])=9 |Average |
||
Disk group [{#NAME}]: Disk group has missing disks | The disk group is online and fault tolerant, but some of it's disks are missing. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",status])=8 |Average |
||
Disk group [{#NAME}]: Disk group is offline | Either the disk group is using offline initialization, or it's disks are down and data may be lost. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",status])=3 |Average |
||
Disk group [{#NAME}]: Disk group is quarantined critical | The disk group is critical with at least one inaccessible disk. For example, two disks are inaccessible in a RAID 6 disk group or one disk is inaccessible for other fault-tolerant RAID levels. If the inaccessible disks come online or if after 60 seconds from being quarantined the disk group is QTCRor QTDN, the disk group is automatically dequarantined. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",status])=4 |Average |
||
Disk group [{#NAME}]: Disk group is quarantined offline | The disk group is offline with multiple inaccessible disks causing user data to be incomplete, or is an NRAID or RAID 0 disk group. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",status])=5 |Average |
||
Disk group [{#NAME}]: Disk group is quarantined unsupported | The disk group contains data in a format that is not supported by this system. For example, this system does not support linear disk groups. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",status])=5 |Average |
||
Disk group [{#NAME}]: Disk group is quarantined with an inaccessible disk | The RAID6 disk group has one inaccessible disk. The disk group is fault tolerant but degraded. If the inaccessible disks come online or if after 60 seconds from being quarantined the disk group is QTCRor QTDN, the disk group is automatically dequarantined. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",status])=6 |Average |
||
Disk group [{#NAME}]: Disk group is stopped | The disk group is stopped. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",status])=7 |Average |
||
Disk group [{#NAME}]: Disk group status is critical | The disk group is online but isn't fault tolerant because some of its disks are down. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",status])=2 |Average |
Name | Description | Type | Key and additional info |
---|---|---|---|
Pools discovery | Discover pools. |
Dependent item | hpe.msa.pools.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Pool [{#NAME}]: Get data | The discovered pool data. |
Dependent item | hpe.msa.get.pools["{#NAME}",data] Preprocessing
|
Pool [{#NAME}]: Health | Pool health. |
Dependent item | hpe.msa.pools["{#NAME}",health] Preprocessing
|
Pool [{#NAME}]: Blocks size | The size of a block, in bytes. |
Dependent item | hpe.msa.pools.blocks["{#NAME}",size] Preprocessing
|
Pool [{#NAME}]: Blocks available | Available space in blocks. |
Dependent item | hpe.msa.pools.blocks["{#NAME}",available] Preprocessing
|
Pool [{#NAME}]: Blocks total | Total space in blocks. |
Dependent item | hpe.msa.pools.blocks["{#NAME}",total] Preprocessing
|
Pool [{#NAME}]: Space free | The free space in the pool. |
Calculated | hpe.msa.pools.space["{#NAME}",free] Preprocessing
|
Pool [{#NAME}]: Space total | The capacity of the pool. |
Calculated | hpe.msa.pools.space["{#NAME}",total] Preprocessing
|
Pool [{#NAME}]: Space utilization | The space utilization percentage in the pool. |
Calculated | hpe.msa.pools.space["{#NAME}",util] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Pool [{#NAME}]: Pool health is in degraded state | Pool health is in degraded state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.pools["{#NAME}",health])=1 |Warning |
||
Pool [{#NAME}]: Pool health is in fault state | Pool health is in fault state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.pools["{#NAME}",health])=2 |Average |
||
Pool [{#NAME}]: Pool health is in unknown state | Pool [{#NAME}] health is in unknown state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.pools["{#NAME}",health])=3 |Info |
||
Pool [{#NAME}]: Pool space is low | Pool is running low on free space (less than {$HPE.MSA.POOL.PUSED.MAX.WARN:"{#NAME}"}% available). |
min(/HPE MSA 2060 Storage by HTTP/hpe.msa.pools.space["{#NAME}",util],5m)>{$HPE.MSA.POOL.PUSED.MAX.WARN:"{#NAME}"} |Warning |
Depends on:
|
|
Pool [{#NAME}]: Pool space is critically low | Pool is running low on free space (less than {$HPE.MSA.POOL.PUSED.MAX.CRIT:"{#NAME}"}% available). |
min(/HPE MSA 2060 Storage by HTTP/hpe.msa.pools.space["{#NAME}",util],5m)>{$HPE.MSA.POOL.PUSED.MAX.CRIT:"{#NAME}"} |Average |
Name | Description | Type | Key and additional info |
---|---|---|---|
Volumes discovery | Discover volumes. |
Dependent item | hpe.msa.volumes.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Volume [{#NAME}]: Get data | The discovered volume data. |
Dependent item | hpe.msa.get.volumes["{#NAME}",data] Preprocessing
|
Volume [{#NAME}]: Health | Volume health status. |
Dependent item | hpe.msa.volumes["{#DURABLE.ID}",health] Preprocessing
|
Volume [{#NAME}]: Get statistics data | The discovered volume statistics data. |
Dependent item | hpe.msa.get.volumes.statistics["{#NAME}",data] Preprocessing
|
Volume [{#NAME}]: Blocks size | The size of a block, in bytes. |
Dependent item | hpe.msa.volumes.blocks["{#NAME}",size] Preprocessing
|
Volume [{#NAME}]: Blocks allocated | The amount of blocks currently allocated to the volume. |
Dependent item | hpe.msa.volumes.blocks["{#NAME}",allocated] Preprocessing
|
Volume [{#NAME}]: Blocks total | Total space in blocks. |
Dependent item | hpe.msa.volumes.blocks["{#NAME}",total] Preprocessing
|
Volume [{#NAME}]: Space allocated | The amount of space currently allocated to the volume. |
Calculated | hpe.msa.volumes.space["{#NAME}",allocated] Preprocessing
|
Volume [{#NAME}]: Space total | The capacity of the volume. |
Calculated | hpe.msa.volumes.space["{#NAME}",total] Preprocessing
|
Volume [{#NAME}]: IOPS, total rate | Total input/output operations per second, calculated over the interval since these statistics were last requested or reset. This value will be zero if it has not been requested or reset since a controller restart. |
Dependent item | hpe.msa.volumes.iops.total["{#NAME}",rate] Preprocessing
|
Volume [{#NAME}]: IOPS, read rate | Number of read operations per second. |
Dependent item | hpe.msa.volumes.iops.read["{#NAME}",rate] Preprocessing
|
Volume [{#NAME}]: IOPS, write rate | Number of write operations per second. |
Dependent item | hpe.msa.volumes.iops.write["{#NAME}",rate] Preprocessing
|
Volume [{#NAME}]: Data transfer rate: Total | The data transfer rate, in bytes per second, calculated over the interval since these statistics were last requested or reset. This value will be zero if it has not been requested or reset since a controller restart. |
Dependent item | hpe.msa.volumes.data_transfer.total["{#NAME}",rate] Preprocessing
|
Volume [{#NAME}]: Data transfer rate: Reads | The data read rate, in bytes per second. |
Dependent item | hpe.msa.volumes.data_transfer.reads["{#NAME}",rate] Preprocessing
|
Volume [{#NAME}]: Data transfer rate: Writes | The data write rate, in bytes per second. |
Dependent item | hpe.msa.volumes.data_transfer.writes["{#NAME}",rate] Preprocessing
|
Volume [{#NAME}]: Cache: Read hits, rate | For the controller that owns the volume, the number of times the block to be read is found in cache per second. |
Dependent item | hpe.msa.volumes.cache.read.hits["{#NAME}",rate] Preprocessing
|
Volume [{#NAME}]: Cache: Read misses, rate | For the controller that owns the volume, the number of times the block to be read is not found in cache per second. |
Dependent item | hpe.msa.volumes.cache.read.misses["{#NAME}",rate] Preprocessing
|
Volume [{#NAME}]: Cache: Write hits, rate | For the controller that owns the volume, the number of times the block written to is found in cache per second. |
Dependent item | hpe.msa.volumes.cache.write.hits["{#NAME}",rate] Preprocessing
|
Volume [{#NAME}]: Cache: Write misses, rate | For the controller that owns the volume, the number of times the block written to is not found in cache per second. |
Dependent item | hpe.msa.volumes.cache.write.misses["{#NAME}",rate] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Volume [{#NAME}]: Volume health is in degraded state | Volume health is in degraded state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.volumes["{#DURABLE.ID}",health])=1 |Warning |
||
Volume [{#NAME}]: Volume health is in fault state | Volume health is in fault state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.volumes["{#DURABLE.ID}",health])=2 |Average |
||
Volume [{#NAME}]: Volume health is in unknown state | Volume health is in unknown state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.volumes["{#DURABLE.ID}",health])=3 |Info |
Name | Description | Type | Key and additional info |
---|---|---|---|
Enclosures discovery | Discover enclosures. |
Dependent item | hpe.msa.enclosures.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Enclosure [{#DURABLE.ID}]: Get data | The discovered enclosure data. |
Dependent item | hpe.msa.get.enclosures["{#DURABLE.ID}",data] Preprocessing
|
Enclosure [{#DURABLE.ID}]: Health | Enclosure health. |
Dependent item | hpe.msa.enclosures["{#DURABLE.ID}",health] Preprocessing
|
Enclosure [{#DURABLE.ID}]: Status | Enclosure status. |
Dependent item | hpe.msa.enclosures["{#DURABLE.ID}",status] Preprocessing
|
Enclosure [{#DURABLE.ID}]: Midplane serial number | Midplane serial number. |
Dependent item | hpe.msa.enclosures["{#DURABLE.ID}",midplaneserialnumber] Preprocessing
|
Enclosure [{#DURABLE.ID}]: Part number | Enclosure part number. |
Dependent item | hpe.msa.enclosures["{#DURABLE.ID}",part_number] Preprocessing
|
Enclosure [{#DURABLE.ID}]: Model | Enclosure model. |
Dependent item | hpe.msa.enclosures["{#DURABLE.ID}",model] Preprocessing
|
Enclosure [{#DURABLE.ID}]: Power | Enclosure power in watts. |
Dependent item | hpe.msa.enclosures["{#DURABLE.ID}",power] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Enclosure [{#DURABLE.ID}]: Enclosure health is in degraded state | Enclosure health is in degraded state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.enclosures["{#DURABLE.ID}",health])=1 |Warning |
||
Enclosure [{#DURABLE.ID}]: Enclosure health is in fault state | Enclosure health is in fault state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.enclosures["{#DURABLE.ID}",health])=2 |Average |
||
Enclosure [{#DURABLE.ID}]: Enclosure health is in unknown state | Enclosure health is in unknown state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.enclosures["{#DURABLE.ID}",health])=3 |Info |
||
Enclosure [{#DURABLE.ID}]: Enclosure has critical status | Enclosure has critical status. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.enclosures["{#DURABLE.ID}",status])=2 |High |
||
Enclosure [{#DURABLE.ID}]: Enclosure has warning status | Enclosure has warning status. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.enclosures["{#DURABLE.ID}",status])=3 |Warning |
||
Enclosure [{#DURABLE.ID}]: Enclosure is unavailable | Enclosure is unavailable. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.enclosures["{#DURABLE.ID}",status])=7 |High |
||
Enclosure [{#DURABLE.ID}]: Enclosure is unrecoverable | Enclosure is unrecoverable. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.enclosures["{#DURABLE.ID}",status])=4 |High |
||
Enclosure [{#DURABLE.ID}]: Enclosure has unknown status | Enclosure has unknown status. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.enclosures["{#DURABLE.ID}",status])=6 |Info |
Name | Description | Type | Key and additional info |
---|---|---|---|
Power supplies discovery | Discover power supplies. |
Dependent item | hpe.msa.power_supplies.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Power supply [{#DURABLE.ID}]: Get data | The discovered power supply data. |
Dependent item | hpe.msa.get.power_supplies["{#DURABLE.ID}",data] Preprocessing
|
Power supply [{#DURABLE.ID}]: Health | Power supply health status. |
Dependent item | hpe.msa.power_supplies["{#DURABLE.ID}",health] Preprocessing
|
Power supply [{#DURABLE.ID}]: Status | Power supply status. |
Dependent item | hpe.msa.power_supplies["{#DURABLE.ID}",status] Preprocessing
|
Power supply [{#DURABLE.ID}]: Part number | Power supply part number. |
Dependent item | hpe.msa.powersupplies["{#DURABLE.ID}",partnumber] Preprocessing
|
Power supply [{#DURABLE.ID}]: Serial number | Power supply serial number. |
Dependent item | hpe.msa.powersupplies["{#DURABLE.ID}",serialnumber] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Power supply [{#DURABLE.ID}]: Power supply health is in degraded state | Power supply health is in degraded state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.power_supplies["{#DURABLE.ID}",health])=1 |Warning |
||
Power supply [{#DURABLE.ID}]: Power supply health is in fault state | Power supply health is in fault state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.power_supplies["{#DURABLE.ID}",health])=2 |Average |
||
Power supply [{#DURABLE.ID}]: Power supply health is in unknown state | Power supply health is in unknown state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.power_supplies["{#DURABLE.ID}",health])=3 |Info |
||
Power supply [{#DURABLE.ID}]: Power supply has error status | Power supply has error status. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.power_supplies["{#DURABLE.ID}",status])=2 |Average |
||
Power supply [{#DURABLE.ID}]: Power supply has warning status | Power supply has warning status. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.power_supplies["{#DURABLE.ID}",status])=1 |Warning |
||
Power supply [{#DURABLE.ID}]: Power supply has unknown status | Power supply has unknown status. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.power_supplies["{#DURABLE.ID}",status])=4 |Info |
Name | Description | Type | Key and additional info |
---|---|---|---|
Ports discovery | Discover ports. |
Dependent item | hpe.msa.ports.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Port [{#NAME}]: Get data | The discovered port data. |
Dependent item | hpe.msa.get.ports["{#NAME}",,data] Preprocessing
|
Port [{#NAME}]: Health | Port health status. |
Dependent item | hpe.msa.ports["{#NAME}",health] Preprocessing
|
Port [{#NAME}]: Status | Port status. |
Dependent item | hpe.msa.ports["{#NAME}",status] Preprocessing
|
Port [{#NAME}]: Type | Port type. |
Dependent item | hpe.msa.ports["{#NAME}",type] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Port [{#NAME}]: Port health is in degraded state | Port health is in degraded state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.ports["{#NAME}",health])=1 |Warning |
||
Port [{#NAME}]: Port health is in fault state | Port health is in fault state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.ports["{#NAME}",health])=2 |Average |
||
Port [{#NAME}]: Port health is in unknown state | Port health is in unknown state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.ports["{#NAME}",health])=3 |Info |
||
Port [{#NAME}]: Port has error status | Port has error status. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.ports["{#NAME}",status])=2 |Average |
||
Port [{#NAME}]: Port has warning status | Port has warning status. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.ports["{#NAME}",status])=1 |Warning |
||
Port [{#NAME}]: Port has unknown status | Port has unknown status. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.ports["{#NAME}",status])=4 |Info |
Name | Description | Type | Key and additional info |
---|---|---|---|
Fans discovery | Discover fans. |
Dependent item | hpe.msa.fans.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Fan [{#DURABLE.ID}]: Get data | The discovered fan data. |
Dependent item | hpe.msa.get.fans["{#DURABLE.ID}",data] Preprocessing
|
Fan [{#DURABLE.ID}]: Health | Fan health status. |
Dependent item | hpe.msa.fans["{#DURABLE.ID}",health] Preprocessing
|
Fan [{#DURABLE.ID}]: Status | Fan status. |
Dependent item | hpe.msa.fans["{#DURABLE.ID}",status] Preprocessing
|
Fan [{#DURABLE.ID}]: Speed | Fan speed (revolutions per minute). |
Dependent item | hpe.msa.fans["{#DURABLE.ID}",speed] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Fan [{#DURABLE.ID}]: Fan health is in degraded state | Fan health is in degraded state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.fans["{#DURABLE.ID}",health])=1 |Warning |
||
Fan [{#DURABLE.ID}]: Fan health is in fault state | Fan health is in fault state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.fans["{#DURABLE.ID}",health])=2 |Average |
||
Fan [{#DURABLE.ID}]: Fan health is in unknown state | Fan health is in unknown state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.fans["{#DURABLE.ID}",health])=3 |Info |
||
Fan [{#DURABLE.ID}]: Fan has error status | Fan has error status. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.fans["{#DURABLE.ID}",status])=1 |Average |
||
Fan [{#DURABLE.ID}]: Fan is missing | Fan is missing. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.fans["{#DURABLE.ID}",status])=3 |Info |
||
Fan [{#DURABLE.ID}]: Fan is off | Fan is off. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.fans["{#DURABLE.ID}",status])=2 |Warning |
Name | Description | Type | Key and additional info |
---|---|---|---|
Disks discovery | Discover disks. |
Dependent item | hpe.msa.disks.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Disk [{#DURABLE.ID}]: Get data | The discovered disk data. |
Dependent item | hpe.msa.get.disks["{#DURABLE.ID}",data] Preprocessing
|
Disk [{#DURABLE.ID}]: Health | Disk health status. |
Dependent item | hpe.msa.disks["{#DURABLE.ID}",health] Preprocessing
|
Disk [{#DURABLE.ID}]: Temperature status | Disk temperature status. |
Dependent item | hpe.msa.disks["{#DURABLE.ID}",temperature_status] Preprocessing
|
Disk [{#DURABLE.ID}]: Temperature | Temperature of the disk. |
Dependent item | hpe.msa.disks["{#DURABLE.ID}",temperature] Preprocessing
|
Disk [{#DURABLE.ID}]: Type | Disk type: SAS: Enterprise SAS spinning disk. SAS MDL: Midline SAS spinning disk. SSD SAS: SAS solit-state disk. |
Dependent item | hpe.msa.disks["{#DURABLE.ID}",type] Preprocessing
|
Disk [{#DURABLE.ID}]: Disk group | If the disk is in a disk group, the disk group name. |
Dependent item | hpe.msa.disks["{#DURABLE.ID}",group] Preprocessing
|
Disk [{#DURABLE.ID}]: Storage pool | If the disk is in a pool, the pool name. |
Dependent item | hpe.msa.disks["{#DURABLE.ID}",pool] Preprocessing
|
Disk [{#DURABLE.ID}]: Vendor | Disk vendor. |
Dependent item | hpe.msa.disks["{#DURABLE.ID}",vendor] Preprocessing
|
Disk [{#DURABLE.ID}]: Model | Disk model. |
Dependent item | hpe.msa.disks["{#DURABLE.ID}",model] Preprocessing
|
Disk [{#DURABLE.ID}]: Serial number | Disk serial number. |
Dependent item | hpe.msa.disks["{#DURABLE.ID}",serial_number] Preprocessing
|
Disk [{#DURABLE.ID}]: Blocks size | The size of a block, in bytes. |
Dependent item | hpe.msa.disks.blocks["{#DURABLE.ID}",size] Preprocessing
|
Disk [{#DURABLE.ID}]: Blocks total | Total space in blocks. |
Dependent item | hpe.msa.disks.blocks["{#DURABLE.ID}",total] Preprocessing
|
Disk [{#DURABLE.ID}]: Space total | Total size of the disk. |
Calculated | hpe.msa.disks.space["{#DURABLE.ID}",total] Preprocessing
|
Disk [{#DURABLE.ID}]: SSD life left | The percentage of disk life remaining. |
Dependent item | hpe.msa.disks.ssd["{#DURABLE.ID}",life_left] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Disk [{#DURABLE.ID}]: Disk health is in degraded state | Disk health is in degraded state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.disks["{#DURABLE.ID}",health])=1 |Warning |
||
Disk [{#DURABLE.ID}]: Disk health is in fault state | Disk health is in fault state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.disks["{#DURABLE.ID}",health])=2 |Average |
||
Disk [{#DURABLE.ID}]: Disk health is in unknown state | Disk health is in unknown state. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.disks["{#DURABLE.ID}",health])=3 |Info |
||
Disk [{#DURABLE.ID}]: Disk temperature is high | Disk temperature is high. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.disks["{#DURABLE.ID}",temperature_status])=3 |Warning |
||
Disk [{#DURABLE.ID}]: Disk temperature is critically high | Disk temperature is critically high. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.disks["{#DURABLE.ID}",temperature_status])=2 |Average |
||
Disk [{#DURABLE.ID}]: Disk temperature is unknown | Disk temperature is unknown. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.disks["{#DURABLE.ID}",temperature_status])=4 |Info |
Name | Description | Type | Key and additional info |
---|---|---|---|
FRU discovery | Discover FRU. |
Dependent item | hpe.msa.frus.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
FRU [{#ENCLOSURE.ID}: {#LOCATION}]: Get data | The discovered FRU data. |
Dependent item | hpe.msa.get.frus["{#ENCLOSURE.ID}:{#LOCATION}",data] Preprocessing
|
FRU [{#ENCLOSURE.ID}: {#LOCATION}]: Status | {#DESCRIPTION}. FRU status: Absent: The FRU is not present. Fault: The FRU's health is Degraded or Fault. Invalid data: The FRU ID data is invalid. The FRU's EEPROM is improperly programmed. OK: The FRU is operating normally. Power off: The FRU is powered off. |
Dependent item | hpe.msa.frus["{#ENCLOSURE.ID}:{#LOCATION}",status] Preprocessing
|
FRU [{#ENCLOSURE.ID}: {#LOCATION}]: Part number | {#DESCRIPTION}. Part number of the FRU. |
Dependent item | hpe.msa.frus["{#ENCLOSURE.ID}:{#LOCATION}",part_number] Preprocessing
|
FRU [{#ENCLOSURE.ID}: {#LOCATION}]: Serial number | {#DESCRIPTION}. FRU serial number. |
Dependent item | hpe.msa.frus["{#ENCLOSURE.ID}:{#LOCATION}",serial_number] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
FRU [{#ENCLOSURE.ID}: {#LOCATION}]: FRU status is Degraded or Fault | FRU status is Degraded or Fault. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.frus["{#ENCLOSURE.ID}:{#LOCATION}",status])=1 |Average |
||
FRU [{#ENCLOSURE.ID}: {#LOCATION}]: FRU ID data is invalid | The FRU ID data is invalid. The FRU's EEPROM is improperly programmed. |
last(/HPE MSA 2060 Storage by HTTP/hpe.msa.frus["{#ENCLOSURE.ID}:{#LOCATION}",status])=0 |Warning |
Please report any issues with the template at https://support.zabbix.com
You can also provide feedback, discuss the template, or ask for help at ZABBIX forums
The template to monitor HPE MSA 2040 by HTTP. It works without any external scripts and uses the script item.
Zabbix version: 6.0 and higher.
This template has been tested on:
Zabbix should be configured according to the instructions in the Templates out of the box section.
Name | Description | Default |
---|---|---|
{$HPE.MSA.API.SCHEME} | Connection scheme for API. |
https |
{$HPE.MSA.API.PORT} | Connection port for API. |
443 |
{$HPE.MSA.DATA.TIMEOUT} | Response timeout for API. |
30s |
{$HPE.MSA.API.USERNAME} | Specify user name for API. |
zabbix |
{$HPE.MSA.API.PASSWORD} | Specify password for API. |
|
{$HPE.MSA.DISKS.GROUP.PUSED.MAX.WARN} | The warning threshold of the disk group space utilization in %. |
80 |
{$HPE.MSA.DISKS.GROUP.PUSED.MAX.CRIT} | The critical threshold of the disk group space utilization in %. |
90 |
{$HPE.MSA.POOL.PUSED.MAX.WARN} | The warning threshold of the pool space utilization in %. |
80 |
{$HPE.MSA.POOL.PUSED.MAX.CRIT} | The critical threshold of the pool space utilization in %. |
90 |
{$HPE.MSA.CONTROLLER.CPU.UTIL.CRIT} | The critical threshold of the CPU utilization expressed in %. |
90 |
Name | Description | Type | Key and additional info |
---|---|---|---|
HPE MSA: Get data | The JSON with result of API requests. |
Script | hpe.msa.get.data |
Get system | The system data. |
Dependent item | hpe.msa.get.system Preprocessing
|
Get FRU | FRU data. |
Dependent item | hpe.msa.get.fru Preprocessing
|
Get fans | Fans data. |
Dependent item | hpe.msa.get.fans Preprocessing
|
Get disks | Disks data. |
Dependent item | hpe.msa.get.disks Preprocessing
|
Get enclosures | Enclosures data. |
Dependent item | hpe.msa.get.enclosures Preprocessing
|
Get ports | Ports data. |
Dependent item | hpe.msa.get.ports Preprocessing
|
Get power supplies | Power supplies data. |
Dependent item | hpe.msa.get.power_supplies Preprocessing
|
Get pools | Pools data. |
Dependent item | hpe.msa.get.pools Preprocessing
|
Get controllers | Controllers data. |
Dependent item | hpe.msa.get.controllers Preprocessing
|
Get controller statistics | Controllers statistics data. |
Dependent item | hpe.msa.get.controller_statistics Preprocessing
|
Get disk groups | Disk groups data. |
Dependent item | hpe.msa.get.disks.groups Preprocessing
|
Get disk group statistics | Disk groups statistics data. |
Dependent item | hpe.msa.disks.get.groups.statistics Preprocessing
|
Get volumes | Volumes data. |
Dependent item | hpe.msa.get.volumes Preprocessing
|
Get volume statistics | Volumes statistics data. |
Dependent item | hpe.msa.get.volumes.statistics Preprocessing
|
Get method errors | A list of method errors from API requests. |
Dependent item | hpe.msa.get.errors Preprocessing
|
Product ID | The product model identifier. |
Dependent item | hpe.msa.system.product_id Preprocessing
|
System contact | The name of the person who administers the system. |
Dependent item | hpe.msa.system.contact Preprocessing
|
System information | A brief description of what the system is used for or how it is configured. |
Dependent item | hpe.msa.system.info Preprocessing
|
System location | The location of the system. |
Dependent item | hpe.msa.system.location Preprocessing
|
System name | The name of the storage system. |
Dependent item | hpe.msa.system.name Preprocessing
|
Vendor name | The vendor name. |
Dependent item | hpe.msa.system.vendor_name Preprocessing
|
System health | System health status. |
Dependent item | hpe.msa.system.health Preprocessing
|
HPE MSA: Service ping | Check if HTTP/HTTPS service accepts TCP connections. |
Simple check | net.tcp.service["{$HPE.MSA.API.SCHEME}","{HOST.CONN}","{$HPE.MSA.API.PORT}"] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
There are errors in method requests to API | There are errors in method requests to API. |
length(last(/HPE MSA 2040 Storage by HTTP/hpe.msa.get.errors))>0 |Average |
Depends on:
|
|
System health is in degraded state | System health is in degraded state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.system.health)=1 |Warning |
||
System health is in fault state | System health is in fault state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.system.health)=2 |Average |
||
System health is in unknown state | System health is in unknown state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.system.health)=3 |Info |
||
Service is down or unavailable | HTTP/HTTPS service is down or unable to establish TCP connection. |
max(/HPE MSA 2040 Storage by HTTP/net.tcp.service["{$HPE.MSA.API.SCHEME}","{HOST.CONN}","{$HPE.MSA.API.PORT}"],5m)=0 |High |
Name | Description | Type | Key and additional info |
---|---|---|---|
Controllers discovery | Discover controllers. |
Dependent item | hpe.msa.controllers.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Controller [{#CONTROLLER.ID}]: Get data | The discovered controller data. |
Dependent item | hpe.msa.get.controllers["{#CONTROLLER.ID}",data] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Get statistics data | The discovered controller statistics data. |
Dependent item | hpe.msa.get.controller_statistics["{#CONTROLLER.ID}",data] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Firmware version | Storage controller firmware version. |
Dependent item | hpe.msa.controllers["{#CONTROLLER.ID}",firmware] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Part number | Part number of the controller. |
Dependent item | hpe.msa.controllers["{#CONTROLLER.ID}",part_number] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Serial number | Storage controller serial number. |
Dependent item | hpe.msa.controllers["{#CONTROLLER.ID}",serial_number] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Health | Controller health status. |
Dependent item | hpe.msa.controllers["{#CONTROLLER.ID}",health] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Status | Storage controller status. |
Dependent item | hpe.msa.controllers["{#CONTROLLER.ID}",status] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Disks | Number of disks in the storage system. |
Dependent item | hpe.msa.controllers["{#CONTROLLER.ID}",disks] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Pools | Number of pools in the storage system. |
Dependent item | hpe.msa.controllers["{#CONTROLLER.ID}",pools] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Disk groups | Number of disk groups in the storage system. |
Dependent item | hpe.msa.controllers["{#CONTROLLER.ID}",disk_groups] Preprocessing
|
Controller [{#CONTROLLER.ID}]: IP address | Controller network port IP address. |
Dependent item | hpe.msa.controllers["{#CONTROLLER.ID}",ip_address] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Cache memory size | Controller cache memory size. |
Dependent item | hpe.msa.controllers.cache["{#CONTROLLER.ID}",total] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Cache: Write utilization | Percentage of write cache in use, from 0 to 100. |
Dependent item | hpe.msa.controllers.cache.write["{#CONTROLLER.ID}",util] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Cache: Read hits, rate | For the controller that owns the volume, the number of times the block to be read is found in cache per second. |
Dependent item | hpe.msa.controllers.cache.read.hits["{#CONTROLLER.ID}",rate] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Cache: Read misses, rate | For the controller that owns the volume, the number of times the block to be read is not found in cache per second. |
Dependent item | hpe.msa.controllers.cache.read.misses["{#CONTROLLER.ID}",rate] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Cache: Write hits, rate | For the controller that owns the volume, the number of times the block written to is found in cache per second. |
Dependent item | hpe.msa.controllers.cache.write.hits["{#CONTROLLER.ID}",rate] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Cache: Write misses, rate | For the controller that owns the volume, the number of times the block written to is not found in cache per second. |
Dependent item | hpe.msa.controllers.cache.write.misses["{#CONTROLLER.ID}",rate] Preprocessing
|
Controller [{#CONTROLLER.ID}]: CPU utilization | Percentage of time the CPU is busy, from 0 to 100. |
Dependent item | hpe.msa.controllers.cpu["{#CONTROLLER.ID}",util] Preprocessing
|
Controller [{#CONTROLLER.ID}]: IOPS, total rate | Input/output operations per second, calculated over the interval since these statistics were last requested or reset. This value will be zero if it has not been requested or reset since a controller restart. |
Dependent item | hpe.msa.controllers.iops.total["{#CONTROLLER.ID}",rate] Preprocessing
|
Controller [{#CONTROLLER.ID}]: IOPS, read rate | Number of read operations per second. |
Dependent item | hpe.msa.controllers.iops.read["{#CONTROLLER.ID}",rate] Preprocessing
|
Controller [{#CONTROLLER.ID}]: IOPS, write rate | Number of write operations per second. |
Dependent item | hpe.msa.controllers.iops.write["{#CONTROLLER.ID}",rate] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Data transfer rate: Total | The data transfer rate, in bytes per second, calculated over the interval since these statistics were last requested or reset. This value will be zero if it has not been requested or reset since a controller restart. |
Dependent item | hpe.msa.controllers.data_transfer.total["{#CONTROLLER.ID}",rate] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Data transfer rate: Reads | The data read rate, in bytes per second. |
Dependent item | hpe.msa.controllers.data_transfer.reads["{#CONTROLLER.ID}",rate] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Data transfer rate: Writes | The data write rate, in bytes per second. |
Dependent item | hpe.msa.controllers.data_transfer.writes["{#CONTROLLER.ID}",rate] Preprocessing
|
Controller [{#CONTROLLER.ID}]: Uptime | Number of seconds since the controller was restarted. |
Dependent item | hpe.msa.controllers["{#CONTROLLER.ID}",uptime] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Controller [{#CONTROLLER.ID}]: Controller health is in degraded state | Controller health is in degraded state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.controllers["{#CONTROLLER.ID}",health])=1 |Warning |
Depends on:
|
|
Controller [{#CONTROLLER.ID}]: Controller health is in fault state | Controller health is in fault state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.controllers["{#CONTROLLER.ID}",health])=2 |Average |
Depends on:
|
|
Controller [{#CONTROLLER.ID}]: Controller health is in unknown state | Controller health is in unknown state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.controllers["{#CONTROLLER.ID}",health])=3 |Info |
Depends on:
|
|
Controller [{#CONTROLLER.ID}]: Controller is down | The controller is down. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.controllers["{#CONTROLLER.ID}",status])=1 |High |
||
Controller [{#CONTROLLER.ID}]: High CPU utilization | Controller CPU utilization is too high. The system might be slow to respond. |
min(/HPE MSA 2040 Storage by HTTP/hpe.msa.controllers.cpu["{#CONTROLLER.ID}",util],5m)>{$HPE.MSA.CONTROLLER.CPU.UTIL.CRIT} |Warning |
||
Controller [{#CONTROLLER.ID}]: Controller has been restarted | The controller uptime is less than 10 minutes. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.controllers["{#CONTROLLER.ID}",uptime])<10m |Warning |
Name | Description | Type | Key and additional info |
---|---|---|---|
Disk groups discovery | Discover disk groups. |
Dependent item | hpe.msa.disks.groups.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Disk group [{#NAME}]: Get data | The discovered disk group data. |
Dependent item | hpe.msa.get.disks.groups["{#NAME}",data] Preprocessing
|
Disk group [{#NAME}]: Get statistics data | The discovered disk group statistics data. |
Dependent item | hpe.msa.get.disks.groups.statistics["{#NAME}",data] Preprocessing
|
Disk group [{#NAME}]: Disks count | Number of disks in the disk group. |
Dependent item | hpe.msa.disks.groups["{#NAME}",disk_count] Preprocessing
|
Disk group [{#NAME}]: Pool space used | The percentage of pool capacity that the disk group occupies. |
Dependent item | hpe.msa.disks.groups.space["{#NAME}",pool_util] Preprocessing
|
Disk group [{#NAME}]: Health | Disk group health. |
Dependent item | hpe.msa.disks.groups["{#NAME}",health] Preprocessing
|
Disk group [{#NAME}]: Space free | The free space in the disk group. |
Dependent item | hpe.msa.disks.groups.space["{#NAME}",free] Preprocessing
|
Disk group [{#NAME}]: Space total | The capacity of the disk group. |
Dependent item | hpe.msa.disks.groups.space["{#NAME}",total] Preprocessing
|
Disk group [{#NAME}]: Space utilization | The space utilization percentage in the disk group. |
Calculated | hpe.msa.disks.groups.space["{#NAME}",util] Preprocessing
|
Disk group [{#NAME}]: RAID type | The RAID level of the disk group. |
Dependent item | hpe.msa.disks.groups.raid["{#NAME}",type] Preprocessing
|
Disk group [{#NAME}]: Status | The status of the disk group: - CRIT: Critical. The disk group is online but isn't fault tolerant because some of it's disks are down. - DMGD: Damaged. The disk group is online and fault tolerant, but some of it's disks are damaged. - FTDN: Fault tolerant with a down disk.The disk group is online and fault tolerant, but some of it's disks are down. - FTOL: Fault tolerant. - MSNG: Missing. The disk group is online and fault tolerant, but some of it's disks are missing. - OFFL: Offline. Either the disk group is using offline initialization, or it's disks are down and data may be lost. - QTCR: Quarantined critical. The disk group is critical with at least one inaccessible disk. For example, two disks are inaccessible in a RAID 6 disk group or one disk is inaccessible for other fault-tolerant RAID levels. If the inaccessible disks come online or if after 60 seconds from being quarantined the disk group is QTCRor QTDN, the disk group is automatically dequarantined. - QTDN: Quarantined with a down disk. The RAID6 disk group has one inaccessible disk. The disk group is fault tolerant but degraded. If the inaccessible disks come online or if after 60 seconds from being quarantined the disk group is QTCRor QTDN, the disk group is automatically dequarantined. - QTOF: Quarantined offline. The disk group is offline with multiple inaccessible disks causing user data to be incomplete, or is an NRAID or RAID 0 disk group. - QTUN: Quarantined unsupported. The disk group contains data in a format that is not supported by this system. For example, this system does not support linear disk groups. - STOP: The disk group is stopped. - UNKN: Unknown. - UP: Up. The disk group is online and does not have fault-tolerant attributes. |
Dependent item | hpe.msa.disks.groups["{#NAME}",status] Preprocessing
|
Disk group [{#NAME}]: IOPS, total rate | Input/output operations per second, calculated over the interval since these statistics were last requested or reset. This value will be zero if it has not been requested or reset since a controller restart. |
Dependent item | hpe.msa.disks.groups.iops.total["{#NAME}",rate] Preprocessing
|
Disk group [{#NAME}]: Average response time: Total | Average response time for read and write operations, calculated over the interval since these statistics were last requested or reset. |
Dependent item | hpe.msa.disks.groups.avgrsptime["{#NAME}",total] Preprocessing
|
Disk group [{#NAME}]: Average response time: Read | Average response time for all read operations, calculated over the interval since these statistics were last requested or reset. |
Dependent item | hpe.msa.disks.groups.avgrsptime["{#NAME}",read] Preprocessing
|
Disk group [{#NAME}]: Average response time: Write | Average response time for all write operations, calculated over the interval since these statistics were last requested or reset. |
Dependent item | hpe.msa.disks.groups.avgrsptime["{#NAME}",write] Preprocessing
|
Disk group [{#NAME}]: IOPS, read rate | Number of read operations per second. |
Dependent item | hpe.msa.disks.groups.iops.read["{#NAME}",rate] Preprocessing
|
Disk group [{#NAME}]: IOPS, write rate | Number of write operations per second. |
Dependent item | hpe.msa.disks.groups.iops.write["{#NAME}",rate] Preprocessing
|
Disk group [{#NAME}]: Data transfer rate: Total | The data transfer rate, in bytes per second, calculated over the interval since these statistics were last requested or reset. This value will be zero if it has not been requested or reset since a controller restart. |
Dependent item | hpe.msa.disks.groups.data_transfer.total["{#NAME}",rate] Preprocessing
|
Disk group [{#NAME}]: Data transfer rate: Reads | The data read rate, in bytes per second. |
Dependent item | hpe.msa.disks.groups.data_transfer.reads["{#NAME}",rate] Preprocessing
|
Disk group [{#NAME}]: Data transfer rate: Writes | The data write rate, in bytes per second. |
Dependent item | hpe.msa.disks.groups.data_transfer.writes["{#NAME}",rate] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Disk group [{#NAME}]: Disk group health is in degraded state | Disk group health is in degraded state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",health])=1 |Warning |
||
Disk group [{#NAME}]: Disk group health is in fault state | Disk group health is in fault state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",health])=2 |Average |
||
Disk group [{#NAME}]: Disk group health is in unknown state | Disk group health is in unknown state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",health])=3 |Info |
||
Disk group [{#NAME}]: Disk group space is low | Disk group is running low on free space (less than {$HPE.MSA.DISKS.GROUP.PUSED.MAX.WARN:"{#NAME}"}% available). |
min(/HPE MSA 2040 Storage by HTTP/hpe.msa.disks.groups.space["{#NAME}",util],5m)>{$HPE.MSA.DISKS.GROUP.PUSED.MAX.WARN:"{#NAME}"} |Warning |
Depends on:
|
|
Disk group [{#NAME}]: Disk group space is critically low | Disk group is running low on free space (less than {$HPE.MSA.DISKS.GROUP.PUSED.MAX.CRIT:"{#NAME}"}% available). |
min(/HPE MSA 2040 Storage by HTTP/hpe.msa.disks.groups.space["{#NAME}",util],5m)>{$HPE.MSA.DISKS.GROUP.PUSED.MAX.CRIT:"{#NAME}"} |Average |
||
Disk group [{#NAME}]: Disk group is fault tolerant with a down disk | The disk group is online and fault tolerant, but some of it's disks are down. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",status])=1 |Average |
||
Disk group [{#NAME}]: Disk group has damaged disks | The disk group is online and fault tolerant, but some of it's disks are damaged. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",status])=9 |Average |
||
Disk group [{#NAME}]: Disk group has missing disks | The disk group is online and fault tolerant, but some of it's disks are missing. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",status])=8 |Average |
||
Disk group [{#NAME}]: Disk group is offline | Either the disk group is using offline initialization, or it's disks are down and data may be lost. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",status])=3 |Average |
||
Disk group [{#NAME}]: Disk group is quarantined critical | The disk group is critical with at least one inaccessible disk. For example, two disks are inaccessible in a RAID 6 disk group or one disk is inaccessible for other fault-tolerant RAID levels. If the inaccessible disks come online or if after 60 seconds from being quarantined the disk group is QTCRor QTDN, the disk group is automatically dequarantined. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",status])=4 |Average |
||
Disk group [{#NAME}]: Disk group is quarantined offline | The disk group is offline with multiple inaccessible disks causing user data to be incomplete, or is an NRAID or RAID 0 disk group. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",status])=5 |Average |
||
Disk group [{#NAME}]: Disk group is quarantined unsupported | The disk group contains data in a format that is not supported by this system. For example, this system does not support linear disk groups. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",status])=5 |Average |
||
Disk group [{#NAME}]: Disk group is quarantined with an inaccessible disk | The RAID6 disk group has one inaccessible disk. The disk group is fault tolerant but degraded. If the inaccessible disks come online or if after 60 seconds from being quarantined the disk group is QTCRor QTDN, the disk group is automatically dequarantined. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",status])=6 |Average |
||
Disk group [{#NAME}]: Disk group is stopped | The disk group is stopped. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",status])=7 |Average |
||
Disk group [{#NAME}]: Disk group status is critical | The disk group is online but isn't fault tolerant because some of its disks are down. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.disks.groups["{#NAME}",status])=2 |Average |
Name | Description | Type | Key and additional info |
---|---|---|---|
Pools discovery | Discover pools. |
Dependent item | hpe.msa.pools.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Pool [{#NAME}]: Get data | The discovered pool data. |
Dependent item | hpe.msa.get.pools["{#NAME}",data] Preprocessing
|
Pool [{#NAME}]: Health | Pool health. |
Dependent item | hpe.msa.pools["{#NAME}",health] Preprocessing
|
Pool [{#NAME}]: Space free | The free space in the pool. |
Dependent item | hpe.msa.pools.space["{#NAME}",free] Preprocessing
|
Pool [{#NAME}]: Space total | The capacity of the pool. |
Dependent item | hpe.msa.pools.space["{#NAME}",total] Preprocessing
|
Pool [{#NAME}]: Space utilization | The space utilization percentage in the pool. |
Calculated | hpe.msa.pools.space["{#NAME}",util] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Pool [{#NAME}]: Pool health is in degraded state | Pool health is in degraded state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.pools["{#NAME}",health])=1 |Warning |
||
Pool [{#NAME}]: Pool health is in fault state | Pool health is in fault state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.pools["{#NAME}",health])=2 |Average |
||
Pool [{#NAME}]: Pool health is in unknown state | Pool [{#NAME}] health is in unknown state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.pools["{#NAME}",health])=3 |Info |
||
Pool [{#NAME}]: Pool space is low | Pool is running low on free space (less than {$HPE.MSA.POOL.PUSED.MAX.WARN:"{#NAME}"}% available). |
min(/HPE MSA 2040 Storage by HTTP/hpe.msa.pools.space["{#NAME}",util],5m)>{$HPE.MSA.POOL.PUSED.MAX.WARN:"{#NAME}"} |Warning |
Depends on:
|
|
Pool [{#NAME}]: Pool space is critically low | Pool is running low on free space (less than {$HPE.MSA.POOL.PUSED.MAX.CRIT:"{#NAME}"}% available). |
min(/HPE MSA 2040 Storage by HTTP/hpe.msa.pools.space["{#NAME}",util],5m)>{$HPE.MSA.POOL.PUSED.MAX.CRIT:"{#NAME}"} |Average |
Name | Description | Type | Key and additional info |
---|---|---|---|
Volumes discovery | Discover volumes. |
Dependent item | hpe.msa.volumes.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Volume [{#NAME}]: Get data | The discovered volume data. |
Dependent item | hpe.msa.get.volumes["{#NAME}",data] Preprocessing
|
Volume [{#NAME}]: Health | Volume health status. |
Dependent item | hpe.msa.volumes["{#DURABLE.ID}",health] Preprocessing
|
Volume [{#NAME}]: Get statistics data | The discovered volume statistics data. |
Dependent item | hpe.msa.get.volumes.statistics["{#NAME}",data] Preprocessing
|
Volume [{#NAME}]: Space allocated | The amount of space currently allocated to the volume. |
Dependent item | hpe.msa.volumes.space["{#NAME}",allocated] Preprocessing
|
Volume [{#NAME}]: Space total | The capacity of the volume. |
Dependent item | hpe.msa.volumes.space["{#NAME}",total] Preprocessing
|
Volume [{#NAME}]: IOPS, total rate | Total input/output operations per second, calculated over the interval since these statistics were last requested or reset. This value will be zero if it has not been requested or reset since a controller restart. |
Dependent item | hpe.msa.volumes.iops.total["{#NAME}",rate] Preprocessing
|
Volume [{#NAME}]: IOPS, read rate | Number of read operations per second. |
Dependent item | hpe.msa.volumes.iops.read["{#NAME}",rate] Preprocessing
|
Volume [{#NAME}]: IOPS, write rate | Number of write operations per second. |
Dependent item | hpe.msa.volumes.iops.write["{#NAME}",rate] Preprocessing
|
Volume [{#NAME}]: Data transfer rate: Total | The data transfer rate, in bytes per second, calculated over the interval since these statistics were last requested or reset. This value will be zero if it has not been requested or reset since a controller restart. |
Dependent item | hpe.msa.volumes.data_transfer.total["{#NAME}",rate] Preprocessing
|
Volume [{#NAME}]: Data transfer rate: Reads | The data read rate, in bytes per second. |
Dependent item | hpe.msa.volumes.data_transfer.reads["{#NAME}",rate] Preprocessing
|
Volume [{#NAME}]: Data transfer rate: Writes | The data write rate, in bytes per second. |
Dependent item | hpe.msa.volumes.data_transfer.writes["{#NAME}",rate] Preprocessing
|
Volume [{#NAME}]: Cache: Read hits, rate | For the controller that owns the volume, the number of times the block to be read is found in cache per second. |
Dependent item | hpe.msa.volumes.cache.read.hits["{#NAME}",rate] Preprocessing
|
Volume [{#NAME}]: Cache: Read misses, rate | For the controller that owns the volume, the number of times the block to be read is not found in cache per second. |
Dependent item | hpe.msa.volumes.cache.read.misses["{#NAME}",rate] Preprocessing
|
Volume [{#NAME}]: Cache: Write hits, rate | For the controller that owns the volume, the number of times the block written to is found in cache per second. |
Dependent item | hpe.msa.volumes.cache.write.hits["{#NAME}",rate] Preprocessing
|
Volume [{#NAME}]: Cache: Write misses, rate | For the controller that owns the volume, the number of times the block written to is not found in cache per second. |
Dependent item | hpe.msa.volumes.cache.write.misses["{#NAME}",rate] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Volume [{#NAME}]: Volume health is in degraded state | Volume health is in degraded state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.volumes["{#DURABLE.ID}",health])=1 |Warning |
||
Volume [{#NAME}]: Volume health is in fault state | Volume health is in fault state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.volumes["{#DURABLE.ID}",health])=2 |Average |
||
Volume [{#NAME}]: Volume health is in unknown state | Volume health is in unknown state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.volumes["{#DURABLE.ID}",health])=3 |Info |
Name | Description | Type | Key and additional info |
---|---|---|---|
Enclosures discovery | Discover enclosures. |
Dependent item | hpe.msa.enclosures.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Enclosure [{#DURABLE.ID}]: Get data | The discovered enclosure data. |
Dependent item | hpe.msa.get.enclosures["{#DURABLE.ID}",data] Preprocessing
|
Enclosure [{#DURABLE.ID}]: Health | Enclosure health. |
Dependent item | hpe.msa.enclosures["{#DURABLE.ID}",health] Preprocessing
|
Enclosure [{#DURABLE.ID}]: Status | Enclosure status. |
Dependent item | hpe.msa.enclosures["{#DURABLE.ID}",status] Preprocessing
|
Enclosure [{#DURABLE.ID}]: Midplane serial number | Midplane serial number. |
Dependent item | hpe.msa.enclosures["{#DURABLE.ID}",midplaneserialnumber] Preprocessing
|
Enclosure [{#DURABLE.ID}]: Part number | Enclosure part number. |
Dependent item | hpe.msa.enclosures["{#DURABLE.ID}",part_number] Preprocessing
|
Enclosure [{#DURABLE.ID}]: Model | Enclosure model. |
Dependent item | hpe.msa.enclosures["{#DURABLE.ID}",model] Preprocessing
|
Enclosure [{#DURABLE.ID}]: Power | Enclosure power in watts. |
Dependent item | hpe.msa.enclosures["{#DURABLE.ID}",power] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Enclosure [{#DURABLE.ID}]: Enclosure health is in degraded state | Enclosure health is in degraded state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.enclosures["{#DURABLE.ID}",health])=1 |Warning |
||
Enclosure [{#DURABLE.ID}]: Enclosure health is in fault state | Enclosure health is in fault state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.enclosures["{#DURABLE.ID}",health])=2 |Average |
||
Enclosure [{#DURABLE.ID}]: Enclosure health is in unknown state | Enclosure health is in unknown state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.enclosures["{#DURABLE.ID}",health])=3 |Info |
||
Enclosure [{#DURABLE.ID}]: Enclosure has critical status | Enclosure has critical status. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.enclosures["{#DURABLE.ID}",status])=2 |High |
||
Enclosure [{#DURABLE.ID}]: Enclosure has warning status | Enclosure has warning status. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.enclosures["{#DURABLE.ID}",status])=3 |Warning |
||
Enclosure [{#DURABLE.ID}]: Enclosure is unavailable | Enclosure is unavailable. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.enclosures["{#DURABLE.ID}",status])=7 |High |
||
Enclosure [{#DURABLE.ID}]: Enclosure is unrecoverable | Enclosure is unrecoverable. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.enclosures["{#DURABLE.ID}",status])=4 |High |
||
Enclosure [{#DURABLE.ID}]: Enclosure has unknown status | Enclosure has unknown status. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.enclosures["{#DURABLE.ID}",status])=6 |Info |
Name | Description | Type | Key and additional info |
---|---|---|---|
Power supplies discovery | Discover power supplies. |
Dependent item | hpe.msa.power_supplies.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Power supply [{#DURABLE.ID}]: Get data | The discovered power supply data. |
Dependent item | hpe.msa.get.power_supplies["{#DURABLE.ID}",data] Preprocessing
|
Power supply [{#DURABLE.ID}]: Health | Power supply health status. |
Dependent item | hpe.msa.power_supplies["{#DURABLE.ID}",health] Preprocessing
|
Power supply [{#DURABLE.ID}]: Status | Power supply status. |
Dependent item | hpe.msa.power_supplies["{#DURABLE.ID}",status] Preprocessing
|
Power supply [{#DURABLE.ID}]: Part number | Power supply part number. |
Dependent item | hpe.msa.powersupplies["{#DURABLE.ID}",partnumber] Preprocessing
|
Power supply [{#DURABLE.ID}]: Serial number | Power supply serial number. |
Dependent item | hpe.msa.powersupplies["{#DURABLE.ID}",serialnumber] Preprocessing
|
Power supply [{#DURABLE.ID}]: Temperature | Power supply temperature. |
Dependent item | hpe.msa.power_supplies["{#DURABLE.ID}",temperature] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Power supply [{#DURABLE.ID}]: Power supply health is in degraded state | Power supply health is in degraded state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.power_supplies["{#DURABLE.ID}",health])=1 |Warning |
||
Power supply [{#DURABLE.ID}]: Power supply health is in fault state | Power supply health is in fault state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.power_supplies["{#DURABLE.ID}",health])=2 |Average |
||
Power supply [{#DURABLE.ID}]: Power supply health is in unknown state | Power supply health is in unknown state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.power_supplies["{#DURABLE.ID}",health])=3 |Info |
||
Power supply [{#DURABLE.ID}]: Power supply has error status | Power supply has error status. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.power_supplies["{#DURABLE.ID}",status])=2 |Average |
||
Power supply [{#DURABLE.ID}]: Power supply has warning status | Power supply has warning status. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.power_supplies["{#DURABLE.ID}",status])=1 |Warning |
||
Power supply [{#DURABLE.ID}]: Power supply has unknown status | Power supply has unknown status. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.power_supplies["{#DURABLE.ID}",status])=4 |Info |
Name | Description | Type | Key and additional info |
---|---|---|---|
Ports discovery | Discover ports. |
Dependent item | hpe.msa.ports.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Port [{#NAME}]: Get data | The discovered port data. |
Dependent item | hpe.msa.get.ports["{#NAME}",,data] Preprocessing
|
Port [{#NAME}]: Health | Port health status. |
Dependent item | hpe.msa.ports["{#NAME}",health] Preprocessing
|
Port [{#NAME}]: Status | Port status. |
Dependent item | hpe.msa.ports["{#NAME}",status] Preprocessing
|
Port [{#NAME}]: Type | Port type. |
Dependent item | hpe.msa.ports["{#NAME}",type] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Port [{#NAME}]: Port health is in degraded state | Port health is in degraded state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.ports["{#NAME}",health])=1 |Warning |
||
Port [{#NAME}]: Port health is in fault state | Port health is in fault state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.ports["{#NAME}",health])=2 |Average |
||
Port [{#NAME}]: Port health is in unknown state | Port health is in unknown state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.ports["{#NAME}",health])=3 |Info |
||
Port [{#NAME}]: Port has error status | Port has error status. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.ports["{#NAME}",status])=2 |Average |
||
Port [{#NAME}]: Port has warning status | Port has warning status. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.ports["{#NAME}",status])=1 |Warning |
||
Port [{#NAME}]: Port has unknown status | Port has unknown status. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.ports["{#NAME}",status])=4 |Info |
Name | Description | Type | Key and additional info |
---|---|---|---|
Fans discovery | Discover fans. |
Dependent item | hpe.msa.fans.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Fan [{#DURABLE.ID}]: Get data | The discovered fan data. |
Dependent item | hpe.msa.get.fans["{#DURABLE.ID}",data] Preprocessing
|
Fan [{#DURABLE.ID}]: Health | Fan health status. |
Dependent item | hpe.msa.fans["{#DURABLE.ID}",health] Preprocessing
|
Fan [{#DURABLE.ID}]: Status | Fan status. |
Dependent item | hpe.msa.fans["{#DURABLE.ID}",status] Preprocessing
|
Fan [{#DURABLE.ID}]: Speed | Fan speed (revolutions per minute). |
Dependent item | hpe.msa.fans["{#DURABLE.ID}",speed] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Fan [{#DURABLE.ID}]: Fan health is in degraded state | Fan health is in degraded state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.fans["{#DURABLE.ID}",health])=1 |Warning |
||
Fan [{#DURABLE.ID}]: Fan health is in fault state | Fan health is in fault state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.fans["{#DURABLE.ID}",health])=2 |Average |
||
Fan [{#DURABLE.ID}]: Fan health is in unknown state | Fan health is in unknown state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.fans["{#DURABLE.ID}",health])=3 |Info |
||
Fan [{#DURABLE.ID}]: Fan has error status | Fan has error status. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.fans["{#DURABLE.ID}",status])=1 |Average |
||
Fan [{#DURABLE.ID}]: Fan is missing | Fan is missing. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.fans["{#DURABLE.ID}",status])=3 |Info |
||
Fan [{#DURABLE.ID}]: Fan is off | Fan is off. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.fans["{#DURABLE.ID}",status])=2 |Warning |
Name | Description | Type | Key and additional info |
---|---|---|---|
Disks discovery | Discover disks. |
Dependent item | hpe.msa.disks.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
Disk [{#DURABLE.ID}]: Get data | The discovered disk data. |
Dependent item | hpe.msa.get.disks["{#DURABLE.ID}",data] Preprocessing
|
Disk [{#DURABLE.ID}]: Health | Disk health status. |
Dependent item | hpe.msa.disks["{#DURABLE.ID}",health] Preprocessing
|
Disk [{#DURABLE.ID}]: Temperature status | Disk temperature status. |
Dependent item | hpe.msa.disks["{#DURABLE.ID}",temperature_status] Preprocessing
|
Disk [{#DURABLE.ID}]: Temperature | Temperature of the disk. |
Dependent item | hpe.msa.disks["{#DURABLE.ID}",temperature] Preprocessing
|
Disk [{#DURABLE.ID}]: Type | Disk type: SAS: Enterprise SAS spinning disk. SAS MDL: Midline SAS spinning disk. SSD SAS: SAS solit-state disk. |
Dependent item | hpe.msa.disks["{#DURABLE.ID}",type] Preprocessing
|
Disk [{#DURABLE.ID}]: Disk group | If the disk is in a disk group, the disk group name. |
Dependent item | hpe.msa.disks["{#DURABLE.ID}",group] Preprocessing
|
Disk [{#DURABLE.ID}]: Storage pool | If the disk is in a pool, the pool name. |
Dependent item | hpe.msa.disks["{#DURABLE.ID}",pool] Preprocessing
|
Disk [{#DURABLE.ID}]: Vendor | Disk vendor. |
Dependent item | hpe.msa.disks["{#DURABLE.ID}",vendor] Preprocessing
|
Disk [{#DURABLE.ID}]: Model | Disk model. |
Dependent item | hpe.msa.disks["{#DURABLE.ID}",model] Preprocessing
|
Disk [{#DURABLE.ID}]: Serial number | Disk serial number. |
Dependent item | hpe.msa.disks["{#DURABLE.ID}",serial_number] Preprocessing
|
Disk [{#DURABLE.ID}]: Space total | Total size of the disk. |
Dependent item | hpe.msa.disks.space["{#DURABLE.ID}",total] Preprocessing
|
Disk [{#DURABLE.ID}]: SSD life left | The percentage of disk life remaining. |
Dependent item | hpe.msa.disks.ssd["{#DURABLE.ID}",life_left] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
Disk [{#DURABLE.ID}]: Disk health is in degraded state | Disk health is in degraded state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.disks["{#DURABLE.ID}",health])=1 |Warning |
||
Disk [{#DURABLE.ID}]: Disk health is in fault state | Disk health is in fault state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.disks["{#DURABLE.ID}",health])=2 |Average |
||
Disk [{#DURABLE.ID}]: Disk health is in unknown state | Disk health is in unknown state. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.disks["{#DURABLE.ID}",health])=3 |Info |
||
Disk [{#DURABLE.ID}]: Disk temperature is high | Disk temperature is high. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.disks["{#DURABLE.ID}",temperature_status])=3 |Warning |
||
Disk [{#DURABLE.ID}]: Disk temperature is critically high | Disk temperature is critically high. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.disks["{#DURABLE.ID}",temperature_status])=2 |Average |
||
Disk [{#DURABLE.ID}]: Disk temperature is unknown | Disk temperature is unknown. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.disks["{#DURABLE.ID}",temperature_status])=4 |Info |
Name | Description | Type | Key and additional info |
---|---|---|---|
FRU discovery | Discover FRU. |
Dependent item | hpe.msa.frus.discovery |
Name | Description | Type | Key and additional info |
---|---|---|---|
FRU [{#ENCLOSURE.ID}: {#LOCATION}]: Get data | The discovered FRU data. |
Dependent item | hpe.msa.get.frus["{#ENCLOSURE.ID}:{#LOCATION}",data] Preprocessing
|
FRU [{#ENCLOSURE.ID}: {#LOCATION}]: Status | {#DESCRIPTION}. FRU status: Absent: Component is not present. Fault: At least one subcomponent has a fault. Invalid data: For a power supply module, the EEPROM is improperly programmed. OK: All subcomponents are operating normally. Not available: Status is not available. |
Dependent item | hpe.msa.frus["{#ENCLOSURE.ID}:{#LOCATION}",status] Preprocessing
|
FRU [{#ENCLOSURE.ID}: {#LOCATION}]: Part number | {#DESCRIPTION}. Part number of the FRU. |
Dependent item | hpe.msa.frus["{#ENCLOSURE.ID}:{#LOCATION}",part_number] Preprocessing
|
FRU [{#ENCLOSURE.ID}: {#LOCATION}]: Serial number | {#DESCRIPTION}. FRU serial number. |
Dependent item | hpe.msa.frus["{#ENCLOSURE.ID}:{#LOCATION}",serial_number] Preprocessing
|
Name | Description | Expression | Severity | Dependencies and additional info |
---|---|---|---|---|
FRU [{#ENCLOSURE.ID}: {#LOCATION}]: FRU status is Degraded or Fault | FRU status is Degraded or Fault. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.frus["{#ENCLOSURE.ID}:{#LOCATION}",status])=1 |Average |
||
FRU [{#ENCLOSURE.ID}: {#LOCATION}]: FRU ID data is invalid | The FRU ID data is invalid. The FRU's EEPROM is improperly programmed. |
last(/HPE MSA 2040 Storage by HTTP/hpe.msa.frus["{#ENCLOSURE.ID}:{#LOCATION}",status])=0 |Warning |
Please report any issues with the template at https://support.zabbix.com
You can also provide feedback, discuss the template, or ask for help at ZABBIX forums