Skip to main content

ZFS Pools

Plugin: go.d.plugin Module: zfspool

Overview

This collector monitors the health and space usage of ZFS pools using the command line tool zpool.

This collector is only supported on the following platforms:

  • Linux
  • BSD

This collector only supports collecting metrics from a single instance of this integration.

Default Behavior

Auto-Detection

This integration doesn't support auto-detection.

Limits

The default configuration for this integration does not impose any limits on data collection.

Performance Impact

The default configuration for this integration is not expected to impose a significant performance impact on the system.

Metrics

Metrics grouped by scope.

The scope defines the instance that the metric belongs to. An instance is uniquely identified by a set of labels.

Per zfs pool

These metrics refer to the ZFS pool.

Labels:

LabelDescription
poolZpool name

Metrics:

MetricDimensionsUnit
zfspool.pool_space_utilizationutilization%
zfspool.pool_space_usagefree, usedbytes
zfspool.pool_fragmentationfragmentation%
zfspool.pool_health_stateonline, degraded, faulted, offline, unavail, removed, suspendedstate

Per zfs pool vdev

These metrics refer to the ZFS pool virtual device.

Labels:

LabelDescription
poolZpool name
vdevUnique identifier for a virtual device (vdev) within a ZFS pool.

Metrics:

MetricDimensionsUnit
zfspool.vdev_health_stateonline, degraded, faulted, offline, unavail, removed, suspendedstate

Alerts

The following alerts are available:

Alert nameOn metricDescription
zfs_pool_space_utilization zfspool.pool_space_utilizationZFS pool ${label:pool} is nearing capacity. Current space usage is above the threshold.
zfs_pool_health_state_warn zfspool.pool_health_stateZFS pool ${label:pool} state is degraded
zfs_pool_health_state_crit zfspool.pool_health_stateZFS pool ${label:pool} state is faulted or unavail
zfs_vdev_health_state zfspool.vdev_health_stateZFS vdev ${label:vdev} state is faulted or degraded

Setup

Prerequisites

No action required.

Configuration

File

The configuration file name for this integration is go.d/zfspool.conf.

You can edit the configuration file using the edit-config script from the Netdata config directory.

cd /etc/netdata 2>/dev/null || cd /opt/netdata/etc/netdata
sudo ./edit-config go.d/zfspool.conf

Options

The following options can be defined globally: update_every.

Config options
NameDescriptionDefaultRequired
update_everyData collection frequency.10no
binary_pathPath to the zpool binary. If an absolute path is provided, the collector will use it directly; otherwise, it will search for the binary in directories specified in the PATH environment variable./usr/bin/zpoolyes
timeoutTimeout for executing the binary, specified in seconds.2no

Examples

Custom binary path

The executable is not in the directories specified in the PATH environment variable.

Config
jobs:
- name: zfspool
binary_path: /usr/local/sbin/zpool

Troubleshooting

Debug Mode

Important: Debug mode is not supported for data collection jobs created via the UI using the Dyncfg feature.

To troubleshoot issues with the zfspool collector, run the go.d.plugin with the debug option enabled. The output should give you clues as to why the collector isn't working.

  • Navigate to the plugins.d directory, usually at /usr/libexec/netdata/plugins.d/. If that's not the case on your system, open netdata.conf and look for the plugins setting under [directories].

    cd /usr/libexec/netdata/plugins.d/
  • Switch to the netdata user.

    sudo -u netdata -s
  • Run the go.d.plugin to debug the collector:

    ./go.d.plugin -d -m zfspool

Getting Logs

If you're encountering problems with the zfspool collector, follow these steps to retrieve logs and identify potential issues:

  • Run the command specific to your system (systemd, non-systemd, or Docker container).
  • Examine the output for any warnings or error messages that might indicate issues. These messages should provide clues about the root cause of the problem.

System with systemd

Use the following command to view logs generated since the last Netdata service restart:

journalctl _SYSTEMD_INVOCATION_ID="$(systemctl show --value --property=InvocationID netdata)" --namespace=netdata --grep zfspool

System without systemd

Locate the collector log file, typically at /var/log/netdata/collector.log, and use grep to filter for collector's name:

grep zfspool /var/log/netdata/collector.log

Note: This method shows logs from all restarts. Focus on the latest entries for troubleshooting current issues.

Docker Container

If your Netdata runs in a Docker container named "netdata" (replace if different), use this command:

docker logs netdata 2>&1 | grep zfspool

Do you have any feedback for this page? If so, you can open a new issue on our netdata/learn repository.