2020-10-09 16:29:21 +00:00
|
|
|
# Influxdb Metrics for ZFS Pools
|
|
|
|
The _zpool_influxdb_ program produces
|
|
|
|
[influxdb](https://github.com/influxdata/influxdb) line protocol
|
|
|
|
compatible metrics from zpools. In the UNIX tradition, _zpool_influxdb_
|
|
|
|
does one thing: read statistics from a pool and print them to
|
|
|
|
stdout. In many ways, this is a metrics-friendly output of
|
|
|
|
statistics normally observed via the `zpool` command.
|
|
|
|
|
|
|
|
## Usage
|
|
|
|
When run without arguments, _zpool_influxdb_ runs once, reading data
|
|
|
|
from all imported pools, and prints to stdout.
|
|
|
|
```shell
|
|
|
|
zpool_influxdb [options] [poolname]
|
|
|
|
```
|
|
|
|
If no poolname is specified, then all pools are sampled.
|
|
|
|
|
|
|
|
| option | short option | description |
|
|
|
|
|---|---|---|
|
|
|
|
| --execd | -e | For use with telegraf's `execd` plugin. When [enter] is pressed, the pools are sampled. To exit, use [ctrl+D] |
|
|
|
|
| --no-histogram | -n | Do not print histogram information |
|
|
|
|
| --signed-int | -i | Use signed integer data type (default=unsigned) |
|
|
|
|
| --sum-histogram-buckets | -s | Sum histogram bucket values |
|
|
|
|
| --tags key=value[,key=value...] | -t | Add tags to data points. No tag sanity checking is performed. |
|
|
|
|
| --help | -h | Print a short usage message |
|
|
|
|
|
|
|
|
#### Histogram Bucket Values
|
|
|
|
The histogram data collected by ZFS is stored as independent bucket values.
|
|
|
|
This works well out-of-the-box with an influxdb data source and grafana's
|
|
|
|
heatmap visualization. The influxdb query for a grafana heatmap
|
|
|
|
visualization looks like:
|
|
|
|
```
|
|
|
|
field(disk_read) last() non_negative_derivative(1s)
|
|
|
|
```
|
|
|
|
|
|
|
|
Another method for storing histogram data sums the values for lower-value
|
|
|
|
buckets. For example, a latency bucket tagged "le=10" includes the values
|
|
|
|
in the bucket "le=1".
|
|
|
|
This method is often used for prometheus histograms.
|
|
|
|
The `zpool_influxdb --sum-histogram-buckets` option presents the data from ZFS
|
|
|
|
as summed values.
|
|
|
|
|
|
|
|
## Measurements
|
|
|
|
The following measurements are collected:
|
|
|
|
|
|
|
|
| measurement | description | zpool equivalent |
|
|
|
|
|---|---|---|
|
|
|
|
| zpool_stats | general size and data | zpool list |
|
|
|
|
| zpool_scan_stats | scrub, rebuild, and resilver statistics (omitted if no scan has been requested) | zpool status |
|
|
|
|
| zpool_vdev_stats | per-vdev statistics | zpool iostat -q |
|
|
|
|
| zpool_io_size | per-vdev I/O size histogram | zpool iostat -r |
|
|
|
|
| zpool_latency | per-vdev I/O latency histogram | zpool iostat -w |
|
|
|
|
| zpool_vdev_queue | per-vdev instantaneous queue depth | zpool iostat -q |
|
|
|
|
|
|
|
|
### zpool_stats Description
|
|
|
|
zpool_stats contains top-level summary statistics for the pool.
|
|
|
|
Performance counters measure the I/Os to the pool's devices.
|
|
|
|
|
|
|
|
#### zpool_stats Tags
|
|
|
|
|
|
|
|
| label | description |
|
|
|
|
|---|---|
|
|
|
|
| name | pool name |
|
|
|
|
| path | for leaf vdevs, the pathname |
|
|
|
|
| state | pool state, as shown by _zpool status_ |
|
|
|
|
| vdev | vdev name (root = entire pool) |
|
|
|
|
|
|
|
|
#### zpool_stats Fields
|
|
|
|
|
|
|
|
| field | units | description |
|
|
|
|
|---|---|---|
|
|
|
|
| alloc | bytes | allocated space |
|
|
|
|
| free | bytes | unallocated space |
|
|
|
|
| size | bytes | total pool size |
|
|
|
|
| read_bytes | bytes | bytes read since pool import |
|
|
|
|
| read_errors | count | number of read errors |
|
|
|
|
| read_ops | count | number of read operations |
|
|
|
|
| write_bytes | bytes | bytes written since pool import |
|
|
|
|
| write_errors | count | number of write errors |
|
|
|
|
| write_ops | count | number of write operations |
|
|
|
|
|
|
|
|
### zpool_scan_stats Description
|
|
|
|
Once a pool has been scrubbed, resilvered, or rebuilt, the zpool_scan_stats
|
|
|
|
contain information about the status and performance of the operation.
|
|
|
|
Otherwise, the zpool_scan_stats do not exist in the kernel, and therefore
|
|
|
|
cannot be reported by this collector.
|
|
|
|
|
|
|
|
#### zpool_scan_stats Tags
|
|
|
|
|
|
|
|
| label | description |
|
|
|
|
|---|---|
|
|
|
|
| name | pool name |
|
|
|
|
| function | name of the scan function running or recently completed |
|
|
|
|
| state | scan state, as shown by _zpool status_ |
|
|
|
|
|
|
|
|
#### zpool_scan_stats Fields
|
|
|
|
|
|
|
|
| field | units | description |
|
|
|
|
|---|---|---|
|
|
|
|
| errors | count | number of errors encountered by scan |
|
|
|
|
| examined | bytes | total data examined during scan |
|
|
|
|
| to_examine | bytes | prediction of total bytes to be scanned |
|
|
|
|
| pass_examined | bytes | data examined during current scan pass |
|
|
|
|
| issued | bytes | size of I/Os issued to disks |
|
|
|
|
| pass_issued | bytes | size of I/Os issued to disks for current pass |
|
|
|
|
| processed | bytes | data reconstructed during scan |
|
|
|
|
| to_process | bytes | total bytes to be repaired |
|
|
|
|
| rate | bytes/sec | examination rate |
|
|
|
|
| start_ts | epoch timestamp | start timestamp for scan |
|
|
|
|
| pause_ts | epoch timestamp | timestamp for a scan pause request |
|
|
|
|
| end_ts | epoch timestamp | completion timestamp for scan |
|
|
|
|
| paused_t | seconds | elapsed time while paused |
|
|
|
|
| remaining_t | seconds | estimate of time remaining for scan |
|
|
|
|
|
|
|
|
### zpool_vdev_stats Description
|
|
|
|
The ZFS I/O (ZIO) scheduler uses five queues to schedule I/Os to each vdev.
|
|
|
|
These queues are further divided into active and pending states.
|
|
|
|
An I/O is pending prior to being issued to the vdev. An active
|
|
|
|
I/O has been issued to the vdev. The scheduler and its tunable
|
|
|
|
parameters are described at the
|
|
|
|
[ZFS documentation for ZIO Scheduler]
|
|
|
|
(https://openzfs.github.io/openzfs-docs/Performance%20and%20Tuning/ZIO%20Scheduler.html)
|
|
|
|
The ZIO scheduler reports the queue depths as gauges where the value
|
|
|
|
represents an instantaneous snapshot of the queue depth at
|
|
|
|
the sample time. Therefore, it is not unusual to see all zeroes
|
|
|
|
for an idle pool.
|
|
|
|
|
|
|
|
#### zpool_vdev_stats Tags
|
|
|
|
| label | description |
|
|
|
|
|---|---|
|
|
|
|
| name | pool name |
|
|
|
|
| vdev | vdev name (root = entire pool) |
|
|
|
|
|
|
|
|
#### zpool_vdev_stats Fields
|
|
|
|
| field | units | description |
|
|
|
|
|---|---|---|
|
|
|
|
| sync_r_active_queue | entries | synchronous read active queue depth |
|
|
|
|
| sync_w_active_queue | entries | synchronous write active queue depth |
|
|
|
|
| async_r_active_queue | entries | asynchronous read active queue depth |
|
|
|
|
| async_w_active_queue | entries | asynchronous write active queue depth |
|
|
|
|
| async_scrub_active_queue | entries | asynchronous scrub active queue depth |
|
|
|
|
| sync_r_pend_queue | entries | synchronous read pending queue depth |
|
|
|
|
| sync_w_pend_queue | entries | synchronous write pending queue depth |
|
|
|
|
| async_r_pend_queue | entries | asynchronous read pending queue depth |
|
|
|
|
| async_w_pend_queue | entries | asynchronous write pending queue depth |
|
|
|
|
| async_scrub_pend_queue | entries | asynchronous scrub pending queue depth |
|
|
|
|
|
|
|
|
### zpool_latency Histogram
|
|
|
|
ZFS tracks the latency of each I/O in the ZIO pipeline. This latency can
|
|
|
|
be useful for observing latency-related issues that are not easily observed
|
|
|
|
using the averaged latency statistics.
|
|
|
|
|
|
|
|
The histogram fields show cumulative values from lowest to highest.
|
|
|
|
The largest bucket is tagged "le=+Inf", representing the total count
|
|
|
|
of I/Os by type and vdev.
|
|
|
|
|
|
|
|
#### zpool_latency Histogram Tags
|
|
|
|
| label | description |
|
|
|
|
|---|---|
|
|
|
|
| le | bucket for histogram, latency is less than or equal to bucket value in seconds |
|
|
|
|
| name | pool name |
|
|
|
|
| path | for leaf vdevs, the device path name, otherwise omitted |
|
|
|
|
| vdev | vdev name (root = entire pool) |
|
|
|
|
|
|
|
|
#### zpool_latency Histogram Fields
|
|
|
|
| field | units | description |
|
|
|
|
|---|---|---|
|
|
|
|
| total_read | operations | read operations of all types |
|
|
|
|
| total_write | operations | write operations of all types |
|
|
|
|
| disk_read | operations | disk read operations |
|
|
|
|
| disk_write | operations | disk write operations |
|
|
|
|
| sync_read | operations | ZIO sync reads |
|
|
|
|
| sync_write | operations | ZIO sync writes |
|
|
|
|
| async_read | operations | ZIO async reads|
|
|
|
|
| async_write | operations | ZIO async writes |
|
|
|
|
| scrub | operations | ZIO scrub/scan reads |
|
|
|
|
| trim | operations | ZIO trim (aka unmap) writes |
|
|
|
|
|
|
|
|
### zpool_io_size Histogram
|
|
|
|
ZFS tracks I/O throughout the ZIO pipeline. The size of each I/O is used
|
|
|
|
to create a histogram of the size by I/O type and vdev. For example, a
|
|
|
|
4KiB write to mirrored pool will show a 4KiB write to the top-level vdev
|
|
|
|
(root) and a 4KiB write to each of the mirror leaf vdevs.
|
|
|
|
|
|
|
|
The ZIO pipeline can aggregate I/O operations. For example, a contiguous
|
|
|
|
series of writes can be aggregated into a single, larger I/O to the leaf
|
|
|
|
vdev. The independent I/O operations reflect the logical operations and
|
|
|
|
the aggregated I/O operations reflect the physical operations.
|
|
|
|
|
|
|
|
The histogram fields show cumulative values from lowest to highest.
|
|
|
|
The largest bucket is tagged "le=+Inf", representing the total count
|
|
|
|
of I/Os by type and vdev.
|
|
|
|
|
|
|
|
Note: trim I/Os can be larger than 16MiB, but the larger sizes are
|
|
|
|
accounted in the 16MiB bucket.
|
|
|
|
|
|
|
|
#### zpool_io_size Histogram Tags
|
|
|
|
| label | description |
|
|
|
|
|---|---|
|
|
|
|
| le | bucket for histogram, I/O size is less than or equal to bucket value in bytes |
|
|
|
|
| name | pool name |
|
|
|
|
| path | for leaf vdevs, the device path name, otherwise omitted |
|
|
|
|
| vdev | vdev name (root = entire pool) |
|
|
|
|
|
|
|
|
#### zpool_io_size Histogram Fields
|
|
|
|
| field | units | description |
|
|
|
|
|---|---|---|
|
|
|
|
| sync_read_ind | blocks | independent sync reads |
|
|
|
|
| sync_write_ind | blocks | independent sync writes |
|
|
|
|
| async_read_ind | blocks | independent async reads |
|
|
|
|
| async_write_ind | blocks | independent async writes |
|
|
|
|
| scrub_read_ind | blocks | independent scrub/scan reads |
|
|
|
|
| trim_write_ind | blocks | independent trim (aka unmap) writes |
|
|
|
|
| sync_read_agg | blocks | aggregated sync reads |
|
|
|
|
| sync_write_agg | blocks | aggregated sync writes |
|
|
|
|
| async_read_agg | blocks | aggregated async reads |
|
|
|
|
| async_write_agg | blocks | aggregated async writes |
|
|
|
|
| scrub_read_agg | blocks | aggregated scrub/scan reads |
|
|
|
|
| trim_write_agg | blocks | aggregated trim (aka unmap) writes |
|
|
|
|
|
|
|
|
#### About unsigned integers
|
|
|
|
Telegraf v1.6.2 and later support unsigned 64-bit integers which more
|
|
|
|
closely matches the uint64_t values used by ZFS. By default, zpool_influxdb
|
|
|
|
uses ZFS' uint64_t values and influxdb line protocol unsigned integer type.
|
|
|
|
If you are using old telegraf or influxdb where unsigned integers are not
|
|
|
|
available, use the `--signed-int` option.
|
|
|
|
|
|
|
|
## Using _zpool_influxdb_
|
|
|
|
|
|
|
|
The simplest method is to use the execd input agent in telegraf. For older
|
|
|
|
versions of telegraf which lack execd, the exec input agent can be used.
|
|
|
|
For convenience, one of the sample config files below can be placed in the
|
|
|
|
telegraf config-directory (often /etc/telegraf/telegraf.d). Telegraf can
|
|
|
|
be restarted to read the config-directory files.
|
|
|
|
|
|
|
|
### Example telegraf execd configuration
|
|
|
|
```toml
|
|
|
|
# # Read metrics from zpool_influxdb
|
|
|
|
[[inputs.execd]]
|
|
|
|
# ## default installation location for zpool_influxdb command
|
2020-11-28 19:15:57 +00:00
|
|
|
command = ["/usr/libexec/zfs/zpool_influxdb", "--execd"]
|
2020-10-09 16:29:21 +00:00
|
|
|
|
|
|
|
## Define how the process is signaled on each collection interval.
|
|
|
|
## Valid values are:
|
|
|
|
## "none" : Do not signal anything. (Recommended for service inputs)
|
|
|
|
## The process must output metrics by itself.
|
|
|
|
## "STDIN" : Send a newline on STDIN. (Recommended for gather inputs)
|
|
|
|
## "SIGHUP" : Send a HUP signal. Not available on Windows. (not recommended)
|
|
|
|
## "SIGUSR1" : Send a USR1 signal. Not available on Windows.
|
|
|
|
## "SIGUSR2" : Send a USR2 signal. Not available on Windows.
|
|
|
|
signal = "STDIN"
|
|
|
|
|
|
|
|
## Delay before the process is restarted after an unexpected termination
|
|
|
|
restart_delay = "10s"
|
|
|
|
|
|
|
|
## Data format to consume.
|
|
|
|
## Each data format has its own unique set of configuration options, read
|
|
|
|
## more about them here:
|
|
|
|
## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md
|
|
|
|
data_format = "influx"
|
|
|
|
```
|
|
|
|
|
|
|
|
### Example telegraf exec configuration
|
|
|
|
```toml
|
|
|
|
# # Read metrics from zpool_influxdb
|
|
|
|
[[inputs.exec]]
|
|
|
|
# ## default installation location for zpool_influxdb command
|
2020-11-28 19:15:57 +00:00
|
|
|
commands = ["/usr/libexec/zfs/zpool_influxdb"]
|
2020-10-09 16:29:21 +00:00
|
|
|
data_format = "influx"
|
|
|
|
```
|
|
|
|
|
|
|
|
## Caveat Emptor
|
|
|
|
* Like the _zpool_ command, _zpool_influxdb_ takes a reader
|
|
|
|
lock on spa_config for each imported pool. If this lock blocks,
|
|
|
|
then the command will also block indefinitely and might be
|
|
|
|
unkillable. This is not a normal condition, but can occur if
|
|
|
|
there are bugs in the kernel modules.
|
|
|
|
For this reason, care should be taken:
|
|
|
|
* avoid spawning many of these commands hoping that one might
|
|
|
|
finish
|
|
|
|
* avoid frequent updates or short sample time
|
|
|
|
intervals, because the locks can interfere with the performance
|
|
|
|
of other instances of _zpool_ or _zpool_influxdb_
|
|
|
|
|
|
|
|
## Other collectors
|
|
|
|
There are a few other collectors for zpool statistics roaming around
|
|
|
|
the Internet. Many attempt to screen-scrape `zpool` output in various
|
|
|
|
ways. The screen-scrape method works poorly for `zpool` output because
|
|
|
|
of its human-friendly nature. Also, they suffer from the same caveats
|
|
|
|
as this implementation. This implementation is optimized for directly
|
|
|
|
collecting the metrics and is much more efficient than the screen-scrapers.
|
|
|
|
|
|
|
|
## Feedback Encouraged
|
|
|
|
Pull requests and issues are greatly appreciated at
|
|
|
|
https://github.com/openzfs/zfs
|