Use udev for partition detection

When ZFS partitions a block device it must wait for udev to create
both a device node and all the device symlinks.  This process takes
a variable length of time and depends on factors such how many links
must be created, the complexity of the rules, etc.  Complicating
the situation further it is not uncommon for udev to create and
then remove a link multiple times while processing the udev rules.

Given the above, the existing scheme of waiting for an expected
partition to appear by name isn't 100% reliable.  At this point
udev may still remove and recreate think link resulting in the
kernel modules being unable to open the device.

In order to address this the zpool_label_disk_wait() function
has been updated to use libudev.  Until the registered system
device acknowledges that it in fully initialized the function
will wait.  Once fully initialized all device links are checked
and allowed to settle for 50ms.  This makes it far more likely
that all the device nodes will exist when the kernel modules
need to open them.

For systems without libudev an alternate zpool_label_disk_wait()
was updated to include a settle time.  In addition, the kernel
modules were updated to include retry logic for this ENOENT case.
Due to the improved checks in the utilities it is unlikely this
logic will be invoked.  However, if the rare event it is needed
it will prevent a failure.

Signed-off-by: Brian Behlendorf <behlendorf1@llnl.gov>
Signed-off-by: Tony Hutter <hutter2@llnl.gov>
Signed-off-by: Richard Laager <rlaager@wiktel.com>
Closes #4523
Closes #3708
Closes #4077
Closes #4144
Closes #4214
Closes #4517
This commit is contained in:
Brian Behlendorf 2016-04-19 11:19:12 -07:00 committed by Ned Bass
parent c9ca152fd1
commit 2cb77346cb
4 changed files with 83 additions and 43 deletions

View File

@ -1206,12 +1206,10 @@ make_disks(zpool_handle_t *zhp, nvlist_t *nv)
/* /*
* Remove any previously existing symlink from a udev path to * Remove any previously existing symlink from a udev path to
* the device before labeling the disk. This makes * the device before labeling the disk. This ensures that
* zpool_label_disk_wait() truly wait for the new link to show * only newly created links are used. Otherwise there is a
* up instead of returning if it finds an old link still in * window between when udev deletes and recreates the link
* place. Otherwise there is a window between when udev * during which access attempts will fail with ENOENT.
* deletes and recreates the link during which access attempts
* will fail with ENOENT.
*/ */
strncpy(udevpath, path, MAXPATHLEN); strncpy(udevpath, path, MAXPATHLEN);
(void) zfs_append_partition(udevpath, MAXPATHLEN); (void) zfs_append_partition(udevpath, MAXPATHLEN);
@ -1235,6 +1233,8 @@ make_disks(zpool_handle_t *zhp, nvlist_t *nv)
* and then block until udev creates the new link. * and then block until udev creates the new link.
*/ */
if (!is_exclusive || !is_spare(NULL, udevpath)) { if (!is_exclusive || !is_spare(NULL, udevpath)) {
char *devnode = strrchr(devpath, '/') + 1;
ret = strncmp(udevpath, UDISK_ROOT, strlen(UDISK_ROOT)); ret = strncmp(udevpath, UDISK_ROOT, strlen(UDISK_ROOT));
if (ret == 0) { if (ret == 0) {
ret = lstat64(udevpath, &statbuf); ret = lstat64(udevpath, &statbuf);
@ -1242,18 +1242,29 @@ make_disks(zpool_handle_t *zhp, nvlist_t *nv)
(void) unlink(udevpath); (void) unlink(udevpath);
} }
if (zpool_label_disk(g_zfs, zhp, /*
strrchr(devpath, '/') + 1) == -1) * When labeling a pool the raw device node name
* is provided as it appears under /dev/.
*/
if (zpool_label_disk(g_zfs, zhp, devnode) == -1)
return (-1); return (-1);
/*
* Wait for udev to signal the device is available
* by the provided path.
*/
ret = zpool_label_disk_wait(udevpath, DISK_LABEL_WAIT); ret = zpool_label_disk_wait(udevpath, DISK_LABEL_WAIT);
if (ret) { if (ret) {
(void) fprintf(stderr, gettext("cannot " (void) fprintf(stderr,
"resolve path '%s': %d\n"), udevpath, ret); gettext("missing link: %s was "
return (-1); "partitioned but %s is missing\n"),
devnode, udevpath);
return (ret);
} }
(void) zero_label(udevpath); ret = zero_label(udevpath);
if (ret)
return (ret);
} }
/* /*

View File

@ -123,6 +123,40 @@ get_devid(const char *path)
return (ret); return (ret);
} }
/*
* Wait up to timeout_ms for udev to set up the device node. The device is
* considered ready when the provided path have been verified to exist and
* it has been allowed to settle. At this point the device the device can
* be accessed reliably. Depending on the complexity of the udev rules thisi
* process could take several seconds.
*/
int
zpool_label_disk_wait(char *path, int timeout_ms)
{
int settle_ms = 50;
long sleep_ms = 10;
hrtime_t start, settle;
struct stat64 statbuf;
start = gethrtime();
settle = 0;
do {
errno = 0;
if ((stat64(path, &statbuf) == 0) && (errno == 0)) {
if (settle == 0)
settle = gethrtime();
else if (NSEC2MSEC(gethrtime() - settle) >= settle_ms)
return (0);
} else if (errno != ENOENT) {
return (errno);
}
usleep(sleep_ms * MILLISEC);
} while (NSEC2MSEC(gethrtime() - start) < timeout_ms);
return (ENODEV);
}
/* /*
* Go through and fix up any path and/or devid information for the given vdev * Go through and fix up any path and/or devid information for the given vdev

View File

@ -4094,29 +4094,6 @@ find_start_block(nvlist_t *config)
return (MAXOFFSET_T); return (MAXOFFSET_T);
} }
int
zpool_label_disk_wait(char *path, int timeout)
{
struct stat64 statbuf;
int i;
/*
* Wait timeout miliseconds for a newly created device to be available
* from the given path. There is a small window when a /dev/ device
* will exist and the udev link will not, so we must wait for the
* symlink. Depending on the udev rules this may take a few seconds.
*/
for (i = 0; i < timeout; i++) {
usleep(1000);
errno = 0;
if ((stat64(path, &statbuf) == 0) && (errno == 0))
return (0);
}
return (ENOENT);
}
int int
zpool_label_disk_check(char *path) zpool_label_disk_check(char *path)
{ {
@ -4282,12 +4259,11 @@ zpool_label_disk(libzfs_handle_t *hdl, zpool_handle_t *zhp, char *name)
(void) close(fd); (void) close(fd);
efi_free(vtoc); efi_free(vtoc);
/* Wait for the first expected partition to appear. */
(void) snprintf(path, sizeof (path), "%s/%s", DISK_ROOT, name); (void) snprintf(path, sizeof (path), "%s/%s", DISK_ROOT, name);
(void) zfs_append_partition(path, MAXPATHLEN); (void) zfs_append_partition(path, MAXPATHLEN);
rval = zpool_label_disk_wait(path, 3000); /* Wait to udev to signal use the device has settled. */
rval = zpool_label_disk_wait(path, DISK_LABEL_WAIT);
if (rval) { if (rval) {
zfs_error_aux(hdl, dgettext(TEXT_DOMAIN, "failed to " zfs_error_aux(hdl, dgettext(TEXT_DOMAIN, "failed to "
"detect device partitions on '%s': %d"), path, rval); "detect device partitions on '%s': %d"), path, rval);

View File

@ -244,12 +244,12 @@ vdev_disk_open(vdev_t *v, uint64_t *psize, uint64_t *max_psize,
{ {
struct block_device *bdev = ERR_PTR(-ENXIO); struct block_device *bdev = ERR_PTR(-ENXIO);
vdev_disk_t *vd; vdev_disk_t *vd;
int mode, block_size; int count = 0, mode, block_size;
/* Must have a pathname and it must be absolute. */ /* Must have a pathname and it must be absolute. */
if (v->vdev_path == NULL || v->vdev_path[0] != '/') { if (v->vdev_path == NULL || v->vdev_path[0] != '/') {
v->vdev_stat.vs_aux = VDEV_AUX_BAD_LABEL; v->vdev_stat.vs_aux = VDEV_AUX_BAD_LABEL;
return (EINVAL); return (SET_ERROR(EINVAL));
} }
/* /*
@ -264,7 +264,7 @@ vdev_disk_open(vdev_t *v, uint64_t *psize, uint64_t *max_psize,
vd = kmem_zalloc(sizeof (vdev_disk_t), KM_SLEEP); vd = kmem_zalloc(sizeof (vdev_disk_t), KM_SLEEP);
if (vd == NULL) if (vd == NULL)
return (ENOMEM); return (SET_ERROR(ENOMEM));
/* /*
* Devices are always opened by the path provided at configuration * Devices are always opened by the path provided at configuration
@ -279,16 +279,35 @@ vdev_disk_open(vdev_t *v, uint64_t *psize, uint64_t *max_psize,
* /dev/[hd]d devices which may be reordered due to probing order. * /dev/[hd]d devices which may be reordered due to probing order.
* Devices in the wrong locations will be detected by the higher * Devices in the wrong locations will be detected by the higher
* level vdev validation. * level vdev validation.
*
* The specified paths may be briefly removed and recreated in
* response to udev events. This should be exceptionally unlikely
* because the zpool command makes every effort to verify these paths
* have already settled prior to reaching this point. Therefore,
* a ENOENT failure at this point is highly likely to be transient
* and it is reasonable to sleep and retry before giving up. In
* practice delays have been observed to be on the order of 100ms.
*/ */
mode = spa_mode(v->vdev_spa); mode = spa_mode(v->vdev_spa);
if (v->vdev_wholedisk && v->vdev_expanding) if (v->vdev_wholedisk && v->vdev_expanding)
bdev = vdev_disk_rrpart(v->vdev_path, mode, vd); bdev = vdev_disk_rrpart(v->vdev_path, mode, vd);
if (IS_ERR(bdev))
while (IS_ERR(bdev) && count < 50) {
bdev = vdev_bdev_open(v->vdev_path, bdev = vdev_bdev_open(v->vdev_path,
vdev_bdev_mode(mode), zfs_vdev_holder); vdev_bdev_mode(mode), zfs_vdev_holder);
if (unlikely(PTR_ERR(bdev) == -ENOENT)) {
msleep(10);
count++;
} else if (IS_ERR(bdev)) {
break;
}
}
if (IS_ERR(bdev)) { if (IS_ERR(bdev)) {
dprintf("failed open v->vdev_path=%s, error=%d count=%d\n",
v->vdev_path, -PTR_ERR(bdev), count);
kmem_free(vd, sizeof (vdev_disk_t)); kmem_free(vd, sizeof (vdev_disk_t));
return (-PTR_ERR(bdev)); return (SET_ERROR(-PTR_ERR(bdev)));
} }
v->vdev_tsd = vd; v->vdev_tsd = vd;