Clean up RAIDZ/DRAID ereport code
The RAIDZ and DRAID code is responsible for reporting checksum errors on their child vdevs. Checksum errors represent events where a disk returned data or parity that should have been correct, but was not. In other words, these are instances of silent data corruption. The checksum errors show up in the vdev stats (and thus `zpool status`'s CKSUM column), and in the event log (`zpool events`). Note, this is in contrast with the more common "noisy" errors where a disk goes offline, in which case ZFS knows that the disk is bad and doesn't try to read it, or the device returns an error on the requested read or write operation. RAIDZ/DRAID generate checksum errors via three code paths: 1. When RAIDZ/DRAID reconstructs a damaged block, checksum errors are reported on any children whose data was not used during the reconstruction. This is handled in `raidz_reconstruct()`. This is the most common type of RAIDZ/DRAID checksum error. 2. When RAIDZ/DRAID is not able to reconstruct a damaged block, that means that the data has been lost. The zio fails and an error is returned to the consumer (e.g. the read(2) system call). This would happen if, for example, three different disks in a RAIDZ2 group are silently damaged. Since the damage is silent, it isn't possible to know which three disks are damaged, so a checksum error is reported against every child that returned data or parity for this read. (For DRAID, typically only one "group" of children is involved in each io.) This case is handled in `vdev_raidz_cksum_finish()`. This is the next most common type of RAIDZ/DRAID checksum error. 3. If RAIDZ/DRAID is not able to reconstruct a damaged block (like in case 2), but there happens to be additional copies of this block due to "ditto blocks" (i.e. multiple DVA's in this blkptr_t), and one of those copies is good, then RAIDZ/DRAID compares each sector of the data or parity that it retrieved with the good data from the other DVA, and if they differ then it reports a checksum error on this child. This differs from case 2 in that the checksum error is reported on only the subset of children that actually have bad data or parity. This case happens very rarely, since normally only metadata has ditto blocks. If the silent damage is extensive, there will be many instances of case 2, and the pool will likely be unrecoverable. The code for handling case 3 is considerably more complicated than the other cases, for two reasons: 1. It needs to run after the main raidz read logic has completed. The data RAIDZ read needs to be preserved until after the alternate DVA has been read, which necessitates refcounts and callbacks managed by the non-raidz-specific zio layer. 2. It's nontrivial to map the sections of data read by RAIDZ to the correct data. For example, the correct data does not include the parity information, so the parity must be recalculated based on the correct data, and then compared to the parity that was read from the RAIDZ children. Due to the complexity of case 3, the rareness of hitting it, and the minimal benefit it provides above case 2, this commit removes the code for case 3. These types of errors will now be handled the same as case 2, i.e. the checksum error will be reported against all children that returned data or parity. Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov> Signed-off-by: Matthew Ahrens <mahrens@delphix.com> Closes #11735
This commit is contained in:
parent
2f385c913f
commit
330c6c0523
|
@ -448,7 +448,6 @@ vdev_raidz_map_alloc_expanded(abd_t *abd, uint64_t size, uint64_t offset,
|
||||||
rr->rr_missingdata = 0;
|
rr->rr_missingdata = 0;
|
||||||
rr->rr_missingparity = 0;
|
rr->rr_missingparity = 0;
|
||||||
rr->rr_firstdatacol = nparity;
|
rr->rr_firstdatacol = nparity;
|
||||||
rr->rr_abd_copy = NULL;
|
|
||||||
rr->rr_abd_empty = NULL;
|
rr->rr_abd_empty = NULL;
|
||||||
rr->rr_nempty = 0;
|
rr->rr_nempty = 0;
|
||||||
|
|
||||||
|
@ -459,7 +458,6 @@ vdev_raidz_map_alloc_expanded(abd_t *abd, uint64_t size, uint64_t offset,
|
||||||
}
|
}
|
||||||
rr->rr_col[c].rc_devidx = child_id;
|
rr->rr_col[c].rc_devidx = child_id;
|
||||||
rr->rr_col[c].rc_offset = child_offset;
|
rr->rr_col[c].rc_offset = child_offset;
|
||||||
rr->rr_col[c].rc_gdata = NULL;
|
|
||||||
rr->rr_col[c].rc_orig_data = NULL;
|
rr->rr_col[c].rc_orig_data = NULL;
|
||||||
rr->rr_col[c].rc_error = 0;
|
rr->rr_col[c].rc_error = 0;
|
||||||
rr->rr_col[c].rc_tried = 0;
|
rr->rr_col[c].rc_tried = 0;
|
||||||
|
|
|
@ -50,6 +50,8 @@ void vdev_raidz_reconstruct(struct raidz_map *, const int *, int);
|
||||||
void vdev_raidz_child_done(zio_t *);
|
void vdev_raidz_child_done(zio_t *);
|
||||||
void vdev_raidz_io_done(zio_t *);
|
void vdev_raidz_io_done(zio_t *);
|
||||||
|
|
||||||
|
extern const zio_vsd_ops_t vdev_raidz_vsd_ops;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* vdev_raidz_math interface
|
* vdev_raidz_math interface
|
||||||
*/
|
*/
|
||||||
|
|
|
@ -108,8 +108,7 @@ typedef struct raidz_col {
|
||||||
uint64_t rc_size; /* I/O size */
|
uint64_t rc_size; /* I/O size */
|
||||||
abd_t rc_abdstruct; /* rc_abd probably points here */
|
abd_t rc_abdstruct; /* rc_abd probably points here */
|
||||||
abd_t *rc_abd; /* I/O data */
|
abd_t *rc_abd; /* I/O data */
|
||||||
void *rc_orig_data; /* pre-reconstruction */
|
abd_t *rc_orig_data; /* pre-reconstruction */
|
||||||
abd_t *rc_gdata; /* used to store the "good" version */
|
|
||||||
int rc_error; /* I/O error for this device */
|
int rc_error; /* I/O error for this device */
|
||||||
uint8_t rc_tried; /* Did we attempt this I/O column? */
|
uint8_t rc_tried; /* Did we attempt this I/O column? */
|
||||||
uint8_t rc_skipped; /* Did we skip this I/O column? */
|
uint8_t rc_skipped; /* Did we skip this I/O column? */
|
||||||
|
@ -124,7 +123,6 @@ typedef struct raidz_row {
|
||||||
uint64_t rr_missingdata; /* Count of missing data devices */
|
uint64_t rr_missingdata; /* Count of missing data devices */
|
||||||
uint64_t rr_missingparity; /* Count of missing parity devices */
|
uint64_t rr_missingparity; /* Count of missing parity devices */
|
||||||
uint64_t rr_firstdatacol; /* First data column/parity count */
|
uint64_t rr_firstdatacol; /* First data column/parity count */
|
||||||
abd_t *rr_abd_copy; /* rm_asize-buffer of copied data */
|
|
||||||
abd_t *rr_abd_empty; /* dRAID empty sector buffer */
|
abd_t *rr_abd_empty; /* dRAID empty sector buffer */
|
||||||
int rr_nempty; /* empty sectors included in parity */
|
int rr_nempty; /* empty sectors included in parity */
|
||||||
#ifdef ZFS_DEBUG
|
#ifdef ZFS_DEBUG
|
||||||
|
@ -135,8 +133,6 @@ typedef struct raidz_row {
|
||||||
} raidz_row_t;
|
} raidz_row_t;
|
||||||
|
|
||||||
typedef struct raidz_map {
|
typedef struct raidz_map {
|
||||||
uintptr_t rm_reports; /* # of referencing checksum reports */
|
|
||||||
boolean_t rm_freed; /* map no longer has referencing ZIO */
|
|
||||||
boolean_t rm_ecksuminjected; /* checksum error was injected */
|
boolean_t rm_ecksuminjected; /* checksum error was injected */
|
||||||
int rm_nrows; /* Regular row count */
|
int rm_nrows; /* Regular row count */
|
||||||
int rm_nskip; /* RAIDZ sectors skipped for padding */
|
int rm_nskip; /* RAIDZ sectors skipped for padding */
|
||||||
|
|
|
@ -382,14 +382,8 @@ struct zio_cksum_report {
|
||||||
struct zio_bad_cksum *zcr_ckinfo; /* information from failure */
|
struct zio_bad_cksum *zcr_ckinfo; /* information from failure */
|
||||||
};
|
};
|
||||||
|
|
||||||
typedef void zio_vsd_cksum_report_f(zio_t *zio, zio_cksum_report_t *zcr,
|
|
||||||
void *arg);
|
|
||||||
|
|
||||||
zio_vsd_cksum_report_f zio_vsd_default_cksum_report;
|
|
||||||
|
|
||||||
typedef struct zio_vsd_ops {
|
typedef struct zio_vsd_ops {
|
||||||
zio_done_func_t *vsd_free;
|
zio_done_func_t *vsd_free;
|
||||||
zio_vsd_cksum_report_f *vsd_cksum_report;
|
|
||||||
} zio_vsd_ops_t;
|
} zio_vsd_ops_t;
|
||||||
|
|
||||||
typedef struct zio_gang_node {
|
typedef struct zio_gang_node {
|
||||||
|
@ -683,7 +677,7 @@ extern hrtime_t zio_handle_io_delay(zio_t *zio);
|
||||||
*/
|
*/
|
||||||
extern int zfs_ereport_start_checksum(spa_t *spa, vdev_t *vd,
|
extern int zfs_ereport_start_checksum(spa_t *spa, vdev_t *vd,
|
||||||
const zbookmark_phys_t *zb, struct zio *zio, uint64_t offset,
|
const zbookmark_phys_t *zb, struct zio *zio, uint64_t offset,
|
||||||
uint64_t length, void *arg, struct zio_bad_cksum *info);
|
uint64_t length, struct zio_bad_cksum *info);
|
||||||
extern void zfs_ereport_finish_checksum(zio_cksum_report_t *report,
|
extern void zfs_ereport_finish_checksum(zio_cksum_report_t *report,
|
||||||
const abd_t *good_data, const abd_t *bad_data, boolean_t drop_if_identical);
|
const abd_t *good_data, const abd_t *bad_data, boolean_t drop_if_identical);
|
||||||
|
|
||||||
|
@ -695,6 +689,8 @@ extern int zfs_ereport_post_checksum(spa_t *spa, vdev_t *vd,
|
||||||
uint64_t length, const abd_t *good_data, const abd_t *bad_data,
|
uint64_t length, const abd_t *good_data, const abd_t *bad_data,
|
||||||
struct zio_bad_cksum *info);
|
struct zio_bad_cksum *info);
|
||||||
|
|
||||||
|
void zio_vsd_default_cksum_report(zio_t *zio, zio_cksum_report_t *zcr);
|
||||||
|
|
||||||
/* Called from spa_sync(), but primarily an injection handler */
|
/* Called from spa_sync(), but primarily an injection handler */
|
||||||
extern void spa_handle_ignored_writes(spa_t *spa);
|
extern void spa_handle_ignored_writes(spa_t *spa);
|
||||||
|
|
||||||
|
|
|
@ -295,6 +295,23 @@ identified by a unique identifier instead of its path since the path was never
|
||||||
correct in the first place.
|
correct in the first place.
|
||||||
.El
|
.El
|
||||||
.Pp
|
.Pp
|
||||||
|
Checksum errors represent events where a disk returned data that was expected
|
||||||
|
to be correct, but was not.
|
||||||
|
In other words, these are instances of silent data corruption.
|
||||||
|
The checksum errors are reported in
|
||||||
|
.Nm zpool Cm status
|
||||||
|
and
|
||||||
|
.Nm zpool Cm events .
|
||||||
|
When a block is stored redundantly, a damaged block may be reconstructed
|
||||||
|
(e.g. from RAIDZ parity or a mirrored copy).
|
||||||
|
In this case, ZFS reports the checksum error against the disks that contained
|
||||||
|
damaged data.
|
||||||
|
If a block is unable to be reconstructed (e.g. due to 3 disks being damaged
|
||||||
|
in a RAIDZ2 group), it is not possible to determine which disks were silently
|
||||||
|
corrupted.
|
||||||
|
In this case, checksum errors are reported for all disks on which the block
|
||||||
|
is stored.
|
||||||
|
.Pp
|
||||||
If a device is removed and later re-attached to the system, ZFS attempts
|
If a device is removed and later re-attached to the system, ZFS attempts
|
||||||
to put the device online automatically.
|
to put the device online automatically.
|
||||||
Device attach detection is hardware-dependent and might not be supported on all
|
Device attach detection is hardware-dependent and might not be supported on all
|
||||||
|
|
|
@ -632,236 +632,6 @@ vdev_draid_group_to_offset(vdev_t *vd, uint64_t group)
|
||||||
return (group * vdc->vdc_groupsz);
|
return (group * vdc->vdc_groupsz);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
static void
|
|
||||||
vdev_draid_map_free_vsd(zio_t *zio)
|
|
||||||
{
|
|
||||||
raidz_map_t *rm = zio->io_vsd;
|
|
||||||
|
|
||||||
ASSERT0(rm->rm_freed);
|
|
||||||
rm->rm_freed = B_TRUE;
|
|
||||||
|
|
||||||
if (rm->rm_reports == 0) {
|
|
||||||
vdev_raidz_map_free(rm);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/*ARGSUSED*/
|
|
||||||
static void
|
|
||||||
vdev_draid_cksum_free(void *arg, size_t ignored)
|
|
||||||
{
|
|
||||||
raidz_map_t *rm = arg;
|
|
||||||
|
|
||||||
ASSERT3U(rm->rm_reports, >, 0);
|
|
||||||
|
|
||||||
if (--rm->rm_reports == 0 && rm->rm_freed)
|
|
||||||
vdev_raidz_map_free(rm);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void
|
|
||||||
vdev_draid_cksum_finish(zio_cksum_report_t *zcr, const abd_t *good_data)
|
|
||||||
{
|
|
||||||
raidz_map_t *rm = zcr->zcr_cbdata;
|
|
||||||
const size_t c = zcr->zcr_cbinfo;
|
|
||||||
uint64_t skip_size = zcr->zcr_sector;
|
|
||||||
uint64_t parity_size;
|
|
||||||
size_t x, offset, size;
|
|
||||||
|
|
||||||
if (good_data == NULL) {
|
|
||||||
zfs_ereport_finish_checksum(zcr, NULL, NULL, B_FALSE);
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Detailed cksum reporting is currently only supported for single
|
|
||||||
* row draid mappings, this covers the vast majority of zios. Only
|
|
||||||
* a dRAID zio which spans groups will have multiple rows.
|
|
||||||
*/
|
|
||||||
if (rm->rm_nrows != 1) {
|
|
||||||
zfs_ereport_finish_checksum(zcr, NULL, NULL, B_FALSE);
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
raidz_row_t *rr = rm->rm_row[0];
|
|
||||||
const abd_t *good = NULL;
|
|
||||||
const abd_t *bad = rr->rr_col[c].rc_abd;
|
|
||||||
|
|
||||||
if (c < rr->rr_firstdatacol) {
|
|
||||||
/*
|
|
||||||
* The first time through, calculate the parity blocks for
|
|
||||||
* the good data (this relies on the fact that the good
|
|
||||||
* data never changes for a given logical zio)
|
|
||||||
*/
|
|
||||||
if (rr->rr_col[0].rc_gdata == NULL) {
|
|
||||||
abd_t *bad_parity[VDEV_DRAID_MAXPARITY];
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Set up the rr_col[]s to generate the parity for
|
|
||||||
* good_data, first saving the parity bufs and
|
|
||||||
* replacing them with buffers to hold the result.
|
|
||||||
*/
|
|
||||||
for (x = 0; x < rr->rr_firstdatacol; x++) {
|
|
||||||
bad_parity[x] = rr->rr_col[x].rc_abd;
|
|
||||||
rr->rr_col[x].rc_abd = rr->rr_col[x].rc_gdata =
|
|
||||||
abd_alloc_sametype(rr->rr_col[x].rc_abd,
|
|
||||||
rr->rr_col[x].rc_size);
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Fill in the data columns from good_data being
|
|
||||||
* careful to pad short columns and empty columns
|
|
||||||
* with a skip sector.
|
|
||||||
*/
|
|
||||||
uint64_t good_size = abd_get_size((abd_t *)good_data);
|
|
||||||
|
|
||||||
offset = 0;
|
|
||||||
for (; x < rr->rr_cols; x++) {
|
|
||||||
abd_free(rr->rr_col[x].rc_abd);
|
|
||||||
|
|
||||||
if (offset == good_size) {
|
|
||||||
/* empty data column (small write) */
|
|
||||||
rr->rr_col[x].rc_abd =
|
|
||||||
abd_get_zeros(skip_size);
|
|
||||||
} else if (x < rr->rr_bigcols) {
|
|
||||||
/* this is a "big column" */
|
|
||||||
size = rr->rr_col[x].rc_size;
|
|
||||||
rr->rr_col[x].rc_abd =
|
|
||||||
abd_get_offset_size(
|
|
||||||
(abd_t *)good_data, offset, size);
|
|
||||||
offset += size;
|
|
||||||
} else {
|
|
||||||
/* short data column, add skip sector */
|
|
||||||
size = rr->rr_col[x].rc_size -skip_size;
|
|
||||||
rr->rr_col[x].rc_abd = abd_alloc(
|
|
||||||
rr->rr_col[x].rc_size, B_TRUE);
|
|
||||||
abd_copy_off(rr->rr_col[x].rc_abd,
|
|
||||||
(abd_t *)good_data, 0, offset,
|
|
||||||
size);
|
|
||||||
abd_zero_off(rr->rr_col[x].rc_abd,
|
|
||||||
size, skip_size);
|
|
||||||
offset += size;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Construct the parity from the good data.
|
|
||||||
*/
|
|
||||||
vdev_raidz_generate_parity_row(rm, rr);
|
|
||||||
|
|
||||||
/* restore everything back to its original state */
|
|
||||||
for (x = 0; x < rr->rr_firstdatacol; x++)
|
|
||||||
rr->rr_col[x].rc_abd = bad_parity[x];
|
|
||||||
|
|
||||||
offset = 0;
|
|
||||||
for (x = rr->rr_firstdatacol; x < rr->rr_cols; x++) {
|
|
||||||
abd_free(rr->rr_col[x].rc_abd);
|
|
||||||
rr->rr_col[x].rc_abd = abd_get_offset_size(
|
|
||||||
rr->rr_abd_copy, offset,
|
|
||||||
rr->rr_col[x].rc_size);
|
|
||||||
offset += rr->rr_col[x].rc_size;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
ASSERT3P(rr->rr_col[c].rc_gdata, !=, NULL);
|
|
||||||
good = abd_get_offset_size(rr->rr_col[c].rc_gdata, 0,
|
|
||||||
rr->rr_col[c].rc_size);
|
|
||||||
} else {
|
|
||||||
/* adjust good_data to point at the start of our column */
|
|
||||||
parity_size = size = rr->rr_col[0].rc_size;
|
|
||||||
if (c >= rr->rr_bigcols) {
|
|
||||||
size -= skip_size;
|
|
||||||
zcr->zcr_length = size;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* empty column */
|
|
||||||
if (size == 0) {
|
|
||||||
zfs_ereport_finish_checksum(zcr, NULL, NULL, B_TRUE);
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
offset = 0;
|
|
||||||
for (x = rr->rr_firstdatacol; x < c; x++) {
|
|
||||||
if (x < rr->rr_bigcols) {
|
|
||||||
offset += parity_size;
|
|
||||||
} else {
|
|
||||||
offset += parity_size - skip_size;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
good = abd_get_offset_size((abd_t *)good_data, offset, size);
|
|
||||||
}
|
|
||||||
|
|
||||||
/* we drop the ereport if it ends up that the data was good */
|
|
||||||
zfs_ereport_finish_checksum(zcr, good, bad, B_TRUE);
|
|
||||||
abd_free((abd_t *)good);
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Invoked indirectly by zfs_ereport_start_checksum(), called
|
|
||||||
* below when our read operation fails completely. The main point
|
|
||||||
* is to keep a copy of everything we read from disk, so that at
|
|
||||||
* vdev_draid_cksum_finish() time we can compare it with the good data.
|
|
||||||
*/
|
|
||||||
static void
|
|
||||||
vdev_draid_cksum_report(zio_t *zio, zio_cksum_report_t *zcr, void *arg)
|
|
||||||
{
|
|
||||||
size_t c = (size_t)(uintptr_t)arg;
|
|
||||||
raidz_map_t *rm = zio->io_vsd;
|
|
||||||
|
|
||||||
/* set up the report and bump the refcount */
|
|
||||||
zcr->zcr_cbdata = rm;
|
|
||||||
zcr->zcr_cbinfo = c;
|
|
||||||
zcr->zcr_finish = vdev_draid_cksum_finish;
|
|
||||||
zcr->zcr_free = vdev_draid_cksum_free;
|
|
||||||
|
|
||||||
rm->rm_reports++;
|
|
||||||
ASSERT3U(rm->rm_reports, >, 0);
|
|
||||||
|
|
||||||
if (rm->rm_row[0]->rr_abd_copy != NULL)
|
|
||||||
return;
|
|
||||||
|
|
||||||
/*
|
|
||||||
* It's the first time we're called for this raidz_map_t, so we need
|
|
||||||
* to copy the data aside; there's no guarantee that our zio's buffer
|
|
||||||
* won't be re-used for something else.
|
|
||||||
*
|
|
||||||
* Our parity data is already in separate buffers, so there's no need
|
|
||||||
* to copy them. Furthermore, all columns should have been expanded
|
|
||||||
* by vdev_draid_map_alloc_empty() when attempting reconstruction.
|
|
||||||
*/
|
|
||||||
for (int i = 0; i < rm->rm_nrows; i++) {
|
|
||||||
raidz_row_t *rr = rm->rm_row[i];
|
|
||||||
size_t offset = 0;
|
|
||||||
size_t size = 0;
|
|
||||||
|
|
||||||
for (c = rr->rr_firstdatacol; c < rr->rr_cols; c++) {
|
|
||||||
ASSERT3U(rr->rr_col[c].rc_size, ==,
|
|
||||||
rr->rr_col[0].rc_size);
|
|
||||||
size += rr->rr_col[c].rc_size;
|
|
||||||
}
|
|
||||||
|
|
||||||
rr->rr_abd_copy = abd_alloc_for_io(size, B_FALSE);
|
|
||||||
|
|
||||||
for (c = rr->rr_firstdatacol; c < rr->rr_cols; c++) {
|
|
||||||
raidz_col_t *col = &rr->rr_col[c];
|
|
||||||
abd_t *tmp = abd_get_offset_size(rr->rr_abd_copy,
|
|
||||||
offset, col->rc_size);
|
|
||||||
|
|
||||||
abd_copy(tmp, col->rc_abd, col->rc_size);
|
|
||||||
abd_free(col->rc_abd);
|
|
||||||
|
|
||||||
col->rc_abd = tmp;
|
|
||||||
offset += col->rc_size;
|
|
||||||
}
|
|
||||||
ASSERT3U(offset, ==, size);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
const zio_vsd_ops_t vdev_draid_vsd_ops = {
|
|
||||||
.vsd_free = vdev_draid_map_free_vsd,
|
|
||||||
.vsd_cksum_report = vdev_draid_cksum_report
|
|
||||||
};
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Full stripe writes. When writing, all columns (D+P) are required. Parity
|
* Full stripe writes. When writing, all columns (D+P) are required. Parity
|
||||||
* is calculated over all the columns, including empty zero filled sectors,
|
* is calculated over all the columns, including empty zero filled sectors,
|
||||||
|
@ -1208,7 +978,6 @@ vdev_draid_map_alloc_row(zio_t *zio, raidz_row_t **rrp, uint64_t io_offset,
|
||||||
rr->rr_missingdata = 0;
|
rr->rr_missingdata = 0;
|
||||||
rr->rr_missingparity = 0;
|
rr->rr_missingparity = 0;
|
||||||
rr->rr_firstdatacol = vdc->vdc_nparity;
|
rr->rr_firstdatacol = vdc->vdc_nparity;
|
||||||
rr->rr_abd_copy = NULL;
|
|
||||||
rr->rr_abd_empty = NULL;
|
rr->rr_abd_empty = NULL;
|
||||||
#ifdef ZFS_DEBUG
|
#ifdef ZFS_DEBUG
|
||||||
rr->rr_offset = io_offset;
|
rr->rr_offset = io_offset;
|
||||||
|
@ -1230,7 +999,6 @@ vdev_draid_map_alloc_row(zio_t *zio, raidz_row_t **rrp, uint64_t io_offset,
|
||||||
rc->rc_devidx = vdev_draid_permute_id(vdc, base, iter, c);
|
rc->rc_devidx = vdev_draid_permute_id(vdc, base, iter, c);
|
||||||
rc->rc_offset = physical_offset;
|
rc->rc_offset = physical_offset;
|
||||||
rc->rc_abd = NULL;
|
rc->rc_abd = NULL;
|
||||||
rc->rc_gdata = NULL;
|
|
||||||
rc->rc_orig_data = NULL;
|
rc->rc_orig_data = NULL;
|
||||||
rc->rc_error = 0;
|
rc->rc_error = 0;
|
||||||
rc->rc_tried = 0;
|
rc->rc_tried = 0;
|
||||||
|
@ -1328,9 +1096,6 @@ vdev_draid_map_alloc(zio_t *zio)
|
||||||
if (nrows == 2)
|
if (nrows == 2)
|
||||||
rm->rm_row[1] = rr[1];
|
rm->rm_row[1] = rr[1];
|
||||||
|
|
||||||
zio->io_vsd = rm;
|
|
||||||
zio->io_vsd_ops = &vdev_draid_vsd_ops;
|
|
||||||
|
|
||||||
return (rm);
|
return (rm);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2183,12 +1948,13 @@ static void
|
||||||
vdev_draid_io_start(zio_t *zio)
|
vdev_draid_io_start(zio_t *zio)
|
||||||
{
|
{
|
||||||
vdev_t *vd __maybe_unused = zio->io_vd;
|
vdev_t *vd __maybe_unused = zio->io_vd;
|
||||||
raidz_map_t *rm;
|
|
||||||
|
|
||||||
ASSERT3P(vd->vdev_ops, ==, &vdev_draid_ops);
|
ASSERT3P(vd->vdev_ops, ==, &vdev_draid_ops);
|
||||||
ASSERT3U(zio->io_offset, ==, vdev_draid_get_astart(vd, zio->io_offset));
|
ASSERT3U(zio->io_offset, ==, vdev_draid_get_astart(vd, zio->io_offset));
|
||||||
|
|
||||||
rm = vdev_draid_map_alloc(zio);
|
raidz_map_t *rm = vdev_draid_map_alloc(zio);
|
||||||
|
zio->io_vsd = rm;
|
||||||
|
zio->io_vsd_ops = &vdev_raidz_vsd_ops;
|
||||||
|
|
||||||
if (zio->io_type == ZIO_TYPE_WRITE) {
|
if (zio->io_type == ZIO_TYPE_WRITE) {
|
||||||
for (int i = 0; i < rm->rm_nrows; i++) {
|
for (int i = 0; i < rm->rm_nrows; i++) {
|
||||||
|
|
|
@ -315,7 +315,6 @@ vdev_indirect_map_free(zio_t *zio)
|
||||||
|
|
||||||
static const zio_vsd_ops_t vdev_indirect_vsd_ops = {
|
static const zio_vsd_ops_t vdev_indirect_vsd_ops = {
|
||||||
.vsd_free = vdev_indirect_map_free,
|
.vsd_free = vdev_indirect_map_free,
|
||||||
.vsd_cksum_report = zio_vsd_default_cksum_report
|
|
||||||
};
|
};
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
|
|
@ -174,7 +174,6 @@ vdev_mirror_map_free(zio_t *zio)
|
||||||
|
|
||||||
static const zio_vsd_ops_t vdev_mirror_vsd_ops = {
|
static const zio_vsd_ops_t vdev_mirror_vsd_ops = {
|
||||||
.vsd_free = vdev_mirror_map_free,
|
.vsd_free = vdev_mirror_map_free,
|
||||||
.vsd_cksum_report = zio_vsd_default_cksum_report
|
|
||||||
};
|
};
|
||||||
|
|
||||||
static int
|
static int
|
||||||
|
@ -379,8 +378,6 @@ vdev_mirror_map_init(zio_t *zio)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
zio->io_vsd = mm;
|
|
||||||
zio->io_vsd_ops = &vdev_mirror_vsd_ops;
|
|
||||||
return (mm);
|
return (mm);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -629,6 +626,8 @@ vdev_mirror_io_start(zio_t *zio)
|
||||||
int c, children;
|
int c, children;
|
||||||
|
|
||||||
mm = vdev_mirror_map_init(zio);
|
mm = vdev_mirror_map_init(zio);
|
||||||
|
zio->io_vsd = mm;
|
||||||
|
zio->io_vsd_ops = &vdev_mirror_vsd_ops;
|
||||||
|
|
||||||
if (mm == NULL) {
|
if (mm == NULL) {
|
||||||
ASSERT(!spa_trust_config(zio->io_spa));
|
ASSERT(!spa_trust_config(zio->io_spa));
|
||||||
|
|
|
@ -143,15 +143,10 @@ vdev_raidz_row_free(raidz_row_t *rr)
|
||||||
|
|
||||||
if (rc->rc_size != 0)
|
if (rc->rc_size != 0)
|
||||||
abd_free(rc->rc_abd);
|
abd_free(rc->rc_abd);
|
||||||
if (rc->rc_gdata != NULL)
|
|
||||||
abd_free(rc->rc_gdata);
|
|
||||||
if (rc->rc_orig_data != NULL)
|
if (rc->rc_orig_data != NULL)
|
||||||
zio_buf_free(rc->rc_orig_data, rc->rc_size);
|
abd_free(rc->rc_orig_data);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (rr->rr_abd_copy != NULL)
|
|
||||||
abd_free(rr->rr_abd_copy);
|
|
||||||
|
|
||||||
if (rr->rr_abd_empty != NULL)
|
if (rr->rr_abd_empty != NULL)
|
||||||
abd_free(rr->rr_abd_empty);
|
abd_free(rr->rr_abd_empty);
|
||||||
|
|
||||||
|
@ -172,175 +167,11 @@ vdev_raidz_map_free_vsd(zio_t *zio)
|
||||||
{
|
{
|
||||||
raidz_map_t *rm = zio->io_vsd;
|
raidz_map_t *rm = zio->io_vsd;
|
||||||
|
|
||||||
ASSERT0(rm->rm_freed);
|
vdev_raidz_map_free(rm);
|
||||||
rm->rm_freed = B_TRUE;
|
|
||||||
|
|
||||||
if (rm->rm_reports == 0) {
|
|
||||||
vdev_raidz_map_free(rm);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/*ARGSUSED*/
|
const zio_vsd_ops_t vdev_raidz_vsd_ops = {
|
||||||
static void
|
|
||||||
vdev_raidz_cksum_free(void *arg, size_t ignored)
|
|
||||||
{
|
|
||||||
raidz_map_t *rm = arg;
|
|
||||||
|
|
||||||
ASSERT3U(rm->rm_reports, >, 0);
|
|
||||||
|
|
||||||
if (--rm->rm_reports == 0 && rm->rm_freed)
|
|
||||||
vdev_raidz_map_free(rm);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void
|
|
||||||
vdev_raidz_cksum_finish(zio_cksum_report_t *zcr, const abd_t *good_data)
|
|
||||||
{
|
|
||||||
raidz_map_t *rm = zcr->zcr_cbdata;
|
|
||||||
const size_t c = zcr->zcr_cbinfo;
|
|
||||||
size_t x, offset;
|
|
||||||
|
|
||||||
if (good_data == NULL) {
|
|
||||||
zfs_ereport_finish_checksum(zcr, NULL, NULL, B_FALSE);
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
ASSERT3U(rm->rm_nrows, ==, 1);
|
|
||||||
raidz_row_t *rr = rm->rm_row[0];
|
|
||||||
|
|
||||||
const abd_t *good = NULL;
|
|
||||||
const abd_t *bad = rr->rr_col[c].rc_abd;
|
|
||||||
|
|
||||||
if (c < rr->rr_firstdatacol) {
|
|
||||||
/*
|
|
||||||
* The first time through, calculate the parity blocks for
|
|
||||||
* the good data (this relies on the fact that the good
|
|
||||||
* data never changes for a given logical ZIO)
|
|
||||||
*/
|
|
||||||
if (rr->rr_col[0].rc_gdata == NULL) {
|
|
||||||
abd_t *bad_parity[VDEV_RAIDZ_MAXPARITY];
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Set up the rr_col[]s to generate the parity for
|
|
||||||
* good_data, first saving the parity bufs and
|
|
||||||
* replacing them with buffers to hold the result.
|
|
||||||
*/
|
|
||||||
for (x = 0; x < rr->rr_firstdatacol; x++) {
|
|
||||||
bad_parity[x] = rr->rr_col[x].rc_abd;
|
|
||||||
rr->rr_col[x].rc_abd = rr->rr_col[x].rc_gdata =
|
|
||||||
abd_alloc_sametype(rr->rr_col[x].rc_abd,
|
|
||||||
rr->rr_col[x].rc_size);
|
|
||||||
}
|
|
||||||
|
|
||||||
/* fill in the data columns from good_data */
|
|
||||||
offset = 0;
|
|
||||||
for (; x < rr->rr_cols; x++) {
|
|
||||||
abd_free(rr->rr_col[x].rc_abd);
|
|
||||||
|
|
||||||
rr->rr_col[x].rc_abd =
|
|
||||||
abd_get_offset_size((abd_t *)good_data,
|
|
||||||
offset, rr->rr_col[x].rc_size);
|
|
||||||
offset += rr->rr_col[x].rc_size;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Construct the parity from the good data.
|
|
||||||
*/
|
|
||||||
vdev_raidz_generate_parity_row(rm, rr);
|
|
||||||
|
|
||||||
/* restore everything back to its original state */
|
|
||||||
for (x = 0; x < rr->rr_firstdatacol; x++)
|
|
||||||
rr->rr_col[x].rc_abd = bad_parity[x];
|
|
||||||
|
|
||||||
offset = 0;
|
|
||||||
for (x = rr->rr_firstdatacol; x < rr->rr_cols; x++) {
|
|
||||||
abd_free(rr->rr_col[x].rc_abd);
|
|
||||||
rr->rr_col[x].rc_abd = abd_get_offset_size(
|
|
||||||
rr->rr_abd_copy, offset,
|
|
||||||
rr->rr_col[x].rc_size);
|
|
||||||
offset += rr->rr_col[x].rc_size;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
ASSERT3P(rr->rr_col[c].rc_gdata, !=, NULL);
|
|
||||||
good = abd_get_offset_size(rr->rr_col[c].rc_gdata, 0,
|
|
||||||
rr->rr_col[c].rc_size);
|
|
||||||
} else {
|
|
||||||
/* adjust good_data to point at the start of our column */
|
|
||||||
offset = 0;
|
|
||||||
for (x = rr->rr_firstdatacol; x < c; x++)
|
|
||||||
offset += rr->rr_col[x].rc_size;
|
|
||||||
|
|
||||||
good = abd_get_offset_size((abd_t *)good_data, offset,
|
|
||||||
rr->rr_col[c].rc_size);
|
|
||||||
}
|
|
||||||
|
|
||||||
/* we drop the ereport if it ends up that the data was good */
|
|
||||||
zfs_ereport_finish_checksum(zcr, good, bad, B_TRUE);
|
|
||||||
abd_free((abd_t *)good);
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Invoked indirectly by zfs_ereport_start_checksum(), called
|
|
||||||
* below when our read operation fails completely. The main point
|
|
||||||
* is to keep a copy of everything we read from disk, so that at
|
|
||||||
* vdev_raidz_cksum_finish() time we can compare it with the good data.
|
|
||||||
*/
|
|
||||||
static void
|
|
||||||
vdev_raidz_cksum_report(zio_t *zio, zio_cksum_report_t *zcr, void *arg)
|
|
||||||
{
|
|
||||||
size_t c = (size_t)(uintptr_t)arg;
|
|
||||||
raidz_map_t *rm = zio->io_vsd;
|
|
||||||
|
|
||||||
/* set up the report and bump the refcount */
|
|
||||||
zcr->zcr_cbdata = rm;
|
|
||||||
zcr->zcr_cbinfo = c;
|
|
||||||
zcr->zcr_finish = vdev_raidz_cksum_finish;
|
|
||||||
zcr->zcr_free = vdev_raidz_cksum_free;
|
|
||||||
|
|
||||||
rm->rm_reports++;
|
|
||||||
ASSERT3U(rm->rm_reports, >, 0);
|
|
||||||
ASSERT3U(rm->rm_nrows, ==, 1);
|
|
||||||
|
|
||||||
if (rm->rm_row[0]->rr_abd_copy != NULL)
|
|
||||||
return;
|
|
||||||
|
|
||||||
/*
|
|
||||||
* It's the first time we're called for this raidz_map_t, so we need
|
|
||||||
* to copy the data aside; there's no guarantee that our zio's buffer
|
|
||||||
* won't be re-used for something else.
|
|
||||||
*
|
|
||||||
* Our parity data is already in separate buffers, so there's no need
|
|
||||||
* to copy them.
|
|
||||||
*/
|
|
||||||
for (int i = 0; i < rm->rm_nrows; i++) {
|
|
||||||
raidz_row_t *rr = rm->rm_row[i];
|
|
||||||
size_t offset = 0;
|
|
||||||
size_t size = 0;
|
|
||||||
|
|
||||||
for (c = rr->rr_firstdatacol; c < rr->rr_cols; c++)
|
|
||||||
size += rr->rr_col[c].rc_size;
|
|
||||||
|
|
||||||
rr->rr_abd_copy = abd_alloc_for_io(size, B_FALSE);
|
|
||||||
|
|
||||||
for (c = rr->rr_firstdatacol; c < rr->rr_cols; c++) {
|
|
||||||
raidz_col_t *col = &rr->rr_col[c];
|
|
||||||
abd_t *tmp = abd_get_offset_size(rr->rr_abd_copy,
|
|
||||||
offset, col->rc_size);
|
|
||||||
|
|
||||||
abd_copy(tmp, col->rc_abd, col->rc_size);
|
|
||||||
|
|
||||||
abd_free(col->rc_abd);
|
|
||||||
col->rc_abd = tmp;
|
|
||||||
|
|
||||||
offset += col->rc_size;
|
|
||||||
}
|
|
||||||
ASSERT3U(offset, ==, size);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
static const zio_vsd_ops_t vdev_raidz_vsd_ops = {
|
|
||||||
.vsd_free = vdev_raidz_map_free_vsd,
|
.vsd_free = vdev_raidz_map_free_vsd,
|
||||||
.vsd_cksum_report = vdev_raidz_cksum_report
|
|
||||||
};
|
};
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -414,7 +245,6 @@ vdev_raidz_map_alloc(zio_t *zio, uint64_t ashift, uint64_t dcols,
|
||||||
rr->rr_missingdata = 0;
|
rr->rr_missingdata = 0;
|
||||||
rr->rr_missingparity = 0;
|
rr->rr_missingparity = 0;
|
||||||
rr->rr_firstdatacol = nparity;
|
rr->rr_firstdatacol = nparity;
|
||||||
rr->rr_abd_copy = NULL;
|
|
||||||
rr->rr_abd_empty = NULL;
|
rr->rr_abd_empty = NULL;
|
||||||
rr->rr_nempty = 0;
|
rr->rr_nempty = 0;
|
||||||
#ifdef ZFS_DEBUG
|
#ifdef ZFS_DEBUG
|
||||||
|
@ -435,7 +265,6 @@ vdev_raidz_map_alloc(zio_t *zio, uint64_t ashift, uint64_t dcols,
|
||||||
rc->rc_devidx = col;
|
rc->rc_devidx = col;
|
||||||
rc->rc_offset = coff;
|
rc->rc_offset = coff;
|
||||||
rc->rc_abd = NULL;
|
rc->rc_abd = NULL;
|
||||||
rc->rc_gdata = NULL;
|
|
||||||
rc->rc_orig_data = NULL;
|
rc->rc_orig_data = NULL;
|
||||||
rc->rc_error = 0;
|
rc->rc_error = 0;
|
||||||
rc->rc_tried = 0;
|
rc->rc_tried = 0;
|
||||||
|
@ -1798,10 +1627,11 @@ vdev_raidz_io_start(zio_t *zio)
|
||||||
vdev_t *vd = zio->io_vd;
|
vdev_t *vd = zio->io_vd;
|
||||||
vdev_t *tvd = vd->vdev_top;
|
vdev_t *tvd = vd->vdev_top;
|
||||||
vdev_raidz_t *vdrz = vd->vdev_tsd;
|
vdev_raidz_t *vdrz = vd->vdev_tsd;
|
||||||
raidz_map_t *rm;
|
|
||||||
|
|
||||||
rm = vdev_raidz_map_alloc(zio, tvd->vdev_ashift,
|
raidz_map_t *rm = vdev_raidz_map_alloc(zio, tvd->vdev_ashift,
|
||||||
vdrz->vd_logical_width, vdrz->vd_nparity);
|
vdrz->vd_logical_width, vdrz->vd_nparity);
|
||||||
|
zio->io_vsd = rm;
|
||||||
|
zio->io_vsd_ops = &vdev_raidz_vsd_ops;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Until raidz expansion is implemented all maps for a raidz vdev
|
* Until raidz expansion is implemented all maps for a raidz vdev
|
||||||
|
@ -1810,9 +1640,6 @@ vdev_raidz_io_start(zio_t *zio)
|
||||||
ASSERT3U(rm->rm_nrows, ==, 1);
|
ASSERT3U(rm->rm_nrows, ==, 1);
|
||||||
raidz_row_t *rr = rm->rm_row[0];
|
raidz_row_t *rr = rm->rm_row[0];
|
||||||
|
|
||||||
zio->io_vsd = rm;
|
|
||||||
zio->io_vsd_ops = &vdev_raidz_vsd_ops;
|
|
||||||
|
|
||||||
if (zio->io_type == ZIO_TYPE_WRITE) {
|
if (zio->io_type == ZIO_TYPE_WRITE) {
|
||||||
vdev_raidz_io_start_write(zio, rr, tvd->vdev_ashift);
|
vdev_raidz_io_start_write(zio, rr, tvd->vdev_ashift);
|
||||||
} else {
|
} else {
|
||||||
|
@ -2008,7 +1835,7 @@ raidz_restore_orig_data(raidz_map_t *rm)
|
||||||
for (int c = 0; c < rr->rr_cols; c++) {
|
for (int c = 0; c < rr->rr_cols; c++) {
|
||||||
raidz_col_t *rc = &rr->rr_col[c];
|
raidz_col_t *rc = &rr->rr_col[c];
|
||||||
if (rc->rc_need_orig_restore) {
|
if (rc->rc_need_orig_restore) {
|
||||||
abd_copy_from_buf(rc->rc_abd,
|
abd_copy(rc->rc_abd,
|
||||||
rc->rc_orig_data, rc->rc_size);
|
rc->rc_orig_data, rc->rc_size);
|
||||||
rc->rc_need_orig_restore = B_FALSE;
|
rc->rc_need_orig_restore = B_FALSE;
|
||||||
}
|
}
|
||||||
|
@ -2049,9 +1876,9 @@ raidz_reconstruct(zio_t *zio, int *ltgts, int ntgts, int nparity)
|
||||||
if (rc->rc_devidx == ltgts[lt]) {
|
if (rc->rc_devidx == ltgts[lt]) {
|
||||||
if (rc->rc_orig_data == NULL) {
|
if (rc->rc_orig_data == NULL) {
|
||||||
rc->rc_orig_data =
|
rc->rc_orig_data =
|
||||||
zio_buf_alloc(rc->rc_size);
|
abd_alloc_linear(
|
||||||
abd_copy_to_buf(
|
rc->rc_size, B_TRUE);
|
||||||
rc->rc_orig_data,
|
abd_copy(rc->rc_orig_data,
|
||||||
rc->rc_abd, rc->rc_size);
|
rc->rc_abd, rc->rc_size);
|
||||||
}
|
}
|
||||||
rc->rc_need_orig_restore = B_TRUE;
|
rc->rc_need_orig_restore = B_TRUE;
|
||||||
|
@ -2096,7 +1923,7 @@ raidz_reconstruct(zio_t *zio, int *ltgts, int ntgts, int nparity)
|
||||||
if (rc->rc_error == 0 &&
|
if (rc->rc_error == 0 &&
|
||||||
c >= rr->rr_firstdatacol) {
|
c >= rr->rr_firstdatacol) {
|
||||||
raidz_checksum_error(zio,
|
raidz_checksum_error(zio,
|
||||||
rc, rc->rc_gdata);
|
rc, rc->rc_orig_data);
|
||||||
rc->rc_error =
|
rc->rc_error =
|
||||||
SET_ERROR(ECKSUM);
|
SET_ERROR(ECKSUM);
|
||||||
}
|
}
|
||||||
|
@ -2431,7 +2258,7 @@ vdev_raidz_io_done_unrecoverable(zio_t *zio)
|
||||||
|
|
||||||
(void) zfs_ereport_start_checksum(zio->io_spa,
|
(void) zfs_ereport_start_checksum(zio->io_spa,
|
||||||
cvd, &zio->io_bookmark, zio, rc->rc_offset,
|
cvd, &zio->io_bookmark, zio, rc->rc_offset,
|
||||||
rc->rc_size, (void *)(uintptr_t)c, &zbc);
|
rc->rc_size, &zbc);
|
||||||
mutex_enter(&cvd->vdev_stat_lock);
|
mutex_enter(&cvd->vdev_stat_lock);
|
||||||
cvd->vdev_stat.vs_checksum_errors++;
|
cvd->vdev_stat.vs_checksum_errors++;
|
||||||
mutex_exit(&cvd->vdev_stat_lock);
|
mutex_exit(&cvd->vdev_stat_lock);
|
||||||
|
|
|
@ -1125,8 +1125,7 @@ zfs_ereport_post(const char *subclass, spa_t *spa, vdev_t *vd,
|
||||||
*/
|
*/
|
||||||
int
|
int
|
||||||
zfs_ereport_start_checksum(spa_t *spa, vdev_t *vd, const zbookmark_phys_t *zb,
|
zfs_ereport_start_checksum(spa_t *spa, vdev_t *vd, const zbookmark_phys_t *zb,
|
||||||
struct zio *zio, uint64_t offset, uint64_t length, void *arg,
|
struct zio *zio, uint64_t offset, uint64_t length, zio_bad_cksum_t *info)
|
||||||
zio_bad_cksum_t *info)
|
|
||||||
{
|
{
|
||||||
zio_cksum_report_t *report;
|
zio_cksum_report_t *report;
|
||||||
|
|
||||||
|
@ -1144,10 +1143,7 @@ zfs_ereport_start_checksum(spa_t *spa, vdev_t *vd, const zbookmark_phys_t *zb,
|
||||||
|
|
||||||
report = kmem_zalloc(sizeof (*report), KM_SLEEP);
|
report = kmem_zalloc(sizeof (*report), KM_SLEEP);
|
||||||
|
|
||||||
if (zio->io_vsd != NULL)
|
zio_vsd_default_cksum_report(zio, report);
|
||||||
zio->io_vsd_ops->vsd_cksum_report(zio, report, arg);
|
|
||||||
else
|
|
||||||
zio_vsd_default_cksum_report(zio, report, arg);
|
|
||||||
|
|
||||||
/* copy the checksum failure information if it was provided */
|
/* copy the checksum failure information if it was provided */
|
||||||
if (info != NULL) {
|
if (info != NULL) {
|
||||||
|
|
|
@ -3919,7 +3919,7 @@ zio_vsd_default_cksum_finish(zio_cksum_report_t *zcr,
|
||||||
|
|
||||||
/*ARGSUSED*/
|
/*ARGSUSED*/
|
||||||
void
|
void
|
||||||
zio_vsd_default_cksum_report(zio_t *zio, zio_cksum_report_t *zcr, void *ignored)
|
zio_vsd_default_cksum_report(zio_t *zio, zio_cksum_report_t *zcr)
|
||||||
{
|
{
|
||||||
void *abd = abd_alloc_sametype(zio->io_abd, zio->io_size);
|
void *abd = abd_alloc_sametype(zio->io_abd, zio->io_size);
|
||||||
|
|
||||||
|
@ -4257,7 +4257,7 @@ zio_checksum_verify(zio_t *zio)
|
||||||
!(zio->io_flags & ZIO_FLAG_SPECULATIVE)) {
|
!(zio->io_flags & ZIO_FLAG_SPECULATIVE)) {
|
||||||
(void) zfs_ereport_start_checksum(zio->io_spa,
|
(void) zfs_ereport_start_checksum(zio->io_spa,
|
||||||
zio->io_vd, &zio->io_bookmark, zio,
|
zio->io_vd, &zio->io_bookmark, zio,
|
||||||
zio->io_offset, zio->io_size, NULL, &info);
|
zio->io_offset, zio->io_size, &info);
|
||||||
mutex_enter(&zio->io_vd->vdev_stat_lock);
|
mutex_enter(&zio->io_vd->vdev_stat_lock);
|
||||||
zio->io_vd->vdev_stat.vs_checksum_errors++;
|
zio->io_vd->vdev_stat.vs_checksum_errors++;
|
||||||
mutex_exit(&zio->io_vd->vdev_stat_lock);
|
mutex_exit(&zio->io_vd->vdev_stat_lock);
|
||||||
|
|
Loading…
Reference in New Issue