Switch refcount tracking from lists to AVL-trees.

With large number of tracked references list searches under the lock
become too expensive, creating enormous lock contention.

On my tests with ZFS_DEBUG enabled this increases write throughput
with 32KB blocks from ~1.2GB/s to ~7.5GB/s.

Reviewed-by: Brian Atkinson <batkinson@lanl.gov>
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
Signed-off-by: Alexander Motin <mav@FreeBSD.org>
Sponsored by:	iXsystems, Inc.
Closes #14970
This commit is contained in:
Alexander Motin 2023-06-14 11:02:27 -04:00 committed by GitHub
parent 8af1104f83
commit d057807ede
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
2 changed files with 110 additions and 97 deletions

View File

@ -27,6 +27,7 @@
#define _SYS_ZFS_REFCOUNT_H #define _SYS_ZFS_REFCOUNT_H
#include <sys/inttypes.h> #include <sys/inttypes.h>
#include <sys/avl.h>
#include <sys/list.h> #include <sys/list.h>
#include <sys/zfs_context.h> #include <sys/zfs_context.h>
@ -43,19 +44,22 @@ extern "C" {
#ifdef ZFS_DEBUG #ifdef ZFS_DEBUG
typedef struct reference { typedef struct reference {
list_node_t ref_link; union {
avl_node_t a;
list_node_t l;
} ref_link;
const void *ref_holder; const void *ref_holder;
uint64_t ref_number; uint64_t ref_number;
uint8_t *ref_removed; boolean_t ref_search;
} reference_t; } reference_t;
typedef struct refcount { typedef struct refcount {
kmutex_t rc_mtx;
boolean_t rc_tracked;
list_t rc_list;
list_t rc_removed;
uint64_t rc_count; uint64_t rc_count;
uint64_t rc_removed_count; kmutex_t rc_mtx;
avl_tree_t rc_tree;
list_t rc_removed;
uint_t rc_removed_count;
boolean_t rc_tracked;
} zfs_refcount_t; } zfs_refcount_t;
/* /*

View File

@ -36,33 +36,40 @@ int reference_tracking_enable = B_FALSE;
static uint_t reference_history = 3; /* tunable */ static uint_t reference_history = 3; /* tunable */
static kmem_cache_t *reference_cache; static kmem_cache_t *reference_cache;
static kmem_cache_t *reference_history_cache;
void void
zfs_refcount_init(void) zfs_refcount_init(void)
{ {
reference_cache = kmem_cache_create("reference_cache", reference_cache = kmem_cache_create("reference_cache",
sizeof (reference_t), 0, NULL, NULL, NULL, NULL, NULL, 0); sizeof (reference_t), 0, NULL, NULL, NULL, NULL, NULL, 0);
reference_history_cache = kmem_cache_create("reference_history_cache",
sizeof (uint64_t), 0, NULL, NULL, NULL, NULL, NULL, 0);
} }
void void
zfs_refcount_fini(void) zfs_refcount_fini(void)
{ {
kmem_cache_destroy(reference_cache); kmem_cache_destroy(reference_cache);
kmem_cache_destroy(reference_history_cache); }
static int
zfs_refcount_compare(const void *x1, const void *x2)
{
const reference_t *r1 = (const reference_t *)x1;
const reference_t *r2 = (const reference_t *)x2;
int cmp1 = TREE_CMP(r1->ref_holder, r2->ref_holder);
int cmp2 = TREE_CMP(r1->ref_number, r2->ref_number);
int cmp = cmp1 ? cmp1 : cmp2;
return ((cmp || r1->ref_search) ? cmp : TREE_PCMP(r1, r2));
} }
void void
zfs_refcount_create(zfs_refcount_t *rc) zfs_refcount_create(zfs_refcount_t *rc)
{ {
mutex_init(&rc->rc_mtx, NULL, MUTEX_DEFAULT, NULL); mutex_init(&rc->rc_mtx, NULL, MUTEX_DEFAULT, NULL);
list_create(&rc->rc_list, sizeof (reference_t), avl_create(&rc->rc_tree, zfs_refcount_compare, sizeof (reference_t),
offsetof(reference_t, ref_link)); offsetof(reference_t, ref_link.a));
list_create(&rc->rc_removed, sizeof (reference_t), list_create(&rc->rc_removed, sizeof (reference_t),
offsetof(reference_t, ref_link)); offsetof(reference_t, ref_link.l));
rc->rc_count = 0; rc->rc_count = 0;
rc->rc_removed_count = 0; rc->rc_removed_count = 0;
rc->rc_tracked = reference_tracking_enable; rc->rc_tracked = reference_tracking_enable;
@ -86,16 +93,15 @@ void
zfs_refcount_destroy_many(zfs_refcount_t *rc, uint64_t number) zfs_refcount_destroy_many(zfs_refcount_t *rc, uint64_t number)
{ {
reference_t *ref; reference_t *ref;
void *cookie = NULL;
ASSERT3U(rc->rc_count, ==, number); ASSERT3U(rc->rc_count, ==, number);
while ((ref = list_remove_head(&rc->rc_list))) while ((ref = avl_destroy_nodes(&rc->rc_tree, &cookie)) != NULL)
kmem_cache_free(reference_cache, ref); kmem_cache_free(reference_cache, ref);
list_destroy(&rc->rc_list); avl_destroy(&rc->rc_tree);
while ((ref = list_remove_head(&rc->rc_removed))) { while ((ref = list_remove_head(&rc->rc_removed)))
kmem_cache_free(reference_history_cache, ref->ref_removed);
kmem_cache_free(reference_cache, ref); kmem_cache_free(reference_cache, ref);
}
list_destroy(&rc->rc_removed); list_destroy(&rc->rc_removed);
mutex_destroy(&rc->rc_mtx); mutex_destroy(&rc->rc_mtx);
} }
@ -121,10 +127,10 @@ zfs_refcount_count(zfs_refcount_t *rc)
int64_t int64_t
zfs_refcount_add_many(zfs_refcount_t *rc, uint64_t number, const void *holder) zfs_refcount_add_many(zfs_refcount_t *rc, uint64_t number, const void *holder)
{ {
reference_t *ref = NULL; reference_t *ref;
int64_t count; int64_t count;
if (!rc->rc_tracked) { if (likely(!rc->rc_tracked)) {
count = atomic_add_64_nv(&(rc)->rc_count, number); count = atomic_add_64_nv(&(rc)->rc_count, number);
ASSERT3U(count, >=, number); ASSERT3U(count, >=, number);
return (count); return (count);
@ -133,8 +139,9 @@ zfs_refcount_add_many(zfs_refcount_t *rc, uint64_t number, const void *holder)
ref = kmem_cache_alloc(reference_cache, KM_SLEEP); ref = kmem_cache_alloc(reference_cache, KM_SLEEP);
ref->ref_holder = holder; ref->ref_holder = holder;
ref->ref_number = number; ref->ref_number = number;
ref->ref_search = B_FALSE;
mutex_enter(&rc->rc_mtx); mutex_enter(&rc->rc_mtx);
list_insert_head(&rc->rc_list, ref); avl_add(&rc->rc_tree, ref);
rc->rc_count += number; rc->rc_count += number;
count = rc->rc_count; count = rc->rc_count;
mutex_exit(&rc->rc_mtx); mutex_exit(&rc->rc_mtx);
@ -151,7 +158,7 @@ zfs_refcount_add(zfs_refcount_t *rc, const void *holder)
void void
zfs_refcount_add_few(zfs_refcount_t *rc, uint64_t number, const void *holder) zfs_refcount_add_few(zfs_refcount_t *rc, uint64_t number, const void *holder)
{ {
if (!rc->rc_tracked) if (likely(!rc->rc_tracked))
(void) zfs_refcount_add_many(rc, number, holder); (void) zfs_refcount_add_many(rc, number, holder);
else for (; number > 0; number--) else for (; number > 0; number--)
(void) zfs_refcount_add(rc, holder); (void) zfs_refcount_add(rc, holder);
@ -161,47 +168,42 @@ int64_t
zfs_refcount_remove_many(zfs_refcount_t *rc, uint64_t number, zfs_refcount_remove_many(zfs_refcount_t *rc, uint64_t number,
const void *holder) const void *holder)
{ {
reference_t *ref; reference_t *ref, s;
int64_t count; int64_t count;
if (!rc->rc_tracked) { if (likely(!rc->rc_tracked)) {
count = atomic_add_64_nv(&(rc)->rc_count, -number); count = atomic_add_64_nv(&(rc)->rc_count, -number);
ASSERT3S(count, >=, 0); ASSERT3S(count, >=, 0);
return (count); return (count);
} }
s.ref_holder = holder;
s.ref_number = number;
s.ref_search = B_TRUE;
mutex_enter(&rc->rc_mtx); mutex_enter(&rc->rc_mtx);
ASSERT3U(rc->rc_count, >=, number); ASSERT3U(rc->rc_count, >=, number);
for (ref = list_head(&rc->rc_list); ref; ref = avl_find(&rc->rc_tree, &s, NULL);
ref = list_next(&rc->rc_list, ref)) { if (unlikely(ref == NULL)) {
if (ref->ref_holder == holder && ref->ref_number == number) { panic("No such hold %p on refcount %llx", holder,
list_remove(&rc->rc_list, ref); (u_longlong_t)(uintptr_t)rc);
if (reference_history > 0) { return (-1);
ref->ref_removed =
kmem_cache_alloc(reference_history_cache,
KM_SLEEP);
list_insert_head(&rc->rc_removed, ref);
rc->rc_removed_count++;
if (rc->rc_removed_count > reference_history) {
ref = list_tail(&rc->rc_removed);
list_remove(&rc->rc_removed, ref);
kmem_cache_free(reference_history_cache,
ref->ref_removed);
kmem_cache_free(reference_cache, ref);
rc->rc_removed_count--;
}
} else {
kmem_cache_free(reference_cache, ref);
}
rc->rc_count -= number;
count = rc->rc_count;
mutex_exit(&rc->rc_mtx);
return (count);
}
} }
panic("No such hold %p on refcount %llx", holder, avl_remove(&rc->rc_tree, ref);
(u_longlong_t)(uintptr_t)rc); if (reference_history > 0) {
return (-1); list_insert_head(&rc->rc_removed, ref);
if (rc->rc_removed_count >= reference_history) {
ref = list_remove_tail(&rc->rc_removed);
kmem_cache_free(reference_cache, ref);
} else {
rc->rc_removed_count++;
}
} else {
kmem_cache_free(reference_cache, ref);
}
rc->rc_count -= number;
count = rc->rc_count;
mutex_exit(&rc->rc_mtx);
return (count);
} }
int64_t int64_t
@ -213,7 +215,7 @@ zfs_refcount_remove(zfs_refcount_t *rc, const void *holder)
void void
zfs_refcount_remove_few(zfs_refcount_t *rc, uint64_t number, const void *holder) zfs_refcount_remove_few(zfs_refcount_t *rc, uint64_t number, const void *holder)
{ {
if (!rc->rc_tracked) if (likely(!rc->rc_tracked))
(void) zfs_refcount_remove_many(rc, number, holder); (void) zfs_refcount_remove_many(rc, number, holder);
else for (; number > 0; number--) else for (; number > 0; number--)
(void) zfs_refcount_remove(rc, holder); (void) zfs_refcount_remove(rc, holder);
@ -222,31 +224,38 @@ zfs_refcount_remove_few(zfs_refcount_t *rc, uint64_t number, const void *holder)
void void
zfs_refcount_transfer(zfs_refcount_t *dst, zfs_refcount_t *src) zfs_refcount_transfer(zfs_refcount_t *dst, zfs_refcount_t *src)
{ {
int64_t count, removed_count; avl_tree_t tree;
list_t list, removed; list_t removed;
reference_t *ref;
void *cookie = NULL;
uint64_t count;
uint_t removed_count;
list_create(&list, sizeof (reference_t), avl_create(&tree, zfs_refcount_compare, sizeof (reference_t),
offsetof(reference_t, ref_link)); offsetof(reference_t, ref_link.a));
list_create(&removed, sizeof (reference_t), list_create(&removed, sizeof (reference_t),
offsetof(reference_t, ref_link)); offsetof(reference_t, ref_link.l));
mutex_enter(&src->rc_mtx); mutex_enter(&src->rc_mtx);
count = src->rc_count; count = src->rc_count;
removed_count = src->rc_removed_count; removed_count = src->rc_removed_count;
src->rc_count = 0; src->rc_count = 0;
src->rc_removed_count = 0; src->rc_removed_count = 0;
list_move_tail(&list, &src->rc_list); avl_swap(&tree, &src->rc_tree);
list_move_tail(&removed, &src->rc_removed); list_move_tail(&removed, &src->rc_removed);
mutex_exit(&src->rc_mtx); mutex_exit(&src->rc_mtx);
mutex_enter(&dst->rc_mtx); mutex_enter(&dst->rc_mtx);
dst->rc_count += count; dst->rc_count += count;
dst->rc_removed_count += removed_count; dst->rc_removed_count += removed_count;
list_move_tail(&dst->rc_list, &list); if (avl_is_empty(&dst->rc_tree))
avl_swap(&dst->rc_tree, &tree);
else while ((ref = avl_destroy_nodes(&tree, &cookie)) != NULL)
avl_add(&dst->rc_tree, ref);
list_move_tail(&dst->rc_removed, &removed); list_move_tail(&dst->rc_removed, &removed);
mutex_exit(&dst->rc_mtx); mutex_exit(&dst->rc_mtx);
list_destroy(&list); avl_destroy(&tree);
list_destroy(&removed); list_destroy(&removed);
} }
@ -254,23 +263,19 @@ void
zfs_refcount_transfer_ownership_many(zfs_refcount_t *rc, uint64_t number, zfs_refcount_transfer_ownership_many(zfs_refcount_t *rc, uint64_t number,
const void *current_holder, const void *new_holder) const void *current_holder, const void *new_holder)
{ {
reference_t *ref; reference_t *ref, s;
boolean_t found = B_FALSE;
if (!rc->rc_tracked) if (likely(!rc->rc_tracked))
return; return;
s.ref_holder = current_holder;
s.ref_number = number;
s.ref_search = B_TRUE;
mutex_enter(&rc->rc_mtx); mutex_enter(&rc->rc_mtx);
for (ref = list_head(&rc->rc_list); ref; ref = avl_find(&rc->rc_tree, &s, NULL);
ref = list_next(&rc->rc_list, ref)) { ASSERT(ref);
if (ref->ref_holder == current_holder && ref->ref_holder = new_holder;
ref->ref_number == number) { avl_update(&rc->rc_tree, ref);
ref->ref_holder = new_holder;
found = B_TRUE;
break;
}
}
ASSERT(found);
mutex_exit(&rc->rc_mtx); mutex_exit(&rc->rc_mtx);
} }
@ -290,21 +295,23 @@ zfs_refcount_transfer_ownership(zfs_refcount_t *rc, const void *current_holder,
boolean_t boolean_t
zfs_refcount_held(zfs_refcount_t *rc, const void *holder) zfs_refcount_held(zfs_refcount_t *rc, const void *holder)
{ {
reference_t *ref; reference_t *ref, s;
avl_index_t idx;
boolean_t res;
if (!rc->rc_tracked) if (likely(!rc->rc_tracked))
return (zfs_refcount_count(rc) > 0); return (zfs_refcount_count(rc) > 0);
s.ref_holder = holder;
s.ref_number = 0;
s.ref_search = B_TRUE;
mutex_enter(&rc->rc_mtx); mutex_enter(&rc->rc_mtx);
for (ref = list_head(&rc->rc_list); ref; ref = avl_find(&rc->rc_tree, &s, &idx);
ref = list_next(&rc->rc_list, ref)) { if (likely(ref == NULL))
if (ref->ref_holder == holder) { ref = avl_nearest(&rc->rc_tree, idx, AVL_AFTER);
mutex_exit(&rc->rc_mtx); res = ref && ref->ref_holder == holder;
return (B_TRUE);
}
}
mutex_exit(&rc->rc_mtx); mutex_exit(&rc->rc_mtx);
return (B_FALSE); return (res);
} }
/* /*
@ -315,21 +322,23 @@ zfs_refcount_held(zfs_refcount_t *rc, const void *holder)
boolean_t boolean_t
zfs_refcount_not_held(zfs_refcount_t *rc, const void *holder) zfs_refcount_not_held(zfs_refcount_t *rc, const void *holder)
{ {
reference_t *ref; reference_t *ref, s;
avl_index_t idx;
boolean_t res;
if (!rc->rc_tracked) if (likely(!rc->rc_tracked))
return (B_TRUE); return (B_TRUE);
mutex_enter(&rc->rc_mtx); mutex_enter(&rc->rc_mtx);
for (ref = list_head(&rc->rc_list); ref; s.ref_holder = holder;
ref = list_next(&rc->rc_list, ref)) { s.ref_number = 0;
if (ref->ref_holder == holder) { s.ref_search = B_TRUE;
mutex_exit(&rc->rc_mtx); ref = avl_find(&rc->rc_tree, &s, &idx);
return (B_FALSE); if (likely(ref == NULL))
} ref = avl_nearest(&rc->rc_tree, idx, AVL_AFTER);
} res = ref == NULL || ref->ref_holder != holder;
mutex_exit(&rc->rc_mtx); mutex_exit(&rc->rc_mtx);
return (B_TRUE); return (res);
} }
EXPORT_SYMBOL(zfs_refcount_create); EXPORT_SYMBOL(zfs_refcount_create);