Add a SPL_AC_TYPE_ATOMIC64_T test to configure for systems which do
already supprt atomic64_t types. * spl-07-kmem-cleanup.patch This moves all the debugging code from sys/kmem.h to spl-kmem.c, because the huge macros were hard to debug and were bloating functions that allocated memory. I also fixed some other minor problems, including 32-bit fixes and a reported memory leak which was just due to using the wrong free function. git-svn-id: https://outreach.scidac.gov/svn/spl/trunk@163 7e1ea52c-4ff2-0310-8f11-9dd32ca42a1c
This commit is contained in:
parent
550f170525
commit
a0f6da3d95
|
@ -328,6 +328,25 @@ AC_DEFUN([SPL_AC_TYPE_UINTPTR_T],
|
||||||
])
|
])
|
||||||
])
|
])
|
||||||
|
|
||||||
|
dnl #
|
||||||
|
dnl # 2.6.x API change,
|
||||||
|
dnl # check if atomic64_t typedef is defined
|
||||||
|
dnl #
|
||||||
|
AC_DEFUN([SPL_AC_TYPE_ATOMIC64_T],
|
||||||
|
[AC_MSG_CHECKING([whether kernel defines atomic64_t])
|
||||||
|
SPL_LINUX_TRY_COMPILE([
|
||||||
|
#include <asm/atomic.h>
|
||||||
|
],[
|
||||||
|
atomic64_t *ptr;
|
||||||
|
],[
|
||||||
|
AC_MSG_RESULT([yes])
|
||||||
|
AC_DEFINE(HAVE_ATOMIC64_T, 1,
|
||||||
|
[kernel defines atomic64_t])
|
||||||
|
],[
|
||||||
|
AC_MSG_RESULT([no])
|
||||||
|
])
|
||||||
|
])
|
||||||
|
|
||||||
dnl #
|
dnl #
|
||||||
dnl # 2.6.20 API change,
|
dnl # 2.6.20 API change,
|
||||||
dnl # INIT_WORK use 2 args and not store data inside
|
dnl # INIT_WORK use 2 args and not store data inside
|
||||||
|
|
|
@ -44,6 +44,7 @@ SPL_AC_DEBUG_MUTEX
|
||||||
SPL_AC_DEBUG_KSTAT
|
SPL_AC_DEBUG_KSTAT
|
||||||
SPL_AC_DEBUG_CALLB
|
SPL_AC_DEBUG_CALLB
|
||||||
SPL_AC_TYPE_UINTPTR_T
|
SPL_AC_TYPE_UINTPTR_T
|
||||||
|
SPL_AC_TYPE_ATOMIC64_T
|
||||||
SPL_AC_3ARGS_INIT_WORK
|
SPL_AC_3ARGS_INIT_WORK
|
||||||
SPL_AC_2ARGS_REGISTER_SYSCTL
|
SPL_AC_2ARGS_REGISTER_SYSCTL
|
||||||
SPL_AC_SET_SHRINKER
|
SPL_AC_SET_SHRINKER
|
||||||
|
|
|
@ -2,6 +2,7 @@
|
||||||
#define _SPL_ATOMIC_COMPAT_H
|
#define _SPL_ATOMIC_COMPAT_H
|
||||||
|
|
||||||
#include <asm/atomic.h>
|
#include <asm/atomic.h>
|
||||||
|
#include "spl_config.h"
|
||||||
|
|
||||||
#ifndef HAVE_ATOMIC64_T
|
#ifndef HAVE_ATOMIC64_T
|
||||||
#include <linux/spinlock.h>
|
#include <linux/spinlock.h>
|
||||||
|
|
|
@ -201,6 +201,7 @@ struct page_collection {
|
||||||
#ifdef NDEBUG
|
#ifdef NDEBUG
|
||||||
|
|
||||||
#define CDEBUG_STACK() (0)
|
#define CDEBUG_STACK() (0)
|
||||||
|
#define CDEBUG_LIMIT(x, y, z, a...) ((void)0)
|
||||||
#define __CDEBUG_LIMIT(x, y, z, a...) ((void)0)
|
#define __CDEBUG_LIMIT(x, y, z, a...) ((void)0)
|
||||||
#define CDEBUG(mask, format, a...) ((void)0)
|
#define CDEBUG(mask, format, a...) ((void)0)
|
||||||
#define CWARN(fmt, a...) ((void)0)
|
#define CWARN(fmt, a...) ((void)0)
|
||||||
|
|
|
@ -64,337 +64,70 @@ extern "C" {
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef DEBUG_KMEM
|
#ifdef DEBUG_KMEM
|
||||||
|
|
||||||
extern atomic64_t kmem_alloc_used;
|
extern atomic64_t kmem_alloc_used;
|
||||||
extern unsigned long kmem_alloc_max;
|
extern unsigned long long kmem_alloc_max;
|
||||||
extern atomic64_t vmem_alloc_used;
|
extern atomic64_t vmem_alloc_used;
|
||||||
extern unsigned long vmem_alloc_max;
|
extern unsigned long long vmem_alloc_max;
|
||||||
extern int kmem_warning_flag;
|
|
||||||
|
# define kmem_alloc(size, flags) __kmem_alloc((size), (flags), 0, 0)
|
||||||
|
# define kmem_zalloc(size, flags) __kmem_alloc((size), ((flags) | \
|
||||||
|
__GFP_ZERO), 0, 0)
|
||||||
|
|
||||||
|
/* The node alloc functions are only used by the SPL code itself */
|
||||||
|
# ifdef HAVE_KMALLOC_NODE
|
||||||
|
# define kmem_alloc_node(size, flags, node) __kmem_alloc((size), (flags), 1, \
|
||||||
|
node)
|
||||||
|
# else
|
||||||
|
# define kmem_alloc_node(size, flags, node) __kmem_alloc((size), (flags), 0, 0)
|
||||||
|
# endif
|
||||||
|
|
||||||
|
# define vmem_zalloc(size, flags) vmem_alloc((size), ((flags) | \
|
||||||
|
__GFP_ZERO))
|
||||||
|
|
||||||
# ifdef DEBUG_KMEM_TRACKING
|
# ifdef DEBUG_KMEM_TRACKING
|
||||||
/* XXX - Not to surprisingly with debugging enabled the xmem_locks are very
|
|
||||||
* highly contended particularly on xfree(). If we want to run with this
|
|
||||||
* detailed debugging enabled for anything other than debugging we need to
|
|
||||||
* minimize the contention by moving to a lock per xmem_table entry model.
|
|
||||||
*/
|
|
||||||
#define KMEM_HASH_BITS 10
|
|
||||||
#define KMEM_TABLE_SIZE (1 << KMEM_HASH_BITS)
|
|
||||||
|
|
||||||
extern struct hlist_head kmem_table[KMEM_TABLE_SIZE];
|
extern void *kmem_alloc_track(size_t size, int flags, const char *func,
|
||||||
extern struct list_head kmem_list;
|
int line, int node_alloc, int node);
|
||||||
extern spinlock_t kmem_lock;
|
extern void kmem_free_track(void *ptr, size_t size);
|
||||||
|
extern void *vmem_alloc_track(size_t size, int flags, const char *func,
|
||||||
|
int line);
|
||||||
|
extern void vmem_free_track(void *ptr, size_t size);
|
||||||
|
|
||||||
#define VMEM_HASH_BITS 10
|
# define __kmem_alloc(size, flags, na, node) kmem_alloc_track((size), \
|
||||||
#define VMEM_TABLE_SIZE (1 << VMEM_HASH_BITS)
|
(flags), __FUNCTION__, \
|
||||||
|
__LINE__, (na), (node))
|
||||||
extern struct hlist_head vmem_table[VMEM_TABLE_SIZE];
|
# define kmem_free(ptr, size) kmem_free_track((ptr), (size))
|
||||||
extern struct list_head vmem_list;
|
# define vmem_alloc(size, flags) vmem_alloc_track((size), \
|
||||||
extern spinlock_t vmem_lock;
|
(flags),__FUNCTION__, \
|
||||||
|
__LINE__)
|
||||||
typedef struct kmem_debug {
|
# define vmem_free(ptr, size) vmem_free_track((ptr), (size))
|
||||||
struct hlist_node kd_hlist; /* Hash node linkage */
|
|
||||||
struct list_head kd_list; /* List of all allocations */
|
|
||||||
void *kd_addr; /* Allocation pointer */
|
|
||||||
size_t kd_size; /* Allocation size */
|
|
||||||
const char *kd_func; /* Allocation function */
|
|
||||||
int kd_line; /* Allocation line */
|
|
||||||
} kmem_debug_t;
|
|
||||||
|
|
||||||
static __inline__ kmem_debug_t *
|
|
||||||
__kmem_del_init(spinlock_t *lock,struct hlist_head *table,int bits,void *addr)
|
|
||||||
{
|
|
||||||
struct hlist_head *head;
|
|
||||||
struct hlist_node *node;
|
|
||||||
struct kmem_debug *p;
|
|
||||||
unsigned long flags;
|
|
||||||
|
|
||||||
spin_lock_irqsave(lock, flags);
|
|
||||||
head = &table[hash_ptr(addr, bits)];
|
|
||||||
hlist_for_each_entry_rcu(p, node, head, kd_hlist) {
|
|
||||||
if (p->kd_addr == addr) {
|
|
||||||
hlist_del_init(&p->kd_hlist);
|
|
||||||
list_del_init(&p->kd_list);
|
|
||||||
spin_unlock_irqrestore(lock, flags);
|
|
||||||
return p;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
spin_unlock_irqrestore(lock, flags);
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
|
|
||||||
#define __kmem_alloc(size, flags, allocator, args...) \
|
|
||||||
({ void *_ptr_ = NULL; \
|
|
||||||
kmem_debug_t *_dptr_; \
|
|
||||||
unsigned long _flags_; \
|
|
||||||
\
|
|
||||||
_dptr_ = (kmem_debug_t *)kmalloc(sizeof(kmem_debug_t), (flags)); \
|
|
||||||
if (_dptr_ == NULL) { \
|
|
||||||
__CDEBUG_LIMIT(S_KMEM, D_WARNING, "Warning " \
|
|
||||||
"kmem_alloc(%d, 0x%x) debug failed\n", \
|
|
||||||
sizeof(kmem_debug_t), (int)(flags)); \
|
|
||||||
} else { \
|
|
||||||
/* Marked unlikely because we should never be doing this, */ \
|
|
||||||
/* we tolerate to up 2 pages but a single page is best. */ \
|
|
||||||
if (unlikely((size) > (PAGE_SIZE * 2)) && kmem_warning_flag) \
|
|
||||||
__CDEBUG_LIMIT(S_KMEM, D_WARNING, "Warning large " \
|
|
||||||
"kmem_alloc(%d, 0x%x) (%ld/%ld)\n", \
|
|
||||||
(int)(size), (int)(flags), \
|
|
||||||
atomic64_read(&kmem_alloc_used), \
|
|
||||||
kmem_alloc_max); \
|
|
||||||
\
|
|
||||||
_ptr_ = (void *)allocator((size), (flags), ## args); \
|
|
||||||
if (_ptr_ == NULL) { \
|
|
||||||
kfree(_dptr_); \
|
|
||||||
__CDEBUG_LIMIT(S_KMEM, D_WARNING, "Warning " \
|
|
||||||
"kmem_alloc(%d, 0x%x) failed (%ld/" \
|
|
||||||
"%ld)\n", (int)(size), (int)(flags), \
|
|
||||||
atomic64_read(&kmem_alloc_used), \
|
|
||||||
kmem_alloc_max); \
|
|
||||||
} else { \
|
|
||||||
atomic64_add((size), &kmem_alloc_used); \
|
|
||||||
if (unlikely(atomic64_read(&kmem_alloc_used) > \
|
|
||||||
kmem_alloc_max)) \
|
|
||||||
kmem_alloc_max = \
|
|
||||||
atomic64_read(&kmem_alloc_used); \
|
|
||||||
\
|
|
||||||
INIT_HLIST_NODE(&_dptr_->kd_hlist); \
|
|
||||||
INIT_LIST_HEAD(&_dptr_->kd_list); \
|
|
||||||
_dptr_->kd_addr = _ptr_; \
|
|
||||||
_dptr_->kd_size = (size); \
|
|
||||||
_dptr_->kd_func = __FUNCTION__; \
|
|
||||||
_dptr_->kd_line = __LINE__; \
|
|
||||||
spin_lock_irqsave(&kmem_lock, _flags_); \
|
|
||||||
hlist_add_head_rcu(&_dptr_->kd_hlist, \
|
|
||||||
&kmem_table[hash_ptr(_ptr_, KMEM_HASH_BITS)]);\
|
|
||||||
list_add_tail(&_dptr_->kd_list, &kmem_list); \
|
|
||||||
spin_unlock_irqrestore(&kmem_lock, _flags_); \
|
|
||||||
\
|
|
||||||
__CDEBUG_LIMIT(S_KMEM, D_INFO, "kmem_alloc(" \
|
|
||||||
"%d, 0x%x) = %p (%ld/%ld)\n", \
|
|
||||||
(int)(size), (int)(flags), _ptr_, \
|
|
||||||
atomic64_read(&kmem_alloc_used), \
|
|
||||||
kmem_alloc_max); \
|
|
||||||
} \
|
|
||||||
} \
|
|
||||||
\
|
|
||||||
_ptr_; \
|
|
||||||
})
|
|
||||||
|
|
||||||
#define kmem_free(ptr, size) \
|
|
||||||
({ \
|
|
||||||
kmem_debug_t *_dptr_; \
|
|
||||||
ASSERT((ptr) || (size > 0)); \
|
|
||||||
\
|
|
||||||
_dptr_ = __kmem_del_init(&kmem_lock, kmem_table, KMEM_HASH_BITS, ptr);\
|
|
||||||
ASSERT(_dptr_); /* Must exist in hash due to kmem_alloc() */ \
|
|
||||||
ASSERTF(_dptr_->kd_size == (size), "kd_size (%d) != size (%d), " \
|
|
||||||
"kd_func = %s, kd_line = %d\n", _dptr_->kd_size, (size), \
|
|
||||||
_dptr_->kd_func, _dptr_->kd_line); /* Size must match */ \
|
|
||||||
atomic64_sub((size), &kmem_alloc_used); \
|
|
||||||
__CDEBUG_LIMIT(S_KMEM, D_INFO, "kmem_free(%p, %d) (%ld/%ld)\n", \
|
|
||||||
(ptr), (int)(size), atomic64_read(&kmem_alloc_used), \
|
|
||||||
kmem_alloc_max); \
|
|
||||||
\
|
|
||||||
memset(_dptr_, 0x5a, sizeof(kmem_debug_t)); \
|
|
||||||
kfree(_dptr_); \
|
|
||||||
\
|
|
||||||
memset(ptr, 0x5a, (size)); \
|
|
||||||
kfree(ptr); \
|
|
||||||
})
|
|
||||||
|
|
||||||
#define __vmem_alloc(size, flags) \
|
|
||||||
({ void *_ptr_ = NULL; \
|
|
||||||
kmem_debug_t *_dptr_; \
|
|
||||||
unsigned long _flags_; \
|
|
||||||
\
|
|
||||||
ASSERT((flags) & KM_SLEEP); \
|
|
||||||
\
|
|
||||||
_dptr_ = (kmem_debug_t *)kmalloc(sizeof(kmem_debug_t), (flags)); \
|
|
||||||
if (_dptr_ == NULL) { \
|
|
||||||
__CDEBUG_LIMIT(S_KMEM, D_WARNING, "Warning " \
|
|
||||||
"vmem_alloc(%d, 0x%x) debug failed\n", \
|
|
||||||
sizeof(kmem_debug_t), (int)(flags)); \
|
|
||||||
} else { \
|
|
||||||
_ptr_ = (void *)__vmalloc((size), (((flags) | \
|
|
||||||
__GFP_HIGHMEM) & ~__GFP_ZERO), \
|
|
||||||
PAGE_KERNEL); \
|
|
||||||
if (_ptr_ == NULL) { \
|
|
||||||
kfree(_dptr_); \
|
|
||||||
__CDEBUG_LIMIT(S_KMEM, D_WARNING, "Warning " \
|
|
||||||
"vmem_alloc(%d, 0x%x) failed (%ld/" \
|
|
||||||
"%ld)\n", (int)(size), (int)(flags), \
|
|
||||||
atomic64_read(&vmem_alloc_used), \
|
|
||||||
vmem_alloc_max); \
|
|
||||||
} else { \
|
|
||||||
if (flags & __GFP_ZERO) \
|
|
||||||
memset(_ptr_, 0, (size)); \
|
|
||||||
\
|
|
||||||
atomic64_add((size), &vmem_alloc_used); \
|
|
||||||
if (unlikely(atomic64_read(&vmem_alloc_used) > \
|
|
||||||
vmem_alloc_max)) \
|
|
||||||
vmem_alloc_max = \
|
|
||||||
atomic64_read(&vmem_alloc_used); \
|
|
||||||
\
|
|
||||||
INIT_HLIST_NODE(&_dptr_->kd_hlist); \
|
|
||||||
INIT_LIST_HEAD(&_dptr_->kd_list); \
|
|
||||||
_dptr_->kd_addr = _ptr_; \
|
|
||||||
_dptr_->kd_size = (size); \
|
|
||||||
_dptr_->kd_func = __FUNCTION__; \
|
|
||||||
_dptr_->kd_line = __LINE__; \
|
|
||||||
spin_lock_irqsave(&vmem_lock, _flags_); \
|
|
||||||
hlist_add_head_rcu(&_dptr_->kd_hlist, \
|
|
||||||
&vmem_table[hash_ptr(_ptr_, VMEM_HASH_BITS)]);\
|
|
||||||
list_add_tail(&_dptr_->kd_list, &vmem_list); \
|
|
||||||
spin_unlock_irqrestore(&vmem_lock, _flags_); \
|
|
||||||
\
|
|
||||||
__CDEBUG_LIMIT(S_KMEM, D_INFO, "vmem_alloc(" \
|
|
||||||
"%d, 0x%x) = %p (%ld/%ld)\n", \
|
|
||||||
(int)(size), (int)(flags), _ptr_, \
|
|
||||||
atomic64_read(&vmem_alloc_used), \
|
|
||||||
vmem_alloc_max); \
|
|
||||||
} \
|
|
||||||
} \
|
|
||||||
\
|
|
||||||
_ptr_; \
|
|
||||||
})
|
|
||||||
|
|
||||||
#define vmem_free(ptr, size) \
|
|
||||||
({ \
|
|
||||||
kmem_debug_t *_dptr_; \
|
|
||||||
ASSERT((ptr) || (size > 0)); \
|
|
||||||
\
|
|
||||||
_dptr_ = __kmem_del_init(&vmem_lock, vmem_table, VMEM_HASH_BITS, ptr);\
|
|
||||||
ASSERT(_dptr_); /* Must exist in hash due to vmem_alloc() */ \
|
|
||||||
ASSERTF(_dptr_->kd_size == (size), "kd_size (%d) != size (%d), " \
|
|
||||||
"kd_func = %s, kd_line = %d\n", _dptr_->kd_size, (size), \
|
|
||||||
_dptr_->kd_func, _dptr_->kd_line); /* Size must match */ \
|
|
||||||
atomic64_sub((size), &vmem_alloc_used); \
|
|
||||||
__CDEBUG_LIMIT(S_KMEM, D_INFO, "vmem_free(%p, %d) (%ld/%ld)\n", \
|
|
||||||
(ptr), (int)(size), atomic64_read(&vmem_alloc_used), \
|
|
||||||
vmem_alloc_max); \
|
|
||||||
\
|
|
||||||
memset(_dptr_, 0x5a, sizeof(kmem_debug_t)); \
|
|
||||||
kfree(_dptr_); \
|
|
||||||
\
|
|
||||||
memset(ptr, 0x5a, (size)); \
|
|
||||||
vfree(ptr); \
|
|
||||||
})
|
|
||||||
|
|
||||||
# else /* DEBUG_KMEM_TRACKING */
|
# else /* DEBUG_KMEM_TRACKING */
|
||||||
|
|
||||||
#define __kmem_alloc(size, flags, allocator, args...) \
|
extern void *kmem_alloc_debug(size_t size, int flags, const char *func,
|
||||||
({ void *_ptr_ = NULL; \
|
int line, int node_alloc, int node);
|
||||||
\
|
extern void kmem_free_debug(void *ptr, size_t size);
|
||||||
/* Marked unlikely because we should never be doing this, */ \
|
extern void *vmem_alloc_debug(size_t size, int flags, const char *func,
|
||||||
/* we tolerate to up 2 pages but a single page is best. */ \
|
int line);
|
||||||
if (unlikely((size) > (PAGE_SIZE * 2)) && kmem_warning_flag) \
|
extern void vmem_free_debug(void *ptr, size_t size);
|
||||||
__CDEBUG_LIMIT(S_KMEM, D_WARNING, "Warning large " \
|
|
||||||
"kmem_alloc(%d, 0x%x) (%ld/%ld)\n", \
|
|
||||||
(int)(size), (int)(flags), \
|
|
||||||
atomic64_read(&kmem_alloc_used), \
|
|
||||||
kmem_alloc_max); \
|
|
||||||
\
|
|
||||||
_ptr_ = (void *)allocator((size), (flags), ## args); \
|
|
||||||
if (_ptr_ == NULL) { \
|
|
||||||
__CDEBUG_LIMIT(S_KMEM, D_WARNING, "Warning " \
|
|
||||||
"kmem_alloc(%d, 0x%x) failed (%ld/" \
|
|
||||||
"%ld)\n", (int)(size), (int)(flags), \
|
|
||||||
atomic64_read(&kmem_alloc_used), \
|
|
||||||
kmem_alloc_max); \
|
|
||||||
} else { \
|
|
||||||
atomic64_add((size), &kmem_alloc_used); \
|
|
||||||
if (unlikely(atomic64_read(&kmem_alloc_used) > \
|
|
||||||
kmem_alloc_max)) \
|
|
||||||
kmem_alloc_max = \
|
|
||||||
atomic64_read(&kmem_alloc_used); \
|
|
||||||
\
|
|
||||||
__CDEBUG_LIMIT(S_KMEM, D_INFO, "kmem_alloc(%d, 0x%x) = %p " \
|
|
||||||
"(%ld/%ld)\n", (int)(size), (int)(flags), \
|
|
||||||
_ptr_, atomic64_read(&kmem_alloc_used), \
|
|
||||||
kmem_alloc_max); \
|
|
||||||
} \
|
|
||||||
\
|
|
||||||
_ptr_; \
|
|
||||||
})
|
|
||||||
|
|
||||||
#define kmem_free(ptr, size) \
|
# define __kmem_alloc(size, flags, na, node) kmem_alloc_debug((size), \
|
||||||
({ \
|
(flags), __FUNCTION__, \
|
||||||
ASSERT((ptr) || (size > 0)); \
|
__LINE__, (na), (node))
|
||||||
\
|
# define kmem_free(ptr, size) kmem_free_debug((ptr), (size))
|
||||||
atomic64_sub((size), &kmem_alloc_used); \
|
# define vmem_alloc(size, flags) vmem_alloc_debug((size), \
|
||||||
__CDEBUG_LIMIT(S_KMEM, D_INFO, "kmem_free(%p, %d) (%ld/%ld)\n", \
|
(flags), __FUNCTION__, \
|
||||||
(ptr), (int)(size), atomic64_read(&kmem_alloc_used), \
|
__LINE__)
|
||||||
kmem_alloc_max); \
|
# define vmem_free(ptr, size) vmem_free_debug((ptr), (size))
|
||||||
memset(ptr, 0x5a, (size)); \
|
|
||||||
kfree(ptr); \
|
|
||||||
})
|
|
||||||
|
|
||||||
#define __vmem_alloc(size, flags) \
|
|
||||||
({ void *_ptr_ = NULL; \
|
|
||||||
\
|
|
||||||
ASSERT((flags) & KM_SLEEP); \
|
|
||||||
\
|
|
||||||
_ptr_ = (void *)__vmalloc((size), (((flags) | \
|
|
||||||
__GFP_HIGHMEM) & ~__GFP_ZERO), PAGE_KERNEL);\
|
|
||||||
if (_ptr_ == NULL) { \
|
|
||||||
__CDEBUG_LIMIT(S_KMEM, D_WARNING, "Warning " \
|
|
||||||
"vmem_alloc(%d, 0x%x) failed (%ld/" \
|
|
||||||
"%ld)\n", (int)(size), (int)(flags), \
|
|
||||||
atomic64_read(&vmem_alloc_used), \
|
|
||||||
vmem_alloc_max); \
|
|
||||||
} else { \
|
|
||||||
if (flags & __GFP_ZERO) \
|
|
||||||
memset(_ptr_, 0, (size)); \
|
|
||||||
\
|
|
||||||
atomic64_add((size), &vmem_alloc_used); \
|
|
||||||
if (unlikely(atomic64_read(&vmem_alloc_used) > \
|
|
||||||
vmem_alloc_max)) \
|
|
||||||
vmem_alloc_max = \
|
|
||||||
atomic64_read(&vmem_alloc_used); \
|
|
||||||
\
|
|
||||||
__CDEBUG_LIMIT(S_KMEM, D_INFO, "vmem_alloc(" \
|
|
||||||
"%d, 0x%x) = %p (%ld/%ld)\n", \
|
|
||||||
(int)(size), (int)(flags), _ptr_, \
|
|
||||||
atomic64_read(&vmem_alloc_used), \
|
|
||||||
vmem_alloc_max); \
|
|
||||||
} \
|
|
||||||
\
|
|
||||||
_ptr_; \
|
|
||||||
})
|
|
||||||
|
|
||||||
#define vmem_free(ptr, size) \
|
|
||||||
({ \
|
|
||||||
ASSERT((ptr) || (size > 0)); \
|
|
||||||
\
|
|
||||||
atomic64_sub((size), &vmem_alloc_used); \
|
|
||||||
__CDEBUG_LIMIT(S_KMEM, D_INFO, "vmem_free(%p, %d) (%ld/%ld)\n", \
|
|
||||||
(ptr), (int)(size), atomic64_read(&vmem_alloc_used), \
|
|
||||||
vmem_alloc_max); \
|
|
||||||
memset(ptr, 0x5a, (size)); \
|
|
||||||
vfree(ptr); \
|
|
||||||
})
|
|
||||||
|
|
||||||
# endif /* DEBUG_KMEM_TRACKING */
|
# endif /* DEBUG_KMEM_TRACKING */
|
||||||
|
|
||||||
#define kmem_alloc(size, flags) __kmem_alloc((size), (flags), kmalloc)
|
|
||||||
#define kmem_zalloc(size, flags) __kmem_alloc((size), (flags), kzalloc)
|
|
||||||
|
|
||||||
#ifdef HAVE_KMALLOC_NODE
|
|
||||||
#define kmem_alloc_node(size, flags, node) \
|
|
||||||
__kmem_alloc((size), (flags), kmalloc_node, node)
|
|
||||||
#else
|
|
||||||
#define kmem_alloc_node(size, flags, node) \
|
|
||||||
__kmem_alloc((size), (flags), kmalloc)
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#define vmem_alloc(size, flags) __vmem_alloc((size), (flags))
|
|
||||||
#define vmem_zalloc(size, flags) __vmem_alloc((size), ((flags) | __GFP_ZERO))
|
|
||||||
|
|
||||||
#else /* DEBUG_KMEM */
|
#else /* DEBUG_KMEM */
|
||||||
|
|
||||||
# define kmem_alloc(size, flags) kmalloc((size), (flags))
|
# define kmem_alloc(size, flags) kmalloc((size), (flags))
|
||||||
# define kmem_zalloc(size, flags) kzalloc((size), (flags))
|
# define kmem_zalloc(size, flags) kzalloc((size), (flags))
|
||||||
#define kmem_free(ptr, size) kfree(ptr)
|
# define kmem_free(ptr, size) (kfree(ptr), (void)(size))
|
||||||
|
|
||||||
# ifdef HAVE_KMALLOC_NODE
|
# ifdef HAVE_KMALLOC_NODE
|
||||||
# define kmem_alloc_node(size, flags, node) \
|
# define kmem_alloc_node(size, flags, node) \
|
||||||
|
@ -413,7 +146,7 @@ __kmem_del_init(spinlock_t *lock,struct hlist_head *table,int bits,void *addr)
|
||||||
memset(_ptr_, 0, (size)); \
|
memset(_ptr_, 0, (size)); \
|
||||||
_ptr_; \
|
_ptr_; \
|
||||||
})
|
})
|
||||||
#define vmem_free(ptr, size) vfree(ptr)
|
# define vmem_free(ptr, size) (vfree(ptr), (void)(size))
|
||||||
|
|
||||||
#endif /* DEBUG_KMEM */
|
#endif /* DEBUG_KMEM */
|
||||||
|
|
||||||
|
|
|
@ -44,9 +44,9 @@
|
||||||
#ifdef DEBUG_KMEM
|
#ifdef DEBUG_KMEM
|
||||||
/* Shim layer memory accounting */
|
/* Shim layer memory accounting */
|
||||||
atomic64_t kmem_alloc_used = ATOMIC64_INIT(0);
|
atomic64_t kmem_alloc_used = ATOMIC64_INIT(0);
|
||||||
unsigned long kmem_alloc_max = 0;
|
unsigned long long kmem_alloc_max = 0;
|
||||||
atomic64_t vmem_alloc_used = ATOMIC64_INIT(0);
|
atomic64_t vmem_alloc_used = ATOMIC64_INIT(0);
|
||||||
unsigned long vmem_alloc_max = 0;
|
unsigned long long vmem_alloc_max = 0;
|
||||||
int kmem_warning_flag = 1;
|
int kmem_warning_flag = 1;
|
||||||
|
|
||||||
EXPORT_SYMBOL(kmem_alloc_used);
|
EXPORT_SYMBOL(kmem_alloc_used);
|
||||||
|
@ -56,6 +56,28 @@ EXPORT_SYMBOL(vmem_alloc_max);
|
||||||
EXPORT_SYMBOL(kmem_warning_flag);
|
EXPORT_SYMBOL(kmem_warning_flag);
|
||||||
|
|
||||||
# ifdef DEBUG_KMEM_TRACKING
|
# ifdef DEBUG_KMEM_TRACKING
|
||||||
|
|
||||||
|
/* XXX - Not to surprisingly with debugging enabled the xmem_locks are very
|
||||||
|
* highly contended particularly on xfree(). If we want to run with this
|
||||||
|
* detailed debugging enabled for anything other than debugging we need to
|
||||||
|
* minimize the contention by moving to a lock per xmem_table entry model.
|
||||||
|
*/
|
||||||
|
|
||||||
|
# define KMEM_HASH_BITS 10
|
||||||
|
# define KMEM_TABLE_SIZE (1 << KMEM_HASH_BITS)
|
||||||
|
|
||||||
|
# define VMEM_HASH_BITS 10
|
||||||
|
# define VMEM_TABLE_SIZE (1 << VMEM_HASH_BITS)
|
||||||
|
|
||||||
|
typedef struct kmem_debug {
|
||||||
|
struct hlist_node kd_hlist; /* Hash node linkage */
|
||||||
|
struct list_head kd_list; /* List of all allocations */
|
||||||
|
void *kd_addr; /* Allocation pointer */
|
||||||
|
size_t kd_size; /* Allocation size */
|
||||||
|
const char *kd_func; /* Allocation function */
|
||||||
|
int kd_line; /* Allocation line */
|
||||||
|
} kmem_debug_t;
|
||||||
|
|
||||||
spinlock_t kmem_lock;
|
spinlock_t kmem_lock;
|
||||||
struct hlist_head kmem_table[KMEM_TABLE_SIZE];
|
struct hlist_head kmem_table[KMEM_TABLE_SIZE];
|
||||||
struct list_head kmem_list;
|
struct list_head kmem_list;
|
||||||
|
@ -90,10 +112,10 @@ EXPORT_SYMBOL(kmem_set_warning);
|
||||||
* kernel have removed support for destructors. This is a deal
|
* kernel have removed support for destructors. This is a deal
|
||||||
* breaker for the SPL which contains particularly expensive
|
* breaker for the SPL which contains particularly expensive
|
||||||
* initializers for mutex's, condition variables, etc. We also
|
* initializers for mutex's, condition variables, etc. We also
|
||||||
* require a minimal level of cleaner for these data types unlike
|
* require a minimal level of cleanup for these data types unlike
|
||||||
* may Linux data type which do need to be explicitly destroyed.
|
* many Linux data type which do need to be explicitly destroyed.
|
||||||
*
|
*
|
||||||
* 2) Virtual address backed slab. Callers of the Solaris slab
|
* 2) Virtual address space backed slab. Callers of the Solaris slab
|
||||||
* expect it to work well for both small are very large allocations.
|
* expect it to work well for both small are very large allocations.
|
||||||
* Because of memory fragmentation the Linux slab which is backed
|
* Because of memory fragmentation the Linux slab which is backed
|
||||||
* by kmalloc'ed memory performs very badly when confronted with
|
* by kmalloc'ed memory performs very badly when confronted with
|
||||||
|
@ -147,6 +169,352 @@ static struct shrinker spl_kmem_cache_shrinker = {
|
||||||
};
|
};
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
#ifdef DEBUG_KMEM
|
||||||
|
# ifdef DEBUG_KMEM_TRACKING
|
||||||
|
|
||||||
|
static kmem_debug_t *
|
||||||
|
kmem_del_init(spinlock_t *lock, struct hlist_head *table, int bits,
|
||||||
|
void *addr)
|
||||||
|
{
|
||||||
|
struct hlist_head *head;
|
||||||
|
struct hlist_node *node;
|
||||||
|
struct kmem_debug *p;
|
||||||
|
unsigned long flags;
|
||||||
|
ENTRY;
|
||||||
|
|
||||||
|
spin_lock_irqsave(lock, flags);
|
||||||
|
|
||||||
|
head = &table[hash_ptr(addr, bits)];
|
||||||
|
hlist_for_each_entry_rcu(p, node, head, kd_hlist) {
|
||||||
|
if (p->kd_addr == addr) {
|
||||||
|
hlist_del_init(&p->kd_hlist);
|
||||||
|
list_del_init(&p->kd_list);
|
||||||
|
spin_unlock_irqrestore(lock, flags);
|
||||||
|
return p;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
spin_unlock_irqrestore(lock, flags);
|
||||||
|
|
||||||
|
RETURN(NULL);
|
||||||
|
}
|
||||||
|
|
||||||
|
void *
|
||||||
|
kmem_alloc_track(size_t size, int flags, const char *func, int line,
|
||||||
|
int node_alloc, int node)
|
||||||
|
{
|
||||||
|
void *ptr = NULL;
|
||||||
|
kmem_debug_t *dptr;
|
||||||
|
unsigned long irq_flags;
|
||||||
|
ENTRY;
|
||||||
|
|
||||||
|
dptr = (kmem_debug_t *) kmalloc(sizeof(kmem_debug_t),
|
||||||
|
flags & ~__GFP_ZERO);
|
||||||
|
|
||||||
|
if (dptr == NULL) {
|
||||||
|
CWARN("kmem_alloc(%ld, 0x%x) debug failed\n",
|
||||||
|
sizeof(kmem_debug_t), flags);
|
||||||
|
} else {
|
||||||
|
/* Marked unlikely because we should never be doing this,
|
||||||
|
* we tolerate to up 2 pages but a single page is best. */
|
||||||
|
if (unlikely((size) > (PAGE_SIZE * 2)) && kmem_warning_flag)
|
||||||
|
CWARN("Large kmem_alloc(%llu, 0x%x) (%lld/%llu)\n",
|
||||||
|
(unsigned long long) size, flags,
|
||||||
|
atomic64_read(&kmem_alloc_used), kmem_alloc_max);
|
||||||
|
|
||||||
|
/* Use the correct allocator */
|
||||||
|
if (node_alloc) {
|
||||||
|
ASSERT(!(flags & __GFP_ZERO));
|
||||||
|
ptr = kmalloc_node(size, flags, node);
|
||||||
|
} else if (flags & __GFP_ZERO) {
|
||||||
|
ptr = kzalloc(size, flags & ~__GFP_ZERO);
|
||||||
|
} else {
|
||||||
|
ptr = kmalloc(size, flags);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (unlikely(ptr == NULL)) {
|
||||||
|
kfree(dptr);
|
||||||
|
CWARN("kmem_alloc(%llu, 0x%x) failed (%lld/%llu)\n",
|
||||||
|
(unsigned long long) size, flags,
|
||||||
|
atomic64_read(&kmem_alloc_used), kmem_alloc_max);
|
||||||
|
goto out;
|
||||||
|
}
|
||||||
|
|
||||||
|
atomic64_add(size, &kmem_alloc_used);
|
||||||
|
if (unlikely(atomic64_read(&kmem_alloc_used) >
|
||||||
|
kmem_alloc_max))
|
||||||
|
kmem_alloc_max =
|
||||||
|
atomic64_read(&kmem_alloc_used);
|
||||||
|
|
||||||
|
INIT_HLIST_NODE(&dptr->kd_hlist);
|
||||||
|
INIT_LIST_HEAD(&dptr->kd_list);
|
||||||
|
|
||||||
|
dptr->kd_addr = ptr;
|
||||||
|
dptr->kd_size = size;
|
||||||
|
dptr->kd_func = func;
|
||||||
|
dptr->kd_line = line;
|
||||||
|
|
||||||
|
spin_lock_irqsave(&kmem_lock, irq_flags);
|
||||||
|
hlist_add_head_rcu(&dptr->kd_hlist,
|
||||||
|
&kmem_table[hash_ptr(ptr, KMEM_HASH_BITS)]);
|
||||||
|
list_add_tail(&dptr->kd_list, &kmem_list);
|
||||||
|
spin_unlock_irqrestore(&kmem_lock, irq_flags);
|
||||||
|
|
||||||
|
CDEBUG_LIMIT(D_INFO, "kmem_alloc(%llu, 0x%x) = %p "
|
||||||
|
"(%lld/%llu)\n", (unsigned long long) size, flags,
|
||||||
|
ptr, atomic64_read(&kmem_alloc_used),
|
||||||
|
kmem_alloc_max);
|
||||||
|
}
|
||||||
|
out:
|
||||||
|
RETURN(ptr);
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(kmem_alloc_track);
|
||||||
|
|
||||||
|
void
|
||||||
|
kmem_free_track(void *ptr, size_t size)
|
||||||
|
{
|
||||||
|
kmem_debug_t *dptr;
|
||||||
|
ENTRY;
|
||||||
|
|
||||||
|
ASSERTF(ptr || size > 0, "ptr: %p, size: %llu", ptr,
|
||||||
|
(unsigned long long) size);
|
||||||
|
|
||||||
|
dptr = kmem_del_init(&kmem_lock, kmem_table, KMEM_HASH_BITS, ptr);
|
||||||
|
|
||||||
|
ASSERT(dptr); /* Must exist in hash due to kmem_alloc() */
|
||||||
|
|
||||||
|
/* Size must match */
|
||||||
|
ASSERTF(dptr->kd_size == size, "kd_size (%llu) != size (%llu), "
|
||||||
|
"kd_func = %s, kd_line = %d\n", (unsigned long long) dptr->kd_size,
|
||||||
|
(unsigned long long) size, dptr->kd_func, dptr->kd_line);
|
||||||
|
|
||||||
|
atomic64_sub(size, &kmem_alloc_used);
|
||||||
|
|
||||||
|
CDEBUG_LIMIT(D_INFO, "kmem_free(%p, %llu) (%lld/%llu)\n", ptr,
|
||||||
|
(unsigned long long) size, atomic64_read(&kmem_alloc_used),
|
||||||
|
kmem_alloc_max);
|
||||||
|
|
||||||
|
memset(dptr, 0x5a, sizeof(kmem_debug_t));
|
||||||
|
kfree(dptr);
|
||||||
|
|
||||||
|
memset(ptr, 0x5a, size);
|
||||||
|
kfree(ptr);
|
||||||
|
|
||||||
|
EXIT;
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(kmem_free_track);
|
||||||
|
|
||||||
|
void *
|
||||||
|
vmem_alloc_track(size_t size, int flags, const char *func, int line)
|
||||||
|
{
|
||||||
|
void *ptr = NULL;
|
||||||
|
kmem_debug_t *dptr;
|
||||||
|
unsigned long irq_flags;
|
||||||
|
ENTRY;
|
||||||
|
|
||||||
|
ASSERT(flags & KM_SLEEP);
|
||||||
|
|
||||||
|
dptr = (kmem_debug_t *) kmalloc(sizeof(kmem_debug_t), flags);
|
||||||
|
if (dptr == NULL) {
|
||||||
|
CWARN("vmem_alloc(%ld, 0x%x) debug failed\n",
|
||||||
|
sizeof(kmem_debug_t), flags);
|
||||||
|
} else {
|
||||||
|
ptr = __vmalloc(size, (flags | __GFP_HIGHMEM) & ~__GFP_ZERO,
|
||||||
|
PAGE_KERNEL);
|
||||||
|
|
||||||
|
if (unlikely(ptr == NULL)) {
|
||||||
|
kfree(dptr);
|
||||||
|
CWARN("vmem_alloc(%llu, 0x%x) failed (%lld/%llu)\n",
|
||||||
|
(unsigned long long) size, flags,
|
||||||
|
atomic64_read(&vmem_alloc_used), vmem_alloc_max);
|
||||||
|
goto out;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (flags & __GFP_ZERO)
|
||||||
|
memset(ptr, 0, size);
|
||||||
|
|
||||||
|
atomic64_add(size, &vmem_alloc_used);
|
||||||
|
if (unlikely(atomic64_read(&vmem_alloc_used) >
|
||||||
|
vmem_alloc_max))
|
||||||
|
vmem_alloc_max =
|
||||||
|
atomic64_read(&vmem_alloc_used);
|
||||||
|
|
||||||
|
INIT_HLIST_NODE(&dptr->kd_hlist);
|
||||||
|
INIT_LIST_HEAD(&dptr->kd_list);
|
||||||
|
|
||||||
|
dptr->kd_addr = ptr;
|
||||||
|
dptr->kd_size = size;
|
||||||
|
dptr->kd_func = func;
|
||||||
|
dptr->kd_line = line;
|
||||||
|
|
||||||
|
spin_lock_irqsave(&vmem_lock, irq_flags);
|
||||||
|
hlist_add_head_rcu(&dptr->kd_hlist,
|
||||||
|
&vmem_table[hash_ptr(ptr, VMEM_HASH_BITS)]);
|
||||||
|
list_add_tail(&dptr->kd_list, &vmem_list);
|
||||||
|
spin_unlock_irqrestore(&vmem_lock, irq_flags);
|
||||||
|
|
||||||
|
CDEBUG_LIMIT(D_INFO, "vmem_alloc(%llu, 0x%x) = %p "
|
||||||
|
"(%lld/%llu)\n", (unsigned long long) size, flags,
|
||||||
|
ptr, atomic64_read(&vmem_alloc_used),
|
||||||
|
vmem_alloc_max);
|
||||||
|
}
|
||||||
|
out:
|
||||||
|
RETURN(ptr);
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(vmem_alloc_track);
|
||||||
|
|
||||||
|
void
|
||||||
|
vmem_free_track(void *ptr, size_t size)
|
||||||
|
{
|
||||||
|
kmem_debug_t *dptr;
|
||||||
|
ENTRY;
|
||||||
|
|
||||||
|
ASSERTF(ptr || size > 0, "ptr: %p, size: %llu", ptr,
|
||||||
|
(unsigned long long) size);
|
||||||
|
|
||||||
|
dptr = kmem_del_init(&vmem_lock, vmem_table, VMEM_HASH_BITS, ptr);
|
||||||
|
ASSERT(dptr); /* Must exist in hash due to vmem_alloc() */
|
||||||
|
|
||||||
|
/* Size must match */
|
||||||
|
ASSERTF(dptr->kd_size == size, "kd_size (%llu) != size (%llu), "
|
||||||
|
"kd_func = %s, kd_line = %d\n", (unsigned long long) dptr->kd_size,
|
||||||
|
(unsigned long long) size, dptr->kd_func, dptr->kd_line);
|
||||||
|
|
||||||
|
atomic64_sub(size, &vmem_alloc_used);
|
||||||
|
CDEBUG_LIMIT(D_INFO, "vmem_free(%p, %llu) (%lld/%llu)\n", ptr,
|
||||||
|
(unsigned long long) size, atomic64_read(&vmem_alloc_used),
|
||||||
|
vmem_alloc_max);
|
||||||
|
|
||||||
|
memset(dptr, 0x5a, sizeof(kmem_debug_t));
|
||||||
|
kfree(dptr);
|
||||||
|
|
||||||
|
memset(ptr, 0x5a, size);
|
||||||
|
vfree(ptr);
|
||||||
|
|
||||||
|
EXIT;
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(vmem_free_track);
|
||||||
|
|
||||||
|
# else /* DEBUG_KMEM_TRACKING */
|
||||||
|
|
||||||
|
void *
|
||||||
|
kmem_alloc_debug(size_t size, int flags, const char *func, int line,
|
||||||
|
int node_alloc, int node)
|
||||||
|
{
|
||||||
|
void *ptr;
|
||||||
|
ENTRY;
|
||||||
|
|
||||||
|
/* Marked unlikely because we should never be doing this,
|
||||||
|
* we tolerate to up 2 pages but a single page is best. */
|
||||||
|
if (unlikely(size > (PAGE_SIZE * 2)) && kmem_warning_flag)
|
||||||
|
CWARN("Large kmem_alloc(%llu, 0x%x) (%lld/%llu)\n",
|
||||||
|
(unsigned long long) size, flags,
|
||||||
|
atomic64_read(&kmem_alloc_used), kmem_alloc_max);
|
||||||
|
|
||||||
|
/* Use the correct allocator */
|
||||||
|
if (node_alloc) {
|
||||||
|
ASSERT(!(flags & __GFP_ZERO));
|
||||||
|
ptr = kmalloc_node(size, flags, node);
|
||||||
|
} else if (flags & __GFP_ZERO) {
|
||||||
|
ptr = kzalloc(size, flags & (~__GFP_ZERO));
|
||||||
|
} else {
|
||||||
|
ptr = kmalloc(size, flags);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (ptr == NULL) {
|
||||||
|
CWARN("kmem_alloc(%llu, 0x%x) failed (%lld/%llu)\n",
|
||||||
|
(unsigned long long) size, flags,
|
||||||
|
atomic64_read(&kmem_alloc_used), kmem_alloc_max);
|
||||||
|
} else {
|
||||||
|
atomic64_add(size, &kmem_alloc_used);
|
||||||
|
if (unlikely(atomic64_read(&kmem_alloc_used) > kmem_alloc_max))
|
||||||
|
kmem_alloc_max = atomic64_read(&kmem_alloc_used);
|
||||||
|
|
||||||
|
CDEBUG_LIMIT(D_INFO, "kmem_alloc(%llu, 0x%x) = %p "
|
||||||
|
"(%lld/%llu)\n", (unsigned long long) size, flags, ptr,
|
||||||
|
atomic64_read(&kmem_alloc_used), kmem_alloc_max);
|
||||||
|
}
|
||||||
|
RETURN(ptr);
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(kmem_alloc_debug);
|
||||||
|
|
||||||
|
void
|
||||||
|
kmem_free_debug(void *ptr, size_t size)
|
||||||
|
{
|
||||||
|
ENTRY;
|
||||||
|
|
||||||
|
ASSERTF(ptr || size > 0, "ptr: %p, size: %llu", ptr,
|
||||||
|
(unsigned long long) size);
|
||||||
|
|
||||||
|
atomic64_sub(size, &kmem_alloc_used);
|
||||||
|
|
||||||
|
CDEBUG_LIMIT(D_INFO, "kmem_free(%p, %llu) (%lld/%llu)\n", ptr,
|
||||||
|
(unsigned long long) size, atomic64_read(&kmem_alloc_used),
|
||||||
|
kmem_alloc_max);
|
||||||
|
|
||||||
|
memset(ptr, 0x5a, size);
|
||||||
|
kfree(ptr);
|
||||||
|
|
||||||
|
EXIT;
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(kmem_free_debug);
|
||||||
|
|
||||||
|
void *
|
||||||
|
vmem_alloc_debug(size_t size, int flags, const char *func, int line)
|
||||||
|
{
|
||||||
|
void *ptr;
|
||||||
|
ENTRY;
|
||||||
|
|
||||||
|
ASSERT(flags & KM_SLEEP);
|
||||||
|
|
||||||
|
ptr = __vmalloc(size, (flags | __GFP_HIGHMEM) & ~__GFP_ZERO,
|
||||||
|
PAGE_KERNEL);
|
||||||
|
if (ptr == NULL) {
|
||||||
|
CWARN("vmem_alloc(%llu, 0x%x) failed (%lld/%llu)\n",
|
||||||
|
(unsigned long long) size, flags,
|
||||||
|
atomic64_read(&vmem_alloc_used), vmem_alloc_max);
|
||||||
|
} else {
|
||||||
|
if (flags & __GFP_ZERO)
|
||||||
|
memset(ptr, 0, size);
|
||||||
|
|
||||||
|
atomic64_add(size, &vmem_alloc_used);
|
||||||
|
|
||||||
|
if (unlikely(atomic64_read(&vmem_alloc_used) > vmem_alloc_max))
|
||||||
|
vmem_alloc_max = atomic64_read(&vmem_alloc_used);
|
||||||
|
|
||||||
|
CDEBUG_LIMIT(D_INFO, "vmem_alloc(%llu, 0x%x) = %p "
|
||||||
|
"(%lld/%llu)\n", (unsigned long long) size, flags, ptr,
|
||||||
|
atomic64_read(&vmem_alloc_used), vmem_alloc_max);
|
||||||
|
}
|
||||||
|
|
||||||
|
RETURN(ptr);
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(vmem_alloc_debug);
|
||||||
|
|
||||||
|
void
|
||||||
|
vmem_free_debug(void *ptr, size_t size)
|
||||||
|
{
|
||||||
|
ENTRY;
|
||||||
|
|
||||||
|
ASSERTF(ptr || size > 0, "ptr: %p, size: %llu", ptr,
|
||||||
|
(unsigned long long) size);
|
||||||
|
|
||||||
|
atomic64_sub(size, &vmem_alloc_used);
|
||||||
|
|
||||||
|
CDEBUG_LIMIT(D_INFO, "vmem_free(%p, %llu) (%lld/%llu)\n", ptr,
|
||||||
|
(unsigned long long) size, atomic64_read(&vmem_alloc_used),
|
||||||
|
vmem_alloc_max);
|
||||||
|
|
||||||
|
memset(ptr, 0x5a, size);
|
||||||
|
vfree(ptr);
|
||||||
|
|
||||||
|
EXIT;
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(vmem_free_debug);
|
||||||
|
|
||||||
|
# endif /* DEBUG_KMEM_TRACKING */
|
||||||
|
#endif /* DEBUG_KMEM */
|
||||||
|
|
||||||
static void *
|
static void *
|
||||||
kv_alloc(spl_kmem_cache_t *skc, int size, int flags)
|
kv_alloc(spl_kmem_cache_t *skc, int size, int flags)
|
||||||
{
|
{
|
||||||
|
@ -386,10 +754,14 @@ spl_magazine_alloc(spl_kmem_cache_t *skc, int node)
|
||||||
static void
|
static void
|
||||||
spl_magazine_free(spl_kmem_magazine_t *skm)
|
spl_magazine_free(spl_kmem_magazine_t *skm)
|
||||||
{
|
{
|
||||||
|
int size = sizeof(spl_kmem_magazine_t) +
|
||||||
|
sizeof(void *) * skm->skm_size;
|
||||||
|
|
||||||
ENTRY;
|
ENTRY;
|
||||||
ASSERT(skm->skm_magic == SKM_MAGIC);
|
ASSERT(skm->skm_magic == SKM_MAGIC);
|
||||||
ASSERT(skm->skm_avail == 0);
|
ASSERT(skm->skm_avail == 0);
|
||||||
kfree(skm);
|
|
||||||
|
kmem_free(skm, size);
|
||||||
EXIT;
|
EXIT;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -976,13 +1348,12 @@ spl_kmem_fini_tracking(struct list_head *list, spinlock_t *lock)
|
||||||
|
|
||||||
spin_lock_irqsave(lock, flags);
|
spin_lock_irqsave(lock, flags);
|
||||||
if (!list_empty(list))
|
if (!list_empty(list))
|
||||||
CDEBUG(D_WARNING, "%-16s %-5s %-16s %s:%s\n",
|
printk(KERN_WARNING "%-16s %-5s %-16s %s:%s\n", "address",
|
||||||
"address", "size", "data", "func", "line");
|
"size", "data", "func", "line");
|
||||||
|
|
||||||
list_for_each_entry(kd, list, kd_list)
|
list_for_each_entry(kd, list, kd_list)
|
||||||
CDEBUG(D_WARNING, "%p %-5d %-16s %s:%d\n",
|
printk(KERN_WARNING "%p %-5d %-16s %s:%d\n", kd->kd_addr,
|
||||||
kd->kd_addr, kd->kd_size,
|
kd->kd_size, spl_sprintf_addr(kd, str, 17, 8),
|
||||||
spl_sprintf_addr(kd, str, 17, 8),
|
|
||||||
kd->kd_func, kd->kd_line);
|
kd->kd_func, kd->kd_line);
|
||||||
|
|
||||||
spin_unlock_irqrestore(lock, flags);
|
spin_unlock_irqrestore(lock, flags);
|
||||||
|
|
Loading…
Reference in New Issue