zfs/modules/spl/spl-debug.c

1294 lines
36 KiB
C
Raw Normal View History

/* -*- mode: c; c-basic-offset: 8; indent-tabs-mode: nil; -*-
* vim:expandtab:shiftwidth=8:tabstop=8:
*
* Copyright (C) 2004 Cluster File Systems, Inc.
* Author: Zach Brown <zab@clusterfs.com>
* Author: Phil Schwan <phil@clusterfs.com>
* Author: Brian Behlendorf <behlendorf1@llnl.gov>
*
* This file was originally part of Lustre, http://www.lustre.org.
* but has subsequently been adapted for use in the SPL in
* accordance with the GPL.
*
* SPL is free software; you can redistribute it and/or
* modify it under the terms of version 2 of the GNU General Public
* License as published by the Free Software Foundation.
*
* SPL is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with SPL; if not, write to the Free Software
* Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
*/
#include <linux/kmod.h>
#include <linux/mm.h>
#include <linux/vmalloc.h>
#include <linux/pagemap.h>
#include <linux/slab.h>
#include <linux/ctype.h>
#include <linux/kthread.h>
#include <linux/hardirq.h>
#include <linux/interrupt.h>
#include <sys/sysmacros.h>
#include <sys/proc.h>
#include <sys/debug.h>
#include <spl-ctl.h>
#include "config.h"
#ifdef DEBUG_SUBSYSTEM
#undef DEBUG_SUBSYSTEM
#endif
#define DEBUG_SUBSYSTEM S_DEBUG
unsigned long spl_debug_subsys = ~0;
EXPORT_SYMBOL(spl_debug_subsys);
module_param(spl_debug_subsys, long, 0644);
MODULE_PARM_DESC(spl_debug_subsys, "Subsystem debugging level mask.");
unsigned long spl_debug_mask = (D_EMERG | D_ERROR | D_WARNING | D_CONSOLE);
EXPORT_SYMBOL(spl_debug_mask);
module_param(spl_debug_mask, long, 0644);
MODULE_PARM_DESC(spl_debug_mask, "Debugging level mask.");
unsigned long spl_debug_printk = D_CANTMASK;
EXPORT_SYMBOL(spl_debug_printk);
module_param(spl_debug_printk, long, 0644);
MODULE_PARM_DESC(spl_debug_printk, "Console printk level mask.");
int spl_debug_mb = -1;
EXPORT_SYMBOL(spl_debug_mb);
module_param(spl_debug_mb, int, 0644);
MODULE_PARM_DESC(spl_debug_mb, "Total debug buffer size.");
unsigned int spl_debug_binary = 1;
EXPORT_SYMBOL(spl_debug_binary);
unsigned int spl_debug_catastrophe;
EXPORT_SYMBOL(spl_debug_catastrophe);
unsigned int spl_debug_panic_on_bug = 1;
EXPORT_SYMBOL(spl_debug_panic_on_bug);
module_param(spl_debug_panic_on_bug, int, 0644);
MODULE_PARM_DESC(spl_debug_panic_on_bug, "Panic on BUG");
static char spl_debug_file_name[PATH_MAX];
char spl_debug_file_path[PATH_MAX] = "/var/dumps/spl-log";
unsigned int spl_console_ratelimit = 1;
EXPORT_SYMBOL(spl_console_ratelimit);
long spl_console_max_delay;
EXPORT_SYMBOL(spl_console_max_delay);
long spl_console_min_delay;
EXPORT_SYMBOL(spl_console_min_delay);
unsigned int spl_console_backoff = SPL_DEFAULT_BACKOFF;
EXPORT_SYMBOL(spl_console_backoff);
unsigned int spl_debug_stack;
EXPORT_SYMBOL(spl_debug_stack);
static int spl_panic_in_progress;
union trace_data_union (*trace_data[TCD_TYPE_MAX])[NR_CPUS] __cacheline_aligned;
char *trace_console_buffers[NR_CPUS][3];
struct rw_semaphore trace_sem;
atomic_t trace_tage_allocated = ATOMIC_INIT(0);
static int panic_notifier(struct notifier_block *, unsigned long, void *);
static int spl_debug_dump_all_pages(char *);
static void trace_fini(void);
/* Memory percentage breakdown by type */
static unsigned int pages_factor[TCD_TYPE_MAX] = {
80, /* 80% pages for TCD_TYPE_PROC */
10, /* 10% pages for TCD_TYPE_SOFTIRQ */
10 /* 10% pages for TCD_TYPE_IRQ */
};
static struct notifier_block spl_panic_notifier = {
notifier_call: panic_notifier,
next: NULL,
priority: 10000
};
const char *
spl_debug_subsys2str(int subsys)
{
switch (subsys) {
default:
return NULL;
case S_UNDEFINED:
return "undefined";
case S_ATOMIC:
return "atomic";
case S_KOBJ:
return "kobj";
case S_VNODE:
return "vnode";
case S_TIME:
return "time";
case S_RWLOCK:
return "rwlock";
case S_THREAD:
return "thread";
case S_CONDVAR:
return "condvar";
case S_MUTEX:
return "mutex";
case S_RNG:
return "rng";
case S_TASKQ:
return "taskq";
case S_KMEM:
return "kmem";
}
}
const char *
spl_debug_dbg2str(int debug)
{
switch (debug) {
default:
return NULL;
case D_TRACE:
return "trace";
case D_INFO:
return "info";
case D_WARNING:
return "warning";
case D_ERROR:
return "error";
case D_EMERG:
return "emerg";
case D_CONSOLE:
return "console";
case D_IOCTL:
return "ioctl";
case D_DPRINTF:
return "dprintf";
case D_OTHER:
return "other";
}
}
int
spl_debug_mask2str(char *str, int size, unsigned long mask, int is_subsys)
{
const char *(*fn)(int bit) = is_subsys ? spl_debug_subsys2str :
spl_debug_dbg2str;
const char *token;
int i, bit, len = 0;
if (mask == 0) { /* "0" */
if (size > 0)
str[0] = '0';
len = 1;
} else { /* space-separated tokens */
for (i = 0; i < 32; i++) {
bit = 1 << i;
if ((mask & bit) == 0)
continue;
token = fn(bit);
if (token == NULL) /* unused bit */
continue;
if (len > 0) { /* separator? */
if (len < size)
str[len] = ' ';
len++;
}
while (*token != 0) {
if (len < size)
str[len] = *token;
token++;
len++;
}
}
}
/* terminate 'str' */
if (len < size)
str[len] = 0;
else
str[size - 1] = 0;
return len;
}
static int
spl_debug_token2mask(int *mask, const char *str, int len, int is_subsys)
{
const char *(*fn)(int bit) = is_subsys ? spl_debug_subsys2str :
spl_debug_dbg2str;
const char *token;
int i, j, bit;
/* match against known tokens */
for (i = 0; i < 32; i++) {
bit = 1 << i;
token = fn(bit);
if (token == NULL) /* unused? */
continue;
/* strcasecmp */
for (j = 0; ; j++) {
if (j == len) { /* end of token */
if (token[j] == 0) {
*mask = bit;
return 0;
}
break;
}
if (token[j] == 0)
break;
if (str[j] == token[j])
continue;
if (str[j] < 'A' || 'Z' < str[j])
break;
if (str[j] - 'A' + 'a' != token[j])
break;
}
}
return -EINVAL; /* no match */
}
int
spl_debug_str2mask(unsigned long *mask, const char *str, int is_subsys)
{
char op = 0;
int m = 0, matched, n, t;
/* Allow a number for backwards compatibility */
for (n = strlen(str); n > 0; n--)
if (!isspace(str[n-1]))
break;
matched = n;
if ((t = sscanf(str, "%i%n", &m, &matched)) >= 1 && matched == n) {
*mask = m;
return 0;
}
/* <str> must be a list of debug tokens or numbers separated by
* whitespace and optionally an operator ('+' or '-'). If an operator
* appears first in <str>, '*mask' is used as the starting point
* (relative), otherwise 0 is used (absolute). An operator applies to
* all following tokens up to the next operator. */
matched = 0;
while (*str != 0) {
while (isspace(*str)) /* skip whitespace */
str++;
if (*str == 0)
break;
if (*str == '+' || *str == '-') {
op = *str++;
/* op on first token == relative */
if (!matched)
m = *mask;
while (isspace(*str)) /* skip whitespace */
str++;
if (*str == 0) /* trailing op */
return -EINVAL;
}
/* find token length */
for (n = 0; str[n] != 0 && !isspace(str[n]); n++);
/* match token */
if (spl_debug_token2mask(&t, str, n, is_subsys) != 0)
return -EINVAL;
matched = 1;
if (op == '-')
m &= ~t;
else
m |= t;
str += n;
}
if (!matched)
return -EINVAL;
*mask = m;
return 0;
}
typedef struct dumplog_priv {
wait_queue_head_t dp_waitq;
pid_t dp_pid;
atomic_t dp_flag;
} dumplog_priv_t;
static void
spl_debug_dumplog_internal(dumplog_priv_t *dp)
{
void *journal_info;
journal_info = current->journal_info;
current->journal_info = NULL;
snprintf(spl_debug_file_name, sizeof(spl_debug_file_path) - 1,
"%s.%ld.%ld", spl_debug_file_path,
get_seconds(), (long)dp->dp_pid);
printk(KERN_ALERT "SPL: dumping log to %s\n", spl_debug_file_name);
spl_debug_dump_all_pages(spl_debug_file_name);
current->journal_info = journal_info;
}
static int
spl_debug_dumplog_thread(void *arg)
{
dumplog_priv_t *dp = (dumplog_priv_t *)arg;
spl_debug_dumplog_internal(dp);
atomic_set(&dp->dp_flag, 1);
wake_up(&dp->dp_waitq);
do_exit(0);
return 0; /* Unreachable */
}
int
spl_debug_dumplog(void)
{
struct task_struct *tsk;
dumplog_priv_t dp;
init_waitqueue_head(&dp.dp_waitq);
dp.dp_pid = current->pid;
atomic_set(&dp.dp_flag, 0);
tsk = kthread_create(spl_debug_dumplog_thread,(void *)&dp,"spl_debug");
if (tsk == NULL)
return -ENOMEM;
wake_up_process(tsk);
wait_event(dp.dp_waitq, atomic_read(&dp.dp_flag));
return 0;
}
EXPORT_SYMBOL(spl_debug_dumplog);
static char *
trace_get_console_buffer(void)
{
int cpu = get_cpu();
int idx;
if (in_irq()) {
idx = 0;
} else if (in_softirq()) {
idx = 1;
} else {
idx = 2;
}
return trace_console_buffers[cpu][idx];
}
static void
trace_put_console_buffer(char *buffer)
{
put_cpu();
}
static struct trace_cpu_data *
trace_get_tcd(void)
{
int cpu;
cpu = get_cpu();
if (in_irq())
return &(*trace_data[TCD_TYPE_IRQ])[cpu].tcd;
else if (in_softirq())
return &(*trace_data[TCD_TYPE_SOFTIRQ])[cpu].tcd;
return &(*trace_data[TCD_TYPE_PROC])[cpu].tcd;
}
static void
trace_put_tcd (struct trace_cpu_data *tcd)
{
put_cpu();
}
static int
trace_lock_tcd(struct trace_cpu_data *tcd)
{
__ASSERT(tcd->tcd_type < TCD_TYPE_MAX);
if (tcd->tcd_type == TCD_TYPE_IRQ)
local_irq_disable();
else if (tcd->tcd_type == TCD_TYPE_SOFTIRQ)
local_bh_disable();
return 1;
}
static void
trace_unlock_tcd(struct trace_cpu_data *tcd)
{
__ASSERT(tcd->tcd_type < TCD_TYPE_MAX);
if (tcd->tcd_type == TCD_TYPE_IRQ)
local_irq_enable();
else if (tcd->tcd_type == TCD_TYPE_SOFTIRQ)
local_bh_enable();
}
static void
trace_set_debug_header(struct spl_debug_header *header, int subsys,
int mask, const int line, unsigned long stack)
{
struct timeval tv;
do_gettimeofday(&tv);
header->ph_subsys = subsys;
header->ph_mask = mask;
header->ph_cpu_id = smp_processor_id();
header->ph_sec = (__u32)tv.tv_sec;
header->ph_usec = tv.tv_usec;
header->ph_stack = stack;
header->ph_pid = current->pid;
header->ph_line_num = line;
return;
}
static void
trace_print_to_console(struct spl_debug_header *hdr, int mask, const char *buf,
int len, const char *file, const char *fn)
{
char *prefix = "SPL", *ptype = NULL;
if ((mask & D_EMERG) != 0) {
prefix = "SPLError";
ptype = KERN_EMERG;
} else if ((mask & D_ERROR) != 0) {
prefix = "SPLError";
ptype = KERN_ERR;
} else if ((mask & D_WARNING) != 0) {
prefix = "SPL";
ptype = KERN_WARNING;
} else if ((mask & (D_CONSOLE | spl_debug_printk)) != 0) {
prefix = "SPL";
ptype = KERN_INFO;
}
if ((mask & D_CONSOLE) != 0) {
printk("%s%s: %.*s", ptype, prefix, len, buf);
} else {
printk("%s%s: %d:%d:(%s:%d:%s()) %.*s", ptype, prefix,
hdr->ph_pid, hdr->ph_stack, file,
hdr->ph_line_num, fn, len, buf);
}
return;
}
static int
trace_max_debug_mb(void)
{
return MAX(512, ((num_physpages >> (20 - PAGE_SHIFT)) * 80) / 100);
}
static void
trace_call_on_all_cpus(void (*fn)(void *arg), void *arg)
{
cpumask_t mask, cpus_allowed = current->cpus_allowed;
int cpu;
for_each_online_cpu(cpu) {
cpus_clear(mask);
cpu_set(cpu, mask);
set_cpus_allowed(current, mask);
fn(arg);
set_cpus_allowed(current, cpus_allowed);
}
}
static struct trace_page *
tage_alloc(int gfp)
{
struct page *page;
struct trace_page *tage;
page = alloc_pages(gfp | __GFP_NOWARN, 0);
if (page == NULL)
return NULL;
tage = kmalloc(sizeof(*tage), gfp);
if (tage == NULL) {
__free_pages(page, 0);
return NULL;
}
tage->page = page;
atomic_inc(&trace_tage_allocated);
return tage;
}
static void
tage_free(struct trace_page *tage)
{
__ASSERT(tage != NULL);
__ASSERT(tage->page != NULL);
__free_pages(tage->page, 0);
kfree(tage);
atomic_dec(&trace_tage_allocated);
}
static struct trace_page *
tage_from_list(struct list_head *list)
{
return list_entry(list, struct trace_page, linkage);
}
static void
tage_to_tail(struct trace_page *tage, struct list_head *queue)
{
__ASSERT(tage != NULL);
__ASSERT(queue != NULL);
list_move_tail(&tage->linkage, queue);
}
/* try to return a page that has 'len' bytes left at the end */
static struct trace_page *
trace_get_tage_try(struct trace_cpu_data *tcd, unsigned long len)
{
struct trace_page *tage;
if (tcd->tcd_cur_pages > 0) {
__ASSERT(!list_empty(&tcd->tcd_pages));
tage = tage_from_list(tcd->tcd_pages.prev);
if (tage->used + len <= PAGE_SIZE)
return tage;
}
if (tcd->tcd_cur_pages < tcd->tcd_max_pages) {
if (tcd->tcd_cur_stock_pages > 0) {
tage = tage_from_list(tcd->tcd_stock_pages.prev);
tcd->tcd_cur_stock_pages--;
list_del_init(&tage->linkage);
} else {
tage = tage_alloc(GFP_ATOMIC);
if (tage == NULL) {
printk(KERN_WARNING
"failure to allocate a tage (%ld)\n",
tcd->tcd_cur_pages);
return NULL;
}
}
tage->used = 0;
tage->cpu = smp_processor_id();
tage->type = tcd->tcd_type;
list_add_tail(&tage->linkage, &tcd->tcd_pages);
tcd->tcd_cur_pages++;
return tage;
}
return NULL;
}
/* return a page that has 'len' bytes left at the end */
static struct trace_page *
trace_get_tage(struct trace_cpu_data *tcd, unsigned long len)
{
struct trace_page *tage;
__ASSERT(len <= PAGE_SIZE);
tage = trace_get_tage_try(tcd, len);
if (tage)
return tage;
if (tcd->tcd_cur_pages > 0) {
tage = tage_from_list(tcd->tcd_pages.next);
tage->used = 0;
tage_to_tail(tage, &tcd->tcd_pages);
}
return tage;
}
int
spl_debug_vmsg(spl_debug_limit_state_t *cdls, int subsys, int mask,
const char *file, const char *fn, const int line,
const char *format1, va_list args, const char *format2, ...)
{
struct trace_cpu_data *tcd = NULL;
struct spl_debug_header header;
struct trace_page *tage;
/* string_buf is used only if tcd != NULL, and is always set then */
char *string_buf = NULL;
char *debug_buf;
int known_size;
int needed = 85; /* average message length */
int max_nob;
va_list ap;
int i;
int remain;
if (strchr(file, '/'))
file = strrchr(file, '/') + 1;
trace_set_debug_header(&header, subsys, mask, line, CDEBUG_STACK());
tcd = trace_get_tcd();
if (tcd == NULL)
goto console;
if (tcd->tcd_shutting_down) {
trace_put_tcd(tcd);
tcd = NULL;
goto console;
}
known_size = strlen(file) + 1;
if (fn)
known_size += strlen(fn) + 1;
if (spl_debug_binary)
known_size += sizeof(header);
/* '2' used because vsnprintf returns real size required for output
* _without_ terminating NULL. */
for (i = 0; i < 2; i++) {
tage = trace_get_tage(tcd, needed + known_size + 1);
if (tage == NULL) {
if (needed + known_size > PAGE_SIZE)
mask |= D_ERROR;
trace_put_tcd(tcd);
tcd = NULL;
goto console;
}
string_buf = (char *)page_address(tage->page) +
tage->used + known_size;
max_nob = PAGE_SIZE - tage->used - known_size;
if (max_nob <= 0) {
printk(KERN_EMERG "negative max_nob: %i\n", max_nob);
mask |= D_ERROR;
trace_put_tcd(tcd);
tcd = NULL;
goto console;
}
needed = 0;
if (format1) {
va_copy(ap, args);
needed = vsnprintf(string_buf, max_nob, format1, ap);
va_end(ap);
}
if (format2) {
remain = max_nob - needed;
if (remain < 0)
remain = 0;
va_start(ap, format2);
needed += vsnprintf(string_buf+needed, remain, format2, ap);
va_end(ap);
}
if (needed < max_nob)
break;
}
if (unlikely(*(string_buf + needed - 1) != '\n'))
printk(KERN_INFO "format at %s:%d:%s doesn't end in newline\n",
file, line, fn);
header.ph_len = known_size + needed;
debug_buf = (char *)page_address(tage->page) + tage->used;
if (spl_debug_binary) {
memcpy(debug_buf, &header, sizeof(header));
tage->used += sizeof(header);
debug_buf += sizeof(header);
}
strcpy(debug_buf, file);
tage->used += strlen(file) + 1;
debug_buf += strlen(file) + 1;
if (fn) {
strcpy(debug_buf, fn);
tage->used += strlen(fn) + 1;
debug_buf += strlen(fn) + 1;
}
__ASSERT(debug_buf == string_buf);
tage->used += needed;
__ASSERT (tage->used <= PAGE_SIZE);
console:
if ((mask & spl_debug_printk) == 0) {
/* no console output requested */
if (tcd != NULL)
trace_put_tcd(tcd);
return 1;
}
if (cdls != NULL) {
if (spl_console_ratelimit && cdls->cdls_next != 0 &&
!time_before(cdls->cdls_next, jiffies)) {
/* skipping a console message */
cdls->cdls_count++;
if (tcd != NULL)
trace_put_tcd(tcd);
return 1;
}
if (time_before(cdls->cdls_next + spl_console_max_delay +
(10 * HZ), jiffies)) {
/* last timeout was a long time ago */
cdls->cdls_delay /= spl_console_backoff * 4;
} else {
cdls->cdls_delay *= spl_console_backoff;
if (cdls->cdls_delay < spl_console_min_delay)
cdls->cdls_delay = spl_console_min_delay;
else if (cdls->cdls_delay > spl_console_max_delay)
cdls->cdls_delay = spl_console_max_delay;
}
/* ensure cdls_next is never zero after it's been seen */
cdls->cdls_next = (jiffies + cdls->cdls_delay) | 1;
}
if (tcd != NULL) {
trace_print_to_console(&header, mask, string_buf, needed, file, fn);
trace_put_tcd(tcd);
} else {
string_buf = trace_get_console_buffer();
needed = 0;
if (format1 != NULL) {
va_copy(ap, args);
needed = vsnprintf(string_buf, TRACE_CONSOLE_BUFFER_SIZE, format1, ap);
va_end(ap);
}
if (format2 != NULL) {
remain = TRACE_CONSOLE_BUFFER_SIZE - needed;
if (remain > 0) {
va_start(ap, format2);
needed += vsnprintf(string_buf+needed, remain, format2, ap);
va_end(ap);
}
}
trace_print_to_console(&header, mask,
string_buf, needed, file, fn);
trace_put_console_buffer(string_buf);
}
if (cdls != NULL && cdls->cdls_count != 0) {
string_buf = trace_get_console_buffer();
needed = snprintf(string_buf, TRACE_CONSOLE_BUFFER_SIZE,
"Skipped %d previous similar message%s\n",
cdls->cdls_count, (cdls->cdls_count > 1) ? "s" : "");
trace_print_to_console(&header, mask,
string_buf, needed, file, fn);
trace_put_console_buffer(string_buf);
cdls->cdls_count = 0;
}
return 0;
}
EXPORT_SYMBOL(spl_debug_vmsg);
/* Do the collect_pages job on a single CPU: assumes that all other
* CPUs have been stopped during a panic. If this isn't true for
* some arch, this will have to be implemented separately in each arch.
*/
static void
panic_collect_pages(struct page_collection *pc)
{
struct trace_cpu_data *tcd;
int i, j;
tcd_for_each(tcd, i, j) {
list_splice_init(&tcd->tcd_pages, &pc->pc_pages);
tcd->tcd_cur_pages = 0;
}
}
static void
collect_pages_on_cpu(void *info)
{
struct trace_cpu_data *tcd;
struct page_collection *pc = info;
int i;
spin_lock(&pc->pc_lock);
tcd_for_each_type_lock(tcd, i) {
list_splice_init(&tcd->tcd_pages, &pc->pc_pages);
tcd->tcd_cur_pages = 0;
}
spin_unlock(&pc->pc_lock);
}
static void
collect_pages(struct page_collection *pc)
{
INIT_LIST_HEAD(&pc->pc_pages);
if (spl_panic_in_progress)
panic_collect_pages(pc);
else
trace_call_on_all_cpus(collect_pages_on_cpu, pc);
}
static void
put_pages_back_on_cpu(void *info)
{
struct page_collection *pc = info;
struct trace_cpu_data *tcd;
struct list_head *cur_head;
struct trace_page *tage;
struct trace_page *tmp;
int i;
spin_lock(&pc->pc_lock);
tcd_for_each_type_lock(tcd, i) {
cur_head = tcd->tcd_pages.next;
list_for_each_entry_safe(tage, tmp, &pc->pc_pages, linkage) {
__ASSERT_TAGE_INVARIANT(tage);
if (tage->cpu != smp_processor_id() || tage->type != i)
continue;
tage_to_tail(tage, cur_head);
tcd->tcd_cur_pages++;
}
}
spin_unlock(&pc->pc_lock);
}
static void
put_pages_back(struct page_collection *pc)
{
if (!spl_panic_in_progress)
trace_call_on_all_cpus(put_pages_back_on_cpu, pc);
}
static struct file *
trace_filp_open (const char *name, int flags, int mode, int *err)
{
struct file *filp = NULL;
int rc;
filp = filp_open(name, flags, mode);
if (IS_ERR(filp)) {
rc = PTR_ERR(filp);
printk(KERN_ERR "SPL: Can't open %s file: %d\n", name, rc);
if (err)
*err = rc;
filp = NULL;
}
return filp;
}
#define trace_filp_write(fp, b, s, p) (fp)->f_op->write((fp), (b), (s), p)
#define trace_filp_fsync(fp) (fp)->f_op->fsync((fp),(fp)->f_dentry,1)
#define trace_filp_close(f) filp_close(f, NULL)
#define trace_filp_poff(f) (&(f)->f_pos)
static int
spl_debug_dump_all_pages(char *filename)
{
struct page_collection pc;
struct file *filp;
struct trace_page *tage;
struct trace_page *tmp;
mm_segment_t oldfs;
int rc = 0;
down_write(&trace_sem);
filp = trace_filp_open(filename, O_CREAT|O_EXCL|O_WRONLY|O_LARGEFILE,
0600, &rc);
if (filp == NULL) {
if (rc != -EEXIST)
printk(KERN_ERR "SPL: Can't open %s for dump: %d\n",
filename, rc);
goto out;
}
spin_lock_init(&pc.pc_lock);
collect_pages(&pc);
if (list_empty(&pc.pc_pages)) {
rc = 0;
goto close;
}
oldfs = get_fs();
set_fs(get_ds());
list_for_each_entry_safe(tage, tmp, &pc.pc_pages, linkage) {
__ASSERT_TAGE_INVARIANT(tage);
rc = trace_filp_write(filp, page_address(tage->page),
tage->used, trace_filp_poff(filp));
if (rc != (int)tage->used) {
printk(KERN_WARNING "SPL: Wanted to write %u "
"but wrote %d\n", tage->used, rc);
put_pages_back(&pc);
__ASSERT(list_empty(&pc.pc_pages));
break;
}
list_del(&tage->linkage);
tage_free(tage);
}
set_fs(oldfs);
rc = trace_filp_fsync(filp);
if (rc)
printk(KERN_ERR "SPL: Unable to sync: %d\n", rc);
close:
trace_filp_close(filp);
out:
up_write(&trace_sem);
return rc;
}
static void
spl_debug_flush_pages(void)
{
struct page_collection pc;
struct trace_page *tage;
struct trace_page *tmp;
spin_lock_init(&pc.pc_lock);
collect_pages(&pc);
list_for_each_entry_safe(tage, tmp, &pc.pc_pages, linkage) {
__ASSERT_TAGE_INVARIANT(tage);
list_del(&tage->linkage);
tage_free(tage);
}
}
unsigned long
spl_debug_set_mask(unsigned long mask) {
spl_debug_mask = mask;
return 0;
}
EXPORT_SYMBOL(spl_debug_set_mask);
unsigned long
spl_debug_get_mask(void) {
return spl_debug_mask;
}
EXPORT_SYMBOL(spl_debug_get_mask);
unsigned long
spl_debug_set_subsys(unsigned long subsys) {
spl_debug_subsys = subsys;
return 0;
}
EXPORT_SYMBOL(spl_debug_set_subsys);
unsigned long
spl_debug_get_subsys(void) {
return spl_debug_subsys;
}
EXPORT_SYMBOL(spl_debug_get_subsys);
int
spl_debug_set_mb(int mb)
{
int i, j, pages;
int limit = trace_max_debug_mb();
struct trace_cpu_data *tcd;
if (mb < num_possible_cpus()) {
printk(KERN_ERR "SPL: Refusing to set debug buffer size to "
"%dMB - lower limit is %d\n", mb, num_possible_cpus());
return -EINVAL;
}
if (mb > limit) {
printk(KERN_ERR "SPL: Refusing to set debug buffer size to "
"%dMB - upper limit is %d\n", mb, limit);
return -EINVAL;
}
mb /= num_possible_cpus();
pages = mb << (20 - PAGE_SHIFT);
down_write(&trace_sem);
tcd_for_each(tcd, i, j)
tcd->tcd_max_pages = (pages * tcd->tcd_pages_factor) / 100;
up_write(&trace_sem);
return 0;
}
EXPORT_SYMBOL(spl_debug_set_mb);
int
spl_debug_get_mb(void)
{
int i, j;
struct trace_cpu_data *tcd;
int total_pages = 0;
down_read(&trace_sem);
tcd_for_each(tcd, i, j)
total_pages += tcd->tcd_max_pages;
up_read(&trace_sem);
return (total_pages >> (20 - PAGE_SHIFT)) + 1;
}
EXPORT_SYMBOL(spl_debug_get_mb);
void spl_debug_dumpstack(struct task_struct *tsk)
{
extern void show_task(struct task_struct *);
if (tsk == NULL)
tsk = current;
printk(KERN_ERR "SPL: Showing stack for process %d\n", tsk->pid);
show_task(tsk);
}
EXPORT_SYMBOL(spl_debug_dumpstack);
void spl_debug_bug(char *file, const char *func, const int line)
{
spl_debug_catastrophe = 1;
spl_debug_msg(NULL, 0, D_EMERG, file, func, line, "SBUG\n");
if (in_interrupt()) {
panic("SBUG in interrupt.\n");
/* not reached */
}
/* Ensure all debug pages and dumped by current cpu */
if (spl_debug_panic_on_bug)
spl_panic_in_progress = 1;
spl_debug_dumpstack(NULL);
spl_debug_dumplog();
if (spl_debug_panic_on_bug)
panic("SBUG");
set_task_state(current, TASK_UNINTERRUPTIBLE);
while (1)
schedule();
}
EXPORT_SYMBOL(spl_debug_bug);
int
spl_debug_clear_buffer(void)
{
spl_debug_flush_pages();
return 0;
}
EXPORT_SYMBOL(spl_debug_clear_buffer);
int
spl_debug_mark_buffer(char *text)
{
CDEBUG(D_WARNING, "*************************************\n");
CDEBUG(D_WARNING, "DEBUG MARKER: %s\n", text);
CDEBUG(D_WARNING, "*************************************\n");
return 0;
}
EXPORT_SYMBOL(spl_debug_mark_buffer);
static int
panic_notifier(struct notifier_block *self,
unsigned long unused1, void *unused2)
{
if (spl_panic_in_progress)
return 0;
spl_panic_in_progress = 1;
mb();
if (!in_interrupt()) {
while (current->lock_depth >= 0)
unlock_kernel();
spl_debug_dumplog_internal((void *)(long)current->pid);
}
return 0;
}
static int
trace_init(int max_pages)
{
struct trace_cpu_data *tcd;
int i, j;
init_rwsem(&trace_sem);
/* initialize trace_data */
memset(trace_data, 0, sizeof(trace_data));
for (i = 0; i < TCD_TYPE_MAX; i++) {
trace_data[i] = kmalloc(sizeof(union trace_data_union) *
NR_CPUS, GFP_KERNEL);
if (trace_data[i] == NULL)
goto out;
}
tcd_for_each(tcd, i, j) {
tcd->tcd_pages_factor = pages_factor[i];
tcd->tcd_type = i;
tcd->tcd_cpu = j;
INIT_LIST_HEAD(&tcd->tcd_pages);
INIT_LIST_HEAD(&tcd->tcd_stock_pages);
tcd->tcd_cur_pages = 0;
tcd->tcd_cur_stock_pages = 0;
tcd->tcd_max_pages = (max_pages * pages_factor[i]) / 100;
tcd->tcd_shutting_down = 0;
}
for (i = 0; i < num_possible_cpus(); i++) {
for (j = 0; j < 3; j++) {
trace_console_buffers[i][j] =
kmalloc(TRACE_CONSOLE_BUFFER_SIZE,
GFP_KERNEL);
if (trace_console_buffers[i][j] == NULL)
goto out;
}
}
return 0;
out:
trace_fini();
printk(KERN_ERR "SPL: Insufficient memory for debug logs\n");
return -ENOMEM;
}
int
debug_init(void)
{
int rc, max = spl_debug_mb;
spl_console_max_delay = SPL_DEFAULT_MAX_DELAY;
spl_console_min_delay = SPL_DEFAULT_MIN_DELAY;
/* If spl_debug_mb is set to an invalid value or uninitialized
* then just make the total buffers smp_num_cpus TCD_MAX_PAGES */
if (max > (num_physpages >> (20 - 2 - PAGE_SHIFT)) / 5 ||
max >= 512 || max < 0) {
max = TCD_MAX_PAGES;
} else {
max = (max / num_online_cpus()) << (20 - PAGE_SHIFT);
}
rc = trace_init(max);
if (rc)
return rc;
atomic_notifier_chain_register(&panic_notifier_list,
&spl_panic_notifier);
return rc;
}
static void
trace_cleanup_on_cpu(void *info)
{
struct trace_cpu_data *tcd;
struct trace_page *tage;
struct trace_page *tmp;
int i;
tcd_for_each_type_lock(tcd, i) {
tcd->tcd_shutting_down = 1;
list_for_each_entry_safe(tage, tmp, &tcd->tcd_pages, linkage) {
__ASSERT_TAGE_INVARIANT(tage);
list_del(&tage->linkage);
tage_free(tage);
}
tcd->tcd_cur_pages = 0;
}
}
static void
trace_fini(void)
{
int i, j;
trace_call_on_all_cpus(trace_cleanup_on_cpu, NULL);
for (i = 0; i < num_possible_cpus(); i++) {
for (j = 0; j < 3; j++) {
if (trace_console_buffers[i][j] != NULL) {
kfree(trace_console_buffers[i][j]);
trace_console_buffers[i][j] = NULL;
}
}
}
for (i = 0; trace_data[i] != NULL; i++) {
kfree(trace_data[i]);
trace_data[i] = NULL;
}
}
void
debug_fini(void)
{
atomic_notifier_chain_unregister(&panic_notifier_list,
&spl_panic_notifier);
trace_fini();
return;
}