324 lines
		
	
	
		
			7.2 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			324 lines
		
	
	
		
			7.2 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 * Copyright (c) 2000-2003,2005 Silicon Graphics, Inc.
 | 
						|
 * All Rights Reserved.
 | 
						|
 *
 | 
						|
 * This program is free software; you can redistribute it and/or
 | 
						|
 * modify it under the terms of the GNU General Public License as
 | 
						|
 * published by the Free Software Foundation.
 | 
						|
 *
 | 
						|
 * This program is distributed in the hope that it would be useful,
 | 
						|
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
						|
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | 
						|
 * GNU General Public License for more details.
 | 
						|
 *
 | 
						|
 * You should have received a copy of the GNU General Public License
 | 
						|
 * along with this program; if not, write the Free Software Foundation,
 | 
						|
 * Inc.,  51 Franklin St, Fifth Floor, Boston, MA  02110-1301  USA
 | 
						|
 */
 | 
						|
#include <xfs.h>
 | 
						|
 | 
						|
static kmem_zone_t *ktrace_hdr_zone;
 | 
						|
static kmem_zone_t *ktrace_ent_zone;
 | 
						|
static int          ktrace_zentries;
 | 
						|
 | 
						|
void __init
 | 
						|
ktrace_init(int zentries)
 | 
						|
{
 | 
						|
	ktrace_zentries = roundup_pow_of_two(zentries);
 | 
						|
 | 
						|
	ktrace_hdr_zone = kmem_zone_init(sizeof(ktrace_t),
 | 
						|
					"ktrace_hdr");
 | 
						|
	ASSERT(ktrace_hdr_zone);
 | 
						|
 | 
						|
	ktrace_ent_zone = kmem_zone_init(ktrace_zentries
 | 
						|
					* sizeof(ktrace_entry_t),
 | 
						|
					"ktrace_ent");
 | 
						|
	ASSERT(ktrace_ent_zone);
 | 
						|
}
 | 
						|
 | 
						|
void __exit
 | 
						|
ktrace_uninit(void)
 | 
						|
{
 | 
						|
	kmem_zone_destroy(ktrace_hdr_zone);
 | 
						|
	kmem_zone_destroy(ktrace_ent_zone);
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * ktrace_alloc()
 | 
						|
 *
 | 
						|
 * Allocate a ktrace header and enough buffering for the given
 | 
						|
 * number of entries. Round the number of entries up to a
 | 
						|
 * power of 2 so we can do fast masking to get the index from
 | 
						|
 * the atomic index counter.
 | 
						|
 */
 | 
						|
ktrace_t *
 | 
						|
ktrace_alloc(int nentries, unsigned int __nocast sleep)
 | 
						|
{
 | 
						|
	ktrace_t        *ktp;
 | 
						|
	ktrace_entry_t  *ktep;
 | 
						|
	int		entries;
 | 
						|
 | 
						|
	ktp = (ktrace_t*)kmem_zone_alloc(ktrace_hdr_zone, sleep);
 | 
						|
 | 
						|
	if (ktp == (ktrace_t*)NULL) {
 | 
						|
		/*
 | 
						|
		 * KM_SLEEP callers don't expect failure.
 | 
						|
		 */
 | 
						|
		if (sleep & KM_SLEEP)
 | 
						|
			panic("ktrace_alloc: NULL memory on KM_SLEEP request!");
 | 
						|
 | 
						|
		return NULL;
 | 
						|
	}
 | 
						|
 | 
						|
	/*
 | 
						|
	 * Special treatment for buffers with the ktrace_zentries entries
 | 
						|
	 */
 | 
						|
	entries = roundup_pow_of_two(nentries);
 | 
						|
	if (entries == ktrace_zentries) {
 | 
						|
		ktep = (ktrace_entry_t*)kmem_zone_zalloc(ktrace_ent_zone,
 | 
						|
							    sleep);
 | 
						|
	} else {
 | 
						|
		ktep = (ktrace_entry_t*)kmem_zalloc((entries * sizeof(*ktep)),
 | 
						|
							    sleep | KM_LARGE);
 | 
						|
	}
 | 
						|
 | 
						|
	if (ktep == NULL) {
 | 
						|
		/*
 | 
						|
		 * KM_SLEEP callers don't expect failure.
 | 
						|
		 */
 | 
						|
		if (sleep & KM_SLEEP)
 | 
						|
			panic("ktrace_alloc: NULL memory on KM_SLEEP request!");
 | 
						|
 | 
						|
		kmem_free(ktp);
 | 
						|
 | 
						|
		return NULL;
 | 
						|
	}
 | 
						|
 | 
						|
	ktp->kt_entries  = ktep;
 | 
						|
	ktp->kt_nentries = entries;
 | 
						|
	ASSERT(is_power_of_2(entries));
 | 
						|
	ktp->kt_index_mask = entries - 1;
 | 
						|
	atomic_set(&ktp->kt_index, 0);
 | 
						|
	ktp->kt_rollover = 0;
 | 
						|
	return ktp;
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
/*
 | 
						|
 * ktrace_free()
 | 
						|
 *
 | 
						|
 * Free up the ktrace header and buffer.  It is up to the caller
 | 
						|
 * to ensure that no-one is referencing it.
 | 
						|
 */
 | 
						|
void
 | 
						|
ktrace_free(ktrace_t *ktp)
 | 
						|
{
 | 
						|
	if (ktp == (ktrace_t *)NULL)
 | 
						|
		return;
 | 
						|
 | 
						|
	/*
 | 
						|
	 * Special treatment for the Vnode trace buffer.
 | 
						|
	 */
 | 
						|
	if (ktp->kt_nentries == ktrace_zentries)
 | 
						|
		kmem_zone_free(ktrace_ent_zone, ktp->kt_entries);
 | 
						|
	else
 | 
						|
		kmem_free(ktp->kt_entries);
 | 
						|
 | 
						|
	kmem_zone_free(ktrace_hdr_zone, ktp);
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
/*
 | 
						|
 * Enter the given values into the "next" entry in the trace buffer.
 | 
						|
 * kt_index is always the index of the next entry to be filled.
 | 
						|
 */
 | 
						|
void
 | 
						|
ktrace_enter(
 | 
						|
	ktrace_t        *ktp,
 | 
						|
	void            *val0,
 | 
						|
	void            *val1,
 | 
						|
	void            *val2,
 | 
						|
	void            *val3,
 | 
						|
	void            *val4,
 | 
						|
	void            *val5,
 | 
						|
	void            *val6,
 | 
						|
	void            *val7,
 | 
						|
	void            *val8,
 | 
						|
	void            *val9,
 | 
						|
	void            *val10,
 | 
						|
	void            *val11,
 | 
						|
	void            *val12,
 | 
						|
	void            *val13,
 | 
						|
	void            *val14,
 | 
						|
	void            *val15)
 | 
						|
{
 | 
						|
	int             index;
 | 
						|
	ktrace_entry_t  *ktep;
 | 
						|
 | 
						|
	ASSERT(ktp != NULL);
 | 
						|
 | 
						|
	/*
 | 
						|
	 * Grab an entry by pushing the index up to the next one.
 | 
						|
	 */
 | 
						|
	index = atomic_add_return(1, &ktp->kt_index);
 | 
						|
	index = (index - 1) & ktp->kt_index_mask;
 | 
						|
	if (!ktp->kt_rollover && index == ktp->kt_nentries - 1)
 | 
						|
		ktp->kt_rollover = 1;
 | 
						|
 | 
						|
	ASSERT((index >= 0) && (index < ktp->kt_nentries));
 | 
						|
 | 
						|
	ktep = &(ktp->kt_entries[index]);
 | 
						|
 | 
						|
	ktep->val[0]  = val0;
 | 
						|
	ktep->val[1]  = val1;
 | 
						|
	ktep->val[2]  = val2;
 | 
						|
	ktep->val[3]  = val3;
 | 
						|
	ktep->val[4]  = val4;
 | 
						|
	ktep->val[5]  = val5;
 | 
						|
	ktep->val[6]  = val6;
 | 
						|
	ktep->val[7]  = val7;
 | 
						|
	ktep->val[8]  = val8;
 | 
						|
	ktep->val[9]  = val9;
 | 
						|
	ktep->val[10] = val10;
 | 
						|
	ktep->val[11] = val11;
 | 
						|
	ktep->val[12] = val12;
 | 
						|
	ktep->val[13] = val13;
 | 
						|
	ktep->val[14] = val14;
 | 
						|
	ktep->val[15] = val15;
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * Return the number of entries in the trace buffer.
 | 
						|
 */
 | 
						|
int
 | 
						|
ktrace_nentries(
 | 
						|
	ktrace_t        *ktp)
 | 
						|
{
 | 
						|
	int	index;
 | 
						|
	if (ktp == NULL)
 | 
						|
		return 0;
 | 
						|
 | 
						|
	index = atomic_read(&ktp->kt_index) & ktp->kt_index_mask;
 | 
						|
	return (ktp->kt_rollover ? ktp->kt_nentries : index);
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * ktrace_first()
 | 
						|
 *
 | 
						|
 * This is used to find the start of the trace buffer.
 | 
						|
 * In conjunction with ktrace_next() it can be used to
 | 
						|
 * iterate through the entire trace buffer.  This code does
 | 
						|
 * not do any locking because it is assumed that it is called
 | 
						|
 * from the debugger.
 | 
						|
 *
 | 
						|
 * The caller must pass in a pointer to a ktrace_snap
 | 
						|
 * structure in which we will keep some state used to
 | 
						|
 * iterate through the buffer.  This state must not touched
 | 
						|
 * by any code outside of this module.
 | 
						|
 */
 | 
						|
ktrace_entry_t *
 | 
						|
ktrace_first(ktrace_t   *ktp, ktrace_snap_t     *ktsp)
 | 
						|
{
 | 
						|
	ktrace_entry_t  *ktep;
 | 
						|
	int             index;
 | 
						|
	int             nentries;
 | 
						|
 | 
						|
	if (ktp->kt_rollover)
 | 
						|
		index = atomic_read(&ktp->kt_index) & ktp->kt_index_mask;
 | 
						|
	else
 | 
						|
		index = 0;
 | 
						|
 | 
						|
	ktsp->ks_start = index;
 | 
						|
	ktep = &(ktp->kt_entries[index]);
 | 
						|
 | 
						|
	nentries = ktrace_nentries(ktp);
 | 
						|
	index++;
 | 
						|
	if (index < nentries) {
 | 
						|
		ktsp->ks_index = index;
 | 
						|
	} else {
 | 
						|
		ktsp->ks_index = 0;
 | 
						|
		if (index > nentries)
 | 
						|
			ktep = NULL;
 | 
						|
	}
 | 
						|
	return ktep;
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * ktrace_next()
 | 
						|
 *
 | 
						|
 * This is used to iterate through the entries of the given
 | 
						|
 * trace buffer.  The caller must pass in the ktrace_snap_t
 | 
						|
 * structure initialized by ktrace_first().  The return value
 | 
						|
 * will be either a pointer to the next ktrace_entry or NULL
 | 
						|
 * if all of the entries have been traversed.
 | 
						|
 */
 | 
						|
ktrace_entry_t *
 | 
						|
ktrace_next(
 | 
						|
	ktrace_t        *ktp,
 | 
						|
	ktrace_snap_t   *ktsp)
 | 
						|
{
 | 
						|
	int             index;
 | 
						|
	ktrace_entry_t  *ktep;
 | 
						|
 | 
						|
	index = ktsp->ks_index;
 | 
						|
	if (index == ktsp->ks_start) {
 | 
						|
		ktep = NULL;
 | 
						|
	} else {
 | 
						|
		ktep = &ktp->kt_entries[index];
 | 
						|
	}
 | 
						|
 | 
						|
	index++;
 | 
						|
	if (index == ktrace_nentries(ktp)) {
 | 
						|
		ktsp->ks_index = 0;
 | 
						|
	} else {
 | 
						|
		ktsp->ks_index = index;
 | 
						|
	}
 | 
						|
 | 
						|
	return ktep;
 | 
						|
}
 | 
						|
 | 
						|
/*
 | 
						|
 * ktrace_skip()
 | 
						|
 *
 | 
						|
 * Skip the next "count" entries and return the entry after that.
 | 
						|
 * Return NULL if this causes us to iterate past the beginning again.
 | 
						|
 */
 | 
						|
ktrace_entry_t *
 | 
						|
ktrace_skip(
 | 
						|
	ktrace_t        *ktp,
 | 
						|
	int             count,
 | 
						|
	ktrace_snap_t   *ktsp)
 | 
						|
{
 | 
						|
	int             index;
 | 
						|
	int             new_index;
 | 
						|
	ktrace_entry_t  *ktep;
 | 
						|
	int             nentries = ktrace_nentries(ktp);
 | 
						|
 | 
						|
	index = ktsp->ks_index;
 | 
						|
	new_index = index + count;
 | 
						|
	while (new_index >= nentries) {
 | 
						|
		new_index -= nentries;
 | 
						|
	}
 | 
						|
	if (index == ktsp->ks_start) {
 | 
						|
		/*
 | 
						|
		 * We've iterated around to the start, so we're done.
 | 
						|
		 */
 | 
						|
		ktep = NULL;
 | 
						|
	} else if ((new_index < index) && (index < ktsp->ks_index)) {
 | 
						|
		/*
 | 
						|
		 * We've skipped past the start again, so we're done.
 | 
						|
		 */
 | 
						|
		ktep = NULL;
 | 
						|
		ktsp->ks_index = ktsp->ks_start;
 | 
						|
	} else {
 | 
						|
		ktep = &(ktp->kt_entries[new_index]);
 | 
						|
		new_index++;
 | 
						|
		if (new_index == nentries) {
 | 
						|
			ktsp->ks_index = 0;
 | 
						|
		} else {
 | 
						|
			ktsp->ks_index = new_index;
 | 
						|
		}
 | 
						|
	}
 | 
						|
	return ktep;
 | 
						|
}
 |