| /* |
| * Copyright (c) 2000-2003,2005 Silicon Graphics, Inc. |
| * All Rights Reserved. |
| * |
| * This program is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU General Public License as |
| * published by the Free Software Foundation. |
| * |
| * This program is distributed in the hope that it would be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| * GNU General Public License for more details. |
| * |
| * You should have received a copy of the GNU General Public License |
| * along with this program; if not, write the Free Software Foundation, |
| * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA |
| */ |
| #include <xfs.h> |
| |
| static kmem_zone_t *ktrace_hdr_zone; |
| static kmem_zone_t *ktrace_ent_zone; |
| static int ktrace_zentries; |
| |
| void __init |
| ktrace_init(int zentries) |
| { |
| ktrace_zentries = roundup_pow_of_two(zentries); |
| |
| ktrace_hdr_zone = kmem_zone_init(sizeof(ktrace_t), |
| "ktrace_hdr"); |
| ASSERT(ktrace_hdr_zone); |
| |
| ktrace_ent_zone = kmem_zone_init(ktrace_zentries |
| * sizeof(ktrace_entry_t), |
| "ktrace_ent"); |
| ASSERT(ktrace_ent_zone); |
| } |
| |
| void __exit |
| ktrace_uninit(void) |
| { |
| kmem_zone_destroy(ktrace_hdr_zone); |
| kmem_zone_destroy(ktrace_ent_zone); |
| } |
| |
| /* |
| * ktrace_alloc() |
| * |
| * Allocate a ktrace header and enough buffering for the given |
| * number of entries. Round the number of entries up to a |
| * power of 2 so we can do fast masking to get the index from |
| * the atomic index counter. |
| */ |
| ktrace_t * |
| ktrace_alloc(int nentries, unsigned int __nocast sleep) |
| { |
| ktrace_t *ktp; |
| ktrace_entry_t *ktep; |
| int entries; |
| |
| ktp = (ktrace_t*)kmem_zone_alloc(ktrace_hdr_zone, sleep); |
| |
| if (ktp == (ktrace_t*)NULL) { |
| /* |
| * KM_SLEEP callers don't expect failure. |
| */ |
| if (sleep & KM_SLEEP) |
| panic("ktrace_alloc: NULL memory on KM_SLEEP request!"); |
| |
| return NULL; |
| } |
| |
| /* |
| * Special treatment for buffers with the ktrace_zentries entries |
| */ |
| entries = roundup_pow_of_two(nentries); |
| if (entries == ktrace_zentries) { |
| ktep = (ktrace_entry_t*)kmem_zone_zalloc(ktrace_ent_zone, |
| sleep); |
| } else { |
| ktep = (ktrace_entry_t*)kmem_zalloc((entries * sizeof(*ktep)), |
| sleep | KM_LARGE); |
| } |
| |
| if (ktep == NULL) { |
| /* |
| * KM_SLEEP callers don't expect failure. |
| */ |
| if (sleep & KM_SLEEP) |
| panic("ktrace_alloc: NULL memory on KM_SLEEP request!"); |
| |
| kmem_free(ktp); |
| |
| return NULL; |
| } |
| |
| ktp->kt_entries = ktep; |
| ktp->kt_nentries = entries; |
| ASSERT(is_power_of_2(entries)); |
| ktp->kt_index_mask = entries - 1; |
| atomic_set(&ktp->kt_index, 0); |
| ktp->kt_rollover = 0; |
| return ktp; |
| } |
| |
| |
| /* |
| * ktrace_free() |
| * |
| * Free up the ktrace header and buffer. It is up to the caller |
| * to ensure that no-one is referencing it. |
| */ |
| void |
| ktrace_free(ktrace_t *ktp) |
| { |
| if (ktp == (ktrace_t *)NULL) |
| return; |
| |
| /* |
| * Special treatment for the Vnode trace buffer. |
| */ |
| if (ktp->kt_nentries == ktrace_zentries) |
| kmem_zone_free(ktrace_ent_zone, ktp->kt_entries); |
| else |
| kmem_free(ktp->kt_entries); |
| |
| kmem_zone_free(ktrace_hdr_zone, ktp); |
| } |
| |
| |
| /* |
| * Enter the given values into the "next" entry in the trace buffer. |
| * kt_index is always the index of the next entry to be filled. |
| */ |
| void |
| ktrace_enter( |
| ktrace_t *ktp, |
| void *val0, |
| void *val1, |
| void *val2, |
| void *val3, |
| void *val4, |
| void *val5, |
| void *val6, |
| void *val7, |
| void *val8, |
| void *val9, |
| void *val10, |
| void *val11, |
| void *val12, |
| void *val13, |
| void *val14, |
| void *val15) |
| { |
| int index; |
| ktrace_entry_t *ktep; |
| |
| ASSERT(ktp != NULL); |
| |
| /* |
| * Grab an entry by pushing the index up to the next one. |
| */ |
| index = atomic_add_return(1, &ktp->kt_index); |
| index = (index - 1) & ktp->kt_index_mask; |
| if (!ktp->kt_rollover && index == ktp->kt_nentries - 1) |
| ktp->kt_rollover = 1; |
| |
| ASSERT((index >= 0) && (index < ktp->kt_nentries)); |
| |
| ktep = &(ktp->kt_entries[index]); |
| |
| ktep->val[0] = val0; |
| ktep->val[1] = val1; |
| ktep->val[2] = val2; |
| ktep->val[3] = val3; |
| ktep->val[4] = val4; |
| ktep->val[5] = val5; |
| ktep->val[6] = val6; |
| ktep->val[7] = val7; |
| ktep->val[8] = val8; |
| ktep->val[9] = val9; |
| ktep->val[10] = val10; |
| ktep->val[11] = val11; |
| ktep->val[12] = val12; |
| ktep->val[13] = val13; |
| ktep->val[14] = val14; |
| ktep->val[15] = val15; |
| } |
| |
| /* |
| * Return the number of entries in the trace buffer. |
| */ |
| int |
| ktrace_nentries( |
| ktrace_t *ktp) |
| { |
| int index; |
| if (ktp == NULL) |
| return 0; |
| |
| index = atomic_read(&ktp->kt_index) & ktp->kt_index_mask; |
| return (ktp->kt_rollover ? ktp->kt_nentries : index); |
| } |
| |
| /* |
| * ktrace_first() |
| * |
| * This is used to find the start of the trace buffer. |
| * In conjunction with ktrace_next() it can be used to |
| * iterate through the entire trace buffer. This code does |
| * not do any locking because it is assumed that it is called |
| * from the debugger. |
| * |
| * The caller must pass in a pointer to a ktrace_snap |
| * structure in which we will keep some state used to |
| * iterate through the buffer. This state must not touched |
| * by any code outside of this module. |
| */ |
| ktrace_entry_t * |
| ktrace_first(ktrace_t *ktp, ktrace_snap_t *ktsp) |
| { |
| ktrace_entry_t *ktep; |
| int index; |
| int nentries; |
| |
| if (ktp->kt_rollover) |
| index = atomic_read(&ktp->kt_index) & ktp->kt_index_mask; |
| else |
| index = 0; |
| |
| ktsp->ks_start = index; |
| ktep = &(ktp->kt_entries[index]); |
| |
| nentries = ktrace_nentries(ktp); |
| index++; |
| if (index < nentries) { |
| ktsp->ks_index = index; |
| } else { |
| ktsp->ks_index = 0; |
| if (index > nentries) |
| ktep = NULL; |
| } |
| return ktep; |
| } |
| |
| /* |
| * ktrace_next() |
| * |
| * This is used to iterate through the entries of the given |
| * trace buffer. The caller must pass in the ktrace_snap_t |
| * structure initialized by ktrace_first(). The return value |
| * will be either a pointer to the next ktrace_entry or NULL |
| * if all of the entries have been traversed. |
| */ |
| ktrace_entry_t * |
| ktrace_next( |
| ktrace_t *ktp, |
| ktrace_snap_t *ktsp) |
| { |
| int index; |
| ktrace_entry_t *ktep; |
| |
| index = ktsp->ks_index; |
| if (index == ktsp->ks_start) { |
| ktep = NULL; |
| } else { |
| ktep = &ktp->kt_entries[index]; |
| } |
| |
| index++; |
| if (index == ktrace_nentries(ktp)) { |
| ktsp->ks_index = 0; |
| } else { |
| ktsp->ks_index = index; |
| } |
| |
| return ktep; |
| } |
| |
| /* |
| * ktrace_skip() |
| * |
| * Skip the next "count" entries and return the entry after that. |
| * Return NULL if this causes us to iterate past the beginning again. |
| */ |
| ktrace_entry_t * |
| ktrace_skip( |
| ktrace_t *ktp, |
| int count, |
| ktrace_snap_t *ktsp) |
| { |
| int index; |
| int new_index; |
| ktrace_entry_t *ktep; |
| int nentries = ktrace_nentries(ktp); |
| |
| index = ktsp->ks_index; |
| new_index = index + count; |
| while (new_index >= nentries) { |
| new_index -= nentries; |
| } |
| if (index == ktsp->ks_start) { |
| /* |
| * We've iterated around to the start, so we're done. |
| */ |
| ktep = NULL; |
| } else if ((new_index < index) && (index < ktsp->ks_index)) { |
| /* |
| * We've skipped past the start again, so we're done. |
| */ |
| ktep = NULL; |
| ktsp->ks_index = ktsp->ks_start; |
| } else { |
| ktep = &(ktp->kt_entries[new_index]); |
| new_index++; |
| if (new_index == nentries) { |
| ktsp->ks_index = 0; |
| } else { |
| ktsp->ks_index = new_index; |
| } |
| } |
| return ktep; |
| } |