bucket 150 dev/raidframe/rf_debugMem.c unsigned long bucket = HASHADDR(addr); bucket 156 dev/raidframe/rf_debugMem.c for (p = mh_table[bucket]; p && (p->address != addr); p = p->next); bucket 160 dev/raidframe/rf_debugMem.c p->next = mh_table[bucket]; bucket 161 dev/raidframe/rf_debugMem.c mh_table[bucket] = p; bucket 182 dev/raidframe/rf_debugMem.c unsigned long bucket = HASHADDR(addr); bucket 186 dev/raidframe/rf_debugMem.c for (p = mh_table[bucket]; p && (p->address != addr); p = p->next); bucket 77 kern/kern_malloc.c struct kmembuckets bucket[MINBUCKET + 16]; bucket 178 kern/kern_malloc.c kbp = &bucket[indx]; bucket 368 kern/kern_malloc.c kbp = &bucket[kup->ku_indx]; bucket 530 kern/kern_malloc.c bucket[indx].kb_elmpercl = 1; bucket 532 kern/kern_malloc.c bucket[indx].kb_elmpercl = PAGE_SIZE / (1 << indx); bucket 533 kern/kern_malloc.c bucket[indx].kb_highwat = 5 * bucket[indx].kb_elmpercl; bucket 576 kern/kern_malloc.c bcopy(&bucket[BUCKETINDX(name[1])], &kb, sizeof(kb)); bucket 269 kern/kern_timeout.c db_show_callout_bucket(struct circq *bucket) bucket 276 kern/kern_timeout.c for (p = CIRCQ_FIRST(bucket); p != bucket; p = CIRCQ_FIRST(p)) { bucket 281 kern/kern_timeout.c (bucket - timeout_wheel) / WHEELSIZE, bucket 282 kern/kern_timeout.c bucket - timeout_wheel, to->to_arg, name); bucket 404 sys/malloc.h struct kmembuckets *kbp = &bucket[BUCKETINDX(kbp_size)]; \ bucket 422 sys/malloc.h kbp = &bucket[kup->ku_indx]; \ bucket 437 sys/malloc.h extern struct kmembuckets bucket[]; bucket 1412 uvm/uvm_aobj.c int bucket; bucket 1415 uvm/uvm_aobj.c for (bucket = aobj->u_swhashmask; bucket >= 0; bucket--) { bucket 1416 uvm/uvm_aobj.c for (elt = LIST_FIRST(&aobj->u_swhash[bucket]);