alp               483 kern/kern_lock.c #define	SLOCK_WHERE(str, alp, id, l)					\
alp               487 kern/kern_lock.c 	lock_printf("lock: %p, currently at: %s:%d\n", (alp), (id), (l)); \
alp               489 kern/kern_lock.c 	if ((alp)->lock_file != NULL)					\
alp               490 kern/kern_lock.c 		lock_printf("last locked: %s:%d\n", (alp)->lock_file,	\
alp               491 kern/kern_lock.c 		    (alp)->lock_line);					\
alp               492 kern/kern_lock.c 	if ((alp)->unlock_file != NULL)					\
alp               493 kern/kern_lock.c 		lock_printf("last unlocked: %s:%d\n", (alp)->unlock_file, \
alp               494 kern/kern_lock.c 		    (alp)->unlock_line);				\
alp               508 kern/kern_lock.c 	__cpu_simple_lock_init(&alp->lock_data);
alp               510 kern/kern_lock.c 	alp->lock_data = __SIMPLELOCK_UNLOCKED;
alp               512 kern/kern_lock.c 	alp->lock_file = NULL;
alp               513 kern/kern_lock.c 	alp->lock_line = 0;
alp               514 kern/kern_lock.c 	alp->unlock_file = NULL;
alp               515 kern/kern_lock.c 	alp->unlock_line = 0;
alp               516 kern/kern_lock.c 	alp->lock_holder = LK_NOCPU;
alp               531 kern/kern_lock.c 	if (alp->lock_data == __SIMPLELOCK_LOCKED) {
alp               533 kern/kern_lock.c 		if (alp->lock_holder == cpu_id) {
alp               535 kern/kern_lock.c 			    alp, id, l);
alp               539 kern/kern_lock.c 		SLOCK_WHERE("simple_lock: lock held\n", alp, id, l);
alp               547 kern/kern_lock.c 	__cpu_simple_lock(&alp->lock_data);
alp               550 kern/kern_lock.c 	alp->lock_data = __SIMPLELOCK_LOCKED;
alp               553 kern/kern_lock.c 	if (alp->lock_holder != LK_NOCPU) {
alp               555 kern/kern_lock.c 		    alp, id, l);
alp               557 kern/kern_lock.c 	alp->lock_file = id;
alp               558 kern/kern_lock.c 	alp->lock_line = l;
alp               559 kern/kern_lock.c 	alp->lock_holder = cpu_id;
alp               563 kern/kern_lock.c 	TAILQ_INSERT_TAIL(&simplelock_list, (struct simplelock *)alp, list);
alp               573 kern/kern_lock.c _simple_lock_held(__volatile struct simplelock *alp)
alp               581 kern/kern_lock.c 	if (__cpu_simple_lock_try(&alp->lock_data) == 0)
alp               582 kern/kern_lock.c 		locked = (alp->lock_holder == cpu_id);
alp               584 kern/kern_lock.c 		__cpu_simple_unlock(&alp->lock_data);
alp               586 kern/kern_lock.c 	if (alp->lock_data == __SIMPLELOCK_LOCKED) {
alp               588 kern/kern_lock.c 		KASSERT(alp->lock_holder == cpu_id);
alp               610 kern/kern_lock.c 	if ((rv = __cpu_simple_lock_try(&alp->lock_data)) == 0) {
alp               611 kern/kern_lock.c 		if (alp->lock_holder == cpu_id)
alp               613 kern/kern_lock.c 			    alp, id, l);
alp               617 kern/kern_lock.c 	if (alp->lock_data == __SIMPLELOCK_LOCKED) {
alp               618 kern/kern_lock.c 		SLOCK_WHERE("simple_lock_try: lock held\n", alp, id, l);
alp               621 kern/kern_lock.c 	alp->lock_data = __SIMPLELOCK_LOCKED;
alp               630 kern/kern_lock.c 	alp->lock_file = id;
alp               631 kern/kern_lock.c 	alp->lock_line = l;
alp               632 kern/kern_lock.c 	alp->lock_holder = cpu_id;
alp               636 kern/kern_lock.c 	TAILQ_INSERT_TAIL(&simplelock_list, (struct simplelock *)alp, list);
alp               657 kern/kern_lock.c 	if (alp->lock_data == __SIMPLELOCK_UNLOCKED) {
alp               659 kern/kern_lock.c 		    alp, id, l);
alp               664 kern/kern_lock.c 	TAILQ_REMOVE(&simplelock_list, alp, list);
alp               669 kern/kern_lock.c 	alp->list.tqe_next = NULL;	/* sanity */
alp               670 kern/kern_lock.c 	alp->list.tqe_prev = NULL;	/* sanity */
alp               672 kern/kern_lock.c 	alp->unlock_file = id;
alp               673 kern/kern_lock.c 	alp->unlock_line = l;
alp               676 kern/kern_lock.c 	alp->lock_holder = LK_NOCPU;
alp               678 kern/kern_lock.c 	__cpu_simple_unlock(&alp->lock_data);
alp               680 kern/kern_lock.c 	alp->lock_data = __SIMPLELOCK_UNLOCKED;
alp               681 kern/kern_lock.c 	KASSERT(alp->lock_holder == CPU_NUMBER());
alp               682 kern/kern_lock.c 	alp->lock_holder = LK_NOCPU;
alp               692 kern/kern_lock.c 	struct simplelock *alp;
alp               698 kern/kern_lock.c 	TAILQ_FOREACH(alp, &simplelock_list, list) {
alp               699 kern/kern_lock.c 		lock_printf("%p CPU %lu %s:%d\n", alp, alp->lock_holder,
alp               700 kern/kern_lock.c 		    alp->lock_file, alp->lock_line);
alp               709 kern/kern_lock.c 	struct simplelock *alp;
alp               714 kern/kern_lock.c 	TAILQ_FOREACH(alp, &simplelock_list, list) {
alp               715 kern/kern_lock.c 		if ((void *)alp >= start && (void *)alp < end) {
alp               717 kern/kern_lock.c 			    alp, alp->lock_holder, alp->lock_file,
alp               718 kern/kern_lock.c 			    alp->lock_line);
alp               742 kern/kern_lock.c 	struct simplelock *alp;
alp               751 kern/kern_lock.c 	TAILQ_FOREACH(alp, &simplelock_list, list) {
alp               752 kern/kern_lock.c 		if (alp == lp)
alp               754 kern/kern_lock.c 		if (alp->lock_holder == cpu_id)
alp               760 kern/kern_lock.c 	if (alp != NULL) {
alp               763 kern/kern_lock.c 		    where, alp, alp->lock_holder, alp->lock_file,
alp               764 kern/kern_lock.c 		    alp->lock_line);