rwl_owner          88 kern/kern_rwlock.c 	unsigned long owner = rwl->rwl_owner;
rwl_owner          91 kern/kern_rwlock.c 	    rw_cas(&rwl->rwl_owner, owner, owner + RWLOCK_READ_INCR)))
rwl_owner         100 kern/kern_rwlock.c 	if (__predict_false(rw_cas(&rwl->rwl_owner, 0,
rwl_owner         108 kern/kern_rwlock.c 	unsigned long owner = rwl->rwl_owner;
rwl_owner         111 kern/kern_rwlock.c 	    rw_cas(&rwl->rwl_owner, owner, owner - RWLOCK_READ_INCR)))
rwl_owner         118 kern/kern_rwlock.c 	unsigned long owner = rwl->rwl_owner;
rwl_owner         121 kern/kern_rwlock.c 	    rw_cas(&rwl->rwl_owner, owner, 0)))
rwl_owner         150 kern/kern_rwlock.c 		if (RW_PROC(curproc) == RW_PROC(rwl->rwl_owner))
rwl_owner         158 kern/kern_rwlock.c 		if ((rwl->rwl_owner & RWLOCK_WRLOCK) == 0)
rwl_owner         161 kern/kern_rwlock.c 		if (RW_PROC(curproc) != RW_PROC(rwl->rwl_owner))
rwl_owner         178 kern/kern_rwlock.c 	rwl->rwl_owner = 0;
rwl_owner         194 kern/kern_rwlock.c 	while (__predict_false(((o = rwl->rwl_owner) & op->check) != 0)) {
rwl_owner         207 kern/kern_rwlock.c 		do_sleep = !rw_cas(&rwl->rwl_owner, o, set);
rwl_owner         217 kern/kern_rwlock.c 	if (__predict_false(rw_cas(&rwl->rwl_owner, o, o + inc)))
rwl_owner         235 kern/kern_rwlock.c 	unsigned long owner = rwl->rwl_owner;
rwl_owner         240 kern/kern_rwlock.c 		owner = rwl->rwl_owner;
rwl_owner         246 kern/kern_rwlock.c 	} while (rw_cas(&rwl->rwl_owner, owner, set));
rwl_owner          75 sys/rwlock.h   	__volatile unsigned long rwl_owner;
rwl_owner          86 sys/rwlock.h   #define RWLOCK_OWNER(rwl)	((struct proc *)((rwl)->rwl_owner & ~RWLOCK_MASK))