rwl_owner 88 kern/kern_rwlock.c unsigned long owner = rwl->rwl_owner;
rwl_owner 91 kern/kern_rwlock.c rw_cas(&rwl->rwl_owner, owner, owner + RWLOCK_READ_INCR)))
rwl_owner 100 kern/kern_rwlock.c if (__predict_false(rw_cas(&rwl->rwl_owner, 0,
rwl_owner 108 kern/kern_rwlock.c unsigned long owner = rwl->rwl_owner;
rwl_owner 111 kern/kern_rwlock.c rw_cas(&rwl->rwl_owner, owner, owner - RWLOCK_READ_INCR)))
rwl_owner 118 kern/kern_rwlock.c unsigned long owner = rwl->rwl_owner;
rwl_owner 121 kern/kern_rwlock.c rw_cas(&rwl->rwl_owner, owner, 0)))
rwl_owner 150 kern/kern_rwlock.c if (RW_PROC(curproc) == RW_PROC(rwl->rwl_owner))
rwl_owner 158 kern/kern_rwlock.c if ((rwl->rwl_owner & RWLOCK_WRLOCK) == 0)
rwl_owner 161 kern/kern_rwlock.c if (RW_PROC(curproc) != RW_PROC(rwl->rwl_owner))
rwl_owner 178 kern/kern_rwlock.c rwl->rwl_owner = 0;
rwl_owner 194 kern/kern_rwlock.c while (__predict_false(((o = rwl->rwl_owner) & op->check) != 0)) {
rwl_owner 207 kern/kern_rwlock.c do_sleep = !rw_cas(&rwl->rwl_owner, o, set);
rwl_owner 217 kern/kern_rwlock.c if (__predict_false(rw_cas(&rwl->rwl_owner, o, o + inc)))
rwl_owner 235 kern/kern_rwlock.c unsigned long owner = rwl->rwl_owner;
rwl_owner 240 kern/kern_rwlock.c owner = rwl->rwl_owner;
rwl_owner 246 kern/kern_rwlock.c } while (rw_cas(&rwl->rwl_owner, owner, set));
rwl_owner 75 sys/rwlock.h __volatile unsigned long rwl_owner;
rwl_owner 86 sys/rwlock.h #define RWLOCK_OWNER(rwl) ((struct proc *)((rwl)->rwl_owner & ~RWLOCK_MASK))