rwl_owner 88 kern/kern_rwlock.c unsigned long owner = rwl->rwl_owner; rwl_owner 91 kern/kern_rwlock.c rw_cas(&rwl->rwl_owner, owner, owner + RWLOCK_READ_INCR))) rwl_owner 100 kern/kern_rwlock.c if (__predict_false(rw_cas(&rwl->rwl_owner, 0, rwl_owner 108 kern/kern_rwlock.c unsigned long owner = rwl->rwl_owner; rwl_owner 111 kern/kern_rwlock.c rw_cas(&rwl->rwl_owner, owner, owner - RWLOCK_READ_INCR))) rwl_owner 118 kern/kern_rwlock.c unsigned long owner = rwl->rwl_owner; rwl_owner 121 kern/kern_rwlock.c rw_cas(&rwl->rwl_owner, owner, 0))) rwl_owner 150 kern/kern_rwlock.c if (RW_PROC(curproc) == RW_PROC(rwl->rwl_owner)) rwl_owner 158 kern/kern_rwlock.c if ((rwl->rwl_owner & RWLOCK_WRLOCK) == 0) rwl_owner 161 kern/kern_rwlock.c if (RW_PROC(curproc) != RW_PROC(rwl->rwl_owner)) rwl_owner 178 kern/kern_rwlock.c rwl->rwl_owner = 0; rwl_owner 194 kern/kern_rwlock.c while (__predict_false(((o = rwl->rwl_owner) & op->check) != 0)) { rwl_owner 207 kern/kern_rwlock.c do_sleep = !rw_cas(&rwl->rwl_owner, o, set); rwl_owner 217 kern/kern_rwlock.c if (__predict_false(rw_cas(&rwl->rwl_owner, o, o + inc))) rwl_owner 235 kern/kern_rwlock.c unsigned long owner = rwl->rwl_owner; rwl_owner 240 kern/kern_rwlock.c owner = rwl->rwl_owner; rwl_owner 246 kern/kern_rwlock.c } while (rw_cas(&rwl->rwl_owner, owner, set)); rwl_owner 75 sys/rwlock.h __volatile unsigned long rwl_owner; rwl_owner 86 sys/rwlock.h #define RWLOCK_OWNER(rwl) ((struct proc *)((rwl)->rwl_owner & ~RWLOCK_MASK))