ufi 1740 dev/usb/umass.c sc->sbl.ufi.asc, sc->sbl.ufi.ascq));
ufi 1742 dev/usb/umass.c if ((sc->sbl.ufi.asc == 0 && sc->sbl.ufi.ascq == 0) ||
ufi 125 dev/usb/umassvar.h } ufi;
ufi 374 uvm/uvm_device.c udv_fault(ufi, vaddr, pps, npages, centeridx, fault_type, access_type, flags)
ufi 375 uvm/uvm_device.c struct uvm_faultinfo *ufi;
ufi 382 uvm/uvm_device.c struct vm_map_entry *entry = ufi->entry;
ufi 403 uvm/uvm_device.c uvmfault_unlockall(ufi, ufi->entry->aref.ar_amap, uobj, NULL);
ufi 445 uvm/uvm_device.c mapprot = ufi->entry->protection;
ufi 448 uvm/uvm_device.c ufi->orig_map->pmap, curr_va, (u_long)paddr, mapprot);
ufi 449 uvm/uvm_device.c if (pmap_enter(ufi->orig_map->pmap, curr_va, paddr,
ufi 461 uvm/uvm_device.c uvmfault_unlockall(ufi, ufi->entry->aref.ar_amap,
ufi 465 uvm/uvm_device.c pmap_update(ufi->orig_map->pmap);
ufi 471 uvm/uvm_device.c uvmfault_unlockall(ufi, ufi->entry->aref.ar_amap, uobj, NULL);
ufi 472 uvm/uvm_device.c pmap_update(ufi->orig_map->pmap);
ufi 233 uvm/uvm_fault.c uvmfault_amapcopy(ufi)
ufi 234 uvm/uvm_fault.c struct uvm_faultinfo *ufi;
ufi 247 uvm/uvm_fault.c if (uvmfault_lookup(ufi, TRUE) == FALSE)
ufi 254 uvm/uvm_fault.c if (UVM_ET_ISNEEDSCOPY(ufi->entry))
ufi 255 uvm/uvm_fault.c amap_copy(ufi->map, ufi->entry, M_NOWAIT, TRUE,
ufi 256 uvm/uvm_fault.c ufi->orig_rvaddr, ufi->orig_rvaddr + 1);
ufi 262 uvm/uvm_fault.c if (UVM_ET_ISNEEDSCOPY(ufi->entry)) {
ufi 263 uvm/uvm_fault.c uvmfault_unlockmaps(ufi, TRUE);
ufi 272 uvm/uvm_fault.c uvmfault_unlockmaps(ufi, TRUE);
ufi 295 uvm/uvm_fault.c uvmfault_anonget(ufi, amap, anon)
ufi 296 uvm/uvm_fault.c struct uvm_faultinfo *ufi;
ufi 357 uvm/uvm_fault.c uvmfault_unlockall(ufi, amap, NULL, anon);
ufi 365 uvm/uvm_fault.c uvmfault_unlockall(ufi, amap, NULL, NULL);
ufi 382 uvm/uvm_fault.c uvmfault_unlockall(ufi, amap, NULL, anon);
ufi 393 uvm/uvm_fault.c uvmfault_unlockall(ufi, amap, NULL, anon);
ufi 418 uvm/uvm_fault.c locked = uvmfault_relock(ufi);
ufi 454 uvm/uvm_fault.c uvmfault_unlockall(ufi, amap, NULL,
ufi 486 uvm/uvm_fault.c uvmfault_unlockall(ufi, amap, NULL,
ufi 519 uvm/uvm_fault.c if (ufi != NULL &&
ufi 520 uvm/uvm_fault.c amap_lookup(&ufi->entry->aref,
ufi 521 uvm/uvm_fault.c ufi->orig_rvaddr - ufi->entry->start) != anon) {
ufi 523 uvm/uvm_fault.c uvmfault_unlockall(ufi, amap, NULL, anon);
ufi 565 uvm/uvm_fault.c struct uvm_faultinfo ufi;
ufi 590 uvm/uvm_fault.c ufi.orig_map = orig_map;
ufi 591 uvm/uvm_fault.c ufi.orig_rvaddr = trunc_page(vaddr);
ufi 592 uvm/uvm_fault.c ufi.orig_size = PAGE_SIZE; /* can't get any smaller than this */
ufi 608 uvm/uvm_fault.c if (uvmfault_lookup(&ufi, FALSE) == FALSE) {
ufi 615 uvm/uvm_fault.c if ((ufi.map->flags & VM_MAP_PAGEABLE) == 0)
ufi 617 uvm/uvm_fault.c ufi.map, vaddr);
ufi 624 uvm/uvm_fault.c if ((ufi.entry->protection & access_type) != access_type) {
ufi 627 uvm/uvm_fault.c ufi.entry->protection, access_type, 0, 0);
ufi 628 uvm/uvm_fault.c uvmfault_unlockmaps(&ufi, FALSE);
ufi 639 uvm/uvm_fault.c enter_prot = ufi.entry->protection;
ufi 640 uvm/uvm_fault.c wired = VM_MAPENT_ISWIRED(ufi.entry) || (fault_type == VM_FAULT_WIRE);
ufi 651 uvm/uvm_fault.c if (UVM_ET_ISNEEDSCOPY(ufi.entry)) {
ufi 653 uvm/uvm_fault.c (ufi.entry->object.uvm_obj == NULL)) {
ufi 657 uvm/uvm_fault.c uvmfault_unlockmaps(&ufi, FALSE);
ufi 658 uvm/uvm_fault.c uvmfault_amapcopy(&ufi);
ufi 677 uvm/uvm_fault.c amap = ufi.entry->aref.ar_amap; /* top layer */
ufi 678 uvm/uvm_fault.c uobj = ufi.entry->object.uvm_obj; /* bottom layer */
ufi 686 uvm/uvm_fault.c uvmfault_unlockmaps(&ufi, FALSE);
ufi 701 uvm/uvm_fault.c KASSERT(uvmadvice[ufi.entry->advice].advice ==
ufi 702 uvm/uvm_fault.c ufi.entry->advice);
ufi 703 uvm/uvm_fault.c nback = min(uvmadvice[ufi.entry->advice].nback,
ufi 704 uvm/uvm_fault.c (ufi.orig_rvaddr - ufi.entry->start) >> PAGE_SHIFT);
ufi 705 uvm/uvm_fault.c startva = ufi.orig_rvaddr - (nback << PAGE_SHIFT);
ufi 706 uvm/uvm_fault.c nforw = min(uvmadvice[ufi.entry->advice].nforw,
ufi 707 uvm/uvm_fault.c ((ufi.entry->end - ufi.orig_rvaddr) >>
ufi 722 uvm/uvm_fault.c startva = ufi.orig_rvaddr;
ufi 731 uvm/uvm_fault.c UVMHIST_LOG(maphist, " entry=%p, amap=%p, obj=%p", ufi.entry,
ufi 740 uvm/uvm_fault.c amap_lookups(&ufi.entry->aref, startva - ufi.entry->start,
ufi 753 uvm/uvm_fault.c if (ufi.entry->advice == MADV_SEQUENTIAL && nback != 0) {
ufi 763 uvm/uvm_fault.c uoff = (startva - ufi.entry->start) + ufi.entry->offset;
ufi 795 uvm/uvm_fault.c pmap_extract(ufi.orig_map->pmap, currva, &pa)) {
ufi 827 uvm/uvm_fault.c ufi.orig_map->pmap, currva, anon->an_page, 0);
ufi 836 uvm/uvm_fault.c (void) pmap_enter(ufi.orig_map->pmap, currva,
ufi 841 uvm/uvm_fault.c (VM_MAPENT_ISWIRED(ufi.entry) ? PMAP_WIRED : 0));
ufi 844 uvm/uvm_fault.c pmap_update(ufi.orig_map->pmap);
ufi 872 uvm/uvm_fault.c result = uobj->pgops->pgo_fault(&ufi, startva, pages, npages,
ufi 906 uvm/uvm_fault.c (void) uobj->pgops->pgo_get(uobj, ufi.entry->offset +
ufi 907 uvm/uvm_fault.c (startva - ufi.entry->start),
ufi 909 uvm/uvm_fault.c access_type & MASK(ufi.entry),
ufi 910 uvm/uvm_fault.c ufi.entry->advice, PGO_LOCKED);
ufi 959 uvm/uvm_fault.c ufi.orig_map->pmap, currva, pages[lcv], 0);
ufi 969 uvm/uvm_fault.c (void) pmap_enter(ufi.orig_map->pmap, currva,
ufi 971 uvm/uvm_fault.c enter_prot & MASK(ufi.entry),
ufi 985 uvm/uvm_fault.c pmap_update(ufi.orig_map->pmap);
ufi 1044 uvm/uvm_fault.c result = uvmfault_anonget(&ufi, amap, anon);
ufi 1111 uvm/uvm_fault.c uvmfault_unlockall(&ufi, amap, uobj,
ufi 1182 uvm/uvm_fault.c uvmfault_unlockall(&ufi, amap, uobj, oanon);
ufi 1202 uvm/uvm_fault.c amap_add(&ufi.entry->aref, ufi.orig_rvaddr - ufi.entry->start,
ufi 1234 uvm/uvm_fault.c ufi.orig_map->pmap, ufi.orig_rvaddr, pg, 0);
ufi 1235 uvm/uvm_fault.c if (pmap_enter(ufi.orig_map->pmap, ufi.orig_rvaddr, VM_PAGE_TO_PHYS(pg),
ufi 1245 uvm/uvm_fault.c uvmfault_unlockall(&ufi, amap, uobj, oanon);
ufi 1286 uvm/uvm_fault.c uvmfault_unlockall(&ufi, amap, uobj, oanon);
ufi 1287 uvm/uvm_fault.c pmap_update(ufi.orig_map->pmap);
ufi 1314 uvm/uvm_fault.c UVM_ET_ISCOPYONWRITE(ufi.entry);
ufi 1336 uvm/uvm_fault.c uvmfault_unlockall(&ufi, amap, NULL, NULL);
ufi 1341 uvm/uvm_fault.c uoff = (ufi.orig_rvaddr - ufi.entry->start) + ufi.entry->offset;
ufi 1343 uvm/uvm_fault.c 0, access_type & MASK(ufi.entry), ufi.entry->advice,
ufi 1374 uvm/uvm_fault.c locked = uvmfault_relock(&ufi);
ufi 1388 uvm/uvm_fault.c amap_lookup(&ufi.entry->aref,
ufi 1389 uvm/uvm_fault.c ufi.orig_rvaddr - ufi.entry->start))) {
ufi 1391 uvm/uvm_fault.c uvmfault_unlockall(&ufi, amap, NULL, NULL);
ufi 1466 uvm/uvm_fault.c if (UVM_ET_ISCOPYONWRITE(ufi.entry))
ufi 1505 uvm/uvm_fault.c uvmfault_unlockall(&ufi, amap, uobj,
ufi 1602 uvm/uvm_fault.c uvmfault_unlockall(&ufi, amap, uobj, NULL);
ufi 1668 uvm/uvm_fault.c amap_add(&ufi.entry->aref, ufi.orig_rvaddr - ufi.entry->start,
ufi 1686 uvm/uvm_fault.c ufi.orig_map->pmap, ufi.orig_rvaddr, pg, promote);
ufi 1687 uvm/uvm_fault.c if (pmap_enter(ufi.orig_map->pmap, ufi.orig_rvaddr, VM_PAGE_TO_PHYS(pg),
ufi 1709 uvm/uvm_fault.c uvmfault_unlockall(&ufi, amap, uobj, NULL);
ufi 1753 uvm/uvm_fault.c uvmfault_unlockall(&ufi, amap, uobj, NULL);
ufi 1754 uvm/uvm_fault.c pmap_update(ufi.orig_map->pmap);
ufi 55 uvm/uvm_fault_i.h uvmfault_unlockmaps(ufi, write_locked)
ufi 56 uvm/uvm_fault_i.h struct uvm_faultinfo *ufi;
ufi 64 uvm/uvm_fault_i.h if (ufi == NULL) {
ufi 69 uvm/uvm_fault_i.h vm_map_unlock(ufi->map);
ufi 71 uvm/uvm_fault_i.h vm_map_unlock_read(ufi->map);
ufi 82 uvm/uvm_fault_i.h uvmfault_unlockall(ufi, amap, uobj, anon)
ufi 83 uvm/uvm_fault_i.h struct uvm_faultinfo *ufi;
ufi 93 uvm/uvm_fault_i.h uvmfault_unlockmaps(ufi, FALSE);
ufi 112 uvm/uvm_fault_i.h uvmfault_lookup(ufi, write_lock)
ufi 113 uvm/uvm_fault_i.h struct uvm_faultinfo *ufi;
ufi 122 uvm/uvm_fault_i.h ufi->map = ufi->orig_map;
ufi 123 uvm/uvm_fault_i.h ufi->size = ufi->orig_size;
ufi 136 uvm/uvm_fault_i.h vm_map_lock(ufi->map);
ufi 138 uvm/uvm_fault_i.h vm_map_lock_read(ufi->map);
ufi 144 uvm/uvm_fault_i.h if (!uvm_map_lookup_entry(ufi->map, ufi->orig_rvaddr,
ufi 145 uvm/uvm_fault_i.h &ufi->entry)) {
ufi 146 uvm/uvm_fault_i.h uvmfault_unlockmaps(ufi, write_lock);
ufi 153 uvm/uvm_fault_i.h if (ufi->entry->end - ufi->orig_rvaddr < ufi->size)
ufi 154 uvm/uvm_fault_i.h ufi->size = ufi->entry->end - ufi->orig_rvaddr;
ufi 160 uvm/uvm_fault_i.h if (UVM_ET_ISSUBMAP(ufi->entry)) {
ufi 161 uvm/uvm_fault_i.h tmpmap = ufi->entry->object.sub_map;
ufi 163 uvm/uvm_fault_i.h vm_map_unlock(ufi->map);
ufi 165 uvm/uvm_fault_i.h vm_map_unlock_read(ufi->map);
ufi 167 uvm/uvm_fault_i.h ufi->map = tmpmap;
ufi 175 uvm/uvm_fault_i.h ufi->mapv = ufi->map->timestamp;
ufi 191 uvm/uvm_fault_i.h uvmfault_relock(ufi)
ufi 192 uvm/uvm_fault_i.h struct uvm_faultinfo *ufi;
ufi 199 uvm/uvm_fault_i.h if (ufi == NULL) {
ufi 210 uvm/uvm_fault_i.h vm_map_lock_read(ufi->map);
ufi 211 uvm/uvm_fault_i.h if (ufi->mapv != ufi->map->timestamp) {
ufi 212 uvm/uvm_fault_i.h vm_map_unlock_read(ufi->map);
ufi 130 uvm/uvm_loan.c uvm_loanentry(ufi, output, flags)
ufi 131 uvm/uvm_loan.c struct uvm_faultinfo *ufi;
ufi 135 uvm/uvm_loan.c vaddr_t curaddr = ufi->orig_rvaddr;
ufi 136 uvm/uvm_loan.c vsize_t togo = ufi->size;
ufi 137 uvm/uvm_loan.c struct vm_aref *aref = &ufi->entry->aref;
ufi 138 uvm/uvm_loan.c struct uvm_object *uobj = ufi->entry->object.uvm_obj;
ufi 158 uvm/uvm_loan.c anon = amap_lookup(aref, curaddr - ufi->entry->start);
ufi 164 uvm/uvm_loan.c rv = uvm_loananon(ufi, output, flags, anon);
ufi 166 uvm/uvm_loan.c rv = uvm_loanuobj(ufi, output, flags, curaddr);
ufi 167 uvm/uvm_loan.c } else if (UVM_ET_ISCOPYONWRITE(ufi->entry)) {
ufi 168 uvm/uvm_loan.c rv = uvm_loanzero(ufi, output, flags);
ufi 192 uvm/uvm_loan.c uvmfault_unlockall(ufi, aref->ar_amap, uobj, NULL);
ufi 219 uvm/uvm_loan.c struct uvm_faultinfo ufi;
ufi 254 uvm/uvm_loan.c ufi.orig_map = map;
ufi 255 uvm/uvm_loan.c ufi.orig_rvaddr = start;
ufi 256 uvm/uvm_loan.c ufi.orig_size = len;
ufi 263 uvm/uvm_loan.c if (!uvmfault_lookup(&ufi, FALSE))
ufi 269 uvm/uvm_loan.c rv = uvm_loanentry(&ufi, &output, flags);
ufi 319 uvm/uvm_loan.c uvm_loananon(ufi, output, flags, anon)
ufi 320 uvm/uvm_loan.c struct uvm_faultinfo *ufi;
ufi 353 uvm/uvm_loan.c result = uvmfault_anonget(ufi, ufi->entry->aref.ar_amap, anon);
ufi 408 uvm/uvm_loan.c uvm_loanuobj(ufi, output, flags, va)
ufi 409 uvm/uvm_loan.c struct uvm_faultinfo *ufi;
ufi 414 uvm/uvm_loan.c struct vm_amap *amap = ufi->entry->aref.ar_amap;
ufi 415 uvm/uvm_loan.c struct uvm_object *uobj = ufi->entry->object.uvm_obj;
ufi 430 uvm/uvm_loan.c result = uobj->pgops->pgo_get(uobj, va - ufi->entry->start,
ufi 442 uvm/uvm_loan.c uvmfault_unlockall(ufi, amap, uobj, NULL);
ufi 451 uvm/uvm_loan.c uvmfault_unlockall(ufi, amap, NULL, NULL);
ufi 455 uvm/uvm_loan.c result = uobj->pgops->pgo_get(uobj, va - ufi->entry->start,
ufi 475 uvm/uvm_loan.c locked = uvmfault_relock(ufi);
ufi 485 uvm/uvm_loan.c (locked && amap && amap_lookup(&ufi->entry->aref,
ufi 486 uvm/uvm_loan.c ufi->orig_rvaddr - ufi->entry->start))) {
ufi 489 uvm/uvm_loan.c uvmfault_unlockall(ufi, amap, NULL, NULL);
ufi 580 uvm/uvm_loan.c uvmfault_unlockall(ufi, amap, uobj, NULL);
ufi 611 uvm/uvm_loan.c uvm_loanzero(ufi, output, flags)
ufi 612 uvm/uvm_loan.c struct uvm_faultinfo *ufi;
ufi 623 uvm/uvm_loan.c uvmfault_unlockall(ufi, ufi->entry->aref.ar_amap,
ufi 624 uvm/uvm_loan.c ufi->entry->object.uvm_obj, NULL);
ufi 626 uvm/uvm_loan.c if (!uvmfault_relock(ufi))
ufi 628 uvm/uvm_loan.c if (ufi->entry->object.uvm_obj)
ufi 630 uvm/uvm_loan.c &ufi->entry->object.uvm_obj->vmobjlock);
ufi 652 uvm/uvm_loan.c uvmfault_unlockall(ufi, ufi->entry->aref.ar_amap,
ufi 653 uvm/uvm_loan.c ufi->entry->object.uvm_obj, NULL);
ufi 662 uvm/uvm_loan.c if (!uvmfault_relock(ufi))
ufi 667 uvm/uvm_loan.c if (ufi->entry->object.uvm_obj)
ufi 668 uvm/uvm_loan.c simple_lock(&ufi->entry->object.uvm_obj->vmobjlock);