kva 190 arch/i386/i386/kvm86.c kvm86_bios_delpage(uint32_t vmva, void *kva)
kva 194 arch/i386/i386/kvm86.c uvm_km_free(kernel_map, (vaddr_t)kva, PAGE_SIZE);
kva 284 arch/i386/i386/kvm86.c unsigned char *kva, insn, trapno;
kva 287 arch/i386/i386/kvm86.c kva = (unsigned char *)((tf->tf_cs << 4) + tf->tf_eip);
kva 288 arch/i386/i386/kvm86.c insn = *kva;
kva 302 arch/i386/i386/kvm86.c trapno = *(kva + 1);
kva 3958 arch/i386/i386/machdep.c _bus_dmamem_unmap(bus_dma_tag_t t, caddr_t kva, size_t size)
kva 3962 arch/i386/i386/machdep.c if ((u_long)kva & PGOFSET)
kva 3967 arch/i386/i386/machdep.c uvm_km_free(kernel_map, (vaddr_t)kva, size);
kva 761 arch/i386/i386/pmap.c vaddr_t kva;
kva 847 arch/i386/i386/pmap.c for (kva = VM_MIN_KERNEL_ADDRESS ; kva < virtual_avail ;
kva 848 arch/i386/i386/pmap.c kva += PAGE_SIZE)
kva 849 arch/i386/i386/pmap.c if (pmap_valid_entry(PTE_BASE[atop(kva)]))
kva 850 arch/i386/i386/pmap.c PTE_BASE[atop(kva)] |= PG_G;
kva 897 arch/i386/include/bus.h void _bus_dmamem_unmap(bus_dma_tag_t tag, caddr_t kva,
kva 989 arch/i386/isa/isa_machdep.c _isa_bus_dmamem_unmap(bus_dma_tag_t t, caddr_t kva, size_t size)
kva 992 arch/i386/isa/isa_machdep.c _bus_dmamem_unmap(t, kva, size);
kva 912 dev/ic/mpi.c char *kva = MPI_DMA_KVA(sc->sc_replies);
kva 921 dev/ic/mpi.c rcb->rcb_reply = kva + MPI_REPLY_SIZE * i;
kva 2295 dev/ic/mpi.c char *kva;
kva 2337 dev/ic/mpi.c kva = ccb->ccb_cmd;
kva 2338 dev/ic/mpi.c kva += sizeof(struct mpi_msg_config_request);
kva 2340 dev/ic/mpi.c bcopy(page, kva, len);
kva 2375 dev/ic/mpi.c bcopy(kva, page, len);
kva 1224 dev/ic/rtl81x9.c caddr_t kva;
kva 1264 dev/ic/rtl81x9.c RL_RXBUFLEN + 32, &kva, BUS_DMA_NOWAIT)) {
kva 1273 dev/ic/rtl81x9.c bus_dmamem_unmap(sc->sc_dmat, kva, RL_RXBUFLEN + 32);
kva 1277 dev/ic/rtl81x9.c if (bus_dmamap_load(sc->sc_dmat, sc->sc_rx_dmamap, kva,
kva 1281 dev/ic/rtl81x9.c bus_dmamem_unmap(sc->sc_dmat, kva, RL_RXBUFLEN + 32);
kva 1285 dev/ic/rtl81x9.c sc->rl_cdata.rl_rx_buf = kva;
kva 73 dev/isa/isadma.c caddr_t kva;
kva 620 dev/isa/isadma.c isa_dmamem_unmap(isadev, chan, kva, size)
kva 623 dev/isa/isadma.c caddr_t kva;
kva 633 dev/isa/isadma.c bus_dmamem_unmap(sc->sc_dmat, kva, size);
kva 683 dev/isa/isadma.c caddr_t kva;
kva 691 dev/isa/isadma.c if (isa_dmamem_map(isadev, chan, addr, size, &kva, bflags)) {
kva 697 dev/isa/isadma.c isa_dmamem_unmap(isadev, chan, kva, size);
kva 705 dev/isa/isadma.c m->kva = kva;
kva 708 dev/isa/isadma.c return (void *)kva;
kva 717 dev/isa/isadma.c caddr_t kva = (caddr_t)addr;
kva 719 dev/isa/isadma.c for(mp = &isa_mem_head; *mp && (*mp)->kva != kva; mp = &(*mp)->next)
kva 727 dev/isa/isadma.c isa_dmamem_unmap(m->isadev, m->chan, kva, m->size);
kva 740 dev/isa/isadma.c for(m = isa_mem_head; m && m->kva != (caddr_t)mem; m = m->next)
kva 608 dev/pci/arc.c char *kva = ARC_DMA_KVA(sc->sc_requests);
kva 635 dev/pci/arc.c cmd = (struct arc_io_cmd *)(kva +
kva 830 dev/pci/arc.c char *kva = ARC_DMA_KVA(sc->sc_requests);
kva 844 dev/pci/arc.c cmd = (struct arc_io_cmd *)(kva +
kva 272 dev/pci/bktr/bktr_os.c caddr_t kva;
kva 289 dev/pci/bktr/bktr_os.c &kva, BUS_DMA_NOWAIT|BUS_DMA_COHERENT)) {
kva 301 dev/pci/bktr/bktr_os.c bus_dmamem_unmap(dmat, kva, size);
kva 305 dev/pci/bktr/bktr_os.c if (bus_dmamap_load(dmat, *dmapp, kva, size, NULL, BUS_DMA_NOWAIT)) {
kva 308 dev/pci/bktr/bktr_os.c bus_dmamem_unmap(dmat, kva, size);
kva 313 dev/pci/bktr/bktr_os.c return (vaddr_t)kva;
kva 317 dev/pci/bktr/bktr_os.c free_bktr_mem(bktr, dmap, kva)
kva 320 dev/pci/bktr/bktr_os.c vaddr_t kva;
kva 324 dev/pci/bktr/bktr_os.c bus_dmamem_unmap(dmat, (caddr_t)kva, dmap->dm_mapsize);
kva 153 dev/pci/hifn7751.c caddr_t kva;
kva 206 dev/pci/hifn7751.c sizeof(*sc->sc_dma), &kva, BUS_DMA_NOWAIT)) {
kva 213 dev/pci/hifn7751.c if (bus_dmamap_load(sc->sc_dmat, sc->sc_dmamap, kva,
kva 216 dev/pci/hifn7751.c bus_dmamem_unmap(sc->sc_dmat, kva, sizeof(*sc->sc_dma));
kva 221 dev/pci/hifn7751.c sc->sc_dma = (struct hifn_dma *)kva;
kva 325 dev/pci/hifn7751.c bus_dmamem_unmap(sc->sc_dmat, kva, sizeof(*sc->sc_dma));
kva 232 dev/pci/if_bce.c caddr_t kva;
kva 321 dev/pci/if_bce.c 2 * PAGE_SIZE, &kva, BUS_DMA_NOWAIT))) {
kva 333 dev/pci/if_bce.c bus_dmamem_unmap(sc->bce_dmatag, kva, 2 * PAGE_SIZE);
kva 338 dev/pci/if_bce.c if (bus_dmamap_load(sc->bce_dmatag, sc->bce_ring_map, kva,
kva 341 dev/pci/if_bce.c bus_dmamem_unmap(sc->bce_dmatag, kva, 2 * PAGE_SIZE);
kva 346 dev/pci/if_bce.c sc->bce_rx_ring = (struct bce_dma_slot *) kva;
kva 347 dev/pci/if_bce.c sc->bce_tx_ring = (struct bce_dma_slot *) (kva + PAGE_SIZE);
kva 621 dev/pci/if_bge.c caddr_t ptr, kva;
kva 636 dev/pci/if_bge.c if (bus_dmamem_map(sc->bge_dmatag, &seg, rseg, BGE_JMEM, &kva,
kva 654 dev/pci/if_bge.c kva, BGE_JMEM, NULL, BUS_DMA_NOWAIT)) {
kva 661 dev/pci/if_bge.c sc->bge_cdata.bge_jumbo_buf = (caddr_t)kva;
kva 697 dev/pci/if_bge.c bus_dmamem_unmap(sc->bge_dmatag, kva, BGE_JMEM);
kva 1696 dev/pci/if_bge.c caddr_t kva;
kva 1908 dev/pci/if_bge.c sizeof(struct bge_ring_data), &kva,
kva 1922 dev/pci/if_bge.c if (bus_dmamap_load(sc->bge_dmatag, sc->bge_ring_map, kva,
kva 1929 dev/pci/if_bge.c sc->bge_rdata = (struct bge_ring_data *)kva;
kva 2079 dev/pci/if_bge.c bus_dmamem_unmap(sc->bge_dmatag, kva,
kva 420 dev/pci/if_cas.c caddr_t kva;
kva 432 dev/pci/if_cas.c CAS_PAGE_SIZE, &kva, BUS_DMA_NOWAIT)) != 0) {
kva 437 dev/pci/if_cas.c sc->sc_rxsoft[i].rxs_kva = kva;
kva 447 dev/pci/if_cas.c sc->sc_rxsoft[i].rxs_dmamap, kva, CAS_PAGE_SIZE, NULL,
kva 413 dev/pci/if_lge.c caddr_t kva;
kva 520 dev/pci/if_lge.c sizeof(struct lge_list_data), &kva,
kva 534 dev/pci/if_lge.c if (bus_dmamap_load(sc->sc_dmatag, dmamap, kva,
kva 541 dev/pci/if_lge.c sc->lge_ldata = (struct lge_list_data *)kva;
kva 611 dev/pci/if_lge.c bus_dmamem_unmap(sc->sc_dmatag, kva,
kva 749 dev/pci/if_lge.c caddr_t ptr, kva;
kva 765 dev/pci/if_lge.c if (bus_dmamem_map(sc->sc_dmatag, &seg, rseg, LGE_JMEM, &kva,
kva 782 dev/pci/if_lge.c if (bus_dmamap_load(sc->sc_dmatag, dmamap, kva, LGE_JMEM,
kva 790 dev/pci/if_lge.c sc->lge_cdata.lge_jumbo_buf = (caddr_t)kva;
kva 826 dev/pci/if_lge.c bus_dmamem_unmap(sc->sc_dmatag, kva, LGE_JMEM);
kva 567 dev/pci/if_msk.c caddr_t ptr, kva;
kva 582 dev/pci/if_msk.c if (bus_dmamem_map(sc->sc_dmatag, &seg, rseg, MSK_JMEM, &kva,
kva 599 dev/pci/if_msk.c kva, MSK_JMEM, NULL, BUS_DMA_NOWAIT)) {
kva 606 dev/pci/if_msk.c sc_if->sk_cdata.sk_jumbo_buf = (caddr_t)kva;
kva 642 dev/pci/if_msk.c bus_dmamem_unmap(sc->sc_dmatag, kva, MSK_JMEM);
kva 990 dev/pci/if_msk.c caddr_t kva;
kva 1044 dev/pci/if_msk.c sizeof(struct msk_ring_data), &kva, BUS_DMA_NOWAIT)) {
kva 1055 dev/pci/if_msk.c if (bus_dmamap_load(sc->sc_dmatag, sc_if->sk_ring_map, kva,
kva 1060 dev/pci/if_msk.c sc_if->sk_rdata = (struct msk_ring_data *)kva;
kva 1129 dev/pci/if_msk.c bus_dmamem_unmap(sc->sc_dmatag, kva, sizeof(struct msk_ring_data));
kva 1166 dev/pci/if_msk.c caddr_t kva;
kva 1257 dev/pci/if_msk.c &kva, BUS_DMA_NOWAIT)) {
kva 1269 dev/pci/if_msk.c if (bus_dmamap_load(sc->sc_dmatag, sc->sk_status_map, kva,
kva 1275 dev/pci/if_msk.c sc->sk_status_ring = (struct msk_status_desc *)kva;
kva 1392 dev/pci/if_msk.c bus_dmamem_unmap(sc->sc_dmatag, kva,
kva 738 dev/pci/if_nge.c caddr_t kva;
kva 849 dev/pci/if_nge.c sizeof(struct nge_list_data), &kva,
kva 863 dev/pci/if_nge.c if (bus_dmamap_load(sc->sc_dmatag, dmamap, kva,
kva 870 dev/pci/if_nge.c sc->nge_ldata = (struct nge_list_data *)kva;
kva 966 dev/pci/if_nge.c bus_dmamem_unmap(sc->sc_dmatag, kva,
kva 1108 dev/pci/if_nge.c caddr_t ptr, kva;
kva 1123 dev/pci/if_nge.c if (bus_dmamem_map(sc->sc_dmatag, &seg, rseg, NGE_JMEM, &kva,
kva 1140 dev/pci/if_nge.c if (bus_dmamap_load(sc->sc_dmatag, dmamap, kva, NGE_JMEM,
kva 1148 dev/pci/if_nge.c sc->nge_cdata.nge_jumbo_buf = (caddr_t)kva;
kva 1190 dev/pci/if_nge.c bus_dmamem_unmap(sc->sc_dmatag, kva, NGE_JMEM);
kva 695 dev/pci/if_sk.c caddr_t ptr, kva;
kva 710 dev/pci/if_sk.c if (bus_dmamem_map(sc->sc_dmatag, &seg, rseg, SK_JMEM, &kva,
kva 727 dev/pci/if_sk.c kva, SK_JMEM, NULL, BUS_DMA_NOWAIT)) {
kva 734 dev/pci/if_sk.c sc_if->sk_cdata.sk_jumbo_buf = (caddr_t)kva;
kva 770 dev/pci/if_sk.c bus_dmamem_unmap(sc->sc_dmatag, kva, SK_JMEM);
kva 1048 dev/pci/if_sk.c caddr_t kva;
kva 1158 dev/pci/if_sk.c sizeof(struct sk_ring_data), &kva, BUS_DMA_NOWAIT)) {
kva 1169 dev/pci/if_sk.c if (bus_dmamap_load(sc->sc_dmatag, sc_if->sk_ring_map, kva,
kva 1174 dev/pci/if_sk.c sc_if->sk_rdata = (struct sk_ring_data *)kva;
kva 1265 dev/pci/if_sk.c bus_dmamem_unmap(sc->sc_dmatag, kva, sizeof(struct sk_ring_data));
kva 587 dev/pci/if_ti.c caddr_t ptr, kva;
kva 602 dev/pci/if_ti.c if (bus_dmamem_map(sc->sc_dmatag, &seg, rseg, TI_JMEM, &kva,
kva 619 dev/pci/if_ti.c if (bus_dmamap_load(sc->sc_dmatag, sc->ti_cdata.ti_rx_jumbo_map, kva,
kva 627 dev/pci/if_ti.c sc->ti_cdata.ti_jumbo_buf = (caddr_t)kva;
kva 663 dev/pci/if_ti.c bus_dmamem_unmap(sc->sc_dmatag, kva, TI_JMEM);
kva 1525 dev/pci/if_ti.c caddr_t kva;
kva 1596 dev/pci/if_ti.c sizeof(struct ti_ring_data), &kva, BUS_DMA_NOWAIT)) {
kva 1607 dev/pci/if_ti.c if (bus_dmamap_load(sc->sc_dmatag, sc->ti_ring_map, kva,
kva 1611 dev/pci/if_ti.c sc->ti_rdata = (struct ti_ring_data *)kva;
kva 1704 dev/pci/if_ti.c bus_dmamem_unmap(sc->sc_dmatag, kva,
kva 2008 dev/pci/if_tl.c caddr_t kva;
kva 2070 dev/pci/if_tl.c &kva, BUS_DMA_NOWAIT)) {
kva 2079 dev/pci/if_tl.c bus_dmamem_unmap(sc->sc_dmat, kva, sizeof(struct tl_list_data));
kva 2084 dev/pci/if_tl.c if (bus_dmamap_load(sc->sc_dmat, dmamap, kva,
kva 2088 dev/pci/if_tl.c bus_dmamem_unmap(sc->sc_dmat, kva, sizeof(struct tl_list_data));
kva 2093 dev/pci/if_tl.c sc->tl_ldata = (struct tl_list_data *)kva;
kva 500 dev/pci/if_vic.c u_int8_t *kva;
kva 527 dev/pci/if_vic.c kva = VIC_DMA_KVA(sc);
kva 538 dev/pci/if_vic.c sc->sc_rxq = (struct vic_rxdesc *)&kva[offset];
kva 546 dev/pci/if_vic.c sc->sc_rxq2 = (struct vic_rxdesc *)&kva[offset];
kva 563 dev/pci/if_vic.c sc->sc_txq = (struct vic_txdesc *)&kva[offset];
kva 625 dev/pci/if_vr.c caddr_t kva;
kva 741 dev/pci/if_vr.c sizeof(struct vr_list_data), &kva, BUS_DMA_NOWAIT)) {
kva 751 dev/pci/if_vr.c if (bus_dmamap_load(sc->sc_dmat, sc->sc_listmap, kva,
kva 756 dev/pci/if_vr.c sc->vr_ldata = (struct vr_list_data *)kva;
kva 799 dev/pci/if_vr.c bus_dmamem_unmap(sc->sc_dmat, kva, sizeof(struct vr_list_data));
kva 723 dev/pci/if_wb.c caddr_t kva;
kva 808 dev/pci/if_wb.c sizeof(struct wb_list_data), &kva, BUS_DMA_NOWAIT)) {
kva 818 dev/pci/if_wb.c if (bus_dmamap_load(pa->pa_dmat, dmamap, kva,
kva 823 dev/pci/if_wb.c sc->wb_ldata = (struct wb_list_data *)kva;
kva 868 dev/pci/if_wb.c bus_dmamem_unmap(pa->pa_dmat, kva,
kva 1199 dev/pci/if_xge.c caddr_t kva;
kva 1208 dev/pci/if_xge.c if (bus_dmamem_map(sc->sc_dmat, &seg, rseg, TXMAPSZ, &kva,
kva 1218 dev/pci/if_xge.c kva, TXMAPSZ, NULL, BUS_DMA_NOWAIT))
kva 1222 dev/pci/if_xge.c txp = (struct txd *)kva;
kva 1224 dev/pci/if_xge.c for (txp = (struct txd *)kva, i = 0; i < NTXDESCS; i++) {
kva 1237 dev/pci/if_xge.c bus_dmamem_unmap(sc->sc_dmat, kva, TXMAPSZ);
kva 1253 dev/pci/if_xge.c caddr_t kva;
kva 1268 dev/pci/if_xge.c if (bus_dmamem_map(sc->sc_dmat, &seg, rseg, RXMAPSZ, &kva,
kva 1278 dev/pci/if_xge.c kva, RXMAPSZ, NULL, BUS_DMA_NOWAIT))
kva 1282 dev/pci/if_xge.c for (rxpp = (struct rxd_4k *)kva, i = 0; i < NRXPAGES; i++, rxpp++) {
kva 1296 dev/pci/if_xge.c bus_dmamem_unmap(sc->sc_dmat, kva, RXMAPSZ);
kva 65 uvm/uvm_io.c vaddr_t baseva, endva, pageoffset, kva;
kva 110 uvm/uvm_io.c error = uvm_map_extract(map, baseva, chunksz, kernel_map, &kva,
kva 132 uvm/uvm_io.c error = uiomove((caddr_t) (kva + pageoffset), sz, uio);
kva 142 uvm/uvm_io.c uvm_unmap_remove(kernel_map, kva, kva+chunksz,
kva 342 uvm/uvm_km.c vaddr_t kva, loopva;
kva 356 uvm/uvm_km.c kva = vm_map_min(map); /* hint */
kva 362 uvm/uvm_km.c if (__predict_false(uvm_map(map, &kva, size, obj, UVM_UNKNOWN_OFFSET,
kva 374 uvm/uvm_km.c UVMHIST_LOG(maphist,"<- done valloc (kva=0x%lx)", kva,0,0,0);
kva 375 uvm/uvm_km.c return(kva);
kva 383 uvm/uvm_km.c offset = kva - vm_map_min(kernel_map);
kva 387 uvm/uvm_km.c UVMHIST_LOG(maphist, " kva=0x%lx, offset=0x%lx", kva, offset,0,0);
kva 394 uvm/uvm_km.c loopva = kva;
kva 395 uvm/uvm_km.c while (loopva != kva + size) {
kva 407 uvm/uvm_km.c uvm_unmap(map, kva, kva + size);
kva 433 uvm/uvm_km.c UVMHIST_LOG(maphist,"<- done (kva=0x%lx)", kva,0,0,0);
kva 434 uvm/uvm_km.c return(kva);
kva 478 uvm/uvm_km.c vaddr_t kva, loopva;
kva 487 uvm/uvm_km.c kva = vm_map_min(map); /* hint */
kva 493 uvm/uvm_km.c if (__predict_false(uvm_map(map, &kva, size, uvm.kernel_object,
kva 504 uvm/uvm_km.c offset = kva - vm_map_min(kernel_map);
kva 505 uvm/uvm_km.c UVMHIST_LOG(maphist," kva=0x%lx, offset=0x%lx", kva, offset,0,0);
kva 511 uvm/uvm_km.c loopva = kva;
kva 543 uvm/uvm_km.c uvm_unmap(map, kva, loopva - kva);
kva 570 uvm/uvm_km.c memset((caddr_t)kva, 0, loopva - kva);
kva 572 uvm/uvm_km.c UVMHIST_LOG(maphist,"<- done (kva=0x%lx)", kva,0,0,0);
kva 573 uvm/uvm_km.c return(kva);
kva 591 uvm/uvm_km.c vaddr_t kva;
kva 598 uvm/uvm_km.c kva = vm_map_min(map); /* hint */
kva 604 uvm/uvm_km.c if (__predict_false(uvm_map(map, &kva, size, uvm.kernel_object,
kva 611 uvm/uvm_km.c UVMHIST_LOG(maphist, "<- done (kva=0x%lx)", kva,0,0,0);
kva 612 uvm/uvm_km.c return(kva);
kva 626 uvm/uvm_km.c vaddr_t kva;
kva 637 uvm/uvm_km.c kva = vm_map_min(map); /* hint */
kva 644 uvm/uvm_km.c if (__predict_true(uvm_map(map, &kva, size, uvm.kernel_object,
kva 647 uvm/uvm_km.c UVMHIST_LOG(maphist,"<- done (kva=0x%lx)", kva,0,0,0);
kva 648 uvm/uvm_km.c return(kva);
kva 121 uvm/uvm_pager.c vaddr_t kva;
kva 140 uvm/uvm_pager.c kva = 0; /* let system choose VA */
kva 142 uvm/uvm_pager.c if (uvm_map(pager_map, &kva, size, NULL,
kva 154 uvm/uvm_pager.c kva = emergva;
kva 172 uvm/uvm_pager.c for (cva = kva ; size != 0 ; size -= PAGE_SIZE, cva += PAGE_SIZE) {
kva 181 uvm/uvm_pager.c UVMHIST_LOG(maphist, "<- done (KVA=0x%lx)", kva,0,0,0);
kva 182 uvm/uvm_pager.c return(kva);
kva 193 uvm/uvm_pager.c uvm_pagermapout(kva, npages)
kva 194 uvm/uvm_pager.c vaddr_t kva;
kva 201 uvm/uvm_pager.c UVMHIST_LOG(maphist, " (kva=0x%lx, npages=%ld)", kva, npages,0,0);
kva 207 uvm/uvm_pager.c if (kva == emergva) {
kva 217 uvm/uvm_pager.c uvm_unmap_remove(pager_map, kva, kva + size, &entries, NULL);
kva 226 uvm/uvm_pager.c pmap_remove(pmap_kernel(), kva, kva + (npages << PAGE_SHIFT));
kva 57 uvm/uvm_pager_i.h uvm_pageratop(kva)
kva 58 uvm/uvm_pager_i.h vaddr_t kva;
kva 64 uvm/uvm_pager_i.h rv = pmap_extract(pmap_kernel(), kva, &pa);
kva 1805 uvm/uvm_swap.c vaddr_t kva;
kva 1834 uvm/uvm_swap.c kva = uvm_pagermapin(pps, npages, mapinflags);
kva 1835 uvm/uvm_swap.c if (kva == 0)
kva 1881 uvm/uvm_swap.c uvm_pagermapout(kva, npages);
kva 1887 uvm/uvm_swap.c uvm_pagermapout(kva, npages);
kva 1892 uvm/uvm_swap.c src = (caddr_t) kva;
kva 1907 uvm/uvm_swap.c uvm_pagermapout(kva, npages);
kva 1914 uvm/uvm_swap.c kva = dstkva;
kva 1939 uvm/uvm_swap.c uvm_pagermapout(kva, npages);
kva 1965 uvm/uvm_swap.c bp->b_data = (caddr_t)kva;
kva 2043 uvm/uvm_swap.c uvm_pagermapout(kva, npages);
kva 1558 uvm/uvm_vnode.c vaddr_t kva;
kva 1607 uvm/uvm_vnode.c kva = uvm_pagermapin(pps, npages, mapinflags);
kva 1608 uvm/uvm_vnode.c if (kva == 0 && waitf == M_NOWAIT) {
kva 1623 uvm/uvm_vnode.c if (kva == 0)
kva 1624 uvm/uvm_vnode.c kva = uvm_pagermapin(pps, npages,
kva 1637 uvm/uvm_vnode.c iov.iov_base = (caddr_t) kva;
kva 1697 uvm/uvm_vnode.c memset((void *) (kva + got), 0,
kva 1705 uvm/uvm_vnode.c uvm_pagermapout(kva, npages);