kva 190 arch/i386/i386/kvm86.c kvm86_bios_delpage(uint32_t vmva, void *kva) kva 194 arch/i386/i386/kvm86.c uvm_km_free(kernel_map, (vaddr_t)kva, PAGE_SIZE); kva 284 arch/i386/i386/kvm86.c unsigned char *kva, insn, trapno; kva 287 arch/i386/i386/kvm86.c kva = (unsigned char *)((tf->tf_cs << 4) + tf->tf_eip); kva 288 arch/i386/i386/kvm86.c insn = *kva; kva 302 arch/i386/i386/kvm86.c trapno = *(kva + 1); kva 3958 arch/i386/i386/machdep.c _bus_dmamem_unmap(bus_dma_tag_t t, caddr_t kva, size_t size) kva 3962 arch/i386/i386/machdep.c if ((u_long)kva & PGOFSET) kva 3967 arch/i386/i386/machdep.c uvm_km_free(kernel_map, (vaddr_t)kva, size); kva 761 arch/i386/i386/pmap.c vaddr_t kva; kva 847 arch/i386/i386/pmap.c for (kva = VM_MIN_KERNEL_ADDRESS ; kva < virtual_avail ; kva 848 arch/i386/i386/pmap.c kva += PAGE_SIZE) kva 849 arch/i386/i386/pmap.c if (pmap_valid_entry(PTE_BASE[atop(kva)])) kva 850 arch/i386/i386/pmap.c PTE_BASE[atop(kva)] |= PG_G; kva 897 arch/i386/include/bus.h void _bus_dmamem_unmap(bus_dma_tag_t tag, caddr_t kva, kva 989 arch/i386/isa/isa_machdep.c _isa_bus_dmamem_unmap(bus_dma_tag_t t, caddr_t kva, size_t size) kva 992 arch/i386/isa/isa_machdep.c _bus_dmamem_unmap(t, kva, size); kva 912 dev/ic/mpi.c char *kva = MPI_DMA_KVA(sc->sc_replies); kva 921 dev/ic/mpi.c rcb->rcb_reply = kva + MPI_REPLY_SIZE * i; kva 2295 dev/ic/mpi.c char *kva; kva 2337 dev/ic/mpi.c kva = ccb->ccb_cmd; kva 2338 dev/ic/mpi.c kva += sizeof(struct mpi_msg_config_request); kva 2340 dev/ic/mpi.c bcopy(page, kva, len); kva 2375 dev/ic/mpi.c bcopy(kva, page, len); kva 1224 dev/ic/rtl81x9.c caddr_t kva; kva 1264 dev/ic/rtl81x9.c RL_RXBUFLEN + 32, &kva, BUS_DMA_NOWAIT)) { kva 1273 dev/ic/rtl81x9.c bus_dmamem_unmap(sc->sc_dmat, kva, RL_RXBUFLEN + 32); kva 1277 dev/ic/rtl81x9.c if (bus_dmamap_load(sc->sc_dmat, sc->sc_rx_dmamap, kva, kva 1281 dev/ic/rtl81x9.c bus_dmamem_unmap(sc->sc_dmat, kva, RL_RXBUFLEN + 32); kva 1285 dev/ic/rtl81x9.c sc->rl_cdata.rl_rx_buf = kva; kva 73 dev/isa/isadma.c caddr_t kva; kva 620 dev/isa/isadma.c isa_dmamem_unmap(isadev, chan, kva, size) kva 623 dev/isa/isadma.c caddr_t kva; kva 633 dev/isa/isadma.c bus_dmamem_unmap(sc->sc_dmat, kva, size); kva 683 dev/isa/isadma.c caddr_t kva; kva 691 dev/isa/isadma.c if (isa_dmamem_map(isadev, chan, addr, size, &kva, bflags)) { kva 697 dev/isa/isadma.c isa_dmamem_unmap(isadev, chan, kva, size); kva 705 dev/isa/isadma.c m->kva = kva; kva 708 dev/isa/isadma.c return (void *)kva; kva 717 dev/isa/isadma.c caddr_t kva = (caddr_t)addr; kva 719 dev/isa/isadma.c for(mp = &isa_mem_head; *mp && (*mp)->kva != kva; mp = &(*mp)->next) kva 727 dev/isa/isadma.c isa_dmamem_unmap(m->isadev, m->chan, kva, m->size); kva 740 dev/isa/isadma.c for(m = isa_mem_head; m && m->kva != (caddr_t)mem; m = m->next) kva 608 dev/pci/arc.c char *kva = ARC_DMA_KVA(sc->sc_requests); kva 635 dev/pci/arc.c cmd = (struct arc_io_cmd *)(kva + kva 830 dev/pci/arc.c char *kva = ARC_DMA_KVA(sc->sc_requests); kva 844 dev/pci/arc.c cmd = (struct arc_io_cmd *)(kva + kva 272 dev/pci/bktr/bktr_os.c caddr_t kva; kva 289 dev/pci/bktr/bktr_os.c &kva, BUS_DMA_NOWAIT|BUS_DMA_COHERENT)) { kva 301 dev/pci/bktr/bktr_os.c bus_dmamem_unmap(dmat, kva, size); kva 305 dev/pci/bktr/bktr_os.c if (bus_dmamap_load(dmat, *dmapp, kva, size, NULL, BUS_DMA_NOWAIT)) { kva 308 dev/pci/bktr/bktr_os.c bus_dmamem_unmap(dmat, kva, size); kva 313 dev/pci/bktr/bktr_os.c return (vaddr_t)kva; kva 317 dev/pci/bktr/bktr_os.c free_bktr_mem(bktr, dmap, kva) kva 320 dev/pci/bktr/bktr_os.c vaddr_t kva; kva 324 dev/pci/bktr/bktr_os.c bus_dmamem_unmap(dmat, (caddr_t)kva, dmap->dm_mapsize); kva 153 dev/pci/hifn7751.c caddr_t kva; kva 206 dev/pci/hifn7751.c sizeof(*sc->sc_dma), &kva, BUS_DMA_NOWAIT)) { kva 213 dev/pci/hifn7751.c if (bus_dmamap_load(sc->sc_dmat, sc->sc_dmamap, kva, kva 216 dev/pci/hifn7751.c bus_dmamem_unmap(sc->sc_dmat, kva, sizeof(*sc->sc_dma)); kva 221 dev/pci/hifn7751.c sc->sc_dma = (struct hifn_dma *)kva; kva 325 dev/pci/hifn7751.c bus_dmamem_unmap(sc->sc_dmat, kva, sizeof(*sc->sc_dma)); kva 232 dev/pci/if_bce.c caddr_t kva; kva 321 dev/pci/if_bce.c 2 * PAGE_SIZE, &kva, BUS_DMA_NOWAIT))) { kva 333 dev/pci/if_bce.c bus_dmamem_unmap(sc->bce_dmatag, kva, 2 * PAGE_SIZE); kva 338 dev/pci/if_bce.c if (bus_dmamap_load(sc->bce_dmatag, sc->bce_ring_map, kva, kva 341 dev/pci/if_bce.c bus_dmamem_unmap(sc->bce_dmatag, kva, 2 * PAGE_SIZE); kva 346 dev/pci/if_bce.c sc->bce_rx_ring = (struct bce_dma_slot *) kva; kva 347 dev/pci/if_bce.c sc->bce_tx_ring = (struct bce_dma_slot *) (kva + PAGE_SIZE); kva 621 dev/pci/if_bge.c caddr_t ptr, kva; kva 636 dev/pci/if_bge.c if (bus_dmamem_map(sc->bge_dmatag, &seg, rseg, BGE_JMEM, &kva, kva 654 dev/pci/if_bge.c kva, BGE_JMEM, NULL, BUS_DMA_NOWAIT)) { kva 661 dev/pci/if_bge.c sc->bge_cdata.bge_jumbo_buf = (caddr_t)kva; kva 697 dev/pci/if_bge.c bus_dmamem_unmap(sc->bge_dmatag, kva, BGE_JMEM); kva 1696 dev/pci/if_bge.c caddr_t kva; kva 1908 dev/pci/if_bge.c sizeof(struct bge_ring_data), &kva, kva 1922 dev/pci/if_bge.c if (bus_dmamap_load(sc->bge_dmatag, sc->bge_ring_map, kva, kva 1929 dev/pci/if_bge.c sc->bge_rdata = (struct bge_ring_data *)kva; kva 2079 dev/pci/if_bge.c bus_dmamem_unmap(sc->bge_dmatag, kva, kva 420 dev/pci/if_cas.c caddr_t kva; kva 432 dev/pci/if_cas.c CAS_PAGE_SIZE, &kva, BUS_DMA_NOWAIT)) != 0) { kva 437 dev/pci/if_cas.c sc->sc_rxsoft[i].rxs_kva = kva; kva 447 dev/pci/if_cas.c sc->sc_rxsoft[i].rxs_dmamap, kva, CAS_PAGE_SIZE, NULL, kva 413 dev/pci/if_lge.c caddr_t kva; kva 520 dev/pci/if_lge.c sizeof(struct lge_list_data), &kva, kva 534 dev/pci/if_lge.c if (bus_dmamap_load(sc->sc_dmatag, dmamap, kva, kva 541 dev/pci/if_lge.c sc->lge_ldata = (struct lge_list_data *)kva; kva 611 dev/pci/if_lge.c bus_dmamem_unmap(sc->sc_dmatag, kva, kva 749 dev/pci/if_lge.c caddr_t ptr, kva; kva 765 dev/pci/if_lge.c if (bus_dmamem_map(sc->sc_dmatag, &seg, rseg, LGE_JMEM, &kva, kva 782 dev/pci/if_lge.c if (bus_dmamap_load(sc->sc_dmatag, dmamap, kva, LGE_JMEM, kva 790 dev/pci/if_lge.c sc->lge_cdata.lge_jumbo_buf = (caddr_t)kva; kva 826 dev/pci/if_lge.c bus_dmamem_unmap(sc->sc_dmatag, kva, LGE_JMEM); kva 567 dev/pci/if_msk.c caddr_t ptr, kva; kva 582 dev/pci/if_msk.c if (bus_dmamem_map(sc->sc_dmatag, &seg, rseg, MSK_JMEM, &kva, kva 599 dev/pci/if_msk.c kva, MSK_JMEM, NULL, BUS_DMA_NOWAIT)) { kva 606 dev/pci/if_msk.c sc_if->sk_cdata.sk_jumbo_buf = (caddr_t)kva; kva 642 dev/pci/if_msk.c bus_dmamem_unmap(sc->sc_dmatag, kva, MSK_JMEM); kva 990 dev/pci/if_msk.c caddr_t kva; kva 1044 dev/pci/if_msk.c sizeof(struct msk_ring_data), &kva, BUS_DMA_NOWAIT)) { kva 1055 dev/pci/if_msk.c if (bus_dmamap_load(sc->sc_dmatag, sc_if->sk_ring_map, kva, kva 1060 dev/pci/if_msk.c sc_if->sk_rdata = (struct msk_ring_data *)kva; kva 1129 dev/pci/if_msk.c bus_dmamem_unmap(sc->sc_dmatag, kva, sizeof(struct msk_ring_data)); kva 1166 dev/pci/if_msk.c caddr_t kva; kva 1257 dev/pci/if_msk.c &kva, BUS_DMA_NOWAIT)) { kva 1269 dev/pci/if_msk.c if (bus_dmamap_load(sc->sc_dmatag, sc->sk_status_map, kva, kva 1275 dev/pci/if_msk.c sc->sk_status_ring = (struct msk_status_desc *)kva; kva 1392 dev/pci/if_msk.c bus_dmamem_unmap(sc->sc_dmatag, kva, kva 738 dev/pci/if_nge.c caddr_t kva; kva 849 dev/pci/if_nge.c sizeof(struct nge_list_data), &kva, kva 863 dev/pci/if_nge.c if (bus_dmamap_load(sc->sc_dmatag, dmamap, kva, kva 870 dev/pci/if_nge.c sc->nge_ldata = (struct nge_list_data *)kva; kva 966 dev/pci/if_nge.c bus_dmamem_unmap(sc->sc_dmatag, kva, kva 1108 dev/pci/if_nge.c caddr_t ptr, kva; kva 1123 dev/pci/if_nge.c if (bus_dmamem_map(sc->sc_dmatag, &seg, rseg, NGE_JMEM, &kva, kva 1140 dev/pci/if_nge.c if (bus_dmamap_load(sc->sc_dmatag, dmamap, kva, NGE_JMEM, kva 1148 dev/pci/if_nge.c sc->nge_cdata.nge_jumbo_buf = (caddr_t)kva; kva 1190 dev/pci/if_nge.c bus_dmamem_unmap(sc->sc_dmatag, kva, NGE_JMEM); kva 695 dev/pci/if_sk.c caddr_t ptr, kva; kva 710 dev/pci/if_sk.c if (bus_dmamem_map(sc->sc_dmatag, &seg, rseg, SK_JMEM, &kva, kva 727 dev/pci/if_sk.c kva, SK_JMEM, NULL, BUS_DMA_NOWAIT)) { kva 734 dev/pci/if_sk.c sc_if->sk_cdata.sk_jumbo_buf = (caddr_t)kva; kva 770 dev/pci/if_sk.c bus_dmamem_unmap(sc->sc_dmatag, kva, SK_JMEM); kva 1048 dev/pci/if_sk.c caddr_t kva; kva 1158 dev/pci/if_sk.c sizeof(struct sk_ring_data), &kva, BUS_DMA_NOWAIT)) { kva 1169 dev/pci/if_sk.c if (bus_dmamap_load(sc->sc_dmatag, sc_if->sk_ring_map, kva, kva 1174 dev/pci/if_sk.c sc_if->sk_rdata = (struct sk_ring_data *)kva; kva 1265 dev/pci/if_sk.c bus_dmamem_unmap(sc->sc_dmatag, kva, sizeof(struct sk_ring_data)); kva 587 dev/pci/if_ti.c caddr_t ptr, kva; kva 602 dev/pci/if_ti.c if (bus_dmamem_map(sc->sc_dmatag, &seg, rseg, TI_JMEM, &kva, kva 619 dev/pci/if_ti.c if (bus_dmamap_load(sc->sc_dmatag, sc->ti_cdata.ti_rx_jumbo_map, kva, kva 627 dev/pci/if_ti.c sc->ti_cdata.ti_jumbo_buf = (caddr_t)kva; kva 663 dev/pci/if_ti.c bus_dmamem_unmap(sc->sc_dmatag, kva, TI_JMEM); kva 1525 dev/pci/if_ti.c caddr_t kva; kva 1596 dev/pci/if_ti.c sizeof(struct ti_ring_data), &kva, BUS_DMA_NOWAIT)) { kva 1607 dev/pci/if_ti.c if (bus_dmamap_load(sc->sc_dmatag, sc->ti_ring_map, kva, kva 1611 dev/pci/if_ti.c sc->ti_rdata = (struct ti_ring_data *)kva; kva 1704 dev/pci/if_ti.c bus_dmamem_unmap(sc->sc_dmatag, kva, kva 2008 dev/pci/if_tl.c caddr_t kva; kva 2070 dev/pci/if_tl.c &kva, BUS_DMA_NOWAIT)) { kva 2079 dev/pci/if_tl.c bus_dmamem_unmap(sc->sc_dmat, kva, sizeof(struct tl_list_data)); kva 2084 dev/pci/if_tl.c if (bus_dmamap_load(sc->sc_dmat, dmamap, kva, kva 2088 dev/pci/if_tl.c bus_dmamem_unmap(sc->sc_dmat, kva, sizeof(struct tl_list_data)); kva 2093 dev/pci/if_tl.c sc->tl_ldata = (struct tl_list_data *)kva; kva 500 dev/pci/if_vic.c u_int8_t *kva; kva 527 dev/pci/if_vic.c kva = VIC_DMA_KVA(sc); kva 538 dev/pci/if_vic.c sc->sc_rxq = (struct vic_rxdesc *)&kva[offset]; kva 546 dev/pci/if_vic.c sc->sc_rxq2 = (struct vic_rxdesc *)&kva[offset]; kva 563 dev/pci/if_vic.c sc->sc_txq = (struct vic_txdesc *)&kva[offset]; kva 625 dev/pci/if_vr.c caddr_t kva; kva 741 dev/pci/if_vr.c sizeof(struct vr_list_data), &kva, BUS_DMA_NOWAIT)) { kva 751 dev/pci/if_vr.c if (bus_dmamap_load(sc->sc_dmat, sc->sc_listmap, kva, kva 756 dev/pci/if_vr.c sc->vr_ldata = (struct vr_list_data *)kva; kva 799 dev/pci/if_vr.c bus_dmamem_unmap(sc->sc_dmat, kva, sizeof(struct vr_list_data)); kva 723 dev/pci/if_wb.c caddr_t kva; kva 808 dev/pci/if_wb.c sizeof(struct wb_list_data), &kva, BUS_DMA_NOWAIT)) { kva 818 dev/pci/if_wb.c if (bus_dmamap_load(pa->pa_dmat, dmamap, kva, kva 823 dev/pci/if_wb.c sc->wb_ldata = (struct wb_list_data *)kva; kva 868 dev/pci/if_wb.c bus_dmamem_unmap(pa->pa_dmat, kva, kva 1199 dev/pci/if_xge.c caddr_t kva; kva 1208 dev/pci/if_xge.c if (bus_dmamem_map(sc->sc_dmat, &seg, rseg, TXMAPSZ, &kva, kva 1218 dev/pci/if_xge.c kva, TXMAPSZ, NULL, BUS_DMA_NOWAIT)) kva 1222 dev/pci/if_xge.c txp = (struct txd *)kva; kva 1224 dev/pci/if_xge.c for (txp = (struct txd *)kva, i = 0; i < NTXDESCS; i++) { kva 1237 dev/pci/if_xge.c bus_dmamem_unmap(sc->sc_dmat, kva, TXMAPSZ); kva 1253 dev/pci/if_xge.c caddr_t kva; kva 1268 dev/pci/if_xge.c if (bus_dmamem_map(sc->sc_dmat, &seg, rseg, RXMAPSZ, &kva, kva 1278 dev/pci/if_xge.c kva, RXMAPSZ, NULL, BUS_DMA_NOWAIT)) kva 1282 dev/pci/if_xge.c for (rxpp = (struct rxd_4k *)kva, i = 0; i < NRXPAGES; i++, rxpp++) { kva 1296 dev/pci/if_xge.c bus_dmamem_unmap(sc->sc_dmat, kva, RXMAPSZ); kva 65 uvm/uvm_io.c vaddr_t baseva, endva, pageoffset, kva; kva 110 uvm/uvm_io.c error = uvm_map_extract(map, baseva, chunksz, kernel_map, &kva, kva 132 uvm/uvm_io.c error = uiomove((caddr_t) (kva + pageoffset), sz, uio); kva 142 uvm/uvm_io.c uvm_unmap_remove(kernel_map, kva, kva+chunksz, kva 342 uvm/uvm_km.c vaddr_t kva, loopva; kva 356 uvm/uvm_km.c kva = vm_map_min(map); /* hint */ kva 362 uvm/uvm_km.c if (__predict_false(uvm_map(map, &kva, size, obj, UVM_UNKNOWN_OFFSET, kva 374 uvm/uvm_km.c UVMHIST_LOG(maphist,"<- done valloc (kva=0x%lx)", kva,0,0,0); kva 375 uvm/uvm_km.c return(kva); kva 383 uvm/uvm_km.c offset = kva - vm_map_min(kernel_map); kva 387 uvm/uvm_km.c UVMHIST_LOG(maphist, " kva=0x%lx, offset=0x%lx", kva, offset,0,0); kva 394 uvm/uvm_km.c loopva = kva; kva 395 uvm/uvm_km.c while (loopva != kva + size) { kva 407 uvm/uvm_km.c uvm_unmap(map, kva, kva + size); kva 433 uvm/uvm_km.c UVMHIST_LOG(maphist,"<- done (kva=0x%lx)", kva,0,0,0); kva 434 uvm/uvm_km.c return(kva); kva 478 uvm/uvm_km.c vaddr_t kva, loopva; kva 487 uvm/uvm_km.c kva = vm_map_min(map); /* hint */ kva 493 uvm/uvm_km.c if (__predict_false(uvm_map(map, &kva, size, uvm.kernel_object, kva 504 uvm/uvm_km.c offset = kva - vm_map_min(kernel_map); kva 505 uvm/uvm_km.c UVMHIST_LOG(maphist," kva=0x%lx, offset=0x%lx", kva, offset,0,0); kva 511 uvm/uvm_km.c loopva = kva; kva 543 uvm/uvm_km.c uvm_unmap(map, kva, loopva - kva); kva 570 uvm/uvm_km.c memset((caddr_t)kva, 0, loopva - kva); kva 572 uvm/uvm_km.c UVMHIST_LOG(maphist,"<- done (kva=0x%lx)", kva,0,0,0); kva 573 uvm/uvm_km.c return(kva); kva 591 uvm/uvm_km.c vaddr_t kva; kva 598 uvm/uvm_km.c kva = vm_map_min(map); /* hint */ kva 604 uvm/uvm_km.c if (__predict_false(uvm_map(map, &kva, size, uvm.kernel_object, kva 611 uvm/uvm_km.c UVMHIST_LOG(maphist, "<- done (kva=0x%lx)", kva,0,0,0); kva 612 uvm/uvm_km.c return(kva); kva 626 uvm/uvm_km.c vaddr_t kva; kva 637 uvm/uvm_km.c kva = vm_map_min(map); /* hint */ kva 644 uvm/uvm_km.c if (__predict_true(uvm_map(map, &kva, size, uvm.kernel_object, kva 647 uvm/uvm_km.c UVMHIST_LOG(maphist,"<- done (kva=0x%lx)", kva,0,0,0); kva 648 uvm/uvm_km.c return(kva); kva 121 uvm/uvm_pager.c vaddr_t kva; kva 140 uvm/uvm_pager.c kva = 0; /* let system choose VA */ kva 142 uvm/uvm_pager.c if (uvm_map(pager_map, &kva, size, NULL, kva 154 uvm/uvm_pager.c kva = emergva; kva 172 uvm/uvm_pager.c for (cva = kva ; size != 0 ; size -= PAGE_SIZE, cva += PAGE_SIZE) { kva 181 uvm/uvm_pager.c UVMHIST_LOG(maphist, "<- done (KVA=0x%lx)", kva,0,0,0); kva 182 uvm/uvm_pager.c return(kva); kva 193 uvm/uvm_pager.c uvm_pagermapout(kva, npages) kva 194 uvm/uvm_pager.c vaddr_t kva; kva 201 uvm/uvm_pager.c UVMHIST_LOG(maphist, " (kva=0x%lx, npages=%ld)", kva, npages,0,0); kva 207 uvm/uvm_pager.c if (kva == emergva) { kva 217 uvm/uvm_pager.c uvm_unmap_remove(pager_map, kva, kva + size, &entries, NULL); kva 226 uvm/uvm_pager.c pmap_remove(pmap_kernel(), kva, kva + (npages << PAGE_SHIFT)); kva 57 uvm/uvm_pager_i.h uvm_pageratop(kva) kva 58 uvm/uvm_pager_i.h vaddr_t kva; kva 64 uvm/uvm_pager_i.h rv = pmap_extract(pmap_kernel(), kva, &pa); kva 1805 uvm/uvm_swap.c vaddr_t kva; kva 1834 uvm/uvm_swap.c kva = uvm_pagermapin(pps, npages, mapinflags); kva 1835 uvm/uvm_swap.c if (kva == 0) kva 1881 uvm/uvm_swap.c uvm_pagermapout(kva, npages); kva 1887 uvm/uvm_swap.c uvm_pagermapout(kva, npages); kva 1892 uvm/uvm_swap.c src = (caddr_t) kva; kva 1907 uvm/uvm_swap.c uvm_pagermapout(kva, npages); kva 1914 uvm/uvm_swap.c kva = dstkva; kva 1939 uvm/uvm_swap.c uvm_pagermapout(kva, npages); kva 1965 uvm/uvm_swap.c bp->b_data = (caddr_t)kva; kva 2043 uvm/uvm_swap.c uvm_pagermapout(kva, npages); kva 1558 uvm/uvm_vnode.c vaddr_t kva; kva 1607 uvm/uvm_vnode.c kva = uvm_pagermapin(pps, npages, mapinflags); kva 1608 uvm/uvm_vnode.c if (kva == 0 && waitf == M_NOWAIT) { kva 1623 uvm/uvm_vnode.c if (kva == 0) kva 1624 uvm/uvm_vnode.c kva = uvm_pagermapin(pps, npages, kva 1637 uvm/uvm_vnode.c iov.iov_base = (caddr_t) kva; kva 1697 uvm/uvm_vnode.c memset((void *) (kva + got), 0, kva 1705 uvm/uvm_vnode.c uvm_pagermapout(kva, npages);