kva               190 arch/i386/i386/kvm86.c kvm86_bios_delpage(uint32_t vmva, void *kva)
kva               194 arch/i386/i386/kvm86.c 	uvm_km_free(kernel_map, (vaddr_t)kva, PAGE_SIZE);
kva               284 arch/i386/i386/kvm86.c 	unsigned char *kva, insn, trapno;
kva               287 arch/i386/i386/kvm86.c 	kva = (unsigned char *)((tf->tf_cs << 4) + tf->tf_eip);
kva               288 arch/i386/i386/kvm86.c 	insn = *kva;
kva               302 arch/i386/i386/kvm86.c 		trapno = *(kva + 1);
kva              3958 arch/i386/i386/machdep.c _bus_dmamem_unmap(bus_dma_tag_t t, caddr_t kva, size_t size)
kva              3962 arch/i386/i386/machdep.c 	if ((u_long)kva & PGOFSET)
kva              3967 arch/i386/i386/machdep.c 	uvm_km_free(kernel_map, (vaddr_t)kva, size);
kva               761 arch/i386/i386/pmap.c 	vaddr_t kva;
kva               847 arch/i386/i386/pmap.c 		for (kva = VM_MIN_KERNEL_ADDRESS ; kva < virtual_avail ;
kva               848 arch/i386/i386/pmap.c 		     kva += PAGE_SIZE)
kva               849 arch/i386/i386/pmap.c 			if (pmap_valid_entry(PTE_BASE[atop(kva)]))
kva               850 arch/i386/i386/pmap.c 				PTE_BASE[atop(kva)] |= PG_G;
kva               897 arch/i386/include/bus.h void	_bus_dmamem_unmap(bus_dma_tag_t tag, caddr_t kva,
kva               989 arch/i386/isa/isa_machdep.c _isa_bus_dmamem_unmap(bus_dma_tag_t t, caddr_t kva, size_t size)
kva               992 arch/i386/isa/isa_machdep.c 	_bus_dmamem_unmap(t, kva, size);
kva               912 dev/ic/mpi.c   	char				*kva = MPI_DMA_KVA(sc->sc_replies);
kva               921 dev/ic/mpi.c   		rcb->rcb_reply = kva + MPI_REPLY_SIZE * i;
kva              2295 dev/ic/mpi.c   	char					*kva;
kva              2337 dev/ic/mpi.c   	kva = ccb->ccb_cmd;
kva              2338 dev/ic/mpi.c   	kva += sizeof(struct mpi_msg_config_request);
kva              2340 dev/ic/mpi.c   		bcopy(page, kva, len);
kva              2375 dev/ic/mpi.c   		bcopy(kva, page, len);
kva              1224 dev/ic/rtl81x9.c 	caddr_t kva;
kva              1264 dev/ic/rtl81x9.c 	    RL_RXBUFLEN + 32, &kva, BUS_DMA_NOWAIT)) {
kva              1273 dev/ic/rtl81x9.c 		bus_dmamem_unmap(sc->sc_dmat, kva, RL_RXBUFLEN + 32);
kva              1277 dev/ic/rtl81x9.c 	if (bus_dmamap_load(sc->sc_dmat, sc->sc_rx_dmamap, kva,
kva              1281 dev/ic/rtl81x9.c 		bus_dmamem_unmap(sc->sc_dmat, kva, RL_RXBUFLEN + 32);
kva              1285 dev/ic/rtl81x9.c 	sc->rl_cdata.rl_rx_buf = kva;
kva                73 dev/isa/isadma.c 	caddr_t kva;
kva               620 dev/isa/isadma.c isa_dmamem_unmap(isadev, chan, kva, size)
kva               623 dev/isa/isadma.c 	caddr_t kva;
kva               633 dev/isa/isadma.c 	bus_dmamem_unmap(sc->sc_dmat, kva, size);
kva               683 dev/isa/isadma.c 	caddr_t kva;
kva               691 dev/isa/isadma.c 	if (isa_dmamem_map(isadev, chan, addr, size, &kva, bflags)) {
kva               697 dev/isa/isadma.c 		isa_dmamem_unmap(isadev, chan, kva, size);
kva               705 dev/isa/isadma.c 	m->kva = kva;
kva               708 dev/isa/isadma.c 	return (void *)kva;
kva               717 dev/isa/isadma.c 	caddr_t kva = (caddr_t)addr;
kva               719 dev/isa/isadma.c 	for(mp = &isa_mem_head; *mp && (*mp)->kva != kva; mp = &(*mp)->next)
kva               727 dev/isa/isadma.c 	isa_dmamem_unmap(m->isadev, m->chan, kva, m->size);
kva               740 dev/isa/isadma.c 	for(m = isa_mem_head; m && m->kva != (caddr_t)mem; m = m->next)
kva               608 dev/pci/arc.c  	char				*kva = ARC_DMA_KVA(sc->sc_requests);
kva               635 dev/pci/arc.c  		cmd = (struct arc_io_cmd *)(kva +
kva               830 dev/pci/arc.c  	char				*kva = ARC_DMA_KVA(sc->sc_requests);
kva               844 dev/pci/arc.c  		cmd = (struct arc_io_cmd *)(kva +
kva               272 dev/pci/bktr/bktr_os.c         caddr_t kva;
kva               289 dev/pci/bktr/bktr_os.c                            &kva, BUS_DMA_NOWAIT|BUS_DMA_COHERENT)) {
kva               301 dev/pci/bktr/bktr_os.c                 bus_dmamem_unmap(dmat, kva, size);
kva               305 dev/pci/bktr/bktr_os.c         if (bus_dmamap_load(dmat, *dmapp, kva, size, NULL, BUS_DMA_NOWAIT)) {
kva               308 dev/pci/bktr/bktr_os.c                 bus_dmamem_unmap(dmat, kva, size);
kva               313 dev/pci/bktr/bktr_os.c         return (vaddr_t)kva;
kva               317 dev/pci/bktr/bktr_os.c free_bktr_mem(bktr, dmap, kva)
kva               320 dev/pci/bktr/bktr_os.c         vaddr_t kva;
kva               324 dev/pci/bktr/bktr_os.c         bus_dmamem_unmap(dmat, (caddr_t)kva, dmap->dm_mapsize);
kva               153 dev/pci/hifn7751.c 	caddr_t kva;
kva               206 dev/pci/hifn7751.c 	    sizeof(*sc->sc_dma), &kva, BUS_DMA_NOWAIT)) {
kva               213 dev/pci/hifn7751.c 	if (bus_dmamap_load(sc->sc_dmat, sc->sc_dmamap, kva,
kva               216 dev/pci/hifn7751.c 		bus_dmamem_unmap(sc->sc_dmat, kva, sizeof(*sc->sc_dma));
kva               221 dev/pci/hifn7751.c 	sc->sc_dma = (struct hifn_dma *)kva;
kva               325 dev/pci/hifn7751.c 	bus_dmamem_unmap(sc->sc_dmat, kva, sizeof(*sc->sc_dma));
kva               232 dev/pci/if_bce.c 	caddr_t         kva;
kva               321 dev/pci/if_bce.c 	    2 * PAGE_SIZE, &kva, BUS_DMA_NOWAIT))) {
kva               333 dev/pci/if_bce.c 		bus_dmamem_unmap(sc->bce_dmatag, kva, 2 * PAGE_SIZE);
kva               338 dev/pci/if_bce.c 	if (bus_dmamap_load(sc->bce_dmatag, sc->bce_ring_map, kva,
kva               341 dev/pci/if_bce.c 		bus_dmamem_unmap(sc->bce_dmatag, kva, 2 * PAGE_SIZE);
kva               346 dev/pci/if_bce.c 	sc->bce_rx_ring = (struct bce_dma_slot *) kva;
kva               347 dev/pci/if_bce.c 	sc->bce_tx_ring = (struct bce_dma_slot *) (kva + PAGE_SIZE);
kva               621 dev/pci/if_bge.c 	caddr_t			ptr, kva;
kva               636 dev/pci/if_bge.c 	if (bus_dmamem_map(sc->bge_dmatag, &seg, rseg, BGE_JMEM, &kva,
kva               654 dev/pci/if_bge.c 			    kva, BGE_JMEM, NULL, BUS_DMA_NOWAIT)) {
kva               661 dev/pci/if_bge.c 	sc->bge_cdata.bge_jumbo_buf = (caddr_t)kva;
kva               697 dev/pci/if_bge.c 			bus_dmamem_unmap(sc->bge_dmatag, kva, BGE_JMEM);
kva              1696 dev/pci/if_bge.c 	caddr_t			kva;
kva              1908 dev/pci/if_bge.c 			   sizeof(struct bge_ring_data), &kva,
kva              1922 dev/pci/if_bge.c 	if (bus_dmamap_load(sc->bge_dmatag, sc->bge_ring_map, kva,
kva              1929 dev/pci/if_bge.c 	sc->bge_rdata = (struct bge_ring_data *)kva;
kva              2079 dev/pci/if_bge.c 	bus_dmamem_unmap(sc->bge_dmatag, kva,
kva               420 dev/pci/if_cas.c 		caddr_t kva;
kva               432 dev/pci/if_cas.c 		    CAS_PAGE_SIZE, &kva, BUS_DMA_NOWAIT)) != 0) {
kva               437 dev/pci/if_cas.c 		sc->sc_rxsoft[i].rxs_kva = kva;
kva               447 dev/pci/if_cas.c 		   sc->sc_rxsoft[i].rxs_dmamap, kva, CAS_PAGE_SIZE, NULL,
kva               413 dev/pci/if_lge.c 	caddr_t			kva;
kva               520 dev/pci/if_lge.c 			   sizeof(struct lge_list_data), &kva,
kva               534 dev/pci/if_lge.c 	if (bus_dmamap_load(sc->sc_dmatag, dmamap, kva,
kva               541 dev/pci/if_lge.c 	sc->lge_ldata = (struct lge_list_data *)kva;
kva               611 dev/pci/if_lge.c 	bus_dmamem_unmap(sc->sc_dmatag, kva,
kva               749 dev/pci/if_lge.c 	caddr_t			ptr, kva;
kva               765 dev/pci/if_lge.c 	if (bus_dmamem_map(sc->sc_dmatag, &seg, rseg, LGE_JMEM, &kva,
kva               782 dev/pci/if_lge.c 	if (bus_dmamap_load(sc->sc_dmatag, dmamap, kva, LGE_JMEM,
kva               790 dev/pci/if_lge.c 	sc->lge_cdata.lge_jumbo_buf = (caddr_t)kva;
kva               826 dev/pci/if_lge.c 			bus_dmamem_unmap(sc->sc_dmatag, kva, LGE_JMEM);
kva               567 dev/pci/if_msk.c 	caddr_t			ptr, kva;
kva               582 dev/pci/if_msk.c 	if (bus_dmamem_map(sc->sc_dmatag, &seg, rseg, MSK_JMEM, &kva,
kva               599 dev/pci/if_msk.c 			    kva, MSK_JMEM, NULL, BUS_DMA_NOWAIT)) {
kva               606 dev/pci/if_msk.c 	sc_if->sk_cdata.sk_jumbo_buf = (caddr_t)kva;
kva               642 dev/pci/if_msk.c 			bus_dmamem_unmap(sc->sc_dmatag, kva, MSK_JMEM);
kva               990 dev/pci/if_msk.c 	caddr_t kva;
kva              1044 dev/pci/if_msk.c 	    sizeof(struct msk_ring_data), &kva, BUS_DMA_NOWAIT)) {
kva              1055 dev/pci/if_msk.c 	if (bus_dmamap_load(sc->sc_dmatag, sc_if->sk_ring_map, kva,
kva              1060 dev/pci/if_msk.c         sc_if->sk_rdata = (struct msk_ring_data *)kva;
kva              1129 dev/pci/if_msk.c 	bus_dmamem_unmap(sc->sc_dmatag, kva, sizeof(struct msk_ring_data));
kva              1166 dev/pci/if_msk.c 	caddr_t kva;
kva              1257 dev/pci/if_msk.c 	    &kva, BUS_DMA_NOWAIT)) {
kva              1269 dev/pci/if_msk.c 	if (bus_dmamap_load(sc->sc_dmatag, sc->sk_status_map, kva,
kva              1275 dev/pci/if_msk.c 	sc->sk_status_ring = (struct msk_status_desc *)kva;
kva              1392 dev/pci/if_msk.c 	bus_dmamem_unmap(sc->sc_dmatag, kva, 
kva               738 dev/pci/if_nge.c 	caddr_t			kva;
kva               849 dev/pci/if_nge.c 			   sizeof(struct nge_list_data), &kva,
kva               863 dev/pci/if_nge.c 	if (bus_dmamap_load(sc->sc_dmatag, dmamap, kva,
kva               870 dev/pci/if_nge.c 	sc->nge_ldata = (struct nge_list_data *)kva;
kva               966 dev/pci/if_nge.c 	bus_dmamem_unmap(sc->sc_dmatag, kva,
kva              1108 dev/pci/if_nge.c 	caddr_t			ptr, kva;
kva              1123 dev/pci/if_nge.c 	if (bus_dmamem_map(sc->sc_dmatag, &seg, rseg, NGE_JMEM, &kva,
kva              1140 dev/pci/if_nge.c 	if (bus_dmamap_load(sc->sc_dmatag, dmamap, kva, NGE_JMEM,
kva              1148 dev/pci/if_nge.c 	sc->nge_cdata.nge_jumbo_buf = (caddr_t)kva;
kva              1190 dev/pci/if_nge.c 			bus_dmamem_unmap(sc->sc_dmatag, kva, NGE_JMEM);
kva               695 dev/pci/if_sk.c 	caddr_t			ptr, kva;
kva               710 dev/pci/if_sk.c 	if (bus_dmamem_map(sc->sc_dmatag, &seg, rseg, SK_JMEM, &kva,
kva               727 dev/pci/if_sk.c 			    kva, SK_JMEM, NULL, BUS_DMA_NOWAIT)) {
kva               734 dev/pci/if_sk.c 	sc_if->sk_cdata.sk_jumbo_buf = (caddr_t)kva;
kva               770 dev/pci/if_sk.c 			bus_dmamem_unmap(sc->sc_dmatag, kva, SK_JMEM);
kva              1048 dev/pci/if_sk.c 	caddr_t kva;
kva              1158 dev/pci/if_sk.c 	    sizeof(struct sk_ring_data), &kva, BUS_DMA_NOWAIT)) {
kva              1169 dev/pci/if_sk.c 	if (bus_dmamap_load(sc->sc_dmatag, sc_if->sk_ring_map, kva,
kva              1174 dev/pci/if_sk.c         sc_if->sk_rdata = (struct sk_ring_data *)kva;
kva              1265 dev/pci/if_sk.c 	bus_dmamem_unmap(sc->sc_dmatag, kva, sizeof(struct sk_ring_data));
kva               587 dev/pci/if_ti.c 	caddr_t ptr, kva;
kva               602 dev/pci/if_ti.c 	if (bus_dmamem_map(sc->sc_dmatag, &seg, rseg, TI_JMEM, &kva,
kva               619 dev/pci/if_ti.c 	if (bus_dmamap_load(sc->sc_dmatag, sc->ti_cdata.ti_rx_jumbo_map, kva,
kva               627 dev/pci/if_ti.c 	sc->ti_cdata.ti_jumbo_buf = (caddr_t)kva;
kva               663 dev/pci/if_ti.c 			bus_dmamem_unmap(sc->sc_dmatag, kva, TI_JMEM);
kva              1525 dev/pci/if_ti.c 	caddr_t kva;
kva              1596 dev/pci/if_ti.c 	    sizeof(struct ti_ring_data), &kva, BUS_DMA_NOWAIT)) {
kva              1607 dev/pci/if_ti.c 	if (bus_dmamap_load(sc->sc_dmatag, sc->ti_ring_map, kva,
kva              1611 dev/pci/if_ti.c 	sc->ti_rdata = (struct ti_ring_data *)kva;
kva              1704 dev/pci/if_ti.c 	bus_dmamem_unmap(sc->sc_dmatag, kva,
kva              2008 dev/pci/if_tl.c 	caddr_t kva;
kva              2070 dev/pci/if_tl.c 	    &kva, BUS_DMA_NOWAIT)) {
kva              2079 dev/pci/if_tl.c 		bus_dmamem_unmap(sc->sc_dmat, kva, sizeof(struct tl_list_data));
kva              2084 dev/pci/if_tl.c 	if (bus_dmamap_load(sc->sc_dmat, dmamap, kva,
kva              2088 dev/pci/if_tl.c 		bus_dmamem_unmap(sc->sc_dmat, kva, sizeof(struct tl_list_data));
kva              2093 dev/pci/if_tl.c 	sc->tl_ldata = (struct tl_list_data *)kva;
kva               500 dev/pci/if_vic.c 	u_int8_t			*kva;
kva               527 dev/pci/if_vic.c 	kva = VIC_DMA_KVA(sc);
kva               538 dev/pci/if_vic.c 	sc->sc_rxq = (struct vic_rxdesc *)&kva[offset];
kva               546 dev/pci/if_vic.c 	sc->sc_rxq2 = (struct vic_rxdesc *)&kva[offset];
kva               563 dev/pci/if_vic.c 	sc->sc_txq = (struct vic_txdesc *)&kva[offset];
kva               625 dev/pci/if_vr.c 	caddr_t kva;
kva               741 dev/pci/if_vr.c 	    sizeof(struct vr_list_data), &kva, BUS_DMA_NOWAIT)) {
kva               751 dev/pci/if_vr.c 	if (bus_dmamap_load(sc->sc_dmat, sc->sc_listmap, kva,
kva               756 dev/pci/if_vr.c 	sc->vr_ldata = (struct vr_list_data *)kva;
kva               799 dev/pci/if_vr.c 	bus_dmamem_unmap(sc->sc_dmat, kva, sizeof(struct vr_list_data));
kva               723 dev/pci/if_wb.c 	caddr_t kva;
kva               808 dev/pci/if_wb.c 	    sizeof(struct wb_list_data), &kva, BUS_DMA_NOWAIT)) {
kva               818 dev/pci/if_wb.c 	if (bus_dmamap_load(pa->pa_dmat, dmamap, kva,
kva               823 dev/pci/if_wb.c 	sc->wb_ldata = (struct wb_list_data *)kva;
kva               868 dev/pci/if_wb.c 	bus_dmamem_unmap(pa->pa_dmat, kva,
kva              1199 dev/pci/if_xge.c 	caddr_t kva;
kva              1208 dev/pci/if_xge.c 	if (bus_dmamem_map(sc->sc_dmat, &seg, rseg, TXMAPSZ, &kva,
kva              1218 dev/pci/if_xge.c 	    kva, TXMAPSZ, NULL, BUS_DMA_NOWAIT))
kva              1222 dev/pci/if_xge.c 	txp = (struct txd *)kva;
kva              1224 dev/pci/if_xge.c 	for (txp = (struct txd *)kva, i = 0; i < NTXDESCS; i++) {
kva              1237 dev/pci/if_xge.c 		bus_dmamem_unmap(sc->sc_dmat, kva, TXMAPSZ);
kva              1253 dev/pci/if_xge.c 	caddr_t kva;
kva              1268 dev/pci/if_xge.c 	if (bus_dmamem_map(sc->sc_dmat, &seg, rseg, RXMAPSZ, &kva,
kva              1278 dev/pci/if_xge.c 	    kva, RXMAPSZ, NULL, BUS_DMA_NOWAIT))
kva              1282 dev/pci/if_xge.c 	for (rxpp = (struct rxd_4k *)kva, i = 0; i < NRXPAGES; i++, rxpp++) {
kva              1296 dev/pci/if_xge.c 		bus_dmamem_unmap(sc->sc_dmat, kva, RXMAPSZ);
kva                65 uvm/uvm_io.c   	vaddr_t baseva, endva, pageoffset, kva;
kva               110 uvm/uvm_io.c   		error = uvm_map_extract(map, baseva, chunksz, kernel_map, &kva,
kva               132 uvm/uvm_io.c   		error = uiomove((caddr_t) (kva + pageoffset), sz, uio);
kva               142 uvm/uvm_io.c   		uvm_unmap_remove(kernel_map, kva, kva+chunksz,
kva               342 uvm/uvm_km.c   	vaddr_t kva, loopva;
kva               356 uvm/uvm_km.c   	kva = vm_map_min(map);	/* hint */
kva               362 uvm/uvm_km.c   	if (__predict_false(uvm_map(map, &kva, size, obj, UVM_UNKNOWN_OFFSET,
kva               374 uvm/uvm_km.c   		UVMHIST_LOG(maphist,"<- done valloc (kva=0x%lx)", kva,0,0,0);
kva               375 uvm/uvm_km.c   		return(kva);
kva               383 uvm/uvm_km.c   		offset = kva - vm_map_min(kernel_map);
kva               387 uvm/uvm_km.c   	UVMHIST_LOG(maphist, "  kva=0x%lx, offset=0x%lx", kva, offset,0,0);
kva               394 uvm/uvm_km.c   	loopva = kva;
kva               395 uvm/uvm_km.c   	while (loopva != kva + size) {
kva               407 uvm/uvm_km.c   				uvm_unmap(map, kva, kva + size);
kva               433 uvm/uvm_km.c   	UVMHIST_LOG(maphist,"<- done (kva=0x%lx)", kva,0,0,0);
kva               434 uvm/uvm_km.c   	return(kva);
kva               478 uvm/uvm_km.c   	vaddr_t kva, loopva;
kva               487 uvm/uvm_km.c   	kva = vm_map_min(map);		/* hint */
kva               493 uvm/uvm_km.c   	if (__predict_false(uvm_map(map, &kva, size, uvm.kernel_object,
kva               504 uvm/uvm_km.c   	offset = kva - vm_map_min(kernel_map);
kva               505 uvm/uvm_km.c   	UVMHIST_LOG(maphist,"  kva=0x%lx, offset=0x%lx", kva, offset,0,0);
kva               511 uvm/uvm_km.c   	loopva = kva;
kva               543 uvm/uvm_km.c   				uvm_unmap(map, kva, loopva - kva);
kva               570 uvm/uvm_km.c   		memset((caddr_t)kva, 0, loopva - kva);
kva               572 uvm/uvm_km.c   	UVMHIST_LOG(maphist,"<- done (kva=0x%lx)", kva,0,0,0);
kva               573 uvm/uvm_km.c   	return(kva);
kva               591 uvm/uvm_km.c   	vaddr_t kva;
kva               598 uvm/uvm_km.c   	kva = vm_map_min(map);		/* hint */
kva               604 uvm/uvm_km.c   	if (__predict_false(uvm_map(map, &kva, size, uvm.kernel_object,
kva               611 uvm/uvm_km.c   	UVMHIST_LOG(maphist, "<- done (kva=0x%lx)", kva,0,0,0);
kva               612 uvm/uvm_km.c   	return(kva);
kva               626 uvm/uvm_km.c   	vaddr_t kva;
kva               637 uvm/uvm_km.c   		kva = vm_map_min(map);		/* hint */
kva               644 uvm/uvm_km.c   		if (__predict_true(uvm_map(map, &kva, size, uvm.kernel_object,
kva               647 uvm/uvm_km.c   			UVMHIST_LOG(maphist,"<- done (kva=0x%lx)", kva,0,0,0);
kva               648 uvm/uvm_km.c   			return(kva);
kva               121 uvm/uvm_pager.c 	vaddr_t kva;
kva               140 uvm/uvm_pager.c 	kva = 0;			/* let system choose VA */
kva               142 uvm/uvm_pager.c 	if (uvm_map(pager_map, &kva, size, NULL, 
kva               154 uvm/uvm_pager.c 			kva = emergva;
kva               172 uvm/uvm_pager.c 	for (cva = kva ; size != 0 ; size -= PAGE_SIZE, cva += PAGE_SIZE) {
kva               181 uvm/uvm_pager.c 	UVMHIST_LOG(maphist, "<- done (KVA=0x%lx)", kva,0,0,0);
kva               182 uvm/uvm_pager.c 	return(kva);
kva               193 uvm/uvm_pager.c uvm_pagermapout(kva, npages)
kva               194 uvm/uvm_pager.c 	vaddr_t kva;
kva               201 uvm/uvm_pager.c 	UVMHIST_LOG(maphist, " (kva=0x%lx, npages=%ld)", kva, npages,0,0);
kva               207 uvm/uvm_pager.c 	if (kva == emergva) {
kva               217 uvm/uvm_pager.c 	uvm_unmap_remove(pager_map, kva, kva + size, &entries, NULL);
kva               226 uvm/uvm_pager.c 	pmap_remove(pmap_kernel(), kva, kva + (npages << PAGE_SHIFT));
kva                57 uvm/uvm_pager_i.h uvm_pageratop(kva)
kva                58 uvm/uvm_pager_i.h 	vaddr_t kva;
kva                64 uvm/uvm_pager_i.h 	rv = pmap_extract(pmap_kernel(), kva, &pa);
kva              1805 uvm/uvm_swap.c 	vaddr_t kva;
kva              1834 uvm/uvm_swap.c 	kva = uvm_pagermapin(pps, npages, mapinflags);
kva              1835 uvm/uvm_swap.c 	if (kva == 0)
kva              1881 uvm/uvm_swap.c 			uvm_pagermapout(kva, npages);
kva              1887 uvm/uvm_swap.c 			uvm_pagermapout(kva, npages);
kva              1892 uvm/uvm_swap.c 		src = (caddr_t) kva;
kva              1907 uvm/uvm_swap.c 		uvm_pagermapout(kva, npages);
kva              1914 uvm/uvm_swap.c 		kva = dstkva;
kva              1939 uvm/uvm_swap.c 			uvm_pagermapout(kva, npages);
kva              1965 uvm/uvm_swap.c 	bp->b_data = (caddr_t)kva;
kva              2043 uvm/uvm_swap.c 	uvm_pagermapout(kva, npages);
kva              1558 uvm/uvm_vnode.c 	vaddr_t kva;
kva              1607 uvm/uvm_vnode.c 	kva = uvm_pagermapin(pps, npages, mapinflags);
kva              1608 uvm/uvm_vnode.c 	if (kva == 0 && waitf == M_NOWAIT) {
kva              1623 uvm/uvm_vnode.c 	if (kva == 0)
kva              1624 uvm/uvm_vnode.c 		kva = uvm_pagermapin(pps, npages,
kva              1637 uvm/uvm_vnode.c 	iov.iov_base = (caddr_t) kva;
kva              1697 uvm/uvm_vnode.c 			memset((void *) (kva + got), 0,
kva              1705 uvm/uvm_vnode.c 	uvm_pagermapout(kva, npages);