kernel_map         54 arch/i386/i386/acpi_machdep.c 	vaddr_t va = uvm_km_valloc(kernel_map, endpa - pgpa);
kernel_map         77 arch/i386/i386/acpi_machdep.c 	uvm_km_free(kernel_map, handle->baseva, handle->vsize);
kernel_map        250 arch/i386/i386/bios.c 			eva = uvm_km_valloc(kernel_map, end-pa);
kernel_map        503 arch/i386/i386/bios.c 	sva = va = uvm_km_valloc(kernel_map, endpa);
kernel_map        232 arch/i386/i386/cpu.c 	kstack = uvm_km_alloc(kernel_map, USPACE);
kernel_map        518 arch/i386/i386/cpu.c 	ci->ci_doubleflt_stack = (char *)uvm_km_alloc(kernel_map, USPACE);
kernel_map        535 arch/i386/i386/cpu.c 	ci->ci_ddbipi_stack = (char *)uvm_km_alloc(kernel_map, USPACE);
kernel_map        138 arch/i386/i386/gdt.c 	gdt = (union descriptor *)uvm_km_valloc(kernel_map, max_len);
kernel_map        164 arch/i386/i386/gdt.c 	ci->ci_gdt = (union descriptor *)uvm_km_valloc(kernel_map, max_len);
kernel_map        165 arch/i386/i386/gdt.c 	uvm_map_pageable(kernel_map, (vaddr_t)ci->ci_gdt,
kernel_map         90 arch/i386/i386/kvm86.c 	if ((buf = (char *)uvm_km_zalloc(kernel_map, vmdsize)) == NULL)
kernel_map        118 arch/i386/i386/kvm86.c 	if ((bioscallscratchpage = (void *)uvm_km_alloc(kernel_map, PAGE_SIZE))
kernel_map        125 arch/i386/i386/kvm86.c 	bioscalltmpva = uvm_km_alloc(kernel_map, PAGE_SIZE);
kernel_map        180 arch/i386/i386/kvm86.c 	if ((mem = (void *)uvm_km_alloc(kernel_map, PAGE_SIZE)) == NULL)
kernel_map        194 arch/i386/i386/kvm86.c 	uvm_km_free(kernel_map, (vaddr_t)kva, PAGE_SIZE);
kernel_map        424 arch/i386/i386/machdep.c 	if ((v = (caddr_t)uvm_km_zalloc(kernel_map, round_page(sz))) == 0)
kernel_map        439 arch/i386/i386/machdep.c 	minaddr = vm_map_min(kernel_map);
kernel_map        440 arch/i386/i386/machdep.c 	exec_map = uvm_km_suballoc(kernel_map, &minaddr, &maxaddr,
kernel_map        446 arch/i386/i386/machdep.c 	phys_map = uvm_km_suballoc(kernel_map, &minaddr, &maxaddr,
kernel_map       2766 arch/i386/i386/machdep.c 	va = uvm_km_zalloc(kernel_map, NBPG*2);
kernel_map       2812 arch/i386/i386/machdep.c 	cpu_ldt = (union descriptor *)uvm_km_alloc(kernel_map, len);
kernel_map       3517 arch/i386/i386/machdep.c 	va = uvm_km_valloc(kernel_map, map_size);
kernel_map       3574 arch/i386/i386/machdep.c 		uvm_km_free(kernel_map, va, endva - va);
kernel_map       3618 arch/i386/i386/machdep.c 		uvm_km_free(kernel_map, va, endva - va);
kernel_map       3931 arch/i386/i386/machdep.c 	va = uvm_km_valloc(kernel_map, size);
kernel_map       3967 arch/i386/i386/machdep.c 	uvm_km_free(kernel_map, (vaddr_t)kva, size);
kernel_map        231 arch/i386/i386/mpbios.c 	vaddr_t va = uvm_km_valloc(kernel_map, endpa - pgpa);
kernel_map        262 arch/i386/i386/mpbios.c 	uvm_km_free(kernel_map, handle->baseva, handle->vsize);
kernel_map        953 arch/i386/i386/pmap.c 	pv_initpage = (struct pv_page *) uvm_km_alloc(kernel_map, PAGE_SIZE);
kernel_map       1257 arch/i386/i386/pmap.c 		map = kernel_map;
kernel_map       1452 arch/i386/i386/pmap.c 	pmap->pm_pdir = (pd_entry_t *) uvm_km_alloc(kernel_map, NBPG);
kernel_map       1562 arch/i386/i386/pmap.c 	uvm_km_free(kernel_map, (vaddr_t)pmap->pm_pdir, NBPG);
kernel_map       1574 arch/i386/i386/pmap.c 		uvm_km_free(kernel_map, (vaddr_t)pmap->pm_ldt,
kernel_map       1611 arch/i386/i386/pmap.c 		new_ldt = (union descriptor *)uvm_km_alloc(kernel_map, len);
kernel_map       1663 arch/i386/i386/pmap.c 		uvm_km_free(kernel_map, (vaddr_t)old_ldt, len);
kernel_map       1038 arch/i386/i386/pmapae.c 	pmap->pm_pdir = uvm_km_alloc(kernel_map, 4 * NBPG);
kernel_map         70 arch/i386/i386/sys_machdep.c extern struct vm_map *kernel_map;
kernel_map        265 arch/i386/i386/sys_machdep.c 		new_ldt = (union descriptor *)uvm_km_alloc(kernel_map,
kernel_map        278 arch/i386/i386/sys_machdep.c 			uvm_km_free(kernel_map, (vaddr_t)new_ldt, new_len);
kernel_map        295 arch/i386/i386/sys_machdep.c 			uvm_km_free(kernel_map, (vaddr_t)old_ldt, old_len);
kernel_map        477 arch/i386/i386/trap.c 			map = kernel_map;
kernel_map        482 arch/i386/i386/trap.c 		if (map == kernel_map && va == 0) {
kernel_map        494 arch/i386/i386/trap.c 			if (map != kernel_map)
kernel_map        302 ddb/db_break.c 		((map1 == NULL) && (map2 == kernel_map)) ||
kernel_map        303 ddb/db_break.c 		((map1 == kernel_map) && (map2 == NULL)));
kernel_map        313 ddb/db_break.c 		(map == kernel_map) ||
kernel_map        339 ddb/db_break.c 	    return kernel_map;
kernel_map         95 dev/ic/pdqvar.h #define	PDQ_OS_MEMFREE_CONTIG(p, n)	kmem_free(kernel_map, (vm_offset_t) p, n)
kernel_map         97 dev/ic/pdqvar.h #define	PDQ_OS_MEMALLOC_CONTIG(n)	uvm_km_alloc(kernel_map, round_page(n))
kernel_map         98 dev/ic/pdqvar.h #define	PDQ_OS_MEMFREE_CONTIG(p, n)	uvm_km_free(kernel_map, (vaddr_t) p, n)
kernel_map        269 dev/ic/sti.c   	if (!(scr->scr_code = uvm_km_alloc(kernel_map, round_page(size)))) {
kernel_map        321 dev/ic/sti.c   	if ((error = uvm_map_protect(kernel_map, scr->scr_code,
kernel_map        324 dev/ic/sti.c   		uvm_km_free(kernel_map, scr->scr_code, round_page(size));
kernel_map        338 dev/ic/sti.c   			uvm_km_free(kernel_map, scr->scr_code,
kernel_map       1126 dev/isa/aha.c  	    uvm_map(kernel_map, &va, size, NULL, UVM_UNKNOWN_OFFSET, 0,
kernel_map        541 dev/ramdisk.c  	addr = uvm_km_zalloc(kernel_map, size);
kernel_map        792 kern/kern_exec.c 		va = vm_map_min(kernel_map);	/* hint */
kernel_map        793 kern/kern_exec.c 		if ((r = uvm_map(kernel_map, &va, round_page(sz), e->e_sigobject,
kernel_map        800 kern/kern_exec.c 		uvm_unmap(kernel_map, va, va + round_page(sz));
kernel_map        211 kern/kern_fork.c 	uaddr = uvm_km_alloc1(kernel_map, USPACE, USPACE_ALIGN, 1);
kernel_map        262 kern/kern_lkm.c 		uvm_km_free(kernel_map, (vaddr_t)curp->syms, curp->sym_size);
kernel_map        270 kern/kern_lkm.c 		uvm_km_free(kernel_map, curp->area, curp->size);
kernel_map        342 kern/kern_lkm.c 		curp->area = uvm_km_zalloc(kernel_map, curp->size);
kernel_map        349 kern/kern_lkm.c 			curp->syms = (caddr_t)uvm_km_zalloc(kernel_map,
kernel_map        519 kern/kern_malloc.c 	base = vm_map_min(kernel_map);
kernel_map        520 kern/kern_malloc.c 	kmem_map = uvm_km_suballoc(kernel_map, &base, &limit,
kernel_map        525 kern/kern_malloc.c 	kmemusage = (struct kmemusage *) uvm_km_zalloc(kernel_map,
kernel_map        250 kern/kern_malloc_debug.c 	offset = va - vm_map_min(kernel_map);
kernel_map       2028 kern/subr_pool.c 	return ((void *)uvm_km_alloc_poolpage1(kernel_map, uvm.kernel_object,
kernel_map       2037 kern/subr_pool.c 	uvm_km_free_poolpage1(kernel_map, (vaddr_t)v);
kernel_map         84 kern/subr_prof.c 	cp = (char *)uvm_km_zalloc(kernel_map, round_page(size));
kernel_map        173 kern/sys_pipe.c 	buffer = (caddr_t)uvm_km_valloc(kernel_map, size);
kernel_map        717 kern/sys_pipe.c 		uvm_km_free(kernel_map, (vaddr_t)cpipe->pipe_buffer.buffer,
kernel_map        346 kern/vfs_bio.c 	minaddr = vm_map_min(kernel_map);
kernel_map        347 kern/vfs_bio.c 	buf_map = uvm_km_suballoc(kernel_map, &minaddr, &maxaddr,
kernel_map        410 uvm/uvm_extern.h extern struct vm_map *kernel_map;
kernel_map        115 uvm/uvm_glue.c 	vm_map_lock_read(kernel_map);
kernel_map        116 uvm/uvm_glue.c 	rv = uvm_map_checkprot(kernel_map, saddr, eaddr, prot);
kernel_map        117 uvm/uvm_glue.c 	vm_map_unlock_read(kernel_map);
kernel_map        285 uvm/uvm_glue.c 	uvm_km_free(kernel_map, (vaddr_t)p->p_addr, USPACE);
kernel_map        159 uvm/uvm_init.c 	if (uvm_map(kernel_map, &kvm_start, 3 * PAGE_SIZE,
kernel_map        166 uvm/uvm_init.c 	if (uvm_map(kernel_map, &kvm_start, 3 * PAGE_SIZE,
kernel_map        110 uvm/uvm_io.c   		error = uvm_map_extract(map, baseva, chunksz, kernel_map, &kva,
kernel_map        141 uvm/uvm_io.c   		vm_map_lock(kernel_map);
kernel_map        142 uvm/uvm_io.c   		uvm_unmap_remove(kernel_map, kva, kva+chunksz,
kernel_map        144 uvm/uvm_io.c   		vm_map_unlock(kernel_map);
kernel_map        147 uvm/uvm_km.c   struct vm_map *kernel_map = NULL;
kernel_map        194 uvm/uvm_km.c   	kernel_map = &kernel_map_store;
kernel_map        383 uvm/uvm_km.c   		offset = kva - vm_map_min(kernel_map);
kernel_map        504 uvm/uvm_km.c   	offset = kva - vm_map_min(kernel_map);
kernel_map        817 uvm/uvm_km.c   		page = (void *)uvm_km_alloc(kernel_map, PAGE_SIZE);
kernel_map        852 uvm/uvm_km.c   			page = (void *)uvm_km_alloc(kernel_map, PAGE_SIZE);
kernel_map        419 uvm/uvm_map.c  	} else if (map == kernel_map) {
kernel_map        430 uvm/uvm_map.c  	    ((map->flags & VM_MAP_INTRSAFE) != 0 || map == kernel_map), 0, 0);
kernel_map        762 uvm/uvm_map.c  		if (map == kernel_map && uvm_maxkaddr < (*startp + size))
kernel_map        790 uvm/uvm_map.c  			uoffset = *startp - vm_map_min(kernel_map);
kernel_map       1498 uvm/uvm_map.c  			    entry->start - vm_map_min(kernel_map),
kernel_map       1499 uvm/uvm_map.c  			    entry->end - vm_map_min(kernel_map));
kernel_map        668 uvm/uvm_page.c 		pgs = (vm_page *)uvm_km_alloc(kernel_map,
kernel_map        808 uvm/uvm_page.c 	newbuckets = (struct pglist *) uvm_km_alloc(kernel_map, newsize);
kernel_map        844 uvm/uvm_page.c 		uvm_km_free(kernel_map, (vaddr_t) oldbuckets, oldsize);
kernel_map         84 uvm/uvm_pager.c 	pager_map = uvm_km_suballoc(kernel_map, &uvm.pager_sva, &uvm.pager_eva,
kernel_map         88 uvm/uvm_pager.c 	emergva = uvm_km_valloc(kernel_map, MAXBSIZE);
kernel_map       1221 xfs/xfs_vnodeops-bsd.c     kernel_upl_map(kernel_map, ap->a_pl, &iov.iov_base);
kernel_map       1239 xfs/xfs_vnodeops-bsd.c     kernel_upl_unmap(kernel_map, ap->a_pl);
kernel_map       1267 xfs/xfs_vnodeops-bsd.c     kernel_upl_map(kernel_map, ap->a_pl, &iov.iov_base);
kernel_map       1281 xfs/xfs_vnodeops-bsd.c     kernel_upl_unmap(kernel_map, ap->a_pl);