kernel_map 54 arch/i386/i386/acpi_machdep.c vaddr_t va = uvm_km_valloc(kernel_map, endpa - pgpa); kernel_map 77 arch/i386/i386/acpi_machdep.c uvm_km_free(kernel_map, handle->baseva, handle->vsize); kernel_map 250 arch/i386/i386/bios.c eva = uvm_km_valloc(kernel_map, end-pa); kernel_map 503 arch/i386/i386/bios.c sva = va = uvm_km_valloc(kernel_map, endpa); kernel_map 232 arch/i386/i386/cpu.c kstack = uvm_km_alloc(kernel_map, USPACE); kernel_map 518 arch/i386/i386/cpu.c ci->ci_doubleflt_stack = (char *)uvm_km_alloc(kernel_map, USPACE); kernel_map 535 arch/i386/i386/cpu.c ci->ci_ddbipi_stack = (char *)uvm_km_alloc(kernel_map, USPACE); kernel_map 138 arch/i386/i386/gdt.c gdt = (union descriptor *)uvm_km_valloc(kernel_map, max_len); kernel_map 164 arch/i386/i386/gdt.c ci->ci_gdt = (union descriptor *)uvm_km_valloc(kernel_map, max_len); kernel_map 165 arch/i386/i386/gdt.c uvm_map_pageable(kernel_map, (vaddr_t)ci->ci_gdt, kernel_map 90 arch/i386/i386/kvm86.c if ((buf = (char *)uvm_km_zalloc(kernel_map, vmdsize)) == NULL) kernel_map 118 arch/i386/i386/kvm86.c if ((bioscallscratchpage = (void *)uvm_km_alloc(kernel_map, PAGE_SIZE)) kernel_map 125 arch/i386/i386/kvm86.c bioscalltmpva = uvm_km_alloc(kernel_map, PAGE_SIZE); kernel_map 180 arch/i386/i386/kvm86.c if ((mem = (void *)uvm_km_alloc(kernel_map, PAGE_SIZE)) == NULL) kernel_map 194 arch/i386/i386/kvm86.c uvm_km_free(kernel_map, (vaddr_t)kva, PAGE_SIZE); kernel_map 424 arch/i386/i386/machdep.c if ((v = (caddr_t)uvm_km_zalloc(kernel_map, round_page(sz))) == 0) kernel_map 439 arch/i386/i386/machdep.c minaddr = vm_map_min(kernel_map); kernel_map 440 arch/i386/i386/machdep.c exec_map = uvm_km_suballoc(kernel_map, &minaddr, &maxaddr, kernel_map 446 arch/i386/i386/machdep.c phys_map = uvm_km_suballoc(kernel_map, &minaddr, &maxaddr, kernel_map 2766 arch/i386/i386/machdep.c va = uvm_km_zalloc(kernel_map, NBPG*2); kernel_map 2812 arch/i386/i386/machdep.c cpu_ldt = (union descriptor *)uvm_km_alloc(kernel_map, len); kernel_map 3517 arch/i386/i386/machdep.c va = uvm_km_valloc(kernel_map, map_size); kernel_map 3574 arch/i386/i386/machdep.c uvm_km_free(kernel_map, va, endva - va); kernel_map 3618 arch/i386/i386/machdep.c uvm_km_free(kernel_map, va, endva - va); kernel_map 3931 arch/i386/i386/machdep.c va = uvm_km_valloc(kernel_map, size); kernel_map 3967 arch/i386/i386/machdep.c uvm_km_free(kernel_map, (vaddr_t)kva, size); kernel_map 231 arch/i386/i386/mpbios.c vaddr_t va = uvm_km_valloc(kernel_map, endpa - pgpa); kernel_map 262 arch/i386/i386/mpbios.c uvm_km_free(kernel_map, handle->baseva, handle->vsize); kernel_map 953 arch/i386/i386/pmap.c pv_initpage = (struct pv_page *) uvm_km_alloc(kernel_map, PAGE_SIZE); kernel_map 1257 arch/i386/i386/pmap.c map = kernel_map; kernel_map 1452 arch/i386/i386/pmap.c pmap->pm_pdir = (pd_entry_t *) uvm_km_alloc(kernel_map, NBPG); kernel_map 1562 arch/i386/i386/pmap.c uvm_km_free(kernel_map, (vaddr_t)pmap->pm_pdir, NBPG); kernel_map 1574 arch/i386/i386/pmap.c uvm_km_free(kernel_map, (vaddr_t)pmap->pm_ldt, kernel_map 1611 arch/i386/i386/pmap.c new_ldt = (union descriptor *)uvm_km_alloc(kernel_map, len); kernel_map 1663 arch/i386/i386/pmap.c uvm_km_free(kernel_map, (vaddr_t)old_ldt, len); kernel_map 1038 arch/i386/i386/pmapae.c pmap->pm_pdir = uvm_km_alloc(kernel_map, 4 * NBPG); kernel_map 70 arch/i386/i386/sys_machdep.c extern struct vm_map *kernel_map; kernel_map 265 arch/i386/i386/sys_machdep.c new_ldt = (union descriptor *)uvm_km_alloc(kernel_map, kernel_map 278 arch/i386/i386/sys_machdep.c uvm_km_free(kernel_map, (vaddr_t)new_ldt, new_len); kernel_map 295 arch/i386/i386/sys_machdep.c uvm_km_free(kernel_map, (vaddr_t)old_ldt, old_len); kernel_map 477 arch/i386/i386/trap.c map = kernel_map; kernel_map 482 arch/i386/i386/trap.c if (map == kernel_map && va == 0) { kernel_map 494 arch/i386/i386/trap.c if (map != kernel_map) kernel_map 302 ddb/db_break.c ((map1 == NULL) && (map2 == kernel_map)) || kernel_map 303 ddb/db_break.c ((map1 == kernel_map) && (map2 == NULL))); kernel_map 313 ddb/db_break.c (map == kernel_map) || kernel_map 339 ddb/db_break.c return kernel_map; kernel_map 95 dev/ic/pdqvar.h #define PDQ_OS_MEMFREE_CONTIG(p, n) kmem_free(kernel_map, (vm_offset_t) p, n) kernel_map 97 dev/ic/pdqvar.h #define PDQ_OS_MEMALLOC_CONTIG(n) uvm_km_alloc(kernel_map, round_page(n)) kernel_map 98 dev/ic/pdqvar.h #define PDQ_OS_MEMFREE_CONTIG(p, n) uvm_km_free(kernel_map, (vaddr_t) p, n) kernel_map 269 dev/ic/sti.c if (!(scr->scr_code = uvm_km_alloc(kernel_map, round_page(size)))) { kernel_map 321 dev/ic/sti.c if ((error = uvm_map_protect(kernel_map, scr->scr_code, kernel_map 324 dev/ic/sti.c uvm_km_free(kernel_map, scr->scr_code, round_page(size)); kernel_map 338 dev/ic/sti.c uvm_km_free(kernel_map, scr->scr_code, kernel_map 1126 dev/isa/aha.c uvm_map(kernel_map, &va, size, NULL, UVM_UNKNOWN_OFFSET, 0, kernel_map 541 dev/ramdisk.c addr = uvm_km_zalloc(kernel_map, size); kernel_map 792 kern/kern_exec.c va = vm_map_min(kernel_map); /* hint */ kernel_map 793 kern/kern_exec.c if ((r = uvm_map(kernel_map, &va, round_page(sz), e->e_sigobject, kernel_map 800 kern/kern_exec.c uvm_unmap(kernel_map, va, va + round_page(sz)); kernel_map 211 kern/kern_fork.c uaddr = uvm_km_alloc1(kernel_map, USPACE, USPACE_ALIGN, 1); kernel_map 262 kern/kern_lkm.c uvm_km_free(kernel_map, (vaddr_t)curp->syms, curp->sym_size); kernel_map 270 kern/kern_lkm.c uvm_km_free(kernel_map, curp->area, curp->size); kernel_map 342 kern/kern_lkm.c curp->area = uvm_km_zalloc(kernel_map, curp->size); kernel_map 349 kern/kern_lkm.c curp->syms = (caddr_t)uvm_km_zalloc(kernel_map, kernel_map 519 kern/kern_malloc.c base = vm_map_min(kernel_map); kernel_map 520 kern/kern_malloc.c kmem_map = uvm_km_suballoc(kernel_map, &base, &limit, kernel_map 525 kern/kern_malloc.c kmemusage = (struct kmemusage *) uvm_km_zalloc(kernel_map, kernel_map 250 kern/kern_malloc_debug.c offset = va - vm_map_min(kernel_map); kernel_map 2028 kern/subr_pool.c return ((void *)uvm_km_alloc_poolpage1(kernel_map, uvm.kernel_object, kernel_map 2037 kern/subr_pool.c uvm_km_free_poolpage1(kernel_map, (vaddr_t)v); kernel_map 84 kern/subr_prof.c cp = (char *)uvm_km_zalloc(kernel_map, round_page(size)); kernel_map 173 kern/sys_pipe.c buffer = (caddr_t)uvm_km_valloc(kernel_map, size); kernel_map 717 kern/sys_pipe.c uvm_km_free(kernel_map, (vaddr_t)cpipe->pipe_buffer.buffer, kernel_map 346 kern/vfs_bio.c minaddr = vm_map_min(kernel_map); kernel_map 347 kern/vfs_bio.c buf_map = uvm_km_suballoc(kernel_map, &minaddr, &maxaddr, kernel_map 410 uvm/uvm_extern.h extern struct vm_map *kernel_map; kernel_map 115 uvm/uvm_glue.c vm_map_lock_read(kernel_map); kernel_map 116 uvm/uvm_glue.c rv = uvm_map_checkprot(kernel_map, saddr, eaddr, prot); kernel_map 117 uvm/uvm_glue.c vm_map_unlock_read(kernel_map); kernel_map 285 uvm/uvm_glue.c uvm_km_free(kernel_map, (vaddr_t)p->p_addr, USPACE); kernel_map 159 uvm/uvm_init.c if (uvm_map(kernel_map, &kvm_start, 3 * PAGE_SIZE, kernel_map 166 uvm/uvm_init.c if (uvm_map(kernel_map, &kvm_start, 3 * PAGE_SIZE, kernel_map 110 uvm/uvm_io.c error = uvm_map_extract(map, baseva, chunksz, kernel_map, &kva, kernel_map 141 uvm/uvm_io.c vm_map_lock(kernel_map); kernel_map 142 uvm/uvm_io.c uvm_unmap_remove(kernel_map, kva, kva+chunksz, kernel_map 144 uvm/uvm_io.c vm_map_unlock(kernel_map); kernel_map 147 uvm/uvm_km.c struct vm_map *kernel_map = NULL; kernel_map 194 uvm/uvm_km.c kernel_map = &kernel_map_store; kernel_map 383 uvm/uvm_km.c offset = kva - vm_map_min(kernel_map); kernel_map 504 uvm/uvm_km.c offset = kva - vm_map_min(kernel_map); kernel_map 817 uvm/uvm_km.c page = (void *)uvm_km_alloc(kernel_map, PAGE_SIZE); kernel_map 852 uvm/uvm_km.c page = (void *)uvm_km_alloc(kernel_map, PAGE_SIZE); kernel_map 419 uvm/uvm_map.c } else if (map == kernel_map) { kernel_map 430 uvm/uvm_map.c ((map->flags & VM_MAP_INTRSAFE) != 0 || map == kernel_map), 0, 0); kernel_map 762 uvm/uvm_map.c if (map == kernel_map && uvm_maxkaddr < (*startp + size)) kernel_map 790 uvm/uvm_map.c uoffset = *startp - vm_map_min(kernel_map); kernel_map 1498 uvm/uvm_map.c entry->start - vm_map_min(kernel_map), kernel_map 1499 uvm/uvm_map.c entry->end - vm_map_min(kernel_map)); kernel_map 668 uvm/uvm_page.c pgs = (vm_page *)uvm_km_alloc(kernel_map, kernel_map 808 uvm/uvm_page.c newbuckets = (struct pglist *) uvm_km_alloc(kernel_map, newsize); kernel_map 844 uvm/uvm_page.c uvm_km_free(kernel_map, (vaddr_t) oldbuckets, oldsize); kernel_map 84 uvm/uvm_pager.c pager_map = uvm_km_suballoc(kernel_map, &uvm.pager_sva, &uvm.pager_eva, kernel_map 88 uvm/uvm_pager.c emergva = uvm_km_valloc(kernel_map, MAXBSIZE); kernel_map 1221 xfs/xfs_vnodeops-bsd.c kernel_upl_map(kernel_map, ap->a_pl, &iov.iov_base); kernel_map 1239 xfs/xfs_vnodeops-bsd.c kernel_upl_unmap(kernel_map, ap->a_pl); kernel_map 1267 xfs/xfs_vnodeops-bsd.c kernel_upl_map(kernel_map, ap->a_pl, &iov.iov_base); kernel_map 1281 xfs/xfs_vnodeops-bsd.c kernel_upl_unmap(kernel_map, ap->a_pl);