kernel_object    2028 kern/subr_pool.c 	return ((void *)uvm_km_alloc_poolpage1(kernel_map, uvm.kernel_object,
kernel_object     120 uvm/uvm.h      	struct uvm_object *kernel_object;
kernel_object     175 uvm/uvm_km.c   	uvm.kernel_object = uao_create(VM_MAX_KERNEL_ADDRESS -
kernel_object     493 uvm/uvm_km.c   	if (__predict_false(uvm_map(map, &kva, size, uvm.kernel_object,
kernel_object     513 uvm/uvm_km.c   		simple_lock(&uvm.kernel_object->vmobjlock);
kernel_object     514 uvm/uvm_km.c   		pg = uvm_pagelookup(uvm.kernel_object, offset);
kernel_object     524 uvm/uvm_km.c   			UVM_UNLOCK_AND_WAIT(pg, &uvm.kernel_object->vmobjlock,
kernel_object     530 uvm/uvm_km.c   		pg = uvm_pagealloc(uvm.kernel_object, offset, NULL, 0);
kernel_object     535 uvm/uvm_km.c   		simple_unlock(&uvm.kernel_object->vmobjlock);
kernel_object     604 uvm/uvm_km.c   	if (__predict_false(uvm_map(map, &kva, size, uvm.kernel_object,
kernel_object     644 uvm/uvm_km.c   		if (__predict_true(uvm_map(map, &kva, size, uvm.kernel_object,
kernel_object     259 uvm/uvm_stat.c 	(*pr)("    objs(kern)=%p\n", uvm.kernel_object);