pg                127 arch/i386/i386/gdt.c 	struct vm_page *pg;
pg                140 arch/i386/i386/gdt.c 		pg = uvm_pagealloc(NULL, 0, NULL, UVM_PGA_ZERO);
pg                141 arch/i386/i386/gdt.c 		if (pg == NULL)
pg                143 arch/i386/i386/gdt.c 		pmap_kenter_pa(va, VM_PAGE_TO_PHYS(pg),
pg                198 arch/i386/i386/gdt.c 	struct vm_page *pg;
pg                210 arch/i386/i386/gdt.c 			    (pg =
pg                215 arch/i386/i386/gdt.c 			pmap_kenter_pa(va, VM_PAGE_TO_PHYS(pg),
pg                159 arch/i386/i386/mpbios.c 	paddr_t 	pg;
pg                235 arch/i386/i386/mpbios.c 	handle->pg = pgpa;
pg                416 arch/i386/i386/pmap.c pmap_sync_flags_pte(struct vm_page *pg, u_long pte)
pg                419 arch/i386/i386/pmap.c 		atomic_setbits_int(&pg->pg_flags, pmap_pte2flags(pte));
pg               1050 arch/i386/i386/pmap.c 	struct vm_page *pg;
pg               1093 arch/i386/i386/pmap.c 	pg = uvm_pagealloc(NULL, 0, NULL, UVM_PGA_USERESERVE);
pg               1094 arch/i386/i386/pmap.c 	if (pg == NULL)
pg               1097 arch/i386/i386/pmap.c 	atomic_clearbits_int(&pg->pg_flags, PG_BUSY);
pg               1106 arch/i386/i386/pmap.c 	pmap_kenter_pa(pv_cachedva, VM_PAGE_TO_PHYS(pg),
pg               1302 arch/i386/i386/pmap.c pmap_enter_pv(struct vm_page *pg, struct pv_entry *pve, struct pmap *pmap,
pg               1308 arch/i386/i386/pmap.c 	pve->pv_next = pg->mdpage.pv_list;	/* add to ... */
pg               1309 arch/i386/i386/pmap.c 	pg->mdpage.pv_list = pve;			/* ... locked list */
pg               1322 arch/i386/i386/pmap.c pmap_remove_pv(struct vm_page *pg, struct pmap *pmap, vaddr_t va)
pg               1326 arch/i386/i386/pmap.c 	prevptr = &pg->mdpage.pv_list;		/* previous pv_entry pointer */
pg               1527 arch/i386/i386/pmap.c 	struct vm_page *pg;
pg               1547 arch/i386/i386/pmap.c 		pg = TAILQ_FIRST(&pmap->pm_obj.memq);
pg               1549 arch/i386/i386/pmap.c 		if (pg->pg_flags & PG_BUSY)
pg               1554 arch/i386/i386/pmap.c 		pg->wire_count = 0;
pg               1555 arch/i386/i386/pmap.c 		uvm_pagefree(pg);
pg               1771 arch/i386/i386/pmap.c pmap_zero_page(struct vm_page *pg)
pg               1773 arch/i386/i386/pmap.c 	pmap_zero_phys(VM_PAGE_TO_PHYS(pg));
pg               1878 arch/i386/i386/pmap.c 	struct vm_page *pg;
pg               1911 arch/i386/i386/pmap.c 		pg = PHYS_TO_VM_PAGE(opte & PG_FRAME);
pg               1918 arch/i386/i386/pmap.c 			if (pg != NULL)
pg               1926 arch/i386/i386/pmap.c 		if (pg == NULL)
pg               1933 arch/i386/i386/pmap.c 		pmap_sync_flags_pte(pg, opte);
pg               1934 arch/i386/i386/pmap.c 		pve = pmap_remove_pv(pg, pmap, startva);
pg               1962 arch/i386/i386/pmap.c 	struct vm_page *pg;
pg               1983 arch/i386/i386/pmap.c 	pg = PHYS_TO_VM_PAGE(opte & PG_FRAME);
pg               1990 arch/i386/i386/pmap.c 		if (pg != NULL)
pg               1998 arch/i386/i386/pmap.c 	if (pg == NULL)
pg               2004 arch/i386/i386/pmap.c 	pmap_sync_flags_pte(pg, opte);
pg               2005 arch/i386/i386/pmap.c 	pve = pmap_remove_pv(pg, pmap, va);
pg               2240 arch/i386/i386/pmap.c pmap_page_remove(struct vm_page *pg)
pg               2247 arch/i386/i386/pmap.c 	if (pg->mdpage.pv_list == NULL)
pg               2254 arch/i386/i386/pmap.c 	for (pve = pg->mdpage.pv_list ; pve != NULL ; pve = pve->pv_next) {
pg               2264 arch/i386/i386/pmap.c 			       pg, pve->pv_va, pve->pv_ptp);
pg               2281 arch/i386/i386/pmap.c 		pmap_sync_flags_pte(pg, opte);
pg               2317 arch/i386/i386/pmap.c 	pmap_free_pvs(NULL, pg->mdpage.pv_list);
pg               2318 arch/i386/i386/pmap.c 	pg->mdpage.pv_list = NULL;
pg               2340 arch/i386/i386/pmap.c pmap_test_attrs(struct vm_page *pg, int testbits)
pg               2348 arch/i386/i386/pmap.c 	if (pg->pg_flags & testflags)
pg               2353 arch/i386/i386/pmap.c 	for (pve = pg->mdpage.pv_list; pve != NULL && mybits == 0;
pg               2365 arch/i386/i386/pmap.c 	atomic_setbits_int(&pg->pg_flags, pmap_pte2flags(mybits));
pg               2377 arch/i386/i386/pmap.c pmap_clear_attrs(struct vm_page *pg, int clearbits)
pg               2388 arch/i386/i386/pmap.c 	result = pg->pg_flags & clearflags;
pg               2390 arch/i386/i386/pmap.c 		atomic_clearbits_int(&pg->pg_flags, clearflags);
pg               2392 arch/i386/i386/pmap.c 	for (pve = pg->mdpage.pv_list; pve != NULL; pve = pve->pv_next) {
pg               2600 arch/i386/i386/pmap.c 	struct vm_page *pg = NULL;
pg               2665 arch/i386/i386/pmap.c 				pg = PHYS_TO_VM_PAGE(pa);
pg               2667 arch/i386/i386/pmap.c 				if (pg == NULL)
pg               2673 arch/i386/i386/pmap.c 				pmap_sync_flags_pte(pg, opte);
pg               2688 arch/i386/i386/pmap.c 			pg = PHYS_TO_VM_PAGE(opte & PG_FRAME);
pg               2690 arch/i386/i386/pmap.c 			if (pg == NULL)
pg               2695 arch/i386/i386/pmap.c 			pmap_sync_flags_pte(pg, opte);
pg               2696 arch/i386/i386/pmap.c 			pve = pmap_remove_pv(pg, pmap, va);
pg               2697 arch/i386/i386/pmap.c 			pg = NULL; /* This is not page we are looking for */
pg               2715 arch/i386/i386/pmap.c 	if (pmap_initialized && pg == NULL)
pg               2716 arch/i386/i386/pmap.c 		pg = PHYS_TO_VM_PAGE(pa);
pg               2718 arch/i386/i386/pmap.c 	if (pg != NULL) {
pg               2733 arch/i386/i386/pmap.c 		pmap_enter_pv(pg, pve, pmap, va, ptp);
pg               2760 arch/i386/i386/pmap.c 	if (pg) {
pg               2762 arch/i386/i386/pmap.c 		pmap_sync_flags_pte(pg, npte);
pg               1570 arch/i386/i386/pmapae.c pmap_page_remove_pae(struct vm_page *pg)
pg               1581 arch/i386/i386/pmapae.c 	bank = vm_physseg_find(atop(VM_PAGE_TO_PHYS(pg)), &off);
pg               1610 arch/i386/i386/pmapae.c 			       pg, pve->pv_va, pve->pv_ptp);
pg               1700 arch/i386/i386/pmapae.c pmap_test_attrs_pae(struct vm_page *pg, int testbits)
pg               1709 arch/i386/i386/pmapae.c 	bank = vm_physseg_find(atop(VM_PAGE_TO_PHYS(pg)), &off);
pg               1761 arch/i386/i386/pmapae.c pmap_change_attrs_pae(struct vm_page *pg, int setbits, int clearbits)
pg               1772 arch/i386/i386/pmapae.c 	bank = vm_physseg_find(atop(VM_PAGE_TO_PHYS(pg)), &off);
pg                353 arch/i386/include/pmap.h #define pmap_clear_modify(pg)		pmap_clear_attrs(pg, PG_M)
pg                354 arch/i386/include/pmap.h #define pmap_clear_reference(pg)	pmap_clear_attrs(pg, PG_U)
pg                356 arch/i386/include/pmap.h #define pmap_is_modified(pg)		pmap_test_attrs(pg, PG_M)
pg                357 arch/i386/include/pmap.h #define pmap_is_referenced(pg)		pmap_test_attrs(pg, PG_U)
pg                399 arch/i386/include/pmap.h #define	PMAP_PAGEIDLEZERO(pg)	pmap_zero_page_uncached(VM_PAGE_TO_PHYS(pg))
pg                428 arch/i386/include/pmap.h pmap_page_protect(pg, prot)
pg                429 arch/i386/include/pmap.h 	struct vm_page *pg;
pg                434 arch/i386/include/pmap.h 			(void) pmap_clear_attrs(pg, PG_RW);
pg                436 arch/i386/include/pmap.h 			pmap_page_remove(pg);
pg                123 arch/i386/include/vmparam.h #define VM_MDPAGE_INIT(pg) do {			\
pg                124 arch/i386/include/vmparam.h 	(pg)->mdpage.pv_list = NULL;	\
pg                249 dev/ic/lemac.c 	int pg, conf;
pg                263 dev/ic/lemac.c 	for (pg = 1; pg <= sc->sc_lastpage; pg++)
pg                264 dev/ic/lemac.c 		LEMAC_OUTB(sc, LEMAC_REG_FMQ, pg);
pg               1046 dev/isa/aha.c  	struct vm_page *pg;
pg               1132 dev/isa/aha.c  	for (pg = TAILQ_FIRST(&pglist); pg != NULL;
pg               1133 dev/isa/aha.c  	    pg = TAILQ_NEXT(pg, pageq)) {
pg               1134 dev/isa/aha.c  		pmap_kenter_pa(va, VM_PAGE_TO_PHYS(pg),
pg                581 dev/isa/if_ie.c         u_short pg, adjust, decode, edecode;
pg                694 dev/isa/if_ie.c 	pg = (kvtop(sc->sc_maddr) & 0x3C000) >> 14;
pg                695 dev/isa/if_ie.c 	adjust = IEE16_MCTRL_FMCS16 | (pg & 0x3) << 2;
pg                696 dev/isa/if_ie.c 	decode = ((1 << (sc->sc_msize / 16384)) - 1) << pg;
pg                293 kern/kern_ktrace.c 	struct pgrp *pg;
pg                350 kern/kern_ktrace.c 		pg = pgfind(-SCARG(uap, pid));
pg                351 kern/kern_ktrace.c 		if (pg == NULL) {
pg                355 kern/kern_ktrace.c 		LIST_FOREACH(p, &pg->pg_members, p_pglist)
pg                234 kern/kern_malloc_debug.c 	struct vm_page *pg;
pg                252 kern/kern_malloc_debug.c 		pg = uvm_pagealloc(NULL, 0, NULL, 0);
pg                253 kern/kern_malloc_debug.c 		if (pg) {
pg                254 kern/kern_malloc_debug.c 			atomic_clearbits_int(&pg->pg_flags, PG_BUSY);
pg                255 kern/kern_malloc_debug.c 			UVM_PAGE_OWN(pg, NULL);
pg                258 kern/kern_malloc_debug.c 		if (pg)
pg                270 kern/kern_malloc_debug.c 	md->md_pa = VM_PAGE_TO_PHYS(pg);
pg                348 kern/kern_proc.c orphanpg(struct pgrp *pg)
pg                352 kern/kern_proc.c 	LIST_FOREACH(p, &pg->pg_members, p_pglist) {
pg                354 kern/kern_proc.c 			LIST_FOREACH(p, &pg->pg_members, p_pglist) {
pg                 87 kern/kern_resource.c 		struct pgrp *pg;
pg                 90 kern/kern_resource.c 			pg = curp->p_pgrp;
pg                 91 kern/kern_resource.c 		else if ((pg = pgfind(SCARG(uap, who))) == NULL)
pg                 93 kern/kern_resource.c 		LIST_FOREACH(p, &pg->pg_members, p_pglist) {
pg                144 kern/kern_resource.c 		struct pgrp *pg;
pg                147 kern/kern_resource.c 			pg = curp->p_pgrp;
pg                148 kern/kern_resource.c 		else if ((pg = pgfind(SCARG(uap, who))) == NULL)
pg                150 kern/kern_resource.c 		LIST_FOREACH(p, &pg->pg_members, p_pglist) {
pg                282 scsi/scsi_disk.h #define	DISK_PGCODE(pg, n)	((pg) != NULL) && (((pg)->pg_code & 0x3f) == n)
pg                755 uvm/uvm_amap.c 	struct vm_page *pg, *npg;
pg                772 uvm/uvm_amap.c 		pg = anon->an_page;
pg                778 uvm/uvm_amap.c 		if (pg == NULL)
pg                792 uvm/uvm_amap.c 		if (anon->an_ref > 1 && pg->loan_count == 0) {
pg                798 uvm/uvm_amap.c 			if (pg->pg_flags & PG_BUSY) {
pg                799 uvm/uvm_amap.c 				atomic_setbits_int(&pg->pg_flags, PG_WANTED);
pg                800 uvm/uvm_amap.c 				UVM_UNLOCK_AND_WAIT(pg, &anon->an_lock, FALSE,
pg                833 uvm/uvm_amap.c 			uvm_pagecopy(pg, npg);		/* old -> new */
pg                 94 uvm/uvm_anon.c 	struct vm_page *pg;
pg                102 uvm/uvm_anon.c 	pg = anon->an_page;
pg                110 uvm/uvm_anon.c 	if (pg && pg->loan_count)
pg                111 uvm/uvm_anon.c 		pg = uvm_anon_lockloanpg(anon);
pg                118 uvm/uvm_anon.c 	if (pg) {
pg                125 uvm/uvm_anon.c 		if (pg->uobject) {
pg                127 uvm/uvm_anon.c 			KASSERT(pg->loan_count > 0);
pg                128 uvm/uvm_anon.c 			pg->loan_count--;
pg                129 uvm/uvm_anon.c 			pg->uanon = NULL;
pg                131 uvm/uvm_anon.c 			simple_unlock(&pg->uobject->vmobjlock);
pg                143 uvm/uvm_anon.c 			if ((pg->pg_flags & PG_BUSY) != 0) {
pg                145 uvm/uvm_anon.c 				atomic_setbits_int(&pg->pg_flags, PG_RELEASED);
pg                148 uvm/uvm_anon.c 				    anon, pg, 0, 0);
pg                151 uvm/uvm_anon.c 			pmap_page_protect(pg, VM_PROT_NONE);
pg                153 uvm/uvm_anon.c 			uvm_pagefree(pg);	/* bye bye */
pg                156 uvm/uvm_anon.c 			    anon, pg, 0, 0);
pg                224 uvm/uvm_anon.c 	struct vm_page *pg;
pg                236 uvm/uvm_anon.c 	while (((pg = anon->an_page) != NULL) && pg->loan_count != 0) {
pg                247 uvm/uvm_anon.c 		if (pg->uobject) {
pg                250 uvm/uvm_anon.c 			if (pg->uobject) {	/* the "real" check */
pg                252 uvm/uvm_anon.c 				    simple_lock_try(&pg->uobject->vmobjlock);
pg                282 uvm/uvm_anon.c 		if (pg->uobject == NULL && (pg->pg_flags & PQ_ANON) == 0) {
pg                284 uvm/uvm_anon.c 			atomic_setbits_int(&pg->pg_flags, PQ_ANON);
pg                285 uvm/uvm_anon.c 			pg->loan_count--;	/* ... and drop our loan */
pg                295 uvm/uvm_anon.c 	return(pg);
pg                309 uvm/uvm_anon.c 	struct vm_page *pg;
pg                348 uvm/uvm_anon.c 	pg = anon->an_page;
pg                349 uvm/uvm_anon.c 	uobj = pg->uobject;
pg                352 uvm/uvm_anon.c 	atomic_clearbits_int(&pg->pg_flags, PG_CLEAN);
pg                358 uvm/uvm_anon.c 	pmap_clear_reference(pg);
pg                359 uvm/uvm_anon.c 	pmap_page_protect(pg, VM_PROT_NONE);
pg                361 uvm/uvm_anon.c 	uvm_pagedeactivate(pg);
pg                661 uvm/uvm_aobj.c 	struct vm_page *pg, *next;
pg                693 uvm/uvm_aobj.c 	for (pg = TAILQ_FIRST(&uobj->memq); pg != NULL; pg = next) {
pg                694 uvm/uvm_aobj.c 		next = TAILQ_NEXT(pg, listq);
pg                695 uvm/uvm_aobj.c 		if (pg->pg_flags & PG_BUSY) {
pg                696 uvm/uvm_aobj.c 			atomic_setbits_int(&pg->pg_flags, PG_RELEASED);
pg                702 uvm/uvm_aobj.c 		pmap_page_protect(pg, VM_PROT_NONE);
pg                703 uvm/uvm_aobj.c 		uao_dropswap(&aobj->u_obj, pg->offset >> PAGE_SHIFT);
pg                705 uvm/uvm_aobj.c 		uvm_pagefree(pg);
pg               1260 uvm/uvm_aobj.c uao_releasepg(pg, nextpgp)
pg               1261 uvm/uvm_aobj.c 	struct vm_page *pg;
pg               1264 uvm/uvm_aobj.c 	struct uvm_aobj *aobj = (struct uvm_aobj *) pg->uobject;
pg               1266 uvm/uvm_aobj.c 	KASSERT(pg->pg_flags & PG_RELEASED);
pg               1271 uvm/uvm_aobj.c 	pmap_page_protect(pg, VM_PROT_NONE);
pg               1272 uvm/uvm_aobj.c 	uao_dropswap(&aobj->u_obj, pg->offset >> PAGE_SHIFT);
pg               1275 uvm/uvm_aobj.c 		*nextpgp = TAILQ_NEXT(pg, pageq); /* next page for daemon */
pg               1276 uvm/uvm_aobj.c 	uvm_pagefree(pg);
pg               1484 uvm/uvm_aobj.c 	struct vm_page *pg;
pg               1487 uvm/uvm_aobj.c 	pg = NULL;
pg               1491 uvm/uvm_aobj.c 		     &pg, &npages, 0, VM_PROT_READ|VM_PROT_WRITE, 0, 0);
pg               1513 uvm/uvm_aobj.c 	KASSERT((pg->pg_flags & PG_RELEASED) == 0);
pg               1521 uvm/uvm_aobj.c 	atomic_clearbits_int(&pg->pg_flags, PG_BUSY|PG_CLEAN|PG_FAKE);
pg               1522 uvm/uvm_aobj.c 	UVM_PAGE_OWN(pg, NULL);
pg               1527 uvm/uvm_aobj.c 	pmap_clear_reference(pg);
pg               1529 uvm/uvm_aobj.c 	pmap_page_protect(pg, VM_PROT_NONE);
pg               1532 uvm/uvm_aobj.c 	uvm_pagedeactivate(pg);
pg                197 uvm/uvm_fault.c 	struct vm_page *pg;
pg                203 uvm/uvm_fault.c 		pg = anons[lcv]->an_page;
pg                204 uvm/uvm_fault.c 		if (pg && (pg->pg_flags & PG_BUSY) == 0 && pg->loan_count == 0) {
pg                206 uvm/uvm_fault.c 			if (pg->wire_count == 0) {
pg                208 uvm/uvm_fault.c 				pmap_clear_reference(pg);
pg                210 uvm/uvm_fault.c 				pmap_page_protect(pg, VM_PROT_NONE);
pg                212 uvm/uvm_fault.c 				uvm_pagedeactivate(pg);
pg                302 uvm/uvm_fault.c 	struct vm_page *pg;
pg                321 uvm/uvm_fault.c 		pg = anon->an_page;
pg                329 uvm/uvm_fault.c 		if (pg && pg->loan_count)
pg                330 uvm/uvm_fault.c 			pg = uvm_anon_lockloanpg(anon);
pg                336 uvm/uvm_fault.c 		if (pg) {
pg                345 uvm/uvm_fault.c 			if ((pg->pg_flags & (PG_BUSY|PG_RELEASED)) == 0) {
pg                349 uvm/uvm_fault.c 			atomic_setbits_int(&pg->pg_flags, PG_WANTED);
pg                356 uvm/uvm_fault.c 			if (pg->uobject) {	/* owner is uobject ? */
pg                360 uvm/uvm_fault.c 				UVM_UNLOCK_AND_WAIT(pg,
pg                361 uvm/uvm_fault.c 				    &pg->uobject->vmobjlock,
pg                368 uvm/uvm_fault.c 				UVM_UNLOCK_AND_WAIT(pg,&anon->an_lock,0,
pg                378 uvm/uvm_fault.c 			pg = uvm_pagealloc(NULL, 0, anon, 0);
pg                380 uvm/uvm_fault.c 			if (pg == NULL) {		/* out of RAM.  */
pg                403 uvm/uvm_fault.c 				result = uvm_swap_get(pg, anon->an_swslot,
pg                435 uvm/uvm_fault.c 			if (pg->pg_flags & PG_WANTED) {
pg                437 uvm/uvm_fault.c 				wakeup(pg);	
pg                440 uvm/uvm_fault.c 			atomic_clearbits_int(&pg->pg_flags,
pg                442 uvm/uvm_fault.c 			UVM_PAGE_OWN(pg, NULL);
pg                449 uvm/uvm_fault.c 			if (pg->pg_flags & PG_RELEASED) {
pg                450 uvm/uvm_fault.c 				pmap_page_protect(pg, VM_PROT_NONE);
pg                482 uvm/uvm_fault.c 				uvm_pagefree(pg);
pg                498 uvm/uvm_fault.c 			pmap_clear_modify(pg);
pg                500 uvm/uvm_fault.c 			uvm_pageactivate(pg);
pg                575 uvm/uvm_fault.c 	struct vm_page *pages[UVM_MAXRANGE], *pg, *uobjpage;
pg                582 uvm/uvm_fault.c 	pg = NULL;
pg               1109 uvm/uvm_fault.c 				pg = uvm_pagealloc(NULL, 0, NULL, 0);
pg               1110 uvm/uvm_fault.c 				if (pg == NULL) {
pg               1122 uvm/uvm_fault.c 				uvm_pagecopy(anon->an_page, pg);
pg               1135 uvm/uvm_fault.c 				uvm_pageactivate(pg);
pg               1143 uvm/uvm_fault.c 				anon->an_page = pg;
pg               1144 uvm/uvm_fault.c 				pg->uanon = anon;
pg               1145 uvm/uvm_fault.c 				atomic_setbits_int(&pg->pg_flags, PQ_ANON);
pg               1146 uvm/uvm_fault.c 				atomic_clearbits_int(&pg->pg_flags,
pg               1148 uvm/uvm_fault.c 				UVM_PAGE_OWN(pg, NULL);
pg               1175 uvm/uvm_fault.c 			pg = uvm_pagealloc(NULL, 0, anon, 0);
pg               1179 uvm/uvm_fault.c 		if (anon == NULL || pg == NULL) {
pg               1198 uvm/uvm_fault.c 		uvm_pagecopy(oanon->an_page, pg);	/* pg now !PG_CLEAN */
pg               1200 uvm/uvm_fault.c 		atomic_clearbits_int(&pg->pg_flags, PG_BUSY|PG_FAKE);
pg               1201 uvm/uvm_fault.c 		UVM_PAGE_OWN(pg, NULL);
pg               1218 uvm/uvm_fault.c 		pg = anon->an_page;
pg               1234 uvm/uvm_fault.c 	    ufi.orig_map->pmap, ufi.orig_rvaddr, pg, 0);
pg               1235 uvm/uvm_fault.c 	if (pmap_enter(ufi.orig_map->pmap, ufi.orig_rvaddr, VM_PAGE_TO_PHYS(pg),
pg               1265 uvm/uvm_fault.c 		uvm_pagewire(pg);
pg               1273 uvm/uvm_fault.c 		atomic_clearbits_int(&pg->pg_flags, PG_CLEAN);
pg               1277 uvm/uvm_fault.c 		uvm_pageactivate(pg);
pg               1468 uvm/uvm_fault.c 		pg = uobjpage;		/* map in the actual object */
pg               1486 uvm/uvm_fault.c 				pg = uvm_pagealloc(NULL, 0, NULL, 0);
pg               1488 uvm/uvm_fault.c 				if (pg == NULL) {
pg               1523 uvm/uvm_fault.c 				uvm_pagecopy(uobjpage, pg);	/* old -> new */
pg               1524 uvm/uvm_fault.c 				atomic_clearbits_int(&pg->pg_flags,
pg               1544 uvm/uvm_fault.c 				uvm_pagerealloc(pg, uobj, uoff);
pg               1552 uvm/uvm_fault.c 				uobjpage = pg;
pg               1576 uvm/uvm_fault.c 			pg = uvm_pagealloc(NULL, 0, anon,
pg               1583 uvm/uvm_fault.c 		if (anon == NULL || pg == NULL) {
pg               1626 uvm/uvm_fault.c 			uvm_pagecopy(uobjpage, pg);
pg               1656 uvm/uvm_fault.c 			    uobjpage, anon, pg, 0);
pg               1665 uvm/uvm_fault.c 			    anon, pg, 0, 0);
pg               1686 uvm/uvm_fault.c 	    ufi.orig_map->pmap, ufi.orig_rvaddr, pg, promote);
pg               1687 uvm/uvm_fault.c 	if (pmap_enter(ufi.orig_map->pmap, ufi.orig_rvaddr, VM_PAGE_TO_PHYS(pg),
pg               1699 uvm/uvm_fault.c 		if (pg->pg_flags & PG_WANTED)
pg               1700 uvm/uvm_fault.c 			wakeup(pg);		/* lock still held */
pg               1707 uvm/uvm_fault.c 		atomic_clearbits_int(&pg->pg_flags, PG_BUSY|PG_FAKE|PG_WANTED);
pg               1708 uvm/uvm_fault.c 		UVM_PAGE_OWN(pg, NULL);
pg               1725 uvm/uvm_fault.c 		uvm_pagewire(pg);
pg               1726 uvm/uvm_fault.c 		if (pg->pg_flags & PQ_AOBJ) {
pg               1734 uvm/uvm_fault.c 			atomic_clearbits_int(&pg->pg_flags, PG_CLEAN);
pg               1735 uvm/uvm_fault.c 			uao_dropswap(uobj, pg->offset >> PAGE_SHIFT);
pg               1739 uvm/uvm_fault.c 		uvm_pageactivate(pg);
pg               1743 uvm/uvm_fault.c 	if (pg->pg_flags & PG_WANTED)
pg               1744 uvm/uvm_fault.c 		wakeup(pg);		/* lock still held */
pg               1751 uvm/uvm_fault.c 	atomic_clearbits_int(&pg->pg_flags, PG_BUSY|PG_FAKE|PG_WANTED);
pg               1752 uvm/uvm_fault.c 	UVM_PAGE_OWN(pg, NULL);
pg               1831 uvm/uvm_fault.c 	struct vm_page *pg;
pg               1871 uvm/uvm_fault.c 		pg = PHYS_TO_VM_PAGE(pa);
pg               1872 uvm/uvm_fault.c 		if (pg)
pg               1873 uvm/uvm_fault.c 			uvm_pageunwire(pg);
pg                311 uvm/uvm_km.c   	struct vm_page *pg;
pg                318 uvm/uvm_km.c   		pg = PHYS_TO_VM_PAGE(pa);
pg                319 uvm/uvm_km.c   		if (pg == NULL)
pg                321 uvm/uvm_km.c   		uvm_pagefree(pg);
pg                344 uvm/uvm_km.c   	struct vm_page *pg;
pg                396 uvm/uvm_km.c   		pg = uvm_pagealloc(obj, offset, NULL, 0);
pg                397 uvm/uvm_km.c   		if (pg) {
pg                398 uvm/uvm_km.c   			atomic_clearbits_int(&pg->pg_flags, PG_BUSY);
pg                399 uvm/uvm_km.c   			UVM_PAGE_OWN(pg, NULL);
pg                402 uvm/uvm_km.c   		if (__predict_false(pg == NULL)) {
pg                421 uvm/uvm_km.c   			pmap_kenter_pa(loopva, VM_PAGE_TO_PHYS(pg),
pg                424 uvm/uvm_km.c   			pmap_enter(map->pmap, loopva, VM_PAGE_TO_PHYS(pg),
pg                480 uvm/uvm_km.c   	struct vm_page *pg;
pg                514 uvm/uvm_km.c   		pg = uvm_pagelookup(uvm.kernel_object, offset);
pg                520 uvm/uvm_km.c   		if (pg) {
pg                521 uvm/uvm_km.c   			if ((pg->pg_flags & PG_RELEASED) == 0)
pg                523 uvm/uvm_km.c   			atomic_setbits_int(&pg->pg_flags, PG_WANTED);
pg                524 uvm/uvm_km.c   			UVM_UNLOCK_AND_WAIT(pg, &uvm.kernel_object->vmobjlock,
pg                530 uvm/uvm_km.c   		pg = uvm_pagealloc(uvm.kernel_object, offset, NULL, 0);
pg                531 uvm/uvm_km.c   		if (pg) {
pg                532 uvm/uvm_km.c   			atomic_clearbits_int(&pg->pg_flags, PG_BUSY);
pg                533 uvm/uvm_km.c   			UVM_PAGE_OWN(pg, NULL);
pg                536 uvm/uvm_km.c   		if (__predict_false(pg == NULL)) {
pg                555 uvm/uvm_km.c   		pmap_enter(map->pmap, loopva, VM_PAGE_TO_PHYS(pg),
pg                679 uvm/uvm_km.c   	struct vm_page *pg;
pg                683 uvm/uvm_km.c   	pg = uvm_pagealloc(NULL, 0, NULL, UVM_PGA_USERESERVE);
pg                684 uvm/uvm_km.c   	if (__predict_false(pg == NULL)) {
pg                691 uvm/uvm_km.c   	va = pmap_map_direct(pg);
pg                693 uvm/uvm_km.c   		uvm_pagefree(pg);
pg                325 uvm/uvm_loan.c 	struct vm_page *pg;
pg                335 uvm/uvm_loan.c 		pg = anon->an_page;
pg                336 uvm/uvm_loan.c 		if (pg && (pg->pg_flags & PQ_ANON) != 0 && anon->an_ref == 1)
pg                338 uvm/uvm_loan.c 			pmap_page_protect(pg, VM_PROT_READ);
pg                380 uvm/uvm_loan.c 	pg = anon->an_page;
pg                382 uvm/uvm_loan.c 	if (pg->loan_count == 0)
pg                383 uvm/uvm_loan.c 		pmap_page_protect(pg, VM_PROT_READ);
pg                384 uvm/uvm_loan.c 	pg->loan_count++;
pg                385 uvm/uvm_loan.c 	uvm_pagewire(pg);	/* always wire it */
pg                387 uvm/uvm_loan.c 	**output = pg;
pg                391 uvm/uvm_loan.c 	if (pg->uobject)
pg                392 uvm/uvm_loan.c 		simple_unlock(&pg->uobject->vmobjlock);
pg                416 uvm/uvm_loan.c 	struct vm_page *pg;
pg                429 uvm/uvm_loan.c 		pg = NULL;
pg                431 uvm/uvm_loan.c 		    &pg, &npages, 0, VM_PROT_READ, MADV_NORMAL, PGO_LOCKED);
pg                456 uvm/uvm_loan.c 		    &pg, &npages, 0, VM_PROT_READ, MADV_NORMAL, 0);
pg                484 uvm/uvm_loan.c 		if ((pg->pg_flags & PG_RELEASED) != 0 ||
pg                499 uvm/uvm_loan.c 			if (pg->pg_flags & PG_WANTED)
pg                501 uvm/uvm_loan.c 				wakeup(pg);
pg                503 uvm/uvm_loan.c 			if (pg->pg_flags & PG_RELEASED) {
pg                509 uvm/uvm_loan.c 				if (uobj->pgops->pgo_releasepg(pg, NULL))
pg                515 uvm/uvm_loan.c 			uvm_pageactivate(pg); /* make sure it is in queues */
pg                517 uvm/uvm_loan.c 			atomic_clearbits_int(&pg->pg_flags, PG_BUSY|PG_WANTED);
pg                518 uvm/uvm_loan.c 			UVM_PAGE_OWN(pg, NULL);
pg                532 uvm/uvm_loan.c 		if (pg->loan_count == 0)
pg                533 uvm/uvm_loan.c 			pmap_page_protect(pg, VM_PROT_READ);
pg                534 uvm/uvm_loan.c 		pg->loan_count++;
pg                535 uvm/uvm_loan.c 		uvm_pagewire(pg);
pg                537 uvm/uvm_loan.c 		**output = pg;
pg                539 uvm/uvm_loan.c 		if (pg->pg_flags & PG_WANTED)
pg                540 uvm/uvm_loan.c 			wakeup(pg);
pg                541 uvm/uvm_loan.c 		atomic_clearbits_int(&pg->pg_flags, PG_WANTED|PG_BUSY);
pg                542 uvm/uvm_loan.c 		UVM_PAGE_OWN(pg, NULL);
pg                553 uvm/uvm_loan.c 	if (pg->uanon) {
pg                554 uvm/uvm_loan.c 		anon = pg->uanon;
pg                561 uvm/uvm_loan.c 		uvm_pageactivate(pg);	/* reactivate */
pg                563 uvm/uvm_loan.c 		if (pg->pg_flags & PG_WANTED)
pg                564 uvm/uvm_loan.c 			wakeup(pg);
pg                565 uvm/uvm_loan.c 		atomic_clearbits_int(&pg->pg_flags, PG_WANTED|PG_BUSY);
pg                566 uvm/uvm_loan.c 		UVM_PAGE_OWN(pg, NULL);
pg                576 uvm/uvm_loan.c 		if (pg->pg_flags & PG_WANTED)
pg                577 uvm/uvm_loan.c 			wakeup(pg);
pg                578 uvm/uvm_loan.c 		atomic_clearbits_int(&pg->pg_flags, PG_WANTED|PG_BUSY);
pg                579 uvm/uvm_loan.c 		UVM_PAGE_OWN(pg, NULL);
pg                583 uvm/uvm_loan.c 	anon->an_page = pg;
pg                584 uvm/uvm_loan.c 	pg->uanon = anon;
pg                586 uvm/uvm_loan.c 	if (pg->loan_count == 0)
pg                587 uvm/uvm_loan.c 		pmap_page_protect(pg, VM_PROT_READ);
pg                588 uvm/uvm_loan.c 	pg->loan_count++;
pg                589 uvm/uvm_loan.c 	uvm_pageactivate(pg);
pg                593 uvm/uvm_loan.c 	if (pg->pg_flags & PG_WANTED)
pg                594 uvm/uvm_loan.c 		wakeup(pg);
pg                595 uvm/uvm_loan.c 	atomic_clearbits_int(&pg->pg_flags, PG_WANTED|PG_BUSY);
pg                596 uvm/uvm_loan.c 	UVM_PAGE_OWN(pg, NULL);
pg                617 uvm/uvm_loan.c 	struct vm_page *pg;
pg                621 uvm/uvm_loan.c 		while ((pg = uvm_pagealloc(NULL, 0, NULL,
pg                635 uvm/uvm_loan.c 		atomic_clearbits_int(&pg->pg_flags, PG_BUSY|PG_FAKE);
pg                636 uvm/uvm_loan.c 		UVM_PAGE_OWN(pg, NULL);
pg                637 uvm/uvm_loan.c 		**output = pg;
pg                641 uvm/uvm_loan.c 		uvm_pagewire(pg);
pg                642 uvm/uvm_loan.c 		pg->loan_count = 1;
pg                649 uvm/uvm_loan.c 	    (pg = uvm_pagealloc(NULL, 0, anon, UVM_PGA_ZERO)) == NULL) {
pg                673 uvm/uvm_loan.c 	atomic_clearbits_int(&pg->pg_flags, PG_BUSY|PG_FAKE);
pg                674 uvm/uvm_loan.c 	UVM_PAGE_OWN(pg, NULL);
pg                676 uvm/uvm_loan.c 	uvm_pageactivate(pg);
pg                722 uvm/uvm_loan.c 	struct vm_page *pg;
pg                727 uvm/uvm_loan.c 		pg = *ploans++;
pg                729 uvm/uvm_loan.c 		if (pg->loan_count < 1)
pg                730 uvm/uvm_loan.c 			panic("uvm_unloanpage: page %p isn't loaned", pg);
pg                732 uvm/uvm_loan.c 		pg->loan_count--;		/* drop loan */
pg                733 uvm/uvm_loan.c 		uvm_pageunwire(pg);		/* and wire */
pg                739 uvm/uvm_loan.c 		if (pg->loan_count == 0 && pg->uobject == NULL &&
pg                740 uvm/uvm_loan.c 		    pg->uanon == NULL) {
pg                742 uvm/uvm_loan.c 			if (pg->pg_flags & PG_BUSY)
pg                743 uvm/uvm_loan.c 	panic("uvm_unloanpage: page %p unowned but PG_BUSY!", pg);
pg                746 uvm/uvm_loan.c 			pmap_page_protect(pg, VM_PROT_NONE);
pg                747 uvm/uvm_loan.c 			uvm_pagefree(pg);	/* pageq locked above */
pg               2865 uvm/uvm_map.c  	struct vm_page *pg;
pg               2930 uvm/uvm_map.c  			pg = anon->an_page;
pg               2931 uvm/uvm_map.c  			if (pg == NULL) {
pg               2950 uvm/uvm_map.c  				if (pg->loan_count != 0 ||
pg               2951 uvm/uvm_map.c  				    pg->wire_count != 0) {
pg               2964 uvm/uvm_map.c  				if ((pg->pg_flags & PQ_ANON) == 0) {
pg               2965 uvm/uvm_map.c  					KASSERT(pg->uobject == NULL);
pg               2970 uvm/uvm_map.c  				KASSERT(pg->uanon == anon);
pg               2974 uvm/uvm_map.c  				pmap_clear_reference(pg);
pg               2977 uvm/uvm_map.c  				pmap_page_protect(pg, VM_PROT_NONE);
pg               2981 uvm/uvm_map.c  				uvm_pagedeactivate(pg);
pg               2998 uvm/uvm_map.c  				if (pg->wire_count != 0) {
pg               3660 uvm/uvm_map.c  	struct vm_page *pg;
pg               3674 uvm/uvm_map.c  	for (pg = TAILQ_FIRST(&uobj->memq);
pg               3675 uvm/uvm_map.c  	     pg != NULL;
pg               3676 uvm/uvm_map.c  	     pg = TAILQ_NEXT(pg, listq), cnt++) {
pg               3677 uvm/uvm_map.c  		(*pr)("<%p,0x%llx> ", pg, (long long)pg->offset);
pg               3698 uvm/uvm_map.c  uvm_page_printit(pg, full, pr)
pg               3699 uvm/uvm_map.c  	struct vm_page *pg;
pg               3709 uvm/uvm_map.c  	(*pr)("PAGE %p:\n", pg);
pg               3710 uvm/uvm_map.c  	snprintf(pgbuf, sizeof(pgbuf), "%b", pg->pg_flags, page_flagbits);
pg               3711 uvm/uvm_map.c  	snprintf(pqbuf, sizeof(pqbuf), "%b", pg->pg_flags, page_pqflagbits);
pg               3713 uvm/uvm_map.c  	    pgbuf, pqbuf, pg->pg_version, pg->wire_count,
pg               3714 uvm/uvm_map.c  	    (long long)pg->phys_addr);
pg               3716 uvm/uvm_map.c  	    pg->uobject, pg->uanon, (long long)pg->offset, pg->loan_count);
pg               3718 uvm/uvm_map.c  	if (pg->pg_flags & PG_BUSY)
pg               3720 uvm/uvm_map.c  		    pg->owner, pg->owner_tag);
pg               3731 uvm/uvm_map.c  	if ((pg->pg_flags & PQ_FREE) == 0) {
pg               3732 uvm/uvm_map.c  		if (pg->pg_flags & PQ_ANON) {
pg               3733 uvm/uvm_map.c  			if (pg->uanon == NULL || pg->uanon->an_page != pg)
pg               3735 uvm/uvm_map.c  				(pg->uanon) ? pg->uanon->an_page : NULL);
pg               3739 uvm/uvm_map.c  			uobj = pg->uobject;
pg               3743 uvm/uvm_map.c  					if (tpg == pg) {
pg               3756 uvm/uvm_map.c  	if (pg->pg_flags & PQ_FREE) {
pg               3757 uvm/uvm_map.c  		int fl = uvm_page_lookup_freelist(pg);
pg               3758 uvm/uvm_map.c  		pgl = &uvm.page_free[fl].pgfl_queues[((pg)->pg_flags & PG_ZERO) ?
pg               3760 uvm/uvm_map.c  	} else if (pg->pg_flags & PQ_INACTIVE) {
pg               3761 uvm/uvm_map.c  		pgl = (pg->pg_flags & PQ_SWAPBACKED) ?
pg               3763 uvm/uvm_map.c  	} else if (pg->pg_flags & PQ_ACTIVE) {
pg               3772 uvm/uvm_map.c  			if (tpg == pg) {
pg                154 uvm/uvm_page.c uvm_pageinsert(struct vm_page *pg)
pg                160 uvm/uvm_page.c 	KASSERT((pg->pg_flags & PG_TABLED) == 0);
pg                161 uvm/uvm_page.c 	buck = &uvm.page_hash[uvm_pagehash(pg->uobject,pg->offset)];
pg                164 uvm/uvm_page.c 	TAILQ_INSERT_TAIL(buck, pg, hashq);	/* put in hash */
pg                168 uvm/uvm_page.c 	TAILQ_INSERT_TAIL(&pg->uobject->memq, pg, listq); /* put in object */
pg                169 uvm/uvm_page.c 	atomic_setbits_int(&pg->pg_flags, PG_TABLED);
pg                170 uvm/uvm_page.c 	pg->uobject->uo_npages++;
pg                181 uvm/uvm_page.c uvm_pageremove(struct vm_page *pg)
pg                187 uvm/uvm_page.c 	KASSERT(pg->pg_flags & PG_TABLED);
pg                188 uvm/uvm_page.c 	buck = &uvm.page_hash[uvm_pagehash(pg->uobject,pg->offset)];
pg                191 uvm/uvm_page.c 	TAILQ_REMOVE(buck, pg, hashq);
pg                196 uvm/uvm_page.c 	if (pg->uobject->pgops == &uvm_vnodeops) {
pg                202 uvm/uvm_page.c 	TAILQ_REMOVE(&pg->uobject->memq, pg, listq);
pg                204 uvm/uvm_page.c 	atomic_clearbits_int(&pg->pg_flags, PG_TABLED);
pg                205 uvm/uvm_page.c 	pg->uobject->uo_npages--;
pg                206 uvm/uvm_page.c 	pg->uobject = NULL;
pg                207 uvm/uvm_page.c 	pg->pg_version++;
pg                775 uvm/uvm_page.c 	struct vm_page *pg;
pg                829 uvm/uvm_page.c 		while ((pg = TAILQ_FIRST(&oldbuckets[lcv])) != NULL) {
pg                830 uvm/uvm_page.c 			TAILQ_REMOVE(&oldbuckets[lcv], pg, hashq);
pg                832 uvm/uvm_page.c 			  &uvm.page_hash[uvm_pagehash(pg->uobject, pg->offset)],
pg                833 uvm/uvm_page.c 			  pg, hashq);
pg                909 uvm/uvm_page.c 	struct vm_page *pg;
pg                978 uvm/uvm_page.c 			if ((pg = TAILQ_FIRST((freeq =
pg                980 uvm/uvm_page.c 			    (pg = TAILQ_FIRST((freeq =
pg                993 uvm/uvm_page.c 		if ((pg = TAILQ_FIRST((freeq =
pg                995 uvm/uvm_page.c 		    (pg = TAILQ_FIRST((freeq =
pg               1014 uvm/uvm_page.c 	TAILQ_REMOVE(freeq, pg, pageq);
pg               1018 uvm/uvm_page.c 	if (pg->pg_flags & PG_ZERO)
pg               1026 uvm/uvm_page.c 		if (pg->pg_flags & PG_ZERO) {
pg               1037 uvm/uvm_page.c 	pg->offset = off;
pg               1038 uvm/uvm_page.c 	pg->uobject = obj;
pg               1039 uvm/uvm_page.c 	pg->uanon = anon;
pg               1040 uvm/uvm_page.c 	pg->pg_flags = PG_BUSY|PG_CLEAN|PG_FAKE;
pg               1041 uvm/uvm_page.c 	pg->pg_version++;
pg               1043 uvm/uvm_page.c 		anon->an_page = pg;
pg               1044 uvm/uvm_page.c 		atomic_setbits_int(&pg->pg_flags, PQ_ANON);
pg               1050 uvm/uvm_page.c 			uvm_pageinsert(pg);
pg               1053 uvm/uvm_page.c 	pg->owner_tag = NULL;
pg               1055 uvm/uvm_page.c 	UVM_PAGE_OWN(pg, "new alloc");
pg               1062 uvm/uvm_page.c 		atomic_clearbits_int(&pg->pg_flags, PG_CLEAN);
pg               1064 uvm/uvm_page.c 			pmap_zero_page(pg);
pg               1067 uvm/uvm_page.c 	UVMHIST_LOG(pghist, "allocated pg %p/%lx", pg,
pg               1068 uvm/uvm_page.c 	    (u_long)VM_PAGE_TO_PHYS(pg), 0, 0);
pg               1069 uvm/uvm_page.c 	return(pg);
pg               1084 uvm/uvm_page.c uvm_pagerealloc(pg, newobj, newoff)
pg               1085 uvm/uvm_page.c 	struct vm_page *pg;
pg               1096 uvm/uvm_page.c 	if (pg->uobject) {
pg               1097 uvm/uvm_page.c 		uvm_pageremove(pg);
pg               1105 uvm/uvm_page.c 		pg->uobject = newobj;
pg               1106 uvm/uvm_page.c 		pg->offset = newoff;
pg               1107 uvm/uvm_page.c 		pg->pg_version++;
pg               1108 uvm/uvm_page.c 		uvm_pageinsert(pg);
pg               1124 uvm/uvm_page.c uvm_pagefree(struct vm_page *pg)
pg               1127 uvm/uvm_page.c 	int saved_loan_count = pg->loan_count;
pg               1131 uvm/uvm_page.c 	if (pg->uobject == (void *)0xdeadbeef &&
pg               1132 uvm/uvm_page.c 	    pg->uanon == (void *)0xdeadbeef) {
pg               1133 uvm/uvm_page.c 		panic("uvm_pagefree: freeing free page %p", pg);
pg               1137 uvm/uvm_page.c 	UVMHIST_LOG(pghist, "freeing pg %p/%lx", pg,
pg               1138 uvm/uvm_page.c 	    (u_long)VM_PAGE_TO_PHYS(pg), 0, 0);
pg               1145 uvm/uvm_page.c 	if (pg->pg_flags & PG_TABLED) {
pg               1157 uvm/uvm_page.c 			atomic_clearbits_int(&pg->pg_flags, PG_CLEAN);
pg               1158 uvm/uvm_page.c 		uvm_pageremove(pg);
pg               1171 uvm/uvm_page.c 	} else if (saved_loan_count && pg->uanon) {
pg               1179 uvm/uvm_page.c 		atomic_clearbits_int(&pg->pg_flags, PQ_ANON);
pg               1180 uvm/uvm_page.c 		pg->uanon->an_page = NULL;
pg               1181 uvm/uvm_page.c 		pg->uanon = NULL;
pg               1190 uvm/uvm_page.c 	if (pg->pg_flags & PQ_ACTIVE) {
pg               1191 uvm/uvm_page.c 		TAILQ_REMOVE(&uvm.page_active, pg, pageq);
pg               1192 uvm/uvm_page.c 		atomic_clearbits_int(&pg->pg_flags, PQ_ACTIVE);
pg               1195 uvm/uvm_page.c 	if (pg->pg_flags & PQ_INACTIVE) {
pg               1196 uvm/uvm_page.c 		if (pg->pg_flags & PQ_SWAPBACKED)
pg               1197 uvm/uvm_page.c 			TAILQ_REMOVE(&uvm.page_inactive_swp, pg, pageq);
pg               1199 uvm/uvm_page.c 			TAILQ_REMOVE(&uvm.page_inactive_obj, pg, pageq);
pg               1200 uvm/uvm_page.c 		atomic_clearbits_int(&pg->pg_flags, PQ_INACTIVE);
pg               1208 uvm/uvm_page.c 	if (pg->wire_count) {
pg               1209 uvm/uvm_page.c 		pg->wire_count = 0;
pg               1212 uvm/uvm_page.c 	if (pg->uanon) {
pg               1213 uvm/uvm_page.c 		pg->uanon->an_page = NULL;
pg               1223 uvm/uvm_page.c 	atomic_clearbits_int(&pg->pg_flags, PG_ZERO);
pg               1227 uvm/uvm_page.c 	    uvm_page_lookup_freelist(pg)].pgfl_queues[PGFL_UNKNOWN], pg, pageq);
pg               1228 uvm/uvm_page.c 	atomic_clearbits_int(&pg->pg_flags, PQ_MASK);
pg               1229 uvm/uvm_page.c 	atomic_setbits_int(&pg->pg_flags, PQ_FREE);
pg               1231 uvm/uvm_page.c 	pg->uobject = (void *)0xdeadbeef;
pg               1232 uvm/uvm_page.c 	pg->offset = 0xdeadbeef;
pg               1233 uvm/uvm_page.c 	pg->uanon = (void *)0xdeadbeef;
pg               1256 uvm/uvm_page.c 	struct vm_page *pg;
pg               1262 uvm/uvm_page.c 		pg = pgs[i];
pg               1264 uvm/uvm_page.c 		if (pg == NULL || pg == PGO_DONTCARE) {
pg               1267 uvm/uvm_page.c 		if (pg->pg_flags & PG_WANTED) {
pg               1268 uvm/uvm_page.c 			wakeup(pg);
pg               1270 uvm/uvm_page.c 		if (pg->pg_flags & PG_RELEASED) {
pg               1271 uvm/uvm_page.c 			UVMHIST_LOG(pdhist, "releasing pg %p", pg,0,0,0);
pg               1272 uvm/uvm_page.c 			uobj = pg->uobject;
pg               1274 uvm/uvm_page.c 				uobj->pgops->pgo_releasepg(pg, NULL);
pg               1276 uvm/uvm_page.c 				atomic_clearbits_int(&pg->pg_flags, PG_BUSY);
pg               1277 uvm/uvm_page.c 				UVM_PAGE_OWN(pg, NULL);
pg               1278 uvm/uvm_page.c 				uvm_anfree(pg->uanon);
pg               1281 uvm/uvm_page.c 			UVMHIST_LOG(pdhist, "unbusying pg %p", pg,0,0,0);
pg               1282 uvm/uvm_page.c 			atomic_clearbits_int(&pg->pg_flags, PG_WANTED|PG_BUSY);
pg               1283 uvm/uvm_page.c 			UVM_PAGE_OWN(pg, NULL);
pg               1299 uvm/uvm_page.c uvm_page_own(pg, tag)
pg               1300 uvm/uvm_page.c 	struct vm_page *pg;
pg               1305 uvm/uvm_page.c 		if (pg->owner_tag) {
pg               1307 uvm/uvm_page.c 			    "by proc %d [%s]\n", pg,
pg               1308 uvm/uvm_page.c 			     pg->owner, pg->owner_tag);
pg               1311 uvm/uvm_page.c 		pg->owner = (curproc) ? curproc->p_pid :  (pid_t) -1;
pg               1312 uvm/uvm_page.c 		pg->owner_tag = tag;
pg               1317 uvm/uvm_page.c 	if (pg->owner_tag == NULL) {
pg               1319 uvm/uvm_page.c 		    "page (%p)\n", pg);
pg               1322 uvm/uvm_page.c 	pg->owner_tag = NULL;
pg               1337 uvm/uvm_page.c 	struct vm_page *pg;
pg               1353 uvm/uvm_page.c 			if ((pg = TAILQ_FIRST(&pgfl->pgfl_queues[
pg               1358 uvm/uvm_page.c 		if (pg == NULL) {
pg               1368 uvm/uvm_page.c 		TAILQ_REMOVE(&pgfl->pgfl_queues[PGFL_UNKNOWN], pg, pageq);
pg               1373 uvm/uvm_page.c 		if (PMAP_PAGEIDLEZERO(pg) == FALSE) {
pg               1382 uvm/uvm_page.c 			    pg, pageq);
pg               1393 uvm/uvm_page.c 		pmap_zero_page(pg);
pg               1395 uvm/uvm_page.c 		atomic_setbits_int(&pg->pg_flags, PG_ZERO);
pg               1398 uvm/uvm_page.c 		TAILQ_INSERT_HEAD(&pgfl->pgfl_queues[PGFL_ZEROS], pg, pageq);
pg                126 uvm/uvm_page_i.h 	struct vm_page *pg;
pg                134 uvm/uvm_page_i.h 	TAILQ_FOREACH(pg, buck, hashq) {
pg                135 uvm/uvm_page_i.h 		if (pg->uobject == obj && pg->offset == off) {
pg                141 uvm/uvm_page_i.h 	return(pg);
pg                151 uvm/uvm_page_i.h uvm_pagewire(struct vm_page *pg)
pg                153 uvm/uvm_page_i.h 	if (pg->wire_count == 0) {
pg                154 uvm/uvm_page_i.h 		if (pg->pg_flags & PQ_ACTIVE) {
pg                155 uvm/uvm_page_i.h 			TAILQ_REMOVE(&uvm.page_active, pg, pageq);
pg                156 uvm/uvm_page_i.h 			atomic_clearbits_int(&pg->pg_flags, PQ_ACTIVE);
pg                159 uvm/uvm_page_i.h 		if (pg->pg_flags & PQ_INACTIVE) {
pg                160 uvm/uvm_page_i.h 			if (pg->pg_flags & PQ_SWAPBACKED)
pg                161 uvm/uvm_page_i.h 				TAILQ_REMOVE(&uvm.page_inactive_swp, pg, pageq);
pg                163 uvm/uvm_page_i.h 				TAILQ_REMOVE(&uvm.page_inactive_obj, pg, pageq);
pg                164 uvm/uvm_page_i.h 			atomic_clearbits_int(&pg->pg_flags, PQ_INACTIVE);
pg                169 uvm/uvm_page_i.h 	pg->wire_count++;
pg                180 uvm/uvm_page_i.h uvm_pageunwire(struct vm_page *pg)
pg                182 uvm/uvm_page_i.h 	pg->wire_count--;
pg                183 uvm/uvm_page_i.h 	if (pg->wire_count == 0) {
pg                184 uvm/uvm_page_i.h 		TAILQ_INSERT_TAIL(&uvm.page_active, pg, pageq);
pg                186 uvm/uvm_page_i.h 		atomic_setbits_int(&pg->pg_flags, PQ_ACTIVE);
pg                200 uvm/uvm_page_i.h uvm_pagedeactivate(struct vm_page *pg)
pg                202 uvm/uvm_page_i.h 	if (pg->pg_flags & PQ_ACTIVE) {
pg                203 uvm/uvm_page_i.h 		TAILQ_REMOVE(&uvm.page_active, pg, pageq);
pg                204 uvm/uvm_page_i.h 		atomic_clearbits_int(&pg->pg_flags, PQ_ACTIVE);
pg                207 uvm/uvm_page_i.h 	if ((pg->pg_flags & PQ_INACTIVE) == 0) {
pg                208 uvm/uvm_page_i.h 		KASSERT(pg->wire_count == 0);
pg                209 uvm/uvm_page_i.h 		if (pg->pg_flags & PQ_SWAPBACKED)
pg                210 uvm/uvm_page_i.h 			TAILQ_INSERT_TAIL(&uvm.page_inactive_swp, pg, pageq);
pg                212 uvm/uvm_page_i.h 			TAILQ_INSERT_TAIL(&uvm.page_inactive_obj, pg, pageq);
pg                213 uvm/uvm_page_i.h 		atomic_setbits_int(&pg->pg_flags, PQ_INACTIVE);
pg                215 uvm/uvm_page_i.h 		pmap_clear_reference(pg);
pg                222 uvm/uvm_page_i.h 		if ((pg->pg_flags & PG_CLEAN) != 0 &&
pg                223 uvm/uvm_page_i.h 		    pmap_is_modified(pg))
pg                224 uvm/uvm_page_i.h 			atomic_clearbits_int(&pg->pg_flags, PG_CLEAN);
pg                235 uvm/uvm_page_i.h uvm_pageactivate(struct vm_page *pg)
pg                237 uvm/uvm_page_i.h 	if (pg->pg_flags & PQ_INACTIVE) {
pg                238 uvm/uvm_page_i.h 		if (pg->pg_flags & PQ_SWAPBACKED)
pg                239 uvm/uvm_page_i.h 			TAILQ_REMOVE(&uvm.page_inactive_swp, pg, pageq);
pg                241 uvm/uvm_page_i.h 			TAILQ_REMOVE(&uvm.page_inactive_obj, pg, pageq);
pg                242 uvm/uvm_page_i.h 		atomic_clearbits_int(&pg->pg_flags, PQ_INACTIVE);
pg                245 uvm/uvm_page_i.h 	if (pg->wire_count == 0) {
pg                252 uvm/uvm_page_i.h 		if (pg->pg_flags & PQ_ACTIVE)
pg                253 uvm/uvm_page_i.h 			TAILQ_REMOVE(&uvm.page_active, pg, pageq);
pg                255 uvm/uvm_page_i.h 			atomic_setbits_int(&pg->pg_flags, PQ_ACTIVE);
pg                259 uvm/uvm_page_i.h 		TAILQ_INSERT_TAIL(&uvm.page_active, pg, pageq);
pg                271 uvm/uvm_page_i.h uvm_pagezero(struct vm_page *pg)
pg                273 uvm/uvm_page_i.h 	atomic_clearbits_int(&pg->pg_flags, PG_CLEAN);
pg                274 uvm/uvm_page_i.h 	pmap_zero_page(pg);
pg                296 uvm/uvm_page_i.h uvm_page_lookup_freelist(struct vm_page *pg)
pg                300 uvm/uvm_page_i.h 	lcv = vm_physseg_find(atop(VM_PAGE_TO_PHYS(pg)), NULL);
pg                417 uvm/uvm_pager.c uvm_pager_put(uobj, pg, ppsp_ptr, npages, flags, start, stop)
pg                419 uvm/uvm_pager.c 	struct vm_page *pg, ***ppsp_ptr;/* IN, IN/OUT */
pg                444 uvm/uvm_pager.c 			    npages, pg, flags, start, stop);
pg                447 uvm/uvm_pager.c 			ppsp[0] = pg;
pg                510 uvm/uvm_pager.c 			if (*npages > 1 || pg == NULL)
pg                511 uvm/uvm_pager.c 				uvm_pager_dropcluster(uobj, pg, ppsp, npages,
pg                525 uvm/uvm_pager.c 	if (*npages > 1 || pg == NULL) {
pg                529 uvm/uvm_pager.c 		uvm_pager_dropcluster(uobj, pg, ppsp, npages, PGO_REALLOCSWAP);
pg                538 uvm/uvm_pager.c 		if (uobj == NULL && pg != NULL) {
pg                541 uvm/uvm_pager.c 			if (pg->pg_flags & PQ_ANON) {
pg                542 uvm/uvm_pager.c 				simple_lock(&pg->uanon->an_lock);
pg                543 uvm/uvm_pager.c 				pg->uanon->an_swslot = nswblk;
pg                544 uvm/uvm_pager.c 				simple_unlock(&pg->uanon->an_lock);
pg                546 uvm/uvm_pager.c 				simple_lock(&pg->uobject->vmobjlock);
pg                547 uvm/uvm_pager.c 				uao_set_swslot(pg->uobject,
pg                548 uvm/uvm_pager.c 					       pg->offset >> PAGE_SHIFT,
pg                550 uvm/uvm_pager.c 				simple_unlock(&pg->uobject->vmobjlock);
pg                561 uvm/uvm_pager.c 				if (pg) {
pg                569 uvm/uvm_pager.c 			if (pg) {
pg                570 uvm/uvm_pager.c 				ppsp[0] = pg;
pg                618 uvm/uvm_pager.c uvm_pager_dropcluster(uobj, pg, ppsp, npages, flags)
pg                620 uvm/uvm_pager.c 	struct vm_page *pg, **ppsp;	/* IN, IN/OUT */
pg                635 uvm/uvm_pager.c 		if (ppsp[lcv] == pg || ppsp[lcv] == NULL)
pg                793 uvm/uvm_pager.c 	struct vm_page *pg, *pgs[npages];
pg                827 uvm/uvm_pager.c 		pg = pgs[i];
pg                830 uvm/uvm_pager.c 			swap = (pg->pg_flags & PQ_SWAPBACKED) != 0;
pg                832 uvm/uvm_pager.c 				uobj = pg->uobject;
pg                836 uvm/uvm_pager.c 		KASSERT(swap || pg->uobject == uobj);
pg                838 uvm/uvm_pager.c 			if (pg->pg_flags & PQ_ANON) {
pg                839 uvm/uvm_pager.c 				simple_lock(&pg->uanon->an_lock);
pg                841 uvm/uvm_pager.c 				simple_lock(&pg->uobject->vmobjlock);
pg                850 uvm/uvm_pager.c 			atomic_setbits_int(&pg->pg_flags, PG_RELEASED);
pg                868 uvm/uvm_pager.c 			if (pg->pg_flags & PQ_ANON) {
pg                869 uvm/uvm_pager.c 				simple_unlock(&pg->uanon->an_lock);
pg                871 uvm/uvm_pager.c 				simple_unlock(&pg->uobject->vmobjlock);
pg                 60 uvm/uvm_pager_i.h 	struct vm_page *pg;
pg                 66 uvm/uvm_pager_i.h 	pg = PHYS_TO_VM_PAGE(pa);
pg                 67 uvm/uvm_pager_i.h 	KASSERT(pg != NULL);
pg                 68 uvm/uvm_pager_i.h 	return (pg);
pg                 74 uvm/uvm_pglist.c 	struct vm_page *pg;
pg                103 uvm/uvm_pglist.c 		pg = &vm_physmem[psi].pgs[idx];
pg                104 uvm/uvm_pglist.c 		if (VM_PAGE_IS_FREE(pg) == 0)
pg                107 uvm/uvm_pglist.c 		free_list = uvm_page_lookup_freelist(pg);
pg                108 uvm/uvm_pglist.c 		pgflidx = (pg->pg_flags & PG_ZERO) ? PGFL_ZEROS : PGFL_UNKNOWN;
pg                113 uvm/uvm_pglist.c 			if (tp == pg)
pg                119 uvm/uvm_pglist.c 		TAILQ_REMOVE(&uvm.page_free[free_list].pgfl_queues[pgflidx], pg, pageq);
pg                121 uvm/uvm_pglist.c 		if (pg->pg_flags & PG_ZERO)
pg                123 uvm/uvm_pglist.c 		pg->pg_flags = PG_CLEAN;
pg                124 uvm/uvm_pglist.c 		pg->uobject = NULL;
pg                125 uvm/uvm_pglist.c 		pg->uanon = NULL;
pg                126 uvm/uvm_pglist.c 		pg->pg_version++;
pg                127 uvm/uvm_pglist.c 		TAILQ_INSERT_TAIL(rlist, pg, pageq);
pg                663 uvm/uvm_vnode.c uvn_releasepg(pg, nextpgp)
pg                664 uvm/uvm_vnode.c 	struct vm_page *pg;
pg                667 uvm/uvm_vnode.c 	struct uvm_vnode *uvn = (struct uvm_vnode *) pg->uobject;
pg                669 uvm/uvm_vnode.c 	if ((pg->pg_flags & PG_RELEASED) == 0)
pg                676 uvm/uvm_vnode.c 	pmap_page_protect(pg, VM_PROT_NONE);
pg                679 uvm/uvm_vnode.c 		*nextpgp = TAILQ_NEXT(pg, pageq); /* next page for daemon */
pg                680 uvm/uvm_vnode.c 	uvm_pagefree(pg);
pg               1116 xfs/xfs_vnodeops-bsd.c 	struct vm_page *pg;
pg               1119 xfs/xfs_vnodeops-bsd.c 	pg = TAILQ_FIRST(&uobj->memq);
pg               1121 xfs/xfs_vnodeops-bsd.c 	while (pg && !dirty) {
pg               1122 xfs/xfs_vnodeops-bsd.c 	    dirty = pmap_is_modified(pg) || (pg->flags & PG_CLEAN) == 0;
pg               1123 xfs/xfs_vnodeops-bsd.c 	    pg = TAILQ_NEXT(pg, listq);