pp 402 arch/i386/i386/ioapic.c struct ioapic_pin *pp; pp 405 arch/i386/i386/ioapic.c pp = &sc->sc_pins[pin]; pp 406 arch/i386/i386/ioapic.c map = pp->ip_map; pp 413 arch/i386/i386/ioapic.c else if (pp->ip_handler == NULL) { pp 416 arch/i386/i386/ioapic.c redlo |= (pp->ip_vector & 0xff); pp 433 arch/i386/i386/ioapic.c if (pp->ip_type == IST_LEVEL) pp 438 arch/i386/i386/ioapic.c if (pp->ip_type == IST_LEVEL) pp 479 arch/i386/i386/ioapic.c struct ioapic_pin *pp = &sc->sc_pins[pin]; pp 483 arch/i386/i386/ioapic.c ovector = pp->ip_vector; pp 487 arch/i386/i386/ioapic.c pp->ip_minlevel = 0xff; /* XXX magic */ pp 488 arch/i386/i386/ioapic.c pp->ip_maxlevel = 0; /* XXX magic */ pp 489 arch/i386/i386/ioapic.c pp->ip_vector = 0; pp 490 arch/i386/i386/ioapic.c } else if (maxlevel != pp->ip_maxlevel) { pp 522 arch/i386/i386/ioapic.c pp->ip_vector = nvector; pp 523 arch/i386/i386/ioapic.c pp->ip_minlevel = minlevel; pp 524 arch/i386/i386/ioapic.c pp->ip_maxlevel = maxlevel; pp 526 arch/i386/i386/ioapic.c apic_intrhand[pp->ip_vector] = pp->ip_handler; pp 149 arch/i386/stand/pxeboot/net.c u_int *pp = parts; pp 173 arch/i386/stand/pxeboot/net.c if (pp >= parts + 3 || val > 0xff) pp 175 arch/i386/stand/pxeboot/net.c *pp++ = val, cp++; pp 189 arch/i386/stand/pxeboot/net.c n = pp - parts + 1; pp 270 ddb/db_input.c char *pp, *pc; pp 275 ddb/db_input.c for (pp = db_history_prev, pc = db_lbuf_start; pp 276 ddb/db_input.c pc != db_le && *pp; ) { pp 277 ddb/db_input.c if (*pp != *pc) pp 279 ddb/db_input.c if (++pp == db_history + db_history_size) pp 280 ddb/db_input.c pp = db_history; pp 283 ddb/db_input.c if (!*pp && pc == db_le) { pp 144 dev/acpi/acpiprt.c struct aml_value res, *pp; pp 167 dev/acpi/acpiprt.c pp = v->v_package[2]; pp 168 dev/acpi/acpiprt.c if (pp->type == AML_OBJTYPE_NAMEREF) { pp 169 dev/acpi/acpiprt.c node = aml_searchname(sc->sc_devnode, pp->v_nameref); pp 174 dev/acpi/acpiprt.c pp = node->value; pp 176 dev/acpi/acpiprt.c if (pp->type == AML_OBJTYPE_OBJREF) { pp 177 dev/acpi/acpiprt.c pp = pp->v_objref.ref; pp 179 dev/acpi/acpiprt.c if (pp->type == AML_OBJTYPE_DEVICE) { pp 180 dev/acpi/acpiprt.c node = pp->node; pp 206 dev/acpi/acpiprt.c DEVNAME(sc), aml_nodename(pp->node), addr, pin, irq); pp 743 dev/acpi/dsdt.c struct aml_node *node, **pp; pp 777 dev/acpi/dsdt.c for (pp = &root->child; *pp; pp = &(*pp)->sibling) pp 781 dev/acpi/dsdt.c *pp = node; pp 2469 dev/audio.c struct audio_params pp, rp; pp 2485 dev/audio.c pp = sc->sc_pparams; /* Temporary encoding storage in */ pp 2490 dev/audio.c pp.sample_rate = p->sample_rate; pp 2498 dev/audio.c pp.encoding = p->encoding; pp 2506 dev/audio.c pp.precision = p->precision; pp 2514 dev/audio.c pp.channels = p->channels; pp 2525 dev/audio.c audio_print_params("Setting play params", &pp); pp 2529 dev/audio.c if (np && (error = audio_check_params(&pp))) pp 2544 dev/audio.c pp.sw_code = 0; pp 2545 dev/audio.c pp.factor = 1; pp 2565 dev/audio.c pp = rp; pp 2567 dev/audio.c rp = pp; pp 2570 dev/audio.c sc->sc_mode & (AUMODE_PLAY | AUMODE_RECORD), &pp, &rp); pp 2575 dev/audio.c pp.sample_rate = rp.sample_rate; pp 2576 dev/audio.c pp.encoding = rp.encoding; pp 2577 dev/audio.c pp.channels = rp.channels; pp 2578 dev/audio.c pp.precision = rp.precision; pp 2580 dev/audio.c rp.sample_rate = pp.sample_rate; pp 2581 dev/audio.c rp.encoding = pp.encoding; pp 2582 dev/audio.c rp.channels = pp.channels; pp 2583 dev/audio.c rp.precision = pp.precision; pp 2587 dev/audio.c sc->sc_pparams = pp; pp 712 dev/ccd.c struct partition *pp; pp 724 dev/ccd.c pp = &cs->sc_dkdev.dk_label->d_partitions[DISKPART(bp->b_dev)]; pp 725 dev/ccd.c bn += DL_GETPOFFSET(pp); pp 1202 dev/pci/autri.c struct autri_dma **pp, *p; pp 1204 dev/pci/autri.c for (pp = &sc->sc_dmas; (p = *pp) != NULL; pp = &p->next) { pp 1207 dev/pci/autri.c *pp = p->next; pp 763 dev/pci/auvia.c struct auvia_dma **pp, *p; pp 765 dev/pci/auvia.c for (pp = &(sc->sc_dmas); (p = *pp) != NULL; pp = &p->next) pp 772 dev/pci/auvia.c *pp = p->next; pp 1455 dev/pci/cs4280.c struct cs4280_dma **pp, *p; pp 1457 dev/pci/cs4280.c for (pp = &sc->sc_dmas; (p = *pp) != NULL; pp = &p->next) { pp 1460 dev/pci/cs4280.c *pp = p->next; pp 1381 dev/pci/cs4281.c struct cs4281_dma **pp, *p; pp 1384 dev/pci/cs4281.c for (pp = &sc->sc_dmas; (p = *pp) != NULL; pp = &p->next) { pp 1391 dev/pci/cs4281.c *pp = p->next; pp 1634 dev/pci/eap.c struct eap_dma **pp, *p; pp 1636 dev/pci/eap.c for (pp = &sc->sc_dmas; (p = *pp) != NULL; pp = &p->next) { pp 1639 dev/pci/eap.c *pp = p->next; pp 518 dev/pci/esa.c struct esa_dma **pp; pp 520 dev/pci/esa.c for (pp = &vc->dma; (p = *pp) != NULL; pp = &p->next) pp 523 dev/pci/esa.c *pp = p->next; pp 1576 dev/pci/eso.c struct eso_dma *p, **pp; pp 1578 dev/pci/eso.c for (pp = &sc->sc_dmas; (p = *pp) != NULL; pp = &p->ed_next) { pp 1581 dev/pci/eso.c *pp = p->ed_next; pp 774 dev/pci/fms.c struct fms_dma **pp, *p; pp 776 dev/pci/fms.c for (pp = &(sc->sc_dmas); (p = *pp) != NULL; pp = &p->next) pp 783 dev/pci/fms.c *pp = p->next; pp 1825 dev/pci/if_sk.c u_int8_t *pp; pp 1829 dev/pci/if_sk.c pp = mtod(m, u_int8_t *); pp 1833 dev/pci/if_sk.c eh = (struct ether_header *)pp; pp 1837 dev/pci/if_sk.c u_int16_t *xp = (u_int16_t *)pp; pp 1839 dev/pci/if_sk.c xp = (u_int16_t *)pp; pp 1844 dev/pci/if_sk.c xp = (u_int16_t *)(pp + sizeof(struct ip)); pp 1847 dev/pci/if_sk.c pp += EVL_ENCAPLEN; pp 1851 dev/pci/if_sk.c pp += sizeof(*eh); pp 1854 dev/pci/if_sk.c ip = (struct ip *)pp; pp 1889 dev/pci/if_sk.c pp += hlen; pp 1893 dev/pci/if_sk.c struct udphdr *uh = (struct udphdr *)pp; pp 125 dev/pci/if_wi_pci.c const struct wi_pci_product *pp; pp 127 dev/pci/if_wi_pci.c for (pp = wi_pci_products; pp->pp_product != 0; pp++) { pp 128 dev/pci/if_wi_pci.c if (PCI_VENDOR(pa->pa_id) == pp->pp_vendor && pp 129 dev/pci/if_wi_pci.c PCI_PRODUCT(pa->pa_id) == pp->pp_product) pp 130 dev/pci/if_wi_pci.c return (pp); pp 148 dev/pci/if_wi_pci.c const struct wi_pci_product *pp; pp 150 dev/pci/if_wi_pci.c pp = wi_pci_lookup(pa); pp 151 dev/pci/if_wi_pci.c if (pp->pp_attach(pa, sc) != 0) pp 1175 dev/pci/pciide.c const struct pciide_product_desc *pp; pp 1188 dev/pci/pciide.c for (pp = vp->ide_products, i = 0; i < vp->ide_nproducts; pp++, i++) pp 1189 dev/pci/pciide.c if (PCI_PRODUCT(id) == pp->ide_product) pp 1194 dev/pci/pciide.c return (pp); pp 1201 dev/pci/pciide.c const struct pciide_product_desc *pp; pp 1215 dev/pci/pciide.c pp = pciide_lookup_product(pa->pa_id); pp 1216 dev/pci/pciide.c if (pp && (pp->ide_flags & IDE_PCI_CLASS_OVERRIDE)) pp 1236 dev/pci/pciide.c if (pp) pp 66 dev/pci/siop_pci.c const struct siop_product_desc *pp; pp 69 dev/pci/siop_pci.c pp = siop_lookup_product(pa->pa_id, PCI_REVISION(pa->pa_class)); pp 70 dev/pci/siop_pci.c if (pp) pp 196 dev/pci/siop_pci_common.c const struct siop_product_desc *pp; pp 202 dev/pci/siop_pci_common.c for (pp = siop_products; pp->product != 0; pp++) { pp 203 dev/pci/siop_pci_common.c if (PCI_PRODUCT(id) == pp->product && pp->revision <= rev) pp 204 dev/pci/siop_pci_common.c if (rp == NULL || pp->revision > rp->revision) pp 205 dev/pci/siop_pci_common.c rp = pp; pp 1722 dev/pci/yds.c struct yds_dma **pp, *p; pp 1724 dev/pci/yds.c for (pp = &sc->sc_dmas; (p = *pp) != NULL; pp = &p->next) { pp 1727 dev/pci/yds.c *pp = p->next; pp 333 dev/pcmcia/if_wi_pcmcia.c const struct wi_pcmcia_product *pp; pp 341 dev/pcmcia/if_wi_pcmcia.c for (pp = wi_pcmcia_products; pp < epp; pp++) { pp 343 dev/pcmcia/if_wi_pcmcia.c pp->pp_cisinfo[0] != NULL && pp 344 dev/pcmcia/if_wi_pcmcia.c strcmp(pa->card->cis1_info[0], pp->pp_cisinfo[0]) == 0 && pp 346 dev/pcmcia/if_wi_pcmcia.c pp->pp_cisinfo[1] != NULL && pp 347 dev/pcmcia/if_wi_pcmcia.c strcmp(pa->card->cis1_info[1], pp->pp_cisinfo[1]) == 0) pp 348 dev/pcmcia/if_wi_pcmcia.c return (pp); pp 352 dev/pcmcia/if_wi_pcmcia.c for (pp = wi_pcmcia_products; pp < epp; pp++) { pp 354 dev/pcmcia/if_wi_pcmcia.c pa->manufacturer == pp->pp_vendor && pp 356 dev/pcmcia/if_wi_pcmcia.c pa->product == pp->pp_product) pp 357 dev/pcmcia/if_wi_pcmcia.c return (pp); pp 1716 dev/raidframe/rf_openbsdkintf.c struct partition *pp; pp 1755 dev/raidframe/rf_openbsdkintf.c pp = &rs->sc_dkdev.dk_label->d_partitions[DISKPART(bp->b_dev)]; pp 1756 dev/raidframe/rf_openbsdkintf.c blocknum += DL_GETPOFFSET(pp); pp 2132 dev/raidframe/rf_openbsdkintf.c struct partition *pp; pp 2168 dev/raidframe/rf_openbsdkintf.c pp = &lp->d_partitions[i]; pp 2169 dev/raidframe/rf_openbsdkintf.c if (DL_GETPOFFSET(pp) + DL_GETPSIZE(pp) > rs->sc_size) pp 1467 dev/sbus/cs4231.c struct cs_dma *p, **pp; pp 1469 dev/sbus/cs4231.c for (pp = &sc->sc_dmas; (p = *pp) != NULL; pp = &(*pp)->next) { pp 1476 dev/sbus/cs4231.c *pp = p->next; pp 70 dev/usb/usbf_subr.c usbf_realloc(void **pp, size_t *sizep, size_t newsize) pp 77 dev/usb/usbf_subr.c free(*pp, M_USB); pp 78 dev/usb/usbf_subr.c *pp = NULL; pp 89 dev/usb/usbf_subr.c bcopy(*pp, p, oldsize); pp 90 dev/usb/usbf_subr.c *pp = p; pp 536 kern/exec_elf.c Elf_Phdr *ph, *pp; pp 579 kern/exec_elf.c pp = &ph[i]; pp 580 kern/exec_elf.c if (pp->p_type == PT_INTERP) { pp 581 kern/exec_elf.c if (pp->p_filesz >= MAXPATHLEN) pp 585 kern/exec_elf.c pp->p_offset, interp, pp->p_filesz)) != 0) { pp 638 kern/exec_elf.c pp = &ph[i]; pp 702 kern/exec_elf.c phdr = pp->p_vaddr; pp 297 kern/kern_exit.c struct proc *pp = p->p_pptr; pp 304 kern/kern_exit.c if (LIST_EMPTY(&pp->p_children)) pp 305 kern/kern_exit.c wakeup(pp); pp 402 kern/kern_proc.c struct proc *p, *pp; pp 437 kern/kern_proc.c pp = p->p_pptr; pp 453 kern/kern_proc.c pp ? pp->p_pid : -1, p->p_pgrp->pg_id, pp 219 kern/subr_disk.c struct partition *pp; pp 295 kern/subr_disk.c pp = &dlp->d_partitions[i]; pp 296 kern/subr_disk.c pp->p_size = swap32(pp->p_size); pp 297 kern/subr_disk.c pp->p_offset = swap32(pp->p_offset); pp 299 kern/subr_disk.c v0pp = (struct __partitionv0 *)pp; pp 302 kern/subr_disk.c pp->p_offseth = swap16(pp->p_offseth); pp 303 kern/subr_disk.c pp->p_sizeh = swap16(pp->p_sizeh); pp 305 kern/subr_disk.c pp->p_cpg = swap16(pp->p_cpg); pp 329 kern/subr_disk.c pp = lp->d_partitions; pp 330 kern/subr_disk.c for (i = 0; i < lp->d_npartitions; i++, pp++, v0pp++) { pp 331 kern/subr_disk.c pp->p_fragblock = DISKLABELV1_FFS_FRAGBLOCK(v0pp-> pp 333 kern/subr_disk.c pp->p_offseth = 0; pp 334 kern/subr_disk.c pp->p_sizeh = 0; pp 442 kern/subr_disk.c struct partition *pp = &lp->d_partitions[8+n]; pp 453 kern/subr_disk.c DL_SETPOFFSET(pp, pp 456 kern/subr_disk.c DL_SETPSIZE(pp, letoh32(dp2->dp_size)); pp 460 kern/subr_disk.c pp->p_fstype = FS_UNUSED; pp 465 kern/subr_disk.c pp->p_fstype = FS_EXT2FS; pp 475 kern/subr_disk.c pp->p_fstype = FS_MSDOS; pp 488 kern/subr_disk.c pp->p_fstype = FS_OTHER; pp 108 kern/subr_pool.c #define POOL_NEEDS_CATCHUP(pp) \ pp 109 kern/subr_pool.c ((pp)->pr_nitems < (pp)->pr_minitems) pp 198 kern/subr_pool.c pr_log(struct pool *pp, void *v, int action, const char *file, long line) pp 200 kern/subr_pool.c int n = pp->pr_curlogentry; pp 203 kern/subr_pool.c if ((pp->pr_roflags & PR_LOGGING) == 0) pp 210 kern/subr_pool.c pl = &pp->pr_log[n]; pp 215 kern/subr_pool.c if (++n >= pp->pr_logsize) pp 217 kern/subr_pool.c pp->pr_curlogentry = n; pp 221 kern/subr_pool.c pr_printlog(struct pool *pp, struct pool_item *pi, pp 224 kern/subr_pool.c int i = pp->pr_logsize; pp 225 kern/subr_pool.c int n = pp->pr_curlogentry; pp 227 kern/subr_pool.c if ((pp->pr_roflags & PR_LOGGING) == 0) pp 234 kern/subr_pool.c struct pool_log *pl = &pp->pr_log[n]; pp 245 kern/subr_pool.c if (++n >= pp->pr_logsize) pp 251 kern/subr_pool.c pr_enter(struct pool *pp, const char *file, long line) pp 254 kern/subr_pool.c if (__predict_false(pp->pr_entered_file != NULL)) { pp 256 kern/subr_pool.c pp->pr_wchan, file, line); pp 258 kern/subr_pool.c pp->pr_entered_file, pp->pr_entered_line); pp 262 kern/subr_pool.c pp->pr_entered_file = file; pp 263 kern/subr_pool.c pp->pr_entered_line = line; pp 267 kern/subr_pool.c pr_leave(struct pool *pp) pp 270 kern/subr_pool.c if (__predict_false(pp->pr_entered_file == NULL)) { pp 271 kern/subr_pool.c printf("pool %s not entered?\n", pp->pr_wchan); pp 275 kern/subr_pool.c pp->pr_entered_file = NULL; pp 276 kern/subr_pool.c pp->pr_entered_line = 0; pp 280 kern/subr_pool.c pr_enter_check(struct pool *pp, int (*pr)(const char *, ...)) pp 283 kern/subr_pool.c if (pp->pr_entered_file != NULL) pp 285 kern/subr_pool.c pp->pr_entered_file, pp->pr_entered_line); pp 288 kern/subr_pool.c #define pr_log(pp, v, action, file, line) pp 289 kern/subr_pool.c #define pr_printlog(pp, pi, pr) pp 290 kern/subr_pool.c #define pr_enter(pp, file, line) pp 291 kern/subr_pool.c #define pr_leave(pp) pp 292 kern/subr_pool.c #define pr_enter_check(pp, pr) pp 313 kern/subr_pool.c pr_find_pagehead(struct pool *pp, caddr_t page) pp 317 kern/subr_pool.c if ((pp->pr_roflags & PR_PHINPAGE) != 0) pp 318 kern/subr_pool.c return ((struct pool_item_header *)(page + pp->pr_phoffset)); pp 321 kern/subr_pool.c ph = SPLAY_FIND(phtree, &pp->pr_phtree, &tmp); pp 329 kern/subr_pool.c pr_rmpage(struct pool *pp, struct pool_item_header *ph, pp 339 kern/subr_pool.c if (pp->pr_nidle == 0) pp 341 kern/subr_pool.c if (pp->pr_nitems < pp->pr_itemsperpage) pp 344 kern/subr_pool.c pp->pr_nidle--; pp 347 kern/subr_pool.c pp->pr_nitems -= pp->pr_itemsperpage; pp 356 kern/subr_pool.c pool_allocator_free(pp, ph->ph_page); pp 357 kern/subr_pool.c if ((pp->pr_roflags & PR_PHINPAGE) == 0) { pp 358 kern/subr_pool.c SPLAY_REMOVE(phtree, &pp->pr_phtree, ph); pp 364 kern/subr_pool.c pp->pr_npages--; pp 365 kern/subr_pool.c pp->pr_npagefree++; pp 367 kern/subr_pool.c pool_update_curpage(pp); pp 377 kern/subr_pool.c pool_init(struct pool *pp, size_t size, u_int align, u_int ioff, int flags, pp 427 kern/subr_pool.c LIST_INIT(&pp->pr_emptypages); pp 428 kern/subr_pool.c LIST_INIT(&pp->pr_fullpages); pp 429 kern/subr_pool.c LIST_INIT(&pp->pr_partpages); pp 430 kern/subr_pool.c TAILQ_INIT(&pp->pr_cachelist); pp 431 kern/subr_pool.c pp->pr_curpage = NULL; pp 432 kern/subr_pool.c pp->pr_npages = 0; pp 433 kern/subr_pool.c pp->pr_minitems = 0; pp 434 kern/subr_pool.c pp->pr_minpages = 0; pp 435 kern/subr_pool.c pp->pr_maxpages = 8; pp 436 kern/subr_pool.c pp->pr_roflags = flags; pp 437 kern/subr_pool.c pp->pr_flags = 0; pp 438 kern/subr_pool.c pp->pr_size = size; pp 439 kern/subr_pool.c pp->pr_align = align; pp 440 kern/subr_pool.c pp->pr_wchan = wchan; pp 441 kern/subr_pool.c pp->pr_alloc = palloc; pp 442 kern/subr_pool.c pp->pr_nitems = 0; pp 443 kern/subr_pool.c pp->pr_nout = 0; pp 444 kern/subr_pool.c pp->pr_hardlimit = UINT_MAX; pp 445 kern/subr_pool.c pp->pr_hardlimit_warning = NULL; pp 446 kern/subr_pool.c pp->pr_hardlimit_ratecap.tv_sec = 0; pp 447 kern/subr_pool.c pp->pr_hardlimit_ratecap.tv_usec = 0; pp 448 kern/subr_pool.c pp->pr_hardlimit_warning_last.tv_sec = 0; pp 449 kern/subr_pool.c pp->pr_hardlimit_warning_last.tv_usec = 0; pp 450 kern/subr_pool.c pp->pr_serial = ++pool_serial; pp 461 kern/subr_pool.c if (pp->pr_size < palloc->pa_pagesz/16) { pp 463 kern/subr_pool.c pp->pr_roflags |= PR_PHINPAGE; pp 464 kern/subr_pool.c pp->pr_phoffset = off = palloc->pa_pagesz - pp 468 kern/subr_pool.c pp->pr_phoffset = 0; pp 470 kern/subr_pool.c SPLAY_INIT(&pp->pr_phtree); pp 480 kern/subr_pool.c pp->pr_itemoffset = ioff = ioff % align; pp 481 kern/subr_pool.c pp->pr_itemsperpage = (off - ((align - ioff) % align)) / pp->pr_size; pp 482 kern/subr_pool.c KASSERT(pp->pr_itemsperpage != 0); pp 488 kern/subr_pool.c slack = off - pp->pr_itemsperpage * pp->pr_size; pp 489 kern/subr_pool.c pp->pr_maxcolor = (slack / align) * align; pp 490 kern/subr_pool.c pp->pr_curcolor = 0; pp 492 kern/subr_pool.c pp->pr_nget = 0; pp 493 kern/subr_pool.c pp->pr_nfail = 0; pp 494 kern/subr_pool.c pp->pr_nput = 0; pp 495 kern/subr_pool.c pp->pr_npagealloc = 0; pp 496 kern/subr_pool.c pp->pr_npagefree = 0; pp 497 kern/subr_pool.c pp->pr_hiwat = 0; pp 498 kern/subr_pool.c pp->pr_nidle = 0; pp 503 kern/subr_pool.c (pp->pr_log = malloc(pool_logsize * sizeof(struct pool_log), pp 505 kern/subr_pool.c pp->pr_roflags &= ~PR_LOGGING; pp 506 kern/subr_pool.c pp->pr_curlogentry = 0; pp 507 kern/subr_pool.c pp->pr_logsize = pool_logsize; pp 511 kern/subr_pool.c pp->pr_entered_file = NULL; pp 512 kern/subr_pool.c pp->pr_entered_line = 0; pp 514 kern/subr_pool.c simple_lock_init(&pp->pr_slock); pp 516 kern/subr_pool.c pp->pr_ipl = -1; pp 534 kern/subr_pool.c TAILQ_INSERT_TAIL(&pool_head, pp, pr_poollist); pp 539 kern/subr_pool.c TAILQ_INSERT_TAIL(&palloc->pa_list, pp, pr_alloc_list); pp 545 kern/subr_pool.c pool_setipl(struct pool *pp, int ipl) pp 547 kern/subr_pool.c pp->pr_ipl = ipl; pp 555 kern/subr_pool.c pool_destroy(struct pool *pp) pp 561 kern/subr_pool.c simple_lock(&pp->pr_alloc->pa_slock); pp 562 kern/subr_pool.c TAILQ_REMOVE(&pp->pr_alloc->pa_list, pp, pr_alloc_list); pp 563 kern/subr_pool.c simple_unlock(&pp->pr_alloc->pa_slock); pp 566 kern/subr_pool.c while ((pc = TAILQ_FIRST(&pp->pr_cachelist)) != NULL) pp 570 kern/subr_pool.c if (pp->pr_nout != 0) { pp 571 kern/subr_pool.c pr_printlog(pp, NULL, printf); pp 573 kern/subr_pool.c pp->pr_nout); pp 578 kern/subr_pool.c while ((ph = LIST_FIRST(&pp->pr_emptypages)) != NULL) pp 579 kern/subr_pool.c pr_rmpage(pp, ph, NULL); pp 580 kern/subr_pool.c KASSERT(LIST_EMPTY(&pp->pr_fullpages)); pp 581 kern/subr_pool.c KASSERT(LIST_EMPTY(&pp->pr_partpages)); pp 585 kern/subr_pool.c TAILQ_REMOVE(&pool_head, pp, pr_poollist); pp 589 kern/subr_pool.c if ((pp->pr_roflags & PR_LOGGING) != 0) pp 590 kern/subr_pool.c free(pp->pr_log, M_TEMP); pp 595 kern/subr_pool.c pool_alloc_item_header(struct pool *pp, caddr_t storage, int flags) pp 600 kern/subr_pool.c LOCK_ASSERT(simple_lock_held(&pp->pr_slock) == 0); pp 602 kern/subr_pool.c if ((pp->pr_roflags & PR_PHINPAGE) != 0) pp 603 kern/subr_pool.c ph = (struct pool_item_header *) (storage + pp->pr_phoffset); pp 618 kern/subr_pool.c _pool_get(struct pool *pp, int flags, const char *file, long line) pp 620 kern/subr_pool.c pool_get(struct pool *pp, int flags) pp 630 kern/subr_pool.c if (pp->pr_ipl != -1) pp 631 kern/subr_pool.c splassert(pp->pr_ipl); pp 634 kern/subr_pool.c panic("pool_get: %s:must have NOWAIT", pp->pr_wchan); pp 643 kern/subr_pool.c if (pp->pr_roflags & PR_DEBUG) { pp 647 kern/subr_pool.c debug_malloc(pp->pr_size, M_DEBUG, pp 653 kern/subr_pool.c simple_lock(&pp->pr_slock); pp 654 kern/subr_pool.c pr_enter(pp, file, line); pp 663 kern/subr_pool.c if (__predict_false(pp->pr_nout > pp->pr_hardlimit)) { pp 664 kern/subr_pool.c pr_leave(pp); pp 665 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 666 kern/subr_pool.c panic("pool_get: %s: crossed hard limit", pp->pr_wchan); pp 669 kern/subr_pool.c if (__predict_false(pp->pr_nout == pp->pr_hardlimit)) { pp 675 kern/subr_pool.c pp->pr_flags |= PR_WANTED; pp 676 kern/subr_pool.c pr_leave(pp); pp 677 kern/subr_pool.c ltsleep(pp, PSWP, pp->pr_wchan, 0, &pp->pr_slock); pp 678 kern/subr_pool.c pr_enter(pp, file, line); pp 685 kern/subr_pool.c if (pp->pr_hardlimit_warning != NULL && pp 686 kern/subr_pool.c ratecheck(&pp->pr_hardlimit_warning_last, pp 687 kern/subr_pool.c &pp->pr_hardlimit_ratecap)) pp 688 kern/subr_pool.c log(LOG_ERR, "%s\n", pp->pr_hardlimit_warning); pp 690 kern/subr_pool.c pp->pr_nfail++; pp 692 kern/subr_pool.c pr_leave(pp); pp 693 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 703 kern/subr_pool.c if ((ph = pp->pr_curpage) == NULL) { pp 705 kern/subr_pool.c if (pp->pr_nitems != 0) { pp 706 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 708 kern/subr_pool.c pp->pr_wchan, pp->pr_nitems); pp 718 kern/subr_pool.c pr_leave(pp); pp 719 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 720 kern/subr_pool.c v = pool_allocator_alloc(pp, flags); pp 722 kern/subr_pool.c ph = pool_alloc_item_header(pp, v, flags); pp 723 kern/subr_pool.c simple_lock(&pp->pr_slock); pp 724 kern/subr_pool.c pr_enter(pp, file, line); pp 728 kern/subr_pool.c pool_allocator_free(pp, v); pp 736 kern/subr_pool.c if (pp->pr_curpage != NULL) pp 740 kern/subr_pool.c pp->pr_nfail++; pp 741 kern/subr_pool.c pr_leave(pp); pp 742 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 752 kern/subr_pool.c pp->pr_flags |= PR_WANTED; pp 754 kern/subr_pool.c pr_leave(pp); pp 755 kern/subr_pool.c ltsleep(pp, PSWP, pp->pr_wchan, 0, &pp->pr_slock); pp 756 kern/subr_pool.c pr_enter(pp, file, line); pp 761 kern/subr_pool.c pool_prime_page(pp, v, ph); pp 762 kern/subr_pool.c pp->pr_npagealloc++; pp 768 kern/subr_pool.c pr_leave(pp); pp 769 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 770 kern/subr_pool.c panic("pool_get: %s: page empty", pp->pr_wchan); pp 773 kern/subr_pool.c if (__predict_false(pp->pr_nitems == 0)) { pp 774 kern/subr_pool.c pr_leave(pp); pp 775 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 777 kern/subr_pool.c pp->pr_wchan, pp->pr_nitems); pp 783 kern/subr_pool.c pr_log(pp, v, PRLOG_GET, file, line); pp 788 kern/subr_pool.c pr_printlog(pp, pi, printf); pp 791 kern/subr_pool.c pp->pr_wchan, pi->pi_magic, ph->ph_page, pi); pp 799 kern/subr_pool.c pp->pr_nitems--; pp 800 kern/subr_pool.c pp->pr_nout++; pp 803 kern/subr_pool.c if (__predict_false(pp->pr_nidle == 0)) pp 806 kern/subr_pool.c pp->pr_nidle--; pp 813 kern/subr_pool.c LIST_INSERT_HEAD(&pp->pr_partpages, ph, ph_pagelist); pp 818 kern/subr_pool.c if (__predict_false(ph->ph_nmissing != pp->pr_itemsperpage)) { pp 819 kern/subr_pool.c pr_leave(pp); pp 820 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 822 kern/subr_pool.c pp->pr_wchan); pp 830 kern/subr_pool.c LIST_INSERT_HEAD(&pp->pr_fullpages, ph, ph_pagelist); pp 831 kern/subr_pool.c pool_update_curpage(pp); pp 834 kern/subr_pool.c pp->pr_nget++; pp 840 kern/subr_pool.c if (POOL_NEEDS_CATCHUP(pp) && pool_catchup(pp) != 0) { pp 848 kern/subr_pool.c pr_leave(pp); pp 849 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 857 kern/subr_pool.c pool_do_put(struct pool *pp, void *v) pp 864 kern/subr_pool.c if (pp->pr_roflags & PR_DEBUG) { pp 870 kern/subr_pool.c LOCK_ASSERT(simple_lock_held(&pp->pr_slock)); pp 872 kern/subr_pool.c page = (caddr_t)((vaddr_t)v & pp->pr_alloc->pa_pagemask); pp 875 kern/subr_pool.c if (pp->pr_ipl != -1) pp 876 kern/subr_pool.c splassert(pp->pr_ipl); pp 878 kern/subr_pool.c if (__predict_false(pp->pr_nout == 0)) { pp 880 kern/subr_pool.c pp->pr_wchan); pp 885 kern/subr_pool.c if (__predict_false((ph = pr_find_pagehead(pp, page)) == NULL)) { pp 886 kern/subr_pool.c pr_printlog(pp, NULL, printf); pp 887 kern/subr_pool.c panic("pool_put: %s: page header missing", pp->pr_wchan); pp 894 kern/subr_pool.c simple_lock_freecheck((caddr_t)pi, ((caddr_t)pi) + pp->pr_size); pp 907 kern/subr_pool.c for (i = 0; i < pp->pr_size / sizeof(int); i++) { pp 915 kern/subr_pool.c pp->pr_nput++; pp 916 kern/subr_pool.c pp->pr_nitems++; pp 917 kern/subr_pool.c pp->pr_nout--; pp 920 kern/subr_pool.c if (pp->pr_curpage == NULL) pp 921 kern/subr_pool.c pp->pr_curpage = ph; pp 923 kern/subr_pool.c if (pp->pr_flags & PR_WANTED) { pp 924 kern/subr_pool.c pp->pr_flags &= ~PR_WANTED; pp 926 kern/subr_pool.c pp->pr_nidle++; pp 927 kern/subr_pool.c wakeup(pp); pp 943 kern/subr_pool.c pp->pr_nidle++; pp 944 kern/subr_pool.c if (pp->pr_nidle > pp->pr_maxpages || pp 945 kern/subr_pool.c (pp->pr_alloc->pa_flags & PA_WANT) != 0) { pp 946 kern/subr_pool.c pr_rmpage(pp, ph, NULL); pp 949 kern/subr_pool.c LIST_INSERT_HEAD(&pp->pr_emptypages, ph, ph_pagelist); pp 951 kern/subr_pool.c pool_update_curpage(pp); pp 960 kern/subr_pool.c else if (ph->ph_nmissing == (pp->pr_itemsperpage - 1)) { pp 962 kern/subr_pool.c LIST_INSERT_HEAD(&pp->pr_partpages, ph, ph_pagelist); pp 963 kern/subr_pool.c pp->pr_curpage = ph; pp 972 kern/subr_pool.c _pool_put(struct pool *pp, void *v, const char *file, long line) pp 975 kern/subr_pool.c simple_lock(&pp->pr_slock); pp 976 kern/subr_pool.c pr_enter(pp, file, line); pp 978 kern/subr_pool.c pr_log(pp, v, PRLOG_PUT, file, line); pp 980 kern/subr_pool.c pool_do_put(pp, v); pp 982 kern/subr_pool.c pr_leave(pp); pp 983 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 989 kern/subr_pool.c pool_put(struct pool *pp, void *v) pp 992 kern/subr_pool.c simple_lock(&pp->pr_slock); pp 994 kern/subr_pool.c pool_do_put(pp, v); pp 996 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 1007 kern/subr_pool.c pool_prime(struct pool *pp, int n) pp 1013 kern/subr_pool.c simple_lock(&pp->pr_slock); pp 1015 kern/subr_pool.c newpages = roundup(n, pp->pr_itemsperpage) / pp->pr_itemsperpage; pp 1018 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 1019 kern/subr_pool.c cp = pool_allocator_alloc(pp, PR_NOWAIT); pp 1021 kern/subr_pool.c ph = pool_alloc_item_header(pp, cp, PR_NOWAIT); pp 1022 kern/subr_pool.c simple_lock(&pp->pr_slock); pp 1026 kern/subr_pool.c pool_allocator_free(pp, cp); pp 1030 kern/subr_pool.c pool_prime_page(pp, cp, ph); pp 1031 kern/subr_pool.c pp->pr_npagealloc++; pp 1032 kern/subr_pool.c pp->pr_minpages++; pp 1035 kern/subr_pool.c if (pp->pr_minpages >= pp->pr_maxpages) pp 1036 kern/subr_pool.c pp->pr_maxpages = pp->pr_minpages + 1; /* XXX */ pp 1038 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 1048 kern/subr_pool.c pool_prime_page(struct pool *pp, caddr_t storage, struct pool_item_header *ph) pp 1052 kern/subr_pool.c unsigned int align = pp->pr_align; pp 1053 kern/subr_pool.c unsigned int ioff = pp->pr_itemoffset; pp 1057 kern/subr_pool.c if (((u_long)cp & (pp->pr_alloc->pa_pagesz - 1)) != 0) pp 1058 kern/subr_pool.c panic("pool_prime_page: %s: unaligned page", pp->pr_wchan); pp 1064 kern/subr_pool.c LIST_INSERT_HEAD(&pp->pr_emptypages, ph, ph_pagelist); pp 1068 kern/subr_pool.c if ((pp->pr_roflags & PR_PHINPAGE) == 0) pp 1069 kern/subr_pool.c SPLAY_INSERT(phtree, &pp->pr_phtree, ph); pp 1071 kern/subr_pool.c pp->pr_nidle++; pp 1076 kern/subr_pool.c cp = (caddr_t)(cp + pp->pr_curcolor); pp 1077 kern/subr_pool.c if ((pp->pr_curcolor += align) > pp->pr_maxcolor) pp 1078 kern/subr_pool.c pp->pr_curcolor = 0; pp 1089 kern/subr_pool.c n = pp->pr_itemsperpage; pp 1090 kern/subr_pool.c pp->pr_nitems += n; pp 1102 kern/subr_pool.c cp = (caddr_t)(cp + pp->pr_size); pp 1108 kern/subr_pool.c if (pp->pr_curpage == NULL) pp 1109 kern/subr_pool.c pp->pr_curpage = ph; pp 1111 kern/subr_pool.c if (++pp->pr_npages > pp->pr_hiwat) pp 1112 kern/subr_pool.c pp->pr_hiwat = pp->pr_npages; pp 1125 kern/subr_pool.c pool_catchup(struct pool *pp) pp 1131 kern/subr_pool.c while (POOL_NEEDS_CATCHUP(pp)) { pp 1138 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 1139 kern/subr_pool.c cp = pool_allocator_alloc(pp, PR_NOWAIT); pp 1141 kern/subr_pool.c ph = pool_alloc_item_header(pp, cp, PR_NOWAIT); pp 1142 kern/subr_pool.c simple_lock(&pp->pr_slock); pp 1145 kern/subr_pool.c pool_allocator_free(pp, cp); pp 1149 kern/subr_pool.c pool_prime_page(pp, cp, ph); pp 1150 kern/subr_pool.c pp->pr_npagealloc++; pp 1157 kern/subr_pool.c pool_update_curpage(struct pool *pp) pp 1160 kern/subr_pool.c pp->pr_curpage = LIST_FIRST(&pp->pr_partpages); pp 1161 kern/subr_pool.c if (pp->pr_curpage == NULL) { pp 1162 kern/subr_pool.c pp->pr_curpage = LIST_FIRST(&pp->pr_emptypages); pp 1167 kern/subr_pool.c pool_setlowat(struct pool *pp, int n) pp 1170 kern/subr_pool.c simple_lock(&pp->pr_slock); pp 1172 kern/subr_pool.c pp->pr_minitems = n; pp 1173 kern/subr_pool.c pp->pr_minpages = (n == 0) pp 1175 kern/subr_pool.c : roundup(n, pp->pr_itemsperpage) / pp->pr_itemsperpage; pp 1178 kern/subr_pool.c if (POOL_NEEDS_CATCHUP(pp) && pool_catchup(pp) != 0) { pp 1186 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 1190 kern/subr_pool.c pool_sethiwat(struct pool *pp, int n) pp 1193 kern/subr_pool.c simple_lock(&pp->pr_slock); pp 1195 kern/subr_pool.c pp->pr_maxpages = (n == 0) pp 1197 kern/subr_pool.c : roundup(n, pp->pr_itemsperpage) / pp->pr_itemsperpage; pp 1199 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 1203 kern/subr_pool.c pool_sethardlimit(struct pool *pp, unsigned n, const char *warnmess, int ratecap) pp 1207 kern/subr_pool.c simple_lock(&pp->pr_slock); pp 1209 kern/subr_pool.c if (n < pp->pr_nout) { pp 1214 kern/subr_pool.c pp->pr_hardlimit = n; pp 1215 kern/subr_pool.c pp->pr_hardlimit_warning = warnmess; pp 1216 kern/subr_pool.c pp->pr_hardlimit_ratecap.tv_sec = ratecap; pp 1217 kern/subr_pool.c pp->pr_hardlimit_warning_last.tv_sec = 0; pp 1218 kern/subr_pool.c pp->pr_hardlimit_warning_last.tv_usec = 0; pp 1224 kern/subr_pool.c pp->pr_maxpages = (n == 0 || n == UINT_MAX) pp 1226 kern/subr_pool.c : roundup(n, pp->pr_itemsperpage) / pp->pr_itemsperpage; pp 1229 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 1241 kern/subr_pool.c _pool_reclaim(struct pool *pp, const char *file, long line) pp 1243 kern/subr_pool.c pool_reclaim(struct pool *pp) pp 1251 kern/subr_pool.c if (simple_lock_try(&pp->pr_slock) == 0) pp 1253 kern/subr_pool.c pr_enter(pp, file, line); pp 1260 kern/subr_pool.c TAILQ_FOREACH(pc, &pp->pr_cachelist, pc_poollist) pp 1263 kern/subr_pool.c for (ph = LIST_FIRST(&pp->pr_emptypages); ph != NULL; ph = phnext) { pp 1267 kern/subr_pool.c if (pp->pr_npages <= pp->pr_minpages) pp 1276 kern/subr_pool.c if ((pp->pr_nitems - pp->pr_itemsperpage) < pp 1277 kern/subr_pool.c pp->pr_minitems) pp 1280 kern/subr_pool.c pr_rmpage(pp, ph, &pq); pp 1283 kern/subr_pool.c pr_leave(pp); pp 1284 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 1289 kern/subr_pool.c pool_allocator_free(pp, ph->ph_page); pp 1290 kern/subr_pool.c if (pp->pr_roflags & PR_PHINPAGE) { pp 1293 kern/subr_pool.c SPLAY_REMOVE(phtree, &pp->pr_phtree, ph); pp 1311 kern/subr_pool.c pool_printit(struct pool *pp, const char *modif, int (*pr)(const char *, ...)) pp 1316 kern/subr_pool.c if (simple_lock_try(&pp->pr_slock) == 0) { pp 1318 kern/subr_pool.c pp->pr_wchan); pp 1322 kern/subr_pool.c pool_print1(pp, modif, pr); pp 1323 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 1350 kern/subr_pool.c pool_print1(struct pool *pp, const char *modif, int (*pr)(const char *, ...)) pp 1369 kern/subr_pool.c pp->pr_wchan, pp->pr_size, pp->pr_align, pp->pr_itemoffset, pp 1370 kern/subr_pool.c pp->pr_roflags); pp 1371 kern/subr_pool.c (*pr)("\talloc %p\n", pp->pr_alloc); pp 1373 kern/subr_pool.c pp->pr_minitems, pp->pr_minpages, pp->pr_maxpages, pp->pr_npages); pp 1375 kern/subr_pool.c pp->pr_itemsperpage, pp->pr_nitems, pp->pr_nout, pp->pr_hardlimit); pp 1378 kern/subr_pool.c pp->pr_nget, pp->pr_nfail, pp->pr_nput); pp 1380 kern/subr_pool.c pp->pr_npagealloc, pp->pr_npagefree, pp->pr_hiwat, pp->pr_nidle); pp 1385 kern/subr_pool.c if ((ph = LIST_FIRST(&pp->pr_emptypages)) != NULL) pp 1387 kern/subr_pool.c pool_print_pagelist(&pp->pr_emptypages, pr); pp 1388 kern/subr_pool.c if ((ph = LIST_FIRST(&pp->pr_fullpages)) != NULL) pp 1390 kern/subr_pool.c pool_print_pagelist(&pp->pr_fullpages, pr); pp 1391 kern/subr_pool.c if ((ph = LIST_FIRST(&pp->pr_partpages)) != NULL) pp 1393 kern/subr_pool.c pool_print_pagelist(&pp->pr_partpages, pr); pp 1395 kern/subr_pool.c if (pp->pr_curpage == NULL) pp 1398 kern/subr_pool.c (*pr)("\tcurpage %p\n", pp->pr_curpage->ph_page); pp 1405 kern/subr_pool.c if ((pp->pr_roflags & PR_LOGGING) == 0) pp 1408 kern/subr_pool.c pr_printlog(pp, NULL, pr); pp 1414 kern/subr_pool.c TAILQ_FOREACH(pc, &pp->pr_cachelist, pc_poollist) { pp 1427 kern/subr_pool.c pr_enter_check(pp, pr); pp 1433 kern/subr_pool.c struct pool *pp; pp 1462 kern/subr_pool.c TAILQ_FOREACH(pp, &pool_head, pr_poollist) { pp 1464 kern/subr_pool.c db_printf("%-10s %18p %18p\n", pp->pr_wchan, pp, pp 1465 kern/subr_pool.c pp->pr_alloc); pp 1469 kern/subr_pool.c if (!pp->pr_nget) pp 1472 kern/subr_pool.c if (pp->pr_maxpages == UINT_MAX) pp 1475 kern/subr_pool.c snprintf(maxp, sizeof maxp, "%u", pp->pr_maxpages); pp 1487 kern/subr_pool.c PRWORD(ovflw, "%-*s", 10, 0, pp->pr_wchan); pp 1488 kern/subr_pool.c PRWORD(ovflw, " %*u", 4, 1, pp->pr_size); pp 1489 kern/subr_pool.c PRWORD(ovflw, " %*lu", 9, 1, pp->pr_nget); pp 1490 kern/subr_pool.c PRWORD(ovflw, " %*lu", 5, 1, pp->pr_nfail); pp 1491 kern/subr_pool.c PRWORD(ovflw, " %*lu", 9, 1, pp->pr_nput); pp 1492 kern/subr_pool.c PRWORD(ovflw, " %*lu", 6, 1, pp->pr_npagealloc); pp 1493 kern/subr_pool.c PRWORD(ovflw, " %*lu", 6, 1, pp->pr_npagefree); pp 1494 kern/subr_pool.c PRWORD(ovflw, " %*d", 6, 1, pp->pr_npages); pp 1495 kern/subr_pool.c PRWORD(ovflw, " %*d", 6, 1, pp->pr_hiwat); pp 1496 kern/subr_pool.c PRWORD(ovflw, " %*d", 6, 1, pp->pr_minpages); pp 1498 kern/subr_pool.c PRWORD(ovflw, " %*lu\n", 5, 1, pp->pr_nidle); pp 1503 kern/subr_pool.c pool_chk_page(struct pool *pp, const char *label, struct pool_item_header *ph) pp 1509 kern/subr_pool.c page = (caddr_t)((u_long)ph & pp->pr_alloc->pa_pagemask); pp 1511 kern/subr_pool.c (pp->pr_roflags & PR_PHINPAGE) != 0) { pp 1515 kern/subr_pool.c " at page head addr %p (p %p)\n", pp, pp 1516 kern/subr_pool.c pp->pr_wchan, ph->ph_page, pp 1532 kern/subr_pool.c pp->pr_wchan, pi->pi_magic, ph->ph_page, pp 1538 kern/subr_pool.c (caddr_t)((u_long)pi & pp->pr_alloc->pa_pagemask); pp 1545 kern/subr_pool.c " item ordinal %d; addr %p (p %p)\n", pp, pp 1546 kern/subr_pool.c pp->pr_wchan, ph->ph_page, pp 1554 kern/subr_pool.c pool_chk(struct pool *pp, const char *label) pp 1559 kern/subr_pool.c simple_lock(&pp->pr_slock); pp 1560 kern/subr_pool.c LIST_FOREACH(ph, &pp->pr_emptypages, ph_pagelist) { pp 1561 kern/subr_pool.c r = pool_chk_page(pp, label, ph); pp 1566 kern/subr_pool.c LIST_FOREACH(ph, &pp->pr_fullpages, ph_pagelist) { pp 1567 kern/subr_pool.c r = pool_chk_page(pp, label, ph); pp 1572 kern/subr_pool.c LIST_FOREACH(ph, &pp->pr_partpages, ph_pagelist) { pp 1573 kern/subr_pool.c r = pool_chk_page(pp, label, ph); pp 1580 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 1594 kern/subr_pool.c pool_cache_init(struct pool_cache *pc, struct pool *pp, pp 1605 kern/subr_pool.c pc->pc_pool = pp; pp 1618 kern/subr_pool.c simple_lock(&pp->pr_slock); pp 1619 kern/subr_pool.c TAILQ_INSERT_TAIL(&pp->pr_cachelist, pc, pc_poollist); pp 1620 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 1631 kern/subr_pool.c struct pool *pp = pc->pc_pool; pp 1637 kern/subr_pool.c simple_lock(&pp->pr_slock); pp 1638 kern/subr_pool.c TAILQ_REMOVE(&pp->pr_cachelist, pc, pc_poollist); pp 1639 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 1876 kern/subr_pool.c struct pool *pp, *foundpool = NULL; pp 1905 kern/subr_pool.c TAILQ_FOREACH(pp, &pool_head, pr_poollist) { pp 1907 kern/subr_pool.c if (lookfor == pp->pr_serial) { pp 1908 kern/subr_pool.c foundpool = pp; pp 1971 kern/subr_pool.c pool_allocator_alloc(struct pool *pp, int flags) pp 1974 kern/subr_pool.c return (pp->pr_alloc->pa_alloc(pp, flags)); pp 1978 kern/subr_pool.c pool_allocator_free(struct pool *pp, void *v) pp 1980 kern/subr_pool.c struct pool_allocator *pa = pp->pr_alloc; pp 1983 kern/subr_pool.c (*pa->pa_free)(pp, v); pp 1993 kern/subr_pool.c TAILQ_FOREACH(pp, &pa->pa_list, pr_alloc_list) { pp 1994 kern/subr_pool.c simple_lock(&pp->pr_slock); pp 1995 kern/subr_pool.c if ((pp->pr_flags & PR_WANTED) != 0) { pp 1996 kern/subr_pool.c pp->pr_flags &= ~PR_WANTED; pp 1997 kern/subr_pool.c wakeup(pp); pp 1999 kern/subr_pool.c simple_unlock(&pp->pr_slock); pp 2007 kern/subr_pool.c pool_page_alloc(struct pool *pp, int flags) pp 2015 kern/subr_pool.c pool_page_free(struct pool *pp, void *v) pp 2022 kern/subr_pool.c pool_page_alloc_oldnointr(struct pool *pp, int flags) pp 2033 kern/subr_pool.c pool_page_free_oldnointr(struct pool *pp, void *v) pp 368 kern/sys_process.c struct proc *pp; pp 370 kern/sys_process.c pp = pfind(t->p_oppid); pp 371 kern/sys_process.c proc_reparent(t, pp ? pp : initproc); pp 73 lib/libsa/cd9660.c #define PTSIZE(pp) roundup(PTFIXSZ + isonum_711((pp)->namlen), 2) pp 78 lib/libsa/cd9660.c pnmatch(char *path, struct ptable_ent *pp) pp 83 lib/libsa/cd9660.c cp = pp->name; pp 84 lib/libsa/cd9660.c for (i = isonum_711(pp->namlen); --i >= 0; path++, cp++) { pp 139 lib/libsa/cd9660.c struct ptable_ent *pp; pp 188 lib/libsa/cd9660.c pp = (struct ptable_ent *)buf; pp 190 lib/libsa/cd9660.c bno = isonum_732(pp->block) + isonum_711(pp->extlen); pp 200 lib/libsa/cd9660.c if ((void *)pp >= buf + psize) pp 202 lib/libsa/cd9660.c if (isonum_722(pp->parent) != parent) pp 204 lib/libsa/cd9660.c if (!pnmatch(path, pp)) { pp 205 lib/libsa/cd9660.c pp = (struct ptable_ent *)((void *)pp + PTSIZE(pp)); pp 209 lib/libsa/cd9660.c path += isonum_711(pp->namlen) + 1; pp 211 lib/libsa/cd9660.c bno = isonum_732(pp->block) + isonum_711(pp->extlen); pp 212 lib/libsa/cd9660.c while ((void *)pp < buf + psize) { pp 213 lib/libsa/cd9660.c if (isonum_722(pp->parent) == parent) pp 215 lib/libsa/cd9660.c pp = (struct ptable_ent *)((void *)pp + PTSIZE(pp)); pp 331 lib/libsa/net.c u_int *pp = parts; pp 355 lib/libsa/net.c if (pp >= parts + 3 || val > 0xff) pp 357 lib/libsa/net.c *pp++ = val, cp++; pp 371 lib/libsa/net.c n = pp - parts + 1; pp 191 miscfs/procfs/procfs_ctl.c struct proc *pp; pp 193 miscfs/procfs/procfs_ctl.c pp = pfind(p->p_oppid); pp 194 miscfs/procfs/procfs_ctl.c if (pp) pp 195 miscfs/procfs/procfs_ctl.c proc_reparent(p, pp); pp 160 net/pf_ioctl.c pool_sethardlimit(pf_pool_limits[PF_LIMIT_STATES].pp, pp 1875 net/pf_ioctl.c pf_pool_limits[pl->index].pp == NULL) { pp 1879 net/pf_ioctl.c if (pool_sethardlimit(pf_pool_limits[pl->index].pp, pp 2061 net/pf_ioctl.c struct pfioc_pooladdr *pp = (struct pfioc_pooladdr *)addr; pp 2064 net/pf_ioctl.c pp->ticket = ++ticket_pabuf; pp 2069 net/pf_ioctl.c struct pfioc_pooladdr *pp = (struct pfioc_pooladdr *)addr; pp 2071 net/pf_ioctl.c if (pp->ticket != ticket_pabuf) { pp 2076 net/pf_ioctl.c if (pp->af == AF_INET) { pp 2082 net/pf_ioctl.c if (pp->af == AF_INET6) { pp 2087 net/pf_ioctl.c if (pp->addr.addr.type != PF_ADDR_ADDRMASK && pp 2088 net/pf_ioctl.c pp->addr.addr.type != PF_ADDR_DYNIFTL && pp 2089 net/pf_ioctl.c pp->addr.addr.type != PF_ADDR_TABLE) { pp 2098 net/pf_ioctl.c bcopy(&pp->addr, pa, sizeof(struct pf_pooladdr)); pp 2108 net/pf_ioctl.c if (pfi_dynaddr_setup(&pa->addr, pp->af)) { pp 2120 net/pf_ioctl.c struct pfioc_pooladdr *pp = (struct pfioc_pooladdr *)addr; pp 2122 net/pf_ioctl.c pp->nr = 0; pp 2123 net/pf_ioctl.c pool = pf_get_pool(pp->anchor, pp->ticket, pp->r_action, pp 2124 net/pf_ioctl.c pp->r_num, 0, 1, 0); pp 2130 net/pf_ioctl.c pp->nr++; pp 2135 net/pf_ioctl.c struct pfioc_pooladdr *pp = (struct pfioc_pooladdr *)addr; pp 2138 net/pf_ioctl.c pool = pf_get_pool(pp->anchor, pp->ticket, pp->r_action, pp 2139 net/pf_ioctl.c pp->r_num, 0, 1, 1); pp 2145 net/pf_ioctl.c while ((pa != NULL) && (nr < pp->nr)) { pp 2153 net/pf_ioctl.c bcopy(pa, &pp->addr, sizeof(struct pf_pooladdr)); pp 2154 net/pf_ioctl.c pfi_dynaddr_copyout(&pp->addr.addr); pp 2155 net/pf_ioctl.c pf_tbladdr_copyout(&pp->addr.addr); pp 2156 net/pf_ioctl.c pf_rtlabel_copyout(&pp->addr.addr); pp 173 net/pfkeyv2.c struct pfkeyv2_socket **pp; pp 175 net/pfkeyv2.c for (pp = &pfkeyv2_sockets; *pp && ((*pp)->socket != socket); pp 176 net/pfkeyv2.c pp = &((*pp)->next)) pp 179 net/pfkeyv2.c if (*pp) { pp 182 net/pfkeyv2.c pfkeyv2_socket = *pp; pp 183 net/pfkeyv2.c *pp = (*pp)->next; pp 1715 net/pfvar.h void *pp; pp 789 uvm/uvm_aobj.c struct vm_page *pp, *ppnext; pp 835 uvm/uvm_aobj.c pp = TAILQ_FIRST(&uobj->memq); pp 838 uvm/uvm_aobj.c pp = uvm_pagelookup(uobj, curoff); pp 845 uvm/uvm_aobj.c for ( ; (by_list && pp != NULL) || pp 846 uvm/uvm_aobj.c (!by_list && curoff < stop) ; pp = ppnext) { pp 848 uvm/uvm_aobj.c ppnext = TAILQ_NEXT(pp, listq); pp 851 uvm/uvm_aobj.c if (pp->offset < start || pp->offset >= stop) pp 859 uvm/uvm_aobj.c if (pp == NULL) pp 875 uvm/uvm_aobj.c if (pp->loan_count != 0 || pp 876 uvm/uvm_aobj.c pp->wire_count != 0) pp 881 uvm/uvm_aobj.c pmap_clear_reference(pp); pp 884 uvm/uvm_aobj.c pmap_page_protect(pp, VM_PROT_NONE); pp 888 uvm/uvm_aobj.c uvm_pagedeactivate(pp); pp 901 uvm/uvm_aobj.c if (pp->loan_count != 0 || pp 902 uvm/uvm_aobj.c pp->wire_count != 0) pp 908 uvm/uvm_aobj.c if (pp->pg_flags & PG_BUSY) { pp 909 uvm/uvm_aobj.c atomic_setbits_int(&pp->pg_flags, PG_RELEASED); pp 914 uvm/uvm_aobj.c pmap_page_protect(pp, VM_PROT_NONE); pp 916 uvm/uvm_aobj.c uao_dropswap(uobj, pp->offset >> PAGE_SHIFT); pp 917 uvm/uvm_aobj.c uvm_pagefree(pp); pp 264 uvm/uvm_km.c struct vm_page *pp; pp 271 uvm/uvm_km.c pp = uvm_pagelookup(uobj, curoff); pp 272 uvm/uvm_km.c if (pp == NULL) pp 275 uvm/uvm_km.c UVMHIST_LOG(maphist," page %p, busy=%ld", pp, pp 276 uvm/uvm_km.c pp->pg_flags & PG_BUSY, 0, 0); pp 278 uvm/uvm_km.c if (pp->pg_flags & PG_BUSY) { pp 280 uvm/uvm_km.c atomic_setbits_int(&pp->pg_flags, PG_RELEASED); pp 290 uvm/uvm_km.c uvm_pagefree(pp); pp 123 uvm/uvm_pager.c struct vm_page *pp; pp 173 uvm/uvm_pager.c pp = *pps++; pp 174 uvm/uvm_pager.c KASSERT(pp); pp 175 uvm/uvm_pager.c KASSERT(pp->pg_flags & PG_BUSY); pp 176 uvm/uvm_pager.c pmap_enter(vm_map_pmap(pager_map), cva, VM_PAGE_TO_PHYS(pp), pp 585 uvm/uvm_vnode.c struct vm_page *pp; pp 586 uvm/uvm_vnode.c TAILQ_FOREACH(pp, &uvn->u_obj.memq, listq) { pp 587 uvm/uvm_vnode.c if ((pp->pg_flags & PG_BUSY) == 0) pp 811 uvm/uvm_vnode.c struct vm_page *pp, *ppnext, *ptmp; pp 860 uvm/uvm_vnode.c TAILQ_FOREACH(pp, &uobj->memq, listq) { pp 862 uvm/uvm_vnode.c (pp->offset < start || pp->offset >= stop)) pp 864 uvm/uvm_vnode.c atomic_clearbits_int(&pp->pg_flags, pp 871 uvm/uvm_vnode.c pp = uvm_pagelookup(uobj, curoff); pp 872 uvm/uvm_vnode.c if (pp) pp 873 uvm/uvm_vnode.c atomic_clearbits_int(&pp->pg_flags, pp 886 uvm/uvm_vnode.c pp = TAILQ_FIRST(&uobj->memq); pp 889 uvm/uvm_vnode.c pp = uvm_pagelookup(uobj, curoff); pp 897 uvm/uvm_vnode.c for ( ; (by_list && pp != NULL) || pp 898 uvm/uvm_vnode.c (!by_list && curoff < stop) ; pp = ppnext) { pp 907 uvm/uvm_vnode.c (pp->offset < start || pp->offset >= stop)) { pp 908 uvm/uvm_vnode.c ppnext = TAILQ_NEXT(pp, listq); pp 919 uvm/uvm_vnode.c if (pp == NULL) { pp 938 uvm/uvm_vnode.c if ((flags & PGO_CLEANIT) == 0 || (pp->pg_flags & PG_BUSY) != 0) { pp 940 uvm/uvm_vnode.c if ((pp->pg_flags & PG_BUSY) != 0 && pp 949 uvm/uvm_vnode.c if ((pp->pg_flags & PG_CLEAN) != 0 && pp 951 uvm/uvm_vnode.c (pp->pg_flags & PQ_ACTIVE) != 0) pp 952 uvm/uvm_vnode.c pmap_page_protect(pp, VM_PROT_NONE); pp 953 uvm/uvm_vnode.c if ((pp->pg_flags & PG_CLEAN) != 0 && pp 954 uvm/uvm_vnode.c pmap_is_modified(pp)) pp 955 uvm/uvm_vnode.c atomic_clearbits_int(&pp->pg_flags, PG_CLEAN); pp 956 uvm/uvm_vnode.c atomic_setbits_int(&pp->pg_flags, PG_CLEANCHK); pp 958 uvm/uvm_vnode.c needs_clean = ((pp->pg_flags & PG_CLEAN) == 0); pp 967 uvm/uvm_vnode.c ppnext = TAILQ_NEXT(pp, listq); pp 975 uvm/uvm_vnode.c if ((pp->pg_flags & PQ_INACTIVE) == 0 && pp 976 uvm/uvm_vnode.c pp->wire_count == 0) { pp 977 uvm/uvm_vnode.c pmap_page_protect(pp, VM_PROT_NONE); pp 978 uvm/uvm_vnode.c uvm_pagedeactivate(pp); pp 982 uvm/uvm_vnode.c if (pp->pg_flags & PG_BUSY) { pp 984 uvm/uvm_vnode.c atomic_setbits_int(&pp->pg_flags, pp 987 uvm/uvm_vnode.c pmap_page_protect(pp, VM_PROT_NONE); pp 989 uvm/uvm_vnode.c uvm_pagefree(pp); pp 1005 uvm/uvm_vnode.c atomic_setbits_int(&pp->pg_flags, PG_BUSY); pp 1006 uvm/uvm_vnode.c UVM_PAGE_OWN(pp, "uvn_flush"); pp 1007 uvm/uvm_vnode.c pmap_page_protect(pp, VM_PROT_READ); pp 1008 uvm/uvm_vnode.c pp_version = pp->pg_version; pp 1014 uvm/uvm_vnode.c result = uvm_pager_put(uobj, pp, &ppsp, &npages, pp 1072 uvm/uvm_vnode.c if (pp->pg_version == pp_version) pp 1073 uvm/uvm_vnode.c ppnext = TAILQ_NEXT(pp, listq); pp 1103 uvm/uvm_vnode.c if (ppsp[lcv] == pp) pp 1107 uvm/uvm_vnode.c ptmp = pp; pp 1111 uvm/uvm_vnode.c if (pp->pg_version == pp_version) pp 1112 uvm/uvm_vnode.c ppnext = TAILQ_NEXT(pp, listq); pp 1167 uvm/uvm_vnode.c if ((pp->pg_flags & PQ_INACTIVE) == 0 && pp 1168 uvm/uvm_vnode.c pp->wire_count == 0) { pp 1185 uvm/uvm_vnode.c pp->uobject, pp 1186 uvm/uvm_vnode.c (long long)pp->offset); pp 342 xfs/xfs_dev-common.c d_thread_t *pp = xfs_uio_to_thread(uiop); pp 344 xfs/xfs_dev-common.c d_thread_t *pp = xfs_uio_to_proc(uiop); pp 351 xfs/xfs_dev-common.c pp);