entry 153 arch/i386/i386/apm.c u_int32_t entry; entry 829 arch/i386/i386/apm.c apm_ep.entry = ap->apm_entry; entry 212 arch/i386/i386/bios.c if (h->entry <= BIOS32_START || h->entry >= BIOS32_END) entry 216 arch/i386/i386/bios.c bios32_entry.offset = (u_int32_t)ISA_HOLE_VADDR(h->entry); entry 217 arch/i386/i386/bios.c printf(", BIOS32 rev. %d @ 0x%lx", h->rev, h->entry); entry 716 arch/i386/i386/mpbios.c const struct mpbios_proc *entry = (const struct mpbios_proc *)ent; entry 721 arch/i386/i386/mpbios.c if (!(entry->cpu_flags & PROCENTRY_FLAG_EN)) entry 725 arch/i386/i386/mpbios.c if (entry->cpu_flags & PROCENTRY_FLAG_BP) entry 731 arch/i386/i386/mpbios.c caa.cpu_number = entry->apic_id; entry 734 arch/i386/i386/mpbios.c caa.cpu_signature = entry->cpu_signature; entry 740 arch/i386/i386/mpbios.c caa.feature_flags = entry->feature_flags; entry 763 arch/i386/i386/mpbios.c mp_cfg_special_intr(const struct mpbios_int *entry, u_int32_t *redir) entry 775 arch/i386/i386/mpbios.c switch (entry->int_type) { entry 793 arch/i386/i386/mpbios.c panic("unknown MPS interrupt type %d", entry->int_type); entry 800 arch/i386/i386/mpbios.c mp_cfg_pci_intr(const struct mpbios_int *entry, u_int32_t *redir) entry 802 arch/i386/i386/mpbios.c int mpspo = entry->int_flags & 0x03; /* XXX magic */ entry 803 arch/i386/i386/mpbios.c int mpstrig = (entry->int_flags >> 2) & 0x03; /* XXX magic */ entry 818 arch/i386/i386/mpbios.c if (entry->int_type != MPS_INTTYPE_INT) { entry 819 arch/i386/i386/mpbios.c mp_cfg_special_intr(entry, redir); entry 838 arch/i386/i386/mpbios.c mp_cfg_eisa_intr (const struct mpbios_int *entry, u_int32_t *redir) entry 840 arch/i386/i386/mpbios.c int mpspo = entry->int_flags & 0x03; /* XXX magic */ entry 841 arch/i386/i386/mpbios.c int mpstrig = (entry->int_flags >> 2) & 0x03; /* XXX magic */ entry 856 arch/i386/i386/mpbios.c if (entry->int_type != MPS_INTTYPE_INT) { entry 857 arch/i386/i386/mpbios.c mp_cfg_special_intr(entry, redir); entry 874 arch/i386/i386/mpbios.c if (mp_busses[entry->src_bus_id].mb_data & entry 875 arch/i386/i386/mpbios.c (1<<entry->src_bus_irq)) { entry 888 arch/i386/i386/mpbios.c mp_cfg_isa_intr(const struct mpbios_int *entry, u_int32_t *redir) entry 890 arch/i386/i386/mpbios.c int mpspo = entry->int_flags & 0x03; /* XXX magic */ entry 891 arch/i386/i386/mpbios.c int mpstrig = (entry->int_flags >> 2) & 0x03; /* XXX magic */ entry 906 arch/i386/i386/mpbios.c if (entry->int_type != MPS_INTTYPE_INT) { entry 907 arch/i386/i386/mpbios.c mp_cfg_special_intr(entry, redir); entry 960 arch/i386/i386/mpbios.c const struct mpbios_bus *entry = (const struct mpbios_bus *)ent; entry 961 arch/i386/i386/mpbios.c int bus_id = entry->bus_id; entry 964 arch/i386/i386/mpbios.c bus_id, entry->bus_type); entry 973 arch/i386/i386/mpbios.c self->dv_xname, bus_id, entry->bus_type); entry 979 arch/i386/i386/mpbios.c if (memcmp(entry->bus_type, "PCI ", 6) == 0) { entry 984 arch/i386/i386/mpbios.c } else if (memcmp(entry->bus_type, "EISA ", 6) == 0) { entry 997 arch/i386/i386/mpbios.c } else if (memcmp(entry->bus_type, "ISA ", 6) == 0) { entry 1009 arch/i386/i386/mpbios.c entry->bus_type); entry 1017 arch/i386/i386/mpbios.c const struct mpbios_ioapic *entry = (const struct mpbios_ioapic *)ent; entry 1021 arch/i386/i386/mpbios.c if (!(entry->apic_flags & IOAPICENTRY_FLAG_EN)) entry 1025 arch/i386/i386/mpbios.c aaa.apic_id = entry->apic_id; entry 1026 arch/i386/i386/mpbios.c aaa.apic_version = entry->apic_version; entry 1027 arch/i386/i386/mpbios.c aaa.apic_address = (u_int32_t)entry->apic_address; entry 1037 arch/i386/i386/mpbios.c const struct mpbios_int *entry = (const struct mpbios_int *)ent; entry 1038 arch/i386/i386/mpbios.c struct mpbios_int rw_entry = *entry; entry 1044 arch/i386/i386/mpbios.c u_int32_t id = IOAPIC_REMAPPED_ID(entry->dst_apic_id); entry 1045 arch/i386/i386/mpbios.c u_int32_t pin = entry->dst_apic_int; entry 1046 arch/i386/i386/mpbios.c u_int32_t bus = entry->src_bus_id; entry 1047 arch/i386/i386/mpbios.c u_int32_t dev = entry->src_bus_irq; entry 1048 arch/i386/i386/mpbios.c u_int32_t type = entry->int_type; entry 1049 arch/i386/i386/mpbios.c u_int32_t flags = entry->int_flags; entry 1084 arch/i386/i386/mpbios.c if (entry->type == MPS_MCT_IOINT) { entry 72 arch/i386/include/biosvar.h u_int32_t entry; /* initialization entry point */ entry 83 arch/i386/include/biosvar.h u_int32_t entry; /* 04: entry point */ entry 337 arch/i386/pci/pci_intr_fixup.c int entry; entry 342 arch/i386/pci/pci_intr_fixup.c for (entry = 0; entry < pcibios_pir_table_nentries; entry++) { entry 343 arch/i386/pci/pci_intr_fixup.c pir = &pcibios_pir_table[entry]; entry 371 arch/i386/pci/pci_intr_fixup.c int entry, pin, link; entry 383 arch/i386/pci/pci_intr_fixup.c for (entry = 0; entry < pcibios_pir_table_nentries; entry++) { entry 384 arch/i386/pci/pci_intr_fixup.c pir = &pcibios_pir_table[entry]; entry 48 arch/i386/stand/libsa/exec_i386.c u_long entry; entry 71 arch/i386/stand/libsa/exec_i386.c entry = marks[MARK_ENTRY] & 0x0fffffff; entry 73 arch/i386/stand/libsa/exec_i386.c printf("entry point at 0x%x\n", (int) entry); entry 75 arch/i386/stand/libsa/exec_i386.c (*(startfuncp)entry)(howto, bootdev, BOOTARG_APIVER, entry 350 dev/acpi/acpi.c struct acpi_q *entry; entry 396 dev/acpi/acpi.c SIMPLEQ_FOREACH(entry, &sc->sc_tables, q_next) { entry 397 dev/acpi/acpi.c if (memcmp(entry->q_table, FADT_SIG, entry 399 dev/acpi/acpi.c sc->sc_fadt = entry->q_table; entry 432 dev/acpi/acpi.c acpi_load_dsdt(sc->sc_fadt->dsdt, &entry); entry 434 dev/acpi/acpi.c acpi_load_dsdt(sc->sc_fadt->x_dsdt, &entry); entry 436 dev/acpi/acpi.c if (entry == NULL) entry 438 dev/acpi/acpi.c SIMPLEQ_INSERT_HEAD(&sc->sc_tables, entry, q_next); entry 440 dev/acpi/acpi.c p_dsdt = entry->q_table; entry 445 dev/acpi/acpi.c SIMPLEQ_FOREACH(entry, &sc->sc_tables, q_next) { entry 446 dev/acpi/acpi.c if (memcmp(entry->q_table, SSDT_SIG, entry 448 dev/acpi/acpi.c p_dsdt = entry->q_table; entry 515 dev/acpi/acpi.c SIMPLEQ_FOREACH(entry, &sc->sc_tables, q_next) { entry 516 dev/acpi/acpi.c printf("%.4s ", entry->q_table); entry 542 dev/acpi/acpi.c SIMPLEQ_FOREACH(entry, &sc->sc_tables, q_next) { entry 552 dev/acpi/acpi.c aaa.aaa_table = entry->q_table; entry 717 dev/acpi/acpi.c struct acpi_q *entry; entry 719 dev/acpi/acpi.c entry = malloc(len + sizeof(struct acpi_q), M_DEVBUF, M_NOWAIT); entry 721 dev/acpi/acpi.c if (entry != NULL) { entry 723 dev/acpi/acpi.c free(entry, M_DEVBUF); entry 726 dev/acpi/acpi.c memcpy(entry->q_data, handle.va, len); entry 727 dev/acpi/acpi.c entry->q_table = entry->q_data; entry 729 dev/acpi/acpi.c SIMPLEQ_INSERT_TAIL(queue, entry, q_next); entry 157 dev/acpi/acpimadt.c union acpi_madt_entry *entry = (union acpi_madt_entry *)addr; entry 159 dev/acpi/acpimadt.c switch (entry->madt_lapic.apic_type) { entry 162 dev/acpi/acpimadt.c self->dv_xname, entry->madt_lapic.acpi_proc_id, entry 163 dev/acpi/acpimadt.c entry->madt_lapic.apic_id, entry 164 dev/acpi/acpimadt.c entry->madt_lapic.flags); entry 166 dev/acpi/acpimadt.c lapic_map[entry->madt_lapic.acpi_proc_id] = entry 167 dev/acpi/acpimadt.c entry->madt_lapic.apic_id; entry 172 dev/acpi/acpimadt.c if ((entry->madt_lapic.flags & ACPI_PROC_ENABLE) == 0) entry 178 dev/acpi/acpimadt.c caa.cpu_number = entry->madt_lapic.apic_id; entry 193 dev/acpi/acpimadt.c self->dv_xname, entry->madt_ioapic.acpi_ioapic_id, entry 194 dev/acpi/acpimadt.c entry->madt_ioapic.address, entry 195 dev/acpi/acpimadt.c entry->madt_ioapic.global_int_base); entry 202 dev/acpi/acpimadt.c aaa.apic_id = entry->madt_ioapic.acpi_ioapic_id; entry 203 dev/acpi/acpimadt.c aaa.apic_address = entry->madt_ioapic.address; entry 204 dev/acpi/acpimadt.c aaa.apic_vecbase = entry->madt_ioapic.global_int_base; entry 213 dev/acpi/acpimadt.c addr += entry->madt_lapic.length; entry 223 dev/acpi/acpimadt.c union acpi_madt_entry *entry = (union acpi_madt_entry *)addr; entry 225 dev/acpi/acpimadt.c switch (entry->madt_lapic.apic_type) { entry 232 dev/acpi/acpimadt.c self->dv_xname, entry->madt_override.bus, entry 233 dev/acpi/acpimadt.c entry->madt_override.source, entry 234 dev/acpi/acpimadt.c entry->madt_override.global_int, entry 235 dev/acpi/acpimadt.c entry->madt_override.flags); entry 237 dev/acpi/acpimadt.c pin = entry->madt_override.global_int; entry 247 dev/acpi/acpimadt.c map->bus_pin = entry->madt_override.source; entry 248 dev/acpi/acpimadt.c map->flags = entry->madt_override.flags; entry 250 dev/acpi/acpimadt.c map->global_int = entry->madt_override.global_int; entry 252 dev/acpi/acpimadt.c acpimadt_cfg_intr(entry->madt_override.flags, &map->redir); entry 266 dev/acpi/acpimadt.c self->dv_xname, entry->madt_lapic_nmi.acpi_proc_id, entry 267 dev/acpi/acpimadt.c entry->madt_lapic_nmi.local_apic_lint, entry 268 dev/acpi/acpimadt.c entry->madt_lapic_nmi.flags); entry 270 dev/acpi/acpimadt.c pin = entry->madt_lapic_nmi.local_apic_lint; entry 274 dev/acpi/acpimadt.c map->cpu_id = lapic_map[entry->madt_lapic_nmi.acpi_proc_id]; entry 276 dev/acpi/acpimadt.c map->flags = entry->madt_lapic_nmi.flags; entry 278 dev/acpi/acpimadt.c acpimadt_cfg_intr(entry->madt_lapic_nmi.flags, &map->redir); entry 285 dev/acpi/acpimadt.c self->dv_xname, entry->madt_lapic.apic_type); entry 288 dev/acpi/acpimadt.c addr += entry->madt_lapic.length; entry 249 dev/i2o/i2o.h struct i2o_hrt_entry entry[1]; entry 285 dev/i2o/i2o.h struct i2o_lct_entry entry[1]; entry 319 dev/i2o/i2o.h struct i2o_systab_entry entry[1]; entry 478 dev/i2o/iop.c for (i = 0, ste = iop_systab->entry; i < iop_cd.cd_ndevs; i++) entry 637 dev/i2o/iop.c le = sc->sc_lct->entry; entry 754 dev/i2o/iop.c for (i = 0, le = sc->sc_lct->entry; i < nent; i++, le++) { entry 277 dev/i2o/iopsp.c for (le = iop->sc_lct->entry; nent != 0; nent--, le++) entry 286 dev/i2o/iopsp.c for (i = 0, le = iop->sc_lct->entry; i < nent; i++, le++) { entry 3522 dev/ic/aic79xx.c struct ahd_phase_table_entry *entry; entry 3530 dev/ic/aic79xx.c for (entry = ahd_phase_table; entry < last_entry; entry++) { entry 3531 dev/ic/aic79xx.c if (phase == entry->phase) entry 3534 dev/ic/aic79xx.c return (entry); entry 8807 dev/ic/aic79xx.c int entry, printed; entry 8823 dev/ic/aic79xx.c for (entry = 0; entry < num_entries; entry++) { entry 8824 dev/ic/aic79xx.c if (((value & table[entry].mask) != table[entry].value) entry 8825 dev/ic/aic79xx.c || ((printed_mask & table[entry].mask) == entry 8826 dev/ic/aic79xx.c table[entry].mask)) entry 8831 dev/ic/aic79xx.c table[entry].name); entry 8832 dev/ic/aic79xx.c printed_mask |= table[entry].mask; entry 8836 dev/ic/aic79xx.c if (entry >= num_entries) entry 2275 dev/ic/aic7xxx.c struct ahc_phase_table_entry *entry; entry 2283 dev/ic/aic7xxx.c for (entry = ahc_phase_table; entry < last_entry; entry++) { entry 2284 dev/ic/aic7xxx.c if (phase == entry->phase) entry 2287 dev/ic/aic7xxx.c return (entry); entry 6444 dev/ic/aic7xxx.c int entry, printed; entry 6460 dev/ic/aic7xxx.c for (entry = 0; entry < num_entries; entry++) { entry 6461 dev/ic/aic7xxx.c if (((value & table[entry].mask) != table[entry].value) entry 6462 dev/ic/aic7xxx.c || ((printed_mask & table[entry].mask) == entry 6463 dev/ic/aic7xxx.c table[entry].mask)) entry 6468 dev/ic/aic7xxx.c table[entry].name); entry 6469 dev/ic/aic7xxx.c printed_mask |= table[entry].mask; entry 6473 dev/ic/aic7xxx.c if (entry >= num_entries) entry 1814 dev/ic/ar5210.c ar5k_ar5210_reset_key(struct ath_hal *hal, u_int16_t entry) entry 1818 dev/ic/ar5210.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5210_KEYTABLE_SIZE); entry 1821 dev/ic/ar5210.c AR5K_REG_WRITE(AR5K_AR5210_KEYTABLE_OFF(entry, i), 0); entry 1827 dev/ic/ar5210.c ar5k_ar5210_is_key_valid(struct ath_hal *hal, u_int16_t entry) entry 1829 dev/ic/ar5210.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5210_KEYTABLE_SIZE); entry 1834 dev/ic/ar5210.c if (AR5K_REG_READ(AR5K_AR5210_KEYTABLE_MAC1(entry)) & entry 1842 dev/ic/ar5210.c ar5k_ar5210_set_key(struct ath_hal *hal, u_int16_t entry, entry 1848 dev/ic/ar5210.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5210_KEYTABLE_SIZE); entry 1883 dev/ic/ar5210.c AR5K_REG_WRITE(AR5K_AR5210_KEYTABLE_OFF(entry, i), key_v[i]); entry 1885 dev/ic/ar5210.c return (ar5k_ar5210_set_key_lladdr(hal, entry, mac)); entry 1889 dev/ic/ar5210.c ar5k_ar5210_set_key_lladdr(struct ath_hal *hal, u_int16_t entry, entry 1898 dev/ic/ar5210.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5210_KEYTABLE_SIZE); entry 1906 dev/ic/ar5210.c AR5K_REG_WRITE(AR5K_AR5210_KEYTABLE_MAC0(entry), low_id); entry 1907 dev/ic/ar5210.c AR5K_REG_WRITE(AR5K_AR5210_KEYTABLE_MAC1(entry), high_id); entry 1919 dev/ic/ar5211.c ar5k_ar5211_reset_key(struct ath_hal *hal, u_int16_t entry) entry 1923 dev/ic/ar5211.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5211_KEYTABLE_SIZE); entry 1926 dev/ic/ar5211.c AR5K_REG_WRITE(AR5K_AR5211_KEYTABLE_OFF(entry, i), 0); entry 1932 dev/ic/ar5211.c ar5k_ar5211_is_key_valid(struct ath_hal *hal, u_int16_t entry) entry 1934 dev/ic/ar5211.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5211_KEYTABLE_SIZE); entry 1939 dev/ic/ar5211.c if (AR5K_REG_READ(AR5K_AR5211_KEYTABLE_MAC1(entry)) & entry 1947 dev/ic/ar5211.c ar5k_ar5211_set_key(struct ath_hal *hal, u_int16_t entry, entry 1953 dev/ic/ar5211.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5211_KEYTABLE_SIZE); entry 1988 dev/ic/ar5211.c AR5K_REG_WRITE(AR5K_AR5211_KEYTABLE_OFF(entry, i), key_v[i]); entry 1990 dev/ic/ar5211.c return (ar5k_ar5211_set_key_lladdr(hal, entry, mac)); entry 1994 dev/ic/ar5211.c ar5k_ar5211_set_key_lladdr(struct ath_hal *hal, u_int16_t entry, entry 2003 dev/ic/ar5211.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5211_KEYTABLE_SIZE); entry 2011 dev/ic/ar5211.c AR5K_REG_WRITE(AR5K_AR5211_KEYTABLE_MAC0(entry), low_id); entry 2012 dev/ic/ar5211.c AR5K_REG_WRITE(AR5K_AR5211_KEYTABLE_MAC1(entry), high_id); entry 2254 dev/ic/ar5212.c ar5k_ar5212_reset_key(struct ath_hal *hal, u_int16_t entry) entry 2258 dev/ic/ar5212.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5212_KEYTABLE_SIZE); entry 2261 dev/ic/ar5212.c AR5K_REG_WRITE(AR5K_AR5212_KEYTABLE_OFF(entry, i), 0); entry 2264 dev/ic/ar5212.c AR5K_REG_WRITE(AR5K_AR5212_KEYTABLE_TYPE(entry), entry 2271 dev/ic/ar5212.c ar5k_ar5212_is_key_valid(struct ath_hal *hal, u_int16_t entry) entry 2273 dev/ic/ar5212.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5212_KEYTABLE_SIZE); entry 2278 dev/ic/ar5212.c if (AR5K_REG_READ(AR5K_AR5212_KEYTABLE_MAC1(entry)) & entry 2286 dev/ic/ar5212.c ar5k_ar5212_set_key(struct ath_hal *hal, u_int16_t entry, entry 2292 dev/ic/ar5212.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5212_KEYTABLE_SIZE); entry 2327 dev/ic/ar5212.c AR5K_REG_WRITE(AR5K_AR5212_KEYTABLE_OFF(entry, i), key_v[i]); entry 2329 dev/ic/ar5212.c return (ar5k_ar5212_set_key_lladdr(hal, entry, mac)); entry 2333 dev/ic/ar5212.c ar5k_ar5212_set_key_lladdr(struct ath_hal *hal, u_int16_t entry, entry 2342 dev/ic/ar5212.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5212_KEYTABLE_SIZE); entry 2350 dev/ic/ar5212.c AR5K_REG_WRITE(AR5K_AR5212_KEYTABLE_MAC0(entry), low_id); entry 2351 dev/ic/ar5212.c AR5K_REG_WRITE(AR5K_AR5212_KEYTABLE_MAC1(entry), high_id); entry 1302 dev/ic/ar5xxx.c u_int32_t mask, entry, last, data, shift, position; entry 1316 dev/ic/ar5xxx.c entry = ((first - 1) / 8) + offset; entry 1322 dev/ic/ar5xxx.c for (i = shift = 0, left = bits; left > 0; position = 0, entry++, i++) { entry 1328 dev/ic/ar5xxx.c rf[entry] &= ~mask; entry 1329 dev/ic/ar5xxx.c rf[entry] |= ((data << position) << (col * 8)) & mask; entry 1332 dev/ic/ar5xxx.c data = (((rf[entry] & mask) >> (col * 8)) >> entry 760 dev/ic/sti.c a.in.entry = i; entry 471 dev/ic/stireg.h u_int32_t entry; entry 799 dev/isa/ad1848.c ad1848_devmap_t *entry; entry 804 dev/isa/ad1848.c if (!(entry = ad1848_mixer_find_dev(map, cnt, cp))) entry 807 dev/isa/ad1848.c dev = entry->dev; entry 809 dev/isa/ad1848.c switch (entry->kind) { entry 873 dev/isa/ad1848.c ad1848_devmap_t *entry; entry 878 dev/isa/ad1848.c if (!(entry = ad1848_mixer_find_dev(map, cnt, cp))) entry 881 dev/isa/ad1848.c dev = entry->dev; entry 883 dev/isa/ad1848.c switch (entry->kind) { entry 659 dev/pci/ahc_pci.c const struct ahc_pci_identity *entry; entry 679 dev/pci/ahc_pci.c entry = &ahc_pci_ident_table[i]; entry 680 dev/pci/ahc_pci.c if (entry->full_id == (full_id & entry->id_mask)) entry 681 dev/pci/ahc_pci.c return (entry); entry 692 dev/pci/ahc_pci.c const struct ahc_pci_identity *entry; entry 696 dev/pci/ahc_pci.c entry = ahc_find_pci_device(pa->pa_id, subid, pa->pa_function); entry 697 dev/pci/ahc_pci.c return (entry != NULL && entry->setup != NULL) ? 1 : 0; entry 706 dev/pci/ahc_pci.c const struct ahc_pci_identity *entry; entry 753 dev/pci/ahc_pci.c entry = ahc_find_pci_device(pa->pa_id, subid, pa->pa_function); entry 754 dev/pci/ahc_pci.c if (entry == NULL) entry 773 dev/pci/ahc_pci.c error = entry->setup(ahc); entry 298 dev/pci/ahd_pci.c const struct ahd_pci_identity *entry; entry 314 dev/pci/ahd_pci.c entry = &ahd_pci_ident_table[i]; entry 315 dev/pci/ahd_pci.c if (entry->full_id == (full_id & entry->id_mask)) { entry 316 dev/pci/ahd_pci.c return (entry); entry 325 dev/pci/ahd_pci.c const struct ahd_pci_identity *entry; entry 330 dev/pci/ahd_pci.c entry = ahd_find_pci_device(pa->pa_id, subid); entry 331 dev/pci/ahd_pci.c return entry != NULL ? 1 : 0; entry 337 dev/pci/ahd_pci.c const struct ahd_pci_identity *entry; entry 353 dev/pci/ahd_pci.c entry = ahd_find_pci_device(pa->pa_id, subid); entry 354 dev/pci/ahd_pci.c if (entry == NULL) entry 371 dev/pci/ahd_pci.c error = entry->setup(ahd, pa); entry 624 dev/pci/if_bge.c struct bge_jpool_entry *entry; entry 675 dev/pci/if_bge.c entry = malloc(sizeof(struct bge_jpool_entry), entry 677 dev/pci/if_bge.c if (entry == NULL) { entry 683 dev/pci/if_bge.c entry->slot = i; entry 685 dev/pci/if_bge.c entry, jpool_entries); entry 715 dev/pci/if_bge.c struct bge_jpool_entry *entry; entry 717 dev/pci/if_bge.c entry = SLIST_FIRST(&sc->bge_jfree_listhead); entry 719 dev/pci/if_bge.c if (entry == NULL) entry 723 dev/pci/if_bge.c SLIST_INSERT_HEAD(&sc->bge_jinuse_listhead, entry, jpool_entries); entry 724 dev/pci/if_bge.c return (sc->bge_cdata.bge_jslots[entry->slot]); entry 733 dev/pci/if_bge.c struct bge_jpool_entry *entry; entry 751 dev/pci/if_bge.c entry = SLIST_FIRST(&sc->bge_jinuse_listhead); entry 752 dev/pci/if_bge.c if (entry == NULL) entry 754 dev/pci/if_bge.c entry->slot = i; entry 756 dev/pci/if_bge.c SLIST_INSERT_HEAD(&sc->bge_jfree_listhead, entry, jpool_entries); entry 753 dev/pci/if_lge.c struct lge_jpool_entry *entry; entry 805 dev/pci/if_lge.c entry = malloc(sizeof(struct lge_jpool_entry), entry 807 dev/pci/if_lge.c if (entry == NULL) { entry 814 dev/pci/if_lge.c entry->slot = i; entry 816 dev/pci/if_lge.c entry, jpool_entries); entry 844 dev/pci/if_lge.c struct lge_jpool_entry *entry; entry 846 dev/pci/if_lge.c entry = LIST_FIRST(&sc->lge_jfree_listhead); entry 848 dev/pci/if_lge.c if (entry == NULL) entry 851 dev/pci/if_lge.c LIST_REMOVE(entry, jpool_entries); entry 852 dev/pci/if_lge.c LIST_INSERT_HEAD(&sc->lge_jinuse_listhead, entry, jpool_entries); entry 853 dev/pci/if_lge.c return (sc->lge_cdata.lge_jslots[entry->slot]); entry 864 dev/pci/if_lge.c struct lge_jpool_entry *entry; entry 878 dev/pci/if_lge.c entry = LIST_FIRST(&sc->lge_jinuse_listhead); entry 879 dev/pci/if_lge.c if (entry == NULL) entry 881 dev/pci/if_lge.c entry->slot = i; entry 882 dev/pci/if_lge.c LIST_REMOVE(entry, jpool_entries); entry 883 dev/pci/if_lge.c LIST_INSERT_HEAD(&sc->lge_jfree_listhead, entry, jpool_entries); entry 467 dev/pci/if_msk.c struct sk_txmap_entry *entry; entry 486 dev/pci/if_msk.c entry = malloc(sizeof(*entry), M_DEVBUF, M_NOWAIT); entry 487 dev/pci/if_msk.c if (!entry) { entry 491 dev/pci/if_msk.c entry->dmamap = dmamap; entry 492 dev/pci/if_msk.c SIMPLEQ_INSERT_HEAD(&sc_if->sk_txmap_head, entry, link); entry 570 dev/pci/if_msk.c struct sk_jpool_entry *entry; entry 620 dev/pci/if_msk.c entry = malloc(sizeof(struct sk_jpool_entry), entry 622 dev/pci/if_msk.c if (entry == NULL) { entry 628 dev/pci/if_msk.c entry->slot = i; entry 630 dev/pci/if_msk.c entry, jpool_entries); entry 660 dev/pci/if_msk.c struct sk_jpool_entry *entry; entry 662 dev/pci/if_msk.c entry = LIST_FIRST(&sc_if->sk_jfree_listhead); entry 664 dev/pci/if_msk.c if (entry == NULL) entry 667 dev/pci/if_msk.c LIST_REMOVE(entry, jpool_entries); entry 668 dev/pci/if_msk.c LIST_INSERT_HEAD(&sc_if->sk_jinuse_listhead, entry, jpool_entries); entry 669 dev/pci/if_msk.c return (sc_if->sk_cdata.sk_jslots[entry->slot]); entry 678 dev/pci/if_msk.c struct sk_jpool_entry *entry; entry 695 dev/pci/if_msk.c entry = LIST_FIRST(&sc->sk_jinuse_listhead); entry 696 dev/pci/if_msk.c if (entry == NULL) entry 698 dev/pci/if_msk.c entry->slot = i; entry 699 dev/pci/if_msk.c LIST_REMOVE(entry, jpool_entries); entry 700 dev/pci/if_msk.c LIST_INSERT_HEAD(&sc->sk_jfree_listhead, entry, jpool_entries); entry 1409 dev/pci/if_msk.c struct sk_txmap_entry *entry; entry 1414 dev/pci/if_msk.c entry = SIMPLEQ_FIRST(&sc_if->sk_txmap_head); entry 1415 dev/pci/if_msk.c if (entry == NULL) { entry 1419 dev/pci/if_msk.c txmap = entry->dmamap; entry 1467 dev/pci/if_msk.c sc_if->sk_cdata.sk_tx_map[cur] = entry; entry 1679 dev/pci/if_msk.c struct sk_txmap_entry *entry; entry 1706 dev/pci/if_msk.c entry = sc_if->sk_cdata.sk_tx_map[idx]; entry 1711 dev/pci/if_msk.c bus_dmamap_sync(sc->sc_dmatag, entry->dmamap, 0, entry 1712 dev/pci/if_msk.c entry->dmamap->dm_mapsize, BUS_DMASYNC_POSTWRITE); entry 1714 dev/pci/if_msk.c bus_dmamap_unload(sc->sc_dmatag, entry->dmamap); entry 1715 dev/pci/if_msk.c SIMPLEQ_INSERT_TAIL(&sc_if->sk_txmap_head, entry, entry 1112 dev/pci/if_nge.c struct nge_jpool_entry *entry; entry 1169 dev/pci/if_nge.c entry = malloc(sizeof(struct nge_jpool_entry), entry 1171 dev/pci/if_nge.c if (entry == NULL) { entry 1178 dev/pci/if_nge.c entry->slot = i; entry 1179 dev/pci/if_nge.c LIST_INSERT_HEAD(&sc->nge_jfree_listhead, entry, entry 1209 dev/pci/if_nge.c struct nge_jpool_entry *entry; entry 1211 dev/pci/if_nge.c entry = LIST_FIRST(&sc->nge_jfree_listhead); entry 1213 dev/pci/if_nge.c if (entry == NULL) entry 1216 dev/pci/if_nge.c LIST_REMOVE(entry, jpool_entries); entry 1217 dev/pci/if_nge.c LIST_INSERT_HEAD(&sc->nge_jinuse_listhead, entry, jpool_entries); entry 1218 dev/pci/if_nge.c sc->nge_cdata.nge_jslots[entry->slot].nge_inuse = 1; entry 1219 dev/pci/if_nge.c return(sc->nge_cdata.nge_jslots[entry->slot].nge_buf); entry 1233 dev/pci/if_nge.c struct nge_jpool_entry *entry; entry 1253 dev/pci/if_nge.c entry = LIST_FIRST(&sc->nge_jinuse_listhead); entry 1254 dev/pci/if_nge.c if (entry == NULL) entry 1256 dev/pci/if_nge.c entry->slot = i; entry 1257 dev/pci/if_nge.c LIST_REMOVE(entry, jpool_entries); entry 1259 dev/pci/if_nge.c entry, jpool_entries); entry 101 dev/pci/if_san_xilinx.c SIMPLEQ_ENTRY(xilinx_rx_buffer) entry; entry 424 dev/pci/if_san_xilinx.c sc->rx_dma_buf, entry); entry 588 dev/pci/if_san_xilinx.c SIMPLEQ_REMOVE_HEAD(&sc->wp_rx_free_list, entry); entry 593 dev/pci/if_san_xilinx.c SIMPLEQ_REMOVE_HEAD(&sc->wp_rx_complete_list, entry); entry 1689 dev/pci/if_san_xilinx.c SIMPLEQ_REMOVE_HEAD(&sc->wp_rx_free_list, entry); entry 2105 dev/pci/if_san_xilinx.c SIMPLEQ_INSERT_TAIL(&sc->wp_rx_complete_list, buf, entry); entry 2413 dev/pci/if_san_xilinx.c SIMPLEQ_REMOVE_HEAD(&sc->wp_rx_free_list, entry); entry 2418 dev/pci/if_san_xilinx.c SIMPLEQ_REMOVE_HEAD(&sc->wp_rx_complete_list, entry); entry 2470 dev/pci/if_san_xilinx.c SIMPLEQ_INSERT_TAIL(&sc->wp_rx_free_list, buf, entry); entry 2493 dev/pci/if_san_xilinx.c SIMPLEQ_INSERT_TAIL(&sc->wp_rx_free_list, buf, entry); entry 2531 dev/pci/if_san_xilinx.c SIMPLEQ_REMOVE_HEAD(&sc->wp_rx_complete_list, entry); entry 596 dev/pci/if_sk.c struct sk_txmap_entry *entry; entry 616 dev/pci/if_sk.c entry = malloc(sizeof(*entry), M_DEVBUF, M_NOWAIT); entry 617 dev/pci/if_sk.c if (!entry) { entry 621 dev/pci/if_sk.c entry->dmamap = dmamap; entry 622 dev/pci/if_sk.c SIMPLEQ_INSERT_HEAD(&sc_if->sk_txmap_head, entry, link); entry 698 dev/pci/if_sk.c struct sk_jpool_entry *entry; entry 748 dev/pci/if_sk.c entry = malloc(sizeof(struct sk_jpool_entry), entry 750 dev/pci/if_sk.c if (entry == NULL) { entry 756 dev/pci/if_sk.c entry->slot = i; entry 758 dev/pci/if_sk.c entry, jpool_entries); entry 788 dev/pci/if_sk.c struct sk_jpool_entry *entry; entry 790 dev/pci/if_sk.c entry = LIST_FIRST(&sc_if->sk_jfree_listhead); entry 792 dev/pci/if_sk.c if (entry == NULL) entry 795 dev/pci/if_sk.c LIST_REMOVE(entry, jpool_entries); entry 796 dev/pci/if_sk.c LIST_INSERT_HEAD(&sc_if->sk_jinuse_listhead, entry, jpool_entries); entry 797 dev/pci/if_sk.c return (sc_if->sk_cdata.sk_jslots[entry->slot]); entry 806 dev/pci/if_sk.c struct sk_jpool_entry *entry; entry 823 dev/pci/if_sk.c entry = LIST_FIRST(&sc->sk_jinuse_listhead); entry 824 dev/pci/if_sk.c if (entry == NULL) entry 826 dev/pci/if_sk.c entry->slot = i; entry 827 dev/pci/if_sk.c LIST_REMOVE(entry, jpool_entries); entry 828 dev/pci/if_sk.c LIST_INSERT_HEAD(&sc->sk_jfree_listhead, entry, jpool_entries); entry 1523 dev/pci/if_sk.c struct sk_txmap_entry *entry; entry 1528 dev/pci/if_sk.c entry = SIMPLEQ_FIRST(&sc_if->sk_txmap_head); entry 1529 dev/pci/if_sk.c if (entry == NULL) { entry 1533 dev/pci/if_sk.c txmap = entry->dmamap; entry 1581 dev/pci/if_sk.c sc_if->sk_cdata.sk_tx_map[cur] = entry; entry 1915 dev/pci/if_sk.c struct sk_txmap_entry *entry; entry 1941 dev/pci/if_sk.c entry = sc_if->sk_cdata.sk_tx_map[idx]; entry 1946 dev/pci/if_sk.c bus_dmamap_sync(sc->sc_dmatag, entry->dmamap, 0, entry 1947 dev/pci/if_sk.c entry->dmamap->dm_mapsize, BUS_DMASYNC_POSTWRITE); entry 1949 dev/pci/if_sk.c bus_dmamap_unload(sc->sc_dmatag, entry->dmamap); entry 1950 dev/pci/if_sk.c SIMPLEQ_INSERT_TAIL(&sc_if->sk_txmap_head, entry, entry 590 dev/pci/if_ti.c struct ti_jpool_entry *entry; entry 641 dev/pci/if_ti.c entry = malloc(sizeof(struct ti_jpool_entry), entry 643 dev/pci/if_ti.c if (entry == NULL) { entry 650 dev/pci/if_ti.c entry->slot = i; entry 651 dev/pci/if_ti.c SLIST_INSERT_HEAD(&sc->ti_jfree_listhead, entry, jpool_entries); entry 681 dev/pci/if_ti.c struct ti_jpool_entry *entry; entry 683 dev/pci/if_ti.c entry = SLIST_FIRST(&sc->ti_jfree_listhead); entry 685 dev/pci/if_ti.c if (entry == NULL) entry 689 dev/pci/if_ti.c SLIST_INSERT_HEAD(&sc->ti_jinuse_listhead, entry, jpool_entries); entry 690 dev/pci/if_ti.c sc->ti_cdata.ti_jslots[entry->slot].ti_inuse = 1; entry 691 dev/pci/if_ti.c return (sc->ti_cdata.ti_jslots[entry->slot].ti_buf); entry 702 dev/pci/if_ti.c struct ti_jpool_entry *entry; entry 720 dev/pci/if_ti.c entry = SLIST_FIRST(&sc->ti_jinuse_listhead); entry 721 dev/pci/if_ti.c if (entry == NULL) entry 723 dev/pci/if_ti.c entry->slot = i; entry 726 dev/pci/if_ti.c entry, jpool_entries); entry 1021 dev/pci/if_ti.c struct ti_txmap_entry *entry; entry 1038 dev/pci/if_ti.c while ((entry = SLIST_FIRST(&sc->ti_tx_map_listhead))) { entry 1040 dev/pci/if_ti.c bus_dmamap_destroy(sc->sc_dmatag, entry->dmamap); entry 1041 dev/pci/if_ti.c free(entry, M_DEVBUF); entry 1050 dev/pci/if_ti.c struct ti_txmap_entry *entry; entry 1063 dev/pci/if_ti.c entry = malloc(sizeof(*entry), M_DEVBUF, M_NOWAIT); entry 1064 dev/pci/if_ti.c if (!entry) { entry 1068 dev/pci/if_ti.c entry->dmamap = dmamap; entry 1069 dev/pci/if_ti.c SLIST_INSERT_HEAD(&sc->ti_tx_map_listhead, entry, link); entry 1843 dev/pci/if_ti.c struct ti_txmap_entry *entry; entry 1867 dev/pci/if_ti.c entry = sc->ti_cdata.ti_tx_map[idx]; entry 1868 dev/pci/if_ti.c bus_dmamap_sync(sc->sc_dmatag, entry->dmamap, 0, entry 1869 dev/pci/if_ti.c entry->dmamap->dm_mapsize, BUS_DMASYNC_POSTWRITE); entry 1871 dev/pci/if_ti.c bus_dmamap_unload(sc->sc_dmatag, entry->dmamap); entry 1872 dev/pci/if_ti.c SLIST_INSERT_HEAD(&sc->ti_tx_map_listhead, entry, entry 1893 dev/pci/if_ti.c struct ti_txmap_entry *entry; entry 1913 dev/pci/if_ti.c entry = sc->ti_cdata.ti_tx_map[idx]; entry 1914 dev/pci/if_ti.c bus_dmamap_sync(sc->sc_dmatag, entry->dmamap, 0, entry 1915 dev/pci/if_ti.c entry->dmamap->dm_mapsize, BUS_DMASYNC_POSTWRITE); entry 1917 dev/pci/if_ti.c bus_dmamap_unload(sc->sc_dmatag, entry->dmamap); entry 1918 dev/pci/if_ti.c SLIST_INSERT_HEAD(&sc->ti_tx_map_listhead, entry, entry 1998 dev/pci/if_ti.c struct ti_txmap_entry *entry; entry 2010 dev/pci/if_ti.c entry = SLIST_FIRST(&sc->ti_tx_map_listhead); entry 2011 dev/pci/if_ti.c if (entry == NULL) entry 2013 dev/pci/if_ti.c txmap = entry->dmamap; entry 2069 dev/pci/if_ti.c sc->ti_cdata.ti_tx_map[cur] = entry; entry 2086 dev/pci/if_ti.c struct ti_txmap_entry *entry; entry 2097 dev/pci/if_ti.c entry = SLIST_FIRST(&sc->ti_tx_map_listhead); entry 2098 dev/pci/if_ti.c if (entry == NULL) entry 2100 dev/pci/if_ti.c txmap = entry->dmamap; entry 2153 dev/pci/if_ti.c sc->ti_cdata.ti_tx_map[cur] = entry; entry 37 dev/sdmmc/sdmmc_io.c TAILQ_ENTRY(sdmmc_intr_handler) entry; entry 634 dev/sdmmc/sdmmc_io.c TAILQ_INSERT_TAIL(&sc->sc_intrq, ih, entry); entry 653 dev/sdmmc/sdmmc_io.c TAILQ_REMOVE(&sc->sc_intrq, ih, entry); entry 689 dev/sdmmc/sdmmc_io.c TAILQ_FOREACH(ih, &sc->sc_intrq, entry) { entry 107 kern/exec_ecoff.c epp->ep_entry = eap->entry; entry 138 kern/exec_ecoff.c epp->ep_entry = eap->entry; entry 178 kern/exec_ecoff.c epp->ep_entry = eap->entry; entry 445 kern/kern_lkm.c curp->entry = (int (*)(struct lkm_table *, int, int)) entry 449 kern/kern_lkm.c printf("LKM: call entrypoint %x\n", curp->entry); entry 453 kern/kern_lkm.c error = (*(curp->entry))(curp, LKM_E_LOAD, curp->ver); entry 493 kern/kern_lkm.c if ((*(curp->entry))(curp, LKM_E_UNLOAD, curp->ver)) { entry 510 kern/kern_lkm.c if ((error = (*curp->entry)(curp, LKM_E_STAT, curp->ver))) entry 269 lib/libsa/loadfile.c marks[MARK_ENTRY] = LOADADDR(coff->a.entry); entry 281 lib/libsa/loadfile.c u_long entry = x->a_entry; entry 295 lib/libsa/loadfile.c minp = maxp = ALIGNENTRY(entry); entry 453 lib/libsa/loadfile.c marks[MARK_ENTRY] = LOADADDR(entry); entry 131 net/if_media.c struct ifmedia_entry *entry; entry 144 net/if_media.c entry = malloc(sizeof(*entry), M_IFADDR, M_NOWAIT); entry 145 net/if_media.c if (entry == NULL) entry 148 net/if_media.c entry->ifm_media = mword; entry 149 net/if_media.c entry->ifm_data = data; entry 150 net/if_media.c entry->ifm_aux = aux; entry 152 net/if_media.c TAILQ_INSERT_TAIL(&ifm->ifm_list, entry, ifm_list); entry 294 net/pf.c RB_GENERATE(pf_src_tree, pf_src_node, entry, pf_src_compare); entry 379 net/pf_if.c TAILQ_INSERT_TAIL(&dyn->pfid_kif->pfik_dynaddrs, dyn, entry); entry 404 net/pf_if.c TAILQ_FOREACH(p, &kif->pfik_dynaddrs, entry) entry 568 net/pf_if.c TAILQ_REMOVE(&aw->p.dyn->pfid_kif->pfik_dynaddrs, aw->p.dyn, entry); entry 265 net/pf_osfp.c struct pf_osfp_entry *entry; entry 276 net/pf_osfp.c SLIST_FOREACH(entry, list, fp_entry) { entry 277 net/pf_osfp.c PF_OSFP_UNPACK(entry->fp_os, en_class, en_version, en_subtype); entry 282 net/pf_osfp.c entry->fp_class_nm, entry->fp_version_nm, entry 283 net/pf_osfp.c entry->fp_subtype_nm, os, entry->fp_os); entry 307 net/pf_osfp.c struct pf_osfp_entry *entry; entry 311 net/pf_osfp.c while ((entry = SLIST_FIRST(&fp->fp_oses))) { entry 313 net/pf_osfp.c pool_put(&pf_osfp_entry_pl, entry); entry 325 net/pf_osfp.c struct pf_osfp_entry *entry; entry 363 net/pf_osfp.c SLIST_FOREACH(entry, &fp->fp_oses, fp_entry) { entry 364 net/pf_osfp.c if (PF_OSFP_ENTRY_EQ(entry, &fpioc->fp_os)) entry 367 net/pf_osfp.c if ((entry = pool_get(&pf_osfp_entry_pl, PR_NOWAIT)) == NULL) entry 382 net/pf_osfp.c if ((entry = pool_get(&pf_osfp_entry_pl, PR_NOWAIT)) == NULL) { entry 388 net/pf_osfp.c memcpy(entry, &fpioc->fp_os, sizeof(*entry)); entry 391 net/pf_osfp.c entry->fp_class_nm[sizeof(entry->fp_class_nm)-1] = '\0'; entry 392 net/pf_osfp.c entry->fp_version_nm[sizeof(entry->fp_version_nm)-1] = '\0'; entry 393 net/pf_osfp.c entry->fp_subtype_nm[sizeof(entry->fp_subtype_nm)-1] = '\0'; entry 395 net/pf_osfp.c SLIST_INSERT_HEAD(&fp->fp_oses, entry, fp_entry); entry 521 net/pf_osfp.c struct pf_osfp_entry *entry; entry 528 net/pf_osfp.c SLIST_FOREACH(entry, &fp->fp_oses, fp_entry) { entry 537 net/pf_osfp.c memcpy(&fpioc->fp_os, entry, entry 172 net/pfvar.h TAILQ_ENTRY(pfi_dynaddr) entry; entry 633 net/pfvar.h RB_ENTRY(pf_src_node) entry; entry 1538 net/pfvar.h RB_PROTOTYPE(pf_src_tree, pf_src_node, entry, pf_src_compare); entry 56 sys/exec_ecoff.h u_long entry; entry 179 sys/lkm.h int (*entry)(struct lkm_table *, int, int); /* entry function */ entry 296 uvm/uvm_amap.c amap_extend(struct vm_map_entry *entry, vsize_t addsize) entry 298 uvm/uvm_amap.c struct vm_amap *amap = entry->aref.ar_amap; entry 299 uvm/uvm_amap.c int slotoff = entry->aref.ar_pageoff; entry 309 uvm/uvm_amap.c UVMHIST_LOG(maphist, " (entry=%p, addsize=%lu)", entry, addsize, 0, 0); entry 317 uvm/uvm_amap.c AMAP_B2SLOT(slotmapped, entry->end - entry->start); /* slots mapped */ entry 475 uvm/uvm_amap.c amap_share_protect(struct vm_map_entry *entry, vm_prot_t prot) entry 477 uvm/uvm_amap.c struct vm_amap *amap = entry->aref.ar_amap; entry 480 uvm/uvm_amap.c AMAP_B2SLOT(slots, (entry->end - entry->start)); entry 481 uvm/uvm_amap.c stop = entry->aref.ar_pageoff + slots; entry 485 uvm/uvm_amap.c for (lcv = entry->aref.ar_pageoff ; lcv < stop ; lcv++) { entry 498 uvm/uvm_amap.c if (slot < entry->aref.ar_pageoff || slot >= stop) entry 581 uvm/uvm_amap.c amap_copy(struct vm_map *map, struct vm_map_entry *entry, int waitf, entry 589 uvm/uvm_amap.c map, entry, waitf, 0); entry 595 uvm/uvm_amap.c if (entry->aref.ar_amap == NULL) { entry 603 uvm/uvm_amap.c if (canchunk && atop(entry->end - entry->start) >= entry 611 uvm/uvm_amap.c entry->start, entry->end, startva, endva); entry 612 uvm/uvm_amap.c UVM_MAP_CLIP_START(map, entry, startva); entry 615 uvm/uvm_amap.c UVM_MAP_CLIP_END(map, entry, endva); entry 619 uvm/uvm_amap.c entry->start, entry->end, 0, 0); entry 620 uvm/uvm_amap.c entry->aref.ar_pageoff = 0; entry 621 uvm/uvm_amap.c entry->aref.ar_amap = amap_alloc(entry->end - entry->start, 0, entry 623 uvm/uvm_amap.c if (entry->aref.ar_amap != NULL) entry 624 uvm/uvm_amap.c entry->etype &= ~UVM_ET_NEEDSCOPY; entry 638 uvm/uvm_amap.c if (entry->aref.ar_amap->am_ref == 1) { entry 639 uvm/uvm_amap.c entry->etype &= ~UVM_ET_NEEDSCOPY; entry 650 uvm/uvm_amap.c entry->aref.ar_amap, entry->aref.ar_amap->am_ref, 0, 0); entry 651 uvm/uvm_amap.c AMAP_B2SLOT(slots, entry->end - entry->start); entry 657 uvm/uvm_amap.c srcamap = entry->aref.ar_amap; entry 668 uvm/uvm_amap.c entry->etype &= ~UVM_ET_NEEDSCOPY; entry 681 uvm/uvm_amap.c srcamap->am_anon[entry->aref.ar_pageoff + lcv]; entry 706 uvm/uvm_amap.c amap_pp_adjref(srcamap, entry->aref.ar_pageoff, entry 707 uvm/uvm_amap.c (entry->end - entry->start) >> PAGE_SHIFT, -1); entry 715 uvm/uvm_amap.c entry->aref.ar_pageoff = 0; entry 716 uvm/uvm_amap.c entry->aref.ar_amap = amap; entry 717 uvm/uvm_amap.c entry->etype &= ~UVM_ET_NEEDSCOPY; entry 750 uvm/uvm_amap.c amap_cow_now(struct vm_map *map, struct vm_map_entry *entry) entry 752 uvm/uvm_amap.c struct vm_amap *amap = entry->aref.ar_amap; entry 382 uvm/uvm_device.c struct vm_map_entry *entry = ufi->entry; entry 383 uvm/uvm_device.c struct uvm_object *uobj = entry->object.uvm_obj; entry 400 uvm/uvm_device.c if (UVM_ET_ISCOPYONWRITE(entry)) { entry 402 uvm/uvm_device.c entry->etype, 0,0,0); entry 403 uvm/uvm_device.c uvmfault_unlockall(ufi, ufi->entry->aref.ar_amap, uobj, NULL); entry 422 uvm/uvm_device.c curr_offset = entry->offset + (vaddr - entry->start); entry 445 uvm/uvm_device.c mapprot = ufi->entry->protection; entry 461 uvm/uvm_device.c uvmfault_unlockall(ufi, ufi->entry->aref.ar_amap, entry 471 uvm/uvm_device.c uvmfault_unlockall(ufi, ufi->entry->aref.ar_amap, uobj, NULL); entry 254 uvm/uvm_fault.c if (UVM_ET_ISNEEDSCOPY(ufi->entry)) entry 255 uvm/uvm_fault.c amap_copy(ufi->map, ufi->entry, M_NOWAIT, TRUE, entry 262 uvm/uvm_fault.c if (UVM_ET_ISNEEDSCOPY(ufi->entry)) { entry 520 uvm/uvm_fault.c amap_lookup(&ufi->entry->aref, entry 521 uvm/uvm_fault.c ufi->orig_rvaddr - ufi->entry->start) != anon) { entry 555 uvm/uvm_fault.c #define MASK(entry) (UVM_ET_ISCOPYONWRITE(entry) ? \ entry 624 uvm/uvm_fault.c if ((ufi.entry->protection & access_type) != access_type) { entry 627 uvm/uvm_fault.c ufi.entry->protection, access_type, 0, 0); entry 639 uvm/uvm_fault.c enter_prot = ufi.entry->protection; entry 640 uvm/uvm_fault.c wired = VM_MAPENT_ISWIRED(ufi.entry) || (fault_type == VM_FAULT_WIRE); entry 651 uvm/uvm_fault.c if (UVM_ET_ISNEEDSCOPY(ufi.entry)) { entry 653 uvm/uvm_fault.c (ufi.entry->object.uvm_obj == NULL)) { entry 677 uvm/uvm_fault.c amap = ufi.entry->aref.ar_amap; /* top layer */ entry 678 uvm/uvm_fault.c uobj = ufi.entry->object.uvm_obj; /* bottom layer */ entry 701 uvm/uvm_fault.c KASSERT(uvmadvice[ufi.entry->advice].advice == entry 702 uvm/uvm_fault.c ufi.entry->advice); entry 703 uvm/uvm_fault.c nback = min(uvmadvice[ufi.entry->advice].nback, entry 704 uvm/uvm_fault.c (ufi.orig_rvaddr - ufi.entry->start) >> PAGE_SHIFT); entry 706 uvm/uvm_fault.c nforw = min(uvmadvice[ufi.entry->advice].nforw, entry 707 uvm/uvm_fault.c ((ufi.entry->end - ufi.orig_rvaddr) >> entry 731 uvm/uvm_fault.c UVMHIST_LOG(maphist, " entry=%p, amap=%p, obj=%p", ufi.entry, entry 740 uvm/uvm_fault.c amap_lookups(&ufi.entry->aref, startva - ufi.entry->start, entry 753 uvm/uvm_fault.c if (ufi.entry->advice == MADV_SEQUENTIAL && nback != 0) { entry 763 uvm/uvm_fault.c uoff = (startva - ufi.entry->start) + ufi.entry->offset; entry 841 uvm/uvm_fault.c (VM_MAPENT_ISWIRED(ufi.entry) ? PMAP_WIRED : 0)); entry 906 uvm/uvm_fault.c (void) uobj->pgops->pgo_get(uobj, ufi.entry->offset + entry 907 uvm/uvm_fault.c (startva - ufi.entry->start), entry 909 uvm/uvm_fault.c access_type & MASK(ufi.entry), entry 910 uvm/uvm_fault.c ufi.entry->advice, PGO_LOCKED); entry 971 uvm/uvm_fault.c enter_prot & MASK(ufi.entry), entry 1202 uvm/uvm_fault.c amap_add(&ufi.entry->aref, ufi.orig_rvaddr - ufi.entry->start, entry 1314 uvm/uvm_fault.c UVM_ET_ISCOPYONWRITE(ufi.entry); entry 1341 uvm/uvm_fault.c uoff = (ufi.orig_rvaddr - ufi.entry->start) + ufi.entry->offset; entry 1343 uvm/uvm_fault.c 0, access_type & MASK(ufi.entry), ufi.entry->advice, entry 1388 uvm/uvm_fault.c amap_lookup(&ufi.entry->aref, entry 1389 uvm/uvm_fault.c ufi.orig_rvaddr - ufi.entry->start))) { entry 1466 uvm/uvm_fault.c if (UVM_ET_ISCOPYONWRITE(ufi.entry)) entry 1668 uvm/uvm_fault.c amap_add(&ufi.entry->aref, ufi.orig_rvaddr - ufi.entry->start, entry 1827 uvm/uvm_fault.c vm_map_entry_t entry; entry 1848 uvm/uvm_fault.c if (uvm_map_lookup_entry(map, start, &entry) == FALSE) entry 1858 uvm/uvm_fault.c KASSERT(va >= entry->start); entry 1859 uvm/uvm_fault.c while (va >= entry->end) { entry 1860 uvm/uvm_fault.c KASSERT(entry->next != &map->header && entry 1861 uvm/uvm_fault.c entry->next->start <= entry->end); entry 1862 uvm/uvm_fault.c entry = entry->next; entry 1868 uvm/uvm_fault.c if (VM_MAPENT_ISWIRED(entry) == 0) entry 65 uvm/uvm_fault.h vm_map_entry_t entry; /* map entry (from 'map') */ entry 145 uvm/uvm_fault_i.h &ufi->entry)) { entry 153 uvm/uvm_fault_i.h if (ufi->entry->end - ufi->orig_rvaddr < ufi->size) entry 154 uvm/uvm_fault_i.h ufi->size = ufi->entry->end - ufi->orig_rvaddr; entry 160 uvm/uvm_fault_i.h if (UVM_ET_ISSUBMAP(ufi->entry)) { entry 161 uvm/uvm_fault_i.h tmpmap = ufi->entry->object.sub_map; entry 137 uvm/uvm_loan.c struct vm_aref *aref = &ufi->entry->aref; entry 138 uvm/uvm_loan.c struct uvm_object *uobj = ufi->entry->object.uvm_obj; entry 158 uvm/uvm_loan.c anon = amap_lookup(aref, curaddr - ufi->entry->start); entry 167 uvm/uvm_loan.c } else if (UVM_ET_ISCOPYONWRITE(ufi->entry)) { entry 353 uvm/uvm_loan.c result = uvmfault_anonget(ufi, ufi->entry->aref.ar_amap, anon); entry 414 uvm/uvm_loan.c struct vm_amap *amap = ufi->entry->aref.ar_amap; entry 415 uvm/uvm_loan.c struct uvm_object *uobj = ufi->entry->object.uvm_obj; entry 430 uvm/uvm_loan.c result = uobj->pgops->pgo_get(uobj, va - ufi->entry->start, entry 455 uvm/uvm_loan.c result = uobj->pgops->pgo_get(uobj, va - ufi->entry->start, entry 485 uvm/uvm_loan.c (locked && amap && amap_lookup(&ufi->entry->aref, entry 486 uvm/uvm_loan.c ufi->orig_rvaddr - ufi->entry->start))) { entry 623 uvm/uvm_loan.c uvmfault_unlockall(ufi, ufi->entry->aref.ar_amap, entry 624 uvm/uvm_loan.c ufi->entry->object.uvm_obj, NULL); entry 628 uvm/uvm_loan.c if (ufi->entry->object.uvm_obj) entry 630 uvm/uvm_loan.c &ufi->entry->object.uvm_obj->vmobjlock); entry 652 uvm/uvm_loan.c uvmfault_unlockall(ufi, ufi->entry->aref.ar_amap, entry 653 uvm/uvm_loan.c ufi->entry->object.uvm_obj, NULL); entry 667 uvm/uvm_loan.c if (ufi->entry->object.uvm_obj) entry 668 uvm/uvm_loan.c simple_lock(&ufi->entry->object.uvm_obj->vmobjlock); entry 143 uvm/uvm_map.c #define uvm_map_entry_link(map, after_where, entry) do { \ entry 145 uvm/uvm_map.c (entry)->prev = (after_where); \ entry 146 uvm/uvm_map.c (entry)->next = (after_where)->next; \ entry 147 uvm/uvm_map.c (entry)->prev->next = (entry); \ entry 148 uvm/uvm_map.c (entry)->next->prev = (entry); \ entry 149 uvm/uvm_map.c uvm_rb_insert(map, entry); \ entry 157 uvm/uvm_map.c #define uvm_map_entry_unlink(map, entry) do { \ entry 159 uvm/uvm_map.c (entry)->next->prev = (entry)->prev; \ entry 160 uvm/uvm_map.c (entry)->prev->next = (entry)->next; \ entry 161 uvm/uvm_map.c uvm_rb_remove(map, entry); \ entry 232 uvm/uvm_map.c uvm_rb_augment(struct vm_map_entry *entry) entry 234 uvm/uvm_map.c entry->space = uvm_rb_subtree_space(entry); entry 242 uvm/uvm_map.c uvm_rb_space(struct vm_map *map, struct vm_map_entry *entry) entry 247 uvm/uvm_map.c if ((next = entry->next) == &map->header) entry 248 uvm/uvm_map.c space = map->max_offset - entry->end; entry 251 uvm/uvm_map.c space = next->start - entry->end; entry 257 uvm/uvm_map.c uvm_rb_subtree_space(struct vm_map_entry *entry) entry 261 uvm/uvm_map.c space = entry->ownspace; entry 262 uvm/uvm_map.c if (RB_LEFT(entry, rb_entry)) { entry 263 uvm/uvm_map.c tmp = RB_LEFT(entry, rb_entry)->space; entry 268 uvm/uvm_map.c if (RB_RIGHT(entry, rb_entry)) { entry 269 uvm/uvm_map.c tmp = RB_RIGHT(entry, rb_entry)->space; entry 278 uvm/uvm_map.c uvm_rb_fixup(struct vm_map *map, struct vm_map_entry *entry) entry 282 uvm/uvm_map.c entry->ownspace = uvm_rb_space(map, entry); entry 283 uvm/uvm_map.c entry->space = uvm_rb_subtree_space(entry); entry 284 uvm/uvm_map.c } while ((entry = RB_PARENT(entry, rb_entry)) != NULL); entry 288 uvm/uvm_map.c uvm_rb_insert(struct vm_map *map, struct vm_map_entry *entry) entry 290 uvm/uvm_map.c vaddr_t space = uvm_rb_space(map, entry); entry 293 uvm/uvm_map.c entry->ownspace = entry->space = space; entry 294 uvm/uvm_map.c tmp = RB_INSERT(uvm_tree, &(map)->rbhead, entry); entry 299 uvm/uvm_map.c uvm_rb_fixup(map, entry); entry 300 uvm/uvm_map.c if (entry->prev != &map->header) entry 301 uvm/uvm_map.c uvm_rb_fixup(map, entry->prev); entry 305 uvm/uvm_map.c uvm_rb_remove(struct vm_map *map, struct vm_map_entry *entry) entry 309 uvm/uvm_map.c parent = RB_PARENT(entry, rb_entry); entry 310 uvm/uvm_map.c RB_REMOVE(uvm_tree, &(map)->rbhead, entry); entry 311 uvm/uvm_map.c if (entry->prev != &map->header) entry 312 uvm/uvm_map.c uvm_rb_fixup(map, entry->prev); entry 482 uvm/uvm_map.c uvm_map_entry_unwire(struct vm_map *map, struct vm_map_entry *entry) entry 485 uvm/uvm_map.c entry->wired_count = 0; entry 486 uvm/uvm_map.c uvm_fault_unwire_locked(map, entry->start, entry->end); entry 494 uvm/uvm_map.c uvm_map_reference_amap(struct vm_map_entry *entry, int flags) entry 496 uvm/uvm_map.c amap_ref(entry->aref.ar_amap, entry->aref.ar_pageoff, entry 497 uvm/uvm_map.c (entry->end - entry->start) >> PAGE_SHIFT, flags); entry 505 uvm/uvm_map.c uvm_map_unreference_amap(struct vm_map_entry *entry, int flags) entry 507 uvm/uvm_map.c amap_unref(entry->aref.ar_amap, entry->aref.ar_pageoff, entry 508 uvm/uvm_map.c (entry->end - entry->start) >> PAGE_SHIFT, flags); entry 581 uvm/uvm_map.c uvm_map_clip_start(struct vm_map *map, struct vm_map_entry *entry, entry 598 uvm/uvm_map.c uvm_mapent_copy(entry, new_entry); /* entry -> new_entry */ entry 602 uvm/uvm_map.c if (entry->object.uvm_obj) entry 603 uvm/uvm_map.c entry->offset += new_adj; /* shift start over */ entry 606 uvm/uvm_map.c entry->start = start; entry 609 uvm/uvm_map.c amap_splitref(&new_entry->aref, &entry->aref, new_adj); entry 612 uvm/uvm_map.c uvm_map_entry_link(map, entry->prev, new_entry); entry 614 uvm/uvm_map.c if (UVM_ET_ISSUBMAP(entry)) { entry 618 uvm/uvm_map.c if (UVM_ET_ISOBJ(entry) && entry 619 uvm/uvm_map.c entry->object.uvm_obj->pgops && entry 620 uvm/uvm_map.c entry->object.uvm_obj->pgops->pgo_reference) entry 621 uvm/uvm_map.c entry->object.uvm_obj->pgops->pgo_reference( entry 622 uvm/uvm_map.c entry->object.uvm_obj); entry 638 uvm/uvm_map.c uvm_map_clip_end(struct vm_map *map, struct vm_map_entry *entry, vaddr_t end) entry 650 uvm/uvm_map.c uvm_mapent_copy(entry, new_entry); /* entry -> new_entry */ entry 652 uvm/uvm_map.c new_entry->start = entry->end = end; entry 653 uvm/uvm_map.c new_adj = end - entry->start; entry 657 uvm/uvm_map.c if (entry->aref.ar_amap) entry 658 uvm/uvm_map.c amap_splitref(&entry->aref, &new_entry->aref, new_adj); entry 660 uvm/uvm_map.c uvm_rb_fixup(map, entry); entry 662 uvm/uvm_map.c uvm_map_entry_link(map, entry, new_entry); entry 664 uvm/uvm_map.c if (UVM_ET_ISSUBMAP(entry)) { entry 668 uvm/uvm_map.c if (UVM_ET_ISOBJ(entry) && entry 669 uvm/uvm_map.c entry->object.uvm_obj->pgops && entry 670 uvm/uvm_map.c entry->object.uvm_obj->pgops->pgo_reference) entry 671 uvm/uvm_map.c entry->object.uvm_obj->pgops->pgo_reference( entry 672 uvm/uvm_map.c entry->object.uvm_obj); entry 953 uvm/uvm_map.c struct vm_map_entry **entry) entry 962 uvm/uvm_map.c map, address, entry, 0); entry 992 uvm/uvm_map.c *entry = cur; entry 1022 uvm/uvm_map.c *entry = cur; entry 1031 uvm/uvm_map.c *entry = prev; entry 1048 uvm/uvm_map.c *entry = cur; entry 1059 uvm/uvm_map.c *entry = cur->prev; entry 1060 uvm/uvm_map.c SAVE_HINT(map, map->hint, *entry); entry 1149 uvm/uvm_map.c struct vm_map_entry *entry, *next, *tmp; entry 1189 uvm/uvm_map.c if ((entry = map->first_free) != &map->header) entry 1190 uvm/uvm_map.c hint = entry->end; entry 1201 uvm/uvm_map.c entry = tmp; entry 1210 uvm/uvm_map.c next = entry->next; entry 1220 uvm/uvm_map.c if (uvm_map_spacefits(map, &hint, length, entry->next, uoffset, align)) entry 1261 uvm/uvm_map.c entry = tmp; entry 1272 uvm/uvm_map.c entry = prev; entry 1292 uvm/uvm_map.c entry = tmp; entry 1307 uvm/uvm_map.c for (;; hint = (entry = next)->end) { entry 1336 uvm/uvm_map.c next = entry->next; entry 1341 uvm/uvm_map.c SAVE_HINT(map, map->hint, entry); entry 1344 uvm/uvm_map.c return (entry); entry 1374 uvm/uvm_map.c struct vm_map_entry *entry, *first_entry, *next; entry 1394 uvm/uvm_map.c entry = first_entry; entry 1395 uvm/uvm_map.c UVM_MAP_CLIP_START(map, entry, start); entry 1397 uvm/uvm_map.c SAVE_HINT(map, entry, entry->prev); entry 1400 uvm/uvm_map.c entry = first_entry->next; entry 1408 uvm/uvm_map.c map->first_free = entry->prev; entry 1434 uvm/uvm_map.c while ((entry != &map->header) && (entry->start < end)) { entry 1436 uvm/uvm_map.c UVM_MAP_CLIP_END(map, entry, end); entry 1437 uvm/uvm_map.c next = entry->next; entry 1438 uvm/uvm_map.c len = entry->end - entry->start; entry 1439 uvm/uvm_map.c if (p && entry->object.uvm_obj == NULL) entry 1447 uvm/uvm_map.c if (VM_MAPENT_ISWIRED(entry)) entry 1448 uvm/uvm_map.c uvm_map_entry_unwire(map, entry); entry 1455 uvm/uvm_map.c uvm_km_pgremove_intrsafe(entry->start, entry->end); entry 1456 uvm/uvm_map.c pmap_kremove(entry->start, len); entry 1457 uvm/uvm_map.c } else if (UVM_ET_ISOBJ(entry) && entry 1458 uvm/uvm_map.c UVM_OBJ_IS_KERN_OBJECT(entry->object.uvm_obj)) { entry 1496 uvm/uvm_map.c pmap_remove(pmap_kernel(), entry->start, entry->end); entry 1497 uvm/uvm_map.c uvm_km_pgremove(entry->object.uvm_obj, entry 1498 uvm/uvm_map.c entry->start - vm_map_min(kernel_map), entry 1499 uvm/uvm_map.c entry->end - vm_map_min(kernel_map)); entry 1505 uvm/uvm_map.c entry->etype &= ~UVM_ET_OBJ; entry 1506 uvm/uvm_map.c entry->object.uvm_obj = NULL; /* to be safe */ entry 1512 uvm/uvm_map.c pmap_remove(map->pmap, entry->start, entry->end); entry 1519 uvm/uvm_map.c UVMHIST_LOG(maphist, " removed map entry %p", entry, 0, 0,0); entry 1522 uvm/uvm_map.c SAVE_HINT(map, entry, entry->prev); entry 1524 uvm/uvm_map.c uvm_map_entry_unlink(map, entry); entry 1526 uvm/uvm_map.c entry->next = first_entry; entry 1527 uvm/uvm_map.c first_entry = entry; entry 1528 uvm/uvm_map.c entry = next; /* next entry, please */ entry 1785 uvm/uvm_map.c struct vm_map_entry *chain, *endchain, *entry, *orig_entry, *newentry; entry 1830 uvm/uvm_map.c if (uvm_map_lookup_entry(srcmap, start, &entry)) { entry 1843 uvm/uvm_map.c fudge = start - entry->start; entry 1851 uvm/uvm_map.c UVM_MAP_CLIP_START(srcmap, entry, start); entry 1852 uvm/uvm_map.c SAVE_HINT(srcmap, srcmap->hint, entry->prev); entry 1863 uvm/uvm_map.c entry = entry->next; entry 1868 uvm/uvm_map.c orig_entry = entry; entry 1876 uvm/uvm_map.c while (entry->start < end && entry != &srcmap->header) { entry 1880 uvm/uvm_map.c UVM_MAP_CLIP_END(srcmap, entry, end); entry 1883 uvm/uvm_map.c if (UVM_ET_ISNEEDSCOPY(entry)) { entry 1885 uvm/uvm_map.c oldstart = entry->start; entry 1888 uvm/uvm_map.c amap_copy(srcmap, entry, M_NOWAIT, TRUE, start, end); entry 1889 uvm/uvm_map.c if (UVM_ET_ISNEEDSCOPY(entry)) { /* failed? */ entry 1896 uvm/uvm_map.c fudge = fudge - (entry->start - oldstart); entry 1902 uvm/uvm_map.c oldoffset = (entry->start + fudge) - start; entry 1916 uvm/uvm_map.c newentry->start + (entry->end - (entry->start + fudge)); entry 1919 uvm/uvm_map.c newentry->object.uvm_obj = entry->object.uvm_obj; entry 1924 uvm/uvm_map.c newentry->offset = entry->offset + fudge; entry 1928 uvm/uvm_map.c newentry->etype = entry->etype; entry 1930 uvm/uvm_map.c entry->max_protection : entry->protection; entry 1931 uvm/uvm_map.c newentry->max_protection = entry->max_protection; entry 1932 uvm/uvm_map.c newentry->inheritance = entry->inheritance; entry 1934 uvm/uvm_map.c newentry->aref.ar_amap = entry->aref.ar_amap; entry 1937 uvm/uvm_map.c entry->aref.ar_pageoff + (fudge >> PAGE_SHIFT); entry 1943 uvm/uvm_map.c newentry->advice = entry->advice; entry 1955 uvm/uvm_map.c if ((flags & UVM_EXTRACT_CONTIG) && entry->end < end && entry 1956 uvm/uvm_map.c (entry->next == &srcmap->header || entry 1957 uvm/uvm_map.c entry->next->start != entry->end)) { entry 1961 uvm/uvm_map.c entry = entry->next; entry 2009 uvm/uvm_map.c entry = orig_entry; entry 2013 uvm/uvm_map.c while (entry->start < end && entry != &srcmap->header) { entry 2015 uvm/uvm_map.c oldoffset = (entry->start + fudge) - start; entry 2016 uvm/uvm_map.c elen = MIN(end, entry->end) - entry 2017 uvm/uvm_map.c (entry->start + fudge); entry 2020 uvm/uvm_map.c entry->start + fudge); entry 2025 uvm/uvm_map.c pmap_remove(srcmap->pmap, entry->start, entry 2026 uvm/uvm_map.c entry->end); entry 2027 uvm/uvm_map.c oldentry = entry; /* save entry */ entry 2028 uvm/uvm_map.c entry = entry->next; /* advance */ entry 2034 uvm/uvm_map.c entry = entry->next; /* advance */ entry 2122 uvm/uvm_map.c struct vm_map_entry *entry; entry 2129 uvm/uvm_map.c if (uvm_map_lookup_entry(map, start, &entry)) { entry 2130 uvm/uvm_map.c UVM_MAP_CLIP_START(map, entry, start); entry 2131 uvm/uvm_map.c UVM_MAP_CLIP_END(map, entry, end); /* to be safe */ entry 2133 uvm/uvm_map.c entry = NULL; entry 2136 uvm/uvm_map.c if (entry != NULL && entry 2137 uvm/uvm_map.c entry->start == start && entry->end == end && entry 2138 uvm/uvm_map.c entry->object.uvm_obj == NULL && entry->aref.ar_amap == NULL && entry 2139 uvm/uvm_map.c !UVM_ET_ISCOPYONWRITE(entry) && !UVM_ET_ISNEEDSCOPY(entry)) { entry 2140 uvm/uvm_map.c entry->etype |= UVM_ET_SUBMAP; entry 2141 uvm/uvm_map.c entry->object.sub_map = submap; entry 2142 uvm/uvm_map.c entry->offset = 0; entry 2160 uvm/uvm_map.c #define MASK(entry) (UVM_ET_ISCOPYONWRITE(entry) ? \ entry 2168 uvm/uvm_map.c struct vm_map_entry *current, *entry; entry 2178 uvm/uvm_map.c if (uvm_map_lookup_entry(map, start, &entry)) { entry 2179 uvm/uvm_map.c UVM_MAP_CLIP_START(map, entry, start); entry 2181 uvm/uvm_map.c entry = entry->next; entry 2188 uvm/uvm_map.c current = entry; entry 2203 uvm/uvm_map.c current = entry; entry 2224 uvm/uvm_map.c if ((current->protection & MASK(entry)) == PROT_NONE && entry 2225 uvm/uvm_map.c VM_MAPENT_ISWIRED(entry)) entry 2228 uvm/uvm_map.c current->protection & MASK(entry)); entry 2238 uvm/uvm_map.c VM_MAPENT_ISWIRED(entry) == 0 && entry 2241 uvm/uvm_map.c if (uvm_map_pageable(map, entry->start, entry->end, entry 2285 uvm/uvm_map.c struct vm_map_entry *entry, *temp_entry; entry 2305 uvm/uvm_map.c entry = temp_entry; entry 2306 uvm/uvm_map.c UVM_MAP_CLIP_START(map, entry, start); entry 2308 uvm/uvm_map.c entry = temp_entry->next; entry 2311 uvm/uvm_map.c while ((entry != &map->header) && (entry->start < end)) { entry 2312 uvm/uvm_map.c UVM_MAP_CLIP_END(map, entry, end); entry 2313 uvm/uvm_map.c entry->inheritance = new_inheritance; entry 2314 uvm/uvm_map.c entry = entry->next; entry 2331 uvm/uvm_map.c struct vm_map_entry *entry, *temp_entry; entry 2339 uvm/uvm_map.c entry = temp_entry; entry 2340 uvm/uvm_map.c UVM_MAP_CLIP_START(map, entry, start); entry 2342 uvm/uvm_map.c entry = temp_entry->next; entry 2349 uvm/uvm_map.c while ((entry != &map->header) && (entry->start < end)) { entry 2350 uvm/uvm_map.c UVM_MAP_CLIP_END(map, entry, end); entry 2364 uvm/uvm_map.c entry->advice = new_advice; entry 2365 uvm/uvm_map.c entry = entry->next; entry 2391 uvm/uvm_map.c struct vm_map_entry *entry, *start_entry, *failed_entry; entry 2421 uvm/uvm_map.c entry = start_entry; entry 2428 uvm/uvm_map.c UVM_MAP_CLIP_START(map, entry, start); entry 2435 uvm/uvm_map.c while ((entry != &map->header) && (entry->start < end)) { entry 2436 uvm/uvm_map.c if (entry->wired_count == 0 || entry 2437 uvm/uvm_map.c (entry->end < end && entry 2438 uvm/uvm_map.c (entry->next == &map->header || entry 2439 uvm/uvm_map.c entry->next->start > entry->end))) { entry 2446 uvm/uvm_map.c entry = entry->next; entry 2455 uvm/uvm_map.c entry = start_entry; entry 2456 uvm/uvm_map.c while ((entry != &map->header) && (entry->start < end)) { entry 2457 uvm/uvm_map.c UVM_MAP_CLIP_END(map, entry, end); entry 2458 uvm/uvm_map.c if (VM_MAPENT_ISWIRED(entry)) entry 2459 uvm/uvm_map.c uvm_map_entry_unwire(map, entry); entry 2460 uvm/uvm_map.c entry = entry->next; entry 2489 uvm/uvm_map.c while ((entry != &map->header) && (entry->start < end)) { entry 2490 uvm/uvm_map.c if (VM_MAPENT_ISWIRED(entry) == 0) { /* not already wired? */ entry 2500 uvm/uvm_map.c if (!UVM_ET_ISSUBMAP(entry)) { /* not submap */ entry 2501 uvm/uvm_map.c if (UVM_ET_ISNEEDSCOPY(entry) && entry 2502 uvm/uvm_map.c ((entry->protection & VM_PROT_WRITE) || entry 2503 uvm/uvm_map.c (entry->object.uvm_obj == NULL))) { entry 2504 uvm/uvm_map.c amap_copy(map, entry, M_WAITOK, TRUE, entry 2510 uvm/uvm_map.c UVM_MAP_CLIP_START(map, entry, start); entry 2511 uvm/uvm_map.c UVM_MAP_CLIP_END(map, entry, end); entry 2512 uvm/uvm_map.c entry->wired_count++; entry 2518 uvm/uvm_map.c if (entry->protection == VM_PROT_NONE || entry 2519 uvm/uvm_map.c (entry->end < end && entry 2520 uvm/uvm_map.c (entry->next == &map->header || entry 2521 uvm/uvm_map.c entry->next->start > entry->end))) { entry 2528 uvm/uvm_map.c while (entry != &map->header && entry->end > start) { entry 2529 uvm/uvm_map.c entry->wired_count--; entry 2530 uvm/uvm_map.c entry = entry->prev; entry 2537 uvm/uvm_map.c entry = entry->next; entry 2551 uvm/uvm_map.c entry = start_entry; entry 2552 uvm/uvm_map.c while (entry != &map->header && entry->start < end) { entry 2553 uvm/uvm_map.c if (entry->wired_count == 1) { entry 2554 uvm/uvm_map.c rv = uvm_fault_wire(map, entry->start, entry->end, entry 2555 uvm/uvm_map.c entry->protection); entry 2565 uvm/uvm_map.c entry = entry->next; entry 2587 uvm/uvm_map.c failed_entry = entry; entry 2588 uvm/uvm_map.c while (entry != &map->header && entry->start < end) { entry 2589 uvm/uvm_map.c entry->wired_count--; entry 2590 uvm/uvm_map.c entry = entry->next; entry 2598 uvm/uvm_map.c entry = start_entry; entry 2599 uvm/uvm_map.c while (entry != failed_entry) { entry 2600 uvm/uvm_map.c entry->wired_count--; entry 2601 uvm/uvm_map.c if (VM_MAPENT_ISWIRED(entry) == 0) entry 2602 uvm/uvm_map.c uvm_map_entry_unwire(map, entry); entry 2603 uvm/uvm_map.c entry = entry->next; entry 2641 uvm/uvm_map.c struct vm_map_entry *entry, *failed_entry; entry 2663 uvm/uvm_map.c for (entry = map->header.next; entry != &map->header; entry 2664 uvm/uvm_map.c entry = entry->next) { entry 2665 uvm/uvm_map.c if (VM_MAPENT_ISWIRED(entry)) entry 2666 uvm/uvm_map.c uvm_map_entry_unwire(map, entry); entry 2717 uvm/uvm_map.c for (size = 0, entry = map->header.next; entry != &map->header; entry 2718 uvm/uvm_map.c entry = entry->next) { entry 2719 uvm/uvm_map.c if (entry->protection != VM_PROT_NONE && entry 2720 uvm/uvm_map.c VM_MAPENT_ISWIRED(entry) == 0) { /* not already wired? */ entry 2721 uvm/uvm_map.c size += entry->end - entry->start; entry 2743 uvm/uvm_map.c for (entry = map->header.next; entry != &map->header; entry 2744 uvm/uvm_map.c entry = entry->next) { entry 2745 uvm/uvm_map.c if (entry->protection == VM_PROT_NONE) entry 2747 uvm/uvm_map.c if (VM_MAPENT_ISWIRED(entry) == 0) { /* not already wired? */ entry 2755 uvm/uvm_map.c if (!UVM_ET_ISSUBMAP(entry)) { /* not submap */ entry 2756 uvm/uvm_map.c if (UVM_ET_ISNEEDSCOPY(entry) && entry 2757 uvm/uvm_map.c ((entry->protection & VM_PROT_WRITE) || entry 2758 uvm/uvm_map.c (entry->object.uvm_obj == NULL))) { entry 2759 uvm/uvm_map.c amap_copy(map, entry, M_WAITOK, TRUE, entry 2760 uvm/uvm_map.c entry->start, entry->end); entry 2765 uvm/uvm_map.c entry->wired_count++; entry 2778 uvm/uvm_map.c for (error = 0, entry = map->header.next; entry 2779 uvm/uvm_map.c entry != &map->header && error == 0; entry 2780 uvm/uvm_map.c entry = entry->next) { entry 2781 uvm/uvm_map.c if (entry->wired_count == 1) { entry 2782 uvm/uvm_map.c error = uvm_fault_wire(map, entry->start, entry->end, entry 2783 uvm/uvm_map.c entry->protection); entry 2805 uvm/uvm_map.c failed_entry = entry; entry 2806 uvm/uvm_map.c for (/* nothing */; entry != &map->header; entry 2807 uvm/uvm_map.c entry = entry->next) { entry 2808 uvm/uvm_map.c if (entry->protection == VM_PROT_NONE) entry 2810 uvm/uvm_map.c entry->wired_count--; entry 2819 uvm/uvm_map.c for (entry = map->header.next; entry != failed_entry; entry 2820 uvm/uvm_map.c entry = entry->next) { entry 2821 uvm/uvm_map.c if (entry->protection == VM_PROT_NONE) entry 2823 uvm/uvm_map.c entry->wired_count--; entry 2824 uvm/uvm_map.c if (VM_MAPENT_ISWIRED(entry)) entry 2825 uvm/uvm_map.c uvm_map_entry_unwire(map, entry); entry 2861 uvm/uvm_map.c struct vm_map_entry *current, *entry; entry 2878 uvm/uvm_map.c if (uvm_map_lookup_entry(map, start, &entry) == FALSE) { entry 2887 uvm/uvm_map.c for (current = entry; current->start < end; current = current->next) { entry 2901 uvm/uvm_map.c for (current = entry; current->start < end; current = current->next) { entry 3027 uvm/uvm_map.c ((entry->max_protection & VM_PROT_WRITE) != 0 && entry 3028 uvm/uvm_map.c (entry->etype & UVM_ET_COPYONWRITE) == 0))) { entry 3055 uvm/uvm_map.c struct vm_map_entry *entry; entry 3061 uvm/uvm_map.c entry = tmp_entry; entry 3063 uvm/uvm_map.c if (entry == &map->header) { entry 3071 uvm/uvm_map.c if (start < entry->start) { entry 3079 uvm/uvm_map.c if ((entry->protection & protection) != protection) { entry 3085 uvm/uvm_map.c start = entry->end; entry 3086 uvm/uvm_map.c entry = entry->next; entry 3618 uvm/uvm_map.c struct vm_map_entry *entry; entry 3633 uvm/uvm_map.c for (entry = map->header.next; entry != &map->header; entry 3634 uvm/uvm_map.c entry = entry->next) { entry 3636 uvm/uvm_map.c entry, entry->start, entry->end, entry->object.uvm_obj, entry 3637 uvm/uvm_map.c (long long)entry->offset, entry->aref.ar_amap, entry 3638 uvm/uvm_map.c entry->aref.ar_pageoff); entry 3642 uvm/uvm_map.c (entry->etype & UVM_ET_SUBMAP) ? 'T' : 'F', entry 3643 uvm/uvm_map.c (entry->etype & UVM_ET_COPYONWRITE) ? 'T' : 'F', entry 3644 uvm/uvm_map.c (entry->etype & UVM_ET_NEEDSCOPY) ? 'T' : 'F', entry 3645 uvm/uvm_map.c entry->protection, entry->max_protection, entry 3646 uvm/uvm_map.c entry->inheritance, entry->wired_count, entry->advice); entry 160 uvm/uvm_map.h #define VM_MAPENT_ISWIRED(entry) ((entry)->wired_count != 0) entry 234 uvm/uvm_meter.c struct vm_map_entry * entry; entry 280 uvm/uvm_meter.c for (map = &p->p_vmspace->vm_map, entry = map->header.next; entry 281 uvm/uvm_meter.c entry != &map->header; entry = entry->next) { entry 282 uvm/uvm_meter.c if (entry->is_a_map || entry->is_sub_map || entry 283 uvm/uvm_meter.c entry->object.uvm_obj == NULL) entry 254 uvm/uvm_mmap.c vm_map_entry_t entry; entry 284 uvm/uvm_mmap.c if (uvm_map_lookup_entry(map, start, &entry) == FALSE) { entry 290 uvm/uvm_mmap.c entry != &map->header && entry->start < end; entry 291 uvm/uvm_mmap.c entry = entry->next) { entry 292 uvm/uvm_mmap.c KASSERT(!UVM_ET_ISSUBMAP(entry)); entry 293 uvm/uvm_mmap.c KASSERT(start >= entry->start); entry 296 uvm/uvm_mmap.c if (entry->end < end && entry 297 uvm/uvm_mmap.c (entry->next == &map->header || entry 298 uvm/uvm_mmap.c entry->next->start > entry->end)) { entry 303 uvm/uvm_mmap.c lim = end < entry->end ? end : entry->end; entry 309 uvm/uvm_mmap.c if (UVM_ET_ISOBJ(entry)) { entry 310 uvm/uvm_mmap.c KASSERT(!UVM_OBJ_IS_KERN_OBJECT(entry->object.uvm_obj)); entry 311 uvm/uvm_mmap.c if (entry->object.uvm_obj->pgops->pgo_releasepg entry 321 uvm/uvm_mmap.c amap = entry->aref.ar_amap; /* top layer */ entry 322 uvm/uvm_mmap.c uobj = entry->object.uvm_obj; /* bottom layer */ entry 331 uvm/uvm_mmap.c anon = amap_lookup(&entry->aref, entry 332 uvm/uvm_mmap.c start - entry->start); entry 346 uvm/uvm_mmap.c entry->offset + (start - entry->start)); entry 678 uvm/uvm_mmap.c vm_map_entry_t entry; entry 681 uvm/uvm_mmap.c rv = uvm_map_lookup_entry(map, addr, &entry); entry 683 uvm/uvm_mmap.c addr = entry->start; entry 684 uvm/uvm_mmap.c size = entry->end - entry->start; entry 311 uvm/uvm_page.h #define VM_PAGE_TO_PHYS(entry) ((entry)->phys_addr) entry 409 uvm/uvm_page.h #define VM_PAGE_IS_FREE(entry) ((entry)->pg_flags & PQ_FREE) entry 188 uvm/uvm_unix.c vm_map_entry_t entry; entry 196 uvm/uvm_unix.c for (entry = map->header.next; entry != &map->header; entry 197 uvm/uvm_unix.c entry = entry->next) { entry 200 uvm/uvm_unix.c if (UVM_ET_ISSUBMAP(entry)) { entry 204 uvm/uvm_unix.c if (!(entry->protection & VM_PROT_WRITE)) entry 210 uvm/uvm_unix.c if (entry->object.uvm_obj != NULL && entry 211 uvm/uvm_unix.c UVM_OBJ_IS_DEVICE(entry->object.uvm_obj)) entry 214 uvm/uvm_unix.c start = entry->start; entry 215 uvm/uvm_unix.c end = entry->end;