entry 153 arch/i386/i386/apm.c u_int32_t entry;
entry 829 arch/i386/i386/apm.c apm_ep.entry = ap->apm_entry;
entry 212 arch/i386/i386/bios.c if (h->entry <= BIOS32_START || h->entry >= BIOS32_END)
entry 216 arch/i386/i386/bios.c bios32_entry.offset = (u_int32_t)ISA_HOLE_VADDR(h->entry);
entry 217 arch/i386/i386/bios.c printf(", BIOS32 rev. %d @ 0x%lx", h->rev, h->entry);
entry 716 arch/i386/i386/mpbios.c const struct mpbios_proc *entry = (const struct mpbios_proc *)ent;
entry 721 arch/i386/i386/mpbios.c if (!(entry->cpu_flags & PROCENTRY_FLAG_EN))
entry 725 arch/i386/i386/mpbios.c if (entry->cpu_flags & PROCENTRY_FLAG_BP)
entry 731 arch/i386/i386/mpbios.c caa.cpu_number = entry->apic_id;
entry 734 arch/i386/i386/mpbios.c caa.cpu_signature = entry->cpu_signature;
entry 740 arch/i386/i386/mpbios.c caa.feature_flags = entry->feature_flags;
entry 763 arch/i386/i386/mpbios.c mp_cfg_special_intr(const struct mpbios_int *entry, u_int32_t *redir)
entry 775 arch/i386/i386/mpbios.c switch (entry->int_type) {
entry 793 arch/i386/i386/mpbios.c panic("unknown MPS interrupt type %d", entry->int_type);
entry 800 arch/i386/i386/mpbios.c mp_cfg_pci_intr(const struct mpbios_int *entry, u_int32_t *redir)
entry 802 arch/i386/i386/mpbios.c int mpspo = entry->int_flags & 0x03; /* XXX magic */
entry 803 arch/i386/i386/mpbios.c int mpstrig = (entry->int_flags >> 2) & 0x03; /* XXX magic */
entry 818 arch/i386/i386/mpbios.c if (entry->int_type != MPS_INTTYPE_INT) {
entry 819 arch/i386/i386/mpbios.c mp_cfg_special_intr(entry, redir);
entry 838 arch/i386/i386/mpbios.c mp_cfg_eisa_intr (const struct mpbios_int *entry, u_int32_t *redir)
entry 840 arch/i386/i386/mpbios.c int mpspo = entry->int_flags & 0x03; /* XXX magic */
entry 841 arch/i386/i386/mpbios.c int mpstrig = (entry->int_flags >> 2) & 0x03; /* XXX magic */
entry 856 arch/i386/i386/mpbios.c if (entry->int_type != MPS_INTTYPE_INT) {
entry 857 arch/i386/i386/mpbios.c mp_cfg_special_intr(entry, redir);
entry 874 arch/i386/i386/mpbios.c if (mp_busses[entry->src_bus_id].mb_data &
entry 875 arch/i386/i386/mpbios.c (1<<entry->src_bus_irq)) {
entry 888 arch/i386/i386/mpbios.c mp_cfg_isa_intr(const struct mpbios_int *entry, u_int32_t *redir)
entry 890 arch/i386/i386/mpbios.c int mpspo = entry->int_flags & 0x03; /* XXX magic */
entry 891 arch/i386/i386/mpbios.c int mpstrig = (entry->int_flags >> 2) & 0x03; /* XXX magic */
entry 906 arch/i386/i386/mpbios.c if (entry->int_type != MPS_INTTYPE_INT) {
entry 907 arch/i386/i386/mpbios.c mp_cfg_special_intr(entry, redir);
entry 960 arch/i386/i386/mpbios.c const struct mpbios_bus *entry = (const struct mpbios_bus *)ent;
entry 961 arch/i386/i386/mpbios.c int bus_id = entry->bus_id;
entry 964 arch/i386/i386/mpbios.c bus_id, entry->bus_type);
entry 973 arch/i386/i386/mpbios.c self->dv_xname, bus_id, entry->bus_type);
entry 979 arch/i386/i386/mpbios.c if (memcmp(entry->bus_type, "PCI ", 6) == 0) {
entry 984 arch/i386/i386/mpbios.c } else if (memcmp(entry->bus_type, "EISA ", 6) == 0) {
entry 997 arch/i386/i386/mpbios.c } else if (memcmp(entry->bus_type, "ISA ", 6) == 0) {
entry 1009 arch/i386/i386/mpbios.c entry->bus_type);
entry 1017 arch/i386/i386/mpbios.c const struct mpbios_ioapic *entry = (const struct mpbios_ioapic *)ent;
entry 1021 arch/i386/i386/mpbios.c if (!(entry->apic_flags & IOAPICENTRY_FLAG_EN))
entry 1025 arch/i386/i386/mpbios.c aaa.apic_id = entry->apic_id;
entry 1026 arch/i386/i386/mpbios.c aaa.apic_version = entry->apic_version;
entry 1027 arch/i386/i386/mpbios.c aaa.apic_address = (u_int32_t)entry->apic_address;
entry 1037 arch/i386/i386/mpbios.c const struct mpbios_int *entry = (const struct mpbios_int *)ent;
entry 1038 arch/i386/i386/mpbios.c struct mpbios_int rw_entry = *entry;
entry 1044 arch/i386/i386/mpbios.c u_int32_t id = IOAPIC_REMAPPED_ID(entry->dst_apic_id);
entry 1045 arch/i386/i386/mpbios.c u_int32_t pin = entry->dst_apic_int;
entry 1046 arch/i386/i386/mpbios.c u_int32_t bus = entry->src_bus_id;
entry 1047 arch/i386/i386/mpbios.c u_int32_t dev = entry->src_bus_irq;
entry 1048 arch/i386/i386/mpbios.c u_int32_t type = entry->int_type;
entry 1049 arch/i386/i386/mpbios.c u_int32_t flags = entry->int_flags;
entry 1084 arch/i386/i386/mpbios.c if (entry->type == MPS_MCT_IOINT) {
entry 72 arch/i386/include/biosvar.h u_int32_t entry; /* initialization entry point */
entry 83 arch/i386/include/biosvar.h u_int32_t entry; /* 04: entry point */
entry 337 arch/i386/pci/pci_intr_fixup.c int entry;
entry 342 arch/i386/pci/pci_intr_fixup.c for (entry = 0; entry < pcibios_pir_table_nentries; entry++) {
entry 343 arch/i386/pci/pci_intr_fixup.c pir = &pcibios_pir_table[entry];
entry 371 arch/i386/pci/pci_intr_fixup.c int entry, pin, link;
entry 383 arch/i386/pci/pci_intr_fixup.c for (entry = 0; entry < pcibios_pir_table_nentries; entry++) {
entry 384 arch/i386/pci/pci_intr_fixup.c pir = &pcibios_pir_table[entry];
entry 48 arch/i386/stand/libsa/exec_i386.c u_long entry;
entry 71 arch/i386/stand/libsa/exec_i386.c entry = marks[MARK_ENTRY] & 0x0fffffff;
entry 73 arch/i386/stand/libsa/exec_i386.c printf("entry point at 0x%x\n", (int) entry);
entry 75 arch/i386/stand/libsa/exec_i386.c (*(startfuncp)entry)(howto, bootdev, BOOTARG_APIVER,
entry 350 dev/acpi/acpi.c struct acpi_q *entry;
entry 396 dev/acpi/acpi.c SIMPLEQ_FOREACH(entry, &sc->sc_tables, q_next) {
entry 397 dev/acpi/acpi.c if (memcmp(entry->q_table, FADT_SIG,
entry 399 dev/acpi/acpi.c sc->sc_fadt = entry->q_table;
entry 432 dev/acpi/acpi.c acpi_load_dsdt(sc->sc_fadt->dsdt, &entry);
entry 434 dev/acpi/acpi.c acpi_load_dsdt(sc->sc_fadt->x_dsdt, &entry);
entry 436 dev/acpi/acpi.c if (entry == NULL)
entry 438 dev/acpi/acpi.c SIMPLEQ_INSERT_HEAD(&sc->sc_tables, entry, q_next);
entry 440 dev/acpi/acpi.c p_dsdt = entry->q_table;
entry 445 dev/acpi/acpi.c SIMPLEQ_FOREACH(entry, &sc->sc_tables, q_next) {
entry 446 dev/acpi/acpi.c if (memcmp(entry->q_table, SSDT_SIG,
entry 448 dev/acpi/acpi.c p_dsdt = entry->q_table;
entry 515 dev/acpi/acpi.c SIMPLEQ_FOREACH(entry, &sc->sc_tables, q_next) {
entry 516 dev/acpi/acpi.c printf("%.4s ", entry->q_table);
entry 542 dev/acpi/acpi.c SIMPLEQ_FOREACH(entry, &sc->sc_tables, q_next) {
entry 552 dev/acpi/acpi.c aaa.aaa_table = entry->q_table;
entry 717 dev/acpi/acpi.c struct acpi_q *entry;
entry 719 dev/acpi/acpi.c entry = malloc(len + sizeof(struct acpi_q), M_DEVBUF, M_NOWAIT);
entry 721 dev/acpi/acpi.c if (entry != NULL) {
entry 723 dev/acpi/acpi.c free(entry, M_DEVBUF);
entry 726 dev/acpi/acpi.c memcpy(entry->q_data, handle.va, len);
entry 727 dev/acpi/acpi.c entry->q_table = entry->q_data;
entry 729 dev/acpi/acpi.c SIMPLEQ_INSERT_TAIL(queue, entry, q_next);
entry 157 dev/acpi/acpimadt.c union acpi_madt_entry *entry = (union acpi_madt_entry *)addr;
entry 159 dev/acpi/acpimadt.c switch (entry->madt_lapic.apic_type) {
entry 162 dev/acpi/acpimadt.c self->dv_xname, entry->madt_lapic.acpi_proc_id,
entry 163 dev/acpi/acpimadt.c entry->madt_lapic.apic_id,
entry 164 dev/acpi/acpimadt.c entry->madt_lapic.flags);
entry 166 dev/acpi/acpimadt.c lapic_map[entry->madt_lapic.acpi_proc_id] =
entry 167 dev/acpi/acpimadt.c entry->madt_lapic.apic_id;
entry 172 dev/acpi/acpimadt.c if ((entry->madt_lapic.flags & ACPI_PROC_ENABLE) == 0)
entry 178 dev/acpi/acpimadt.c caa.cpu_number = entry->madt_lapic.apic_id;
entry 193 dev/acpi/acpimadt.c self->dv_xname, entry->madt_ioapic.acpi_ioapic_id,
entry 194 dev/acpi/acpimadt.c entry->madt_ioapic.address,
entry 195 dev/acpi/acpimadt.c entry->madt_ioapic.global_int_base);
entry 202 dev/acpi/acpimadt.c aaa.apic_id = entry->madt_ioapic.acpi_ioapic_id;
entry 203 dev/acpi/acpimadt.c aaa.apic_address = entry->madt_ioapic.address;
entry 204 dev/acpi/acpimadt.c aaa.apic_vecbase = entry->madt_ioapic.global_int_base;
entry 213 dev/acpi/acpimadt.c addr += entry->madt_lapic.length;
entry 223 dev/acpi/acpimadt.c union acpi_madt_entry *entry = (union acpi_madt_entry *)addr;
entry 225 dev/acpi/acpimadt.c switch (entry->madt_lapic.apic_type) {
entry 232 dev/acpi/acpimadt.c self->dv_xname, entry->madt_override.bus,
entry 233 dev/acpi/acpimadt.c entry->madt_override.source,
entry 234 dev/acpi/acpimadt.c entry->madt_override.global_int,
entry 235 dev/acpi/acpimadt.c entry->madt_override.flags);
entry 237 dev/acpi/acpimadt.c pin = entry->madt_override.global_int;
entry 247 dev/acpi/acpimadt.c map->bus_pin = entry->madt_override.source;
entry 248 dev/acpi/acpimadt.c map->flags = entry->madt_override.flags;
entry 250 dev/acpi/acpimadt.c map->global_int = entry->madt_override.global_int;
entry 252 dev/acpi/acpimadt.c acpimadt_cfg_intr(entry->madt_override.flags, &map->redir);
entry 266 dev/acpi/acpimadt.c self->dv_xname, entry->madt_lapic_nmi.acpi_proc_id,
entry 267 dev/acpi/acpimadt.c entry->madt_lapic_nmi.local_apic_lint,
entry 268 dev/acpi/acpimadt.c entry->madt_lapic_nmi.flags);
entry 270 dev/acpi/acpimadt.c pin = entry->madt_lapic_nmi.local_apic_lint;
entry 274 dev/acpi/acpimadt.c map->cpu_id = lapic_map[entry->madt_lapic_nmi.acpi_proc_id];
entry 276 dev/acpi/acpimadt.c map->flags = entry->madt_lapic_nmi.flags;
entry 278 dev/acpi/acpimadt.c acpimadt_cfg_intr(entry->madt_lapic_nmi.flags, &map->redir);
entry 285 dev/acpi/acpimadt.c self->dv_xname, entry->madt_lapic.apic_type);
entry 288 dev/acpi/acpimadt.c addr += entry->madt_lapic.length;
entry 249 dev/i2o/i2o.h struct i2o_hrt_entry entry[1];
entry 285 dev/i2o/i2o.h struct i2o_lct_entry entry[1];
entry 319 dev/i2o/i2o.h struct i2o_systab_entry entry[1];
entry 478 dev/i2o/iop.c for (i = 0, ste = iop_systab->entry; i < iop_cd.cd_ndevs; i++)
entry 637 dev/i2o/iop.c le = sc->sc_lct->entry;
entry 754 dev/i2o/iop.c for (i = 0, le = sc->sc_lct->entry; i < nent; i++, le++) {
entry 277 dev/i2o/iopsp.c for (le = iop->sc_lct->entry; nent != 0; nent--, le++)
entry 286 dev/i2o/iopsp.c for (i = 0, le = iop->sc_lct->entry; i < nent; i++, le++) {
entry 3522 dev/ic/aic79xx.c struct ahd_phase_table_entry *entry;
entry 3530 dev/ic/aic79xx.c for (entry = ahd_phase_table; entry < last_entry; entry++) {
entry 3531 dev/ic/aic79xx.c if (phase == entry->phase)
entry 3534 dev/ic/aic79xx.c return (entry);
entry 8807 dev/ic/aic79xx.c int entry, printed;
entry 8823 dev/ic/aic79xx.c for (entry = 0; entry < num_entries; entry++) {
entry 8824 dev/ic/aic79xx.c if (((value & table[entry].mask) != table[entry].value)
entry 8825 dev/ic/aic79xx.c || ((printed_mask & table[entry].mask) ==
entry 8826 dev/ic/aic79xx.c table[entry].mask))
entry 8831 dev/ic/aic79xx.c table[entry].name);
entry 8832 dev/ic/aic79xx.c printed_mask |= table[entry].mask;
entry 8836 dev/ic/aic79xx.c if (entry >= num_entries)
entry 2275 dev/ic/aic7xxx.c struct ahc_phase_table_entry *entry;
entry 2283 dev/ic/aic7xxx.c for (entry = ahc_phase_table; entry < last_entry; entry++) {
entry 2284 dev/ic/aic7xxx.c if (phase == entry->phase)
entry 2287 dev/ic/aic7xxx.c return (entry);
entry 6444 dev/ic/aic7xxx.c int entry, printed;
entry 6460 dev/ic/aic7xxx.c for (entry = 0; entry < num_entries; entry++) {
entry 6461 dev/ic/aic7xxx.c if (((value & table[entry].mask) != table[entry].value)
entry 6462 dev/ic/aic7xxx.c || ((printed_mask & table[entry].mask) ==
entry 6463 dev/ic/aic7xxx.c table[entry].mask))
entry 6468 dev/ic/aic7xxx.c table[entry].name);
entry 6469 dev/ic/aic7xxx.c printed_mask |= table[entry].mask;
entry 6473 dev/ic/aic7xxx.c if (entry >= num_entries)
entry 1814 dev/ic/ar5210.c ar5k_ar5210_reset_key(struct ath_hal *hal, u_int16_t entry)
entry 1818 dev/ic/ar5210.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5210_KEYTABLE_SIZE);
entry 1821 dev/ic/ar5210.c AR5K_REG_WRITE(AR5K_AR5210_KEYTABLE_OFF(entry, i), 0);
entry 1827 dev/ic/ar5210.c ar5k_ar5210_is_key_valid(struct ath_hal *hal, u_int16_t entry)
entry 1829 dev/ic/ar5210.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5210_KEYTABLE_SIZE);
entry 1834 dev/ic/ar5210.c if (AR5K_REG_READ(AR5K_AR5210_KEYTABLE_MAC1(entry)) &
entry 1842 dev/ic/ar5210.c ar5k_ar5210_set_key(struct ath_hal *hal, u_int16_t entry,
entry 1848 dev/ic/ar5210.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5210_KEYTABLE_SIZE);
entry 1883 dev/ic/ar5210.c AR5K_REG_WRITE(AR5K_AR5210_KEYTABLE_OFF(entry, i), key_v[i]);
entry 1885 dev/ic/ar5210.c return (ar5k_ar5210_set_key_lladdr(hal, entry, mac));
entry 1889 dev/ic/ar5210.c ar5k_ar5210_set_key_lladdr(struct ath_hal *hal, u_int16_t entry,
entry 1898 dev/ic/ar5210.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5210_KEYTABLE_SIZE);
entry 1906 dev/ic/ar5210.c AR5K_REG_WRITE(AR5K_AR5210_KEYTABLE_MAC0(entry), low_id);
entry 1907 dev/ic/ar5210.c AR5K_REG_WRITE(AR5K_AR5210_KEYTABLE_MAC1(entry), high_id);
entry 1919 dev/ic/ar5211.c ar5k_ar5211_reset_key(struct ath_hal *hal, u_int16_t entry)
entry 1923 dev/ic/ar5211.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5211_KEYTABLE_SIZE);
entry 1926 dev/ic/ar5211.c AR5K_REG_WRITE(AR5K_AR5211_KEYTABLE_OFF(entry, i), 0);
entry 1932 dev/ic/ar5211.c ar5k_ar5211_is_key_valid(struct ath_hal *hal, u_int16_t entry)
entry 1934 dev/ic/ar5211.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5211_KEYTABLE_SIZE);
entry 1939 dev/ic/ar5211.c if (AR5K_REG_READ(AR5K_AR5211_KEYTABLE_MAC1(entry)) &
entry 1947 dev/ic/ar5211.c ar5k_ar5211_set_key(struct ath_hal *hal, u_int16_t entry,
entry 1953 dev/ic/ar5211.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5211_KEYTABLE_SIZE);
entry 1988 dev/ic/ar5211.c AR5K_REG_WRITE(AR5K_AR5211_KEYTABLE_OFF(entry, i), key_v[i]);
entry 1990 dev/ic/ar5211.c return (ar5k_ar5211_set_key_lladdr(hal, entry, mac));
entry 1994 dev/ic/ar5211.c ar5k_ar5211_set_key_lladdr(struct ath_hal *hal, u_int16_t entry,
entry 2003 dev/ic/ar5211.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5211_KEYTABLE_SIZE);
entry 2011 dev/ic/ar5211.c AR5K_REG_WRITE(AR5K_AR5211_KEYTABLE_MAC0(entry), low_id);
entry 2012 dev/ic/ar5211.c AR5K_REG_WRITE(AR5K_AR5211_KEYTABLE_MAC1(entry), high_id);
entry 2254 dev/ic/ar5212.c ar5k_ar5212_reset_key(struct ath_hal *hal, u_int16_t entry)
entry 2258 dev/ic/ar5212.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5212_KEYTABLE_SIZE);
entry 2261 dev/ic/ar5212.c AR5K_REG_WRITE(AR5K_AR5212_KEYTABLE_OFF(entry, i), 0);
entry 2264 dev/ic/ar5212.c AR5K_REG_WRITE(AR5K_AR5212_KEYTABLE_TYPE(entry),
entry 2271 dev/ic/ar5212.c ar5k_ar5212_is_key_valid(struct ath_hal *hal, u_int16_t entry)
entry 2273 dev/ic/ar5212.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5212_KEYTABLE_SIZE);
entry 2278 dev/ic/ar5212.c if (AR5K_REG_READ(AR5K_AR5212_KEYTABLE_MAC1(entry)) &
entry 2286 dev/ic/ar5212.c ar5k_ar5212_set_key(struct ath_hal *hal, u_int16_t entry,
entry 2292 dev/ic/ar5212.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5212_KEYTABLE_SIZE);
entry 2327 dev/ic/ar5212.c AR5K_REG_WRITE(AR5K_AR5212_KEYTABLE_OFF(entry, i), key_v[i]);
entry 2329 dev/ic/ar5212.c return (ar5k_ar5212_set_key_lladdr(hal, entry, mac));
entry 2333 dev/ic/ar5212.c ar5k_ar5212_set_key_lladdr(struct ath_hal *hal, u_int16_t entry,
entry 2342 dev/ic/ar5212.c AR5K_ASSERT_ENTRY(entry, AR5K_AR5212_KEYTABLE_SIZE);
entry 2350 dev/ic/ar5212.c AR5K_REG_WRITE(AR5K_AR5212_KEYTABLE_MAC0(entry), low_id);
entry 2351 dev/ic/ar5212.c AR5K_REG_WRITE(AR5K_AR5212_KEYTABLE_MAC1(entry), high_id);
entry 1302 dev/ic/ar5xxx.c u_int32_t mask, entry, last, data, shift, position;
entry 1316 dev/ic/ar5xxx.c entry = ((first - 1) / 8) + offset;
entry 1322 dev/ic/ar5xxx.c for (i = shift = 0, left = bits; left > 0; position = 0, entry++, i++) {
entry 1328 dev/ic/ar5xxx.c rf[entry] &= ~mask;
entry 1329 dev/ic/ar5xxx.c rf[entry] |= ((data << position) << (col * 8)) & mask;
entry 1332 dev/ic/ar5xxx.c data = (((rf[entry] & mask) >> (col * 8)) >>
entry 760 dev/ic/sti.c a.in.entry = i;
entry 471 dev/ic/stireg.h u_int32_t entry;
entry 799 dev/isa/ad1848.c ad1848_devmap_t *entry;
entry 804 dev/isa/ad1848.c if (!(entry = ad1848_mixer_find_dev(map, cnt, cp)))
entry 807 dev/isa/ad1848.c dev = entry->dev;
entry 809 dev/isa/ad1848.c switch (entry->kind) {
entry 873 dev/isa/ad1848.c ad1848_devmap_t *entry;
entry 878 dev/isa/ad1848.c if (!(entry = ad1848_mixer_find_dev(map, cnt, cp)))
entry 881 dev/isa/ad1848.c dev = entry->dev;
entry 883 dev/isa/ad1848.c switch (entry->kind) {
entry 659 dev/pci/ahc_pci.c const struct ahc_pci_identity *entry;
entry 679 dev/pci/ahc_pci.c entry = &ahc_pci_ident_table[i];
entry 680 dev/pci/ahc_pci.c if (entry->full_id == (full_id & entry->id_mask))
entry 681 dev/pci/ahc_pci.c return (entry);
entry 692 dev/pci/ahc_pci.c const struct ahc_pci_identity *entry;
entry 696 dev/pci/ahc_pci.c entry = ahc_find_pci_device(pa->pa_id, subid, pa->pa_function);
entry 697 dev/pci/ahc_pci.c return (entry != NULL && entry->setup != NULL) ? 1 : 0;
entry 706 dev/pci/ahc_pci.c const struct ahc_pci_identity *entry;
entry 753 dev/pci/ahc_pci.c entry = ahc_find_pci_device(pa->pa_id, subid, pa->pa_function);
entry 754 dev/pci/ahc_pci.c if (entry == NULL)
entry 773 dev/pci/ahc_pci.c error = entry->setup(ahc);
entry 298 dev/pci/ahd_pci.c const struct ahd_pci_identity *entry;
entry 314 dev/pci/ahd_pci.c entry = &ahd_pci_ident_table[i];
entry 315 dev/pci/ahd_pci.c if (entry->full_id == (full_id & entry->id_mask)) {
entry 316 dev/pci/ahd_pci.c return (entry);
entry 325 dev/pci/ahd_pci.c const struct ahd_pci_identity *entry;
entry 330 dev/pci/ahd_pci.c entry = ahd_find_pci_device(pa->pa_id, subid);
entry 331 dev/pci/ahd_pci.c return entry != NULL ? 1 : 0;
entry 337 dev/pci/ahd_pci.c const struct ahd_pci_identity *entry;
entry 353 dev/pci/ahd_pci.c entry = ahd_find_pci_device(pa->pa_id, subid);
entry 354 dev/pci/ahd_pci.c if (entry == NULL)
entry 371 dev/pci/ahd_pci.c error = entry->setup(ahd, pa);
entry 624 dev/pci/if_bge.c struct bge_jpool_entry *entry;
entry 675 dev/pci/if_bge.c entry = malloc(sizeof(struct bge_jpool_entry),
entry 677 dev/pci/if_bge.c if (entry == NULL) {
entry 683 dev/pci/if_bge.c entry->slot = i;
entry 685 dev/pci/if_bge.c entry, jpool_entries);
entry 715 dev/pci/if_bge.c struct bge_jpool_entry *entry;
entry 717 dev/pci/if_bge.c entry = SLIST_FIRST(&sc->bge_jfree_listhead);
entry 719 dev/pci/if_bge.c if (entry == NULL)
entry 723 dev/pci/if_bge.c SLIST_INSERT_HEAD(&sc->bge_jinuse_listhead, entry, jpool_entries);
entry 724 dev/pci/if_bge.c return (sc->bge_cdata.bge_jslots[entry->slot]);
entry 733 dev/pci/if_bge.c struct bge_jpool_entry *entry;
entry 751 dev/pci/if_bge.c entry = SLIST_FIRST(&sc->bge_jinuse_listhead);
entry 752 dev/pci/if_bge.c if (entry == NULL)
entry 754 dev/pci/if_bge.c entry->slot = i;
entry 756 dev/pci/if_bge.c SLIST_INSERT_HEAD(&sc->bge_jfree_listhead, entry, jpool_entries);
entry 753 dev/pci/if_lge.c struct lge_jpool_entry *entry;
entry 805 dev/pci/if_lge.c entry = malloc(sizeof(struct lge_jpool_entry),
entry 807 dev/pci/if_lge.c if (entry == NULL) {
entry 814 dev/pci/if_lge.c entry->slot = i;
entry 816 dev/pci/if_lge.c entry, jpool_entries);
entry 844 dev/pci/if_lge.c struct lge_jpool_entry *entry;
entry 846 dev/pci/if_lge.c entry = LIST_FIRST(&sc->lge_jfree_listhead);
entry 848 dev/pci/if_lge.c if (entry == NULL)
entry 851 dev/pci/if_lge.c LIST_REMOVE(entry, jpool_entries);
entry 852 dev/pci/if_lge.c LIST_INSERT_HEAD(&sc->lge_jinuse_listhead, entry, jpool_entries);
entry 853 dev/pci/if_lge.c return (sc->lge_cdata.lge_jslots[entry->slot]);
entry 864 dev/pci/if_lge.c struct lge_jpool_entry *entry;
entry 878 dev/pci/if_lge.c entry = LIST_FIRST(&sc->lge_jinuse_listhead);
entry 879 dev/pci/if_lge.c if (entry == NULL)
entry 881 dev/pci/if_lge.c entry->slot = i;
entry 882 dev/pci/if_lge.c LIST_REMOVE(entry, jpool_entries);
entry 883 dev/pci/if_lge.c LIST_INSERT_HEAD(&sc->lge_jfree_listhead, entry, jpool_entries);
entry 467 dev/pci/if_msk.c struct sk_txmap_entry *entry;
entry 486 dev/pci/if_msk.c entry = malloc(sizeof(*entry), M_DEVBUF, M_NOWAIT);
entry 487 dev/pci/if_msk.c if (!entry) {
entry 491 dev/pci/if_msk.c entry->dmamap = dmamap;
entry 492 dev/pci/if_msk.c SIMPLEQ_INSERT_HEAD(&sc_if->sk_txmap_head, entry, link);
entry 570 dev/pci/if_msk.c struct sk_jpool_entry *entry;
entry 620 dev/pci/if_msk.c entry = malloc(sizeof(struct sk_jpool_entry),
entry 622 dev/pci/if_msk.c if (entry == NULL) {
entry 628 dev/pci/if_msk.c entry->slot = i;
entry 630 dev/pci/if_msk.c entry, jpool_entries);
entry 660 dev/pci/if_msk.c struct sk_jpool_entry *entry;
entry 662 dev/pci/if_msk.c entry = LIST_FIRST(&sc_if->sk_jfree_listhead);
entry 664 dev/pci/if_msk.c if (entry == NULL)
entry 667 dev/pci/if_msk.c LIST_REMOVE(entry, jpool_entries);
entry 668 dev/pci/if_msk.c LIST_INSERT_HEAD(&sc_if->sk_jinuse_listhead, entry, jpool_entries);
entry 669 dev/pci/if_msk.c return (sc_if->sk_cdata.sk_jslots[entry->slot]);
entry 678 dev/pci/if_msk.c struct sk_jpool_entry *entry;
entry 695 dev/pci/if_msk.c entry = LIST_FIRST(&sc->sk_jinuse_listhead);
entry 696 dev/pci/if_msk.c if (entry == NULL)
entry 698 dev/pci/if_msk.c entry->slot = i;
entry 699 dev/pci/if_msk.c LIST_REMOVE(entry, jpool_entries);
entry 700 dev/pci/if_msk.c LIST_INSERT_HEAD(&sc->sk_jfree_listhead, entry, jpool_entries);
entry 1409 dev/pci/if_msk.c struct sk_txmap_entry *entry;
entry 1414 dev/pci/if_msk.c entry = SIMPLEQ_FIRST(&sc_if->sk_txmap_head);
entry 1415 dev/pci/if_msk.c if (entry == NULL) {
entry 1419 dev/pci/if_msk.c txmap = entry->dmamap;
entry 1467 dev/pci/if_msk.c sc_if->sk_cdata.sk_tx_map[cur] = entry;
entry 1679 dev/pci/if_msk.c struct sk_txmap_entry *entry;
entry 1706 dev/pci/if_msk.c entry = sc_if->sk_cdata.sk_tx_map[idx];
entry 1711 dev/pci/if_msk.c bus_dmamap_sync(sc->sc_dmatag, entry->dmamap, 0,
entry 1712 dev/pci/if_msk.c entry->dmamap->dm_mapsize, BUS_DMASYNC_POSTWRITE);
entry 1714 dev/pci/if_msk.c bus_dmamap_unload(sc->sc_dmatag, entry->dmamap);
entry 1715 dev/pci/if_msk.c SIMPLEQ_INSERT_TAIL(&sc_if->sk_txmap_head, entry,
entry 1112 dev/pci/if_nge.c struct nge_jpool_entry *entry;
entry 1169 dev/pci/if_nge.c entry = malloc(sizeof(struct nge_jpool_entry),
entry 1171 dev/pci/if_nge.c if (entry == NULL) {
entry 1178 dev/pci/if_nge.c entry->slot = i;
entry 1179 dev/pci/if_nge.c LIST_INSERT_HEAD(&sc->nge_jfree_listhead, entry,
entry 1209 dev/pci/if_nge.c struct nge_jpool_entry *entry;
entry 1211 dev/pci/if_nge.c entry = LIST_FIRST(&sc->nge_jfree_listhead);
entry 1213 dev/pci/if_nge.c if (entry == NULL)
entry 1216 dev/pci/if_nge.c LIST_REMOVE(entry, jpool_entries);
entry 1217 dev/pci/if_nge.c LIST_INSERT_HEAD(&sc->nge_jinuse_listhead, entry, jpool_entries);
entry 1218 dev/pci/if_nge.c sc->nge_cdata.nge_jslots[entry->slot].nge_inuse = 1;
entry 1219 dev/pci/if_nge.c return(sc->nge_cdata.nge_jslots[entry->slot].nge_buf);
entry 1233 dev/pci/if_nge.c struct nge_jpool_entry *entry;
entry 1253 dev/pci/if_nge.c entry = LIST_FIRST(&sc->nge_jinuse_listhead);
entry 1254 dev/pci/if_nge.c if (entry == NULL)
entry 1256 dev/pci/if_nge.c entry->slot = i;
entry 1257 dev/pci/if_nge.c LIST_REMOVE(entry, jpool_entries);
entry 1259 dev/pci/if_nge.c entry, jpool_entries);
entry 101 dev/pci/if_san_xilinx.c SIMPLEQ_ENTRY(xilinx_rx_buffer) entry;
entry 424 dev/pci/if_san_xilinx.c sc->rx_dma_buf, entry);
entry 588 dev/pci/if_san_xilinx.c SIMPLEQ_REMOVE_HEAD(&sc->wp_rx_free_list, entry);
entry 593 dev/pci/if_san_xilinx.c SIMPLEQ_REMOVE_HEAD(&sc->wp_rx_complete_list, entry);
entry 1689 dev/pci/if_san_xilinx.c SIMPLEQ_REMOVE_HEAD(&sc->wp_rx_free_list, entry);
entry 2105 dev/pci/if_san_xilinx.c SIMPLEQ_INSERT_TAIL(&sc->wp_rx_complete_list, buf, entry);
entry 2413 dev/pci/if_san_xilinx.c SIMPLEQ_REMOVE_HEAD(&sc->wp_rx_free_list, entry);
entry 2418 dev/pci/if_san_xilinx.c SIMPLEQ_REMOVE_HEAD(&sc->wp_rx_complete_list, entry);
entry 2470 dev/pci/if_san_xilinx.c SIMPLEQ_INSERT_TAIL(&sc->wp_rx_free_list, buf, entry);
entry 2493 dev/pci/if_san_xilinx.c SIMPLEQ_INSERT_TAIL(&sc->wp_rx_free_list, buf, entry);
entry 2531 dev/pci/if_san_xilinx.c SIMPLEQ_REMOVE_HEAD(&sc->wp_rx_complete_list, entry);
entry 596 dev/pci/if_sk.c struct sk_txmap_entry *entry;
entry 616 dev/pci/if_sk.c entry = malloc(sizeof(*entry), M_DEVBUF, M_NOWAIT);
entry 617 dev/pci/if_sk.c if (!entry) {
entry 621 dev/pci/if_sk.c entry->dmamap = dmamap;
entry 622 dev/pci/if_sk.c SIMPLEQ_INSERT_HEAD(&sc_if->sk_txmap_head, entry, link);
entry 698 dev/pci/if_sk.c struct sk_jpool_entry *entry;
entry 748 dev/pci/if_sk.c entry = malloc(sizeof(struct sk_jpool_entry),
entry 750 dev/pci/if_sk.c if (entry == NULL) {
entry 756 dev/pci/if_sk.c entry->slot = i;
entry 758 dev/pci/if_sk.c entry, jpool_entries);
entry 788 dev/pci/if_sk.c struct sk_jpool_entry *entry;
entry 790 dev/pci/if_sk.c entry = LIST_FIRST(&sc_if->sk_jfree_listhead);
entry 792 dev/pci/if_sk.c if (entry == NULL)
entry 795 dev/pci/if_sk.c LIST_REMOVE(entry, jpool_entries);
entry 796 dev/pci/if_sk.c LIST_INSERT_HEAD(&sc_if->sk_jinuse_listhead, entry, jpool_entries);
entry 797 dev/pci/if_sk.c return (sc_if->sk_cdata.sk_jslots[entry->slot]);
entry 806 dev/pci/if_sk.c struct sk_jpool_entry *entry;
entry 823 dev/pci/if_sk.c entry = LIST_FIRST(&sc->sk_jinuse_listhead);
entry 824 dev/pci/if_sk.c if (entry == NULL)
entry 826 dev/pci/if_sk.c entry->slot = i;
entry 827 dev/pci/if_sk.c LIST_REMOVE(entry, jpool_entries);
entry 828 dev/pci/if_sk.c LIST_INSERT_HEAD(&sc->sk_jfree_listhead, entry, jpool_entries);
entry 1523 dev/pci/if_sk.c struct sk_txmap_entry *entry;
entry 1528 dev/pci/if_sk.c entry = SIMPLEQ_FIRST(&sc_if->sk_txmap_head);
entry 1529 dev/pci/if_sk.c if (entry == NULL) {
entry 1533 dev/pci/if_sk.c txmap = entry->dmamap;
entry 1581 dev/pci/if_sk.c sc_if->sk_cdata.sk_tx_map[cur] = entry;
entry 1915 dev/pci/if_sk.c struct sk_txmap_entry *entry;
entry 1941 dev/pci/if_sk.c entry = sc_if->sk_cdata.sk_tx_map[idx];
entry 1946 dev/pci/if_sk.c bus_dmamap_sync(sc->sc_dmatag, entry->dmamap, 0,
entry 1947 dev/pci/if_sk.c entry->dmamap->dm_mapsize, BUS_DMASYNC_POSTWRITE);
entry 1949 dev/pci/if_sk.c bus_dmamap_unload(sc->sc_dmatag, entry->dmamap);
entry 1950 dev/pci/if_sk.c SIMPLEQ_INSERT_TAIL(&sc_if->sk_txmap_head, entry,
entry 590 dev/pci/if_ti.c struct ti_jpool_entry *entry;
entry 641 dev/pci/if_ti.c entry = malloc(sizeof(struct ti_jpool_entry),
entry 643 dev/pci/if_ti.c if (entry == NULL) {
entry 650 dev/pci/if_ti.c entry->slot = i;
entry 651 dev/pci/if_ti.c SLIST_INSERT_HEAD(&sc->ti_jfree_listhead, entry, jpool_entries);
entry 681 dev/pci/if_ti.c struct ti_jpool_entry *entry;
entry 683 dev/pci/if_ti.c entry = SLIST_FIRST(&sc->ti_jfree_listhead);
entry 685 dev/pci/if_ti.c if (entry == NULL)
entry 689 dev/pci/if_ti.c SLIST_INSERT_HEAD(&sc->ti_jinuse_listhead, entry, jpool_entries);
entry 690 dev/pci/if_ti.c sc->ti_cdata.ti_jslots[entry->slot].ti_inuse = 1;
entry 691 dev/pci/if_ti.c return (sc->ti_cdata.ti_jslots[entry->slot].ti_buf);
entry 702 dev/pci/if_ti.c struct ti_jpool_entry *entry;
entry 720 dev/pci/if_ti.c entry = SLIST_FIRST(&sc->ti_jinuse_listhead);
entry 721 dev/pci/if_ti.c if (entry == NULL)
entry 723 dev/pci/if_ti.c entry->slot = i;
entry 726 dev/pci/if_ti.c entry, jpool_entries);
entry 1021 dev/pci/if_ti.c struct ti_txmap_entry *entry;
entry 1038 dev/pci/if_ti.c while ((entry = SLIST_FIRST(&sc->ti_tx_map_listhead))) {
entry 1040 dev/pci/if_ti.c bus_dmamap_destroy(sc->sc_dmatag, entry->dmamap);
entry 1041 dev/pci/if_ti.c free(entry, M_DEVBUF);
entry 1050 dev/pci/if_ti.c struct ti_txmap_entry *entry;
entry 1063 dev/pci/if_ti.c entry = malloc(sizeof(*entry), M_DEVBUF, M_NOWAIT);
entry 1064 dev/pci/if_ti.c if (!entry) {
entry 1068 dev/pci/if_ti.c entry->dmamap = dmamap;
entry 1069 dev/pci/if_ti.c SLIST_INSERT_HEAD(&sc->ti_tx_map_listhead, entry, link);
entry 1843 dev/pci/if_ti.c struct ti_txmap_entry *entry;
entry 1867 dev/pci/if_ti.c entry = sc->ti_cdata.ti_tx_map[idx];
entry 1868 dev/pci/if_ti.c bus_dmamap_sync(sc->sc_dmatag, entry->dmamap, 0,
entry 1869 dev/pci/if_ti.c entry->dmamap->dm_mapsize, BUS_DMASYNC_POSTWRITE);
entry 1871 dev/pci/if_ti.c bus_dmamap_unload(sc->sc_dmatag, entry->dmamap);
entry 1872 dev/pci/if_ti.c SLIST_INSERT_HEAD(&sc->ti_tx_map_listhead, entry,
entry 1893 dev/pci/if_ti.c struct ti_txmap_entry *entry;
entry 1913 dev/pci/if_ti.c entry = sc->ti_cdata.ti_tx_map[idx];
entry 1914 dev/pci/if_ti.c bus_dmamap_sync(sc->sc_dmatag, entry->dmamap, 0,
entry 1915 dev/pci/if_ti.c entry->dmamap->dm_mapsize, BUS_DMASYNC_POSTWRITE);
entry 1917 dev/pci/if_ti.c bus_dmamap_unload(sc->sc_dmatag, entry->dmamap);
entry 1918 dev/pci/if_ti.c SLIST_INSERT_HEAD(&sc->ti_tx_map_listhead, entry,
entry 1998 dev/pci/if_ti.c struct ti_txmap_entry *entry;
entry 2010 dev/pci/if_ti.c entry = SLIST_FIRST(&sc->ti_tx_map_listhead);
entry 2011 dev/pci/if_ti.c if (entry == NULL)
entry 2013 dev/pci/if_ti.c txmap = entry->dmamap;
entry 2069 dev/pci/if_ti.c sc->ti_cdata.ti_tx_map[cur] = entry;
entry 2086 dev/pci/if_ti.c struct ti_txmap_entry *entry;
entry 2097 dev/pci/if_ti.c entry = SLIST_FIRST(&sc->ti_tx_map_listhead);
entry 2098 dev/pci/if_ti.c if (entry == NULL)
entry 2100 dev/pci/if_ti.c txmap = entry->dmamap;
entry 2153 dev/pci/if_ti.c sc->ti_cdata.ti_tx_map[cur] = entry;
entry 37 dev/sdmmc/sdmmc_io.c TAILQ_ENTRY(sdmmc_intr_handler) entry;
entry 634 dev/sdmmc/sdmmc_io.c TAILQ_INSERT_TAIL(&sc->sc_intrq, ih, entry);
entry 653 dev/sdmmc/sdmmc_io.c TAILQ_REMOVE(&sc->sc_intrq, ih, entry);
entry 689 dev/sdmmc/sdmmc_io.c TAILQ_FOREACH(ih, &sc->sc_intrq, entry) {
entry 107 kern/exec_ecoff.c epp->ep_entry = eap->entry;
entry 138 kern/exec_ecoff.c epp->ep_entry = eap->entry;
entry 178 kern/exec_ecoff.c epp->ep_entry = eap->entry;
entry 445 kern/kern_lkm.c curp->entry = (int (*)(struct lkm_table *, int, int))
entry 449 kern/kern_lkm.c printf("LKM: call entrypoint %x\n", curp->entry);
entry 453 kern/kern_lkm.c error = (*(curp->entry))(curp, LKM_E_LOAD, curp->ver);
entry 493 kern/kern_lkm.c if ((*(curp->entry))(curp, LKM_E_UNLOAD, curp->ver)) {
entry 510 kern/kern_lkm.c if ((error = (*curp->entry)(curp, LKM_E_STAT, curp->ver)))
entry 269 lib/libsa/loadfile.c marks[MARK_ENTRY] = LOADADDR(coff->a.entry);
entry 281 lib/libsa/loadfile.c u_long entry = x->a_entry;
entry 295 lib/libsa/loadfile.c minp = maxp = ALIGNENTRY(entry);
entry 453 lib/libsa/loadfile.c marks[MARK_ENTRY] = LOADADDR(entry);
entry 131 net/if_media.c struct ifmedia_entry *entry;
entry 144 net/if_media.c entry = malloc(sizeof(*entry), M_IFADDR, M_NOWAIT);
entry 145 net/if_media.c if (entry == NULL)
entry 148 net/if_media.c entry->ifm_media = mword;
entry 149 net/if_media.c entry->ifm_data = data;
entry 150 net/if_media.c entry->ifm_aux = aux;
entry 152 net/if_media.c TAILQ_INSERT_TAIL(&ifm->ifm_list, entry, ifm_list);
entry 294 net/pf.c RB_GENERATE(pf_src_tree, pf_src_node, entry, pf_src_compare);
entry 379 net/pf_if.c TAILQ_INSERT_TAIL(&dyn->pfid_kif->pfik_dynaddrs, dyn, entry);
entry 404 net/pf_if.c TAILQ_FOREACH(p, &kif->pfik_dynaddrs, entry)
entry 568 net/pf_if.c TAILQ_REMOVE(&aw->p.dyn->pfid_kif->pfik_dynaddrs, aw->p.dyn, entry);
entry 265 net/pf_osfp.c struct pf_osfp_entry *entry;
entry 276 net/pf_osfp.c SLIST_FOREACH(entry, list, fp_entry) {
entry 277 net/pf_osfp.c PF_OSFP_UNPACK(entry->fp_os, en_class, en_version, en_subtype);
entry 282 net/pf_osfp.c entry->fp_class_nm, entry->fp_version_nm,
entry 283 net/pf_osfp.c entry->fp_subtype_nm, os, entry->fp_os);
entry 307 net/pf_osfp.c struct pf_osfp_entry *entry;
entry 311 net/pf_osfp.c while ((entry = SLIST_FIRST(&fp->fp_oses))) {
entry 313 net/pf_osfp.c pool_put(&pf_osfp_entry_pl, entry);
entry 325 net/pf_osfp.c struct pf_osfp_entry *entry;
entry 363 net/pf_osfp.c SLIST_FOREACH(entry, &fp->fp_oses, fp_entry) {
entry 364 net/pf_osfp.c if (PF_OSFP_ENTRY_EQ(entry, &fpioc->fp_os))
entry 367 net/pf_osfp.c if ((entry = pool_get(&pf_osfp_entry_pl, PR_NOWAIT)) == NULL)
entry 382 net/pf_osfp.c if ((entry = pool_get(&pf_osfp_entry_pl, PR_NOWAIT)) == NULL) {
entry 388 net/pf_osfp.c memcpy(entry, &fpioc->fp_os, sizeof(*entry));
entry 391 net/pf_osfp.c entry->fp_class_nm[sizeof(entry->fp_class_nm)-1] = '\0';
entry 392 net/pf_osfp.c entry->fp_version_nm[sizeof(entry->fp_version_nm)-1] = '\0';
entry 393 net/pf_osfp.c entry->fp_subtype_nm[sizeof(entry->fp_subtype_nm)-1] = '\0';
entry 395 net/pf_osfp.c SLIST_INSERT_HEAD(&fp->fp_oses, entry, fp_entry);
entry 521 net/pf_osfp.c struct pf_osfp_entry *entry;
entry 528 net/pf_osfp.c SLIST_FOREACH(entry, &fp->fp_oses, fp_entry) {
entry 537 net/pf_osfp.c memcpy(&fpioc->fp_os, entry,
entry 172 net/pfvar.h TAILQ_ENTRY(pfi_dynaddr) entry;
entry 633 net/pfvar.h RB_ENTRY(pf_src_node) entry;
entry 1538 net/pfvar.h RB_PROTOTYPE(pf_src_tree, pf_src_node, entry, pf_src_compare);
entry 56 sys/exec_ecoff.h u_long entry;
entry 179 sys/lkm.h int (*entry)(struct lkm_table *, int, int); /* entry function */
entry 296 uvm/uvm_amap.c amap_extend(struct vm_map_entry *entry, vsize_t addsize)
entry 298 uvm/uvm_amap.c struct vm_amap *amap = entry->aref.ar_amap;
entry 299 uvm/uvm_amap.c int slotoff = entry->aref.ar_pageoff;
entry 309 uvm/uvm_amap.c UVMHIST_LOG(maphist, " (entry=%p, addsize=%lu)", entry, addsize, 0, 0);
entry 317 uvm/uvm_amap.c AMAP_B2SLOT(slotmapped, entry->end - entry->start); /* slots mapped */
entry 475 uvm/uvm_amap.c amap_share_protect(struct vm_map_entry *entry, vm_prot_t prot)
entry 477 uvm/uvm_amap.c struct vm_amap *amap = entry->aref.ar_amap;
entry 480 uvm/uvm_amap.c AMAP_B2SLOT(slots, (entry->end - entry->start));
entry 481 uvm/uvm_amap.c stop = entry->aref.ar_pageoff + slots;
entry 485 uvm/uvm_amap.c for (lcv = entry->aref.ar_pageoff ; lcv < stop ; lcv++) {
entry 498 uvm/uvm_amap.c if (slot < entry->aref.ar_pageoff || slot >= stop)
entry 581 uvm/uvm_amap.c amap_copy(struct vm_map *map, struct vm_map_entry *entry, int waitf,
entry 589 uvm/uvm_amap.c map, entry, waitf, 0);
entry 595 uvm/uvm_amap.c if (entry->aref.ar_amap == NULL) {
entry 603 uvm/uvm_amap.c if (canchunk && atop(entry->end - entry->start) >=
entry 611 uvm/uvm_amap.c entry->start, entry->end, startva, endva);
entry 612 uvm/uvm_amap.c UVM_MAP_CLIP_START(map, entry, startva);
entry 615 uvm/uvm_amap.c UVM_MAP_CLIP_END(map, entry, endva);
entry 619 uvm/uvm_amap.c entry->start, entry->end, 0, 0);
entry 620 uvm/uvm_amap.c entry->aref.ar_pageoff = 0;
entry 621 uvm/uvm_amap.c entry->aref.ar_amap = amap_alloc(entry->end - entry->start, 0,
entry 623 uvm/uvm_amap.c if (entry->aref.ar_amap != NULL)
entry 624 uvm/uvm_amap.c entry->etype &= ~UVM_ET_NEEDSCOPY;
entry 638 uvm/uvm_amap.c if (entry->aref.ar_amap->am_ref == 1) {
entry 639 uvm/uvm_amap.c entry->etype &= ~UVM_ET_NEEDSCOPY;
entry 650 uvm/uvm_amap.c entry->aref.ar_amap, entry->aref.ar_amap->am_ref, 0, 0);
entry 651 uvm/uvm_amap.c AMAP_B2SLOT(slots, entry->end - entry->start);
entry 657 uvm/uvm_amap.c srcamap = entry->aref.ar_amap;
entry 668 uvm/uvm_amap.c entry->etype &= ~UVM_ET_NEEDSCOPY;
entry 681 uvm/uvm_amap.c srcamap->am_anon[entry->aref.ar_pageoff + lcv];
entry 706 uvm/uvm_amap.c amap_pp_adjref(srcamap, entry->aref.ar_pageoff,
entry 707 uvm/uvm_amap.c (entry->end - entry->start) >> PAGE_SHIFT, -1);
entry 715 uvm/uvm_amap.c entry->aref.ar_pageoff = 0;
entry 716 uvm/uvm_amap.c entry->aref.ar_amap = amap;
entry 717 uvm/uvm_amap.c entry->etype &= ~UVM_ET_NEEDSCOPY;
entry 750 uvm/uvm_amap.c amap_cow_now(struct vm_map *map, struct vm_map_entry *entry)
entry 752 uvm/uvm_amap.c struct vm_amap *amap = entry->aref.ar_amap;
entry 382 uvm/uvm_device.c struct vm_map_entry *entry = ufi->entry;
entry 383 uvm/uvm_device.c struct uvm_object *uobj = entry->object.uvm_obj;
entry 400 uvm/uvm_device.c if (UVM_ET_ISCOPYONWRITE(entry)) {
entry 402 uvm/uvm_device.c entry->etype, 0,0,0);
entry 403 uvm/uvm_device.c uvmfault_unlockall(ufi, ufi->entry->aref.ar_amap, uobj, NULL);
entry 422 uvm/uvm_device.c curr_offset = entry->offset + (vaddr - entry->start);
entry 445 uvm/uvm_device.c mapprot = ufi->entry->protection;
entry 461 uvm/uvm_device.c uvmfault_unlockall(ufi, ufi->entry->aref.ar_amap,
entry 471 uvm/uvm_device.c uvmfault_unlockall(ufi, ufi->entry->aref.ar_amap, uobj, NULL);
entry 254 uvm/uvm_fault.c if (UVM_ET_ISNEEDSCOPY(ufi->entry))
entry 255 uvm/uvm_fault.c amap_copy(ufi->map, ufi->entry, M_NOWAIT, TRUE,
entry 262 uvm/uvm_fault.c if (UVM_ET_ISNEEDSCOPY(ufi->entry)) {
entry 520 uvm/uvm_fault.c amap_lookup(&ufi->entry->aref,
entry 521 uvm/uvm_fault.c ufi->orig_rvaddr - ufi->entry->start) != anon) {
entry 555 uvm/uvm_fault.c #define MASK(entry) (UVM_ET_ISCOPYONWRITE(entry) ? \
entry 624 uvm/uvm_fault.c if ((ufi.entry->protection & access_type) != access_type) {
entry 627 uvm/uvm_fault.c ufi.entry->protection, access_type, 0, 0);
entry 639 uvm/uvm_fault.c enter_prot = ufi.entry->protection;
entry 640 uvm/uvm_fault.c wired = VM_MAPENT_ISWIRED(ufi.entry) || (fault_type == VM_FAULT_WIRE);
entry 651 uvm/uvm_fault.c if (UVM_ET_ISNEEDSCOPY(ufi.entry)) {
entry 653 uvm/uvm_fault.c (ufi.entry->object.uvm_obj == NULL)) {
entry 677 uvm/uvm_fault.c amap = ufi.entry->aref.ar_amap; /* top layer */
entry 678 uvm/uvm_fault.c uobj = ufi.entry->object.uvm_obj; /* bottom layer */
entry 701 uvm/uvm_fault.c KASSERT(uvmadvice[ufi.entry->advice].advice ==
entry 702 uvm/uvm_fault.c ufi.entry->advice);
entry 703 uvm/uvm_fault.c nback = min(uvmadvice[ufi.entry->advice].nback,
entry 704 uvm/uvm_fault.c (ufi.orig_rvaddr - ufi.entry->start) >> PAGE_SHIFT);
entry 706 uvm/uvm_fault.c nforw = min(uvmadvice[ufi.entry->advice].nforw,
entry 707 uvm/uvm_fault.c ((ufi.entry->end - ufi.orig_rvaddr) >>
entry 731 uvm/uvm_fault.c UVMHIST_LOG(maphist, " entry=%p, amap=%p, obj=%p", ufi.entry,
entry 740 uvm/uvm_fault.c amap_lookups(&ufi.entry->aref, startva - ufi.entry->start,
entry 753 uvm/uvm_fault.c if (ufi.entry->advice == MADV_SEQUENTIAL && nback != 0) {
entry 763 uvm/uvm_fault.c uoff = (startva - ufi.entry->start) + ufi.entry->offset;
entry 841 uvm/uvm_fault.c (VM_MAPENT_ISWIRED(ufi.entry) ? PMAP_WIRED : 0));
entry 906 uvm/uvm_fault.c (void) uobj->pgops->pgo_get(uobj, ufi.entry->offset +
entry 907 uvm/uvm_fault.c (startva - ufi.entry->start),
entry 909 uvm/uvm_fault.c access_type & MASK(ufi.entry),
entry 910 uvm/uvm_fault.c ufi.entry->advice, PGO_LOCKED);
entry 971 uvm/uvm_fault.c enter_prot & MASK(ufi.entry),
entry 1202 uvm/uvm_fault.c amap_add(&ufi.entry->aref, ufi.orig_rvaddr - ufi.entry->start,
entry 1314 uvm/uvm_fault.c UVM_ET_ISCOPYONWRITE(ufi.entry);
entry 1341 uvm/uvm_fault.c uoff = (ufi.orig_rvaddr - ufi.entry->start) + ufi.entry->offset;
entry 1343 uvm/uvm_fault.c 0, access_type & MASK(ufi.entry), ufi.entry->advice,
entry 1388 uvm/uvm_fault.c amap_lookup(&ufi.entry->aref,
entry 1389 uvm/uvm_fault.c ufi.orig_rvaddr - ufi.entry->start))) {
entry 1466 uvm/uvm_fault.c if (UVM_ET_ISCOPYONWRITE(ufi.entry))
entry 1668 uvm/uvm_fault.c amap_add(&ufi.entry->aref, ufi.orig_rvaddr - ufi.entry->start,
entry 1827 uvm/uvm_fault.c vm_map_entry_t entry;
entry 1848 uvm/uvm_fault.c if (uvm_map_lookup_entry(map, start, &entry) == FALSE)
entry 1858 uvm/uvm_fault.c KASSERT(va >= entry->start);
entry 1859 uvm/uvm_fault.c while (va >= entry->end) {
entry 1860 uvm/uvm_fault.c KASSERT(entry->next != &map->header &&
entry 1861 uvm/uvm_fault.c entry->next->start <= entry->end);
entry 1862 uvm/uvm_fault.c entry = entry->next;
entry 1868 uvm/uvm_fault.c if (VM_MAPENT_ISWIRED(entry) == 0)
entry 65 uvm/uvm_fault.h vm_map_entry_t entry; /* map entry (from 'map') */
entry 145 uvm/uvm_fault_i.h &ufi->entry)) {
entry 153 uvm/uvm_fault_i.h if (ufi->entry->end - ufi->orig_rvaddr < ufi->size)
entry 154 uvm/uvm_fault_i.h ufi->size = ufi->entry->end - ufi->orig_rvaddr;
entry 160 uvm/uvm_fault_i.h if (UVM_ET_ISSUBMAP(ufi->entry)) {
entry 161 uvm/uvm_fault_i.h tmpmap = ufi->entry->object.sub_map;
entry 137 uvm/uvm_loan.c struct vm_aref *aref = &ufi->entry->aref;
entry 138 uvm/uvm_loan.c struct uvm_object *uobj = ufi->entry->object.uvm_obj;
entry 158 uvm/uvm_loan.c anon = amap_lookup(aref, curaddr - ufi->entry->start);
entry 167 uvm/uvm_loan.c } else if (UVM_ET_ISCOPYONWRITE(ufi->entry)) {
entry 353 uvm/uvm_loan.c result = uvmfault_anonget(ufi, ufi->entry->aref.ar_amap, anon);
entry 414 uvm/uvm_loan.c struct vm_amap *amap = ufi->entry->aref.ar_amap;
entry 415 uvm/uvm_loan.c struct uvm_object *uobj = ufi->entry->object.uvm_obj;
entry 430 uvm/uvm_loan.c result = uobj->pgops->pgo_get(uobj, va - ufi->entry->start,
entry 455 uvm/uvm_loan.c result = uobj->pgops->pgo_get(uobj, va - ufi->entry->start,
entry 485 uvm/uvm_loan.c (locked && amap && amap_lookup(&ufi->entry->aref,
entry 486 uvm/uvm_loan.c ufi->orig_rvaddr - ufi->entry->start))) {
entry 623 uvm/uvm_loan.c uvmfault_unlockall(ufi, ufi->entry->aref.ar_amap,
entry 624 uvm/uvm_loan.c ufi->entry->object.uvm_obj, NULL);
entry 628 uvm/uvm_loan.c if (ufi->entry->object.uvm_obj)
entry 630 uvm/uvm_loan.c &ufi->entry->object.uvm_obj->vmobjlock);
entry 652 uvm/uvm_loan.c uvmfault_unlockall(ufi, ufi->entry->aref.ar_amap,
entry 653 uvm/uvm_loan.c ufi->entry->object.uvm_obj, NULL);
entry 667 uvm/uvm_loan.c if (ufi->entry->object.uvm_obj)
entry 668 uvm/uvm_loan.c simple_lock(&ufi->entry->object.uvm_obj->vmobjlock);
entry 143 uvm/uvm_map.c #define uvm_map_entry_link(map, after_where, entry) do { \
entry 145 uvm/uvm_map.c (entry)->prev = (after_where); \
entry 146 uvm/uvm_map.c (entry)->next = (after_where)->next; \
entry 147 uvm/uvm_map.c (entry)->prev->next = (entry); \
entry 148 uvm/uvm_map.c (entry)->next->prev = (entry); \
entry 149 uvm/uvm_map.c uvm_rb_insert(map, entry); \
entry 157 uvm/uvm_map.c #define uvm_map_entry_unlink(map, entry) do { \
entry 159 uvm/uvm_map.c (entry)->next->prev = (entry)->prev; \
entry 160 uvm/uvm_map.c (entry)->prev->next = (entry)->next; \
entry 161 uvm/uvm_map.c uvm_rb_remove(map, entry); \
entry 232 uvm/uvm_map.c uvm_rb_augment(struct vm_map_entry *entry)
entry 234 uvm/uvm_map.c entry->space = uvm_rb_subtree_space(entry);
entry 242 uvm/uvm_map.c uvm_rb_space(struct vm_map *map, struct vm_map_entry *entry)
entry 247 uvm/uvm_map.c if ((next = entry->next) == &map->header)
entry 248 uvm/uvm_map.c space = map->max_offset - entry->end;
entry 251 uvm/uvm_map.c space = next->start - entry->end;
entry 257 uvm/uvm_map.c uvm_rb_subtree_space(struct vm_map_entry *entry)
entry 261 uvm/uvm_map.c space = entry->ownspace;
entry 262 uvm/uvm_map.c if (RB_LEFT(entry, rb_entry)) {
entry 263 uvm/uvm_map.c tmp = RB_LEFT(entry, rb_entry)->space;
entry 268 uvm/uvm_map.c if (RB_RIGHT(entry, rb_entry)) {
entry 269 uvm/uvm_map.c tmp = RB_RIGHT(entry, rb_entry)->space;
entry 278 uvm/uvm_map.c uvm_rb_fixup(struct vm_map *map, struct vm_map_entry *entry)
entry 282 uvm/uvm_map.c entry->ownspace = uvm_rb_space(map, entry);
entry 283 uvm/uvm_map.c entry->space = uvm_rb_subtree_space(entry);
entry 284 uvm/uvm_map.c } while ((entry = RB_PARENT(entry, rb_entry)) != NULL);
entry 288 uvm/uvm_map.c uvm_rb_insert(struct vm_map *map, struct vm_map_entry *entry)
entry 290 uvm/uvm_map.c vaddr_t space = uvm_rb_space(map, entry);
entry 293 uvm/uvm_map.c entry->ownspace = entry->space = space;
entry 294 uvm/uvm_map.c tmp = RB_INSERT(uvm_tree, &(map)->rbhead, entry);
entry 299 uvm/uvm_map.c uvm_rb_fixup(map, entry);
entry 300 uvm/uvm_map.c if (entry->prev != &map->header)
entry 301 uvm/uvm_map.c uvm_rb_fixup(map, entry->prev);
entry 305 uvm/uvm_map.c uvm_rb_remove(struct vm_map *map, struct vm_map_entry *entry)
entry 309 uvm/uvm_map.c parent = RB_PARENT(entry, rb_entry);
entry 310 uvm/uvm_map.c RB_REMOVE(uvm_tree, &(map)->rbhead, entry);
entry 311 uvm/uvm_map.c if (entry->prev != &map->header)
entry 312 uvm/uvm_map.c uvm_rb_fixup(map, entry->prev);
entry 482 uvm/uvm_map.c uvm_map_entry_unwire(struct vm_map *map, struct vm_map_entry *entry)
entry 485 uvm/uvm_map.c entry->wired_count = 0;
entry 486 uvm/uvm_map.c uvm_fault_unwire_locked(map, entry->start, entry->end);
entry 494 uvm/uvm_map.c uvm_map_reference_amap(struct vm_map_entry *entry, int flags)
entry 496 uvm/uvm_map.c amap_ref(entry->aref.ar_amap, entry->aref.ar_pageoff,
entry 497 uvm/uvm_map.c (entry->end - entry->start) >> PAGE_SHIFT, flags);
entry 505 uvm/uvm_map.c uvm_map_unreference_amap(struct vm_map_entry *entry, int flags)
entry 507 uvm/uvm_map.c amap_unref(entry->aref.ar_amap, entry->aref.ar_pageoff,
entry 508 uvm/uvm_map.c (entry->end - entry->start) >> PAGE_SHIFT, flags);
entry 581 uvm/uvm_map.c uvm_map_clip_start(struct vm_map *map, struct vm_map_entry *entry,
entry 598 uvm/uvm_map.c uvm_mapent_copy(entry, new_entry); /* entry -> new_entry */
entry 602 uvm/uvm_map.c if (entry->object.uvm_obj)
entry 603 uvm/uvm_map.c entry->offset += new_adj; /* shift start over */
entry 606 uvm/uvm_map.c entry->start = start;
entry 609 uvm/uvm_map.c amap_splitref(&new_entry->aref, &entry->aref, new_adj);
entry 612 uvm/uvm_map.c uvm_map_entry_link(map, entry->prev, new_entry);
entry 614 uvm/uvm_map.c if (UVM_ET_ISSUBMAP(entry)) {
entry 618 uvm/uvm_map.c if (UVM_ET_ISOBJ(entry) &&
entry 619 uvm/uvm_map.c entry->object.uvm_obj->pgops &&
entry 620 uvm/uvm_map.c entry->object.uvm_obj->pgops->pgo_reference)
entry 621 uvm/uvm_map.c entry->object.uvm_obj->pgops->pgo_reference(
entry 622 uvm/uvm_map.c entry->object.uvm_obj);
entry 638 uvm/uvm_map.c uvm_map_clip_end(struct vm_map *map, struct vm_map_entry *entry, vaddr_t end)
entry 650 uvm/uvm_map.c uvm_mapent_copy(entry, new_entry); /* entry -> new_entry */
entry 652 uvm/uvm_map.c new_entry->start = entry->end = end;
entry 653 uvm/uvm_map.c new_adj = end - entry->start;
entry 657 uvm/uvm_map.c if (entry->aref.ar_amap)
entry 658 uvm/uvm_map.c amap_splitref(&entry->aref, &new_entry->aref, new_adj);
entry 660 uvm/uvm_map.c uvm_rb_fixup(map, entry);
entry 662 uvm/uvm_map.c uvm_map_entry_link(map, entry, new_entry);
entry 664 uvm/uvm_map.c if (UVM_ET_ISSUBMAP(entry)) {
entry 668 uvm/uvm_map.c if (UVM_ET_ISOBJ(entry) &&
entry 669 uvm/uvm_map.c entry->object.uvm_obj->pgops &&
entry 670 uvm/uvm_map.c entry->object.uvm_obj->pgops->pgo_reference)
entry 671 uvm/uvm_map.c entry->object.uvm_obj->pgops->pgo_reference(
entry 672 uvm/uvm_map.c entry->object.uvm_obj);
entry 953 uvm/uvm_map.c struct vm_map_entry **entry)
entry 962 uvm/uvm_map.c map, address, entry, 0);
entry 992 uvm/uvm_map.c *entry = cur;
entry 1022 uvm/uvm_map.c *entry = cur;
entry 1031 uvm/uvm_map.c *entry = prev;
entry 1048 uvm/uvm_map.c *entry = cur;
entry 1059 uvm/uvm_map.c *entry = cur->prev;
entry 1060 uvm/uvm_map.c SAVE_HINT(map, map->hint, *entry);
entry 1149 uvm/uvm_map.c struct vm_map_entry *entry, *next, *tmp;
entry 1189 uvm/uvm_map.c if ((entry = map->first_free) != &map->header)
entry 1190 uvm/uvm_map.c hint = entry->end;
entry 1201 uvm/uvm_map.c entry = tmp;
entry 1210 uvm/uvm_map.c next = entry->next;
entry 1220 uvm/uvm_map.c if (uvm_map_spacefits(map, &hint, length, entry->next, uoffset, align))
entry 1261 uvm/uvm_map.c entry = tmp;
entry 1272 uvm/uvm_map.c entry = prev;
entry 1292 uvm/uvm_map.c entry = tmp;
entry 1307 uvm/uvm_map.c for (;; hint = (entry = next)->end) {
entry 1336 uvm/uvm_map.c next = entry->next;
entry 1341 uvm/uvm_map.c SAVE_HINT(map, map->hint, entry);
entry 1344 uvm/uvm_map.c return (entry);
entry 1374 uvm/uvm_map.c struct vm_map_entry *entry, *first_entry, *next;
entry 1394 uvm/uvm_map.c entry = first_entry;
entry 1395 uvm/uvm_map.c UVM_MAP_CLIP_START(map, entry, start);
entry 1397 uvm/uvm_map.c SAVE_HINT(map, entry, entry->prev);
entry 1400 uvm/uvm_map.c entry = first_entry->next;
entry 1408 uvm/uvm_map.c map->first_free = entry->prev;
entry 1434 uvm/uvm_map.c while ((entry != &map->header) && (entry->start < end)) {
entry 1436 uvm/uvm_map.c UVM_MAP_CLIP_END(map, entry, end);
entry 1437 uvm/uvm_map.c next = entry->next;
entry 1438 uvm/uvm_map.c len = entry->end - entry->start;
entry 1439 uvm/uvm_map.c if (p && entry->object.uvm_obj == NULL)
entry 1447 uvm/uvm_map.c if (VM_MAPENT_ISWIRED(entry))
entry 1448 uvm/uvm_map.c uvm_map_entry_unwire(map, entry);
entry 1455 uvm/uvm_map.c uvm_km_pgremove_intrsafe(entry->start, entry->end);
entry 1456 uvm/uvm_map.c pmap_kremove(entry->start, len);
entry 1457 uvm/uvm_map.c } else if (UVM_ET_ISOBJ(entry) &&
entry 1458 uvm/uvm_map.c UVM_OBJ_IS_KERN_OBJECT(entry->object.uvm_obj)) {
entry 1496 uvm/uvm_map.c pmap_remove(pmap_kernel(), entry->start, entry->end);
entry 1497 uvm/uvm_map.c uvm_km_pgremove(entry->object.uvm_obj,
entry 1498 uvm/uvm_map.c entry->start - vm_map_min(kernel_map),
entry 1499 uvm/uvm_map.c entry->end - vm_map_min(kernel_map));
entry 1505 uvm/uvm_map.c entry->etype &= ~UVM_ET_OBJ;
entry 1506 uvm/uvm_map.c entry->object.uvm_obj = NULL; /* to be safe */
entry 1512 uvm/uvm_map.c pmap_remove(map->pmap, entry->start, entry->end);
entry 1519 uvm/uvm_map.c UVMHIST_LOG(maphist, " removed map entry %p", entry, 0, 0,0);
entry 1522 uvm/uvm_map.c SAVE_HINT(map, entry, entry->prev);
entry 1524 uvm/uvm_map.c uvm_map_entry_unlink(map, entry);
entry 1526 uvm/uvm_map.c entry->next = first_entry;
entry 1527 uvm/uvm_map.c first_entry = entry;
entry 1528 uvm/uvm_map.c entry = next; /* next entry, please */
entry 1785 uvm/uvm_map.c struct vm_map_entry *chain, *endchain, *entry, *orig_entry, *newentry;
entry 1830 uvm/uvm_map.c if (uvm_map_lookup_entry(srcmap, start, &entry)) {
entry 1843 uvm/uvm_map.c fudge = start - entry->start;
entry 1851 uvm/uvm_map.c UVM_MAP_CLIP_START(srcmap, entry, start);
entry 1852 uvm/uvm_map.c SAVE_HINT(srcmap, srcmap->hint, entry->prev);
entry 1863 uvm/uvm_map.c entry = entry->next;
entry 1868 uvm/uvm_map.c orig_entry = entry;
entry 1876 uvm/uvm_map.c while (entry->start < end && entry != &srcmap->header) {
entry 1880 uvm/uvm_map.c UVM_MAP_CLIP_END(srcmap, entry, end);
entry 1883 uvm/uvm_map.c if (UVM_ET_ISNEEDSCOPY(entry)) {
entry 1885 uvm/uvm_map.c oldstart = entry->start;
entry 1888 uvm/uvm_map.c amap_copy(srcmap, entry, M_NOWAIT, TRUE, start, end);
entry 1889 uvm/uvm_map.c if (UVM_ET_ISNEEDSCOPY(entry)) { /* failed? */
entry 1896 uvm/uvm_map.c fudge = fudge - (entry->start - oldstart);
entry 1902 uvm/uvm_map.c oldoffset = (entry->start + fudge) - start;
entry 1916 uvm/uvm_map.c newentry->start + (entry->end - (entry->start + fudge));
entry 1919 uvm/uvm_map.c newentry->object.uvm_obj = entry->object.uvm_obj;
entry 1924 uvm/uvm_map.c newentry->offset = entry->offset + fudge;
entry 1928 uvm/uvm_map.c newentry->etype = entry->etype;
entry 1930 uvm/uvm_map.c entry->max_protection : entry->protection;
entry 1931 uvm/uvm_map.c newentry->max_protection = entry->max_protection;
entry 1932 uvm/uvm_map.c newentry->inheritance = entry->inheritance;
entry 1934 uvm/uvm_map.c newentry->aref.ar_amap = entry->aref.ar_amap;
entry 1937 uvm/uvm_map.c entry->aref.ar_pageoff + (fudge >> PAGE_SHIFT);
entry 1943 uvm/uvm_map.c newentry->advice = entry->advice;
entry 1955 uvm/uvm_map.c if ((flags & UVM_EXTRACT_CONTIG) && entry->end < end &&
entry 1956 uvm/uvm_map.c (entry->next == &srcmap->header ||
entry 1957 uvm/uvm_map.c entry->next->start != entry->end)) {
entry 1961 uvm/uvm_map.c entry = entry->next;
entry 2009 uvm/uvm_map.c entry = orig_entry;
entry 2013 uvm/uvm_map.c while (entry->start < end && entry != &srcmap->header) {
entry 2015 uvm/uvm_map.c oldoffset = (entry->start + fudge) - start;
entry 2016 uvm/uvm_map.c elen = MIN(end, entry->end) -
entry 2017 uvm/uvm_map.c (entry->start + fudge);
entry 2020 uvm/uvm_map.c entry->start + fudge);
entry 2025 uvm/uvm_map.c pmap_remove(srcmap->pmap, entry->start,
entry 2026 uvm/uvm_map.c entry->end);
entry 2027 uvm/uvm_map.c oldentry = entry; /* save entry */
entry 2028 uvm/uvm_map.c entry = entry->next; /* advance */
entry 2034 uvm/uvm_map.c entry = entry->next; /* advance */
entry 2122 uvm/uvm_map.c struct vm_map_entry *entry;
entry 2129 uvm/uvm_map.c if (uvm_map_lookup_entry(map, start, &entry)) {
entry 2130 uvm/uvm_map.c UVM_MAP_CLIP_START(map, entry, start);
entry 2131 uvm/uvm_map.c UVM_MAP_CLIP_END(map, entry, end); /* to be safe */
entry 2133 uvm/uvm_map.c entry = NULL;
entry 2136 uvm/uvm_map.c if (entry != NULL &&
entry 2137 uvm/uvm_map.c entry->start == start && entry->end == end &&
entry 2138 uvm/uvm_map.c entry->object.uvm_obj == NULL && entry->aref.ar_amap == NULL &&
entry 2139 uvm/uvm_map.c !UVM_ET_ISCOPYONWRITE(entry) && !UVM_ET_ISNEEDSCOPY(entry)) {
entry 2140 uvm/uvm_map.c entry->etype |= UVM_ET_SUBMAP;
entry 2141 uvm/uvm_map.c entry->object.sub_map = submap;
entry 2142 uvm/uvm_map.c entry->offset = 0;
entry 2160 uvm/uvm_map.c #define MASK(entry) (UVM_ET_ISCOPYONWRITE(entry) ? \
entry 2168 uvm/uvm_map.c struct vm_map_entry *current, *entry;
entry 2178 uvm/uvm_map.c if (uvm_map_lookup_entry(map, start, &entry)) {
entry 2179 uvm/uvm_map.c UVM_MAP_CLIP_START(map, entry, start);
entry 2181 uvm/uvm_map.c entry = entry->next;
entry 2188 uvm/uvm_map.c current = entry;
entry 2203 uvm/uvm_map.c current = entry;
entry 2224 uvm/uvm_map.c if ((current->protection & MASK(entry)) == PROT_NONE &&
entry 2225 uvm/uvm_map.c VM_MAPENT_ISWIRED(entry))
entry 2228 uvm/uvm_map.c current->protection & MASK(entry));
entry 2238 uvm/uvm_map.c VM_MAPENT_ISWIRED(entry) == 0 &&
entry 2241 uvm/uvm_map.c if (uvm_map_pageable(map, entry->start, entry->end,
entry 2285 uvm/uvm_map.c struct vm_map_entry *entry, *temp_entry;
entry 2305 uvm/uvm_map.c entry = temp_entry;
entry 2306 uvm/uvm_map.c UVM_MAP_CLIP_START(map, entry, start);
entry 2308 uvm/uvm_map.c entry = temp_entry->next;
entry 2311 uvm/uvm_map.c while ((entry != &map->header) && (entry->start < end)) {
entry 2312 uvm/uvm_map.c UVM_MAP_CLIP_END(map, entry, end);
entry 2313 uvm/uvm_map.c entry->inheritance = new_inheritance;
entry 2314 uvm/uvm_map.c entry = entry->next;
entry 2331 uvm/uvm_map.c struct vm_map_entry *entry, *temp_entry;
entry 2339 uvm/uvm_map.c entry = temp_entry;
entry 2340 uvm/uvm_map.c UVM_MAP_CLIP_START(map, entry, start);
entry 2342 uvm/uvm_map.c entry = temp_entry->next;
entry 2349 uvm/uvm_map.c while ((entry != &map->header) && (entry->start < end)) {
entry 2350 uvm/uvm_map.c UVM_MAP_CLIP_END(map, entry, end);
entry 2364 uvm/uvm_map.c entry->advice = new_advice;
entry 2365 uvm/uvm_map.c entry = entry->next;
entry 2391 uvm/uvm_map.c struct vm_map_entry *entry, *start_entry, *failed_entry;
entry 2421 uvm/uvm_map.c entry = start_entry;
entry 2428 uvm/uvm_map.c UVM_MAP_CLIP_START(map, entry, start);
entry 2435 uvm/uvm_map.c while ((entry != &map->header) && (entry->start < end)) {
entry 2436 uvm/uvm_map.c if (entry->wired_count == 0 ||
entry 2437 uvm/uvm_map.c (entry->end < end &&
entry 2438 uvm/uvm_map.c (entry->next == &map->header ||
entry 2439 uvm/uvm_map.c entry->next->start > entry->end))) {
entry 2446 uvm/uvm_map.c entry = entry->next;
entry 2455 uvm/uvm_map.c entry = start_entry;
entry 2456 uvm/uvm_map.c while ((entry != &map->header) && (entry->start < end)) {
entry 2457 uvm/uvm_map.c UVM_MAP_CLIP_END(map, entry, end);
entry 2458 uvm/uvm_map.c if (VM_MAPENT_ISWIRED(entry))
entry 2459 uvm/uvm_map.c uvm_map_entry_unwire(map, entry);
entry 2460 uvm/uvm_map.c entry = entry->next;
entry 2489 uvm/uvm_map.c while ((entry != &map->header) && (entry->start < end)) {
entry 2490 uvm/uvm_map.c if (VM_MAPENT_ISWIRED(entry) == 0) { /* not already wired? */
entry 2500 uvm/uvm_map.c if (!UVM_ET_ISSUBMAP(entry)) { /* not submap */
entry 2501 uvm/uvm_map.c if (UVM_ET_ISNEEDSCOPY(entry) &&
entry 2502 uvm/uvm_map.c ((entry->protection & VM_PROT_WRITE) ||
entry 2503 uvm/uvm_map.c (entry->object.uvm_obj == NULL))) {
entry 2504 uvm/uvm_map.c amap_copy(map, entry, M_WAITOK, TRUE,
entry 2510 uvm/uvm_map.c UVM_MAP_CLIP_START(map, entry, start);
entry 2511 uvm/uvm_map.c UVM_MAP_CLIP_END(map, entry, end);
entry 2512 uvm/uvm_map.c entry->wired_count++;
entry 2518 uvm/uvm_map.c if (entry->protection == VM_PROT_NONE ||
entry 2519 uvm/uvm_map.c (entry->end < end &&
entry 2520 uvm/uvm_map.c (entry->next == &map->header ||
entry 2521 uvm/uvm_map.c entry->next->start > entry->end))) {
entry 2528 uvm/uvm_map.c while (entry != &map->header && entry->end > start) {
entry 2529 uvm/uvm_map.c entry->wired_count--;
entry 2530 uvm/uvm_map.c entry = entry->prev;
entry 2537 uvm/uvm_map.c entry = entry->next;
entry 2551 uvm/uvm_map.c entry = start_entry;
entry 2552 uvm/uvm_map.c while (entry != &map->header && entry->start < end) {
entry 2553 uvm/uvm_map.c if (entry->wired_count == 1) {
entry 2554 uvm/uvm_map.c rv = uvm_fault_wire(map, entry->start, entry->end,
entry 2555 uvm/uvm_map.c entry->protection);
entry 2565 uvm/uvm_map.c entry = entry->next;
entry 2587 uvm/uvm_map.c failed_entry = entry;
entry 2588 uvm/uvm_map.c while (entry != &map->header && entry->start < end) {
entry 2589 uvm/uvm_map.c entry->wired_count--;
entry 2590 uvm/uvm_map.c entry = entry->next;
entry 2598 uvm/uvm_map.c entry = start_entry;
entry 2599 uvm/uvm_map.c while (entry != failed_entry) {
entry 2600 uvm/uvm_map.c entry->wired_count--;
entry 2601 uvm/uvm_map.c if (VM_MAPENT_ISWIRED(entry) == 0)
entry 2602 uvm/uvm_map.c uvm_map_entry_unwire(map, entry);
entry 2603 uvm/uvm_map.c entry = entry->next;
entry 2641 uvm/uvm_map.c struct vm_map_entry *entry, *failed_entry;
entry 2663 uvm/uvm_map.c for (entry = map->header.next; entry != &map->header;
entry 2664 uvm/uvm_map.c entry = entry->next) {
entry 2665 uvm/uvm_map.c if (VM_MAPENT_ISWIRED(entry))
entry 2666 uvm/uvm_map.c uvm_map_entry_unwire(map, entry);
entry 2717 uvm/uvm_map.c for (size = 0, entry = map->header.next; entry != &map->header;
entry 2718 uvm/uvm_map.c entry = entry->next) {
entry 2719 uvm/uvm_map.c if (entry->protection != VM_PROT_NONE &&
entry 2720 uvm/uvm_map.c VM_MAPENT_ISWIRED(entry) == 0) { /* not already wired? */
entry 2721 uvm/uvm_map.c size += entry->end - entry->start;
entry 2743 uvm/uvm_map.c for (entry = map->header.next; entry != &map->header;
entry 2744 uvm/uvm_map.c entry = entry->next) {
entry 2745 uvm/uvm_map.c if (entry->protection == VM_PROT_NONE)
entry 2747 uvm/uvm_map.c if (VM_MAPENT_ISWIRED(entry) == 0) { /* not already wired? */
entry 2755 uvm/uvm_map.c if (!UVM_ET_ISSUBMAP(entry)) { /* not submap */
entry 2756 uvm/uvm_map.c if (UVM_ET_ISNEEDSCOPY(entry) &&
entry 2757 uvm/uvm_map.c ((entry->protection & VM_PROT_WRITE) ||
entry 2758 uvm/uvm_map.c (entry->object.uvm_obj == NULL))) {
entry 2759 uvm/uvm_map.c amap_copy(map, entry, M_WAITOK, TRUE,
entry 2760 uvm/uvm_map.c entry->start, entry->end);
entry 2765 uvm/uvm_map.c entry->wired_count++;
entry 2778 uvm/uvm_map.c for (error = 0, entry = map->header.next;
entry 2779 uvm/uvm_map.c entry != &map->header && error == 0;
entry 2780 uvm/uvm_map.c entry = entry->next) {
entry 2781 uvm/uvm_map.c if (entry->wired_count == 1) {
entry 2782 uvm/uvm_map.c error = uvm_fault_wire(map, entry->start, entry->end,
entry 2783 uvm/uvm_map.c entry->protection);
entry 2805 uvm/uvm_map.c failed_entry = entry;
entry 2806 uvm/uvm_map.c for (/* nothing */; entry != &map->header;
entry 2807 uvm/uvm_map.c entry = entry->next) {
entry 2808 uvm/uvm_map.c if (entry->protection == VM_PROT_NONE)
entry 2810 uvm/uvm_map.c entry->wired_count--;
entry 2819 uvm/uvm_map.c for (entry = map->header.next; entry != failed_entry;
entry 2820 uvm/uvm_map.c entry = entry->next) {
entry 2821 uvm/uvm_map.c if (entry->protection == VM_PROT_NONE)
entry 2823 uvm/uvm_map.c entry->wired_count--;
entry 2824 uvm/uvm_map.c if (VM_MAPENT_ISWIRED(entry))
entry 2825 uvm/uvm_map.c uvm_map_entry_unwire(map, entry);
entry 2861 uvm/uvm_map.c struct vm_map_entry *current, *entry;
entry 2878 uvm/uvm_map.c if (uvm_map_lookup_entry(map, start, &entry) == FALSE) {
entry 2887 uvm/uvm_map.c for (current = entry; current->start < end; current = current->next) {
entry 2901 uvm/uvm_map.c for (current = entry; current->start < end; current = current->next) {
entry 3027 uvm/uvm_map.c ((entry->max_protection & VM_PROT_WRITE) != 0 &&
entry 3028 uvm/uvm_map.c (entry->etype & UVM_ET_COPYONWRITE) == 0))) {
entry 3055 uvm/uvm_map.c struct vm_map_entry *entry;
entry 3061 uvm/uvm_map.c entry = tmp_entry;
entry 3063 uvm/uvm_map.c if (entry == &map->header) {
entry 3071 uvm/uvm_map.c if (start < entry->start) {
entry 3079 uvm/uvm_map.c if ((entry->protection & protection) != protection) {
entry 3085 uvm/uvm_map.c start = entry->end;
entry 3086 uvm/uvm_map.c entry = entry->next;
entry 3618 uvm/uvm_map.c struct vm_map_entry *entry;
entry 3633 uvm/uvm_map.c for (entry = map->header.next; entry != &map->header;
entry 3634 uvm/uvm_map.c entry = entry->next) {
entry 3636 uvm/uvm_map.c entry, entry->start, entry->end, entry->object.uvm_obj,
entry 3637 uvm/uvm_map.c (long long)entry->offset, entry->aref.ar_amap,
entry 3638 uvm/uvm_map.c entry->aref.ar_pageoff);
entry 3642 uvm/uvm_map.c (entry->etype & UVM_ET_SUBMAP) ? 'T' : 'F',
entry 3643 uvm/uvm_map.c (entry->etype & UVM_ET_COPYONWRITE) ? 'T' : 'F',
entry 3644 uvm/uvm_map.c (entry->etype & UVM_ET_NEEDSCOPY) ? 'T' : 'F',
entry 3645 uvm/uvm_map.c entry->protection, entry->max_protection,
entry 3646 uvm/uvm_map.c entry->inheritance, entry->wired_count, entry->advice);
entry 160 uvm/uvm_map.h #define VM_MAPENT_ISWIRED(entry) ((entry)->wired_count != 0)
entry 234 uvm/uvm_meter.c struct vm_map_entry * entry;
entry 280 uvm/uvm_meter.c for (map = &p->p_vmspace->vm_map, entry = map->header.next;
entry 281 uvm/uvm_meter.c entry != &map->header; entry = entry->next) {
entry 282 uvm/uvm_meter.c if (entry->is_a_map || entry->is_sub_map ||
entry 283 uvm/uvm_meter.c entry->object.uvm_obj == NULL)
entry 254 uvm/uvm_mmap.c vm_map_entry_t entry;
entry 284 uvm/uvm_mmap.c if (uvm_map_lookup_entry(map, start, &entry) == FALSE) {
entry 290 uvm/uvm_mmap.c entry != &map->header && entry->start < end;
entry 291 uvm/uvm_mmap.c entry = entry->next) {
entry 292 uvm/uvm_mmap.c KASSERT(!UVM_ET_ISSUBMAP(entry));
entry 293 uvm/uvm_mmap.c KASSERT(start >= entry->start);
entry 296 uvm/uvm_mmap.c if (entry->end < end &&
entry 297 uvm/uvm_mmap.c (entry->next == &map->header ||
entry 298 uvm/uvm_mmap.c entry->next->start > entry->end)) {
entry 303 uvm/uvm_mmap.c lim = end < entry->end ? end : entry->end;
entry 309 uvm/uvm_mmap.c if (UVM_ET_ISOBJ(entry)) {
entry 310 uvm/uvm_mmap.c KASSERT(!UVM_OBJ_IS_KERN_OBJECT(entry->object.uvm_obj));
entry 311 uvm/uvm_mmap.c if (entry->object.uvm_obj->pgops->pgo_releasepg
entry 321 uvm/uvm_mmap.c amap = entry->aref.ar_amap; /* top layer */
entry 322 uvm/uvm_mmap.c uobj = entry->object.uvm_obj; /* bottom layer */
entry 331 uvm/uvm_mmap.c anon = amap_lookup(&entry->aref,
entry 332 uvm/uvm_mmap.c start - entry->start);
entry 346 uvm/uvm_mmap.c entry->offset + (start - entry->start));
entry 678 uvm/uvm_mmap.c vm_map_entry_t entry;
entry 681 uvm/uvm_mmap.c rv = uvm_map_lookup_entry(map, addr, &entry);
entry 683 uvm/uvm_mmap.c addr = entry->start;
entry 684 uvm/uvm_mmap.c size = entry->end - entry->start;
entry 311 uvm/uvm_page.h #define VM_PAGE_TO_PHYS(entry) ((entry)->phys_addr)
entry 409 uvm/uvm_page.h #define VM_PAGE_IS_FREE(entry) ((entry)->pg_flags & PQ_FREE)
entry 188 uvm/uvm_unix.c vm_map_entry_t entry;
entry 196 uvm/uvm_unix.c for (entry = map->header.next; entry != &map->header;
entry 197 uvm/uvm_unix.c entry = entry->next) {
entry 200 uvm/uvm_unix.c if (UVM_ET_ISSUBMAP(entry)) {
entry 204 uvm/uvm_unix.c if (!(entry->protection & VM_PROT_WRITE))
entry 210 uvm/uvm_unix.c if (entry->object.uvm_obj != NULL &&
entry 211 uvm/uvm_unix.c UVM_OBJ_IS_DEVICE(entry->object.uvm_obj))
entry 214 uvm/uvm_unix.c start = entry->start;
entry 215 uvm/uvm_unix.c end = entry->end;