entry             153 arch/i386/i386/apm.c 	u_int32_t entry;
entry             829 arch/i386/i386/apm.c 		apm_ep.entry = ap->apm_entry;
entry             212 arch/i386/i386/bios.c 			if (h->entry <= BIOS32_START || h->entry >= BIOS32_END)
entry             216 arch/i386/i386/bios.c 			bios32_entry.offset = (u_int32_t)ISA_HOLE_VADDR(h->entry);
entry             217 arch/i386/i386/bios.c 			printf(", BIOS32 rev. %d @ 0x%lx", h->rev, h->entry);
entry             716 arch/i386/i386/mpbios.c 	const struct mpbios_proc *entry = (const struct mpbios_proc *)ent;
entry             721 arch/i386/i386/mpbios.c 	if (!(entry->cpu_flags & PROCENTRY_FLAG_EN))
entry             725 arch/i386/i386/mpbios.c 	if (entry->cpu_flags & PROCENTRY_FLAG_BP)
entry             731 arch/i386/i386/mpbios.c 	caa.cpu_number = entry->apic_id;
entry             734 arch/i386/i386/mpbios.c 	caa.cpu_signature = entry->cpu_signature;
entry             740 arch/i386/i386/mpbios.c 	caa.feature_flags = entry->feature_flags;
entry             763 arch/i386/i386/mpbios.c mp_cfg_special_intr(const struct mpbios_int *entry, u_int32_t *redir)
entry             775 arch/i386/i386/mpbios.c 	switch (entry->int_type) {
entry             793 arch/i386/i386/mpbios.c 		panic("unknown MPS interrupt type %d", entry->int_type);
entry             800 arch/i386/i386/mpbios.c mp_cfg_pci_intr(const struct mpbios_int *entry, u_int32_t *redir)
entry             802 arch/i386/i386/mpbios.c 	int mpspo = entry->int_flags & 0x03; /* XXX magic */
entry             803 arch/i386/i386/mpbios.c 	int mpstrig = (entry->int_flags >> 2) & 0x03; /* XXX magic */
entry             818 arch/i386/i386/mpbios.c 	if (entry->int_type != MPS_INTTYPE_INT) {
entry             819 arch/i386/i386/mpbios.c 		mp_cfg_special_intr(entry, redir);
entry             838 arch/i386/i386/mpbios.c mp_cfg_eisa_intr (const struct mpbios_int *entry, u_int32_t *redir)
entry             840 arch/i386/i386/mpbios.c 	int mpspo = entry->int_flags & 0x03; /* XXX magic */
entry             841 arch/i386/i386/mpbios.c 	int mpstrig = (entry->int_flags >> 2) & 0x03; /* XXX magic */
entry             856 arch/i386/i386/mpbios.c 	if (entry->int_type != MPS_INTTYPE_INT) {
entry             857 arch/i386/i386/mpbios.c 		mp_cfg_special_intr(entry, redir);
entry             874 arch/i386/i386/mpbios.c 		if (mp_busses[entry->src_bus_id].mb_data &
entry             875 arch/i386/i386/mpbios.c 		    (1<<entry->src_bus_irq)) {
entry             888 arch/i386/i386/mpbios.c mp_cfg_isa_intr(const struct mpbios_int *entry, u_int32_t *redir)
entry             890 arch/i386/i386/mpbios.c 	int mpspo = entry->int_flags & 0x03; /* XXX magic */
entry             891 arch/i386/i386/mpbios.c 	int mpstrig = (entry->int_flags >> 2) & 0x03; /* XXX magic */
entry             906 arch/i386/i386/mpbios.c 	if (entry->int_type != MPS_INTTYPE_INT) {
entry             907 arch/i386/i386/mpbios.c 		mp_cfg_special_intr(entry, redir);
entry             960 arch/i386/i386/mpbios.c 	const struct mpbios_bus *entry = (const struct mpbios_bus *)ent;
entry             961 arch/i386/i386/mpbios.c 	int bus_id = entry->bus_id;
entry             964 arch/i386/i386/mpbios.c 	    bus_id, entry->bus_type);
entry             973 arch/i386/i386/mpbios.c 		    self->dv_xname, bus_id, entry->bus_type);
entry             979 arch/i386/i386/mpbios.c 	if (memcmp(entry->bus_type, "PCI   ", 6) == 0) {
entry             984 arch/i386/i386/mpbios.c 	} else if (memcmp(entry->bus_type, "EISA  ", 6) == 0) {
entry             997 arch/i386/i386/mpbios.c 	} else if (memcmp(entry->bus_type, "ISA   ", 6) == 0) {
entry            1009 arch/i386/i386/mpbios.c 		    entry->bus_type);
entry            1017 arch/i386/i386/mpbios.c 	const struct mpbios_ioapic *entry = (const struct mpbios_ioapic *)ent;
entry            1021 arch/i386/i386/mpbios.c 	if (!(entry->apic_flags & IOAPICENTRY_FLAG_EN))
entry            1025 arch/i386/i386/mpbios.c 	aaa.apic_id = entry->apic_id;
entry            1026 arch/i386/i386/mpbios.c 	aaa.apic_version = entry->apic_version;
entry            1027 arch/i386/i386/mpbios.c 	aaa.apic_address = (u_int32_t)entry->apic_address;
entry            1037 arch/i386/i386/mpbios.c 	const struct mpbios_int *entry = (const struct mpbios_int *)ent;
entry            1038 arch/i386/i386/mpbios.c 	struct mpbios_int rw_entry = *entry;
entry            1044 arch/i386/i386/mpbios.c 	u_int32_t id = IOAPIC_REMAPPED_ID(entry->dst_apic_id);
entry            1045 arch/i386/i386/mpbios.c 	u_int32_t pin = entry->dst_apic_int;
entry            1046 arch/i386/i386/mpbios.c 	u_int32_t bus = entry->src_bus_id;
entry            1047 arch/i386/i386/mpbios.c 	u_int32_t dev = entry->src_bus_irq;
entry            1048 arch/i386/i386/mpbios.c 	u_int32_t type = entry->int_type;
entry            1049 arch/i386/i386/mpbios.c 	u_int32_t flags = entry->int_flags;
entry            1084 arch/i386/i386/mpbios.c 	if (entry->type == MPS_MCT_IOINT) {
entry              72 arch/i386/include/biosvar.h 	u_int32_t	entry;		/* initialization entry point */
entry              83 arch/i386/include/biosvar.h 	u_int32_t	entry;		/* 04: entry point */
entry             337 arch/i386/pci/pci_intr_fixup.c 	int entry;
entry             342 arch/i386/pci/pci_intr_fixup.c 	for (entry = 0; entry < pcibios_pir_table_nentries; entry++) {
entry             343 arch/i386/pci/pci_intr_fixup.c 		pir = &pcibios_pir_table[entry];
entry             371 arch/i386/pci/pci_intr_fixup.c 	int entry, pin, link;
entry             383 arch/i386/pci/pci_intr_fixup.c 	for (entry = 0; entry < pcibios_pir_table_nentries; entry++) {
entry             384 arch/i386/pci/pci_intr_fixup.c 		pir = &pcibios_pir_table[entry];
entry              48 arch/i386/stand/libsa/exec_i386.c 	u_long entry;
entry              71 arch/i386/stand/libsa/exec_i386.c 	entry = marks[MARK_ENTRY] & 0x0fffffff;
entry              73 arch/i386/stand/libsa/exec_i386.c 	printf("entry point at 0x%x\n", (int) entry);
entry              75 arch/i386/stand/libsa/exec_i386.c 	(*(startfuncp)entry)(howto, bootdev, BOOTARG_APIVER,
entry             350 dev/acpi/acpi.c 	struct acpi_q *entry;
entry             396 dev/acpi/acpi.c 	SIMPLEQ_FOREACH(entry, &sc->sc_tables, q_next) {
entry             397 dev/acpi/acpi.c 		if (memcmp(entry->q_table, FADT_SIG,
entry             399 dev/acpi/acpi.c 			sc->sc_fadt = entry->q_table;
entry             432 dev/acpi/acpi.c 		acpi_load_dsdt(sc->sc_fadt->dsdt, &entry);
entry             434 dev/acpi/acpi.c 		acpi_load_dsdt(sc->sc_fadt->x_dsdt, &entry);
entry             436 dev/acpi/acpi.c 	if (entry == NULL)
entry             438 dev/acpi/acpi.c 	SIMPLEQ_INSERT_HEAD(&sc->sc_tables, entry, q_next);
entry             440 dev/acpi/acpi.c 	p_dsdt = entry->q_table;
entry             445 dev/acpi/acpi.c 	SIMPLEQ_FOREACH(entry, &sc->sc_tables, q_next) {
entry             446 dev/acpi/acpi.c 		if (memcmp(entry->q_table, SSDT_SIG,
entry             448 dev/acpi/acpi.c 			p_dsdt = entry->q_table;
entry             515 dev/acpi/acpi.c 	SIMPLEQ_FOREACH(entry, &sc->sc_tables, q_next) {
entry             516 dev/acpi/acpi.c 		printf("%.4s ", entry->q_table);
entry             542 dev/acpi/acpi.c 	SIMPLEQ_FOREACH(entry, &sc->sc_tables, q_next) {
entry             552 dev/acpi/acpi.c 		aaa.aaa_table = entry->q_table;
entry             717 dev/acpi/acpi.c 	struct acpi_q *entry;
entry             719 dev/acpi/acpi.c 	entry = malloc(len + sizeof(struct acpi_q), M_DEVBUF, M_NOWAIT);
entry             721 dev/acpi/acpi.c 	if (entry != NULL) {
entry             723 dev/acpi/acpi.c 			free(entry, M_DEVBUF);
entry             726 dev/acpi/acpi.c 		memcpy(entry->q_data, handle.va, len);
entry             727 dev/acpi/acpi.c 		entry->q_table = entry->q_data;
entry             729 dev/acpi/acpi.c 		SIMPLEQ_INSERT_TAIL(queue, entry, q_next);
entry             157 dev/acpi/acpimadt.c 		union acpi_madt_entry *entry = (union acpi_madt_entry *)addr;
entry             159 dev/acpi/acpimadt.c 		switch (entry->madt_lapic.apic_type) {
entry             162 dev/acpi/acpimadt.c 			    self->dv_xname, entry->madt_lapic.acpi_proc_id,
entry             163 dev/acpi/acpimadt.c 			    entry->madt_lapic.apic_id,
entry             164 dev/acpi/acpimadt.c 			    entry->madt_lapic.flags);
entry             166 dev/acpi/acpimadt.c 			lapic_map[entry->madt_lapic.acpi_proc_id] =
entry             167 dev/acpi/acpimadt.c 			    entry->madt_lapic.apic_id;
entry             172 dev/acpi/acpimadt.c 				if ((entry->madt_lapic.flags & ACPI_PROC_ENABLE) == 0)
entry             178 dev/acpi/acpimadt.c 				caa.cpu_number = entry->madt_lapic.apic_id;
entry             193 dev/acpi/acpimadt.c 			    self->dv_xname, entry->madt_ioapic.acpi_ioapic_id,
entry             194 dev/acpi/acpimadt.c 			    entry->madt_ioapic.address,
entry             195 dev/acpi/acpimadt.c 			    entry->madt_ioapic.global_int_base);
entry             202 dev/acpi/acpimadt.c 				aaa.apic_id = entry->madt_ioapic.acpi_ioapic_id;
entry             203 dev/acpi/acpimadt.c 				aaa.apic_address = entry->madt_ioapic.address;
entry             204 dev/acpi/acpimadt.c 				aaa.apic_vecbase = entry->madt_ioapic.global_int_base;
entry             213 dev/acpi/acpimadt.c 		addr += entry->madt_lapic.length;
entry             223 dev/acpi/acpimadt.c 		union acpi_madt_entry *entry = (union acpi_madt_entry *)addr;
entry             225 dev/acpi/acpimadt.c 		switch (entry->madt_lapic.apic_type) {
entry             232 dev/acpi/acpimadt.c 			    self->dv_xname, entry->madt_override.bus,
entry             233 dev/acpi/acpimadt.c 			    entry->madt_override.source,
entry             234 dev/acpi/acpimadt.c 			    entry->madt_override.global_int,
entry             235 dev/acpi/acpimadt.c 			    entry->madt_override.flags);
entry             237 dev/acpi/acpimadt.c 			pin = entry->madt_override.global_int;
entry             247 dev/acpi/acpimadt.c 			map->bus_pin = entry->madt_override.source;
entry             248 dev/acpi/acpimadt.c 			map->flags = entry->madt_override.flags;
entry             250 dev/acpi/acpimadt.c 			map->global_int = entry->madt_override.global_int;
entry             252 dev/acpi/acpimadt.c 			acpimadt_cfg_intr(entry->madt_override.flags, &map->redir);
entry             266 dev/acpi/acpimadt.c 			    self->dv_xname, entry->madt_lapic_nmi.acpi_proc_id,
entry             267 dev/acpi/acpimadt.c 			    entry->madt_lapic_nmi.local_apic_lint,
entry             268 dev/acpi/acpimadt.c 			    entry->madt_lapic_nmi.flags);
entry             270 dev/acpi/acpimadt.c 			pin = entry->madt_lapic_nmi.local_apic_lint;
entry             274 dev/acpi/acpimadt.c 			map->cpu_id = lapic_map[entry->madt_lapic_nmi.acpi_proc_id];
entry             276 dev/acpi/acpimadt.c 			map->flags = entry->madt_lapic_nmi.flags;
entry             278 dev/acpi/acpimadt.c 			acpimadt_cfg_intr(entry->madt_lapic_nmi.flags, &map->redir);
entry             285 dev/acpi/acpimadt.c 			    self->dv_xname, entry->madt_lapic.apic_type);
entry             288 dev/acpi/acpimadt.c 		addr += entry->madt_lapic.length;
entry             249 dev/i2o/i2o.h  	struct i2o_hrt_entry	entry[1];
entry             285 dev/i2o/i2o.h  	struct i2o_lct_entry	entry[1];
entry             319 dev/i2o/i2o.h  	struct	i2o_systab_entry entry[1];
entry             478 dev/i2o/iop.c  		for (i = 0, ste = iop_systab->entry; i < iop_cd.cd_ndevs; i++)
entry             637 dev/i2o/iop.c  		le = sc->sc_lct->entry;
entry             754 dev/i2o/iop.c  	for (i = 0, le = sc->sc_lct->entry; i < nent; i++, le++) {
entry             277 dev/i2o/iopsp.c 	for (le = iop->sc_lct->entry; nent != 0; nent--, le++)
entry             286 dev/i2o/iopsp.c 	for (i = 0, le = iop->sc_lct->entry; i < nent; i++, le++) {
entry            3522 dev/ic/aic79xx.c 	struct ahd_phase_table_entry *entry;
entry            3530 dev/ic/aic79xx.c 	for (entry = ahd_phase_table; entry < last_entry; entry++) {
entry            3531 dev/ic/aic79xx.c 		if (phase == entry->phase)
entry            3534 dev/ic/aic79xx.c 	return (entry);
entry            8807 dev/ic/aic79xx.c 	int entry, printed;
entry            8823 dev/ic/aic79xx.c 		for (entry = 0; entry < num_entries; entry++) {
entry            8824 dev/ic/aic79xx.c 			if (((value & table[entry].mask) != table[entry].value)
entry            8825 dev/ic/aic79xx.c 			    || ((printed_mask & table[entry].mask) ==
entry            8826 dev/ic/aic79xx.c 			    table[entry].mask))
entry            8831 dev/ic/aic79xx.c 					  table[entry].name);
entry            8832 dev/ic/aic79xx.c 			printed_mask |= table[entry].mask;
entry            8836 dev/ic/aic79xx.c 		if (entry >= num_entries)
entry            2275 dev/ic/aic7xxx.c 	struct ahc_phase_table_entry *entry;
entry            2283 dev/ic/aic7xxx.c 	for (entry = ahc_phase_table; entry < last_entry; entry++) {
entry            2284 dev/ic/aic7xxx.c 		if (phase == entry->phase)
entry            2287 dev/ic/aic7xxx.c 	return (entry);
entry            6444 dev/ic/aic7xxx.c 	int entry, printed;
entry            6460 dev/ic/aic7xxx.c 		for (entry = 0; entry < num_entries; entry++) {
entry            6461 dev/ic/aic7xxx.c 			if (((value & table[entry].mask) != table[entry].value)
entry            6462 dev/ic/aic7xxx.c 			    || ((printed_mask & table[entry].mask) ==
entry            6463 dev/ic/aic7xxx.c 			    table[entry].mask))
entry            6468 dev/ic/aic7xxx.c 					  table[entry].name);
entry            6469 dev/ic/aic7xxx.c 			printed_mask |= table[entry].mask;
entry            6473 dev/ic/aic7xxx.c 		if (entry >= num_entries)
entry            1814 dev/ic/ar5210.c ar5k_ar5210_reset_key(struct ath_hal *hal, u_int16_t entry)
entry            1818 dev/ic/ar5210.c 	AR5K_ASSERT_ENTRY(entry, AR5K_AR5210_KEYTABLE_SIZE);
entry            1821 dev/ic/ar5210.c 		AR5K_REG_WRITE(AR5K_AR5210_KEYTABLE_OFF(entry, i), 0);
entry            1827 dev/ic/ar5210.c ar5k_ar5210_is_key_valid(struct ath_hal *hal, u_int16_t entry)
entry            1829 dev/ic/ar5210.c 	AR5K_ASSERT_ENTRY(entry, AR5K_AR5210_KEYTABLE_SIZE);
entry            1834 dev/ic/ar5210.c 	if (AR5K_REG_READ(AR5K_AR5210_KEYTABLE_MAC1(entry)) &
entry            1842 dev/ic/ar5210.c ar5k_ar5210_set_key(struct ath_hal *hal, u_int16_t entry,
entry            1848 dev/ic/ar5210.c 	AR5K_ASSERT_ENTRY(entry, AR5K_AR5210_KEYTABLE_SIZE);
entry            1883 dev/ic/ar5210.c 		AR5K_REG_WRITE(AR5K_AR5210_KEYTABLE_OFF(entry, i), key_v[i]);
entry            1885 dev/ic/ar5210.c 	return (ar5k_ar5210_set_key_lladdr(hal, entry, mac));
entry            1889 dev/ic/ar5210.c ar5k_ar5210_set_key_lladdr(struct ath_hal *hal, u_int16_t entry,
entry            1898 dev/ic/ar5210.c 	AR5K_ASSERT_ENTRY(entry, AR5K_AR5210_KEYTABLE_SIZE);
entry            1906 dev/ic/ar5210.c 	AR5K_REG_WRITE(AR5K_AR5210_KEYTABLE_MAC0(entry), low_id);
entry            1907 dev/ic/ar5210.c 	AR5K_REG_WRITE(AR5K_AR5210_KEYTABLE_MAC1(entry), high_id);
entry            1919 dev/ic/ar5211.c ar5k_ar5211_reset_key(struct ath_hal *hal, u_int16_t entry)
entry            1923 dev/ic/ar5211.c 	AR5K_ASSERT_ENTRY(entry, AR5K_AR5211_KEYTABLE_SIZE);
entry            1926 dev/ic/ar5211.c 		AR5K_REG_WRITE(AR5K_AR5211_KEYTABLE_OFF(entry, i), 0);
entry            1932 dev/ic/ar5211.c ar5k_ar5211_is_key_valid(struct ath_hal *hal, u_int16_t entry)
entry            1934 dev/ic/ar5211.c 	AR5K_ASSERT_ENTRY(entry, AR5K_AR5211_KEYTABLE_SIZE);
entry            1939 dev/ic/ar5211.c 	if (AR5K_REG_READ(AR5K_AR5211_KEYTABLE_MAC1(entry)) &
entry            1947 dev/ic/ar5211.c ar5k_ar5211_set_key(struct ath_hal *hal, u_int16_t entry,
entry            1953 dev/ic/ar5211.c 	AR5K_ASSERT_ENTRY(entry, AR5K_AR5211_KEYTABLE_SIZE);
entry            1988 dev/ic/ar5211.c 		AR5K_REG_WRITE(AR5K_AR5211_KEYTABLE_OFF(entry, i), key_v[i]);
entry            1990 dev/ic/ar5211.c 	return (ar5k_ar5211_set_key_lladdr(hal, entry, mac));
entry            1994 dev/ic/ar5211.c ar5k_ar5211_set_key_lladdr(struct ath_hal *hal, u_int16_t entry,
entry            2003 dev/ic/ar5211.c 	AR5K_ASSERT_ENTRY(entry, AR5K_AR5211_KEYTABLE_SIZE);
entry            2011 dev/ic/ar5211.c 	AR5K_REG_WRITE(AR5K_AR5211_KEYTABLE_MAC0(entry), low_id);
entry            2012 dev/ic/ar5211.c 	AR5K_REG_WRITE(AR5K_AR5211_KEYTABLE_MAC1(entry), high_id);
entry            2254 dev/ic/ar5212.c ar5k_ar5212_reset_key(struct ath_hal *hal, u_int16_t entry)
entry            2258 dev/ic/ar5212.c 	AR5K_ASSERT_ENTRY(entry, AR5K_AR5212_KEYTABLE_SIZE);
entry            2261 dev/ic/ar5212.c 		AR5K_REG_WRITE(AR5K_AR5212_KEYTABLE_OFF(entry, i), 0);
entry            2264 dev/ic/ar5212.c 	AR5K_REG_WRITE(AR5K_AR5212_KEYTABLE_TYPE(entry),
entry            2271 dev/ic/ar5212.c ar5k_ar5212_is_key_valid(struct ath_hal *hal, u_int16_t entry)
entry            2273 dev/ic/ar5212.c 	AR5K_ASSERT_ENTRY(entry, AR5K_AR5212_KEYTABLE_SIZE);
entry            2278 dev/ic/ar5212.c 	if (AR5K_REG_READ(AR5K_AR5212_KEYTABLE_MAC1(entry)) &
entry            2286 dev/ic/ar5212.c ar5k_ar5212_set_key(struct ath_hal *hal, u_int16_t entry,
entry            2292 dev/ic/ar5212.c 	AR5K_ASSERT_ENTRY(entry, AR5K_AR5212_KEYTABLE_SIZE);
entry            2327 dev/ic/ar5212.c 		AR5K_REG_WRITE(AR5K_AR5212_KEYTABLE_OFF(entry, i), key_v[i]);
entry            2329 dev/ic/ar5212.c 	return (ar5k_ar5212_set_key_lladdr(hal, entry, mac));
entry            2333 dev/ic/ar5212.c ar5k_ar5212_set_key_lladdr(struct ath_hal *hal, u_int16_t entry,
entry            2342 dev/ic/ar5212.c 	AR5K_ASSERT_ENTRY(entry, AR5K_AR5212_KEYTABLE_SIZE);
entry            2350 dev/ic/ar5212.c 	AR5K_REG_WRITE(AR5K_AR5212_KEYTABLE_MAC0(entry), low_id);
entry            2351 dev/ic/ar5212.c 	AR5K_REG_WRITE(AR5K_AR5212_KEYTABLE_MAC1(entry), high_id);
entry            1302 dev/ic/ar5xxx.c 	u_int32_t mask, entry, last, data, shift, position;
entry            1316 dev/ic/ar5xxx.c 	entry = ((first - 1) / 8) + offset;
entry            1322 dev/ic/ar5xxx.c 	for (i = shift = 0, left = bits; left > 0; position = 0, entry++, i++) {
entry            1328 dev/ic/ar5xxx.c 			rf[entry] &= ~mask;
entry            1329 dev/ic/ar5xxx.c 			rf[entry] |= ((data << position) << (col * 8)) & mask;
entry            1332 dev/ic/ar5xxx.c 			data = (((rf[entry] & mask) >> (col * 8)) >>
entry             760 dev/ic/sti.c   	a.in.entry = i;
entry             471 dev/ic/stireg.h 	u_int32_t	entry;
entry             799 dev/isa/ad1848.c   ad1848_devmap_t *entry;
entry             804 dev/isa/ad1848.c   if (!(entry = ad1848_mixer_find_dev(map, cnt, cp)))
entry             807 dev/isa/ad1848.c   dev = entry->dev;
entry             809 dev/isa/ad1848.c   switch (entry->kind) {
entry             873 dev/isa/ad1848.c   ad1848_devmap_t *entry;
entry             878 dev/isa/ad1848.c   if (!(entry = ad1848_mixer_find_dev(map, cnt, cp)))
entry             881 dev/isa/ad1848.c   dev = entry->dev;
entry             883 dev/isa/ad1848.c   switch (entry->kind) {
entry             659 dev/pci/ahc_pci.c 	const struct	   ahc_pci_identity *entry;
entry             679 dev/pci/ahc_pci.c 		entry = &ahc_pci_ident_table[i];
entry             680 dev/pci/ahc_pci.c 		if (entry->full_id == (full_id & entry->id_mask))
entry             681 dev/pci/ahc_pci.c 			return (entry);
entry             692 dev/pci/ahc_pci.c 	const struct	   ahc_pci_identity *entry;
entry             696 dev/pci/ahc_pci.c 	entry = ahc_find_pci_device(pa->pa_id, subid, pa->pa_function);
entry             697 dev/pci/ahc_pci.c 	return (entry != NULL && entry->setup != NULL) ? 1 : 0;
entry             706 dev/pci/ahc_pci.c 	const struct	   ahc_pci_identity *entry;
entry             753 dev/pci/ahc_pci.c 	entry = ahc_find_pci_device(pa->pa_id, subid, pa->pa_function);
entry             754 dev/pci/ahc_pci.c 	if (entry == NULL)
entry             773 dev/pci/ahc_pci.c 	error = entry->setup(ahc);
entry             298 dev/pci/ahd_pci.c 	const struct ahd_pci_identity *entry;
entry             314 dev/pci/ahd_pci.c 		entry = &ahd_pci_ident_table[i];
entry             315 dev/pci/ahd_pci.c 		if (entry->full_id == (full_id & entry->id_mask)) {
entry             316 dev/pci/ahd_pci.c 			return (entry);
entry             325 dev/pci/ahd_pci.c 	const struct ahd_pci_identity *entry;
entry             330 dev/pci/ahd_pci.c 	entry = ahd_find_pci_device(pa->pa_id, subid);
entry             331 dev/pci/ahd_pci.c 	return entry != NULL ? 1 : 0;
entry             337 dev/pci/ahd_pci.c 	const struct ahd_pci_identity *entry;
entry             353 dev/pci/ahd_pci.c 	entry = ahd_find_pci_device(pa->pa_id, subid);
entry             354 dev/pci/ahd_pci.c 	if (entry == NULL)
entry             371 dev/pci/ahd_pci.c 	error = entry->setup(ahd, pa);
entry             624 dev/pci/if_bge.c 	struct bge_jpool_entry   *entry;
entry             675 dev/pci/if_bge.c 		entry = malloc(sizeof(struct bge_jpool_entry),
entry             677 dev/pci/if_bge.c 		if (entry == NULL) {
entry             683 dev/pci/if_bge.c 		entry->slot = i;
entry             685 dev/pci/if_bge.c 				 entry, jpool_entries);
entry             715 dev/pci/if_bge.c 	struct bge_jpool_entry   *entry;
entry             717 dev/pci/if_bge.c 	entry = SLIST_FIRST(&sc->bge_jfree_listhead);
entry             719 dev/pci/if_bge.c 	if (entry == NULL)
entry             723 dev/pci/if_bge.c 	SLIST_INSERT_HEAD(&sc->bge_jinuse_listhead, entry, jpool_entries);
entry             724 dev/pci/if_bge.c 	return (sc->bge_cdata.bge_jslots[entry->slot]);
entry             733 dev/pci/if_bge.c 	struct bge_jpool_entry *entry;
entry             751 dev/pci/if_bge.c 	entry = SLIST_FIRST(&sc->bge_jinuse_listhead);
entry             752 dev/pci/if_bge.c 	if (entry == NULL)
entry             754 dev/pci/if_bge.c 	entry->slot = i;
entry             756 dev/pci/if_bge.c 	SLIST_INSERT_HEAD(&sc->bge_jfree_listhead, entry, jpool_entries);
entry             753 dev/pci/if_lge.c 	struct lge_jpool_entry   *entry;
entry             805 dev/pci/if_lge.c 		entry = malloc(sizeof(struct lge_jpool_entry), 
entry             807 dev/pci/if_lge.c 		if (entry == NULL) {
entry             814 dev/pci/if_lge.c 		entry->slot = i;
entry             816 dev/pci/if_lge.c 				 entry, jpool_entries);
entry             844 dev/pci/if_lge.c 	struct lge_jpool_entry   *entry;
entry             846 dev/pci/if_lge.c 	entry = LIST_FIRST(&sc->lge_jfree_listhead);
entry             848 dev/pci/if_lge.c 	if (entry == NULL)
entry             851 dev/pci/if_lge.c 	LIST_REMOVE(entry, jpool_entries);
entry             852 dev/pci/if_lge.c 	LIST_INSERT_HEAD(&sc->lge_jinuse_listhead, entry, jpool_entries);
entry             853 dev/pci/if_lge.c 	return (sc->lge_cdata.lge_jslots[entry->slot]);
entry             864 dev/pci/if_lge.c 	struct lge_jpool_entry   *entry;
entry             878 dev/pci/if_lge.c 	entry = LIST_FIRST(&sc->lge_jinuse_listhead);
entry             879 dev/pci/if_lge.c 	if (entry == NULL)
entry             881 dev/pci/if_lge.c 	entry->slot = i;
entry             882 dev/pci/if_lge.c 	LIST_REMOVE(entry, jpool_entries);
entry             883 dev/pci/if_lge.c 	LIST_INSERT_HEAD(&sc->lge_jfree_listhead, entry, jpool_entries);
entry             467 dev/pci/if_msk.c 	struct sk_txmap_entry	*entry;
entry             486 dev/pci/if_msk.c 		entry = malloc(sizeof(*entry), M_DEVBUF, M_NOWAIT);
entry             487 dev/pci/if_msk.c 		if (!entry) {
entry             491 dev/pci/if_msk.c 		entry->dmamap = dmamap;
entry             492 dev/pci/if_msk.c 		SIMPLEQ_INSERT_HEAD(&sc_if->sk_txmap_head, entry, link);
entry             570 dev/pci/if_msk.c 	struct sk_jpool_entry   *entry;
entry             620 dev/pci/if_msk.c 		entry = malloc(sizeof(struct sk_jpool_entry),
entry             622 dev/pci/if_msk.c 		if (entry == NULL) {
entry             628 dev/pci/if_msk.c 		entry->slot = i;
entry             630 dev/pci/if_msk.c 				 entry, jpool_entries);
entry             660 dev/pci/if_msk.c 	struct sk_jpool_entry   *entry;
entry             662 dev/pci/if_msk.c 	entry = LIST_FIRST(&sc_if->sk_jfree_listhead);
entry             664 dev/pci/if_msk.c 	if (entry == NULL)
entry             667 dev/pci/if_msk.c 	LIST_REMOVE(entry, jpool_entries);
entry             668 dev/pci/if_msk.c 	LIST_INSERT_HEAD(&sc_if->sk_jinuse_listhead, entry, jpool_entries);
entry             669 dev/pci/if_msk.c 	return (sc_if->sk_cdata.sk_jslots[entry->slot]);
entry             678 dev/pci/if_msk.c 	struct sk_jpool_entry *entry;
entry             695 dev/pci/if_msk.c 	entry = LIST_FIRST(&sc->sk_jinuse_listhead);
entry             696 dev/pci/if_msk.c 	if (entry == NULL)
entry             698 dev/pci/if_msk.c 	entry->slot = i;
entry             699 dev/pci/if_msk.c 	LIST_REMOVE(entry, jpool_entries);
entry             700 dev/pci/if_msk.c 	LIST_INSERT_HEAD(&sc->sk_jfree_listhead, entry, jpool_entries);
entry            1409 dev/pci/if_msk.c 	struct sk_txmap_entry	*entry;
entry            1414 dev/pci/if_msk.c 	entry = SIMPLEQ_FIRST(&sc_if->sk_txmap_head);
entry            1415 dev/pci/if_msk.c 	if (entry == NULL) {
entry            1419 dev/pci/if_msk.c 	txmap = entry->dmamap;
entry            1467 dev/pci/if_msk.c 	sc_if->sk_cdata.sk_tx_map[cur] = entry;
entry            1679 dev/pci/if_msk.c 	struct sk_txmap_entry	*entry;
entry            1706 dev/pci/if_msk.c 			entry = sc_if->sk_cdata.sk_tx_map[idx];
entry            1711 dev/pci/if_msk.c 			bus_dmamap_sync(sc->sc_dmatag, entry->dmamap, 0,
entry            1712 dev/pci/if_msk.c 			    entry->dmamap->dm_mapsize, BUS_DMASYNC_POSTWRITE);
entry            1714 dev/pci/if_msk.c 			bus_dmamap_unload(sc->sc_dmatag, entry->dmamap);
entry            1715 dev/pci/if_msk.c 			SIMPLEQ_INSERT_TAIL(&sc_if->sk_txmap_head, entry,
entry            1112 dev/pci/if_nge.c 	struct nge_jpool_entry	*entry;
entry            1169 dev/pci/if_nge.c 		entry = malloc(sizeof(struct nge_jpool_entry),
entry            1171 dev/pci/if_nge.c 		if (entry == NULL) {
entry            1178 dev/pci/if_nge.c 		entry->slot = i;
entry            1179 dev/pci/if_nge.c 		LIST_INSERT_HEAD(&sc->nge_jfree_listhead, entry,
entry            1209 dev/pci/if_nge.c 	struct nge_jpool_entry   *entry;
entry            1211 dev/pci/if_nge.c 	entry = LIST_FIRST(&sc->nge_jfree_listhead);
entry            1213 dev/pci/if_nge.c 	if (entry == NULL)
entry            1216 dev/pci/if_nge.c 	LIST_REMOVE(entry, jpool_entries);
entry            1217 dev/pci/if_nge.c 	LIST_INSERT_HEAD(&sc->nge_jinuse_listhead, entry, jpool_entries);
entry            1218 dev/pci/if_nge.c 	sc->nge_cdata.nge_jslots[entry->slot].nge_inuse = 1;
entry            1219 dev/pci/if_nge.c 	return(sc->nge_cdata.nge_jslots[entry->slot].nge_buf);
entry            1233 dev/pci/if_nge.c 	struct nge_jpool_entry *entry;
entry            1253 dev/pci/if_nge.c 			entry = LIST_FIRST(&sc->nge_jinuse_listhead);
entry            1254 dev/pci/if_nge.c 			if (entry == NULL)
entry            1256 dev/pci/if_nge.c 			entry->slot = i;
entry            1257 dev/pci/if_nge.c 			LIST_REMOVE(entry, jpool_entries);
entry            1259 dev/pci/if_nge.c 					 entry, jpool_entries);
entry             101 dev/pci/if_san_xilinx.c 	SIMPLEQ_ENTRY(xilinx_rx_buffer) entry;
entry             424 dev/pci/if_san_xilinx.c 		    sc->rx_dma_buf, entry);
entry             588 dev/pci/if_san_xilinx.c 		SIMPLEQ_REMOVE_HEAD(&sc->wp_rx_free_list, entry);
entry             593 dev/pci/if_san_xilinx.c 		SIMPLEQ_REMOVE_HEAD(&sc->wp_rx_complete_list, entry);
entry            1689 dev/pci/if_san_xilinx.c 	SIMPLEQ_REMOVE_HEAD(&sc->wp_rx_free_list, entry);
entry            2105 dev/pci/if_san_xilinx.c 	SIMPLEQ_INSERT_TAIL(&sc->wp_rx_complete_list, buf, entry);
entry            2413 dev/pci/if_san_xilinx.c 		SIMPLEQ_REMOVE_HEAD(&sc->wp_rx_free_list, entry);
entry            2418 dev/pci/if_san_xilinx.c 		SIMPLEQ_REMOVE_HEAD(&sc->wp_rx_complete_list, entry);
entry            2470 dev/pci/if_san_xilinx.c 		SIMPLEQ_INSERT_TAIL(&sc->wp_rx_free_list, buf, entry);
entry            2493 dev/pci/if_san_xilinx.c 	SIMPLEQ_INSERT_TAIL(&sc->wp_rx_free_list, buf, entry);
entry            2531 dev/pci/if_san_xilinx.c 		SIMPLEQ_REMOVE_HEAD(&sc->wp_rx_complete_list, entry);
entry             596 dev/pci/if_sk.c 	struct sk_txmap_entry	*entry;
entry             616 dev/pci/if_sk.c 		entry = malloc(sizeof(*entry), M_DEVBUF, M_NOWAIT);
entry             617 dev/pci/if_sk.c 		if (!entry) {
entry             621 dev/pci/if_sk.c 		entry->dmamap = dmamap;
entry             622 dev/pci/if_sk.c 		SIMPLEQ_INSERT_HEAD(&sc_if->sk_txmap_head, entry, link);
entry             698 dev/pci/if_sk.c 	struct sk_jpool_entry   *entry;
entry             748 dev/pci/if_sk.c 		entry = malloc(sizeof(struct sk_jpool_entry),
entry             750 dev/pci/if_sk.c 		if (entry == NULL) {
entry             756 dev/pci/if_sk.c 		entry->slot = i;
entry             758 dev/pci/if_sk.c 				 entry, jpool_entries);
entry             788 dev/pci/if_sk.c 	struct sk_jpool_entry   *entry;
entry             790 dev/pci/if_sk.c 	entry = LIST_FIRST(&sc_if->sk_jfree_listhead);
entry             792 dev/pci/if_sk.c 	if (entry == NULL)
entry             795 dev/pci/if_sk.c 	LIST_REMOVE(entry, jpool_entries);
entry             796 dev/pci/if_sk.c 	LIST_INSERT_HEAD(&sc_if->sk_jinuse_listhead, entry, jpool_entries);
entry             797 dev/pci/if_sk.c 	return (sc_if->sk_cdata.sk_jslots[entry->slot]);
entry             806 dev/pci/if_sk.c 	struct sk_jpool_entry *entry;
entry             823 dev/pci/if_sk.c 	entry = LIST_FIRST(&sc->sk_jinuse_listhead);
entry             824 dev/pci/if_sk.c 	if (entry == NULL)
entry             826 dev/pci/if_sk.c 	entry->slot = i;
entry             827 dev/pci/if_sk.c 	LIST_REMOVE(entry, jpool_entries);
entry             828 dev/pci/if_sk.c 	LIST_INSERT_HEAD(&sc->sk_jfree_listhead, entry, jpool_entries);
entry            1523 dev/pci/if_sk.c 	struct sk_txmap_entry	*entry;
entry            1528 dev/pci/if_sk.c 	entry = SIMPLEQ_FIRST(&sc_if->sk_txmap_head);
entry            1529 dev/pci/if_sk.c 	if (entry == NULL) {
entry            1533 dev/pci/if_sk.c 	txmap = entry->dmamap;
entry            1581 dev/pci/if_sk.c 	sc_if->sk_cdata.sk_tx_map[cur] = entry;
entry            1915 dev/pci/if_sk.c 	struct sk_txmap_entry	*entry;
entry            1941 dev/pci/if_sk.c 			entry = sc_if->sk_cdata.sk_tx_map[idx];
entry            1946 dev/pci/if_sk.c 			bus_dmamap_sync(sc->sc_dmatag, entry->dmamap, 0,
entry            1947 dev/pci/if_sk.c 			    entry->dmamap->dm_mapsize, BUS_DMASYNC_POSTWRITE);
entry            1949 dev/pci/if_sk.c 			bus_dmamap_unload(sc->sc_dmatag, entry->dmamap);
entry            1950 dev/pci/if_sk.c 			SIMPLEQ_INSERT_TAIL(&sc_if->sk_txmap_head, entry,
entry             590 dev/pci/if_ti.c 	struct ti_jpool_entry *entry;
entry             641 dev/pci/if_ti.c 		entry = malloc(sizeof(struct ti_jpool_entry),
entry             643 dev/pci/if_ti.c 		if (entry == NULL) {
entry             650 dev/pci/if_ti.c 		entry->slot = i;
entry             651 dev/pci/if_ti.c 		SLIST_INSERT_HEAD(&sc->ti_jfree_listhead, entry, jpool_entries);
entry             681 dev/pci/if_ti.c 	struct ti_jpool_entry   *entry;
entry             683 dev/pci/if_ti.c 	entry = SLIST_FIRST(&sc->ti_jfree_listhead);
entry             685 dev/pci/if_ti.c 	if (entry == NULL)
entry             689 dev/pci/if_ti.c 	SLIST_INSERT_HEAD(&sc->ti_jinuse_listhead, entry, jpool_entries);
entry             690 dev/pci/if_ti.c 	sc->ti_cdata.ti_jslots[entry->slot].ti_inuse = 1;
entry             691 dev/pci/if_ti.c 	return (sc->ti_cdata.ti_jslots[entry->slot].ti_buf);
entry             702 dev/pci/if_ti.c 	struct ti_jpool_entry	*entry;
entry             720 dev/pci/if_ti.c 		entry = SLIST_FIRST(&sc->ti_jinuse_listhead);
entry             721 dev/pci/if_ti.c 		if (entry == NULL)
entry             723 dev/pci/if_ti.c 		entry->slot = i;
entry             726 dev/pci/if_ti.c 				  entry, jpool_entries);
entry            1021 dev/pci/if_ti.c 	struct ti_txmap_entry *entry;
entry            1038 dev/pci/if_ti.c 	while ((entry = SLIST_FIRST(&sc->ti_tx_map_listhead))) {
entry            1040 dev/pci/if_ti.c 		bus_dmamap_destroy(sc->sc_dmatag, entry->dmamap);
entry            1041 dev/pci/if_ti.c 		free(entry, M_DEVBUF);
entry            1050 dev/pci/if_ti.c 	struct ti_txmap_entry *entry;
entry            1063 dev/pci/if_ti.c 		entry = malloc(sizeof(*entry), M_DEVBUF, M_NOWAIT);
entry            1064 dev/pci/if_ti.c 		if (!entry) {
entry            1068 dev/pci/if_ti.c 		entry->dmamap = dmamap;
entry            1069 dev/pci/if_ti.c 		SLIST_INSERT_HEAD(&sc->ti_tx_map_listhead, entry, link);
entry            1843 dev/pci/if_ti.c 	struct ti_txmap_entry	*entry;
entry            1867 dev/pci/if_ti.c 			entry = sc->ti_cdata.ti_tx_map[idx];
entry            1868 dev/pci/if_ti.c 			bus_dmamap_sync(sc->sc_dmatag, entry->dmamap, 0,
entry            1869 dev/pci/if_ti.c 			    entry->dmamap->dm_mapsize, BUS_DMASYNC_POSTWRITE);
entry            1871 dev/pci/if_ti.c 			bus_dmamap_unload(sc->sc_dmatag, entry->dmamap);
entry            1872 dev/pci/if_ti.c 			SLIST_INSERT_HEAD(&sc->ti_tx_map_listhead, entry,
entry            1893 dev/pci/if_ti.c 	struct ti_txmap_entry	*entry;
entry            1913 dev/pci/if_ti.c 			entry = sc->ti_cdata.ti_tx_map[idx];
entry            1914 dev/pci/if_ti.c 			bus_dmamap_sync(sc->sc_dmatag, entry->dmamap, 0,
entry            1915 dev/pci/if_ti.c 			    entry->dmamap->dm_mapsize, BUS_DMASYNC_POSTWRITE);
entry            1917 dev/pci/if_ti.c 			bus_dmamap_unload(sc->sc_dmatag, entry->dmamap);
entry            1918 dev/pci/if_ti.c 			SLIST_INSERT_HEAD(&sc->ti_tx_map_listhead, entry,
entry            1998 dev/pci/if_ti.c 	struct ti_txmap_entry	*entry;
entry            2010 dev/pci/if_ti.c 	entry = SLIST_FIRST(&sc->ti_tx_map_listhead);
entry            2011 dev/pci/if_ti.c 	if (entry == NULL)
entry            2013 dev/pci/if_ti.c 	txmap = entry->dmamap;
entry            2069 dev/pci/if_ti.c 	sc->ti_cdata.ti_tx_map[cur] = entry;
entry            2086 dev/pci/if_ti.c 	struct ti_txmap_entry	*entry;
entry            2097 dev/pci/if_ti.c 	entry = SLIST_FIRST(&sc->ti_tx_map_listhead);
entry            2098 dev/pci/if_ti.c 	if (entry == NULL)
entry            2100 dev/pci/if_ti.c 	txmap = entry->dmamap;
entry            2153 dev/pci/if_ti.c 	sc->ti_cdata.ti_tx_map[cur] = entry;
entry              37 dev/sdmmc/sdmmc_io.c 	TAILQ_ENTRY(sdmmc_intr_handler) entry;
entry             634 dev/sdmmc/sdmmc_io.c 	TAILQ_INSERT_TAIL(&sc->sc_intrq, ih, entry);
entry             653 dev/sdmmc/sdmmc_io.c 	TAILQ_REMOVE(&sc->sc_intrq, ih, entry);
entry             689 dev/sdmmc/sdmmc_io.c 	TAILQ_FOREACH(ih, &sc->sc_intrq, entry) {
entry             107 kern/exec_ecoff.c 	epp->ep_entry = eap->entry;
entry             138 kern/exec_ecoff.c 	epp->ep_entry = eap->entry;
entry             178 kern/exec_ecoff.c 	epp->ep_entry = eap->entry;
entry             445 kern/kern_lkm.c 		curp->entry = (int (*)(struct lkm_table *, int, int))
entry             449 kern/kern_lkm.c 		printf("LKM: call entrypoint %x\n", curp->entry);
entry             453 kern/kern_lkm.c 		error = (*(curp->entry))(curp, LKM_E_LOAD, curp->ver);
entry             493 kern/kern_lkm.c 		if ((*(curp->entry))(curp, LKM_E_UNLOAD, curp->ver)) {
entry             510 kern/kern_lkm.c 		if ((error = (*curp->entry)(curp, LKM_E_STAT, curp->ver)))
entry             269 lib/libsa/loadfile.c 	marks[MARK_ENTRY] = LOADADDR(coff->a.entry);
entry             281 lib/libsa/loadfile.c 	u_long entry = x->a_entry;
entry             295 lib/libsa/loadfile.c 	minp = maxp = ALIGNENTRY(entry);
entry             453 lib/libsa/loadfile.c 	marks[MARK_ENTRY] = LOADADDR(entry);
entry             131 net/if_media.c 	struct ifmedia_entry *entry;
entry             144 net/if_media.c 	entry = malloc(sizeof(*entry), M_IFADDR, M_NOWAIT);
entry             145 net/if_media.c 	if (entry == NULL)
entry             148 net/if_media.c 	entry->ifm_media = mword;
entry             149 net/if_media.c 	entry->ifm_data = data;
entry             150 net/if_media.c 	entry->ifm_aux = aux;
entry             152 net/if_media.c 	TAILQ_INSERT_TAIL(&ifm->ifm_list, entry, ifm_list);
entry             294 net/pf.c       RB_GENERATE(pf_src_tree, pf_src_node, entry, pf_src_compare);
entry             379 net/pf_if.c    	TAILQ_INSERT_TAIL(&dyn->pfid_kif->pfik_dynaddrs, dyn, entry);
entry             404 net/pf_if.c    	TAILQ_FOREACH(p, &kif->pfik_dynaddrs, entry)
entry             568 net/pf_if.c    	TAILQ_REMOVE(&aw->p.dyn->pfid_kif->pfik_dynaddrs, aw->p.dyn, entry);
entry             265 net/pf_osfp.c  	struct pf_osfp_entry *entry;
entry             276 net/pf_osfp.c  	SLIST_FOREACH(entry, list, fp_entry) {
entry             277 net/pf_osfp.c  		PF_OSFP_UNPACK(entry->fp_os, en_class, en_version, en_subtype);
entry             282 net/pf_osfp.c  			    entry->fp_class_nm, entry->fp_version_nm,
entry             283 net/pf_osfp.c  			    entry->fp_subtype_nm, os, entry->fp_os);
entry             307 net/pf_osfp.c  	struct pf_osfp_entry *entry;
entry             311 net/pf_osfp.c  		while ((entry = SLIST_FIRST(&fp->fp_oses))) {
entry             313 net/pf_osfp.c  			pool_put(&pf_osfp_entry_pl, entry);
entry             325 net/pf_osfp.c  	struct pf_osfp_entry *entry;
entry             363 net/pf_osfp.c  		 SLIST_FOREACH(entry, &fp->fp_oses, fp_entry) {
entry             364 net/pf_osfp.c  			if (PF_OSFP_ENTRY_EQ(entry, &fpioc->fp_os))
entry             367 net/pf_osfp.c  		if ((entry = pool_get(&pf_osfp_entry_pl, PR_NOWAIT)) == NULL)
entry             382 net/pf_osfp.c  		if ((entry = pool_get(&pf_osfp_entry_pl, PR_NOWAIT)) == NULL) {
entry             388 net/pf_osfp.c  	memcpy(entry, &fpioc->fp_os, sizeof(*entry));
entry             391 net/pf_osfp.c  	entry->fp_class_nm[sizeof(entry->fp_class_nm)-1] = '\0';
entry             392 net/pf_osfp.c  	entry->fp_version_nm[sizeof(entry->fp_version_nm)-1] = '\0';
entry             393 net/pf_osfp.c  	entry->fp_subtype_nm[sizeof(entry->fp_subtype_nm)-1] = '\0';
entry             395 net/pf_osfp.c  	SLIST_INSERT_HEAD(&fp->fp_oses, entry, fp_entry);
entry             521 net/pf_osfp.c  	struct pf_osfp_entry *entry;
entry             528 net/pf_osfp.c  		SLIST_FOREACH(entry, &fp->fp_oses, fp_entry) {
entry             537 net/pf_osfp.c  				memcpy(&fpioc->fp_os, entry,
entry             172 net/pfvar.h    	TAILQ_ENTRY(pfi_dynaddr)	 entry;
entry             633 net/pfvar.h    	RB_ENTRY(pf_src_node) entry;
entry            1538 net/pfvar.h    RB_PROTOTYPE(pf_src_tree, pf_src_node, entry, pf_src_compare);
entry              56 sys/exec_ecoff.h 	u_long  entry;
entry             179 sys/lkm.h      	int	(*entry)(struct lkm_table *, int, int);	/* entry function */
entry             296 uvm/uvm_amap.c amap_extend(struct vm_map_entry *entry, vsize_t addsize)
entry             298 uvm/uvm_amap.c 	struct vm_amap *amap = entry->aref.ar_amap;
entry             299 uvm/uvm_amap.c 	int slotoff = entry->aref.ar_pageoff;
entry             309 uvm/uvm_amap.c 	UVMHIST_LOG(maphist, "  (entry=%p, addsize=%lu)", entry, addsize, 0, 0);
entry             317 uvm/uvm_amap.c 	AMAP_B2SLOT(slotmapped, entry->end - entry->start); /* slots mapped */
entry             475 uvm/uvm_amap.c amap_share_protect(struct vm_map_entry *entry, vm_prot_t prot)
entry             477 uvm/uvm_amap.c 	struct vm_amap *amap = entry->aref.ar_amap;
entry             480 uvm/uvm_amap.c 	AMAP_B2SLOT(slots, (entry->end - entry->start));
entry             481 uvm/uvm_amap.c 	stop = entry->aref.ar_pageoff + slots;
entry             485 uvm/uvm_amap.c 		for (lcv = entry->aref.ar_pageoff ; lcv < stop ; lcv++) {
entry             498 uvm/uvm_amap.c 		if (slot < entry->aref.ar_pageoff || slot >= stop)
entry             581 uvm/uvm_amap.c amap_copy(struct vm_map *map, struct vm_map_entry *entry, int waitf,
entry             589 uvm/uvm_amap.c 		    map, entry, waitf, 0);
entry             595 uvm/uvm_amap.c 	if (entry->aref.ar_amap == NULL) {
entry             603 uvm/uvm_amap.c 		if (canchunk && atop(entry->end - entry->start) >=
entry             611 uvm/uvm_amap.c 			    entry->start, entry->end, startva, endva);
entry             612 uvm/uvm_amap.c 			UVM_MAP_CLIP_START(map, entry, startva);
entry             615 uvm/uvm_amap.c 				UVM_MAP_CLIP_END(map, entry, endva);
entry             619 uvm/uvm_amap.c 		    entry->start, entry->end, 0, 0);
entry             620 uvm/uvm_amap.c 		entry->aref.ar_pageoff = 0;
entry             621 uvm/uvm_amap.c 		entry->aref.ar_amap = amap_alloc(entry->end - entry->start, 0,
entry             623 uvm/uvm_amap.c 		if (entry->aref.ar_amap != NULL)
entry             624 uvm/uvm_amap.c 			entry->etype &= ~UVM_ET_NEEDSCOPY;
entry             638 uvm/uvm_amap.c 	if (entry->aref.ar_amap->am_ref == 1) {
entry             639 uvm/uvm_amap.c 		entry->etype &= ~UVM_ET_NEEDSCOPY;
entry             650 uvm/uvm_amap.c 	    entry->aref.ar_amap, entry->aref.ar_amap->am_ref, 0, 0);
entry             651 uvm/uvm_amap.c 	AMAP_B2SLOT(slots, entry->end - entry->start);
entry             657 uvm/uvm_amap.c 	srcamap = entry->aref.ar_amap;
entry             668 uvm/uvm_amap.c 		entry->etype &= ~UVM_ET_NEEDSCOPY;
entry             681 uvm/uvm_amap.c 		    srcamap->am_anon[entry->aref.ar_pageoff + lcv];
entry             706 uvm/uvm_amap.c 		amap_pp_adjref(srcamap, entry->aref.ar_pageoff, 
entry             707 uvm/uvm_amap.c 		    (entry->end - entry->start) >> PAGE_SHIFT, -1);
entry             715 uvm/uvm_amap.c 	entry->aref.ar_pageoff = 0;
entry             716 uvm/uvm_amap.c 	entry->aref.ar_amap = amap;
entry             717 uvm/uvm_amap.c 	entry->etype &= ~UVM_ET_NEEDSCOPY;
entry             750 uvm/uvm_amap.c amap_cow_now(struct vm_map *map, struct vm_map_entry *entry)
entry             752 uvm/uvm_amap.c 	struct vm_amap *amap = entry->aref.ar_amap;
entry             382 uvm/uvm_device.c 	struct vm_map_entry *entry = ufi->entry;
entry             383 uvm/uvm_device.c 	struct uvm_object *uobj = entry->object.uvm_obj;
entry             400 uvm/uvm_device.c 	if (UVM_ET_ISCOPYONWRITE(entry)) {
entry             402 uvm/uvm_device.c 		    entry->etype, 0,0,0);
entry             403 uvm/uvm_device.c 		uvmfault_unlockall(ufi, ufi->entry->aref.ar_amap, uobj, NULL);
entry             422 uvm/uvm_device.c 	curr_offset = entry->offset + (vaddr - entry->start);
entry             445 uvm/uvm_device.c 		mapprot = ufi->entry->protection;
entry             461 uvm/uvm_device.c 			uvmfault_unlockall(ufi, ufi->entry->aref.ar_amap,
entry             471 uvm/uvm_device.c 	uvmfault_unlockall(ufi, ufi->entry->aref.ar_amap, uobj, NULL);
entry             254 uvm/uvm_fault.c 		if (UVM_ET_ISNEEDSCOPY(ufi->entry))
entry             255 uvm/uvm_fault.c 			amap_copy(ufi->map, ufi->entry, M_NOWAIT, TRUE, 
entry             262 uvm/uvm_fault.c 		if (UVM_ET_ISNEEDSCOPY(ufi->entry)) {
entry             520 uvm/uvm_fault.c 		    amap_lookup(&ufi->entry->aref, 
entry             521 uvm/uvm_fault.c 				ufi->orig_rvaddr - ufi->entry->start) != anon) {
entry             555 uvm/uvm_fault.c #define MASK(entry)     (UVM_ET_ISCOPYONWRITE(entry) ? \
entry             624 uvm/uvm_fault.c 	if ((ufi.entry->protection & access_type) != access_type) {
entry             627 uvm/uvm_fault.c 		    ufi.entry->protection, access_type, 0, 0);
entry             639 uvm/uvm_fault.c 	enter_prot = ufi.entry->protection;
entry             640 uvm/uvm_fault.c 	wired = VM_MAPENT_ISWIRED(ufi.entry) || (fault_type == VM_FAULT_WIRE);
entry             651 uvm/uvm_fault.c 	if (UVM_ET_ISNEEDSCOPY(ufi.entry)) {
entry             653 uvm/uvm_fault.c 		    (ufi.entry->object.uvm_obj == NULL)) {
entry             677 uvm/uvm_fault.c 	amap = ufi.entry->aref.ar_amap;		/* top layer */
entry             678 uvm/uvm_fault.c 	uobj = ufi.entry->object.uvm_obj;	/* bottom layer */
entry             701 uvm/uvm_fault.c 		KASSERT(uvmadvice[ufi.entry->advice].advice ==
entry             702 uvm/uvm_fault.c 			 ufi.entry->advice);
entry             703 uvm/uvm_fault.c 		nback = min(uvmadvice[ufi.entry->advice].nback,
entry             704 uvm/uvm_fault.c 			    (ufi.orig_rvaddr - ufi.entry->start) >> PAGE_SHIFT);
entry             706 uvm/uvm_fault.c 		nforw = min(uvmadvice[ufi.entry->advice].nforw,
entry             707 uvm/uvm_fault.c 			    ((ufi.entry->end - ufi.orig_rvaddr) >>
entry             731 uvm/uvm_fault.c 	UVMHIST_LOG(maphist, "  entry=%p, amap=%p, obj=%p", ufi.entry,
entry             740 uvm/uvm_fault.c 		amap_lookups(&ufi.entry->aref, startva - ufi.entry->start,
entry             753 uvm/uvm_fault.c 	if (ufi.entry->advice == MADV_SEQUENTIAL && nback != 0) {
entry             763 uvm/uvm_fault.c 			uoff = (startva - ufi.entry->start) + ufi.entry->offset;
entry             841 uvm/uvm_fault.c 			     (VM_MAPENT_ISWIRED(ufi.entry) ? PMAP_WIRED : 0));
entry             906 uvm/uvm_fault.c 		(void) uobj->pgops->pgo_get(uobj, ufi.entry->offset +
entry             907 uvm/uvm_fault.c 				(startva - ufi.entry->start),
entry             909 uvm/uvm_fault.c 				access_type & MASK(ufi.entry),
entry             910 uvm/uvm_fault.c 				ufi.entry->advice, PGO_LOCKED);
entry             971 uvm/uvm_fault.c 				    enter_prot & MASK(ufi.entry),
entry            1202 uvm/uvm_fault.c 		amap_add(&ufi.entry->aref, ufi.orig_rvaddr - ufi.entry->start,
entry            1314 uvm/uvm_fault.c 		     UVM_ET_ISCOPYONWRITE(ufi.entry);
entry            1341 uvm/uvm_fault.c 		uoff = (ufi.orig_rvaddr - ufi.entry->start) + ufi.entry->offset;
entry            1343 uvm/uvm_fault.c 		    0, access_type & MASK(ufi.entry), ufi.entry->advice,
entry            1388 uvm/uvm_fault.c 		    amap_lookup(&ufi.entry->aref,
entry            1389 uvm/uvm_fault.c 		      ufi.orig_rvaddr - ufi.entry->start))) {
entry            1466 uvm/uvm_fault.c 		if (UVM_ET_ISCOPYONWRITE(ufi.entry))
entry            1668 uvm/uvm_fault.c 		amap_add(&ufi.entry->aref, ufi.orig_rvaddr - ufi.entry->start,
entry            1827 uvm/uvm_fault.c 	vm_map_entry_t entry;
entry            1848 uvm/uvm_fault.c 	if (uvm_map_lookup_entry(map, start, &entry) == FALSE)
entry            1858 uvm/uvm_fault.c 		KASSERT(va >= entry->start);
entry            1859 uvm/uvm_fault.c 		while (va >= entry->end) {
entry            1860 uvm/uvm_fault.c 			KASSERT(entry->next != &map->header &&
entry            1861 uvm/uvm_fault.c 				entry->next->start <= entry->end);
entry            1862 uvm/uvm_fault.c 			entry = entry->next;
entry            1868 uvm/uvm_fault.c 		if (VM_MAPENT_ISWIRED(entry) == 0)
entry              65 uvm/uvm_fault.h 	vm_map_entry_t entry;		/* map entry (from 'map') */
entry             145 uvm/uvm_fault_i.h 								&ufi->entry)) {
entry             153 uvm/uvm_fault_i.h 		if (ufi->entry->end - ufi->orig_rvaddr < ufi->size)
entry             154 uvm/uvm_fault_i.h 			ufi->size = ufi->entry->end - ufi->orig_rvaddr;
entry             160 uvm/uvm_fault_i.h 		if (UVM_ET_ISSUBMAP(ufi->entry)) {
entry             161 uvm/uvm_fault_i.h 			tmpmap = ufi->entry->object.sub_map;
entry             137 uvm/uvm_loan.c 	struct vm_aref *aref = &ufi->entry->aref;
entry             138 uvm/uvm_loan.c 	struct uvm_object *uobj = ufi->entry->object.uvm_obj;
entry             158 uvm/uvm_loan.c 			anon = amap_lookup(aref, curaddr - ufi->entry->start);
entry             167 uvm/uvm_loan.c 		} else if (UVM_ET_ISCOPYONWRITE(ufi->entry)) {
entry             353 uvm/uvm_loan.c 	result = uvmfault_anonget(ufi, ufi->entry->aref.ar_amap, anon);
entry             414 uvm/uvm_loan.c 	struct vm_amap *amap = ufi->entry->aref.ar_amap;
entry             415 uvm/uvm_loan.c 	struct uvm_object *uobj = ufi->entry->object.uvm_obj;
entry             430 uvm/uvm_loan.c 		result = uobj->pgops->pgo_get(uobj, va - ufi->entry->start,
entry             455 uvm/uvm_loan.c 		result = uobj->pgops->pgo_get(uobj, va - ufi->entry->start,
entry             485 uvm/uvm_loan.c 		    (locked && amap && amap_lookup(&ufi->entry->aref,
entry             486 uvm/uvm_loan.c 		    ufi->orig_rvaddr - ufi->entry->start))) {
entry             623 uvm/uvm_loan.c 			uvmfault_unlockall(ufi, ufi->entry->aref.ar_amap, 
entry             624 uvm/uvm_loan.c 			    ufi->entry->object.uvm_obj, NULL);
entry             628 uvm/uvm_loan.c 			if (ufi->entry->object.uvm_obj)
entry             630 uvm/uvm_loan.c 				    &ufi->entry->object.uvm_obj->vmobjlock);
entry             652 uvm/uvm_loan.c 		uvmfault_unlockall(ufi, ufi->entry->aref.ar_amap,
entry             653 uvm/uvm_loan.c 		       ufi->entry->object.uvm_obj, NULL);
entry             667 uvm/uvm_loan.c 		if (ufi->entry->object.uvm_obj)
entry             668 uvm/uvm_loan.c 			simple_lock(&ufi->entry->object.uvm_obj->vmobjlock);
entry             143 uvm/uvm_map.c  #define uvm_map_entry_link(map, after_where, entry) do { \
entry             145 uvm/uvm_map.c  	(entry)->prev = (after_where); \
entry             146 uvm/uvm_map.c  	(entry)->next = (after_where)->next; \
entry             147 uvm/uvm_map.c  	(entry)->prev->next = (entry); \
entry             148 uvm/uvm_map.c  	(entry)->next->prev = (entry); \
entry             149 uvm/uvm_map.c  	uvm_rb_insert(map, entry); \
entry             157 uvm/uvm_map.c  #define uvm_map_entry_unlink(map, entry) do { \
entry             159 uvm/uvm_map.c  	(entry)->next->prev = (entry)->prev; \
entry             160 uvm/uvm_map.c  	(entry)->prev->next = (entry)->next; \
entry             161 uvm/uvm_map.c  	uvm_rb_remove(map, entry); \
entry             232 uvm/uvm_map.c  uvm_rb_augment(struct vm_map_entry *entry)
entry             234 uvm/uvm_map.c  	entry->space = uvm_rb_subtree_space(entry);
entry             242 uvm/uvm_map.c  uvm_rb_space(struct vm_map *map, struct vm_map_entry *entry)
entry             247 uvm/uvm_map.c  	if ((next = entry->next) == &map->header)
entry             248 uvm/uvm_map.c  		space = map->max_offset - entry->end;
entry             251 uvm/uvm_map.c  		space = next->start - entry->end;
entry             257 uvm/uvm_map.c  uvm_rb_subtree_space(struct vm_map_entry *entry)
entry             261 uvm/uvm_map.c  	space = entry->ownspace;
entry             262 uvm/uvm_map.c  	if (RB_LEFT(entry, rb_entry)) {
entry             263 uvm/uvm_map.c  		tmp = RB_LEFT(entry, rb_entry)->space;
entry             268 uvm/uvm_map.c  	if (RB_RIGHT(entry, rb_entry)) {
entry             269 uvm/uvm_map.c  		tmp = RB_RIGHT(entry, rb_entry)->space;
entry             278 uvm/uvm_map.c  uvm_rb_fixup(struct vm_map *map, struct vm_map_entry *entry)
entry             282 uvm/uvm_map.c  		entry->ownspace = uvm_rb_space(map, entry);
entry             283 uvm/uvm_map.c  		entry->space = uvm_rb_subtree_space(entry);
entry             284 uvm/uvm_map.c  	} while ((entry = RB_PARENT(entry, rb_entry)) != NULL);
entry             288 uvm/uvm_map.c  uvm_rb_insert(struct vm_map *map, struct vm_map_entry *entry)
entry             290 uvm/uvm_map.c  	vaddr_t space = uvm_rb_space(map, entry);
entry             293 uvm/uvm_map.c  	entry->ownspace = entry->space = space;
entry             294 uvm/uvm_map.c  	tmp = RB_INSERT(uvm_tree, &(map)->rbhead, entry);
entry             299 uvm/uvm_map.c  	uvm_rb_fixup(map, entry);
entry             300 uvm/uvm_map.c  	if (entry->prev != &map->header)
entry             301 uvm/uvm_map.c  		uvm_rb_fixup(map, entry->prev);
entry             305 uvm/uvm_map.c  uvm_rb_remove(struct vm_map *map, struct vm_map_entry *entry)
entry             309 uvm/uvm_map.c  	parent = RB_PARENT(entry, rb_entry);
entry             310 uvm/uvm_map.c  	RB_REMOVE(uvm_tree, &(map)->rbhead, entry);
entry             311 uvm/uvm_map.c  	if (entry->prev != &map->header)
entry             312 uvm/uvm_map.c  		uvm_rb_fixup(map, entry->prev);
entry             482 uvm/uvm_map.c  uvm_map_entry_unwire(struct vm_map *map, struct vm_map_entry *entry)
entry             485 uvm/uvm_map.c  	entry->wired_count = 0;
entry             486 uvm/uvm_map.c  	uvm_fault_unwire_locked(map, entry->start, entry->end);
entry             494 uvm/uvm_map.c  uvm_map_reference_amap(struct vm_map_entry *entry, int flags)
entry             496 uvm/uvm_map.c  	amap_ref(entry->aref.ar_amap, entry->aref.ar_pageoff,
entry             497 uvm/uvm_map.c  	    (entry->end - entry->start) >> PAGE_SHIFT, flags);
entry             505 uvm/uvm_map.c  uvm_map_unreference_amap(struct vm_map_entry *entry, int flags)
entry             507 uvm/uvm_map.c  	amap_unref(entry->aref.ar_amap, entry->aref.ar_pageoff,
entry             508 uvm/uvm_map.c  	    (entry->end - entry->start) >> PAGE_SHIFT, flags);
entry             581 uvm/uvm_map.c  uvm_map_clip_start(struct vm_map *map, struct vm_map_entry *entry,
entry             598 uvm/uvm_map.c  	uvm_mapent_copy(entry, new_entry); /* entry -> new_entry */
entry             602 uvm/uvm_map.c  	if (entry->object.uvm_obj)
entry             603 uvm/uvm_map.c  		entry->offset += new_adj;	/* shift start over */
entry             606 uvm/uvm_map.c  	entry->start = start;
entry             609 uvm/uvm_map.c  		amap_splitref(&new_entry->aref, &entry->aref, new_adj);
entry             612 uvm/uvm_map.c  	uvm_map_entry_link(map, entry->prev, new_entry);
entry             614 uvm/uvm_map.c  	if (UVM_ET_ISSUBMAP(entry)) {
entry             618 uvm/uvm_map.c  		if (UVM_ET_ISOBJ(entry) && 
entry             619 uvm/uvm_map.c  		    entry->object.uvm_obj->pgops &&
entry             620 uvm/uvm_map.c  		    entry->object.uvm_obj->pgops->pgo_reference)
entry             621 uvm/uvm_map.c  			entry->object.uvm_obj->pgops->pgo_reference(
entry             622 uvm/uvm_map.c  			    entry->object.uvm_obj);
entry             638 uvm/uvm_map.c  uvm_map_clip_end(struct vm_map *map, struct vm_map_entry *entry, vaddr_t end)
entry             650 uvm/uvm_map.c  	uvm_mapent_copy(entry, new_entry); /* entry -> new_entry */
entry             652 uvm/uvm_map.c  	new_entry->start = entry->end = end;
entry             653 uvm/uvm_map.c  	new_adj = end - entry->start;
entry             657 uvm/uvm_map.c  	if (entry->aref.ar_amap)
entry             658 uvm/uvm_map.c  		amap_splitref(&entry->aref, &new_entry->aref, new_adj);
entry             660 uvm/uvm_map.c  	uvm_rb_fixup(map, entry);
entry             662 uvm/uvm_map.c  	uvm_map_entry_link(map, entry, new_entry);
entry             664 uvm/uvm_map.c  	if (UVM_ET_ISSUBMAP(entry)) {
entry             668 uvm/uvm_map.c  		if (UVM_ET_ISOBJ(entry) &&
entry             669 uvm/uvm_map.c  		    entry->object.uvm_obj->pgops &&
entry             670 uvm/uvm_map.c  		    entry->object.uvm_obj->pgops->pgo_reference)
entry             671 uvm/uvm_map.c  			entry->object.uvm_obj->pgops->pgo_reference(
entry             672 uvm/uvm_map.c  			    entry->object.uvm_obj);
entry             953 uvm/uvm_map.c      struct vm_map_entry **entry)
entry             962 uvm/uvm_map.c  	    map, address, entry, 0);
entry             992 uvm/uvm_map.c  			*entry = cur;
entry            1022 uvm/uvm_map.c  					*entry = cur;
entry            1031 uvm/uvm_map.c  		*entry = prev;
entry            1048 uvm/uvm_map.c  				*entry = cur;
entry            1059 uvm/uvm_map.c  	*entry = cur->prev;
entry            1060 uvm/uvm_map.c  	SAVE_HINT(map, map->hint, *entry);
entry            1149 uvm/uvm_map.c  	struct vm_map_entry *entry, *next, *tmp;
entry            1189 uvm/uvm_map.c  		if ((entry = map->first_free) != &map->header) 
entry            1190 uvm/uvm_map.c  			hint = entry->end;
entry            1201 uvm/uvm_map.c  		entry = tmp;
entry            1210 uvm/uvm_map.c  		next = entry->next;
entry            1220 uvm/uvm_map.c  	if (uvm_map_spacefits(map, &hint, length, entry->next, uoffset, align))
entry            1261 uvm/uvm_map.c  			entry = tmp;
entry            1272 uvm/uvm_map.c  		entry = prev;
entry            1292 uvm/uvm_map.c  		entry = tmp;
entry            1307 uvm/uvm_map.c  	for (;; hint = (entry = next)->end) {
entry            1336 uvm/uvm_map.c  		next = entry->next;
entry            1341 uvm/uvm_map.c  	SAVE_HINT(map, map->hint, entry);
entry            1344 uvm/uvm_map.c  	return (entry);
entry            1374 uvm/uvm_map.c  	struct vm_map_entry *entry, *first_entry, *next;
entry            1394 uvm/uvm_map.c  		entry = first_entry;
entry            1395 uvm/uvm_map.c  		UVM_MAP_CLIP_START(map, entry, start);
entry            1397 uvm/uvm_map.c  		SAVE_HINT(map, entry, entry->prev);
entry            1400 uvm/uvm_map.c  		entry = first_entry->next;
entry            1408 uvm/uvm_map.c  		map->first_free = entry->prev;
entry            1434 uvm/uvm_map.c  	while ((entry != &map->header) && (entry->start < end)) {
entry            1436 uvm/uvm_map.c  		UVM_MAP_CLIP_END(map, entry, end); 
entry            1437 uvm/uvm_map.c  		next = entry->next;
entry            1438 uvm/uvm_map.c  		len = entry->end - entry->start;
entry            1439 uvm/uvm_map.c  		if (p && entry->object.uvm_obj == NULL)
entry            1447 uvm/uvm_map.c  		if (VM_MAPENT_ISWIRED(entry))
entry            1448 uvm/uvm_map.c  			uvm_map_entry_unwire(map, entry);
entry            1455 uvm/uvm_map.c  			uvm_km_pgremove_intrsafe(entry->start, entry->end);
entry            1456 uvm/uvm_map.c  			pmap_kremove(entry->start, len);
entry            1457 uvm/uvm_map.c  		} else if (UVM_ET_ISOBJ(entry) &&
entry            1458 uvm/uvm_map.c  		    UVM_OBJ_IS_KERN_OBJECT(entry->object.uvm_obj)) {
entry            1496 uvm/uvm_map.c  			pmap_remove(pmap_kernel(), entry->start, entry->end);
entry            1497 uvm/uvm_map.c  			uvm_km_pgremove(entry->object.uvm_obj,
entry            1498 uvm/uvm_map.c  			    entry->start - vm_map_min(kernel_map),
entry            1499 uvm/uvm_map.c  			    entry->end - vm_map_min(kernel_map));
entry            1505 uvm/uvm_map.c  			entry->etype &= ~UVM_ET_OBJ;
entry            1506 uvm/uvm_map.c  			entry->object.uvm_obj = NULL;	/* to be safe */
entry            1512 uvm/uvm_map.c  			pmap_remove(map->pmap, entry->start, entry->end);
entry            1519 uvm/uvm_map.c  		UVMHIST_LOG(maphist, "  removed map entry %p", entry, 0, 0,0);
entry            1522 uvm/uvm_map.c  		SAVE_HINT(map, entry, entry->prev);
entry            1524 uvm/uvm_map.c  		uvm_map_entry_unlink(map, entry);
entry            1526 uvm/uvm_map.c  		entry->next = first_entry;
entry            1527 uvm/uvm_map.c  		first_entry = entry;
entry            1528 uvm/uvm_map.c  		entry = next;		/* next entry, please */
entry            1785 uvm/uvm_map.c  	struct vm_map_entry *chain, *endchain, *entry, *orig_entry, *newentry;
entry            1830 uvm/uvm_map.c  	if (uvm_map_lookup_entry(srcmap, start, &entry)) {
entry            1843 uvm/uvm_map.c  			fudge = start - entry->start;
entry            1851 uvm/uvm_map.c  			UVM_MAP_CLIP_START(srcmap, entry, start);
entry            1852 uvm/uvm_map.c  			SAVE_HINT(srcmap, srcmap->hint, entry->prev);
entry            1863 uvm/uvm_map.c  		entry = entry->next;
entry            1868 uvm/uvm_map.c  	orig_entry = entry;
entry            1876 uvm/uvm_map.c  	while (entry->start < end && entry != &srcmap->header) {
entry            1880 uvm/uvm_map.c  			UVM_MAP_CLIP_END(srcmap, entry, end);
entry            1883 uvm/uvm_map.c  		if (UVM_ET_ISNEEDSCOPY(entry)) {
entry            1885 uvm/uvm_map.c  				oldstart = entry->start;
entry            1888 uvm/uvm_map.c  			amap_copy(srcmap, entry, M_NOWAIT, TRUE, start, end);
entry            1889 uvm/uvm_map.c  			if (UVM_ET_ISNEEDSCOPY(entry)) {  /* failed? */
entry            1896 uvm/uvm_map.c  				fudge = fudge - (entry->start - oldstart);
entry            1902 uvm/uvm_map.c  		oldoffset = (entry->start + fudge) - start;
entry            1916 uvm/uvm_map.c  		    newentry->start + (entry->end - (entry->start + fudge));
entry            1919 uvm/uvm_map.c  		newentry->object.uvm_obj = entry->object.uvm_obj;
entry            1924 uvm/uvm_map.c  			newentry->offset = entry->offset + fudge;
entry            1928 uvm/uvm_map.c  		newentry->etype = entry->etype;
entry            1930 uvm/uvm_map.c  			entry->max_protection : entry->protection; 
entry            1931 uvm/uvm_map.c  		newentry->max_protection = entry->max_protection;
entry            1932 uvm/uvm_map.c  		newentry->inheritance = entry->inheritance;
entry            1934 uvm/uvm_map.c  		newentry->aref.ar_amap = entry->aref.ar_amap;
entry            1937 uvm/uvm_map.c  			    entry->aref.ar_pageoff + (fudge >> PAGE_SHIFT);
entry            1943 uvm/uvm_map.c  		newentry->advice = entry->advice;
entry            1955 uvm/uvm_map.c  		if ((flags & UVM_EXTRACT_CONTIG) && entry->end < end && 
entry            1956 uvm/uvm_map.c  		    (entry->next == &srcmap->header ||
entry            1957 uvm/uvm_map.c  		    entry->next->start != entry->end)) {
entry            1961 uvm/uvm_map.c  		entry = entry->next;
entry            2009 uvm/uvm_map.c  		entry = orig_entry;
entry            2013 uvm/uvm_map.c  		while (entry->start < end && entry != &srcmap->header) {
entry            2015 uvm/uvm_map.c  				oldoffset = (entry->start + fudge) - start;
entry            2016 uvm/uvm_map.c  				elen = MIN(end, entry->end) -
entry            2017 uvm/uvm_map.c  				    (entry->start + fudge);
entry            2020 uvm/uvm_map.c  				    entry->start + fudge);
entry            2025 uvm/uvm_map.c  				pmap_remove(srcmap->pmap, entry->start, 
entry            2026 uvm/uvm_map.c  						entry->end);
entry            2027 uvm/uvm_map.c          			oldentry = entry;	/* save entry */
entry            2028 uvm/uvm_map.c          			entry = entry->next;	/* advance */
entry            2034 uvm/uvm_map.c          			entry = entry->next;		/* advance */
entry            2122 uvm/uvm_map.c  	struct vm_map_entry *entry;
entry            2129 uvm/uvm_map.c  	if (uvm_map_lookup_entry(map, start, &entry)) {
entry            2130 uvm/uvm_map.c  		UVM_MAP_CLIP_START(map, entry, start);
entry            2131 uvm/uvm_map.c  		UVM_MAP_CLIP_END(map, entry, end);		/* to be safe */
entry            2133 uvm/uvm_map.c  		entry = NULL;
entry            2136 uvm/uvm_map.c  	if (entry != NULL && 
entry            2137 uvm/uvm_map.c  	    entry->start == start && entry->end == end &&
entry            2138 uvm/uvm_map.c  	    entry->object.uvm_obj == NULL && entry->aref.ar_amap == NULL &&
entry            2139 uvm/uvm_map.c  	    !UVM_ET_ISCOPYONWRITE(entry) && !UVM_ET_ISNEEDSCOPY(entry)) {
entry            2140 uvm/uvm_map.c  		entry->etype |= UVM_ET_SUBMAP;
entry            2141 uvm/uvm_map.c  		entry->object.sub_map = submap;
entry            2142 uvm/uvm_map.c  		entry->offset = 0;
entry            2160 uvm/uvm_map.c  #define MASK(entry)     (UVM_ET_ISCOPYONWRITE(entry) ? \
entry            2168 uvm/uvm_map.c  	struct vm_map_entry *current, *entry;
entry            2178 uvm/uvm_map.c  	if (uvm_map_lookup_entry(map, start, &entry)) {
entry            2179 uvm/uvm_map.c  		UVM_MAP_CLIP_START(map, entry, start);
entry            2181 uvm/uvm_map.c  		entry = entry->next;
entry            2188 uvm/uvm_map.c  	current = entry;
entry            2203 uvm/uvm_map.c  	current = entry;
entry            2224 uvm/uvm_map.c  			if ((current->protection & MASK(entry)) == PROT_NONE &&
entry            2225 uvm/uvm_map.c  			    VM_MAPENT_ISWIRED(entry))
entry            2228 uvm/uvm_map.c  			    current->protection & MASK(entry));
entry            2238 uvm/uvm_map.c  		    VM_MAPENT_ISWIRED(entry) == 0 &&
entry            2241 uvm/uvm_map.c  			if (uvm_map_pageable(map, entry->start, entry->end,
entry            2285 uvm/uvm_map.c  	struct vm_map_entry *entry, *temp_entry;
entry            2305 uvm/uvm_map.c  		entry = temp_entry;
entry            2306 uvm/uvm_map.c  		UVM_MAP_CLIP_START(map, entry, start);
entry            2308 uvm/uvm_map.c  		entry = temp_entry->next;
entry            2311 uvm/uvm_map.c  	while ((entry != &map->header) && (entry->start < end)) {
entry            2312 uvm/uvm_map.c  		UVM_MAP_CLIP_END(map, entry, end);
entry            2313 uvm/uvm_map.c  		entry->inheritance = new_inheritance;
entry            2314 uvm/uvm_map.c  		entry = entry->next;
entry            2331 uvm/uvm_map.c  	struct vm_map_entry *entry, *temp_entry;
entry            2339 uvm/uvm_map.c  		entry = temp_entry;
entry            2340 uvm/uvm_map.c  		UVM_MAP_CLIP_START(map, entry, start);
entry            2342 uvm/uvm_map.c  		entry = temp_entry->next;
entry            2349 uvm/uvm_map.c  	while ((entry != &map->header) && (entry->start < end)) {
entry            2350 uvm/uvm_map.c  		UVM_MAP_CLIP_END(map, entry, end);
entry            2364 uvm/uvm_map.c  		entry->advice = new_advice;
entry            2365 uvm/uvm_map.c  		entry = entry->next;
entry            2391 uvm/uvm_map.c  	struct vm_map_entry *entry, *start_entry, *failed_entry;
entry            2421 uvm/uvm_map.c  	entry = start_entry;
entry            2428 uvm/uvm_map.c  		UVM_MAP_CLIP_START(map, entry, start);
entry            2435 uvm/uvm_map.c  		while ((entry != &map->header) && (entry->start < end)) {
entry            2436 uvm/uvm_map.c  			if (entry->wired_count == 0 ||
entry            2437 uvm/uvm_map.c  			    (entry->end < end &&
entry            2438 uvm/uvm_map.c  			     (entry->next == &map->header ||
entry            2439 uvm/uvm_map.c  			      entry->next->start > entry->end))) {
entry            2446 uvm/uvm_map.c  			entry = entry->next;
entry            2455 uvm/uvm_map.c  		entry = start_entry;
entry            2456 uvm/uvm_map.c  		while ((entry != &map->header) && (entry->start < end)) {
entry            2457 uvm/uvm_map.c  			UVM_MAP_CLIP_END(map, entry, end);
entry            2458 uvm/uvm_map.c  			if (VM_MAPENT_ISWIRED(entry))
entry            2459 uvm/uvm_map.c  				uvm_map_entry_unwire(map, entry);
entry            2460 uvm/uvm_map.c  			entry = entry->next;
entry            2489 uvm/uvm_map.c  	while ((entry != &map->header) && (entry->start < end)) {
entry            2490 uvm/uvm_map.c  		if (VM_MAPENT_ISWIRED(entry) == 0) { /* not already wired? */
entry            2500 uvm/uvm_map.c  			if (!UVM_ET_ISSUBMAP(entry)) {  /* not submap */
entry            2501 uvm/uvm_map.c  				if (UVM_ET_ISNEEDSCOPY(entry) && 
entry            2502 uvm/uvm_map.c  				    ((entry->protection & VM_PROT_WRITE) ||
entry            2503 uvm/uvm_map.c  				     (entry->object.uvm_obj == NULL))) {
entry            2504 uvm/uvm_map.c  					amap_copy(map, entry, M_WAITOK, TRUE,
entry            2510 uvm/uvm_map.c  		UVM_MAP_CLIP_START(map, entry, start);
entry            2511 uvm/uvm_map.c  		UVM_MAP_CLIP_END(map, entry, end);
entry            2512 uvm/uvm_map.c  		entry->wired_count++;
entry            2518 uvm/uvm_map.c  		if (entry->protection == VM_PROT_NONE ||
entry            2519 uvm/uvm_map.c  		    (entry->end < end &&
entry            2520 uvm/uvm_map.c  		     (entry->next == &map->header ||
entry            2521 uvm/uvm_map.c  		      entry->next->start > entry->end))) {
entry            2528 uvm/uvm_map.c  			while (entry != &map->header && entry->end > start) {
entry            2529 uvm/uvm_map.c  				entry->wired_count--;
entry            2530 uvm/uvm_map.c  				entry = entry->prev;
entry            2537 uvm/uvm_map.c  		entry = entry->next;
entry            2551 uvm/uvm_map.c  	entry = start_entry;
entry            2552 uvm/uvm_map.c  	while (entry != &map->header && entry->start < end) {
entry            2553 uvm/uvm_map.c  		if (entry->wired_count == 1) {
entry            2554 uvm/uvm_map.c  			rv = uvm_fault_wire(map, entry->start, entry->end,
entry            2555 uvm/uvm_map.c  			    entry->protection);
entry            2565 uvm/uvm_map.c  		entry = entry->next;
entry            2587 uvm/uvm_map.c  		failed_entry = entry;
entry            2588 uvm/uvm_map.c  		while (entry != &map->header && entry->start < end) {
entry            2589 uvm/uvm_map.c  			entry->wired_count--;
entry            2590 uvm/uvm_map.c  			entry = entry->next;
entry            2598 uvm/uvm_map.c  		entry = start_entry;
entry            2599 uvm/uvm_map.c  		while (entry != failed_entry) {
entry            2600 uvm/uvm_map.c  			entry->wired_count--;
entry            2601 uvm/uvm_map.c  			if (VM_MAPENT_ISWIRED(entry) == 0)
entry            2602 uvm/uvm_map.c  				uvm_map_entry_unwire(map, entry);
entry            2603 uvm/uvm_map.c  			entry = entry->next;
entry            2641 uvm/uvm_map.c  	struct vm_map_entry *entry, *failed_entry;
entry            2663 uvm/uvm_map.c  		for (entry = map->header.next; entry != &map->header;
entry            2664 uvm/uvm_map.c  		     entry = entry->next) {
entry            2665 uvm/uvm_map.c  			if (VM_MAPENT_ISWIRED(entry))
entry            2666 uvm/uvm_map.c  				uvm_map_entry_unwire(map, entry);
entry            2717 uvm/uvm_map.c  	for (size = 0, entry = map->header.next; entry != &map->header;
entry            2718 uvm/uvm_map.c  	     entry = entry->next) {
entry            2719 uvm/uvm_map.c  		if (entry->protection != VM_PROT_NONE &&
entry            2720 uvm/uvm_map.c  		    VM_MAPENT_ISWIRED(entry) == 0) { /* not already wired? */
entry            2721 uvm/uvm_map.c  			size += entry->end - entry->start;
entry            2743 uvm/uvm_map.c  	for (entry = map->header.next; entry != &map->header;
entry            2744 uvm/uvm_map.c  	     entry = entry->next) {
entry            2745 uvm/uvm_map.c  		if (entry->protection == VM_PROT_NONE)
entry            2747 uvm/uvm_map.c  		if (VM_MAPENT_ISWIRED(entry) == 0) { /* not already wired? */
entry            2755 uvm/uvm_map.c  			if (!UVM_ET_ISSUBMAP(entry)) {	/* not submap */
entry            2756 uvm/uvm_map.c  				if (UVM_ET_ISNEEDSCOPY(entry) && 
entry            2757 uvm/uvm_map.c  				    ((entry->protection & VM_PROT_WRITE) ||
entry            2758 uvm/uvm_map.c  				     (entry->object.uvm_obj == NULL))) {
entry            2759 uvm/uvm_map.c  					amap_copy(map, entry, M_WAITOK, TRUE,
entry            2760 uvm/uvm_map.c  					    entry->start, entry->end);
entry            2765 uvm/uvm_map.c  		entry->wired_count++;
entry            2778 uvm/uvm_map.c  	for (error = 0, entry = map->header.next;
entry            2779 uvm/uvm_map.c  	    entry != &map->header && error == 0;
entry            2780 uvm/uvm_map.c  	    entry = entry->next) {
entry            2781 uvm/uvm_map.c  		if (entry->wired_count == 1) {
entry            2782 uvm/uvm_map.c  			error = uvm_fault_wire(map, entry->start, entry->end,
entry            2783 uvm/uvm_map.c  			     entry->protection);
entry            2805 uvm/uvm_map.c  		failed_entry = entry;
entry            2806 uvm/uvm_map.c  		for (/* nothing */; entry != &map->header;
entry            2807 uvm/uvm_map.c  		     entry = entry->next) {
entry            2808 uvm/uvm_map.c  			if (entry->protection == VM_PROT_NONE)
entry            2810 uvm/uvm_map.c  			entry->wired_count--;
entry            2819 uvm/uvm_map.c  		for (entry = map->header.next; entry != failed_entry;
entry            2820 uvm/uvm_map.c  		     entry = entry->next) {
entry            2821 uvm/uvm_map.c  			if (entry->protection == VM_PROT_NONE)
entry            2823 uvm/uvm_map.c  			entry->wired_count--;
entry            2824 uvm/uvm_map.c  			if (VM_MAPENT_ISWIRED(entry))
entry            2825 uvm/uvm_map.c  				uvm_map_entry_unwire(map, entry);
entry            2861 uvm/uvm_map.c  	struct vm_map_entry *current, *entry;
entry            2878 uvm/uvm_map.c  	if (uvm_map_lookup_entry(map, start, &entry) == FALSE) {
entry            2887 uvm/uvm_map.c  	for (current = entry; current->start < end; current = current->next) {
entry            2901 uvm/uvm_map.c  	for (current = entry; current->start < end; current = current->next) {
entry            3027 uvm/uvm_map.c  		     ((entry->max_protection & VM_PROT_WRITE) != 0 &&
entry            3028 uvm/uvm_map.c  		      (entry->etype & UVM_ET_COPYONWRITE) == 0))) {
entry            3055 uvm/uvm_map.c  	struct vm_map_entry *entry;
entry            3061 uvm/uvm_map.c  	entry = tmp_entry;
entry            3063 uvm/uvm_map.c  		if (entry == &map->header) {
entry            3071 uvm/uvm_map.c  		if (start < entry->start) {
entry            3079 uvm/uvm_map.c  		if ((entry->protection & protection) != protection) {
entry            3085 uvm/uvm_map.c  		start = entry->end;
entry            3086 uvm/uvm_map.c  		entry = entry->next;
entry            3618 uvm/uvm_map.c  	struct vm_map_entry *entry;
entry            3633 uvm/uvm_map.c  	for (entry = map->header.next; entry != &map->header;
entry            3634 uvm/uvm_map.c  	    entry = entry->next) {
entry            3636 uvm/uvm_map.c  		    entry, entry->start, entry->end, entry->object.uvm_obj,
entry            3637 uvm/uvm_map.c  		    (long long)entry->offset, entry->aref.ar_amap,
entry            3638 uvm/uvm_map.c  		    entry->aref.ar_pageoff);
entry            3642 uvm/uvm_map.c  		    (entry->etype & UVM_ET_SUBMAP) ? 'T' : 'F',
entry            3643 uvm/uvm_map.c  		    (entry->etype & UVM_ET_COPYONWRITE) ? 'T' : 'F', 
entry            3644 uvm/uvm_map.c  		    (entry->etype & UVM_ET_NEEDSCOPY) ? 'T' : 'F',
entry            3645 uvm/uvm_map.c  		    entry->protection, entry->max_protection,
entry            3646 uvm/uvm_map.c  		    entry->inheritance, entry->wired_count, entry->advice);
entry             160 uvm/uvm_map.h  #define	VM_MAPENT_ISWIRED(entry)	((entry)->wired_count != 0)
entry             234 uvm/uvm_meter.c 	struct vm_map_entry *	entry;
entry             280 uvm/uvm_meter.c 		for (map = &p->p_vmspace->vm_map, entry = map->header.next;
entry             281 uvm/uvm_meter.c 		    entry != &map->header; entry = entry->next) {
entry             282 uvm/uvm_meter.c 			if (entry->is_a_map || entry->is_sub_map ||
entry             283 uvm/uvm_meter.c 			    entry->object.uvm_obj == NULL)
entry             254 uvm/uvm_mmap.c 	vm_map_entry_t entry;
entry             284 uvm/uvm_mmap.c 	if (uvm_map_lookup_entry(map, start, &entry) == FALSE) {
entry             290 uvm/uvm_mmap.c 	     entry != &map->header && entry->start < end;
entry             291 uvm/uvm_mmap.c 	     entry = entry->next) {
entry             292 uvm/uvm_mmap.c 		KASSERT(!UVM_ET_ISSUBMAP(entry));
entry             293 uvm/uvm_mmap.c 		KASSERT(start >= entry->start);
entry             296 uvm/uvm_mmap.c 		if (entry->end < end &&
entry             297 uvm/uvm_mmap.c 		     (entry->next == &map->header ||
entry             298 uvm/uvm_mmap.c 		      entry->next->start > entry->end)) {
entry             303 uvm/uvm_mmap.c 		lim = end < entry->end ? end : entry->end;
entry             309 uvm/uvm_mmap.c 		if (UVM_ET_ISOBJ(entry)) {
entry             310 uvm/uvm_mmap.c 			KASSERT(!UVM_OBJ_IS_KERN_OBJECT(entry->object.uvm_obj));
entry             311 uvm/uvm_mmap.c 			if (entry->object.uvm_obj->pgops->pgo_releasepg
entry             321 uvm/uvm_mmap.c 		amap = entry->aref.ar_amap;	/* top layer */
entry             322 uvm/uvm_mmap.c 		uobj = entry->object.uvm_obj;	/* bottom layer */
entry             331 uvm/uvm_mmap.c 				anon = amap_lookup(&entry->aref,
entry             332 uvm/uvm_mmap.c 				    start - entry->start);
entry             346 uvm/uvm_mmap.c 				    entry->offset + (start - entry->start));
entry             678 uvm/uvm_mmap.c 		vm_map_entry_t entry;
entry             681 uvm/uvm_mmap.c 		rv = uvm_map_lookup_entry(map, addr, &entry);
entry             683 uvm/uvm_mmap.c 			addr = entry->start;
entry             684 uvm/uvm_mmap.c 			size = entry->end - entry->start;
entry             311 uvm/uvm_page.h #define VM_PAGE_TO_PHYS(entry)	((entry)->phys_addr)
entry             409 uvm/uvm_page.h #define VM_PAGE_IS_FREE(entry)  ((entry)->pg_flags & PQ_FREE)
entry             188 uvm/uvm_unix.c 	vm_map_entry_t entry;
entry             196 uvm/uvm_unix.c 	for (entry = map->header.next; entry != &map->header;
entry             197 uvm/uvm_unix.c 	    entry = entry->next) {
entry             200 uvm/uvm_unix.c 		if (UVM_ET_ISSUBMAP(entry)) {
entry             204 uvm/uvm_unix.c 		if (!(entry->protection & VM_PROT_WRITE))
entry             210 uvm/uvm_unix.c 		if (entry->object.uvm_obj != NULL &&
entry             211 uvm/uvm_unix.c 		    UVM_OBJ_IS_DEVICE(entry->object.uvm_obj))
entry             214 uvm/uvm_unix.c 		start = entry->start;
entry             215 uvm/uvm_unix.c 		end = entry->end;