map               403 arch/i386/i386/ioapic.c 	struct mp_intr_map *map;
map               406 arch/i386/i386/ioapic.c 	map = pp->ip_map;
map               407 arch/i386/i386/ioapic.c 	redlo = (map == NULL) ? IOAPIC_REDLO_MASK : map->redir;
map               437 arch/i386/i386/ioapic.c 		if (map != NULL && ((map->flags & 3) == MPS_INTPO_DEF)) {
map              3651 arch/i386/i386/machdep.c 	struct i386_bus_dmamap *map;
map              3674 arch/i386/i386/machdep.c 	map = (struct i386_bus_dmamap *)mapstore;
map              3675 arch/i386/i386/machdep.c 	map->_dm_size = size;
map              3676 arch/i386/i386/machdep.c 	map->_dm_segcnt = nsegments;
map              3677 arch/i386/i386/machdep.c 	map->_dm_maxsegsz = maxsegsz;
map              3678 arch/i386/i386/machdep.c 	map->_dm_boundary = boundary;
map              3679 arch/i386/i386/machdep.c 	map->_dm_flags = flags & ~(BUS_DMA_WAITOK|BUS_DMA_NOWAIT);
map              3680 arch/i386/i386/machdep.c 	map->dm_mapsize = 0;		/* no valid mappings */
map              3681 arch/i386/i386/machdep.c 	map->dm_nsegs = 0;
map              3683 arch/i386/i386/machdep.c 	*dmamp = map;
map              3692 arch/i386/i386/machdep.c _bus_dmamap_destroy(bus_dma_tag_t t, bus_dmamap_t map)
map              3695 arch/i386/i386/machdep.c 	free(map, M_DEVBUF);
map              3703 arch/i386/i386/machdep.c _bus_dmamap_load(bus_dma_tag_t t, bus_dmamap_t map, void *buf,
map              3712 arch/i386/i386/machdep.c 	map->dm_mapsize = 0;
map              3713 arch/i386/i386/machdep.c 	map->dm_nsegs = 0;
map              3715 arch/i386/i386/machdep.c 	if (buflen > map->_dm_size)
map              3719 arch/i386/i386/machdep.c 	error = _bus_dmamap_load_buffer(t, map, buf, buflen, p, flags,
map              3722 arch/i386/i386/machdep.c 		map->dm_mapsize = buflen;
map              3723 arch/i386/i386/machdep.c 		map->dm_nsegs = seg + 1;
map              3732 arch/i386/i386/machdep.c _bus_dmamap_load_mbuf(bus_dma_tag_t t, bus_dmamap_t map, struct mbuf *m0,
map              3742 arch/i386/i386/machdep.c 	map->dm_mapsize = 0;
map              3743 arch/i386/i386/machdep.c 	map->dm_nsegs = 0;
map              3750 arch/i386/i386/machdep.c 	if (m0->m_pkthdr.len > map->_dm_size)
map              3759 arch/i386/i386/machdep.c 		error = _bus_dmamap_load_buffer(t, map, m->m_data, m->m_len,
map              3764 arch/i386/i386/machdep.c 		map->dm_mapsize = m0->m_pkthdr.len;
map              3765 arch/i386/i386/machdep.c 		map->dm_nsegs = seg + 1;
map              3774 arch/i386/i386/machdep.c _bus_dmamap_load_uio(bus_dma_tag_t t, bus_dmamap_t map, struct uio *uio,
map              3787 arch/i386/i386/machdep.c 	map->dm_mapsize = 0;
map              3788 arch/i386/i386/machdep.c 	map->dm_nsegs = 0;
map              3793 arch/i386/i386/machdep.c 	if (resid > map->_dm_size)
map              3815 arch/i386/i386/machdep.c 		error = _bus_dmamap_load_buffer(t, map, addr, minlen,
map              3822 arch/i386/i386/machdep.c 		map->dm_mapsize = uio->uio_resid;
map              3823 arch/i386/i386/machdep.c 		map->dm_nsegs = seg + 1;
map              3833 arch/i386/i386/machdep.c _bus_dmamap_load_raw(bus_dma_tag_t t, bus_dmamap_t map, bus_dma_segment_t *segs,
map              3836 arch/i386/i386/machdep.c 	if (nsegs > map->_dm_segcnt || size > map->_dm_size)
map              3842 arch/i386/i386/machdep.c 	if (map->_dm_boundary) {
map              3843 arch/i386/i386/machdep.c 		bus_addr_t bmask = ~(map->_dm_boundary - 1);
map              3847 arch/i386/i386/machdep.c 			if (segs[i].ds_len > map->_dm_maxsegsz)
map              3855 arch/i386/i386/machdep.c 	bcopy(segs, map->dm_segs, nsegs * sizeof(*segs));
map              3856 arch/i386/i386/machdep.c 	map->dm_nsegs = nsegs;
map              3865 arch/i386/i386/machdep.c _bus_dmamap_unload(bus_dma_tag_t t, bus_dmamap_t map)
map              3872 arch/i386/i386/machdep.c 	map->dm_mapsize = 0;
map              3873 arch/i386/i386/machdep.c 	map->dm_nsegs = 0;
map              4012 arch/i386/i386/machdep.c _bus_dmamap_load_buffer(bus_dma_tag_t t, bus_dmamap_t map, void *buf,
map              4028 arch/i386/i386/machdep.c 	bmask  = ~(map->_dm_boundary - 1);
map              4046 arch/i386/i386/machdep.c 		if (map->_dm_boundary > 0) {
map              4047 arch/i386/i386/machdep.c 			baddr = (curaddr + map->_dm_boundary) & bmask;
map              4057 arch/i386/i386/machdep.c 			map->dm_segs[seg].ds_addr = curaddr;
map              4058 arch/i386/i386/machdep.c 			map->dm_segs[seg].ds_len = sgsize;
map              4062 arch/i386/i386/machdep.c 			    (map->dm_segs[seg].ds_len + sgsize) <=
map              4063 arch/i386/i386/machdep.c 			     map->_dm_maxsegsz &&
map              4064 arch/i386/i386/machdep.c 			    (map->_dm_boundary == 0 ||
map              4065 arch/i386/i386/machdep.c 			     (map->dm_segs[seg].ds_addr & bmask) ==
map              4067 arch/i386/i386/machdep.c 				map->dm_segs[seg].ds_len += sgsize;
map              4069 arch/i386/i386/machdep.c 				if (++seg >= map->_dm_segcnt)
map              4071 arch/i386/i386/machdep.c 				map->dm_segs[seg].ds_addr = curaddr;
map              4072 arch/i386/i386/machdep.c 				map->dm_segs[seg].ds_len = sgsize;
map               410 arch/i386/i386/mpbios.c mpbios_search(struct device *self, paddr_t start, int count, struct mp_map *map)
map               431 arch/i386/i386/mpbios.c 			return (mpbios_map(start + i, len, map));
map               613 arch/i386/i386/pmap.c pmap_exec_fixup(struct vm_map *map, struct trapframe *tf, struct pcb *pcb)
map               616 arch/i386/i386/pmap.c 	struct pmap *pm = vm_map_pmap(map);
map               619 arch/i386/i386/pmap.c 	vm_map_lock(map);
map               620 arch/i386/i386/pmap.c 	for (ent = (&map->header)->next; ent != &map->header; ent = ent->next) {
map               628 arch/i386/i386/pmap.c 	vm_map_unlock(map);
map              1244 arch/i386/i386/pmap.c 	struct vm_map *map;
map              1257 arch/i386/i386/pmap.c 		map = kernel_map;
map              1259 arch/i386/i386/pmap.c 		map = kmem_map;
map              1260 arch/i386/i386/pmap.c 	if (vm_map_lock_try(map)) {
map              1267 arch/i386/i386/pmap.c 		uvm_unmap_remove(map, (vaddr_t)pvp, ((vaddr_t)pvp) + PAGE_SIZE,
map              1269 arch/i386/i386/pmap.c 		vm_map_unlock(map);
map               457 arch/i386/i386/trap.c 		struct vm_map *map;
map               477 arch/i386/i386/trap.c 			map = kernel_map;
map               479 arch/i386/i386/trap.c 			map = &vm->vm_map;
map               482 arch/i386/i386/trap.c 		if (map == kernel_map && va == 0) {
map               490 arch/i386/i386/trap.c 		rv = uvm_fault(map, va, 0, ftype);
map               494 arch/i386/i386/trap.c 			if (map != kernel_map)
map               510 arch/i386/i386/trap.c 			    map, va, ftype, rv);
map               663 arch/i386/isa/isa_machdep.c 	bus_dmamap_t map;
map               674 arch/i386/isa/isa_machdep.c 	map = *dmamp;
map               675 arch/i386/isa/isa_machdep.c 	map->_dm_cookie = NULL;
map               704 arch/i386/isa/isa_machdep.c 	    ((map->_dm_size / NBPG) + 1) > map->_dm_segcnt) {
map               706 arch/i386/isa/isa_machdep.c 		cookiesize += (sizeof(bus_dma_segment_t) * map->_dm_segcnt);
map               720 arch/i386/isa/isa_machdep.c 	map->_dm_cookie = cookie;
map               730 arch/i386/isa/isa_machdep.c 		error = _isa_dma_alloc_bouncebuf(t, map, size, flags);
map               735 arch/i386/isa/isa_machdep.c 		if (map->_dm_cookie != NULL)
map               736 arch/i386/isa/isa_machdep.c 			free(map->_dm_cookie, M_DEVBUF);
map               737 arch/i386/isa/isa_machdep.c 		_bus_dmamap_destroy(t, map);
map               746 arch/i386/isa/isa_machdep.c _isa_bus_dmamap_destroy(bus_dma_tag_t t, bus_dmamap_t map)
map               748 arch/i386/isa/isa_machdep.c 	struct i386_isa_dma_cookie *cookie = map->_dm_cookie;
map               754 arch/i386/isa/isa_machdep.c 		_isa_dma_free_bouncebuf(t, map);
map               757 arch/i386/isa/isa_machdep.c 	_bus_dmamap_destroy(t, map);
map               764 arch/i386/isa/isa_machdep.c _isa_bus_dmamap_load(bus_dma_tag_t t, bus_dmamap_t map, void *buf,
map               767 arch/i386/isa/isa_machdep.c 	struct i386_isa_dma_cookie *cookie = map->_dm_cookie;
map               781 arch/i386/isa/isa_machdep.c 		    map->_dm_segcnt, map->_dm_boundary, p) == 0)
map               782 arch/i386/isa/isa_machdep.c 			return (_bus_dmamap_load(t, map, buf, buflen,
map               791 arch/i386/isa/isa_machdep.c 			error = _isa_dma_alloc_bouncebuf(t, map, buflen,
map               803 arch/i386/isa/isa_machdep.c 		error = _bus_dmamap_load(t, map, cookie->id_bouncebuf,
map               811 arch/i386/isa/isa_machdep.c 			if ((map->_dm_flags & BUS_DMA_ALLOCNOW) == 0)
map               812 arch/i386/isa/isa_machdep.c 				_isa_dma_free_bouncebuf(t, map);
map               821 arch/i386/isa/isa_machdep.c 		error = _bus_dmamap_load(t, map, buf, buflen, p, flags); 
map               831 arch/i386/isa/isa_machdep.c _isa_bus_dmamap_load_mbuf(bus_dma_tag_t t, bus_dmamap_t map, struct mbuf *m,
map               842 arch/i386/isa/isa_machdep.c _isa_bus_dmamap_load_uio(bus_dma_tag_t t, bus_dmamap_t map, struct uio *uio,
map               854 arch/i386/isa/isa_machdep.c _isa_bus_dmamap_load_raw(bus_dma_tag_t t, bus_dmamap_t map,
map               865 arch/i386/isa/isa_machdep.c _isa_bus_dmamap_unload(bus_dma_tag_t t, bus_dmamap_t map)
map               867 arch/i386/isa/isa_machdep.c 	struct i386_isa_dma_cookie *cookie = map->_dm_cookie;
map               874 arch/i386/isa/isa_machdep.c 	    (map->_dm_flags & BUS_DMA_ALLOCNOW) == 0)
map               875 arch/i386/isa/isa_machdep.c 		_isa_dma_free_bouncebuf(t, map);
map               882 arch/i386/isa/isa_machdep.c 	_bus_dmamap_unload(t, map);
map               889 arch/i386/isa/isa_machdep.c _isa_bus_dmamap_sync(bus_dma_tag_t t, bus_dmamap_t map, bus_addr_t offset,
map               892 arch/i386/isa/isa_machdep.c 	struct i386_isa_dma_cookie *cookie = map->_dm_cookie;
map               896 arch/i386/isa/isa_machdep.c 		if (offset >= map->dm_mapsize)
map               898 arch/i386/isa/isa_machdep.c 		if (len == 0 || (offset + len) > map->dm_mapsize)
map               941 arch/i386/isa/isa_machdep.c 	_bus_dmamap_sync(t, map, op);
map              1072 arch/i386/isa/isa_machdep.c _isa_dma_alloc_bouncebuf(bus_dma_tag_t t, bus_dmamap_t map, bus_size_t size, int flags)
map              1074 arch/i386/isa/isa_machdep.c 	struct i386_isa_dma_cookie *cookie = map->_dm_cookie;
map              1079 arch/i386/isa/isa_machdep.c 	    NBPG, map->_dm_boundary, cookie->id_bouncesegs,
map              1080 arch/i386/isa/isa_machdep.c 	    map->_dm_segcnt, &cookie->id_nbouncesegs, flags);
map              1102 arch/i386/isa/isa_machdep.c _isa_dma_free_bouncebuf(bus_dma_tag_t t, bus_dmamap_t map)
map              1104 arch/i386/isa/isa_machdep.c 	struct i386_isa_dma_cookie *cookie = map->_dm_cookie;
map                82 ddb/db_break.c db_set_breakpoint(struct vm_map *map, db_addr_t addr, int count)
map                86 ddb/db_break.c 	if (db_find_breakpoint(map, addr)) {
map               104 ddb/db_break.c 	bkpt->map = map;
map               115 ddb/db_break.c db_delete_breakpoint(struct vm_map *map, db_addr_t addr)
map               123 ddb/db_break.c 	    if (db_map_equal(bkpt->map, map) &&
map               138 ddb/db_break.c db_find_breakpoint(struct vm_map *map, db_addr_t addr)
map               146 ddb/db_break.c 	    if (db_map_equal(bkpt->map, map) &&
map               171 ddb/db_break.c 		if (db_map_current(bkpt->map)) {
map               191 ddb/db_break.c 		if (db_map_current(bkpt->map)) {
map               221 ddb/db_break.c 	bkpt->map = NULL;
map               258 ddb/db_break.c 		      db_map_current(bkpt->map) ? "*" : " ",
map               259 ddb/db_break.c 		      bkpt->map, bkpt->init_count);
map               307 ddb/db_break.c db_map_current(struct vm_map *map)
map               312 ddb/db_break.c 	return ((map == NULL) ||
map               313 ddb/db_break.c 		(map == kernel_map) ||
map               315 ddb/db_break.c 		 (map == thread->proc->map)));
map               336 ddb/db_break.c 	    return thread->proc->map;
map                42 ddb/db_break.h 	struct vm_map *map;			/* in this map */
map                87 ddb/db_watch.c db_set_watchpoint(struct vm_map *map, db_addr_t addr, vsize_t size)
map                91 ddb/db_watch.c 	if (map == NULL) {
map               103 ddb/db_watch.c 	    if (db_map_equal(watch->map, map) &&
map               116 ddb/db_watch.c 	watch->map = map;
map               127 ddb/db_watch.c db_delete_watchpoint(struct vm_map *map, db_addr_t addr)
map               135 ddb/db_watch.c 	    if (db_map_equal(watch->map, map) &&
map               161 ddb/db_watch.c 		      db_map_current(watch->map) ? "*" : " ",
map               162 ddb/db_watch.c 		      watch->map, watch->loaddr,
map               208 ddb/db_watch.c 		pmap_protect(watch->map->pmap,
map               212 ddb/db_watch.c 	    pmap_update(watch->map->pmap);
map               224 ddb/db_watch.c db_find_watchpoint(struct vm_map *map, db_addr_t addr, db_regs_t *regs)
map               232 ddb/db_watch.c 	    if (db_map_equal(watch->map, map)) {
map                40 ddb/db_watch.h 	struct vm_map *map;		/* in this map */
map               130 dev/acpi/acpimadt.c 	struct mp_intr_map *map;
map               240 dev/acpi/acpimadt.c 			map = malloc(sizeof (struct mp_intr_map), M_DEVBUF, M_NOWAIT);
map               241 dev/acpi/acpimadt.c 			if (map == NULL)
map               244 dev/acpi/acpimadt.c 			memset(map, 0, sizeof *map);
map               245 dev/acpi/acpimadt.c 			map->ioapic = apic;
map               246 dev/acpi/acpimadt.c 			map->ioapic_pin = pin - apic->sc_apic_vecbase;
map               247 dev/acpi/acpimadt.c 			map->bus_pin = entry->madt_override.source;
map               248 dev/acpi/acpimadt.c 			map->flags = entry->madt_override.flags;
map               250 dev/acpi/acpimadt.c 			map->global_int = entry->madt_override.global_int;
map               252 dev/acpi/acpimadt.c 			acpimadt_cfg_intr(entry->madt_override.flags, &map->redir);
map               254 dev/acpi/acpimadt.c 			map->ioapic_ih = APIC_INT_VIA_APIC |
map               258 dev/acpi/acpimadt.c 			apic->sc_pins[pin].ip_map = map;
map               260 dev/acpi/acpimadt.c 			map->next = mp_isa_bus->mb_intrs;
map               261 dev/acpi/acpimadt.c 			mp_isa_bus->mb_intrs = map;
map               272 dev/acpi/acpimadt.c 			map = &mp_intrs[mp_nintrs++];
map               273 dev/acpi/acpimadt.c 			memset(map, 0, sizeof *map);
map               274 dev/acpi/acpimadt.c 			map->cpu_id = lapic_map[entry->madt_lapic_nmi.acpi_proc_id];
map               275 dev/acpi/acpimadt.c 			map->ioapic_pin = pin;
map               276 dev/acpi/acpimadt.c 			map->flags = entry->madt_lapic_nmi.flags;
map               278 dev/acpi/acpimadt.c 			acpimadt_cfg_intr(entry->madt_lapic_nmi.flags, &map->redir);
map               279 dev/acpi/acpimadt.c 			map->redir &= ~IOAPIC_REDLO_DEL_MASK;
map               280 dev/acpi/acpimadt.c 			map->redir |= (IOAPIC_REDLO_DEL_NMI << IOAPIC_REDLO_DEL_SHIFT);
map               296 dev/acpi/acpimadt.c 		map = malloc(sizeof (struct mp_intr_map), M_DEVBUF, M_NOWAIT);
map               297 dev/acpi/acpimadt.c 		if (map == NULL)
map               300 dev/acpi/acpimadt.c 		memset(map, 0, sizeof *map);
map               301 dev/acpi/acpimadt.c 		map->ioapic = apic;
map               302 dev/acpi/acpimadt.c 		map->ioapic_pin = pin;
map               303 dev/acpi/acpimadt.c 		map->bus_pin = pin;
map               305 dev/acpi/acpimadt.c 		map->global_int = -1;
map               307 dev/acpi/acpimadt.c 		map->redir = (IOAPIC_REDLO_DEL_LOPRI << IOAPIC_REDLO_DEL_SHIFT);
map               309 dev/acpi/acpimadt.c 		map->ioapic_ih = APIC_INT_VIA_APIC |
map               313 dev/acpi/acpimadt.c 		apic->sc_pins[pin].ip_map = map;
map               315 dev/acpi/acpimadt.c 		map->next = mp_isa_bus->mb_intrs;
map               316 dev/acpi/acpimadt.c 		mp_isa_bus->mb_intrs = map;
map               148 dev/acpi/acpiprt.c 	struct mp_intr_map *map;
map               217 dev/acpi/acpiprt.c 		map = malloc(sizeof (struct mp_intr_map), M_DEVBUF, M_NOWAIT);
map               218 dev/acpi/acpiprt.c 		if (map == NULL)
map               221 dev/acpi/acpiprt.c 		memset(map, 0, sizeof *map);
map               222 dev/acpi/acpiprt.c 		map->ioapic = apic;
map               223 dev/acpi/acpiprt.c 		map->ioapic_pin = irq - apic->sc_apic_vecbase;
map               224 dev/acpi/acpiprt.c 		map->bus_pin = ((addr >> 14) & 0x7c) | (pin & 0x3);
map               225 dev/acpi/acpiprt.c 		map->redir = IOAPIC_REDLO_ACTLO | IOAPIC_REDLO_LEVEL;
map               226 dev/acpi/acpiprt.c 		map->redir |= (IOAPIC_REDLO_DEL_LOPRI << IOAPIC_REDLO_DEL_SHIFT);
map               228 dev/acpi/acpiprt.c 		map->ioapic_ih = APIC_INT_VIA_APIC |
map               230 dev/acpi/acpiprt.c 		    (map->ioapic_pin << APIC_INT_PIN_SHIFT));
map               232 dev/acpi/acpiprt.c 		apic->sc_pins[map->ioapic_pin].ip_map = map;
map               234 dev/acpi/acpiprt.c 		map->next = mp_busses[sc->sc_bus].mb_intrs;
map               235 dev/acpi/acpiprt.c 		mp_busses[sc->sc_bus].mb_intrs = map;
map              3420 dev/acpi/dsdt.c 	static union acpi_resource map;
map              3424 dev/acpi/dsdt.c 	if (rlen >= sizeof(map))
map              3427 dev/acpi/dsdt.c 	memset(&map, 0, sizeof(map));
map              3428 dev/acpi/dsdt.c 	memcpy(&map, crs, rlen);
map              3430 dev/acpi/dsdt.c 	return &map;
map               502 dev/adb/akbdmap.h #define KBD_MAP(name, base, map) \
map               503 dev/adb/akbdmap.h 			{ name, base, sizeof(map)/sizeof(keysym_t), map }
map               151 dev/gpio/gpio.c gpio_pin_map(void *gpio, int offset, u_int32_t mask, struct gpio_pinmap *map)
map               168 dev/gpio/gpio.c 			map->pm_map[npins++] = pin;
map               170 dev/gpio/gpio.c 	map->pm_size = npins;
map               176 dev/gpio/gpio.c gpio_pin_unmap(void *gpio, struct gpio_pinmap *map)
map               181 dev/gpio/gpio.c 	for (i = 0; i < map->pm_size; i++) {
map               182 dev/gpio/gpio.c 		pin = map->pm_map[i];
map               188 dev/gpio/gpio.c gpio_pin_read(void *gpio, struct gpio_pinmap *map, int pin)
map               192 dev/gpio/gpio.c 	return (gpiobus_pin_read(sc->sc_gc, map->pm_map[pin]));
map               196 dev/gpio/gpio.c gpio_pin_write(void *gpio, struct gpio_pinmap *map, int pin, int value)
map               200 dev/gpio/gpio.c 	return (gpiobus_pin_write(sc->sc_gc, map->pm_map[pin], value));
map               204 dev/gpio/gpio.c gpio_pin_ctl(void *gpio, struct gpio_pinmap *map, int pin, int flags)
map               208 dev/gpio/gpio.c 	return (gpiobus_pin_ctl(sc->sc_gc, map->pm_map[pin], flags));
map               212 dev/gpio/gpio.c gpio_pin_caps(void *gpio, struct gpio_pinmap *map, int pin)
map               216 dev/gpio/gpio.c 	return (sc->sc_pins[map->pm_map[pin]].pin_caps);
map               661 dev/hil/hilkbdmap.c #define KBD_MAP(name, base, map) \
map               662 dev/hil/hilkbdmap.c 			{ name, base, sizeof(map)/sizeof(keysym_t), map }
map                47 dev/i2o/iopspvar.h #define	IOPSP_TIDMAP(map, t, l)	(map[(t) * IOPSP_MAX_LUN + (l)])
map              2127 dev/ic/acx.c   	bus_dmamap_t map;
map              2159 dev/ic/acx.c   	map = rb->rb_mbuf_dmamap;
map              2161 dev/ic/acx.c   	bd->mbuf_tmp_dmamap = map;
map               649 dev/ic/acx100.c 	const uint8_t *map;
map               653 dev/ic/acx100.c 		map = acx100_txpower_maxim;
map               657 dev/ic/acx100.c 		map = acx100_txpower_rfmd;
map               665 dev/ic/acx100.c 	acx_write_phyreg(sc, ACXRV_PHYREG_TXPOWER, map[ACX100_TXPOWER]);
map              9997 dev/ic/aic79xx.c ahd_createdmamem(struct ahd_softc *ahd, size_t size, struct map_node *map,
map              10003 dev/ic/aic79xx.c         bzero(map, sizeof(*map));
map              10006 dev/ic/aic79xx.c 	    &map->dmamap)) != 0) {
map              10012 dev/ic/aic79xx.c 	if ((error = bus_dmamem_alloc(tag, size, PAGE_SIZE, 0, &map->dmaseg,
map              10019 dev/ic/aic79xx.c 	if ((error = bus_dmamem_map(tag, &map->dmaseg, nseg, size,
map              10020 dev/ic/aic79xx.c 	    (caddr_t *)&map->vaddr, BUS_DMA_NOWAIT|BUS_DMA_COHERENT)) != 0) {
map              10026 dev/ic/aic79xx.c 	if ((error = bus_dmamap_load(tag, map->dmamap, map->vaddr, size, NULL,
map              10033 dev/ic/aic79xx.c 	map->size = size;
map              10034 dev/ic/aic79xx.c 	map->busaddr = map->dmamap->dm_segs[0].ds_addr;
map              10038 dev/ic/aic79xx.c 	bus_dmamem_unmap(tag, map->vaddr, size);
map              10040 dev/ic/aic79xx.c 	bus_dmamem_free(tag, &map->dmaseg, 1);
map              10042 dev/ic/aic79xx.c 	bus_dmamap_destroy(tag, map->dmamap);
map              10044 dev/ic/aic79xx.c 	bzero(map, sizeof(*map));
map              10049 dev/ic/aic79xx.c ahd_freedmamem(struct ahd_softc* ahd, struct map_node *map)
map              10053 dev/ic/aic79xx.c 	bus_dmamap_unload(tag, map->dmamap);
map              10054 dev/ic/aic79xx.c 	bus_dmamem_unmap(tag, map->vaddr, map->size);
map              10055 dev/ic/aic79xx.c 	bus_dmamem_free(tag, &map->dmaseg, 1);
map              10056 dev/ic/aic79xx.c 	bus_dmamap_destroy(tag, map->dmamap);
map               260 dev/ic/aic7xxx.c 			   bus_dmamap_t map,
map              7350 dev/ic/aic7xxx.c ahc_freedmamem(bus_dma_tag_t tag, int size, bus_dmamap_t map, caddr_t vaddr,
map              7354 dev/ic/aic7xxx.c 	bus_dmamap_unload(tag, map);
map              7357 dev/ic/aic7xxx.c 	bus_dmamap_destroy(tag, map);
map               758 dev/ic/ciss.c  	lmap->size = htobe32(sc->maxunits * sizeof(lmap->map));
map               759 dev/ic/ciss.c  	total = sizeof(*lmap) + (sc->maxunits - 1) * sizeof(lmap->map);
map               782 dev/ic/ciss.c  	    lmap->map[0].tgt, lmap->map[0].tgt2));
map               129 dev/ic/cissreg.h 	} map[1];
map              1942 dev/ic/dc.c    	bus_dmamap_t map;
map              1962 dev/ic/dc.c    		map = sc->dc_cdata.dc_rx_chain[i].sd_map;
map              1964 dev/ic/dc.c    		sc->sc_rx_sparemap = map;
map              2358 dev/ic/dc.c    			bus_dmamap_t map = sc->dc_cdata.dc_tx_chain[idx].sd_map;
map              2360 dev/ic/dc.c    			bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map              2362 dev/ic/dc.c    			bus_dmamap_unload(sc->sc_dmat, map);
map              2604 dev/ic/dc.c    	bus_dmamap_t map;
map              2611 dev/ic/dc.c    	map = sc->sc_tx_sparemap;
map              2613 dev/ic/dc.c    	if (bus_dmamap_load_mbuf(sc->sc_dmat, map,
map              2619 dev/ic/dc.c    	for (i = 0; i < map->dm_nsegs; i++) {
map              2623 dev/ic/dc.c    				bus_dmamap_unload(sc->sc_dmat, map);
map              2629 dev/ic/dc.c    			bus_dmamap_unload(sc->sc_dmat, map);
map              2634 dev/ic/dc.c    		f->dc_ctl = htole32(DC_TXCTL_TLINK | map->dm_segs[i].ds_len);
map              2640 dev/ic/dc.c    		f->dc_data = htole32(map->dm_segs[i].ds_addr);
map              2649 dev/ic/dc.c    	sc->dc_cdata.dc_tx_chain[cur].sd_map = map;
map              2664 dev/ic/dc.c    	bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map              3168 dev/ic/dc.c    			bus_dmamap_t map = sc->dc_cdata.dc_rx_chain[i].sd_map;
map              3170 dev/ic/dc.c    			bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map              3172 dev/ic/dc.c    			bus_dmamap_unload(sc->sc_dmat, map);
map              3187 dev/ic/dc.c    			bus_dmamap_t map = sc->dc_cdata.dc_tx_chain[i].sd_map;
map              3189 dev/ic/dc.c    			bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map              3191 dev/ic/dc.c    			bus_dmamap_unload(sc->sc_dmat, map);
map               166 dev/ic/fxpvar.h #define	FXP_RXMAP_PUT(sc,map)	((sc)->sc_rxmaps[--(sc)->sc_rxfree] = (map))
map              1608 dev/ic/gem.c   	bus_dmamap_t map;
map              1611 dev/ic/gem.c   	map = sc->sc_txd[cur].sd_map;
map              1613 dev/ic/gem.c   	if (bus_dmamap_load_mbuf(sc->sc_dmatag, map, mhead,
map              1618 dev/ic/gem.c   	if ((sc->sc_tx_cnt + map->dm_nsegs) > (GEM_NTXDESC - 2)) {
map              1619 dev/ic/gem.c   		bus_dmamap_unload(sc->sc_dmatag, map);
map              1623 dev/ic/gem.c   	bus_dmamap_sync(sc->sc_dmatag, map, 0, map->dm_mapsize,
map              1626 dev/ic/gem.c   	for (i = 0; i < map->dm_nsegs; i++) {
map              1628 dev/ic/gem.c   		    GEM_DMA_WRITE(sc, map->dm_segs[i].ds_addr);
map              1629 dev/ic/gem.c   		flags = (map->dm_segs[i].ds_len & GEM_TD_BUFSIZE) |
map              1631 dev/ic/gem.c   		    ((i == (map->dm_nsegs - 1)) ? GEM_TD_END_OF_PACKET : 0);
map              1641 dev/ic/gem.c   	sc->sc_tx_cnt += map->dm_nsegs;
map              1643 dev/ic/gem.c   	sc->sc_txd[cur].sd_map = map;
map              1495 dev/ic/hme.c   	bus_dmamap_t map;
map              1538 dev/ic/hme.c   	map = d->sd_map;
map              1540 dev/ic/hme.c   	sc->sc_rxmap_spare = map;
map              1420 dev/ic/isp.c   isp_getmap(struct ispsoftc *isp, fcpos_map_t *map)
map              1441 dev/ic/isp.c   		MEMCPY(map, fcp->isp_scratch, sizeof (fcpos_map_t));
map              1442 dev/ic/isp.c   		map->fwmap = mbs.param[1] != 0;
map               714 dev/ic/ispmbox.h 	u_int8_t	map[127];
map                70 dev/ic/malo.c  	bus_dmamap_t	map;
map                75 dev/ic/malo.c  	bus_dmamap_t		map;
map               565 dev/ic/malo.c  	    BUS_DMA_NOWAIT, &ring->map);
map               590 dev/ic/malo.c  	error = bus_dmamap_load(sc->sc_dmat, ring->map, ring->desc,
map               599 dev/ic/malo.c  	ring->physaddr = ring->map->dm_segs->ds_addr;
map               619 dev/ic/malo.c  		    0, BUS_DMA_NOWAIT, &data->map);
map               642 dev/ic/malo.c  		error = bus_dmamap_load(sc->sc_dmat, data->map,
map               651 dev/ic/malo.c  		desc->physdata = htole32(data->map->dm_segs->ds_addr);
map               656 dev/ic/malo.c  	bus_dmamap_sync(sc->sc_dmat, ring->map, 0, ring->map->dm_mapsize,
map               673 dev/ic/malo.c  	bus_dmamap_sync(sc->sc_dmat, ring->map, 0, ring->map->dm_mapsize,
map               686 dev/ic/malo.c  		bus_dmamap_sync(sc->sc_dmat, ring->map, 0,
map               687 dev/ic/malo.c  		    ring->map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
map               688 dev/ic/malo.c  		bus_dmamap_unload(sc->sc_dmat, ring->map);
map               699 dev/ic/malo.c  				bus_dmamap_sync(sc->sc_dmat, data->map, 0,
map               700 dev/ic/malo.c  				    data->map->dm_mapsize,
map               702 dev/ic/malo.c  				bus_dmamap_unload(sc->sc_dmat, data->map);
map               706 dev/ic/malo.c  			if (data->map != NULL)
map               707 dev/ic/malo.c  				bus_dmamap_destroy(sc->sc_dmat, data->map);
map               725 dev/ic/malo.c  	    count * sizeof(struct malo_tx_desc), 0, BUS_DMA_NOWAIT, &ring->map);
map               750 dev/ic/malo.c  	error = bus_dmamap_load(sc->sc_dmat, ring->map, ring->desc,
map               759 dev/ic/malo.c  	ring->physaddr = ring->map->dm_segs->ds_addr;
map               774 dev/ic/malo.c  		    &ring->data[i].map);
map               802 dev/ic/malo.c  			bus_dmamap_sync(sc->sc_dmat, data->map, 0,
map               803 dev/ic/malo.c  			    data->map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
map               804 dev/ic/malo.c  			bus_dmamap_unload(sc->sc_dmat, data->map);
map               818 dev/ic/malo.c  	bus_dmamap_sync(sc->sc_dmat, ring->map, 0, ring->map->dm_mapsize,
map               832 dev/ic/malo.c  		bus_dmamap_sync(sc->sc_dmat, ring->map, 0,
map               833 dev/ic/malo.c  		    ring->map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
map               834 dev/ic/malo.c  		bus_dmamap_unload(sc->sc_dmat, ring->map);
map               845 dev/ic/malo.c  				bus_dmamap_sync(sc->sc_dmat, data->map, 0,
map               846 dev/ic/malo.c  				    data->map->dm_mapsize,
map               848 dev/ic/malo.c  				bus_dmamap_unload(sc->sc_dmat, data->map);
map               858 dev/ic/malo.c  			if (data->map != NULL)
map               859 dev/ic/malo.c  				bus_dmamap_destroy(sc->sc_dmat, data->map);
map              1393 dev/ic/malo.c  		bus_dmamap_sync(sc->sc_dmat, data->map, 0,
map              1394 dev/ic/malo.c  		    data->map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
map              1395 dev/ic/malo.c  		bus_dmamap_unload(sc->sc_dmat, data->map);
map              1494 dev/ic/malo.c  	error = bus_dmamap_load_mbuf(sc->sc_dmat, data->map, m0,
map              1508 dev/ic/malo.c  	    data->map->dm_segs, data->map->dm_nsegs);
map              1510 dev/ic/malo.c  	bus_dmamap_sync(sc->sc_dmat, data->map, 0, data->map->dm_mapsize,
map              1512 dev/ic/malo.c  	bus_dmamap_sync(sc->sc_dmat, sc->sc_txring.map,
map              1612 dev/ic/malo.c  	error = bus_dmamap_load_mbuf(sc->sc_dmat, data->map, m0,
map              1626 dev/ic/malo.c  	    data->map->dm_segs, data->map->dm_nsegs);
map              1628 dev/ic/malo.c  	bus_dmamap_sync(sc->sc_dmat, data->map, 0, data->map->dm_mapsize,
map              1630 dev/ic/malo.c  	bus_dmamap_sync(sc->sc_dmat, sc->sc_txring.map,
map              1677 dev/ic/malo.c  		bus_dmamap_sync(sc->sc_dmat, sc->sc_rxring.map,
map              1705 dev/ic/malo.c  		bus_dmamap_sync(sc->sc_dmat, data->map, 0,
map              1706 dev/ic/malo.c  		    data->map->dm_mapsize, BUS_DMASYNC_POSTREAD);
map              1707 dev/ic/malo.c  		bus_dmamap_unload(sc->sc_dmat, data->map);
map              1709 dev/ic/malo.c  		error = bus_dmamap_load(sc->sc_dmat, data->map,
map              1714 dev/ic/malo.c  			error = bus_dmamap_load(sc->sc_dmat, data->map,
map              1730 dev/ic/malo.c  		desc->physdata = htole32(data->map->dm_segs->ds_addr);
map              1783 dev/ic/malo.c  		bus_dmamap_sync(sc->sc_dmat, sc->sc_rxring.map,
map                24 dev/ic/malo.h  	bus_dmamap_t		map;
map                38 dev/ic/malo.h  	bus_dmamap_t		map;
map               370 dev/ic/mtd8xx.c 	bus_dmamap_t map;
map               377 dev/ic/mtd8xx.c 	map = sc->sc_tx_sparemap;
map               379 dev/ic/mtd8xx.c 	if (bus_dmamap_load_mbuf(sc->sc_dmat, map,
map               385 dev/ic/mtd8xx.c 	for (i = 0; i < map->dm_nsegs; i++) {
map               388 dev/ic/mtd8xx.c 			bus_dmamap_unload(sc->sc_dmat, map);
map               393 dev/ic/mtd8xx.c 		f->td_tcw = htole32(map->dm_segs[i].ds_len);
map               394 dev/ic/mtd8xx.c 		total_len += map->dm_segs[i].ds_len;
map               400 dev/ic/mtd8xx.c 		f->td_buf = htole32(map->dm_segs[i].ds_addr);
map               409 dev/ic/mtd8xx.c 	sc->mtd_cdata.mtd_tx_chain[cur].sd_map = map;
map               415 dev/ic/mtd8xx.c 	bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map               497 dev/ic/mtd8xx.c 	bus_dmamap_t map;
map               518 dev/ic/mtd8xx.c 		map = sc->mtd_cdata.mtd_rx_chain[i].sd_map;
map               520 dev/ic/mtd8xx.c 		sc->sc_rx_sparemap = map;
map               785 dev/ic/mtd8xx.c 			bus_dmamap_t map = sc->mtd_cdata.mtd_rx_chain[i].sd_map;
map               787 dev/ic/mtd8xx.c 			bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map               789 dev/ic/mtd8xx.c 			bus_dmamap_unload(sc->sc_dmat, map);
map               804 dev/ic/mtd8xx.c 			bus_dmamap_t map = sc->mtd_cdata.mtd_tx_chain[i].sd_map;
map               806 dev/ic/mtd8xx.c 			bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map               808 dev/ic/mtd8xx.c 			bus_dmamap_unload(sc->sc_dmat, map);
map              1074 dev/ic/mtd8xx.c 			bus_dmamap_t map =
map              1076 dev/ic/mtd8xx.c 			bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map              1078 dev/ic/mtd8xx.c 			bus_dmamap_unload(sc->sc_dmat, map);
map              1107 dev/ic/re.c    	bus_dmamap_t	map;
map              1136 dev/ic/re.c    	map = rxs->rxs_dmamap;
map              1137 dev/ic/re.c    	error = bus_dmamap_load_mbuf(sc->sc_dmat, map, m,
map              1143 dev/ic/re.c    	bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map              1159 dev/ic/re.c    	cmdstat = map->dm_segs[0].ds_len;
map              1162 dev/ic/re.c    	re_set_bufaddr(d, map->dm_segs[0].ds_addr);
map              1530 dev/ic/re.c    	bus_dmamap_t	map;
map              1569 dev/ic/re.c    	map = txq->txq_dmamap;
map              1570 dev/ic/re.c    	error = bus_dmamap_load_mbuf(sc->sc_dmat, map, m,
map              1579 dev/ic/re.c    	nsegs = map->dm_nsegs;
map              1596 dev/ic/re.c    	bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map              1612 dev/ic/re.c    	for (seg = 0; seg < map->dm_nsegs;
map              1634 dev/ic/re.c    		re_set_bufaddr(d, map->dm_segs[seg].ds_addr);
map              1635 dev/ic/re.c    		cmdstat = rl_flags | map->dm_segs[seg].ds_len;
map              1703 dev/ic/re.c    	bus_dmamap_unload(sc->sc_dmat, map);
map               353 dev/ic/rt2560.c 	    count * RT2560_TX_DESC_SIZE, 0, BUS_DMA_NOWAIT, &ring->map);
map               377 dev/ic/rt2560.c 	error = bus_dmamap_load(sc->sc_dmat, ring->map, ring->desc,
map               386 dev/ic/rt2560.c 	ring->physaddr = ring->map->dm_segs->ds_addr;
map               401 dev/ic/rt2560.c 		    &ring->data[i].map);
map               425 dev/ic/rt2560.c 			bus_dmamap_sync(sc->sc_dmat, data->map, 0,
map               426 dev/ic/rt2560.c 			    data->map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
map               427 dev/ic/rt2560.c 			bus_dmamap_unload(sc->sc_dmat, data->map);
map               441 dev/ic/rt2560.c 	bus_dmamap_sync(sc->sc_dmat, ring->map, 0, ring->map->dm_mapsize,
map               455 dev/ic/rt2560.c 		bus_dmamap_sync(sc->sc_dmat, ring->map, 0,
map               456 dev/ic/rt2560.c 		    ring->map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
map               457 dev/ic/rt2560.c 		bus_dmamap_unload(sc->sc_dmat, ring->map);
map               468 dev/ic/rt2560.c 				bus_dmamap_sync(sc->sc_dmat, data->map, 0,
map               469 dev/ic/rt2560.c 				    data->map->dm_mapsize,
map               471 dev/ic/rt2560.c 				bus_dmamap_unload(sc->sc_dmat, data->map);
map               481 dev/ic/rt2560.c 			if (data->map != NULL)
map               482 dev/ic/rt2560.c 				bus_dmamap_destroy(sc->sc_dmat, data->map);
map               499 dev/ic/rt2560.c 	    count * RT2560_RX_DESC_SIZE, 0, BUS_DMA_NOWAIT, &ring->map);
map               523 dev/ic/rt2560.c 	error = bus_dmamap_load(sc->sc_dmat, ring->map, ring->desc,
map               532 dev/ic/rt2560.c 	ring->physaddr = ring->map->dm_segs->ds_addr;
map               552 dev/ic/rt2560.c 		    0, BUS_DMA_NOWAIT, &data->map);
map               574 dev/ic/rt2560.c 		error = bus_dmamap_load(sc->sc_dmat, data->map,
map               583 dev/ic/rt2560.c 		desc->physaddr = htole32(data->map->dm_segs->ds_addr);
map               586 dev/ic/rt2560.c 	bus_dmamap_sync(sc->sc_dmat, ring->map, 0, ring->map->dm_mapsize,
map               605 dev/ic/rt2560.c 	bus_dmamap_sync(sc->sc_dmat, ring->map, 0, ring->map->dm_mapsize,
map               618 dev/ic/rt2560.c 		bus_dmamap_sync(sc->sc_dmat, ring->map, 0,
map               619 dev/ic/rt2560.c 		    ring->map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
map               620 dev/ic/rt2560.c 		bus_dmamap_unload(sc->sc_dmat, ring->map);
map               631 dev/ic/rt2560.c 				bus_dmamap_sync(sc->sc_dmat, data->map, 0,
map               632 dev/ic/rt2560.c 				    data->map->dm_mapsize,
map               634 dev/ic/rt2560.c 				bus_dmamap_unload(sc->sc_dmat, data->map);
map               638 dev/ic/rt2560.c 			if (data->map != NULL)
map               639 dev/ic/rt2560.c 				bus_dmamap_destroy(sc->sc_dmat, data->map);
map               893 dev/ic/rt2560.c 		bus_dmamap_sync(sc->sc_dmat, sc->txq.map,
map               909 dev/ic/rt2560.c 		bus_dmamap_sync(sc->sc_dmat, sc->txq.map,
map               935 dev/ic/rt2560.c 		bus_dmamap_sync(sc->sc_dmat, sc->txq.map,
map               977 dev/ic/rt2560.c 		bus_dmamap_sync(sc->sc_dmat, data->map, 0,
map               978 dev/ic/rt2560.c 		    data->map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
map               979 dev/ic/rt2560.c 		bus_dmamap_unload(sc->sc_dmat, data->map);
map               988 dev/ic/rt2560.c 		bus_dmamap_sync(sc->sc_dmat, sc->txq.map,
map              1013 dev/ic/rt2560.c 		bus_dmamap_sync(sc->sc_dmat, sc->prioq.map,
map              1043 dev/ic/rt2560.c 		bus_dmamap_sync(sc->sc_dmat, data->map, 0,
map              1044 dev/ic/rt2560.c 		    data->map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
map              1045 dev/ic/rt2560.c 		bus_dmamap_unload(sc->sc_dmat, data->map);
map              1054 dev/ic/rt2560.c 		bus_dmamap_sync(sc->sc_dmat, sc->prioq.map,
map              1093 dev/ic/rt2560.c 		bus_dmamap_sync(sc->sc_dmat, sc->rxq.map,
map              1131 dev/ic/rt2560.c 		bus_dmamap_sync(sc->sc_dmat, data->map, 0,
map              1132 dev/ic/rt2560.c 		    data->map->dm_mapsize, BUS_DMASYNC_POSTREAD);
map              1133 dev/ic/rt2560.c 		bus_dmamap_unload(sc->sc_dmat, data->map);
map              1135 dev/ic/rt2560.c 		error = bus_dmamap_load(sc->sc_dmat, data->map,
map              1141 dev/ic/rt2560.c 			error = bus_dmamap_load(sc->sc_dmat, data->map,
map              1159 dev/ic/rt2560.c 		desc->physaddr = htole32(data->map->dm_segs->ds_addr);
map              1206 dev/ic/rt2560.c 		bus_dmamap_sync(sc->sc_dmat, sc->rxq.map,
map              1235 dev/ic/rt2560.c 		bus_dmamap_sync(sc->sc_dmat, sc->rxq.map,
map              1264 dev/ic/rt2560.c 		bus_dmamap_sync(sc->sc_dmat, sc->rxq.map,
map              1304 dev/ic/rt2560.c 		bus_dmamap_sync(sc->sc_dmat, data->map, 0,
map              1305 dev/ic/rt2560.c 		    data->map->dm_mapsize, BUS_DMASYNC_PREWRITE);
map              1553 dev/ic/rt2560.c 	error = bus_dmamap_load_mbuf(sc->sc_dmat, data->map, m0,
map              1567 dev/ic/rt2560.c 	    data->map->dm_segs->ds_addr);
map              1569 dev/ic/rt2560.c 	bus_dmamap_sync(sc->sc_dmat, data->map, 0, data->map->dm_mapsize,
map              1571 dev/ic/rt2560.c 	bus_dmamap_sync(sc->sc_dmat, sc->bcnq.map,
map              1624 dev/ic/rt2560.c 	error = bus_dmamap_load_mbuf(sc->sc_dmat, data->map, m0,
map              1674 dev/ic/rt2560.c 	    data->map->dm_segs->ds_addr);
map              1676 dev/ic/rt2560.c 	bus_dmamap_sync(sc->sc_dmat, data->map, 0, data->map->dm_mapsize,
map              1678 dev/ic/rt2560.c 	bus_dmamap_sync(sc->sc_dmat, sc->prioq.map,
map              1785 dev/ic/rt2560.c 		error = bus_dmamap_load_mbuf(sc->sc_dmat, data->map, mprot,
map              1804 dev/ic/rt2560.c 		    data->map->dm_segs->ds_addr);
map              1806 dev/ic/rt2560.c 		bus_dmamap_sync(sc->sc_dmat, data->map, 0,
map              1807 dev/ic/rt2560.c 		    data->map->dm_mapsize, BUS_DMASYNC_PREWRITE);
map              1808 dev/ic/rt2560.c 		bus_dmamap_sync(sc->sc_dmat, txq->map,
map              1822 dev/ic/rt2560.c 	error = bus_dmamap_load_mbuf(sc->sc_dmat, data->map, m0,
map              1853 dev/ic/rt2560.c 		error = bus_dmamap_load_mbuf(sc->sc_dmat, data->map, m0,
map              1899 dev/ic/rt2560.c 	    data->map->dm_segs->ds_addr);
map              1901 dev/ic/rt2560.c 	bus_dmamap_sync(sc->sc_dmat, data->map, 0, data->map->dm_mapsize,
map              1903 dev/ic/rt2560.c 	bus_dmamap_sync(sc->sc_dmat, txq->map,
map                55 dev/ic/rt2560var.h 	bus_dmamap_t			map;
map                61 dev/ic/rt2560var.h 	bus_dmamap_t		map;
map                75 dev/ic/rt2560var.h 	bus_dmamap_t	map;
map                81 dev/ic/rt2560var.h 	bus_dmamap_t		map;
map               386 dev/ic/rt2661.c 	    count * RT2661_TX_DESC_SIZE, 0, BUS_DMA_NOWAIT, &ring->map);
map               410 dev/ic/rt2661.c 	error = bus_dmamap_load(sc->sc_dmat, ring->map, ring->desc,
map               419 dev/ic/rt2661.c 	ring->physaddr = ring->map->dm_segs->ds_addr;
map               434 dev/ic/rt2661.c 		    &ring->data[i].map);
map               458 dev/ic/rt2661.c 			bus_dmamap_sync(sc->sc_dmat, data->map, 0,
map               459 dev/ic/rt2661.c 			    data->map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
map               460 dev/ic/rt2661.c 			bus_dmamap_unload(sc->sc_dmat, data->map);
map               474 dev/ic/rt2661.c 	bus_dmamap_sync(sc->sc_dmat, ring->map, 0, ring->map->dm_mapsize,
map               487 dev/ic/rt2661.c 		bus_dmamap_sync(sc->sc_dmat, ring->map, 0,
map               488 dev/ic/rt2661.c 		    ring->map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
map               489 dev/ic/rt2661.c 		bus_dmamap_unload(sc->sc_dmat, ring->map);
map               500 dev/ic/rt2661.c 				bus_dmamap_sync(sc->sc_dmat, data->map, 0,
map               501 dev/ic/rt2661.c 				    data->map->dm_mapsize,
map               503 dev/ic/rt2661.c 				bus_dmamap_unload(sc->sc_dmat, data->map);
map               512 dev/ic/rt2661.c 			if (data->map != NULL)
map               513 dev/ic/rt2661.c 				bus_dmamap_destroy(sc->sc_dmat, data->map);
map               529 dev/ic/rt2661.c 	    count * RT2661_RX_DESC_SIZE, 0, BUS_DMA_NOWAIT, &ring->map);
map               553 dev/ic/rt2661.c 	error = bus_dmamap_load(sc->sc_dmat, ring->map, ring->desc,
map               562 dev/ic/rt2661.c 	ring->physaddr = ring->map->dm_segs->ds_addr;
map               582 dev/ic/rt2661.c 		    0, BUS_DMA_NOWAIT, &data->map);
map               604 dev/ic/rt2661.c 		error = bus_dmamap_load(sc->sc_dmat, data->map,
map               613 dev/ic/rt2661.c 		desc->physaddr = htole32(data->map->dm_segs->ds_addr);
map               616 dev/ic/rt2661.c 	bus_dmamap_sync(sc->sc_dmat, ring->map, 0, ring->map->dm_mapsize,
map               633 dev/ic/rt2661.c 	bus_dmamap_sync(sc->sc_dmat, ring->map, 0, ring->map->dm_mapsize,
map               645 dev/ic/rt2661.c 		bus_dmamap_sync(sc->sc_dmat, ring->map, 0,
map               646 dev/ic/rt2661.c 		    ring->map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
map               647 dev/ic/rt2661.c 		bus_dmamap_unload(sc->sc_dmat, ring->map);
map               658 dev/ic/rt2661.c 				bus_dmamap_sync(sc->sc_dmat, data->map, 0,
map               659 dev/ic/rt2661.c 				    data->map->dm_mapsize,
map               661 dev/ic/rt2661.c 				bus_dmamap_unload(sc->sc_dmat, data->map);
map               665 dev/ic/rt2661.c 			if (data->map != NULL)
map               666 dev/ic/rt2661.c 				bus_dmamap_destroy(sc->sc_dmat, data->map);
map               968 dev/ic/rt2661.c 		bus_dmamap_sync(sc->sc_dmat, txq->map,
map               976 dev/ic/rt2661.c 		bus_dmamap_sync(sc->sc_dmat, data->map, 0,
map               977 dev/ic/rt2661.c 		    data->map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
map               978 dev/ic/rt2661.c 		bus_dmamap_unload(sc->sc_dmat, data->map);
map               986 dev/ic/rt2661.c 		bus_dmamap_sync(sc->sc_dmat, txq->map,
map              1011 dev/ic/rt2661.c 		bus_dmamap_sync(sc->sc_dmat, sc->rxq.map,
map              1054 dev/ic/rt2661.c 		bus_dmamap_sync(sc->sc_dmat, data->map, 0,
map              1055 dev/ic/rt2661.c 		    data->map->dm_mapsize, BUS_DMASYNC_POSTREAD);
map              1056 dev/ic/rt2661.c 		bus_dmamap_unload(sc->sc_dmat, data->map);
map              1058 dev/ic/rt2661.c 		error = bus_dmamap_load(sc->sc_dmat, data->map,
map              1064 dev/ic/rt2661.c 			error = bus_dmamap_load(sc->sc_dmat, data->map,
map              1082 dev/ic/rt2661.c 		desc->physaddr = htole32(data->map->dm_segs->ds_addr);
map              1136 dev/ic/rt2661.c 		bus_dmamap_sync(sc->sc_dmat, sc->rxq.map,
map              1474 dev/ic/rt2661.c 	error = bus_dmamap_load_mbuf(sc->sc_dmat, data->map, m0,
map              1521 dev/ic/rt2661.c 	    m0->m_pkthdr.len, rate, data->map->dm_segs, data->map->dm_nsegs,
map              1524 dev/ic/rt2661.c 	bus_dmamap_sync(sc->sc_dmat, data->map, 0, data->map->dm_mapsize,
map              1526 dev/ic/rt2661.c 	bus_dmamap_sync(sc->sc_dmat, sc->mgtq.map,
map              1631 dev/ic/rt2661.c 		error = bus_dmamap_load_mbuf(sc->sc_dmat, data->map, mprot,
map              1649 dev/ic/rt2661.c 		    0, mprot->m_pkthdr.len, protrate, data->map->dm_segs,
map              1650 dev/ic/rt2661.c 		    data->map->dm_nsegs, ac);
map              1652 dev/ic/rt2661.c 		bus_dmamap_sync(sc->sc_dmat, data->map, 0,
map              1653 dev/ic/rt2661.c 		    data->map->dm_mapsize, BUS_DMASYNC_PREWRITE);
map              1654 dev/ic/rt2661.c 		bus_dmamap_sync(sc->sc_dmat, txq->map,
map              1667 dev/ic/rt2661.c 	error = bus_dmamap_load_mbuf(sc->sc_dmat, data->map, m0,
map              1698 dev/ic/rt2661.c 		error = bus_dmamap_load_mbuf(sc->sc_dmat, data->map, m0,
map              1743 dev/ic/rt2661.c 	    data->map->dm_segs, data->map->dm_nsegs, ac);
map              1745 dev/ic/rt2661.c 	bus_dmamap_sync(sc->sc_dmat, data->map, 0, data->map->dm_mapsize,
map              1747 dev/ic/rt2661.c 	bus_dmamap_sync(sc->sc_dmat, txq->map, txq->cur * RT2661_TX_DESC_SIZE,
map                51 dev/ic/rt2661var.h 	bus_dmamap_t			map;
map                57 dev/ic/rt2661var.h 	bus_dmamap_t		map;
map                70 dev/ic/rt2661var.h 	bus_dmamap_t	map;
map                75 dev/ic/rt2661var.h 	bus_dmamap_t		map;
map               516 dev/ic/vga.c   vga_extended_attach(self, iot, memt, type, map)
map               520 dev/ic/vga.c   	paddr_t (*map)(void *, off_t, int);
map               541 dev/ic/vga.c   	vc->vc_mmap = map;
map              1100 dev/ic/xl.c    	bus_dmamap_t	map;
map              1120 dev/ic/xl.c    	if (c->map->dm_nsegs != 0) {
map              1121 dev/ic/xl.c    		bus_dmamap_sync(sc->sc_dmat, c->map,
map              1122 dev/ic/xl.c    		    0, c->map->dm_mapsize, BUS_DMASYNC_POSTREAD);
map              1123 dev/ic/xl.c    		bus_dmamap_unload(sc->sc_dmat, c->map);
map              1126 dev/ic/xl.c    	map = c->map;
map              1127 dev/ic/xl.c    	c->map = sc->sc_rx_sparemap;
map              1128 dev/ic/xl.c    	sc->sc_rx_sparemap = map;
map              1133 dev/ic/xl.c    	bus_dmamap_sync(sc->sc_dmat, c->map, 0, c->map->dm_mapsize,
map              1138 dev/ic/xl.c    	    htole32(c->map->dm_segs[0].ds_addr + ETHER_ALIGN);
map              1140 dev/ic/xl.c    	    htole32(c->map->dm_segs[0].ds_len | XL_LAST_FRAG);
map              1348 dev/ic/xl.c    		if (cur_tx->map->dm_nsegs != 0) {
map              1349 dev/ic/xl.c    			bus_dmamap_t map = cur_tx->map;
map              1351 dev/ic/xl.c    			bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map              1353 dev/ic/xl.c    			bus_dmamap_unload(sc->sc_dmat, map);
map              1401 dev/ic/xl.c    		if (cur_tx->map->dm_nsegs != 0) {
map              1402 dev/ic/xl.c    			bus_dmamap_sync(sc->sc_dmat, cur_tx->map,
map              1403 dev/ic/xl.c    			    0, cur_tx->map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
map              1404 dev/ic/xl.c    			bus_dmamap_unload(sc->sc_dmat, cur_tx->map);
map              1612 dev/ic/xl.c    	bus_dmamap_t	map;
map              1614 dev/ic/xl.c    	map = sc->sc_tx_sparemap;
map              1617 dev/ic/xl.c    	error = bus_dmamap_load_mbuf(sc->sc_dmat, map,
map              1630 dev/ic/xl.c    	for (frag = 0, total_len = 0; frag < map->dm_nsegs; frag++) {
map              1633 dev/ic/xl.c    		total_len += map->dm_segs[frag].ds_len;
map              1635 dev/ic/xl.c    		    htole32(map->dm_segs[frag].ds_addr);
map              1637 dev/ic/xl.c    		    htole32(map->dm_segs[frag].ds_len);
map              1672 dev/ic/xl.c    	bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map              1675 dev/ic/xl.c    	if (c->map->dm_nsegs != 0) {
map              1676 dev/ic/xl.c    		bus_dmamap_sync(sc->sc_dmat, c->map,
map              1677 dev/ic/xl.c    		    0, c->map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
map              1678 dev/ic/xl.c    		bus_dmamap_unload(sc->sc_dmat, c->map);
map              1682 dev/ic/xl.c    	sc->sc_tx_sparemap = c->map;
map              1683 dev/ic/xl.c    	c->map = map;
map              2384 dev/ic/xl.c    	bus_dmamap_t	map;
map              2391 dev/ic/xl.c    		if (sc->xl_cdata.xl_rx_chain[i].map->dm_nsegs != 0) {
map              2392 dev/ic/xl.c    			map = sc->xl_cdata.xl_rx_chain[i].map;
map              2394 dev/ic/xl.c    			bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map              2396 dev/ic/xl.c    			bus_dmamap_unload(sc->sc_dmat, map);
map              2409 dev/ic/xl.c    		if (sc->xl_cdata.xl_tx_chain[i].map->dm_nsegs != 0) {
map              2410 dev/ic/xl.c    			map = sc->xl_cdata.xl_tx_chain[i].map;
map              2412 dev/ic/xl.c    			bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map              2414 dev/ic/xl.c    			bus_dmamap_unload(sc->sc_dmat, map);
map              2520 dev/ic/xl.c    		    &sc->xl_cdata.xl_rx_chain[i].map) != 0) {
map              2534 dev/ic/xl.c    		    &sc->xl_cdata.xl_tx_chain[i].map) != 0) {
map               469 dev/ic/xlreg.h 	bus_dmamap_t		map;
map               476 dev/ic/xlreg.h 	bus_dmamap_t		map;
map               776 dev/isa/ad1848.c ad1848_mixer_find_dev(map, cnt, cp)
map               777 dev/isa/ad1848.c   ad1848_devmap_t *map;
map               785 dev/isa/ad1848.c     if (map[idx].id == cp->dev) {
map               786 dev/isa/ad1848.c       return (&map[idx]);
map               793 dev/isa/ad1848.c ad1848_mixer_get_port(ac, map, cnt, cp)
map               795 dev/isa/ad1848.c   struct ad1848_devmap *map;
map               804 dev/isa/ad1848.c   if (!(entry = ad1848_mixer_find_dev(map, cnt, cp)))
map               867 dev/isa/ad1848.c ad1848_mixer_set_port(ac, map, cnt, cp)
map               869 dev/isa/ad1848.c   struct ad1848_devmap *map;
map               878 dev/isa/ad1848.c   if (!(entry = ad1848_mixer_find_dev(map, cnt, cp)))
map               727 dev/pci/agp.c  agp_free_dmamem(bus_dma_tag_t tag, size_t size, bus_dmamap_t map,
map               731 dev/pci/agp.c  	bus_dmamap_unload(tag, map);
map               732 dev/pci/agp.c  	bus_dmamap_destroy(tag, map);
map               152 dev/pci/auich.c 	bus_dmamap_t map;
map              1093 dev/pci/auich.c 	    p->size, 0, BUS_DMA_NOWAIT, &p->map)) != 0) {
map              1102 dev/pci/auich.c 	if ((error = bus_dmamap_load(sc->dmat, p->map, p->addr, p->size,
map              1106 dev/pci/auich.c 		bus_dmamap_destroy(sc->dmat, p->map);
map              1134 dev/pci/auich.c 	bus_dmamap_unload(sc->dmat, p->map);
map              1135 dev/pci/auich.c 	bus_dmamap_destroy(sc->dmat, p->map);
map              1488 dev/pci/auich.c 		sc->dmalist_pcmi[i].base = p->map->dm_segs[0].ds_addr;
map                90 dev/pci/auixp.c #define	DMAADDR(p)	((p)->map->dm_segs[0].ds_addr)
map              1131 dev/pci/auixp.c 				  BUS_DMA_NOWAIT, &dma->map);
map              1139 dev/pci/auixp.c 	error = bus_dmamap_load(sc->sc_dmat, dma->map, dma->addr, dma->size, NULL,
map              1147 dev/pci/auixp.c 	bus_dmamap_destroy(sc->sc_dmat, dma->map);
map              1162 dev/pci/auixp.c 	bus_dmamap_unload(sc->sc_dmat, p->map);
map              1163 dev/pci/auixp.c 	bus_dmamap_destroy(sc->sc_dmat, p->map);
map                53 dev/pci/auixpvar.h 	bus_dmamap_t		 map;
map                78 dev/pci/autri.c #define DMAADDR(p) ((p)->map->dm_segs[0].ds_addr)
map               887 dev/pci/autri.c 	    0, BUS_DMA_NOWAIT, &p->map);
map               891 dev/pci/autri.c 	error = bus_dmamap_load(sc->sc_dmatag, p->map, p->addr, p->size, NULL,
map               898 dev/pci/autri.c 	bus_dmamap_destroy(sc->sc_dmatag, p->map);
map               911 dev/pci/autri.c 	bus_dmamap_unload(sc->sc_dmatag, p->map);
map               912 dev/pci/autri.c 	bus_dmamap_destroy(sc->sc_dmatag, p->map);
map                35 dev/pci/autrivar.h 	bus_dmamap_t		map;
map                69 dev/pci/auvia.c 	bus_dmamap_t map;
map               728 dev/pci/auvia.c 	    BUS_DMA_NOWAIT, &p->map)) != 0) {
map               734 dev/pci/auvia.c 	if ((error = bus_dmamap_load(sc->sc_dmat, p->map, p->addr, size, NULL,
map               748 dev/pci/auvia.c 	bus_dmamap_destroy(sc->sc_dmat, p->map);
map               767 dev/pci/auvia.c 			bus_dmamap_unload(sc->sc_dmat, p->map);
map               768 dev/pci/auvia.c 			bus_dmamap_destroy(sc->sc_dmat, p->map);
map               818 dev/pci/auvia.c 	s = p->map->dm_segs[0].ds_addr;
map               883 dev/pci/auvia.c 	    ch->sc_dma_ops_dma->map->dm_segs[0].ds_addr);
map               928 dev/pci/auvia.c 	    ch->sc_dma_ops_dma->map->dm_segs[0].ds_addr);
map               145 dev/pci/azalia.c 	bus_dmamap_t map;
map               150 dev/pci/azalia.c #define AZALIA_DMA_DMAADDR(p)	((p)->map->dm_segs[0].ds_addr)
map              1007 dev/pci/azalia.c 	    BUS_DMA_NOWAIT, &d->map);
map              1010 dev/pci/azalia.c 	err = bus_dmamap_load(az->dmat, d->map, d->addr, size,
map              1022 dev/pci/azalia.c 	bus_dmamap_destroy(az->dmat, d->map);
map              1036 dev/pci/azalia.c 	bus_dmamap_unload(az->dmat, d->map);
map              1037 dev/pci/azalia.c 	bus_dmamap_destroy(az->dmat, d->map);
map                93 dev/pci/cs4280.c 	bus_dmamap_t map;
map               101 dev/pci/cs4280.c #define DMAADDR(p) ((p)->map->dm_segs[0].ds_addr)
map               892 dev/pci/cs4280.c 		err = cs4280_download(sc, &BA1Struct->map[offset],
map               943 dev/pci/cs4280.c 		err = cs4280_checkimage(sc, &BA1Struct->map[offset],
map              1347 dev/pci/cs4280.c 	bus_dmamap_unload(sc->sc_dmatag, p->map);
map              1348 dev/pci/cs4280.c 	bus_dmamap_destroy(sc->sc_dmatag, p->map);
map              1383 dev/pci/cs4280.c 				  0, BUS_DMA_NOWAIT, &p->map);
map              1390 dev/pci/cs4280.c 	error = bus_dmamap_load(sc->sc_dmatag, p->map, p->addr, p->size, NULL, 
map              1400 dev/pci/cs4280.c 	bus_dmamap_destroy(sc->sc_dmatag, p->map);
map               274 dev/pci/cs4280reg.h 	u_int32_t map[BA1_DWORD_SIZE];
map                69 dev/pci/cs4281.c 	bus_dmamap_t map;
map                77 dev/pci/cs4281.c #define DMAADDR(p) ((p)->map->dm_segs[0].ds_addr)
map              1386 dev/pci/cs4281.c 			bus_dmamap_unload(sc->sc_dmatag, p->map);
map              1387 dev/pci/cs4281.c 			bus_dmamap_destroy(sc->sc_dmatag, p->map);
map              1556 dev/pci/cs4281.c 				  0, BUS_DMA_NOWAIT, &p->map);
map              1563 dev/pci/cs4281.c 	error = bus_dmamap_load(sc->sc_dmatag, p->map, p->addr, p->size, NULL,
map              1573 dev/pci/cs4281.c 	bus_dmamap_destroy(sc->sc_dmatag, p->map);
map               102 dev/pci/eap.c  	bus_dmamap_t map;
map               110 dev/pci/eap.c  #define DMAADDR(p) ((p)->map->dm_segs[0].ds_addr)
map               827 dev/pci/eap.c  	    0, BUS_DMA_NOWAIT, &p->map);
map               831 dev/pci/eap.c  	error = bus_dmamap_load(sc->sc_dmatag, p->map, p->addr, p->size, NULL,
map               838 dev/pci/eap.c  	bus_dmamap_destroy(sc->sc_dmatag, p->map);
map               849 dev/pci/eap.c  	bus_dmamap_unload(sc->sc_dmatag, p->map);
map               850 dev/pci/eap.c  	bus_dmamap_destroy(sc->sc_dmatag, p->map);
map               323 dev/pci/emuxki.c 			      mem->bound, bus_dma_flags, &(mem->map))) {
map               330 dev/pci/emuxki.c 	if (bus_dmamap_load(dmat, mem->map, mem->kaddr, 
map               332 dev/pci/emuxki.c 		bus_dmamap_destroy(dmat, mem->map);
map               345 dev/pci/emuxki.c 	bus_dmamap_unload(mem->dmat, mem->map);
map               346 dev/pci/emuxki.c 	bus_dmamap_destroy(mem->dmat, mem->map);
map                59 dev/pci/emuxkivar.h 	bus_dmamap_t    map;
map               959 dev/pci/esa.c  				  BUS_DMA_NOWAIT, &p->map);
map               963 dev/pci/esa.c  	error = bus_dmamap_load(sc->sc_dmat, p->map, p->addr, p->size, NULL,
map               971 dev/pci/esa.c  	bus_dmamap_destroy(sc->sc_dmat, p->map);
map               984 dev/pci/esa.c  	bus_dmamap_unload(sc->sc_dmat, p->map);
map               985 dev/pci/esa.c  	bus_dmamap_destroy(sc->sc_dmat, p->map);
map                51 dev/pci/esavar.h #define	DMAADDR(p)	((p)->map->dm_segs[0].ds_addr)
map                64 dev/pci/esavar.h 	bus_dmamap_t		map;
map                76 dev/pci/fms.c  	bus_dmamap_t map;
map               737 dev/pci/fms.c  				       BUS_DMA_NOWAIT, &p->map)) != 0) {
map               743 dev/pci/fms.c  	if ((error = bus_dmamap_load(sc->sc_dmat, p->map, p->addr, size, NULL,
map               757 dev/pci/fms.c  	bus_dmamap_destroy(sc->sc_dmat, p->map);
map               778 dev/pci/fms.c  			bus_dmamap_unload(sc->sc_dmat, p->map);
map               779 dev/pci/fms.c  			bus_dmamap_destroy(sc->sc_dmat, p->map);
map               853 dev/pci/fms.c  	sc->sc_play_start = p->map->dm_segs[0].ds_addr;
map               891 dev/pci/fms.c  	sc->sc_rec_start = p->map->dm_segs[0].ds_addr;
map              1258 dev/pci/hifn7751.c hifn_dmamap_aligned(bus_dmamap_t map)
map              1262 dev/pci/hifn7751.c 	for (i = 0; i < map->dm_nsegs; i++) {
map              1263 dev/pci/hifn7751.c 		if (map->dm_segs[i].ds_addr & 3)
map              1265 dev/pci/hifn7751.c 		if ((i != (map->dm_nsegs - 1)) &&
map              1266 dev/pci/hifn7751.c 		    (map->dm_segs[i].ds_len & 3))
map              1276 dev/pci/hifn7751.c 	bus_dmamap_t map = cmd->dst_map;
map              1281 dev/pci/hifn7751.c 	for (i = 0; i < map->dm_nsegs - 1; i++) {
map              1282 dev/pci/hifn7751.c 		dma->dstr[idx].p = htole32(map->dm_segs[i].ds_addr);
map              1284 dev/pci/hifn7751.c 		    HIFN_D_MASKDONEIRQ | map->dm_segs[i].ds_len);
map              1299 dev/pci/hifn7751.c 		p = map->dm_segs[i].ds_addr;
map              1301 dev/pci/hifn7751.c 		    map->dm_segs[i].ds_len;
map              1308 dev/pci/hifn7751.c 		if ((map->dm_segs[i].ds_len - cmd->sloplen) != 0) {
map              1309 dev/pci/hifn7751.c 			dma->dstr[idx].p = htole32(map->dm_segs[i].ds_addr);
map              1312 dev/pci/hifn7751.c 			    (map->dm_segs[i].ds_len - cmd->sloplen));
map              1348 dev/pci/hifn7751.c 	bus_dmamap_t map = cmd->src_map;
map              1353 dev/pci/hifn7751.c 	for (i = 0; i < map->dm_nsegs; i++) {
map              1354 dev/pci/hifn7751.c 		if (i == map->dm_nsegs - 1)
map              1357 dev/pci/hifn7751.c 		dma->srcr[idx].p = htole32(map->dm_segs[i].ds_addr);
map              1358 dev/pci/hifn7751.c 		dma->srcr[idx].l = htole32(map->dm_segs[i].ds_len |
map              1372 dev/pci/hifn7751.c 	dma->srcu += map->dm_nsegs;
map              3243 dev/pci/if_bnx.c 	bus_dmamap_t		map;
map              3310 dev/pci/if_bnx.c 	map = sc->rx_mbuf_map[*chain_prod];
map              3312 dev/pci/if_bnx.c 	if (bus_dmamap_load_mbuf(sc->bnx_dmatag, map, m_new, BUS_DMA_NOWAIT)) {
map              3335 dev/pci/if_bnx.c 	addr = (u_int32_t)(map->dm_segs[0].ds_addr);
map              3337 dev/pci/if_bnx.c 	addr = (u_int32_t)((u_int64_t)map->dm_segs[0].ds_addr >> 32);
map              3339 dev/pci/if_bnx.c 	rxbd->rx_bd_len = htole32(map->dm_segs[0].ds_len);
map              3341 dev/pci/if_bnx.c 	*prod_bseq += map->dm_segs[0].ds_len;
map              3343 dev/pci/if_bnx.c 	for (i = 1; i < map->dm_nsegs; i++) {
map              3350 dev/pci/if_bnx.c 		addr = (u_int32_t)(map->dm_segs[i].ds_addr);
map              3352 dev/pci/if_bnx.c 		addr = (u_int32_t)((u_int64_t)map->dm_segs[i].ds_addr >> 32);
map              3354 dev/pci/if_bnx.c 		rxbd->rx_bd_len = htole32(map->dm_segs[i].ds_len);
map              3356 dev/pci/if_bnx.c 		*prod_bseq += map->dm_segs[i].ds_len;
map              3368 dev/pci/if_bnx.c 	sc->rx_mbuf_map[*chain_prod] = map;
map              3369 dev/pci/if_bnx.c 	sc->free_rx_bd -= map->dm_nsegs;
map              3372 dev/pci/if_bnx.c 	    map->dm_nsegs));
map              4309 dev/pci/if_bnx.c 	bus_dmamap_t		map;
map              4343 dev/pci/if_bnx.c 	map = sc->tx_mbuf_map[chain_prod];
map              4346 dev/pci/if_bnx.c 	error = bus_dmamap_load_mbuf(sc->bnx_dmatag, map, m0, BUS_DMA_NOWAIT);
map              4361 dev/pci/if_bnx.c 	if (map->dm_nsegs > (USABLE_TX_BD - sc->used_tx_bd - BNX_TX_SLACK_SPACE)) {
map              4362 dev/pci/if_bnx.c 		bus_dmamap_unload(sc->bnx_dmatag, map);
map              4383 dev/pci/if_bnx.c 	for (i = 0; i < map->dm_nsegs ; i++) {
map              4387 dev/pci/if_bnx.c 		addr = (u_int32_t)(map->dm_segs[i].ds_addr);
map              4389 dev/pci/if_bnx.c 		addr = (u_int32_t)((u_int64_t)map->dm_segs[i].ds_addr >> 32);
map              4391 dev/pci/if_bnx.c 		txbd->tx_bd_mss_nbytes = htole16(map->dm_segs[i].ds_len);
map              4394 dev/pci/if_bnx.c 		prod_bseq += map->dm_segs[i].ds_len;
map              4420 dev/pci/if_bnx.c 	sc->used_tx_bd += map->dm_nsegs;
map              1843 dev/pci/if_cas.c 	bus_dmamap_t map;
map              1846 dev/pci/if_cas.c 	map = sc->sc_txd[cur].sd_map;
map              1848 dev/pci/if_cas.c 	if (bus_dmamap_load_mbuf(sc->sc_dmatag, map, mhead,
map              1853 dev/pci/if_cas.c 	if ((sc->sc_tx_cnt + map->dm_nsegs) > (CAS_NTXDESC - 2)) {
map              1854 dev/pci/if_cas.c 		bus_dmamap_unload(sc->sc_dmatag, map);
map              1858 dev/pci/if_cas.c 	bus_dmamap_sync(sc->sc_dmatag, map, 0, map->dm_mapsize,
map              1861 dev/pci/if_cas.c 	for (i = 0; i < map->dm_nsegs; i++) {
map              1863 dev/pci/if_cas.c 		    CAS_DMA_WRITE(map->dm_segs[i].ds_addr);
map              1864 dev/pci/if_cas.c 		flags = (map->dm_segs[i].ds_len & CAS_TD_BUFSIZE) |
map              1866 dev/pci/if_cas.c 		    ((i == (map->dm_nsegs - 1)) ? CAS_TD_END_OF_PACKET : 0);
map              1876 dev/pci/if_cas.c 	sc->sc_tx_cnt += map->dm_nsegs;
map              1878 dev/pci/if_cas.c 	sc->sc_txd[cur].sd_map = map;
map              3061 dev/pci/if_de.c 	bus_dmamap_t map;
map              3066 dev/pci/if_de.c 	map = TULIP_GETCTX(m, bus_dmamap_t);
map              3067 dev/pci/if_de.c 	bus_dmamap_unload(sc->tulip_dmatag, map);
map              3068 dev/pci/if_de.c 	sc->tulip_txmaps[sc->tulip_txmaps_free++] = map;
map              3099 dev/pci/if_de.c 	bus_dmamap_t map;
map              3104 dev/pci/if_de.c 	map = TULIP_GETCTX(m, bus_dmamap_t);
map              3105 dev/pci/if_de.c 	bus_dmamap_unload(sc->tulip_dmatag, map);
map              3106 dev/pci/if_de.c 	sc->tulip_rxmaps[sc->tulip_rxmaps_free++] = map;
map              3197 dev/pci/if_de.c 	bus_dmamap_t map;
map              3261 dev/pci/if_de.c 		map = TULIP_GETCTX(me, bus_dmamap_t);
map              3262 dev/pci/if_de.c 		TULIP_RXMAP_POSTSYNC(sc, map);
map              3263 dev/pci/if_de.c 		bus_dmamap_unload(sc->tulip_dmatag, map);
map              3264 dev/pci/if_de.c 		sc->tulip_rxmaps[sc->tulip_rxmaps_free++] = map;
map              3283 dev/pci/if_de.c 	    map = TULIP_GETCTX(me, bus_dmamap_t);
map              3284 dev/pci/if_de.c 	    bus_dmamap_sync(sc->tulip_dmatag, map, 0, me->m_len,
map              3286 dev/pci/if_de.c 	    bus_dmamap_unload(sc->tulip_dmatag, map);
map              3287 dev/pci/if_de.c 	    sc->tulip_rxmaps[sc->tulip_rxmaps_free++] = map;
map              3333 dev/pci/if_de.c 	    map = TULIP_GETCTX(me, bus_dmamap_t);
map              3334 dev/pci/if_de.c 	    bus_dmamap_unload(sc->tulip_dmatag, map);
map              3335 dev/pci/if_de.c 	    sc->tulip_rxmaps[sc->tulip_rxmaps_free++] = map;
map              3417 dev/pci/if_de.c 		map = sc->tulip_rxmaps[--sc->tulip_rxmaps_free];
map              3426 dev/pci/if_de.c 	    TULIP_SETCTX(ms, map);
map              3427 dev/pci/if_de.c 	    error = bus_dmamap_load(sc->tulip_dmatag, map, mtod(ms, void *),
map              3434 dev/pci/if_de.c 	    nextout->d_addr1 = map->dm_segs[0].ds_addr;
map              3435 dev/pci/if_de.c 	    nextout->d_length1 = map->dm_segs[0].ds_len;
map              3436 dev/pci/if_de.c 	    if (map->dm_nsegs == 2) {
map              3437 dev/pci/if_de.c 		nextout->d_addr2 = map->dm_segs[1].ds_addr;
map              3438 dev/pci/if_de.c 		nextout->d_length2 = map->dm_segs[1].ds_len;
map              3509 dev/pci/if_de.c 		    bus_dmamap_t map = TULIP_GETCTX(m, bus_dmamap_t);
map              3510 dev/pci/if_de.c 		    TULIP_TXMAP_POSTSYNC(sc, map);
map              3511 dev/pci/if_de.c 		    sc->tulip_txmaps[sc->tulip_txmaps_free++] = map;
map              3816 dev/pci/if_de.c     bus_dmamap_t map;
map              3865 dev/pci/if_de.c 	map = sc->tulip_txmaps[sc->tulip_txmaps_free-1];
map              3873 dev/pci/if_de.c     error = bus_dmamap_load_mbuf(sc->tulip_dmatag, map, m, BUS_DMA_NOWAIT);
map              3908 dev/pci/if_de.c 	    error = bus_dmamap_load_mbuf(sc->tulip_dmatag, map, m, BUS_DMA_NOWAIT);
map              3919 dev/pci/if_de.c     if ((freedescs -= (map->dm_nsegs + 1) / 2) <= 0
map              3934 dev/pci/if_de.c 	bus_dmamap_unload(sc->tulip_dmatag, map);
map              3937 dev/pci/if_de.c     for (; map->dm_nsegs - segcnt > 1; segcnt += 2) {
map              3941 dev/pci/if_de.c 	eop->d_addr1   = map->dm_segs[segcnt].ds_addr;
map              3942 dev/pci/if_de.c 	eop->d_length1 = map->dm_segs[segcnt].ds_len;
map              3943 dev/pci/if_de.c 	eop->d_addr2   = map->dm_segs[segcnt+1].ds_addr;
map              3944 dev/pci/if_de.c 	eop->d_length2 = map->dm_segs[segcnt+1].ds_len;
map              3949 dev/pci/if_de.c     if (segcnt < map->dm_nsegs) {
map              3953 dev/pci/if_de.c 	eop->d_addr1   = map->dm_segs[segcnt].ds_addr;
map              3954 dev/pci/if_de.c 	eop->d_length1 = map->dm_segs[segcnt].ds_len;
map              3960 dev/pci/if_de.c     TULIP_TXMAP_PRESYNC(sc, map);
map              3961 dev/pci/if_de.c     TULIP_SETCTX(m, map);
map              3962 dev/pci/if_de.c     map = NULL;
map              4400 dev/pci/if_de.c 	    bus_dmamap_t map;
map              4402 dev/pci/if_de.c 				      BUS_DMA_NOWAIT, &map);
map              4404 dev/pci/if_de.c 		error = bus_dmamap_load(sc->tulip_dmatag, map, desc,
map              4407 dev/pci/if_de.c 		    bus_dmamap_destroy(sc->tulip_dmatag, map);
map              4409 dev/pci/if_de.c 		    *map_p = map;
map              4454 dev/pci/if_de.c 	    bus_dmamap_t map;
map              4455 dev/pci/if_de.c 	    if ((error = TULIP_TXMAP_CREATE(sc, &map)) == 0)
map              4456 dev/pci/if_de.c 		sc->tulip_txmaps[sc->tulip_txmaps_free++] = map;
map              4479 dev/pci/if_de.c 	    bus_dmamap_t map;
map              4480 dev/pci/if_de.c 	    if ((error = TULIP_RXMAP_CREATE(sc, &map)) == 0)
map              4481 dev/pci/if_de.c 		sc->tulip_rxmaps[sc->tulip_rxmaps_free++] = map;
map               764 dev/pci/if_devar.h #define	TULIP_RXMAP_PRESYNC(sc, map) \
map               765 dev/pci/if_devar.h 	bus_dmamap_sync((sc)->tulip_dmatag, (map), 0, (map)->dm_mapsize, \
map               767 dev/pci/if_devar.h #define	TULIP_RXMAP_POSTSYNC(sc, map) \
map               768 dev/pci/if_devar.h 	bus_dmamap_sync((sc)->tulip_dmatag, (map), 0, (map)->dm_mapsize, \
map               783 dev/pci/if_devar.h #define	TULIP_TXMAP_PRESYNC(sc, map) \
map               784 dev/pci/if_devar.h 	bus_dmamap_sync((sc)->tulip_dmatag, (map), 0, (map)->dm_mapsize, \
map               786 dev/pci/if_devar.h #define	TULIP_TXMAP_POSTSYNC(sc, map) \
map               787 dev/pci/if_devar.h 	bus_dmamap_sync((sc)->tulip_dmatag, (map), 0, (map)->dm_mapsize, \
map               938 dev/pci/if_em.c 	bus_dmamap_t	map;
map               972 dev/pci/if_em.c 	map = tx_buffer->map;
map               974 dev/pci/if_em.c 	error = bus_dmamap_load_mbuf(sc->txtag, map, m_head, BUS_DMA_NOWAIT);
map               979 dev/pci/if_em.c 	EM_KASSERT(map->dm_nsegs!= 0, ("em_encap: empty packet"));
map               981 dev/pci/if_em.c 	if (map->dm_nsegs > sc->num_tx_desc_avail - 2)
map               997 dev/pci/if_em.c 	for (j = 0; j < map->dm_nsegs; j++) {
map              1004 dev/pci/if_em.c 			array_elements = em_fill_descriptors(map->dm_segs[j].ds_addr,
map              1005 dev/pci/if_em.c 							     map->dm_segs[j].ds_len,
map              1032 dev/pci/if_em.c 			current_tx_desc->buffer_addr = htole64(map->dm_segs[j].ds_addr);
map              1034 dev/pci/if_em.c 				sc->txd_cmd | txd_lower | map->dm_segs[j].ds_len);
map              1049 dev/pci/if_em.c 		sc->num_tx_desc_avail -= map->dm_nsegs;
map              1052 dev/pci/if_em.c 	tx_buffer_mapped->map = tx_buffer->map;
map              1053 dev/pci/if_em.c 	tx_buffer->map = map;
map              1054 dev/pci/if_em.c 	bus_dmamap_sync(sc->txtag, map, 0, map->dm_mapsize,
map              1093 dev/pci/if_em.c 	bus_dmamap_unload(sc->txtag, map);
map              1882 dev/pci/if_em.c 			    BUS_DMA_NOWAIT, &tx_buffer->map);
map              2000 dev/pci/if_em.c 			if (tx_buffer->map != NULL &&
map              2001 dev/pci/if_em.c 			    tx_buffer->map->dm_nsegs > 0) {
map              2002 dev/pci/if_em.c 				bus_dmamap_sync(sc->txtag, tx_buffer->map,
map              2003 dev/pci/if_em.c 				    0, tx_buffer->map->dm_mapsize,
map              2006 dev/pci/if_em.c 				    tx_buffer->map);
map              2012 dev/pci/if_em.c 			if (tx_buffer->map != NULL) {
map              2014 dev/pci/if_em.c 				    tx_buffer->map);
map              2015 dev/pci/if_em.c 				tx_buffer->map = NULL;
map              2156 dev/pci/if_em.c 				if (tx_buffer->map->dm_nsegs > 0) {
map              2158 dev/pci/if_em.c 					    tx_buffer->map, 0,
map              2159 dev/pci/if_em.c 					    tx_buffer->map->dm_mapsize,
map              2162 dev/pci/if_em.c 					    tx_buffer->map);
map              2219 dev/pci/if_em.c 	bus_dmamap_t	map;
map              2252 dev/pci/if_em.c 		bus_dmamap_unload(sc->rxtag, rx_buffer->map);
map              2254 dev/pci/if_em.c 	map = rx_buffer->map;
map              2255 dev/pci/if_em.c 	rx_buffer->map = sc->rx_sparemap;
map              2256 dev/pci/if_em.c 	sc->rx_sparemap = map;
map              2258 dev/pci/if_em.c 	bus_dmamap_sync(sc->rxtag, rx_buffer->map, 0,
map              2259 dev/pci/if_em.c 	    rx_buffer->map->dm_mapsize, BUS_DMASYNC_PREREAD);
map              2263 dev/pci/if_em.c 	sc->rx_desc_base[i].buffer_addr = htole64(rx_buffer->map->dm_segs[0].ds_addr);
map              2309 dev/pci/if_em.c 					&rx_buffer->map);
map              2453 dev/pci/if_em.c 			if (rx_buffer->map != NULL &&
map              2454 dev/pci/if_em.c 			    rx_buffer->map->dm_nsegs > 0) {
map              2455 dev/pci/if_em.c 				bus_dmamap_sync(sc->rxtag, rx_buffer->map,
map              2456 dev/pci/if_em.c 				    0, rx_buffer->map->dm_mapsize,
map              2459 dev/pci/if_em.c 				    rx_buffer->map);
map              2465 dev/pci/if_em.c 			if (rx_buffer->map != NULL) {
map              2467 dev/pci/if_em.c 				    rx_buffer->map);
map              2468 dev/pci/if_em.c 				rx_buffer->map = NULL;
map              2523 dev/pci/if_em.c 		bus_dmamap_sync(sc->rxtag, sc->rx_buffer_area[i].map,
map              2524 dev/pci/if_em.c 		    0, sc->rx_buffer_area[i].map->dm_mapsize,
map               277 dev/pci/if_em.h 	bus_dmamap_t	map;		/* bus_dma map for packet */
map               451 dev/pci/if_ipw.c 		    1, sizeof (struct ipw_hdr), 0, BUS_DMA_NOWAIT, &shdr->map);
map               467 dev/pci/if_ipw.c 		    MCLBYTES, 0, BUS_DMA_NOWAIT, &sbuf->map);
map               511 dev/pci/if_ipw.c 		    0, BUS_DMA_NOWAIT, &sbuf->map);
map               518 dev/pci/if_ipw.c 		error = bus_dmamap_load(sc->sc_dmat, sbuf->map,
map               528 dev/pci/if_ipw.c 		sbd->bd->physaddr = htole32(sbuf->map->dm_segs[0].ds_addr);
map               581 dev/pci/if_ipw.c 		bus_dmamap_destroy(sc->sc_dmat, sc->shdr_list[i].map);
map               584 dev/pci/if_ipw.c 		bus_dmamap_destroy(sc->sc_dmat, sc->tx_sbuf_list[i].map);
map               588 dev/pci/if_ipw.c 		if (sbuf->map != NULL) {
map               590 dev/pci/if_ipw.c 				bus_dmamap_unload(sc->sc_dmat, sbuf->map);
map               593 dev/pci/if_ipw.c 			bus_dmamap_destroy(sc->sc_dmat, sbuf->map);
map               761 dev/pci/if_ipw.c 	bus_dmamap_sync(sc->sc_dmat, sbuf->map, 0, sizeof (struct ipw_cmd),
map               780 dev/pci/if_ipw.c 	bus_dmamap_sync(sc->sc_dmat, sbuf->map, 0, sizeof state,
map               847 dev/pci/if_ipw.c 	bus_dmamap_sync(sc->sc_dmat, sbuf->map, 0, letoh32(status->len),
map               849 dev/pci/if_ipw.c 	bus_dmamap_unload(sc->sc_dmat, sbuf->map);
map               851 dev/pci/if_ipw.c 	error = bus_dmamap_load(sc->sc_dmat, sbuf->map, mtod(mnew, void *),
map               857 dev/pci/if_ipw.c 		error = bus_dmamap_load(sc->sc_dmat, sbuf->map,
map               870 dev/pci/if_ipw.c 	sbd->bd->physaddr = htole32(sbuf->map->dm_segs[0].ds_addr);
map               984 dev/pci/if_ipw.c 		bus_dmamap_unload(sc->sc_dmat, shdr->map);
map               990 dev/pci/if_ipw.c 		bus_dmamap_unload(sc->sc_dmat, sbuf->map);
map              1180 dev/pci/if_ipw.c 	error = bus_dmamap_load_mbuf(sc->sc_dmat, sbuf->map, m, BUS_DMA_NOWAIT);
map              1211 dev/pci/if_ipw.c 		error = bus_dmamap_load_mbuf(sc->sc_dmat, sbuf->map, m,
map              1221 dev/pci/if_ipw.c 	error = bus_dmamap_load(sc->sc_dmat, shdr->map, &shdr->hdr,
map              1226 dev/pci/if_ipw.c 		bus_dmamap_unload(sc->sc_dmat, sbuf->map);
map              1237 dev/pci/if_ipw.c 	sbd->bd->physaddr = htole32(shdr->map->dm_segs[0].ds_addr);
map              1239 dev/pci/if_ipw.c 	sbd->bd->nfrag = 1 + sbuf->map->dm_nsegs;
map              1258 dev/pci/if_ipw.c 	for (i = 0; i < sbuf->map->dm_nsegs; i++) {
map              1260 dev/pci/if_ipw.c 		sbd->bd->physaddr = htole32(sbuf->map->dm_segs[i].ds_addr);
map              1261 dev/pci/if_ipw.c 		sbd->bd->len = htole32(sbuf->map->dm_segs[i].ds_len);
map              1264 dev/pci/if_ipw.c 		if (i == sbuf->map->dm_nsegs - 1) {
map              1274 dev/pci/if_ipw.c 		    sbuf->map->dm_segs[i].ds_len));
map              1284 dev/pci/if_ipw.c 	bus_dmamap_sync(sc->sc_dmat, sbuf->map, 0, sbuf->map->dm_mapsize,
map              1286 dev/pci/if_ipw.c 	bus_dmamap_sync(sc->sc_dmat, shdr->map, 0, sizeof (struct ipw_hdr),
map                50 dev/pci/if_ipwvar.h 	bus_dmamap_t			map;
map                57 dev/pci/if_ipwvar.h 	bus_dmamap_t			map;
map               400 dev/pci/if_iwi.c 	    BUS_DMA_NOWAIT, &ring->map);
map               425 dev/pci/if_iwi.c 	error = bus_dmamap_load(sc->sc_dmat, ring->map, ring->desc,
map               451 dev/pci/if_iwi.c 	if (ring->map != NULL) {
map               453 dev/pci/if_iwi.c 			bus_dmamap_unload(sc->sc_dmat, ring->map);
map               458 dev/pci/if_iwi.c 		bus_dmamap_destroy(sc->sc_dmat, ring->map);
map               477 dev/pci/if_iwi.c 	    &ring->map);
map               502 dev/pci/if_iwi.c 	error = bus_dmamap_load(sc->sc_dmat, ring->map, ring->desc,
map               517 dev/pci/if_iwi.c 		    IWI_MAX_SCATTER, MCLBYTES, 0, BUS_DMA_NOWAIT, &data->map);
map               541 dev/pci/if_iwi.c 			bus_dmamap_unload(sc->sc_dmat, data->map);
map               557 dev/pci/if_iwi.c 	if (ring->map != NULL) {
map               559 dev/pci/if_iwi.c 			bus_dmamap_unload(sc->sc_dmat, ring->map);
map               564 dev/pci/if_iwi.c 		bus_dmamap_destroy(sc->sc_dmat, ring->map);
map               571 dev/pci/if_iwi.c 			bus_dmamap_unload(sc->sc_dmat, data->map);
map               574 dev/pci/if_iwi.c 		bus_dmamap_destroy(sc->sc_dmat, data->map);
map               590 dev/pci/if_iwi.c 		    0, BUS_DMA_NOWAIT, &data->map);
map               615 dev/pci/if_iwi.c 		error = bus_dmamap_load(sc->sc_dmat, data->map,
map               648 dev/pci/if_iwi.c 			bus_dmamap_unload(sc->sc_dmat, data->map);
map               651 dev/pci/if_iwi.c 		bus_dmamap_destroy(sc->sc_dmat, data->map);
map               904 dev/pci/if_iwi.c 	bus_dmamap_unload(sc->sc_dmat, data->map);
map               906 dev/pci/if_iwi.c 	error = bus_dmamap_load(sc->sc_dmat, data->map, mtod(mnew, void *),
map               912 dev/pci/if_iwi.c 		error = bus_dmamap_load(sc->sc_dmat, data->map,
map               925 dev/pci/if_iwi.c 	CSR_WRITE_4(sc, data->reg, data->map->dm_segs[0].ds_addr);
map              1104 dev/pci/if_iwi.c 		bus_dmamap_sync(sc->sc_dmat, data->map, 0, MCLBYTES,
map              1146 dev/pci/if_iwi.c 		bus_dmamap_unload(sc->sc_dmat, data->map);
map              1238 dev/pci/if_iwi.c 	bus_dmamap_sync(sc->sc_dmat, sc->cmdq.map,
map              1303 dev/pci/if_iwi.c 	error = bus_dmamap_load_mbuf(sc->sc_dmat, data->map, m0,
map              1335 dev/pci/if_iwi.c 		error = bus_dmamap_load_mbuf(sc->sc_dmat, data->map, m0,
map              1374 dev/pci/if_iwi.c 	desc->nseg = htole32(data->map->dm_nsegs);
map              1375 dev/pci/if_iwi.c 	for (i = 0; i < data->map->dm_nsegs; i++) {
map              1376 dev/pci/if_iwi.c 		desc->seg_addr[i] = htole32(data->map->dm_segs[i].ds_addr);
map              1377 dev/pci/if_iwi.c 		desc->seg_len[i]  = htole16(data->map->dm_segs[i].ds_len);
map              1380 dev/pci/if_iwi.c 	bus_dmamap_sync(sc->sc_dmat, data->map, 0, data->map->dm_mapsize,
map              1382 dev/pci/if_iwi.c 	bus_dmamap_sync(sc->sc_dmat, txq->map,
map              1387 dev/pci/if_iwi.c 	    letoh16(desc->len), data->map->dm_nsegs));
map              1668 dev/pci/if_iwi.c 	bus_dmamap_t map;
map              1677 dev/pci/if_iwi.c 	    BUS_DMA_NOWAIT, &map);
map              1700 dev/pci/if_iwi.c 	error = bus_dmamap_load(sc->sc_dmat, map, virtaddr, size, NULL,
map              1712 dev/pci/if_iwi.c 	bus_dmamap_sync(sc->sc_dmat, map, 0, size, BUS_DMASYNC_PREWRITE);
map              1722 dev/pci/if_iwi.c 	src = map->dm_segs[0].ds_addr;
map              1791 dev/pci/if_iwi.c fail5:	bus_dmamap_sync(sc->sc_dmat, map, 0, size, BUS_DMASYNC_POSTWRITE);
map              1792 dev/pci/if_iwi.c 	bus_dmamap_unload(sc->sc_dmat, map);
map              1795 dev/pci/if_iwi.c fail2:	bus_dmamap_destroy(sc->sc_dmat, map);
map              2200 dev/pci/if_iwi.c 	CSR_WRITE_4(sc, IWI_CSR_CMD_BASE, sc->cmdq.map->dm_segs[0].ds_addr);
map              2204 dev/pci/if_iwi.c 	CSR_WRITE_4(sc, IWI_CSR_TX1_BASE, sc->txq[0].map->dm_segs[0].ds_addr);
map              2208 dev/pci/if_iwi.c 	CSR_WRITE_4(sc, IWI_CSR_TX2_BASE, sc->txq[1].map->dm_segs[0].ds_addr);
map              2212 dev/pci/if_iwi.c 	CSR_WRITE_4(sc, IWI_CSR_TX3_BASE, sc->txq[2].map->dm_segs[0].ds_addr);
map              2216 dev/pci/if_iwi.c 	CSR_WRITE_4(sc, IWI_CSR_TX4_BASE, sc->txq[3].map->dm_segs[0].ds_addr);
map              2222 dev/pci/if_iwi.c 		CSR_WRITE_4(sc, data->reg, data->map->dm_segs[0].ds_addr);
map                60 dev/pci/if_iwivar.h 	bus_dmamap_t		map;
map                69 dev/pci/if_iwivar.h 	bus_dmamap_t		map;
map                75 dev/pci/if_iwivar.h 	bus_dmamap_t		map;
map                87 dev/pci/if_iwivar.h 	bus_dmamap_t		map;
map               658 dev/pci/if_ixgb.c 	bus_dmamap_t	map;
map               680 dev/pci/if_ixgb.c 	map = tx_buffer->map;
map               682 dev/pci/if_ixgb.c 	error = bus_dmamap_load_mbuf(sc->txtag, map,
map               688 dev/pci/if_ixgb.c 	IXGB_KASSERT(map->dm_nsegs != 0, ("ixgb_encap: empty packet"));
map               690 dev/pci/if_ixgb.c 	if (map->dm_nsegs > sc->num_tx_desc_avail)
map               700 dev/pci/if_ixgb.c 	for (j = 0; j < map->dm_nsegs; j++) {
map               704 dev/pci/if_ixgb.c 		current_tx_desc->buff_addr = htole64(map->dm_segs[j].ds_addr);
map               705 dev/pci/if_ixgb.c 		current_tx_desc->cmd_type_len = htole32((sc->txd_cmd | map->dm_segs[j].ds_len));
map               713 dev/pci/if_ixgb.c 	sc->num_tx_desc_avail -= map->dm_nsegs;
map               717 dev/pci/if_ixgb.c 	bus_dmamap_sync(sc->txtag, map, 0, map->dm_mapsize,
map               738 dev/pci/if_ixgb.c 	bus_dmamap_unload(sc->txtag, map);
map              1194 dev/pci/if_ixgb.c 			    BUS_DMA_NOWAIT, &tx_buffer->map);
map              1278 dev/pci/if_ixgb.c 			if (tx_buffer->map != NULL &&
map              1279 dev/pci/if_ixgb.c 			    tx_buffer->map->dm_nsegs > 0) {
map              1280 dev/pci/if_ixgb.c 				bus_dmamap_sync(sc->txtag, tx_buffer->map,
map              1281 dev/pci/if_ixgb.c 				    0, tx_buffer->map->dm_mapsize,
map              1284 dev/pci/if_ixgb.c 				    tx_buffer->map);
map              1291 dev/pci/if_ixgb.c 			if (tx_buffer->map != NULL) {
map              1293 dev/pci/if_ixgb.c 				    tx_buffer->map);
map              1294 dev/pci/if_ixgb.c 				tx_buffer->map = NULL;
map              1415 dev/pci/if_ixgb.c 			if (tx_buffer->map->dm_nsegs > 0) {
map              1416 dev/pci/if_ixgb.c 				bus_dmamap_sync(sc->txtag, tx_buffer->map,
map              1417 dev/pci/if_ixgb.c 				    0, tx_buffer->map->dm_mapsize,
map              1419 dev/pci/if_ixgb.c 				bus_dmamap_unload(sc->txtag, tx_buffer->map);
map              1500 dev/pci/if_ixgb.c 	error = bus_dmamap_load_mbuf(sc->rxtag, rx_buffer->map,
map              1507 dev/pci/if_ixgb.c 	sc->rx_desc_base[i].buff_addr = htole64(rx_buffer->map->dm_segs[0].ds_addr);
map              1508 dev/pci/if_ixgb.c 	bus_dmamap_sync(sc->rxtag, rx_buffer->map, 0,
map              1509 dev/pci/if_ixgb.c 	    rx_buffer->map->dm_mapsize, BUS_DMASYNC_PREREAD);
map              1546 dev/pci/if_ixgb.c 					  &rx_buffer->map);
map              1701 dev/pci/if_ixgb.c 			if (rx_buffer->map != NULL &&
map              1702 dev/pci/if_ixgb.c 			    rx_buffer->map->dm_nsegs > 0) {
map              1703 dev/pci/if_ixgb.c 				bus_dmamap_sync(sc->rxtag, rx_buffer->map,
map              1704 dev/pci/if_ixgb.c 				    0, rx_buffer->map->dm_mapsize,
map              1707 dev/pci/if_ixgb.c 				    rx_buffer->map);
map              1713 dev/pci/if_ixgb.c 			if (rx_buffer->map != NULL) {
map              1715 dev/pci/if_ixgb.c 				    rx_buffer->map);
map              1716 dev/pci/if_ixgb.c 				rx_buffer->map = NULL;
map              1769 dev/pci/if_ixgb.c 		bus_dmamap_sync(sc->rxtag, sc->rx_buffer_area[i].map,
map              1770 dev/pci/if_ixgb.c 		    0, sc->rx_buffer_area[i].map->dm_mapsize,
map              1772 dev/pci/if_ixgb.c 		bus_dmamap_unload(sc->rxtag, sc->rx_buffer_area[i].map);
map               207 dev/pci/if_ixgb.h 	bus_dmamap_t    map;	/* bus_dma map for packet */
map               510 dev/pci/if_lmc.c 		bus_dmamap_t map;
map               575 dev/pci/if_lmc.c 				map = LMC_GETCTX(me, bus_dmamap_t);
map               576 dev/pci/if_lmc.c 				LMC_RXMAP_POSTSYNC(sc, map);
map               577 dev/pci/if_lmc.c 				bus_dmamap_unload(sc->lmc_dmatag, map);
map               578 dev/pci/if_lmc.c 				sc->lmc_rxmaps[sc->lmc_rxmaps_free++] = map;
map               606 dev/pci/if_lmc.c 			map = LMC_GETCTX(me, bus_dmamap_t);
map               607 dev/pci/if_lmc.c 			bus_dmamap_sync(sc->lmc_dmatag, map, 0, me->m_len,
map               609 dev/pci/if_lmc.c 			bus_dmamap_unload(sc->lmc_dmatag, map);
map               610 dev/pci/if_lmc.c 			sc->lmc_rxmaps[sc->lmc_rxmaps_free++] = map;
map               632 dev/pci/if_lmc.c 			map = LMC_GETCTX(me, bus_dmamap_t);
map               633 dev/pci/if_lmc.c 			bus_dmamap_unload(sc->lmc_dmatag, map);
map               634 dev/pci/if_lmc.c 			sc->lmc_rxmaps[sc->lmc_rxmaps_free++] = map;
map               694 dev/pci/if_lmc.c 				map = sc->lmc_rxmaps[--sc->lmc_rxmaps_free];
map               703 dev/pci/if_lmc.c 			LMC_SETCTX(ms, map);
map               704 dev/pci/if_lmc.c 			error = bus_dmamap_load(sc->lmc_dmatag, map,
map               720 dev/pci/if_lmc.c 			nextout->d_addr1 = htole32(map->dm_segs[0].ds_addr);
map               721 dev/pci/if_lmc.c 			if (map->dm_nsegs == 2) {
map               722 dev/pci/if_lmc.c 				nextout->d_addr2 = htole32(map->dm_segs[1].ds_addr);
map               725 dev/pci/if_lmc.c 						map->dm_segs[0].ds_len,
map               726 dev/pci/if_lmc.c 						map->dm_segs[1].ds_len));
map               731 dev/pci/if_lmc.c 						map->dm_segs[0].ds_len, 0));
map               771 dev/pci/if_lmc.c 		    bus_dmamap_t map = LMC_GETCTX(m, bus_dmamap_t);
map               772 dev/pci/if_lmc.c 		    LMC_TXMAP_POSTSYNC(sc, map);
map               773 dev/pci/if_lmc.c 		    sc->lmc_txmaps[sc->lmc_txmaps_free++] = map;
map               991 dev/pci/if_lmc.c 	bus_dmamap_t map;
map              1034 dev/pci/if_lmc.c 		map = sc->lmc_txmaps[sc->lmc_txmaps_free-1];
map              1042 dev/pci/if_lmc.c 	error = bus_dmamap_load_mbuf(sc->lmc_dmatag, map, m, BUS_DMA_NOWAIT);
map              1057 dev/pci/if_lmc.c 			error = bus_dmamap_load_mbuf(sc->lmc_dmatag, map, m,
map              1069 dev/pci/if_lmc.c 	if ((free -= (map->dm_nsegs + 1) / 2) <= 0
map              1084 dev/pci/if_lmc.c 		bus_dmamap_unload(sc->lmc_dmatag, map);
map              1087 dev/pci/if_lmc.c 	for (; map->dm_nsegs - segcnt > 1; segcnt += 2) {
map              1097 dev/pci/if_lmc.c 		eop->d_addr1   = htole32(map->dm_segs[segcnt].ds_addr);
map              1098 dev/pci/if_lmc.c 		eop->d_addr2   = htole32(map->dm_segs[segcnt+1].ds_addr);
map              1100 dev/pci/if_lmc.c 				 map->dm_segs[segcnt].ds_len,
map              1101 dev/pci/if_lmc.c 				 map->dm_segs[segcnt+1].ds_len));
map              1106 dev/pci/if_lmc.c 	if (segcnt < map->dm_nsegs) {
map              1116 dev/pci/if_lmc.c 		eop->d_addr1   = htole32(map->dm_segs[segcnt].ds_addr);
map              1119 dev/pci/if_lmc.c 				 map->dm_segs[segcnt].ds_len, 0));
map              1123 dev/pci/if_lmc.c 	LMC_TXMAP_PRESYNC(sc, map);
map              1124 dev/pci/if_lmc.c 	LMC_SETCTX(m, map);
map              1125 dev/pci/if_lmc.c 	map = NULL;
map               271 dev/pci/if_lmc_common.c 		bus_dmamap_t map;
map               277 dev/pci/if_lmc_common.c 		map = LMC_GETCTX(m, bus_dmamap_t);
map               278 dev/pci/if_lmc_common.c 		bus_dmamap_unload(sc->lmc_dmatag, map);
map               279 dev/pci/if_lmc_common.c 		sc->lmc_txmaps[sc->lmc_txmaps_free++] = map;
map               315 dev/pci/if_lmc_common.c 		bus_dmamap_t map;
map               320 dev/pci/if_lmc_common.c 		map = LMC_GETCTX(m, bus_dmamap_t);
map               321 dev/pci/if_lmc_common.c 		bus_dmamap_unload(sc->lmc_dmatag, map);
map               322 dev/pci/if_lmc_common.c 		sc->lmc_rxmaps[sc->lmc_rxmaps_free++] = map;
map               349 dev/pci/if_lmc_obsd.c 	    bus_dmamap_t map;
map               351 dev/pci/if_lmc_obsd.c 				      BUS_DMA_NOWAIT, &map);
map               353 dev/pci/if_lmc_obsd.c 		error = bus_dmamap_load(sc->lmc_dmatag, map, desc,
map               356 dev/pci/if_lmc_obsd.c 		    bus_dmamap_destroy(sc->lmc_dmatag, map);
map               358 dev/pci/if_lmc_obsd.c 		    *map_p = map;
map               391 dev/pci/if_lmc_obsd.c 	    bus_dmamap_t map;
map               392 dev/pci/if_lmc_obsd.c 	    if ((error = LMC_TXMAP_CREATE(sc, &map)) == 0)
map               393 dev/pci/if_lmc_obsd.c 		sc->lmc_txmaps[sc->lmc_txmaps_free++] = map;
map               416 dev/pci/if_lmc_obsd.c 	    bus_dmamap_t map;
map               417 dev/pci/if_lmc_obsd.c 	    if ((error = LMC_RXMAP_CREATE(sc, &map)) == 0)
map               418 dev/pci/if_lmc_obsd.c 		sc->lmc_rxmaps[sc->lmc_rxmaps_free++] = map;
map               463 dev/pci/if_lmcvar.h #define	LMC_RXMAP_PRESYNC(sc, map) \
map               464 dev/pci/if_lmcvar.h 	bus_dmamap_sync((sc)->lmc_dmatag, (map), 0, (map)->dm_mapsize, \
map               466 dev/pci/if_lmcvar.h #define	LMC_RXMAP_POSTSYNC(sc, map) \
map               467 dev/pci/if_lmcvar.h 	bus_dmamap_sync((sc)->lmc_dmatag, (map), 0, (map)->dm_mapsize, \
map               482 dev/pci/if_lmcvar.h #define	LMC_TXMAP_PRESYNC(sc, map) \
map               483 dev/pci/if_lmcvar.h 	bus_dmamap_sync((sc)->lmc_dmatag, (map), 0, (map)->dm_mapsize, \
map               485 dev/pci/if_lmcvar.h #define	LMC_TXMAP_POSTSYNC(sc, map) \
map               486 dev/pci/if_lmcvar.h 	bus_dmamap_sync((sc)->lmc_dmatag, (map), 0, (map)->dm_mapsize, \
map               554 dev/pci/if_myx.c 	bus_dmamap_t		 map = sc->sc_cmddma.mxm_map;
map               598 dev/pci/if_myx.c 	mc->mc_addr_high = htobe32(MYX_ADDRHIGH(map->dm_segs[0].ds_addr));
map               599 dev/pci/if_myx.c 	mc->mc_addr_low = htobe32(MYX_ADDRLOW(map->dm_segs[0].ds_addr));
map               608 dev/pci/if_myx.c 		bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map               633 dev/pci/if_myx.c 	bus_dmamap_t		 map = sc->sc_cmddma.mxm_map;
map               637 dev/pci/if_myx.c 	bc->bc_addr_high = htobe32(MYX_ADDRHIGH(map->dm_segs[0].ds_addr));
map               638 dev/pci/if_myx.c 	bc->bc_addr_low = htobe32(MYX_ADDRLOW(map->dm_segs[0].ds_addr));
map               652 dev/pci/if_myx.c 		bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map               672 dev/pci/if_myx.c 	bus_dmamap_t		 map = sc->sc_cmddma.mxm_map;
map               681 dev/pci/if_myx.c 	rc.rc_addr_high = htobe32(MYX_ADDRHIGH(map->dm_segs[0].ds_addr));
map               682 dev/pci/if_myx.c 	rc.rc_addr_low = htobe32(MYX_ADDRLOW(map->dm_segs[0].ds_addr));
map               695 dev/pci/if_myx.c 		bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map               963 dev/pci/if_myx.c 	bus_dmamap_t		 map = sc->sc_stsdma.mxm_map;
map               968 dev/pci/if_myx.c 	bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map              1026 dev/pci/if_myx.c 	bus_dmamap_t		 map;
map              1077 dev/pci/if_myx.c 	map = sc->sc_rxdma.mxm_map;
map              1078 dev/pci/if_myx.c 	mc.mc_data0 = MYX_ADDRLOW(map->dm_segs[0].ds_addr);
map              1079 dev/pci/if_myx.c 	mc.mc_data1 = MYX_ADDRHIGH(map->dm_segs[0].ds_addr);
map              1132 dev/pci/if_myx.c 		map = mb->mb_dmamap;
map              1133 dev/pci/if_myx.c 		mb->mb_m = myx_getbuf(sc, map, 1);
map              1135 dev/pci/if_myx.c 			bus_dmamap_destroy(sc->sc_dmat, map);
map              1139 dev/pci/if_myx.c 		bus_dmamap_sync(sc->sc_dmat, map, 0,
map              1143 dev/pci/if_myx.c 		    htobe32(MYX_ADDRHIGH(map->dm_segs[0].ds_addr));
map              1145 dev/pci/if_myx.c 		    htobe32(MYX_ADDRHIGH(map->dm_segs[0].ds_addr));
map              1163 dev/pci/if_myx.c 		map = mb->mb_dmamap;
map              1164 dev/pci/if_myx.c 		mb->mb_m = myx_getbuf(sc, map, 1);
map              1166 dev/pci/if_myx.c 			bus_dmamap_destroy(sc->sc_dmat, map);
map              1170 dev/pci/if_myx.c 		bus_dmamap_sync(sc->sc_dmat, map, 0,
map              1174 dev/pci/if_myx.c 		    htobe32(MYX_ADDRHIGH(map->dm_segs[0].ds_addr));
map              1176 dev/pci/if_myx.c 		    htobe32(MYX_ADDRHIGH(map->dm_segs[0].ds_addr));
map              1199 dev/pci/if_myx.c 	map = sc->sc_stsdma.mxm_map;
map              1202 dev/pci/if_myx.c 	mc.mc_data0 = MYX_ADDRLOW(map->dm_segs[0].ds_addr);
map              1203 dev/pci/if_myx.c 	mc.mc_data1 = MYX_ADDRHIGH(map->dm_segs[0].ds_addr);
map              1210 dev/pci/if_myx.c 	bus_dmamap_sync(sc->sc_dmat, map, 0,
map              1211 dev/pci/if_myx.c 	    map->dm_mapsize, BUS_DMASYNC_PREWRITE);
map              1250 dev/pci/if_myx.c myx_getbuf(struct myx_softc *sc, bus_dmamap_t map, int wait)
map              1263 dev/pci/if_myx.c 	if (bus_dmamap_load_mbuf(sc->sc_dmat, map, m,
map               561 dev/pci/if_nfe.c 	bus_dmamap_sync(sc->sc_dmat, sc->txq.map,
map               569 dev/pci/if_nfe.c 	bus_dmamap_sync(sc->sc_dmat, sc->txq.map,
map               578 dev/pci/if_nfe.c 		bus_dmamap_sync(sc->sc_dmat, sc->txq.map,
map               585 dev/pci/if_nfe.c 	bus_dmamap_sync(sc->sc_dmat, sc->txq.map,
map               591 dev/pci/if_nfe.c 	bus_dmamap_sync(sc->sc_dmat, sc->txq.map, 0,
map               599 dev/pci/if_nfe.c 		bus_dmamap_sync(sc->sc_dmat, sc->txq.map,
map               606 dev/pci/if_nfe.c 	bus_dmamap_sync(sc->sc_dmat, sc->txq.map,
map               612 dev/pci/if_nfe.c 	bus_dmamap_sync(sc->sc_dmat, sc->txq.map, 0,
map               619 dev/pci/if_nfe.c 	bus_dmamap_sync(sc->sc_dmat, sc->rxq.map,
map               627 dev/pci/if_nfe.c 	bus_dmamap_sync(sc->sc_dmat, sc->rxq.map,
map               722 dev/pci/if_nfe.c 			bus_dmamap_sync(sc->sc_dmat, data->map, 0,
map               723 dev/pci/if_nfe.c 			    data->map->dm_mapsize, BUS_DMASYNC_POSTREAD);
map               724 dev/pci/if_nfe.c 			bus_dmamap_unload(sc->sc_dmat, data->map);
map               726 dev/pci/if_nfe.c 			error = bus_dmamap_load(sc->sc_dmat, data->map,
map               733 dev/pci/if_nfe.c 				error = bus_dmamap_load(sc->sc_dmat, data->map,
map               744 dev/pci/if_nfe.c 			physaddr = data->map->dm_segs[0].ds_addr;
map               880 dev/pci/if_nfe.c 	bus_dmamap_t map;
map               887 dev/pci/if_nfe.c 	map = sc->txq.data[first].map;
map               889 dev/pci/if_nfe.c 	error = bus_dmamap_load_mbuf(sc->sc_dmat, map, m0, BUS_DMA_NOWAIT);
map               896 dev/pci/if_nfe.c 	if (sc->txq.queued + map->dm_nsegs >= NFE_TX_RING_COUNT - 1) {
map               897 dev/pci/if_nfe.c 		bus_dmamap_unload(sc->sc_dmat, map);
map               914 dev/pci/if_nfe.c 	for (i = 0; i < map->dm_nsegs; i++) {
map               921 dev/pci/if_nfe.c 			    htole32(map->dm_segs[i].ds_addr >> 32);
map               924 dev/pci/if_nfe.c 			    htole32(map->dm_segs[i].ds_addr & 0xffffffff);
map               925 dev/pci/if_nfe.c 			desc64->length = htole16(map->dm_segs[i].ds_len - 1);
map               933 dev/pci/if_nfe.c 			desc32->physaddr = htole32(map->dm_segs[i].ds_addr);
map               934 dev/pci/if_nfe.c 			desc32->length = htole16(map->dm_segs[i].ds_len - 1);
map               938 dev/pci/if_nfe.c 		if (map->dm_nsegs > 1) {
map               979 dev/pci/if_nfe.c 	data->active = map;
map               981 dev/pci/if_nfe.c 	bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map              1201 dev/pci/if_nfe.c 	    NFE_RX_RING_COUNT * descsize, 0, BUS_DMA_NOWAIT, &ring->map);
map              1224 dev/pci/if_nfe.c 	error = bus_dmamap_load(sc->sc_dmat, ring->map, *desc,
map              1233 dev/pci/if_nfe.c 	ring->physaddr = ring->map->dm_segs[0].ds_addr;
map              1270 dev/pci/if_nfe.c 			    MCLBYTES, 0, BUS_DMA_NOWAIT, &data->map);
map              1284 dev/pci/if_nfe.c 			error = bus_dmamap_load(sc->sc_dmat, data->map,
map              1292 dev/pci/if_nfe.c 			physaddr = data->map->dm_segs[0].ds_addr;
map              1311 dev/pci/if_nfe.c 	bus_dmamap_sync(sc->sc_dmat, ring->map, 0, ring->map->dm_mapsize,
map              1335 dev/pci/if_nfe.c 	bus_dmamap_sync(sc->sc_dmat, ring->map, 0, ring->map->dm_mapsize,
map              1357 dev/pci/if_nfe.c 		bus_dmamap_sync(sc->sc_dmat, ring->map, 0,
map              1358 dev/pci/if_nfe.c 		    ring->map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
map              1359 dev/pci/if_nfe.c 		bus_dmamap_unload(sc->sc_dmat, ring->map);
map              1368 dev/pci/if_nfe.c 		if (data->map != NULL) {
map              1369 dev/pci/if_nfe.c 			bus_dmamap_sync(sc->sc_dmat, data->map, 0,
map              1370 dev/pci/if_nfe.c 			    data->map->dm_mapsize, BUS_DMASYNC_POSTREAD);
map              1371 dev/pci/if_nfe.c 			bus_dmamap_unload(sc->sc_dmat, data->map);
map              1372 dev/pci/if_nfe.c 			bus_dmamap_destroy(sc->sc_dmat, data->map);
map              1519 dev/pci/if_nfe.c 	    NFE_TX_RING_COUNT * descsize, 0, BUS_DMA_NOWAIT, &ring->map);
map              1543 dev/pci/if_nfe.c 	error = bus_dmamap_load(sc->sc_dmat, ring->map, *desc,
map              1552 dev/pci/if_nfe.c 	ring->physaddr = ring->map->dm_segs[0].ds_addr;
map              1557 dev/pci/if_nfe.c 		    &ring->data[i].map);
map              1594 dev/pci/if_nfe.c 	bus_dmamap_sync(sc->sc_dmat, ring->map, 0, ring->map->dm_mapsize,
map              1617 dev/pci/if_nfe.c 		bus_dmamap_sync(sc->sc_dmat, ring->map, 0,
map              1618 dev/pci/if_nfe.c 		    ring->map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
map              1619 dev/pci/if_nfe.c 		bus_dmamap_unload(sc->sc_dmat, ring->map);
map              1639 dev/pci/if_nfe.c 		if (data->map == NULL)
map              1641 dev/pci/if_nfe.c 		bus_dmamap_destroy(sc->sc_dmat, data->map);
map                22 dev/pci/if_nfevar.h 	bus_dmamap_t	map;
map                28 dev/pci/if_nfevar.h 	bus_dmamap_t		map;
map                46 dev/pci/if_nfevar.h 	bus_dmamap_t	map;
map                51 dev/pci/if_nfevar.h 	bus_dmamap_t		map;
map              1110 dev/pci/if_sis.c 		    BUS_DMA_NOWAIT, &sc->sis_ldata->sis_rx_list[i].map) != 0) {
map              1124 dev/pci/if_sis.c 		    &sc->sis_ldata->sis_tx_list[i].map) != 0) {
map              1239 dev/pci/if_sis.c 	bus_dmamap_t		map;
map              1267 dev/pci/if_sis.c 	map = c->map;
map              1268 dev/pci/if_sis.c 	c->map = sc->sc_rx_sparemap;
map              1269 dev/pci/if_sis.c 	sc->sc_rx_sparemap = map;
map              1271 dev/pci/if_sis.c 	bus_dmamap_sync(sc->sc_dmat, c->map, 0, c->map->dm_mapsize,
map              1277 dev/pci/if_sis.c 	c->sis_ptr = c->map->dm_segs[0].ds_addr + sizeof(u_int64_t);
map              1330 dev/pci/if_sis.c 		bus_dmamap_sync(sc->sc_dmat, cur_rx->map, 0,
map              1331 dev/pci/if_sis.c 		    cur_rx->map->dm_mapsize, BUS_DMASYNC_POSTREAD);
map              1425 dev/pci/if_sis.c 		if (cur_tx->map->dm_nsegs != 0) {
map              1426 dev/pci/if_sis.c 			bus_dmamap_t map = cur_tx->map;
map              1428 dev/pci/if_sis.c 			bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map              1430 dev/pci/if_sis.c 			bus_dmamap_unload(sc->sc_dmat, map);
map              1547 dev/pci/if_sis.c 	bus_dmamap_t		map;
map              1549 dev/pci/if_sis.c 	map = sc->sc_tx_sparemap;
map              1550 dev/pci/if_sis.c 	if (bus_dmamap_load_mbuf(sc->sc_dmat, map,
map              1561 dev/pci/if_sis.c 	for (i = 0; i < map->dm_nsegs; i++) {
map              1565 dev/pci/if_sis.c 		f->sis_ctl = SIS_CMDSTS_MORE | map->dm_segs[i].ds_len;
map              1566 dev/pci/if_sis.c 		f->sis_ptr = map->dm_segs[i].ds_addr;
map              1573 dev/pci/if_sis.c 	bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map              2033 dev/pci/if_sis.c 		if (sc->sis_ldata->sis_rx_list[i].map->dm_nsegs != 0) {
map              2034 dev/pci/if_sis.c 			bus_dmamap_t map = sc->sis_ldata->sis_rx_list[i].map;
map              2036 dev/pci/if_sis.c 			bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map              2038 dev/pci/if_sis.c 			bus_dmamap_unload(sc->sc_dmat, map);
map              2052 dev/pci/if_sis.c 		if (sc->sis_ldata->sis_tx_list[i].map->dm_nsegs != 0) {
map              2053 dev/pci/if_sis.c 			bus_dmamap_t map = sc->sis_ldata->sis_tx_list[i].map;
map              2055 dev/pci/if_sis.c 			bus_dmamap_sync(sc->sc_dmat, map, 0, map->dm_mapsize,
map              2057 dev/pci/if_sis.c 			bus_dmamap_unload(sc->sc_dmat, map);
map               333 dev/pci/if_sisreg.h 	bus_dmamap_t		map;
map              1269 dev/pci/if_vic.c vic_alloc_mbuf(struct vic_softc *sc, bus_dmamap_t map)
map              1284 dev/pci/if_vic.c 	if (bus_dmamap_load_mbuf(sc->sc_dmat, map, m, BUS_DMA_NOWAIT) != 0) {
map              1636 dev/pci/if_vr.c 	bus_dmamap_t	map;
map              1662 dev/pci/if_vr.c 		map = sc->vr_cdata.vr_rx_chain[i].vr_map;
map              1663 dev/pci/if_vr.c 		if (map != NULL) {
map              1664 dev/pci/if_vr.c 			if (map->dm_nsegs > 0)
map              1665 dev/pci/if_vr.c 				bus_dmamap_unload(sc->sc_dmat, map);
map              1666 dev/pci/if_vr.c 			bus_dmamap_destroy(sc->sc_dmat, map);
map              1677 dev/pci/if_vr.c 		bus_dmamap_t map;
map              1683 dev/pci/if_vr.c 		map = sc->vr_cdata.vr_tx_chain[i].vr_map;
map              1684 dev/pci/if_vr.c 		if (map != NULL) {
map              1685 dev/pci/if_vr.c 			if (map->dm_nsegs > 0)
map              1686 dev/pci/if_vr.c 				bus_dmamap_unload(sc->sc_dmat, map);
map              1687 dev/pci/if_vr.c 			bus_dmamap_destroy(sc->sc_dmat, map);
map               391 dev/pci/if_wpi.c 	error = bus_dmamap_create(tag, size, 1, size, 0, flags, &dma->map);
map               404 dev/pci/if_wpi.c 	error = bus_dmamap_load_raw(tag, dma->map, &dma->seg, 1, size, flags);
map               410 dev/pci/if_wpi.c 	dma->paddr = dma->map->dm_segs[0].ds_addr;
map               423 dev/pci/if_wpi.c 	if (dma->map != NULL) {
map               425 dev/pci/if_wpi.c 			bus_dmamap_unload(dma->tag, dma->map);
map               430 dev/pci/if_wpi.c 		bus_dmamap_destroy(dma->tag, dma->map);
map               431 dev/pci/if_wpi.c 		dma->map = NULL;
map               665 dev/pci/if_wpi.c 		    &data->map);
map               704 dev/pci/if_wpi.c 			bus_dmamap_unload(sc->sc_dmat, data->map);
map               727 dev/pci/if_wpi.c 				bus_dmamap_unload(sc->sc_dmat, data->map);
map              1348 dev/pci/if_wpi.c 	bus_dmamap_unload(sc->sc_dmat, data->map);
map              1374 dev/pci/if_wpi.c 		bus_dmamap_unload(sc->sc_dmat, data->map);
map              1693 dev/pci/if_wpi.c 	error = bus_dmamap_load_mbuf(sc->sc_dmat, data->map, m0,
map              1724 dev/pci/if_wpi.c 		error = bus_dmamap_load_mbuf(sc->sc_dmat, data->map, m0,
map              1738 dev/pci/if_wpi.c 	    ring->qid, ring->cur, m0->m_pkthdr.len, data->map->dm_nsegs));
map              1742 dev/pci/if_wpi.c 	    (1 + data->map->dm_nsegs) << 24);
map              1746 dev/pci/if_wpi.c 	for (i = 1; i <= data->map->dm_nsegs; i++) {
map              1748 dev/pci/if_wpi.c 		    htole32(data->map->dm_segs[i - 1].ds_addr);
map              1750 dev/pci/if_wpi.c 		    htole32(data->map->dm_segs[i - 1].ds_len);
map              2317 dev/pci/if_wpi.c 	error = bus_dmamap_load(sc->sc_dmat, data->map, mtod(m0, void *),
map              2332 dev/pci/if_wpi.c 	desc->segs[1].addr = htole32(data->map->dm_segs[0].ds_addr);
map              2333 dev/pci/if_wpi.c 	desc->segs[1].len  = htole32(data->map->dm_segs[0].ds_len);
map              2545 dev/pci/if_wpi.c 	error = bus_dmamap_load(sc->sc_dmat, data->map, cmd, pktlen, NULL,
map              2556 dev/pci/if_wpi.c 	desc->segs[0].addr = htole32(data->map->dm_segs[0].ds_addr);
map              2557 dev/pci/if_wpi.c 	desc->segs[0].len  = htole32(data->map->dm_segs[0].ds_len);
map                58 dev/pci/if_wpivar.h 	bus_dmamap_t		map;
map                66 dev/pci/if_wpivar.h 	bus_dmamap_t		map;
map               315 dev/pci/musycc.c 		    MCLBYTES, 0, BUS_DMA_NOWAIT, &dd->map)) {
map               858 dev/pci/musycc.c 		if (dd->map->dm_nsegs != 0) {
map               859 dev/pci/musycc.c 			bus_dmamap_t map = dd->map;
map               861 dev/pci/musycc.c 			bus_dmamap_unload(mg->mg_dmat, map);
map               889 dev/pci/musycc.c 		if (dd->map->dm_nsegs != 0) {
map               890 dev/pci/musycc.c 			bus_dmamap_t map = dd->map;
map               892 dev/pci/musycc.c 			bus_dmamap_unload(mg->mg_dmat, map);
map               936 dev/pci/musycc.c 	bus_dmamap_t	 map;
map               963 dev/pci/musycc.c 	map = c->map;
map               964 dev/pci/musycc.c 	c->map = mg->mg_rx_sparemap;
map               965 dev/pci/musycc.c 	mg->mg_rx_sparemap = map;
map               967 dev/pci/musycc.c 	bus_dmamap_sync(mg->mg_dmat, c->map, 0, c->map->dm_mapsize,
map               971 dev/pci/musycc.c 	c->data = htole32(c->map->dm_segs[0].ds_addr);
map               990 dev/pci/musycc.c 	bus_dmamap_t	 map;
map               997 dev/pci/musycc.c 	map = mg->mg_tx_sparemap;
map               998 dev/pci/musycc.c 	if (bus_dmamap_load_mbuf(mg->mg_dmat, map, m_head,
map              1008 dev/pci/musycc.c 	if (map->dm_nsegs + mg->mg_dma_d[c].tx_use >= mg->mg_dma_d[c].tx_cnt) {
map              1015 dev/pci/musycc.c 	while (i < map->dm_nsegs) {
map              1017 dev/pci/musycc.c 		    MUSYCC_STATUS_LEN(map->dm_segs[i].ds_len);
map              1022 dev/pci/musycc.c 		cur->data = htole32(map->dm_segs[i].ds_addr);
map              1028 dev/pci/musycc.c 		if (++i >= map->dm_nsegs)
map              1033 dev/pci/musycc.c 	bus_dmamap_sync(mg->mg_dmat, map, 0, map->dm_mapsize,
map              1037 dev/pci/musycc.c 	mg->mg_tx_sparemap = cur->map;
map              1038 dev/pci/musycc.c 	cur->map = map;
map              1198 dev/pci/musycc.c 		bus_dmamap_sync(mg->mg_dmat, cur_rx->map, 0,
map              1199 dev/pci/musycc.c 		    cur_rx->map->dm_mapsize, BUS_DMASYNC_POSTREAD);
map              1273 dev/pci/musycc.c 		if (dd->map->dm_nsegs != 0) {
map              1274 dev/pci/musycc.c 			bus_dmamap_sync(mg->mg_dmat, dd->map, 0,
map              1275 dev/pci/musycc.c 			    dd->map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
map              1276 dev/pci/musycc.c 			bus_dmamap_unload(mg->mg_dmat, dd->map);
map                49 dev/pci/musyccvar.h 	bus_dmamap_t		 map;
map              1467 dev/pci/safe.c 	for (i = 0; i < op->map->dm_nsegs; i++) {
map              1468 dev/pci/safe.c 		if (op->map->dm_segs[i].ds_addr & 3)
map              1470 dev/pci/safe.c 		if (i != (op->map->dm_nsegs - 1) &&
map              1471 dev/pci/safe.c 		    (op->map->dm_segs[i].ds_len & 3))
map              1567 dev/pci/safe.c 	if (op->map->dm_nsegs <= 0)
map              1570 dev/pci/safe.c 	for (i = 0; i < op->map->dm_nsegs-1; i++) {
map              1571 dev/pci/safe.c 		if (op->map->dm_segs[i].ds_len % SAFE_MAX_DSIZE)
map              1573 dev/pci/safe.c 		if (op->map->dm_segs[i].ds_len != SAFE_MAX_DSIZE)
map                86 dev/pci/safevar.h 	bus_dmamap_t		map;
map               122 dev/pci/safevar.h #define	re_src_map	re_src.map
map               123 dev/pci/safevar.h #define	re_src_nsegs	re_src.map->dm_nsegs
map               124 dev/pci/safevar.h #define	re_src_segs	re_src.map->dm_segs
map               125 dev/pci/safevar.h #define	re_src_mapsize	re_src.map->dm_mapsize
map               129 dev/pci/safevar.h #define	re_dst_map	re_dst.map
map               130 dev/pci/safevar.h #define	re_dst_nsegs	re_dst.map->dm_nsegs
map               131 dev/pci/safevar.h #define	re_dst_segs	re_dst.map->dm_segs
map               132 dev/pci/safevar.h #define	re_dst_mapsize	re_dst.map->dm_mapsize
map                80 dev/pci/sv.c   	bus_dmamap_t map;
map                87 dev/pci/sv.c   #define DMAADDR(map) ((map)->segs[0].ds_addr)
map                88 dev/pci/sv.c   #define KERNADDR(map) ((void *)((map)->addr))
map               473 dev/pci/sv.c   				  0, BUS_DMA_NOWAIT, &p->map);
map               477 dev/pci/sv.c   	error = bus_dmamap_load(sc->sc_dmatag, p->map, p->addr, p->size, NULL, 
map               484 dev/pci/sv.c   	bus_dmamap_destroy(sc->sc_dmatag, p->map);
map               497 dev/pci/sv.c   	bus_dmamap_unload(sc->sc_dmatag, p->map);
map               498 dev/pci/sv.c   	bus_dmamap_destroy(sc->sc_dmatag, p->map);
map              1693 dev/pci/ubsec.c ubsec_dmamap_aligned(bus_dmamap_t map)
map              1697 dev/pci/ubsec.c 	for (i = 0; i < map->dm_nsegs; i++) {
map              1698 dev/pci/ubsec.c 		if (map->dm_segs[i].ds_addr & 3)
map              1700 dev/pci/ubsec.c 		if ((i != (map->dm_nsegs - 1)) &&
map              1701 dev/pci/ubsec.c 		    (map->dm_segs[i].ds_len & 3))
map                96 dev/pci/yds.c  #define DMAADDR(p) ((p)->map->dm_segs[0].ds_addr)
map               470 dev/pci/yds.c  		 mp, (void *) sc->sc_ctrldata.map->dm_segs[0].ds_addr));
map               509 dev/pci/yds.c  	bus_dmamap_sync(sc->sc_dmatag, p->map,
map               963 dev/pci/yds.c  			bus_dmamap_sync(sc->sc_dmatag, sc->sc_ctrldata.map,
map               980 dev/pci/yds.c  						sc->sc_play.dma->map,
map               994 dev/pci/yds.c  						sc->sc_play.dma->map,
map              1003 dev/pci/yds.c  			bus_dmamap_sync(sc->sc_dmatag, sc->sc_ctrldata.map,
map              1020 dev/pci/yds.c  						sc->sc_rec.dma->map,
map              1034 dev/pci/yds.c  						sc->sc_rec.dma->map,
map              1066 dev/pci/yds.c  				  0, BUS_DMA_NOWAIT, &p->map);
map              1070 dev/pci/yds.c  	error = bus_dmamap_load(sc->sc_dmatag, p->map, p->addr, p->size, NULL, 
map              1077 dev/pci/yds.c  	bus_dmamap_destroy(sc->sc_dmatag, p->map);
map              1090 dev/pci/yds.c  	bus_dmamap_unload(sc->sc_dmatag, p->map);
map              1091 dev/pci/yds.c  	bus_dmamap_destroy(sc->sc_dmatag, p->map);
map              1456 dev/pci/yds.c  	bus_dmamap_sync(sc->sc_dmatag, sc->sc_ctrldata.map,
map              1462 dev/pci/yds.c  	bus_dmamap_sync(sc->sc_dmatag, p->map, 0, blksize,
map              1548 dev/pci/yds.c  	bus_dmamap_sync(sc->sc_dmatag, sc->sc_ctrldata.map,
map              1555 dev/pci/yds.c  	bus_dmamap_sync(sc->sc_dmatag, p->map, 0, blksize,
map              1595 dev/pci/yds.c  		bus_dmamap_sync(sc->sc_dmatag, sc->sc_ctrldata.map,
map              1606 dev/pci/yds.c  		bus_dmamap_sync(sc->sc_dmatag, sc->sc_play.dma->map,
map              1625 dev/pci/yds.c  		bus_dmamap_sync(sc->sc_dmatag, sc->sc_ctrldata.map,
map              1631 dev/pci/yds.c  		bus_dmamap_sync(sc->sc_dmatag, sc->sc_rec.dma->map,
map               324 dev/pci/ydsreg.h 	bus_dmamap_t map;
map               982 dev/pckbc/wskbdmap_mfii.c #define KBD_MAP(name, base, map) \
map               983 dev/pckbc/wskbdmap_mfii.c 			{ name, base, sizeof(map)/sizeof(keysym_t), map }
map               263 dev/raidframe/rf_aselect.c 		(raidPtr->Layout.map->SelectionFunc) (raidPtr, type, asm_p,
map               322 dev/raidframe/rf_aselect.c 				(raidPtr->Layout.map->SelectionFunc) (raidPtr,
map               401 dev/raidframe/rf_aselect.c 						(raidPtr->Layout.map->
map               320 dev/raidframe/rf_copyback.c 			(raidPtr->Layout.map->MapSector) (raidPtr, addr,
map               341 dev/raidframe/rf_copyback.c 			(raidPtr->Layout.map->MapParity) (raidPtr, stripeAddr,
map               374 dev/raidframe/rf_copyback.c 	if (raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE) {
map               376 dev/raidframe/rf_copyback.c 			raidPtr->Layout.map->MapSector(raidPtr, addr, &spRow,
map               379 dev/raidframe/rf_copyback.c 			raidPtr->Layout.map->MapParity(raidPtr, addr, &spRow,
map               467 dev/raidframe/rf_copyback.c 		if (raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE) {
map               468 dev/raidframe/rf_copyback.c 			RF_ASSERT(raidPtr->Layout.map->parityConfig == 'D');
map               851 dev/raidframe/rf_dagdegrd.c 		(raidPtr->Layout.map->MapSector) (raidPtr, pda_p->raidAddress,
map               881 dev/raidframe/rf_dagdegrd.c 			(raidPtr->Layout.map->MapSector) (raidPtr,
map               915 dev/raidframe/rf_dagdegrd.c 				(raidPtr->Layout.map->MapSector) (raidPtr,
map               932 dev/raidframe/rf_dagdegrd.c 				(raidPtr->Layout.map->MapSector) (raidPtr,
map               948 dev/raidframe/rf_dagdegrd.c 				(raidPtr->Layout.map->MapSector) (raidPtr,
map               961 dev/raidframe/rf_dagdegrd.c 					(raidPtr->Layout.map->MapSector)
map               984 dev/raidframe/rf_dagdegrd.c 					(raidPtr->Layout.map->MapSector)
map              1004 dev/raidframe/rf_dagdegrd.c 					(raidPtr->Layout.map->MapSector)
map              1026 dev/raidframe/rf_dagdegrd.c 					(raidPtr->Layout.map->MapSector)
map              1046 dev/raidframe/rf_dagdegrd.c 					(raidPtr->Layout.map->MapSector)
map              1069 dev/raidframe/rf_dagdegrd.c 		(raidPtr->Layout.map->MapSector) (raidPtr, pda_p->raidAddress,
map              1099 dev/raidframe/rf_dagdegrd.c 			(raidPtr->Layout.map->MapSector) (raidPtr,
map               659 dev/raidframe/rf_dagdegwr.c 		(raidPtr->Layout.map->MapSector) (raidPtr, pda_p->raidAddress,
map               686 dev/raidframe/rf_dagdegwr.c 			(raidPtr->Layout.map->MapSector) (raidPtr,
map               759 dev/raidframe/rf_dagutils.c 	int ds = (raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE) ? 1 : 0;
map               778 dev/raidframe/rf_dagutils.c 				raidPtr->Layout.map->MapSector(raidPtr,
map               798 dev/raidframe/rf_dagutils.c 			(raidPtr->Layout.map->MapParity) (raidPtr,
map               100 dev/raidframe/rf_decluster.c 	if (layoutPtr->map->flags & RF_DISTRIBUTE_SPARE) {
map               123 dev/raidframe/rf_decluster.c 	if ((layoutPtr->map->flags & RF_DISTRIBUTE_SPARE) &&
map               154 dev/raidframe/rf_decluster.c 	if ((raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE)) {
map               182 dev/raidframe/rf_decluster.c 	if (raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE) {
map               273 dev/raidframe/rf_decluster.c 	if (raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE) {
map               412 dev/raidframe/rf_decluster.c 	if (raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE) {
map               494 dev/raidframe/rf_decluster.c 	if ((raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE)) {
map               102 dev/raidframe/rf_declusterPQ.c 	if ((raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE) &&
map               132 dev/raidframe/rf_declusterPQ.c 	if ((raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE)) {
map               159 dev/raidframe/rf_declusterPQ.c 	if ((raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE)) {
map               248 dev/raidframe/rf_declusterPQ.c 	if ((raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE)) {
map               360 dev/raidframe/rf_declusterPQ.c 	if ((raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE)) {
map               443 dev/raidframe/rf_declusterPQ.c 	if ((raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE)) {
map               501 dev/raidframe/rf_declusterPQ.c 	if ((raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE)) {
map               741 dev/raidframe/rf_driver.c 	    flags, cbF, cbA, raidPtr->Layout.map->states);
map               760 dev/raidframe/rf_driver.c 	if (!(raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE)) {
map               774 dev/raidframe/rf_driver.c 	if (raidPtr->Layout.map->flags & RF_BD_DECLUSTERED)
map               795 dev/raidframe/rf_evenodd_dagfuncs.c 			(raidPtr->Layout.map->MapSector) (raidPtr,
map              1017 dev/raidframe/rf_evenodd_dagfuncs.c 		(raidPtr->Layout.map->MapSector) (raidPtr, npda.raidAddress,
map               466 dev/raidframe/rf_layout.c 	layoutPtr->map = p;
map               496 dev/raidframe/rf_layout.c 		    " with head sep limit %ld.\n", layoutPtr->map->configName,
map               501 dev/raidframe/rf_layout.c 		    " with no head sep limit.\n", layoutPtr->map->configName,
map               524 dev/raidframe/rf_layout.c 	if ((layoutPtr->SUsPerPU == 1) || !layoutPtr->map->MapSIDToPSID) {
map               528 dev/raidframe/rf_layout.c 		(layoutPtr->map->MapSIDToPSID) (layoutPtr, stripeID,
map               220 dev/raidframe/rf_layout.h 	RF_LayoutSW_t		*map;		/*
map                96 dev/raidframe/rf_map.c 	int faultsTolerated = layoutPtr->map->faultsTolerated;
map               179 dev/raidframe/rf_map.c 			(layoutPtr->map->MapSector) (raidPtr, raidAddress,
map               230 dev/raidframe/rf_map.c 			(layoutPtr->map->MapParity) (raidPtr,
map               259 dev/raidframe/rf_map.c 			(layoutPtr->map->MapParity) (raidPtr,
map               263 dev/raidframe/rf_map.c 			(layoutPtr->map->MapQ) (raidPtr,
map               692 dev/raidframe/rf_map.c 	(layoutPtr->map->IdentifyStripe) (raidPtr, asmap->raidAddress,
map               694 dev/raidframe/rf_map.c 	(layoutPtr->map->MapParity) (raidPtr, asmap->raidAddress,
map               713 dev/raidframe/rf_map.c 				layoutPtr->map->MapSector(raidPtr,
map               756 dev/raidframe/rf_map.c 		(layoutPtr->map->MapSector) (raidPtr, sosAddr + i *
map               968 dev/raidframe/rf_map.c 				raidPtr->Layout.map->MapSector(raidPtr,
map               972 dev/raidframe/rf_map.c 				raidPtr->Layout.map->MapParity(raidPtr,
map              1176 dev/raidframe/rf_openbsdkintf.c 		if (raidPtr->Layout.map->faultsTolerated == 0) {
map              1232 dev/raidframe/rf_openbsdkintf.c 		if (raidPtr->Layout.map->faultsTolerated == 0) {
map              1357 dev/raidframe/rf_openbsdkintf.c 		if (raidPtr->Layout.map->faultsTolerated == 0) {
map              1374 dev/raidframe/rf_openbsdkintf.c 		if (raidPtr->Layout.map->faultsTolerated == 0) {
map              1411 dev/raidframe/rf_openbsdkintf.c 		if (raidPtr->Layout.map->faultsTolerated == 0) {
map              1447 dev/raidframe/rf_openbsdkintf.c 		if (raidPtr->Layout.map->faultsTolerated == 0) {
map              3404 dev/raidframe/rf_openbsdkintf.c 	clabel->parityConfig = raidPtr->Layout.map->parityConfig;
map              3502 dev/raidframe/rf_openbsdkintf.c 	    (raidPtrs[raidID]->Layout).map->configName,
map               988 dev/raidframe/rf_paritylogging.c 				if (layoutPtr->map->flags &
map               995 dev/raidframe/rf_paritylogging.c 						(layoutPtr->map->MapParity)
map              1036 dev/raidframe/rf_paritylogging.c 							(layoutPtr->map
map                71 dev/raidframe/rf_parityscan.c 	if (raidPtr->Layout.map->faultsTolerated == 0) {
map               149 dev/raidframe/rf_parityscan.c 	lp = raidPtr->Layout.map;
map               337 dev/raidframe/rf_parityscan.c 			if (raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE) {
map               341 dev/raidframe/rf_parityscan.c 					(raidPtr->Layout.map->MapParity)
map               353 dev/raidframe/rf_parityscan.c 					(raidPtr->Layout.map->MapSector)
map               271 dev/raidframe/rf_pqdegdags.c 			(raidPtr->Layout.map->MapSector) (raidPtr,
map               374 dev/raidframe/rf_pqdegdags.c 		(raidPtr->Layout.map->MapSector) (raidPtr, npda.raidAddress,
map               222 dev/raidframe/rf_raid5.c 				if (layoutPtr->map->flags &
map               229 dev/raidframe/rf_raid5.c 						(layoutPtr->map->MapParity)
map               271 dev/raidframe/rf_raid5.c 							(layoutPtr->map
map               112 dev/raidframe/rf_reconbuffer.c 	lp = rbuf->raidPtr->Layout.map;
map               338 dev/raidframe/rf_reconstruct.c 	lp = raidPtr->Layout.map;
map               386 dev/raidframe/rf_reconstruct.c 	if (raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE) {
map               481 dev/raidframe/rf_reconstruct.c 	lp = raidPtr->Layout.map;
map               546 dev/raidframe/rf_reconstruct.c 		if (raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE) {
map               881 dev/raidframe/rf_reconstruct.c 		ds = (raidPtr->Layout.map->flags & RF_DISTRIBUTE_SPARE);
map              1337 dev/raidframe/rf_reconstruct.c 	(layoutPtr->map->IdentifyStripe) (raidPtr, sosRaidAddress, &diskids,
map              1361 dev/raidframe/rf_reconstruct.c 	(layoutPtr->map->MapParity) (raidPtr, sosRaidAddress, &prow, &pcol,
map              1403 dev/raidframe/rf_reconstruct.c 		layoutPtr->map->MapParity(raidPtr, sosRaidAddress + i_offset *
map              1407 dev/raidframe/rf_reconstruct.c 		layoutPtr->map->MapSector(raidPtr, sosRaidAddress + i_offset *
map              1414 dev/raidframe/rf_reconstruct.c 		layoutPtr->map->MapParity(raidPtr, sosRaidAddress + j_offset *
map              1418 dev/raidframe/rf_reconstruct.c 		layoutPtr->map->MapSector(raidPtr, sosRaidAddress + j_offset *
map              1424 dev/raidframe/rf_reconstruct.c 	if (layoutPtr->map->flags & RF_DISTRIBUTE_SPARE) {
map              1426 dev/raidframe/rf_reconstruct.c 			layoutPtr->map->MapParity(raidPtr, sosRaidAddress +
map              1430 dev/raidframe/rf_reconstruct.c 			layoutPtr->map->MapSector(raidPtr, sosRaidAddress +
map              1811 dev/raidframe/rf_reconstruct.c 		(raidPtr->Layout.map->IdentifyStripe) (raidPtr,
map                70 dev/raidframe/rf_reconutil.c 	lp = raidPtr->Layout.map;
map               212 dev/raidframe/rf_reconutil.c 	lp = raidPtr->Layout.map;
map               229 dev/raidframe/rf_reconutil.c 	lp = raidPtr->Layout.map;
map               354 dev/raidframe/rf_states.c 	if (!(raidPtr->Layout.map->flags & RF_NO_STRIPE_LOCKS)) {
map               676 dev/raidframe/rf_states.c 	if (!(raidPtr->Layout.map->flags & RF_NO_STRIPE_LOCKS)) {
map               862 dev/sun/sunkbdmap.c #define KBD_MAP(name, base, map) \
map               863 dev/sun/sunkbdmap.c     { name, base, sizeof(map)/sizeof(keysym_t), map }
map              1018 dev/usb/ukbdmap.c #define KBD_MAP(name, base, map) \
map              1019 dev/usb/ukbdmap.c 			{ name, base, sizeof(map)/sizeof(keysym_t), map }
map               157 dev/usb/usb_mem.c 				  0, BUS_DMA_NOWAIT, &p->map);
map               161 dev/usb/usb_mem.c 	error = bus_dmamap_load(tag, p->map, p->kaddr, p->size, NULL,
map               170 dev/usb/usb_mem.c 	bus_dmamap_destroy(tag, p->map);
map               190 dev/usb/usb_mem.c 	bus_dmamap_unload(p->tag, p->map);
map               191 dev/usb/usb_mem.c 	bus_dmamap_destroy(p->tag, p->map);
map                44 dev/usb/usb_mem.h 	bus_dmamap_t map;
map                54 dev/usb/usb_mem.h #define DMAADDR(dma, o) ((dma)->block->map->dm_segs[0].ds_addr + (dma)->offs + (o))
map               175 dev/wscons/wsconsio.h 	struct wscons_keymap *map;		/* map to get or set */
map              1085 dev/wscons/wskbd.c 		error = copyin(umdp->map, buf, len);
map              1102 dev/wscons/wskbd.c 		error = copyout(sc->sc_map, umdp->map,
map               363 dev/wscons/wskbdutil.c 		for (kp = mp->map; kp < mp->map + mp->map_size; kp++) {
map               373 dev/wscons/wskbdutil.c 				for (l = 0; kp + l < mp->map + mp->map_size;
map               392 dev/wscons/wskbdutil.c wskbd_init_keymap(newlen, map, maplen)
map               394 dev/wscons/wskbdutil.c 	struct wscons_keymap **map;
map               401 dev/wscons/wskbdutil.c 			free(*map, M_TEMP);
map               403 dev/wscons/wskbdutil.c 		*map = malloc(newlen*sizeof(struct wscons_keymap),
map               408 dev/wscons/wskbdutil.c 		(*map)[i].command = KS_voidSymbol;
map               409 dev/wscons/wskbdutil.c 		(*map)[i].group1[0] = KS_voidSymbol;
map               410 dev/wscons/wskbdutil.c 		(*map)[i].group1[1] = KS_voidSymbol;
map               411 dev/wscons/wskbdutil.c 		(*map)[i].group2[0] = KS_voidSymbol;
map               412 dev/wscons/wskbdutil.c 		(*map)[i].group2[1] = KS_voidSymbol;
map               417 dev/wscons/wskbdutil.c wskbd_load_keymap(mapdata, map, maplen)
map               419 dev/wscons/wskbdutil.c 	struct wscons_keymap **map;
map               450 dev/wscons/wskbdutil.c 		for (kp = mp->map; kp < mp->map + mp->map_size; kp++) {
map               457 dev/wscons/wskbdutil.c 	wskbd_init_keymap(i + 1, map, maplen);
map               461 dev/wscons/wskbdutil.c 		for (kp = mp->map; kp < mp->map + mp->map_size; ) {
map               472 dev/wscons/wskbdutil.c 				(*map)[kc].command = *kp;
map               476 dev/wscons/wskbdutil.c 			for (i = 0; kp + i < mp->map + mp->map_size; i++) {
map               486 dev/wscons/wskbdutil.c 			fillmapentry(kp, i, &(*map)[kc]);
map                60 dev/wscons/wsksymvar.h 	const keysym_t *map;			/* the map itself */
map               271 isofs/udf/ecma167-udf.h 	uint32_t		map;
map              1325 isofs/udf/udf_vnops.c 				   letoh32(ump->um_stbl->entries[i].map) +
map               693 msdosfs/msdosfs_fat.c 	u_int map;
map               699 msdosfs/msdosfs_fat.c 	map = pmp->pm_inusemap[idx];
map               700 msdosfs/msdosfs_fat.c 	map &= ~((1 << start) - 1);
map               701 msdosfs/msdosfs_fat.c 	if (map) {
map               702 msdosfs/msdosfs_fat.c 		len = ffs(map) - 1 - start;
map               711 msdosfs/msdosfs_fat.c 		if ((map = pmp->pm_inusemap[idx]) != 0) {
map               712 msdosfs/msdosfs_fat.c 			len +=  ffs(map) - 1;
map               781 msdosfs/msdosfs_fat.c 	u_int map;
map               809 msdosfs/msdosfs_fat.c 		map = pmp->pm_inusemap[idx];
map               810 msdosfs/msdosfs_fat.c 		map |= (1 << (cn % N_INUSEBITS)) - 1;
map               811 msdosfs/msdosfs_fat.c 		if (map != (u_int)-1) {
map               812 msdosfs/msdosfs_fat.c 			cn = idx * N_INUSEBITS + ffs(map^(u_int)-1) - 1;
map               826 msdosfs/msdosfs_fat.c 		map = pmp->pm_inusemap[idx];
map               827 msdosfs/msdosfs_fat.c 		map |= (1 << (cn % N_INUSEBITS)) - 1;
map               828 msdosfs/msdosfs_fat.c 		if (map != (u_int)-1) {
map               829 msdosfs/msdosfs_fat.c 			cn = idx * N_INUSEBITS + ffs(map^(u_int)-1) - 1;
map              1534 netbt/hci.h    	uint8_t		map[10];    /* AFH Channel Map */
map               404 ufs/ext2fs/ext2fs_alloc.c 	int error, start, len, loc, map, i;
map               438 ufs/ext2fs/ext2fs_alloc.c 	map = ibp[i];
map               441 ufs/ext2fs/ext2fs_alloc.c 		if ((map & i) == 0) {
map               558 ufs/ext2fs/ext2fs_alloc.c 	int start, len, loc, i, map;
map               582 ufs/ext2fs/ext2fs_alloc.c 	map = bbp[i];
map               585 ufs/ext2fs/ext2fs_alloc.c 		if ((map & i) == 0)
map              1420 ufs/ffs/ffs_alloc.c 	int i, got, run, bno, bit, map;
map              1472 ufs/ffs/ffs_alloc.c 	map = *mapp++;
map              1475 ufs/ffs/ffs_alloc.c 		if ((map & bit) == 0) {
map              1485 ufs/ffs/ffs_alloc.c 			map = *mapp++;
map              1526 ufs/ffs/ffs_alloc.c 	int error, start, len, loc, map, i;
map              1605 ufs/ffs/ffs_alloc.c 	map = cg_inosused(cgp)[i];
map              1608 ufs/ffs/ffs_alloc.c 		if ((map & i) == 0) {
map              1984 ufs/ffs/ffs_alloc.c 	int i, start, end, forw, back, map, bit;
map              2005 ufs/ffs/ffs_alloc.c 	map = *mapp++;
map              2008 ufs/ffs/ffs_alloc.c 		if ((map & bit) == 0)
map              2013 ufs/ffs/ffs_alloc.c 			map = *mapp++;
map              2026 ufs/ffs/ffs_alloc.c 	map = *mapp--;
map              2029 ufs/ffs/ffs_alloc.c 		if ((map & bit) == 0)
map              2034 ufs/ffs/ffs_alloc.c 			map = *mapp--;
map               497 ufs/ffs/fs.h   #define blkmap(fs, map, loc) \
map               498 ufs/ffs/fs.h       (((map)[(loc) / NBBY] >> ((loc) % NBBY)) & (0xff >> (NBBY - (fs)->fs_frag)))
map               581 uvm/uvm_amap.c amap_copy(struct vm_map *map, struct vm_map_entry *entry, int waitf,
map               589 uvm/uvm_amap.c 		    map, entry, waitf, 0);
map               612 uvm/uvm_amap.c 			UVM_MAP_CLIP_START(map, entry, startva);
map               615 uvm/uvm_amap.c 				UVM_MAP_CLIP_END(map, entry, endva);
map               750 uvm/uvm_amap.c amap_cow_now(struct vm_map *map, struct vm_map_entry *entry)
map               255 uvm/uvm_fault.c 			amap_copy(ufi->map, ufi->entry, M_NOWAIT, TRUE, 
map               615 uvm/uvm_fault.c 	if ((ufi.map->flags & VM_MAP_PAGEABLE) == 0)
map               617 uvm/uvm_fault.c 		    ufi.map, vaddr);
map              1771 uvm/uvm_fault.c uvm_fault_wire(map, start, end, access_type)
map              1772 uvm/uvm_fault.c 	vm_map_t map;
map              1780 uvm/uvm_fault.c 	pmap = vm_map_pmap(map);
map              1789 uvm/uvm_fault.c 		rv = uvm_fault(map, va, VM_FAULT_WIRE, access_type);
map              1792 uvm/uvm_fault.c 				uvm_fault_unwire(map, start, va);
map              1806 uvm/uvm_fault.c uvm_fault_unwire(map, start, end)
map              1807 uvm/uvm_fault.c 	vm_map_t map;
map              1811 uvm/uvm_fault.c 	vm_map_lock_read(map);
map              1812 uvm/uvm_fault.c 	uvm_fault_unwire_locked(map, start, end);
map              1813 uvm/uvm_fault.c 	vm_map_unlock_read(map);
map              1823 uvm/uvm_fault.c uvm_fault_unwire_locked(map, start, end)
map              1824 uvm/uvm_fault.c 	vm_map_t map;
map              1828 uvm/uvm_fault.c 	pmap_t pmap = vm_map_pmap(map);
map              1833 uvm/uvm_fault.c 	KASSERT((map->flags & VM_MAP_INTRSAFE) == 0);
map              1847 uvm/uvm_fault.c 	KASSERT(start >= vm_map_min(map) && end <= vm_map_max(map));
map              1848 uvm/uvm_fault.c 	if (uvm_map_lookup_entry(map, start, &entry) == FALSE)
map              1860 uvm/uvm_fault.c 			KASSERT(entry->next != &map->header &&
map                63 uvm/uvm_fault.h 	vm_map_t map;			/* map (could be a submap) */
map                69 uvm/uvm_fault_i.h 		vm_map_unlock(ufi->map);
map                71 uvm/uvm_fault_i.h 		vm_map_unlock_read(ufi->map);
map               122 uvm/uvm_fault_i.h 	ufi->map = ufi->orig_map;
map               136 uvm/uvm_fault_i.h 			vm_map_lock(ufi->map);
map               138 uvm/uvm_fault_i.h 			vm_map_lock_read(ufi->map);
map               144 uvm/uvm_fault_i.h 		if (!uvm_map_lookup_entry(ufi->map, ufi->orig_rvaddr, 
map               163 uvm/uvm_fault_i.h 				vm_map_unlock(ufi->map);
map               165 uvm/uvm_fault_i.h 				vm_map_unlock_read(ufi->map);
map               167 uvm/uvm_fault_i.h 			ufi->map = tmpmap;
map               175 uvm/uvm_fault_i.h 		ufi->mapv = ufi->map->timestamp;
map               210 uvm/uvm_fault_i.h 	vm_map_lock_read(ufi->map);
map               211 uvm/uvm_fault_i.h 	if (ufi->mapv != ufi->map->timestamp) {
map               212 uvm/uvm_fault_i.h 		vm_map_unlock_read(ufi->map);
map               176 uvm/uvm_glue.c 	vm_map_t map;
map               180 uvm/uvm_glue.c 	map = &p->p_vmspace->vm_map;
map               186 uvm/uvm_glue.c 	rv = uvm_fault_wire(map, start, end, access_type);
map                63 uvm/uvm_io.c   uvm_io(vm_map_t map, struct uio *uio, int flags)
map               110 uvm/uvm_io.c   		error = uvm_map_extract(map, baseva, chunksz, kernel_map, &kva,
map               208 uvm/uvm_km.c   uvm_km_suballoc(struct vm_map *map, vaddr_t *min, vaddr_t *max, vsize_t size,
map               219 uvm/uvm_km.c   	if (uvm_map(map, min, size, NULL, UVM_UNKNOWN_OFFSET, 0,
map               235 uvm/uvm_km.c   	pmap_reference(vm_map_pmap(map));
map               237 uvm/uvm_km.c   		submap = uvm_map_create(vm_map_pmap(map), *min, *max, flags);
map               242 uvm/uvm_km.c   		submap->pmap = vm_map_pmap(map);
map               249 uvm/uvm_km.c   	if (uvm_map_submap(map, *min, *max, submap) != 0)
map               339 uvm/uvm_km.c   uvm_km_kmemalloc(struct vm_map *map, struct uvm_object *obj, vsize_t size,
map               348 uvm/uvm_km.c   		    map, obj, size, flags);
map               349 uvm/uvm_km.c   	KASSERT(vm_map_pmap(map) == pmap_kernel());
map               356 uvm/uvm_km.c   	kva = vm_map_min(map);	/* hint */
map               362 uvm/uvm_km.c   	if (__predict_false(uvm_map(map, &kva, size, obj, UVM_UNKNOWN_OFFSET,
map               407 uvm/uvm_km.c   				uvm_unmap(map, kva, kva + size);
map               424 uvm/uvm_km.c   			pmap_enter(map->pmap, loopva, VM_PAGE_TO_PHYS(pg),
map               442 uvm/uvm_km.c   uvm_km_free(struct vm_map *map, vaddr_t addr, vsize_t size)
map               444 uvm/uvm_km.c   	uvm_unmap(map, trunc_page(addr), round_page(addr+size));
map               455 uvm/uvm_km.c   uvm_km_free_wakeup(struct vm_map *map, vaddr_t addr, vsize_t size)
map               459 uvm/uvm_km.c   	vm_map_lock(map);
map               460 uvm/uvm_km.c   	uvm_unmap_remove(map, trunc_page(addr), round_page(addr+size), 
map               462 uvm/uvm_km.c   	wakeup(map);
map               463 uvm/uvm_km.c   	vm_map_unlock(map);
map               476 uvm/uvm_km.c   uvm_km_alloc1(struct vm_map *map, vsize_t size, vsize_t align, boolean_t zeroit)
map               483 uvm/uvm_km.c   	UVMHIST_LOG(maphist,"(map=%p, size=0x%lx)", map, size,0,0);
map               484 uvm/uvm_km.c   	KASSERT(vm_map_pmap(map) == pmap_kernel());
map               487 uvm/uvm_km.c   	kva = vm_map_min(map);		/* hint */
map               493 uvm/uvm_km.c   	if (__predict_false(uvm_map(map, &kva, size, uvm.kernel_object,
map               543 uvm/uvm_km.c   				uvm_unmap(map, kva, loopva - kva);
map               555 uvm/uvm_km.c   		pmap_enter(map->pmap, loopva, VM_PAGE_TO_PHYS(pg),
map               562 uvm/uvm_km.c   	pmap_update(map->pmap);
map               583 uvm/uvm_km.c   uvm_km_valloc(struct vm_map *map, vsize_t size)
map               585 uvm/uvm_km.c   	return(uvm_km_valloc_align(map, size, 0));
map               589 uvm/uvm_km.c   uvm_km_valloc_align(struct vm_map *map, vsize_t size, vsize_t align)
map               594 uvm/uvm_km.c   	UVMHIST_LOG(maphist, "(map=%p, size=0x%lx)", map, size, 0,0);
map               595 uvm/uvm_km.c   	KASSERT(vm_map_pmap(map) == pmap_kernel());
map               598 uvm/uvm_km.c   	kva = vm_map_min(map);		/* hint */
map               604 uvm/uvm_km.c   	if (__predict_false(uvm_map(map, &kva, size, uvm.kernel_object,
map               624 uvm/uvm_km.c   uvm_km_valloc_prefer_wait(struct vm_map *map, vsize_t size, voff_t prefer)
map               629 uvm/uvm_km.c   	UVMHIST_LOG(maphist, "(map=%p, size=0x%lx)", map, size, 0,0);
map               630 uvm/uvm_km.c   	KASSERT(vm_map_pmap(map) == pmap_kernel());
map               633 uvm/uvm_km.c   	if (size > vm_map_max(map) - vm_map_min(map))
map               637 uvm/uvm_km.c   		kva = vm_map_min(map);		/* hint */
map               644 uvm/uvm_km.c   		if (__predict_true(uvm_map(map, &kva, size, uvm.kernel_object,
map               656 uvm/uvm_km.c   		tsleep((caddr_t)map, PVM, "vallocwait", 0);
map               662 uvm/uvm_km.c   uvm_km_valloc_wait(struct vm_map *map, vsize_t size)
map               664 uvm/uvm_km.c   	return uvm_km_valloc_prefer_wait(map, size, UVM_UNKNOWN_OFFSET);
map               675 uvm/uvm_km.c   uvm_km_alloc_poolpage1(struct vm_map *map, struct uvm_object *obj,
map               710 uvm/uvm_km.c   	va = uvm_km_kmemalloc(map, obj, PAGE_SIZE, waitok ? 0 : UVM_KMF_NOWAIT);
map               724 uvm/uvm_km.c   uvm_km_free_poolpage1(struct vm_map *map, vaddr_t addr)
map               742 uvm/uvm_km.c   	uvm_km_free(map, addr, PAGE_SIZE);
map               212 uvm/uvm_loan.c uvm_loan(map, start, len, result, flags)
map               213 uvm/uvm_loan.c 	struct vm_map *map;
map               224 uvm/uvm_loan.c 	if (map->flags & VM_MAP_INTRSAFE)
map               254 uvm/uvm_loan.c 		ufi.orig_map = map;
map               143 uvm/uvm_map.c  #define uvm_map_entry_link(map, after_where, entry) do { \
map               144 uvm/uvm_map.c  	(map)->nentries++; \
map               149 uvm/uvm_map.c  	uvm_rb_insert(map, entry); \
map               157 uvm/uvm_map.c  #define uvm_map_entry_unlink(map, entry) do { \
map               158 uvm/uvm_map.c  	(map)->nentries--; \
map               161 uvm/uvm_map.c  	uvm_rb_remove(map, entry); \
map               169 uvm/uvm_map.c  #define SAVE_HINT(map,check,value) do { \
map               170 uvm/uvm_map.c  	simple_lock(&(map)->hint_lock); \
map               171 uvm/uvm_map.c  	if ((map)->hint == (check)) \
map               172 uvm/uvm_map.c  		(map)->hint = (value); \
map               173 uvm/uvm_map.c  	simple_unlock(&(map)->hint_lock); \
map               182 uvm/uvm_map.c  #define VM_MAP_RANGE_CHECK(map, start, end) do { \
map               183 uvm/uvm_map.c  	if (start < vm_map_min(map)) 		\
map               184 uvm/uvm_map.c  		start = vm_map_min(map);        \
map               185 uvm/uvm_map.c  	if (end > vm_map_max(map))              \
map               186 uvm/uvm_map.c  		end = vm_map_max(map);          \
map               214 uvm/uvm_map.c  int _uvm_tree_sanity(struct vm_map *map, const char *name);
map               242 uvm/uvm_map.c  uvm_rb_space(struct vm_map *map, struct vm_map_entry *entry)
map               247 uvm/uvm_map.c  	if ((next = entry->next) == &map->header)
map               248 uvm/uvm_map.c  		space = map->max_offset - entry->end;
map               278 uvm/uvm_map.c  uvm_rb_fixup(struct vm_map *map, struct vm_map_entry *entry)
map               282 uvm/uvm_map.c  		entry->ownspace = uvm_rb_space(map, entry);
map               288 uvm/uvm_map.c  uvm_rb_insert(struct vm_map *map, struct vm_map_entry *entry)
map               290 uvm/uvm_map.c  	vaddr_t space = uvm_rb_space(map, entry);
map               294 uvm/uvm_map.c  	tmp = RB_INSERT(uvm_tree, &(map)->rbhead, entry);
map               299 uvm/uvm_map.c  	uvm_rb_fixup(map, entry);
map               300 uvm/uvm_map.c  	if (entry->prev != &map->header)
map               301 uvm/uvm_map.c  		uvm_rb_fixup(map, entry->prev);
map               305 uvm/uvm_map.c  uvm_rb_remove(struct vm_map *map, struct vm_map_entry *entry)
map               310 uvm/uvm_map.c  	RB_REMOVE(uvm_tree, &(map)->rbhead, entry);
map               311 uvm/uvm_map.c  	if (entry->prev != &map->header)
map               312 uvm/uvm_map.c  		uvm_rb_fixup(map, entry->prev);
map               314 uvm/uvm_map.c  		uvm_rb_fixup(map, parent);
map               325 uvm/uvm_map.c  _uvm_tree_sanity(struct vm_map *map, const char *name)
map               330 uvm/uvm_map.c  	RB_FOREACH(tmp, uvm_tree, &map->rbhead) {
map               331 uvm/uvm_map.c  		if (tmp->ownspace != uvm_rb_space(map, tmp)) {
map               333 uvm/uvm_map.c  			    name, n + 1, map->nentries,
map               334 uvm/uvm_map.c  			    tmp->ownspace, uvm_rb_space(map, tmp),
map               335 uvm/uvm_map.c  			    tmp->next == &map->header ? "(last)" : "");
map               340 uvm/uvm_map.c  	RB_FOREACH(tmp, uvm_tree, &map->rbhead) {
map               356 uvm/uvm_map.c  	if (n != map->nentries) {
map               358 uvm/uvm_map.c  		    name, n, map->nentries);
map               362 uvm/uvm_map.c  	for (tmp = map->header.next; tmp && tmp != &map->header;
map               364 uvm/uvm_map.c  		trtmp = RB_FIND(uvm_tree, &map->rbhead, tmp);
map               388 uvm/uvm_map.c  uvm_mapent_alloc(struct vm_map *map)
map               394 uvm/uvm_map.c  	if (map->flags & VM_MAP_INTRSAFE || cold) {
map               419 uvm/uvm_map.c  	} else if (map == kernel_map) {
map               430 uvm/uvm_map.c  	    ((map->flags & VM_MAP_INTRSAFE) != 0 || map == kernel_map), 0, 0);
map               482 uvm/uvm_map.c  uvm_map_entry_unwire(struct vm_map *map, struct vm_map_entry *entry)
map               486 uvm/uvm_map.c  	uvm_fault_unwire_locked(map, entry->start, entry->end);
map               581 uvm/uvm_map.c  uvm_map_clip_start(struct vm_map *map, struct vm_map_entry *entry,
map               589 uvm/uvm_map.c  	uvm_tree_sanity(map, "clip_start entry");
map               597 uvm/uvm_map.c  	new_entry = uvm_mapent_alloc(map);
map               612 uvm/uvm_map.c  	uvm_map_entry_link(map, entry->prev, new_entry);
map               625 uvm/uvm_map.c  	uvm_tree_sanity(map, "clip_start leave");
map               638 uvm/uvm_map.c  uvm_map_clip_end(struct vm_map *map, struct vm_map_entry *entry, vaddr_t end)
map               643 uvm/uvm_map.c  	uvm_tree_sanity(map, "clip_end entry");
map               649 uvm/uvm_map.c  	new_entry = uvm_mapent_alloc(map);
map               660 uvm/uvm_map.c  	uvm_rb_fixup(map, entry);
map               662 uvm/uvm_map.c  	uvm_map_entry_link(map, entry, new_entry);
map               674 uvm/uvm_map.c  	uvm_tree_sanity(map, "clip_end leave");
map               708 uvm/uvm_map.c  uvm_map_p(struct vm_map *map, vaddr_t *startp, vsize_t size,
map               722 uvm/uvm_map.c  	    map, *startp, size, flags);
map               725 uvm/uvm_map.c  	uvm_tree_sanity(map, "map entry");
map               727 uvm/uvm_map.c  	if ((map->flags & VM_MAP_INTRSAFE) == 0)
map               744 uvm/uvm_map.c  	if (vm_map_lock_try(map) == FALSE) {
map               747 uvm/uvm_map.c  		vm_map_lock(map); /* could sleep here */
map               749 uvm/uvm_map.c  	if ((prev_entry = uvm_map_findspace(map, *startp, size, startp, 
map               752 uvm/uvm_map.c  		vm_map_unlock(map);
map               762 uvm/uvm_map.c  		if (map == kernel_map && uvm_maxkaddr < (*startp + size))
map               802 uvm/uvm_map.c  	    prev_entry->end == *startp && prev_entry != &map->header &&
map               838 uvm/uvm_map.c  				vm_map_unlock(map);
map               855 uvm/uvm_map.c  		uvm_rb_fixup(map, prev_entry);
map               856 uvm/uvm_map.c  		map->size += size;
map               860 uvm/uvm_map.c  		uvm_tree_sanity(map, "map leave 2");
map               863 uvm/uvm_map.c  		vm_map_unlock(map);
map               876 uvm/uvm_map.c  	    prev_entry->next != &map->header && 
map               884 uvm/uvm_map.c  	new_entry = uvm_mapent_alloc(map);
map               921 uvm/uvm_map.c  	uvm_map_entry_link(map, prev_entry, new_entry);
map               923 uvm/uvm_map.c  	map->size += size;
map               932 uvm/uvm_map.c  	if ((map->first_free == prev_entry) &&
map               934 uvm/uvm_map.c  		map->first_free = new_entry;
map               936 uvm/uvm_map.c  	uvm_tree_sanity(map, "map leave");
map               939 uvm/uvm_map.c  	vm_map_unlock(map);
map               952 uvm/uvm_map.c  uvm_map_lookup_entry(struct vm_map *map, vaddr_t address,
map               962 uvm/uvm_map.c  	    map, address, entry, 0);
map               969 uvm/uvm_map.c  	simple_lock(&map->hint_lock);
map               970 uvm/uvm_map.c  	cur = map->hint;
map               971 uvm/uvm_map.c  	simple_unlock(&map->hint_lock);
map               973 uvm/uvm_map.c  	if (cur == &map->header)
map               989 uvm/uvm_map.c  		last = &map->header;
map               998 uvm/uvm_map.c  		if (map->nentries > 30)
map              1005 uvm/uvm_map.c  		cur = map->header.next;
map              1009 uvm/uvm_map.c  	uvm_tree_sanity(map, __func__);
map              1012 uvm/uvm_map.c  		struct vm_map_entry *prev = &map->header;
map              1013 uvm/uvm_map.c  		cur = RB_ROOT(&map->rbhead);
map              1023 uvm/uvm_map.c  					SAVE_HINT(map, map->hint, cur);
map              1049 uvm/uvm_map.c  				SAVE_HINT(map, map->hint, cur);
map              1060 uvm/uvm_map.c  	SAVE_HINT(map, map->hint, *entry);
map              1072 uvm/uvm_map.c  uvm_map_spacefits(struct vm_map *map, vaddr_t *phint, vsize_t length,
map              1092 uvm/uvm_map.c  	if (end > map->max_offset || end < hint)
map              1094 uvm/uvm_map.c  	if (after != NULL && after != &map->header && after->start < end)
map              1145 uvm/uvm_map.c  uvm_map_findspace(struct vm_map *map, vaddr_t hint, vsize_t length,
map              1157 uvm/uvm_map.c  		    map, hint, length, flags);
map              1161 uvm/uvm_map.c  	uvm_tree_sanity(map, "map_findspace entry");
map              1170 uvm/uvm_map.c  	if (hint < map->min_offset) {	/* check ranges ... */
map              1175 uvm/uvm_map.c  		hint = map->min_offset;
map              1177 uvm/uvm_map.c  	if (hint > map->max_offset) {
map              1179 uvm/uvm_map.c  				hint, map->min_offset, map->max_offset, 0);
map              1188 uvm/uvm_map.c  	if ((flags & UVM_FLAG_FIXED) == 0 && hint == map->min_offset) {
map              1189 uvm/uvm_map.c  		if ((entry = map->first_free) != &map->header) 
map              1192 uvm/uvm_map.c  		if (uvm_map_lookup_entry(map, hint, &tmp)) {
map              1206 uvm/uvm_map.c  		if (end > map->max_offset || end < hint) {
map              1211 uvm/uvm_map.c  		if (next == &map->header || next->start >= end)
map              1220 uvm/uvm_map.c  	if (uvm_map_spacefits(map, &hint, length, entry->next, uoffset, align))
map              1224 uvm/uvm_map.c  	tmp = RB_ROOT(&map->rbhead);
map              1259 uvm/uvm_map.c  		if (uvm_map_spacefits(map, &hint, length, tmp->next, uoffset,
map              1270 uvm/uvm_map.c  	if (uvm_map_spacefits(map, &hint, length, prev->next, uoffset,
map              1291 uvm/uvm_map.c  	if (uvm_map_spacefits(map, &hint, length, tmp->next, uoffset, align)) {
map              1332 uvm/uvm_map.c  		if (end > map->max_offset || end < hint) {
map              1337 uvm/uvm_map.c  		if (next == &map->header || next->start >= end)
map              1341 uvm/uvm_map.c  	SAVE_HINT(map, map->hint, entry);
map              1351 uvm/uvm_map.c  		return (uvm_map_findspace(map, orig_hint,
map              1371 uvm/uvm_map.c  uvm_unmap_remove(struct vm_map *map, vaddr_t start, vaddr_t end,
map              1380 uvm/uvm_map.c  	    map, start, end, 0);
map              1382 uvm/uvm_map.c  	VM_MAP_RANGE_CHECK(map, start, end);
map              1384 uvm/uvm_map.c  	uvm_tree_sanity(map, "unmap_remove entry");
map              1386 uvm/uvm_map.c  	if ((map->flags & VM_MAP_INTRSAFE) == 0)
map              1392 uvm/uvm_map.c  	if (uvm_map_lookup_entry(map, start, &first_entry) == TRUE) {
map              1395 uvm/uvm_map.c  		UVM_MAP_CLIP_START(map, entry, start);
map              1397 uvm/uvm_map.c  		SAVE_HINT(map, entry, entry->prev);
map              1407 uvm/uvm_map.c  	if (map->first_free->start >= start)
map              1408 uvm/uvm_map.c  		map->first_free = entry->prev;
map              1434 uvm/uvm_map.c  	while ((entry != &map->header) && (entry->start < end)) {
map              1436 uvm/uvm_map.c  		UVM_MAP_CLIP_END(map, entry, end); 
map              1448 uvm/uvm_map.c  			uvm_map_entry_unwire(map, entry);
map              1454 uvm/uvm_map.c  		if (map->flags & VM_MAP_INTRSAFE) {
map              1459 uvm/uvm_map.c  			KASSERT(vm_map_pmap(map) == pmap_kernel());
map              1512 uvm/uvm_map.c  			pmap_remove(map->pmap, entry->start, entry->end);
map              1522 uvm/uvm_map.c  		SAVE_HINT(map, entry, entry->prev);
map              1524 uvm/uvm_map.c  		uvm_map_entry_unlink(map, entry);
map              1525 uvm/uvm_map.c  		map->size -= len;
map              1531 uvm/uvm_map.c  		pmap_update(vm_map_pmap(map));
map              1535 uvm/uvm_map.c  	uvm_tree_sanity(map, "unmap_remove leave");
map              1609 uvm/uvm_map.c  uvm_map_reserve(struct vm_map *map, vsize_t size, vaddr_t offset,
map              1615 uvm/uvm_map.c  	      map,size,offset,raddr);
map              1618 uvm/uvm_map.c  	if (*raddr < vm_map_min(map))
map              1619 uvm/uvm_map.c  		*raddr = vm_map_min(map);                /* hint */
map              1625 uvm/uvm_map.c  	if (uvm_map(map, raddr, size, NULL, offset, 0,
map              1648 uvm/uvm_map.c  uvm_map_replace(struct vm_map *map, vaddr_t start, vaddr_t end,
map              1653 uvm/uvm_map.c  	uvm_tree_sanity(map, "map_replace entry");
map              1659 uvm/uvm_map.c  	if (!uvm_map_lookup_entry(map, start, &oldent)) {
map              1714 uvm/uvm_map.c  		SAVE_HINT(map, map->hint, newents);
map              1715 uvm/uvm_map.c  		if (map->first_free == oldent)
map              1716 uvm/uvm_map.c  			map->first_free = last;
map              1722 uvm/uvm_map.c  		uvm_rb_remove(map, oldent);
map              1726 uvm/uvm_map.c  		map->nentries = map->nentries + (nnewents - 1);
map              1735 uvm/uvm_map.c  				uvm_rb_insert(map, tmp);
map              1742 uvm/uvm_map.c  		SAVE_HINT(map, map->hint, oldent->prev);
map              1743 uvm/uvm_map.c  		if (map->first_free == oldent)
map              1744 uvm/uvm_map.c  			map->first_free = oldent->prev;
map              1747 uvm/uvm_map.c  		uvm_map_entry_unlink(map, oldent);
map              1751 uvm/uvm_map.c  	uvm_tree_sanity(map, "map_replace leave");
map              2119 uvm/uvm_map.c  uvm_map_submap(struct vm_map *map, vaddr_t start, vaddr_t end,
map              2125 uvm/uvm_map.c  	vm_map_lock(map);
map              2127 uvm/uvm_map.c  	VM_MAP_RANGE_CHECK(map, start, end);
map              2129 uvm/uvm_map.c  	if (uvm_map_lookup_entry(map, start, &entry)) {
map              2130 uvm/uvm_map.c  		UVM_MAP_CLIP_START(map, entry, start);
map              2131 uvm/uvm_map.c  		UVM_MAP_CLIP_END(map, entry, end);		/* to be safe */
map              2148 uvm/uvm_map.c  	vm_map_unlock(map);
map              2165 uvm/uvm_map.c  uvm_map_protect(struct vm_map *map, vaddr_t start, vaddr_t end,
map              2172 uvm/uvm_map.c  		    map, start, end, new_prot);
map              2174 uvm/uvm_map.c  	vm_map_lock(map);
map              2176 uvm/uvm_map.c  	VM_MAP_RANGE_CHECK(map, start, end);
map              2178 uvm/uvm_map.c  	if (uvm_map_lookup_entry(map, start, &entry)) {
map              2179 uvm/uvm_map.c  		UVM_MAP_CLIP_START(map, entry, start);
map              2189 uvm/uvm_map.c  	while ((current != &map->header) && (current->start < end)) {
map              2205 uvm/uvm_map.c  	while ((current != &map->header) && (current->start < end)) {
map              2208 uvm/uvm_map.c  		UVM_MAP_CLIP_END(map, current, end);
map              2227 uvm/uvm_map.c  			pmap_protect(map->pmap, current->start, current->end,
map              2237 uvm/uvm_map.c  		if ((map->flags & VM_MAP_WIREFUTURE) != 0 &&
map              2241 uvm/uvm_map.c  			if (uvm_map_pageable(map, entry->start, entry->end,
map              2262 uvm/uvm_map.c  	pmap_update(map->pmap);
map              2265 uvm/uvm_map.c  	vm_map_unlock(map);
map              2282 uvm/uvm_map.c  uvm_map_inherit(struct vm_map *map, vaddr_t start, vaddr_t end,
map              2288 uvm/uvm_map.c  	    map, start, end, new_inheritance);
map              2300 uvm/uvm_map.c  	vm_map_lock(map);
map              2302 uvm/uvm_map.c  	VM_MAP_RANGE_CHECK(map, start, end);
map              2304 uvm/uvm_map.c  	if (uvm_map_lookup_entry(map, start, &temp_entry)) {
map              2306 uvm/uvm_map.c  		UVM_MAP_CLIP_START(map, entry, start);
map              2311 uvm/uvm_map.c  	while ((entry != &map->header) && (entry->start < end)) {
map              2312 uvm/uvm_map.c  		UVM_MAP_CLIP_END(map, entry, end);
map              2317 uvm/uvm_map.c  	vm_map_unlock(map);
map              2329 uvm/uvm_map.c  uvm_map_advice(struct vm_map *map, vaddr_t start, vaddr_t end, int new_advice)
map              2334 uvm/uvm_map.c  	    map, start, end, new_advice);
map              2336 uvm/uvm_map.c  	vm_map_lock(map);
map              2337 uvm/uvm_map.c  	VM_MAP_RANGE_CHECK(map, start, end);
map              2338 uvm/uvm_map.c  	if (uvm_map_lookup_entry(map, start, &temp_entry)) {
map              2340 uvm/uvm_map.c  		UVM_MAP_CLIP_START(map, entry, start);
map              2349 uvm/uvm_map.c  	while ((entry != &map->header) && (entry->start < end)) {
map              2350 uvm/uvm_map.c  		UVM_MAP_CLIP_END(map, entry, end);
map              2360 uvm/uvm_map.c  			vm_map_unlock(map);
map              2368 uvm/uvm_map.c  	vm_map_unlock(map);
map              2388 uvm/uvm_map.c  uvm_map_pageable(struct vm_map *map, vaddr_t start, vaddr_t end,
map              2398 uvm/uvm_map.c  		    map, start, end, new_pageable);
map              2399 uvm/uvm_map.c  	KASSERT(map->flags & VM_MAP_PAGEABLE);
map              2402 uvm/uvm_map.c  		vm_map_lock(map);
map              2404 uvm/uvm_map.c  	VM_MAP_RANGE_CHECK(map, start, end);
map              2414 uvm/uvm_map.c  	if (uvm_map_lookup_entry(map, start, &start_entry) == FALSE) {
map              2416 uvm/uvm_map.c  			vm_map_unlock(map);
map              2428 uvm/uvm_map.c  		UVM_MAP_CLIP_START(map, entry, start);
map              2435 uvm/uvm_map.c  		while ((entry != &map->header) && (entry->start < end)) {
map              2438 uvm/uvm_map.c  			     (entry->next == &map->header ||
map              2441 uvm/uvm_map.c  					vm_map_unlock(map);
map              2456 uvm/uvm_map.c  		while ((entry != &map->header) && (entry->start < end)) {
map              2457 uvm/uvm_map.c  			UVM_MAP_CLIP_END(map, entry, end);
map              2459 uvm/uvm_map.c  				uvm_map_entry_unwire(map, entry);
map              2463 uvm/uvm_map.c  			vm_map_unlock(map);
map              2489 uvm/uvm_map.c  	while ((entry != &map->header) && (entry->start < end)) {
map              2504 uvm/uvm_map.c  					amap_copy(map, entry, M_WAITOK, TRUE,
map              2510 uvm/uvm_map.c  		UVM_MAP_CLIP_START(map, entry, start);
map              2511 uvm/uvm_map.c  		UVM_MAP_CLIP_END(map, entry, end);
map              2520 uvm/uvm_map.c  		     (entry->next == &map->header ||
map              2528 uvm/uvm_map.c  			while (entry != &map->header && entry->end > start) {
map              2533 uvm/uvm_map.c  				vm_map_unlock(map);
map              2545 uvm/uvm_map.c  	timestamp_save = map->timestamp;
map              2547 uvm/uvm_map.c  	vm_map_busy(map);
map              2548 uvm/uvm_map.c  	vm_map_downgrade(map);
map              2552 uvm/uvm_map.c  	while (entry != &map->header && entry->start < end) {
map              2554 uvm/uvm_map.c  			rv = uvm_fault_wire(map, entry->start, entry->end,
map              2574 uvm/uvm_map.c  		vm_map_upgrade(map);
map              2575 uvm/uvm_map.c  		vm_map_unbusy(map);
map              2578 uvm/uvm_map.c  		if (timestamp_save != map->timestamp)
map              2588 uvm/uvm_map.c  		while (entry != &map->header && entry->start < end) {
map              2602 uvm/uvm_map.c  				uvm_map_entry_unwire(map, entry);
map              2606 uvm/uvm_map.c  			vm_map_unlock(map);
map              2613 uvm/uvm_map.c  		vm_map_unbusy(map);
map              2614 uvm/uvm_map.c  		vm_map_unlock_read(map);
map              2621 uvm/uvm_map.c  		vm_map_upgrade(map);
map              2622 uvm/uvm_map.c  		vm_map_unbusy(map);
map              2639 uvm/uvm_map.c  uvm_map_pageable_all(struct vm_map *map, int flags, vsize_t limit)
map              2648 uvm/uvm_map.c  	UVMHIST_LOG(maphist,"(map=%p,flags=0x%lx)", map, flags, 0, 0);
map              2650 uvm/uvm_map.c  	KASSERT(map->flags & VM_MAP_PAGEABLE);
map              2652 uvm/uvm_map.c  	vm_map_lock(map);
map              2663 uvm/uvm_map.c  		for (entry = map->header.next; entry != &map->header;
map              2666 uvm/uvm_map.c  				uvm_map_entry_unwire(map, entry);
map              2668 uvm/uvm_map.c  		vm_map_modflags(map, 0, VM_MAP_WIREFUTURE);
map              2669 uvm/uvm_map.c  		vm_map_unlock(map);
map              2682 uvm/uvm_map.c  		vm_map_modflags(map, VM_MAP_WIREFUTURE, 0);
map              2690 uvm/uvm_map.c  		vm_map_unlock(map);
map              2717 uvm/uvm_map.c  	for (size = 0, entry = map->header.next; entry != &map->header;
map              2726 uvm/uvm_map.c  		vm_map_unlock(map);
map              2733 uvm/uvm_map.c  	    (size + ptoa(pmap_wired_count(vm_map_pmap(map))) > limit)) {
map              2734 uvm/uvm_map.c  		vm_map_unlock(map);
map              2743 uvm/uvm_map.c  	for (entry = map->header.next; entry != &map->header;
map              2759 uvm/uvm_map.c  					amap_copy(map, entry, M_WAITOK, TRUE,
map              2773 uvm/uvm_map.c  	timestamp_save = map->timestamp;
map              2775 uvm/uvm_map.c  	vm_map_busy(map);
map              2776 uvm/uvm_map.c  	vm_map_downgrade(map);
map              2778 uvm/uvm_map.c  	for (error = 0, entry = map->header.next;
map              2779 uvm/uvm_map.c  	    entry != &map->header && error == 0;
map              2782 uvm/uvm_map.c  			error = uvm_fault_wire(map, entry->start, entry->end,
map              2791 uvm/uvm_map.c  		vm_map_upgrade(map);
map              2792 uvm/uvm_map.c  		vm_map_unbusy(map);
map              2795 uvm/uvm_map.c  		if (timestamp_save != map->timestamp)
map              2806 uvm/uvm_map.c  		for (/* nothing */; entry != &map->header;
map              2819 uvm/uvm_map.c  		for (entry = map->header.next; entry != failed_entry;
map              2825 uvm/uvm_map.c  				uvm_map_entry_unwire(map, entry);
map              2827 uvm/uvm_map.c  		vm_map_unlock(map);
map              2833 uvm/uvm_map.c  	vm_map_unbusy(map);
map              2834 uvm/uvm_map.c  	vm_map_unlock_read(map);
map              2859 uvm/uvm_map.c  uvm_map_clean(struct vm_map *map, vaddr_t start, vaddr_t end, int flags)
map              2872 uvm/uvm_map.c  		    map, start, end, flags);
map              2876 uvm/uvm_map.c  	vm_map_lock_read(map);
map              2877 uvm/uvm_map.c  	VM_MAP_RANGE_CHECK(map, start, end);
map              2878 uvm/uvm_map.c  	if (uvm_map_lookup_entry(map, start, &entry) == FALSE) {
map              2879 uvm/uvm_map.c  		vm_map_unlock_read(map);
map              2889 uvm/uvm_map.c  			vm_map_unlock_read(map);
map              2892 uvm/uvm_map.c  		if (end > current->end && (current->next == &map->header ||
map              2894 uvm/uvm_map.c  			vm_map_unlock_read(map);
map              3039 uvm/uvm_map.c  	vm_map_unlock_read(map);
map              3052 uvm/uvm_map.c  uvm_map_checkprot(struct vm_map *map, vaddr_t start, vaddr_t end,
map              3058 uvm/uvm_map.c  	if (!uvm_map_lookup_entry(map, start, &tmp_entry)) {
map              3063 uvm/uvm_map.c  		if (entry == &map->header) {
map              3190 uvm/uvm_map.c  	struct vm_map *map = &ovm->vm_map;
map              3217 uvm/uvm_map.c  		vm_map_lock(map);
map              3218 uvm/uvm_map.c  		vm_map_modflags(map, 0, VM_MAP_WIREFUTURE);
map              3219 uvm/uvm_map.c  		vm_map_unlock(map);
map              3224 uvm/uvm_map.c  		uvm_unmap(map, map->min_offset, map->max_offset);
map              3229 uvm/uvm_map.c  		vm_map_lock(map);
map              3230 uvm/uvm_map.c  		map->min_offset = start;
map              3231 uvm/uvm_map.c  		uvm_tree_sanity(map, "resize enter");
map              3232 uvm/uvm_map.c  		map->max_offset = end;
map              3233 uvm/uvm_map.c  		if (map->header.prev != &map->header)
map              3234 uvm/uvm_map.c  			uvm_rb_fixup(map, map->header.prev);
map              3235 uvm/uvm_map.c  		uvm_tree_sanity(map, "resize leave");
map              3236 uvm/uvm_map.c  		vm_map_unlock(map);
map              3247 uvm/uvm_map.c  			 (map->flags & VM_MAP_PAGEABLE) ? TRUE : FALSE);
map              3615 uvm/uvm_map.c  uvm_map_printit(struct vm_map *map, boolean_t full,
map              3620 uvm/uvm_map.c  	(*pr)("MAP %p: [0x%lx->0x%lx]\n", map, map->min_offset,map->max_offset);
map              3622 uvm/uvm_map.c  	    map->nentries, map->size, map->ref_count, map->timestamp,
map              3623 uvm/uvm_map.c  	    map->flags);
map              3625 uvm/uvm_map.c  	(*pr)("\tpmap=%p(resident=%d)\n", map->pmap, 
map              3626 uvm/uvm_map.c  	    pmap_resident_count(map->pmap));
map              3629 uvm/uvm_map.c  	(*pr)("\tpmap=%p(resident=<<NOT SUPPORTED!!!>>)\n", map->pmap);
map              3633 uvm/uvm_map.c  	for (entry = map->header.next; entry != &map->header;
map               246 uvm/uvm_map.h  #define	vm_map_modflags(map, set, clear)				\
map               248 uvm/uvm_map.h  	(map)->flags = ((map)->flags | (set)) & ~(clear);		\
map               358 uvm/uvm_map.h  vm_map_lock_try(struct vm_map *map)
map               362 uvm/uvm_map.h  	if (map->flags & VM_MAP_INTRSAFE) {
map               365 uvm/uvm_map.h  		if (map->flags & VM_MAP_BUSY) {
map               368 uvm/uvm_map.h  		rv = (rw_enter(&map->lock, RW_WRITE|RW_NOSLEEP) == 0);
map               372 uvm/uvm_map.h  		map->timestamp++;
map               378 uvm/uvm_map.h  vm_map_lock(struct vm_map *map)
map               380 uvm/uvm_map.h  	if (map->flags & VM_MAP_INTRSAFE)
map               384 uvm/uvm_map.h  		while (map->flags & VM_MAP_BUSY) {
map               385 uvm/uvm_map.h  			map->flags |= VM_MAP_WANTLOCK;
map               386 uvm/uvm_map.h  			tsleep(&map->flags, PVM, (char *)vmmapbsy, 0);
map               388 uvm/uvm_map.h  	} while (rw_enter(&map->lock, RW_WRITE|RW_SLEEPFAIL) != 0);
map               390 uvm/uvm_map.h  	map->timestamp++;
map               393 uvm/uvm_map.h  #define	vm_map_lock_read(map) rw_enter_read(&(map)->lock)
map               395 uvm/uvm_map.h  #define	vm_map_unlock(map)						\
map               397 uvm/uvm_map.h  	if (((map)->flags & VM_MAP_INTRSAFE) == 0)			\
map               398 uvm/uvm_map.h  		rw_exit(&(map)->lock);					\
map               401 uvm/uvm_map.h  #define	vm_map_unlock_read(map)	rw_exit_read(&(map)->lock)
map               403 uvm/uvm_map.h  #define	vm_map_downgrade(map) rw_enter(&(map)->lock, RW_DOWNGRADE)
map               405 uvm/uvm_map.h  #define	vm_map_upgrade(map)						\
map               407 uvm/uvm_map.h  	rw_exit_read(&(map)->lock);					\
map               408 uvm/uvm_map.h  	rw_enter_write(&(map)->lock);					\
map               411 uvm/uvm_map.h  #define	vm_map_busy(map)						\
map               413 uvm/uvm_map.h  	(map)->flags |= VM_MAP_BUSY;					\
map               416 uvm/uvm_map.h  #define	vm_map_unbusy(map)						\
map               420 uvm/uvm_map.h  	oflags = (map)->flags;						\
map               421 uvm/uvm_map.h  	(map)->flags &= ~(VM_MAP_BUSY|VM_MAP_WANTLOCK);			\
map               423 uvm/uvm_map.h  		wakeup(&(map)->flags);					\
map               430 uvm/uvm_map.h  #define		vm_map_min(map)		((map)->min_offset)
map               431 uvm/uvm_map.h  #define		vm_map_max(map)		((map)->max_offset)
map               432 uvm/uvm_map.h  #define		vm_map_pmap(map)	((map)->pmap)
map               108 uvm/uvm_map_i.h uvm_map_setup(map, min, max, flags)
map               109 uvm/uvm_map_i.h 	vm_map_t map;
map               114 uvm/uvm_map_i.h 	RB_INIT(&map->rbhead);
map               115 uvm/uvm_map_i.h 	map->header.next = map->header.prev = &map->header;
map               116 uvm/uvm_map_i.h 	map->nentries = 0;
map               117 uvm/uvm_map_i.h 	map->size = 0;
map               118 uvm/uvm_map_i.h 	map->ref_count = 1;
map               119 uvm/uvm_map_i.h 	map->min_offset = min;
map               120 uvm/uvm_map_i.h 	map->max_offset = max;
map               121 uvm/uvm_map_i.h 	map->flags = flags;
map               122 uvm/uvm_map_i.h 	map->first_free = &map->header;
map               123 uvm/uvm_map_i.h 	map->hint = &map->header;
map               124 uvm/uvm_map_i.h 	map->timestamp = 0;
map               125 uvm/uvm_map_i.h 	rw_init(&map->lock, "vmmaplk");
map               126 uvm/uvm_map_i.h 	simple_lock_init(&map->ref_lock);
map               127 uvm/uvm_map_i.h 	simple_lock_init(&map->hint_lock);
map               143 uvm/uvm_map_i.h uvm_unmap_p(map, start, end, p)
map               144 uvm/uvm_map_i.h 	vm_map_t map;
map               152 uvm/uvm_map_i.h 	    map, start, end, 0);
map               157 uvm/uvm_map_i.h 	vm_map_lock(map);
map               158 uvm/uvm_map_i.h 	uvm_unmap_remove(map, start, end, &dead_entries, p);
map               159 uvm/uvm_map_i.h 	vm_map_unlock(map);
map               175 uvm/uvm_map_i.h uvm_map_reference(map)
map               176 uvm/uvm_map_i.h 	vm_map_t map;
map               178 uvm/uvm_map_i.h 	simple_lock(&map->ref_lock);
map               179 uvm/uvm_map_i.h 	map->ref_count++; 
map               180 uvm/uvm_map_i.h 	simple_unlock(&map->ref_lock);
map               191 uvm/uvm_map_i.h uvm_map_deallocate(map)
map               192 uvm/uvm_map_i.h 	vm_map_t map;
map               196 uvm/uvm_map_i.h 	simple_lock(&map->ref_lock);
map               197 uvm/uvm_map_i.h 	c = --map->ref_count;
map               198 uvm/uvm_map_i.h 	simple_unlock(&map->ref_lock);
map               207 uvm/uvm_map_i.h 	uvm_unmap(map, map->min_offset, map->max_offset);
map               208 uvm/uvm_map_i.h 	pmap_destroy(map->pmap);
map               209 uvm/uvm_map_i.h 	FREE(map, M_VMMAP);
map               235 uvm/uvm_meter.c 	struct vm_map *map;
map               279 uvm/uvm_meter.c 		vm_map_lock(map);
map               280 uvm/uvm_meter.c 		for (map = &p->p_vmspace->vm_map, entry = map->header.next;
map               281 uvm/uvm_meter.c 		    entry != &map->header; entry = entry->next) {
map               287 uvm/uvm_meter.c 		vm_map_unlock(map);
map               256 uvm/uvm_mmap.c 	vm_map_t map;
map               260 uvm/uvm_mmap.c 	map = &p->p_vmspace->vm_map;
map               282 uvm/uvm_mmap.c 	vm_map_lock_read(map);
map               284 uvm/uvm_mmap.c 	if (uvm_map_lookup_entry(map, start, &entry) == FALSE) {
map               290 uvm/uvm_mmap.c 	     entry != &map->header && entry->start < end;
map               297 uvm/uvm_mmap.c 		     (entry->next == &map->header ||
map               364 uvm/uvm_mmap.c 	vm_map_unlock_read(map);
map               635 uvm/uvm_mmap.c 	vm_map_t map;
map               665 uvm/uvm_mmap.c 	map = &p->p_vmspace->vm_map;
map               680 uvm/uvm_mmap.c 		vm_map_lock_read(map);
map               681 uvm/uvm_mmap.c 		rv = uvm_map_lookup_entry(map, addr, &entry);
map               686 uvm/uvm_mmap.c 		vm_map_unlock_read(map);
map               702 uvm/uvm_mmap.c 	return (uvm_map_clean(map, addr, addr+size, uvmflags));
map               721 uvm/uvm_mmap.c 	vm_map_t map;
map               747 uvm/uvm_mmap.c 	map = &p->p_vmspace->vm_map;
map               750 uvm/uvm_mmap.c 	vm_map_lock(map);	/* lock map so we can checkprot */
map               757 uvm/uvm_mmap.c 	if (!uvm_map_checkprot(map, addr, addr + size, VM_PROT_NONE)) {
map               758 uvm/uvm_mmap.c 		vm_map_unlock(map);
map               765 uvm/uvm_mmap.c 	uvm_unmap_remove(map, addr, addr + size, &dead_entries, p);
map               767 uvm/uvm_mmap.c 	vm_map_unlock(map);	/* and unlock */
map              1089 uvm/uvm_mmap.c uvm_mmap(map, addr, size, prot, maxprot, flags, handle, foff, locklimit, p)
map              1090 uvm/uvm_mmap.c 	vm_map_t map;
map              1129 uvm/uvm_mmap.c 		uvm_unmap_p(map, *addr, *addr + size, p);	/* zap! */
map              1223 uvm/uvm_mmap.c 	error = uvm_map_p(map, addr, size, uobj, foff, align, uvmflag, p);
map              1237 uvm/uvm_mmap.c 		vm_map_lock(map);
map              1239 uvm/uvm_mmap.c 		if (map->flags & VM_MAP_WIREFUTURE) {
map              1243 uvm/uvm_mmap.c 			         ptoa(pmap_wired_count(vm_map_pmap(map)))) >
map              1248 uvm/uvm_mmap.c 				vm_map_unlock(map);
map              1250 uvm/uvm_mmap.c 				uvm_unmap(map, *addr, *addr + size);
map              1257 uvm/uvm_mmap.c 			error = uvm_map_pageable(map, *addr, *addr + size,
map              1261 uvm/uvm_mmap.c 				uvm_unmap(map, *addr, *addr + size);
map              1267 uvm/uvm_mmap.c 		vm_map_unlock(map);
map               187 uvm/uvm_unix.c 	vm_map_t map = &vm->vm_map;
map               196 uvm/uvm_unix.c 	for (entry = map->header.next; entry != &map->header;
map                54 uvm/uvm_user.c uvm_deallocate(struct vm_map *map, vaddr_t start, vsize_t size)
map                57 uvm/uvm_user.c 	if (map == NULL)
map                63 uvm/uvm_user.c 	uvm_unmap(map, trunc_page(start), round_page(start+size));