paddr 2058 dev/ic/acx.c uint32_t paddr; paddr 2062 dev/ic/acx.c paddr = rd->tx_ring_paddr; paddr 2064 dev/ic/acx.c paddr += sizeof(struct acx_host_desc); paddr 2072 dev/ic/acx.c rd->tx_ring[i].h_next_desc = htole32(paddr); paddr 2091 dev/ic/acx.c uint32_t paddr; paddr 2096 dev/ic/acx.c paddr = rd->rx_ring_paddr; paddr 2101 dev/ic/acx.c paddr += sizeof(struct acx_host_desc); paddr 2111 dev/ic/acx.c rd->rx_ring[i].h_next_desc = htole32(paddr); paddr 2128 dev/ic/acx.c uint32_t paddr; paddr 2162 dev/ic/acx.c paddr = rb->rb_mbuf_dmamap->dm_segs[0].ds_addr; paddr 2165 dev/ic/acx.c rb->rb_desc->h_data_paddr = htole32(paddr); paddr 2181 dev/ic/acx.c uint32_t paddr; paddr 2280 dev/ic/acx.c paddr = txbuf->tb_mbuf_dmamap->dm_segs[0].ds_addr; paddr 2281 dev/ic/acx.c txbuf->tb_desc1->h_data_paddr = htole32(paddr); paddr 2282 dev/ic/acx.c txbuf->tb_desc2->h_data_paddr = htole32(paddr + ACX_FRAME_HDRLEN); paddr 410 dev/pci/if_wpi.c dma->paddr = dma->map->dm_segs[0].ds_addr; paddr 519 dev/pci/if_wpi.c rbuf->paddr = ring->buf_dma.paddr + i * WPI_RBUF_SIZE; paddr 574 dev/pci/if_wpi.c ring->desc[i] = htole32(rbuf->paddr); paddr 639 dev/pci/if_wpi.c sc->shared->txbase[qid] = htole32(ring->desc_dma.paddr); paddr 1077 dev/pci/if_wpi.c wpi_mem_write(sc, WPI_MEM_DATA_BASE, dma->paddr); paddr 1080 dev/pci/if_wpi.c dma->paddr + WPI_FW_INIT_DATA_MAXSZ); paddr 1108 dev/pci/if_wpi.c wpi_mem_write(sc, WPI_MEM_DATA_BASE, dma->paddr); paddr 1111 dev/pci/if_wpi.c dma->paddr + WPI_FW_MAIN_DATA_MAXSZ); paddr 1254 dev/pci/if_wpi.c ring->desc[ring->cur] = htole32(rbuf->paddr); paddr 1743 dev/pci/if_wpi.c desc->segs[0].addr = htole32(ring->cmd_dma.paddr + paddr 2043 dev/pci/if_wpi.c desc->segs[0].addr = htole32(ring->cmd_dma.paddr + paddr 2329 dev/pci/if_wpi.c desc->segs[0].addr = htole32(ring->cmd_dma.paddr + paddr 2805 dev/pci/if_wpi.c WPI_WRITE(sc, WPI_RX_BASE, sc->rxq.desc_dma.paddr); paddr 2806 dev/pci/if_wpi.c WPI_WRITE(sc, WPI_RX_RIDX_PTR, sc->shared_dma.paddr + paddr 2822 dev/pci/if_wpi.c WPI_WRITE(sc, WPI_TX_BASE_PTR, sc->shared_dma.paddr); paddr 60 dev/pci/if_wpivar.h bus_addr_t paddr; paddr 90 dev/pci/if_wpivar.h bus_addr_t paddr; paddr 266 dev/sbus/qec.c bus_addr_t paddr; paddr 273 dev/sbus/qec.c paddr = sc->sc_range[i].poffset + offset; paddr 276 dev/sbus/qec.c (t, t0, BUS_ADDR(iospace, paddr), size, flags, hp)); paddr 545 kern/uipc_socket.c soreceive(struct socket *so, struct mbuf **paddr, struct uio *uio, paddr 558 kern/uipc_socket.c if (paddr) paddr 559 kern/uipc_socket.c *paddr = 0; paddr 673 kern/uipc_socket.c if (paddr) paddr 674 kern/uipc_socket.c *paddr = m_copy(m, 0, m->m_len); paddr 678 kern/uipc_socket.c if (paddr) { paddr 679 kern/uipc_socket.c *paddr = m; paddr 299 sys/socketvar.h int soreceive(struct socket *so, struct mbuf **paddr, struct uio *uio, paddr 387 uvm/uvm_device.c paddr_t paddr, mdpgno; paddr 444 uvm/uvm_device.c paddr = pmap_phys_address(mdpgno); paddr 448 uvm/uvm_device.c ufi->orig_map->pmap, curr_va, (u_long)paddr, mapprot); paddr 449 uvm/uvm_device.c if (pmap_enter(ufi->orig_map->pmap, curr_va, paddr, paddr 223 uvm/uvm_page.c paddr_t paddr; paddr 320 uvm/uvm_page.c paddr = ptoa(vm_physmem[lcv].start); paddr 321 uvm/uvm_page.c for (i = 0 ; i < n ; i++, paddr += PAGE_SIZE) { paddr 322 uvm/uvm_page.c vm_physmem[lcv].pgs[i].phys_addr = paddr; paddr 326 uvm/uvm_page.c if (atop(paddr) >= vm_physmem[lcv].avail_start && paddr 327 uvm/uvm_page.c atop(paddr) <= vm_physmem[lcv].avail_end) { paddr 422 uvm/uvm_page.c paddr_t paddr; paddr 470 uvm/uvm_page.c if (!uvm_page_physget(&paddr)) paddr 477 uvm/uvm_page.c pmap_kenter_pa(vaddr, paddr, VM_PROT_READ|VM_PROT_WRITE); paddr 666 uvm/uvm_page.c paddr_t paddr; paddr 678 uvm/uvm_page.c for (lcv = 0, paddr = ptoa(start) ; paddr 679 uvm/uvm_page.c lcv < npages ; lcv++, paddr += PAGE_SIZE) { paddr 680 uvm/uvm_page.c pgs[lcv].phys_addr = paddr; paddr 682 uvm/uvm_page.c if (atop(paddr) >= avail_start && paddr 683 uvm/uvm_page.c atop(paddr) <= avail_end)