rd 81 arch/i386/i386/conf.c bdev_disk_init(NRD,rd), /* 17: ram disk driver */ rd 261 arch/i386/i386/conf.c cdev_disk_init(NRD,rd), /* 47: ram disk driver */ rd 2721 arch/i386/i386/machdep.c setregion(struct region_descriptor *rd, void *base, size_t limit) rd 2724 arch/i386/i386/machdep.c rd->rd_limit = (int)limit; rd 2725 arch/i386/i386/machdep.c rd->rd_base = (int)base; rd 544 dev/ic/acx.c struct acx_ring_data *rd = &sc->sc_ring_data; rd 586 dev/ic/acx.c bzero(rd->tx_ring, ACX_TX_RING_SIZE); rd 599 dev/ic/acx.c bzero(rd->rx_ring, ACX_RX_RING_SIZE); rd 1280 dev/ic/acx.c struct acx_ring_data *rd = &sc->sc_ring_data; rd 1285 dev/ic/acx.c bus_dmamap_sync(sc->sc_dmat, rd->rx_ring_dmamap, 0, rd 1286 dev/ic/acx.c rd->rx_ring_dmamap->dm_mapsize, BUS_DMASYNC_POSTREAD); rd 1405 dev/ic/acx.c bus_dmamap_sync(sc->sc_dmat, rd->rx_ring_dmamap, 0, rd 1406 dev/ic/acx.c rd->rx_ring_dmamap->dm_mapsize, BUS_DMASYNC_PREWRITE); rd 1887 dev/ic/acx.c struct acx_ring_data *rd = &sc->sc_ring_data; rd 1894 dev/ic/acx.c ACX_RX_RING_SIZE, 0, BUS_DMA_NOWAIT, &rd->rx_ring_dmamap); rd 1903 dev/ic/acx.c 0, &rd->rx_ring_seg, 1, &nsegs, BUS_DMA_NOWAIT); rd 1911 dev/ic/acx.c error = bus_dmamem_map(sc->sc_dmat, &rd->rx_ring_seg, nsegs, rd 1912 dev/ic/acx.c ACX_RX_RING_SIZE, (caddr_t *)&rd->rx_ring, rd 1921 dev/ic/acx.c error = bus_dmamap_load(sc->sc_dmat, rd->rx_ring_dmamap, rd 1922 dev/ic/acx.c rd->rx_ring, ACX_RX_RING_SIZE, NULL, BUS_DMA_WAITOK); rd 1927 dev/ic/acx.c bus_dmamem_free(sc->sc_dmat, &rd->rx_ring_seg, 1); rd 1931 dev/ic/acx.c rd->rx_ring_paddr = rd->rx_ring_dmamap->dm_segs[0].ds_addr; rd 1935 dev/ic/acx.c ACX_TX_RING_SIZE, 0, BUS_DMA_NOWAIT, &rd->tx_ring_dmamap); rd 1943 dev/ic/acx.c 0, &rd->tx_ring_seg, 1, &nsegs, BUS_DMA_NOWAIT); rd 1951 dev/ic/acx.c error = bus_dmamem_map(sc->sc_dmat, &rd->tx_ring_seg, nsegs, rd 1952 dev/ic/acx.c ACX_TX_RING_SIZE, (caddr_t *)&rd->tx_ring, BUS_DMA_NOWAIT); rd 1960 dev/ic/acx.c error = bus_dmamap_load(sc->sc_dmat, rd->tx_ring_dmamap, rd 1961 dev/ic/acx.c rd->tx_ring, ACX_TX_RING_SIZE, NULL, BUS_DMA_WAITOK); rd 1965 dev/ic/acx.c bus_dmamem_free(sc->sc_dmat, &rd->tx_ring_seg, 1); rd 1969 dev/ic/acx.c rd->tx_ring_paddr = rd->tx_ring_dmamap->dm_segs[0].ds_addr; rd 1989 dev/ic/acx.c bd->rx_buf[i].rb_desc = &rd->rx_ring[i]; rd 2001 dev/ic/acx.c bd->tx_buf[i].tb_desc1 = &rd->tx_ring[i * 2]; rd 2002 dev/ic/acx.c bd->tx_buf[i].tb_desc2 = &rd->tx_ring[(i * 2) + 1]; rd 2011 dev/ic/acx.c struct acx_ring_data *rd = &sc->sc_ring_data; rd 2015 dev/ic/acx.c if (rd->rx_ring != NULL) { rd 2016 dev/ic/acx.c bus_dmamap_unload(sc->sc_dmat, rd->rx_ring_dmamap); rd 2017 dev/ic/acx.c bus_dmamem_free(sc->sc_dmat, &rd->rx_ring_seg, 1); rd 2020 dev/ic/acx.c if (rd->tx_ring != NULL) { rd 2021 dev/ic/acx.c bus_dmamap_unload(sc->sc_dmat, rd->tx_ring_dmamap); rd 2022 dev/ic/acx.c bus_dmamem_free(sc->sc_dmat, &rd->tx_ring_seg, 1); rd 2056 dev/ic/acx.c struct acx_ring_data *rd; rd 2061 dev/ic/acx.c rd = &sc->sc_ring_data; rd 2062 dev/ic/acx.c paddr = rd->tx_ring_paddr; rd 2066 dev/ic/acx.c bzero(&rd->tx_ring[i], sizeof(struct acx_host_desc)); rd 2067 dev/ic/acx.c rd->tx_ring[i].h_ctrl = htole16(DESC_CTRL_HOSTOWN); rd 2070 dev/ic/acx.c rd->tx_ring[i].h_next_desc = htole32(rd->tx_ring_paddr); rd 2072 dev/ic/acx.c rd->tx_ring[i].h_next_desc = htole32(paddr); rd 2075 dev/ic/acx.c bus_dmamap_sync(sc->sc_dmat, rd->tx_ring_dmamap, 0, rd 2076 dev/ic/acx.c rd->tx_ring_dmamap->dm_mapsize, BUS_DMASYNC_PREWRITE); rd 2089 dev/ic/acx.c struct acx_ring_data *rd; rd 2095 dev/ic/acx.c rd = &sc->sc_ring_data; rd 2096 dev/ic/acx.c paddr = rd->rx_ring_paddr; rd 2102 dev/ic/acx.c bzero(&rd->rx_ring[i], sizeof(struct acx_host_desc)); rd 2109 dev/ic/acx.c rd->rx_ring[i].h_next_desc = htole32(rd->rx_ring_paddr); rd 2111 dev/ic/acx.c rd->rx_ring[i].h_next_desc = htole32(paddr); rd 2114 dev/ic/acx.c bus_dmamap_sync(sc->sc_dmat, rd->rx_ring_dmamap, 0, rd 2115 dev/ic/acx.c rd->rx_ring_dmamap->dm_mapsize, BUS_DMASYNC_PREWRITE); rd 2178 dev/ic/acx.c struct acx_ring_data *rd = &sc->sc_ring_data; rd 2308 dev/ic/acx.c bus_dmamap_sync(sc->sc_dmat, rd->tx_ring_dmamap, 0, rd 2309 dev/ic/acx.c rd->tx_ring_dmamap->dm_mapsize, BUS_DMASYNC_PREWRITE); rd 203 dev/ic/bt463.c bt463_register(v, sched_update, wr, rd) rd 207 dev/ic/bt463.c u_int8_t (*rd)(void *, u_int); rd 221 dev/ic/bt463.c data->ramdac_rd = rd; rd 231 dev/ic/bt463.c bt463_cninit(v, sched_update, wr, rd) rd 235 dev/ic/bt463.c u_int8_t (*rd)(void *, u_int); rd 241 dev/ic/bt463.c data->ramdac_rd = rd; rd 147 dev/ic/bt485.c bt485_register(v, sched_update, wr, rd) rd 151 dev/ic/bt485.c u_int8_t (*rd)(void *, u_int); rd 165 dev/ic/bt485.c data->ramdac_rd = rd; rd 175 dev/ic/bt485.c bt485_cninit(v, sched_update, wr, rd) rd 179 dev/ic/bt485.c u_int8_t (*rd)(void *, u_int); rd 185 dev/ic/bt485.c data->ramdac_rd = rd; rd 153 dev/ic/ibm561.c ibm561_register(v, sched_update, wr, rd) rd 157 dev/ic/ibm561.c u_int8_t (*rd)(void *, u_int); rd 169 dev/ic/ibm561.c data->ramdac_rd = rd; rd 180 dev/ic/ibm561.c ibm561_cninit(v, sched_update, wr, rd, dotclock) rd 184 dev/ic/ibm561.c u_int8_t (*rd)(void *, u_int); rd 191 dev/ic/ibm561.c data->ramdac_rd = rd; rd 1054 dev/ic/rtw.c struct rtw_rxdesc *rd = &rdb->rdb_desc[idx]; rd 1056 dev/ic/rtw.c obuf = rd->rd_buf; rd 1057 dev/ic/rtw.c rd->rd_buf = htole32(rs->rs_dmamap->dm_segs[0].ds_addr); rd 1065 dev/ic/rtw.c octl = rd->rd_ctl; rd 1066 dev/ic/rtw.c rd->rd_ctl = htole32(ctl); rd 1070 dev/ic/rtw.c ("%s: rd %p buf %08x -> %08x ctl %08x -> %08x\n", __func__, rd, rd 1071 dev/ic/rtw.c letoh32(obuf), letoh32(rd->rd_buf), letoh32(octl), rd 1072 dev/ic/rtw.c letoh32(rd->rd_ctl))); rd 1089 dev/ic/rtw.c struct rtw_rxdesc *rd; rd 1093 dev/ic/rtw.c rd = &rdb->rdb_desc[i]; rd 1137 dev/ic/rtw.c struct rtw_rxdesc *rd; rd 1152 dev/ic/rtw.c rd = &rdb->rdb_desc[next]; rd 1155 dev/ic/rtw.c hstat = letoh32(rd->rd_stat); rd 1156 dev/ic/rtw.c hrssi = letoh32(rd->rd_rssi); rd 1157 dev/ic/rtw.c htsfth = letoh32(rd->rd_tsfth); rd 1158 dev/ic/rtw.c htsftl = letoh32(rd->rd_tsftl); rd 1174 dev/ic/rtw.c rd = &rdb->rdb_desc[0]; rd 1175 dev/ic/rtw.c if ((rd->rd_stat & htole32(RTW_RXSTAT_OWN)) != 0) rd 1574 dev/ic/rtw.c struct rtw_rxdesc *rd; rd 1592 dev/ic/rtw.c rd = &rdb->rdb_desc[desc]; rd 1596 dev/ic/rtw.c letoh32(rd->rd_ctl), letoh32(rd->rd_rssi), rd 1597 dev/ic/rtw.c letoh32(rd->rd_buf), letoh32(rd->rd_tsfth)); rd 729 dev/pci/eso.c int mode, r[2], rd[2], clk; rd 784 dev/pci/eso.c (128 - (rd[0] = 128 - ESO_CLK0 / p->sample_rate)); rd 786 dev/pci/eso.c (128 - (rd[1] = 128 - ESO_CLK1 / p->sample_rate)); rd 789 dev/pci/eso.c srg = rd[clk] | (clk == 1 ? ESO_CLK1_SELECT : 0x00); rd 430 dev/pci/if_msk.c struct msk_ring_data *rd = sc_if->sk_rdata; rd 433 dev/pci/if_msk.c bzero((char *)rd->sk_rx_ring, rd 437 dev/pci/if_msk.c cd->sk_rx_chain[i].sk_le = &rd->sk_rx_ring[i]; rd 465 dev/pci/if_msk.c struct msk_ring_data *rd = sc_if->sk_rdata; rd 475 dev/pci/if_msk.c cd->sk_tx_chain[i].sk_le = &rd->sk_tx_ring[i]; rd 555 dev/pci/if_sk.c struct sk_ring_data *rd = sc_if->sk_rdata; rd 558 dev/pci/if_sk.c bzero((char *)rd->sk_rx_ring, rd 562 dev/pci/if_sk.c cd->sk_rx_chain[i].sk_desc = &rd->sk_rx_ring[i]; rd 568 dev/pci/if_sk.c rd->sk_rx_ring[i].sk_next = htole32(SK_RX_RING_ADDR(sc_if, nexti)); rd 569 dev/pci/if_sk.c rd->sk_rx_ring[i].sk_csum1_start = htole16(ETHER_HDR_LEN); rd 570 dev/pci/if_sk.c rd->sk_rx_ring[i].sk_csum2_start = htole16(ETHER_HDR_LEN + rd 594 dev/pci/if_sk.c struct sk_ring_data *rd = sc_if->sk_rdata; rd 604 dev/pci/if_sk.c cd->sk_tx_chain[i].sk_desc = &rd->sk_tx_ring[i]; rd 610 dev/pci/if_sk.c rd->sk_tx_ring[i].sk_next = htole32(SK_TX_RING_ADDR(sc_if, nexti)); rd 140 dev/pci/lofnreg.h #define LOFN_INSTR(done,op,rd,ra,rb,rm) \ rd 142 dev/pci/lofnreg.h (((rd) & OP_R_MASK) << OP_RD_SHIFT) | \ rd 147 dev/pci/lofnreg.h #define LOFN_INSTR2(done,op,rd,ra,len) \ rd 149 dev/pci/lofnreg.h (((rd) & OP_R_MASK) << OP_RD_SHIFT) | \ rd 702 dev/pci/noct.c u_int32_t reg, rd, wr; rd 706 dev/pci/noct.c rd = (reg & RNGQPTR_READ_M) >> RNGQPTR_READ_S; rd 709 dev/pci/noct.c while (rd != wr && cons < 32) { rd 710 dev/pci/noct.c val = sc->sc_rngbuf[rd]; rd 713 dev/pci/noct.c if (++rd == NOCT_RNG_ENTRIES) rd 714 dev/pci/noct.c rd = 0; rd 719 dev/pci/noct.c NOCT_WRITE_4(sc, NOCT_RNG_Q_PTR, rd); rd 307 dev/pci/nofnreg.h #define NOFN_PK_INSTR(done,op,rd,ra,rb,rm) \ rd 309 dev/pci/nofnreg.h (((rd) & PK_OP_R_MASK) << PK_OP_RD_SHIFT) | \ rd 315 dev/pci/nofnreg.h #define NOFN_PK_INSTR2(done,op,rd,ra,len) \ rd 317 dev/pci/nofnreg.h (((rd) & PK_OP_R_MASK) << PK_OP_RD_SHIFT) | \ rd 55 dev/rd.c rd_attach_hook(unit, rd) rd 57 dev/rd.c struct rd_conf *rd; rd 61 dev/rd.c rd->rd_addr = (caddr_t) rd_root_image; rd 62 dev/rd.c rd->rd_size = (size_t) rd_root_size; rd 63 dev/rd.c rd->rd_type = RD_KMEM_FIXED; rd 72 dev/rd.c rd_open_hook(unit, rd) rd 74 dev/rd.c struct rd_conf *rd; rd 2229 dev/usb/ehci.c int rd, usbd_xfer_handle xfer, ehci_soft_qtd_t **sp, ehci_soft_qtd_t **ep) rd 2247 dev/usb/ehci.c EHCI_QTD_SET_PID(rd ? EHCI_QTD_PID_IN : EHCI_QTD_PID_OUT) | rd 445 dev/usb/ohci.c int alen, int rd, usbd_xfer_handle xfer, rd 462 dev/usb/ohci.c (rd ? OHCI_TD_IN : OHCI_TD_OUT) | rd 509 dev/usb/ohci.c if (!rd && (flags & USBD_FORCE_SHORT_XFER) && rd 1686 dev/usb/uhci.c int rd, u_int16_t flags, usb_dma_t *dma, rd 1745 dev/usb/uhci.c htole32(rd ? UHCI_TD_IN (l, endpt, addr, tog) : rd 2528 dev/usb/uhci.c int rd = UE_GET_DIR(endpt) == UE_DIR_IN; rd 2538 dev/usb/uhci.c token = rd ? UHCI_TD_IN (0, endpt, addr, 0) : rd 418 dev/usb/uhid.c struct usb_ctl_report_desc *rd; rd 455 dev/usb/uhid.c rd = (struct usb_ctl_report_desc *)addr; rd 456 dev/usb/uhid.c size = min(size, sizeof rd->ucrd_data); rd 457 dev/usb/uhid.c rd->ucrd_size = size; rd 458 dev/usb/uhid.c memcpy(rd->ucrd_data, desc, size); rd 600 sys/conf.h bdev_decl(rd); rd 601 sys/conf.h cdev_decl(rd);