rd 81 arch/i386/i386/conf.c bdev_disk_init(NRD,rd), /* 17: ram disk driver */
rd 261 arch/i386/i386/conf.c cdev_disk_init(NRD,rd), /* 47: ram disk driver */
rd 2721 arch/i386/i386/machdep.c setregion(struct region_descriptor *rd, void *base, size_t limit)
rd 2724 arch/i386/i386/machdep.c rd->rd_limit = (int)limit;
rd 2725 arch/i386/i386/machdep.c rd->rd_base = (int)base;
rd 544 dev/ic/acx.c struct acx_ring_data *rd = &sc->sc_ring_data;
rd 586 dev/ic/acx.c bzero(rd->tx_ring, ACX_TX_RING_SIZE);
rd 599 dev/ic/acx.c bzero(rd->rx_ring, ACX_RX_RING_SIZE);
rd 1280 dev/ic/acx.c struct acx_ring_data *rd = &sc->sc_ring_data;
rd 1285 dev/ic/acx.c bus_dmamap_sync(sc->sc_dmat, rd->rx_ring_dmamap, 0,
rd 1286 dev/ic/acx.c rd->rx_ring_dmamap->dm_mapsize, BUS_DMASYNC_POSTREAD);
rd 1405 dev/ic/acx.c bus_dmamap_sync(sc->sc_dmat, rd->rx_ring_dmamap, 0,
rd 1406 dev/ic/acx.c rd->rx_ring_dmamap->dm_mapsize, BUS_DMASYNC_PREWRITE);
rd 1887 dev/ic/acx.c struct acx_ring_data *rd = &sc->sc_ring_data;
rd 1894 dev/ic/acx.c ACX_RX_RING_SIZE, 0, BUS_DMA_NOWAIT, &rd->rx_ring_dmamap);
rd 1903 dev/ic/acx.c 0, &rd->rx_ring_seg, 1, &nsegs, BUS_DMA_NOWAIT);
rd 1911 dev/ic/acx.c error = bus_dmamem_map(sc->sc_dmat, &rd->rx_ring_seg, nsegs,
rd 1912 dev/ic/acx.c ACX_RX_RING_SIZE, (caddr_t *)&rd->rx_ring,
rd 1921 dev/ic/acx.c error = bus_dmamap_load(sc->sc_dmat, rd->rx_ring_dmamap,
rd 1922 dev/ic/acx.c rd->rx_ring, ACX_RX_RING_SIZE, NULL, BUS_DMA_WAITOK);
rd 1927 dev/ic/acx.c bus_dmamem_free(sc->sc_dmat, &rd->rx_ring_seg, 1);
rd 1931 dev/ic/acx.c rd->rx_ring_paddr = rd->rx_ring_dmamap->dm_segs[0].ds_addr;
rd 1935 dev/ic/acx.c ACX_TX_RING_SIZE, 0, BUS_DMA_NOWAIT, &rd->tx_ring_dmamap);
rd 1943 dev/ic/acx.c 0, &rd->tx_ring_seg, 1, &nsegs, BUS_DMA_NOWAIT);
rd 1951 dev/ic/acx.c error = bus_dmamem_map(sc->sc_dmat, &rd->tx_ring_seg, nsegs,
rd 1952 dev/ic/acx.c ACX_TX_RING_SIZE, (caddr_t *)&rd->tx_ring, BUS_DMA_NOWAIT);
rd 1960 dev/ic/acx.c error = bus_dmamap_load(sc->sc_dmat, rd->tx_ring_dmamap,
rd 1961 dev/ic/acx.c rd->tx_ring, ACX_TX_RING_SIZE, NULL, BUS_DMA_WAITOK);
rd 1965 dev/ic/acx.c bus_dmamem_free(sc->sc_dmat, &rd->tx_ring_seg, 1);
rd 1969 dev/ic/acx.c rd->tx_ring_paddr = rd->tx_ring_dmamap->dm_segs[0].ds_addr;
rd 1989 dev/ic/acx.c bd->rx_buf[i].rb_desc = &rd->rx_ring[i];
rd 2001 dev/ic/acx.c bd->tx_buf[i].tb_desc1 = &rd->tx_ring[i * 2];
rd 2002 dev/ic/acx.c bd->tx_buf[i].tb_desc2 = &rd->tx_ring[(i * 2) + 1];
rd 2011 dev/ic/acx.c struct acx_ring_data *rd = &sc->sc_ring_data;
rd 2015 dev/ic/acx.c if (rd->rx_ring != NULL) {
rd 2016 dev/ic/acx.c bus_dmamap_unload(sc->sc_dmat, rd->rx_ring_dmamap);
rd 2017 dev/ic/acx.c bus_dmamem_free(sc->sc_dmat, &rd->rx_ring_seg, 1);
rd 2020 dev/ic/acx.c if (rd->tx_ring != NULL) {
rd 2021 dev/ic/acx.c bus_dmamap_unload(sc->sc_dmat, rd->tx_ring_dmamap);
rd 2022 dev/ic/acx.c bus_dmamem_free(sc->sc_dmat, &rd->tx_ring_seg, 1);
rd 2056 dev/ic/acx.c struct acx_ring_data *rd;
rd 2061 dev/ic/acx.c rd = &sc->sc_ring_data;
rd 2062 dev/ic/acx.c paddr = rd->tx_ring_paddr;
rd 2066 dev/ic/acx.c bzero(&rd->tx_ring[i], sizeof(struct acx_host_desc));
rd 2067 dev/ic/acx.c rd->tx_ring[i].h_ctrl = htole16(DESC_CTRL_HOSTOWN);
rd 2070 dev/ic/acx.c rd->tx_ring[i].h_next_desc = htole32(rd->tx_ring_paddr);
rd 2072 dev/ic/acx.c rd->tx_ring[i].h_next_desc = htole32(paddr);
rd 2075 dev/ic/acx.c bus_dmamap_sync(sc->sc_dmat, rd->tx_ring_dmamap, 0,
rd 2076 dev/ic/acx.c rd->tx_ring_dmamap->dm_mapsize, BUS_DMASYNC_PREWRITE);
rd 2089 dev/ic/acx.c struct acx_ring_data *rd;
rd 2095 dev/ic/acx.c rd = &sc->sc_ring_data;
rd 2096 dev/ic/acx.c paddr = rd->rx_ring_paddr;
rd 2102 dev/ic/acx.c bzero(&rd->rx_ring[i], sizeof(struct acx_host_desc));
rd 2109 dev/ic/acx.c rd->rx_ring[i].h_next_desc = htole32(rd->rx_ring_paddr);
rd 2111 dev/ic/acx.c rd->rx_ring[i].h_next_desc = htole32(paddr);
rd 2114 dev/ic/acx.c bus_dmamap_sync(sc->sc_dmat, rd->rx_ring_dmamap, 0,
rd 2115 dev/ic/acx.c rd->rx_ring_dmamap->dm_mapsize, BUS_DMASYNC_PREWRITE);
rd 2178 dev/ic/acx.c struct acx_ring_data *rd = &sc->sc_ring_data;
rd 2308 dev/ic/acx.c bus_dmamap_sync(sc->sc_dmat, rd->tx_ring_dmamap, 0,
rd 2309 dev/ic/acx.c rd->tx_ring_dmamap->dm_mapsize, BUS_DMASYNC_PREWRITE);
rd 203 dev/ic/bt463.c bt463_register(v, sched_update, wr, rd)
rd 207 dev/ic/bt463.c u_int8_t (*rd)(void *, u_int);
rd 221 dev/ic/bt463.c data->ramdac_rd = rd;
rd 231 dev/ic/bt463.c bt463_cninit(v, sched_update, wr, rd)
rd 235 dev/ic/bt463.c u_int8_t (*rd)(void *, u_int);
rd 241 dev/ic/bt463.c data->ramdac_rd = rd;
rd 147 dev/ic/bt485.c bt485_register(v, sched_update, wr, rd)
rd 151 dev/ic/bt485.c u_int8_t (*rd)(void *, u_int);
rd 165 dev/ic/bt485.c data->ramdac_rd = rd;
rd 175 dev/ic/bt485.c bt485_cninit(v, sched_update, wr, rd)
rd 179 dev/ic/bt485.c u_int8_t (*rd)(void *, u_int);
rd 185 dev/ic/bt485.c data->ramdac_rd = rd;
rd 153 dev/ic/ibm561.c ibm561_register(v, sched_update, wr, rd)
rd 157 dev/ic/ibm561.c u_int8_t (*rd)(void *, u_int);
rd 169 dev/ic/ibm561.c data->ramdac_rd = rd;
rd 180 dev/ic/ibm561.c ibm561_cninit(v, sched_update, wr, rd, dotclock)
rd 184 dev/ic/ibm561.c u_int8_t (*rd)(void *, u_int);
rd 191 dev/ic/ibm561.c data->ramdac_rd = rd;
rd 1054 dev/ic/rtw.c struct rtw_rxdesc *rd = &rdb->rdb_desc[idx];
rd 1056 dev/ic/rtw.c obuf = rd->rd_buf;
rd 1057 dev/ic/rtw.c rd->rd_buf = htole32(rs->rs_dmamap->dm_segs[0].ds_addr);
rd 1065 dev/ic/rtw.c octl = rd->rd_ctl;
rd 1066 dev/ic/rtw.c rd->rd_ctl = htole32(ctl);
rd 1070 dev/ic/rtw.c ("%s: rd %p buf %08x -> %08x ctl %08x -> %08x\n", __func__, rd,
rd 1071 dev/ic/rtw.c letoh32(obuf), letoh32(rd->rd_buf), letoh32(octl),
rd 1072 dev/ic/rtw.c letoh32(rd->rd_ctl)));
rd 1089 dev/ic/rtw.c struct rtw_rxdesc *rd;
rd 1093 dev/ic/rtw.c rd = &rdb->rdb_desc[i];
rd 1137 dev/ic/rtw.c struct rtw_rxdesc *rd;
rd 1152 dev/ic/rtw.c rd = &rdb->rdb_desc[next];
rd 1155 dev/ic/rtw.c hstat = letoh32(rd->rd_stat);
rd 1156 dev/ic/rtw.c hrssi = letoh32(rd->rd_rssi);
rd 1157 dev/ic/rtw.c htsfth = letoh32(rd->rd_tsfth);
rd 1158 dev/ic/rtw.c htsftl = letoh32(rd->rd_tsftl);
rd 1174 dev/ic/rtw.c rd = &rdb->rdb_desc[0];
rd 1175 dev/ic/rtw.c if ((rd->rd_stat & htole32(RTW_RXSTAT_OWN)) != 0)
rd 1574 dev/ic/rtw.c struct rtw_rxdesc *rd;
rd 1592 dev/ic/rtw.c rd = &rdb->rdb_desc[desc];
rd 1596 dev/ic/rtw.c letoh32(rd->rd_ctl), letoh32(rd->rd_rssi),
rd 1597 dev/ic/rtw.c letoh32(rd->rd_buf), letoh32(rd->rd_tsfth));
rd 729 dev/pci/eso.c int mode, r[2], rd[2], clk;
rd 784 dev/pci/eso.c (128 - (rd[0] = 128 - ESO_CLK0 / p->sample_rate));
rd 786 dev/pci/eso.c (128 - (rd[1] = 128 - ESO_CLK1 / p->sample_rate));
rd 789 dev/pci/eso.c srg = rd[clk] | (clk == 1 ? ESO_CLK1_SELECT : 0x00);
rd 430 dev/pci/if_msk.c struct msk_ring_data *rd = sc_if->sk_rdata;
rd 433 dev/pci/if_msk.c bzero((char *)rd->sk_rx_ring,
rd 437 dev/pci/if_msk.c cd->sk_rx_chain[i].sk_le = &rd->sk_rx_ring[i];
rd 465 dev/pci/if_msk.c struct msk_ring_data *rd = sc_if->sk_rdata;
rd 475 dev/pci/if_msk.c cd->sk_tx_chain[i].sk_le = &rd->sk_tx_ring[i];
rd 555 dev/pci/if_sk.c struct sk_ring_data *rd = sc_if->sk_rdata;
rd 558 dev/pci/if_sk.c bzero((char *)rd->sk_rx_ring,
rd 562 dev/pci/if_sk.c cd->sk_rx_chain[i].sk_desc = &rd->sk_rx_ring[i];
rd 568 dev/pci/if_sk.c rd->sk_rx_ring[i].sk_next = htole32(SK_RX_RING_ADDR(sc_if, nexti));
rd 569 dev/pci/if_sk.c rd->sk_rx_ring[i].sk_csum1_start = htole16(ETHER_HDR_LEN);
rd 570 dev/pci/if_sk.c rd->sk_rx_ring[i].sk_csum2_start = htole16(ETHER_HDR_LEN +
rd 594 dev/pci/if_sk.c struct sk_ring_data *rd = sc_if->sk_rdata;
rd 604 dev/pci/if_sk.c cd->sk_tx_chain[i].sk_desc = &rd->sk_tx_ring[i];
rd 610 dev/pci/if_sk.c rd->sk_tx_ring[i].sk_next = htole32(SK_TX_RING_ADDR(sc_if, nexti));
rd 140 dev/pci/lofnreg.h #define LOFN_INSTR(done,op,rd,ra,rb,rm) \
rd 142 dev/pci/lofnreg.h (((rd) & OP_R_MASK) << OP_RD_SHIFT) | \
rd 147 dev/pci/lofnreg.h #define LOFN_INSTR2(done,op,rd,ra,len) \
rd 149 dev/pci/lofnreg.h (((rd) & OP_R_MASK) << OP_RD_SHIFT) | \
rd 702 dev/pci/noct.c u_int32_t reg, rd, wr;
rd 706 dev/pci/noct.c rd = (reg & RNGQPTR_READ_M) >> RNGQPTR_READ_S;
rd 709 dev/pci/noct.c while (rd != wr && cons < 32) {
rd 710 dev/pci/noct.c val = sc->sc_rngbuf[rd];
rd 713 dev/pci/noct.c if (++rd == NOCT_RNG_ENTRIES)
rd 714 dev/pci/noct.c rd = 0;
rd 719 dev/pci/noct.c NOCT_WRITE_4(sc, NOCT_RNG_Q_PTR, rd);
rd 307 dev/pci/nofnreg.h #define NOFN_PK_INSTR(done,op,rd,ra,rb,rm) \
rd 309 dev/pci/nofnreg.h (((rd) & PK_OP_R_MASK) << PK_OP_RD_SHIFT) | \
rd 315 dev/pci/nofnreg.h #define NOFN_PK_INSTR2(done,op,rd,ra,len) \
rd 317 dev/pci/nofnreg.h (((rd) & PK_OP_R_MASK) << PK_OP_RD_SHIFT) | \
rd 55 dev/rd.c rd_attach_hook(unit, rd)
rd 57 dev/rd.c struct rd_conf *rd;
rd 61 dev/rd.c rd->rd_addr = (caddr_t) rd_root_image;
rd 62 dev/rd.c rd->rd_size = (size_t) rd_root_size;
rd 63 dev/rd.c rd->rd_type = RD_KMEM_FIXED;
rd 72 dev/rd.c rd_open_hook(unit, rd)
rd 74 dev/rd.c struct rd_conf *rd;
rd 2229 dev/usb/ehci.c int rd, usbd_xfer_handle xfer, ehci_soft_qtd_t **sp, ehci_soft_qtd_t **ep)
rd 2247 dev/usb/ehci.c EHCI_QTD_SET_PID(rd ? EHCI_QTD_PID_IN : EHCI_QTD_PID_OUT) |
rd 445 dev/usb/ohci.c int alen, int rd, usbd_xfer_handle xfer,
rd 462 dev/usb/ohci.c (rd ? OHCI_TD_IN : OHCI_TD_OUT) |
rd 509 dev/usb/ohci.c if (!rd && (flags & USBD_FORCE_SHORT_XFER) &&
rd 1686 dev/usb/uhci.c int rd, u_int16_t flags, usb_dma_t *dma,
rd 1745 dev/usb/uhci.c htole32(rd ? UHCI_TD_IN (l, endpt, addr, tog) :
rd 2528 dev/usb/uhci.c int rd = UE_GET_DIR(endpt) == UE_DIR_IN;
rd 2538 dev/usb/uhci.c token = rd ? UHCI_TD_IN (0, endpt, addr, 0) :
rd 418 dev/usb/uhid.c struct usb_ctl_report_desc *rd;
rd 455 dev/usb/uhid.c rd = (struct usb_ctl_report_desc *)addr;
rd 456 dev/usb/uhid.c size = min(size, sizeof rd->ucrd_data);
rd 457 dev/usb/uhid.c rd->ucrd_size = size;
rd 458 dev/usb/uhid.c memcpy(rd->ucrd_data, desc, size);
rd 600 sys/conf.h bdev_decl(rd);
rd 601 sys/conf.h cdev_decl(rd);