cur 959 arch/i386/i386/est.c u_int16_t idhi, idlo, cur;
cur 977 arch/i386/i386/est.c cur = msr & 0xffff;
cur 980 arch/i386/i386/est.c crcur = (cur >> 8) & 0xff;
cur 1020 arch/i386/i386/est.c if (cur == idhi || cur == idlo) {
cur 1030 arch/i386/i386/est.c fake_table[1] = cur;
cur 1039 arch/i386/i386/est.c mhz = MSR2MHZ(cur, bus_clock);
cur 1040 arch/i386/i386/est.c mv = MSR2MV(cur);
cur 1047 arch/i386/i386/est.c if (cur == est_fqlist->table[i])
cur 378 compat/ibcs2/ibcs2_fcntl.c off_t off, cur;
cur 405 compat/ibcs2/ibcs2_fcntl.c if ((error = sys_lseek(p, &ols, (register_t *)&cur)) != 0)
cur 417 compat/ibcs2/ibcs2_fcntl.c off = cur - (off_t)ifl.l_start;
cur 495 dev/ic/an.c int cur, id;
cur 508 dev/ic/an.c cur = sc->sc_txcur;
cur 509 dev/ic/an.c if (sc->sc_txd[cur].d_fid == id) {
cur 510 dev/ic/an.c sc->sc_txd[cur].d_inuse = 0;
cur 511 dev/ic/an.c DPRINTF2(("an_txeof: sent %x/%d\n", id, cur));
cur 512 dev/ic/an.c AN_INC(cur, AN_TX_RING_CNT);
cur 513 dev/ic/an.c sc->sc_txcur = cur;
cur 515 dev/ic/an.c for (cur = 0; cur < AN_TX_RING_CNT; cur++) {
cur 516 dev/ic/an.c if (id == sc->sc_txd[cur].d_fid) {
cur 517 dev/ic/an.c sc->sc_txd[cur].d_inuse = 0;
cur 526 dev/ic/an.c id, cur);
cur 1096 dev/ic/an.c int cur, fid;
cur 1105 dev/ic/an.c cur = sc->sc_txnext;
cur 1116 dev/ic/an.c if (sc->sc_txd[cur].d_inuse) {
cur 1118 dev/ic/an.c sc->sc_txd[cur].d_fid, cur));
cur 1206 dev/ic/an.c fid = sc->sc_txd[cur].d_fid;
cur 1219 dev/ic/an.c fid, cur));
cur 1220 dev/ic/an.c sc->sc_txd[cur].d_inuse = 1;
cur 1223 dev/ic/an.c sc->sc_txd[cur].d_inuse = 0;
cur 1228 dev/ic/an.c AN_INC(cur, AN_TX_RING_CNT);
cur 1229 dev/ic/an.c sc->sc_txnext = cur;
cur 635 dev/ic/bt463.c int bt463_set_cursor (rc, cur)
cur 637 dev/ic/bt463.c struct wsdisplay_cursor *cur;
cur 640 dev/ic/bt463.c return tga_builtin_set_cursor(data->cookie, cur);
cur 643 dev/ic/bt463.c int bt463_get_cursor (rc, cur)
cur 645 dev/ic/bt463.c struct wsdisplay_cursor *cur;
cur 648 dev/ic/bt463.c return tga_builtin_get_cursor(data->cookie, cur);
cur 651 dev/ic/bt463.c int bt463_set_curpos (rc, cur)
cur 653 dev/ic/bt463.c struct wsdisplay_curpos *cur;
cur 656 dev/ic/bt463.c return tga_builtin_set_curpos(data->cookie, cur);
cur 659 dev/ic/bt463.c int bt463_get_curpos (rc, cur)
cur 661 dev/ic/bt463.c struct wsdisplay_curpos *cur;
cur 664 dev/ic/bt463.c return tga_builtin_get_curpos(data->cookie, cur);
cur 667 dev/ic/bt463.c int bt463_get_curmax (rc, cur)
cur 669 dev/ic/bt463.c struct wsdisplay_curpos *cur;
cur 672 dev/ic/bt463.c return tga_builtin_get_curmax(data->cookie, cur);
cur 2603 dev/ic/dc.c int frag, cur, cnt = 0, i;
cur 2617 dev/ic/dc.c cur = frag = *txidx;
cur 2641 dev/ic/dc.c cur = frag;
cur 2647 dev/ic/dc.c sc->dc_cdata.dc_tx_chain[cur].sd_mbuf = m_head;
cur 2648 dev/ic/dc.c sc->sc_tx_sparemap = sc->dc_cdata.dc_tx_chain[cur].sd_map;
cur 2649 dev/ic/dc.c sc->dc_cdata.dc_tx_chain[cur].sd_map = map;
cur 2650 dev/ic/dc.c sc->dc_ldata->dc_tx_list[cur].dc_ctl |= htole32(DC_TXCTL_LASTFRAG);
cur 2655 dev/ic/dc.c sc->dc_ldata->dc_tx_list[cur].dc_ctl |=
cur 2658 dev/ic/dc.c sc->dc_ldata->dc_tx_list[cur].dc_ctl |=
cur 2662 dev/ic/dc.c sc->dc_ldata->dc_tx_list[cur].dc_ctl |=
cur 1607 dev/ic/gem.c u_int32_t cur, frag, i;
cur 1610 dev/ic/gem.c cur = frag = *bixp;
cur 1611 dev/ic/gem.c map = sc->sc_txd[cur].sd_map;
cur 1636 dev/ic/gem.c cur = frag;
cur 1642 dev/ic/gem.c sc->sc_txd[*bixp].sd_map = sc->sc_txd[cur].sd_map;
cur 1643 dev/ic/gem.c sc->sc_txd[cur].sd_map = map;
cur 1644 dev/ic/gem.c sc->sc_txd[cur].sd_mbuf = mhead;
cur 1409 dev/ic/hme.c int frag, cur, cnt = 0;
cur 1413 dev/ic/hme.c cur = frag = *bixp;
cur 1443 dev/ic/hme.c cur = frag;
cur 1453 dev/ic/hme.c flags = HME_XD_GETFLAGS(sc->sc_pci, hr->rb_txd, cur);
cur 1455 dev/ic/hme.c HME_XD_SETFLAGS(sc->sc_pci, hr->rb_txd, cur, flags);
cur 1456 dev/ic/hme.c sc->sc_txd[cur].sd_mbuf = mhead;
cur 1150 dev/ic/i82596.c int off, cur, prev;
cur 1152 dev/ic/i82596.c cur = sc->xctail;
cur 1156 dev/ic/i82596.c printf("%s: xmit buffer %d\n", sc->sc_dev.dv_xname, cur);
cur 1162 dev/ic/i82596.c sc->ie_bus_write16(sc, IE_CMD_XMIT_DESC(sc->xmit_cmds, cur),
cur 1163 dev/ic/i82596.c IE_XBD_ADDR(sc->xbds, cur));
cur 1165 dev/ic/i82596.c sc->ie_bus_write16(sc, IE_CMD_XMIT_STATUS(sc->xmit_cmds, cur), 0);
cur 1171 dev/ic/i82596.c sc->ie_bus_write16(sc, IE_CMD_XMIT_LINK(sc->xmit_cmds, cur),
cur 1172 dev/ic/i82596.c IE_CMD_NOP_ADDR(sc->nop_cmds, cur));
cur 1173 dev/ic/i82596.c sc->ie_bus_write16(sc, IE_CMD_XMIT_CMD(sc->xmit_cmds, cur),
cur 1179 dev/ic/i82596.c sc->ie_bus_write16(sc, IE_CMD_NOP_STATUS(sc->nop_cmds, cur), 0);
cur 1180 dev/ic/i82596.c sc->ie_bus_write16(sc, IE_CMD_NOP_LINK(sc->nop_cmds, cur),
cur 1181 dev/ic/i82596.c IE_CMD_NOP_ADDR(sc->nop_cmds, cur));
cur 1186 dev/ic/i82596.c prev = (cur + NTXBUF - 1) % NTXBUF;
cur 1189 dev/ic/i82596.c IE_CMD_XMIT_ADDR(sc->xmit_cmds, cur));
cur 1199 dev/ic/i82596.c sc->ie_bus_write16(sc, IE_CMD_XMIT_LINK(sc->xmit_cmds,cur),
cur 1202 dev/ic/i82596.c sc->ie_bus_write16(sc, IE_CMD_XMIT_CMD(sc->xmit_cmds, cur),
cur 1206 dev/ic/i82596.c sc->ie_bus_write16(sc, off, IE_CMD_XMIT_ADDR(sc->xmit_cmds, cur));
cur 560 dev/ic/malo.c ring->cur = ring->next = 0;
cur 676 dev/ic/malo.c ring->cur = ring->next = 0;
cur 721 dev/ic/malo.c ring->cur = ring->next = ring->stat = 0;
cur 822 dev/ic/malo.c ring->cur = ring->next = ring->stat = 0;
cur 1431 dev/ic/malo.c desc = &sc->sc_txring.desc[sc->sc_txring.cur];
cur 1432 dev/ic/malo.c data = &sc->sc_txring.data[sc->sc_txring.cur];
cur 1513 dev/ic/malo.c sc->sc_txring.cur * sizeof(struct malo_tx_desc),
cur 1517 dev/ic/malo.c sc->sc_dev.dv_xname, m0->m_pkthdr.len, sc->sc_txring.cur));
cur 1520 dev/ic/malo.c sc->sc_txring.cur = (sc->sc_txring.cur + 1) % MALO_TX_RING_COUNT;
cur 1543 dev/ic/malo.c desc = &sc->sc_txring.desc[sc->sc_txring.cur];
cur 1544 dev/ic/malo.c data = &sc->sc_txring.data[sc->sc_txring.cur];
cur 1631 dev/ic/malo.c sc->sc_txring.cur * sizeof(struct malo_tx_desc),
cur 1635 dev/ic/malo.c sc->sc_dev.dv_xname, m0->m_pkthdr.len, sc->sc_txring.cur));
cur 1638 dev/ic/malo.c sc->sc_txring.cur = (sc->sc_txring.cur + 1) % MALO_TX_RING_COUNT;
cur 1674 dev/ic/malo.c desc = &sc->sc_rxring.desc[sc->sc_rxring.cur];
cur 1675 dev/ic/malo.c data = &sc->sc_rxring.data[sc->sc_rxring.cur];
cur 1678 dev/ic/malo.c sc->sc_rxring.cur * sizeof(struct malo_rx_desc),
cur 1684 dev/ic/malo.c sc->sc_rxring.cur, desc->rxctrl, desc->rssi, desc->status,
cur 1784 dev/ic/malo.c sc->sc_rxring.cur * sizeof(struct malo_rx_desc),
cur 1787 dev/ic/malo.c sc->sc_rxring.cur = (sc->sc_rxring.cur + 1) %
cur 30 dev/ic/malo.h int cur;
cur 45 dev/ic/malo.h int cur;
cur 1424 dev/ic/midway.c loc = sc->txslot[slot].cur = sc->txslot[slot].start;
cur 1472 dev/ic/midway.c sc->rxslot[slot].cur = sc->rxslot[slot].start;
cur 1988 dev/ic/midway.c u_int32_t cur = sc->txslot[chan].cur,
cur 2015 dev/ic/midway.c dma = cur;
cur 2025 dev/ic/midway.c sc->sc_dev.dv_xname, chan, l->t, cur, (cur-start)/4, need, addtail);
cur 2042 dev/ic/midway.c EN_WRITE(sc, cur, l->tbd1);
cur 2043 dev/ic/midway.c EN_WRAPADD(start, stop, cur, 4);
cur 2044 dev/ic/midway.c EN_WRITE(sc, cur, l->tbd2);
cur 2045 dev/ic/midway.c EN_WRAPADD(start, stop, cur, 4);
cur 2065 dev/ic/midway.c (len % 4) == 0 && ((unsigned long) data % 4) == 0 && (cur % 4) == 0)) {
cur 2079 dev/ic/midway.c EN_WRITEDAT(sc, cur, *data);
cur 2081 dev/ic/midway.c EN_WRAPADD(start, stop, cur, 4);
cur 2086 dev/ic/midway.c sc->sc_dev.dv_xname, chan, len, need, cur);
cur 2092 dev/ic/midway.c if (dma != cur) {
cur 2093 dev/ic/midway.c EN_DTQADD(sc, WORD_IDX(start,cur), chan, MIDDMA_JK, 0, 0, 0);
cur 2096 dev/ic/midway.c sc->sc_dev.dv_xname, chan, cur);
cur 2121 dev/ic/midway.c EN_WRAPADD(start, stop, cur, len);
cur 2124 dev/ic/midway.c sc->sc_dev.dv_xname, chan, len, need, cur);
cur 2130 dev/ic/midway.c dma = cur; /* update dma pointer */
cur 2155 dev/ic/midway.c EN_WRAPADD(start, stop, cur, cnt);
cur 2158 dev/ic/midway.c sc->sc_dev.dv_xname, chan, cnt, need, cur);
cur 2184 dev/ic/midway.c EN_WRAPADD(start, stop, cur, cnt);
cur 2187 dev/ic/midway.c sc->sc_dev.dv_xname, chan, cnt, need, cur);
cur 2203 dev/ic/midway.c EN_WRAPADD(start, stop, cur, cnt);
cur 2206 dev/ic/midway.c sc->sc_dev.dv_xname, chan, cnt, need, cur);
cur 2223 dev/ic/midway.c EN_WRAPADD(start, stop, cur, cnt);
cur 2226 dev/ic/midway.c sc->sc_dev.dv_xname, chan, cnt, need, cur);
cur 2247 dev/ic/midway.c EN_WRAPADD(start, stop, cur, len);
cur 2250 dev/ic/midway.c sc->sc_dev.dv_xname, chan, len, need, cur);
cur 2258 dev/ic/midway.c dma = cur; /* update dma pointer */
cur 2287 dev/ic/midway.c EN_WRAPADD(start, stop, cur, pad);
cur 2292 dev/ic/midway.c sc->sc_dev.dv_xname, chan, pad, need, cur);
cur 2302 dev/ic/midway.c sc->sc_dev.dv_xname, chan, pad * sizeof(u_int32_t), cur);
cur 2305 dev/ic/midway.c EN_WRITEDAT(sc, cur, 0); /* no byte order issues with zero */
cur 2306 dev/ic/midway.c EN_WRAPADD(start, stop, cur, 4);
cur 2309 dev/ic/midway.c EN_WRITE(sc, cur, l->pdu1); /* in host byte order */
cur 2310 dev/ic/midway.c EN_WRAPADD(start, stop, cur, 8);
cur 2314 dev/ic/midway.c if (addtail || dma != cur) {
cur 2316 dev/ic/midway.c EN_DTQADD(sc, WORD_IDX(start,cur), chan, MIDDMA_JK, 0,
cur 2323 dev/ic/midway.c sc->txslot[chan].cur = cur;
cur 2326 dev/ic/midway.c sc->sc_dev.dv_xname, chan, cur);
cur 2394 dev/ic/midway.c if (val > sc->txslot[lcv].cur)
cur 2395 dev/ic/midway.c sc->txslot[lcv].bfree = val - sc->txslot[lcv].cur;
cur 2397 dev/ic/midway.c sc->txslot[lcv].bfree = (val + (EN_TXSZ*1024)) - sc->txslot[lcv].cur;
cur 2628 dev/ic/midway.c u_int32_t cur, dstart, rbd, pdu, *sav, dma, bcode, count, *data, *datastop;
cur 2658 dev/ic/midway.c cur = sc->rxslot[slot].cur;
cur 2662 dev/ic/midway.c sc->sc_dev.dv_xname, slot, vci, raw, start, stop, cur);
cur 2670 dev/ic/midway.c if (dstart == cur) {
cur 2692 dev/ic/midway.c if (dstart > cur)
cur 2693 dev/ic/midway.c mlen = dstart - cur;
cur 2695 dev/ic/midway.c mlen = (dstart + (EN_RXSZ*1024)) - cur;
cur 2705 dev/ic/midway.c rbd = EN_READ(sc, cur);
cur 2718 dev/ic/midway.c pdu = cur + tlen - MID_PDU_SIZE;
cur 2753 dev/ic/midway.c if (sav[0] != cur) {
cur 2808 dev/ic/midway.c sav[0] = cur;
cur 2833 dev/ic/midway.c dma = cur; /* dma = last location we told chip about */
cur 2851 dev/ic/midway.c *data = EN_READDAT(sc, cur);
cur 2853 dev/ic/midway.c EN_WRAPADD(start, stop, cur, 4);
cur 2864 dev/ic/midway.c if (dma != cur) {
cur 2865 dev/ic/midway.c EN_DRQADD(sc, WORD_IDX(start,cur), vci, MIDDMA_JK, 0, 0, 0, 0);
cur 2868 dev/ic/midway.c sc->sc_dev.dv_xname, slot, vci, cur);
cur 2880 dev/ic/midway.c EN_WRAPADD(start, stop, cur, tlen);
cur 2889 dev/ic/midway.c dma = cur; /* update dma pointer */
cur 2915 dev/ic/midway.c EN_WRAPADD(start, stop, cur, cnt);
cur 2934 dev/ic/midway.c EN_WRAPADD(start, stop, cur, cnt);
cur 2953 dev/ic/midway.c EN_WRAPADD(start, stop, cur, tlen);
cur 2964 dev/ic/midway.c dma = cur; /* update dma pointer */
cur 2971 dev/ic/midway.c if (fill || dma != cur) {
cur 2978 dev/ic/midway.c sc->sc_dev.dv_xname, slot, vci, dma, cur);
cur 2980 dev/ic/midway.c EN_WRAPADD(start, stop, cur, fill);
cur 2981 dev/ic/midway.c EN_DRQADD(sc, WORD_IDX(start,cur), vci, MIDDMA_JK, 0, mlen,
cur 3004 dev/ic/midway.c sc->rxslot[slot].cur = cur; /* update master copy of 'cur' */
cur 3008 dev/ic/midway.c sc->sc_dev.dv_xname, slot, vci, cur);
cur 3127 dev/ic/midway.c sc->txslot[slot].start, sc->txslot[slot].stop, sc->txslot[slot].cur,
cur 3128 dev/ic/midway.c (sc->txslot[slot].cur - sc->txslot[slot].start)/4);
cur 3143 dev/ic/midway.c sc->rxslot[slot].stop, sc->rxslot[slot].cur);
cur 136 dev/ic/midwayvar.h u_int32_t cur; /* next free area (byte offset) */
cur 155 dev/ic/midwayvar.h u_int32_t cur; /* where I am at */
cur 369 dev/ic/mtd8xx.c int frag, cur, cnt = 0, i, total_len = 0;
cur 383 dev/ic/mtd8xx.c cur = frag = *txidx;
cur 401 dev/ic/mtd8xx.c cur = frag;
cur 407 dev/ic/mtd8xx.c sc->mtd_cdata.mtd_tx_chain[cur].sd_mbuf = m_head;
cur 408 dev/ic/mtd8xx.c sc->sc_tx_sparemap = sc->mtd_cdata.mtd_tx_chain[cur].sd_map;
cur 409 dev/ic/mtd8xx.c sc->mtd_cdata.mtd_tx_chain[cur].sd_map = map;
cur 410 dev/ic/mtd8xx.c sc->mtd_ldata->mtd_tx_list[cur].td_tcw |= htole32(TCW_LD | TCW_IC);
cur 412 dev/ic/mtd8xx.c sc->mtd_ldata->mtd_tx_list[cur].td_tcw |=
cur 2707 dev/ic/ncr53c9x.c struct timeval wait, cur;
cur 2718 dev/ic/ncr53c9x.c microtime(&cur);
cur 2719 dev/ic/ncr53c9x.c } while (timercmp(&cur, &wait, <=));
cur 349 dev/ic/rt2560.c ring->cur = ring->next = 0;
cur 445 dev/ic/rt2560.c ring->cur = ring->next = 0;
cur 495 dev/ic/rt2560.c ring->cur = ring->next = 0;
cur 608 dev/ic/rt2560.c ring->cur = ring->next = 0;
cur 1232 dev/ic/rt2560.c struct rt2560_rx_desc *desc = &sc->rxq.desc[sc->rxq.cur];
cur 1233 dev/ic/rt2560.c struct rt2560_rx_data *data = &sc->rxq.data[sc->rxq.cur];
cur 1236 dev/ic/rt2560.c sc->rxq.cur * RT2560_RX_DESC_SIZE, RT2560_RX_DESC_SIZE,
cur 1265 dev/ic/rt2560.c sc->rxq.cur * RT2560_RX_DESC_SIZE, RT2560_RX_DESC_SIZE,
cur 1268 dev/ic/rt2560.c DPRINTFN(15, ("rx done idx=%u\n", sc->rxq.cur));
cur 1270 dev/ic/rt2560.c sc->rxq.cur = (sc->rxq.cur + 1) % RT2560_RX_RING_COUNT;
cur 1550 dev/ic/rt2560.c desc = &sc->bcnq.desc[sc->bcnq.cur];
cur 1551 dev/ic/rt2560.c data = &sc->bcnq.data[sc->bcnq.cur];
cur 1572 dev/ic/rt2560.c sc->bcnq.cur * RT2560_TX_DESC_SIZE, RT2560_TX_DESC_SIZE,
cur 1610 dev/ic/rt2560.c desc = &sc->prioq.desc[sc->prioq.cur];
cur 1611 dev/ic/rt2560.c data = &sc->prioq.data[sc->prioq.cur];
cur 1679 dev/ic/rt2560.c sc->prioq.cur * RT2560_TX_DESC_SIZE, RT2560_TX_DESC_SIZE,
cur 1683 dev/ic/rt2560.c m0->m_pkthdr.len, sc->prioq.cur, rate));
cur 1687 dev/ic/rt2560.c sc->prioq.cur = (sc->prioq.cur + 1) % RT2560_PRIO_RING_COUNT;
cur 68 dev/ic/rt2560var.h int cur;
cur 87 dev/ic/rt2560var.h int cur;
cur 383 dev/ic/rt2661.c ring->cur = ring->next = ring->stat = 0;
cur 478 dev/ic/rt2661.c ring->cur = ring->next = ring->stat = 0;
cur 526 dev/ic/rt2661.c ring->cur = ring->next = 0;
cur 636 dev/ic/rt2661.c ring->cur = ring->next = 0;
cur 1008 dev/ic/rt2661.c struct rt2661_rx_desc *desc = &sc->rxq.desc[sc->rxq.cur];
cur 1009 dev/ic/rt2661.c struct rt2661_rx_data *data = &sc->rxq.data[sc->rxq.cur];
cur 1012 dev/ic/rt2661.c sc->rxq.cur * RT2661_RX_DESC_SIZE, RT2661_RX_DESC_SIZE,
cur 1137 dev/ic/rt2661.c sc->rxq.cur * RT2661_RX_DESC_SIZE, RT2661_RX_DESC_SIZE,
cur 1140 dev/ic/rt2661.c DPRINTFN(15, ("rx intr idx=%u\n", sc->rxq.cur));
cur 1142 dev/ic/rt2661.c sc->rxq.cur = (sc->rxq.cur + 1) % RT2661_RX_RING_COUNT;
cur 1457 dev/ic/rt2661.c desc = &sc->mgtq.desc[sc->mgtq.cur];
cur 1458 dev/ic/rt2661.c data = &sc->mgtq.data[sc->mgtq.cur];
cur 1527 dev/ic/rt2661.c sc->mgtq.cur * RT2661_TX_DESC_SIZE, RT2661_TX_DESC_SIZE,
cur 1531 dev/ic/rt2661.c m0->m_pkthdr.len, sc->mgtq.cur, rate));
cur 1535 dev/ic/rt2661.c sc->mgtq.cur = (sc->mgtq.cur + 1) % RT2661_MGT_RING_COUNT;
cur 1628 dev/ic/rt2661.c desc = &txq->desc[txq->cur];
cur 1629 dev/ic/rt2661.c data = &txq->data[txq->cur];
cur 1655 dev/ic/rt2661.c txq->cur * RT2661_TX_DESC_SIZE, RT2661_TX_DESC_SIZE,
cur 1659 dev/ic/rt2661.c txq->cur = (txq->cur + 1) % RT2661_TX_RING_COUNT;
cur 1664 dev/ic/rt2661.c data = &txq->data[txq->cur];
cur 1665 dev/ic/rt2661.c desc = &txq->desc[txq->cur];
cur 1747 dev/ic/rt2661.c bus_dmamap_sync(sc->sc_dmat, txq->map, txq->cur * RT2661_TX_DESC_SIZE,
cur 1751 dev/ic/rt2661.c m0->m_pkthdr.len, txq->cur, rate));
cur 1755 dev/ic/rt2661.c txq->cur = (txq->cur + 1) % RT2661_TX_RING_COUNT;
cur 64 dev/ic/rt2661var.h int cur;
cur 81 dev/ic/rt2661var.h int cur;
cur 1193 dev/pci/azalia.c this->dacs.cur = -1;
cur 1194 dev/pci/azalia.c this->adcs.cur = -1;
cur 1227 dev/pci/azalia.c prev_dac = this->dacs.cur;
cur 1228 dev/pci/azalia.c this->dacs.cur = newdac;
cur 1229 dev/pci/azalia.c group = &this->dacs.groups[this->dacs.cur];
cur 1249 dev/pci/azalia.c prev_adc = this->adcs.cur;
cur 1250 dev/pci/azalia.c this->adcs.cur = newadc;
cur 1251 dev/pci/azalia.c group = &this->adcs.groups[this->adcs.cur];
cur 1285 dev/pci/azalia.c group = &this->dacs.groups[this->dacs.cur];
cur 1296 dev/pci/azalia.c group = &this->adcs.groups[this->adcs.cur];
cur 1411 dev/pci/azalia.c group = &this->adcs.groups[this->adcs.cur];
cur 1413 dev/pci/azalia.c group = &this->dacs.groups[this->dacs.cur];
cur 524 dev/pci/azalia.h int cur;
cur 252 dev/pci/azalia_codec.c this->dacs.cur = 0;
cur 265 dev/pci/azalia_codec.c this->adcs.cur = 0;
cur 1009 dev/pci/azalia_codec.c mc->un.ord = this->dacs.cur;
cur 1014 dev/pci/azalia_codec.c mc->un.ord = this->adcs.cur;
cur 1255 dev/pci/azalia_codec.c mc->un.ord, this->adcs.cur);
cur 1265 dev/pci/azalia_codec.c this->dacs.cur, mc->un.ord);
cur 2732 dev/pci/if_bge.c u_int32_t frag, cur;
cur 2745 dev/pci/if_bge.c cur = frag = *txidx;
cur 2810 dev/pci/if_bge.c cur = frag;
cur 2823 dev/pci/if_bge.c sc->bge_rdata->bge_tx_ring[cur].bge_flags |= BGE_TXBDFLAG_END;
cur 2824 dev/pci/if_bge.c sc->bge_cdata.bge_tx_chain[cur] = m_head;
cur 2826 dev/pci/if_bge.c sc->txdma[cur] = dma;
cur 1842 dev/pci/if_cas.c u_int32_t cur, frag, i;
cur 1845 dev/pci/if_cas.c cur = frag = *bixp;
cur 1846 dev/pci/if_cas.c map = sc->sc_txd[cur].sd_map;
cur 1871 dev/pci/if_cas.c cur = frag;
cur 1877 dev/pci/if_cas.c sc->sc_txd[*bixp].sd_map = sc->sc_txd[cur].sd_map;
cur 1878 dev/pci/if_cas.c sc->sc_txd[cur].sd_map = map;
cur 1879 dev/pci/if_cas.c sc->sc_txd[cur].sd_mbuf = mhead;
cur 395 dev/pci/if_iwi.c ring->cur = ring->next = 0;
cur 445 dev/pci/if_iwi.c ring->cur = ring->next = 0;
cur 470 dev/pci/if_iwi.c ring->cur = ring->next = 0;
cur 548 dev/pci/if_iwi.c ring->cur = ring->next = 0;
cur 584 dev/pci/if_iwi.c ring->cur = 0;
cur 635 dev/pci/if_iwi.c ring->cur = 0;
cur 1101 dev/pci/if_iwi.c for (; sc->rxq.cur != hw;) {
cur 1102 dev/pci/if_iwi.c data = &sc->rxq.data[sc->rxq.cur];
cur 1125 dev/pci/if_iwi.c sc->rxq.cur = (sc->rxq.cur + 1) % IWI_RX_RING_COUNT;
cur 1231 dev/pci/if_iwi.c desc = &sc->cmdq.desc[sc->cmdq.cur];
cur 1239 dev/pci/if_iwi.c sc->cmdq.cur * sizeof (struct iwi_cmd_desc),
cur 1242 dev/pci/if_iwi.c DPRINTFN(2, ("sending command idx=%u type=%u len=%u\n", sc->cmdq.cur,
cur 1245 dev/pci/if_iwi.c sc->cmdq.cur = (sc->cmdq.cur + 1) % IWI_CMD_RING_COUNT;
cur 1249 dev/pci/if_iwi.c sc->cmdq.next = sc->cmdq.cur;
cur 1286 dev/pci/if_iwi.c data = &txq->data[txq->cur];
cur 1287 dev/pci/if_iwi.c desc = &txq->desc[txq->cur];
cur 1383 dev/pci/if_iwi.c txq->cur * sizeof (struct iwi_tx_desc),
cur 1386 dev/pci/if_iwi.c DPRINTFN(5, ("sending data frame idx=%u len=%u nseg=%u\n", txq->cur,
cur 1390 dev/pci/if_iwi.c txq->cur = (txq->cur + 1) % IWI_TX_RING_COUNT;
cur 1391 dev/pci/if_iwi.c CSR_WRITE_4(sc, txq->csr_widx, txq->cur);
cur 2202 dev/pci/if_iwi.c CSR_WRITE_4(sc, IWI_CSR_CMD_WIDX, sc->cmdq.cur);
cur 2206 dev/pci/if_iwi.c CSR_WRITE_4(sc, IWI_CSR_TX1_WIDX, sc->txq[0].cur);
cur 2210 dev/pci/if_iwi.c CSR_WRITE_4(sc, IWI_CSR_TX2_WIDX, sc->txq[1].cur);
cur 2214 dev/pci/if_iwi.c CSR_WRITE_4(sc, IWI_CSR_TX3_WIDX, sc->txq[2].cur);
cur 2218 dev/pci/if_iwi.c CSR_WRITE_4(sc, IWI_CSR_TX4_WIDX, sc->txq[3].cur);
cur 64 dev/pci/if_iwivar.h int cur;
cur 82 dev/pci/if_iwivar.h int cur;
cur 94 dev/pci/if_iwivar.h int cur;
cur 1407 dev/pci/if_msk.c u_int32_t frag, cur;
cur 1421 dev/pci/if_msk.c cur = frag = *txidx;
cur 1460 dev/pci/if_msk.c cur = frag;
cur 1464 dev/pci/if_msk.c sc_if->sk_cdata.sk_tx_chain[cur].sk_mbuf = m_head;
cur 1467 dev/pci/if_msk.c sc_if->sk_cdata.sk_tx_map[cur] = entry;
cur 1468 dev/pci/if_msk.c sc_if->sk_rdata->sk_tx_ring[cur].sk_ctl |= SK_Y2_TXCTL_LASTFRAG;
cur 1610 dev/pci/if_msk.c int cur, total_len = len;
cur 1615 dev/pci/if_msk.c cur = sc_if->sk_cdata.sk_rx_cons;
cur 1620 dev/pci/if_msk.c MSK_CDRXSYNC(sc_if, cur, BUS_DMASYNC_POSTREAD|BUS_DMASYNC_POSTWRITE);
cur 1622 dev/pci/if_msk.c cur_rx = &sc_if->sk_cdata.sk_rx_chain[cur];
cur 1635 dev/pci/if_msk.c msk_newbuf(sc_if, cur, m, dmamap);
cur 1645 dev/pci/if_msk.c if (msk_newbuf(sc_if, cur, NULL, dmamap) == ENOBUFS) {
cur 1649 dev/pci/if_msk.c msk_newbuf(sc_if, cur, m, dmamap);
cur 646 dev/pci/if_nfe.c data = &sc->rxq.data[sc->rxq.cur];
cur 649 dev/pci/if_nfe.c desc64 = &sc->rxq.desc64[sc->rxq.cur];
cur 655 dev/pci/if_nfe.c desc32 = &sc->rxq.desc32[sc->rxq.cur];
cur 796 dev/pci/if_nfe.c sc->rxq.cur = (sc->rxq.cur + 1) % NFE_RX_RING_COUNT;
cur 809 dev/pci/if_nfe.c while (sc->txq.next != sc->txq.cur) {
cur 885 dev/pci/if_nfe.c int error, i, first = sc->txq.cur;
cur 915 dev/pci/if_nfe.c data = &sc->txq.data[sc->txq.cur];
cur 918 dev/pci/if_nfe.c desc64 = &sc->txq.desc64[sc->txq.cur];
cur 931 dev/pci/if_nfe.c desc32 = &sc->txq.desc32[sc->txq.cur];
cur 955 dev/pci/if_nfe.c sc->txq.cur = (sc->txq.cur + 1) % NFE_TX_RING_COUNT;
cur 991 dev/pci/if_nfe.c int old = sc->txq.cur;
cur 1012 dev/pci/if_nfe.c if (sc->txq.cur == old) /* nothing sent */
cur 1016 dev/pci/if_nfe.c nfe_txdesc64_rsync(sc, old, sc->txq.cur, BUS_DMASYNC_PREWRITE);
cur 1018 dev/pci/if_nfe.c nfe_txdesc32_rsync(sc, old, sc->txq.cur, BUS_DMASYNC_PREWRITE);
cur 1197 dev/pci/if_nfe.c ring->cur = ring->next = 0;
cur 1338 dev/pci/if_nfe.c ring->cur = ring->next = 0;
cur 1516 dev/pci/if_nfe.c ring->cur = ring->next = 0;
cur 1598 dev/pci/if_nfe.c ring->cur = ring->next = 0;
cur 35 dev/pci/if_nfevar.h int cur;
cur 63 dev/pci/if_nfevar.h int cur;
cur 1601 dev/pci/if_nge.c int frag, cur, cnt = 0;
cur 1616 dev/pci/if_nge.c cur = frag = *txidx;
cur 1630 dev/pci/if_nge.c cur = frag;
cur 1643 dev/pci/if_nge.c sc->nge_ldata->nge_tx_list[cur].nge_extsts |=
cur 1648 dev/pci/if_nge.c sc->nge_ldata->nge_tx_list[cur].nge_mbuf = m_head;
cur 1649 dev/pci/if_nge.c sc->nge_ldata->nge_tx_list[cur].nge_ctl &= ~NGE_CMDSTS_MORE;
cur 1546 dev/pci/if_sis.c int frag, cur, i;
cur 1559 dev/pci/if_sis.c cur = frag = *txidx;
cur 1569 dev/pci/if_sis.c cur = frag;
cur 1576 dev/pci/if_sis.c sc->sis_ldata->sis_tx_list[cur].sis_mbuf = m_head;
cur 1577 dev/pci/if_sis.c sc->sis_ldata->sis_tx_list[cur].sis_ctl &= ~SIS_CMDSTS_MORE;
cur 1521 dev/pci/if_sk.c u_int32_t frag, cur, sk_ctl;
cur 1535 dev/pci/if_sk.c cur = frag = *txidx;
cur 1574 dev/pci/if_sk.c cur = frag;
cur 1578 dev/pci/if_sk.c sc_if->sk_cdata.sk_tx_chain[cur].sk_mbuf = m_head;
cur 1581 dev/pci/if_sk.c sc_if->sk_cdata.sk_tx_map[cur] = entry;
cur 1582 dev/pci/if_sk.c sc_if->sk_rdata->sk_tx_ring[cur].sk_ctl |=
cur 1731 dev/pci/if_sk.c int i, cur, total_len = 0;
cur 1741 dev/pci/if_sk.c cur = i;
cur 1744 dev/pci/if_sk.c SK_CDRXSYNC(sc_if, cur,
cur 1750 dev/pci/if_sk.c SK_CDRXSYNC(sc_if, cur, BUS_DMASYNC_PREREAD);
cur 1755 dev/pci/if_sk.c cur_rx = &sc_if->sk_cdata.sk_rx_chain[cur];
cur 1756 dev/pci/if_sk.c cur_desc = &sc_if->sk_rdata->sk_rx_ring[cur];
cur 1779 dev/pci/if_sk.c sk_newbuf(sc_if, cur, m, dmamap);
cur 1790 dev/pci/if_sk.c if (sk_newbuf(sc_if, cur, NULL, dmamap) == ENOBUFS) {
cur 1794 dev/pci/if_sk.c sk_newbuf(sc_if, cur, m, dmamap);
cur 1997 dev/pci/if_ti.c u_int32_t frag, cur, cnt = 0;
cur 2015 dev/pci/if_ti.c cur = frag = *txidx;
cur 2052 dev/pci/if_ti.c cur = frag;
cur 2061 dev/pci/if_ti.c ti_mem_write(sc, TI_TX_RING_BASE + cur * sizeof(txdesc),
cur 2067 dev/pci/if_ti.c sc->ti_cdata.ti_tx_chain[cur] = m_head;
cur 2069 dev/pci/if_ti.c sc->ti_cdata.ti_tx_map[cur] = entry;
cur 2085 dev/pci/if_ti.c u_int32_t frag, cur, cnt = 0;
cur 2102 dev/pci/if_ti.c cur = frag = *txidx;
cur 2136 dev/pci/if_ti.c cur = frag;
cur 2144 dev/pci/if_ti.c sc->ti_rdata->ti_tx_ring[cur].ti_flags |= TI_BDFLAG_END;
cur 2149 dev/pci/if_ti.c TI_RING_DMASYNC(sc, ti_tx_ring[cur], BUS_DMASYNC_POSTREAD);
cur 2151 dev/pci/if_ti.c sc->ti_cdata.ti_tx_chain[cur] = m_head;
cur 2153 dev/pci/if_ti.c sc->ti_cdata.ti_tx_map[cur] = entry;
cur 537 dev/pci/if_wpi.c ring->cur = 0;
cur 602 dev/pci/if_wpi.c ring->cur = 0;
cur 627 dev/pci/if_wpi.c ring->cur = 0;
cur 711 dev/pci/if_wpi.c ring->cur = 0;
cur 1222 dev/pci/if_wpi.c "chan=%d tstamp=%llu\n", ring->cur, letoh32(desc->len),
cur 1254 dev/pci/if_wpi.c ring->desc[ring->cur] = htole32(rbuf->paddr);
cur 1390 dev/pci/if_wpi.c while (sc->rxq.cur != hw) {
cur 1391 dev/pci/if_wpi.c struct wpi_rx_data *data = &sc->rxq.data[sc->rxq.cur];
cur 1479 dev/pci/if_wpi.c sc->rxq.cur = (sc->rxq.cur + 1) % WPI_RX_RING_COUNT;
cur 1569 dev/pci/if_wpi.c desc = &ring->desc[ring->cur];
cur 1570 dev/pci/if_wpi.c data = &ring->data[ring->cur];
cur 1610 dev/pci/if_wpi.c cmd = &ring->cmd[ring->cur];
cur 1614 dev/pci/if_wpi.c cmd->idx = ring->cur;
cur 1738 dev/pci/if_wpi.c ring->qid, ring->cur, m0->m_pkthdr.len, data->map->dm_nsegs));
cur 1744 dev/pci/if_wpi.c ring->cur * sizeof (struct wpi_tx_cmd));
cur 1756 dev/pci/if_wpi.c ring->cur = (ring->cur + 1) % WPI_TX_RING_COUNT;
cur 1757 dev/pci/if_wpi.c WPI_WRITE(sc, WPI_TX_WIDX, ring->qid << 8 | ring->cur);
cur 2033 dev/pci/if_wpi.c desc = &ring->desc[ring->cur];
cur 2034 dev/pci/if_wpi.c cmd = &ring->cmd[ring->cur];
cur 2039 dev/pci/if_wpi.c cmd->idx = ring->cur;
cur 2044 dev/pci/if_wpi.c ring->cur * sizeof (struct wpi_tx_cmd));
cur 2048 dev/pci/if_wpi.c ring->cur = (ring->cur + 1) % WPI_CMD_RING_COUNT;
cur 2049 dev/pci/if_wpi.c WPI_WRITE(sc, WPI_TX_WIDX, ring->qid << 8 | ring->cur);
cur 2285 dev/pci/if_wpi.c desc = &ring->desc[ring->cur];
cur 2286 dev/pci/if_wpi.c data = &ring->data[ring->cur];
cur 2295 dev/pci/if_wpi.c cmd = &ring->cmd[ring->cur];
cur 2299 dev/pci/if_wpi.c cmd->idx = ring->cur;
cur 2330 dev/pci/if_wpi.c ring->cur * sizeof (struct wpi_tx_cmd));
cur 2336 dev/pci/if_wpi.c ring->cur = (ring->cur + 1) % WPI_CMD_RING_COUNT;
cur 2337 dev/pci/if_wpi.c WPI_WRITE(sc, WPI_TX_WIDX, ring->qid << 8 | ring->cur);
cur 2428 dev/pci/if_wpi.c desc = &ring->desc[ring->cur];
cur 2429 dev/pci/if_wpi.c data = &ring->data[ring->cur];
cur 2450 dev/pci/if_wpi.c cmd->idx = ring->cur;
cur 2560 dev/pci/if_wpi.c ring->cur = (ring->cur + 1) % WPI_CMD_RING_COUNT;
cur 2561 dev/pci/if_wpi.c WPI_WRITE(sc, WPI_TX_WIDX, ring->qid << 8 | ring->cur);
cur 80 dev/pci/if_wpivar.h int cur;
cur 105 dev/pci/if_wpivar.h int cur;
cur 989 dev/pci/musycc.c struct dma_desc *cur, *tmp;
cur 1005 dev/pci/musycc.c cur = mg->mg_dma_d[c].tx_cur;
cur 1018 dev/pci/musycc.c if (cur != mg->mg_dma_d[c].tx_cur)
cur 1021 dev/pci/musycc.c cur->status = htole32(status);
cur 1022 dev/pci/musycc.c cur->data = htole32(map->dm_segs[i].ds_addr);
cur 1025 dev/pci/musycc.c ((caddr_t)cur - mg->mg_listkva), sizeof(struct dma_desc),
cur 1030 dev/pci/musycc.c cur = cur->nextdesc;
cur 1036 dev/pci/musycc.c cur->mbuf = m_head;
cur 1037 dev/pci/musycc.c mg->mg_tx_sparemap = cur->map;
cur 1038 dev/pci/musycc.c cur->map = map;
cur 1039 dev/pci/musycc.c cur->status |= htole32(MUSYCC_STATUS_EOM);
cur 1041 dev/pci/musycc.c mg->mg_dma_d[c].tx_cur = cur->nextdesc;
cur 2231 dev/usb/ehci.c ehci_soft_qtd_t *next, *cur;
cur 2259 dev/usb/ehci.c cur = ehci_alloc_sqtd(sc);
cur 2260 dev/usb/ehci.c *sp = cur;
cur 2261 dev/usb/ehci.c if (cur == NULL)
cur 2318 dev/usb/ehci.c cur->qtd.qtd_buffer[i] = htole32(a);
cur 2319 dev/usb/ehci.c cur->qtd.qtd_buffer_hi[i] = 0;
cur 2327 dev/usb/ehci.c cur->nextqtd = next;
cur 2328 dev/usb/ehci.c cur->qtd.qtd_next = cur->qtd.qtd_altnext = nextphys;
cur 2329 dev/usb/ehci.c cur->qtd.qtd_status = htole32(qtdstatus |
cur 2331 dev/usb/ehci.c cur->xfer = xfer;
cur 2332 dev/usb/ehci.c cur->len = curlen;
cur 2351 dev/usb/ehci.c cur = next;
cur 2353 dev/usb/ehci.c cur->qtd.qtd_status |= htole32(EHCI_QTD_IOC);
cur 2354 dev/usb/ehci.c *ep = cur;
cur 2423 dev/usb/ehci.c ehci_physaddr_t cur, us, next;
cur 2525 dev/usb/ehci.c cur = EHCI_LINK_ADDR(letoh32(sqh->qh.qh_curqtd));
cur 2542 dev/usb/ehci.c hit |= (cur == sqtd->physaddr);
cur 2567 dev/usb/ehci.c if (cur == sqtd->physaddr) {
cur 67 dev/usb/hid.c struct hid_item cur;
cur 113 dev/usb/hid.c while (s->cur.next != NULL) {
cur 114 dev/usb/hid.c struct hid_item *hi = s->cur.next->next;
cur 115 dev/usb/hid.c free(s->cur.next, M_TEMP);
cur 116 dev/usb/hid.c s->cur.next = hi;
cur 124 dev/usb/hid.c struct hid_item *c = &s->cur;
cur 302 dev/usb/hid.c *hi = s->cur;
cur 308 dev/usb/hid.c s->cur = *hi;
cur 448 dev/usb/ohci.c ohci_soft_td_t *next, *cur;
cur 458 dev/usb/ohci.c cur = sp;
cur 493 dev/usb/ohci.c cur->td.td_flags = tdflags;
cur 494 dev/usb/ohci.c cur->td.td_cbp = htole32(dataphys);
cur 495 dev/usb/ohci.c cur->nexttd = next;
cur 496 dev/usb/ohci.c cur->td.td_nexttd = htole32(next->physaddr);
cur 497 dev/usb/ohci.c cur->td.td_be = htole32(dataphys + curlen - 1);
cur 498 dev/usb/ohci.c cur->len = curlen;
cur 499 dev/usb/ohci.c cur->flags = OHCI_ADD_LEN;
cur 500 dev/usb/ohci.c cur->xfer = xfer;
cur 507 dev/usb/ohci.c cur = next;
cur 513 dev/usb/ohci.c cur = next;
cur 518 dev/usb/ohci.c cur->td.td_flags = tdflags;
cur 519 dev/usb/ohci.c cur->td.td_cbp = 0; /* indicate 0 length packet */
cur 520 dev/usb/ohci.c cur->nexttd = next;
cur 521 dev/usb/ohci.c cur->td.td_nexttd = htole32(next->physaddr);
cur 522 dev/usb/ohci.c cur->td.td_be = ~0;
cur 523 dev/usb/ohci.c cur->len = 0;
cur 524 dev/usb/ohci.c cur->flags = 0;
cur 525 dev/usb/ohci.c cur->xfer = xfer;
cur 528 dev/usb/ohci.c *ep = cur;
cur 139 dev/usb/uaudio.c u_char *cur; /* current position in upper layer buffer */
cur 2663 dev/usb/uaudio.c n = min(total, ch->end - ch->cur);
cur 2664 dev/usb/uaudio.c memcpy(cb->buffer, ch->cur, n);
cur 2665 dev/usb/uaudio.c ch->cur += n;
cur 2666 dev/usb/uaudio.c if (ch->cur >= ch->end)
cur 2667 dev/usb/uaudio.c ch->cur = ch->start;
cur 2670 dev/usb/uaudio.c memcpy(cb->buffer + n, ch->cur, total);
cur 2671 dev/usb/uaudio.c ch->cur += total;
cur 2808 dev/usb/uaudio.c n = min(frsize, ch->end - ch->cur);
cur 2809 dev/usb/uaudio.c memcpy(ch->cur, cb->buffer + cb->offsets[i], n);
cur 2810 dev/usb/uaudio.c ch->cur += n;
cur 2811 dev/usb/uaudio.c if (ch->cur >= ch->end)
cur 2812 dev/usb/uaudio.c ch->cur = ch->start;
cur 2814 dev/usb/uaudio.c memcpy(ch->cur, cb->buffer + cb->offsets[i] + n,
cur 2816 dev/usb/uaudio.c ch->cur += frsize - n;
cur 2861 dev/usb/uaudio.c ch->cur = start;
cur 92 dev/usb/ugen.c u_char *cur; /* current read location (isoc) */
cur 375 dev/usb/ugen.c sce->cur = sce->fill = sce->ibuf;
cur 590 dev/usb/ugen.c while (sce->cur == sce->fill) {
cur 607 dev/usb/ugen.c while (sce->cur != sce->fill && uio->uio_resid > 0 && !error) {
cur 608 dev/usb/ugen.c if(sce->fill > sce->cur)
cur 609 dev/usb/ugen.c n = min(sce->fill - sce->cur, uio->uio_resid);
cur 611 dev/usb/ugen.c n = min(sce->limit - sce->cur, uio->uio_resid);
cur 616 dev/usb/ugen.c error = uiomove(sce->cur, n, uio);
cur 619 dev/usb/ugen.c sce->cur += n;
cur 620 dev/usb/ugen.c if(sce->cur >= sce->limit)
cur 621 dev/usb/ugen.c sce->cur = sce->ibuf;
cur 864 dev/usb/ugen.c if(sce->fill < sce->cur && sce->cur <= sce->fill + count) {
cur 865 dev/usb/ugen.c sce->cur += count;
cur 866 dev/usb/ugen.c if(sce->cur >= sce->limit)
cur 867 dev/usb/ugen.c sce->cur = sce->ibuf + (sce->limit - sce->cur);
cur 1333 dev/usb/ugen.c if (sce->cur != sce->fill)
cur 1385 dev/usb/ugen.c if (sce->cur == sce->fill)
cur 1388 dev/usb/ugen.c if (sce->cur < sce->fill)
cur 1389 dev/usb/ugen.c kn->kn_data = sce->fill - sce->cur;
cur 1391 dev/usb/ugen.c kn->kn_data = (sce->limit - sce->cur) +
cur 1112 dev/usb/usbdi.c iter->cur = (const uByte *)cd;
cur 1121 dev/usb/usbdi.c if (iter->cur + sizeof(usb_descriptor_t) >= iter->end) {
cur 1122 dev/usb/usbdi.c if (iter->cur != iter->end)
cur 1126 dev/usb/usbdi.c desc = (const usb_descriptor_t *)iter->cur;
cur 1131 dev/usb/usbdi.c iter->cur += desc->bLength;
cur 1132 dev/usb/usbdi.c if (iter->cur > iter->end) {
cur 177 dev/usb/usbdi.h const uByte *cur;
cur 339 dev/wscons/wskbdutil.c kbd_t cur;
cur 351 dev/wscons/wskbdutil.c for (cur = mapdata->layout & ~KB_HANDLEDBYWSKBD; cur != 0; ) {
cur 354 dev/wscons/wskbdutil.c if (mp->name == cur)
cur 387 dev/wscons/wskbdutil.c cur = mp->base;
cur 425 dev/wscons/wskbdutil.c kbd_t cur;
cur 428 dev/wscons/wskbdutil.c for (cur = mapdata->layout & ~KB_HANDLEDBYWSKBD, stack_ptr = 0;
cur 429 dev/wscons/wskbdutil.c cur != 0; stack_ptr++) {
cur 432 dev/wscons/wskbdutil.c if (cur == 0 || mp->name == cur) {
cur 445 dev/wscons/wskbdutil.c cur = mp->base;
cur 810 net/pf.c struct pf_state_key *cur;
cur 816 net/pf.c if ((cur = RB_INSERT(pf_state_tree_lan_ext, &pf_statetbl_lan_ext,
cur 819 net/pf.c TAILQ_FOREACH(sp, &cur->states, next)
cur 825 net/pf.c pf_attach_state(cur, s, kif == pfi_all ? 1 : 0);
cur 829 net/pf.c if (cur == NULL && (cur = RB_INSERT(pf_state_tree_ext_gwy,
cur 928 net/pf.c struct pf_src_node *cur, *next;
cur 931 net/pf.c for (cur = RB_MIN(pf_src_tree, &tree_src_tracking); cur; cur = next) {
cur 932 net/pf.c next = RB_NEXT(pf_src_tree, &tree_src_tracking, cur);
cur 934 net/pf.c if (cur->states <= 0 && cur->expire <= time_second) {
cur 938 net/pf.c &tree_src_tracking, cur);
cur 941 net/pf.c if (cur->rule.ptr != NULL) {
cur 942 net/pf.c cur->rule.ptr->src_nodes--;
cur 943 net/pf.c if (cur->rule.ptr->states <= 0 &&
cur 944 net/pf.c cur->rule.ptr->max_src_nodes <= 0)
cur 945 net/pf.c pf_rm_rule(NULL, cur->rule.ptr);
cur 947 net/pf.c RB_REMOVE(pf_src_tree, &tree_src_tracking, cur);
cur 950 net/pf.c pool_put(&pf_src_tree_pl, cur);
cur 990 net/pf.c pf_unlink_state(struct pf_state *cur)
cur 992 net/pf.c if (cur->src.state == PF_TCPS_PROXY_DST) {
cur 993 net/pf.c pf_send_tcp(cur->rule.ptr, cur->state_key->af,
cur 994 net/pf.c &cur->state_key->ext.addr, &cur->state_key->lan.addr,
cur 995 net/pf.c cur->state_key->ext.port, cur->state_key->lan.port,
cur 996 net/pf.c cur->src.seqhi, cur->src.seqlo + 1,
cur 997 net/pf.c TH_RST|TH_ACK, 0, 0, 0, 1, cur->tag, NULL, NULL);
cur 999 net/pf.c RB_REMOVE(pf_state_tree_id, &tree_id, cur);
cur 1001 net/pf.c if (cur->creatorid == pf_status.hostid)
cur 1002 net/pf.c pfsync_delete_state(cur);
cur 1004 net/pf.c cur->timeout = PFTM_UNLINKED;
cur 1005 net/pf.c pf_src_tree_remove_state(cur);
cur 1006 net/pf.c pf_detach_state(cur, 0);
cur 1012 net/pf.c pf_free_state(struct pf_state *cur)
cur 1016 net/pf.c (pfsyncif->sc_bulk_send_next == cur ||
cur 1017 net/pf.c pfsyncif->sc_bulk_terminator == cur))
cur 1020 net/pf.c KASSERT(cur->timeout == PFTM_UNLINKED);
cur 1021 net/pf.c if (--cur->rule.ptr->states <= 0 &&
cur 1022 net/pf.c cur->rule.ptr->src_nodes <= 0)
cur 1023 net/pf.c pf_rm_rule(NULL, cur->rule.ptr);
cur 1024 net/pf.c if (cur->nat_rule.ptr != NULL)
cur 1025 net/pf.c if (--cur->nat_rule.ptr->states <= 0 &&
cur 1026 net/pf.c cur->nat_rule.ptr->src_nodes <= 0)
cur 1027 net/pf.c pf_rm_rule(NULL, cur->nat_rule.ptr);
cur 1028 net/pf.c if (cur->anchor.ptr != NULL)
cur 1029 net/pf.c if (--cur->anchor.ptr->states <= 0)
cur 1030 net/pf.c pf_rm_rule(NULL, cur->anchor.ptr);
cur 1031 net/pf.c pf_normalize_tcp_cleanup(cur);
cur 1032 net/pf.c pfi_kif_unref(cur->kif, PFI_KIF_REF_STATE);
cur 1033 net/pf.c TAILQ_REMOVE(&state_list, cur, entry_list);
cur 1034 net/pf.c if (cur->tag)
cur 1035 net/pf.c pf_tag_unref(cur->tag);
cur 1036 net/pf.c pool_put(&pf_state_pl, cur);
cur 1044 net/pf.c static struct pf_state *cur = NULL;
cur 1050 net/pf.c if (cur == NULL) {
cur 1051 net/pf.c cur = TAILQ_FIRST(&state_list);
cur 1052 net/pf.c if (cur == NULL)
cur 1057 net/pf.c next = TAILQ_NEXT(cur, entry_list);
cur 1059 net/pf.c if (cur->timeout == PFTM_UNLINKED) {
cur 1065 net/pf.c pf_free_state(cur);
cur 1066 net/pf.c } else if (pf_state_expires(cur) <= time_second) {
cur 1068 net/pf.c pf_unlink_state(cur);
cur 1073 net/pf.c pf_free_state(cur);
cur 1075 net/pf.c cur = next;
cur 1243 net/pf.c while (head[i] != cur) { \
cur 1244 net/pf.c head[i]->skip[i].ptr = cur; \
cur 1252 net/pf.c struct pf_rule *cur, *prev, *head[PF_SKIP_COUNT];
cur 1255 net/pf.c cur = TAILQ_FIRST(rules);
cur 1256 net/pf.c prev = cur;
cur 1258 net/pf.c head[i] = cur;
cur 1259 net/pf.c while (cur != NULL) {
cur 1261 net/pf.c if (cur->kif != prev->kif || cur->ifnot != prev->ifnot)
cur 1263 net/pf.c if (cur->direction != prev->direction)
cur 1265 net/pf.c if (cur->af != prev->af)
cur 1267 net/pf.c if (cur->proto != prev->proto)
cur 1269 net/pf.c if (cur->src.neg != prev->src.neg ||
cur 1270 net/pf.c pf_addr_wrap_neq(&cur->src.addr, &prev->src.addr))
cur 1272 net/pf.c if (cur->src.port[0] != prev->src.port[0] ||
cur 1273 net/pf.c cur->src.port[1] != prev->src.port[1] ||
cur 1274 net/pf.c cur->src.port_op != prev->src.port_op)
cur 1276 net/pf.c if (cur->dst.neg != prev->dst.neg ||
cur 1277 net/pf.c pf_addr_wrap_neq(&cur->dst.addr, &prev->dst.addr))
cur 1279 net/pf.c if (cur->dst.port[0] != prev->dst.port[0] ||
cur 1280 net/pf.c cur->dst.port[1] != prev->dst.port[1] ||
cur 1281 net/pf.c cur->dst.port_op != prev->dst.port_op)
cur 1284 net/pf.c prev = cur;
cur 1285 net/pf.c cur = TAILQ_NEXT(cur, entries);
cur 2062 net/pf.c struct pf_addr *raddr = &rpool->cur->addr.v.a.addr;
cur 2063 net/pf.c struct pf_addr *rmask = &rpool->cur->addr.v.a.mask;
cur 2064 net/pf.c struct pf_pooladdr *acur = rpool->cur;
cur 2091 net/pf.c if (rpool->cur->addr.type == PF_ADDR_NOROUTE)
cur 2093 net/pf.c if (rpool->cur->addr.type == PF_ADDR_DYNIFTL) {
cur 2097 net/pf.c if (rpool->cur->addr.p.dyn->pfid_acnt4 < 1 &&
cur 2101 net/pf.c raddr = &rpool->cur->addr.p.dyn->pfid_addr4;
cur 2102 net/pf.c rmask = &rpool->cur->addr.p.dyn->pfid_mask4;
cur 2107 net/pf.c if (rpool->cur->addr.p.dyn->pfid_acnt6 < 1 &&
cur 2111 net/pf.c raddr = &rpool->cur->addr.p.dyn->pfid_addr6;
cur 2112 net/pf.c rmask = &rpool->cur->addr.p.dyn->pfid_mask6;
cur 2116 net/pf.c } else if (rpool->cur->addr.type == PF_ADDR_TABLE) {
cur 2120 net/pf.c raddr = &rpool->cur->addr.v.a.addr;
cur 2121 net/pf.c rmask = &rpool->cur->addr.v.a.mask;
cur 2175 net/pf.c if (rpool->cur->addr.type == PF_ADDR_TABLE) {
cur 2176 net/pf.c if (!pfr_pool_get(rpool->cur->addr.p.tbl,
cur 2180 net/pf.c } else if (rpool->cur->addr.type == PF_ADDR_DYNIFTL) {
cur 2181 net/pf.c if (!pfr_pool_get(rpool->cur->addr.p.dyn->pfid_kt,
cur 2189 net/pf.c if ((rpool->cur = TAILQ_NEXT(rpool->cur, entries)) == NULL)
cur 2190 net/pf.c rpool->cur = TAILQ_FIRST(&rpool->list);
cur 2191 net/pf.c if (rpool->cur->addr.type == PF_ADDR_TABLE) {
cur 2193 net/pf.c if (pfr_pool_get(rpool->cur->addr.p.tbl,
cur 2197 net/pf.c if (rpool->cur != acur)
cur 2201 net/pf.c } else if (rpool->cur->addr.type == PF_ADDR_DYNIFTL) {
cur 2203 net/pf.c if (pfr_pool_get(rpool->cur->addr.p.dyn->pfid_kt,
cur 2207 net/pf.c if (rpool->cur != acur)
cur 2212 net/pf.c raddr = &rpool->cur->addr.v.a.addr;
cur 2213 net/pf.c rmask = &rpool->cur->addr.v.a.mask;
cur 2346 net/pf.c if (r->rpool.cur != NULL)
cur 2347 net/pf.c xdst = &r->rpool.cur->addr;
cur 2453 net/pf.c if (r->rpool.cur->addr.type == PF_ADDR_DYNIFTL){
cur 2457 net/pf.c if (r->rpool.cur->addr.p.dyn->
cur 2461 net/pf.c &r->rpool.cur->addr.p.dyn->
cur 2463 net/pf.c &r->rpool.cur->addr.p.dyn->
cur 2470 net/pf.c if (r->rpool.cur->addr.p.dyn->
cur 2474 net/pf.c &r->rpool.cur->addr.p.dyn->
cur 2476 net/pf.c &r->rpool.cur->addr.p.dyn->
cur 2484 net/pf.c &r->rpool.cur->addr.v.a.addr,
cur 2485 net/pf.c &r->rpool.cur->addr.v.a.mask,
cur 2532 net/pf.c &r->rpool.cur->addr.v.a.mask, daddr,
cur 2778 net/pf.c s->rt_kif = r->rpool.cur->kif;
cur 2785 net/pf.c s->rt_kif = r->rpool.cur->kif;
cur 5014 net/pf.c ifp = r->rpool.cur->kif ?
cur 5015 net/pf.c r->rpool.cur->kif->pfik_ifp : NULL;
cur 5199 net/pf.c ifp = r->rpool.cur->kif ? r->rpool.cur->kif->pfik_ifp : NULL;
cur 1252 net/pf_ioctl.c rule->rpool.cur = TAILQ_FIRST(&rule->rpool.list);
cur 1499 net/pf_ioctl.c newrule->rpool.cur = TAILQ_FIRST(&newrule->rpool.list);
cur 2264 net/pf_ioctl.c pool->cur = TAILQ_FIRST(&pool->list);
cur 2265 net/pf_ioctl.c PF_ACPY(&pool->counter, &pool->cur->addr.v.a.addr,
cur 524 net/pf_norm.c struct pf_frcache *frp, *fra, *cur = NULL;
cur 543 net/pf_norm.c cur = pool_get(&pf_cent_pl, PR_NOWAIT);
cur 544 net/pf_norm.c if (cur == NULL) {
cur 559 net/pf_norm.c cur->fr_off = off;
cur 560 net/pf_norm.c cur->fr_end = max;
cur 562 net/pf_norm.c LIST_INSERT_HEAD(&(*frag)->fr_cache, cur, fr_next);
cur 659 net/pf_norm.c cur = pool_get(&pf_cent_pl, PR_NOWAIT);
cur 660 net/pf_norm.c if (cur == NULL)
cur 664 net/pf_norm.c cur->fr_off = off;
cur 665 net/pf_norm.c cur->fr_end = max;
cur 666 net/pf_norm.c LIST_INSERT_AFTER(frp, cur, fr_next);
cur 713 net/pf_norm.c cur = pool_get(&pf_cent_pl, PR_NOWAIT);
cur 714 net/pf_norm.c if (cur == NULL)
cur 718 net/pf_norm.c cur->fr_off = off;
cur 719 net/pf_norm.c cur->fr_end = max;
cur 720 net/pf_norm.c LIST_INSERT_BEFORE(fra, cur, fr_next);
cur 726 net/pf_norm.c if (cur && fra->fr_off <= cur->fr_end) {
cur 730 net/pf_norm.c h->ip_id, cur->fr_off, cur->fr_end, off,
cur 732 net/pf_norm.c fra->fr_off = cur->fr_off;
cur 733 net/pf_norm.c LIST_REMOVE(cur, fr_next);
cur 734 net/pf_norm.c pool_put(&pf_cent_pl, cur);
cur 736 net/pf_norm.c cur = NULL;
cur 740 net/pf_norm.c KASSERT(cur == NULL);
cur 379 net/pfvar.h struct pf_pooladdr *cur;
cur 245 netinet/ip_carp.c struct in_addr last, cur, in;
cur 277 netinet/ip_carp.c cur.s_addr = 0;
cur 280 netinet/ip_carp.c last = cur;
cur 281 netinet/ip_carp.c cur.s_addr = 0xffffffff;
cur 286 netinet/ip_carp.c ntohl(in.s_addr) < ntohl(cur.s_addr)) {
cur 287 netinet/ip_carp.c cur.s_addr = in.s_addr;
cur 293 netinet/ip_carp.c (void *)&cur, sizeof(cur));
cur 1332 netinet/ip_carp.c int cur, last, count, found;
cur 1367 netinet/ip_carp.c cur = 0;
cur 1372 netinet/ip_carp.c last = cur;
cur 1373 netinet/ip_carp.c cur = 255;
cur 1391 netinet/ip_carp.c if (ifa && vh->sc_vhid > last && vh->sc_vhid < cur) {
cur 1392 netinet/ip_carp.c cur = vh->sc_vhid;
cur 2537 netinet/tcp_input.c struct sackhole *cur, *p, *temp;
cur 2591 netinet/tcp_input.c cur = tp->snd_holes;
cur 2592 netinet/tcp_input.c cur->start = th->th_ack;
cur 2593 netinet/tcp_input.c cur->end = sack.start;
cur 2594 netinet/tcp_input.c cur->rxmit = cur->start;
cur 2595 netinet/tcp_input.c cur->next = NULL;
cur 2602 netinet/tcp_input.c cur->dups = min(tcprexmtthresh,
cur 2603 netinet/tcp_input.c ((sack.end - cur->end)/tp->t_maxseg));
cur 2604 netinet/tcp_input.c if (cur->dups < 1)
cur 2605 netinet/tcp_input.c cur->dups = 1;
cur 2609 netinet/tcp_input.c p = cur = tp->snd_holes;
cur 2610 netinet/tcp_input.c while (cur) {
cur 2611 netinet/tcp_input.c if (SEQ_LEQ(sack.end, cur->start))
cur 2614 netinet/tcp_input.c if (SEQ_GEQ(sack.start, cur->end)) {
cur 2616 netinet/tcp_input.c cur->dups++;
cur 2617 netinet/tcp_input.c if (((sack.end - cur->end)/tp->t_maxseg) >=
cur 2619 netinet/tcp_input.c cur->dups = tcprexmtthresh;
cur 2620 netinet/tcp_input.c p = cur;
cur 2621 netinet/tcp_input.c cur = cur->next;
cur 2624 netinet/tcp_input.c if (SEQ_LEQ(sack.start, cur->start)) {
cur 2627 netinet/tcp_input.c if (SEQ_GT(sack.end, cur->rxmit))
cur 2629 netinet/tcp_input.c tcp_seq_subtract(cur->rxmit,
cur 2630 netinet/tcp_input.c cur->start);
cur 2634 netinet/tcp_input.c cur->start);
cur 2636 netinet/tcp_input.c if (SEQ_GEQ(sack.end, cur->end)) {
cur 2638 netinet/tcp_input.c if (p != cur) {
cur 2639 netinet/tcp_input.c p->next = cur->next;
cur 2640 netinet/tcp_input.c pool_put(&sackhl_pool, cur);
cur 2641 netinet/tcp_input.c cur = p->next;
cur 2643 netinet/tcp_input.c cur = cur->next;
cur 2645 netinet/tcp_input.c p = cur;
cur 2652 netinet/tcp_input.c cur->start = sack.end;
cur 2653 netinet/tcp_input.c cur->rxmit = SEQ_MAX(cur->rxmit, cur->start);
cur 2654 netinet/tcp_input.c p = cur;
cur 2655 netinet/tcp_input.c cur = cur->next;
cur 2659 netinet/tcp_input.c if (SEQ_GEQ(sack.end, cur->end)) {
cur 2661 netinet/tcp_input.c if (SEQ_GT(cur->rxmit, sack.start))
cur 2663 netinet/tcp_input.c tcp_seq_subtract(cur->rxmit,
cur 2666 netinet/tcp_input.c cur->end = sack.start;
cur 2667 netinet/tcp_input.c cur->rxmit = SEQ_MIN(cur->rxmit, cur->end);
cur 2668 netinet/tcp_input.c cur->dups++;
cur 2669 netinet/tcp_input.c if (((sack.end - cur->end)/tp->t_maxseg) >=
cur 2671 netinet/tcp_input.c cur->dups = tcprexmtthresh;
cur 2672 netinet/tcp_input.c p = cur;
cur 2673 netinet/tcp_input.c cur = cur->next;
cur 2676 netinet/tcp_input.c if (SEQ_LT(cur->start, sack.start) &&
cur 2677 netinet/tcp_input.c SEQ_GT(cur->end, sack.end)) {
cur 2687 netinet/tcp_input.c if (SEQ_GT(cur->rxmit, sack.end))
cur 2691 netinet/tcp_input.c else if (SEQ_GT(cur->rxmit, sack.start))
cur 2693 netinet/tcp_input.c tcp_seq_subtract(cur->rxmit,
cur 2696 netinet/tcp_input.c temp->next = cur->next;
cur 2698 netinet/tcp_input.c temp->end = cur->end;
cur 2699 netinet/tcp_input.c temp->dups = cur->dups;
cur 2700 netinet/tcp_input.c temp->rxmit = SEQ_MAX(cur->rxmit, temp->start);
cur 2701 netinet/tcp_input.c cur->end = sack.start;
cur 2702 netinet/tcp_input.c cur->rxmit = SEQ_MIN(cur->rxmit, cur->end);
cur 2703 netinet/tcp_input.c cur->dups++;
cur 2704 netinet/tcp_input.c if (((sack.end - cur->end)/tp->t_maxseg) >=
cur 2706 netinet/tcp_input.c cur->dups = tcprexmtthresh;
cur 2707 netinet/tcp_input.c cur->next = temp;
cur 2709 netinet/tcp_input.c cur = p->next;
cur 2743 netinet/tcp_input.c cur = tp->snd_holes;
cur 2744 netinet/tcp_input.c while (cur) {
cur 2745 netinet/tcp_input.c tp->retran_data += cur->rxmit - cur->start;
cur 2746 netinet/tcp_input.c cur = cur->next;
cur 2769 netinet/tcp_input.c struct sackhole *cur = tp->snd_holes;
cur 2771 netinet/tcp_input.c while (cur)
cur 2772 netinet/tcp_input.c if (SEQ_LEQ(cur->end, lastack)) {
cur 2773 netinet/tcp_input.c prev = cur;
cur 2774 netinet/tcp_input.c cur = cur->next;
cur 2777 netinet/tcp_input.c } else if (SEQ_LT(cur->start, lastack)) {
cur 2778 netinet/tcp_input.c cur->start = lastack;
cur 2779 netinet/tcp_input.c if (SEQ_LT(cur->rxmit, cur->start))
cur 2780 netinet/tcp_input.c cur->rxmit = cur->start;
cur 2784 netinet/tcp_input.c tp->snd_holes = cur;
cur 177 netinet/tcp_output.c struct sackhole *cur = tp->snd_holes;
cur 178 netinet/tcp_output.c if (cur == NULL)
cur 187 netinet/tcp_output.c while (cur->next) {
cur 188 netinet/tcp_output.c if (SEQ_LT(tp->snd_nxt, cur->end))
cur 190 netinet/tcp_output.c if (SEQ_GEQ(tp->snd_nxt, cur->next->start))
cur 191 netinet/tcp_output.c cur = cur->next;
cur 193 netinet/tcp_output.c tp->snd_nxt = cur->next->start;
cur 197 netinet/tcp_output.c if (SEQ_LT(tp->snd_nxt, cur->end))
cur 955 uvm/uvm_map.c struct vm_map_entry *cur;
cur 970 uvm/uvm_map.c cur = map->hint;
cur 973 uvm/uvm_map.c if (cur == &map->header)
cur 974 uvm/uvm_map.c cur = cur->next;
cur 977 uvm/uvm_map.c if (address >= cur->start) {
cur 990 uvm/uvm_map.c if ((cur != last) && (cur->end > address)) {
cur 992 uvm/uvm_map.c *entry = cur;
cur 994 uvm/uvm_map.c cur, 0, 0, 0);
cur 1004 uvm/uvm_map.c last = cur->next;
cur 1005 uvm/uvm_map.c cur = map->header.next;
cur 1013 uvm/uvm_map.c cur = RB_ROOT(&map->rbhead);
cur 1019 uvm/uvm_map.c while (cur) {
cur 1020 uvm/uvm_map.c if (address >= cur->start) {
cur 1021 uvm/uvm_map.c if (address < cur->end) {
cur 1022 uvm/uvm_map.c *entry = cur;
cur 1023 uvm/uvm_map.c SAVE_HINT(map, map->hint, cur);
cur 1026 uvm/uvm_map.c prev = cur;
cur 1027 uvm/uvm_map.c cur = RB_RIGHT(cur, rb_entry);
cur 1029 uvm/uvm_map.c cur = RB_LEFT(cur, rb_entry);
cur 1040 uvm/uvm_map.c while (cur != last) {
cur 1041 uvm/uvm_map.c if (cur->end > address) {
cur 1042 uvm/uvm_map.c if (address >= cur->start) {
cur 1048 uvm/uvm_map.c *entry = cur;
cur 1049 uvm/uvm_map.c SAVE_HINT(map, map->hint, cur);
cur 1051 uvm/uvm_map.c cur, 0, 0, 0);
cur 1056 uvm/uvm_map.c cur = cur->next;
cur 1059 uvm/uvm_map.c *entry = cur->prev;
cur 1679 uvm/uvm_map.c vaddr_t cur = start;
cur 1683 uvm/uvm_map.c if (tmpent->start < cur)
cur 1690 uvm/uvm_map.c cur = tmpent->end;
cur 102 uvm/uvm_stat.c int cur[MAXHISTS];
cur 110 uvm/uvm_stat.c cur[lcv] = hists[lcv]->f;
cur 125 uvm/uvm_stat.c if (cur[lcv] == -1)
cur 132 uvm/uvm_stat.c if (hists[lcv]->e[cur[lcv]].fmt == NULL) {
cur 133 uvm/uvm_stat.c cur[lcv] = (cur[lcv] + 1) % (hists[lcv]->n);
cur 134 uvm/uvm_stat.c if (cur[lcv] == hists[lcv]->f)
cur 135 uvm/uvm_stat.c cur[lcv] = -1;
cur 145 uvm/uvm_stat.c timercmp(&hists[lcv]->e[cur[lcv]].tv, &tv, <)) {
cur 146 uvm/uvm_stat.c tv = hists[lcv]->e[cur[lcv]].tv;
cur 156 uvm/uvm_stat.c uvmhist_print(&hists[hi]->e[cur[hi]]);
cur 157 uvm/uvm_stat.c cur[hi] = (cur[hi] + 1) % (hists[hi]->n);
cur 158 uvm/uvm_stat.c if (cur[hi] == hists[hi]->f)
cur 159 uvm/uvm_stat.c cur[hi] = -1;