cur 959 arch/i386/i386/est.c u_int16_t idhi, idlo, cur; cur 977 arch/i386/i386/est.c cur = msr & 0xffff; cur 980 arch/i386/i386/est.c crcur = (cur >> 8) & 0xff; cur 1020 arch/i386/i386/est.c if (cur == idhi || cur == idlo) { cur 1030 arch/i386/i386/est.c fake_table[1] = cur; cur 1039 arch/i386/i386/est.c mhz = MSR2MHZ(cur, bus_clock); cur 1040 arch/i386/i386/est.c mv = MSR2MV(cur); cur 1047 arch/i386/i386/est.c if (cur == est_fqlist->table[i]) cur 378 compat/ibcs2/ibcs2_fcntl.c off_t off, cur; cur 405 compat/ibcs2/ibcs2_fcntl.c if ((error = sys_lseek(p, &ols, (register_t *)&cur)) != 0) cur 417 compat/ibcs2/ibcs2_fcntl.c off = cur - (off_t)ifl.l_start; cur 495 dev/ic/an.c int cur, id; cur 508 dev/ic/an.c cur = sc->sc_txcur; cur 509 dev/ic/an.c if (sc->sc_txd[cur].d_fid == id) { cur 510 dev/ic/an.c sc->sc_txd[cur].d_inuse = 0; cur 511 dev/ic/an.c DPRINTF2(("an_txeof: sent %x/%d\n", id, cur)); cur 512 dev/ic/an.c AN_INC(cur, AN_TX_RING_CNT); cur 513 dev/ic/an.c sc->sc_txcur = cur; cur 515 dev/ic/an.c for (cur = 0; cur < AN_TX_RING_CNT; cur++) { cur 516 dev/ic/an.c if (id == sc->sc_txd[cur].d_fid) { cur 517 dev/ic/an.c sc->sc_txd[cur].d_inuse = 0; cur 526 dev/ic/an.c id, cur); cur 1096 dev/ic/an.c int cur, fid; cur 1105 dev/ic/an.c cur = sc->sc_txnext; cur 1116 dev/ic/an.c if (sc->sc_txd[cur].d_inuse) { cur 1118 dev/ic/an.c sc->sc_txd[cur].d_fid, cur)); cur 1206 dev/ic/an.c fid = sc->sc_txd[cur].d_fid; cur 1219 dev/ic/an.c fid, cur)); cur 1220 dev/ic/an.c sc->sc_txd[cur].d_inuse = 1; cur 1223 dev/ic/an.c sc->sc_txd[cur].d_inuse = 0; cur 1228 dev/ic/an.c AN_INC(cur, AN_TX_RING_CNT); cur 1229 dev/ic/an.c sc->sc_txnext = cur; cur 635 dev/ic/bt463.c int bt463_set_cursor (rc, cur) cur 637 dev/ic/bt463.c struct wsdisplay_cursor *cur; cur 640 dev/ic/bt463.c return tga_builtin_set_cursor(data->cookie, cur); cur 643 dev/ic/bt463.c int bt463_get_cursor (rc, cur) cur 645 dev/ic/bt463.c struct wsdisplay_cursor *cur; cur 648 dev/ic/bt463.c return tga_builtin_get_cursor(data->cookie, cur); cur 651 dev/ic/bt463.c int bt463_set_curpos (rc, cur) cur 653 dev/ic/bt463.c struct wsdisplay_curpos *cur; cur 656 dev/ic/bt463.c return tga_builtin_set_curpos(data->cookie, cur); cur 659 dev/ic/bt463.c int bt463_get_curpos (rc, cur) cur 661 dev/ic/bt463.c struct wsdisplay_curpos *cur; cur 664 dev/ic/bt463.c return tga_builtin_get_curpos(data->cookie, cur); cur 667 dev/ic/bt463.c int bt463_get_curmax (rc, cur) cur 669 dev/ic/bt463.c struct wsdisplay_curpos *cur; cur 672 dev/ic/bt463.c return tga_builtin_get_curmax(data->cookie, cur); cur 2603 dev/ic/dc.c int frag, cur, cnt = 0, i; cur 2617 dev/ic/dc.c cur = frag = *txidx; cur 2641 dev/ic/dc.c cur = frag; cur 2647 dev/ic/dc.c sc->dc_cdata.dc_tx_chain[cur].sd_mbuf = m_head; cur 2648 dev/ic/dc.c sc->sc_tx_sparemap = sc->dc_cdata.dc_tx_chain[cur].sd_map; cur 2649 dev/ic/dc.c sc->dc_cdata.dc_tx_chain[cur].sd_map = map; cur 2650 dev/ic/dc.c sc->dc_ldata->dc_tx_list[cur].dc_ctl |= htole32(DC_TXCTL_LASTFRAG); cur 2655 dev/ic/dc.c sc->dc_ldata->dc_tx_list[cur].dc_ctl |= cur 2658 dev/ic/dc.c sc->dc_ldata->dc_tx_list[cur].dc_ctl |= cur 2662 dev/ic/dc.c sc->dc_ldata->dc_tx_list[cur].dc_ctl |= cur 1607 dev/ic/gem.c u_int32_t cur, frag, i; cur 1610 dev/ic/gem.c cur = frag = *bixp; cur 1611 dev/ic/gem.c map = sc->sc_txd[cur].sd_map; cur 1636 dev/ic/gem.c cur = frag; cur 1642 dev/ic/gem.c sc->sc_txd[*bixp].sd_map = sc->sc_txd[cur].sd_map; cur 1643 dev/ic/gem.c sc->sc_txd[cur].sd_map = map; cur 1644 dev/ic/gem.c sc->sc_txd[cur].sd_mbuf = mhead; cur 1409 dev/ic/hme.c int frag, cur, cnt = 0; cur 1413 dev/ic/hme.c cur = frag = *bixp; cur 1443 dev/ic/hme.c cur = frag; cur 1453 dev/ic/hme.c flags = HME_XD_GETFLAGS(sc->sc_pci, hr->rb_txd, cur); cur 1455 dev/ic/hme.c HME_XD_SETFLAGS(sc->sc_pci, hr->rb_txd, cur, flags); cur 1456 dev/ic/hme.c sc->sc_txd[cur].sd_mbuf = mhead; cur 1150 dev/ic/i82596.c int off, cur, prev; cur 1152 dev/ic/i82596.c cur = sc->xctail; cur 1156 dev/ic/i82596.c printf("%s: xmit buffer %d\n", sc->sc_dev.dv_xname, cur); cur 1162 dev/ic/i82596.c sc->ie_bus_write16(sc, IE_CMD_XMIT_DESC(sc->xmit_cmds, cur), cur 1163 dev/ic/i82596.c IE_XBD_ADDR(sc->xbds, cur)); cur 1165 dev/ic/i82596.c sc->ie_bus_write16(sc, IE_CMD_XMIT_STATUS(sc->xmit_cmds, cur), 0); cur 1171 dev/ic/i82596.c sc->ie_bus_write16(sc, IE_CMD_XMIT_LINK(sc->xmit_cmds, cur), cur 1172 dev/ic/i82596.c IE_CMD_NOP_ADDR(sc->nop_cmds, cur)); cur 1173 dev/ic/i82596.c sc->ie_bus_write16(sc, IE_CMD_XMIT_CMD(sc->xmit_cmds, cur), cur 1179 dev/ic/i82596.c sc->ie_bus_write16(sc, IE_CMD_NOP_STATUS(sc->nop_cmds, cur), 0); cur 1180 dev/ic/i82596.c sc->ie_bus_write16(sc, IE_CMD_NOP_LINK(sc->nop_cmds, cur), cur 1181 dev/ic/i82596.c IE_CMD_NOP_ADDR(sc->nop_cmds, cur)); cur 1186 dev/ic/i82596.c prev = (cur + NTXBUF - 1) % NTXBUF; cur 1189 dev/ic/i82596.c IE_CMD_XMIT_ADDR(sc->xmit_cmds, cur)); cur 1199 dev/ic/i82596.c sc->ie_bus_write16(sc, IE_CMD_XMIT_LINK(sc->xmit_cmds,cur), cur 1202 dev/ic/i82596.c sc->ie_bus_write16(sc, IE_CMD_XMIT_CMD(sc->xmit_cmds, cur), cur 1206 dev/ic/i82596.c sc->ie_bus_write16(sc, off, IE_CMD_XMIT_ADDR(sc->xmit_cmds, cur)); cur 560 dev/ic/malo.c ring->cur = ring->next = 0; cur 676 dev/ic/malo.c ring->cur = ring->next = 0; cur 721 dev/ic/malo.c ring->cur = ring->next = ring->stat = 0; cur 822 dev/ic/malo.c ring->cur = ring->next = ring->stat = 0; cur 1431 dev/ic/malo.c desc = &sc->sc_txring.desc[sc->sc_txring.cur]; cur 1432 dev/ic/malo.c data = &sc->sc_txring.data[sc->sc_txring.cur]; cur 1513 dev/ic/malo.c sc->sc_txring.cur * sizeof(struct malo_tx_desc), cur 1517 dev/ic/malo.c sc->sc_dev.dv_xname, m0->m_pkthdr.len, sc->sc_txring.cur)); cur 1520 dev/ic/malo.c sc->sc_txring.cur = (sc->sc_txring.cur + 1) % MALO_TX_RING_COUNT; cur 1543 dev/ic/malo.c desc = &sc->sc_txring.desc[sc->sc_txring.cur]; cur 1544 dev/ic/malo.c data = &sc->sc_txring.data[sc->sc_txring.cur]; cur 1631 dev/ic/malo.c sc->sc_txring.cur * sizeof(struct malo_tx_desc), cur 1635 dev/ic/malo.c sc->sc_dev.dv_xname, m0->m_pkthdr.len, sc->sc_txring.cur)); cur 1638 dev/ic/malo.c sc->sc_txring.cur = (sc->sc_txring.cur + 1) % MALO_TX_RING_COUNT; cur 1674 dev/ic/malo.c desc = &sc->sc_rxring.desc[sc->sc_rxring.cur]; cur 1675 dev/ic/malo.c data = &sc->sc_rxring.data[sc->sc_rxring.cur]; cur 1678 dev/ic/malo.c sc->sc_rxring.cur * sizeof(struct malo_rx_desc), cur 1684 dev/ic/malo.c sc->sc_rxring.cur, desc->rxctrl, desc->rssi, desc->status, cur 1784 dev/ic/malo.c sc->sc_rxring.cur * sizeof(struct malo_rx_desc), cur 1787 dev/ic/malo.c sc->sc_rxring.cur = (sc->sc_rxring.cur + 1) % cur 30 dev/ic/malo.h int cur; cur 45 dev/ic/malo.h int cur; cur 1424 dev/ic/midway.c loc = sc->txslot[slot].cur = sc->txslot[slot].start; cur 1472 dev/ic/midway.c sc->rxslot[slot].cur = sc->rxslot[slot].start; cur 1988 dev/ic/midway.c u_int32_t cur = sc->txslot[chan].cur, cur 2015 dev/ic/midway.c dma = cur; cur 2025 dev/ic/midway.c sc->sc_dev.dv_xname, chan, l->t, cur, (cur-start)/4, need, addtail); cur 2042 dev/ic/midway.c EN_WRITE(sc, cur, l->tbd1); cur 2043 dev/ic/midway.c EN_WRAPADD(start, stop, cur, 4); cur 2044 dev/ic/midway.c EN_WRITE(sc, cur, l->tbd2); cur 2045 dev/ic/midway.c EN_WRAPADD(start, stop, cur, 4); cur 2065 dev/ic/midway.c (len % 4) == 0 && ((unsigned long) data % 4) == 0 && (cur % 4) == 0)) { cur 2079 dev/ic/midway.c EN_WRITEDAT(sc, cur, *data); cur 2081 dev/ic/midway.c EN_WRAPADD(start, stop, cur, 4); cur 2086 dev/ic/midway.c sc->sc_dev.dv_xname, chan, len, need, cur); cur 2092 dev/ic/midway.c if (dma != cur) { cur 2093 dev/ic/midway.c EN_DTQADD(sc, WORD_IDX(start,cur), chan, MIDDMA_JK, 0, 0, 0); cur 2096 dev/ic/midway.c sc->sc_dev.dv_xname, chan, cur); cur 2121 dev/ic/midway.c EN_WRAPADD(start, stop, cur, len); cur 2124 dev/ic/midway.c sc->sc_dev.dv_xname, chan, len, need, cur); cur 2130 dev/ic/midway.c dma = cur; /* update dma pointer */ cur 2155 dev/ic/midway.c EN_WRAPADD(start, stop, cur, cnt); cur 2158 dev/ic/midway.c sc->sc_dev.dv_xname, chan, cnt, need, cur); cur 2184 dev/ic/midway.c EN_WRAPADD(start, stop, cur, cnt); cur 2187 dev/ic/midway.c sc->sc_dev.dv_xname, chan, cnt, need, cur); cur 2203 dev/ic/midway.c EN_WRAPADD(start, stop, cur, cnt); cur 2206 dev/ic/midway.c sc->sc_dev.dv_xname, chan, cnt, need, cur); cur 2223 dev/ic/midway.c EN_WRAPADD(start, stop, cur, cnt); cur 2226 dev/ic/midway.c sc->sc_dev.dv_xname, chan, cnt, need, cur); cur 2247 dev/ic/midway.c EN_WRAPADD(start, stop, cur, len); cur 2250 dev/ic/midway.c sc->sc_dev.dv_xname, chan, len, need, cur); cur 2258 dev/ic/midway.c dma = cur; /* update dma pointer */ cur 2287 dev/ic/midway.c EN_WRAPADD(start, stop, cur, pad); cur 2292 dev/ic/midway.c sc->sc_dev.dv_xname, chan, pad, need, cur); cur 2302 dev/ic/midway.c sc->sc_dev.dv_xname, chan, pad * sizeof(u_int32_t), cur); cur 2305 dev/ic/midway.c EN_WRITEDAT(sc, cur, 0); /* no byte order issues with zero */ cur 2306 dev/ic/midway.c EN_WRAPADD(start, stop, cur, 4); cur 2309 dev/ic/midway.c EN_WRITE(sc, cur, l->pdu1); /* in host byte order */ cur 2310 dev/ic/midway.c EN_WRAPADD(start, stop, cur, 8); cur 2314 dev/ic/midway.c if (addtail || dma != cur) { cur 2316 dev/ic/midway.c EN_DTQADD(sc, WORD_IDX(start,cur), chan, MIDDMA_JK, 0, cur 2323 dev/ic/midway.c sc->txslot[chan].cur = cur; cur 2326 dev/ic/midway.c sc->sc_dev.dv_xname, chan, cur); cur 2394 dev/ic/midway.c if (val > sc->txslot[lcv].cur) cur 2395 dev/ic/midway.c sc->txslot[lcv].bfree = val - sc->txslot[lcv].cur; cur 2397 dev/ic/midway.c sc->txslot[lcv].bfree = (val + (EN_TXSZ*1024)) - sc->txslot[lcv].cur; cur 2628 dev/ic/midway.c u_int32_t cur, dstart, rbd, pdu, *sav, dma, bcode, count, *data, *datastop; cur 2658 dev/ic/midway.c cur = sc->rxslot[slot].cur; cur 2662 dev/ic/midway.c sc->sc_dev.dv_xname, slot, vci, raw, start, stop, cur); cur 2670 dev/ic/midway.c if (dstart == cur) { cur 2692 dev/ic/midway.c if (dstart > cur) cur 2693 dev/ic/midway.c mlen = dstart - cur; cur 2695 dev/ic/midway.c mlen = (dstart + (EN_RXSZ*1024)) - cur; cur 2705 dev/ic/midway.c rbd = EN_READ(sc, cur); cur 2718 dev/ic/midway.c pdu = cur + tlen - MID_PDU_SIZE; cur 2753 dev/ic/midway.c if (sav[0] != cur) { cur 2808 dev/ic/midway.c sav[0] = cur; cur 2833 dev/ic/midway.c dma = cur; /* dma = last location we told chip about */ cur 2851 dev/ic/midway.c *data = EN_READDAT(sc, cur); cur 2853 dev/ic/midway.c EN_WRAPADD(start, stop, cur, 4); cur 2864 dev/ic/midway.c if (dma != cur) { cur 2865 dev/ic/midway.c EN_DRQADD(sc, WORD_IDX(start,cur), vci, MIDDMA_JK, 0, 0, 0, 0); cur 2868 dev/ic/midway.c sc->sc_dev.dv_xname, slot, vci, cur); cur 2880 dev/ic/midway.c EN_WRAPADD(start, stop, cur, tlen); cur 2889 dev/ic/midway.c dma = cur; /* update dma pointer */ cur 2915 dev/ic/midway.c EN_WRAPADD(start, stop, cur, cnt); cur 2934 dev/ic/midway.c EN_WRAPADD(start, stop, cur, cnt); cur 2953 dev/ic/midway.c EN_WRAPADD(start, stop, cur, tlen); cur 2964 dev/ic/midway.c dma = cur; /* update dma pointer */ cur 2971 dev/ic/midway.c if (fill || dma != cur) { cur 2978 dev/ic/midway.c sc->sc_dev.dv_xname, slot, vci, dma, cur); cur 2980 dev/ic/midway.c EN_WRAPADD(start, stop, cur, fill); cur 2981 dev/ic/midway.c EN_DRQADD(sc, WORD_IDX(start,cur), vci, MIDDMA_JK, 0, mlen, cur 3004 dev/ic/midway.c sc->rxslot[slot].cur = cur; /* update master copy of 'cur' */ cur 3008 dev/ic/midway.c sc->sc_dev.dv_xname, slot, vci, cur); cur 3127 dev/ic/midway.c sc->txslot[slot].start, sc->txslot[slot].stop, sc->txslot[slot].cur, cur 3128 dev/ic/midway.c (sc->txslot[slot].cur - sc->txslot[slot].start)/4); cur 3143 dev/ic/midway.c sc->rxslot[slot].stop, sc->rxslot[slot].cur); cur 136 dev/ic/midwayvar.h u_int32_t cur; /* next free area (byte offset) */ cur 155 dev/ic/midwayvar.h u_int32_t cur; /* where I am at */ cur 369 dev/ic/mtd8xx.c int frag, cur, cnt = 0, i, total_len = 0; cur 383 dev/ic/mtd8xx.c cur = frag = *txidx; cur 401 dev/ic/mtd8xx.c cur = frag; cur 407 dev/ic/mtd8xx.c sc->mtd_cdata.mtd_tx_chain[cur].sd_mbuf = m_head; cur 408 dev/ic/mtd8xx.c sc->sc_tx_sparemap = sc->mtd_cdata.mtd_tx_chain[cur].sd_map; cur 409 dev/ic/mtd8xx.c sc->mtd_cdata.mtd_tx_chain[cur].sd_map = map; cur 410 dev/ic/mtd8xx.c sc->mtd_ldata->mtd_tx_list[cur].td_tcw |= htole32(TCW_LD | TCW_IC); cur 412 dev/ic/mtd8xx.c sc->mtd_ldata->mtd_tx_list[cur].td_tcw |= cur 2707 dev/ic/ncr53c9x.c struct timeval wait, cur; cur 2718 dev/ic/ncr53c9x.c microtime(&cur); cur 2719 dev/ic/ncr53c9x.c } while (timercmp(&cur, &wait, <=)); cur 349 dev/ic/rt2560.c ring->cur = ring->next = 0; cur 445 dev/ic/rt2560.c ring->cur = ring->next = 0; cur 495 dev/ic/rt2560.c ring->cur = ring->next = 0; cur 608 dev/ic/rt2560.c ring->cur = ring->next = 0; cur 1232 dev/ic/rt2560.c struct rt2560_rx_desc *desc = &sc->rxq.desc[sc->rxq.cur]; cur 1233 dev/ic/rt2560.c struct rt2560_rx_data *data = &sc->rxq.data[sc->rxq.cur]; cur 1236 dev/ic/rt2560.c sc->rxq.cur * RT2560_RX_DESC_SIZE, RT2560_RX_DESC_SIZE, cur 1265 dev/ic/rt2560.c sc->rxq.cur * RT2560_RX_DESC_SIZE, RT2560_RX_DESC_SIZE, cur 1268 dev/ic/rt2560.c DPRINTFN(15, ("rx done idx=%u\n", sc->rxq.cur)); cur 1270 dev/ic/rt2560.c sc->rxq.cur = (sc->rxq.cur + 1) % RT2560_RX_RING_COUNT; cur 1550 dev/ic/rt2560.c desc = &sc->bcnq.desc[sc->bcnq.cur]; cur 1551 dev/ic/rt2560.c data = &sc->bcnq.data[sc->bcnq.cur]; cur 1572 dev/ic/rt2560.c sc->bcnq.cur * RT2560_TX_DESC_SIZE, RT2560_TX_DESC_SIZE, cur 1610 dev/ic/rt2560.c desc = &sc->prioq.desc[sc->prioq.cur]; cur 1611 dev/ic/rt2560.c data = &sc->prioq.data[sc->prioq.cur]; cur 1679 dev/ic/rt2560.c sc->prioq.cur * RT2560_TX_DESC_SIZE, RT2560_TX_DESC_SIZE, cur 1683 dev/ic/rt2560.c m0->m_pkthdr.len, sc->prioq.cur, rate)); cur 1687 dev/ic/rt2560.c sc->prioq.cur = (sc->prioq.cur + 1) % RT2560_PRIO_RING_COUNT; cur 68 dev/ic/rt2560var.h int cur; cur 87 dev/ic/rt2560var.h int cur; cur 383 dev/ic/rt2661.c ring->cur = ring->next = ring->stat = 0; cur 478 dev/ic/rt2661.c ring->cur = ring->next = ring->stat = 0; cur 526 dev/ic/rt2661.c ring->cur = ring->next = 0; cur 636 dev/ic/rt2661.c ring->cur = ring->next = 0; cur 1008 dev/ic/rt2661.c struct rt2661_rx_desc *desc = &sc->rxq.desc[sc->rxq.cur]; cur 1009 dev/ic/rt2661.c struct rt2661_rx_data *data = &sc->rxq.data[sc->rxq.cur]; cur 1012 dev/ic/rt2661.c sc->rxq.cur * RT2661_RX_DESC_SIZE, RT2661_RX_DESC_SIZE, cur 1137 dev/ic/rt2661.c sc->rxq.cur * RT2661_RX_DESC_SIZE, RT2661_RX_DESC_SIZE, cur 1140 dev/ic/rt2661.c DPRINTFN(15, ("rx intr idx=%u\n", sc->rxq.cur)); cur 1142 dev/ic/rt2661.c sc->rxq.cur = (sc->rxq.cur + 1) % RT2661_RX_RING_COUNT; cur 1457 dev/ic/rt2661.c desc = &sc->mgtq.desc[sc->mgtq.cur]; cur 1458 dev/ic/rt2661.c data = &sc->mgtq.data[sc->mgtq.cur]; cur 1527 dev/ic/rt2661.c sc->mgtq.cur * RT2661_TX_DESC_SIZE, RT2661_TX_DESC_SIZE, cur 1531 dev/ic/rt2661.c m0->m_pkthdr.len, sc->mgtq.cur, rate)); cur 1535 dev/ic/rt2661.c sc->mgtq.cur = (sc->mgtq.cur + 1) % RT2661_MGT_RING_COUNT; cur 1628 dev/ic/rt2661.c desc = &txq->desc[txq->cur]; cur 1629 dev/ic/rt2661.c data = &txq->data[txq->cur]; cur 1655 dev/ic/rt2661.c txq->cur * RT2661_TX_DESC_SIZE, RT2661_TX_DESC_SIZE, cur 1659 dev/ic/rt2661.c txq->cur = (txq->cur + 1) % RT2661_TX_RING_COUNT; cur 1664 dev/ic/rt2661.c data = &txq->data[txq->cur]; cur 1665 dev/ic/rt2661.c desc = &txq->desc[txq->cur]; cur 1747 dev/ic/rt2661.c bus_dmamap_sync(sc->sc_dmat, txq->map, txq->cur * RT2661_TX_DESC_SIZE, cur 1751 dev/ic/rt2661.c m0->m_pkthdr.len, txq->cur, rate)); cur 1755 dev/ic/rt2661.c txq->cur = (txq->cur + 1) % RT2661_TX_RING_COUNT; cur 64 dev/ic/rt2661var.h int cur; cur 81 dev/ic/rt2661var.h int cur; cur 1193 dev/pci/azalia.c this->dacs.cur = -1; cur 1194 dev/pci/azalia.c this->adcs.cur = -1; cur 1227 dev/pci/azalia.c prev_dac = this->dacs.cur; cur 1228 dev/pci/azalia.c this->dacs.cur = newdac; cur 1229 dev/pci/azalia.c group = &this->dacs.groups[this->dacs.cur]; cur 1249 dev/pci/azalia.c prev_adc = this->adcs.cur; cur 1250 dev/pci/azalia.c this->adcs.cur = newadc; cur 1251 dev/pci/azalia.c group = &this->adcs.groups[this->adcs.cur]; cur 1285 dev/pci/azalia.c group = &this->dacs.groups[this->dacs.cur]; cur 1296 dev/pci/azalia.c group = &this->adcs.groups[this->adcs.cur]; cur 1411 dev/pci/azalia.c group = &this->adcs.groups[this->adcs.cur]; cur 1413 dev/pci/azalia.c group = &this->dacs.groups[this->dacs.cur]; cur 524 dev/pci/azalia.h int cur; cur 252 dev/pci/azalia_codec.c this->dacs.cur = 0; cur 265 dev/pci/azalia_codec.c this->adcs.cur = 0; cur 1009 dev/pci/azalia_codec.c mc->un.ord = this->dacs.cur; cur 1014 dev/pci/azalia_codec.c mc->un.ord = this->adcs.cur; cur 1255 dev/pci/azalia_codec.c mc->un.ord, this->adcs.cur); cur 1265 dev/pci/azalia_codec.c this->dacs.cur, mc->un.ord); cur 2732 dev/pci/if_bge.c u_int32_t frag, cur; cur 2745 dev/pci/if_bge.c cur = frag = *txidx; cur 2810 dev/pci/if_bge.c cur = frag; cur 2823 dev/pci/if_bge.c sc->bge_rdata->bge_tx_ring[cur].bge_flags |= BGE_TXBDFLAG_END; cur 2824 dev/pci/if_bge.c sc->bge_cdata.bge_tx_chain[cur] = m_head; cur 2826 dev/pci/if_bge.c sc->txdma[cur] = dma; cur 1842 dev/pci/if_cas.c u_int32_t cur, frag, i; cur 1845 dev/pci/if_cas.c cur = frag = *bixp; cur 1846 dev/pci/if_cas.c map = sc->sc_txd[cur].sd_map; cur 1871 dev/pci/if_cas.c cur = frag; cur 1877 dev/pci/if_cas.c sc->sc_txd[*bixp].sd_map = sc->sc_txd[cur].sd_map; cur 1878 dev/pci/if_cas.c sc->sc_txd[cur].sd_map = map; cur 1879 dev/pci/if_cas.c sc->sc_txd[cur].sd_mbuf = mhead; cur 395 dev/pci/if_iwi.c ring->cur = ring->next = 0; cur 445 dev/pci/if_iwi.c ring->cur = ring->next = 0; cur 470 dev/pci/if_iwi.c ring->cur = ring->next = 0; cur 548 dev/pci/if_iwi.c ring->cur = ring->next = 0; cur 584 dev/pci/if_iwi.c ring->cur = 0; cur 635 dev/pci/if_iwi.c ring->cur = 0; cur 1101 dev/pci/if_iwi.c for (; sc->rxq.cur != hw;) { cur 1102 dev/pci/if_iwi.c data = &sc->rxq.data[sc->rxq.cur]; cur 1125 dev/pci/if_iwi.c sc->rxq.cur = (sc->rxq.cur + 1) % IWI_RX_RING_COUNT; cur 1231 dev/pci/if_iwi.c desc = &sc->cmdq.desc[sc->cmdq.cur]; cur 1239 dev/pci/if_iwi.c sc->cmdq.cur * sizeof (struct iwi_cmd_desc), cur 1242 dev/pci/if_iwi.c DPRINTFN(2, ("sending command idx=%u type=%u len=%u\n", sc->cmdq.cur, cur 1245 dev/pci/if_iwi.c sc->cmdq.cur = (sc->cmdq.cur + 1) % IWI_CMD_RING_COUNT; cur 1249 dev/pci/if_iwi.c sc->cmdq.next = sc->cmdq.cur; cur 1286 dev/pci/if_iwi.c data = &txq->data[txq->cur]; cur 1287 dev/pci/if_iwi.c desc = &txq->desc[txq->cur]; cur 1383 dev/pci/if_iwi.c txq->cur * sizeof (struct iwi_tx_desc), cur 1386 dev/pci/if_iwi.c DPRINTFN(5, ("sending data frame idx=%u len=%u nseg=%u\n", txq->cur, cur 1390 dev/pci/if_iwi.c txq->cur = (txq->cur + 1) % IWI_TX_RING_COUNT; cur 1391 dev/pci/if_iwi.c CSR_WRITE_4(sc, txq->csr_widx, txq->cur); cur 2202 dev/pci/if_iwi.c CSR_WRITE_4(sc, IWI_CSR_CMD_WIDX, sc->cmdq.cur); cur 2206 dev/pci/if_iwi.c CSR_WRITE_4(sc, IWI_CSR_TX1_WIDX, sc->txq[0].cur); cur 2210 dev/pci/if_iwi.c CSR_WRITE_4(sc, IWI_CSR_TX2_WIDX, sc->txq[1].cur); cur 2214 dev/pci/if_iwi.c CSR_WRITE_4(sc, IWI_CSR_TX3_WIDX, sc->txq[2].cur); cur 2218 dev/pci/if_iwi.c CSR_WRITE_4(sc, IWI_CSR_TX4_WIDX, sc->txq[3].cur); cur 64 dev/pci/if_iwivar.h int cur; cur 82 dev/pci/if_iwivar.h int cur; cur 94 dev/pci/if_iwivar.h int cur; cur 1407 dev/pci/if_msk.c u_int32_t frag, cur; cur 1421 dev/pci/if_msk.c cur = frag = *txidx; cur 1460 dev/pci/if_msk.c cur = frag; cur 1464 dev/pci/if_msk.c sc_if->sk_cdata.sk_tx_chain[cur].sk_mbuf = m_head; cur 1467 dev/pci/if_msk.c sc_if->sk_cdata.sk_tx_map[cur] = entry; cur 1468 dev/pci/if_msk.c sc_if->sk_rdata->sk_tx_ring[cur].sk_ctl |= SK_Y2_TXCTL_LASTFRAG; cur 1610 dev/pci/if_msk.c int cur, total_len = len; cur 1615 dev/pci/if_msk.c cur = sc_if->sk_cdata.sk_rx_cons; cur 1620 dev/pci/if_msk.c MSK_CDRXSYNC(sc_if, cur, BUS_DMASYNC_POSTREAD|BUS_DMASYNC_POSTWRITE); cur 1622 dev/pci/if_msk.c cur_rx = &sc_if->sk_cdata.sk_rx_chain[cur]; cur 1635 dev/pci/if_msk.c msk_newbuf(sc_if, cur, m, dmamap); cur 1645 dev/pci/if_msk.c if (msk_newbuf(sc_if, cur, NULL, dmamap) == ENOBUFS) { cur 1649 dev/pci/if_msk.c msk_newbuf(sc_if, cur, m, dmamap); cur 646 dev/pci/if_nfe.c data = &sc->rxq.data[sc->rxq.cur]; cur 649 dev/pci/if_nfe.c desc64 = &sc->rxq.desc64[sc->rxq.cur]; cur 655 dev/pci/if_nfe.c desc32 = &sc->rxq.desc32[sc->rxq.cur]; cur 796 dev/pci/if_nfe.c sc->rxq.cur = (sc->rxq.cur + 1) % NFE_RX_RING_COUNT; cur 809 dev/pci/if_nfe.c while (sc->txq.next != sc->txq.cur) { cur 885 dev/pci/if_nfe.c int error, i, first = sc->txq.cur; cur 915 dev/pci/if_nfe.c data = &sc->txq.data[sc->txq.cur]; cur 918 dev/pci/if_nfe.c desc64 = &sc->txq.desc64[sc->txq.cur]; cur 931 dev/pci/if_nfe.c desc32 = &sc->txq.desc32[sc->txq.cur]; cur 955 dev/pci/if_nfe.c sc->txq.cur = (sc->txq.cur + 1) % NFE_TX_RING_COUNT; cur 991 dev/pci/if_nfe.c int old = sc->txq.cur; cur 1012 dev/pci/if_nfe.c if (sc->txq.cur == old) /* nothing sent */ cur 1016 dev/pci/if_nfe.c nfe_txdesc64_rsync(sc, old, sc->txq.cur, BUS_DMASYNC_PREWRITE); cur 1018 dev/pci/if_nfe.c nfe_txdesc32_rsync(sc, old, sc->txq.cur, BUS_DMASYNC_PREWRITE); cur 1197 dev/pci/if_nfe.c ring->cur = ring->next = 0; cur 1338 dev/pci/if_nfe.c ring->cur = ring->next = 0; cur 1516 dev/pci/if_nfe.c ring->cur = ring->next = 0; cur 1598 dev/pci/if_nfe.c ring->cur = ring->next = 0; cur 35 dev/pci/if_nfevar.h int cur; cur 63 dev/pci/if_nfevar.h int cur; cur 1601 dev/pci/if_nge.c int frag, cur, cnt = 0; cur 1616 dev/pci/if_nge.c cur = frag = *txidx; cur 1630 dev/pci/if_nge.c cur = frag; cur 1643 dev/pci/if_nge.c sc->nge_ldata->nge_tx_list[cur].nge_extsts |= cur 1648 dev/pci/if_nge.c sc->nge_ldata->nge_tx_list[cur].nge_mbuf = m_head; cur 1649 dev/pci/if_nge.c sc->nge_ldata->nge_tx_list[cur].nge_ctl &= ~NGE_CMDSTS_MORE; cur 1546 dev/pci/if_sis.c int frag, cur, i; cur 1559 dev/pci/if_sis.c cur = frag = *txidx; cur 1569 dev/pci/if_sis.c cur = frag; cur 1576 dev/pci/if_sis.c sc->sis_ldata->sis_tx_list[cur].sis_mbuf = m_head; cur 1577 dev/pci/if_sis.c sc->sis_ldata->sis_tx_list[cur].sis_ctl &= ~SIS_CMDSTS_MORE; cur 1521 dev/pci/if_sk.c u_int32_t frag, cur, sk_ctl; cur 1535 dev/pci/if_sk.c cur = frag = *txidx; cur 1574 dev/pci/if_sk.c cur = frag; cur 1578 dev/pci/if_sk.c sc_if->sk_cdata.sk_tx_chain[cur].sk_mbuf = m_head; cur 1581 dev/pci/if_sk.c sc_if->sk_cdata.sk_tx_map[cur] = entry; cur 1582 dev/pci/if_sk.c sc_if->sk_rdata->sk_tx_ring[cur].sk_ctl |= cur 1731 dev/pci/if_sk.c int i, cur, total_len = 0; cur 1741 dev/pci/if_sk.c cur = i; cur 1744 dev/pci/if_sk.c SK_CDRXSYNC(sc_if, cur, cur 1750 dev/pci/if_sk.c SK_CDRXSYNC(sc_if, cur, BUS_DMASYNC_PREREAD); cur 1755 dev/pci/if_sk.c cur_rx = &sc_if->sk_cdata.sk_rx_chain[cur]; cur 1756 dev/pci/if_sk.c cur_desc = &sc_if->sk_rdata->sk_rx_ring[cur]; cur 1779 dev/pci/if_sk.c sk_newbuf(sc_if, cur, m, dmamap); cur 1790 dev/pci/if_sk.c if (sk_newbuf(sc_if, cur, NULL, dmamap) == ENOBUFS) { cur 1794 dev/pci/if_sk.c sk_newbuf(sc_if, cur, m, dmamap); cur 1997 dev/pci/if_ti.c u_int32_t frag, cur, cnt = 0; cur 2015 dev/pci/if_ti.c cur = frag = *txidx; cur 2052 dev/pci/if_ti.c cur = frag; cur 2061 dev/pci/if_ti.c ti_mem_write(sc, TI_TX_RING_BASE + cur * sizeof(txdesc), cur 2067 dev/pci/if_ti.c sc->ti_cdata.ti_tx_chain[cur] = m_head; cur 2069 dev/pci/if_ti.c sc->ti_cdata.ti_tx_map[cur] = entry; cur 2085 dev/pci/if_ti.c u_int32_t frag, cur, cnt = 0; cur 2102 dev/pci/if_ti.c cur = frag = *txidx; cur 2136 dev/pci/if_ti.c cur = frag; cur 2144 dev/pci/if_ti.c sc->ti_rdata->ti_tx_ring[cur].ti_flags |= TI_BDFLAG_END; cur 2149 dev/pci/if_ti.c TI_RING_DMASYNC(sc, ti_tx_ring[cur], BUS_DMASYNC_POSTREAD); cur 2151 dev/pci/if_ti.c sc->ti_cdata.ti_tx_chain[cur] = m_head; cur 2153 dev/pci/if_ti.c sc->ti_cdata.ti_tx_map[cur] = entry; cur 537 dev/pci/if_wpi.c ring->cur = 0; cur 602 dev/pci/if_wpi.c ring->cur = 0; cur 627 dev/pci/if_wpi.c ring->cur = 0; cur 711 dev/pci/if_wpi.c ring->cur = 0; cur 1222 dev/pci/if_wpi.c "chan=%d tstamp=%llu\n", ring->cur, letoh32(desc->len), cur 1254 dev/pci/if_wpi.c ring->desc[ring->cur] = htole32(rbuf->paddr); cur 1390 dev/pci/if_wpi.c while (sc->rxq.cur != hw) { cur 1391 dev/pci/if_wpi.c struct wpi_rx_data *data = &sc->rxq.data[sc->rxq.cur]; cur 1479 dev/pci/if_wpi.c sc->rxq.cur = (sc->rxq.cur + 1) % WPI_RX_RING_COUNT; cur 1569 dev/pci/if_wpi.c desc = &ring->desc[ring->cur]; cur 1570 dev/pci/if_wpi.c data = &ring->data[ring->cur]; cur 1610 dev/pci/if_wpi.c cmd = &ring->cmd[ring->cur]; cur 1614 dev/pci/if_wpi.c cmd->idx = ring->cur; cur 1738 dev/pci/if_wpi.c ring->qid, ring->cur, m0->m_pkthdr.len, data->map->dm_nsegs)); cur 1744 dev/pci/if_wpi.c ring->cur * sizeof (struct wpi_tx_cmd)); cur 1756 dev/pci/if_wpi.c ring->cur = (ring->cur + 1) % WPI_TX_RING_COUNT; cur 1757 dev/pci/if_wpi.c WPI_WRITE(sc, WPI_TX_WIDX, ring->qid << 8 | ring->cur); cur 2033 dev/pci/if_wpi.c desc = &ring->desc[ring->cur]; cur 2034 dev/pci/if_wpi.c cmd = &ring->cmd[ring->cur]; cur 2039 dev/pci/if_wpi.c cmd->idx = ring->cur; cur 2044 dev/pci/if_wpi.c ring->cur * sizeof (struct wpi_tx_cmd)); cur 2048 dev/pci/if_wpi.c ring->cur = (ring->cur + 1) % WPI_CMD_RING_COUNT; cur 2049 dev/pci/if_wpi.c WPI_WRITE(sc, WPI_TX_WIDX, ring->qid << 8 | ring->cur); cur 2285 dev/pci/if_wpi.c desc = &ring->desc[ring->cur]; cur 2286 dev/pci/if_wpi.c data = &ring->data[ring->cur]; cur 2295 dev/pci/if_wpi.c cmd = &ring->cmd[ring->cur]; cur 2299 dev/pci/if_wpi.c cmd->idx = ring->cur; cur 2330 dev/pci/if_wpi.c ring->cur * sizeof (struct wpi_tx_cmd)); cur 2336 dev/pci/if_wpi.c ring->cur = (ring->cur + 1) % WPI_CMD_RING_COUNT; cur 2337 dev/pci/if_wpi.c WPI_WRITE(sc, WPI_TX_WIDX, ring->qid << 8 | ring->cur); cur 2428 dev/pci/if_wpi.c desc = &ring->desc[ring->cur]; cur 2429 dev/pci/if_wpi.c data = &ring->data[ring->cur]; cur 2450 dev/pci/if_wpi.c cmd->idx = ring->cur; cur 2560 dev/pci/if_wpi.c ring->cur = (ring->cur + 1) % WPI_CMD_RING_COUNT; cur 2561 dev/pci/if_wpi.c WPI_WRITE(sc, WPI_TX_WIDX, ring->qid << 8 | ring->cur); cur 80 dev/pci/if_wpivar.h int cur; cur 105 dev/pci/if_wpivar.h int cur; cur 989 dev/pci/musycc.c struct dma_desc *cur, *tmp; cur 1005 dev/pci/musycc.c cur = mg->mg_dma_d[c].tx_cur; cur 1018 dev/pci/musycc.c if (cur != mg->mg_dma_d[c].tx_cur) cur 1021 dev/pci/musycc.c cur->status = htole32(status); cur 1022 dev/pci/musycc.c cur->data = htole32(map->dm_segs[i].ds_addr); cur 1025 dev/pci/musycc.c ((caddr_t)cur - mg->mg_listkva), sizeof(struct dma_desc), cur 1030 dev/pci/musycc.c cur = cur->nextdesc; cur 1036 dev/pci/musycc.c cur->mbuf = m_head; cur 1037 dev/pci/musycc.c mg->mg_tx_sparemap = cur->map; cur 1038 dev/pci/musycc.c cur->map = map; cur 1039 dev/pci/musycc.c cur->status |= htole32(MUSYCC_STATUS_EOM); cur 1041 dev/pci/musycc.c mg->mg_dma_d[c].tx_cur = cur->nextdesc; cur 2231 dev/usb/ehci.c ehci_soft_qtd_t *next, *cur; cur 2259 dev/usb/ehci.c cur = ehci_alloc_sqtd(sc); cur 2260 dev/usb/ehci.c *sp = cur; cur 2261 dev/usb/ehci.c if (cur == NULL) cur 2318 dev/usb/ehci.c cur->qtd.qtd_buffer[i] = htole32(a); cur 2319 dev/usb/ehci.c cur->qtd.qtd_buffer_hi[i] = 0; cur 2327 dev/usb/ehci.c cur->nextqtd = next; cur 2328 dev/usb/ehci.c cur->qtd.qtd_next = cur->qtd.qtd_altnext = nextphys; cur 2329 dev/usb/ehci.c cur->qtd.qtd_status = htole32(qtdstatus | cur 2331 dev/usb/ehci.c cur->xfer = xfer; cur 2332 dev/usb/ehci.c cur->len = curlen; cur 2351 dev/usb/ehci.c cur = next; cur 2353 dev/usb/ehci.c cur->qtd.qtd_status |= htole32(EHCI_QTD_IOC); cur 2354 dev/usb/ehci.c *ep = cur; cur 2423 dev/usb/ehci.c ehci_physaddr_t cur, us, next; cur 2525 dev/usb/ehci.c cur = EHCI_LINK_ADDR(letoh32(sqh->qh.qh_curqtd)); cur 2542 dev/usb/ehci.c hit |= (cur == sqtd->physaddr); cur 2567 dev/usb/ehci.c if (cur == sqtd->physaddr) { cur 67 dev/usb/hid.c struct hid_item cur; cur 113 dev/usb/hid.c while (s->cur.next != NULL) { cur 114 dev/usb/hid.c struct hid_item *hi = s->cur.next->next; cur 115 dev/usb/hid.c free(s->cur.next, M_TEMP); cur 116 dev/usb/hid.c s->cur.next = hi; cur 124 dev/usb/hid.c struct hid_item *c = &s->cur; cur 302 dev/usb/hid.c *hi = s->cur; cur 308 dev/usb/hid.c s->cur = *hi; cur 448 dev/usb/ohci.c ohci_soft_td_t *next, *cur; cur 458 dev/usb/ohci.c cur = sp; cur 493 dev/usb/ohci.c cur->td.td_flags = tdflags; cur 494 dev/usb/ohci.c cur->td.td_cbp = htole32(dataphys); cur 495 dev/usb/ohci.c cur->nexttd = next; cur 496 dev/usb/ohci.c cur->td.td_nexttd = htole32(next->physaddr); cur 497 dev/usb/ohci.c cur->td.td_be = htole32(dataphys + curlen - 1); cur 498 dev/usb/ohci.c cur->len = curlen; cur 499 dev/usb/ohci.c cur->flags = OHCI_ADD_LEN; cur 500 dev/usb/ohci.c cur->xfer = xfer; cur 507 dev/usb/ohci.c cur = next; cur 513 dev/usb/ohci.c cur = next; cur 518 dev/usb/ohci.c cur->td.td_flags = tdflags; cur 519 dev/usb/ohci.c cur->td.td_cbp = 0; /* indicate 0 length packet */ cur 520 dev/usb/ohci.c cur->nexttd = next; cur 521 dev/usb/ohci.c cur->td.td_nexttd = htole32(next->physaddr); cur 522 dev/usb/ohci.c cur->td.td_be = ~0; cur 523 dev/usb/ohci.c cur->len = 0; cur 524 dev/usb/ohci.c cur->flags = 0; cur 525 dev/usb/ohci.c cur->xfer = xfer; cur 528 dev/usb/ohci.c *ep = cur; cur 139 dev/usb/uaudio.c u_char *cur; /* current position in upper layer buffer */ cur 2663 dev/usb/uaudio.c n = min(total, ch->end - ch->cur); cur 2664 dev/usb/uaudio.c memcpy(cb->buffer, ch->cur, n); cur 2665 dev/usb/uaudio.c ch->cur += n; cur 2666 dev/usb/uaudio.c if (ch->cur >= ch->end) cur 2667 dev/usb/uaudio.c ch->cur = ch->start; cur 2670 dev/usb/uaudio.c memcpy(cb->buffer + n, ch->cur, total); cur 2671 dev/usb/uaudio.c ch->cur += total; cur 2808 dev/usb/uaudio.c n = min(frsize, ch->end - ch->cur); cur 2809 dev/usb/uaudio.c memcpy(ch->cur, cb->buffer + cb->offsets[i], n); cur 2810 dev/usb/uaudio.c ch->cur += n; cur 2811 dev/usb/uaudio.c if (ch->cur >= ch->end) cur 2812 dev/usb/uaudio.c ch->cur = ch->start; cur 2814 dev/usb/uaudio.c memcpy(ch->cur, cb->buffer + cb->offsets[i] + n, cur 2816 dev/usb/uaudio.c ch->cur += frsize - n; cur 2861 dev/usb/uaudio.c ch->cur = start; cur 92 dev/usb/ugen.c u_char *cur; /* current read location (isoc) */ cur 375 dev/usb/ugen.c sce->cur = sce->fill = sce->ibuf; cur 590 dev/usb/ugen.c while (sce->cur == sce->fill) { cur 607 dev/usb/ugen.c while (sce->cur != sce->fill && uio->uio_resid > 0 && !error) { cur 608 dev/usb/ugen.c if(sce->fill > sce->cur) cur 609 dev/usb/ugen.c n = min(sce->fill - sce->cur, uio->uio_resid); cur 611 dev/usb/ugen.c n = min(sce->limit - sce->cur, uio->uio_resid); cur 616 dev/usb/ugen.c error = uiomove(sce->cur, n, uio); cur 619 dev/usb/ugen.c sce->cur += n; cur 620 dev/usb/ugen.c if(sce->cur >= sce->limit) cur 621 dev/usb/ugen.c sce->cur = sce->ibuf; cur 864 dev/usb/ugen.c if(sce->fill < sce->cur && sce->cur <= sce->fill + count) { cur 865 dev/usb/ugen.c sce->cur += count; cur 866 dev/usb/ugen.c if(sce->cur >= sce->limit) cur 867 dev/usb/ugen.c sce->cur = sce->ibuf + (sce->limit - sce->cur); cur 1333 dev/usb/ugen.c if (sce->cur != sce->fill) cur 1385 dev/usb/ugen.c if (sce->cur == sce->fill) cur 1388 dev/usb/ugen.c if (sce->cur < sce->fill) cur 1389 dev/usb/ugen.c kn->kn_data = sce->fill - sce->cur; cur 1391 dev/usb/ugen.c kn->kn_data = (sce->limit - sce->cur) + cur 1112 dev/usb/usbdi.c iter->cur = (const uByte *)cd; cur 1121 dev/usb/usbdi.c if (iter->cur + sizeof(usb_descriptor_t) >= iter->end) { cur 1122 dev/usb/usbdi.c if (iter->cur != iter->end) cur 1126 dev/usb/usbdi.c desc = (const usb_descriptor_t *)iter->cur; cur 1131 dev/usb/usbdi.c iter->cur += desc->bLength; cur 1132 dev/usb/usbdi.c if (iter->cur > iter->end) { cur 177 dev/usb/usbdi.h const uByte *cur; cur 339 dev/wscons/wskbdutil.c kbd_t cur; cur 351 dev/wscons/wskbdutil.c for (cur = mapdata->layout & ~KB_HANDLEDBYWSKBD; cur != 0; ) { cur 354 dev/wscons/wskbdutil.c if (mp->name == cur) cur 387 dev/wscons/wskbdutil.c cur = mp->base; cur 425 dev/wscons/wskbdutil.c kbd_t cur; cur 428 dev/wscons/wskbdutil.c for (cur = mapdata->layout & ~KB_HANDLEDBYWSKBD, stack_ptr = 0; cur 429 dev/wscons/wskbdutil.c cur != 0; stack_ptr++) { cur 432 dev/wscons/wskbdutil.c if (cur == 0 || mp->name == cur) { cur 445 dev/wscons/wskbdutil.c cur = mp->base; cur 810 net/pf.c struct pf_state_key *cur; cur 816 net/pf.c if ((cur = RB_INSERT(pf_state_tree_lan_ext, &pf_statetbl_lan_ext, cur 819 net/pf.c TAILQ_FOREACH(sp, &cur->states, next) cur 825 net/pf.c pf_attach_state(cur, s, kif == pfi_all ? 1 : 0); cur 829 net/pf.c if (cur == NULL && (cur = RB_INSERT(pf_state_tree_ext_gwy, cur 928 net/pf.c struct pf_src_node *cur, *next; cur 931 net/pf.c for (cur = RB_MIN(pf_src_tree, &tree_src_tracking); cur; cur = next) { cur 932 net/pf.c next = RB_NEXT(pf_src_tree, &tree_src_tracking, cur); cur 934 net/pf.c if (cur->states <= 0 && cur->expire <= time_second) { cur 938 net/pf.c &tree_src_tracking, cur); cur 941 net/pf.c if (cur->rule.ptr != NULL) { cur 942 net/pf.c cur->rule.ptr->src_nodes--; cur 943 net/pf.c if (cur->rule.ptr->states <= 0 && cur 944 net/pf.c cur->rule.ptr->max_src_nodes <= 0) cur 945 net/pf.c pf_rm_rule(NULL, cur->rule.ptr); cur 947 net/pf.c RB_REMOVE(pf_src_tree, &tree_src_tracking, cur); cur 950 net/pf.c pool_put(&pf_src_tree_pl, cur); cur 990 net/pf.c pf_unlink_state(struct pf_state *cur) cur 992 net/pf.c if (cur->src.state == PF_TCPS_PROXY_DST) { cur 993 net/pf.c pf_send_tcp(cur->rule.ptr, cur->state_key->af, cur 994 net/pf.c &cur->state_key->ext.addr, &cur->state_key->lan.addr, cur 995 net/pf.c cur->state_key->ext.port, cur->state_key->lan.port, cur 996 net/pf.c cur->src.seqhi, cur->src.seqlo + 1, cur 997 net/pf.c TH_RST|TH_ACK, 0, 0, 0, 1, cur->tag, NULL, NULL); cur 999 net/pf.c RB_REMOVE(pf_state_tree_id, &tree_id, cur); cur 1001 net/pf.c if (cur->creatorid == pf_status.hostid) cur 1002 net/pf.c pfsync_delete_state(cur); cur 1004 net/pf.c cur->timeout = PFTM_UNLINKED; cur 1005 net/pf.c pf_src_tree_remove_state(cur); cur 1006 net/pf.c pf_detach_state(cur, 0); cur 1012 net/pf.c pf_free_state(struct pf_state *cur) cur 1016 net/pf.c (pfsyncif->sc_bulk_send_next == cur || cur 1017 net/pf.c pfsyncif->sc_bulk_terminator == cur)) cur 1020 net/pf.c KASSERT(cur->timeout == PFTM_UNLINKED); cur 1021 net/pf.c if (--cur->rule.ptr->states <= 0 && cur 1022 net/pf.c cur->rule.ptr->src_nodes <= 0) cur 1023 net/pf.c pf_rm_rule(NULL, cur->rule.ptr); cur 1024 net/pf.c if (cur->nat_rule.ptr != NULL) cur 1025 net/pf.c if (--cur->nat_rule.ptr->states <= 0 && cur 1026 net/pf.c cur->nat_rule.ptr->src_nodes <= 0) cur 1027 net/pf.c pf_rm_rule(NULL, cur->nat_rule.ptr); cur 1028 net/pf.c if (cur->anchor.ptr != NULL) cur 1029 net/pf.c if (--cur->anchor.ptr->states <= 0) cur 1030 net/pf.c pf_rm_rule(NULL, cur->anchor.ptr); cur 1031 net/pf.c pf_normalize_tcp_cleanup(cur); cur 1032 net/pf.c pfi_kif_unref(cur->kif, PFI_KIF_REF_STATE); cur 1033 net/pf.c TAILQ_REMOVE(&state_list, cur, entry_list); cur 1034 net/pf.c if (cur->tag) cur 1035 net/pf.c pf_tag_unref(cur->tag); cur 1036 net/pf.c pool_put(&pf_state_pl, cur); cur 1044 net/pf.c static struct pf_state *cur = NULL; cur 1050 net/pf.c if (cur == NULL) { cur 1051 net/pf.c cur = TAILQ_FIRST(&state_list); cur 1052 net/pf.c if (cur == NULL) cur 1057 net/pf.c next = TAILQ_NEXT(cur, entry_list); cur 1059 net/pf.c if (cur->timeout == PFTM_UNLINKED) { cur 1065 net/pf.c pf_free_state(cur); cur 1066 net/pf.c } else if (pf_state_expires(cur) <= time_second) { cur 1068 net/pf.c pf_unlink_state(cur); cur 1073 net/pf.c pf_free_state(cur); cur 1075 net/pf.c cur = next; cur 1243 net/pf.c while (head[i] != cur) { \ cur 1244 net/pf.c head[i]->skip[i].ptr = cur; \ cur 1252 net/pf.c struct pf_rule *cur, *prev, *head[PF_SKIP_COUNT]; cur 1255 net/pf.c cur = TAILQ_FIRST(rules); cur 1256 net/pf.c prev = cur; cur 1258 net/pf.c head[i] = cur; cur 1259 net/pf.c while (cur != NULL) { cur 1261 net/pf.c if (cur->kif != prev->kif || cur->ifnot != prev->ifnot) cur 1263 net/pf.c if (cur->direction != prev->direction) cur 1265 net/pf.c if (cur->af != prev->af) cur 1267 net/pf.c if (cur->proto != prev->proto) cur 1269 net/pf.c if (cur->src.neg != prev->src.neg || cur 1270 net/pf.c pf_addr_wrap_neq(&cur->src.addr, &prev->src.addr)) cur 1272 net/pf.c if (cur->src.port[0] != prev->src.port[0] || cur 1273 net/pf.c cur->src.port[1] != prev->src.port[1] || cur 1274 net/pf.c cur->src.port_op != prev->src.port_op) cur 1276 net/pf.c if (cur->dst.neg != prev->dst.neg || cur 1277 net/pf.c pf_addr_wrap_neq(&cur->dst.addr, &prev->dst.addr)) cur 1279 net/pf.c if (cur->dst.port[0] != prev->dst.port[0] || cur 1280 net/pf.c cur->dst.port[1] != prev->dst.port[1] || cur 1281 net/pf.c cur->dst.port_op != prev->dst.port_op) cur 1284 net/pf.c prev = cur; cur 1285 net/pf.c cur = TAILQ_NEXT(cur, entries); cur 2062 net/pf.c struct pf_addr *raddr = &rpool->cur->addr.v.a.addr; cur 2063 net/pf.c struct pf_addr *rmask = &rpool->cur->addr.v.a.mask; cur 2064 net/pf.c struct pf_pooladdr *acur = rpool->cur; cur 2091 net/pf.c if (rpool->cur->addr.type == PF_ADDR_NOROUTE) cur 2093 net/pf.c if (rpool->cur->addr.type == PF_ADDR_DYNIFTL) { cur 2097 net/pf.c if (rpool->cur->addr.p.dyn->pfid_acnt4 < 1 && cur 2101 net/pf.c raddr = &rpool->cur->addr.p.dyn->pfid_addr4; cur 2102 net/pf.c rmask = &rpool->cur->addr.p.dyn->pfid_mask4; cur 2107 net/pf.c if (rpool->cur->addr.p.dyn->pfid_acnt6 < 1 && cur 2111 net/pf.c raddr = &rpool->cur->addr.p.dyn->pfid_addr6; cur 2112 net/pf.c rmask = &rpool->cur->addr.p.dyn->pfid_mask6; cur 2116 net/pf.c } else if (rpool->cur->addr.type == PF_ADDR_TABLE) { cur 2120 net/pf.c raddr = &rpool->cur->addr.v.a.addr; cur 2121 net/pf.c rmask = &rpool->cur->addr.v.a.mask; cur 2175 net/pf.c if (rpool->cur->addr.type == PF_ADDR_TABLE) { cur 2176 net/pf.c if (!pfr_pool_get(rpool->cur->addr.p.tbl, cur 2180 net/pf.c } else if (rpool->cur->addr.type == PF_ADDR_DYNIFTL) { cur 2181 net/pf.c if (!pfr_pool_get(rpool->cur->addr.p.dyn->pfid_kt, cur 2189 net/pf.c if ((rpool->cur = TAILQ_NEXT(rpool->cur, entries)) == NULL) cur 2190 net/pf.c rpool->cur = TAILQ_FIRST(&rpool->list); cur 2191 net/pf.c if (rpool->cur->addr.type == PF_ADDR_TABLE) { cur 2193 net/pf.c if (pfr_pool_get(rpool->cur->addr.p.tbl, cur 2197 net/pf.c if (rpool->cur != acur) cur 2201 net/pf.c } else if (rpool->cur->addr.type == PF_ADDR_DYNIFTL) { cur 2203 net/pf.c if (pfr_pool_get(rpool->cur->addr.p.dyn->pfid_kt, cur 2207 net/pf.c if (rpool->cur != acur) cur 2212 net/pf.c raddr = &rpool->cur->addr.v.a.addr; cur 2213 net/pf.c rmask = &rpool->cur->addr.v.a.mask; cur 2346 net/pf.c if (r->rpool.cur != NULL) cur 2347 net/pf.c xdst = &r->rpool.cur->addr; cur 2453 net/pf.c if (r->rpool.cur->addr.type == PF_ADDR_DYNIFTL){ cur 2457 net/pf.c if (r->rpool.cur->addr.p.dyn-> cur 2461 net/pf.c &r->rpool.cur->addr.p.dyn-> cur 2463 net/pf.c &r->rpool.cur->addr.p.dyn-> cur 2470 net/pf.c if (r->rpool.cur->addr.p.dyn-> cur 2474 net/pf.c &r->rpool.cur->addr.p.dyn-> cur 2476 net/pf.c &r->rpool.cur->addr.p.dyn-> cur 2484 net/pf.c &r->rpool.cur->addr.v.a.addr, cur 2485 net/pf.c &r->rpool.cur->addr.v.a.mask, cur 2532 net/pf.c &r->rpool.cur->addr.v.a.mask, daddr, cur 2778 net/pf.c s->rt_kif = r->rpool.cur->kif; cur 2785 net/pf.c s->rt_kif = r->rpool.cur->kif; cur 5014 net/pf.c ifp = r->rpool.cur->kif ? cur 5015 net/pf.c r->rpool.cur->kif->pfik_ifp : NULL; cur 5199 net/pf.c ifp = r->rpool.cur->kif ? r->rpool.cur->kif->pfik_ifp : NULL; cur 1252 net/pf_ioctl.c rule->rpool.cur = TAILQ_FIRST(&rule->rpool.list); cur 1499 net/pf_ioctl.c newrule->rpool.cur = TAILQ_FIRST(&newrule->rpool.list); cur 2264 net/pf_ioctl.c pool->cur = TAILQ_FIRST(&pool->list); cur 2265 net/pf_ioctl.c PF_ACPY(&pool->counter, &pool->cur->addr.v.a.addr, cur 524 net/pf_norm.c struct pf_frcache *frp, *fra, *cur = NULL; cur 543 net/pf_norm.c cur = pool_get(&pf_cent_pl, PR_NOWAIT); cur 544 net/pf_norm.c if (cur == NULL) { cur 559 net/pf_norm.c cur->fr_off = off; cur 560 net/pf_norm.c cur->fr_end = max; cur 562 net/pf_norm.c LIST_INSERT_HEAD(&(*frag)->fr_cache, cur, fr_next); cur 659 net/pf_norm.c cur = pool_get(&pf_cent_pl, PR_NOWAIT); cur 660 net/pf_norm.c if (cur == NULL) cur 664 net/pf_norm.c cur->fr_off = off; cur 665 net/pf_norm.c cur->fr_end = max; cur 666 net/pf_norm.c LIST_INSERT_AFTER(frp, cur, fr_next); cur 713 net/pf_norm.c cur = pool_get(&pf_cent_pl, PR_NOWAIT); cur 714 net/pf_norm.c if (cur == NULL) cur 718 net/pf_norm.c cur->fr_off = off; cur 719 net/pf_norm.c cur->fr_end = max; cur 720 net/pf_norm.c LIST_INSERT_BEFORE(fra, cur, fr_next); cur 726 net/pf_norm.c if (cur && fra->fr_off <= cur->fr_end) { cur 730 net/pf_norm.c h->ip_id, cur->fr_off, cur->fr_end, off, cur 732 net/pf_norm.c fra->fr_off = cur->fr_off; cur 733 net/pf_norm.c LIST_REMOVE(cur, fr_next); cur 734 net/pf_norm.c pool_put(&pf_cent_pl, cur); cur 736 net/pf_norm.c cur = NULL; cur 740 net/pf_norm.c KASSERT(cur == NULL); cur 379 net/pfvar.h struct pf_pooladdr *cur; cur 245 netinet/ip_carp.c struct in_addr last, cur, in; cur 277 netinet/ip_carp.c cur.s_addr = 0; cur 280 netinet/ip_carp.c last = cur; cur 281 netinet/ip_carp.c cur.s_addr = 0xffffffff; cur 286 netinet/ip_carp.c ntohl(in.s_addr) < ntohl(cur.s_addr)) { cur 287 netinet/ip_carp.c cur.s_addr = in.s_addr; cur 293 netinet/ip_carp.c (void *)&cur, sizeof(cur)); cur 1332 netinet/ip_carp.c int cur, last, count, found; cur 1367 netinet/ip_carp.c cur = 0; cur 1372 netinet/ip_carp.c last = cur; cur 1373 netinet/ip_carp.c cur = 255; cur 1391 netinet/ip_carp.c if (ifa && vh->sc_vhid > last && vh->sc_vhid < cur) { cur 1392 netinet/ip_carp.c cur = vh->sc_vhid; cur 2537 netinet/tcp_input.c struct sackhole *cur, *p, *temp; cur 2591 netinet/tcp_input.c cur = tp->snd_holes; cur 2592 netinet/tcp_input.c cur->start = th->th_ack; cur 2593 netinet/tcp_input.c cur->end = sack.start; cur 2594 netinet/tcp_input.c cur->rxmit = cur->start; cur 2595 netinet/tcp_input.c cur->next = NULL; cur 2602 netinet/tcp_input.c cur->dups = min(tcprexmtthresh, cur 2603 netinet/tcp_input.c ((sack.end - cur->end)/tp->t_maxseg)); cur 2604 netinet/tcp_input.c if (cur->dups < 1) cur 2605 netinet/tcp_input.c cur->dups = 1; cur 2609 netinet/tcp_input.c p = cur = tp->snd_holes; cur 2610 netinet/tcp_input.c while (cur) { cur 2611 netinet/tcp_input.c if (SEQ_LEQ(sack.end, cur->start)) cur 2614 netinet/tcp_input.c if (SEQ_GEQ(sack.start, cur->end)) { cur 2616 netinet/tcp_input.c cur->dups++; cur 2617 netinet/tcp_input.c if (((sack.end - cur->end)/tp->t_maxseg) >= cur 2619 netinet/tcp_input.c cur->dups = tcprexmtthresh; cur 2620 netinet/tcp_input.c p = cur; cur 2621 netinet/tcp_input.c cur = cur->next; cur 2624 netinet/tcp_input.c if (SEQ_LEQ(sack.start, cur->start)) { cur 2627 netinet/tcp_input.c if (SEQ_GT(sack.end, cur->rxmit)) cur 2629 netinet/tcp_input.c tcp_seq_subtract(cur->rxmit, cur 2630 netinet/tcp_input.c cur->start); cur 2634 netinet/tcp_input.c cur->start); cur 2636 netinet/tcp_input.c if (SEQ_GEQ(sack.end, cur->end)) { cur 2638 netinet/tcp_input.c if (p != cur) { cur 2639 netinet/tcp_input.c p->next = cur->next; cur 2640 netinet/tcp_input.c pool_put(&sackhl_pool, cur); cur 2641 netinet/tcp_input.c cur = p->next; cur 2643 netinet/tcp_input.c cur = cur->next; cur 2645 netinet/tcp_input.c p = cur; cur 2652 netinet/tcp_input.c cur->start = sack.end; cur 2653 netinet/tcp_input.c cur->rxmit = SEQ_MAX(cur->rxmit, cur->start); cur 2654 netinet/tcp_input.c p = cur; cur 2655 netinet/tcp_input.c cur = cur->next; cur 2659 netinet/tcp_input.c if (SEQ_GEQ(sack.end, cur->end)) { cur 2661 netinet/tcp_input.c if (SEQ_GT(cur->rxmit, sack.start)) cur 2663 netinet/tcp_input.c tcp_seq_subtract(cur->rxmit, cur 2666 netinet/tcp_input.c cur->end = sack.start; cur 2667 netinet/tcp_input.c cur->rxmit = SEQ_MIN(cur->rxmit, cur->end); cur 2668 netinet/tcp_input.c cur->dups++; cur 2669 netinet/tcp_input.c if (((sack.end - cur->end)/tp->t_maxseg) >= cur 2671 netinet/tcp_input.c cur->dups = tcprexmtthresh; cur 2672 netinet/tcp_input.c p = cur; cur 2673 netinet/tcp_input.c cur = cur->next; cur 2676 netinet/tcp_input.c if (SEQ_LT(cur->start, sack.start) && cur 2677 netinet/tcp_input.c SEQ_GT(cur->end, sack.end)) { cur 2687 netinet/tcp_input.c if (SEQ_GT(cur->rxmit, sack.end)) cur 2691 netinet/tcp_input.c else if (SEQ_GT(cur->rxmit, sack.start)) cur 2693 netinet/tcp_input.c tcp_seq_subtract(cur->rxmit, cur 2696 netinet/tcp_input.c temp->next = cur->next; cur 2698 netinet/tcp_input.c temp->end = cur->end; cur 2699 netinet/tcp_input.c temp->dups = cur->dups; cur 2700 netinet/tcp_input.c temp->rxmit = SEQ_MAX(cur->rxmit, temp->start); cur 2701 netinet/tcp_input.c cur->end = sack.start; cur 2702 netinet/tcp_input.c cur->rxmit = SEQ_MIN(cur->rxmit, cur->end); cur 2703 netinet/tcp_input.c cur->dups++; cur 2704 netinet/tcp_input.c if (((sack.end - cur->end)/tp->t_maxseg) >= cur 2706 netinet/tcp_input.c cur->dups = tcprexmtthresh; cur 2707 netinet/tcp_input.c cur->next = temp; cur 2709 netinet/tcp_input.c cur = p->next; cur 2743 netinet/tcp_input.c cur = tp->snd_holes; cur 2744 netinet/tcp_input.c while (cur) { cur 2745 netinet/tcp_input.c tp->retran_data += cur->rxmit - cur->start; cur 2746 netinet/tcp_input.c cur = cur->next; cur 2769 netinet/tcp_input.c struct sackhole *cur = tp->snd_holes; cur 2771 netinet/tcp_input.c while (cur) cur 2772 netinet/tcp_input.c if (SEQ_LEQ(cur->end, lastack)) { cur 2773 netinet/tcp_input.c prev = cur; cur 2774 netinet/tcp_input.c cur = cur->next; cur 2777 netinet/tcp_input.c } else if (SEQ_LT(cur->start, lastack)) { cur 2778 netinet/tcp_input.c cur->start = lastack; cur 2779 netinet/tcp_input.c if (SEQ_LT(cur->rxmit, cur->start)) cur 2780 netinet/tcp_input.c cur->rxmit = cur->start; cur 2784 netinet/tcp_input.c tp->snd_holes = cur; cur 177 netinet/tcp_output.c struct sackhole *cur = tp->snd_holes; cur 178 netinet/tcp_output.c if (cur == NULL) cur 187 netinet/tcp_output.c while (cur->next) { cur 188 netinet/tcp_output.c if (SEQ_LT(tp->snd_nxt, cur->end)) cur 190 netinet/tcp_output.c if (SEQ_GEQ(tp->snd_nxt, cur->next->start)) cur 191 netinet/tcp_output.c cur = cur->next; cur 193 netinet/tcp_output.c tp->snd_nxt = cur->next->start; cur 197 netinet/tcp_output.c if (SEQ_LT(tp->snd_nxt, cur->end)) cur 955 uvm/uvm_map.c struct vm_map_entry *cur; cur 970 uvm/uvm_map.c cur = map->hint; cur 973 uvm/uvm_map.c if (cur == &map->header) cur 974 uvm/uvm_map.c cur = cur->next; cur 977 uvm/uvm_map.c if (address >= cur->start) { cur 990 uvm/uvm_map.c if ((cur != last) && (cur->end > address)) { cur 992 uvm/uvm_map.c *entry = cur; cur 994 uvm/uvm_map.c cur, 0, 0, 0); cur 1004 uvm/uvm_map.c last = cur->next; cur 1005 uvm/uvm_map.c cur = map->header.next; cur 1013 uvm/uvm_map.c cur = RB_ROOT(&map->rbhead); cur 1019 uvm/uvm_map.c while (cur) { cur 1020 uvm/uvm_map.c if (address >= cur->start) { cur 1021 uvm/uvm_map.c if (address < cur->end) { cur 1022 uvm/uvm_map.c *entry = cur; cur 1023 uvm/uvm_map.c SAVE_HINT(map, map->hint, cur); cur 1026 uvm/uvm_map.c prev = cur; cur 1027 uvm/uvm_map.c cur = RB_RIGHT(cur, rb_entry); cur 1029 uvm/uvm_map.c cur = RB_LEFT(cur, rb_entry); cur 1040 uvm/uvm_map.c while (cur != last) { cur 1041 uvm/uvm_map.c if (cur->end > address) { cur 1042 uvm/uvm_map.c if (address >= cur->start) { cur 1048 uvm/uvm_map.c *entry = cur; cur 1049 uvm/uvm_map.c SAVE_HINT(map, map->hint, cur); cur 1051 uvm/uvm_map.c cur, 0, 0, 0); cur 1056 uvm/uvm_map.c cur = cur->next; cur 1059 uvm/uvm_map.c *entry = cur->prev; cur 1679 uvm/uvm_map.c vaddr_t cur = start; cur 1683 uvm/uvm_map.c if (tmpent->start < cur) cur 1690 uvm/uvm_map.c cur = tmpent->end; cur 102 uvm/uvm_stat.c int cur[MAXHISTS]; cur 110 uvm/uvm_stat.c cur[lcv] = hists[lcv]->f; cur 125 uvm/uvm_stat.c if (cur[lcv] == -1) cur 132 uvm/uvm_stat.c if (hists[lcv]->e[cur[lcv]].fmt == NULL) { cur 133 uvm/uvm_stat.c cur[lcv] = (cur[lcv] + 1) % (hists[lcv]->n); cur 134 uvm/uvm_stat.c if (cur[lcv] == hists[lcv]->f) cur 135 uvm/uvm_stat.c cur[lcv] = -1; cur 145 uvm/uvm_stat.c timercmp(&hists[lcv]->e[cur[lcv]].tv, &tv, <)) { cur 146 uvm/uvm_stat.c tv = hists[lcv]->e[cur[lcv]].tv; cur 156 uvm/uvm_stat.c uvmhist_print(&hists[hi]->e[cur[hi]]); cur 157 uvm/uvm_stat.c cur[hi] = (cur[hi] + 1) % (hists[hi]->n); cur 158 uvm/uvm_stat.c if (cur[hi] == hists[hi]->f) cur 159 uvm/uvm_stat.c cur[hi] = -1;