me 3195 dev/pci/if_de.c struct mbuf *ms = NULL, *me = NULL;
me 3221 dev/pci/if_de.c me = ms;
me 3260 dev/pci/if_de.c for (me = ms; total_len > 0; total_len--) {
me 3261 dev/pci/if_de.c map = TULIP_GETCTX(me, bus_dmamap_t);
me 3266 dev/pci/if_de.c TULIP_SETCTX(me, NULL);
me 3268 dev/pci/if_de.c me->m_len = TULIP_RX_BUFLEN;
me 3270 dev/pci/if_de.c IF_DEQUEUE(&sc->tulip_rxq, me->m_next);
me 3271 dev/pci/if_de.c me = me->m_next;
me 3281 dev/pci/if_de.c me->m_len = total_len - last_offset;
me 3283 dev/pci/if_de.c map = TULIP_GETCTX(me, bus_dmamap_t);
me 3284 dev/pci/if_de.c bus_dmamap_sync(sc->tulip_dmatag, map, 0, me->m_len,
me 3289 dev/pci/if_de.c TULIP_SETCTX(me, NULL);
me 3294 dev/pci/if_de.c if (me == ms) {
me 3333 dev/pci/if_de.c map = TULIP_GETCTX(me, bus_dmamap_t);
me 3337 dev/pci/if_de.c TULIP_SETCTX(me, NULL);
me 3448 dev/pci/if_de.c me = ms->m_next;
me 3451 dev/pci/if_de.c } while ((ms = me) != NULL);
me 508 dev/pci/if_lmc.c struct mbuf *ms = NULL, *me = NULL;
me 533 dev/pci/if_lmc.c me = ms;
me 574 dev/pci/if_lmc.c for (me = ms; total_len > 0; total_len--) {
me 575 dev/pci/if_lmc.c map = LMC_GETCTX(me, bus_dmamap_t);
me 580 dev/pci/if_lmc.c LMC_SETCTX(me, NULL);
me 582 dev/pci/if_lmc.c me->m_len = LMC_RX_BUFLEN;
me 584 dev/pci/if_lmc.c IF_DEQUEUE(&sc->lmc_rxq, me->m_next);
me 585 dev/pci/if_lmc.c me = me->m_next;
me 606 dev/pci/if_lmc.c map = LMC_GETCTX(me, bus_dmamap_t);
me 607 dev/pci/if_lmc.c bus_dmamap_sync(sc->lmc_dmatag, map, 0, me->m_len,
me 612 dev/pci/if_lmc.c LMC_SETCTX(me, NULL);
me 615 dev/pci/if_lmc.c me->m_len = total_len - last_offset;
me 618 dev/pci/if_lmc.c if (me == ms)
me 632 dev/pci/if_lmc.c map = LMC_GETCTX(me, bus_dmamap_t);
me 636 dev/pci/if_lmc.c LMC_SETCTX(me, NULL);
me 738 dev/pci/if_lmc.c me = ms->m_next;
me 741 dev/pci/if_lmc.c } while ((ms = me) != NULL);
me 1377 dev/pci/ubsec.c struct ubsec_q2_modexp *me = (struct ubsec_q2_modexp *)q;
me 1380 dev/pci/ubsec.c krp = me->me_krp;
me 1381 dev/pci/ubsec.c rlen = (me->me_modbits + 7) / 8;
me 1384 dev/pci/ubsec.c bus_dmamap_sync(sc->sc_dmat, me->me_M.dma_map,
me 1385 dev/pci/ubsec.c 0, me->me_M.dma_map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
me 1386 dev/pci/ubsec.c bus_dmamap_sync(sc->sc_dmat, me->me_E.dma_map,
me 1387 dev/pci/ubsec.c 0, me->me_E.dma_map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
me 1388 dev/pci/ubsec.c bus_dmamap_sync(sc->sc_dmat, me->me_C.dma_map,
me 1389 dev/pci/ubsec.c 0, me->me_C.dma_map->dm_mapsize, BUS_DMASYNC_POSTREAD);
me 1390 dev/pci/ubsec.c bus_dmamap_sync(sc->sc_dmat, me->me_epb.dma_map,
me 1391 dev/pci/ubsec.c 0, me->me_epb.dma_map->dm_mapsize, BUS_DMASYNC_POSTWRITE);
me 1400 dev/pci/ubsec.c bcopy(me->me_C.dma_vaddr,
me 1402 dev/pci/ubsec.c (me->me_modbits + 7) / 8);
me 1404 dev/pci/ubsec.c ubsec_kshift_l(me->me_shiftbits,
me 1405 dev/pci/ubsec.c me->me_C.dma_vaddr, me->me_normbits,
me 1412 dev/pci/ubsec.c bzero(me->me_E.dma_vaddr, me->me_E.dma_size);
me 1413 dev/pci/ubsec.c bzero(me->me_M.dma_vaddr, me->me_M.dma_size);
me 1414 dev/pci/ubsec.c bzero(me->me_C.dma_vaddr, me->me_C.dma_size);
me 1415 dev/pci/ubsec.c bzero(me->me_q.q_ctx.dma_vaddr, me->me_q.q_ctx.dma_size);
me 1418 dev/pci/ubsec.c SIMPLEQ_INSERT_TAIL(&sc->sc_q2free, &me->me_q, q_next);
me 1728 dev/pci/ubsec.c struct ubsec_q2_modexp *me = (struct ubsec_q2_modexp *)q;
me 1730 dev/pci/ubsec.c ubsec_dma_free(sc, &me->me_q.q_mcr);
me 1731 dev/pci/ubsec.c ubsec_dma_free(sc, &me->me_q.q_ctx);
me 1732 dev/pci/ubsec.c ubsec_dma_free(sc, &me->me_M);
me 1733 dev/pci/ubsec.c ubsec_dma_free(sc, &me->me_E);
me 1734 dev/pci/ubsec.c ubsec_dma_free(sc, &me->me_C);
me 1735 dev/pci/ubsec.c ubsec_dma_free(sc, &me->me_epb);
me 1736 dev/pci/ubsec.c free(me, M_DEVBUF);
me 1801 dev/pci/ubsec.c struct ubsec_q2_modexp *me;
me 1808 dev/pci/ubsec.c me = (struct ubsec_q2_modexp *)malloc(sizeof *me, M_DEVBUF, M_NOWAIT);
me 1809 dev/pci/ubsec.c if (me == NULL) {
me 1813 dev/pci/ubsec.c bzero(me, sizeof *me);
me 1814 dev/pci/ubsec.c me->me_krp = krp;
me 1815 dev/pci/ubsec.c me->me_q.q_type = UBS_CTXOP_MODEXP;
me 1835 dev/pci/ubsec.c me->me_modbits = nbits;
me 1836 dev/pci/ubsec.c me->me_shiftbits = shiftbits;
me 1837 dev/pci/ubsec.c me->me_normbits = normbits;
me 1846 dev/pci/ubsec.c &me->me_q.q_mcr, 0)) {
me 1850 dev/pci/ubsec.c mcr = (struct ubsec_mcr *)me->me_q.q_mcr.dma_vaddr;
me 1853 dev/pci/ubsec.c &me->me_q.q_ctx, 0)) {
me 1863 dev/pci/ubsec.c if (ubsec_dma_malloc(sc, normbits / 8, &me->me_M, 0)) {
me 1869 dev/pci/ubsec.c me->me_M.dma_vaddr, normbits);
me 1871 dev/pci/ubsec.c if (ubsec_dma_malloc(sc, normbits / 8, &me->me_C, 0)) {
me 1875 dev/pci/ubsec.c bzero(me->me_C.dma_vaddr, me->me_C.dma_size);
me 1882 dev/pci/ubsec.c if (ubsec_dma_malloc(sc, normbits / 8, &me->me_E, 0)) {
me 1888 dev/pci/ubsec.c me->me_E.dma_vaddr, normbits);
me 1891 dev/pci/ubsec.c &me->me_epb, 0)) {
me 1895 dev/pci/ubsec.c epb = (struct ubsec_pktbuf *)me->me_epb.dma_vaddr;
me 1896 dev/pci/ubsec.c epb->pb_addr = htole32(me->me_E.dma_paddr);
me 1907 dev/pci/ubsec.c mcr->mcr_cmdctxp = htole32(me->me_q.q_ctx.dma_paddr);
me 1911 dev/pci/ubsec.c mcr->mcr_ipktbuf.pb_addr = htole32(me->me_M.dma_paddr);
me 1913 dev/pci/ubsec.c mcr->mcr_ipktbuf.pb_next = htole32(me->me_epb.dma_paddr);
me 1915 dev/pci/ubsec.c mcr->mcr_opktbuf.pb_addr = htole32(me->me_C.dma_paddr);
me 1929 dev/pci/ubsec.c ctx = (struct ubsec_ctx_modexp *)me->me_q.q_ctx.dma_vaddr;
me 1948 dev/pci/ubsec.c bus_dmamap_sync(sc->sc_dmat, me->me_M.dma_map,
me 1949 dev/pci/ubsec.c 0, me->me_M.dma_map->dm_mapsize, BUS_DMASYNC_PREWRITE);
me 1950 dev/pci/ubsec.c bus_dmamap_sync(sc->sc_dmat, me->me_E.dma_map,
me 1951 dev/pci/ubsec.c 0, me->me_E.dma_map->dm_mapsize, BUS_DMASYNC_PREWRITE);
me 1952 dev/pci/ubsec.c bus_dmamap_sync(sc->sc_dmat, me->me_C.dma_map,
me 1953 dev/pci/ubsec.c 0, me->me_C.dma_map->dm_mapsize, BUS_DMASYNC_PREREAD);
me 1954 dev/pci/ubsec.c bus_dmamap_sync(sc->sc_dmat, me->me_epb.dma_map,
me 1955 dev/pci/ubsec.c 0, me->me_epb.dma_map->dm_mapsize, BUS_DMASYNC_PREWRITE);
me 1959 dev/pci/ubsec.c SIMPLEQ_INSERT_TAIL(&sc->sc_queue2, &me->me_q, q_next);
me 1966 dev/pci/ubsec.c if (me != NULL) {
me 1967 dev/pci/ubsec.c if (me->me_q.q_mcr.dma_map != NULL)
me 1968 dev/pci/ubsec.c ubsec_dma_free(sc, &me->me_q.q_mcr);
me 1969 dev/pci/ubsec.c if (me->me_q.q_ctx.dma_map != NULL) {
me 1970 dev/pci/ubsec.c bzero(me->me_q.q_ctx.dma_vaddr, me->me_q.q_ctx.dma_size);
me 1971 dev/pci/ubsec.c ubsec_dma_free(sc, &me->me_q.q_ctx);
me 1973 dev/pci/ubsec.c if (me->me_M.dma_map != NULL) {
me 1974 dev/pci/ubsec.c bzero(me->me_M.dma_vaddr, me->me_M.dma_size);
me 1975 dev/pci/ubsec.c ubsec_dma_free(sc, &me->me_M);
me 1977 dev/pci/ubsec.c if (me->me_E.dma_map != NULL) {
me 1978 dev/pci/ubsec.c bzero(me->me_E.dma_vaddr, me->me_E.dma_size);
me 1979 dev/pci/ubsec.c ubsec_dma_free(sc, &me->me_E);
me 1981 dev/pci/ubsec.c if (me->me_C.dma_map != NULL) {
me 1982 dev/pci/ubsec.c bzero(me->me_C.dma_vaddr, me->me_C.dma_size);
me 1983 dev/pci/ubsec.c ubsec_dma_free(sc, &me->me_C);
me 1985 dev/pci/ubsec.c if (me->me_epb.dma_map != NULL)
me 1986 dev/pci/ubsec.c ubsec_dma_free(sc, &me->me_epb);
me 1987 dev/pci/ubsec.c free(me, M_DEVBUF);
me 2000 dev/pci/ubsec.c struct ubsec_q2_modexp *me;
me 2007 dev/pci/ubsec.c me = (struct ubsec_q2_modexp *)malloc(sizeof *me, M_DEVBUF, M_NOWAIT);
me 2008 dev/pci/ubsec.c if (me == NULL) {
me 2012 dev/pci/ubsec.c bzero(me, sizeof *me);
me 2013 dev/pci/ubsec.c me->me_krp = krp;
me 2014 dev/pci/ubsec.c me->me_q.q_type = UBS_CTXOP_MODEXP;
me 2035 dev/pci/ubsec.c me->me_modbits = nbits;
me 2036 dev/pci/ubsec.c me->me_shiftbits = shiftbits;
me 2037 dev/pci/ubsec.c me->me_normbits = normbits;
me 2046 dev/pci/ubsec.c &me->me_q.q_mcr, 0)) {
me 2050 dev/pci/ubsec.c mcr = (struct ubsec_mcr *)me->me_q.q_mcr.dma_vaddr;
me 2053 dev/pci/ubsec.c &me->me_q.q_ctx, 0)) {
me 2063 dev/pci/ubsec.c if (ubsec_dma_malloc(sc, normbits / 8, &me->me_M, 0)) {
me 2067 dev/pci/ubsec.c bzero(me->me_M.dma_vaddr, normbits / 8);
me 2069 dev/pci/ubsec.c me->me_M.dma_vaddr, (mbits + 7) / 8);
me 2071 dev/pci/ubsec.c if (ubsec_dma_malloc(sc, normbits / 8, &me->me_C, 0)) {
me 2075 dev/pci/ubsec.c bzero(me->me_C.dma_vaddr, me->me_C.dma_size);
me 2082 dev/pci/ubsec.c if (ubsec_dma_malloc(sc, normbits / 8, &me->me_E, 0)) {
me 2086 dev/pci/ubsec.c bzero(me->me_E.dma_vaddr, normbits / 8);
me 2088 dev/pci/ubsec.c me->me_E.dma_vaddr, (ebits + 7) / 8);
me 2091 dev/pci/ubsec.c &me->me_epb, 0)) {
me 2095 dev/pci/ubsec.c epb = (struct ubsec_pktbuf *)me->me_epb.dma_vaddr;
me 2096 dev/pci/ubsec.c epb->pb_addr = htole32(me->me_E.dma_paddr);
me 2107 dev/pci/ubsec.c mcr->mcr_cmdctxp = htole32(me->me_q.q_ctx.dma_paddr);
me 2111 dev/pci/ubsec.c mcr->mcr_ipktbuf.pb_addr = htole32(me->me_M.dma_paddr);
me 2113 dev/pci/ubsec.c mcr->mcr_ipktbuf.pb_next = htole32(me->me_epb.dma_paddr);
me 2115 dev/pci/ubsec.c mcr->mcr_opktbuf.pb_addr = htole32(me->me_C.dma_paddr);
me 2129 dev/pci/ubsec.c ctx = (struct ubsec_ctx_modexp *)me->me_q.q_ctx.dma_vaddr;
me 2147 dev/pci/ubsec.c bus_dmamap_sync(sc->sc_dmat, me->me_M.dma_map,
me 2148 dev/pci/ubsec.c 0, me->me_M.dma_map->dm_mapsize, BUS_DMASYNC_PREWRITE);
me 2149 dev/pci/ubsec.c bus_dmamap_sync(sc->sc_dmat, me->me_E.dma_map,
me 2150 dev/pci/ubsec.c 0, me->me_E.dma_map->dm_mapsize, BUS_DMASYNC_PREWRITE);
me 2151 dev/pci/ubsec.c bus_dmamap_sync(sc->sc_dmat, me->me_C.dma_map,
me 2152 dev/pci/ubsec.c 0, me->me_C.dma_map->dm_mapsize, BUS_DMASYNC_PREREAD);
me 2153 dev/pci/ubsec.c bus_dmamap_sync(sc->sc_dmat, me->me_epb.dma_map,
me 2154 dev/pci/ubsec.c 0, me->me_epb.dma_map->dm_mapsize, BUS_DMASYNC_PREWRITE);
me 2158 dev/pci/ubsec.c SIMPLEQ_INSERT_TAIL(&sc->sc_queue2, &me->me_q, q_next);
me 2165 dev/pci/ubsec.c if (me != NULL) {
me 2166 dev/pci/ubsec.c if (me->me_q.q_mcr.dma_map != NULL)
me 2167 dev/pci/ubsec.c ubsec_dma_free(sc, &me->me_q.q_mcr);
me 2168 dev/pci/ubsec.c if (me->me_q.q_ctx.dma_map != NULL) {
me 2169 dev/pci/ubsec.c bzero(me->me_q.q_ctx.dma_vaddr, me->me_q.q_ctx.dma_size);
me 2170 dev/pci/ubsec.c ubsec_dma_free(sc, &me->me_q.q_ctx);
me 2172 dev/pci/ubsec.c if (me->me_M.dma_map != NULL) {
me 2173 dev/pci/ubsec.c bzero(me->me_M.dma_vaddr, me->me_M.dma_size);
me 2174 dev/pci/ubsec.c ubsec_dma_free(sc, &me->me_M);
me 2176 dev/pci/ubsec.c if (me->me_E.dma_map != NULL) {
me 2177 dev/pci/ubsec.c bzero(me->me_E.dma_vaddr, me->me_E.dma_size);
me 2178 dev/pci/ubsec.c ubsec_dma_free(sc, &me->me_E);
me 2180 dev/pci/ubsec.c if (me->me_C.dma_map != NULL) {
me 2181 dev/pci/ubsec.c bzero(me->me_C.dma_vaddr, me->me_C.dma_size);
me 2182 dev/pci/ubsec.c ubsec_dma_free(sc, &me->me_C);
me 2184 dev/pci/ubsec.c if (me->me_epb.dma_map != NULL)
me 2185 dev/pci/ubsec.c ubsec_dma_free(sc, &me->me_epb);
me 2186 dev/pci/ubsec.c free(me, M_DEVBUF);
me 769 dev/wscons/wskbd.c wskbd_mux_open(struct wsevsrc *me, struct wseventvar *evp)
me 771 dev/wscons/wskbd.c struct wskbd_softc *sc = (struct wskbd_softc *)me;
me 863 dev/wscons/wskbd.c wskbd_mux_close(struct wsevsrc *me)
me 865 dev/wscons/wskbd.c struct wskbd_softc *sc = (struct wskbd_softc *)me;
me 1188 dev/wscons/wskbd.c wskbd_set_console_display(struct device *displaydv, struct wsevsrc *me)
me 1196 dev/wscons/wskbd.c (void)wsmux_attach_sc((struct wsmux_softc *)me, &sc->sc_base);
me 675 dev/wscons/wsmouse.c wsmouse_mux_open(struct wsevsrc *me, struct wseventvar *evp)
me 677 dev/wscons/wsmouse.c struct wsmouse_softc *sc = (struct wsmouse_softc *)me;
me 686 dev/wscons/wsmouse.c wsmouse_mux_close(struct wsevsrc *me)
me 688 dev/wscons/wsmouse.c struct wsmouse_softc *sc = (struct wsmouse_softc *)me;
me 217 dev/wscons/wsmux.c wsmux_mux_open(struct wsevsrc *me, struct wseventvar *evar)
me 219 dev/wscons/wsmux.c struct wsmux_softc *sc = (struct wsmux_softc *)me;
me 241 dev/wscons/wsmux.c struct wsevsrc *me;
me 249 dev/wscons/wsmux.c CIRCLEQ_FOREACH(me, &sc->sc_cld, me_next) {
me 251 dev/wscons/wsmux.c sc->sc_base.me_dv.dv_xname, me, me->me_dv.dv_xname));
me 253 dev/wscons/wsmux.c if (me->me_evp != NULL) {
me 257 dev/wscons/wsmux.c if (me->me_parent != sc) {
me 258 dev/wscons/wsmux.c printf("wsmux_do_open: bad child=%p\n", me);
me 261 dev/wscons/wsmux.c error = wsevsrc_open(me, evar);
me 267 dev/wscons/wsmux.c (void)wsevsrc_open(me, evar);
me 296 dev/wscons/wsmux.c wsmux_mux_close(struct wsevsrc *me)
me 298 dev/wscons/wsmux.c me->me_evp = NULL;
me 299 dev/wscons/wsmux.c wsmux_do_close((struct wsmux_softc *)me);
me 307 dev/wscons/wsmux.c struct wsevsrc *me;
me 312 dev/wscons/wsmux.c CIRCLEQ_FOREACH(me, &sc->sc_cld, me_next) {
me 314 dev/wscons/wsmux.c sc->sc_base.me_dv.dv_xname, me, me->me_dv.dv_xname));
me 316 dev/wscons/wsmux.c if (me->me_parent != sc) {
me 317 dev/wscons/wsmux.c printf("wsmuxclose: bad child=%p\n", me);
me 321 dev/wscons/wsmux.c (void)wsevsrc_close(me);
me 322 dev/wscons/wsmux.c me->me_evp = NULL;
me 370 dev/wscons/wsmux.c struct wsevsrc *me;
me 441 dev/wscons/wsmux.c CIRCLEQ_FOREACH(me, &sc->sc_cld, me_next) {
me 442 dev/wscons/wsmux.c if (me->me_ops->type == d->type &&
me 443 dev/wscons/wsmux.c me->me_dv.dv_unit == d->idx) {
me 445 dev/wscons/wsmux.c wsmux_detach_sc(me);
me 456 dev/wscons/wsmux.c CIRCLEQ_FOREACH(me, &sc->sc_cld, me_next) {
me 459 dev/wscons/wsmux.c l->devices[n].type = me->me_ops->type;
me 460 dev/wscons/wsmux.c l->devices[n].idx = me->me_dv.dv_unit;
me 514 dev/wscons/wsmux.c CIRCLEQ_FOREACH(me, &sc->sc_cld, me_next) {
me 517 dev/wscons/wsmux.c if (me->me_parent != sc) {
me 518 dev/wscons/wsmux.c printf("wsmux_do_ioctl: bad child %p\n", me);
me 522 dev/wscons/wsmux.c error = wsevsrc_ioctl(me, cmd, data, flag, p);
me 524 dev/wscons/wsmux.c sc->sc_base.me_dv.dv_xname, me, me->me_dv.dv_xname,
me 607 dev/wscons/wsmux.c wsmux_attach_sc(struct wsmux_softc *sc, struct wsevsrc *me)
me 615 dev/wscons/wsmux.c sc->sc_base.me_dv.dv_xname, sc, me->me_ops->type));
me 618 dev/wscons/wsmux.c if (me->me_parent != NULL) {
me 623 dev/wscons/wsmux.c me->me_parent = sc;
me 624 dev/wscons/wsmux.c CIRCLEQ_INSERT_TAIL(&sc->sc_cld, me, me_next);
me 632 dev/wscons/wsmux.c if (me->me_ops->dsetdisplay != NULL) {
me 633 dev/wscons/wsmux.c error = wsevsrc_set_display(me, sc->sc_displaydv);
me 640 dev/wscons/wsmux.c me->me_dv.dv_xname, sc->sc_rawkbd));
me 641 dev/wscons/wsmux.c (void)wsevsrc_ioctl(me, WSKBDIO_SETMODE,
me 645 dev/wscons/wsmux.c (void)wsevsrc_ioctl(me,
me 655 dev/wscons/wsmux.c sc->sc_base.me_dv.dv_xname, me->me_dv.dv_xname));
me 656 dev/wscons/wsmux.c error = wsevsrc_open(me, sc->sc_base.me_evp);
me 663 dev/wscons/wsmux.c me->me_parent = NULL;
me 664 dev/wscons/wsmux.c CIRCLEQ_REMOVE(&sc->sc_cld, me, me_next);
me 674 dev/wscons/wsmux.c wsmux_detach_sc(struct wsevsrc *me)
me 676 dev/wscons/wsmux.c struct wsmux_softc *sc = me->me_parent;
me 679 dev/wscons/wsmux.c me->me_dv.dv_xname, me, sc));
me 684 dev/wscons/wsmux.c me->me_dv.dv_xname);
me 691 dev/wscons/wsmux.c if (me->me_ops->dsetdisplay != NULL)
me 693 dev/wscons/wsmux.c (void)wsevsrc_set_display(me, NULL);
me 696 dev/wscons/wsmux.c if (me->me_evp != NULL) {
me 699 dev/wscons/wsmux.c (void)wsevsrc_close(me);
me 702 dev/wscons/wsmux.c CIRCLEQ_REMOVE(&sc->sc_cld, me, me_next);
me 703 dev/wscons/wsmux.c me->me_parent = NULL;
me 716 dev/wscons/wsmux.c struct wsevsrc *me;
me 735 dev/wscons/wsmux.c CIRCLEQ_FOREACH(me, &sc->sc_cld, me_next) {
me 736 dev/wscons/wsmux.c DPRINTF(("wsmux_displayioctl: me=%p\n", me));
me 738 dev/wscons/wsmux.c if (me->me_parent != sc) {
me 739 dev/wscons/wsmux.c printf("wsmux_displayioctl: bad child %p\n", me);
me 743 dev/wscons/wsmux.c if (me->me_ops->ddispioctl != NULL) {
me 744 dev/wscons/wsmux.c error = wsevsrc_display_ioctl(me, cmd, data, flag, p);
me 746 dev/wscons/wsmux.c me, me->me_dv.dv_xname, error));
me 784 dev/wscons/wsmux.c struct wsevsrc *me;
me 797 dev/wscons/wsmux.c CIRCLEQ_FOREACH(me, &sc->sc_cld,me_next) {
me 799 dev/wscons/wsmux.c if (me->me_parent != sc) {
me 800 dev/wscons/wsmux.c printf("wsmux_set_display: bad child parent %p\n", me);
me 804 dev/wscons/wsmux.c if (me->me_ops->dsetdisplay != NULL) {
me 805 dev/wscons/wsmux.c error = wsevsrc_set_display(me, nsc->sc_displaydv);
me 807 dev/wscons/wsmux.c me, me->me_dv.dv_xname, error));
me 813 dev/wscons/wsmux.c me->me_dv.dv_xname, sc->sc_rawkbd));
me 814 dev/wscons/wsmux.c (void)wsevsrc_ioctl(me, WSKBDIO_SETMODE,
me 71 dev/wscons/wsmuxvar.h #define wsevsrc_open(me, evp) \
me 72 dev/wscons/wsmuxvar.h ((me)->me_ops->dopen((me), evp))
me 73 dev/wscons/wsmuxvar.h #define wsevsrc_close(me) \
me 74 dev/wscons/wsmuxvar.h ((me)->me_ops->dclose((me)))
me 75 dev/wscons/wsmuxvar.h #define wsevsrc_ioctl(me, cmd, data, flag, p) \
me 76 dev/wscons/wsmuxvar.h ((me)->me_ops->dioctl(&(me)->me_dv, cmd, (caddr_t)data, flag, p))
me 77 dev/wscons/wsmuxvar.h #define wsevsrc_display_ioctl(me, cmd, data, flag, p) \
me 78 dev/wscons/wsmuxvar.h ((me)->me_ops->ddispioctl(&(me)->me_dv, cmd, (caddr_t)data, flag, p))
me 79 dev/wscons/wsmuxvar.h #define wsevsrc_set_display(me, arg) \
me 80 dev/wscons/wsmuxvar.h ((me)->me_ops->dsetdisplay(&(me)->me_dv, arg))
me 1051 net/if_spppsubr.c u_int32_t me, mymask;
me 1111 net/if_spppsubr.c sppp_get_ip_addrs(sp, &me, 0, &mymask);
me 1112 net/if_spppsubr.c if (me != 0)
me 1113 net/if_spppsubr.c sppp_cisco_send(sp, CISCO_ADDR_REPLY, me, mymask);
me 390 uvm/uvm_map.c struct vm_map_entry *me, *ne;
me 397 uvm/uvm_map.c me = uvm.kentry_free;
me 398 uvm/uvm_map.c if (me == NULL) {
me 408 uvm/uvm_map.c me = ne;
me 414 uvm/uvm_map.c uvm.kentry_free = me->next;
me 418 uvm/uvm_map.c me->flags = UVM_MAP_STATIC;
me 421 uvm/uvm_map.c me = pool_get(&uvm_map_entry_kmem_pool, PR_WAITOK);
me 422 uvm/uvm_map.c me->flags = UVM_MAP_KMEM;
me 425 uvm/uvm_map.c me = pool_get(&uvm_map_entry_pool, PR_WAITOK);
me 426 uvm/uvm_map.c me->flags = 0;
me 429 uvm/uvm_map.c UVMHIST_LOG(maphist, "<- new entry=%p [kentry=%ld]", me,
me 431 uvm/uvm_map.c return(me);
me 441 uvm/uvm_map.c uvm_mapent_free(struct vm_map_entry *me)
me 447 uvm/uvm_map.c me, me->flags, 0, 0);
me 448 uvm/uvm_map.c if (me->flags & UVM_MAP_STATIC) {
me 451 uvm/uvm_map.c me->next = uvm.kentry_free;
me 452 uvm/uvm_map.c uvm.kentry_free = me;
me 456 uvm/uvm_map.c } else if (me->flags & UVM_MAP_KMEM) {
me 458 uvm/uvm_map.c pool_put(&uvm_map_entry_kmem_pool, me);
me 461 uvm/uvm_map.c pool_put(&uvm_map_entry_pool, me);