rng 1362 dev/pci/ubsec.c struct ubsec_q2_rng *rng = (struct ubsec_q2_rng *)q;
rng 1366 dev/pci/ubsec.c bus_dmamap_sync(sc->sc_dmat, rng->rng_buf.dma_map, 0,
rng 1367 dev/pci/ubsec.c rng->rng_buf.dma_map->dm_mapsize, BUS_DMASYNC_POSTREAD);
rng 1368 dev/pci/ubsec.c p = (u_int32_t *)rng->rng_buf.dma_vaddr;
rng 1371 dev/pci/ubsec.c rng->rng_used = 0;
rng 1457 dev/pci/ubsec.c struct ubsec_q2_rng *rng = &sc->sc_rng;
rng 1463 dev/pci/ubsec.c if (rng->rng_used) {
rng 1471 dev/pci/ubsec.c mcr = (struct ubsec_mcr *)rng->rng_q.q_mcr.dma_vaddr;
rng 1472 dev/pci/ubsec.c ctx = (struct ubsec_ctx_rngbypass *)rng->rng_q.q_ctx.dma_vaddr;
rng 1476 dev/pci/ubsec.c mcr->mcr_cmdctxp = htole32(rng->rng_q.q_ctx.dma_paddr);
rng 1480 dev/pci/ubsec.c mcr->mcr_opktbuf.pb_addr = htole32(rng->rng_buf.dma_paddr);
rng 1487 dev/pci/ubsec.c rng->rng_q.q_type = UBS_CTXOP_RNGSHA1;
rng 1489 dev/pci/ubsec.c bus_dmamap_sync(sc->sc_dmat, rng->rng_buf.dma_map, 0,
rng 1490 dev/pci/ubsec.c rng->rng_buf.dma_map->dm_mapsize, BUS_DMASYNC_PREREAD);
rng 1492 dev/pci/ubsec.c SIMPLEQ_INSERT_TAIL(&sc->sc_queue2, &rng->rng_q, q_next);
rng 1493 dev/pci/ubsec.c rng->rng_used = 1;