crd 87 arch/i386/i386/via.c #define VIAC3_SID(crd,ses) (((crd) << 28) | ((ses) & 0x0fffffff))
crd 342 arch/i386/i386/via.c viac3_crypto_swauth(struct cryptop *crp, struct cryptodesc *crd,
crd 352 arch/i386/i386/via.c return (swcr_authcompute(crp, crd, sw, buf, type));
crd 356 arch/i386/i386/via.c viac3_crypto_encdec(struct cryptop *crp, struct cryptodesc *crd,
crd 362 arch/i386/i386/via.c if ((crd->crd_len % 16) != 0) {
crd 367 arch/i386/i386/via.c sc->op_buf = malloc(crd->crd_len, M_DEVBUF, M_NOWAIT);
crd 373 arch/i386/i386/via.c if (crd->crd_flags & CRD_F_ENCRYPT) {
crd 376 arch/i386/i386/via.c if (crd->crd_flags & CRD_F_IV_EXPLICIT)
crd 377 arch/i386/i386/via.c bcopy(crd->crd_iv, sc->op_iv, 16);
crd 381 arch/i386/i386/via.c if ((crd->crd_flags & CRD_F_IV_PRESENT) == 0) {
crd 384 arch/i386/i386/via.c crd->crd_inject, 16, sc->op_iv);
crd 387 arch/i386/i386/via.c crd->crd_inject, 16, sc->op_iv);
crd 390 arch/i386/i386/via.c crp->crp_buf + crd->crd_inject, 16);
crd 395 arch/i386/i386/via.c if (crd->crd_flags & CRD_F_IV_EXPLICIT)
crd 396 arch/i386/i386/via.c bcopy(crd->crd_iv, sc->op_iv, 16);
crd 400 arch/i386/i386/via.c crd->crd_inject, 16, sc->op_iv);
crd 403 arch/i386/i386/via.c crd->crd_inject, 16, sc->op_iv);
crd 405 arch/i386/i386/via.c bcopy(crp->crp_buf + crd->crd_inject,
crd 412 arch/i386/i386/via.c crd->crd_skip, crd->crd_len, sc->op_buf);
crd 415 arch/i386/i386/via.c crd->crd_skip, crd->crd_len, sc->op_buf);
crd 417 arch/i386/i386/via.c bcopy(crp->crp_buf + crd->crd_skip, sc->op_buf, crd->crd_len);
crd 421 arch/i386/i386/via.c crd->crd_len / 16, sc->op_iv);
crd 425 arch/i386/i386/via.c crd->crd_skip, crd->crd_len, sc->op_buf);
crd 428 arch/i386/i386/via.c crd->crd_skip, crd->crd_len, sc->op_buf);
crd 430 arch/i386/i386/via.c bcopy(sc->op_buf, crp->crp_buf + crd->crd_skip,
crd 431 arch/i386/i386/via.c crd->crd_len);
crd 434 arch/i386/i386/via.c if (crd->crd_flags & CRD_F_ENCRYPT) {
crd 437 arch/i386/i386/via.c crd->crd_skip + crd->crd_len - 16, 16,
crd 441 arch/i386/i386/via.c crd->crd_skip + crd->crd_len - 16, 16,
crd 444 arch/i386/i386/via.c bcopy(crp->crp_buf + crd->crd_skip +
crd 445 arch/i386/i386/via.c crd->crd_len - 16, ses->ses_iv, 16);
crd 449 arch/i386/i386/via.c bzero(sc->op_buf, crd->crd_len);
crd 462 arch/i386/i386/via.c struct cryptodesc *crd;
crd 477 arch/i386/i386/via.c for (crd = crp->crp_desc; crd; crd = crd->crd_next) {
crd 478 arch/i386/i386/via.c switch (crd->crd_alg) {
crd 480 arch/i386/i386/via.c if ((err = viac3_crypto_encdec(crp, crd, ses, sc,
crd 491 arch/i386/i386/via.c if ((err = viac3_crypto_swauth(crp, crd, ses->swd,
crd 191 arch/i386/pci/glxsb.c #define GLXSB_SID(crd,ses) (((crd) << 28) | ((ses) & 0x0fffffff))
crd 569 arch/i386/pci/glxsb.c glxsb_crypto_swauth(struct cryptop *crp, struct cryptodesc *crd,
crd 579 arch/i386/pci/glxsb.c return (swcr_authcompute(crp, crd, sw, buf, type));
crd 583 arch/i386/pci/glxsb.c glxsb_crypto_encdec(struct cryptop *crp, struct cryptodesc *crd,
crd 594 arch/i386/pci/glxsb.c if (crd == NULL || (crd->crd_len % SB_AES_BLOCK_SIZE) != 0) {
crd 600 arch/i386/pci/glxsb.c xlen = crd->crd_len > GLXSB_MAX_AES_LEN ?
crd 601 arch/i386/pci/glxsb.c GLXSB_MAX_AES_LEN : crd->crd_len;
crd 613 arch/i386/pci/glxsb.c if (crd->crd_flags & CRD_F_ENCRYPT) {
crd 615 arch/i386/pci/glxsb.c if (crd->crd_flags & CRD_F_IV_EXPLICIT)
crd 616 arch/i386/pci/glxsb.c bcopy(crd->crd_iv, op_iv, sizeof(op_iv));
crd 620 arch/i386/pci/glxsb.c if ((crd->crd_flags & CRD_F_IV_PRESENT) == 0) {
crd 623 arch/i386/pci/glxsb.c crd->crd_inject, sizeof(op_iv), op_iv);
crd 626 arch/i386/pci/glxsb.c crd->crd_inject, sizeof(op_iv), op_iv);
crd 629 arch/i386/pci/glxsb.c crp->crp_buf + crd->crd_inject, sizeof(op_iv));
crd 633 arch/i386/pci/glxsb.c if (crd->crd_flags & CRD_F_IV_EXPLICIT)
crd 634 arch/i386/pci/glxsb.c bcopy(crd->crd_iv, op_iv, sizeof(op_iv));
crd 638 arch/i386/pci/glxsb.c crd->crd_inject, sizeof(op_iv), op_iv);
crd 641 arch/i386/pci/glxsb.c crd->crd_inject, sizeof(op_iv), op_iv);
crd 643 arch/i386/pci/glxsb.c bcopy(crp->crp_buf + crd->crd_inject,
crd 649 arch/i386/pci/glxsb.c tlen = crd->crd_len;
crd 658 arch/i386/pci/glxsb.c crd->crd_skip + offset, len, op_src);
crd 661 arch/i386/pci/glxsb.c crd->crd_skip + offset, len, op_src);
crd 663 arch/i386/pci/glxsb.c bcopy(crp->crp_buf + crd->crd_skip + offset, op_src,
crd 675 arch/i386/pci/glxsb.c crd->crd_skip + offset, len, op_dst);
crd 678 arch/i386/pci/glxsb.c crd->crd_skip + offset, len, op_dst);
crd 680 arch/i386/pci/glxsb.c bcopy(op_dst, crp->crp_buf + crd->crd_skip + offset,
crd 698 arch/i386/pci/glxsb.c if (crd->crd_flags & CRD_F_ENCRYPT) {
crd 721 arch/i386/pci/glxsb.c struct cryptodesc *crd;
crd 731 arch/i386/pci/glxsb.c crd = crp->crp_desc;
crd 732 arch/i386/pci/glxsb.c if (crd == NULL) {
crd 744 arch/i386/pci/glxsb.c for (crd = crp->crp_desc; crd; crd = crd->crd_next) {
crd 745 arch/i386/pci/glxsb.c switch (crd->crd_alg) {
crd 747 arch/i386/pci/glxsb.c if ((err = glxsb_crypto_encdec(crp, crd, ses, sc,
crd 758 arch/i386/pci/glxsb.c if ((err = glxsb_crypto_swauth(crp, crd, ses->ses_swd,
crd 506 crypto/crypto.c struct cryptodesc *crd;
crd 554 crypto/crypto.c for (crd = crp->crp_desc; crd->crd_next; crd = crd->crd_next)
crd 555 crypto/crypto.c crd->CRD_INI.cri_next = &(crd->crd_next->CRD_INI);
crd 571 crypto/crypto.c struct cryptodesc *crd;
crd 579 crypto/crypto.c while ((crd = crp->crp_desc) != NULL) {
crd 580 crypto/crypto.c crp->crp_desc = crd->crd_next;
crd 581 crypto/crypto.c pool_put(&cryptodesc_pool, crd);
crd 594 crypto/crypto.c struct cryptodesc *crd;
crd 616 crypto/crypto.c crd = pool_get(&cryptodesc_pool, PR_NOWAIT);
crd 617 crypto/crypto.c if (crd == NULL) {
crd 623 crypto/crypto.c bzero(crd, sizeof(struct cryptodesc));
crd 624 crypto/crypto.c crd->crd_next = crp->crp_desc;
crd 625 crypto/crypto.c crp->crp_desc = crd;
crd 79 crypto/cryptosoft.c swcr_encdec(struct cryptodesc *crd, struct swcr_data *sw, caddr_t buf,
crd 94 crypto/cryptosoft.c if (crd->crd_len % blks)
crd 103 crypto/cryptosoft.c if (crd->crd_flags & CRD_F_ENCRYPT) {
crd 105 crypto/cryptosoft.c if (crd->crd_flags & CRD_F_IV_EXPLICIT)
crd 106 crypto/cryptosoft.c bcopy(crd->crd_iv, iv, ivlen);
crd 111 crypto/cryptosoft.c if (!(crd->crd_flags & CRD_F_IV_PRESENT)) {
crd 112 crypto/cryptosoft.c COPYBACK(outtype, buf, crd->crd_inject, ivlen, iv);
crd 117 crypto/cryptosoft.c if (crd->crd_flags & CRD_F_IV_EXPLICIT)
crd 118 crypto/cryptosoft.c bcopy(crd->crd_iv, iv, ivlen);
crd 121 crypto/cryptosoft.c COPYDATA(outtype, buf, crd->crd_inject, ivlen, iv);
crd 132 crypto/cryptosoft.c m = m_getptr(m, crd->crd_skip, &k);
crd 136 crypto/cryptosoft.c i = crd->crd_len;
crd 149 crypto/cryptosoft.c } else if (crd->crd_flags & CRD_F_ENCRYPT) {
crd 221 crypto/cryptosoft.c } else if (crd->crd_flags & CRD_F_ENCRYPT) {
crd 257 crypto/cryptosoft.c count = crd->crd_skip;
crd 262 crypto/cryptosoft.c i = crd->crd_len;
crd 276 crypto/cryptosoft.c } else if (crd->crd_flags & CRD_F_ENCRYPT) {
crd 339 crypto/cryptosoft.c } else if (crd->crd_flags & CRD_F_ENCRYPT) {
crd 383 crypto/cryptosoft.c swcr_authcompute(struct cryptop *crp, struct cryptodesc *crd,
crd 399 crypto/cryptosoft.c err = m_apply((struct mbuf *) buf, crd->crd_skip, crd->crd_len,
crd 403 crypto/cryptosoft.c err = cuio_apply((struct uio *) buf, crd->crd_skip,
crd 404 crypto/cryptosoft.c crd->crd_len,
crd 444 crypto/cryptosoft.c COPYBACK(outtype, buf, crd->crd_inject, axf->authsize, aalg);
crd 455 crypto/cryptosoft.c swcr_compdec(struct cryptodesc *crd, struct swcr_data *sw,
crd 470 crypto/cryptosoft.c MALLOC(data, u_int8_t *, crd->crd_len, M_CRYPTO_DATA, M_NOWAIT);
crd 473 crypto/cryptosoft.c COPYDATA(outtype, buf, crd->crd_skip, crd->crd_len, data);
crd 475 crypto/cryptosoft.c if (crd->crd_flags & CRD_F_COMP)
crd 476 crypto/cryptosoft.c result = cxf->compress(data, crd->crd_len, &out);
crd 478 crypto/cryptosoft.c result = cxf->decompress(data, crd->crd_len, &out);
crd 489 crypto/cryptosoft.c if (crd->crd_flags & CRD_F_COMP) {
crd 490 crypto/cryptosoft.c if (result > crd->crd_len) {
crd 497 crypto/cryptosoft.c COPYBACK(outtype, buf, crd->crd_skip, result, out);
crd 498 crypto/cryptosoft.c if (result < crd->crd_len) {
crd 499 crypto/cryptosoft.c adj = result - crd->crd_len;
crd 501 crypto/cryptosoft.c adj = result - crd->crd_len;
crd 507 crypto/cryptosoft.c adj = crd->crd_len - result;
crd 834 crypto/cryptosoft.c struct cryptodesc *crd;
crd 860 crypto/cryptosoft.c for (crd = crp->crp_desc; crd; crd = crd->crd_next) {
crd 872 crypto/cryptosoft.c sw && sw->sw_alg != crd->crd_alg;
crd 892 crypto/cryptosoft.c if ((crp->crp_etype = swcr_encdec(crd, sw,
crd 906 crypto/cryptosoft.c if ((crp->crp_etype = swcr_authcompute(crp, crd, sw,
crd 912 crypto/cryptosoft.c if ((crp->crp_etype = swcr_compdec(crd, sw,
crd 2257 dev/pci/hifn7751.c struct cryptodesc *crd;
crd 2323 dev/pci/hifn7751.c for (crd = crp->crp_desc; crd; crd = crd->crd_next) {
crd 2324 dev/pci/hifn7751.c if (crd->crd_alg != CRYPTO_DES_CBC &&
crd 2325 dev/pci/hifn7751.c crd->crd_alg != CRYPTO_3DES_CBC &&
crd 2326 dev/pci/hifn7751.c crd->crd_alg != CRYPTO_AES_CBC)
crd 2328 dev/pci/hifn7751.c ivlen = ((crd->crd_alg == CRYPTO_AES_CBC) ?
crd 2332 dev/pci/hifn7751.c crd->crd_skip + crd->crd_len - ivlen, ivlen,
crd 2336 dev/pci/hifn7751.c crd->crd_skip + crd->crd_len - ivlen, ivlen,
crd 2351 dev/pci/hifn7751.c for (crd = crp->crp_desc; crd; crd = crd->crd_next) {
crd 2354 dev/pci/hifn7751.c if (crd->crd_alg == CRYPTO_MD5)
crd 2356 dev/pci/hifn7751.c else if (crd->crd_alg == CRYPTO_SHA1)
crd 2358 dev/pci/hifn7751.c else if (crd->crd_alg == CRYPTO_MD5_HMAC ||
crd 2359 dev/pci/hifn7751.c crd->crd_alg == CRYPTO_SHA1_HMAC)
crd 2366 dev/pci/hifn7751.c crd->crd_inject, len, macbuf);
crd 2387 dev/pci/hifn7751.c struct cryptodesc *crd = crp->crp_desc;
crd 2390 dev/pci/hifn7751.c cmd->compcrd = crd;
crd 2402 dev/pci/hifn7751.c if ((crd->crd_flags & CRD_F_COMP) == 0)
crd 2404 dev/pci/hifn7751.c if (crd->crd_alg == CRYPTO_LZS_COMP)
crd 2434 dev/pci/hifn7751.c if ((crd->crd_flags & CRD_F_COMP) == 0)
crd 311 dev/pci/hifn7751var.h #define HIFN_SID(crd,ses) (((crd) << 28) | ((ses) & 0x7ff))
crd 1141 dev/pci/ises.c struct cryptodesc *maccrd, *enccrd, *crd;
crd 1206 dev/pci/ises.c for (crd = crp->crp_desc; crd; crd = crd->crd_next) {
crd 1207 dev/pci/ises.c switch (crd->crd_alg) {
crd 1214 dev/pci/ises.c maccrd = crd;
crd 1219 dev/pci/ises.c (maccrd && (crd->crd_flags & CRD_F_ENCRYPT)))
crd 1221 dev/pci/ises.c enccrd = crd;
crd 1300 dev/pci/ises.c switch (crd->crd_alg) {
crd 1465 dev/pci/ises.c struct cryptodesc *crd;
crd 1478 dev/pci/ises.c for (crd = crp->crp_desc; crd; crd = crd->crd_next) {
crd 1479 dev/pci/ises.c if (crd->crd_alg != CRYPTO_DES_CBC &&
crd 1480 dev/pci/ises.c crd->crd_alg != CRYPTO_3DES_CBC)
crd 1484 dev/pci/ises.c crd->crd_skip + crd->crd_len - 8, 8, sccr);
crd 1487 dev/pci/ises.c crd->crd_skip + crd->crd_len - 8, 8, sccr);
crd 1491 dev/pci/ises.c for (crd = crp->crp_desc; crd; crd = crd->crd_next) {
crd 1492 dev/pci/ises.c if (crd->crd_alg != CRYPTO_MD5_HMAC &&
crd 1493 dev/pci/ises.c crd->crd_alg != CRYPTO_SHA1_HMAC &&
crd 1494 dev/pci/ises.c crd->crd_alg != CRYPTO_RIPEMD160_HMAC)
crd 1498 dev/pci/ises.c crd->crd_inject, 12, (u_int8_t *)q->q_macbuf);
crd 348 dev/pci/isesreg.h #define ISES_SID(crd,ses) (((crd) << 28) | ((ses) & 0x0fffffff))
crd 878 dev/pci/noct.c struct cryptodesc *crd;
crd 893 dev/pci/noct.c crd = crp->crp_desc;
crd 894 dev/pci/noct.c switch (crd->crd_alg) {
crd 913 dev/pci/noct.c crd->crd_inject, len,
crd 919 dev/pci/noct.c if (crd->crd_alg == CRYPTO_DES_CBC ||
crd 920 dev/pci/noct.c crd->crd_alg == CRYPTO_3DES_CBC) {
crd 923 dev/pci/noct.c crd->crd_skip, crd->crd_len,
crd 927 dev/pci/noct.c crd->crd_skip, crd->crd_len,
crd 933 dev/pci/noct.c bus_dmamem_unmap(sc->sc_dmat, q->q_buf, crd->crd_len);
crd 962 dev/pci/noct.c struct cryptodesc *crd;
crd 966 dev/pci/noct.c crd = crp->crp_desc;
crd 969 dev/pci/noct.c if (crd->crd_next != NULL) {
crd 974 dev/pci/noct.c switch (crd->crd_alg) {
crd 977 dev/pci/noct.c noct_ea_start_hash(sc, q, crp, crd);
crd 981 dev/pci/noct.c noct_ea_start_des(sc, q, crp, crd);
crd 999 dev/pci/noct.c noct_ea_start_hash(sc, q, crp, crd)
crd 1003 dev/pci/noct.c struct cryptodesc *crd;
crd 1009 dev/pci/noct.c if (crd->crd_len > 0x4800) {
crd 1014 dev/pci/noct.c if ((err = bus_dmamem_alloc(sc->sc_dmat, crd->crd_len, PAGE_SIZE, 0,
crd 1019 dev/pci/noct.c crd->crd_len, (caddr_t *)&q->q_buf, BUS_DMA_WAITOK)) != 0)
crd 1022 dev/pci/noct.c if ((err = bus_dmamap_create(sc->sc_dmat, crd->crd_len, 1,
crd 1023 dev/pci/noct.c crd->crd_len, 0, BUS_DMA_WAITOK, &q->q_dmamap)) != 0)
crd 1027 dev/pci/noct.c rseg, crd->crd_len, BUS_DMA_WAITOK)) != 0)
crd 1032 dev/pci/noct.c crd->crd_skip, crd->crd_len, q->q_buf);
crd 1035 dev/pci/noct.c crd->crd_skip, crd->crd_len, q->q_buf);
crd 1055 dev/pci/noct.c switch (crd->crd_alg) {
crd 1065 dev/pci/noct.c sc->sc_eacmd[wp].buf[1] |= htole32(crd->crd_len);
crd 1096 dev/pci/noct.c bus_dmamem_unmap(sc->sc_dmat, q->q_buf, crd->crd_len);
crd 1108 dev/pci/noct.c noct_ea_start_des(sc, q, crp, crd)
crd 1112 dev/pci/noct.c struct cryptodesc *crd;
crd 1120 dev/pci/noct.c if (crd->crd_len > 0x4800) {
crd 1125 dev/pci/noct.c if ((crd->crd_len & 3) != 0) {
crd 1130 dev/pci/noct.c if (crd->crd_alg == CRYPTO_DES_CBC) {
crd 1132 dev/pci/noct.c key[i] = key[i + 8] = key[i + 16] = crd->crd_key[i];
crd 1135 dev/pci/noct.c key[i] = crd->crd_key[i];
crd 1138 dev/pci/noct.c if (crd->crd_flags & CRD_F_ENCRYPT) {
crd 1139 dev/pci/noct.c if (crd->crd_flags & CRD_F_IV_EXPLICIT)
crd 1140 dev/pci/noct.c bcopy(crd->crd_iv, iv, 8);
crd 1144 dev/pci/noct.c if (!(crd->crd_flags & CRD_F_IV_PRESENT)) {
crd 1147 dev/pci/noct.c crd->crd_inject, 8, iv);
crd 1150 dev/pci/noct.c crd->crd_inject, 8, iv);
crd 1153 dev/pci/noct.c if (crd->crd_flags & CRD_F_IV_EXPLICIT)
crd 1154 dev/pci/noct.c bcopy(crd->crd_iv, iv, 8);
crd 1157 dev/pci/noct.c crd->crd_inject, 8, iv);
crd 1160 dev/pci/noct.c crd->crd_inject, 8, iv);
crd 1163 dev/pci/noct.c if ((err = bus_dmamem_alloc(sc->sc_dmat, crd->crd_len, PAGE_SIZE, 0,
crd 1168 dev/pci/noct.c crd->crd_len, (caddr_t *)&q->q_buf, BUS_DMA_WAITOK)) != 0)
crd 1171 dev/pci/noct.c if ((err = bus_dmamap_create(sc->sc_dmat, crd->crd_len, 1,
crd 1172 dev/pci/noct.c crd->crd_len, 0, BUS_DMA_WAITOK, &q->q_dmamap)) != 0)
crd 1176 dev/pci/noct.c rseg, crd->crd_len, BUS_DMA_WAITOK)) != 0)
crd 1181 dev/pci/noct.c crd->crd_skip, crd->crd_len, q->q_buf);
crd 1184 dev/pci/noct.c crd->crd_skip, crd->crd_len, q->q_buf);
crd 1207 dev/pci/noct.c if (crd->crd_flags & CRD_F_ENCRYPT)
crd 1213 dev/pci/noct.c sc->sc_eacmd[wp].buf[1] |= htole32(crd->crd_len);
crd 1257 dev/pci/noct.c bus_dmamem_unmap(sc->sc_dmat, q->q_buf, crd->crd_len);
crd 107 dev/pci/noctvar.h #define NOCT_SID(crd, sesn) (((crd) << 28) | ((sesn) & 0x0fffffff))
crd 1638 dev/pci/safe.c struct cryptodesc *crd;
crd 1675 dev/pci/safe.c for (crd = crp->crp_desc; crd; crd = crd->crd_next) {
crd 1678 dev/pci/safe.c if (crd->crd_alg == CRYPTO_DES_CBC ||
crd 1679 dev/pci/safe.c crd->crd_alg == CRYPTO_3DES_CBC) {
crd 1681 dev/pci/safe.c } else if (crd->crd_alg == CRYPTO_AES_CBC) {
crd 1687 dev/pci/safe.c crd->crd_skip + crd->crd_len - ivsize,
crd 1692 dev/pci/safe.c crd->crd_skip + crd->crd_len - ivsize,
crd 1702 dev/pci/safe.c for (crd = crp->crp_desc; crd; crd = crd->crd_next) {
crd 1703 dev/pci/safe.c if (!(crd->crd_alg == CRYPTO_MD5_HMAC ||
crd 1704 dev/pci/safe.c crd->crd_alg == CRYPTO_SHA1_HMAC))
crd 1706 dev/pci/safe.c if (crd->crd_alg == CRYPTO_SHA1_HMAC) {
crd 1717 dev/pci/safe.c crd->crd_inject, 12,
crd 57 dev/pci/safevar.h #define SAFE_SID(crd, sesn) (((crd) << 28) | ((sesn) & 0x0fffffff))
crd 1220 dev/pci/ubsec.c struct cryptodesc *crd;
crd 1247 dev/pci/ubsec.c for (crd = crp->crp_desc; crd; crd = crd->crd_next) {
crd 1248 dev/pci/ubsec.c if (crd->crd_alg != CRYPTO_DES_CBC &&
crd 1249 dev/pci/ubsec.c crd->crd_alg != CRYPTO_3DES_CBC)
crd 1253 dev/pci/ubsec.c crd->crd_skip + crd->crd_len - 8, 8,
crd 1257 dev/pci/ubsec.c crd->crd_skip + crd->crd_len - 8, 8,
crd 1264 dev/pci/ubsec.c for (crd = crp->crp_desc; crd; crd = crd->crd_next) {
crd 1265 dev/pci/ubsec.c if (crd->crd_alg != CRYPTO_MD5_HMAC &&
crd 1266 dev/pci/ubsec.c crd->crd_alg != CRYPTO_SHA1_HMAC)
crd 1270 dev/pci/ubsec.c crd->crd_inject, 12,
crd 47 dev/pci/ubsecvar.h #define UBSEC_SID(crd, sesn) (((crd) << 28) | ((sesn) & 0x0fffffff))
crd 2908 dev/softraid.c struct cryptodesc *crd;
crd 2938 dev/softraid.c for (i = 0, crd = crp->crp_desc; crd; i++, crd = crd->crd_next) {
crd 2939 dev/softraid.c crd->crd_skip = 512 * i;
crd 2940 dev/softraid.c crd->crd_len = 512;
crd 2941 dev/softraid.c crd->crd_inject = 0;
crd 2942 dev/softraid.c crd->crd_flags = flags;
crd 2943 dev/softraid.c crd->crd_alg = CRYPTO_AES_CBC;
crd 2944 dev/softraid.c crd->crd_klen = 256;
crd 2945 dev/softraid.c crd->crd_rnd = 14;
crd 2946 dev/softraid.c crd->crd_key = sd->mds.mdd_raidc.src_key;
crd 2947 dev/softraid.c memset(crd->crd_iv, blk + i, sizeof(crd->crd_iv));
crd 745 netinet/ip_ah.c struct cryptodesc *crd;
crd 756 netinet/ip_ah.c crd = crp->crp_desc;
crd 215 netinet/ip_ipcomp.c struct cryptodesc *crd;
crd 224 netinet/ip_ipcomp.c crd = crp->crp_desc;