tctx 898 arch/arm64/crypto/aes-glue.c struct mac_tfm_ctx *tctx = crypto_shash_ctx(desc->tfm); tctx 911 arch/arm64/crypto/aes-glue.c mac_do_update(&tctx->key, p, blocks, ctx->dg, tctx 938 arch/arm64/crypto/aes-glue.c struct mac_tfm_ctx *tctx = crypto_shash_ctx(desc->tfm); tctx 941 arch/arm64/crypto/aes-glue.c mac_do_update(&tctx->key, NULL, 0, ctx->dg, (ctx->len != 0), 0); tctx 950 arch/arm64/crypto/aes-glue.c struct mac_tfm_ctx *tctx = crypto_shash_ctx(desc->tfm); tctx 952 arch/arm64/crypto/aes-glue.c u8 *consts = tctx->consts; tctx 959 arch/arm64/crypto/aes-glue.c mac_do_update(&tctx->key, consts, 1, ctx->dg, 0, 1); tctx 120 crypto/adiantum.c struct adiantum_tfm_ctx *tctx = crypto_skcipher_ctx(tfm); tctx 132 crypto/adiantum.c crypto_skcipher_clear_flags(tctx->streamcipher, CRYPTO_TFM_REQ_MASK); tctx 133 crypto/adiantum.c crypto_skcipher_set_flags(tctx->streamcipher, tctx 136 crypto/adiantum.c err = crypto_skcipher_setkey(tctx->streamcipher, key, keylen); tctx 138 crypto/adiantum.c crypto_skcipher_get_flags(tctx->streamcipher) & tctx 145 crypto/adiantum.c crypto_skcipher_reqsize(tctx->streamcipher), GFP_KERNEL); tctx 151 crypto/adiantum.c skcipher_request_set_tfm(&data->req, tctx->streamcipher); tctx 163 crypto/adiantum.c crypto_cipher_clear_flags(tctx->blockcipher, CRYPTO_TFM_REQ_MASK); tctx 164 crypto/adiantum.c crypto_cipher_set_flags(tctx->blockcipher, tctx 167 crypto/adiantum.c err = crypto_cipher_setkey(tctx->blockcipher, keyp, tctx 170 crypto/adiantum.c crypto_cipher_get_flags(tctx->blockcipher) & tctx 177 crypto/adiantum.c poly1305_core_setkey(&tctx->header_hash_key, keyp); tctx 180 crypto/adiantum.c crypto_shash_clear_flags(tctx->hash, CRYPTO_TFM_REQ_MASK); tctx 181 crypto/adiantum.c crypto_shash_set_flags(tctx->hash, crypto_skcipher_get_flags(tfm) & tctx 183 crypto/adiantum.c err = crypto_shash_setkey(tctx->hash, keyp, NHPOLY1305_KEY_SIZE); tctx 184 crypto/adiantum.c crypto_skcipher_set_flags(tfm, crypto_shash_get_flags(tctx->hash) & tctx 230 crypto/adiantum.c const struct adiantum_tfm_ctx *tctx = crypto_skcipher_ctx(tfm); tctx 244 crypto/adiantum.c poly1305_core_blocks(&state, &tctx->header_hash_key, tctx 248 crypto/adiantum.c poly1305_core_blocks(&state, &tctx->header_hash_key, req->iv, tctx 259 crypto/adiantum.c const struct adiantum_tfm_ctx *tctx = crypto_skcipher_ctx(tfm); tctx 267 crypto/adiantum.c hash_desc->tfm = tctx->hash; tctx 293 crypto/adiantum.c const struct adiantum_tfm_ctx *tctx = crypto_skcipher_ctx(tfm); tctx 301 crypto/adiantum.c crypto_cipher_decrypt_one(tctx->blockcipher, rctx->rbuf.bytes, tctx 333 crypto/adiantum.c const struct adiantum_tfm_ctx *tctx = crypto_skcipher_ctx(tfm); tctx 361 crypto/adiantum.c crypto_cipher_encrypt_one(tctx->blockcipher, rctx->rbuf.bytes, tctx 385 crypto/adiantum.c skcipher_request_set_tfm(&rctx->u.streamcipher_req, tctx->streamcipher); tctx 409 crypto/adiantum.c struct adiantum_tfm_ctx *tctx = crypto_skcipher_ctx(tfm); tctx 432 crypto/adiantum.c tctx->streamcipher = streamcipher; tctx 433 crypto/adiantum.c tctx->blockcipher = blockcipher; tctx 434 crypto/adiantum.c tctx->hash = hash; tctx 459 crypto/adiantum.c struct adiantum_tfm_ctx *tctx = crypto_skcipher_ctx(tfm); tctx 461 crypto/adiantum.c crypto_free_skcipher(tctx->streamcipher); tctx 462 crypto/adiantum.c crypto_free_cipher(tctx->blockcipher); tctx 463 crypto/adiantum.c crypto_free_shash(tctx->hash); tctx 856 crypto/ccm.c struct cbcmac_tfm_ctx *tctx = crypto_shash_ctx(parent); tctx 858 crypto/ccm.c struct crypto_cipher *tfm = tctx->child; tctx 882 crypto/ccm.c struct cbcmac_tfm_ctx *tctx = crypto_shash_ctx(parent); tctx 884 crypto/ccm.c struct crypto_cipher *tfm = tctx->child; tctx 122 crypto/cmac.c struct cmac_tfm_ctx *tctx = crypto_shash_ctx(parent); tctx 124 crypto/cmac.c struct crypto_cipher *tfm = tctx->child; tctx 168 crypto/cmac.c struct cmac_tfm_ctx *tctx = crypto_shash_ctx(parent); tctx 170 crypto/cmac.c struct crypto_cipher *tfm = tctx->child; tctx 172 crypto/cmac.c u8 *consts = PTR_ALIGN((void *)tctx->ctx, tctx 68 crypto/essiv.c struct essiv_tfm_ctx *tctx = crypto_skcipher_ctx(tfm); tctx 69 crypto/essiv.c SHASH_DESC_ON_STACK(desc, tctx->hash); tctx 73 crypto/essiv.c crypto_skcipher_clear_flags(tctx->u.skcipher, CRYPTO_TFM_REQ_MASK); tctx 74 crypto/essiv.c crypto_skcipher_set_flags(tctx->u.skcipher, tctx 77 crypto/essiv.c err = crypto_skcipher_setkey(tctx->u.skcipher, key, keylen); tctx 79 crypto/essiv.c crypto_skcipher_get_flags(tctx->u.skcipher) & tctx 84 crypto/essiv.c desc->tfm = tctx->hash; tctx 89 crypto/essiv.c crypto_cipher_clear_flags(tctx->essiv_cipher, CRYPTO_TFM_REQ_MASK); tctx 90 crypto/essiv.c crypto_cipher_set_flags(tctx->essiv_cipher, tctx 93 crypto/essiv.c err = crypto_cipher_setkey(tctx->essiv_cipher, salt, tctx 94 crypto/essiv.c crypto_shash_digestsize(tctx->hash)); tctx 96 crypto/essiv.c crypto_cipher_get_flags(tctx->essiv_cipher) & tctx 105 crypto/essiv.c struct essiv_tfm_ctx *tctx = crypto_aead_ctx(tfm); tctx 106 crypto/essiv.c SHASH_DESC_ON_STACK(desc, tctx->hash); tctx 111 crypto/essiv.c crypto_aead_clear_flags(tctx->u.aead, CRYPTO_TFM_REQ_MASK); tctx 112 crypto/essiv.c crypto_aead_set_flags(tctx->u.aead, crypto_aead_get_flags(tfm) & tctx 114 crypto/essiv.c err = crypto_aead_setkey(tctx->u.aead, key, keylen); tctx 115 crypto/essiv.c crypto_aead_set_flags(tfm, crypto_aead_get_flags(tctx->u.aead) & tctx 125 crypto/essiv.c desc->tfm = tctx->hash; tctx 132 crypto/essiv.c crypto_cipher_clear_flags(tctx->essiv_cipher, CRYPTO_TFM_REQ_MASK); tctx 133 crypto/essiv.c crypto_cipher_set_flags(tctx->essiv_cipher, crypto_aead_get_flags(tfm) & tctx 135 crypto/essiv.c err = crypto_cipher_setkey(tctx->essiv_cipher, salt, tctx 136 crypto/essiv.c crypto_shash_digestsize(tctx->hash)); tctx 137 crypto/essiv.c crypto_aead_set_flags(tfm, crypto_cipher_get_flags(tctx->essiv_cipher) & tctx 146 crypto/essiv.c struct essiv_tfm_ctx *tctx = crypto_aead_ctx(tfm); tctx 148 crypto/essiv.c return crypto_aead_setauthsize(tctx->u.aead, authsize); tctx 161 crypto/essiv.c const struct essiv_tfm_ctx *tctx = crypto_skcipher_ctx(tfm); tctx 164 crypto/essiv.c crypto_cipher_encrypt_one(tctx->essiv_cipher, req->iv, req->iv); tctx 166 crypto/essiv.c skcipher_request_set_tfm(subreq, tctx->u.skcipher); tctx 199 crypto/essiv.c const struct essiv_tfm_ctx *tctx = crypto_aead_ctx(tfm); tctx 205 crypto/essiv.c crypto_cipher_encrypt_one(tctx->essiv_cipher, req->iv, req->iv); tctx 218 crypto/essiv.c u8 *iv = (u8 *)aead_request_ctx(req) + tctx->ivoffset; tctx 259 crypto/essiv.c aead_request_set_tfm(subreq, tctx->u.aead); tctx 284 crypto/essiv.c struct essiv_tfm_ctx *tctx) tctx 300 crypto/essiv.c tctx->essiv_cipher = essiv_cipher; tctx 301 crypto/essiv.c tctx->hash = hash; tctx 314 crypto/essiv.c struct essiv_tfm_ctx *tctx = crypto_skcipher_ctx(tfm); tctx 325 crypto/essiv.c err = essiv_init_tfm(ictx, tctx); tctx 331 crypto/essiv.c tctx->u.skcipher = skcipher; tctx 339 crypto/essiv.c struct essiv_tfm_ctx *tctx = crypto_aead_ctx(tfm); tctx 354 crypto/essiv.c tctx->ivoffset = offsetof(struct essiv_aead_request_ctx, aead_req) + tctx 356 crypto/essiv.c crypto_aead_set_reqsize(tfm, tctx->ivoffset + crypto_aead_ivsize(aead)); tctx 358 crypto/essiv.c err = essiv_init_tfm(ictx, tctx); tctx 364 crypto/essiv.c tctx->u.aead = aead; tctx 370 crypto/essiv.c struct essiv_tfm_ctx *tctx = crypto_skcipher_ctx(tfm); tctx 372 crypto/essiv.c crypto_free_skcipher(tctx->u.skcipher); tctx 373 crypto/essiv.c crypto_free_cipher(tctx->essiv_cipher); tctx 374 crypto/essiv.c crypto_free_shash(tctx->hash); tctx 379 crypto/essiv.c struct essiv_tfm_ctx *tctx = crypto_aead_ctx(tfm); tctx 381 crypto/essiv.c crypto_free_aead(tctx->u.aead); tctx 382 crypto/essiv.c crypto_free_cipher(tctx->essiv_cipher); tctx 383 crypto/essiv.c crypto_free_shash(tctx->hash); tctx 462 crypto/tgr192.c static void tgr192_transform(struct tgr192_ctx *tctx, const u8 * data) tctx 472 crypto/tgr192.c a = aa = tctx->a; tctx 473 crypto/tgr192.c b = bb = tctx->b; tctx 474 crypto/tgr192.c c = cc = tctx->c; tctx 488 crypto/tgr192.c tctx->a = a; tctx 489 crypto/tgr192.c tctx->b = b; tctx 490 crypto/tgr192.c tctx->c = c; tctx 495 crypto/tgr192.c struct tgr192_ctx *tctx = shash_desc_ctx(desc); tctx 497 crypto/tgr192.c tctx->a = 0x0123456789abcdefULL; tctx 498 crypto/tgr192.c tctx->b = 0xfedcba9876543210ULL; tctx 499 crypto/tgr192.c tctx->c = 0xf096a5b4c3b2e187ULL; tctx 500 crypto/tgr192.c tctx->nblocks = 0; tctx 501 crypto/tgr192.c tctx->count = 0; tctx 512 crypto/tgr192.c struct tgr192_ctx *tctx = shash_desc_ctx(desc); tctx 514 crypto/tgr192.c if (tctx->count == 64) { /* flush the buffer */ tctx 515 crypto/tgr192.c tgr192_transform(tctx, tctx->hash); tctx 516 crypto/tgr192.c tctx->count = 0; tctx 517 crypto/tgr192.c tctx->nblocks++; tctx 522 crypto/tgr192.c if (tctx->count) { tctx 523 crypto/tgr192.c for (; len && tctx->count < 64; len--) { tctx 524 crypto/tgr192.c tctx->hash[tctx->count++] = *inbuf++; tctx 534 crypto/tgr192.c tgr192_transform(tctx, inbuf); tctx 535 crypto/tgr192.c tctx->count = 0; tctx 536 crypto/tgr192.c tctx->nblocks++; tctx 540 crypto/tgr192.c for (; len && tctx->count < 64; len--) { tctx 541 crypto/tgr192.c tctx->hash[tctx->count++] = *inbuf++; tctx 552 crypto/tgr192.c struct tgr192_ctx *tctx = shash_desc_ctx(desc); tctx 561 crypto/tgr192.c t = tctx->nblocks; tctx 567 crypto/tgr192.c if ((lsb = t + tctx->count) < t) { /* add the count */ tctx 576 crypto/tgr192.c if (tctx->count < 56) { /* enough room */ tctx 577 crypto/tgr192.c tctx->hash[tctx->count++] = 0x01; /* pad */ tctx 578 crypto/tgr192.c while (tctx->count < 56) { tctx 579 crypto/tgr192.c tctx->hash[tctx->count++] = 0; /* pad */ tctx 582 crypto/tgr192.c tctx->hash[tctx->count++] = 0x01; /* pad character */ tctx 583 crypto/tgr192.c while (tctx->count < 64) { tctx 584 crypto/tgr192.c tctx->hash[tctx->count++] = 0; tctx 587 crypto/tgr192.c memset(tctx->hash, 0, 56); /* fill next block with zeroes */ tctx 590 crypto/tgr192.c le32p = (__le32 *)&tctx->hash[56]; tctx 594 crypto/tgr192.c tgr192_transform(tctx, tctx->hash); tctx 596 crypto/tgr192.c be64p = (__be64 *)tctx->hash; tctx 597 crypto/tgr192.c dst[0] = be64p[0] = cpu_to_be64(tctx->a); tctx 598 crypto/tgr192.c dst[1] = be64p[1] = cpu_to_be64(tctx->b); tctx 599 crypto/tgr192.c dst[2] = be64p[2] = cpu_to_be64(tctx->c); tctx 398 crypto/vmac.c static void vhash_blocks(const struct vmac_tfm_ctx *tctx, tctx 402 crypto/vmac.c const u64 *kptr = tctx->nhkey; tctx 403 crypto/vmac.c const u64 pkh = tctx->polykey[0]; tctx 404 crypto/vmac.c const u64 pkl = tctx->polykey[1]; tctx 432 crypto/vmac.c struct vmac_tfm_ctx *tctx = crypto_shash_ctx(tfm); tctx 443 crypto/vmac.c err = crypto_cipher_setkey(tctx->cipher, key, keylen); tctx 449 crypto/vmac.c for (i = 0; i < ARRAY_SIZE(tctx->nhkey); i += 2) { tctx 450 crypto/vmac.c crypto_cipher_encrypt_one(tctx->cipher, (u8 *)out, in); tctx 451 crypto/vmac.c tctx->nhkey[i] = be64_to_cpu(out[0]); tctx 452 crypto/vmac.c tctx->nhkey[i+1] = be64_to_cpu(out[1]); tctx 459 crypto/vmac.c for (i = 0; i < ARRAY_SIZE(tctx->polykey); i += 2) { tctx 460 crypto/vmac.c crypto_cipher_encrypt_one(tctx->cipher, (u8 *)out, in); tctx 461 crypto/vmac.c tctx->polykey[i] = be64_to_cpu(out[0]) & mpoly; tctx 462 crypto/vmac.c tctx->polykey[i+1] = be64_to_cpu(out[1]) & mpoly; tctx 469 crypto/vmac.c for (i = 0; i < ARRAY_SIZE(tctx->l3key); i += 2) { tctx 471 crypto/vmac.c crypto_cipher_encrypt_one(tctx->cipher, (u8 *)out, in); tctx 472 crypto/vmac.c tctx->l3key[i] = be64_to_cpu(out[0]); tctx 473 crypto/vmac.c tctx->l3key[i+1] = be64_to_cpu(out[1]); tctx 475 crypto/vmac.c } while (tctx->l3key[i] >= p64 || tctx->l3key[i+1] >= p64); tctx 483 crypto/vmac.c const struct vmac_tfm_ctx *tctx = crypto_shash_ctx(desc->tfm); tctx 488 crypto/vmac.c memcpy(dctx->polytmp, tctx->polykey, sizeof(dctx->polytmp)); tctx 495 crypto/vmac.c const struct vmac_tfm_ctx *tctx = crypto_shash_ctx(desc->tfm); tctx 515 crypto/vmac.c vhash_blocks(tctx, dctx, dctx->partial_words, 1); tctx 523 crypto/vmac.c vhash_blocks(tctx, dctx, (const __le64 *)p, n / VMAC_NHBYTES); tctx 536 crypto/vmac.c static u64 vhash_final(const struct vmac_tfm_ctx *tctx, tctx 550 crypto/vmac.c nh_16(dctx->partial_words, tctx->nhkey, n / 8, rh, rl); tctx 553 crypto/vmac.c poly_step(ch, cl, tctx->polykey[0], tctx->polykey[1], tctx 560 crypto/vmac.c return l3hash(ch, cl, tctx->l3key[0], tctx->l3key[1], partial * 8); tctx 565 crypto/vmac.c const struct vmac_tfm_ctx *tctx = crypto_shash_ctx(desc->tfm); tctx 583 crypto/vmac.c hash = vhash_final(tctx, dctx); tctx 589 crypto/vmac.c crypto_cipher_encrypt_one(tctx->cipher, dctx->nonce.bytes, tctx 602 crypto/vmac.c struct vmac_tfm_ctx *tctx = crypto_tfm_ctx(tfm); tctx 609 crypto/vmac.c tctx->cipher = cipher; tctx 615 crypto/vmac.c struct vmac_tfm_ctx *tctx = crypto_tfm_ctx(tfm); tctx 617 crypto/vmac.c crypto_free_cipher(tctx->cipher); tctx 89 crypto/xcbc.c struct xcbc_tfm_ctx *tctx = crypto_shash_ctx(parent); tctx 91 crypto/xcbc.c struct crypto_cipher *tfm = tctx->child; tctx 135 crypto/xcbc.c struct xcbc_tfm_ctx *tctx = crypto_shash_ctx(parent); tctx 137 crypto/xcbc.c struct crypto_cipher *tfm = tctx->child; tctx 139 crypto/xcbc.c u8 *consts = PTR_ALIGN(&tctx->ctx[0], alignmask + 1); tctx 23 crypto/xxhash_generic.c struct xxhash64_tfm_ctx *tctx = crypto_shash_ctx(tfm); tctx 25 crypto/xxhash_generic.c if (keylen != sizeof(tctx->seed)) { tctx 29 crypto/xxhash_generic.c tctx->seed = get_unaligned_le64(key); tctx 35 crypto/xxhash_generic.c struct xxhash64_tfm_ctx *tctx = crypto_shash_ctx(desc->tfm); tctx 38 crypto/xxhash_generic.c xxh64_reset(&dctx->xxhstate, tctx->seed); tctx 65 crypto/xxhash_generic.c struct xxhash64_tfm_ctx *tctx = crypto_shash_ctx(desc->tfm); tctx 67 crypto/xxhash_generic.c put_unaligned_le64(xxh64(data, length, tctx->seed), out); tctx 404 drivers/crypto/atmel-sha.c static struct atmel_sha_dev *atmel_sha_find_dev(struct atmel_sha_ctx *tctx) tctx 410 drivers/crypto/atmel-sha.c if (!tctx->dd) { tctx 415 drivers/crypto/atmel-sha.c tctx->dd = dd; tctx 417 drivers/crypto/atmel-sha.c dd = tctx->dd; tctx 428 drivers/crypto/atmel-sha.c struct atmel_sha_ctx *tctx = crypto_ahash_ctx(tfm); tctx 430 drivers/crypto/atmel-sha.c struct atmel_sha_dev *dd = atmel_sha_find_dev(tctx); tctx 1150 drivers/crypto/atmel-sha.c struct atmel_sha_ctx *tctx = crypto_tfm_ctx(req->base.tfm); tctx 1151 drivers/crypto/atmel-sha.c struct atmel_sha_dev *dd = tctx->dd; tctx 2282 drivers/crypto/atmel-sha.c struct atmel_sha_ctx *tctx; tctx 2316 drivers/crypto/atmel-sha.c tctx = crypto_ahash_ctx(tfm); tctx 2317 drivers/crypto/atmel-sha.c tctx->start = atmel_sha_authenc_start; tctx 2318 drivers/crypto/atmel-sha.c tctx->flags = mode; tctx 2368 drivers/crypto/atmel-sha.c struct atmel_sha_ctx *tctx = crypto_ahash_ctx(tfm); tctx 2375 drivers/crypto/atmel-sha.c dd = atmel_sha_find_dev(tctx); tctx 69 drivers/crypto/geode-aes.c geode_aes_crypt(const struct geode_aes_tfm_ctx *tctx, const void *src, tctx 96 drivers/crypto/geode-aes.c _writefield(AES_WRITEKEY0_REG, tctx->key); tctx 112 drivers/crypto/geode-aes.c struct geode_aes_tfm_ctx *tctx = crypto_tfm_ctx(tfm); tctx 115 drivers/crypto/geode-aes.c tctx->keylen = len; tctx 118 drivers/crypto/geode-aes.c memcpy(tctx->key, key, len); tctx 131 drivers/crypto/geode-aes.c tctx->fallback.cip->base.crt_flags &= ~CRYPTO_TFM_REQ_MASK; tctx 132 drivers/crypto/geode-aes.c tctx->fallback.cip->base.crt_flags |= tctx 135 drivers/crypto/geode-aes.c ret = crypto_cipher_setkey(tctx->fallback.cip, key, len); tctx 138 drivers/crypto/geode-aes.c tfm->crt_flags |= (tctx->fallback.cip->base.crt_flags & tctx 147 drivers/crypto/geode-aes.c struct geode_aes_tfm_ctx *tctx = crypto_skcipher_ctx(tfm); tctx 150 drivers/crypto/geode-aes.c tctx->keylen = len; tctx 153 drivers/crypto/geode-aes.c memcpy(tctx->key, key, len); tctx 166 drivers/crypto/geode-aes.c crypto_skcipher_clear_flags(tctx->fallback.skcipher, tctx 168 drivers/crypto/geode-aes.c crypto_skcipher_set_flags(tctx->fallback.skcipher, tctx 171 drivers/crypto/geode-aes.c ret = crypto_skcipher_setkey(tctx->fallback.skcipher, key, len); tctx 173 drivers/crypto/geode-aes.c crypto_skcipher_get_flags(tctx->fallback.skcipher) & tctx 181 drivers/crypto/geode-aes.c const struct geode_aes_tfm_ctx *tctx = crypto_tfm_ctx(tfm); tctx 183 drivers/crypto/geode-aes.c if (unlikely(tctx->keylen != AES_KEYSIZE_128)) { tctx 184 drivers/crypto/geode-aes.c crypto_cipher_encrypt_one(tctx->fallback.cip, out, in); tctx 188 drivers/crypto/geode-aes.c geode_aes_crypt(tctx, in, out, AES_BLOCK_SIZE, NULL, tctx 196 drivers/crypto/geode-aes.c const struct geode_aes_tfm_ctx *tctx = crypto_tfm_ctx(tfm); tctx 198 drivers/crypto/geode-aes.c if (unlikely(tctx->keylen != AES_KEYSIZE_128)) { tctx 199 drivers/crypto/geode-aes.c crypto_cipher_decrypt_one(tctx->fallback.cip, out, in); tctx 203 drivers/crypto/geode-aes.c geode_aes_crypt(tctx, in, out, AES_BLOCK_SIZE, NULL, tctx 210 drivers/crypto/geode-aes.c struct geode_aes_tfm_ctx *tctx = crypto_tfm_ctx(tfm); tctx 212 drivers/crypto/geode-aes.c tctx->fallback.cip = crypto_alloc_cipher(name, 0, tctx 215 drivers/crypto/geode-aes.c if (IS_ERR(tctx->fallback.cip)) { tctx 217 drivers/crypto/geode-aes.c return PTR_ERR(tctx->fallback.cip); tctx 225 drivers/crypto/geode-aes.c struct geode_aes_tfm_ctx *tctx = crypto_tfm_ctx(tfm); tctx 227 drivers/crypto/geode-aes.c crypto_free_cipher(tctx->fallback.cip); tctx 256 drivers/crypto/geode-aes.c struct geode_aes_tfm_ctx *tctx = crypto_skcipher_ctx(tfm); tctx 258 drivers/crypto/geode-aes.c tctx->fallback.skcipher = tctx 261 drivers/crypto/geode-aes.c if (IS_ERR(tctx->fallback.skcipher)) { tctx 263 drivers/crypto/geode-aes.c return PTR_ERR(tctx->fallback.skcipher); tctx 267 drivers/crypto/geode-aes.c crypto_skcipher_reqsize(tctx->fallback.skcipher)); tctx 273 drivers/crypto/geode-aes.c struct geode_aes_tfm_ctx *tctx = crypto_skcipher_ctx(tfm); tctx 275 drivers/crypto/geode-aes.c crypto_free_skcipher(tctx->fallback.skcipher); tctx 281 drivers/crypto/geode-aes.c const struct geode_aes_tfm_ctx *tctx = crypto_skcipher_ctx(tfm); tctx 286 drivers/crypto/geode-aes.c if (unlikely(tctx->keylen != AES_KEYSIZE_128)) { tctx 290 drivers/crypto/geode-aes.c skcipher_request_set_tfm(subreq, tctx->fallback.skcipher); tctx 300 drivers/crypto/geode-aes.c geode_aes_crypt(tctx, walk.src.virt.addr, walk.dst.virt.addr, tctx 619 drivers/crypto/img-hash.c struct img_hash_ctx *tctx = crypto_ahash_ctx(tfm); tctx 626 drivers/crypto/img-hash.c if (!tctx->hdev) { tctx 631 drivers/crypto/img-hash.c tctx->hdev = hdev; tctx 634 drivers/crypto/img-hash.c hdev = tctx->hdev; tctx 667 drivers/crypto/img-hash.c err = img_hash_handle_queue(tctx->hdev, req); tctx 717 drivers/crypto/img-hash.c struct img_hash_ctx *tctx = crypto_tfm_ctx(tfm); tctx 719 drivers/crypto/img-hash.c crypto_free_ahash(tctx->fallback); tctx 153 drivers/crypto/mediatek/mtk-sha.c static struct mtk_cryp *mtk_sha_find_dev(struct mtk_sha_ctx *tctx) tctx 159 drivers/crypto/mediatek/mtk-sha.c if (!tctx->cryp) { tctx 164 drivers/crypto/mediatek/mtk-sha.c tctx->cryp = cryp; tctx 166 drivers/crypto/mediatek/mtk-sha.c cryp = tctx->cryp; tctx 173 drivers/crypto/mediatek/mtk-sha.c tctx->id = cryp->rec; tctx 358 drivers/crypto/mediatek/mtk-sha.c struct mtk_sha_ctx *tctx = crypto_tfm_ctx(req->base.tfm); tctx 359 drivers/crypto/mediatek/mtk-sha.c struct mtk_sha_hmac_ctx *bctx = tctx->base; tctx 375 drivers/crypto/mediatek/mtk-sha.c struct mtk_sha_ctx *tctx = crypto_ahash_ctx(tfm); tctx 408 drivers/crypto/mediatek/mtk-sha.c ctx->buffer = tctx->buf; tctx 410 drivers/crypto/mediatek/mtk-sha.c if (tctx->flags & SHA_FLAGS_HMAC) { tctx 411 drivers/crypto/mediatek/mtk-sha.c struct mtk_sha_hmac_ctx *bctx = tctx->base; tctx 706 drivers/crypto/mediatek/mtk-sha.c struct mtk_sha_ctx *tctx = crypto_tfm_ctx(req->base.tfm); tctx 710 drivers/crypto/mediatek/mtk-sha.c return mtk_sha_handle_queue(tctx->cryp, tctx->id, req); tctx 802 drivers/crypto/mediatek/mtk-sha.c struct mtk_sha_ctx *tctx = crypto_ahash_ctx(tfm); tctx 803 drivers/crypto/mediatek/mtk-sha.c struct mtk_sha_hmac_ctx *bctx = tctx->base; tctx 851 drivers/crypto/mediatek/mtk-sha.c struct mtk_sha_ctx *tctx = crypto_tfm_ctx(tfm); tctx 854 drivers/crypto/mediatek/mtk-sha.c cryp = mtk_sha_find_dev(tctx); tctx 862 drivers/crypto/mediatek/mtk-sha.c struct mtk_sha_hmac_ctx *bctx = tctx->base; tctx 864 drivers/crypto/mediatek/mtk-sha.c tctx->flags |= SHA_FLAGS_HMAC; tctx 909 drivers/crypto/mediatek/mtk-sha.c struct mtk_sha_ctx *tctx = crypto_tfm_ctx(tfm); tctx 911 drivers/crypto/mediatek/mtk-sha.c if (tctx->flags & SHA_FLAGS_HMAC) { tctx 912 drivers/crypto/mediatek/mtk-sha.c struct mtk_sha_hmac_ctx *bctx = tctx->base; tctx 304 drivers/crypto/omap-sham.c struct omap_sham_ctx *tctx = crypto_ahash_ctx(tfm); tctx 305 drivers/crypto/omap-sham.c struct omap_sham_hmac_ctx *bctx = tctx->base; tctx 466 drivers/crypto/omap-sham.c struct omap_sham_ctx *tctx = crypto_ahash_ctx(tfm); tctx 467 drivers/crypto/omap-sham.c struct omap_sham_hmac_ctx *bctx = tctx->base; tctx 924 drivers/crypto/omap-sham.c struct omap_sham_ctx *tctx = crypto_ahash_ctx(tfm); tctx 930 drivers/crypto/omap-sham.c if (!tctx->dd) { tctx 935 drivers/crypto/omap-sham.c tctx->dd = dd; tctx 937 drivers/crypto/omap-sham.c dd = tctx->dd; tctx 981 drivers/crypto/omap-sham.c if (tctx->flags & BIT(FLAGS_HMAC)) { tctx 983 drivers/crypto/omap-sham.c struct omap_sham_hmac_ctx *bctx = tctx->base; tctx 1048 drivers/crypto/omap-sham.c struct omap_sham_ctx *tctx = crypto_tfm_ctx(req->base.tfm); tctx 1049 drivers/crypto/omap-sham.c struct omap_sham_hmac_ctx *bctx = tctx->base; tctx 1190 drivers/crypto/omap-sham.c struct omap_sham_ctx *tctx = crypto_tfm_ctx(req->base.tfm); tctx 1191 drivers/crypto/omap-sham.c struct omap_sham_dev *dd = tctx->dd; tctx 1231 drivers/crypto/omap-sham.c struct omap_sham_ctx *tctx = crypto_tfm_ctx(req->base.tfm); tctx 1244 drivers/crypto/omap-sham.c return omap_sham_shash_digest(tctx->fallback, req->base.flags, tctx 1301 drivers/crypto/omap-sham.c struct omap_sham_ctx *tctx = crypto_ahash_ctx(tfm); tctx 1302 drivers/crypto/omap-sham.c struct omap_sham_hmac_ctx *bctx = tctx->base; tctx 1309 drivers/crypto/omap-sham.c if (!tctx->dd) { tctx 1314 drivers/crypto/omap-sham.c tctx->dd = dd; tctx 1316 drivers/crypto/omap-sham.c dd = tctx->dd; tctx 1320 drivers/crypto/omap-sham.c err = crypto_shash_setkey(tctx->fallback, key, keylen); tctx 1351 drivers/crypto/omap-sham.c struct omap_sham_ctx *tctx = crypto_tfm_ctx(tfm); tctx 1355 drivers/crypto/omap-sham.c tctx->fallback = crypto_alloc_shash(alg_name, 0, tctx 1357 drivers/crypto/omap-sham.c if (IS_ERR(tctx->fallback)) { tctx 1360 drivers/crypto/omap-sham.c return PTR_ERR(tctx->fallback); tctx 1367 drivers/crypto/omap-sham.c struct omap_sham_hmac_ctx *bctx = tctx->base; tctx 1368 drivers/crypto/omap-sham.c tctx->flags |= BIT(FLAGS_HMAC); tctx 1374 drivers/crypto/omap-sham.c crypto_free_shash(tctx->fallback); tctx 1420 drivers/crypto/omap-sham.c struct omap_sham_ctx *tctx = crypto_tfm_ctx(tfm); tctx 1422 drivers/crypto/omap-sham.c crypto_free_shash(tctx->fallback); tctx 1423 drivers/crypto/omap-sham.c tctx->fallback = NULL; tctx 1425 drivers/crypto/omap-sham.c if (tctx->flags & BIT(FLAGS_HMAC)) { tctx 1426 drivers/crypto/omap-sham.c struct omap_sham_hmac_ctx *bctx = tctx->base; tctx 166 drivers/crypto/rockchip/rk3288_crypto_ahash.c struct rk_ahash_ctx *tctx = crypto_tfm_ctx(req->base.tfm); tctx 167 drivers/crypto/rockchip/rk3288_crypto_ahash.c struct rk_crypto_info *dev = tctx->dev; tctx 274 drivers/crypto/rockchip/rk3288_crypto_ahash.c struct rk_ahash_ctx *tctx = crypto_tfm_ctx(tfm); tctx 282 drivers/crypto/rockchip/rk3288_crypto_ahash.c tctx->dev = algt->dev; tctx 283 drivers/crypto/rockchip/rk3288_crypto_ahash.c tctx->dev->addr_vir = (void *)__get_free_page(GFP_KERNEL); tctx 284 drivers/crypto/rockchip/rk3288_crypto_ahash.c if (!tctx->dev->addr_vir) { tctx 285 drivers/crypto/rockchip/rk3288_crypto_ahash.c dev_err(tctx->dev->dev, "failed to kmalloc for addr_vir\n"); tctx 288 drivers/crypto/rockchip/rk3288_crypto_ahash.c tctx->dev->start = rk_ahash_start; tctx 289 drivers/crypto/rockchip/rk3288_crypto_ahash.c tctx->dev->update = rk_ahash_crypto_rx; tctx 290 drivers/crypto/rockchip/rk3288_crypto_ahash.c tctx->dev->complete = rk_ahash_crypto_complete; tctx 293 drivers/crypto/rockchip/rk3288_crypto_ahash.c tctx->fallback_tfm = crypto_alloc_ahash(alg_name, 0, tctx 295 drivers/crypto/rockchip/rk3288_crypto_ahash.c if (IS_ERR(tctx->fallback_tfm)) { tctx 296 drivers/crypto/rockchip/rk3288_crypto_ahash.c dev_err(tctx->dev->dev, "Could not load fallback driver.\n"); tctx 297 drivers/crypto/rockchip/rk3288_crypto_ahash.c return PTR_ERR(tctx->fallback_tfm); tctx 301 drivers/crypto/rockchip/rk3288_crypto_ahash.c crypto_ahash_reqsize(tctx->fallback_tfm)); tctx 303 drivers/crypto/rockchip/rk3288_crypto_ahash.c return tctx->dev->enable_clk(tctx->dev); tctx 308 drivers/crypto/rockchip/rk3288_crypto_ahash.c struct rk_ahash_ctx *tctx = crypto_tfm_ctx(tfm); tctx 310 drivers/crypto/rockchip/rk3288_crypto_ahash.c free_page((unsigned long)tctx->dev->addr_vir); tctx 311 drivers/crypto/rockchip/rk3288_crypto_ahash.c return tctx->dev->disable_clk(tctx->dev); tctx 1490 drivers/crypto/s5p-sss.c struct s5p_hash_ctx *tctx = crypto_tfm_ctx(req->base.tfm); tctx 1494 drivers/crypto/s5p-sss.c return s5p_hash_handle_queue(tctx->dd, req); tctx 1547 drivers/crypto/s5p-sss.c struct s5p_hash_ctx *tctx = crypto_tfm_ctx(req->base.tfm); tctx 1550 drivers/crypto/s5p-sss.c return s5p_hash_shash_digest(tctx->fallback, req->base.flags, tctx 1628 drivers/crypto/s5p-sss.c struct s5p_hash_ctx *tctx = crypto_ahash_ctx(tfm); tctx 1630 drivers/crypto/s5p-sss.c ctx->dd = tctx->dd; tctx 1638 drivers/crypto/s5p-sss.c dev_dbg(tctx->dd->dev, "init: digest size: %d\n", tctx 1679 drivers/crypto/s5p-sss.c struct s5p_hash_ctx *tctx = crypto_tfm_ctx(tfm); tctx 1682 drivers/crypto/s5p-sss.c tctx->dd = s5p_dev; tctx 1684 drivers/crypto/s5p-sss.c tctx->fallback = crypto_alloc_shash(alg_name, 0, tctx 1686 drivers/crypto/s5p-sss.c if (IS_ERR(tctx->fallback)) { tctx 1688 drivers/crypto/s5p-sss.c return PTR_ERR(tctx->fallback); tctx 1714 drivers/crypto/s5p-sss.c struct s5p_hash_ctx *tctx = crypto_tfm_ctx(tfm); tctx 1716 drivers/crypto/s5p-sss.c crypto_free_shash(tctx->fallback); tctx 1717 drivers/crypto/s5p-sss.c tctx->fallback = NULL; tctx 1743 drivers/crypto/s5p-sss.c struct s5p_hash_ctx *tctx = crypto_ahash_ctx(tfm); tctx 1752 drivers/crypto/s5p-sss.c ctx->dd = tctx->dd; tctx 477 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c struct ttm_operation_ctx tctx = { false, false }; tctx 500 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c r = ttm_bo_validate(&bo->tbo, &bo->placement, &tctx);