tctx              898 arch/arm64/crypto/aes-glue.c 	struct mac_tfm_ctx *tctx = crypto_shash_ctx(desc->tfm);
tctx              911 arch/arm64/crypto/aes-glue.c 			mac_do_update(&tctx->key, p, blocks, ctx->dg,
tctx              938 arch/arm64/crypto/aes-glue.c 	struct mac_tfm_ctx *tctx = crypto_shash_ctx(desc->tfm);
tctx              941 arch/arm64/crypto/aes-glue.c 	mac_do_update(&tctx->key, NULL, 0, ctx->dg, (ctx->len != 0), 0);
tctx              950 arch/arm64/crypto/aes-glue.c 	struct mac_tfm_ctx *tctx = crypto_shash_ctx(desc->tfm);
tctx              952 arch/arm64/crypto/aes-glue.c 	u8 *consts = tctx->consts;
tctx              959 arch/arm64/crypto/aes-glue.c 	mac_do_update(&tctx->key, consts, 1, ctx->dg, 0, 1);
tctx              120 crypto/adiantum.c 	struct adiantum_tfm_ctx *tctx = crypto_skcipher_ctx(tfm);
tctx              132 crypto/adiantum.c 	crypto_skcipher_clear_flags(tctx->streamcipher, CRYPTO_TFM_REQ_MASK);
tctx              133 crypto/adiantum.c 	crypto_skcipher_set_flags(tctx->streamcipher,
tctx              136 crypto/adiantum.c 	err = crypto_skcipher_setkey(tctx->streamcipher, key, keylen);
tctx              138 crypto/adiantum.c 				crypto_skcipher_get_flags(tctx->streamcipher) &
tctx              145 crypto/adiantum.c 		       crypto_skcipher_reqsize(tctx->streamcipher), GFP_KERNEL);
tctx              151 crypto/adiantum.c 	skcipher_request_set_tfm(&data->req, tctx->streamcipher);
tctx              163 crypto/adiantum.c 	crypto_cipher_clear_flags(tctx->blockcipher, CRYPTO_TFM_REQ_MASK);
tctx              164 crypto/adiantum.c 	crypto_cipher_set_flags(tctx->blockcipher,
tctx              167 crypto/adiantum.c 	err = crypto_cipher_setkey(tctx->blockcipher, keyp,
tctx              170 crypto/adiantum.c 				  crypto_cipher_get_flags(tctx->blockcipher) &
tctx              177 crypto/adiantum.c 	poly1305_core_setkey(&tctx->header_hash_key, keyp);
tctx              180 crypto/adiantum.c 	crypto_shash_clear_flags(tctx->hash, CRYPTO_TFM_REQ_MASK);
tctx              181 crypto/adiantum.c 	crypto_shash_set_flags(tctx->hash, crypto_skcipher_get_flags(tfm) &
tctx              183 crypto/adiantum.c 	err = crypto_shash_setkey(tctx->hash, keyp, NHPOLY1305_KEY_SIZE);
tctx              184 crypto/adiantum.c 	crypto_skcipher_set_flags(tfm, crypto_shash_get_flags(tctx->hash) &
tctx              230 crypto/adiantum.c 	const struct adiantum_tfm_ctx *tctx = crypto_skcipher_ctx(tfm);
tctx              244 crypto/adiantum.c 	poly1305_core_blocks(&state, &tctx->header_hash_key,
tctx              248 crypto/adiantum.c 	poly1305_core_blocks(&state, &tctx->header_hash_key, req->iv,
tctx              259 crypto/adiantum.c 	const struct adiantum_tfm_ctx *tctx = crypto_skcipher_ctx(tfm);
tctx              267 crypto/adiantum.c 	hash_desc->tfm = tctx->hash;
tctx              293 crypto/adiantum.c 	const struct adiantum_tfm_ctx *tctx = crypto_skcipher_ctx(tfm);
tctx              301 crypto/adiantum.c 		crypto_cipher_decrypt_one(tctx->blockcipher, rctx->rbuf.bytes,
tctx              333 crypto/adiantum.c 	const struct adiantum_tfm_ctx *tctx = crypto_skcipher_ctx(tfm);
tctx              361 crypto/adiantum.c 		crypto_cipher_encrypt_one(tctx->blockcipher, rctx->rbuf.bytes,
tctx              385 crypto/adiantum.c 	skcipher_request_set_tfm(&rctx->u.streamcipher_req, tctx->streamcipher);
tctx              409 crypto/adiantum.c 	struct adiantum_tfm_ctx *tctx = crypto_skcipher_ctx(tfm);
tctx              432 crypto/adiantum.c 	tctx->streamcipher = streamcipher;
tctx              433 crypto/adiantum.c 	tctx->blockcipher = blockcipher;
tctx              434 crypto/adiantum.c 	tctx->hash = hash;
tctx              459 crypto/adiantum.c 	struct adiantum_tfm_ctx *tctx = crypto_skcipher_ctx(tfm);
tctx              461 crypto/adiantum.c 	crypto_free_skcipher(tctx->streamcipher);
tctx              462 crypto/adiantum.c 	crypto_free_cipher(tctx->blockcipher);
tctx              463 crypto/adiantum.c 	crypto_free_shash(tctx->hash);
tctx              856 crypto/ccm.c   	struct cbcmac_tfm_ctx *tctx = crypto_shash_ctx(parent);
tctx              858 crypto/ccm.c   	struct crypto_cipher *tfm = tctx->child;
tctx              882 crypto/ccm.c   	struct cbcmac_tfm_ctx *tctx = crypto_shash_ctx(parent);
tctx              884 crypto/ccm.c   	struct crypto_cipher *tfm = tctx->child;
tctx              122 crypto/cmac.c  	struct cmac_tfm_ctx *tctx = crypto_shash_ctx(parent);
tctx              124 crypto/cmac.c  	struct crypto_cipher *tfm = tctx->child;
tctx              168 crypto/cmac.c  	struct cmac_tfm_ctx *tctx = crypto_shash_ctx(parent);
tctx              170 crypto/cmac.c  	struct crypto_cipher *tfm = tctx->child;
tctx              172 crypto/cmac.c  	u8 *consts = PTR_ALIGN((void *)tctx->ctx,
tctx               68 crypto/essiv.c 	struct essiv_tfm_ctx *tctx = crypto_skcipher_ctx(tfm);
tctx               69 crypto/essiv.c 	SHASH_DESC_ON_STACK(desc, tctx->hash);
tctx               73 crypto/essiv.c 	crypto_skcipher_clear_flags(tctx->u.skcipher, CRYPTO_TFM_REQ_MASK);
tctx               74 crypto/essiv.c 	crypto_skcipher_set_flags(tctx->u.skcipher,
tctx               77 crypto/essiv.c 	err = crypto_skcipher_setkey(tctx->u.skcipher, key, keylen);
tctx               79 crypto/essiv.c 				  crypto_skcipher_get_flags(tctx->u.skcipher) &
tctx               84 crypto/essiv.c 	desc->tfm = tctx->hash;
tctx               89 crypto/essiv.c 	crypto_cipher_clear_flags(tctx->essiv_cipher, CRYPTO_TFM_REQ_MASK);
tctx               90 crypto/essiv.c 	crypto_cipher_set_flags(tctx->essiv_cipher,
tctx               93 crypto/essiv.c 	err = crypto_cipher_setkey(tctx->essiv_cipher, salt,
tctx               94 crypto/essiv.c 				   crypto_shash_digestsize(tctx->hash));
tctx               96 crypto/essiv.c 				  crypto_cipher_get_flags(tctx->essiv_cipher) &
tctx              105 crypto/essiv.c 	struct essiv_tfm_ctx *tctx = crypto_aead_ctx(tfm);
tctx              106 crypto/essiv.c 	SHASH_DESC_ON_STACK(desc, tctx->hash);
tctx              111 crypto/essiv.c 	crypto_aead_clear_flags(tctx->u.aead, CRYPTO_TFM_REQ_MASK);
tctx              112 crypto/essiv.c 	crypto_aead_set_flags(tctx->u.aead, crypto_aead_get_flags(tfm) &
tctx              114 crypto/essiv.c 	err = crypto_aead_setkey(tctx->u.aead, key, keylen);
tctx              115 crypto/essiv.c 	crypto_aead_set_flags(tfm, crypto_aead_get_flags(tctx->u.aead) &
tctx              125 crypto/essiv.c 	desc->tfm = tctx->hash;
tctx              132 crypto/essiv.c 	crypto_cipher_clear_flags(tctx->essiv_cipher, CRYPTO_TFM_REQ_MASK);
tctx              133 crypto/essiv.c 	crypto_cipher_set_flags(tctx->essiv_cipher, crypto_aead_get_flags(tfm) &
tctx              135 crypto/essiv.c 	err = crypto_cipher_setkey(tctx->essiv_cipher, salt,
tctx              136 crypto/essiv.c 				   crypto_shash_digestsize(tctx->hash));
tctx              137 crypto/essiv.c 	crypto_aead_set_flags(tfm, crypto_cipher_get_flags(tctx->essiv_cipher) &
tctx              146 crypto/essiv.c 	struct essiv_tfm_ctx *tctx = crypto_aead_ctx(tfm);
tctx              148 crypto/essiv.c 	return crypto_aead_setauthsize(tctx->u.aead, authsize);
tctx              161 crypto/essiv.c 	const struct essiv_tfm_ctx *tctx = crypto_skcipher_ctx(tfm);
tctx              164 crypto/essiv.c 	crypto_cipher_encrypt_one(tctx->essiv_cipher, req->iv, req->iv);
tctx              166 crypto/essiv.c 	skcipher_request_set_tfm(subreq, tctx->u.skcipher);
tctx              199 crypto/essiv.c 	const struct essiv_tfm_ctx *tctx = crypto_aead_ctx(tfm);
tctx              205 crypto/essiv.c 	crypto_cipher_encrypt_one(tctx->essiv_cipher, req->iv, req->iv);
tctx              218 crypto/essiv.c 		u8 *iv = (u8 *)aead_request_ctx(req) + tctx->ivoffset;
tctx              259 crypto/essiv.c 	aead_request_set_tfm(subreq, tctx->u.aead);
tctx              284 crypto/essiv.c 			  struct essiv_tfm_ctx *tctx)
tctx              300 crypto/essiv.c 	tctx->essiv_cipher = essiv_cipher;
tctx              301 crypto/essiv.c 	tctx->hash = hash;
tctx              314 crypto/essiv.c 	struct essiv_tfm_ctx *tctx = crypto_skcipher_ctx(tfm);
tctx              325 crypto/essiv.c 	err = essiv_init_tfm(ictx, tctx);
tctx              331 crypto/essiv.c 	tctx->u.skcipher = skcipher;
tctx              339 crypto/essiv.c 	struct essiv_tfm_ctx *tctx = crypto_aead_ctx(tfm);
tctx              354 crypto/essiv.c 	tctx->ivoffset = offsetof(struct essiv_aead_request_ctx, aead_req) +
tctx              356 crypto/essiv.c 	crypto_aead_set_reqsize(tfm, tctx->ivoffset + crypto_aead_ivsize(aead));
tctx              358 crypto/essiv.c 	err = essiv_init_tfm(ictx, tctx);
tctx              364 crypto/essiv.c 	tctx->u.aead = aead;
tctx              370 crypto/essiv.c 	struct essiv_tfm_ctx *tctx = crypto_skcipher_ctx(tfm);
tctx              372 crypto/essiv.c 	crypto_free_skcipher(tctx->u.skcipher);
tctx              373 crypto/essiv.c 	crypto_free_cipher(tctx->essiv_cipher);
tctx              374 crypto/essiv.c 	crypto_free_shash(tctx->hash);
tctx              379 crypto/essiv.c 	struct essiv_tfm_ctx *tctx = crypto_aead_ctx(tfm);
tctx              381 crypto/essiv.c 	crypto_free_aead(tctx->u.aead);
tctx              382 crypto/essiv.c 	crypto_free_cipher(tctx->essiv_cipher);
tctx              383 crypto/essiv.c 	crypto_free_shash(tctx->hash);
tctx              462 crypto/tgr192.c static void tgr192_transform(struct tgr192_ctx *tctx, const u8 * data)
tctx              472 crypto/tgr192.c 	a = aa = tctx->a;
tctx              473 crypto/tgr192.c 	b = bb = tctx->b;
tctx              474 crypto/tgr192.c 	c = cc = tctx->c;
tctx              488 crypto/tgr192.c 	tctx->a = a;
tctx              489 crypto/tgr192.c 	tctx->b = b;
tctx              490 crypto/tgr192.c 	tctx->c = c;
tctx              495 crypto/tgr192.c 	struct tgr192_ctx *tctx = shash_desc_ctx(desc);
tctx              497 crypto/tgr192.c 	tctx->a = 0x0123456789abcdefULL;
tctx              498 crypto/tgr192.c 	tctx->b = 0xfedcba9876543210ULL;
tctx              499 crypto/tgr192.c 	tctx->c = 0xf096a5b4c3b2e187ULL;
tctx              500 crypto/tgr192.c 	tctx->nblocks = 0;
tctx              501 crypto/tgr192.c 	tctx->count = 0;
tctx              512 crypto/tgr192.c 	struct tgr192_ctx *tctx = shash_desc_ctx(desc);
tctx              514 crypto/tgr192.c 	if (tctx->count == 64) {	/* flush the buffer */
tctx              515 crypto/tgr192.c 		tgr192_transform(tctx, tctx->hash);
tctx              516 crypto/tgr192.c 		tctx->count = 0;
tctx              517 crypto/tgr192.c 		tctx->nblocks++;
tctx              522 crypto/tgr192.c 	if (tctx->count) {
tctx              523 crypto/tgr192.c 		for (; len && tctx->count < 64; len--) {
tctx              524 crypto/tgr192.c 			tctx->hash[tctx->count++] = *inbuf++;
tctx              534 crypto/tgr192.c 		tgr192_transform(tctx, inbuf);
tctx              535 crypto/tgr192.c 		tctx->count = 0;
tctx              536 crypto/tgr192.c 		tctx->nblocks++;
tctx              540 crypto/tgr192.c 	for (; len && tctx->count < 64; len--) {
tctx              541 crypto/tgr192.c 		tctx->hash[tctx->count++] = *inbuf++;
tctx              552 crypto/tgr192.c 	struct tgr192_ctx *tctx = shash_desc_ctx(desc);
tctx              561 crypto/tgr192.c 	t = tctx->nblocks;
tctx              567 crypto/tgr192.c 	if ((lsb = t + tctx->count) < t) {	/* add the count */
tctx              576 crypto/tgr192.c 	if (tctx->count < 56) {	/* enough room */
tctx              577 crypto/tgr192.c 		tctx->hash[tctx->count++] = 0x01;	/* pad */
tctx              578 crypto/tgr192.c 		while (tctx->count < 56) {
tctx              579 crypto/tgr192.c 			tctx->hash[tctx->count++] = 0;	/* pad */
tctx              582 crypto/tgr192.c 		tctx->hash[tctx->count++] = 0x01;	/* pad character */
tctx              583 crypto/tgr192.c 		while (tctx->count < 64) {
tctx              584 crypto/tgr192.c 			tctx->hash[tctx->count++] = 0;
tctx              587 crypto/tgr192.c 		memset(tctx->hash, 0, 56);    /* fill next block with zeroes */
tctx              590 crypto/tgr192.c 	le32p = (__le32 *)&tctx->hash[56];
tctx              594 crypto/tgr192.c 	tgr192_transform(tctx, tctx->hash);
tctx              596 crypto/tgr192.c 	be64p = (__be64 *)tctx->hash;
tctx              597 crypto/tgr192.c 	dst[0] = be64p[0] = cpu_to_be64(tctx->a);
tctx              598 crypto/tgr192.c 	dst[1] = be64p[1] = cpu_to_be64(tctx->b);
tctx              599 crypto/tgr192.c 	dst[2] = be64p[2] = cpu_to_be64(tctx->c);
tctx              398 crypto/vmac.c  static void vhash_blocks(const struct vmac_tfm_ctx *tctx,
tctx              402 crypto/vmac.c  	const u64 *kptr = tctx->nhkey;
tctx              403 crypto/vmac.c  	const u64 pkh = tctx->polykey[0];
tctx              404 crypto/vmac.c  	const u64 pkl = tctx->polykey[1];
tctx              432 crypto/vmac.c  	struct vmac_tfm_ctx *tctx = crypto_shash_ctx(tfm);
tctx              443 crypto/vmac.c  	err = crypto_cipher_setkey(tctx->cipher, key, keylen);
tctx              449 crypto/vmac.c  	for (i = 0; i < ARRAY_SIZE(tctx->nhkey); i += 2) {
tctx              450 crypto/vmac.c  		crypto_cipher_encrypt_one(tctx->cipher, (u8 *)out, in);
tctx              451 crypto/vmac.c  		tctx->nhkey[i] = be64_to_cpu(out[0]);
tctx              452 crypto/vmac.c  		tctx->nhkey[i+1] = be64_to_cpu(out[1]);
tctx              459 crypto/vmac.c  	for (i = 0; i < ARRAY_SIZE(tctx->polykey); i += 2) {
tctx              460 crypto/vmac.c  		crypto_cipher_encrypt_one(tctx->cipher, (u8 *)out, in);
tctx              461 crypto/vmac.c  		tctx->polykey[i] = be64_to_cpu(out[0]) & mpoly;
tctx              462 crypto/vmac.c  		tctx->polykey[i+1] = be64_to_cpu(out[1]) & mpoly;
tctx              469 crypto/vmac.c  	for (i = 0; i < ARRAY_SIZE(tctx->l3key); i += 2) {
tctx              471 crypto/vmac.c  			crypto_cipher_encrypt_one(tctx->cipher, (u8 *)out, in);
tctx              472 crypto/vmac.c  			tctx->l3key[i] = be64_to_cpu(out[0]);
tctx              473 crypto/vmac.c  			tctx->l3key[i+1] = be64_to_cpu(out[1]);
tctx              475 crypto/vmac.c  		} while (tctx->l3key[i] >= p64 || tctx->l3key[i+1] >= p64);
tctx              483 crypto/vmac.c  	const struct vmac_tfm_ctx *tctx = crypto_shash_ctx(desc->tfm);
tctx              488 crypto/vmac.c  	memcpy(dctx->polytmp, tctx->polykey, sizeof(dctx->polytmp));
tctx              495 crypto/vmac.c  	const struct vmac_tfm_ctx *tctx = crypto_shash_ctx(desc->tfm);
tctx              515 crypto/vmac.c  			vhash_blocks(tctx, dctx, dctx->partial_words, 1);
tctx              523 crypto/vmac.c  		vhash_blocks(tctx, dctx, (const __le64 *)p, n / VMAC_NHBYTES);
tctx              536 crypto/vmac.c  static u64 vhash_final(const struct vmac_tfm_ctx *tctx,
tctx              550 crypto/vmac.c  		nh_16(dctx->partial_words, tctx->nhkey, n / 8, rh, rl);
tctx              553 crypto/vmac.c  			poly_step(ch, cl, tctx->polykey[0], tctx->polykey[1],
tctx              560 crypto/vmac.c  	return l3hash(ch, cl, tctx->l3key[0], tctx->l3key[1], partial * 8);
tctx              565 crypto/vmac.c  	const struct vmac_tfm_ctx *tctx = crypto_shash_ctx(desc->tfm);
tctx              583 crypto/vmac.c  	hash = vhash_final(tctx, dctx);
tctx              589 crypto/vmac.c  	crypto_cipher_encrypt_one(tctx->cipher, dctx->nonce.bytes,
tctx              602 crypto/vmac.c  	struct vmac_tfm_ctx *tctx = crypto_tfm_ctx(tfm);
tctx              609 crypto/vmac.c  	tctx->cipher = cipher;
tctx              615 crypto/vmac.c  	struct vmac_tfm_ctx *tctx = crypto_tfm_ctx(tfm);
tctx              617 crypto/vmac.c  	crypto_free_cipher(tctx->cipher);
tctx               89 crypto/xcbc.c  	struct xcbc_tfm_ctx *tctx = crypto_shash_ctx(parent);
tctx               91 crypto/xcbc.c  	struct crypto_cipher *tfm = tctx->child;
tctx              135 crypto/xcbc.c  	struct xcbc_tfm_ctx *tctx = crypto_shash_ctx(parent);
tctx              137 crypto/xcbc.c  	struct crypto_cipher *tfm = tctx->child;
tctx              139 crypto/xcbc.c  	u8 *consts = PTR_ALIGN(&tctx->ctx[0], alignmask + 1);
tctx               23 crypto/xxhash_generic.c 	struct xxhash64_tfm_ctx *tctx = crypto_shash_ctx(tfm);
tctx               25 crypto/xxhash_generic.c 	if (keylen != sizeof(tctx->seed)) {
tctx               29 crypto/xxhash_generic.c 	tctx->seed = get_unaligned_le64(key);
tctx               35 crypto/xxhash_generic.c 	struct xxhash64_tfm_ctx *tctx = crypto_shash_ctx(desc->tfm);
tctx               38 crypto/xxhash_generic.c 	xxh64_reset(&dctx->xxhstate, tctx->seed);
tctx               65 crypto/xxhash_generic.c 	struct xxhash64_tfm_ctx *tctx = crypto_shash_ctx(desc->tfm);
tctx               67 crypto/xxhash_generic.c 	put_unaligned_le64(xxh64(data, length, tctx->seed), out);
tctx              404 drivers/crypto/atmel-sha.c static struct atmel_sha_dev *atmel_sha_find_dev(struct atmel_sha_ctx *tctx)
tctx              410 drivers/crypto/atmel-sha.c 	if (!tctx->dd) {
tctx              415 drivers/crypto/atmel-sha.c 		tctx->dd = dd;
tctx              417 drivers/crypto/atmel-sha.c 		dd = tctx->dd;
tctx              428 drivers/crypto/atmel-sha.c 	struct atmel_sha_ctx *tctx = crypto_ahash_ctx(tfm);
tctx              430 drivers/crypto/atmel-sha.c 	struct atmel_sha_dev *dd = atmel_sha_find_dev(tctx);
tctx             1150 drivers/crypto/atmel-sha.c 	struct atmel_sha_ctx *tctx = crypto_tfm_ctx(req->base.tfm);
tctx             1151 drivers/crypto/atmel-sha.c 	struct atmel_sha_dev *dd = tctx->dd;
tctx             2282 drivers/crypto/atmel-sha.c 	struct atmel_sha_ctx *tctx;
tctx             2316 drivers/crypto/atmel-sha.c 	tctx = crypto_ahash_ctx(tfm);
tctx             2317 drivers/crypto/atmel-sha.c 	tctx->start = atmel_sha_authenc_start;
tctx             2318 drivers/crypto/atmel-sha.c 	tctx->flags = mode;
tctx             2368 drivers/crypto/atmel-sha.c 	struct atmel_sha_ctx *tctx = crypto_ahash_ctx(tfm);
tctx             2375 drivers/crypto/atmel-sha.c 	dd = atmel_sha_find_dev(tctx);
tctx               69 drivers/crypto/geode-aes.c geode_aes_crypt(const struct geode_aes_tfm_ctx *tctx, const void *src,
tctx               96 drivers/crypto/geode-aes.c 	_writefield(AES_WRITEKEY0_REG, tctx->key);
tctx              112 drivers/crypto/geode-aes.c 	struct geode_aes_tfm_ctx *tctx = crypto_tfm_ctx(tfm);
tctx              115 drivers/crypto/geode-aes.c 	tctx->keylen = len;
tctx              118 drivers/crypto/geode-aes.c 		memcpy(tctx->key, key, len);
tctx              131 drivers/crypto/geode-aes.c 	tctx->fallback.cip->base.crt_flags &= ~CRYPTO_TFM_REQ_MASK;
tctx              132 drivers/crypto/geode-aes.c 	tctx->fallback.cip->base.crt_flags |=
tctx              135 drivers/crypto/geode-aes.c 	ret = crypto_cipher_setkey(tctx->fallback.cip, key, len);
tctx              138 drivers/crypto/geode-aes.c 		tfm->crt_flags |= (tctx->fallback.cip->base.crt_flags &
tctx              147 drivers/crypto/geode-aes.c 	struct geode_aes_tfm_ctx *tctx = crypto_skcipher_ctx(tfm);
tctx              150 drivers/crypto/geode-aes.c 	tctx->keylen = len;
tctx              153 drivers/crypto/geode-aes.c 		memcpy(tctx->key, key, len);
tctx              166 drivers/crypto/geode-aes.c 	crypto_skcipher_clear_flags(tctx->fallback.skcipher,
tctx              168 drivers/crypto/geode-aes.c 	crypto_skcipher_set_flags(tctx->fallback.skcipher,
tctx              171 drivers/crypto/geode-aes.c 	ret = crypto_skcipher_setkey(tctx->fallback.skcipher, key, len);
tctx              173 drivers/crypto/geode-aes.c 				  crypto_skcipher_get_flags(tctx->fallback.skcipher) &
tctx              181 drivers/crypto/geode-aes.c 	const struct geode_aes_tfm_ctx *tctx = crypto_tfm_ctx(tfm);
tctx              183 drivers/crypto/geode-aes.c 	if (unlikely(tctx->keylen != AES_KEYSIZE_128)) {
tctx              184 drivers/crypto/geode-aes.c 		crypto_cipher_encrypt_one(tctx->fallback.cip, out, in);
tctx              188 drivers/crypto/geode-aes.c 	geode_aes_crypt(tctx, in, out, AES_BLOCK_SIZE, NULL,
tctx              196 drivers/crypto/geode-aes.c 	const struct geode_aes_tfm_ctx *tctx = crypto_tfm_ctx(tfm);
tctx              198 drivers/crypto/geode-aes.c 	if (unlikely(tctx->keylen != AES_KEYSIZE_128)) {
tctx              199 drivers/crypto/geode-aes.c 		crypto_cipher_decrypt_one(tctx->fallback.cip, out, in);
tctx              203 drivers/crypto/geode-aes.c 	geode_aes_crypt(tctx, in, out, AES_BLOCK_SIZE, NULL,
tctx              210 drivers/crypto/geode-aes.c 	struct geode_aes_tfm_ctx *tctx = crypto_tfm_ctx(tfm);
tctx              212 drivers/crypto/geode-aes.c 	tctx->fallback.cip = crypto_alloc_cipher(name, 0,
tctx              215 drivers/crypto/geode-aes.c 	if (IS_ERR(tctx->fallback.cip)) {
tctx              217 drivers/crypto/geode-aes.c 		return PTR_ERR(tctx->fallback.cip);
tctx              225 drivers/crypto/geode-aes.c 	struct geode_aes_tfm_ctx *tctx = crypto_tfm_ctx(tfm);
tctx              227 drivers/crypto/geode-aes.c 	crypto_free_cipher(tctx->fallback.cip);
tctx              256 drivers/crypto/geode-aes.c 	struct geode_aes_tfm_ctx *tctx = crypto_skcipher_ctx(tfm);
tctx              258 drivers/crypto/geode-aes.c 	tctx->fallback.skcipher =
tctx              261 drivers/crypto/geode-aes.c 	if (IS_ERR(tctx->fallback.skcipher)) {
tctx              263 drivers/crypto/geode-aes.c 		return PTR_ERR(tctx->fallback.skcipher);
tctx              267 drivers/crypto/geode-aes.c 				    crypto_skcipher_reqsize(tctx->fallback.skcipher));
tctx              273 drivers/crypto/geode-aes.c 	struct geode_aes_tfm_ctx *tctx = crypto_skcipher_ctx(tfm);
tctx              275 drivers/crypto/geode-aes.c 	crypto_free_skcipher(tctx->fallback.skcipher);
tctx              281 drivers/crypto/geode-aes.c 	const struct geode_aes_tfm_ctx *tctx = crypto_skcipher_ctx(tfm);
tctx              286 drivers/crypto/geode-aes.c 	if (unlikely(tctx->keylen != AES_KEYSIZE_128)) {
tctx              290 drivers/crypto/geode-aes.c 		skcipher_request_set_tfm(subreq, tctx->fallback.skcipher);
tctx              300 drivers/crypto/geode-aes.c 		geode_aes_crypt(tctx, walk.src.virt.addr, walk.dst.virt.addr,
tctx              619 drivers/crypto/img-hash.c 	struct img_hash_ctx *tctx = crypto_ahash_ctx(tfm);
tctx              626 drivers/crypto/img-hash.c 	if (!tctx->hdev) {
tctx              631 drivers/crypto/img-hash.c 		tctx->hdev = hdev;
tctx              634 drivers/crypto/img-hash.c 		hdev = tctx->hdev;
tctx              667 drivers/crypto/img-hash.c 	err = img_hash_handle_queue(tctx->hdev, req);
tctx              717 drivers/crypto/img-hash.c 	struct img_hash_ctx *tctx = crypto_tfm_ctx(tfm);
tctx              719 drivers/crypto/img-hash.c 	crypto_free_ahash(tctx->fallback);
tctx              153 drivers/crypto/mediatek/mtk-sha.c static struct mtk_cryp *mtk_sha_find_dev(struct mtk_sha_ctx *tctx)
tctx              159 drivers/crypto/mediatek/mtk-sha.c 	if (!tctx->cryp) {
tctx              164 drivers/crypto/mediatek/mtk-sha.c 		tctx->cryp = cryp;
tctx              166 drivers/crypto/mediatek/mtk-sha.c 		cryp = tctx->cryp;
tctx              173 drivers/crypto/mediatek/mtk-sha.c 	tctx->id = cryp->rec;
tctx              358 drivers/crypto/mediatek/mtk-sha.c 	struct mtk_sha_ctx *tctx = crypto_tfm_ctx(req->base.tfm);
tctx              359 drivers/crypto/mediatek/mtk-sha.c 	struct mtk_sha_hmac_ctx *bctx = tctx->base;
tctx              375 drivers/crypto/mediatek/mtk-sha.c 	struct mtk_sha_ctx *tctx = crypto_ahash_ctx(tfm);
tctx              408 drivers/crypto/mediatek/mtk-sha.c 	ctx->buffer = tctx->buf;
tctx              410 drivers/crypto/mediatek/mtk-sha.c 	if (tctx->flags & SHA_FLAGS_HMAC) {
tctx              411 drivers/crypto/mediatek/mtk-sha.c 		struct mtk_sha_hmac_ctx *bctx = tctx->base;
tctx              706 drivers/crypto/mediatek/mtk-sha.c 	struct mtk_sha_ctx *tctx = crypto_tfm_ctx(req->base.tfm);
tctx              710 drivers/crypto/mediatek/mtk-sha.c 	return mtk_sha_handle_queue(tctx->cryp, tctx->id, req);
tctx              802 drivers/crypto/mediatek/mtk-sha.c 	struct mtk_sha_ctx *tctx = crypto_ahash_ctx(tfm);
tctx              803 drivers/crypto/mediatek/mtk-sha.c 	struct mtk_sha_hmac_ctx *bctx = tctx->base;
tctx              851 drivers/crypto/mediatek/mtk-sha.c 	struct mtk_sha_ctx *tctx = crypto_tfm_ctx(tfm);
tctx              854 drivers/crypto/mediatek/mtk-sha.c 	cryp = mtk_sha_find_dev(tctx);
tctx              862 drivers/crypto/mediatek/mtk-sha.c 		struct mtk_sha_hmac_ctx *bctx = tctx->base;
tctx              864 drivers/crypto/mediatek/mtk-sha.c 		tctx->flags |= SHA_FLAGS_HMAC;
tctx              909 drivers/crypto/mediatek/mtk-sha.c 	struct mtk_sha_ctx *tctx = crypto_tfm_ctx(tfm);
tctx              911 drivers/crypto/mediatek/mtk-sha.c 	if (tctx->flags & SHA_FLAGS_HMAC) {
tctx              912 drivers/crypto/mediatek/mtk-sha.c 		struct mtk_sha_hmac_ctx *bctx = tctx->base;
tctx              304 drivers/crypto/omap-sham.c 		struct omap_sham_ctx *tctx = crypto_ahash_ctx(tfm);
tctx              305 drivers/crypto/omap-sham.c 		struct omap_sham_hmac_ctx *bctx = tctx->base;
tctx              466 drivers/crypto/omap-sham.c 		struct omap_sham_ctx *tctx = crypto_ahash_ctx(tfm);
tctx              467 drivers/crypto/omap-sham.c 		struct omap_sham_hmac_ctx *bctx = tctx->base;
tctx              924 drivers/crypto/omap-sham.c 	struct omap_sham_ctx *tctx = crypto_ahash_ctx(tfm);
tctx              930 drivers/crypto/omap-sham.c 	if (!tctx->dd) {
tctx              935 drivers/crypto/omap-sham.c 		tctx->dd = dd;
tctx              937 drivers/crypto/omap-sham.c 		dd = tctx->dd;
tctx              981 drivers/crypto/omap-sham.c 	if (tctx->flags & BIT(FLAGS_HMAC)) {
tctx              983 drivers/crypto/omap-sham.c 			struct omap_sham_hmac_ctx *bctx = tctx->base;
tctx             1048 drivers/crypto/omap-sham.c 	struct omap_sham_ctx *tctx = crypto_tfm_ctx(req->base.tfm);
tctx             1049 drivers/crypto/omap-sham.c 	struct omap_sham_hmac_ctx *bctx = tctx->base;
tctx             1190 drivers/crypto/omap-sham.c 	struct omap_sham_ctx *tctx = crypto_tfm_ctx(req->base.tfm);
tctx             1191 drivers/crypto/omap-sham.c 	struct omap_sham_dev *dd = tctx->dd;
tctx             1231 drivers/crypto/omap-sham.c 	struct omap_sham_ctx *tctx = crypto_tfm_ctx(req->base.tfm);
tctx             1244 drivers/crypto/omap-sham.c 	return omap_sham_shash_digest(tctx->fallback, req->base.flags,
tctx             1301 drivers/crypto/omap-sham.c 	struct omap_sham_ctx *tctx = crypto_ahash_ctx(tfm);
tctx             1302 drivers/crypto/omap-sham.c 	struct omap_sham_hmac_ctx *bctx = tctx->base;
tctx             1309 drivers/crypto/omap-sham.c 	if (!tctx->dd) {
tctx             1314 drivers/crypto/omap-sham.c 		tctx->dd = dd;
tctx             1316 drivers/crypto/omap-sham.c 		dd = tctx->dd;
tctx             1320 drivers/crypto/omap-sham.c 	err = crypto_shash_setkey(tctx->fallback, key, keylen);
tctx             1351 drivers/crypto/omap-sham.c 	struct omap_sham_ctx *tctx = crypto_tfm_ctx(tfm);
tctx             1355 drivers/crypto/omap-sham.c 	tctx->fallback = crypto_alloc_shash(alg_name, 0,
tctx             1357 drivers/crypto/omap-sham.c 	if (IS_ERR(tctx->fallback)) {
tctx             1360 drivers/crypto/omap-sham.c 		return PTR_ERR(tctx->fallback);
tctx             1367 drivers/crypto/omap-sham.c 		struct omap_sham_hmac_ctx *bctx = tctx->base;
tctx             1368 drivers/crypto/omap-sham.c 		tctx->flags |= BIT(FLAGS_HMAC);
tctx             1374 drivers/crypto/omap-sham.c 			crypto_free_shash(tctx->fallback);
tctx             1420 drivers/crypto/omap-sham.c 	struct omap_sham_ctx *tctx = crypto_tfm_ctx(tfm);
tctx             1422 drivers/crypto/omap-sham.c 	crypto_free_shash(tctx->fallback);
tctx             1423 drivers/crypto/omap-sham.c 	tctx->fallback = NULL;
tctx             1425 drivers/crypto/omap-sham.c 	if (tctx->flags & BIT(FLAGS_HMAC)) {
tctx             1426 drivers/crypto/omap-sham.c 		struct omap_sham_hmac_ctx *bctx = tctx->base;
tctx              166 drivers/crypto/rockchip/rk3288_crypto_ahash.c 	struct rk_ahash_ctx *tctx = crypto_tfm_ctx(req->base.tfm);
tctx              167 drivers/crypto/rockchip/rk3288_crypto_ahash.c 	struct rk_crypto_info *dev = tctx->dev;
tctx              274 drivers/crypto/rockchip/rk3288_crypto_ahash.c 	struct rk_ahash_ctx *tctx = crypto_tfm_ctx(tfm);
tctx              282 drivers/crypto/rockchip/rk3288_crypto_ahash.c 	tctx->dev = algt->dev;
tctx              283 drivers/crypto/rockchip/rk3288_crypto_ahash.c 	tctx->dev->addr_vir = (void *)__get_free_page(GFP_KERNEL);
tctx              284 drivers/crypto/rockchip/rk3288_crypto_ahash.c 	if (!tctx->dev->addr_vir) {
tctx              285 drivers/crypto/rockchip/rk3288_crypto_ahash.c 		dev_err(tctx->dev->dev, "failed to kmalloc for addr_vir\n");
tctx              288 drivers/crypto/rockchip/rk3288_crypto_ahash.c 	tctx->dev->start = rk_ahash_start;
tctx              289 drivers/crypto/rockchip/rk3288_crypto_ahash.c 	tctx->dev->update = rk_ahash_crypto_rx;
tctx              290 drivers/crypto/rockchip/rk3288_crypto_ahash.c 	tctx->dev->complete = rk_ahash_crypto_complete;
tctx              293 drivers/crypto/rockchip/rk3288_crypto_ahash.c 	tctx->fallback_tfm = crypto_alloc_ahash(alg_name, 0,
tctx              295 drivers/crypto/rockchip/rk3288_crypto_ahash.c 	if (IS_ERR(tctx->fallback_tfm)) {
tctx              296 drivers/crypto/rockchip/rk3288_crypto_ahash.c 		dev_err(tctx->dev->dev, "Could not load fallback driver.\n");
tctx              297 drivers/crypto/rockchip/rk3288_crypto_ahash.c 		return PTR_ERR(tctx->fallback_tfm);
tctx              301 drivers/crypto/rockchip/rk3288_crypto_ahash.c 				 crypto_ahash_reqsize(tctx->fallback_tfm));
tctx              303 drivers/crypto/rockchip/rk3288_crypto_ahash.c 	return tctx->dev->enable_clk(tctx->dev);
tctx              308 drivers/crypto/rockchip/rk3288_crypto_ahash.c 	struct rk_ahash_ctx *tctx = crypto_tfm_ctx(tfm);
tctx              310 drivers/crypto/rockchip/rk3288_crypto_ahash.c 	free_page((unsigned long)tctx->dev->addr_vir);
tctx              311 drivers/crypto/rockchip/rk3288_crypto_ahash.c 	return tctx->dev->disable_clk(tctx->dev);
tctx             1490 drivers/crypto/s5p-sss.c 	struct s5p_hash_ctx *tctx = crypto_tfm_ctx(req->base.tfm);
tctx             1494 drivers/crypto/s5p-sss.c 	return s5p_hash_handle_queue(tctx->dd, req);
tctx             1547 drivers/crypto/s5p-sss.c 	struct s5p_hash_ctx *tctx = crypto_tfm_ctx(req->base.tfm);
tctx             1550 drivers/crypto/s5p-sss.c 	return s5p_hash_shash_digest(tctx->fallback, req->base.flags,
tctx             1628 drivers/crypto/s5p-sss.c 	struct s5p_hash_ctx *tctx = crypto_ahash_ctx(tfm);
tctx             1630 drivers/crypto/s5p-sss.c 	ctx->dd = tctx->dd;
tctx             1638 drivers/crypto/s5p-sss.c 	dev_dbg(tctx->dd->dev, "init: digest size: %d\n",
tctx             1679 drivers/crypto/s5p-sss.c 	struct s5p_hash_ctx *tctx = crypto_tfm_ctx(tfm);
tctx             1682 drivers/crypto/s5p-sss.c 	tctx->dd = s5p_dev;
tctx             1684 drivers/crypto/s5p-sss.c 	tctx->fallback = crypto_alloc_shash(alg_name, 0,
tctx             1686 drivers/crypto/s5p-sss.c 	if (IS_ERR(tctx->fallback)) {
tctx             1688 drivers/crypto/s5p-sss.c 		return PTR_ERR(tctx->fallback);
tctx             1714 drivers/crypto/s5p-sss.c 	struct s5p_hash_ctx *tctx = crypto_tfm_ctx(tfm);
tctx             1716 drivers/crypto/s5p-sss.c 	crypto_free_shash(tctx->fallback);
tctx             1717 drivers/crypto/s5p-sss.c 	tctx->fallback = NULL;
tctx             1743 drivers/crypto/s5p-sss.c 	struct s5p_hash_ctx *tctx = crypto_ahash_ctx(tfm);
tctx             1752 drivers/crypto/s5p-sss.c 	ctx->dd = tctx->dd;
tctx              477 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 	struct ttm_operation_ctx tctx = { false, false };
tctx              500 drivers/gpu/drm/amd/amdgpu/amdgpu_uvd.c 		r = ttm_bo_validate(&bo->tbo, &bo->placement, &tctx);