Lines Matching refs:req
173 struct aead_request *req, in generate_pat() argument
203 if (!req->assoclen) { in generate_pat()
205 } else if (req->assoclen <= 14) { in generate_pat()
211 iauth_len = req->assoclen; in generate_pat()
212 } else if (req->assoclen <= 65280) { in generate_pat()
226 rc = generate_b0(iv, req->assoclen, authsize, nbytes, b0); in generate_pat()
236 if (req->assoclen <= 65280) { in generate_pat()
237 *(u16 *)b1 = (u16)req->assoclen; in generate_pat()
238 scatterwalk_map_and_copy(b1 + 2, req->assoc, 0, in generate_pat()
242 *(u32 *)&b1[2] = (u32)req->assoclen; in generate_pat()
243 scatterwalk_map_and_copy(b1 + 6, req->assoc, 0, in generate_pat()
249 if (!req->assoclen) { in generate_pat()
251 } else if (req->assoclen <= 14) { in generate_pat()
278 req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP); in generate_pat()
283 atomic64_add(req->assoclen, &(nx_ctx->stats->aes_bytes)); in generate_pat()
297 to_process = min_t(u32, req->assoclen - processed, in generate_pat()
302 req->assoc, processed, in generate_pat()
305 if ((to_process + processed) < req->assoclen) { in generate_pat()
320 req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP); in generate_pat()
331 atomic64_add(req->assoclen, in generate_pat()
335 } while (processed < req->assoclen); in generate_pat()
345 static int ccm_nx_decrypt(struct aead_request *req, in ccm_nx_decrypt() argument
348 struct nx_crypto_ctx *nx_ctx = crypto_tfm_ctx(req->base.tfm); in ccm_nx_decrypt()
350 unsigned int nbytes = req->cryptlen; in ccm_nx_decrypt()
351 unsigned int authsize = crypto_aead_authsize(crypto_aead_reqtfm(req)); in ccm_nx_decrypt()
363 req->src, nbytes, authsize, in ccm_nx_decrypt()
366 rc = generate_pat(desc->info, req, nx_ctx, authsize, nbytes, in ccm_nx_decrypt()
385 rc = nx_build_sg_lists(nx_ctx, desc, req->dst, req->src, in ccm_nx_decrypt()
392 req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP); in ccm_nx_decrypt()
422 static int ccm_nx_encrypt(struct aead_request *req, in ccm_nx_encrypt() argument
425 struct nx_crypto_ctx *nx_ctx = crypto_tfm_ctx(req->base.tfm); in ccm_nx_encrypt()
427 unsigned int nbytes = req->cryptlen; in ccm_nx_encrypt()
428 unsigned int authsize = crypto_aead_authsize(crypto_aead_reqtfm(req)); in ccm_nx_encrypt()
435 rc = generate_pat(desc->info, req, nx_ctx, authsize, nbytes, in ccm_nx_encrypt()
453 rc = nx_build_sg_lists(nx_ctx, desc, req->dst, req->src, in ccm_nx_encrypt()
460 req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP); in ccm_nx_encrypt()
486 req->dst, nbytes, authsize, in ccm_nx_encrypt()
494 static int ccm4309_aes_nx_encrypt(struct aead_request *req) in ccm4309_aes_nx_encrypt() argument
496 struct nx_crypto_ctx *nx_ctx = crypto_tfm_ctx(req->base.tfm); in ccm4309_aes_nx_encrypt()
497 struct nx_gcm_rctx *rctx = aead_request_ctx(req); in ccm4309_aes_nx_encrypt()
503 memcpy(iv + 4, req->iv, 8); in ccm4309_aes_nx_encrypt()
506 desc.tfm = (struct crypto_blkcipher *)req->base.tfm; in ccm4309_aes_nx_encrypt()
508 return ccm_nx_encrypt(req, &desc); in ccm4309_aes_nx_encrypt()
511 static int ccm_aes_nx_encrypt(struct aead_request *req) in ccm_aes_nx_encrypt() argument
516 desc.info = req->iv; in ccm_aes_nx_encrypt()
517 desc.tfm = (struct crypto_blkcipher *)req->base.tfm; in ccm_aes_nx_encrypt()
523 return ccm_nx_encrypt(req, &desc); in ccm_aes_nx_encrypt()
526 static int ccm4309_aes_nx_decrypt(struct aead_request *req) in ccm4309_aes_nx_decrypt() argument
528 struct nx_crypto_ctx *nx_ctx = crypto_tfm_ctx(req->base.tfm); in ccm4309_aes_nx_decrypt()
529 struct nx_gcm_rctx *rctx = aead_request_ctx(req); in ccm4309_aes_nx_decrypt()
535 memcpy(iv + 4, req->iv, 8); in ccm4309_aes_nx_decrypt()
538 desc.tfm = (struct crypto_blkcipher *)req->base.tfm; in ccm4309_aes_nx_decrypt()
540 return ccm_nx_decrypt(req, &desc); in ccm4309_aes_nx_decrypt()
543 static int ccm_aes_nx_decrypt(struct aead_request *req) in ccm_aes_nx_decrypt() argument
548 desc.info = req->iv; in ccm_aes_nx_decrypt()
549 desc.tfm = (struct crypto_blkcipher *)req->base.tfm; in ccm_aes_nx_decrypt()
555 return ccm_nx_decrypt(req, &desc); in ccm_aes_nx_decrypt()