| /linux-4.4.14/drivers/crypto/qce/ | 
| D | core.c | 62 static int qce_handle_request(struct crypto_async_request *async_req)  in qce_handle_request()  argument 66 	u32 type = crypto_tfm_alg_type(async_req->tfm);  in qce_handle_request() 72 		ret = ops->async_req_handle(async_req);  in qce_handle_request() 82 	struct crypto_async_request *async_req, *backlog;  in qce_handle_queue()  local 98 	async_req = crypto_dequeue_request(&qce->queue);  in qce_handle_queue() 99 	if (async_req)  in qce_handle_queue() 100 		qce->req = async_req;  in qce_handle_queue() 104 	if (!async_req)  in qce_handle_queue() 113 	err = qce_handle_request(async_req);  in qce_handle_queue()
  | 
| D | common.c | 230 static int qce_setup_regs_ahash(struct crypto_async_request *async_req,  in qce_setup_regs_ahash()  argument 233 	struct ahash_request *req = ahash_request_cast(async_req);  in qce_setup_regs_ahash() 234 	struct crypto_ahash *ahash = __crypto_ahash_cast(async_req->tfm);  in qce_setup_regs_ahash() 236 	struct qce_alg_template *tmpl = to_ahash_tmpl(async_req->tfm);  in qce_setup_regs_ahash() 239 	unsigned int blocksize = crypto_tfm_alg_blocksize(async_req->tfm);  in qce_setup_regs_ahash() 315 static int qce_setup_regs_ablkcipher(struct crypto_async_request *async_req,  in qce_setup_regs_ablkcipher()  argument 318 	struct ablkcipher_request *req = ablkcipher_request_cast(async_req);  in qce_setup_regs_ablkcipher() 320 	struct qce_cipher_ctx *ctx = crypto_tfm_ctx(async_req->tfm);  in qce_setup_regs_ablkcipher() 321 	struct qce_alg_template *tmpl = to_cipher_tmpl(async_req->tfm);  in qce_setup_regs_ablkcipher() 396 int qce_start(struct crypto_async_request *async_req, u32 type, u32 totallen,  in qce_start()  argument [all …] 
 | 
| D | ablkcipher.c | 27 	struct crypto_async_request *async_req = data;  in qce_ablkcipher_done()  local 28 	struct ablkcipher_request *req = ablkcipher_request_cast(async_req);  in qce_ablkcipher_done() 30 	struct qce_alg_template *tmpl = to_cipher_tmpl(async_req->tfm);  in qce_ablkcipher_done() 60 qce_ablkcipher_async_req_handle(struct crypto_async_request *async_req)  in qce_ablkcipher_async_req_handle()  argument 62 	struct ablkcipher_request *req = ablkcipher_request_cast(async_req);  in qce_ablkcipher_async_req_handle() 65 	struct qce_alg_template *tmpl = to_cipher_tmpl(async_req->tfm);  in qce_ablkcipher_async_req_handle() 128 			       qce_ablkcipher_done, async_req);  in qce_ablkcipher_async_req_handle() 134 	ret = qce_start(async_req, tmpl->crypto_alg_type, req->nbytes, 0);  in qce_ablkcipher_async_req_handle()
  | 
| D | sha.c | 39 	struct crypto_async_request *async_req = data;  in qce_ahash_done()  local 40 	struct ahash_request *req = ahash_request_cast(async_req);  in qce_ahash_done() 43 	struct qce_alg_template *tmpl = to_ahash_tmpl(async_req->tfm);  in qce_ahash_done() 76 static int qce_ahash_async_req_handle(struct crypto_async_request *async_req)  in qce_ahash_async_req_handle()  argument 78 	struct ahash_request *req = ahash_request_cast(async_req);  in qce_ahash_async_req_handle() 80 	struct qce_sha_ctx *ctx = crypto_tfm_ctx(async_req->tfm);  in qce_ahash_async_req_handle() 81 	struct qce_alg_template *tmpl = to_ahash_tmpl(async_req->tfm);  in qce_ahash_async_req_handle() 106 			       &rctx->result_sg, 1, qce_ahash_done, async_req);  in qce_ahash_async_req_handle() 112 	ret = qce_start(async_req, tmpl->crypto_alg_type, 0, 0);  in qce_ahash_async_req_handle()
  | 
| D | core.h | 65 	int (*async_req_handle)(struct crypto_async_request *async_req);
  | 
| D | common.h | 99 int qce_start(struct crypto_async_request *async_req, u32 type, u32 totallen,
  | 
| /linux-4.4.14/drivers/crypto/ccp/ | 
| D | ccp-crypto-aes.c | 25 static int ccp_aes_complete(struct crypto_async_request *async_req, int ret)  in ccp_aes_complete()  argument 27 	struct ablkcipher_request *req = ablkcipher_request_cast(async_req);  in ccp_aes_complete() 143 static int ccp_aes_rfc3686_complete(struct crypto_async_request *async_req,  in ccp_aes_rfc3686_complete()  argument 146 	struct ablkcipher_request *req = ablkcipher_request_cast(async_req);  in ccp_aes_rfc3686_complete() 152 	return ccp_aes_complete(async_req, ret);  in ccp_aes_rfc3686_complete()
  | 
| D | ccp-crypto-aes-xts.c | 84 static int ccp_aes_xts_complete(struct crypto_async_request *async_req, int ret)  in ccp_aes_xts_complete()  argument 86 	struct ablkcipher_request *req = ablkcipher_request_cast(async_req);  in ccp_aes_xts_complete()
  | 
| D | ccp-crypto-aes-cmac.c | 26 static int ccp_aes_cmac_complete(struct crypto_async_request *async_req,  in ccp_aes_cmac_complete()  argument 29 	struct ahash_request *req = ahash_request_cast(async_req);  in ccp_aes_cmac_complete()
  | 
| D | ccp-crypto-sha.c | 26 static int ccp_sha_complete(struct crypto_async_request *async_req, int ret)  in ccp_sha_complete()  argument 28 	struct ahash_request *req = ahash_request_cast(async_req);  in ccp_sha_complete()
  | 
| /linux-4.4.14/drivers/crypto/ | 
| D | mv_cesa.c | 602 		struct crypto_async_request *async_req = NULL;  in queue_manag()  local 613 			async_req = crypto_dequeue_request(&cpg->queue);  in queue_manag() 614 			if (async_req) {  in queue_manag() 626 		if (async_req) {  in queue_manag() 627 			if (crypto_tfm_alg_type(async_req->tfm) !=  in queue_manag() 630 				    ablkcipher_request_cast(async_req);  in queue_manag() 634 				    ahash_request_cast(async_req);  in queue_manag() 637 			async_req = NULL;  in queue_manag()
  | 
| D | bfin_crc.c | 282 	struct crypto_async_request *async_req, *backlog;  in bfin_crypto_crc_handle_queue()  local 299 	async_req = crypto_dequeue_request(&crc->queue);  in bfin_crypto_crc_handle_queue() 300 	if (async_req)  in bfin_crypto_crc_handle_queue() 304 	if (!async_req)  in bfin_crypto_crc_handle_queue() 310 	req = ahash_request_cast(async_req);  in bfin_crypto_crc_handle_queue()
  | 
| D | s5p-sss.c | 466 	struct crypto_async_request *async_req, *backlog;  in s5p_tasklet_cb()  local 472 	async_req = crypto_dequeue_request(&dev->queue);  in s5p_tasklet_cb() 474 	if (!async_req) {  in s5p_tasklet_cb() 484 	dev->req = ablkcipher_request_cast(async_req);  in s5p_tasklet_cb()
  | 
| D | sahara.c | 1038 	struct crypto_async_request *async_req;  in sahara_queue_manage()  local 1047 		async_req = crypto_dequeue_request(&dev->queue);  in sahara_queue_manage() 1053 		if (async_req) {  in sahara_queue_manage() 1054 			if (crypto_tfm_alg_type(async_req->tfm) ==  in sahara_queue_manage() 1057 					ahash_request_cast(async_req);  in sahara_queue_manage() 1062 					ablkcipher_request_cast(async_req);  in sahara_queue_manage() 1067 			async_req->complete(async_req, ret);  in sahara_queue_manage()
  | 
| D | img-hash.c | 497 	struct crypto_async_request *async_req, *backlog;  in img_hash_handle_queue()  local 513 	async_req = crypto_dequeue_request(&hdev->queue);  in img_hash_handle_queue() 514 	if (async_req)  in img_hash_handle_queue() 519 	if (!async_req)  in img_hash_handle_queue() 525 	req = ahash_request_cast(async_req);  in img_hash_handle_queue()
  | 
| D | omap-des.c | 592 	struct crypto_async_request *async_req, *backlog;  in omap_des_handle_queue()  local 606 	async_req = crypto_dequeue_request(&dd->queue);  in omap_des_handle_queue() 607 	if (async_req)  in omap_des_handle_queue() 611 	if (!async_req)  in omap_des_handle_queue() 617 	req = ablkcipher_request_cast(async_req);  in omap_des_handle_queue()
  | 
| D | omap-aes.c | 611 	struct crypto_async_request *async_req, *backlog;  in omap_aes_handle_queue()  local 625 	async_req = crypto_dequeue_request(&dd->queue);  in omap_aes_handle_queue() 626 	if (async_req)  in omap_aes_handle_queue() 630 	if (!async_req)  in omap_aes_handle_queue() 636 	req = ablkcipher_request_cast(async_req);  in omap_aes_handle_queue()
  | 
| D | atmel-aes.c | 565 	struct crypto_async_request *async_req, *backlog;  in atmel_aes_handle_queue()  local 579 	async_req = crypto_dequeue_request(&dd->queue);  in atmel_aes_handle_queue() 580 	if (async_req)  in atmel_aes_handle_queue() 584 	if (!async_req)  in atmel_aes_handle_queue() 590 	req = ablkcipher_request_cast(async_req);  in atmel_aes_handle_queue()
  | 
| D | atmel-tdes.c | 591 	struct crypto_async_request *async_req, *backlog;  in atmel_tdes_handle_queue()  local 605 	async_req = crypto_dequeue_request(&dd->queue);  in atmel_tdes_handle_queue() 606 	if (async_req)  in atmel_tdes_handle_queue() 610 	if (!async_req)  in atmel_tdes_handle_queue() 616 	req = ablkcipher_request_cast(async_req);  in atmel_tdes_handle_queue()
  | 
| D | atmel-sha.c | 832 	struct crypto_async_request *async_req, *backlog;  in atmel_sha_handle_queue()  local 847 	async_req = crypto_dequeue_request(&dd->queue);  in atmel_sha_handle_queue() 848 	if (async_req)  in atmel_sha_handle_queue() 853 	if (!async_req)  in atmel_sha_handle_queue() 859 	req = ahash_request_cast(async_req);  in atmel_sha_handle_queue()
  | 
| D | hifn_795x.c | 2018 	struct crypto_async_request *async_req;  in hifn_flush()  local 2034 	while ((async_req = crypto_dequeue_request(&dev->queue))) {  in hifn_flush() 2035 		req = container_of(async_req, struct ablkcipher_request, base);  in hifn_flush() 2137 	struct crypto_async_request *async_req, *backlog;  in hifn_process_queue()  local 2145 		async_req = crypto_dequeue_request(&dev->queue);  in hifn_process_queue() 2148 		if (!async_req)  in hifn_process_queue() 2154 		req = container_of(async_req, struct ablkcipher_request, base);  in hifn_process_queue()
  | 
| D | omap-sham.c | 1016 	struct crypto_async_request *async_req, *backlog;  in omap_sham_handle_queue()  local 1029 	async_req = crypto_dequeue_request(&dd->queue);  in omap_sham_handle_queue() 1030 	if (async_req)  in omap_sham_handle_queue() 1034 	if (!async_req)  in omap_sham_handle_queue() 1040 	req = ahash_request_cast(async_req);  in omap_sham_handle_queue()
  | 
| /linux-4.4.14/drivers/net/usb/ | 
| D | rtl8150.c | 144 struct async_req {  struct 172 	struct async_req *req = (struct async_req *)urb->context;  in async_set_reg_cb() 185 	struct async_req *req;  in async_set_registers() 187 	req = kmalloc(sizeof(struct async_req), GFP_ATOMIC);  in async_set_registers()
  | 
| /linux-4.4.14/drivers/crypto/amcc/ | 
| D | crypto4xx_core.h | 66 	struct crypto_async_request *async_req; 	/* base crypto request  member
  | 
| D | crypto4xx_core.c | 635 	ablk_req = ablkcipher_request_cast(pd_uinfo->async_req);  in crypto4xx_ablkcipher_done() 659 	ahash_req = ahash_request_cast(pd_uinfo->async_req);  in crypto4xx_ahash_done() 679 	if (crypto_tfm_alg_type(pd_uinfo->async_req->tfm) ==  in crypto4xx_pd_done() 837 	pd_uinfo->async_req = req;  in crypto4xx_build_pd()
  | 
| /linux-4.4.14/drivers/md/ | 
| D | dm-crypt.c | 883 static void kcryptd_async_done(struct crypto_async_request *async_req, 1354 static void kcryptd_async_done(struct crypto_async_request *async_req,  in kcryptd_async_done()  argument 1357 	struct dm_crypt_request *dmreq = async_req->data;  in kcryptd_async_done()
  |