Lines Matching refs:req

112 	struct icp_qat_fw_pke_request req;  member
120 struct qat_rsa_request *req = PTR_ALIGN(akcipher_request_ctx(areq), 64); in qat_rsa_cb() local
121 struct device *dev = &GET_DEV(req->ctx->inst->accel_dev); in qat_rsa_cb()
127 if (req->src_align) in qat_rsa_cb()
128 dma_free_coherent(dev, req->ctx->key_sz, req->src_align, in qat_rsa_cb()
129 req->in.enc.m); in qat_rsa_cb()
131 dma_unmap_single(dev, req->in.enc.m, req->ctx->key_sz, in qat_rsa_cb()
134 areq->dst_len = req->ctx->key_sz; in qat_rsa_cb()
135 if (req->dst_align) { in qat_rsa_cb()
136 char *ptr = req->dst_align; in qat_rsa_cb()
143 if (areq->dst_len != req->ctx->key_sz) in qat_rsa_cb()
144 memmove(req->dst_align, ptr, areq->dst_len); in qat_rsa_cb()
146 scatterwalk_map_and_copy(req->dst_align, areq->dst, 0, in qat_rsa_cb()
149 dma_free_coherent(dev, req->ctx->key_sz, req->dst_align, in qat_rsa_cb()
150 req->out.enc.c); in qat_rsa_cb()
162 dma_unmap_single(dev, req->out.enc.c, req->ctx->key_sz, in qat_rsa_cb()
166 dma_unmap_single(dev, req->phy_in, sizeof(struct qat_rsa_input_params), in qat_rsa_cb()
168 dma_unmap_single(dev, req->phy_out, in qat_rsa_cb()
240 static int qat_rsa_enc(struct akcipher_request *req) in qat_rsa_enc() argument
242 struct crypto_akcipher *tfm = crypto_akcipher_reqtfm(req); in qat_rsa_enc()
247 PTR_ALIGN(akcipher_request_ctx(req), 64); in qat_rsa_enc()
248 struct icp_qat_fw_pke_request *msg = &qat_req->req; in qat_rsa_enc()
254 if (req->dst_len < ctx->key_sz) { in qat_rsa_enc()
255 req->dst_len = ctx->key_sz; in qat_rsa_enc()
282 if (sg_is_last(req->src) && req->src_len == ctx->key_sz) { in qat_rsa_enc()
284 qat_req->in.enc.m = dma_map_single(dev, sg_virt(req->src), in qat_rsa_enc()
285 req->src_len, DMA_TO_DEVICE); in qat_rsa_enc()
290 int shift = ctx->key_sz - req->src_len; in qat_rsa_enc()
298 scatterwalk_map_and_copy(qat_req->src_align + shift, req->src, in qat_rsa_enc()
299 0, req->src_len, 0); in qat_rsa_enc()
301 if (sg_is_last(req->dst) && req->dst_len == ctx->key_sz) { in qat_rsa_enc()
303 qat_req->out.enc.c = dma_map_single(dev, sg_virt(req->dst), in qat_rsa_enc()
304 req->dst_len, in qat_rsa_enc()
334 msg->pke_mid.opaque = (uint64_t)(__force long)req; in qat_rsa_enc()
371 static int qat_rsa_dec(struct akcipher_request *req) in qat_rsa_dec() argument
373 struct crypto_akcipher *tfm = crypto_akcipher_reqtfm(req); in qat_rsa_dec()
378 PTR_ALIGN(akcipher_request_ctx(req), 64); in qat_rsa_dec()
379 struct icp_qat_fw_pke_request *msg = &qat_req->req; in qat_rsa_dec()
385 if (req->dst_len < ctx->key_sz) { in qat_rsa_dec()
386 req->dst_len = ctx->key_sz; in qat_rsa_dec()
413 if (sg_is_last(req->src) && req->src_len == ctx->key_sz) { in qat_rsa_dec()
415 qat_req->in.dec.c = dma_map_single(dev, sg_virt(req->src), in qat_rsa_dec()
416 req->dst_len, DMA_TO_DEVICE); in qat_rsa_dec()
421 int shift = ctx->key_sz - req->src_len; in qat_rsa_dec()
429 scatterwalk_map_and_copy(qat_req->src_align + shift, req->src, in qat_rsa_dec()
430 0, req->src_len, 0); in qat_rsa_dec()
432 if (sg_is_last(req->dst) && req->dst_len == ctx->key_sz) { in qat_rsa_dec()
434 qat_req->out.dec.m = dma_map_single(dev, sg_virt(req->dst), in qat_rsa_dec()
435 req->dst_len, in qat_rsa_dec()
466 msg->pke_mid.opaque = (uint64_t)(__force long)req; in qat_rsa_dec()