/linux-4.1.27/drivers/crypto/nx/ |
H A D | nx-aes-xcbc.c | 75 struct nx_sg *in_sg, *out_sg; nx_xcbc_empty() local 99 out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *) keys, &len, nx_xcbc_empty() 106 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); nx_xcbc_empty() 128 out_sg = nx_build_sg_list(nx_ctx->out_sg, out, &len, nx_xcbc_empty() 135 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); nx_xcbc_empty() 187 struct nx_sg *out_sg; nx_xcbc_update() local 216 out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *)sctx->state, nx_xcbc_update() 224 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); nx_xcbc_update() 312 struct nx_sg *in_sg, *out_sg; nx_xcbc_final() local 348 out_sg = nx_build_sg_list(nx_ctx->out_sg, out, &len, nx_xcbc_final() 357 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); nx_xcbc_final()
|
H A D | nx-sha256.c | 74 struct nx_sg *out_sg; nx_sha256_update() local 105 out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *)sctx->state, nx_sha256_update() 107 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); nx_sha256_update() 194 struct nx_sg *in_sg, *out_sg; nx_sha256_final() local 232 out_sg = nx_build_sg_list(nx_ctx->out_sg, out, &len, max_sg_len); nx_sha256_final() 240 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); nx_sha256_final()
|
H A D | nx-sha512.c | 74 struct nx_sg *out_sg; nx_sha512_update() local 105 out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *)sctx->state, nx_sha512_update() 107 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); nx_sha512_update() 197 struct nx_sg *in_sg, *out_sg; nx_sha512_final() local 241 out_sg = nx_build_sg_list(nx_ctx->out_sg, out, &len, nx_sha512_final() 245 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); nx_sha512_final()
|
H A D | nx-aes-gcm.c | 272 struct nx_sg *in_sg, *out_sg; gcm_empty() local 299 out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *) out, &len, gcm_empty() 306 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); gcm_empty()
|
H A D | nx.c | 282 struct nx_sg *nx_outsg = nx_ctx->out_sg; nx_build_sg_lists() 307 nx_ctx->op.outlen = trim_sg_list(nx_ctx->out_sg, nx_outsg, delta, nbytes); nx_build_sg_lists() 327 nx_ctx->op.out = __pa(nx_ctx->out_sg); nx_ctx_init() 335 nx_ctx->op_aead.out = __pa(nx_ctx->out_sg); nx_ctx_init() 619 nx_ctx->out_sg = (struct nx_sg *)((u8 *)nx_ctx->in_sg + NX_PAGE_SIZE); nx_crypto_ctx_init() 623 (struct nx_csbcpb *)((u8 *)nx_ctx->out_sg + nx_crypto_ctx_init() 695 nx_ctx->out_sg = NULL; nx_crypto_ctx_exit()
|
H A D | nx.h | 137 struct nx_sg *out_sg; /* aligned pointer into kmem to an sg list */ member in struct:nx_crypto_ctx
|
H A D | nx-aes-ccm.c | 180 struct nx_sg *nx_outsg = nx_ctx->out_sg; generate_pat() 269 nx_ctx->op.outlen = (nx_ctx->out_sg - nx_outsg) * generate_pat()
|
/linux-4.1.27/drivers/crypto/ |
H A D | omap-aes.c | 170 struct scatterlist *out_sg; member in struct:omap_aes_dev 417 struct scatterlist *in_sg, struct scatterlist *out_sg, omap_aes_crypt_dma() 428 scatterwalk_start(&dd->out_walk, dd->out_sg); omap_aes_crypt_dma() 474 tx_out = dmaengine_prep_slave_sg(dd->dma_lch_out, out_sg, out_sg_len, omap_aes_crypt_dma() 513 err = dma_map_sg(dd->dev, dd->out_sg, dd->out_sg_len, omap_aes_crypt_dma_start() 521 err = omap_aes_crypt_dma(tfm, dd->in_sg, dd->out_sg, dd->in_sg_len, omap_aes_crypt_dma_start() 525 dma_unmap_sg(dd->dev, dd->out_sg, dd->out_sg_len, omap_aes_crypt_dma_start() 592 dd->orig_out = dd->out_sg; omap_aes_copy_sgs() 602 dd->out_sg = &dd->out_sgl; omap_aes_copy_sgs() 642 dd->out_sg = req->dst; omap_aes_handle_queue() 645 omap_aes_check_aligned(dd->out_sg, dd->total)) { omap_aes_handle_queue() 654 dd->out_sg_len = scatterwalk_bytes_sglen(dd->out_sg, dd->total); omap_aes_handle_queue() 686 dma_sync_sg_for_device(dd->dev, dd->out_sg, dd->out_sg_len, omap_aes_done_task() 689 dma_unmap_sg(dd->dev, dd->out_sg, dd->out_sg_len, omap_aes_done_task() 1027 BUG_ON(!dd->out_sg); omap_aes_irq() 1029 BUG_ON(_calc_walked(out) > dd->out_sg->length); omap_aes_irq() 1031 dst = sg_virt(dd->out_sg) + _calc_walked(out); omap_aes_irq() 1036 if (dd->out_sg->length == _calc_walked(out)) { omap_aes_irq() 1037 dd->out_sg = sg_next(dd->out_sg); omap_aes_irq() 1038 if (dd->out_sg) { omap_aes_irq() 1040 dd->out_sg); omap_aes_irq() 1041 dst = sg_virt(dd->out_sg) + omap_aes_irq() 416 omap_aes_crypt_dma(struct crypto_tfm *tfm, struct scatterlist *in_sg, struct scatterlist *out_sg, int in_sg_len, int out_sg_len) omap_aes_crypt_dma() argument
|
H A D | omap-des.c | 151 struct scatterlist *out_sg; member in struct:omap_des_dev 401 struct scatterlist *in_sg, struct scatterlist *out_sg, omap_des_crypt_dma() 412 scatterwalk_start(&dd->out_walk, dd->out_sg); omap_des_crypt_dma() 458 tx_out = dmaengine_prep_slave_sg(dd->dma_lch_out, out_sg, out_sg_len, omap_des_crypt_dma() 497 err = dma_map_sg(dd->dev, dd->out_sg, dd->out_sg_len, omap_des_crypt_dma_start() 505 err = omap_des_crypt_dma(tfm, dd->in_sg, dd->out_sg, dd->in_sg_len, omap_des_crypt_dma_start() 509 dma_unmap_sg(dd->dev, dd->out_sg, dd->out_sg_len, omap_des_crypt_dma_start() 574 dd->orig_out = dd->out_sg; omap_des_copy_sgs() 584 dd->out_sg = &dd->out_sgl; omap_des_copy_sgs() 624 dd->out_sg = req->dst; omap_des_handle_queue() 627 omap_des_copy_needed(dd->out_sg)) { omap_des_handle_queue() 636 dd->out_sg_len = scatterwalk_bytes_sglen(dd->out_sg, dd->total); omap_des_handle_queue() 668 dma_sync_sg_for_device(dd->dev, dd->out_sg, dd->out_sg_len, omap_des_done_task() 671 dma_unmap_sg(dd->dev, dd->out_sg, dd->out_sg_len, omap_des_done_task() 943 BUG_ON(!dd->out_sg); omap_des_irq() 945 BUG_ON(_calc_walked(out) > dd->out_sg->length); omap_des_irq() 947 dst = sg_virt(dd->out_sg) + _calc_walked(out); omap_des_irq() 952 if (dd->out_sg->length == _calc_walked(out)) { omap_des_irq() 953 dd->out_sg = sg_next(dd->out_sg); omap_des_irq() 954 if (dd->out_sg) { omap_des_irq() 956 dd->out_sg); omap_des_irq() 957 dst = sg_virt(dd->out_sg) + omap_des_irq() 400 omap_des_crypt_dma(struct crypto_tfm *tfm, struct scatterlist *in_sg, struct scatterlist *out_sg, int in_sg_len, int out_sg_len) omap_des_crypt_dma() argument
|
H A D | atmel-aes.c | 125 struct scatterlist *out_sg; member in struct:atmel_aes_dev 408 dd->nb_out_sg = atmel_aes_sg_length(dd->req, dd->out_sg); atmel_aes_crypt_cpu_start() 437 out = IS_ALIGNED((u32)dd->out_sg->offset, sizeof(u32)) && atmel_aes_crypt_dma_start() 438 IS_ALIGNED(dd->out_sg->length, dd->ctx->block_size); atmel_aes_crypt_dma_start() 441 if (sg_dma_len(dd->in_sg) != sg_dma_len(dd->out_sg)) atmel_aes_crypt_dma_start() 448 count = min(count, sg_dma_len(dd->out_sg)); atmel_aes_crypt_dma_start() 456 err = dma_map_sg(dd->dev, dd->out_sg, 1, atmel_aes_crypt_dma_start() 466 addr_out = sg_dma_address(dd->out_sg); atmel_aes_crypt_dma_start() 490 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_TO_DEVICE); atmel_aes_crypt_dma_start() 598 dd->out_sg = req->dst; atmel_aes_handle_queue() 631 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_FROM_DEVICE); atmel_aes_crypt_dma_stop() 638 count = atmel_aes_sg_copy(&dd->out_sg, &dd->out_offset, atmel_aes_crypt_dma_stop() 1167 if (sg_copy_from_buffer(dd->out_sg, dd->nb_out_sg, atmel_aes_done_task() 1183 dd->out_sg = sg_next(dd->out_sg); atmel_aes_done_task() 1184 if (!dd->in_sg || !dd->out_sg) atmel_aes_done_task()
|
H A D | atmel-tdes.c | 118 struct scatterlist *out_sg; member in struct:atmel_tdes_dev 328 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_FROM_DEVICE); atmel_tdes_crypt_pdc_stop() 335 count = atmel_tdes_sg_copy(&dd->out_sg, &dd->out_offset, atmel_tdes_crypt_pdc_stop() 518 out = IS_ALIGNED((u32)dd->out_sg->offset, sizeof(u32)) && atmel_tdes_crypt_start() 519 IS_ALIGNED(dd->out_sg->length, dd->ctx->block_size); atmel_tdes_crypt_start() 522 if (sg_dma_len(dd->in_sg) != sg_dma_len(dd->out_sg)) atmel_tdes_crypt_start() 529 count = min(count, sg_dma_len(dd->out_sg)); atmel_tdes_crypt_start() 537 err = dma_map_sg(dd->dev, dd->out_sg, 1, atmel_tdes_crypt_start() 547 addr_out = sg_dma_address(dd->out_sg); atmel_tdes_crypt_start() 571 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_TO_DEVICE); atmel_tdes_crypt_start() 624 dd->out_sg = req->dst; atmel_tdes_handle_queue() 653 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_FROM_DEVICE); atmel_tdes_crypt_dma_stop() 660 count = atmel_tdes_sg_copy(&dd->out_sg, &dd->out_offset, atmel_tdes_crypt_dma_stop() 1236 dd->out_sg = sg_next(dd->out_sg); atmel_tdes_done_task() 1237 if (!dd->in_sg || !dd->out_sg) atmel_tdes_done_task()
|
H A D | sahara.c | 234 struct scatterlist *out_sg; member in struct:sahara_dev 506 dev->nb_out_sg = sahara_sg_length(dev->out_sg, dev->total); sahara_hw_descriptor_create() 519 ret = dma_map_sg(dev->device, dev->out_sg, dev->nb_out_sg, sahara_hw_descriptor_create() 542 sg = dev->out_sg; sahara_hw_descriptor_create() 568 dma_unmap_sg(dev->device, dev->out_sg, dev->nb_out_sg, sahara_hw_descriptor_create() 593 dev->out_sg = req->dst; sahara_aes_process() 619 dma_unmap_sg(dev->device, dev->out_sg, dev->nb_out_sg, sahara_aes_process()
|
/linux-4.1.27/net/ceph/ |
H A D | crypto.c | 208 goto out_sg; ceph_aes_encrypt() 215 out_sg: ceph_aes_encrypt() 271 goto out_sg; ceph_aes_encrypt2() 278 out_sg: ceph_aes_encrypt2() 323 goto out_sg; ceph_aes_decrypt() 342 out_sg: ceph_aes_decrypt() 389 goto out_sg; ceph_aes_decrypt2() 419 out_sg: ceph_aes_decrypt2()
|
/linux-4.1.27/drivers/usb/wusbcore/ |
H A D | wa-xfer.c | 1023 struct scatterlist *out_sg; wa_xfer_create_subset_sg() local 1053 out_sg = kmalloc((sizeof(struct scatterlist) * nents), GFP_ATOMIC); wa_xfer_create_subset_sg() 1054 if (out_sg) { wa_xfer_create_subset_sg() 1055 sg_init_table(out_sg, nents); wa_xfer_create_subset_sg() 1059 last_seg_sg = current_seg_sg = out_sg; wa_xfer_create_subset_sg() 1093 return out_sg; wa_xfer_create_subset_sg()
|