out_sg 467 crypto/asymmetric_keys/asym_tpm.c struct scatterlist in_sg, out_sg; out_sg 495 crypto/asymmetric_keys/asym_tpm.c sg_init_one(&out_sg, out, params->out_len); out_sg 496 crypto/asymmetric_keys/asym_tpm.c akcipher_request_set_crypt(req, &in_sg, &out_sg, params->in_len, out_sg 170 crypto/asymmetric_keys/public_key.c struct scatterlist in_sg, out_sg; out_sg 211 crypto/asymmetric_keys/public_key.c sg_init_one(&out_sg, out, params->out_len); out_sg 212 crypto/asymmetric_keys/public_key.c akcipher_request_set_crypt(req, &in_sg, &out_sg, params->in_len, out_sg 101 crypto/rsa-pkcs1pad.c struct scatterlist in_sg[2], out_sg[1]; out_sg 359 crypto/rsa-pkcs1pad.c pkcs1pad_sg_set_buf(req_ctx->out_sg, req_ctx->out_buf, out_sg 368 crypto/rsa-pkcs1pad.c req_ctx->out_sg, req->src_len, out_sg 547 crypto/rsa-pkcs1pad.c pkcs1pad_sg_set_buf(req_ctx->out_sg, req_ctx->out_buf, out_sg 556 crypto/rsa-pkcs1pad.c req_ctx->out_sg, req->src_len, out_sg 115 drivers/crypto/atmel-tdes.c struct scatterlist *out_sg; out_sg 325 drivers/crypto/atmel-tdes.c dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_FROM_DEVICE); out_sg 332 drivers/crypto/atmel-tdes.c count = atmel_tdes_sg_copy(&dd->out_sg, &dd->out_offset, out_sg 515 drivers/crypto/atmel-tdes.c out = IS_ALIGNED((u32)dd->out_sg->offset, sizeof(u32)) && out_sg 516 drivers/crypto/atmel-tdes.c IS_ALIGNED(dd->out_sg->length, dd->ctx->block_size); out_sg 519 drivers/crypto/atmel-tdes.c if (sg_dma_len(dd->in_sg) != sg_dma_len(dd->out_sg)) out_sg 526 drivers/crypto/atmel-tdes.c count = min_t(size_t, count, sg_dma_len(dd->out_sg)); out_sg 534 drivers/crypto/atmel-tdes.c err = dma_map_sg(dd->dev, dd->out_sg, 1, out_sg 544 drivers/crypto/atmel-tdes.c addr_out = sg_dma_address(dd->out_sg); out_sg 568 drivers/crypto/atmel-tdes.c dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_TO_DEVICE); out_sg 621 drivers/crypto/atmel-tdes.c dd->out_sg = req->dst; out_sg 650 drivers/crypto/atmel-tdes.c dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_FROM_DEVICE); out_sg 657 drivers/crypto/atmel-tdes.c count = atmel_tdes_sg_copy(&dd->out_sg, &dd->out_offset, out_sg 1114 drivers/crypto/atmel-tdes.c dd->out_sg = sg_next(dd->out_sg); out_sg 1115 drivers/crypto/atmel-tdes.c if (!dd->in_sg || !dd->out_sg) out_sg 165 drivers/crypto/nx/nx-aes-ccm.c struct nx_sg *nx_outsg = nx_ctx->out_sg; out_sg 254 drivers/crypto/nx/nx-aes-ccm.c nx_ctx->op.outlen = (nx_ctx->out_sg - nx_outsg) * out_sg 251 drivers/crypto/nx/nx-aes-gcm.c struct nx_sg *in_sg, *out_sg; out_sg 278 drivers/crypto/nx/nx-aes-gcm.c out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *) out, &len, out_sg 285 drivers/crypto/nx/nx-aes-gcm.c nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); out_sg 63 drivers/crypto/nx/nx-aes-xcbc.c struct nx_sg *in_sg, *out_sg; out_sg 87 drivers/crypto/nx/nx-aes-xcbc.c out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *) keys, &len, out_sg 94 drivers/crypto/nx/nx-aes-xcbc.c nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); out_sg 115 drivers/crypto/nx/nx-aes-xcbc.c out_sg = nx_build_sg_list(nx_ctx->out_sg, out, &len, out_sg 122 drivers/crypto/nx/nx-aes-xcbc.c nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); out_sg 173 drivers/crypto/nx/nx-aes-xcbc.c struct nx_sg *out_sg; out_sg 202 drivers/crypto/nx/nx-aes-xcbc.c out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *)sctx->state, out_sg 210 drivers/crypto/nx/nx-aes-xcbc.c nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); out_sg 297 drivers/crypto/nx/nx-aes-xcbc.c struct nx_sg *in_sg, *out_sg; out_sg 333 drivers/crypto/nx/nx-aes-xcbc.c out_sg = nx_build_sg_list(nx_ctx->out_sg, out, &len, out_sg 342 drivers/crypto/nx/nx-aes-xcbc.c nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); out_sg 62 drivers/crypto/nx/nx-sha256.c struct nx_sg *out_sg; out_sg 93 drivers/crypto/nx/nx-sha256.c out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *)sctx->state, out_sg 95 drivers/crypto/nx/nx-sha256.c nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); out_sg 181 drivers/crypto/nx/nx-sha256.c struct nx_sg *in_sg, *out_sg; out_sg 219 drivers/crypto/nx/nx-sha256.c out_sg = nx_build_sg_list(nx_ctx->out_sg, out, &len, max_sg_len); out_sg 227 drivers/crypto/nx/nx-sha256.c nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); out_sg 62 drivers/crypto/nx/nx-sha512.c struct nx_sg *out_sg; out_sg 93 drivers/crypto/nx/nx-sha512.c out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *)sctx->state, out_sg 95 drivers/crypto/nx/nx-sha512.c nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); out_sg 184 drivers/crypto/nx/nx-sha512.c struct nx_sg *in_sg, *out_sg; out_sg 228 drivers/crypto/nx/nx-sha512.c out_sg = nx_build_sg_list(nx_ctx->out_sg, out, &len, out_sg 232 drivers/crypto/nx/nx-sha512.c nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); out_sg 269 drivers/crypto/nx/nx.c struct nx_sg *nx_outsg = nx_ctx->out_sg; out_sg 294 drivers/crypto/nx/nx.c nx_ctx->op.outlen = trim_sg_list(nx_ctx->out_sg, nx_outsg, delta, nbytes); out_sg 314 drivers/crypto/nx/nx.c nx_ctx->op.out = __pa(nx_ctx->out_sg); out_sg 322 drivers/crypto/nx/nx.c nx_ctx->op_aead.out = __pa(nx_ctx->out_sg); out_sg 676 drivers/crypto/nx/nx.c nx_ctx->out_sg = (struct nx_sg *)((u8 *)nx_ctx->in_sg + NX_PAGE_SIZE); out_sg 680 drivers/crypto/nx/nx.c (struct nx_csbcpb *)((u8 *)nx_ctx->out_sg + out_sg 752 drivers/crypto/nx/nx.c nx_ctx->out_sg = NULL; out_sg 127 drivers/crypto/nx/nx.h struct nx_sg *out_sg; /* aligned pointer into kmem to an sg list */ out_sg 34 drivers/crypto/omap-aes-gcm.c dd->out_sg = NULL; out_sg 51 drivers/crypto/omap-aes-gcm.c dma_sync_sg_for_device(dd->dev, dd->out_sg, dd->out_sg_len, out_sg 54 drivers/crypto/omap-aes-gcm.c dma_unmap_sg(dd->dev, dd->out_sg, dd->out_sg_len, DMA_FROM_DEVICE); out_sg 57 drivers/crypto/omap-aes-gcm.c omap_crypto_cleanup(dd->out_sg, dd->orig_out, out_sg 142 drivers/crypto/omap-aes-gcm.c dd->out_sg = req->dst; out_sg 145 drivers/crypto/omap-aes-gcm.c dd->out_sg = scatterwalk_ffwd(sg_arr, req->dst, assoclen); out_sg 148 drivers/crypto/omap-aes-gcm.c if (req->src == req->dst || dd->out_sg == sg_arr) out_sg 151 drivers/crypto/omap-aes-gcm.c ret = omap_crypto_align_sg(&dd->out_sg, cryptlen, out_sg 159 drivers/crypto/omap-aes-gcm.c dd->out_sg_len = sg_nents_for_len(dd->out_sg, clen); out_sg 269 drivers/crypto/omap-aes.c struct scatterlist *out_sg, out_sg 278 drivers/crypto/omap-aes.c scatterwalk_start(&dd->out_walk, dd->out_sg); out_sg 324 drivers/crypto/omap-aes.c tx_out = dmaengine_prep_slave_sg(dd->dma_lch_out, out_sg, out_sg_len, out_sg 364 drivers/crypto/omap-aes.c err = dma_map_sg(dd->dev, dd->out_sg, dd->out_sg_len, out_sg 372 drivers/crypto/omap-aes.c err = omap_aes_crypt_dma(dd, dd->in_sg, dd->out_sg, dd->in_sg_len, out_sg 376 drivers/crypto/omap-aes.c dma_unmap_sg(dd->dev, dd->out_sg, dd->out_sg_len, out_sg 433 drivers/crypto/omap-aes.c dd->out_sg = req->dst; out_sg 446 drivers/crypto/omap-aes.c ret = omap_crypto_align_sg(&dd->out_sg, dd->total, AES_BLOCK_SIZE, out_sg 456 drivers/crypto/omap-aes.c dd->out_sg_len = sg_nents_for_len(dd->out_sg, dd->total); out_sg 489 drivers/crypto/omap-aes.c dma_sync_sg_for_device(dd->dev, dd->out_sg, dd->out_sg_len, out_sg 492 drivers/crypto/omap-aes.c dma_unmap_sg(dd->dev, dd->out_sg, dd->out_sg_len, out_sg 925 drivers/crypto/omap-aes.c BUG_ON(!dd->out_sg); out_sg 927 drivers/crypto/omap-aes.c BUG_ON(_calc_walked(out) > dd->out_sg->length); out_sg 929 drivers/crypto/omap-aes.c dst = sg_virt(dd->out_sg) + _calc_walked(out); out_sg 934 drivers/crypto/omap-aes.c if (dd->out_sg->length == _calc_walked(out)) { out_sg 935 drivers/crypto/omap-aes.c dd->out_sg = sg_next(dd->out_sg); out_sg 936 drivers/crypto/omap-aes.c if (dd->out_sg) { out_sg 938 drivers/crypto/omap-aes.c dd->out_sg); out_sg 939 drivers/crypto/omap-aes.c dst = sg_virt(dd->out_sg) + out_sg 179 drivers/crypto/omap-aes.h struct scatterlist *out_sg; out_sg 152 drivers/crypto/omap-des.c struct scatterlist *out_sg; out_sg 377 drivers/crypto/omap-des.c struct scatterlist *in_sg, struct scatterlist *out_sg, out_sg 388 drivers/crypto/omap-des.c scatterwalk_start(&dd->out_walk, dd->out_sg); out_sg 434 drivers/crypto/omap-des.c tx_out = dmaengine_prep_slave_sg(dd->dma_lch_out, out_sg, out_sg_len, out_sg 473 drivers/crypto/omap-des.c err = dma_map_sg(dd->dev, dd->out_sg, dd->out_sg_len, out_sg 481 drivers/crypto/omap-des.c err = omap_des_crypt_dma(tfm, dd->in_sg, dd->out_sg, dd->in_sg_len, out_sg 485 drivers/crypto/omap-des.c dma_unmap_sg(dd->dev, dd->out_sg, dd->out_sg_len, out_sg 544 drivers/crypto/omap-des.c dd->out_sg = req->dst; out_sg 557 drivers/crypto/omap-des.c ret = omap_crypto_align_sg(&dd->out_sg, dd->total, DES_BLOCK_SIZE, out_sg 567 drivers/crypto/omap-des.c dd->out_sg_len = sg_nents_for_len(dd->out_sg, dd->total); out_sg 603 drivers/crypto/omap-des.c dma_sync_sg_for_device(dd->dev, dd->out_sg, dd->out_sg_len, out_sg 606 drivers/crypto/omap-des.c dma_unmap_sg(dd->dev, dd->out_sg, dd->out_sg_len, out_sg 896 drivers/crypto/omap-des.c BUG_ON(!dd->out_sg); out_sg 898 drivers/crypto/omap-des.c BUG_ON(_calc_walked(out) > dd->out_sg->length); out_sg 900 drivers/crypto/omap-des.c dst = sg_virt(dd->out_sg) + _calc_walked(out); out_sg 905 drivers/crypto/omap-des.c if (dd->out_sg->length == _calc_walked(out)) { out_sg 906 drivers/crypto/omap-des.c dd->out_sg = sg_next(dd->out_sg); out_sg 907 drivers/crypto/omap-des.c if (dd->out_sg) { out_sg 909 drivers/crypto/omap-des.c dd->out_sg); out_sg 910 drivers/crypto/omap-des.c dst = sg_virt(dd->out_sg) + out_sg 223 drivers/crypto/sahara.c struct scatterlist *out_sg; out_sg 475 drivers/crypto/sahara.c dev->nb_out_sg = sg_nents_for_len(dev->out_sg, dev->total); out_sg 492 drivers/crypto/sahara.c ret = dma_map_sg(dev->device, dev->out_sg, dev->nb_out_sg, out_sg 515 drivers/crypto/sahara.c sg = dev->out_sg; out_sg 541 drivers/crypto/sahara.c dma_unmap_sg(dev->device, dev->out_sg, dev->nb_out_sg, out_sg 566 drivers/crypto/sahara.c dev->out_sg = req->dst; out_sg 592 drivers/crypto/sahara.c dma_unmap_sg(dev->device, dev->out_sg, dev->nb_out_sg, out_sg 108 drivers/crypto/stm32/stm32-cryp.c #define _walked_out (cryp->out_walk.offset - cryp->out_sg->offset) out_sg 152 drivers/crypto/stm32/stm32-cryp.c struct scatterlist *out_sg; out_sg 322 drivers/crypto/stm32/stm32-cryp.c ret = stm32_cryp_check_aligned(cryp->out_sg, cryp->total_out, out_sg 373 drivers/crypto/stm32/stm32-cryp.c cryp->out_sg_save = cryp->out_sg; out_sg 374 drivers/crypto/stm32/stm32-cryp.c cryp->out_sg = &cryp->out_sgl; out_sg 982 drivers/crypto/stm32/stm32-cryp.c cryp->out_sg = req ? req->dst : areq->dst; out_sg 983 drivers/crypto/stm32/stm32-cryp.c cryp->out_sg_save = cryp->out_sg; out_sg 992 drivers/crypto/stm32/stm32-cryp.c cryp->out_sg_len = sg_nents_for_len(cryp->out_sg, cryp->total_out); out_sg 1004 drivers/crypto/stm32/stm32-cryp.c scatterwalk_start(&cryp->out_walk, cryp->out_sg); out_sg 1074 drivers/crypto/stm32/stm32-cryp.c if (unlikely(cryp->out_sg->length == _walked_out)) { out_sg 1075 drivers/crypto/stm32/stm32-cryp.c cryp->out_sg = sg_next(cryp->out_sg); out_sg 1076 drivers/crypto/stm32/stm32-cryp.c if (cryp->out_sg) { out_sg 1077 drivers/crypto/stm32/stm32-cryp.c scatterwalk_start(&cryp->out_walk, cryp->out_sg); out_sg 1078 drivers/crypto/stm32/stm32-cryp.c return (sg_virt(cryp->out_sg) + _walked_out); out_sg 1159 drivers/crypto/stm32/stm32-cryp.c dst = sg_virt(cryp->out_sg) + _walked_out; out_sg 1246 drivers/crypto/stm32/stm32-cryp.c dst = sg_virt(cryp->out_sg) + _walked_out; out_sg 1318 drivers/crypto/stm32/stm32-cryp.c struct scatterlist *out_sg_ori = cryp->out_sg; out_sg 1412 drivers/crypto/stm32/stm32-cryp.c struct scatterlist *out_sg_ori = cryp->out_sg; out_sg 150 drivers/crypto/sunxi-ss/sun4i-ss-cipher.c struct scatterlist *out_sg = areq->dst; out_sg 195 drivers/crypto/sunxi-ss/sun4i-ss-cipher.c while (out_sg && no_chunk == 1) { out_sg 196 drivers/crypto/sunxi-ss/sun4i-ss-cipher.c if (out_sg->length % 4) out_sg 198 drivers/crypto/sunxi-ss/sun4i-ss-cipher.c out_sg = sg_next(out_sg); out_sg 1009 drivers/staging/wusbcore/wa-xfer.c struct scatterlist *out_sg; out_sg 1039 drivers/staging/wusbcore/wa-xfer.c out_sg = kmalloc((sizeof(struct scatterlist) * nents), GFP_ATOMIC); out_sg 1040 drivers/staging/wusbcore/wa-xfer.c if (out_sg) { out_sg 1041 drivers/staging/wusbcore/wa-xfer.c sg_init_table(out_sg, nents); out_sg 1045 drivers/staging/wusbcore/wa-xfer.c last_seg_sg = current_seg_sg = out_sg; out_sg 1079 drivers/staging/wusbcore/wa-xfer.c return out_sg; out_sg 17 lib/sg_split.c struct scatterlist *out_sg; out_sg 80 lib/sg_split.c struct scatterlist *in_sg, *out_sg; out_sg 85 lib/sg_split.c out_sg = split->out_sg; out_sg 86 lib/sg_split.c for (j = 0; j < split->nents; j++, out_sg++) { out_sg 87 lib/sg_split.c *out_sg = *in_sg; out_sg 89 lib/sg_split.c out_sg->offset += split->skip_sg0; out_sg 90 lib/sg_split.c out_sg->length -= split->skip_sg0; out_sg 92 lib/sg_split.c out_sg->offset = 0; out_sg 94 lib/sg_split.c sg_dma_address(out_sg) = 0; out_sg 95 lib/sg_split.c sg_dma_len(out_sg) = 0; out_sg 98 lib/sg_split.c out_sg[-1].length = split->length_last_sg; out_sg 99 lib/sg_split.c sg_mark_end(out_sg - 1); out_sg 106 lib/sg_split.c struct scatterlist *in_sg, *out_sg; out_sg 111 lib/sg_split.c out_sg = split->out_sg; out_sg 112 lib/sg_split.c for (j = 0; j < split->nents; j++, out_sg++) { out_sg 113 lib/sg_split.c sg_dma_address(out_sg) = sg_dma_address(in_sg); out_sg 114 lib/sg_split.c sg_dma_len(out_sg) = sg_dma_len(in_sg); out_sg 116 lib/sg_split.c sg_dma_address(out_sg) += split->skip_sg0; out_sg 117 lib/sg_split.c sg_dma_len(out_sg) -= split->skip_sg0; out_sg 121 lib/sg_split.c sg_dma_len(--out_sg) = split->length_last_sg; out_sg 168 lib/sg_split.c splitters[i].out_sg = kmalloc_array(splitters[i].nents, out_sg 171 lib/sg_split.c if (!splitters[i].out_sg) out_sg 188 lib/sg_split.c out[i] = splitters[i].out_sg; out_sg 198 lib/sg_split.c kfree(splitters[i].out_sg); out_sg 1392 net/tls/tls_sw.c struct scatterlist *out_sg, out_sg 1409 net/tls/tls_sw.c if (*zc && (out_iov || out_sg)) { out_sg 1413 net/tls/tls_sw.c n_sgout = sg_nents(out_sg); out_sg 1500 net/tls/tls_sw.c } else if (out_sg) { out_sg 1501 net/tls/tls_sw.c memcpy(sgout, out_sg, n_sgout * sizeof(*sgout)); out_sg 149 net/vmw_vsock/virtio_transport.c int ret, in_sg = 0, out_sg = 0; out_sg 168 net/vmw_vsock/virtio_transport.c sgs[out_sg++] = &hdr; out_sg 171 net/vmw_vsock/virtio_transport.c sgs[out_sg++] = &buf; out_sg 174 net/vmw_vsock/virtio_transport.c ret = virtqueue_add_sgs(vq, sgs, out_sg, in_sg, pkt, GFP_KERNEL);