dsg 281 drivers/crypto/mediatek/mtk-aes.c struct scatterlist *ssg = aes->src.sg, *dsg = aes->dst.sg; dsg 306 drivers/crypto/mediatek/mtk-aes.c for (nents = 0; nents < dlen; ++nents, dsg = sg_next(dsg)) { dsg 308 drivers/crypto/mediatek/mtk-aes.c res->hdr = MTK_DESC_BUF_LEN(dsg->length); dsg 309 drivers/crypto/mediatek/mtk-aes.c res->buf = cpu_to_le32(sg_dma_address(dsg)); dsg 1266 drivers/dma/amba-pl08x.c struct pl08x_sg *dsg; dsg 1284 drivers/dma/amba-pl08x.c list_for_each_entry(dsg, &txd->dsg_list, node) { dsg 1288 drivers/dma/amba-pl08x.c bd.srcbus.addr = dsg->src_addr; dsg 1289 drivers/dma/amba-pl08x.c bd.dstbus.addr = dsg->dst_addr; dsg 1290 drivers/dma/amba-pl08x.c bd.remainder = dsg->len; dsg 1456 drivers/dma/amba-pl08x.c if (total_bytes != dsg->len) { dsg 1459 drivers/dma/amba-pl08x.c __func__, total_bytes, dsg->len); dsg 1495 drivers/dma/amba-pl08x.c struct pl08x_sg *dsg, *_dsg; dsg 1500 drivers/dma/amba-pl08x.c list_for_each_entry_safe(dsg, _dsg, &txd->dsg_list, node) { dsg 1501 drivers/dma/amba-pl08x.c list_del(&dsg->node); dsg 1502 drivers/dma/amba-pl08x.c kfree(dsg); dsg 1581 drivers/dma/amba-pl08x.c struct pl08x_sg *dsg; dsg 1583 drivers/dma/amba-pl08x.c list_for_each_entry(dsg, &txd->dsg_list, node) dsg 1584 drivers/dma/amba-pl08x.c bytes += dsg->len; dsg 1896 drivers/dma/amba-pl08x.c struct pl08x_sg *dsg; dsg 1906 drivers/dma/amba-pl08x.c dsg = kzalloc(sizeof(struct pl08x_sg), GFP_NOWAIT); dsg 1907 drivers/dma/amba-pl08x.c if (!dsg) { dsg 1911 drivers/dma/amba-pl08x.c list_add_tail(&dsg->node, &txd->dsg_list); dsg 1913 drivers/dma/amba-pl08x.c dsg->src_addr = src; dsg 1914 drivers/dma/amba-pl08x.c dsg->dst_addr = dest; dsg 1915 drivers/dma/amba-pl08x.c dsg->len = len; dsg 2029 drivers/dma/amba-pl08x.c struct pl08x_sg *dsg; dsg 2031 drivers/dma/amba-pl08x.c dsg = kzalloc(sizeof(struct pl08x_sg), GFP_NOWAIT); dsg 2032 drivers/dma/amba-pl08x.c if (!dsg) dsg 2035 drivers/dma/amba-pl08x.c list_add_tail(&dsg->node, &txd->dsg_list); dsg 2037 drivers/dma/amba-pl08x.c dsg->len = len; dsg 2039 drivers/dma/amba-pl08x.c dsg->src_addr = buf_addr; dsg 2040 drivers/dma/amba-pl08x.c dsg->dst_addr = slave_addr; dsg 2042 drivers/dma/amba-pl08x.c dsg->src_addr = slave_addr; dsg 2043 drivers/dma/amba-pl08x.c dsg->dst_addr = buf_addr; dsg 533 drivers/dma/dma-axi-dmac.c struct axi_dmac_sg *dsg; dsg 549 drivers/dma/dma-axi-dmac.c dsg = desc->sg; dsg 558 drivers/dma/dma-axi-dmac.c dsg = axi_dmac_fill_linear_sg(chan, direction, sg_dma_address(sg), 1, dsg 559 drivers/dma/dma-axi-dmac.c sg_dma_len(sg), dsg); dsg 429 drivers/dma/s3c24xx-dma.c struct s3c24xx_sg *dsg, *_dsg; dsg 431 drivers/dma/s3c24xx-dma.c list_for_each_entry_safe(dsg, _dsg, &txd->dsg_list, node) { dsg 432 drivers/dma/s3c24xx-dma.c list_del(&dsg->node); dsg 433 drivers/dma/s3c24xx-dma.c kfree(dsg); dsg 445 drivers/dma/s3c24xx-dma.c struct s3c24xx_sg *dsg = list_entry(txd->at, struct s3c24xx_sg, node); dsg 452 drivers/dma/s3c24xx-dma.c dcon |= S3C24XX_DCON_DSZ_BYTE | dsg->len; dsg 455 drivers/dma/s3c24xx-dma.c dcon |= S3C24XX_DCON_DSZ_HALFWORD | (dsg->len / 2); dsg 458 drivers/dma/s3c24xx-dma.c dcon |= S3C24XX_DCON_DSZ_WORD | (dsg->len / 4); dsg 483 drivers/dma/s3c24xx-dma.c writel_relaxed(dsg->src_addr, phy->base + S3C24XX_DISRC); dsg 485 drivers/dma/s3c24xx-dma.c writel_relaxed(dsg->dst_addr, phy->base + S3C24XX_DIDST); dsg 762 drivers/dma/s3c24xx-dma.c struct s3c24xx_sg *dsg; dsg 785 drivers/dma/s3c24xx-dma.c list_for_each_entry(dsg, &txd->dsg_list, node) dsg 786 drivers/dma/s3c24xx-dma.c bytes += dsg->len; dsg 794 drivers/dma/s3c24xx-dma.c dsg = list_entry(txd->at, struct s3c24xx_sg, node); dsg 795 drivers/dma/s3c24xx-dma.c list_for_each_entry_from(dsg, &txd->dsg_list, node) dsg 796 drivers/dma/s3c24xx-dma.c bytes += dsg->len; dsg 822 drivers/dma/s3c24xx-dma.c struct s3c24xx_sg *dsg; dsg 837 drivers/dma/s3c24xx-dma.c dsg = kzalloc(sizeof(*dsg), GFP_NOWAIT); dsg 838 drivers/dma/s3c24xx-dma.c if (!dsg) { dsg 842 drivers/dma/s3c24xx-dma.c list_add_tail(&dsg->node, &txd->dsg_list); dsg 844 drivers/dma/s3c24xx-dma.c dsg->src_addr = src; dsg 845 drivers/dma/s3c24xx-dma.c dsg->dst_addr = dest; dsg 846 drivers/dma/s3c24xx-dma.c dsg->len = len; dsg 886 drivers/dma/s3c24xx-dma.c struct s3c24xx_sg *dsg; dsg 951 drivers/dma/s3c24xx-dma.c dsg = kzalloc(sizeof(*dsg), GFP_NOWAIT); dsg 952 drivers/dma/s3c24xx-dma.c if (!dsg) { dsg 956 drivers/dma/s3c24xx-dma.c list_add_tail(&dsg->node, &txd->dsg_list); dsg 958 drivers/dma/s3c24xx-dma.c dsg->len = period; dsg 961 drivers/dma/s3c24xx-dma.c dsg->len = size - period * i; dsg 963 drivers/dma/s3c24xx-dma.c dsg->src_addr = addr + period * i; dsg 964 drivers/dma/s3c24xx-dma.c dsg->dst_addr = slave_addr; dsg 966 drivers/dma/s3c24xx-dma.c dsg->src_addr = slave_addr; dsg 967 drivers/dma/s3c24xx-dma.c dsg->dst_addr = addr + period * i; dsg 984 drivers/dma/s3c24xx-dma.c struct s3c24xx_sg *dsg; dsg 1043 drivers/dma/s3c24xx-dma.c dsg = kzalloc(sizeof(*dsg), GFP_NOWAIT); dsg 1044 drivers/dma/s3c24xx-dma.c if (!dsg) { dsg 1048 drivers/dma/s3c24xx-dma.c list_add_tail(&dsg->node, &txd->dsg_list); dsg 1050 drivers/dma/s3c24xx-dma.c dsg->len = sg_dma_len(sg); dsg 1052 drivers/dma/s3c24xx-dma.c dsg->src_addr = sg_dma_address(sg); dsg 1053 drivers/dma/s3c24xx-dma.c dsg->dst_addr = slave_addr; dsg 1055 drivers/dma/s3c24xx-dma.c dsg->src_addr = slave_addr; dsg 1056 drivers/dma/s3c24xx-dma.c dsg->dst_addr = sg_dma_address(sg); dsg 1226 drivers/target/target_core_sbc.c struct scatterlist *dsg = cmd->t_data_sg, *psg; dsg 1234 drivers/target/target_core_sbc.c daddr = kmap_atomic(sg_page(dsg)) + dsg->offset; dsg 1241 drivers/target/target_core_sbc.c if (offset >= dsg->length) { dsg 1242 drivers/target/target_core_sbc.c offset -= dsg->length; dsg 1243 drivers/target/target_core_sbc.c kunmap_atomic(daddr - dsg->offset); dsg 1244 drivers/target/target_core_sbc.c dsg = sg_next(dsg); dsg 1245 drivers/target/target_core_sbc.c if (!dsg) { dsg 1249 drivers/target/target_core_sbc.c daddr = kmap_atomic(sg_page(dsg)) + dsg->offset; dsg 1253 drivers/target/target_core_sbc.c avail = min(block_size, dsg->length - offset); dsg 1256 drivers/target/target_core_sbc.c kunmap_atomic(daddr - dsg->offset); dsg 1257 drivers/target/target_core_sbc.c dsg = sg_next(dsg); dsg 1258 drivers/target/target_core_sbc.c if (!dsg) { dsg 1262 drivers/target/target_core_sbc.c daddr = kmap_atomic(sg_page(dsg)) + dsg->offset; dsg 1284 drivers/target/target_core_sbc.c kunmap_atomic(daddr - dsg->offset); dsg 1381 drivers/target/target_core_sbc.c struct scatterlist *dsg = cmd->t_data_sg; dsg 1391 drivers/target/target_core_sbc.c daddr = kmap_atomic(sg_page(dsg)) + dsg->offset; dsg 1399 drivers/target/target_core_sbc.c if (dsg_off >= dsg->length) { dsg 1400 drivers/target/target_core_sbc.c dsg_off -= dsg->length; dsg 1401 drivers/target/target_core_sbc.c kunmap_atomic(daddr - dsg->offset); dsg 1402 drivers/target/target_core_sbc.c dsg = sg_next(dsg); dsg 1403 drivers/target/target_core_sbc.c if (!dsg) { dsg 1407 drivers/target/target_core_sbc.c daddr = kmap_atomic(sg_page(dsg)) + dsg->offset; dsg 1422 drivers/target/target_core_sbc.c avail = min(block_size, dsg->length - dsg_off); dsg 1425 drivers/target/target_core_sbc.c kunmap_atomic(daddr - dsg->offset); dsg 1426 drivers/target/target_core_sbc.c dsg = sg_next(dsg); dsg 1427 drivers/target/target_core_sbc.c if (!dsg) { dsg 1431 drivers/target/target_core_sbc.c daddr = kmap_atomic(sg_page(dsg)) + dsg->offset; dsg 1440 drivers/target/target_core_sbc.c kunmap_atomic(daddr - dsg->offset); dsg 1451 drivers/target/target_core_sbc.c kunmap_atomic(daddr - dsg->offset); dsg 377 net/ipv4/esp4.c struct scatterlist *sg, *dsg; dsg 403 net/ipv4/esp4.c dsg = sg; dsg 405 net/ipv4/esp4.c dsg = &sg[esp->nfrags]; dsg 438 net/ipv4/esp4.c sg_init_table(dsg, skb_shinfo(skb)->nr_frags + 1); dsg 439 net/ipv4/esp4.c err = skb_to_sgvec(skb, dsg, dsg 451 net/ipv4/esp4.c aead_request_set_crypt(req, sg, dsg, ivlen + esp->clen, iv); dsg 474 net/ipv4/esp4.c if (sg != dsg) dsg 320 net/ipv6/esp6.c struct scatterlist *sg, *dsg; dsg 345 net/ipv6/esp6.c dsg = sg; dsg 347 net/ipv6/esp6.c dsg = &sg[esp->nfrags]; dsg 379 net/ipv6/esp6.c sg_init_table(dsg, skb_shinfo(skb)->nr_frags + 1); dsg 380 net/ipv6/esp6.c err = skb_to_sgvec(skb, dsg, dsg 392 net/ipv6/esp6.c aead_request_set_crypt(req, sg, dsg, ivlen + esp->clen, iv); dsg 415 net/ipv6/esp6.c if (sg != dsg)