| /linux-4.4.14/net/sunrpc/xprtrdma/ |
| D | frwr_ops.c | 361 frmr->sg_nents = i; in frwr_op_map() 363 dma_nents = ib_dma_map_sg(device, frmr->sg, frmr->sg_nents, direction); in frwr_op_map() 366 __func__, frmr->sg, frmr->sg_nents); in frwr_op_map() 370 n = ib_map_mr_sg(mr, frmr->sg, frmr->sg_nents, PAGE_SIZE); in frwr_op_map() 371 if (unlikely(n != frmr->sg_nents)) { in frwr_op_map() 373 __func__, frmr->fr_mr, n, frmr->sg_nents); in frwr_op_map() 379 __func__, mw, frmr->sg_nents, mr->length); in frwr_op_map() 404 seg1->mr_nsegs = frmr->sg_nents; in frwr_op_map() 407 return frmr->sg_nents; in frwr_op_map() 440 ib_dma_unmap_sg(ia->ri_device, frmr->sg, frmr->sg_nents, seg1->mr_dir); in frwr_op_unmap()
|
| D | svc_rdma_recvfrom.c | 243 frmr->sg_nents = nents; in rdma_read_chunk_frmr() 275 frmr->sg, frmr->sg_nents, in rdma_read_chunk_frmr() 284 n = ib_map_mr_sg(frmr->mr, frmr->sg, frmr->sg_nents, PAGE_SIZE); in rdma_read_chunk_frmr() 285 if (unlikely(n != frmr->sg_nents)) { in rdma_read_chunk_frmr() 287 frmr->mr, n, frmr->sg_nents); in rdma_read_chunk_frmr() 350 frmr->sg, frmr->sg_nents, frmr->direction); in rdma_read_chunk_frmr()
|
| D | xprt_rdma.h | 205 int sg_nents; member
|
| D | svc_rdma_transport.c | 848 frmr->sg_nents = 0; in svc_rdma_get_frmr() 862 frmr->sg, frmr->sg_nents, frmr->direction); in svc_rdma_put_frmr()
|
| /linux-4.4.14/drivers/crypto/caam/ |
| D | sg_sw_sec4.h | 75 int sg_nents = sg_nents_for_len(sg_list, nbytes); in sg_count() local 77 if (likely(sg_nents == 1)) in sg_count() 80 return sg_nents; in sg_count()
|
| /linux-4.4.14/drivers/crypto/ |
| D | bfin_crc.c | 82 unsigned int sg_nents; member 143 if (sg_nents(req->src) > CRC_MAX_DMA_DESC) { in bfin_crypto_crc_init() 176 dma_map_sg(crc->dev, ctx->sg, ctx->sg_nents, DMA_TO_DEVICE); in bfin_crypto_crc_config_dma() 178 for_each_sg(ctx->sg, sg, ctx->sg_nents, j) { in bfin_crypto_crc_config_dma() 315 ctx->sg_nents = 0; in bfin_crypto_crc_handle_queue() 359 nsg = sg_nents(ctx->sg); in bfin_crypto_crc_handle_queue() 360 ctx->sg_nents = nsg; in bfin_crypto_crc_handle_queue() 375 ctx->sg_nents--; in bfin_crypto_crc_handle_queue()
|
| D | mxs-dcp.c | 259 const int nents = sg_nents(req->src); in mxs_dcp_aes_block_crypt() 568 const int nents = sg_nents(req->src); in dcp_sha_req_to_buf()
|
| D | img-hash.c | 256 ctx->bufcnt = sg_copy_to_buffer(hdev->req->src, sg_nents(ctx->sg), in img_hash_write_via_cpu() 639 ctx->nents = sg_nents(ctx->sg); in img_hash_digest()
|
| /linux-4.4.14/net/rds/ |
| D | iw_send.c | 766 int sg_nents) in rds_iw_build_send_reg() argument 770 n = ib_map_mr_sg(send->s_mr, sg, sg_nents, PAGE_SIZE); in rds_iw_build_send_reg() 771 if (unlikely(n != sg_nents)) in rds_iw_build_send_reg() 804 int sg_nents; in rds_iw_xmit_rdma() local 858 sg_nents = 0; in rds_iw_xmit_rdma() 902 sg_nents++; in rds_iw_xmit_rdma() 950 &op->op_sg[0], sg_nents); in rds_iw_xmit_rdma()
|
| /linux-4.4.14/drivers/ide/ |
| D | ide-dma.c | 144 i = dma_map_sg(hwif->dev, sg, cmd->sg_nents, cmd->sg_dma_direction); in ide_dma_map_sg() 146 cmd->orig_sg_nents = cmd->sg_nents; in ide_dma_map_sg() 147 cmd->sg_nents = i; in ide_dma_map_sg()
|
| D | ide-dma-sff.c | 123 for_each_sg(hwif->sg_table, sg, cmd->sg_nents, i) { in ide_build_dmatable()
|
| D | sgiioc4.c | 418 unsigned int count = 0, i = cmd->sg_nents; in sgiioc4_build_dmatable()
|
| D | au1xxx-ide.c | 220 int i = cmd->sg_nents, count = 0; in auide_build_dmatable()
|
| D | icside.c | 344 set_dma_sg(ec->dma, hwif->sg_table, cmd->sg_nents); in icside_dma_setup()
|
| D | tx4939ide.c | 243 for_each_sg(hwif->sg_table, sg, cmd->sg_nents, i) { in tx4939ide_build_dmatable()
|
| D | ide-io.c | 229 cmd->sg_nents = blk_rq_map_sg(drive->queue, rq, sg); in ide_map_sg()
|
| D | pmac.c | 1465 int i = cmd->sg_nents, count = 0; in pmac_ide_build_dmatable()
|
| /linux-4.4.14/drivers/crypto/sunxi-ss/ |
| D | sun4i-ss-cipher.c | 66 sg_miter_start(&mi, areq->src, sg_nents(areq->src), in sun4i_ss_opti_poll() 68 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst), in sun4i_ss_opti_poll() 199 sg_miter_start(&mi, areq->src, sg_nents(areq->src), in sun4i_ss_cipher_poll() 201 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst), in sun4i_ss_cipher_poll()
|
| D | sun4i-ss-hash.c | 200 copied = sg_pcopy_to_buffer(areq->src, sg_nents(areq->src), in sun4i_hash_update() 229 sg_miter_start(&mi, areq->src, sg_nents(areq->src), in sun4i_hash_update()
|
| /linux-4.4.14/drivers/usb/storage/ |
| D | protocol.c | 144 nents = sg_nents(sg); in usb_stor_access_xfer_buf()
|
| /linux-4.4.14/drivers/infiniband/hw/ocrdma/ |
| D | ocrdma_verbs.h | 130 int sg_nents);
|
| D | ocrdma_verbs.c | 3246 int sg_nents) in ocrdma_map_mr_sg() argument 3252 return ib_sg_to_pages(ibmr, sg, sg_nents, ocrdma_set_page); in ocrdma_map_mr_sg()
|
| /linux-4.4.14/include/linux/sunrpc/ |
| D | svc_rdma.h | 109 int sg_nents; member
|
| /linux-4.4.14/lib/ |
| D | scatterlist.c | 50 int sg_nents(struct scatterlist *sg) in sg_nents() function 57 EXPORT_SYMBOL(sg_nents);
|
| D | sg_split.c | 163 ret = sg_calculate_split(in, sg_nents(in), nb_splits, skip, split_sizes, in sg_split()
|
| /linux-4.4.14/drivers/infiniband/hw/qib/ |
| D | qib_mr.c | 369 int sg_nents) in qib_map_mr_sg() argument 375 return ib_sg_to_pages(ibmr, sg, sg_nents, qib_set_page); in qib_map_mr_sg()
|
| D | qib_verbs.h | 1051 int sg_nents);
|
| /linux-4.4.14/drivers/infiniband/ulp/isert/ |
| D | ib_isert.c | 1285 int imm_data, imm_data_len, unsol_data, sg_nents, rc; in isert_handle_scsi_cmd() local 1312 sg_nents = max(1UL, DIV_ROUND_UP(imm_data_len, PAGE_SIZE)); in isert_handle_scsi_cmd() 1313 sg_copy_from_buffer(cmd->se_cmd.t_data_sg, sg_nents, in isert_handle_scsi_cmd() 1316 sg_nents, imm_data_len); in isert_handle_scsi_cmd() 1355 int rc, sg_nents, sg_off, page_off; in isert_handle_iscsi_dataout() local 1378 sg_nents = max(1UL, DIV_ROUND_UP(unsol_data_len, PAGE_SIZE)); in isert_handle_iscsi_dataout() 1390 sg_nents, &rx_desc->data[0], unsol_data_len); in isert_handle_iscsi_dataout() 1392 sg_copy_from_buffer(sg_start, sg_nents, &rx_desc->data[0], in isert_handle_iscsi_dataout() 2376 int i = 0, sg_nents; in isert_build_rdma_wr() local 2380 sg_nents = min(cmd->se_cmd.t_data_nents - sg_off, isert_conn->max_sge); in isert_build_rdma_wr() [all …]
|
| /linux-4.4.14/include/linux/ |
| D | scatterlist.h | 244 int sg_nents(struct scatterlist *sg);
|
| D | ide.h | 311 int sg_nents; /* number of sg entries */ member
|
| /linux-4.4.14/drivers/infiniband/hw/mlx4/ |
| D | mr.c | 553 int sg_nents) in mlx4_ib_map_mr_sg() argument 564 rc = ib_sg_to_pages(ibmr, sg, sg_nents, mlx4_set_page); in mlx4_ib_map_mr_sg()
|
| D | mlx4_ib.h | 715 int sg_nents);
|
| /linux-4.4.14/drivers/infiniband/core/ |
| D | verbs.c | 1499 int sg_nents, in ib_map_mr_sg() argument 1507 return mr->device->map_mr_sg(mr, sg, sg_nents); in ib_map_mr_sg() 1529 int sg_nents, in ib_sg_to_pages() argument 1541 for_each_sg(sgl, sg, sg_nents, i) { in ib_sg_to_pages()
|
| /linux-4.4.14/include/rdma/ |
| D | ib_verbs.h | 1752 int sg_nents); 3031 int sg_nents, 3037 int sg_nents, in ib_map_mr_sg_zbva() argument 3042 n = ib_map_mr_sg(mr, sg, sg_nents, page_size); in ib_map_mr_sg_zbva() 3050 int sg_nents,
|
| /linux-4.4.14/lib/mpi/ |
| D | mpicoder.c | 446 ents = sg_nents(sgl); in mpi_read_raw_from_sgl()
|
| /linux-4.4.14/drivers/crypto/qat/qat_common/ |
| D | qat_algs.c | 669 int n = sg_nents(sgl); in qat_alg_sgl_to_bufl() 712 n = sg_nents(sglout); in qat_alg_sgl_to_bufl() 762 n = sg_nents(sglout); in qat_alg_sgl_to_bufl()
|
| /linux-4.4.14/crypto/ |
| D | algif_skcipher.c | 88 n = sg_nents(sgl); in skcipher_free_async_sgls() 494 nents += sg_nents(sg); in skcipher_all_sg_nents()
|
| /linux-4.4.14/drivers/memstick/core/ |
| D | ms_block.c | 104 sg_miter_start(&miter, sg, sg_nents(sg), in msb_sg_compare_to_buffer() 872 sg_miter_start(&miter, sg, sg_nents(sg), in msb_read_page() 1636 sg_copy_to_buffer(sg_tmp, sg_nents(sg_tmp), in msb_cache_write() 1659 sg_copy_from_buffer(sg_tmp, sg_nents(sg_tmp), in msb_cache_read()
|
| /linux-4.4.14/drivers/crypto/ccp/ |
| D | ccp-crypto-aes-cmac.c | 110 sg_count = (nbytes) ? sg_nents(req->src) + 2 : 2; in ccp_do_cmac_update()
|
| D | ccp-crypto-sha.c | 103 sg_count = sg_nents(req->src) + 1; in ccp_do_sha_update()
|
| /linux-4.4.14/drivers/infiniband/hw/cxgb4/ |
| D | mem.c | 939 int sg_nents) in c4iw_map_mr_sg() argument 945 return ib_sg_to_pages(ibmr, sg, sg_nents, c4iw_set_page); in c4iw_map_mr_sg()
|
| D | iw_cxgb4.h | 964 int sg_nents);
|
| /linux-4.4.14/drivers/infiniband/hw/mlx5/ |
| D | mlx5_ib.h | 560 int sg_nents);
|
| D | mr.c | 1455 int sg_nents) in mlx5_ib_map_mr_sg() argument 1466 n = ib_sg_to_pages(ibmr, sg, sg_nents, mlx5_set_page); in mlx5_ib_map_mr_sg()
|
| /linux-4.4.14/drivers/infiniband/ulp/srp/ |
| D | ib_srp.c | 1316 struct srp_rdma_ch *ch, int sg_nents) in srp_map_finish_fr() argument 1331 if (sg_nents == 0) in srp_map_finish_fr() 1334 if (sg_nents == 1 && target->global_mr) { in srp_map_finish_fr() 1348 n = ib_map_mr_sg(desc->mr, state->sg, sg_nents, dev->mr_page_size); in srp_map_finish_fr()
|
| /linux-4.4.14/drivers/infiniband/hw/cxgb3/ |
| D | iwch_provider.c | 878 int sg_nents) in iwch_map_mr_sg() argument 884 return ib_sg_to_pages(ibmr, sg, sg_nents, iwch_set_page); in iwch_map_mr_sg()
|
| /linux-4.4.14/drivers/infiniband/hw/nes/ |
| D | nes_verbs.c | 480 int sg_nents) in nes_map_mr_sg() argument 486 return ib_sg_to_pages(ibmr, sg, sg_nents, nes_set_page); in nes_map_mr_sg()
|
| /linux-4.4.14/drivers/gpu/drm/i915/ |
| D | i915_gem_gtt.c | 829 __sg_page_iter_start(&sg_iter, pages->sgl, sg_nents(pages->sgl), 0); in gen8_ppgtt_insert_entries()
|