/linux-4.4.14/drivers/crypto/caam/ |
H A D | sg_sw_sec4.h | 75 int sg_nents = sg_nents_for_len(sg_list, nbytes); sg_count() local 77 if (likely(sg_nents == 1)) sg_count() 80 return sg_nents; sg_count()
|
/linux-4.4.14/net/sunrpc/xprtrdma/ |
H A D | frwr_ops.c | 361 frmr->sg_nents = i; frwr_op_map() 363 dma_nents = ib_dma_map_sg(device, frmr->sg, frmr->sg_nents, direction); frwr_op_map() 365 pr_err("RPC: %s: failed to dma map sg %p sg_nents %u\n", frwr_op_map() 366 __func__, frmr->sg, frmr->sg_nents); frwr_op_map() 370 n = ib_map_mr_sg(mr, frmr->sg, frmr->sg_nents, PAGE_SIZE); frwr_op_map() 371 if (unlikely(n != frmr->sg_nents)) { frwr_op_map() 373 __func__, frmr->fr_mr, n, frmr->sg_nents); frwr_op_map() 379 __func__, mw, frmr->sg_nents, mr->length); frwr_op_map() 404 seg1->mr_nsegs = frmr->sg_nents; frwr_op_map() 407 return frmr->sg_nents; frwr_op_map() 440 ib_dma_unmap_sg(ia->ri_device, frmr->sg, frmr->sg_nents, seg1->mr_dir); frwr_op_unmap()
|
H A D | svc_rdma_recvfrom.c | 243 frmr->sg_nents = nents; rdma_read_chunk_frmr() 275 frmr->sg, frmr->sg_nents, rdma_read_chunk_frmr() 284 n = ib_map_mr_sg(frmr->mr, frmr->sg, frmr->sg_nents, PAGE_SIZE); rdma_read_chunk_frmr() 285 if (unlikely(n != frmr->sg_nents)) { rdma_read_chunk_frmr() 287 frmr->mr, n, frmr->sg_nents); rdma_read_chunk_frmr() 350 frmr->sg, frmr->sg_nents, frmr->direction); rdma_read_chunk_frmr()
|
H A D | svc_rdma_transport.c | 848 frmr->sg_nents = 0; svc_rdma_get_frmr() 862 frmr->sg, frmr->sg_nents, frmr->direction); svc_rdma_put_frmr()
|
H A D | xprt_rdma.h | 205 int sg_nents; member in struct:rpcrdma_frmr
|
/linux-4.4.14/drivers/crypto/ |
H A D | bfin_crc.c | 82 unsigned int sg_nents; member in struct:bfin_crypto_crc_reqctx 143 if (sg_nents(req->src) > CRC_MAX_DMA_DESC) { bfin_crypto_crc_init() 176 dma_map_sg(crc->dev, ctx->sg, ctx->sg_nents, DMA_TO_DEVICE); bfin_crypto_crc_config_dma() 178 for_each_sg(ctx->sg, sg, ctx->sg_nents, j) { bfin_crypto_crc_config_dma() 315 ctx->sg_nents = 0; bfin_crypto_crc_handle_queue() 359 nsg = sg_nents(ctx->sg); bfin_crypto_crc_handle_queue() 360 ctx->sg_nents = nsg; bfin_crypto_crc_handle_queue() 375 ctx->sg_nents--; bfin_crypto_crc_handle_queue()
|
H A D | img-hash.c | 256 ctx->bufcnt = sg_copy_to_buffer(hdev->req->src, sg_nents(ctx->sg), img_hash_write_via_cpu() 639 ctx->nents = sg_nents(ctx->sg); img_hash_digest()
|
H A D | mxs-dcp.c | 259 const int nents = sg_nents(req->src); mxs_dcp_aes_block_crypt() 568 const int nents = sg_nents(req->src); dcp_sha_req_to_buf()
|
/linux-4.4.14/drivers/crypto/sunxi-ss/ |
H A D | sun4i-ss-cipher.c | 66 sg_miter_start(&mi, areq->src, sg_nents(areq->src), sun4i_ss_opti_poll() 68 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst), sun4i_ss_opti_poll() 199 sg_miter_start(&mi, areq->src, sg_nents(areq->src), sun4i_ss_cipher_poll() 201 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst), sun4i_ss_cipher_poll()
|
H A D | sun4i-ss-hash.c | 200 copied = sg_pcopy_to_buffer(areq->src, sg_nents(areq->src), sun4i_hash_update() 229 sg_miter_start(&mi, areq->src, sg_nents(areq->src), sun4i_hash_update()
|
/linux-4.4.14/drivers/infiniband/core/ |
H A D | verbs.c | 1480 * @sg_nents: number of entries in sg 1489 * - If sg_nents total byte length exceeds the mr max_num_sge * page_size 1499 int sg_nents, ib_map_mr_sg() 1507 return mr->device->map_mr_sg(mr, sg, sg_nents); ib_map_mr_sg() 1516 * @sg_nents: number of entries in sg 1529 int sg_nents, ib_sg_to_pages() 1541 for_each_sg(sgl, sg, sg_nents, i) { for_each_sg() 1497 ib_map_mr_sg(struct ib_mr *mr, struct scatterlist *sg, int sg_nents, unsigned int page_size) ib_map_mr_sg() argument 1527 ib_sg_to_pages(struct ib_mr *mr, struct scatterlist *sgl, int sg_nents, int (*set_page)(struct ib_mr *, u64)) ib_sg_to_pages() argument
|
/linux-4.4.14/net/rds/ |
H A D | iw_send.c | 766 int sg_nents) rds_iw_build_send_reg() 770 n = ib_map_mr_sg(send->s_mr, sg, sg_nents, PAGE_SIZE); rds_iw_build_send_reg() 771 if (unlikely(n != sg_nents)) rds_iw_build_send_reg() 804 int sg_nents; rds_iw_xmit_rdma() local 858 sg_nents = 0; rds_iw_xmit_rdma() 902 sg_nents++; rds_iw_xmit_rdma() 950 &op->op_sg[0], sg_nents); rds_iw_xmit_rdma() 764 rds_iw_build_send_reg(struct rds_iw_send_work *send, struct scatterlist *sg, int sg_nents) rds_iw_build_send_reg() argument
|
/linux-4.4.14/drivers/ide/ |
H A D | ide-dma.c | 144 i = dma_map_sg(hwif->dev, sg, cmd->sg_nents, cmd->sg_dma_direction); ide_dma_map_sg() 146 cmd->orig_sg_nents = cmd->sg_nents; ide_dma_map_sg() 147 cmd->sg_nents = i; ide_dma_map_sg()
|
H A D | ide-dma-sff.c | 123 for_each_sg(hwif->sg_table, sg, cmd->sg_nents, i) { ide_build_dmatable()
|
H A D | au1xxx-ide.c | 220 int i = cmd->sg_nents, count = 0; auide_build_dmatable()
|
H A D | icside.c | 344 set_dma_sg(ec->dma, hwif->sg_table, cmd->sg_nents); icside_dma_setup()
|
H A D | sgiioc4.c | 418 unsigned int count = 0, i = cmd->sg_nents; sgiioc4_build_dmatable()
|
H A D | tx4939ide.c | 243 for_each_sg(hwif->sg_table, sg, cmd->sg_nents, i) { tx4939ide_build_dmatable()
|
H A D | ide-io.c | 229 cmd->sg_nents = blk_rq_map_sg(drive->queue, rq, sg); ide_map_sg()
|
H A D | pmac.c | 1465 int i = cmd->sg_nents, count = 0; pmac_ide_build_dmatable()
|
/linux-4.4.14/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_verbs.h | 130 int sg_nents);
|
H A D | ocrdma_verbs.c | 3246 int sg_nents) ocrdma_map_mr_sg() 3252 return ib_sg_to_pages(ibmr, sg, sg_nents, ocrdma_set_page); ocrdma_map_mr_sg() 3244 ocrdma_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents) ocrdma_map_mr_sg() argument
|
/linux-4.4.14/lib/ |
H A D | sg_split.c | 163 ret = sg_calculate_split(in, sg_nents(in), nb_splits, skip, split_sizes, sg_split()
|
H A D | scatterlist.c | 42 * sg_nents - return total count of entries in scatterlist 50 int sg_nents(struct scatterlist *sg) sg_nents() function 57 EXPORT_SYMBOL(sg_nents); variable
|
/linux-4.4.14/drivers/infiniband/ulp/isert/ |
H A D | ib_isert.c | 1285 int imm_data, imm_data_len, unsol_data, sg_nents, rc; isert_handle_scsi_cmd() local 1312 sg_nents = max(1UL, DIV_ROUND_UP(imm_data_len, PAGE_SIZE)); isert_handle_scsi_cmd() 1313 sg_copy_from_buffer(cmd->se_cmd.t_data_sg, sg_nents, isert_handle_scsi_cmd() 1315 isert_dbg("Copy Immediate sg_nents: %u imm_data_len: %d\n", isert_handle_scsi_cmd() 1316 sg_nents, imm_data_len); isert_handle_scsi_cmd() 1355 int rc, sg_nents, sg_off, page_off; isert_handle_iscsi_dataout() local 1378 sg_nents = max(1UL, DIV_ROUND_UP(unsol_data_len, PAGE_SIZE)); isert_handle_iscsi_dataout() 1389 "sg_nents: %u from %p %u\n", sg_start, sg_off, isert_handle_iscsi_dataout() 1390 sg_nents, &rx_desc->data[0], unsol_data_len); isert_handle_iscsi_dataout() 1392 sg_copy_from_buffer(sg_start, sg_nents, &rx_desc->data[0], isert_handle_iscsi_dataout() 1655 isert_dbg("Mapped cmd: %p count: %u sg: %p sg_nents: %u rdma_len %d\n", isert_map_data_buf() 2376 int i = 0, sg_nents; isert_build_rdma_wr() local 2380 sg_nents = min(cmd->se_cmd.t_data_nents - sg_off, isert_conn->max_sge); isert_build_rdma_wr() 2388 for_each_sg(sg_start, tmp_sg, sg_nents, i) { for_each_sg() 2557 isert_dbg("Use fr_desc %p sg_nents %d offset %u\n", isert_fast_reg_mr()
|
/linux-4.4.14/drivers/infiniband/hw/mlx4/ |
H A D | mr.c | 553 int sg_nents) mlx4_ib_map_mr_sg() 564 rc = ib_sg_to_pages(ibmr, sg, sg_nents, mlx4_set_page); mlx4_ib_map_mr_sg() 551 mlx4_ib_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents) mlx4_ib_map_mr_sg() argument
|
H A D | mlx4_ib.h | 715 int sg_nents);
|
/linux-4.4.14/drivers/infiniband/hw/qib/ |
H A D | qib_mr.c | 369 int sg_nents) qib_map_mr_sg() 375 return ib_sg_to_pages(ibmr, sg, sg_nents, qib_set_page); qib_map_mr_sg() 367 qib_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents) qib_map_mr_sg() argument
|
H A D | qib_verbs.h | 1051 int sg_nents);
|
/linux-4.4.14/drivers/usb/storage/ |
H A D | protocol.c | 144 nents = sg_nents(sg); usb_stor_access_xfer_buf()
|
/linux-4.4.14/include/linux/sunrpc/ |
H A D | svc_rdma.h | 109 int sg_nents; member in struct:svc_rdma_fastreg_mr
|
/linux-4.4.14/crypto/ |
H A D | algif_skcipher.c | 88 n = sg_nents(sgl); skcipher_free_async_sgls() 494 nents += sg_nents(sg); skcipher_all_sg_nents()
|
/linux-4.4.14/drivers/crypto/ccp/ |
H A D | ccp-crypto-aes-cmac.c | 110 sg_count = (nbytes) ? sg_nents(req->src) + 2 : 2; ccp_do_cmac_update()
|
H A D | ccp-crypto-sha.c | 103 sg_count = sg_nents(req->src) + 1; ccp_do_sha_update()
|
/linux-4.4.14/lib/mpi/ |
H A D | mpicoder.c | 446 ents = sg_nents(sgl); mpi_read_raw_from_sgl()
|
/linux-4.4.14/drivers/crypto/qat/qat_common/ |
H A D | qat_algs.c | 669 int n = sg_nents(sgl); qat_alg_sgl_to_bufl() 712 n = sg_nents(sglout); 762 n = sg_nents(sglout);
|
/linux-4.4.14/drivers/memstick/core/ |
H A D | ms_block.c | 104 sg_miter_start(&miter, sg, sg_nents(sg), msb_sg_compare_to_buffer() 872 sg_miter_start(&miter, sg, sg_nents(sg), msb_read_page() 1636 sg_copy_to_buffer(sg_tmp, sg_nents(sg_tmp), msb_cache_write() 1659 sg_copy_from_buffer(sg_tmp, sg_nents(sg_tmp), msb_cache_read()
|
/linux-4.4.14/include/linux/ |
H A D | scatterlist.h | 244 int sg_nents(struct scatterlist *sg);
|
H A D | ide.h | 311 int sg_nents; /* number of sg entries */ member in struct:ide_cmd
|
/linux-4.4.14/drivers/infiniband/hw/cxgb4/ |
H A D | mem.c | 939 int sg_nents) c4iw_map_mr_sg() 945 return ib_sg_to_pages(ibmr, sg, sg_nents, c4iw_set_page); c4iw_map_mr_sg() 937 c4iw_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents) c4iw_map_mr_sg() argument
|
H A D | iw_cxgb4.h | 964 int sg_nents);
|
/linux-4.4.14/drivers/infiniband/hw/mlx5/ |
H A D | mr.c | 1455 int sg_nents) mlx5_ib_map_mr_sg() 1466 n = ib_sg_to_pages(ibmr, sg, sg_nents, mlx5_set_page); mlx5_ib_map_mr_sg() 1453 mlx5_ib_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents) mlx5_ib_map_mr_sg() argument
|
H A D | mlx5_ib.h | 560 int sg_nents);
|
/linux-4.4.14/include/rdma/ |
H A D | ib_verbs.h | 1752 int sg_nents); 3031 int sg_nents, 3037 int sg_nents, ib_map_mr_sg_zbva() 3042 n = ib_map_mr_sg(mr, sg, sg_nents, page_size); ib_map_mr_sg_zbva() 3050 int sg_nents, 3035 ib_map_mr_sg_zbva(struct ib_mr *mr, struct scatterlist *sg, int sg_nents, unsigned int page_size) ib_map_mr_sg_zbva() argument
|
/linux-4.4.14/drivers/infiniband/ulp/srp/ |
H A D | ib_srp.c | 1316 struct srp_rdma_ch *ch, int sg_nents) srp_map_finish_fr() 1331 if (sg_nents == 0) srp_map_finish_fr() 1334 if (sg_nents == 1 && target->global_mr) { srp_map_finish_fr() 1348 n = ib_map_mr_sg(desc->mr, state->sg, sg_nents, dev->mr_page_size); srp_map_finish_fr() 1315 srp_map_finish_fr(struct srp_map_state *state, struct srp_rdma_ch *ch, int sg_nents) srp_map_finish_fr() argument
|
/linux-4.4.14/drivers/infiniband/hw/cxgb3/ |
H A D | iwch_provider.c | 878 int sg_nents) iwch_map_mr_sg() 884 return ib_sg_to_pages(ibmr, sg, sg_nents, iwch_set_page); iwch_map_mr_sg() 876 iwch_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents) iwch_map_mr_sg() argument
|
/linux-4.4.14/drivers/infiniband/hw/nes/ |
H A D | nes_verbs.c | 480 int sg_nents) nes_map_mr_sg() 486 return ib_sg_to_pages(ibmr, sg, sg_nents, nes_set_page); nes_map_mr_sg() 478 nes_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents) nes_map_mr_sg() argument
|
/linux-4.4.14/drivers/gpu/drm/i915/ |
H A D | i915_gem_gtt.c | 829 __sg_page_iter_start(&sg_iter, pages->sgl, sg_nents(pages->sgl), 0); gen8_ppgtt_insert_entries()
|