Searched refs:sg_nents (Results 1 - 48 of 48) sorted by relevance

/linux-4.4.14/drivers/crypto/caam/
H A Dsg_sw_sec4.h75 int sg_nents = sg_nents_for_len(sg_list, nbytes); sg_count() local
77 if (likely(sg_nents == 1)) sg_count()
80 return sg_nents; sg_count()
/linux-4.4.14/net/sunrpc/xprtrdma/
H A Dfrwr_ops.c361 frmr->sg_nents = i; frwr_op_map()
363 dma_nents = ib_dma_map_sg(device, frmr->sg, frmr->sg_nents, direction); frwr_op_map()
365 pr_err("RPC: %s: failed to dma map sg %p sg_nents %u\n", frwr_op_map()
366 __func__, frmr->sg, frmr->sg_nents); frwr_op_map()
370 n = ib_map_mr_sg(mr, frmr->sg, frmr->sg_nents, PAGE_SIZE); frwr_op_map()
371 if (unlikely(n != frmr->sg_nents)) { frwr_op_map()
373 __func__, frmr->fr_mr, n, frmr->sg_nents); frwr_op_map()
379 __func__, mw, frmr->sg_nents, mr->length); frwr_op_map()
404 seg1->mr_nsegs = frmr->sg_nents; frwr_op_map()
407 return frmr->sg_nents; frwr_op_map()
440 ib_dma_unmap_sg(ia->ri_device, frmr->sg, frmr->sg_nents, seg1->mr_dir); frwr_op_unmap()
H A Dsvc_rdma_recvfrom.c243 frmr->sg_nents = nents; rdma_read_chunk_frmr()
275 frmr->sg, frmr->sg_nents, rdma_read_chunk_frmr()
284 n = ib_map_mr_sg(frmr->mr, frmr->sg, frmr->sg_nents, PAGE_SIZE); rdma_read_chunk_frmr()
285 if (unlikely(n != frmr->sg_nents)) { rdma_read_chunk_frmr()
287 frmr->mr, n, frmr->sg_nents); rdma_read_chunk_frmr()
350 frmr->sg, frmr->sg_nents, frmr->direction); rdma_read_chunk_frmr()
H A Dsvc_rdma_transport.c848 frmr->sg_nents = 0; svc_rdma_get_frmr()
862 frmr->sg, frmr->sg_nents, frmr->direction); svc_rdma_put_frmr()
H A Dxprt_rdma.h205 int sg_nents; member in struct:rpcrdma_frmr
/linux-4.4.14/drivers/crypto/
H A Dbfin_crc.c82 unsigned int sg_nents; member in struct:bfin_crypto_crc_reqctx
143 if (sg_nents(req->src) > CRC_MAX_DMA_DESC) { bfin_crypto_crc_init()
176 dma_map_sg(crc->dev, ctx->sg, ctx->sg_nents, DMA_TO_DEVICE); bfin_crypto_crc_config_dma()
178 for_each_sg(ctx->sg, sg, ctx->sg_nents, j) { bfin_crypto_crc_config_dma()
315 ctx->sg_nents = 0; bfin_crypto_crc_handle_queue()
359 nsg = sg_nents(ctx->sg); bfin_crypto_crc_handle_queue()
360 ctx->sg_nents = nsg; bfin_crypto_crc_handle_queue()
375 ctx->sg_nents--; bfin_crypto_crc_handle_queue()
H A Dimg-hash.c256 ctx->bufcnt = sg_copy_to_buffer(hdev->req->src, sg_nents(ctx->sg), img_hash_write_via_cpu()
639 ctx->nents = sg_nents(ctx->sg); img_hash_digest()
H A Dmxs-dcp.c259 const int nents = sg_nents(req->src); mxs_dcp_aes_block_crypt()
568 const int nents = sg_nents(req->src); dcp_sha_req_to_buf()
/linux-4.4.14/drivers/crypto/sunxi-ss/
H A Dsun4i-ss-cipher.c66 sg_miter_start(&mi, areq->src, sg_nents(areq->src), sun4i_ss_opti_poll()
68 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst), sun4i_ss_opti_poll()
199 sg_miter_start(&mi, areq->src, sg_nents(areq->src), sun4i_ss_cipher_poll()
201 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst), sun4i_ss_cipher_poll()
H A Dsun4i-ss-hash.c200 copied = sg_pcopy_to_buffer(areq->src, sg_nents(areq->src), sun4i_hash_update()
229 sg_miter_start(&mi, areq->src, sg_nents(areq->src), sun4i_hash_update()
/linux-4.4.14/drivers/infiniband/core/
H A Dverbs.c1480 * @sg_nents: number of entries in sg
1489 * - If sg_nents total byte length exceeds the mr max_num_sge * page_size
1499 int sg_nents, ib_map_mr_sg()
1507 return mr->device->map_mr_sg(mr, sg, sg_nents); ib_map_mr_sg()
1516 * @sg_nents: number of entries in sg
1529 int sg_nents, ib_sg_to_pages()
1541 for_each_sg(sgl, sg, sg_nents, i) { for_each_sg()
1497 ib_map_mr_sg(struct ib_mr *mr, struct scatterlist *sg, int sg_nents, unsigned int page_size) ib_map_mr_sg() argument
1527 ib_sg_to_pages(struct ib_mr *mr, struct scatterlist *sgl, int sg_nents, int (*set_page)(struct ib_mr *, u64)) ib_sg_to_pages() argument
/linux-4.4.14/net/rds/
H A Diw_send.c766 int sg_nents) rds_iw_build_send_reg()
770 n = ib_map_mr_sg(send->s_mr, sg, sg_nents, PAGE_SIZE); rds_iw_build_send_reg()
771 if (unlikely(n != sg_nents)) rds_iw_build_send_reg()
804 int sg_nents; rds_iw_xmit_rdma() local
858 sg_nents = 0; rds_iw_xmit_rdma()
902 sg_nents++; rds_iw_xmit_rdma()
950 &op->op_sg[0], sg_nents); rds_iw_xmit_rdma()
764 rds_iw_build_send_reg(struct rds_iw_send_work *send, struct scatterlist *sg, int sg_nents) rds_iw_build_send_reg() argument
/linux-4.4.14/drivers/ide/
H A Dide-dma.c144 i = dma_map_sg(hwif->dev, sg, cmd->sg_nents, cmd->sg_dma_direction); ide_dma_map_sg()
146 cmd->orig_sg_nents = cmd->sg_nents; ide_dma_map_sg()
147 cmd->sg_nents = i; ide_dma_map_sg()
H A Dide-dma-sff.c123 for_each_sg(hwif->sg_table, sg, cmd->sg_nents, i) { ide_build_dmatable()
H A Dau1xxx-ide.c220 int i = cmd->sg_nents, count = 0; auide_build_dmatable()
H A Dicside.c344 set_dma_sg(ec->dma, hwif->sg_table, cmd->sg_nents); icside_dma_setup()
H A Dsgiioc4.c418 unsigned int count = 0, i = cmd->sg_nents; sgiioc4_build_dmatable()
H A Dtx4939ide.c243 for_each_sg(hwif->sg_table, sg, cmd->sg_nents, i) { tx4939ide_build_dmatable()
H A Dide-io.c229 cmd->sg_nents = blk_rq_map_sg(drive->queue, rq, sg); ide_map_sg()
H A Dpmac.c1465 int i = cmd->sg_nents, count = 0; pmac_ide_build_dmatable()
/linux-4.4.14/drivers/infiniband/hw/ocrdma/
H A Docrdma_verbs.h130 int sg_nents);
H A Docrdma_verbs.c3246 int sg_nents) ocrdma_map_mr_sg()
3252 return ib_sg_to_pages(ibmr, sg, sg_nents, ocrdma_set_page); ocrdma_map_mr_sg()
3244 ocrdma_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents) ocrdma_map_mr_sg() argument
/linux-4.4.14/lib/
H A Dsg_split.c163 ret = sg_calculate_split(in, sg_nents(in), nb_splits, skip, split_sizes, sg_split()
H A Dscatterlist.c42 * sg_nents - return total count of entries in scatterlist
50 int sg_nents(struct scatterlist *sg) sg_nents() function
57 EXPORT_SYMBOL(sg_nents); variable
/linux-4.4.14/drivers/infiniband/ulp/isert/
H A Dib_isert.c1285 int imm_data, imm_data_len, unsol_data, sg_nents, rc; isert_handle_scsi_cmd() local
1312 sg_nents = max(1UL, DIV_ROUND_UP(imm_data_len, PAGE_SIZE)); isert_handle_scsi_cmd()
1313 sg_copy_from_buffer(cmd->se_cmd.t_data_sg, sg_nents, isert_handle_scsi_cmd()
1315 isert_dbg("Copy Immediate sg_nents: %u imm_data_len: %d\n", isert_handle_scsi_cmd()
1316 sg_nents, imm_data_len); isert_handle_scsi_cmd()
1355 int rc, sg_nents, sg_off, page_off; isert_handle_iscsi_dataout() local
1378 sg_nents = max(1UL, DIV_ROUND_UP(unsol_data_len, PAGE_SIZE)); isert_handle_iscsi_dataout()
1389 "sg_nents: %u from %p %u\n", sg_start, sg_off, isert_handle_iscsi_dataout()
1390 sg_nents, &rx_desc->data[0], unsol_data_len); isert_handle_iscsi_dataout()
1392 sg_copy_from_buffer(sg_start, sg_nents, &rx_desc->data[0], isert_handle_iscsi_dataout()
1655 isert_dbg("Mapped cmd: %p count: %u sg: %p sg_nents: %u rdma_len %d\n", isert_map_data_buf()
2376 int i = 0, sg_nents; isert_build_rdma_wr() local
2380 sg_nents = min(cmd->se_cmd.t_data_nents - sg_off, isert_conn->max_sge); isert_build_rdma_wr()
2388 for_each_sg(sg_start, tmp_sg, sg_nents, i) { for_each_sg()
2557 isert_dbg("Use fr_desc %p sg_nents %d offset %u\n", isert_fast_reg_mr()
/linux-4.4.14/drivers/infiniband/hw/mlx4/
H A Dmr.c553 int sg_nents) mlx4_ib_map_mr_sg()
564 rc = ib_sg_to_pages(ibmr, sg, sg_nents, mlx4_set_page); mlx4_ib_map_mr_sg()
551 mlx4_ib_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents) mlx4_ib_map_mr_sg() argument
H A Dmlx4_ib.h715 int sg_nents);
/linux-4.4.14/drivers/infiniband/hw/qib/
H A Dqib_mr.c369 int sg_nents) qib_map_mr_sg()
375 return ib_sg_to_pages(ibmr, sg, sg_nents, qib_set_page); qib_map_mr_sg()
367 qib_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents) qib_map_mr_sg() argument
H A Dqib_verbs.h1051 int sg_nents);
/linux-4.4.14/drivers/usb/storage/
H A Dprotocol.c144 nents = sg_nents(sg); usb_stor_access_xfer_buf()
/linux-4.4.14/include/linux/sunrpc/
H A Dsvc_rdma.h109 int sg_nents; member in struct:svc_rdma_fastreg_mr
/linux-4.4.14/crypto/
H A Dalgif_skcipher.c88 n = sg_nents(sgl); skcipher_free_async_sgls()
494 nents += sg_nents(sg); skcipher_all_sg_nents()
/linux-4.4.14/drivers/crypto/ccp/
H A Dccp-crypto-aes-cmac.c110 sg_count = (nbytes) ? sg_nents(req->src) + 2 : 2; ccp_do_cmac_update()
H A Dccp-crypto-sha.c103 sg_count = sg_nents(req->src) + 1; ccp_do_sha_update()
/linux-4.4.14/lib/mpi/
H A Dmpicoder.c446 ents = sg_nents(sgl); mpi_read_raw_from_sgl()
/linux-4.4.14/drivers/crypto/qat/qat_common/
H A Dqat_algs.c669 int n = sg_nents(sgl); qat_alg_sgl_to_bufl()
712 n = sg_nents(sglout);
762 n = sg_nents(sglout);
/linux-4.4.14/drivers/memstick/core/
H A Dms_block.c104 sg_miter_start(&miter, sg, sg_nents(sg), msb_sg_compare_to_buffer()
872 sg_miter_start(&miter, sg, sg_nents(sg), msb_read_page()
1636 sg_copy_to_buffer(sg_tmp, sg_nents(sg_tmp), msb_cache_write()
1659 sg_copy_from_buffer(sg_tmp, sg_nents(sg_tmp), msb_cache_read()
/linux-4.4.14/include/linux/
H A Dscatterlist.h244 int sg_nents(struct scatterlist *sg);
H A Dide.h311 int sg_nents; /* number of sg entries */ member in struct:ide_cmd
/linux-4.4.14/drivers/infiniband/hw/cxgb4/
H A Dmem.c939 int sg_nents) c4iw_map_mr_sg()
945 return ib_sg_to_pages(ibmr, sg, sg_nents, c4iw_set_page); c4iw_map_mr_sg()
937 c4iw_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents) c4iw_map_mr_sg() argument
H A Diw_cxgb4.h964 int sg_nents);
/linux-4.4.14/drivers/infiniband/hw/mlx5/
H A Dmr.c1455 int sg_nents) mlx5_ib_map_mr_sg()
1466 n = ib_sg_to_pages(ibmr, sg, sg_nents, mlx5_set_page); mlx5_ib_map_mr_sg()
1453 mlx5_ib_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents) mlx5_ib_map_mr_sg() argument
H A Dmlx5_ib.h560 int sg_nents);
/linux-4.4.14/include/rdma/
H A Dib_verbs.h1752 int sg_nents);
3031 int sg_nents,
3037 int sg_nents, ib_map_mr_sg_zbva()
3042 n = ib_map_mr_sg(mr, sg, sg_nents, page_size); ib_map_mr_sg_zbva()
3050 int sg_nents,
3035 ib_map_mr_sg_zbva(struct ib_mr *mr, struct scatterlist *sg, int sg_nents, unsigned int page_size) ib_map_mr_sg_zbva() argument
/linux-4.4.14/drivers/infiniband/ulp/srp/
H A Dib_srp.c1316 struct srp_rdma_ch *ch, int sg_nents) srp_map_finish_fr()
1331 if (sg_nents == 0) srp_map_finish_fr()
1334 if (sg_nents == 1 && target->global_mr) { srp_map_finish_fr()
1348 n = ib_map_mr_sg(desc->mr, state->sg, sg_nents, dev->mr_page_size); srp_map_finish_fr()
1315 srp_map_finish_fr(struct srp_map_state *state, struct srp_rdma_ch *ch, int sg_nents) srp_map_finish_fr() argument
/linux-4.4.14/drivers/infiniband/hw/cxgb3/
H A Diwch_provider.c878 int sg_nents) iwch_map_mr_sg()
884 return ib_sg_to_pages(ibmr, sg, sg_nents, iwch_set_page); iwch_map_mr_sg()
876 iwch_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents) iwch_map_mr_sg() argument
/linux-4.4.14/drivers/infiniband/hw/nes/
H A Dnes_verbs.c480 int sg_nents) nes_map_mr_sg()
486 return ib_sg_to_pages(ibmr, sg, sg_nents, nes_set_page); nes_map_mr_sg()
478 nes_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents) nes_map_mr_sg() argument
/linux-4.4.14/drivers/gpu/drm/i915/
H A Di915_gem_gtt.c829 __sg_page_iter_start(&sg_iter, pages->sgl, sg_nents(pages->sgl), 0); gen8_ppgtt_insert_entries()

Completed in 1706 milliseconds