sg_count 316 drivers/block/sx8.c u8 sg_count; sg_count 759 drivers/block/sx8.c msg->sg_count = n_elem; sg_count 64 drivers/crypto/ccp/ccp-crypto-aes-cmac.c unsigned int need_pad, sg_count; sg_count 107 drivers/crypto/ccp/ccp-crypto-aes-cmac.c sg_count = (nbytes) ? sg_nents(req->src) + 2 : 2; sg_count 110 drivers/crypto/ccp/ccp-crypto-aes-cmac.c ret = sg_alloc_table(&rctx->data_sg, sg_count, gfp); sg_count 65 drivers/crypto/ccp/ccp-crypto-sha.c unsigned int sg_count; sg_count 102 drivers/crypto/ccp/ccp-crypto-sha.c sg_count = sg_nents(req->src) + 1; sg_count 103 drivers/crypto/ccp/ccp-crypto-sha.c ret = sg_alloc_table(&rctx->data_sg, sg_count, gfp); sg_count 1098 drivers/crypto/talitos.c static int sg_to_link_tbl_offset(struct scatterlist *sg, int sg_count, sg_count 1102 drivers/crypto/talitos.c int n_sg = elen ? sg_count + 1 : sg_count; sg_count 1149 drivers/crypto/talitos.c struct talitos_ptr *ptr, int sg_count, sg_count 1161 drivers/crypto/talitos.c if (sg_count == 1 && !force) { sg_count 1163 drivers/crypto/talitos.c return sg_count; sg_count 1167 drivers/crypto/talitos.c return sg_count; sg_count 1169 drivers/crypto/talitos.c sg_count = sg_to_link_tbl_offset(src, sg_count, offset, len, elen, sg_count 1171 drivers/crypto/talitos.c if (sg_count == 1 && !force) { sg_count 1174 drivers/crypto/talitos.c return sg_count; sg_count 1180 drivers/crypto/talitos.c return sg_count; sg_count 1185 drivers/crypto/talitos.c struct talitos_ptr *ptr, int sg_count, sg_count 1188 drivers/crypto/talitos.c return talitos_sg_map_ext(dev, src, len, edesc, ptr, sg_count, offset, sg_count 1209 drivers/crypto/talitos.c int sg_count, ret; sg_count 1222 drivers/crypto/talitos.c sg_count = edesc->src_nents ?: 1; sg_count 1223 drivers/crypto/talitos.c if (is_sec1 && sg_count > 1) sg_count 1224 drivers/crypto/talitos.c sg_copy_to_buffer(areq->src, sg_count, edesc->buf, sg_count 1227 drivers/crypto/talitos.c sg_count = dma_map_sg(dev, areq->src, sg_count, sg_count 1233 drivers/crypto/talitos.c &desc->ptr[1], sg_count, 0, tbl_off); sg_count 1257 drivers/crypto/talitos.c sg_count, areq->assoclen, tbl_off, elen, sg_count 1267 drivers/crypto/talitos.c sg_count = edesc->dst_nents ? : 1; sg_count 1268 drivers/crypto/talitos.c if (!is_sec1 || sg_count == 1) sg_count 1269 drivers/crypto/talitos.c dma_map_sg(dev, areq->dst, sg_count, DMA_FROM_DEVICE); sg_count 1277 drivers/crypto/talitos.c sg_count, areq->assoclen, tbl_off, elen, sg_count 1297 drivers/crypto/talitos.c sg_count, areq->assoclen + cryptlen, tbl_off); sg_count 1582 drivers/crypto/talitos.c int sg_count, ret; sg_count 1595 drivers/crypto/talitos.c sg_count = edesc->src_nents ?: 1; sg_count 1596 drivers/crypto/talitos.c if (is_sec1 && sg_count > 1) sg_count 1597 drivers/crypto/talitos.c sg_copy_to_buffer(areq->src, sg_count, edesc->buf, sg_count 1600 drivers/crypto/talitos.c sg_count = dma_map_sg(dev, areq->src, sg_count, sg_count 1606 drivers/crypto/talitos.c sg_count = talitos_sg_map(dev, areq->src, cryptlen, edesc, sg_count 1607 drivers/crypto/talitos.c &desc->ptr[3], sg_count, 0, 0); sg_count 1608 drivers/crypto/talitos.c if (sg_count > 1) sg_count 1613 drivers/crypto/talitos.c sg_count = edesc->dst_nents ? : 1; sg_count 1614 drivers/crypto/talitos.c if (!is_sec1 || sg_count == 1) sg_count 1615 drivers/crypto/talitos.c dma_map_sg(dev, areq->dst, sg_count, DMA_FROM_DEVICE); sg_count 1619 drivers/crypto/talitos.c sg_count, 0, (edesc->src_nents + 1)); sg_count 1799 drivers/crypto/talitos.c int sg_count; sg_count 1822 drivers/crypto/talitos.c sg_count = edesc->src_nents ?: 1; sg_count 1823 drivers/crypto/talitos.c if (is_sec1 && sg_count > 1) sg_count 1824 drivers/crypto/talitos.c sg_copy_to_buffer(req_ctx->psrc, sg_count, edesc->buf, length); sg_count 1826 drivers/crypto/talitos.c sg_count = dma_map_sg(dev, req_ctx->psrc, sg_count, sg_count 1836 drivers/crypto/talitos.c sg_count = talitos_sg_map(dev, req_ctx->psrc, length, edesc, sg_count 1837 drivers/crypto/talitos.c &desc->ptr[3], sg_count, 0, 0); sg_count 1838 drivers/crypto/talitos.c if (sg_count > 1) sg_count 1882 drivers/crypto/talitos.c sg_count = talitos_sg_map(dev, req_ctx->psrc, length, edesc, sg_count 1883 drivers/crypto/talitos.c &desc2->ptr[3], sg_count, 0, 0); sg_count 1884 drivers/crypto/talitos.c if (sg_count > 1) sg_count 219 drivers/firmware/efi/capsule.c unsigned int count, sg_count; sg_count 230 drivers/firmware/efi/capsule.c sg_count = sg_pages_num(count); sg_count 232 drivers/firmware/efi/capsule.c sg_pages = kcalloc(sg_count, sizeof(*sg_pages), GFP_KERNEL); sg_count 236 drivers/firmware/efi/capsule.c for (i = 0; i < sg_count; i++) { sg_count 244 drivers/firmware/efi/capsule.c for (i = 0; i < sg_count; i++) { sg_count 263 drivers/firmware/efi/capsule.c if (i + 1 == sg_count) sg_count 276 drivers/firmware/efi/capsule.c for (i = 0; rv && i < sg_count; i++) { sg_count 235 drivers/infiniband/hw/qib/qib.h int sg_count; sg_count 534 drivers/infiniband/hw/qib/qib_sdma.c if (tx->txreq.sg_count > qib_sdma_descq_freecnt(ppd)) { sg_count 609 drivers/infiniband/hw/qib/qib_sdma.c ppd->sdma_descq_added += tx->txreq.sg_count; sg_count 687 drivers/infiniband/hw/qib/qib_verbs.c if (qpp->s_tx->txreq.sg_count > avail) sg_count 689 drivers/infiniband/hw/qib/qib_verbs.c avail -= qpp->s_tx->txreq.sg_count; sg_count 829 drivers/infiniband/hw/qib/qib_verbs.c tx->txreq.sg_count = ndesc; sg_count 853 drivers/infiniband/hw/qib/qib_verbs.c tx->txreq.sg_count = 1; sg_count 277 drivers/memstick/host/r592.c int len, sg_count; sg_count 295 drivers/memstick/host/r592.c sg_count = dma_map_sg(&dev->pci_dev->dev, &dev->req->sg, 1, is_write ? sg_count 298 drivers/memstick/host/r592.c if (sg_count != 1 || sg_dma_len(&dev->req->sg) < R592_LFIFO_SIZE) { sg_count 58 drivers/mmc/host/alcor.c int sg_count; sg_count 127 drivers/mmc/host/alcor.c if (!host->sg_count) sg_count 145 drivers/mmc/host/alcor.c host->sg_count--; sg_count 169 drivers/mmc/host/alcor.c alcor_write32(priv, data->sg_count * 0x1000, sg_count 252 drivers/mmc/host/alcor.c host->sg_count = data->sg_count; sg_count 254 drivers/mmc/host/alcor.c host->sg_count, host->blocks); sg_count 503 drivers/mmc/host/alcor.c if (!host->sg_count) sg_count 543 drivers/mmc/host/alcor.c (host->dma_on && !host->sg_count)) sg_count 822 drivers/mmc/host/alcor.c data->sg_count = sg_len; sg_count 425 drivers/mmc/host/cqhci.c int sg_count; sg_count 431 drivers/mmc/host/cqhci.c sg_count = dma_map_sg(mmc_dev(host), data->sg, sg_count 435 drivers/mmc/host/cqhci.c if (!sg_count) { sg_count 440 drivers/mmc/host/cqhci.c return sg_count; sg_count 469 drivers/mmc/host/cqhci.c int i, sg_count, len; sg_count 476 drivers/mmc/host/cqhci.c sg_count = cqhci_dma_map(mrq->host, mrq); sg_count 477 drivers/mmc/host/cqhci.c if (sg_count < 0) { sg_count 479 drivers/mmc/host/cqhci.c mmc_hostname(mrq->host), __func__, sg_count); sg_count 480 drivers/mmc/host/cqhci.c return sg_count; sg_count 485 drivers/mmc/host/cqhci.c for_each_sg(data->sg, sg, sg_count, i) { sg_count 489 drivers/mmc/host/cqhci.c if ((i+1) == sg_count) sg_count 255 drivers/mmc/host/jz4740_mmc.c int sg_count; sg_count 258 drivers/mmc/host/jz4740_mmc.c return data->sg_count; sg_count 260 drivers/mmc/host/jz4740_mmc.c sg_count = dma_map_sg(chan->device->dev, sg_count 265 drivers/mmc/host/jz4740_mmc.c if (sg_count <= 0) { sg_count 271 drivers/mmc/host/jz4740_mmc.c data->sg_count = sg_count; sg_count 274 drivers/mmc/host/jz4740_mmc.c return data->sg_count; sg_count 288 drivers/mmc/host/jz4740_mmc.c int sg_count; sg_count 298 drivers/mmc/host/jz4740_mmc.c sg_count = jz4740_mmc_prepare_dma_data(host, data, COOKIE_MAPPED); sg_count 299 drivers/mmc/host/jz4740_mmc.c if (sg_count < 0) sg_count 300 drivers/mmc/host/jz4740_mmc.c return sg_count; sg_count 303 drivers/mmc/host/jz4740_mmc.c desc = dmaengine_prep_slave_sg(chan, data->sg, sg_count, sg_count 286 drivers/mmc/host/meson-gx-mmc.c data->sg_count = dma_map_sg(mmc_dev(mmc), data->sg, data->sg_len, sg_count 288 drivers/mmc/host/meson-gx-mmc.c if (!data->sg_count) sg_count 297 drivers/mmc/host/meson-gx-mmc.c if (data && meson_mmc_desc_chain_mode(data) && data->sg_count) sg_count 719 drivers/mmc/host/meson-gx-mmc.c for_each_sg(data->sg, sg, data->sg_count, i) { sg_count 733 drivers/mmc/host/meson-gx-mmc.c desc[data->sg_count - 1].cmd_cfg |= CMD_CFG_END_OF_CHAIN; sg_count 634 drivers/mmc/host/mtk-sd.c for_each_sg(data->sg, sg, data->sg_count, j) { sg_count 656 drivers/mmc/host/mtk-sd.c if (j == data->sg_count - 1) /* the last bd */ sg_count 683 drivers/mmc/host/mtk-sd.c data->sg_count = dma_map_sg(host->dev, data->sg, data->sg_len, sg_count 44 drivers/mmc/host/rtsx_pci_sdmmc.c int sg_count; sg_count 175 drivers/mmc/host/rtsx_pci_sdmmc.c host->sg_count = count; sg_count 475 drivers/mmc/host/rtsx_pci_sdmmc.c err = rtsx_pci_dma_transfer(pcr, data->sg, host->sg_count, 1, 10000); sg_count 533 drivers/mmc/host/rtsx_pci_sdmmc.c err = rtsx_pci_dma_transfer(pcr, data->sg, host->sg_count, 0, 10000); sg_count 546 drivers/mmc/host/rtsx_pci_sdmmc.c if (host->sg_count < 0) { sg_count 547 drivers/mmc/host/rtsx_pci_sdmmc.c data->error = host->sg_count; sg_count 549 drivers/mmc/host/rtsx_pci_sdmmc.c __func__, host->sg_count); sg_count 603 drivers/mmc/host/sdhci.c int sg_count; sg_count 610 drivers/mmc/host/sdhci.c return data->sg_count; sg_count 634 drivers/mmc/host/sdhci.c sg_count = 1; sg_count 637 drivers/mmc/host/sdhci.c sg_count = dma_map_sg(mmc_dev(host->mmc), sg_count 642 drivers/mmc/host/sdhci.c if (sg_count == 0) sg_count 645 drivers/mmc/host/sdhci.c data->sg_count = sg_count; sg_count 648 drivers/mmc/host/sdhci.c return sg_count; sg_count 699 drivers/mmc/host/sdhci.c struct mmc_data *data, int sg_count) sg_count 713 drivers/mmc/host/sdhci.c host->sg_count = sg_count; sg_count 720 drivers/mmc/host/sdhci.c for_each_sg(data->sg, sg, host->sg_count, i) { sg_count 791 drivers/mmc/host/sdhci.c for_each_sg(data->sg, sg, host->sg_count, i) sg_count 803 drivers/mmc/host/sdhci.c for_each_sg(data->sg, sg, host->sg_count, i) { sg_count 548 drivers/mmc/host/sdhci.h int sg_count; /* Mapped sg entries */ sg_count 612 drivers/nvme/target/rdma.c int sg_count = num_pages(len); sg_count 617 drivers/nvme/target/rdma.c for (i = 0; i < sg_count; i++, sg++) { sg_count 618 drivers/nvme/target/rdma.c if (i < sg_count - 1) sg_count 630 drivers/nvme/target/rdma.c rsp->req.sg_cnt = sg_count; sg_count 1810 drivers/scsi/3w-9xxx.c int i, sg_count; sg_count 1862 drivers/scsi/3w-9xxx.c sg_count = scsi_dma_map(srb); sg_count 1863 drivers/scsi/3w-9xxx.c if (sg_count < 0) sg_count 1866 drivers/scsi/3w-9xxx.c scsi_for_each_sg(srb, sg, sg_count, i) { sg_count 296 drivers/scsi/3w-sas.c int i, sg_count; sg_count 338 drivers/scsi/3w-sas.c sg_count = scsi_dma_map(srb); sg_count 339 drivers/scsi/3w-sas.c if (sg_count <= 0) sg_count 342 drivers/scsi/3w-sas.c scsi_for_each_sg(srb, sg, sg_count, i) { sg_count 1862 drivers/scsi/53c700.c int sg_count; sg_count 1867 drivers/scsi/53c700.c sg_count = scsi_dma_map(SCp); sg_count 1868 drivers/scsi/53c700.c BUG_ON(sg_count < 0); sg_count 1870 drivers/scsi/53c700.c scsi_for_each_sg(SCp, sg, sg_count, i) { sg_count 4953 drivers/scsi/FlashPoint.c unsigned char sg_count, i; sg_count 4962 drivers/scsi/FlashPoint.c sg_count = 0; sg_count 4972 drivers/scsi/FlashPoint.c while ((sg_count < (unsigned char)SG_BUF_CNT) && sg_count 4982 drivers/scsi/FlashPoint.c if ((!sg_count) && (pcurrSCCB->Sccb_SGoffset)) { sg_count 4998 drivers/scsi/FlashPoint.c sg_count++; sg_count 5004 drivers/scsi/FlashPoint.c WR_HARPOON(p_port + hp_sg_addr, (sg_count << 4)); sg_count 479 drivers/scsi/aacraid/commctrl.c u32 sg_count[HBA_MAX_SG_EMBEDDED]; sg_count 645 drivers/scsi/aacraid/commctrl.c sg_count[i] = (actual_fibsize64 == fibsize) ? sg_count 647 drivers/scsi/aacraid/commctrl.c if (sg_count[i] > sg_count 650 drivers/scsi/aacraid/commctrl.c i, sg_count[i], sg_count 656 drivers/scsi/aacraid/commctrl.c p = kmalloc(sg_count[i], GFP_KERNEL); sg_count 675 drivers/scsi/aacraid/commctrl.c sg_count[i])) { sg_count 680 drivers/scsi/aacraid/commctrl.c addr = pci_map_single(dev->pdev, p, sg_count[i], sg_count 685 drivers/scsi/aacraid/commctrl.c hbacmd->sge[i].len = cpu_to_le32(sg_count[i]); sg_count 687 drivers/scsi/aacraid/commctrl.c byte_count += sg_count[i]; sg_count 711 drivers/scsi/aacraid/commctrl.c sg_count[i] = upsg->sg[i].count; sg_count 712 drivers/scsi/aacraid/commctrl.c if (sg_count[i] > sg_count 721 drivers/scsi/aacraid/commctrl.c p = kmalloc(sg_count[i], GFP_KERNEL); sg_count 724 drivers/scsi/aacraid/commctrl.c sg_count[i], i, upsg->count)); sg_count 736 drivers/scsi/aacraid/commctrl.c sg_count[i])){ sg_count 743 drivers/scsi/aacraid/commctrl.c sg_count[i], data_dir); sg_count 747 drivers/scsi/aacraid/commctrl.c byte_count += sg_count[i]; sg_count 748 drivers/scsi/aacraid/commctrl.c psg->sg[i].count = cpu_to_le32(sg_count[i]); sg_count 766 drivers/scsi/aacraid/commctrl.c sg_count[i] = usg->sg[i].count; sg_count 767 drivers/scsi/aacraid/commctrl.c if (sg_count[i] > sg_count 777 drivers/scsi/aacraid/commctrl.c p = kmalloc(sg_count[i], GFP_KERNEL); sg_count 780 drivers/scsi/aacraid/commctrl.c sg_count[i], i, usg->count)); sg_count 791 drivers/scsi/aacraid/commctrl.c sg_count[i])) { sg_count 799 drivers/scsi/aacraid/commctrl.c sg_count[i], data_dir); sg_count 803 drivers/scsi/aacraid/commctrl.c byte_count += sg_count[i]; sg_count 804 drivers/scsi/aacraid/commctrl.c psg->sg[i].count = cpu_to_le32(sg_count[i]); sg_count 824 drivers/scsi/aacraid/commctrl.c sg_count[i] = usg->sg[i].count; sg_count 825 drivers/scsi/aacraid/commctrl.c if (sg_count[i] > sg_count 833 drivers/scsi/aacraid/commctrl.c p = kmalloc(sg_count[i], GFP_KERNEL); sg_count 836 drivers/scsi/aacraid/commctrl.c sg_count[i], i, usg->count)); sg_count 848 drivers/scsi/aacraid/commctrl.c sg_count[i])){ sg_count 858 drivers/scsi/aacraid/commctrl.c psg->sg[i].count = cpu_to_le32(sg_count[i]); sg_count 865 drivers/scsi/aacraid/commctrl.c sg_count[i] = upsg->sg[i].count; sg_count 866 drivers/scsi/aacraid/commctrl.c if (sg_count[i] > sg_count 874 drivers/scsi/aacraid/commctrl.c p = kmalloc(sg_count[i], GFP_KERNEL); sg_count 877 drivers/scsi/aacraid/commctrl.c sg_count[i], i, upsg->count)); sg_count 887 drivers/scsi/aacraid/commctrl.c sg_count[i])) { sg_count 894 drivers/scsi/aacraid/commctrl.c sg_count[i], data_dir); sg_count 897 drivers/scsi/aacraid/commctrl.c byte_count += sg_count[i]; sg_count 898 drivers/scsi/aacraid/commctrl.c psg->sg[i].count = cpu_to_le32(sg_count[i]); sg_count 922 drivers/scsi/aacraid/commctrl.c if (copy_to_user(sg_user[i], sg_list[i], sg_count[i])) { sg_count 1732 drivers/scsi/advansys.c __le32 sg_count; /* SG element count. */ sg_count 2565 drivers/scsi/advansys.c (u32)le32_to_cpu(b->sg_list[i].sg_count)); sg_count 7700 drivers/scsi/advansys.c sg_block->sg_list[i].sg_count = sg_count 401 drivers/scsi/aha1542.c int mbo, sg_count; sg_count 423 drivers/scsi/aha1542.c sg_count = scsi_dma_map(cmd); sg_count 424 drivers/scsi/aha1542.c if (sg_count) { sg_count 425 drivers/scsi/aha1542.c size_t len = sg_count * sizeof(struct chain); sg_count 484 drivers/scsi/aha1542.c scsi_for_each_sg(cmd, sg, sg_count, i) { sg_count 488 drivers/scsi/aha1542.c any2scsi(ccb[mbo].datalen, sg_count * sizeof(struct chain)); sg_count 618 drivers/scsi/aic7xxx/aic79xx.h u_int sg_count;/* How full ahd_dma_seg is */ sg_count 409 drivers/scsi/aic7xxx/aic79xx_core.c scb->sg_count++; sg_count 530 drivers/scsi/aic7xxx/aic79xx_core.c if (scb->sg_count == 0) sg_count 536 drivers/scsi/aic7xxx/aic79xx_core.c /*len*/ahd_sg_size(ahd) * scb->sg_count, op); sg_count 855 drivers/scsi/aic7xxx/aic79xx_core.c if (scb->sg_count != 0) sg_count 1731 drivers/scsi/aic7xxx/aic79xx_core.c if (scb->sg_count > 0) { sg_count 1736 drivers/scsi/aic7xxx/aic79xx_core.c for (i = 0; i < scb->sg_count; i++) { sg_count 1754 drivers/scsi/aic7xxx/aic79xx_core.c for (i = 0; i < scb->sg_count; i++) { sg_count 2226 drivers/scsi/aic7xxx/aic79xx_core.c ahd_get_transfer_length(scb), scb->sg_count); sg_count 9015 drivers/scsi/aic7xxx/aic79xx_core.c scb->sg_count = 0; sg_count 835 drivers/scsi/aic7xxx/aic79xx_osm.c reset_scb->sg_count = 0; sg_count 1632 drivers/scsi/aic7xxx/aic79xx_osm.c scb->sg_count = 0; sg_count 575 drivers/scsi/aic7xxx/aic7xxx.h u_int sg_count;/* How full ahc_dma_seg is */ sg_count 406 drivers/scsi/aic7xxx/aic7xxx_core.c if (scb->sg_count == 0) sg_count 412 drivers/scsi/aic7xxx/aic7xxx_core.c /*len*/sizeof(struct ahc_dma_seg) * scb->sg_count, op); sg_count 1136 drivers/scsi/aic7xxx/aic7xxx_core.c scb->sg_count = 1; sg_count 1398 drivers/scsi/aic7xxx/aic7xxx_core.c ahc_get_transfer_length(scb), scb->sg_count); sg_count 1399 drivers/scsi/aic7xxx/aic7xxx_core.c if (scb->sg_count > 0) { sg_count 1400 drivers/scsi/aic7xxx/aic7xxx_core.c for (i = 0; i < scb->sg_count; i++) { sg_count 2114 drivers/scsi/aic7xxx/aic7xxx_core.c if (scb->sg_count > 0) { sg_count 2115 drivers/scsi/aic7xxx/aic7xxx_core.c for (i = 0; i < scb->sg_count; i++) { sg_count 475 drivers/scsi/aic7xxx/aic7xxx_osm.c if ((scb->sg_count + 1) > AHC_NSEG) sg_count 1517 drivers/scsi/aic7xxx/aic7xxx_osm.c scb->sg_count = 0; sg_count 1540 drivers/scsi/aic7xxx/aic7xxx_osm.c scb->sg_count += consumed; sg_count 1561 drivers/scsi/aic7xxx/aic7xxx_osm.c scb->sg_count = 0; sg_count 98 drivers/scsi/arm/scsi.h unsigned i, sg_count = scsi_sg_count(SCpnt); sg_count 100 drivers/scsi/arm/scsi.h scsi_for_each_sg(SCpnt, sg, sg_count, i) sg_count 689 drivers/scsi/atp870u.c unsigned long sg_count; sg_count 803 drivers/scsi/atp870u.c sg_count = scsi_dma_map(workreq); sg_count 854 drivers/scsi/atp870u.c scsi_for_each_sg(workreq, sgpnt, sg_count, j) { sg_count 1652 drivers/scsi/bnx2fc/bnx2fc_io.c int sg_count = 0; sg_count 1664 drivers/scsi/bnx2fc/bnx2fc_io.c sg_count = dma_map_sg(&hba->pcidev->dev, scsi_sglist(sc), sg_count 1666 drivers/scsi/bnx2fc/bnx2fc_io.c scsi_for_each_sg(sc, sg, sg_count, i) { sg_count 150 drivers/scsi/bnx2i/bnx2i_iscsi.c int sg_count; sg_count 157 drivers/scsi/bnx2i/bnx2i_iscsi.c sg_count = scsi_dma_map(sc); sg_count 159 drivers/scsi/bnx2i/bnx2i_iscsi.c scsi_for_each_sg(sc, sg, sg_count, i) { sg_count 236 drivers/scsi/dc395x.c u8 sg_count; /* No of HW sg entries for this request */ sg_count 888 drivers/scsi/dc395x.c srb->sg_count = 0; sg_count 917 drivers/scsi/dc395x.c srb->sg_count = nseg; sg_count 922 drivers/scsi/dc395x.c srb->sg_count); sg_count 924 drivers/scsi/dc395x.c scsi_for_each_sg(cmd, sg, srb->sg_count, i) { sg_count 931 drivers/scsi/dc395x.c sgp += srb->sg_count - 1; sg_count 1109 drivers/scsi/dc395x.c srb->segment_x, srb->sg_count, srb->sg_index, sg_count 1836 drivers/scsi/dc395x.c for (; idx < srb->sg_count; psge++, idx++) sg_count 1866 drivers/scsi/dc395x.c for (idx = srb->sg_index; idx < srb->sg_count; idx++) { sg_count 2182 drivers/scsi/dc395x.c srb->sg_count, &offset, &len); sg_count 2309 drivers/scsi/dc395x.c srb->total_xfer_length, srb->sg_index, srb->sg_count); sg_count 2312 drivers/scsi/dc395x.c if (srb->sg_index >= srb->sg_count) { sg_count 2344 drivers/scsi/dc395x.c ((u32)(srb->sg_count - sg_count 2399 drivers/scsi/dc395x.c srb->sg_count, &offset, &len); sg_count 2434 drivers/scsi/dc395x.c if (srb->sg_count) { sg_count 3220 drivers/scsi/dc395x.c srb, scsi_sg_count(cmd), srb->sg_index, srb->sg_count, sg_count 3596 drivers/scsi/dc395x.c srb->sg_count = 1; sg_count 1666 drivers/scsi/dpt_i2o.c u32 sg_count = 0; sg_count 1718 drivers/scsi/dpt_i2o.c sg_count = (size - sg_offset*4) / sizeof(struct sg_simple_element); sg_count 1719 drivers/scsi/dpt_i2o.c if (sg_count > pHba->sg_tablesize){ sg_count 1720 drivers/scsi/dpt_i2o.c printk(KERN_DEBUG"%s:IOCTL SG List too large (%u)\n", pHba->name,sg_count); sg_count 1725 drivers/scsi/dpt_i2o.c for(i = 0; i < sg_count; i++) { sg_count 1738 drivers/scsi/dpt_i2o.c pHba->name,sg_size,i,sg_count); sg_count 1805 drivers/scsi/dpt_i2o.c sg_count = (size - sg_offset*4) / sizeof(struct sg_simple_element); sg_count 1809 drivers/scsi/dpt_i2o.c for (j = 0; j < sg_count; j++) { sg_count 320 drivers/scsi/fnic/fnic_scsi.c int sg_count) sg_count 333 drivers/scsi/fnic/fnic_scsi.c if (sg_count) { sg_count 336 drivers/scsi/fnic/fnic_scsi.c for_each_sg(scsi_sglist(sc), sg, sg_count, i) { sg_count 345 drivers/scsi/fnic/fnic_scsi.c sizeof(io_req->sgl_list[0]) * sg_count, sg_count 359 drivers/scsi/fnic/fnic_scsi.c sizeof(io_req->sgl_list[0]) * sg_count, sg_count 432 drivers/scsi/fnic/fnic_scsi.c int sg_count = 0; sg_count 510 drivers/scsi/fnic/fnic_scsi.c sg_count = scsi_dma_map(sc); sg_count 511 drivers/scsi/fnic/fnic_scsi.c if (sg_count < 0) { sg_count 514 drivers/scsi/fnic/fnic_scsi.c sg_count, CMD_STATE(sc)); sg_count 520 drivers/scsi/fnic/fnic_scsi.c io_req->sgl_cnt = sg_count; sg_count 522 drivers/scsi/fnic/fnic_scsi.c if (sg_count > FNIC_DFLT_SG_DESC_CNT) sg_count 525 drivers/scsi/fnic/fnic_scsi.c if (sg_count) { sg_count 566 drivers/scsi/fnic/fnic_scsi.c ret = fnic_queue_wq_copy_desc(fnic, wq, io_req, sc, sg_count); sg_count 606 drivers/scsi/fnic/fnic_scsi.c sg_count, cmd_trace, sg_count 1105 drivers/scsi/hpsa.c c->busaddr |= (h->ioaccel2_blockFetchTable[cp->sg_count]); sg_count 4992 drivers/scsi/hpsa.c cp->sg_count = 1; sg_count 5000 drivers/scsi/hpsa.c cp->sg_count = (u8) use_sg; sg_count 9925 drivers/scsi/hpsa.c VERIFY_OFFSET(sg_count, 45); sg_count 587 drivers/scsi/hpsa_cmd.h u8 sg_count; /* Number of sg elements */ sg_count 1003 drivers/scsi/hptiop.c int sg_count = 0; sg_count 1041 drivers/scsi/hptiop.c sg_count = hptiop_buildsgl(scp, req->sg_list); sg_count 1042 drivers/scsi/hptiop.c if (!sg_count) sg_count 1055 drivers/scsi/hptiop.c + sg_count * sizeof(struct hpt_iopsg)); sg_count 2660 drivers/scsi/ips.c scb->sg_count = scsi_dma_map(SC); sg_count 2661 drivers/scsi/ips.c BUG_ON(scb->sg_count < 0); sg_count 2662 drivers/scsi/ips.c if (scb->sg_count) { sg_count 2668 drivers/scsi/ips.c scsi_for_each_sg(SC, sg, scb->sg_count, i) { sg_count 3573 drivers/scsi/ips.c scb->cmd.basic_io.sg_count = scb->sg_len; sg_count 3619 drivers/scsi/ips.c scb->cmd.basic_io.sg_count = scb->sg_len; sg_count 3767 drivers/scsi/ips.c tapeDCDB->sg_count = scb->sg_len; sg_count 3806 drivers/scsi/ips.c scb->dcdb.sg_count = scb->sg_len; sg_count 5672 drivers/scsi/ips.c scb->cmd.basic_io.sg_count = 0; sg_count 5715 drivers/scsi/ips.c scb->cmd.basic_io.sg_count = 0; sg_count 412 drivers/scsi/ips.h uint8_t sg_count; sg_count 621 drivers/scsi/ips.h uint8_t sg_count; sg_count 636 drivers/scsi/ips.h uint16_t sg_count; sg_count 1087 drivers/scsi/ips.h int sg_count; sg_count 371 drivers/scsi/libiscsi_tcp.c struct scatterlist *sg_list, unsigned int sg_count, sg_count 380 drivers/scsi/libiscsi_tcp.c for_each_sg(sg_list, sg, sg_count, i) { sg_count 190 drivers/scsi/mvumi.c void *sgl_p, unsigned char *sg_count) sg_count 198 drivers/scsi/mvumi.c *sg_count = dma_map_sg(&mhba->pdev->dev, scsi_sglist(scmd), sgnum, sg_count 200 drivers/scsi/mvumi.c if (*sg_count > mhba->max_sge) { sg_count 203 drivers/scsi/mvumi.c *sg_count, mhba->max_sge); sg_count 208 drivers/scsi/mvumi.c scsi_for_each_sg(scmd, sg, *sg_count, i) { sg_count 214 drivers/scsi/mvumi.c if ((i + 1) == *sg_count) sg_count 1618 drivers/scsi/myrb.c mbox->type5.sg_count = nsge; sg_count 652 drivers/scsi/myrb.h unsigned char sg_count:6; /* Byte 12 Bits 0-5 */ sg_count 288 drivers/scsi/qedf/qedf.h unsigned int sg_count; /* Number of SG elements */ sg_count 264 drivers/scsi/qedf/qedf_debugfs.c seq_printf(s, "%d:", io_log->sg_count); sg_count 487 drivers/scsi/qedf/qedf_io.c int sg_count = 0; sg_count 493 drivers/scsi/qedf/qedf_io.c sg_count = dma_map_sg(&qedf->pdev->dev, scsi_sglist(sc), sg_count 499 drivers/scsi/qedf/qedf_io.c if (sg_count <= 8 || io_req->io_req_flags == QEDF_READ) sg_count 502 drivers/scsi/qedf/qedf_io.c scsi_for_each_sg(sc, sg, sg_count, i) { sg_count 513 drivers/scsi/qedf/qedf_io.c (i != (sg_count - 1)) && sg_len < QEDF_PAGE_SIZE) sg_count 830 drivers/scsi/qedf/qedf_io.c io_log->sg_count = scsi_sg_count(sc_cmd); sg_count 230 drivers/scsi/qedi/qedi.h unsigned int sg_count; /* Number of SG elements */ sg_count 197 drivers/scsi/qedi/qedi_debugfs.c seq_printf(s, "sgcnt-%d:", io_log->sg_count); sg_count 1845 drivers/scsi/qedi/qedi_fw.c int sg_count; sg_count 1853 drivers/scsi/qedi/qedi_fw.c sg_count = dma_map_sg(&qedi->pdev->dev, scsi_sglist(sc), sg_count 1861 drivers/scsi/qedi/qedi_fw.c if ((sg_count == 1) && (sg_dma_len(sg) <= MAX_SGLEN_FOR_CACHESGL)) { sg_count 1871 drivers/scsi/qedi/qedi_fw.c sg_count, addr, sg_len); sg_count 1876 drivers/scsi/qedi/qedi_fw.c scsi_for_each_sg(sc, sg, sg_count, i) { sg_count 1885 drivers/scsi/qedi/qedi_fw.c if ((i == 0) && (sg_count > 1) && (end_addr % QEDI_PAGE_SIZE)) sg_count 1892 drivers/scsi/qedi/qedi_fw.c else if ((i == (sg_count - 1)) && sg_count 1893 drivers/scsi/qedi/qedi_fw.c (sg_count > 1) && (addr % QEDI_PAGE_SIZE)) sg_count 1900 drivers/scsi/qedi/qedi_fw.c else if ((i != 0) && (i != (sg_count - 1)) && sg_count 1995 drivers/scsi/qedi/qedi_fw.c io_log->sg_count = scsi_sg_count(sc_cmd); sg_count 906 drivers/scsi/qlogicpti.c int sg_count; sg_count 909 drivers/scsi/qlogicpti.c sg_count = dma_map_sg(&qpti->op->dev, sg, sg_count 914 drivers/scsi/qlogicpti.c cmd->segment_cnt = sg_count; sg_count 917 drivers/scsi/qlogicpti.c n = sg_count; sg_count 924 drivers/scsi/qlogicpti.c sg_count -= 4; sg_count 926 drivers/scsi/qlogicpti.c while (sg_count > 0) { sg_count 941 drivers/scsi/qlogicpti.c n = sg_count; sg_count 948 drivers/scsi/qlogicpti.c sg_count -= n; sg_count 2856 drivers/scsi/scsi_lib.c void *scsi_kmap_atomic_sg(struct scatterlist *sgl, int sg_count, sg_count 2866 drivers/scsi/scsi_lib.c for_each_sg(sgl, sg, sg_count, i) { sg_count 2873 drivers/scsi/scsi_lib.c if (unlikely(i == sg_count)) { sg_count 2876 drivers/scsi/scsi_lib.c __func__, sg_len, *offset, sg_count); sg_count 4793 drivers/scsi/smartpqi/smartpqi_init.c int sg_count; sg_count 4800 drivers/scsi/smartpqi/smartpqi_init.c sg_count = scsi_dma_map(scmd); sg_count 4801 drivers/scsi/smartpqi/smartpqi_init.c if (sg_count < 0) sg_count 4802 drivers/scsi/smartpqi/smartpqi_init.c return sg_count; sg_count 4807 drivers/scsi/smartpqi/smartpqi_init.c if (sg_count == 0) sg_count 4822 drivers/scsi/smartpqi/smartpqi_init.c if (i == sg_count) sg_count 4829 drivers/scsi/smartpqi/smartpqi_init.c put_unaligned_le32((sg_count - num_sg_in_iu) sg_count 4857 drivers/scsi/smartpqi/smartpqi_init.c int sg_count; sg_count 4864 drivers/scsi/smartpqi/smartpqi_init.c sg_count = scsi_dma_map(scmd); sg_count 4865 drivers/scsi/smartpqi/smartpqi_init.c if (sg_count < 0) sg_count 4866 drivers/scsi/smartpqi/smartpqi_init.c return sg_count; sg_count 4872 drivers/scsi/smartpqi/smartpqi_init.c if (sg_count == 0) sg_count 4886 drivers/scsi/smartpqi/smartpqi_init.c if (i == sg_count) sg_count 4893 drivers/scsi/smartpqi/smartpqi_init.c put_unaligned_le32((sg_count - num_sg_in_iu) sg_count 7590 drivers/scsi/smartpqi/smartpqi_init.c u32 sg_count; sg_count 7599 drivers/scsi/smartpqi/smartpqi_init.c sg_count = (total_size + chunk_size - 1); sg_count 7600 drivers/scsi/smartpqi/smartpqi_init.c sg_count /= chunk_size; sg_count 7604 drivers/scsi/smartpqi/smartpqi_init.c if (sg_count*chunk_size < total_size) sg_count 7608 drivers/scsi/smartpqi/smartpqi_init.c kcalloc(sg_count, sizeof(void *), GFP_KERNEL); sg_count 7631 drivers/scsi/smartpqi/smartpqi_init.c put_unaligned_le16(sg_count, &ofap->num_memory_descriptors); sg_count 458 drivers/scsi/sr.c int i, size = 0, sg_count = scsi_sg_count(SCpnt); sg_count 460 drivers/scsi/sr.c scsi_for_each_sg(SCpnt, sg, sg_count, i) sg_count 198 drivers/scsi/stex.c __le16 sg_count; sg_count 301 drivers/scsi/stex.c int sg_count; sg_count 440 drivers/scsi/stex.c ccb->sg_count = nseg; sg_count 441 drivers/scsi/stex.c dst->sg_count = cpu_to_le16((u16)nseg); sg_count 472 drivers/scsi/stex.c ccb->sg_count = nseg; sg_count 473 drivers/scsi/stex.c dst->sg_count = cpu_to_le16((u16)nseg); sg_count 549 drivers/scsi/stex.c addr += (hba->ccb[tag].sg_count+4)/11; sg_count 719 drivers/scsi/stex.c hba->ccb[tag].sg_count = 0; sg_count 1896 drivers/scsi/stex.c hba->ccb[tag].sg_count = 0; sg_count 1566 drivers/scsi/storvsc_drv.c unsigned int sg_count = 0; sg_count 1639 drivers/scsi/storvsc_drv.c sg_count = scsi_sg_count(scmnd); sg_count 1645 drivers/scsi/storvsc_drv.c if (sg_count) { sg_count 1646 drivers/scsi/storvsc_drv.c if (sg_count > MAX_PAGE_BUFFER_COUNT) { sg_count 1648 drivers/scsi/storvsc_drv.c payload_sz = (sg_count * sizeof(u64) + sg_count 1659 drivers/scsi/storvsc_drv.c for (i = 0; i < sg_count; i++) { sg_count 173 drivers/spi/spi-mxs.c int sg_count; sg_count 201 drivers/spi/spi-mxs.c for (sg_count = 0; sg_count < sgs; sg_count++) { sg_count 209 drivers/spi/spi-mxs.c if ((sg_count + 1 == sgs) && (flags & TXRX_DEASSERT_CS)) sg_count 217 drivers/spi/spi-mxs.c dma_xfer[sg_count].pio[0] = ctrl0; sg_count 218 drivers/spi/spi-mxs.c dma_xfer[sg_count].pio[3] = min; sg_count 227 drivers/spi/spi-mxs.c sg_init_table(&dma_xfer[sg_count].sg, 1); sg_count 228 drivers/spi/spi-mxs.c sg_set_page(&dma_xfer[sg_count].sg, vm_page, sg_count 231 drivers/spi/spi-mxs.c sg_init_one(&dma_xfer[sg_count].sg, buf, min); sg_count 234 drivers/spi/spi-mxs.c ret = dma_map_sg(ssp->dev, &dma_xfer[sg_count].sg, 1, sg_count 242 drivers/spi/spi-mxs.c (struct scatterlist *)dma_xfer[sg_count].pio, sg_count 245 drivers/spi/spi-mxs.c sg_count ? DMA_PREP_INTERRUPT : 0); sg_count 254 drivers/spi/spi-mxs.c &dma_xfer[sg_count].sg, 1, sg_count 288 drivers/spi/spi-mxs.c while (--sg_count >= 0) { sg_count 290 drivers/spi/spi-mxs.c dma_unmap_sg(ssp->dev, &dma_xfer[sg_count].sg, 1, sg_count 673 drivers/vhost/scsi.c struct scatterlist *sg, int sg_count) sg_count 141 include/linux/mmc/core.h int sg_count; /* mapped sg entries */ sg_count 106 include/scsi/libiscsi_tcp.h struct scatterlist *sg_list, unsigned int sg_count, sg_count 164 include/scsi/scsi_cmnd.h extern void *scsi_kmap_atomic_sg(struct scatterlist *sg, int sg_count, sg_count 376 sound/soc/intel/atom/sst-mfld-dsp.h __u16 sg_count; sg_count 402 sound/soc/intel/atom/sst-mfld-dsp.h u8 sg_count; sg_count 151 sound/soc/intel/atom/sst-mfld-platform-compress.c str_params.aparams.sg_count = 1; sg_count 137 sound/soc/intel/atom/sst-mfld-platform-pcm.c alloc_param->sg_count = 1; sg_count 47 sound/soc/intel/atom/sst/sst_stream.c str_info->alloc_param.sg_count = str_params->aparams.sg_count;