nvecs 2398 drivers/irqchip/irq-gic-v3-its.c int nvecs, bool alloc_lpis) nvecs 2413 drivers/irqchip/irq-gic-v3-its.c if (WARN_ON(!is_power_of_2(nvecs))) nvecs 2414 drivers/irqchip/irq-gic-v3-its.c nvecs = roundup_pow_of_two(nvecs); nvecs 2421 drivers/irqchip/irq-gic-v3-its.c nr_ites = max(2, nvecs); nvecs 2426 drivers/irqchip/irq-gic-v3-its.c lpi_map = its_lpi_alloc(nvecs, &lpi_base, &nr_lpis); nvecs 2478 drivers/irqchip/irq-gic-v3-its.c static int its_alloc_device_irq(struct its_device *dev, int nvecs, irq_hw_number_t *hwirq) nvecs 2485 drivers/irqchip/irq-gic-v3-its.c get_count_order(nvecs)); nvecs 1614 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c static void bnx2x_free_msix_irqs(struct bnx2x *bp, int nvecs) nvecs 1618 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c if (nvecs == offset) nvecs 1630 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c if (nvecs == offset) nvecs 1636 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c if (nvecs == offset) nvecs 1649 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c int nvecs = BNX2X_NUM_ETH_QUEUES(bp) + CNIC_SUPPORT(bp); nvecs 1653 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c nvecs++; nvecs 1655 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c bnx2x_free_msix_irqs(bp, nvecs); nvecs 95 drivers/net/ethernet/brocade/bna/bnad.c int vector, nvecs; nvecs 98 drivers/net/ethernet/brocade/bna/bnad.c nvecs = unmap->nvecs; nvecs 102 drivers/net/ethernet/brocade/bna/bnad.c unmap->nvecs = 0; nvecs 107 drivers/net/ethernet/brocade/bna/bnad.c nvecs--; nvecs 110 drivers/net/ethernet/brocade/bna/bnad.c while (nvecs) { nvecs 123 drivers/net/ethernet/brocade/bna/bnad.c nvecs--; nvecs 187 drivers/net/ethernet/brocade/bna/bnad.c unmap_wis = BNA_TXQ_WI_NEEDED(unmap->nvecs); nvecs 500 drivers/net/ethernet/brocade/bna/bnad.c u32 sop_ci, u32 nvecs) nvecs 507 drivers/net/ethernet/brocade/bna/bnad.c for (vec = 0, ci = sop_ci; vec < nvecs; vec++) { nvecs 519 drivers/net/ethernet/brocade/bna/bnad.c bnad_cq_setup_skb_frags(struct bna_ccb *ccb, struct sk_buff *skb, u32 nvecs) nvecs 540 drivers/net/ethernet/brocade/bna/bnad.c while (nvecs--) { nvecs 597 drivers/net/ethernet/brocade/bna/bnad.c u32 pi, vec, sop_ci = 0, nvecs = 0; nvecs 642 drivers/net/ethernet/brocade/bna/bnad.c nvecs = 1; nvecs 670 drivers/net/ethernet/brocade/bna/bnad.c nvecs++; nvecs 683 drivers/net/ethernet/brocade/bna/bnad.c bnad_cq_drop_packet(bnad, rcb, sop_ci, nvecs); nvecs 692 drivers/net/ethernet/brocade/bna/bnad.c bnad_cq_setup_skb_frags(ccb, skb, nvecs); nvecs 720 drivers/net/ethernet/brocade/bna/bnad.c BNA_QE_INDX_ADD(rcb->consumer_index, nvecs, rcb->q_depth); nvecs 721 drivers/net/ethernet/brocade/bna/bnad.c for (vec = 0; vec < nvecs; vec++) { nvecs 3018 drivers/net/ethernet/brocade/bna/bnad.c head_unmap->nvecs = 0; nvecs 3032 drivers/net/ethernet/brocade/bna/bnad.c head_unmap->nvecs++; nvecs 3074 drivers/net/ethernet/brocade/bna/bnad.c head_unmap->nvecs++; nvecs 225 drivers/net/ethernet/brocade/bna/bnad.h u32 nvecs; nvecs 466 drivers/net/ethernet/marvell/octeontx2/af/rvu.c static void rvu_check_min_msix_vec(struct rvu *rvu, int nvecs, int pf, int vf) nvecs 473 drivers/net/ethernet/marvell/octeontx2/af/rvu.c if (!nvecs) { nvecs 476 drivers/net/ethernet/marvell/octeontx2/af/rvu.c pf, vf - 1, nvecs); nvecs 486 drivers/net/ethernet/marvell/octeontx2/af/rvu.c if (!(nvecs < min_vecs)) nvecs 490 drivers/net/ethernet/marvell/octeontx2/af/rvu.c pf, nvecs, min_vecs); nvecs 497 drivers/net/ethernet/marvell/octeontx2/af/rvu.c int nvecs, offset, max_msix; nvecs 538 drivers/net/ethernet/marvell/octeontx2/af/rvu.c nvecs = (cfg >> 12) & 0xFF; nvecs 540 drivers/net/ethernet/marvell/octeontx2/af/rvu.c offset = rvu_alloc_rsrc_contig(&pfvf->msix, nvecs); nvecs 570 drivers/net/ethernet/marvell/octeontx2/af/rvu.c nvecs = (cfg >> 12) & 0xFF; nvecs 572 drivers/net/ethernet/marvell/octeontx2/af/rvu.c offset = rvu_alloc_rsrc_contig(&pfvf->msix, nvecs); nvecs 1251 drivers/net/ethernet/marvell/octeontx2/af/rvu.c u16 nvecs, vec, offset; nvecs 1256 drivers/net/ethernet/marvell/octeontx2/af/rvu.c nvecs = (cfg >> 12) & 0xFF; nvecs 1259 drivers/net/ethernet/marvell/octeontx2/af/rvu.c if (!rvu_rsrc_check_contig(&pfvf->msix, nvecs)) nvecs 1262 drivers/net/ethernet/marvell/octeontx2/af/rvu.c offset = rvu_alloc_rsrc_contig(&pfvf->msix, nvecs); nvecs 1269 drivers/net/ethernet/marvell/octeontx2/af/rvu.c for (vec = 0; vec < nvecs; vec++) nvecs 1276 drivers/net/ethernet/marvell/octeontx2/af/rvu.c u16 nvecs, vec, offset; nvecs 1281 drivers/net/ethernet/marvell/octeontx2/af/rvu.c nvecs = (cfg >> 12) & 0xFF; nvecs 1290 drivers/net/ethernet/marvell/octeontx2/af/rvu.c for (vec = 0; vec < nvecs; vec++) nvecs 1294 drivers/net/ethernet/marvell/octeontx2/af/rvu.c rvu_free_rsrc_contig(&pfvf->msix, nvecs, offset); nvecs 1265 drivers/pci/switch/switchtec.c int nvecs; nvecs 1273 drivers/pci/switch/switchtec.c nvecs = pci_alloc_irq_vectors(stdev->pdev, 1, nirqs, nvecs 1276 drivers/pci/switch/switchtec.c if (nvecs < 0) nvecs 1277 drivers/pci/switch/switchtec.c return nvecs; nvecs 1280 drivers/pci/switch/switchtec.c if (event_irq < 0 || event_irq >= nvecs) nvecs 1298 drivers/pci/switch/switchtec.c if (dma_mrpc_irq < 0 || dma_mrpc_irq >= nvecs) nvecs 169 drivers/scsi/bfa/bfa.h int nvecs; nvecs 180 drivers/scsi/bfa/bfa.h void (*hw_msix_init)(struct bfa_s *bfa, int nvecs); nvecs 186 drivers/scsi/bfa/bfa.h u32 *nvecs, u32 *maxvec); nvecs 314 drivers/scsi/bfa/bfa.h void bfa_hwcb_msix_init(struct bfa_s *bfa, int nvecs); nvecs 319 drivers/scsi/bfa/bfa.h void bfa_hwcb_msix_getvecs(struct bfa_s *bfa, u32 *vecmap, u32 *nvecs, nvecs 328 drivers/scsi/bfa/bfa.h void bfa_hwct_msix_init(struct bfa_s *bfa, int nvecs); nvecs 333 drivers/scsi/bfa/bfa.h void bfa_hwct_msix_getvecs(struct bfa_s *bfa, u32 *vecmap, u32 *nvecs, nvecs 866 drivers/scsi/bfa/bfa_core.c bfa_isr_mode_set(bfa, bfa->msix.nvecs != 0); nvecs 989 drivers/scsi/bfa/bfa_core.c if (bfa->msix.nvecs == 1) nvecs 1085 drivers/scsi/bfa/bfa_core.c bfa->msix.nvecs = 0; nvecs 106 drivers/scsi/bfa/bfa_hw_cb.c bfa_hwcb_msix_init(struct bfa_s *bfa, int nvecs) nvecs 108 drivers/scsi/bfa/bfa_hw_cb.c WARN_ON((nvecs != 1) && (nvecs != __HFN_NUMINTS)); nvecs 110 drivers/scsi/bfa/bfa_hw_cb.c bfa->msix.nvecs = nvecs; nvecs 119 drivers/scsi/bfa/bfa_hw_cb.c if (bfa->msix.nvecs == 0) nvecs 122 drivers/scsi/bfa/bfa_hw_cb.c if (bfa->msix.nvecs == 1) { nvecs 137 drivers/scsi/bfa/bfa_hw_cb.c if (bfa->msix.nvecs == 0) nvecs 140 drivers/scsi/bfa/bfa_hw_cb.c if (bfa->msix.nvecs == 1) { nvecs 103 drivers/scsi/bfa/bfa_hw_ct.c bfa_hwct_msix_init(struct bfa_s *bfa, int nvecs) nvecs 105 drivers/scsi/bfa/bfa_hw_ct.c WARN_ON((nvecs != 1) && (nvecs != BFI_MSIX_CT_MAX)); nvecs 106 drivers/scsi/bfa/bfa_hw_ct.c bfa_trc(bfa, nvecs); nvecs 108 drivers/scsi/bfa/bfa_hw_ct.c bfa->msix.nvecs = nvecs; nvecs 115 drivers/scsi/bfa/bfa_hw_ct.c if (bfa->msix.nvecs == 0) nvecs 118 drivers/scsi/bfa/bfa_hw_ct.c if (bfa->msix.nvecs == 1) nvecs 129 drivers/scsi/bfa/bfa_hw_ct.c if (bfa->msix.nvecs == 0) nvecs 132 drivers/scsi/bfa/bfa_hw_ct.c if (bfa->msix.nvecs == 1) { nvecs 477 drivers/scsi/csiostor/csio_isr.c static void csio_calc_sets(struct irq_affinity *affd, unsigned int nvecs) nvecs 482 drivers/scsi/csiostor/csio_isr.c if (!nvecs) nvecs 485 drivers/scsi/csiostor/csio_isr.c if (nvecs < hw->num_pports) { nvecs 487 drivers/scsi/csiostor/csio_isr.c affd->set_size[0] = nvecs; nvecs 493 drivers/scsi/csiostor/csio_isr.c affd->set_size[i] = nvecs / hw->num_pports; nvecs 2347 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c static void setup_reply_map_v3_hw(struct hisi_hba *hisi_hba, int nvecs) nvecs 2352 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c for (queue = 0; queue < nvecs; queue++) { nvecs 2075 fs/cifs/smbdirect.c int nvecs; nvecs 2147 fs/cifs/smbdirect.c nvecs = (buflen+max_iov_size-1)/max_iov_size; nvecs 2151 fs/cifs/smbdirect.c buflen, nvecs); nvecs 2152 fs/cifs/smbdirect.c for (j = 0; j < nvecs; j++) { nvecs 2157 fs/cifs/smbdirect.c if (j == nvecs-1) nvecs 2160 fs/cifs/smbdirect.c max_iov_size*(nvecs-1); nvecs 2205 fs/cifs/smbdirect.c nvecs = (buflen + max_iov_size - 1) / max_iov_size; nvecs 2207 fs/cifs/smbdirect.c buflen, nvecs); nvecs 2208 fs/cifs/smbdirect.c for (j = 0; j < nvecs; j++) { nvecs 2210 fs/cifs/smbdirect.c if (j == nvecs-1) nvecs 190 fs/nfsd/nfs3proc.c unsigned int nvecs; nvecs 200 fs/nfsd/nfs3proc.c nvecs = svc_fill_write_vector(rqstp, rqstp->rq_arg.pages, nvecs 202 fs/nfsd/nfs3proc.c if (!nvecs) nvecs 205 fs/nfsd/nfs3proc.c rqstp->rq_vec, nvecs, &cnt, nvecs 993 fs/nfsd/nfs4proc.c int nvecs; nvecs 1011 fs/nfsd/nfs4proc.c nvecs = svc_fill_write_vector(rqstp, write->wr_pagelist, nvecs 1013 fs/nfsd/nfs4proc.c WARN_ON_ONCE(nvecs > ARRAY_SIZE(rqstp->rq_vec)); nvecs 1016 fs/nfsd/nfs4proc.c write->wr_offset, rqstp->rq_vec, nvecs, &cnt, nvecs 217 fs/nfsd/nfsproc.c unsigned int nvecs; nvecs 223 fs/nfsd/nfsproc.c nvecs = svc_fill_write_vector(rqstp, rqstp->rq_arg.pages, nvecs 225 fs/nfsd/nfsproc.c if (!nvecs) nvecs 228 fs/nfsd/nfsproc.c argp->offset, rqstp->rq_vec, nvecs, nvecs 63 fs/xfs/xfs_bmap_item.c int *nvecs, nvecs 68 fs/xfs/xfs_bmap_item.c *nvecs += 1; nvecs 163 fs/xfs/xfs_bmap_item.c int *nvecs, nvecs 166 fs/xfs/xfs_bmap_item.c *nvecs += 1; nvecs 52 fs/xfs/xfs_buf_item.c int *nvecs, nvecs 67 fs/xfs/xfs_buf_item.c *nvecs += 2; nvecs 88 fs/xfs/xfs_buf_item.c (*nvecs)++; nvecs 93 fs/xfs/xfs_buf_item.c (*nvecs)++; nvecs 121 fs/xfs/xfs_buf_item.c int *nvecs, nvecs 136 fs/xfs/xfs_buf_item.c *nvecs += bip->bli_format_count; nvecs 152 fs/xfs/xfs_buf_item.c *nvecs = XFS_LOG_VEC_ORDERED; nvecs 167 fs/xfs/xfs_buf_item.c nvecs, nbytes); nvecs 32 fs/xfs/xfs_dquot_item.c int *nvecs, nvecs 35 fs/xfs/xfs_dquot_item.c *nvecs += 2; nvecs 268 fs/xfs/xfs_dquot_item.c int *nvecs, nvecs 271 fs/xfs/xfs_dquot_item.c *nvecs += 1; nvecs 79 fs/xfs/xfs_extfree_item.c int *nvecs, nvecs 82 fs/xfs/xfs_extfree_item.c *nvecs += 1; nvecs 265 fs/xfs/xfs_extfree_item.c int *nvecs, nvecs 268 fs/xfs/xfs_extfree_item.c *nvecs += 1; nvecs 30 fs/xfs/xfs_icreate_item.c int *nvecs, nvecs 33 fs/xfs/xfs_icreate_item.c *nvecs += 1; nvecs 33 fs/xfs/xfs_inode_item.c int *nvecs, nvecs 45 fs/xfs/xfs_inode_item.c *nvecs += 1; nvecs 52 fs/xfs/xfs_inode_item.c *nvecs += 1; nvecs 59 fs/xfs/xfs_inode_item.c *nvecs += 1; nvecs 74 fs/xfs/xfs_inode_item.c int *nvecs, nvecs 86 fs/xfs/xfs_inode_item.c *nvecs += 1; nvecs 93 fs/xfs/xfs_inode_item.c *nvecs += 1; nvecs 100 fs/xfs/xfs_inode_item.c *nvecs += 1; nvecs 119 fs/xfs/xfs_inode_item.c int *nvecs, nvecs 125 fs/xfs/xfs_inode_item.c *nvecs += 2; nvecs 129 fs/xfs/xfs_inode_item.c xfs_inode_item_data_fork_size(iip, nvecs, nbytes); nvecs 131 fs/xfs/xfs_inode_item.c xfs_inode_item_attr_fork_size(iip, nvecs, nbytes); nvecs 412 fs/xfs/xfs_log_cil.c ctx->nvecs += diff_iovecs; nvecs 244 fs/xfs/xfs_log_priv.h int nvecs; /* number of regions */ nvecs 62 fs/xfs/xfs_refcount_item.c int *nvecs, nvecs 67 fs/xfs/xfs_refcount_item.c *nvecs += 1; nvecs 168 fs/xfs/xfs_refcount_item.c int *nvecs, nvecs 171 fs/xfs/xfs_refcount_item.c *nvecs += 1; nvecs 61 fs/xfs/xfs_rmap_item.c int *nvecs, nvecs 66 fs/xfs/xfs_rmap_item.c *nvecs += 1; nvecs 189 fs/xfs/xfs_rmap_item.c int *nvecs, nvecs 192 fs/xfs/xfs_rmap_item.c *nvecs += 1; nvecs 284 include/linux/interrupt.h void (*calc_sets)(struct irq_affinity *, unsigned int nvecs); nvecs 416 kernel/irq/affinity.c irq_create_affinity_masks(unsigned int nvecs, struct irq_affinity *affd) nvecs 427 kernel/irq/affinity.c if (nvecs > affd->pre_vectors + affd->post_vectors) nvecs 428 kernel/irq/affinity.c affvecs = nvecs - affd->pre_vectors - affd->post_vectors; nvecs 449 kernel/irq/affinity.c masks = kcalloc(nvecs, sizeof(*masks), GFP_KERNEL); nvecs 480 kernel/irq/affinity.c for (; curvec < nvecs; curvec++) nvecs 484 kernel/irq/affinity.c for (i = affd->pre_vectors; i < nvecs - affd->post_vectors; i++)