nvecs            2398 drivers/irqchip/irq-gic-v3-its.c 					    int nvecs, bool alloc_lpis)
nvecs            2413 drivers/irqchip/irq-gic-v3-its.c 	if (WARN_ON(!is_power_of_2(nvecs)))
nvecs            2414 drivers/irqchip/irq-gic-v3-its.c 		nvecs = roundup_pow_of_two(nvecs);
nvecs            2421 drivers/irqchip/irq-gic-v3-its.c 	nr_ites = max(2, nvecs);
nvecs            2426 drivers/irqchip/irq-gic-v3-its.c 		lpi_map = its_lpi_alloc(nvecs, &lpi_base, &nr_lpis);
nvecs            2478 drivers/irqchip/irq-gic-v3-its.c static int its_alloc_device_irq(struct its_device *dev, int nvecs, irq_hw_number_t *hwirq)
nvecs            2485 drivers/irqchip/irq-gic-v3-its.c 				      get_count_order(nvecs));
nvecs            1614 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c static void bnx2x_free_msix_irqs(struct bnx2x *bp, int nvecs)
nvecs            1618 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	if (nvecs == offset)
nvecs            1630 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		if (nvecs == offset)
nvecs            1636 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		if (nvecs == offset)
nvecs            1649 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		int nvecs = BNX2X_NUM_ETH_QUEUES(bp) + CNIC_SUPPORT(bp);
nvecs            1653 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 			nvecs++;
nvecs            1655 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		bnx2x_free_msix_irqs(bp, nvecs);
nvecs              95 drivers/net/ethernet/brocade/bna/bnad.c 	int vector, nvecs;
nvecs              98 drivers/net/ethernet/brocade/bna/bnad.c 	nvecs = unmap->nvecs;
nvecs             102 drivers/net/ethernet/brocade/bna/bnad.c 	unmap->nvecs = 0;
nvecs             107 drivers/net/ethernet/brocade/bna/bnad.c 	nvecs--;
nvecs             110 drivers/net/ethernet/brocade/bna/bnad.c 	while (nvecs) {
nvecs             123 drivers/net/ethernet/brocade/bna/bnad.c 		nvecs--;
nvecs             187 drivers/net/ethernet/brocade/bna/bnad.c 		unmap_wis = BNA_TXQ_WI_NEEDED(unmap->nvecs);
nvecs             500 drivers/net/ethernet/brocade/bna/bnad.c 		    u32 sop_ci, u32 nvecs)
nvecs             507 drivers/net/ethernet/brocade/bna/bnad.c 	for (vec = 0, ci = sop_ci; vec < nvecs; vec++) {
nvecs             519 drivers/net/ethernet/brocade/bna/bnad.c bnad_cq_setup_skb_frags(struct bna_ccb *ccb, struct sk_buff *skb, u32 nvecs)
nvecs             540 drivers/net/ethernet/brocade/bna/bnad.c 	while (nvecs--) {
nvecs             597 drivers/net/ethernet/brocade/bna/bnad.c 	u32 pi, vec, sop_ci = 0, nvecs = 0;
nvecs             642 drivers/net/ethernet/brocade/bna/bnad.c 		nvecs = 1;
nvecs             670 drivers/net/ethernet/brocade/bna/bnad.c 				nvecs++;
nvecs             683 drivers/net/ethernet/brocade/bna/bnad.c 			bnad_cq_drop_packet(bnad, rcb, sop_ci, nvecs);
nvecs             692 drivers/net/ethernet/brocade/bna/bnad.c 			bnad_cq_setup_skb_frags(ccb, skb, nvecs);
nvecs             720 drivers/net/ethernet/brocade/bna/bnad.c 		BNA_QE_INDX_ADD(rcb->consumer_index, nvecs, rcb->q_depth);
nvecs             721 drivers/net/ethernet/brocade/bna/bnad.c 		for (vec = 0; vec < nvecs; vec++) {
nvecs            3018 drivers/net/ethernet/brocade/bna/bnad.c 	head_unmap->nvecs = 0;
nvecs            3032 drivers/net/ethernet/brocade/bna/bnad.c 	head_unmap->nvecs++;
nvecs            3074 drivers/net/ethernet/brocade/bna/bnad.c 		head_unmap->nvecs++;
nvecs             225 drivers/net/ethernet/brocade/bna/bnad.h 	u32			nvecs;
nvecs             466 drivers/net/ethernet/marvell/octeontx2/af/rvu.c static void rvu_check_min_msix_vec(struct rvu *rvu, int nvecs, int pf, int vf)
nvecs             473 drivers/net/ethernet/marvell/octeontx2/af/rvu.c 	if (!nvecs) {
nvecs             476 drivers/net/ethernet/marvell/octeontx2/af/rvu.c 			 pf, vf - 1, nvecs);
nvecs             486 drivers/net/ethernet/marvell/octeontx2/af/rvu.c 	if (!(nvecs < min_vecs))
nvecs             490 drivers/net/ethernet/marvell/octeontx2/af/rvu.c 		 pf, nvecs, min_vecs);
nvecs             497 drivers/net/ethernet/marvell/octeontx2/af/rvu.c 	int nvecs, offset, max_msix;
nvecs             538 drivers/net/ethernet/marvell/octeontx2/af/rvu.c 		nvecs = (cfg >> 12) & 0xFF;
nvecs             540 drivers/net/ethernet/marvell/octeontx2/af/rvu.c 		offset = rvu_alloc_rsrc_contig(&pfvf->msix, nvecs);
nvecs             570 drivers/net/ethernet/marvell/octeontx2/af/rvu.c 			nvecs = (cfg >> 12) & 0xFF;
nvecs             572 drivers/net/ethernet/marvell/octeontx2/af/rvu.c 			offset = rvu_alloc_rsrc_contig(&pfvf->msix, nvecs);
nvecs            1251 drivers/net/ethernet/marvell/octeontx2/af/rvu.c 	u16 nvecs, vec, offset;
nvecs            1256 drivers/net/ethernet/marvell/octeontx2/af/rvu.c 	nvecs = (cfg >> 12) & 0xFF;
nvecs            1259 drivers/net/ethernet/marvell/octeontx2/af/rvu.c 	if (!rvu_rsrc_check_contig(&pfvf->msix, nvecs))
nvecs            1262 drivers/net/ethernet/marvell/octeontx2/af/rvu.c 	offset = rvu_alloc_rsrc_contig(&pfvf->msix, nvecs);
nvecs            1269 drivers/net/ethernet/marvell/octeontx2/af/rvu.c 	for (vec = 0; vec < nvecs; vec++)
nvecs            1276 drivers/net/ethernet/marvell/octeontx2/af/rvu.c 	u16 nvecs, vec, offset;
nvecs            1281 drivers/net/ethernet/marvell/octeontx2/af/rvu.c 	nvecs = (cfg >> 12) & 0xFF;
nvecs            1290 drivers/net/ethernet/marvell/octeontx2/af/rvu.c 	for (vec = 0; vec < nvecs; vec++)
nvecs            1294 drivers/net/ethernet/marvell/octeontx2/af/rvu.c 	rvu_free_rsrc_contig(&pfvf->msix, nvecs, offset);
nvecs            1265 drivers/pci/switch/switchtec.c 	int nvecs;
nvecs            1273 drivers/pci/switch/switchtec.c 	nvecs = pci_alloc_irq_vectors(stdev->pdev, 1, nirqs,
nvecs            1276 drivers/pci/switch/switchtec.c 	if (nvecs < 0)
nvecs            1277 drivers/pci/switch/switchtec.c 		return nvecs;
nvecs            1280 drivers/pci/switch/switchtec.c 	if (event_irq < 0 || event_irq >= nvecs)
nvecs            1298 drivers/pci/switch/switchtec.c 	if (dma_mrpc_irq < 0 || dma_mrpc_irq >= nvecs)
nvecs             169 drivers/scsi/bfa/bfa.h 	int	nvecs;
nvecs             180 drivers/scsi/bfa/bfa.h 	void (*hw_msix_init)(struct bfa_s *bfa, int nvecs);
nvecs             186 drivers/scsi/bfa/bfa.h 				u32 *nvecs, u32 *maxvec);
nvecs             314 drivers/scsi/bfa/bfa.h void bfa_hwcb_msix_init(struct bfa_s *bfa, int nvecs);
nvecs             319 drivers/scsi/bfa/bfa.h void bfa_hwcb_msix_getvecs(struct bfa_s *bfa, u32 *vecmap, u32 *nvecs,
nvecs             328 drivers/scsi/bfa/bfa.h void bfa_hwct_msix_init(struct bfa_s *bfa, int nvecs);
nvecs             333 drivers/scsi/bfa/bfa.h void bfa_hwct_msix_getvecs(struct bfa_s *bfa, u32 *vecmap, u32 *nvecs,
nvecs             866 drivers/scsi/bfa/bfa_core.c 	bfa_isr_mode_set(bfa, bfa->msix.nvecs != 0);
nvecs             989 drivers/scsi/bfa/bfa_core.c 	if (bfa->msix.nvecs == 1)
nvecs            1085 drivers/scsi/bfa/bfa_core.c 	bfa->msix.nvecs = 0;
nvecs             106 drivers/scsi/bfa/bfa_hw_cb.c bfa_hwcb_msix_init(struct bfa_s *bfa, int nvecs)
nvecs             108 drivers/scsi/bfa/bfa_hw_cb.c 	WARN_ON((nvecs != 1) && (nvecs != __HFN_NUMINTS));
nvecs             110 drivers/scsi/bfa/bfa_hw_cb.c 	bfa->msix.nvecs = nvecs;
nvecs             119 drivers/scsi/bfa/bfa_hw_cb.c 	if (bfa->msix.nvecs == 0)
nvecs             122 drivers/scsi/bfa/bfa_hw_cb.c 	if (bfa->msix.nvecs == 1) {
nvecs             137 drivers/scsi/bfa/bfa_hw_cb.c 	if (bfa->msix.nvecs == 0)
nvecs             140 drivers/scsi/bfa/bfa_hw_cb.c 	if (bfa->msix.nvecs == 1) {
nvecs             103 drivers/scsi/bfa/bfa_hw_ct.c bfa_hwct_msix_init(struct bfa_s *bfa, int nvecs)
nvecs             105 drivers/scsi/bfa/bfa_hw_ct.c 	WARN_ON((nvecs != 1) && (nvecs != BFI_MSIX_CT_MAX));
nvecs             106 drivers/scsi/bfa/bfa_hw_ct.c 	bfa_trc(bfa, nvecs);
nvecs             108 drivers/scsi/bfa/bfa_hw_ct.c 	bfa->msix.nvecs = nvecs;
nvecs             115 drivers/scsi/bfa/bfa_hw_ct.c 	if (bfa->msix.nvecs == 0)
nvecs             118 drivers/scsi/bfa/bfa_hw_ct.c 	if (bfa->msix.nvecs == 1)
nvecs             129 drivers/scsi/bfa/bfa_hw_ct.c 	if (bfa->msix.nvecs == 0)
nvecs             132 drivers/scsi/bfa/bfa_hw_ct.c 	if (bfa->msix.nvecs == 1) {
nvecs             477 drivers/scsi/csiostor/csio_isr.c static void csio_calc_sets(struct irq_affinity *affd, unsigned int nvecs)
nvecs             482 drivers/scsi/csiostor/csio_isr.c 	if (!nvecs)
nvecs             485 drivers/scsi/csiostor/csio_isr.c 	if (nvecs < hw->num_pports) {
nvecs             487 drivers/scsi/csiostor/csio_isr.c 		affd->set_size[0] = nvecs;
nvecs             493 drivers/scsi/csiostor/csio_isr.c 		affd->set_size[i] = nvecs / hw->num_pports;
nvecs            2347 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c static void setup_reply_map_v3_hw(struct hisi_hba *hisi_hba, int nvecs)
nvecs            2352 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 	for (queue = 0; queue < nvecs; queue++) {
nvecs            2075 fs/cifs/smbdirect.c 	int nvecs;
nvecs            2147 fs/cifs/smbdirect.c 				nvecs = (buflen+max_iov_size-1)/max_iov_size;
nvecs            2151 fs/cifs/smbdirect.c 					buflen, nvecs);
nvecs            2152 fs/cifs/smbdirect.c 				for (j = 0; j < nvecs; j++) {
nvecs            2157 fs/cifs/smbdirect.c 					if (j == nvecs-1)
nvecs            2160 fs/cifs/smbdirect.c 							max_iov_size*(nvecs-1);
nvecs            2205 fs/cifs/smbdirect.c 		nvecs = (buflen + max_iov_size - 1) / max_iov_size;
nvecs            2207 fs/cifs/smbdirect.c 			buflen, nvecs);
nvecs            2208 fs/cifs/smbdirect.c 		for (j = 0; j < nvecs; j++) {
nvecs            2210 fs/cifs/smbdirect.c 			if (j == nvecs-1)
nvecs             190 fs/nfsd/nfs3proc.c 	unsigned int nvecs;
nvecs             200 fs/nfsd/nfs3proc.c 	nvecs = svc_fill_write_vector(rqstp, rqstp->rq_arg.pages,
nvecs             202 fs/nfsd/nfs3proc.c 	if (!nvecs)
nvecs             205 fs/nfsd/nfs3proc.c 			    rqstp->rq_vec, nvecs, &cnt,
nvecs             993 fs/nfsd/nfs4proc.c 	int nvecs;
nvecs            1011 fs/nfsd/nfs4proc.c 	nvecs = svc_fill_write_vector(rqstp, write->wr_pagelist,
nvecs            1013 fs/nfsd/nfs4proc.c 	WARN_ON_ONCE(nvecs > ARRAY_SIZE(rqstp->rq_vec));
nvecs            1016 fs/nfsd/nfs4proc.c 				write->wr_offset, rqstp->rq_vec, nvecs, &cnt,
nvecs             217 fs/nfsd/nfsproc.c 	unsigned int nvecs;
nvecs             223 fs/nfsd/nfsproc.c 	nvecs = svc_fill_write_vector(rqstp, rqstp->rq_arg.pages,
nvecs             225 fs/nfsd/nfsproc.c 	if (!nvecs)
nvecs             228 fs/nfsd/nfsproc.c 			    argp->offset, rqstp->rq_vec, nvecs,
nvecs              63 fs/xfs/xfs_bmap_item.c 	int			*nvecs,
nvecs              68 fs/xfs/xfs_bmap_item.c 	*nvecs += 1;
nvecs             163 fs/xfs/xfs_bmap_item.c 	int			*nvecs,
nvecs             166 fs/xfs/xfs_bmap_item.c 	*nvecs += 1;
nvecs              52 fs/xfs/xfs_buf_item.c 	int				*nvecs,
nvecs              67 fs/xfs/xfs_buf_item.c 	*nvecs += 2;
nvecs              88 fs/xfs/xfs_buf_item.c 			(*nvecs)++;
nvecs              93 fs/xfs/xfs_buf_item.c 			(*nvecs)++;
nvecs             121 fs/xfs/xfs_buf_item.c 	int			*nvecs,
nvecs             136 fs/xfs/xfs_buf_item.c 		*nvecs += bip->bli_format_count;
nvecs             152 fs/xfs/xfs_buf_item.c 		*nvecs = XFS_LOG_VEC_ORDERED;
nvecs             167 fs/xfs/xfs_buf_item.c 					  nvecs, nbytes);
nvecs              32 fs/xfs/xfs_dquot_item.c 	int			*nvecs,
nvecs              35 fs/xfs/xfs_dquot_item.c 	*nvecs += 2;
nvecs             268 fs/xfs/xfs_dquot_item.c 	int			*nvecs,
nvecs             271 fs/xfs/xfs_dquot_item.c 	*nvecs += 1;
nvecs              79 fs/xfs/xfs_extfree_item.c 	int			*nvecs,
nvecs              82 fs/xfs/xfs_extfree_item.c 	*nvecs += 1;
nvecs             265 fs/xfs/xfs_extfree_item.c 	int			*nvecs,
nvecs             268 fs/xfs/xfs_extfree_item.c 	*nvecs += 1;
nvecs              30 fs/xfs/xfs_icreate_item.c 	int			*nvecs,
nvecs              33 fs/xfs/xfs_icreate_item.c 	*nvecs += 1;
nvecs              33 fs/xfs/xfs_inode_item.c 	int			*nvecs,
nvecs              45 fs/xfs/xfs_inode_item.c 			*nvecs += 1;
nvecs              52 fs/xfs/xfs_inode_item.c 			*nvecs += 1;
nvecs              59 fs/xfs/xfs_inode_item.c 			*nvecs += 1;
nvecs              74 fs/xfs/xfs_inode_item.c 	int			*nvecs,
nvecs              86 fs/xfs/xfs_inode_item.c 			*nvecs += 1;
nvecs              93 fs/xfs/xfs_inode_item.c 			*nvecs += 1;
nvecs             100 fs/xfs/xfs_inode_item.c 			*nvecs += 1;
nvecs             119 fs/xfs/xfs_inode_item.c 	int			*nvecs,
nvecs             125 fs/xfs/xfs_inode_item.c 	*nvecs += 2;
nvecs             129 fs/xfs/xfs_inode_item.c 	xfs_inode_item_data_fork_size(iip, nvecs, nbytes);
nvecs             131 fs/xfs/xfs_inode_item.c 		xfs_inode_item_attr_fork_size(iip, nvecs, nbytes);
nvecs             412 fs/xfs/xfs_log_cil.c 	ctx->nvecs += diff_iovecs;
nvecs             244 fs/xfs/xfs_log_priv.h 	int			nvecs;		/* number of regions */
nvecs              62 fs/xfs/xfs_refcount_item.c 	int			*nvecs,
nvecs              67 fs/xfs/xfs_refcount_item.c 	*nvecs += 1;
nvecs             168 fs/xfs/xfs_refcount_item.c 	int			*nvecs,
nvecs             171 fs/xfs/xfs_refcount_item.c 	*nvecs += 1;
nvecs              61 fs/xfs/xfs_rmap_item.c 	int			*nvecs,
nvecs              66 fs/xfs/xfs_rmap_item.c 	*nvecs += 1;
nvecs             189 fs/xfs/xfs_rmap_item.c 	int			*nvecs,
nvecs             192 fs/xfs/xfs_rmap_item.c 	*nvecs += 1;
nvecs             284 include/linux/interrupt.h 	void		(*calc_sets)(struct irq_affinity *, unsigned int nvecs);
nvecs             416 kernel/irq/affinity.c irq_create_affinity_masks(unsigned int nvecs, struct irq_affinity *affd)
nvecs             427 kernel/irq/affinity.c 	if (nvecs > affd->pre_vectors + affd->post_vectors)
nvecs             428 kernel/irq/affinity.c 		affvecs = nvecs - affd->pre_vectors - affd->post_vectors;
nvecs             449 kernel/irq/affinity.c 	masks = kcalloc(nvecs, sizeof(*masks), GFP_KERNEL);
nvecs             480 kernel/irq/affinity.c 	for (; curvec < nvecs; curvec++)
nvecs             484 kernel/irq/affinity.c 	for (i = affd->pre_vectors; i < nvecs - affd->post_vectors; i++)