Home
last modified time | relevance | path

Searched refs:nent (Results 1 – 37 of 37) sorted by relevance

/linux-4.4.14/arch/x86/kvm/
Dcpuid.c171 if (cpuid->nent > KVM_MAX_CPUID_ENTRIES) in kvm_vcpu_ioctl_set_cpuid()
174 cpuid_entries = vmalloc(sizeof(struct kvm_cpuid_entry) * cpuid->nent); in kvm_vcpu_ioctl_set_cpuid()
179 cpuid->nent * sizeof(struct kvm_cpuid_entry))) in kvm_vcpu_ioctl_set_cpuid()
181 for (i = 0; i < cpuid->nent; i++) { in kvm_vcpu_ioctl_set_cpuid()
193 vcpu->arch.cpuid_nent = cpuid->nent; in kvm_vcpu_ioctl_set_cpuid()
212 if (cpuid->nent > KVM_MAX_CPUID_ENTRIES) in kvm_vcpu_ioctl_set_cpuid2()
216 cpuid->nent * sizeof(struct kvm_cpuid_entry2))) in kvm_vcpu_ioctl_set_cpuid2()
218 vcpu->arch.cpuid_nent = cpuid->nent; in kvm_vcpu_ioctl_set_cpuid2()
233 if (cpuid->nent < vcpu->arch.cpuid_nent) in kvm_vcpu_ioctl_get_cpuid2()
242 cpuid->nent = vcpu->arch.cpuid_nent; in kvm_vcpu_ioctl_get_cpuid2()
[all …]
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/engine/gr/
Dgk20a.c46 const int nent = (fuc->size / sizeof(struct gk20a_fw_av)); in gk20a_gr_av_to_init() local
49 pack = vzalloc((sizeof(*pack) * 2) + (sizeof(*init) * (nent + 1))); in gk20a_gr_av_to_init()
57 for (i = 0; i < nent; i++) { in gk20a_gr_av_to_init()
82 const int nent = (fuc->size / sizeof(struct gk20a_fw_aiv)); in gk20a_gr_aiv_to_init() local
85 pack = vzalloc((sizeof(*pack) * 2) + (sizeof(*init) * (nent + 1))); in gk20a_gr_aiv_to_init()
93 for (i = 0; i < nent; i++) { in gk20a_gr_aiv_to_init()
113 const int nent = (fuc->size / sizeof(struct gk20a_fw_av)); in gk20a_gr_av_to_method() local
118 (sizeof(*init) * (nent + 1))); in gk20a_gr_av_to_method()
124 for (i = 0; i < nent; i++) { in gk20a_gr_av_to_method()
/linux-4.4.14/scripts/
Dconmakehash.c85 int i, nuni, nent; in main() local
275 nent = 0; in main()
278 while ( nent >= unicount[fp0] ) in main()
281 nent = 0; in main()
283 printf("0x%04x", unitable[fp0][nent++]); in main()
/linux-4.4.14/drivers/net/ethernet/mellanox/mlx5/core/
Deq.c112 struct mlx5_eqe *eqe = get_eqe(eq, eq->cons_index & (eq->nent - 1)); in next_eqe_sw()
114 return ((eqe->owner & 1) ^ !!(eq->cons_index & eq->nent)) ? NULL : eqe; in next_eqe_sw()
333 for (i = 0; i < eq->nent; i++) { in init_eq_buf()
340 int nent, u64 mask, const char *name, struct mlx5_uar *uar) in mlx5_create_map_eq() argument
348 eq->nent = roundup_pow_of_two(nent + MLX5_NUM_SPARE_EQE); in mlx5_create_map_eq()
350 err = mlx5_buf_alloc(dev, eq->nent * MLX5_EQE_SIZE, &eq->buf); in mlx5_create_map_eq()
367 in->ctx.log_sz_usr_page = cpu_to_be32(ilog2(eq->nent) << 24 | uar->index); in mlx5_create_map_eq()
Dmain.c617 int nent; in alloc_comp_eqs() local
623 nent = MLX5_COMP_EQ_SIZE; in alloc_comp_eqs()
633 i + MLX5_EQ_VEC_COMP_BASE, nent, 0, in alloc_comp_eqs()
/linux-4.4.14/drivers/infiniband/hw/mthca/
Dmthca_allocator.c160 int mthca_array_init(struct mthca_array *array, int nent) in mthca_array_init() argument
162 int npage = (nent * sizeof (void *) + PAGE_SIZE - 1) / PAGE_SIZE; in mthca_array_init()
177 void mthca_array_cleanup(struct mthca_array *array, int nent) in mthca_array_cleanup() argument
181 for (i = 0; i < (nent * sizeof (void *) + PAGE_SIZE - 1) / PAGE_SIZE; ++i) in mthca_array_cleanup()
Dmthca_eq.c184 mthca_write64(MTHCA_EQ_DB_SET_CI | eq->eqn, ci & (eq->nent - 1), in tavor_set_eq_ci()
230 unsigned long off = (entry & (eq->nent - 1)) * MTHCA_EQ_ENTRY_SIZE; in get_eqe()
466 int nent, in mthca_create_eq() argument
479 eq->nent = roundup_pow_of_two(max(nent, 2)); in mthca_create_eq()
480 npages = ALIGN(eq->nent * MTHCA_EQ_ENTRY_SIZE, PAGE_SIZE) / PAGE_SIZE; in mthca_create_eq()
511 for (i = 0; i < eq->nent; ++i) in mthca_create_eq()
535 eq_context->logsize_usrpage = cpu_to_be32((ffs(eq->nent) - 1) << 24); in mthca_create_eq()
560 eq->eqn, eq->nent); in mthca_create_eq()
593 int npages = (eq->nent * MTHCA_EQ_ENTRY_SIZE + PAGE_SIZE - 1) / in mthca_free_eq()
Dmthca_cq.c353 int mthca_alloc_cq_buf(struct mthca_dev *dev, struct mthca_cq_buf *buf, int nent) in mthca_alloc_cq_buf() argument
358 ret = mthca_buf_alloc(dev, nent * MTHCA_CQ_ENTRY_SIZE, in mthca_alloc_cq_buf()
365 for (i = 0; i < nent; ++i) in mthca_alloc_cq_buf()
776 int mthca_init_cq(struct mthca_dev *dev, int nent, in mthca_init_cq() argument
784 cq->ibcq.cqe = nent - 1; in mthca_init_cq()
820 err = mthca_alloc_cq_buf(dev, &cq->buf, nent); in mthca_init_cq()
834 cq_context->logsize_usrpage = cpu_to_be32((ffs(nent) - 1) << 24); in mthca_init_cq()
Dmthca_dev.h422 int mthca_array_init(struct mthca_array *array, int nent);
423 void mthca_array_cleanup(struct mthca_array *array, int nent);
498 int mthca_init_cq(struct mthca_dev *dev, int nent,
509 int mthca_alloc_cq_buf(struct mthca_dev *dev, struct mthca_cq_buf *buf, int nent);
Dmthca_cmd.c662 int nent = 0; in mthca_map_cmd() local
692 pages[nent * 2] = cpu_to_be64(virt); in mthca_map_cmd()
696 pages[nent * 2 + 1] = in mthca_map_cmd()
702 if (++nent == MTHCA_MAILBOX_SIZE / 16) { in mthca_map_cmd()
703 err = mthca_cmd(dev, mailbox->dma, nent, 0, op, in mthca_map_cmd()
707 nent = 0; in mthca_map_cmd()
712 if (nent) in mthca_map_cmd()
713 err = mthca_cmd(dev, mailbox->dma, nent, 0, op, in mthca_map_cmd()
Dmthca_provider.h113 int nent; member
Dmthca_provider.c656 int nent; in mthca_create_cq() local
694 for (nent = 1; nent <= entries; nent <<= 1) in mthca_create_cq()
697 err = mthca_init_cq(to_mdev(ibdev), nent, in mthca_create_cq()
/linux-4.4.14/drivers/iommu/
Domap-iommu.c688 int nent = 1; in iopgtable_clear_entry_core() local
699 nent *= 16; in iopgtable_clear_entry_core()
703 bytes *= nent; in iopgtable_clear_entry_core()
704 memset(iopte, 0, nent * sizeof(*iopte)); in iopgtable_clear_entry_core()
705 flush_iopte_range(iopte, iopte + (nent - 1) * sizeof(*iopte)); in iopgtable_clear_entry_core()
716 nent = 1; /* for the next L1 entry */ in iopgtable_clear_entry_core()
720 nent *= 16; in iopgtable_clear_entry_core()
724 bytes *= nent; in iopgtable_clear_entry_core()
726 memset(iopgd, 0, nent * sizeof(*iopgd)); in iopgtable_clear_entry_core()
727 flush_iopgd_range(iopgd, iopgd + (nent - 1) * sizeof(*iopgd)); in iopgtable_clear_entry_core()
Darm-smmu-v3.c1098 static void arm_smmu_init_bypass_stes(u64 *strtab, unsigned int nent) in arm_smmu_init_bypass_stes() argument
1106 for (i = 0; i < nent; ++i) { in arm_smmu_init_bypass_stes()
/linux-4.4.14/drivers/net/ethernet/mellanox/mlx4/
Deq.c110 unsigned long offset = (entry & (eq->nent - 1)) * eqe_size; in get_eqe()
124 return !!(eqe->owner & 0x80) ^ !!(eq->cons_index & eq->nent) ? NULL : eqe; in next_eqe_sw()
759 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int()
762 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); in mlx4_eq_int()
784 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int()
786 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); in mlx4_eq_int()
796 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int()
799 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); in mlx4_eq_int()
952 static int mlx4_create_eq(struct mlx4_dev *dev, int nent, in mlx4_create_eq() argument
966 eq->nent = roundup_pow_of_two(max(nent, 2)); in mlx4_create_eq()
[all …]
Dcq.c283 int mlx4_cq_alloc(struct mlx4_dev *dev, int nent, in mlx4_cq_alloc() argument
321 cq_context->logsize_usrpage = cpu_to_be32((ilog2(nent) << 24) | uar->index); in mlx4_cq_alloc()
Dfw.c1370 int nent = 0; in mlx4_map_cmd() local
1400 pages[nent * 2] = cpu_to_be64(virt); in mlx4_map_cmd()
1404 pages[nent * 2 + 1] = in mlx4_map_cmd()
1410 if (++nent == MLX4_MAILBOX_SIZE / 16) { in mlx4_map_cmd()
1411 err = mlx4_cmd(dev, mailbox->dma, nent, 0, op, in mlx4_map_cmd()
1416 nent = 0; in mlx4_map_cmd()
1421 if (nent) in mlx4_map_cmd()
1422 err = mlx4_cmd(dev, mailbox->dma, nent, 0, op, in mlx4_map_cmd()
Dmlx4.h398 int nent; member
/linux-4.4.14/drivers/infiniband/hw/mlx5/
Dcq.c78 static u8 sw_ownership_bit(int n, int nent) in sw_ownership_bit() argument
80 return (n & nent) ? 1 : 0; in sw_ownership_bit()
597 int nent, int cqe_size) in alloc_cq_buf() argument
601 err = mlx5_buf_alloc(dev->mdev, nent * cqe_size, &buf->buf); in alloc_cq_buf()
606 buf->nent = nent; in alloc_cq_buf()
691 for (i = 0; i < buf->nent; i++) { in init_cq_buf()
1055 (i + 1) & (cq->resize_buf->nent), in copy_resize_cqes()
1058 sw_own = sw_ownership_bit(i + 1, cq->resize_buf->nent); in copy_resize_cqes()
Dmlx5_ib.h239 int nent; member
/linux-4.4.14/drivers/tty/serial/
Dpch_uart.c253 int nent; member
801 for (i = 0; i < priv->nent; i++, sg++) { in pch_dma_tx_complete()
807 dma_unmap_sg(port->dev, sg, priv->nent, DMA_TO_DEVICE); in pch_dma_tx_complete()
809 priv->nent = 0; in pch_dma_tx_complete()
950 int nent; in dma_handle_tx() local
1029 nent = dma_map_sg(port->dev, sg, num, DMA_TO_DEVICE); in dma_handle_tx()
1030 if (!nent) { in dma_handle_tx()
1034 priv->nent = nent; in dma_handle_tx()
1036 for (i = 0; i < nent; i++, sg++) { in dma_handle_tx()
1041 if (i == (nent - 1)) in dma_handle_tx()
[all …]
Datmel_serial.c919 int ret, nent; in atmel_prepare_tx_dma() local
938 nent = dma_map_sg(port->dev, in atmel_prepare_tx_dma()
943 if (!nent) { in atmel_prepare_tx_dma()
1099 int ret, nent; in atmel_prepare_rx_dma() local
1120 nent = dma_map_sg(port->dev, in atmel_prepare_rx_dma()
1125 if (!nent) { in atmel_prepare_rx_dma()
/linux-4.4.14/drivers/infiniband/hw/qib/
Dqib_pcie.c275 int qib_pcie_params(struct qib_devdata *dd, u32 minw, u32 *nent, in qib_pcie_params() argument
290 if (nent && *nent && pos) { in qib_pcie_params()
291 qib_msix_setup(dd, pos, nent, entry); in qib_pcie_params()
/linux-4.4.14/arch/x86/include/uapi/asm/
Dkvm.h203 __u32 nent; member
225 __u32 nent; member
/linux-4.4.14/drivers/net/ethernet/mellanox/mlxsw/
Dpci.c1285 int nent = 0; in mlxsw_pci_fw_area_init() local
1307 mlxsw_cmd_mbox_map_fa_pa_set(mbox, nent, mem_item->mapaddr); in mlxsw_pci_fw_area_init()
1308 mlxsw_cmd_mbox_map_fa_log2size_set(mbox, nent, 0); /* 1 page */ in mlxsw_pci_fw_area_init()
1309 if (++nent == MLXSW_CMD_MAP_FA_VPM_ENTRIES_MAX) { in mlxsw_pci_fw_area_init()
1310 err = mlxsw_cmd_map_fa(mlxsw_pci->core, mbox, nent); in mlxsw_pci_fw_area_init()
1313 nent = 0; in mlxsw_pci_fw_area_init()
1318 if (nent) { in mlxsw_pci_fw_area_init()
1319 err = mlxsw_cmd_map_fa(mlxsw_pci->core, mbox, nent); in mlxsw_pci_fw_area_init()
/linux-4.4.14/drivers/staging/rdma/hfi1/
Dpcie.c412 void request_msix(struct hfi1_devdata *dd, u32 *nent, in request_msix() argument
418 if (*nent && pos) { in request_msix()
419 msix_setup(dd, pos, nent, entry); in request_msix()
422 *nent = 0; in request_msix()
/linux-4.4.14/arch/alpha/kernel/
Dpci_iommu.c125 long i, p, nent; in iommu_arena_find_pages() local
140 nent = arena->size >> PAGE_SHIFT; in iommu_arena_find_pages()
145 while (i < n && p+i < nent) { in iommu_arena_find_pages()
/linux-4.4.14/drivers/spi/
Dspi-topcliff-pch.c125 int nent; member
803 dma_sync_sg_for_cpu(&data->master->dev, dma->sg_rx_p, dma->nent, in pch_spi_start_transfer()
806 dma_sync_sg_for_cpu(&data->master->dev, dma->sg_tx_p, dma->nent, in pch_spi_start_transfer()
1053 dma->nent = num; in pch_spi_handle_dma()
1112 dma->nent = num; in pch_spi_handle_dma()
/linux-4.4.14/drivers/infiniband/hw/mlx4/
Dcq.c100 static int mlx4_ib_alloc_cq_buf(struct mlx4_ib_dev *dev, struct mlx4_ib_cq_buf *buf, int nent) in mlx4_ib_alloc_cq_buf() argument
104 err = mlx4_buf_alloc(dev->dev, nent * dev->dev->caps.cqe_size, in mlx4_ib_alloc_cq_buf()
126 mlx4_buf_free(dev->dev, nent * buf->entry_size, &buf->buf); in mlx4_ib_alloc_cq_buf()
/linux-4.4.14/include/xen/interface/
Dplatform.h253 uint32_t nent; member
/linux-4.4.14/include/linux/mlx5/
Ddriver.h308 int nent; member
761 int nent, u64 mask, const char *name, struct mlx5_uar *uar);
/linux-4.4.14/drivers/target/
Dtarget_core_transport.c2290 unsigned int nent; in target_alloc_sgl() local
2293 nent = DIV_ROUND_UP(length, PAGE_SIZE); in target_alloc_sgl()
2294 sg = kmalloc(sizeof(struct scatterlist) * nent, GFP_KERNEL); in target_alloc_sgl()
2298 sg_init_table(sg, nent); in target_alloc_sgl()
2311 *nents = nent; in target_alloc_sgl()
/linux-4.4.14/drivers/xen/
Defi.c342 efi_systab_xen.nr_tables = info->cfg.nent; in xen_efi_probe()
/linux-4.4.14/arch/powerpc/xmon/
Dxmon.c2932 int nent, assoc, new_cc = 1; in dump_tlb_book3e() local
2951 nent = tlbcfg & 0xfff; in dump_tlb_book3e()
2953 for (i = 0; i < nent; i++) { in dump_tlb_book3e()
/linux-4.4.14/drivers/scsi/pm8001/
Dpm80xx_hwi.h860 __le32 nent; member
/linux-4.4.14/Documentation/virtual/kvm/
Dapi.txt537 __u32 nent;
1210 __u32 nent;
1238 with the 'nent' field indicating the number of entries in the variable-size
1241 the 'nent' field is adjusted and an error (ENOMEM) is returned. If the
1242 number is just right, the 'nent' field is adjusted to the number of valid
2740 __u32 nent;
2767 structure with the 'nent' field indicating the number of entries in
2770 number is too high, the 'nent' field is adjusted and an error (ENOMEM)
2771 is returned. If the number is just right, the 'nent' field is adjusted
/linux-4.4.14/include/linux/mlx4/
Ddevice.h1094 int mlx4_cq_alloc(struct mlx4_dev *dev, int nent, struct mlx4_mtt *mtt,