| /linux-4.4.14/arch/x86/kvm/ |
| D | cpuid.c | 171 if (cpuid->nent > KVM_MAX_CPUID_ENTRIES) in kvm_vcpu_ioctl_set_cpuid() 174 cpuid_entries = vmalloc(sizeof(struct kvm_cpuid_entry) * cpuid->nent); in kvm_vcpu_ioctl_set_cpuid() 179 cpuid->nent * sizeof(struct kvm_cpuid_entry))) in kvm_vcpu_ioctl_set_cpuid() 181 for (i = 0; i < cpuid->nent; i++) { in kvm_vcpu_ioctl_set_cpuid() 193 vcpu->arch.cpuid_nent = cpuid->nent; in kvm_vcpu_ioctl_set_cpuid() 212 if (cpuid->nent > KVM_MAX_CPUID_ENTRIES) in kvm_vcpu_ioctl_set_cpuid2() 216 cpuid->nent * sizeof(struct kvm_cpuid_entry2))) in kvm_vcpu_ioctl_set_cpuid2() 218 vcpu->arch.cpuid_nent = cpuid->nent; in kvm_vcpu_ioctl_set_cpuid2() 233 if (cpuid->nent < vcpu->arch.cpuid_nent) in kvm_vcpu_ioctl_get_cpuid2() 242 cpuid->nent = vcpu->arch.cpuid_nent; in kvm_vcpu_ioctl_get_cpuid2() [all …]
|
| /linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/engine/gr/ |
| D | gk20a.c | 46 const int nent = (fuc->size / sizeof(struct gk20a_fw_av)); in gk20a_gr_av_to_init() local 49 pack = vzalloc((sizeof(*pack) * 2) + (sizeof(*init) * (nent + 1))); in gk20a_gr_av_to_init() 57 for (i = 0; i < nent; i++) { in gk20a_gr_av_to_init() 82 const int nent = (fuc->size / sizeof(struct gk20a_fw_aiv)); in gk20a_gr_aiv_to_init() local 85 pack = vzalloc((sizeof(*pack) * 2) + (sizeof(*init) * (nent + 1))); in gk20a_gr_aiv_to_init() 93 for (i = 0; i < nent; i++) { in gk20a_gr_aiv_to_init() 113 const int nent = (fuc->size / sizeof(struct gk20a_fw_av)); in gk20a_gr_av_to_method() local 118 (sizeof(*init) * (nent + 1))); in gk20a_gr_av_to_method() 124 for (i = 0; i < nent; i++) { in gk20a_gr_av_to_method()
|
| /linux-4.4.14/scripts/ |
| D | conmakehash.c | 85 int i, nuni, nent; in main() local 275 nent = 0; in main() 278 while ( nent >= unicount[fp0] ) in main() 281 nent = 0; in main() 283 printf("0x%04x", unitable[fp0][nent++]); in main()
|
| /linux-4.4.14/drivers/net/ethernet/mellanox/mlx5/core/ |
| D | eq.c | 112 struct mlx5_eqe *eqe = get_eqe(eq, eq->cons_index & (eq->nent - 1)); in next_eqe_sw() 114 return ((eqe->owner & 1) ^ !!(eq->cons_index & eq->nent)) ? NULL : eqe; in next_eqe_sw() 333 for (i = 0; i < eq->nent; i++) { in init_eq_buf() 340 int nent, u64 mask, const char *name, struct mlx5_uar *uar) in mlx5_create_map_eq() argument 348 eq->nent = roundup_pow_of_two(nent + MLX5_NUM_SPARE_EQE); in mlx5_create_map_eq() 350 err = mlx5_buf_alloc(dev, eq->nent * MLX5_EQE_SIZE, &eq->buf); in mlx5_create_map_eq() 367 in->ctx.log_sz_usr_page = cpu_to_be32(ilog2(eq->nent) << 24 | uar->index); in mlx5_create_map_eq()
|
| D | main.c | 617 int nent; in alloc_comp_eqs() local 623 nent = MLX5_COMP_EQ_SIZE; in alloc_comp_eqs() 633 i + MLX5_EQ_VEC_COMP_BASE, nent, 0, in alloc_comp_eqs()
|
| /linux-4.4.14/drivers/infiniband/hw/mthca/ |
| D | mthca_allocator.c | 160 int mthca_array_init(struct mthca_array *array, int nent) in mthca_array_init() argument 162 int npage = (nent * sizeof (void *) + PAGE_SIZE - 1) / PAGE_SIZE; in mthca_array_init() 177 void mthca_array_cleanup(struct mthca_array *array, int nent) in mthca_array_cleanup() argument 181 for (i = 0; i < (nent * sizeof (void *) + PAGE_SIZE - 1) / PAGE_SIZE; ++i) in mthca_array_cleanup()
|
| D | mthca_eq.c | 184 mthca_write64(MTHCA_EQ_DB_SET_CI | eq->eqn, ci & (eq->nent - 1), in tavor_set_eq_ci() 230 unsigned long off = (entry & (eq->nent - 1)) * MTHCA_EQ_ENTRY_SIZE; in get_eqe() 466 int nent, in mthca_create_eq() argument 479 eq->nent = roundup_pow_of_two(max(nent, 2)); in mthca_create_eq() 480 npages = ALIGN(eq->nent * MTHCA_EQ_ENTRY_SIZE, PAGE_SIZE) / PAGE_SIZE; in mthca_create_eq() 511 for (i = 0; i < eq->nent; ++i) in mthca_create_eq() 535 eq_context->logsize_usrpage = cpu_to_be32((ffs(eq->nent) - 1) << 24); in mthca_create_eq() 560 eq->eqn, eq->nent); in mthca_create_eq() 593 int npages = (eq->nent * MTHCA_EQ_ENTRY_SIZE + PAGE_SIZE - 1) / in mthca_free_eq()
|
| D | mthca_cq.c | 353 int mthca_alloc_cq_buf(struct mthca_dev *dev, struct mthca_cq_buf *buf, int nent) in mthca_alloc_cq_buf() argument 358 ret = mthca_buf_alloc(dev, nent * MTHCA_CQ_ENTRY_SIZE, in mthca_alloc_cq_buf() 365 for (i = 0; i < nent; ++i) in mthca_alloc_cq_buf() 776 int mthca_init_cq(struct mthca_dev *dev, int nent, in mthca_init_cq() argument 784 cq->ibcq.cqe = nent - 1; in mthca_init_cq() 820 err = mthca_alloc_cq_buf(dev, &cq->buf, nent); in mthca_init_cq() 834 cq_context->logsize_usrpage = cpu_to_be32((ffs(nent) - 1) << 24); in mthca_init_cq()
|
| D | mthca_dev.h | 422 int mthca_array_init(struct mthca_array *array, int nent); 423 void mthca_array_cleanup(struct mthca_array *array, int nent); 498 int mthca_init_cq(struct mthca_dev *dev, int nent, 509 int mthca_alloc_cq_buf(struct mthca_dev *dev, struct mthca_cq_buf *buf, int nent);
|
| D | mthca_cmd.c | 662 int nent = 0; in mthca_map_cmd() local 692 pages[nent * 2] = cpu_to_be64(virt); in mthca_map_cmd() 696 pages[nent * 2 + 1] = in mthca_map_cmd() 702 if (++nent == MTHCA_MAILBOX_SIZE / 16) { in mthca_map_cmd() 703 err = mthca_cmd(dev, mailbox->dma, nent, 0, op, in mthca_map_cmd() 707 nent = 0; in mthca_map_cmd() 712 if (nent) in mthca_map_cmd() 713 err = mthca_cmd(dev, mailbox->dma, nent, 0, op, in mthca_map_cmd()
|
| D | mthca_provider.h | 113 int nent; member
|
| D | mthca_provider.c | 656 int nent; in mthca_create_cq() local 694 for (nent = 1; nent <= entries; nent <<= 1) in mthca_create_cq() 697 err = mthca_init_cq(to_mdev(ibdev), nent, in mthca_create_cq()
|
| /linux-4.4.14/drivers/iommu/ |
| D | omap-iommu.c | 688 int nent = 1; in iopgtable_clear_entry_core() local 699 nent *= 16; in iopgtable_clear_entry_core() 703 bytes *= nent; in iopgtable_clear_entry_core() 704 memset(iopte, 0, nent * sizeof(*iopte)); in iopgtable_clear_entry_core() 705 flush_iopte_range(iopte, iopte + (nent - 1) * sizeof(*iopte)); in iopgtable_clear_entry_core() 716 nent = 1; /* for the next L1 entry */ in iopgtable_clear_entry_core() 720 nent *= 16; in iopgtable_clear_entry_core() 724 bytes *= nent; in iopgtable_clear_entry_core() 726 memset(iopgd, 0, nent * sizeof(*iopgd)); in iopgtable_clear_entry_core() 727 flush_iopgd_range(iopgd, iopgd + (nent - 1) * sizeof(*iopgd)); in iopgtable_clear_entry_core()
|
| D | arm-smmu-v3.c | 1098 static void arm_smmu_init_bypass_stes(u64 *strtab, unsigned int nent) in arm_smmu_init_bypass_stes() argument 1106 for (i = 0; i < nent; ++i) { in arm_smmu_init_bypass_stes()
|
| /linux-4.4.14/drivers/net/ethernet/mellanox/mlx4/ |
| D | eq.c | 110 unsigned long offset = (entry & (eq->nent - 1)) * eqe_size; in get_eqe() 124 return !!(eqe->owner & 0x80) ^ !!(eq->cons_index & eq->nent) ? NULL : eqe; in next_eqe_sw() 759 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int() 762 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); in mlx4_eq_int() 784 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int() 786 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); in mlx4_eq_int() 796 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int() 799 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); in mlx4_eq_int() 952 static int mlx4_create_eq(struct mlx4_dev *dev, int nent, in mlx4_create_eq() argument 966 eq->nent = roundup_pow_of_two(max(nent, 2)); in mlx4_create_eq() [all …]
|
| D | cq.c | 283 int mlx4_cq_alloc(struct mlx4_dev *dev, int nent, in mlx4_cq_alloc() argument 321 cq_context->logsize_usrpage = cpu_to_be32((ilog2(nent) << 24) | uar->index); in mlx4_cq_alloc()
|
| D | fw.c | 1370 int nent = 0; in mlx4_map_cmd() local 1400 pages[nent * 2] = cpu_to_be64(virt); in mlx4_map_cmd() 1404 pages[nent * 2 + 1] = in mlx4_map_cmd() 1410 if (++nent == MLX4_MAILBOX_SIZE / 16) { in mlx4_map_cmd() 1411 err = mlx4_cmd(dev, mailbox->dma, nent, 0, op, in mlx4_map_cmd() 1416 nent = 0; in mlx4_map_cmd() 1421 if (nent) in mlx4_map_cmd() 1422 err = mlx4_cmd(dev, mailbox->dma, nent, 0, op, in mlx4_map_cmd()
|
| D | mlx4.h | 398 int nent; member
|
| /linux-4.4.14/drivers/infiniband/hw/mlx5/ |
| D | cq.c | 78 static u8 sw_ownership_bit(int n, int nent) in sw_ownership_bit() argument 80 return (n & nent) ? 1 : 0; in sw_ownership_bit() 597 int nent, int cqe_size) in alloc_cq_buf() argument 601 err = mlx5_buf_alloc(dev->mdev, nent * cqe_size, &buf->buf); in alloc_cq_buf() 606 buf->nent = nent; in alloc_cq_buf() 691 for (i = 0; i < buf->nent; i++) { in init_cq_buf() 1055 (i + 1) & (cq->resize_buf->nent), in copy_resize_cqes() 1058 sw_own = sw_ownership_bit(i + 1, cq->resize_buf->nent); in copy_resize_cqes()
|
| D | mlx5_ib.h | 239 int nent; member
|
| /linux-4.4.14/drivers/tty/serial/ |
| D | pch_uart.c | 253 int nent; member 801 for (i = 0; i < priv->nent; i++, sg++) { in pch_dma_tx_complete() 807 dma_unmap_sg(port->dev, sg, priv->nent, DMA_TO_DEVICE); in pch_dma_tx_complete() 809 priv->nent = 0; in pch_dma_tx_complete() 950 int nent; in dma_handle_tx() local 1029 nent = dma_map_sg(port->dev, sg, num, DMA_TO_DEVICE); in dma_handle_tx() 1030 if (!nent) { in dma_handle_tx() 1034 priv->nent = nent; in dma_handle_tx() 1036 for (i = 0; i < nent; i++, sg++) { in dma_handle_tx() 1041 if (i == (nent - 1)) in dma_handle_tx() [all …]
|
| D | atmel_serial.c | 919 int ret, nent; in atmel_prepare_tx_dma() local 938 nent = dma_map_sg(port->dev, in atmel_prepare_tx_dma() 943 if (!nent) { in atmel_prepare_tx_dma() 1099 int ret, nent; in atmel_prepare_rx_dma() local 1120 nent = dma_map_sg(port->dev, in atmel_prepare_rx_dma() 1125 if (!nent) { in atmel_prepare_rx_dma()
|
| /linux-4.4.14/drivers/infiniband/hw/qib/ |
| D | qib_pcie.c | 275 int qib_pcie_params(struct qib_devdata *dd, u32 minw, u32 *nent, in qib_pcie_params() argument 290 if (nent && *nent && pos) { in qib_pcie_params() 291 qib_msix_setup(dd, pos, nent, entry); in qib_pcie_params()
|
| /linux-4.4.14/arch/x86/include/uapi/asm/ |
| D | kvm.h | 203 __u32 nent; member 225 __u32 nent; member
|
| /linux-4.4.14/drivers/net/ethernet/mellanox/mlxsw/ |
| D | pci.c | 1285 int nent = 0; in mlxsw_pci_fw_area_init() local 1307 mlxsw_cmd_mbox_map_fa_pa_set(mbox, nent, mem_item->mapaddr); in mlxsw_pci_fw_area_init() 1308 mlxsw_cmd_mbox_map_fa_log2size_set(mbox, nent, 0); /* 1 page */ in mlxsw_pci_fw_area_init() 1309 if (++nent == MLXSW_CMD_MAP_FA_VPM_ENTRIES_MAX) { in mlxsw_pci_fw_area_init() 1310 err = mlxsw_cmd_map_fa(mlxsw_pci->core, mbox, nent); in mlxsw_pci_fw_area_init() 1313 nent = 0; in mlxsw_pci_fw_area_init() 1318 if (nent) { in mlxsw_pci_fw_area_init() 1319 err = mlxsw_cmd_map_fa(mlxsw_pci->core, mbox, nent); in mlxsw_pci_fw_area_init()
|
| /linux-4.4.14/drivers/staging/rdma/hfi1/ |
| D | pcie.c | 412 void request_msix(struct hfi1_devdata *dd, u32 *nent, in request_msix() argument 418 if (*nent && pos) { in request_msix() 419 msix_setup(dd, pos, nent, entry); in request_msix() 422 *nent = 0; in request_msix()
|
| /linux-4.4.14/arch/alpha/kernel/ |
| D | pci_iommu.c | 125 long i, p, nent; in iommu_arena_find_pages() local 140 nent = arena->size >> PAGE_SHIFT; in iommu_arena_find_pages() 145 while (i < n && p+i < nent) { in iommu_arena_find_pages()
|
| /linux-4.4.14/drivers/spi/ |
| D | spi-topcliff-pch.c | 125 int nent; member 803 dma_sync_sg_for_cpu(&data->master->dev, dma->sg_rx_p, dma->nent, in pch_spi_start_transfer() 806 dma_sync_sg_for_cpu(&data->master->dev, dma->sg_tx_p, dma->nent, in pch_spi_start_transfer() 1053 dma->nent = num; in pch_spi_handle_dma() 1112 dma->nent = num; in pch_spi_handle_dma()
|
| /linux-4.4.14/drivers/infiniband/hw/mlx4/ |
| D | cq.c | 100 static int mlx4_ib_alloc_cq_buf(struct mlx4_ib_dev *dev, struct mlx4_ib_cq_buf *buf, int nent) in mlx4_ib_alloc_cq_buf() argument 104 err = mlx4_buf_alloc(dev->dev, nent * dev->dev->caps.cqe_size, in mlx4_ib_alloc_cq_buf() 126 mlx4_buf_free(dev->dev, nent * buf->entry_size, &buf->buf); in mlx4_ib_alloc_cq_buf()
|
| /linux-4.4.14/include/xen/interface/ |
| D | platform.h | 253 uint32_t nent; member
|
| /linux-4.4.14/include/linux/mlx5/ |
| D | driver.h | 308 int nent; member 761 int nent, u64 mask, const char *name, struct mlx5_uar *uar);
|
| /linux-4.4.14/drivers/target/ |
| D | target_core_transport.c | 2290 unsigned int nent; in target_alloc_sgl() local 2293 nent = DIV_ROUND_UP(length, PAGE_SIZE); in target_alloc_sgl() 2294 sg = kmalloc(sizeof(struct scatterlist) * nent, GFP_KERNEL); in target_alloc_sgl() 2298 sg_init_table(sg, nent); in target_alloc_sgl() 2311 *nents = nent; in target_alloc_sgl()
|
| /linux-4.4.14/drivers/xen/ |
| D | efi.c | 342 efi_systab_xen.nr_tables = info->cfg.nent; in xen_efi_probe()
|
| /linux-4.4.14/arch/powerpc/xmon/ |
| D | xmon.c | 2932 int nent, assoc, new_cc = 1; in dump_tlb_book3e() local 2951 nent = tlbcfg & 0xfff; in dump_tlb_book3e() 2953 for (i = 0; i < nent; i++) { in dump_tlb_book3e()
|
| /linux-4.4.14/drivers/scsi/pm8001/ |
| D | pm80xx_hwi.h | 860 __le32 nent; member
|
| /linux-4.4.14/Documentation/virtual/kvm/ |
| D | api.txt | 537 __u32 nent; 1210 __u32 nent; 1238 with the 'nent' field indicating the number of entries in the variable-size 1241 the 'nent' field is adjusted and an error (ENOMEM) is returned. If the 1242 number is just right, the 'nent' field is adjusted to the number of valid 2740 __u32 nent; 2767 structure with the 'nent' field indicating the number of entries in 2770 number is too high, the 'nent' field is adjusted and an error (ENOMEM) 2771 is returned. If the number is just right, the 'nent' field is adjusted
|
| /linux-4.4.14/include/linux/mlx4/ |
| D | device.h | 1094 int mlx4_cq_alloc(struct mlx4_dev *dev, int nent, struct mlx4_mtt *mtt,
|