nelems 253 arch/powerpc/include/asm/iommu.h struct scatterlist *sglist, int nelems, nelems 259 arch/powerpc/include/asm/iommu.h int nelems, nelems 91 arch/powerpc/kernel/dma-iommu.c int nelems, enum dma_data_direction direction, nelems 95 arch/powerpc/kernel/dma-iommu.c return dma_direct_map_sg(dev, sglist, nelems, direction, attrs); nelems 96 arch/powerpc/kernel/dma-iommu.c return ppc_iommu_map_sg(dev, get_iommu_table_base(dev), sglist, nelems, nelems 101 arch/powerpc/kernel/dma-iommu.c int nelems, enum dma_data_direction direction, nelems 105 arch/powerpc/kernel/dma-iommu.c ppc_iommu_unmap_sg(get_iommu_table_base(dev), sglist, nelems, nelems 108 arch/powerpc/kernel/dma-iommu.c dma_direct_unmap_sg(dev, sglist, nelems, direction, attrs); nelems 419 arch/powerpc/kernel/iommu.c struct scatterlist *sglist, int nelems, nelems 432 arch/powerpc/kernel/iommu.c if ((nelems == 0) || !tbl) nelems 437 arch/powerpc/kernel/iommu.c incount = nelems; nelems 443 arch/powerpc/kernel/iommu.c DBG("sg mapping %d elements:\n", nelems); nelems 446 arch/powerpc/kernel/iommu.c for_each_sg(sglist, s, nelems, i) { nelems 545 arch/powerpc/kernel/iommu.c for_each_sg(sglist, s, nelems, i) { nelems 564 arch/powerpc/kernel/iommu.c int nelems, enum dma_data_direction direction, nelems 575 arch/powerpc/kernel/iommu.c while (nelems--) { nelems 549 arch/powerpc/platforms/pseries/vio.c int nelems, enum dma_data_direction direction, nelems 558 arch/powerpc/platforms/pseries/vio.c for_each_sg(sglist, sgl, nelems, count) nelems 563 arch/powerpc/platforms/pseries/vio.c ret = ppc_iommu_map_sg(dev, tbl, sglist, nelems, dma_get_mask(dev), nelems 582 arch/powerpc/platforms/pseries/vio.c struct scatterlist *sglist, int nelems, nelems 592 arch/powerpc/platforms/pseries/vio.c for_each_sg(sglist, sgl, nelems, count) nelems 595 arch/powerpc/platforms/pseries/vio.c ppc_iommu_unmap_sg(tbl, sglist, nelems, direction, attrs); nelems 433 arch/sparc/kernel/iommu.c int nelems, enum dma_data_direction direction, nelems 450 arch/sparc/kernel/iommu.c if (nelems == 0 || !iommu) nelems 468 arch/sparc/kernel/iommu.c incount = nelems; nelems 478 arch/sparc/kernel/iommu.c for_each_sg(sglist, s, nelems, i) { nelems 557 arch/sparc/kernel/iommu.c for_each_sg(sglist, s, nelems, i) { nelems 609 arch/sparc/kernel/iommu.c int nelems, enum dma_data_direction direction, nelems 627 arch/sparc/kernel/iommu.c while (nelems--) { nelems 699 arch/sparc/kernel/iommu.c struct scatterlist *sglist, int nelems, nelems 731 arch/sparc/kernel/iommu.c for_each_sg(sglist, sg, nelems, i) { nelems 468 arch/sparc/kernel/pci_sun4v.c int nelems, enum dma_data_direction direction, nelems 487 arch/sparc/kernel/pci_sun4v.c if (nelems == 0 || !iommu) nelems 500 arch/sparc/kernel/pci_sun4v.c incount = nelems; nelems 522 arch/sparc/kernel/pci_sun4v.c for_each_sg(sglist, s, nelems, i) { nelems 604 arch/sparc/kernel/pci_sun4v.c for_each_sg(sglist, s, nelems, i) { nelems 626 arch/sparc/kernel/pci_sun4v.c int nelems, enum dma_data_direction direction, nelems 647 arch/sparc/kernel/pci_sun4v.c while (nelems--) { nelems 326 arch/x86/kernel/amd_gart_64.c int nelems, struct scatterlist *sout, nelems 337 arch/x86/kernel/amd_gart_64.c for_each_sg(start, s, nelems, i) { nelems 364 arch/x86/kernel/amd_gart_64.c dma_map_cont(struct device *dev, struct scatterlist *start, int nelems, nelems 368 arch/x86/kernel/amd_gart_64.c BUG_ON(nelems != 1); nelems 373 arch/x86/kernel/amd_gart_64.c return __dma_map_cont(dev, start, nelems, sout, pages); nelems 326 arch/x86/kernel/pci-calgary_64.c int nelems,enum dma_data_direction dir, nelems 336 arch/x86/kernel/pci-calgary_64.c for_each_sg(sglist, s, nelems, i) { nelems 350 arch/x86/kernel/pci-calgary_64.c int nelems, enum dma_data_direction dir, nelems 360 arch/x86/kernel/pci-calgary_64.c for_each_sg(sg, s, nelems, i) { nelems 381 arch/x86/kernel/pci-calgary_64.c return nelems; nelems 383 arch/x86/kernel/pci-calgary_64.c calgary_unmap_sg(dev, sg, nelems, dir, 0); nelems 384 arch/x86/kernel/pci-calgary_64.c for_each_sg(sg, s, nelems, i) { nelems 11201 drivers/infiniband/hw/hfi1/chip.c static void get_vlarb_preempt(struct hfi1_devdata *dd, u32 nelems, nelems 11206 drivers/infiniband/hw/hfi1/chip.c for (i = 0; i < nelems; i++, vl++) { nelems 2547 drivers/iommu/amd_iommu.c int nelems) nelems 2557 drivers/iommu/amd_iommu.c for_each_sg(sglist, s, nelems, i) { nelems 2576 drivers/iommu/amd_iommu.c int nelems, enum dma_data_direction direction, nelems 2594 drivers/iommu/amd_iommu.c npages = sg_num_pages(dev, sglist, nelems); nelems 2603 drivers/iommu/amd_iommu.c for_each_sg(sglist, s, nelems, i) { nelems 2622 drivers/iommu/amd_iommu.c for_each_sg(sglist, s, nelems, i) { nelems 2635 drivers/iommu/amd_iommu.c return nelems; nelems 2641 drivers/iommu/amd_iommu.c for_each_sg(sglist, s, nelems, i) { nelems 2667 drivers/iommu/amd_iommu.c int nelems, enum dma_data_direction dir, nelems 2681 drivers/iommu/amd_iommu.c npages = sg_num_pages(dev, sglist, nelems); nelems 677 drivers/iommu/dma-iommu.c struct scatterlist *sgl, int nelems, nelems 686 drivers/iommu/dma-iommu.c for_each_sg(sgl, sg, nelems, i) nelems 691 drivers/iommu/dma-iommu.c struct scatterlist *sgl, int nelems, nelems 700 drivers/iommu/dma-iommu.c for_each_sg(sgl, sg, nelems, i) nelems 3688 drivers/iommu/intel-iommu.c int nelems, enum dma_data_direction dir, nelems 3697 drivers/iommu/intel-iommu.c return dma_direct_unmap_sg(dev, sglist, nelems, dir, attrs); nelems 3699 drivers/iommu/intel-iommu.c for_each_sg(sglist, sg, nelems, i) { nelems 3708 drivers/iommu/intel-iommu.c static int intel_map_sg(struct device *dev, struct scatterlist *sglist, int nelems, nelems 3723 drivers/iommu/intel-iommu.c return dma_direct_map_sg(dev, sglist, nelems, dir, attrs); nelems 3731 drivers/iommu/intel-iommu.c for_each_sg(sglist, sg, nelems, i) nelems 3765 drivers/iommu/intel-iommu.c return nelems; nelems 3948 drivers/iommu/intel-iommu.c bounce_unmap_sg(struct device *dev, struct scatterlist *sglist, int nelems, nelems 3954 drivers/iommu/intel-iommu.c for_each_sg(sglist, sg, nelems, i) nelems 3960 drivers/iommu/intel-iommu.c bounce_map_sg(struct device *dev, struct scatterlist *sglist, int nelems, nelems 3966 drivers/iommu/intel-iommu.c for_each_sg(sglist, sg, nelems, i) { nelems 3975 drivers/iommu/intel-iommu.c return nelems; nelems 3998 drivers/iommu/intel-iommu.c int nelems, enum dma_data_direction dir) nelems 4003 drivers/iommu/intel-iommu.c for_each_sg(sglist, sg, nelems, i) nelems 4010 drivers/iommu/intel-iommu.c int nelems, enum dma_data_direction dir) nelems 4015 drivers/iommu/intel-iommu.c for_each_sg(sglist, sg, nelems, i) nelems 2039 drivers/mtd/nand/raw/nand_base.c for (i = 0, ninstrs = 0; i < pat->nelems && instr < end; i++) { nelems 2088 drivers/mtd/nand/raw/nand_base.c for (; i < pat->nelems; i++) { nelems 1543 drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c num_flows = atomic_read(&tc_info->flow_table.nelems); nelems 4172 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c return atomic_read(&tc_ht->nelems); nelems 600 drivers/scsi/hpsa.c static int board_id_in_array(u32 a[], int nelems, u32 board_id) nelems 604 drivers/scsi/hpsa.c for (i = 0; i < nelems; i++) nelems 466 drivers/xen/swiotlb-xen.c xen_swiotlb_unmap_sg(struct device *hwdev, struct scatterlist *sgl, int nelems, nelems 474 drivers/xen/swiotlb-xen.c for_each_sg(sgl, sg, nelems, i) nelems 481 drivers/xen/swiotlb-xen.c xen_swiotlb_map_sg(struct device *dev, struct scatterlist *sgl, int nelems, nelems 489 drivers/xen/swiotlb-xen.c for_each_sg(sgl, sg, nelems, i) { nelems 497 drivers/xen/swiotlb-xen.c return nelems; nelems 506 drivers/xen/swiotlb-xen.c int nelems, enum dma_data_direction dir) nelems 511 drivers/xen/swiotlb-xen.c for_each_sg(sgl, sg, nelems, i) { nelems 519 drivers/xen/swiotlb-xen.c int nelems, enum dma_data_direction dir) nelems 524 drivers/xen/swiotlb-xen.c for_each_sg(sgl, sg, nelems, i) { nelems 37 include/linux/dma-debug.h int nelems, int dir); nelems 62 include/linux/dma-debug.h int nelems, int direction); nelems 66 include/linux/dma-debug.h int nelems, int direction); nelems 106 include/linux/dma-debug.h int nelems, int dir) nelems 146 include/linux/dma-debug.h int nelems, int direction) nelems 152 include/linux/dma-debug.h int nelems, int direction) nelems 409 include/linux/dma-mapping.h int nelems, enum dma_data_direction dir) nelems 415 include/linux/dma-mapping.h dma_direct_sync_sg_for_cpu(dev, sg, nelems, dir); nelems 417 include/linux/dma-mapping.h ops->sync_sg_for_cpu(dev, sg, nelems, dir); nelems 418 include/linux/dma-mapping.h debug_dma_sync_sg_for_cpu(dev, sg, nelems, dir); nelems 423 include/linux/dma-mapping.h int nelems, enum dma_data_direction dir) nelems 429 include/linux/dma-mapping.h dma_direct_sync_sg_for_device(dev, sg, nelems, dir); nelems 431 include/linux/dma-mapping.h ops->sync_sg_for_device(dev, sg, nelems, dir); nelems 432 include/linux/dma-mapping.h debug_dma_sync_sg_for_device(dev, sg, nelems, dir); nelems 508 include/linux/dma-mapping.h struct scatterlist *sg, int nelems, enum dma_data_direction dir) nelems 512 include/linux/dma-mapping.h struct scatterlist *sg, int nelems, enum dma_data_direction dir) nelems 814 include/linux/mtd/rawnand.h unsigned int nelems; nelems 822 include/linux/mtd/rawnand.h .nelems = sizeof((struct nand_op_parser_pattern_elem[]) { __VA_ARGS__ }) / \ nelems 94 include/linux/pci-dma-compat.h int nelems, int direction) nelems 96 include/linux/pci-dma-compat.h dma_sync_sg_for_cpu(&hwdev->dev, sg, nelems, (enum dma_data_direction)direction); nelems 101 include/linux/pci-dma-compat.h int nelems, int direction) nelems 103 include/linux/pci-dma-compat.h dma_sync_sg_for_device(&hwdev->dev, sg, nelems, (enum dma_data_direction)direction); nelems 90 include/linux/rhashtable-types.h atomic_t nelems; nelems 186 include/linux/rhashtable.h return atomic_read(&ht->nelems) > (tbl->size / 4 * 3) && nelems 199 include/linux/rhashtable.h return atomic_read(&ht->nelems) < (tbl->size * 3 / 10) && nelems 211 include/linux/rhashtable.h return atomic_read(&ht->nelems) > tbl->size && nelems 223 include/linux/rhashtable.h return atomic_read(&ht->nelems) >= ht->max_elems; nelems 794 include/linux/rhashtable.h atomic_inc(&ht->nelems); nelems 1063 include/linux/rhashtable.h atomic_dec(&ht->nelems); nelems 434 include/net/netfilter/nf_tables.h atomic_t nelems; nelems 107 include/uapi/linux/btf.h __u32 nelems; nelems 1978 kernel/bpf/btf.c if (array->nelems && elem_size > U32_MAX / array->nelems) { nelems 1984 kernel/bpf/btf.c env_stack_pop_resolved(env, elem_type_id, elem_size * array->nelems); nelems 1995 kernel/bpf/btf.c array->type, array->index_type, array->nelems); nelems 2011 kernel/bpf/btf.c for (i = 0; i < array->nelems; i++) { nelems 3023 kernel/bpf/btf.c (array->nelems * elem_size == nelems 1404 kernel/dma/debug.c int nelems, int dir) nelems 1412 kernel/dma/debug.c for_each_sg(sglist, s, nelems, i) { nelems 1422 kernel/dma/debug.c .sg_call_ents = nelems, nelems 1580 kernel/dma/debug.c int nelems, int direction) nelems 1588 kernel/dma/debug.c for_each_sg(sg, s, nelems, i) { nelems 1598 kernel/dma/debug.c .sg_call_ents = nelems, nelems 1613 kernel/dma/debug.c int nelems, int direction) nelems 1621 kernel/dma/debug.c for_each_sg(sg, s, nelems, i) { nelems 1631 kernel/dma/debug.c .sg_call_ents = nelems, nelems 22 kernel/dma/dummy.c int nelems, enum dma_data_direction dir, nelems 381 lib/rhashtable.c unsigned int nelems = atomic_read(&ht->nelems); nelems 384 lib/rhashtable.c if (nelems) nelems 385 lib/rhashtable.c size = roundup_pow_of_two(nelems * 3 / 2); nelems 573 lib/rhashtable.c atomic_inc(&ht->nelems); nelems 1060 lib/rhashtable.c atomic_set(&ht->nelems, 0); nelems 202 lib/test_rhashtable.c total, atomic_read(&ht->nelems), entries, chain_len); nelems 204 lib/test_rhashtable.c if (total != atomic_read(&ht->nelems) || total != entries) nelems 449 net/bridge/br_multicast.c if (atomic_read(&br->mdb_hash_tbl.nelems) >= br->hash_max) { nelems 71 net/ipv4/proc.c atomic_read(&net->ipv4.fqdir->rhashtable.nelems), nelems 47 net/ipv6/proc.c atomic_read(&net->ipv6.fqdir->rhashtable.nelems), nelems 3821 net/netfilter/nf_tables_api.c (nlh->nlmsg_flags & NLM_F_NONREC && atomic_read(&set->nelems) > 0)) { nelems 4697 net/netfilter/nf_tables_api.c !atomic_add_unless(&set->nelems, 1, set->size + set->ndeact)) { nelems 6924 net/netfilter/nf_tables_api.c atomic_dec(&te->set->nelems); nelems 7089 net/netfilter/nf_tables_api.c atomic_dec(&te->set->nelems); nelems 52 net/netfilter/nft_dynset.c if (!atomic_add_unless(&set->nelems, 1, set->size)) nelems 74 net/netfilter/nft_dynset.c atomic_dec(&set->nelems); nelems 334 net/netfilter/nft_set_hash.c atomic_dec(&set->nelems); nelems 418 net/netfilter/nft_set_rbtree.c atomic_dec(&set->nelems); nelems 423 net/netfilter/nft_set_rbtree.c atomic_dec(&set->nelems); nelems 138 tools/bpf/bpftool/btf.c jsonw_uint_field(w, "nr_elems", arr->nelems); nelems 141 tools/bpf/bpftool/btf.c arr->type, arr->index_type, arr->nelems); nelems 65 tools/bpf/bpftool/btf_dumper.c for (i = 0; i < arr->nelems; i++) { nelems 456 tools/bpf/bpftool/btf_dumper.c BTF_PRINT_ARG("[%d]", array->nelems); nelems 107 tools/include/uapi/linux/btf.h __u32 nelems; nelems 233 tools/lib/bpf/btf.c __u32 nelems = 1; nelems 260 tools/lib/bpf/btf.c if (nelems && array->nelems > UINT32_MAX / nelems) nelems 262 tools/lib/bpf/btf.c nelems *= array->nelems; nelems 275 tools/lib/bpf/btf.c if (nelems && size > UINT32_MAX / nelems) nelems 278 tools/lib/bpf/btf.c return nelems * size; nelems 1872 tools/lib/bpf/btf.c h = hash_combine(h, info->nelems); nelems 1894 tools/lib/bpf/btf.c info1->nelems == info2->nelems; nelems 1907 tools/lib/bpf/btf.c return btf_array(t1)->nelems == btf_array(t2)->nelems; nelems 1253 tools/lib/bpf/btf_dump.c btf_dump_printf(d, "[%u]", a->nelems); nelems 1269 tools/lib/bpf/btf_dump.c btf_dump_printf(d, "[%u]", a->nelems); nelems 1068 tools/lib/bpf/libbpf.c *res = arr_info->nelems; nelems 2449 tools/lib/bpf/libbpf.c if (!t || access_idx >= a->nelems) nelems 2751 tools/lib/bpf/libbpf.c if (local_acc->idx >= a->nelems)