num_counters       59 arch/alpha/oprofile/common.c 	for (i = e = 0; i < model->num_counters; ++i)
num_counters      113 arch/alpha/oprofile/common.c 	for (i = 0; i < model->num_counters; ++i) {
num_counters       51 arch/alpha/oprofile/op_impl.h 	unsigned char num_counters;
num_counters      112 arch/alpha/oprofile/op_model_ev4.c 	.num_counters		= 2,
num_counters      197 arch/alpha/oprofile/op_model_ev5.c 	.num_counters		= 3,
num_counters      207 arch/alpha/oprofile/op_model_ev5.c 	.num_counters		= 3,
num_counters       99 arch/alpha/oprofile/op_model_ev6.c 	.num_counters		= 2,
num_counters      259 arch/alpha/oprofile/op_model_ev67.c 	.num_counters		= 20,
num_counters      435 arch/ia64/kernel/perfmon.c 	unsigned int  num_counters;	/* PMC/PMD counting pairs : computed at init time */
num_counters     6512 arch/ia64/kernel/perfmon.c 	pmu_conf->num_counters  = n_counters;
num_counters     6534 arch/ia64/kernel/perfmon.c 	       pmu_conf->num_counters,
num_counters       90 arch/mips/kernel/perf_event_mipsxx.c 	unsigned int	num_counters;
num_counters      296 arch/mips/kernel/perf_event_mipsxx.c 	for (i = mipspmu.num_counters - 1; i >= 0; i--) {
num_counters      321 arch/mips/kernel/perf_event_mipsxx.c 	WARN_ON(idx < 0 || idx >= mipspmu.num_counters);
num_counters      361 arch/mips/kernel/perf_event_mipsxx.c 	WARN_ON(idx < 0 || idx >= mipspmu.num_counters);
num_counters      498 arch/mips/kernel/perf_event_mipsxx.c 	WARN_ON(idx < 0 || idx >= mipspmu.num_counters);
num_counters      603 arch/mips/kernel/perf_event_mipsxx.c 			(void *)(long)mipspmu.num_counters, 1);
num_counters     1357 arch/mips/kernel/perf_event_mipsxx.c 	int ctr = mipspmu.num_counters;
num_counters     1373 arch/mips/kernel/perf_event_mipsxx.c 	int ctr = mipspmu.num_counters;
num_counters     1385 arch/mips/kernel/perf_event_mipsxx.c 	unsigned int counters = mipspmu.num_counters;
num_counters     1802 arch/mips/kernel/perf_event_mipsxx.c 	mipspmu.num_counters = counters;
num_counters       42 arch/mips/oprofile/common.c 	for (i = 0; i < model->num_counters; ++i) {
num_counters       36 arch/mips/oprofile/op_impl.h 	unsigned char num_counters;
num_counters      160 arch/mips/oprofile/op_model_loongson2.c 	.num_counters = 2
num_counters      212 arch/mips/oprofile/op_model_loongson3.c 	.num_counters	= 2
num_counters      137 arch/mips/oprofile/op_model_mipsxx.c 	unsigned int counters = op_model_mipsxx_ops.num_counters;
num_counters      166 arch/mips/oprofile/op_model_mipsxx.c 	unsigned int counters = op_model_mipsxx_ops.num_counters;
num_counters      193 arch/mips/oprofile/op_model_mipsxx.c 	unsigned int counters = op_model_mipsxx_ops.num_counters;
num_counters      216 arch/mips/oprofile/op_model_mipsxx.c 	unsigned int counters = op_model_mipsxx_ops.num_counters;
num_counters      238 arch/mips/oprofile/op_model_mipsxx.c 	unsigned int counters = op_model_mipsxx_ops.num_counters;
num_counters      346 arch/mips/oprofile/op_model_mipsxx.c 	op_model_mipsxx_ops.num_counters = counters;
num_counters      459 arch/mips/oprofile/op_model_mipsxx.c 	int counters = op_model_mipsxx_ops.num_counters;
num_counters       46 arch/powerpc/include/asm/oprofile_impl.h 			   int num_counters);
num_counters       56 arch/powerpc/include/asm/oprofile_impl.h 	int num_counters;
num_counters       56 arch/powerpc/oprofile/common.c 	op_per_cpu_rc = model->reg_setup(ctr, &sys, model->num_counters);
num_counters      154 arch/powerpc/oprofile/common.c 	for (i = 0; i < model->num_counters; ++i) {
num_counters      225 arch/powerpc/oprofile/common.c 	model->num_counters = cur_cpu_spec->num_pmcs;
num_counters      179 arch/powerpc/oprofile/op_model_cell.c static int num_counters;
num_counters      479 arch/powerpc/oprofile/op_model_cell.c 	for (i = 0; i < num_counters; i++)
num_counters      498 arch/powerpc/oprofile/op_model_cell.c 		for (i = 0; i < num_counters; i++) {
num_counters      527 arch/powerpc/oprofile/op_model_cell.c 		for (i = 0; i < num_counters; i++) {
num_counters      735 arch/powerpc/oprofile/op_model_cell.c 	num_counters = 1;  /* Only support one SPU event at a time */
num_counters      756 arch/powerpc/oprofile/op_model_cell.c 	num_counters = num_ctrs;
num_counters      809 arch/powerpc/oprofile/op_model_cell.c 	for (i = 0; i < num_counters; ++i) {
num_counters      825 arch/powerpc/oprofile/op_model_cell.c 		for (i = 0; i < num_counters; ++i) {
num_counters      921 arch/powerpc/oprofile/op_model_cell.c 	for (i = 0; i < num_counters; ++i) {
num_counters     1409 arch/powerpc/oprofile/op_model_cell.c 		for (i = 0; i < num_counters; ++i) {
num_counters     1636 arch/powerpc/oprofile/op_model_cell.c 		for (i = 0; i < num_counters; ++i) {
num_counters       24 arch/powerpc/oprofile/op_model_fsl_emb.c static int num_counters;
num_counters      262 arch/powerpc/oprofile/op_model_fsl_emb.c 	for (i = 0;i < num_counters;i++) {
num_counters      279 arch/powerpc/oprofile/op_model_fsl_emb.c 	num_counters = num_ctrs;
num_counters      286 arch/powerpc/oprofile/op_model_fsl_emb.c 	for (i = 0; i < num_counters; ++i)
num_counters      298 arch/powerpc/oprofile/op_model_fsl_emb.c 	for (i = 0; i < num_counters; ++i) {
num_counters      351 arch/powerpc/oprofile/op_model_fsl_emb.c 	for (i = 0; i < num_counters; ++i) {
num_counters      293 arch/powerpc/perf/core-fsl-emb.c 	int num_counters = ppmu->n_counter;
num_counters      301 arch/powerpc/perf/core-fsl-emb.c 		num_counters = ppmu->n_restricted;
num_counters      307 arch/powerpc/perf/core-fsl-emb.c 	for (i = num_counters - 1; i >= 0; i--) {
num_counters       74 arch/riscv/include/asm/perf_event.h 	int		num_counters;
num_counters      302 arch/riscv/kernel/perf_event.c 	if (cpuc->n_events == riscv_pmu->num_counters)
num_counters      455 arch/riscv/kernel/perf_event.c 	.num_counters = RISCV_BASE_COUNTERS + 0,
num_counters      389 arch/x86/events/amd/core.c 	for (i = 0; i < x86_pmu.num_counters; i++) {
num_counters      456 arch/x86/events/amd/core.c 	for_each_set_bit(idx, c->idxmsk, x86_pmu.num_counters) {
num_counters      499 arch/x86/events/amd/core.c 	for (i = 0; i < x86_pmu.num_counters; i++) {
num_counters      622 arch/x86/events/amd/core.c 	for (idx = 0; idx < x86_pmu.num_counters; idx++) {
num_counters      915 arch/x86/events/amd/core.c 	.num_counters		= AMD64_NUM_COUNTERS,
num_counters      952 arch/x86/events/amd/core.c 	x86_pmu.num_counters	= AMD64_NUM_COUNTERS_CORE;
num_counters      971 arch/x86/events/amd/core.c 		for (i = 0; i < x86_pmu.num_counters - 1; i += 2)
num_counters      976 arch/x86/events/amd/core.c 				    x86_pmu.num_counters / 2, 0,
num_counters       44 arch/x86/events/amd/uncore.c 	int num_counters;
num_counters      136 arch/x86/events/amd/uncore.c 	for (i = 0; i < uncore->num_counters; i++) {
num_counters      145 arch/x86/events/amd/uncore.c 	for (i = 0; i < uncore->num_counters; i++) {
num_counters      175 arch/x86/events/amd/uncore.c 	for (i = 0; i < uncore->num_counters; i++) {
num_counters      335 arch/x86/events/amd/uncore.c 		uncore_nb->num_counters = num_counters_nb;
num_counters      349 arch/x86/events/amd/uncore.c 		uncore_llc->num_counters = num_counters_llc;
num_counters      151 arch/x86/events/core.c 	for (i = 0; i < x86_pmu.num_counters; i++) {
num_counters      156 arch/x86/events/core.c 	for (i = 0; i < x86_pmu.num_counters; i++) {
num_counters      167 arch/x86/events/core.c 	i = x86_pmu.num_counters;
num_counters      180 arch/x86/events/core.c 	for (i = 0; i < x86_pmu.num_counters; i++) {
num_counters      204 arch/x86/events/core.c 	for (i = 0; i < x86_pmu.num_counters; i++) {
num_counters      619 arch/x86/events/core.c 	for (idx = 0; idx < x86_pmu.num_counters; idx++) {
num_counters      667 arch/x86/events/core.c 	for (idx = 0; idx < x86_pmu.num_counters; idx++) {
num_counters      944 arch/x86/events/core.c 		int gpmax = x86_pmu.num_counters;
num_counters     1009 arch/x86/events/core.c 	max_count = x86_pmu.num_counters + x86_pmu.num_counters_fixed;
num_counters     1360 arch/x86/events/core.c 	if (!x86_pmu.num_counters)
num_counters     1390 arch/x86/events/core.c 	for (idx = 0; idx < x86_pmu.num_counters; idx++) {
num_counters     1511 arch/x86/events/core.c 	for (idx = 0; idx < x86_pmu.num_counters; idx++) {
num_counters     1819 arch/x86/events/core.c 		x86_pmu.intel_ctrl = (1 << x86_pmu.num_counters) - 1;
num_counters     1825 arch/x86/events/core.c 		__EVENT_CONSTRAINT(0, (1ULL << x86_pmu.num_counters) - 1,
num_counters     1826 arch/x86/events/core.c 				   0, x86_pmu.num_counters, 0, 0);
num_counters     1837 arch/x86/events/core.c 	pr_info("... generic registers:      %d\n",     x86_pmu.num_counters);
num_counters     2595 arch/x86/events/core.c 	cap->num_counters_gp	= x86_pmu.num_counters;
num_counters     2298 arch/x86/events/intel/core.c 	if (!x86_pmu.num_counters)
num_counters     2305 arch/x86/events/intel/core.c 	for (idx = 0; idx < x86_pmu.num_counters; idx++) {
num_counters     3373 arch/x86/events/intel/core.c 	for (idx = 0; idx < x86_pmu.num_counters; idx++)  {
num_counters     3391 arch/x86/events/intel/core.c 	*nr = x86_pmu.num_counters;
num_counters     3406 arch/x86/events/intel/core.c 	for (idx = 0; idx < x86_pmu.num_counters; idx++) {
num_counters     4557 arch/x86/events/intel/core.c 	x86_pmu.num_counters		= eax.split.num_counters;
num_counters     4564 arch/x86/events/intel/core.c 	x86_pmu.max_pebs_events		= min_t(unsigned, MAX_PEBS_EVENTS, x86_pmu.num_counters);
num_counters     5096 arch/x86/events/intel/core.c 	if (x86_pmu.num_counters > INTEL_PMC_MAX_GENERIC) {
num_counters     5098 arch/x86/events/intel/core.c 		     x86_pmu.num_counters, INTEL_PMC_MAX_GENERIC);
num_counters     5099 arch/x86/events/intel/core.c 		x86_pmu.num_counters = INTEL_PMC_MAX_GENERIC;
num_counters     5101 arch/x86/events/intel/core.c 	x86_pmu.intel_ctrl = (1ULL << x86_pmu.num_counters) - 1;
num_counters     5120 arch/x86/events/intel/core.c 				c->idxmsk64 |= (1ULL << x86_pmu.num_counters) - 1;
num_counters      306 arch/x86/events/intel/knc.c 	.num_counters		= 2,
num_counters      921 arch/x86/events/intel/p4.c 	for (idx = 0; idx < x86_pmu.num_counters; idx++) {
num_counters      990 arch/x86/events/intel/p4.c 	for (idx = 0; idx < x86_pmu.num_counters; idx++) {
num_counters     1009 arch/x86/events/intel/p4.c 	for (idx = 0; idx < x86_pmu.num_counters; idx++) {
num_counters     1319 arch/x86/events/intel/p4.c 	.num_counters		= ARCH_P4_MAX_CCCR,
num_counters     1370 arch/x86/events/intel/p4.c 	for (i = 0; i < x86_pmu.num_counters; i++) {
num_counters      217 arch/x86/events/intel/p6.c 	.num_counters		= 2,
num_counters      359 arch/x86/events/intel/uncore.c 	max_count = box->pmu->type->num_counters;
num_counters      931 arch/x86/events/intel/uncore.c 		__EVENT_CONSTRAINT(0, (1ULL << type->num_counters) - 1,
num_counters      932 arch/x86/events/intel/uncore.c 				0, type->num_counters, 0, 0);
num_counters       48 arch/x86/events/intel/uncore.h 	int num_counters;
num_counters      155 arch/x86/events/intel/uncore.h 	unsigned int num_counters;
num_counters      402 arch/x86/events/intel/uncore.h 	return box->pmu->type->freerunning[type].num_counters;
num_counters      423 arch/x86/events/intel/uncore.h 	return box->pmu->type->num_counters;
num_counters      283 arch/x86/events/intel/uncore_nhmex.c 	.num_counters	= 1,
num_counters      315 arch/x86/events/intel/uncore_nhmex.c 	.num_counters		= 6,
num_counters      335 arch/x86/events/intel/uncore_nhmex.c 	.num_counters		= 4,
num_counters      426 arch/x86/events/intel/uncore_nhmex.c 	.num_counters		= 4,
num_counters      503 arch/x86/events/intel/uncore_nhmex.c 	.num_counters		= 4,
num_counters      929 arch/x86/events/intel/uncore_nhmex.c 	.num_counters		= 6,
num_counters     1192 arch/x86/events/intel/uncore_nhmex.c 	.num_counters		= 8,
num_counters      192 arch/x86/events/intel/uncore_snb.c 	.num_counters   = 2,
num_counters      210 arch/x86/events/intel/uncore_snb.c 	.num_counters   = 2,
num_counters      270 arch/x86/events/intel/uncore_snb.c 	.num_counters   = 4,
num_counters      302 arch/x86/events/intel/uncore_snb.c 	.num_counters   = 4,
num_counters      329 arch/x86/events/intel/uncore_snb.c 	.num_counters	= 1,
num_counters      581 arch/x86/events/intel/uncore_snb.c 	.num_counters   = 2,
num_counters      980 arch/x86/events/intel/uncore_snb.c 	.num_counters   = 8,
num_counters      799 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 2,
num_counters      956 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     1066 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     1160 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 4,
num_counters     1168 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 4,
num_counters     1180 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     1196 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 4,
num_counters     1205 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 3,
num_counters     1528 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 2,
num_counters     1659 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     1682 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     1710 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 4,
num_counters     1718 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 4,
num_counters     1772 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     1795 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     1809 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 4,
num_counters     1818 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 3,
num_counters     1964 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 2,
num_counters     2080 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     2114 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     2171 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     2187 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     2203 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     2219 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     2240 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 4,
num_counters     2264 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 2,
num_counters     2478 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 2,
num_counters     2643 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     2698 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     2733 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     2775 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 4,
num_counters     2794 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 4,
num_counters     2829 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     2840 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     2876 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 4,
num_counters     2922 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 3,
num_counters     3059 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 2,
num_counters     3083 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     3099 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     3153 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 4,
num_counters     3161 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 4,
num_counters     3173 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     3184 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     3211 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 4,
num_counters     3254 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 3,
num_counters     3398 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 2,
num_counters     3522 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     3580 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     3667 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 17,
num_counters     3692 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 2,
num_counters     3733 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     3786 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 4,
num_counters     3834 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 4,
num_counters     3865 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 4,
num_counters     3883 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 4,
num_counters     3909 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 3,
num_counters     4039 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 2,
num_counters     4103 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     4134 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     4149 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 2,
num_counters     4163 arch/x86/events/intel/uncore_snbep.c 	.num_counters	= 4,
num_counters     4198 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 4,
num_counters     4255 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 9,
num_counters     4314 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 4,
num_counters     4478 arch/x86/events/intel/uncore_snbep.c 	.num_counters   = 4,
num_counters     4527 arch/x86/events/intel/uncore_snbep.c 	.num_counters		= 3,
num_counters      588 arch/x86/events/perf_event.h 	int		num_counters;
num_counters      105 arch/x86/include/asm/perf_event.h 		unsigned int num_counters:8;
num_counters      605 arch/x86/kvm/cpuid.c 		eax.split.num_counters = cap.num_counters_gp;
num_counters      135 arch/x86/kvm/vmx/pmu_intel.c 	unsigned int num_counters;
num_counters      140 arch/x86/kvm/vmx/pmu_intel.c 		num_counters = pmu->nr_arch_fixed_counters;
num_counters      143 arch/x86/kvm/vmx/pmu_intel.c 		num_counters = pmu->nr_arch_gp_counters;
num_counters      145 arch/x86/kvm/vmx/pmu_intel.c 	if (idx >= num_counters)
num_counters      148 arch/x86/kvm/vmx/pmu_intel.c 	return &counters[array_index_nospec(idx, num_counters)];
num_counters      297 arch/x86/kvm/vmx/pmu_intel.c 	pmu->nr_arch_gp_counters = min_t(int, eax.split.num_counters,
num_counters       81 arch/x86/oprofile/nmi_int.c 	for (i = 0; i < model->num_counters; ++i) {
num_counters      145 arch/x86/oprofile/nmi_int.c 	return virt % model->num_counters;
num_counters      206 arch/x86/oprofile/nmi_int.c 	for (i = 0; i < model->num_counters; ++i) {
num_counters      219 arch/x86/oprofile/nmi_int.c 	for (i = 0; i < model->num_counters; ++i) {
num_counters      236 arch/x86/oprofile/nmi_int.c 	si += model->num_counters;
num_counters      256 arch/x86/oprofile/nmi_int.c 	return counter_config[model->num_counters].count ? 0 : -EINVAL;
num_counters      318 arch/x86/oprofile/nmi_int.c 	size_t counters_size = sizeof(struct op_msr) * model->num_counters;
num_counters      367 arch/x86/oprofile/nmi_int.c 	for (i = 0; i < model->num_counters; ++i) {
num_counters      468 arch/x86/oprofile/nmi_int.c 		       sizeof(struct op_msr) * model->num_counters);
num_counters      767 arch/x86/oprofile/nmi_int.c 		model->num_virt_counters = model->num_counters;
num_counters       42 arch/x86/oprofile/op_model_amd.c static int num_counters;
num_counters      275 arch/x86/oprofile/op_model_amd.c 	for (i = 0; i < num_counters; ++i) {
num_counters      294 arch/x86/oprofile/op_model_amd.c 	for (i = 0; i < num_counters; ++i) {
num_counters      306 arch/x86/oprofile/op_model_amd.c 	for (i = 0; i < num_counters; i++) {
num_counters      314 arch/x86/oprofile/op_model_amd.c 		if (num_counters == AMD64_NUM_COUNTERS_CORE) {
num_counters      349 arch/x86/oprofile/op_model_amd.c 	for (i = 0; i < num_counters; ++i) {
num_counters      365 arch/x86/oprofile/op_model_amd.c 	for (i = 0; i < num_counters; ++i) {
num_counters      387 arch/x86/oprofile/op_model_amd.c 	for (i = 0; i < num_counters; ++i) {
num_counters      410 arch/x86/oprofile/op_model_amd.c 	for (i = 0; i < num_counters; ++i) {
num_counters      430 arch/x86/oprofile/op_model_amd.c 	for (i = 0; i < num_counters; ++i) {
num_counters      516 arch/x86/oprofile/op_model_amd.c 		num_counters = AMD64_NUM_COUNTERS_CORE;
num_counters      518 arch/x86/oprofile/op_model_amd.c 		num_counters = AMD64_NUM_COUNTERS;
num_counters      521 arch/x86/oprofile/op_model_amd.c 	op_amd_spec.num_counters = num_counters;
num_counters      522 arch/x86/oprofile/op_model_amd.c 	op_amd_spec.num_controls = num_counters;
num_counters      523 arch/x86/oprofile/op_model_amd.c 	op_amd_spec.num_virt_counters = max(num_counters, NUM_VIRT_COUNTERS);
num_counters       37 arch/x86/oprofile/op_model_p4.c static unsigned int num_counters = NUM_COUNTERS_NON_HT;
num_counters       47 arch/x86/oprofile/op_model_p4.c 		num_counters = NUM_COUNTERS_HT2;
num_counters      384 arch/x86/oprofile/op_model_p4.c #define VIRT_CTR(stagger, i) ((i) + ((num_counters) * (stagger)))
num_counters      392 arch/x86/oprofile/op_model_p4.c 	for (i = 0; i < num_counters; ++i) {
num_counters      401 arch/x86/oprofile/op_model_p4.c 	for (i = num_counters; i < num_controls; ++i) {
num_counters      416 arch/x86/oprofile/op_model_p4.c 	for (i = 0; i < num_counters; ++i) {
num_counters      468 arch/x86/oprofile/op_model_p4.c 	if (num_counters == NUM_COUNTERS_NON_HT) {
num_counters      490 arch/x86/oprofile/op_model_p4.c 	for (i = 0; i < num_counters; ++i) {
num_counters      585 arch/x86/oprofile/op_model_p4.c 	for (i = 0; i < num_counters; i++) {
num_counters      595 arch/x86/oprofile/op_model_p4.c 	for (i = num_counters; i < num_controls; i++) {
num_counters      602 arch/x86/oprofile/op_model_p4.c 	for (i = 0; i < num_counters; ++i) {
num_counters      623 arch/x86/oprofile/op_model_p4.c 	for (i = 0; i < num_counters; ++i) {
num_counters      675 arch/x86/oprofile/op_model_p4.c 	for (i = 0; i < num_counters; ++i) {
num_counters      692 arch/x86/oprofile/op_model_p4.c 	for (i = 0; i < num_counters; ++i) {
num_counters      703 arch/x86/oprofile/op_model_p4.c 	.num_counters		= NUM_COUNTERS_HT2,
num_counters      715 arch/x86/oprofile/op_model_p4.c 	.num_counters		= NUM_COUNTERS_NON_HT,
num_counters       26 arch/x86/oprofile/op_model_ppro.c static int num_counters = 2;
num_counters       37 arch/x86/oprofile/op_model_ppro.c 	for (i = 0; i < num_counters; ++i) {
num_counters       49 arch/x86/oprofile/op_model_ppro.c 	for (i = 0; i < num_counters; i++) {
num_counters       96 arch/x86/oprofile/op_model_ppro.c 	for (i = 0; i < num_counters; ++i) {
num_counters      112 arch/x86/oprofile/op_model_ppro.c 	for (i = 0; i < num_counters; ++i) {
num_counters      133 arch/x86/oprofile/op_model_ppro.c 	for (i = 0; i < num_counters; ++i) {
num_counters      163 arch/x86/oprofile/op_model_ppro.c 	for (i = 0; i < num_counters; ++i) {
num_counters      178 arch/x86/oprofile/op_model_ppro.c 	for (i = 0; i < num_counters; ++i) {
num_counters      188 arch/x86/oprofile/op_model_ppro.c 	.num_counters		= 2,
num_counters      218 arch/x86/oprofile/op_model_ppro.c 		eax.split.num_counters = 2;
num_counters      222 arch/x86/oprofile/op_model_ppro.c 	num_counters = min((int)eax.split.num_counters, OP_MAX_COUNTER);
num_counters      224 arch/x86/oprofile/op_model_ppro.c 	op_arch_perfmon_spec.num_counters = num_counters;
num_counters      225 arch/x86/oprofile/op_model_ppro.c 	op_arch_perfmon_spec.num_controls = num_counters;
num_counters       37 arch/x86/oprofile/op_x86_model.h 	unsigned int	num_counters;
num_counters      577 drivers/acpi/sysfs.c static u32 num_counters;
num_counters      745 drivers/acpi/sysfs.c 		for (i = 0; i < num_counters; ++i)
num_counters      861 drivers/acpi/sysfs.c 	num_counters = num_gpes + ACPI_NUM_FIXED_EVENTS + NUM_COUNTERS_EXTRA;
num_counters      863 drivers/acpi/sysfs.c 	all_attrs = kcalloc(num_counters + 1, sizeof(struct attribute *),
num_counters      868 drivers/acpi/sysfs.c 	all_counters = kcalloc(num_counters, sizeof(struct event_counter),
num_counters      877 drivers/acpi/sysfs.c 	counter_attrs = kcalloc(num_counters, sizeof(struct kobj_attribute),
num_counters      882 drivers/acpi/sysfs.c 	for (i = 0; i < num_counters; ++i) {
num_counters      203 drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c 		  unsigned int num_counters)
num_counters      239 drivers/gpu/drm/amd/amdgpu/amdgpu_pmu.c 			pmu_type_name, num_counters);
num_counters      400 drivers/gpu/drm/amd/amdkfd/kfd_topology.c 	KFD_PERF_DESC(num_counters, 0),
num_counters       54 drivers/infiniband/core/counters.c 		if (port_counter->num_counters)
num_counters       97 drivers/infiniband/core/counters.c 	port_counter->num_counters++;
num_counters      120 drivers/infiniband/core/counters.c 	port_counter->num_counters--;
num_counters      121 drivers/infiniband/core/counters.c 	if (!port_counter->num_counters &&
num_counters      208 drivers/infiniband/core/counters.c 	for (i = 0; i < counter->stats->num_counters; i++)
num_counters      761 drivers/infiniband/core/nldev.c 	for (i = 0; i < st->num_counters; i++)
num_counters      808 drivers/infiniband/core/sysfs.c 	if (ret == stats->num_counters)
num_counters      973 drivers/infiniband/core/sysfs.c 	if (!stats->names || stats->num_counters <= 0)
num_counters      981 drivers/infiniband/core/sysfs.c 		       sizeof(void *) * (stats->num_counters + 2),
num_counters      987 drivers/infiniband/core/sysfs.c 				       stats->num_counters);
num_counters      988 drivers/infiniband/core/sysfs.c 	if (ret != stats->num_counters)
num_counters      996 drivers/infiniband/core/sysfs.c 	for (i = 0; i < stats->num_counters; i++) {
num_counters     2935 drivers/infiniband/hw/bnxt_re/roce_hsi.h 	__le32	num_counters;
num_counters     1160 drivers/infiniband/hw/cxgb3/iwch_provider.c 	return stats->num_counters;
num_counters      415 drivers/infiniband/hw/cxgb4/provider.c 	return stats->num_counters;
num_counters     2552 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	int num_counters = I40IW_HW_STAT_INDEX_MAX_32 +
num_counters     2566 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	return rdma_alloc_hw_stats_struct(i40iw_hw_stat_names, num_counters,
num_counters     2595 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	return stats->num_counters;
num_counters     2127 drivers/infiniband/hw/mlx4/main.c 					  diag[!!port_num].num_counters,
num_counters     2145 drivers/infiniband/hw/mlx4/main.c 				       diag[!!port].num_counters, port);
num_counters     2150 drivers/infiniband/hw/mlx4/main.c 	for (i = 0; i < diag[!!port].num_counters; i++)
num_counters     2153 drivers/infiniband/hw/mlx4/main.c 	return diag[!!port].num_counters;
num_counters     2162 drivers/infiniband/hw/mlx4/main.c 	u32 num_counters;
num_counters     2164 drivers/infiniband/hw/mlx4/main.c 	num_counters = ARRAY_SIZE(diag_basic);
num_counters     2167 drivers/infiniband/hw/mlx4/main.c 		num_counters += ARRAY_SIZE(diag_ext);
num_counters     2170 drivers/infiniband/hw/mlx4/main.c 		num_counters += ARRAY_SIZE(diag_device_only);
num_counters     2172 drivers/infiniband/hw/mlx4/main.c 	*name = kcalloc(num_counters, sizeof(**name), GFP_KERNEL);
num_counters     2176 drivers/infiniband/hw/mlx4/main.c 	*offset = kcalloc(num_counters, sizeof(**offset), GFP_KERNEL);
num_counters     2180 drivers/infiniband/hw/mlx4/main.c 	*num = num_counters;
num_counters     2240 drivers/infiniband/hw/mlx4/main.c 						    &diag[i].num_counters, i);
num_counters      579 drivers/infiniband/hw/mlx4/mlx4_ib.h 	u32 num_counters;
num_counters     5357 drivers/infiniband/hw/mlx5/main.c 	u32 num_counters;
num_counters     5359 drivers/infiniband/hw/mlx5/main.c 	num_counters = ARRAY_SIZE(basic_q_cnts);
num_counters     5362 drivers/infiniband/hw/mlx5/main.c 		num_counters += ARRAY_SIZE(out_of_seq_q_cnts);
num_counters     5365 drivers/infiniband/hw/mlx5/main.c 		num_counters += ARRAY_SIZE(retrans_q_cnts);
num_counters     5368 drivers/infiniband/hw/mlx5/main.c 		num_counters += ARRAY_SIZE(extended_err_cnts);
num_counters     5370 drivers/infiniband/hw/mlx5/main.c 	cnts->num_q_counters = num_counters;
num_counters     5374 drivers/infiniband/hw/mlx5/main.c 		num_counters += ARRAY_SIZE(cong_cnts);
num_counters     5378 drivers/infiniband/hw/mlx5/main.c 		num_counters += ARRAY_SIZE(ext_ppcnt_cnts);
num_counters     5380 drivers/infiniband/hw/mlx5/main.c 	cnts->names = kcalloc(num_counters, sizeof(cnts->names), GFP_KERNEL);
num_counters     5384 drivers/infiniband/hw/mlx5/main.c 	cnts->offsets = kcalloc(num_counters,
num_counters     5585 drivers/infiniband/hw/mlx5/main.c 	int ret, num_counters;
num_counters     5591 drivers/infiniband/hw/mlx5/main.c 	num_counters = cnts->num_q_counters +
num_counters     5629 drivers/infiniband/hw/mlx5/main.c 	return num_counters;
num_counters      161 drivers/net/ethernet/google/gve/gve_adminq.c 					  u32 num_counters,
num_counters      172 drivers/net/ethernet/google/gve/gve_adminq.c 		.num_counters = cpu_to_be32(num_counters),
num_counters       90 drivers/net/ethernet/google/gve/gve_adminq.h 	__be32 num_counters;
num_counters      205 drivers/net/ethernet/google/gve/gve_adminq.h 					  u32 num_counters,
num_counters      729 drivers/net/ethernet/mellanox/mlx5/core/lag.c 				 int num_counters,
num_counters      743 drivers/net/ethernet/mellanox/mlx5/core/lag.c 	memset(values, 0, sizeof(*values) * num_counters);
num_counters      761 drivers/net/ethernet/mellanox/mlx5/core/lag.c 		for (j = 0; j < num_counters; ++j)
num_counters       31 drivers/oprofile/oprofile_perf.c static int num_counters;
num_counters       42 drivers/oprofile/oprofile_perf.c 	for (id = 0; id < num_counters; ++id)
num_counters       46 drivers/oprofile/oprofile_perf.c 	if (id != num_counters)
num_counters       64 drivers/oprofile/oprofile_perf.c 	for (i = 0; i < num_counters; ++i) {
num_counters      120 drivers/oprofile/oprofile_perf.c 		for (event = 0; event < num_counters; ++event) {
num_counters      138 drivers/oprofile/oprofile_perf.c 		for (event = 0; event < num_counters; ++event)
num_counters      146 drivers/oprofile/oprofile_perf.c 	for (i = 0; i < num_counters; i++) {
num_counters      261 drivers/oprofile/oprofile_perf.c 		for (id = 0; id < num_counters; ++id) {
num_counters      282 drivers/oprofile/oprofile_perf.c 	num_counters = perf_num_counters();
num_counters      283 drivers/oprofile/oprofile_perf.c 	if (num_counters <= 0) {
num_counters      289 drivers/oprofile/oprofile_perf.c 	counter_config = kcalloc(num_counters,
num_counters      294 drivers/oprofile/oprofile_perf.c 				"counters\n", num_counters);
num_counters      296 drivers/oprofile/oprofile_perf.c 		num_counters = 0;
num_counters      301 drivers/oprofile/oprofile_perf.c 		per_cpu(perf_events, cpu) = kcalloc(num_counters,
num_counters      305 drivers/oprofile/oprofile_perf.c 					"for cpu %d\n", num_counters, cpu);
num_counters      120 drivers/perf/arm_dsu_pmu.c 	s8				num_counters;
num_counters      244 drivers/perf/arm_dsu_pmu.c 	return (idx < dsu_pmu->num_counters) ||
num_counters      313 drivers/perf/arm_dsu_pmu.c 	idx = find_first_zero_bit(used_mask, dsu_pmu->num_counters);
num_counters      314 drivers/perf/arm_dsu_pmu.c 	if (idx >= dsu_pmu->num_counters)
num_counters      601 drivers/perf/arm_dsu_pmu.c 	dsu_pmu->num_counters = -1;
num_counters      639 drivers/perf/arm_dsu_pmu.c 	u64 num_counters;
num_counters      642 drivers/perf/arm_dsu_pmu.c 	num_counters = (__dsu_pmu_read_pmcr() >> CLUSTERPMCR_N_SHIFT) &
num_counters      645 drivers/perf/arm_dsu_pmu.c 	if (WARN_ON(num_counters > 31))
num_counters      646 drivers/perf/arm_dsu_pmu.c 		num_counters = 31;
num_counters      647 drivers/perf/arm_dsu_pmu.c 	dsu_pmu->num_counters = num_counters;
num_counters      648 drivers/perf/arm_dsu_pmu.c 	if (!dsu_pmu->num_counters)
num_counters      669 drivers/perf/arm_dsu_pmu.c 	if (dsu_pmu->num_counters == -1)
num_counters      109 drivers/perf/arm_smmuv3_pmu.c 	unsigned int num_counters;
num_counters      278 drivers/perf/arm_smmuv3_pmu.c 	unsigned int num_ctrs = smmu_pmu->num_counters;
num_counters      307 drivers/perf/arm_smmuv3_pmu.c 	unsigned int num_ctrs = smmu_pmu->num_counters;
num_counters      376 drivers/perf/arm_smmuv3_pmu.c 		if (++group_num_events > smmu_pmu->num_counters)
num_counters      387 drivers/perf/arm_smmuv3_pmu.c 		if (++group_num_events > smmu_pmu->num_counters)
num_counters      617 drivers/perf/arm_smmuv3_pmu.c 	for_each_set_bit(idx, (unsigned long *)&ovsr, smmu_pmu->num_counters) {
num_counters      698 drivers/perf/arm_smmuv3_pmu.c 	u64 counter_present_mask = GENMASK_ULL(smmu_pmu->num_counters - 1, 0);
num_counters      784 drivers/perf/arm_smmuv3_pmu.c 	smmu_pmu->num_counters = FIELD_GET(SMMU_PMCG_CFGR_NCTR, cfgr) + 1;
num_counters      829 drivers/perf/arm_smmuv3_pmu.c 		 &res_0->start, smmu_pmu->num_counters,
num_counters      347 drivers/perf/hisilicon/hisi_uncore_ddrc_pmu.c 	ddrc_pmu->num_counters = DDRC_NR_COUNTERS;
num_counters      358 drivers/perf/hisilicon/hisi_uncore_hha_pmu.c 	hha_pmu->num_counters = HHA_NR_COUNTERS;
num_counters      348 drivers/perf/hisilicon/hisi_uncore_l3c_pmu.c 	l3c_pmu->num_counters = L3C_NR_COUNTERS;
num_counters       92 drivers/perf/hisilicon/hisi_uncore_pmu.c 	return counters <= hisi_pmu->num_counters;
num_counters       97 drivers/perf/hisilicon/hisi_uncore_pmu.c 	return idx >= 0 && idx < hisi_pmu->num_counters;
num_counters      104 drivers/perf/hisilicon/hisi_uncore_pmu.c 	u32 num_counters = hisi_pmu->num_counters;
num_counters      107 drivers/perf/hisilicon/hisi_uncore_pmu.c 	idx = find_first_zero_bit(used_mask, num_counters);
num_counters      108 drivers/perf/hisilicon/hisi_uncore_pmu.c 	if (idx == num_counters)
num_counters      324 drivers/perf/hisilicon/hisi_uncore_pmu.c 				    hisi_pmu->num_counters);
num_counters       73 drivers/perf/hisilicon/hisi_uncore_pmu.h 	int num_counters;
num_counters      155 drivers/perf/qcom_l2_pmu.c 	int num_counters;
num_counters      386 drivers/perf/qcom_l2_pmu.c 	int num_ctrs = cluster->l2cache_pmu->num_counters - 1;
num_counters      430 drivers/perf/qcom_l2_pmu.c 	int num_counters = cluster->l2cache_pmu->num_counters;
num_counters      438 drivers/perf/qcom_l2_pmu.c 	for_each_set_bit(idx, cluster->used_counters, num_counters) {
num_counters      968 drivers/perf/qcom_l2_pmu.c 	l2cache_pmu->num_counters = get_num_counters();
num_counters      975 drivers/perf/qcom_l2_pmu.c 	l2_cycle_ctr_idx = l2cache_pmu->num_counters - 1;
num_counters      976 drivers/perf/qcom_l2_pmu.c 	l2_counter_present_mask = GENMASK(l2cache_pmu->num_counters - 2, 0) |
num_counters     1770 drivers/staging/comedi/drivers/ni_tio.c 			 unsigned int num_counters,
num_counters     1778 drivers/staging/comedi/drivers/ni_tio.c 	if (num_counters == 0 || counters_per_chip == 0)
num_counters     1793 drivers/staging/comedi/drivers/ni_tio.c 	counter_dev->num_counters = num_counters;
num_counters     1794 drivers/staging/comedi/drivers/ni_tio.c 	counter_dev->num_chips = DIV_ROUND_UP(num_counters, counters_per_chip);
num_counters     1796 drivers/staging/comedi/drivers/ni_tio.c 	counter_dev->counters = kcalloc(num_counters, sizeof(*counter),
num_counters     1807 drivers/staging/comedi/drivers/ni_tio.c 	for (i = 0; i < num_counters; ++i) {
num_counters      109 drivers/staging/comedi/drivers/ni_tio.h 	unsigned int num_counters;
num_counters      124 drivers/staging/comedi/drivers/ni_tio.h 			 unsigned int num_counters,
num_counters     1088 include/linux/mlx5/driver.h 				 int num_counters,
num_counters      315 include/linux/netfilter/x_tables.h 				       unsigned int num_counters,
num_counters      496 include/linux/netfilter/x_tables.h 	compat_uint_t num_counters;
num_counters      557 include/rdma/ib_verbs.h 	int		num_counters;
num_counters      570 include/rdma/ib_verbs.h 		const char * const *names, int num_counters,
num_counters      575 include/rdma/ib_verbs.h 	stats = kzalloc(sizeof(*stats) + num_counters * sizeof(u64),
num_counters      580 include/rdma/ib_verbs.h 	stats->num_counters = num_counters;
num_counters       31 include/rdma/rdma_counter.h 	unsigned int num_counters;
num_counters      119 include/uapi/linux/netfilter/x_tables.h 	unsigned int num_counters;
num_counters      179 include/uapi/linux/netfilter_arp/arp_tables.h 	unsigned int num_counters;
num_counters       54 include/uapi/linux/netfilter_bridge/ebtables.h 	unsigned int num_counters;
num_counters       70 include/uapi/linux/netfilter_bridge/ebtables.h 	unsigned int num_counters;
num_counters      201 include/uapi/linux/netfilter_ipv4/ip_tables.h 	unsigned int num_counters;
num_counters      241 include/uapi/linux/netfilter_ipv6/ip6_tables.h 	unsigned int num_counters;
num_counters       76 kernel/gcov/clang.c 	u32 num_counters;
num_counters      137 kernel/gcov/clang.c void llvm_gcda_emit_arcs(u32 num_counters, u64 *counters)
num_counters      142 kernel/gcov/clang.c 	info->num_counters = num_counters;
num_counters      240 kernel/gcov/clang.c 				sizeof(fn->counters[0]) * fn->num_counters);
num_counters      293 kernel/gcov/clang.c 		for (i = 0; i < sfn_ptr->num_counters; i++)
num_counters      311 kernel/gcov/clang.c 	cv_size = fn->num_counters * sizeof(fn->counters[0]);
num_counters      476 kernel/gcov/clang.c 		pos += store_gcov_u32(buffer, pos, fi_ptr->num_counters * 2);
num_counters      477 kernel/gcov/clang.c 		for (i = 0; i < fi_ptr->num_counters; i++)
num_counters      979 net/bridge/netfilter/ebtables.c 	if (repl->num_counters) {
num_counters      980 net/bridge/netfilter/ebtables.c 		unsigned long size = repl->num_counters * sizeof(*counterstmp);
num_counters     1006 net/bridge/netfilter/ebtables.c 	if (repl->num_counters && repl->num_counters != t->private->nentries) {
num_counters     1021 net/bridge/netfilter/ebtables.c 	if (repl->num_counters)
num_counters     1033 net/bridge/netfilter/ebtables.c 	if (repl->num_counters &&
num_counters     1035 net/bridge/netfilter/ebtables.c 	   repl->num_counters * sizeof(struct ebt_counter))) {
num_counters     1092 net/bridge/netfilter/ebtables.c 	if (tmp.num_counters >= INT_MAX / sizeof(struct ebt_counter))
num_counters     1250 net/bridge/netfilter/ebtables.c 				unsigned int num_counters,
num_counters     1257 net/bridge/netfilter/ebtables.c 	if (num_counters == 0)
num_counters     1260 net/bridge/netfilter/ebtables.c 	tmp = vmalloc(array_size(num_counters, sizeof(*tmp)));
num_counters     1268 net/bridge/netfilter/ebtables.c 	if (num_counters != t->private->nentries) {
num_counters     1273 net/bridge/netfilter/ebtables.c 	if (copy_from_user(tmp, counters, num_counters * sizeof(*counters))) {
num_counters     1282 net/bridge/netfilter/ebtables.c 	for (i = 0; i < num_counters; i++)
num_counters     1302 net/bridge/netfilter/ebtables.c 	if (len != sizeof(hlp) + hlp.num_counters * sizeof(struct ebt_counter))
num_counters     1306 net/bridge/netfilter/ebtables.c 				hlp.num_counters, user, len);
num_counters     1385 net/bridge/netfilter/ebtables.c 				 void __user *user, unsigned int num_counters,
num_counters     1392 net/bridge/netfilter/ebtables.c 	if (num_counters == 0)
num_counters     1395 net/bridge/netfilter/ebtables.c 	if (num_counters != nentries)
num_counters     1439 net/bridge/netfilter/ebtables.c 	   (tmp.num_counters ? nentries * sizeof(struct ebt_counter) : 0))
num_counters     1449 net/bridge/netfilter/ebtables.c 					tmp.num_counters, nentries);
num_counters     1548 net/bridge/netfilter/ebtables.c 	compat_uint_t num_counters;
num_counters     1827 net/bridge/netfilter/ebtables.c 	   (tmp.num_counters && tmp.num_counters != tinfo.nentries))
num_counters     1839 net/bridge/netfilter/ebtables.c 	   (tmp.num_counters? tinfo.nentries * sizeof(struct ebt_counter): 0)) {
num_counters     1847 net/bridge/netfilter/ebtables.c 					tmp.num_counters, tinfo.nentries);
num_counters     2187 net/bridge/netfilter/ebtables.c 	if (tmp.num_counters >= INT_MAX / sizeof(struct ebt_counter))
num_counters     2196 net/bridge/netfilter/ebtables.c 	repl->num_counters = tmp.num_counters;
num_counters     2314 net/bridge/netfilter/ebtables.c 	if (len != sizeof(hlp) + hlp.num_counters * sizeof(struct ebt_counter))
num_counters     2318 net/bridge/netfilter/ebtables.c 					hlp.num_counters, user, len);
num_counters      883 net/ipv4/netfilter/arp_tables.c 			unsigned int num_counters,
num_counters      894 net/ipv4/netfilter/arp_tables.c 	counters = xt_counters_alloc(num_counters);
num_counters      912 net/ipv4/netfilter/arp_tables.c 	oldinfo = xt_replace_table(t, num_counters, newinfo, &ret);
num_counters      935 net/ipv4/netfilter/arp_tables.c 			 sizeof(struct xt_counters) * num_counters) != 0) {
num_counters      964 net/ipv4/netfilter/arp_tables.c 	if (tmp.num_counters >= INT_MAX / sizeof(struct xt_counters))
num_counters      966 net/ipv4/netfilter/arp_tables.c 	if (tmp.num_counters == 0)
num_counters      987 net/ipv4/netfilter/arp_tables.c 			   tmp.num_counters, tmp.counters);
num_counters     1024 net/ipv4/netfilter/arp_tables.c 	if (private->number != tmp.num_counters) {
num_counters     1058 net/ipv4/netfilter/arp_tables.c 	u32				num_counters;
num_counters     1223 net/ipv4/netfilter/arp_tables.c 	repl.num_counters = 0;
num_counters     1262 net/ipv4/netfilter/arp_tables.c 	if (tmp.num_counters >= INT_MAX / sizeof(struct xt_counters))
num_counters     1264 net/ipv4/netfilter/arp_tables.c 	if (tmp.num_counters == 0)
num_counters     1284 net/ipv4/netfilter/arp_tables.c 			   tmp.num_counters, compat_ptr(tmp.counters));
num_counters     1039 net/ipv4/netfilter/ip_tables.c 	     struct xt_table_info *newinfo, unsigned int num_counters,
num_counters     1049 net/ipv4/netfilter/ip_tables.c 	counters = xt_counters_alloc(num_counters);
num_counters     1067 net/ipv4/netfilter/ip_tables.c 	oldinfo = xt_replace_table(t, num_counters, newinfo, &ret);
num_counters     1089 net/ipv4/netfilter/ip_tables.c 			 sizeof(struct xt_counters) * num_counters) != 0) {
num_counters     1118 net/ipv4/netfilter/ip_tables.c 	if (tmp.num_counters >= INT_MAX / sizeof(struct xt_counters))
num_counters     1120 net/ipv4/netfilter/ip_tables.c 	if (tmp.num_counters == 0)
num_counters     1141 net/ipv4/netfilter/ip_tables.c 			   tmp.num_counters, tmp.counters);
num_counters     1179 net/ipv4/netfilter/ip_tables.c 	if (private->number != tmp.num_counters) {
num_counters     1212 net/ipv4/netfilter/ip_tables.c 	u32			num_counters;
num_counters     1462 net/ipv4/netfilter/ip_tables.c 	repl.num_counters = 0;
num_counters     1501 net/ipv4/netfilter/ip_tables.c 	if (tmp.num_counters >= INT_MAX / sizeof(struct xt_counters))
num_counters     1503 net/ipv4/netfilter/ip_tables.c 	if (tmp.num_counters == 0)
num_counters     1524 net/ipv4/netfilter/ip_tables.c 			   tmp.num_counters, compat_ptr(tmp.counters));
num_counters     1056 net/ipv6/netfilter/ip6_tables.c 	     struct xt_table_info *newinfo, unsigned int num_counters,
num_counters     1066 net/ipv6/netfilter/ip6_tables.c 	counters = xt_counters_alloc(num_counters);
num_counters     1084 net/ipv6/netfilter/ip6_tables.c 	oldinfo = xt_replace_table(t, num_counters, newinfo, &ret);
num_counters     1106 net/ipv6/netfilter/ip6_tables.c 			 sizeof(struct xt_counters) * num_counters) != 0) {
num_counters     1135 net/ipv6/netfilter/ip6_tables.c 	if (tmp.num_counters >= INT_MAX / sizeof(struct xt_counters))
num_counters     1137 net/ipv6/netfilter/ip6_tables.c 	if (tmp.num_counters == 0)
num_counters     1158 net/ipv6/netfilter/ip6_tables.c 			   tmp.num_counters, tmp.counters);
num_counters     1195 net/ipv6/netfilter/ip6_tables.c 	if (private->number != tmp.num_counters) {
num_counters     1228 net/ipv6/netfilter/ip6_tables.c 	u32			num_counters;
num_counters     1471 net/ipv6/netfilter/ip6_tables.c 	repl.num_counters = 0;
num_counters     1510 net/ipv6/netfilter/ip6_tables.c 	if (tmp.num_counters >= INT_MAX / sizeof(struct xt_counters))
num_counters     1512 net/ipv6/netfilter/ip6_tables.c 	if (tmp.num_counters == 0)
num_counters     1533 net/ipv6/netfilter/ip6_tables.c 			   tmp.num_counters, compat_ptr(tmp.counters));
num_counters     1070 net/netfilter/x_tables.c 		info->num_counters = compat_tmp.num_counters;
num_counters     1087 net/netfilter/x_tables.c 	size *= info->num_counters;
num_counters     1354 net/netfilter/x_tables.c 	      unsigned int num_counters,
num_counters     1373 net/netfilter/x_tables.c 	if (num_counters != private->number) {
num_counters     1375 net/netfilter/x_tables.c 			 num_counters, private->number);