nmsk               12 kernel/irq/affinity.c static void irq_spread_init_one(struct cpumask *irqmsk, struct cpumask *nmsk,
nmsk               19 kernel/irq/affinity.c 		cpu = cpumask_first(nmsk);
nmsk               25 kernel/irq/affinity.c 		cpumask_clear_cpu(cpu, nmsk);
nmsk               35 kernel/irq/affinity.c 			if (!cpumask_test_and_clear_cpu(sibl, nmsk))
nmsk              132 kernel/irq/affinity.c 				struct cpumask *nmsk,
nmsk              145 kernel/irq/affinity.c 		cpumask_and(nmsk, cpu_mask, node_to_cpumask[n]);
nmsk              146 kernel/irq/affinity.c 		ncpus = cpumask_weight(nmsk);
nmsk              252 kernel/irq/affinity.c 				      struct cpumask *nmsk,
nmsk              288 kernel/irq/affinity.c 			    nodemsk, nmsk, node_vectors);
nmsk              298 kernel/irq/affinity.c 		cpumask_and(nmsk, cpu_mask, node_to_cpumask[nv->id]);
nmsk              299 kernel/irq/affinity.c 		ncpus = cpumask_weight(nmsk);
nmsk              324 kernel/irq/affinity.c 			irq_spread_init_one(&masks[curvec].mask, nmsk,
nmsk              344 kernel/irq/affinity.c 	cpumask_var_t nmsk, npresmsk;
nmsk              347 kernel/irq/affinity.c 	if (!zalloc_cpumask_var(&nmsk, GFP_KERNEL))
nmsk              364 kernel/irq/affinity.c 					 nmsk, masks);
nmsk              381 kernel/irq/affinity.c 					 node_to_cpumask, npresmsk, nmsk,
nmsk              398 kernel/irq/affinity.c 	free_cpumask_var(nmsk);
nmsk             1562 kernel/irq/manage.c 		unsigned int nmsk = new->flags & IRQF_TRIGGER_MASK;
nmsk             1565 kernel/irq/manage.c 		if (nmsk != omsk)
nmsk             1568 kernel/irq/manage.c 				irq, omsk, nmsk);