nmsk 12 kernel/irq/affinity.c static void irq_spread_init_one(struct cpumask *irqmsk, struct cpumask *nmsk, nmsk 19 kernel/irq/affinity.c cpu = cpumask_first(nmsk); nmsk 25 kernel/irq/affinity.c cpumask_clear_cpu(cpu, nmsk); nmsk 35 kernel/irq/affinity.c if (!cpumask_test_and_clear_cpu(sibl, nmsk)) nmsk 132 kernel/irq/affinity.c struct cpumask *nmsk, nmsk 145 kernel/irq/affinity.c cpumask_and(nmsk, cpu_mask, node_to_cpumask[n]); nmsk 146 kernel/irq/affinity.c ncpus = cpumask_weight(nmsk); nmsk 252 kernel/irq/affinity.c struct cpumask *nmsk, nmsk 288 kernel/irq/affinity.c nodemsk, nmsk, node_vectors); nmsk 298 kernel/irq/affinity.c cpumask_and(nmsk, cpu_mask, node_to_cpumask[nv->id]); nmsk 299 kernel/irq/affinity.c ncpus = cpumask_weight(nmsk); nmsk 324 kernel/irq/affinity.c irq_spread_init_one(&masks[curvec].mask, nmsk, nmsk 344 kernel/irq/affinity.c cpumask_var_t nmsk, npresmsk; nmsk 347 kernel/irq/affinity.c if (!zalloc_cpumask_var(&nmsk, GFP_KERNEL)) nmsk 364 kernel/irq/affinity.c nmsk, masks); nmsk 381 kernel/irq/affinity.c node_to_cpumask, npresmsk, nmsk, nmsk 398 kernel/irq/affinity.c free_cpumask_var(nmsk); nmsk 1562 kernel/irq/manage.c unsigned int nmsk = new->flags & IRQF_TRIGGER_MASK; nmsk 1565 kernel/irq/manage.c if (nmsk != omsk) nmsk 1568 kernel/irq/manage.c irq, omsk, nmsk);