affinity 22 arch/alpha/include/asm/gct.h gct6_handle affinity; affinity 137 arch/alpha/kernel/sys_dp264.c cpu_set_irq_affinity(unsigned int irq, cpumask_t affinity) affinity 143 arch/alpha/kernel/sys_dp264.c if (cpumask_test_cpu(cpu, &affinity)) affinity 152 arch/alpha/kernel/sys_dp264.c dp264_set_affinity(struct irq_data *d, const struct cpumask *affinity, affinity 156 arch/alpha/kernel/sys_dp264.c cpu_set_irq_affinity(d->irq, *affinity); affinity 164 arch/alpha/kernel/sys_dp264.c clipper_set_affinity(struct irq_data *d, const struct cpumask *affinity, affinity 168 arch/alpha/kernel/sys_dp264.c cpu_set_irq_affinity(d->irq - 16, *affinity); affinity 136 arch/alpha/kernel/sys_titan.c titan_cpu_set_irq_affinity(unsigned int irq, cpumask_t affinity) affinity 141 arch/alpha/kernel/sys_titan.c if (cpumask_test_cpu(cpu, &affinity)) affinity 150 arch/alpha/kernel/sys_titan.c titan_set_irq_affinity(struct irq_data *d, const struct cpumask *affinity, affinity 155 arch/alpha/kernel/sys_titan.c titan_cpu_set_irq_affinity(irq - 16, *affinity); affinity 615 arch/arm/kernel/setup.c u32 i, affinity; affinity 629 arch/arm/kernel/setup.c affinity = MPIDR_AFFINITY_LEVEL(mask, i); affinity 635 arch/arm/kernel/setup.c ls = fls(affinity); affinity 636 arch/arm/kernel/setup.c fs[i] = affinity ? ffs(affinity) - 1 : 0; affinity 114 arch/arm64/kernel/setup.c u32 i, affinity, fs[4], bits[4], ls; affinity 128 arch/arm64/kernel/setup.c affinity = MPIDR_AFFINITY_LEVEL(mask, i); affinity 134 arch/arm64/kernel/setup.c ls = fls(affinity); affinity 135 arch/arm64/kernel/setup.c fs[i] = affinity ? ffs(affinity) - 1 : 0; affinity 19 arch/mips/loongson64/loongson-3/irq.c int plat_set_irq_affinity(struct irq_data *d, const struct cpumask *affinity, affinity 26 arch/mips/loongson64/loongson-3/irq.c cpumask_copy(&new_affinity, affinity); affinity 27 arch/mips/loongson64/loongson-3/irq.c for_each_cpu(cpu, affinity) affinity 34 arch/mips/loongson64/loongson-3/irq.c cpumask_copy(d->common->affinity, &new_affinity); affinity 43 arch/mips/loongson64/loongson-3/irq.c struct cpumask affinity; affinity 59 arch/mips/loongson64/loongson-3/irq.c cpumask_and(&affinity, irqd->common->affinity, cpu_active_mask); affinity 60 arch/mips/loongson64/loongson-3/irq.c if (cpumask_empty(&affinity)) { affinity 65 arch/mips/loongson64/loongson-3/irq.c irq_cpu[ht_irq[i]] = cpumask_next(irq_cpu[ht_irq[i]], &affinity); affinity 67 arch/mips/loongson64/loongson-3/irq.c irq_cpu[ht_irq[i]] = cpumask_first(&affinity); affinity 85 arch/mips/pmcs-msp71xx/msp_irq_per.c const struct cpumask *affinity, bool force) affinity 152 arch/mips/sgi-ip27/ip27-irq.c cpumask_copy(desc->irq_common_data.affinity, &hub->h_cpus); affinity 409 arch/powerpc/platforms/cell/spufs/inode.c int affinity; affinity 427 arch/powerpc/platforms/cell/spufs/inode.c affinity = flags & (SPU_CREATE_AFFINITY_MEM | SPU_CREATE_AFFINITY_SPU); affinity 428 arch/powerpc/platforms/cell/spufs/inode.c if (affinity) { affinity 444 arch/powerpc/platforms/cell/spufs/inode.c if (affinity) { affinity 456 arch/powerpc/platforms/cell/spufs/inode.c if (affinity) affinity 556 arch/powerpc/sysdev/xive/common.c const struct cpumask *affinity) affinity 570 arch/powerpc/sysdev/xive/common.c for_each_cpu_and(cpu, affinity, cpu_online_mask) { affinity 587 arch/powerpc/sysdev/xive/common.c return xive_find_target_in_mask(affinity, fuzz++); affinity 280 arch/s390/pci/pci_irq.c msi->affinity : NULL); affinity 291 arch/s390/pci/pci_irq.c msg.address_lo |= msi->affinity ? affinity 292 arch/s390/pci/pci_irq.c (cpumask_first(&msi->affinity->mask) << 8) : 0; affinity 349 arch/sparc/kernel/irq_64.c static int irq_choose_cpu(unsigned int irq, const struct cpumask *affinity) affinity 354 arch/sparc/kernel/irq_64.c cpumask_copy(&mask, affinity); affinity 367 arch/sparc/kernel/irq_64.c #define irq_choose_cpu(irq, affinity) \ affinity 109 arch/sparc/kernel/leon_kernel.c static int irq_choose_cpu(const struct cpumask *affinity) affinity 113 arch/sparc/kernel/leon_kernel.c cpumask_and(&mask, cpu_online_mask, affinity); affinity 120 arch/sparc/kernel/leon_kernel.c #define irq_choose_cpu(affinity) boot_cpu_id affinity 559 drivers/firmware/arm_sdei.c void *arg, u64 flags, u64 affinity) affinity 563 drivers/firmware/arm_sdei.c flags, affinity, NULL); affinity 128 drivers/gpu/drm/panfrost/panfrost_job.c u64 affinity; affinity 135 drivers/gpu/drm/panfrost/panfrost_job.c affinity = pfdev->features.shader_present; affinity 137 drivers/gpu/drm/panfrost/panfrost_job.c job_write(pfdev, JS_AFFINITY_NEXT_LO(js), affinity & 0xFFFFFFFF); affinity 138 drivers/gpu/drm/panfrost/panfrost_job.c job_write(pfdev, JS_AFFINITY_NEXT_HI(js), affinity >> 32); affinity 1007 drivers/infiniband/hw/hfi1/affinity.c struct hfi1_affinity_node_list *affinity) affinity 1011 drivers/infiniband/hw/hfi1/affinity.c affinity->num_core_siblings / affinity 1014 drivers/infiniband/hw/hfi1/affinity.c cpumask_copy(hw_thread_mask, &affinity->proc.mask); affinity 1015 drivers/infiniband/hw/hfi1/affinity.c if (affinity->num_core_siblings > 0) { affinity 1044 drivers/infiniband/hw/hfi1/affinity.c struct hfi1_affinity_node_list *affinity = &node_affinity; affinity 1045 drivers/infiniband/hw/hfi1/affinity.c struct cpu_mask_set *set = &affinity->proc; affinity 1103 drivers/infiniband/hw/hfi1/affinity.c mutex_lock(&affinity->lock); affinity 1133 drivers/infiniband/hw/hfi1/affinity.c if (affinity->num_core_siblings > 0) { affinity 1134 drivers/infiniband/hw/hfi1/affinity.c for (i = 0; i < affinity->num_core_siblings; i++) { affinity 1135 drivers/infiniband/hw/hfi1/affinity.c find_hw_thread_mask(i, hw_thread_mask, affinity); affinity 1209 drivers/infiniband/hw/hfi1/affinity.c mutex_unlock(&affinity->lock); affinity 1225 drivers/infiniband/hw/hfi1/affinity.c struct hfi1_affinity_node_list *affinity = &node_affinity; affinity 1226 drivers/infiniband/hw/hfi1/affinity.c struct cpu_mask_set *set = &affinity->proc; affinity 1231 drivers/infiniband/hw/hfi1/affinity.c mutex_lock(&affinity->lock); affinity 1234 drivers/infiniband/hw/hfi1/affinity.c mutex_unlock(&affinity->lock); affinity 111 drivers/iommu/hyperv-iommu.c cpumask_copy(desc->irq_common_data.affinity, &ioapic_max_cpumask); affinity 42 drivers/irqchip/irq-bcm7038-l1.c u8 affinity[MAX_WORDS * IRQS_PER_WORD]; affinity 176 drivers/irqchip/irq-bcm7038-l1.c __bcm7038_l1_unmask(d, intc->affinity[d->hwirq]); affinity 186 drivers/irqchip/irq-bcm7038-l1.c __bcm7038_l1_mask(d, intc->affinity[d->hwirq]); affinity 204 drivers/irqchip/irq-bcm7038-l1.c was_disabled = !!(intc->cpus[intc->affinity[hw]]->mask_cache[word] & affinity 206 drivers/irqchip/irq-bcm7038-l1.c __bcm7038_l1_mask(d, intc->affinity[hw]); affinity 207 drivers/irqchip/irq-bcm7038-l1.c intc->affinity[hw] = first_cpu; affinity 715 drivers/irqchip/irq-gic-v3.c u64 affinity; affinity 757 drivers/irqchip/irq-gic-v3.c affinity = gic_mpidr_to_affinity(cpu_logical_map(smp_processor_id())); affinity 759 drivers/irqchip/irq-gic-v3.c gic_write_irouter(affinity, base + GICD_IROUTER + i * 8); affinity 762 drivers/irqchip/irq-gic-v3.c gic_write_irouter(affinity, base + GICD_IROUTERnE + i * 8); affinity 1298 drivers/pci/controller/pci-hyperv.c struct pci_create_interrupt *int_pkt, struct cpumask *affinity, affinity 1317 drivers/pci/controller/pci-hyperv.c struct pci_create_interrupt2 *int_pkt, struct cpumask *affinity, affinity 1332 drivers/pci/controller/pci-hyperv.c cpu = cpumask_first_and(affinity, cpu_online_mask); affinity 1300 drivers/pci/msi.c return &entry->affinity->mask; affinity 1308 drivers/pci/msi.c if (WARN_ON_ONCE(!entry || !entry->affinity || affinity 1312 drivers/pci/msi.c return &entry->affinity[nr].mask; affinity 441 drivers/xen/evtchn.c desc->irq_common_data.affinity, cpu_online_mask); affinity 444 drivers/xen/evtchn.c selected_cpu = cpumask_first_and(desc->irq_common_data.affinity, affinity 40 include/linux/cpu_rmap.h const struct cpumask *affinity); affinity 150 include/linux/irq.h cpumask_var_t affinity; affinity 676 include/linux/irq.h const struct cpumask *affinity); affinity 678 include/linux/irq.h struct cpumask *affinity); affinity 835 include/linux/irq.h return d ? d->common->affinity : NULL; affinity 840 include/linux/irq.h return d->common->affinity; affinity 862 include/linux/irq.h return d->common->affinity; affinity 870 include/linux/irq.h const struct irq_affinity_desc *affinity); affinity 874 include/linux/irq.h const struct irq_affinity_desc *affinity); affinity 279 include/linux/irqdomain.h const struct irq_affinity_desc *affinity); affinity 468 include/linux/irqdomain.h const struct irq_affinity_desc *affinity); affinity 92 include/linux/msi.h struct irq_affinity_desc *affinity; affinity 185 include/linux/msi.h const struct irq_affinity_desc *affinity); affinity 182 include/ras/ras_event.h __field(u8, affinity) affinity 187 include/ras/ras_event.h __entry->affinity = proc->affinity_level; affinity 189 include/ras/ras_event.h __entry->affinity = ~0; affinity 206 include/ras/ras_event.h __entry->affinity, __entry->mpidr, __entry->midr, affinity 57 kernel/irq/cpuhotplug.c const struct cpumask *affinity; affinity 104 kernel/irq/cpuhotplug.c affinity = irq_desc_get_pending_mask(desc); affinity 106 kernel/irq/cpuhotplug.c affinity = irq_data_get_affinity_mask(d); affinity 112 kernel/irq/cpuhotplug.c if (cpumask_any_and(affinity, cpu_online_mask) >= nr_cpu_ids) { affinity 122 kernel/irq/cpuhotplug.c affinity = cpu_online_mask; affinity 131 kernel/irq/cpuhotplug.c err = irq_do_set_affinity(d, affinity, false); affinity 177 kernel/irq/cpuhotplug.c const struct cpumask *affinity = irq_data_get_affinity_mask(data); affinity 180 kernel/irq/cpuhotplug.c !irq_data_get_irq_chip(data) || !cpumask_test_cpu(cpu, affinity)) affinity 194 kernel/irq/cpuhotplug.c irq_set_affinity_locked(data, affinity, false); affinity 180 kernel/irq/devres.c const struct irq_affinity_desc *affinity) affinity 189 kernel/irq/devres.c base = __irq_alloc_descs(irq, from, cnt, node, owner, affinity); affinity 94 kernel/irq/ipi.c cpumask_copy(data->common->affinity, dest); affinity 57 kernel/irq/irqdesc.c if (!zalloc_cpumask_var_node(&desc->irq_common_data.affinity, affinity 64 kernel/irq/irqdesc.c free_cpumask_var(desc->irq_common_data.affinity); affinity 74 kernel/irq/irqdesc.c free_cpumask_var(desc->irq_common_data.affinity); affinity 82 kernel/irq/irqdesc.c const struct cpumask *affinity) affinity 84 kernel/irq/irqdesc.c if (!affinity) affinity 85 kernel/irq/irqdesc.c affinity = irq_default_affinity; affinity 86 kernel/irq/irqdesc.c cpumask_copy(desc->irq_common_data.affinity, affinity); affinity 100 kernel/irq/irqdesc.c desc_smp_init(struct irq_desc *desc, int node, const struct cpumask *affinity) { } affinity 104 kernel/irq/irqdesc.c const struct cpumask *affinity, struct module *owner) affinity 127 kernel/irq/irqdesc.c desc_smp_init(desc, node, affinity); affinity 368 kernel/irq/irqdesc.c free_cpumask_var(desc->irq_common_data.affinity); affinity 388 kernel/irq/irqdesc.c const struct cpumask *affinity, affinity 409 kernel/irq/irqdesc.c desc_set_defaults(irq, desc, node, affinity, owner); affinity 467 kernel/irq/irqdesc.c const struct irq_affinity_desc *affinity, affinity 474 kernel/irq/irqdesc.c if (affinity) { affinity 476 kernel/irq/irqdesc.c if (cpumask_empty(&affinity[i].mask)) affinity 485 kernel/irq/irqdesc.c if (affinity) { affinity 486 kernel/irq/irqdesc.c if (affinity->is_managed) { affinity 490 kernel/irq/irqdesc.c mask = &affinity->mask; affinity 492 kernel/irq/irqdesc.c affinity++; affinity 598 kernel/irq/irqdesc.c const struct irq_affinity_desc *affinity, affinity 767 kernel/irq/irqdesc.c struct module *owner, const struct irq_affinity_desc *affinity) affinity 800 kernel/irq/irqdesc.c ret = alloc_descs(start, cnt, node, affinity, owner); affinity 901 kernel/irq/irqdesc.c const struct cpumask *affinity) affinity 916 kernel/irq/irqdesc.c if (affinity) affinity 917 kernel/irq/irqdesc.c desc->percpu_affinity = affinity; affinity 930 kernel/irq/irqdesc.c int irq_get_percpu_devid_partition(unsigned int irq, struct cpumask *affinity) affinity 937 kernel/irq/irqdesc.c if (affinity) affinity 938 kernel/irq/irqdesc.c cpumask_copy(affinity, desc->percpu_affinity); affinity 1011 kernel/irq/irqdomain.c int node, const struct irq_affinity_desc *affinity) affinity 1017 kernel/irq/irqdomain.c affinity); affinity 1023 kernel/irq/irqdomain.c affinity); affinity 1026 kernel/irq/irqdomain.c affinity); affinity 1328 kernel/irq/irqdomain.c bool realloc, const struct irq_affinity_desc *affinity) affinity 1342 kernel/irq/irqdomain.c affinity); affinity 224 kernel/irq/manage.c cpumask_copy(desc->irq_common_data.affinity, mask); affinity 344 kernel/irq/manage.c cpumask_copy(cpumask, desc->irq_common_data.affinity); affinity 424 kernel/irq/manage.c if (cpumask_intersects(desc->irq_common_data.affinity, affinity 426 kernel/irq/manage.c set = desc->irq_common_data.affinity; affinity 961 kernel/irq/manage.c if (cpumask_available(desc->irq_common_data.affinity)) { affinity 30 kernel/irq/msi.c const struct irq_affinity_desc *affinity) affinity 41 kernel/irq/msi.c if (affinity) { affinity 42 kernel/irq/msi.c desc->affinity = kmemdup(affinity, affinity 43 kernel/irq/msi.c nvec * sizeof(*desc->affinity), GFP_KERNEL); affinity 44 kernel/irq/msi.c if (!desc->affinity) { affinity 55 kernel/irq/msi.c kfree(entry->affinity); affinity 419 kernel/irq/msi.c desc->affinity); affinity 54 kernel/irq/proc.c mask = desc->irq_common_data.affinity; affinity 156 lib/cpu_rmap.c const struct cpumask *affinity) affinity 179 lib/cpu_rmap.c for_each_cpu(cpu, affinity) { affinity 1040 net/ceph/osdmap.c u32 *affinity; affinity 1042 net/ceph/osdmap.c affinity = ceph_kvmalloc(array_size(max, sizeof(*affinity)), affinity 1044 net/ceph/osdmap.c if (!affinity) affinity 1047 net/ceph/osdmap.c memcpy(affinity, map->osd_primary_affinity, affinity 1048 net/ceph/osdmap.c to_copy * sizeof(*affinity)); affinity 1051 net/ceph/osdmap.c map->osd_primary_affinity = affinity; affinity 708 tools/perf/builtin-record.c if (opts->affinity != PERF_AFFINITY_SYS) affinity 714 tools/perf/builtin-record.c opts->nr_cblocks, opts->affinity, affinity 907 tools/perf/builtin-record.c if (rec->opts.affinity != PERF_AFFINITY_SYS && affinity 1932 tools/perf/builtin-record.c opts->affinity = PERF_AFFINITY_NODE; affinity 1934 tools/perf/builtin-record.c opts->affinity = PERF_AFFINITY_CPU; affinity 2300 tools/perf/builtin-record.c rec->opts.affinity = PERF_AFFINITY_SYS; affinity 2467 tools/perf/builtin-record.c pr_debug("affinity: %s\n", affinity_tags[rec->opts.affinity]); affinity 890 tools/perf/util/evlist.c bool auxtrace_overwrite, int nr_cblocks, int affinity, int flush, affinity 901 tools/perf/util/evlist.c struct mmap_params mp = { .nr_cblocks = nr_cblocks, .affinity = affinity, .flush = flush, affinity 190 tools/perf/util/evlist.h int affinity, int flush, int comp_level); affinity 191 tools/perf/util/mmap.c static int perf_mmap__aio_bind(struct mmap *map, int idx, int cpu, int affinity) affinity 197 tools/perf/util/mmap.c if (affinity != PERF_AFFINITY_SYS && cpu__max_node() > 1) { affinity 226 tools/perf/util/mmap.c int cpu __maybe_unused, int affinity __maybe_unused) affinity 260 tools/perf/util/mmap.c ret = perf_mmap__aio_bind(map, i, map->core.cpu, mp->affinity); affinity 350 tools/perf/util/mmap.c if (mp->affinity == PERF_AFFINITY_NODE && cpu__max_node() > 1) affinity 352 tools/perf/util/mmap.c else if (mp->affinity == PERF_AFFINITY_CPU) affinity 40 tools/perf/util/mmap.h int prot, mask, nr_cblocks, affinity, flush, comp_level; affinity 64 tools/perf/util/record.h int affinity; affinity 18 tools/testing/selftests/rseq/basic_test.c cpu_set_t affinity, test_affinity; affinity 21 tools/testing/selftests/rseq/basic_test.c sched_getaffinity(0, sizeof(affinity), &affinity); affinity 24 tools/testing/selftests/rseq/basic_test.c if (CPU_ISSET(i, &affinity)) { affinity 35 tools/testing/selftests/rseq/basic_test.c sched_setaffinity(0, sizeof(affinity), &affinity); affinity 841 virt/kvm/arm/vgic/vgic-mmio-v3.c unsigned long affinity; affinity 848 virt/kvm/arm/vgic/vgic-mmio-v3.c affinity = kvm_vcpu_get_mpidr_aff(vcpu); affinity 849 virt/kvm/arm/vgic/vgic-mmio-v3.c level0 = MPIDR_AFFINITY_LEVEL(affinity, 0); affinity 850 virt/kvm/arm/vgic/vgic-mmio-v3.c affinity &= ~MPIDR_LEVEL_MASK; affinity 853 virt/kvm/arm/vgic/vgic-mmio-v3.c if (sgi_aff != affinity)