emu_cores 67 arch/s390/numa/mode_emu.c } *emu_cores; emu_cores 74 arch/s390/numa/mode_emu.c if (emu_cores->to_node_id[core_id] == NODE_ID_FREE) { emu_cores 75 arch/s390/numa/mode_emu.c emu_cores->per_node[node_id]++; emu_cores 76 arch/s390/numa/mode_emu.c emu_cores->to_node_id[core_id] = node_id; emu_cores 77 arch/s390/numa/mode_emu.c emu_cores->total++; emu_cores 79 arch/s390/numa/mode_emu.c WARN_ON(emu_cores->to_node_id[core_id] != node_id); emu_cores 88 arch/s390/numa/mode_emu.c return emu_cores->per_node[node->id]; emu_cores 96 arch/s390/numa/mode_emu.c return emu_cores->to_node_id[core->id]; emu_cores 196 arch/s390/numa/mode_emu.c cores_target = emu_cores->per_node_target + extra; emu_cores 240 arch/s390/numa/mode_emu.c int cores_free, cores_target = emu_cores->per_node_target; emu_cores 315 arch/s390/numa/mode_emu.c emu_cores = memblock_alloc(sizeof(*emu_cores), 8); emu_cores 316 arch/s390/numa/mode_emu.c if (!emu_cores) emu_cores 318 arch/s390/numa/mode_emu.c __func__, sizeof(*emu_cores), 8); emu_cores 319 arch/s390/numa/mode_emu.c for (i = 0; i < ARRAY_SIZE(emu_cores->to_node_id); i++) emu_cores 320 arch/s390/numa/mode_emu.c emu_cores->to_node_id[i] = NODE_ID_FREE; emu_cores 333 arch/s390/numa/mode_emu.c cores_total = emu_cores->total + cores_free(phys); emu_cores 334 arch/s390/numa/mode_emu.c emu_cores->per_node_target = cores_total / emu_nodes; emu_cores 422 arch/s390/numa/mode_emu.c for (cid = 0; cid < ARRAY_SIZE(emu_cores->to_node_id); cid++) { emu_cores 423 arch/s390/numa/mode_emu.c if (emu_cores->to_node_id[cid] == nid) emu_cores 441 arch/s390/numa/mode_emu.c if (emu_cores->to_node_id[core_id] != NODE_ID_FREE) emu_cores 461 arch/s390/numa/mode_emu.c if (emu_cores == NULL)