emu_cores          67 arch/s390/numa/mode_emu.c } *emu_cores;
emu_cores          74 arch/s390/numa/mode_emu.c 	if (emu_cores->to_node_id[core_id] == NODE_ID_FREE) {
emu_cores          75 arch/s390/numa/mode_emu.c 		emu_cores->per_node[node_id]++;
emu_cores          76 arch/s390/numa/mode_emu.c 		emu_cores->to_node_id[core_id] = node_id;
emu_cores          77 arch/s390/numa/mode_emu.c 		emu_cores->total++;
emu_cores          79 arch/s390/numa/mode_emu.c 		WARN_ON(emu_cores->to_node_id[core_id] != node_id);
emu_cores          88 arch/s390/numa/mode_emu.c 	return emu_cores->per_node[node->id];
emu_cores          96 arch/s390/numa/mode_emu.c 	return emu_cores->to_node_id[core->id];
emu_cores         196 arch/s390/numa/mode_emu.c 	cores_target = emu_cores->per_node_target + extra;
emu_cores         240 arch/s390/numa/mode_emu.c 	int cores_free, cores_target = emu_cores->per_node_target;
emu_cores         315 arch/s390/numa/mode_emu.c 	emu_cores = memblock_alloc(sizeof(*emu_cores), 8);
emu_cores         316 arch/s390/numa/mode_emu.c 	if (!emu_cores)
emu_cores         318 arch/s390/numa/mode_emu.c 		      __func__, sizeof(*emu_cores), 8);
emu_cores         319 arch/s390/numa/mode_emu.c 	for (i = 0; i < ARRAY_SIZE(emu_cores->to_node_id); i++)
emu_cores         320 arch/s390/numa/mode_emu.c 		emu_cores->to_node_id[i] = NODE_ID_FREE;
emu_cores         333 arch/s390/numa/mode_emu.c 	cores_total = emu_cores->total + cores_free(phys);
emu_cores         334 arch/s390/numa/mode_emu.c 	emu_cores->per_node_target = cores_total / emu_nodes;
emu_cores         422 arch/s390/numa/mode_emu.c 		for (cid = 0; cid < ARRAY_SIZE(emu_cores->to_node_id); cid++) {
emu_cores         423 arch/s390/numa/mode_emu.c 			if (emu_cores->to_node_id[cid] == nid)
emu_cores         441 arch/s390/numa/mode_emu.c 		if (emu_cores->to_node_id[core_id] != NODE_ID_FREE)
emu_cores         461 arch/s390/numa/mode_emu.c 	if (emu_cores == NULL)