max_ctx            40 arch/ia64/include/asm/mmu_context.h 	unsigned int max_ctx;   /* max. context value supported by all CPUs */
max_ctx            95 arch/ia64/include/asm/mmu_context.h 					ia64_ctx.max_ctx, ia64_ctx.next);
max_ctx            97 arch/ia64/include/asm/mmu_context.h 					ia64_ctx.max_ctx, ia64_ctx.next);
max_ctx            98 arch/ia64/include/asm/mmu_context.h 			if (ia64_ctx.next >= ia64_ctx.max_ctx)
max_ctx           946 arch/ia64/kernel/setup.c 	unsigned int max_ctx;
max_ctx          1052 arch/ia64/kernel/setup.c 		max_ctx = (1U << (vmi.pal_vm_info_2_s.rid_size - 3)) - 1;
max_ctx          1056 arch/ia64/kernel/setup.c 		max_ctx = (1U << 15) - 1;	/* use architected minimum */
max_ctx          1058 arch/ia64/kernel/setup.c 	while (max_ctx < ia64_ctx.max_ctx) {
max_ctx          1059 arch/ia64/kernel/setup.c 		unsigned int old = ia64_ctx.max_ctx;
max_ctx          1060 arch/ia64/kernel/setup.c 		if (cmpxchg(&ia64_ctx.max_ctx, old, max_ctx) == old)
max_ctx            46 arch/ia64/mm/tlb.c 	.max_ctx = ~0U
max_ctx            63 arch/ia64/mm/tlb.c 	ia64_ctx.bitmap = memblock_alloc((ia64_ctx.max_ctx + 1) >> 3,
max_ctx            67 arch/ia64/mm/tlb.c 		      (ia64_ctx.max_ctx + 1) >> 3);
max_ctx            68 arch/ia64/mm/tlb.c 	ia64_ctx.flushmap = memblock_alloc((ia64_ctx.max_ctx + 1) >> 3,
max_ctx            72 arch/ia64/mm/tlb.c 		      (ia64_ctx.max_ctx + 1) >> 3);
max_ctx            84 arch/ia64/mm/tlb.c 	for (i=0; i <= ia64_ctx.max_ctx / BITS_PER_LONG; i++) {
max_ctx            91 arch/ia64/mm/tlb.c 				ia64_ctx.max_ctx, 300);
max_ctx            93 arch/ia64/mm/tlb.c 				ia64_ctx.max_ctx, ia64_ctx.next);