max_ctx 40 arch/ia64/include/asm/mmu_context.h unsigned int max_ctx; /* max. context value supported by all CPUs */ max_ctx 95 arch/ia64/include/asm/mmu_context.h ia64_ctx.max_ctx, ia64_ctx.next); max_ctx 97 arch/ia64/include/asm/mmu_context.h ia64_ctx.max_ctx, ia64_ctx.next); max_ctx 98 arch/ia64/include/asm/mmu_context.h if (ia64_ctx.next >= ia64_ctx.max_ctx) max_ctx 946 arch/ia64/kernel/setup.c unsigned int max_ctx; max_ctx 1052 arch/ia64/kernel/setup.c max_ctx = (1U << (vmi.pal_vm_info_2_s.rid_size - 3)) - 1; max_ctx 1056 arch/ia64/kernel/setup.c max_ctx = (1U << 15) - 1; /* use architected minimum */ max_ctx 1058 arch/ia64/kernel/setup.c while (max_ctx < ia64_ctx.max_ctx) { max_ctx 1059 arch/ia64/kernel/setup.c unsigned int old = ia64_ctx.max_ctx; max_ctx 1060 arch/ia64/kernel/setup.c if (cmpxchg(&ia64_ctx.max_ctx, old, max_ctx) == old) max_ctx 46 arch/ia64/mm/tlb.c .max_ctx = ~0U max_ctx 63 arch/ia64/mm/tlb.c ia64_ctx.bitmap = memblock_alloc((ia64_ctx.max_ctx + 1) >> 3, max_ctx 67 arch/ia64/mm/tlb.c (ia64_ctx.max_ctx + 1) >> 3); max_ctx 68 arch/ia64/mm/tlb.c ia64_ctx.flushmap = memblock_alloc((ia64_ctx.max_ctx + 1) >> 3, max_ctx 72 arch/ia64/mm/tlb.c (ia64_ctx.max_ctx + 1) >> 3); max_ctx 84 arch/ia64/mm/tlb.c for (i=0; i <= ia64_ctx.max_ctx / BITS_PER_LONG; i++) { max_ctx 91 arch/ia64/mm/tlb.c ia64_ctx.max_ctx, 300); max_ctx 93 arch/ia64/mm/tlb.c ia64_ctx.max_ctx, ia64_ctx.next);