ia64_rse_slot_num   27 arch/ia64/include/uapi/asm/rse.h 	return ia64_rse_slot_num(addr) == 0x3f;
ia64_rse_slot_num   50 arch/ia64/include/uapi/asm/rse.h 	return slots - (ia64_rse_slot_num(bspstore) + slots)/0x40;
ia64_rse_slot_num   60 arch/ia64/include/uapi/asm/rse.h 	long delta = ia64_rse_slot_num(addr) + num_regs;
ia64_rse_slot_num 1145 arch/ia64/kernel/mca.c 		nat = (new_rnat >> ia64_rse_slot_num(new_bspstore)) & 1UL;
ia64_rse_slot_num 1146 arch/ia64/kernel/mca.c 		old_rnat &= ~(1UL << ia64_rse_slot_num(old_bspstore));
ia64_rse_slot_num 1147 arch/ia64/kernel/mca.c 		old_rnat |= (nat << ia64_rse_slot_num(old_bspstore));
ia64_rse_slot_num  279 arch/ia64/kernel/ptrace.c 	shift = ia64_rse_slot_num(slot0_kaddr);
ia64_rse_slot_num  285 arch/ia64/kernel/ptrace.c 		umask = MASK(ia64_rse_slot_num(ubspstore)) & mask;
ia64_rse_slot_num  353 arch/ia64/kernel/ptrace.c 	shift = ia64_rse_slot_num(slot0_kaddr);
ia64_rse_slot_num  359 arch/ia64/kernel/ptrace.c 		umask = MASK(ia64_rse_slot_num(ubspstore)) & mask;
ia64_rse_slot_num  435 arch/ia64/kernel/ptrace.c 		if (((1UL << ia64_rse_slot_num(laddr)) & ret) != 0) {
ia64_rse_slot_num  339 arch/ia64/kernel/unaligned.c 		nat_mask = 1UL << ia64_rse_slot_num(addr);
ia64_rse_slot_num  367 arch/ia64/kernel/unaligned.c 	       (void *) rnat_addr, rnats, nat, (rnats >> ia64_rse_slot_num(addr)) & 1);
ia64_rse_slot_num  369 arch/ia64/kernel/unaligned.c 	nat_mask = 1UL << ia64_rse_slot_num(addr);
ia64_rse_slot_num  414 arch/ia64/kernel/unaligned.c 			nat_mask = 1UL << ia64_rse_slot_num(addr);
ia64_rse_slot_num  436 arch/ia64/kernel/unaligned.c 		nat_mask = 1UL << ia64_rse_slot_num(addr);
ia64_rse_slot_num  351 arch/ia64/kernel/unwind.c 					nat_mask = (1UL << ia64_rse_slot_num(addr));
ia64_rse_slot_num  382 arch/ia64/kernel/unwind.c 		nat_mask = (1UL << ia64_rse_slot_num(addr));