ia64_rse_slot_num 27 arch/ia64/include/uapi/asm/rse.h return ia64_rse_slot_num(addr) == 0x3f; ia64_rse_slot_num 50 arch/ia64/include/uapi/asm/rse.h return slots - (ia64_rse_slot_num(bspstore) + slots)/0x40; ia64_rse_slot_num 60 arch/ia64/include/uapi/asm/rse.h long delta = ia64_rse_slot_num(addr) + num_regs; ia64_rse_slot_num 1145 arch/ia64/kernel/mca.c nat = (new_rnat >> ia64_rse_slot_num(new_bspstore)) & 1UL; ia64_rse_slot_num 1146 arch/ia64/kernel/mca.c old_rnat &= ~(1UL << ia64_rse_slot_num(old_bspstore)); ia64_rse_slot_num 1147 arch/ia64/kernel/mca.c old_rnat |= (nat << ia64_rse_slot_num(old_bspstore)); ia64_rse_slot_num 279 arch/ia64/kernel/ptrace.c shift = ia64_rse_slot_num(slot0_kaddr); ia64_rse_slot_num 285 arch/ia64/kernel/ptrace.c umask = MASK(ia64_rse_slot_num(ubspstore)) & mask; ia64_rse_slot_num 353 arch/ia64/kernel/ptrace.c shift = ia64_rse_slot_num(slot0_kaddr); ia64_rse_slot_num 359 arch/ia64/kernel/ptrace.c umask = MASK(ia64_rse_slot_num(ubspstore)) & mask; ia64_rse_slot_num 435 arch/ia64/kernel/ptrace.c if (((1UL << ia64_rse_slot_num(laddr)) & ret) != 0) { ia64_rse_slot_num 339 arch/ia64/kernel/unaligned.c nat_mask = 1UL << ia64_rse_slot_num(addr); ia64_rse_slot_num 367 arch/ia64/kernel/unaligned.c (void *) rnat_addr, rnats, nat, (rnats >> ia64_rse_slot_num(addr)) & 1); ia64_rse_slot_num 369 arch/ia64/kernel/unaligned.c nat_mask = 1UL << ia64_rse_slot_num(addr); ia64_rse_slot_num 414 arch/ia64/kernel/unaligned.c nat_mask = 1UL << ia64_rse_slot_num(addr); ia64_rse_slot_num 436 arch/ia64/kernel/unaligned.c nat_mask = 1UL << ia64_rse_slot_num(addr); ia64_rse_slot_num 351 arch/ia64/kernel/unwind.c nat_mask = (1UL << ia64_rse_slot_num(addr)); ia64_rse_slot_num 382 arch/ia64/kernel/unwind.c nat_mask = (1UL << ia64_rse_slot_num(addr));