Searched refs:SID_SHIFT (Results 1 - 16 of 16) sorted by relevance
/linux-4.4.14/arch/powerpc/kvm/ |
H A D | book3s_32_mmu.c | 48 #ifndef SID_SHIFT 49 #define SID_SHIFT 28 macro 101 kvmppc_mmu_book3s_32_esid_to_vsid(vcpu, eaddr >> SID_SHIFT, &vsid); kvmppc_mmu_book3s_32_ea_to_vp() 175 eaddr >> SID_SHIFT, &vsid); kvmppc_mmu_book3s_32_xlate_bat() 350 kvmppc_mmu_map_segment(vcpu, srnum << SID_SHIFT); kvmppc_mmu_book3s_32_mtsrin() 366 ulong ea = esid << SID_SHIFT; kvmppc_mmu_book3s_32_esid_to_vsid()
|
H A D | book3s_64_mmu.c | 82 return slbe->tb ? SID_SHIFT_1T : SID_SHIFT; kvmppc_slb_sid_shift() 427 kvmppc_mmu_map_segment(vcpu, esid << SID_SHIFT); kvmppc_mmu_book3s_64_slbmte() 573 (mp_ea >> SID_SHIFT) == esid; segment_contains_magic_page() 580 ulong ea = esid << SID_SHIFT; kvmppc_mmu_book3s_64_esid_to_vsid() 593 gvsid <<= SID_SHIFT_1T - SID_SHIFT; kvmppc_mmu_book3s_64_esid_to_vsid() 594 gvsid |= esid & ((1ul << (SID_SHIFT_1T - SID_SHIFT)) - 1); kvmppc_mmu_book3s_64_esid_to_vsid() 642 unlikely(esid == (mp_ea >> SID_SHIFT)) && kvmppc_mmu_book3s_64_esid_to_vsid()
|
H A D | book3s_32_mmu_host.c | 170 vcpu->arch.mmu.esid_to_vsid(vcpu, orig_pte->eaddr >> SID_SHIFT, &vsid); kvmppc_mmu_map_page() 179 vpn = (vsid << (SID_SHIFT - VPN_SHIFT)) | kvmppc_mmu_map_page() 316 u32 esid = eaddr >> SID_SHIFT; kvmppc_mmu_map_segment()
|
H A D | book3s_64_mmu_host.c | 118 vcpu->arch.mmu.esid_to_vsid(vcpu, orig_pte->eaddr >> SID_SHIFT, &vsid); kvmppc_mmu_map_page() 222 vcpu->arch.mmu.esid_to_vsid(vcpu, pte->eaddr >> SID_SHIFT, &vsid); kvmppc_mmu_unmap_page() 312 u64 esid = eaddr >> SID_SHIFT; kvmppc_mmu_map_segment()
|
H A D | book3s_pr.c | 563 pte.vpage |= ((u64)VSID_REAL << (SID_SHIFT - 12)); kvmppc_handle_pagefault() 572 vcpu->arch.mmu.esid_to_vsid(vcpu, eaddr >> SID_SHIFT, &vsid); kvmppc_handle_pagefault() 575 pte.vpage |= ((u64)VSID_REAL_DR << (SID_SHIFT - 12)); kvmppc_handle_pagefault() 577 pte.vpage |= ((u64)VSID_REAL_IR << (SID_SHIFT - 12)); kvmppc_handle_pagefault() 914 sr = svcpu->sr[kvmppc_get_pc(vcpu) >> SID_SHIFT]; kvmppc_handle_exit_pr() 962 sr = svcpu->sr[dar >> SID_SHIFT]; kvmppc_handle_exit_pr()
|
/linux-4.4.14/arch/powerpc/include/asm/ |
H A D | kvm_book3s_32.h | 42 #define SID_SHIFT 28 macro
|
H A D | page_64.h | 29 #define SID_SHIFT 28 macro 32 #define GET_ESID(x) (((x) >> SID_SHIFT) & SID_MASK)
|
H A D | mmu-hash64.h | 205 return SID_SHIFT; segment_shift() 309 mask = (1ul << (SID_SHIFT - VPN_SHIFT)) - 1; hpt_hash() 310 hash = (vpn >> (SID_SHIFT - VPN_SHIFT)) ^ hpt_hash() 442 #define USER_VSID_RANGE (1UL << (ESID_BITS + SID_SHIFT)) 586 | (ea >> SID_SHIFT), 256M); get_vsid()
|
H A D | kvm_book3s_64.h | 176 * right shift it with (SID_SHIFT - (23 - 7)) compute_tlbie_rb() 179 va_low ^= v >> (SID_SHIFT - 16); compute_tlbie_rb()
|
/linux-4.4.14/arch/powerpc/mm/ |
H A D | hash_low_64.S | 116 sldi r29,r5,SID_SHIFT - VPN_SHIFT 117 rldicl r28,r3,64 - VPN_SHIFT,64 - (SID_SHIFT - VPN_SHIFT) 417 sldi r29,r5,SID_SHIFT - VPN_SHIFT 419 * clrldi r3,r3,64 - SID_SHIFT --> ea & 0xfffffff 422 rldicl r28,r3,64 - VPN_SHIFT,64 - (SID_SHIFT - VPN_SHIFT) 778 sldi r29,r5,SID_SHIFT - VPN_SHIFT 779 rldicl r28,r3,64 - VPN_SHIFT,64 - (SID_SHIFT - VPN_SHIFT)
|
H A D | slb_low.S | 42 srdi r10,r3,SID_SHIFT /* get esid */ 307 srdi r10,r10,(SID_SHIFT_1T - SID_SHIFT) /* get 1T ESID */
|
H A D | slb.c | 215 << SID_SHIFT; /* EA */ switch_slb()
|
H A D | hash_native_64.c | 561 *vpn = vsid << (SID_SHIFT - VPN_SHIFT) | seg_off >> VPN_SHIFT; hpte_decode()
|
H A D | pgtable_64.c | 67 #if TASK_SIZE_USER64 > (1UL << (ESID_BITS + SID_SHIFT))
|
/linux-4.4.14/drivers/misc/cxl/ |
H A D | fault.c | 46 hash = (slb->esid >> SID_SHIFT) & mask; find_free_sste()
|
/linux-4.4.14/arch/powerpc/kernel/ |
H A D | setup_64.c | 601 return 1UL << SID_SHIFT; safe_stack_limit()
|
Completed in 301 milliseconds