va_bits 26 arch/arm64/include/asm/pgtable-hwdef.h #define ARM64_HW_PGTABLE_LEVELS(va_bits) (((va_bits) - 4) / (PAGE_SHIFT - 3)) va_bits 769 arch/powerpc/include/asm/book3s/64/mmu-hash.h unsigned long va_bits = VA_BITS; va_bits 780 arch/powerpc/include/asm/book3s/64/mmu-hash.h va_bits = 65; va_bits 783 arch/powerpc/include/asm/book3s/64/mmu-hash.h vsid_bits = va_bits - SID_SHIFT; va_bits 789 arch/powerpc/include/asm/book3s/64/mmu-hash.h vsid_bits = va_bits - SID_SHIFT_1T; va_bits 1915 arch/powerpc/platforms/pseries/lpar.c unsigned long va_bits = VA_BITS; va_bits 1920 arch/powerpc/platforms/pseries/lpar.c va_bits = 65; va_bits 1924 arch/powerpc/platforms/pseries/lpar.c vsid_modulus = ((1UL << (va_bits - SID_SHIFT)) - 1); va_bits 1927 arch/powerpc/platforms/pseries/lpar.c vsid_modulus = ((1UL << (va_bits - SID_SHIFT_1T)) - 1); va_bits 713 arch/x86/kvm/emulate.c u8 va_bits; va_bits 720 arch/x86/kvm/emulate.c va_bits = ctxt_virt_addr_bits(ctxt); va_bits 721 arch/x86/kvm/emulate.c if (get_canonical(la, va_bits) != la) va_bits 724 arch/x86/kvm/emulate.c *max_size = min_t(u64, ~0u, (1ull << va_bits) - la); va_bits 746 drivers/iommu/io-pgtable-arm.c unsigned long va_bits, pgd_bits; va_bits 772 drivers/iommu/io-pgtable-arm.c va_bits = cfg->ias - data->pg_shift; va_bits 773 drivers/iommu/io-pgtable-arm.c data->levels = DIV_ROUND_UP(va_bits, data->bits_per_level); va_bits 776 drivers/iommu/io-pgtable-arm.c pgd_bits = va_bits - (data->bits_per_level * (data->levels - 1)); va_bits 329 tools/testing/selftests/kvm/include/x86_64/processor.h void kvm_get_cpu_address_width(unsigned int *pa_bits, unsigned int *va_bits); va_bits 27 tools/testing/selftests/kvm/lib/aarch64/processor.c uint64_t mask = (1UL << (vm->va_bits - shift)) - 1; va_bits 62 tools/testing/selftests/kvm/lib/aarch64/processor.c uint64_t mask = ((1UL << (vm->va_bits - vm->page_shift)) - 1) << vm->page_shift; va_bits 69 tools/testing/selftests/kvm/lib/aarch64/processor.c return 1 << (vm->va_bits - shift); va_bits 297 tools/testing/selftests/kvm/lib/aarch64/processor.c tcr_el1 |= (64 - vm->va_bits) /* T0SZ */; va_bits 54 tools/testing/selftests/kvm/lib/aarch64/ucall.c bits = vm->va_bits - 1; va_bits 152 tools/testing/selftests/kvm/lib/kvm_util.c vm->va_bits = 48; va_bits 159 tools/testing/selftests/kvm/lib/kvm_util.c vm->va_bits = 48; va_bits 166 tools/testing/selftests/kvm/lib/kvm_util.c vm->va_bits = 48; va_bits 173 tools/testing/selftests/kvm/lib/kvm_util.c vm->va_bits = 48; va_bits 180 tools/testing/selftests/kvm/lib/kvm_util.c vm->va_bits = 48; va_bits 187 tools/testing/selftests/kvm/lib/kvm_util.c vm->va_bits = 48; va_bits 193 tools/testing/selftests/kvm/lib/kvm_util.c kvm_get_cpu_address_width(&vm->pa_bits, &vm->va_bits); va_bits 194 tools/testing/selftests/kvm/lib/kvm_util.c TEST_ASSERT(vm->va_bits == 48, "Linear address width " va_bits 195 tools/testing/selftests/kvm/lib/kvm_util.c "(%d bits) not supported", vm->va_bits); va_bits 220 tools/testing/selftests/kvm/lib/kvm_util.c 0, (1ULL << (vm->va_bits - 1)) >> vm->page_shift); va_bits 222 tools/testing/selftests/kvm/lib/kvm_util.c (~((1ULL << (vm->va_bits - 1)) - 1)) >> vm->page_shift, va_bits 223 tools/testing/selftests/kvm/lib/kvm_util.c (1ULL << (vm->va_bits - 1)) >> vm->page_shift); va_bits 53 tools/testing/selftests/kvm/lib/kvm_util_internal.h unsigned int va_bits; va_bits 1166 tools/testing/selftests/kvm/lib/x86_64/processor.c void kvm_get_cpu_address_width(unsigned int *pa_bits, unsigned int *va_bits) va_bits 1175 tools/testing/selftests/kvm/lib/x86_64/processor.c *va_bits = 32; va_bits 1179 tools/testing/selftests/kvm/lib/x86_64/processor.c *va_bits = (entry->eax >> 8) & 0xff;