hpa 286 arch/ia64/hp/common/sba_iommu.c sba_dump_tlb(char *hpa) hpa 288 arch/ia64/hp/common/sba_iommu.c DBG_INIT("IO TLB at 0x%p\n", (void *)hpa); hpa 289 arch/ia64/hp/common/sba_iommu.c DBG_INIT("IOC_IBASE : %016lx\n", READ_REG(hpa+IOC_IBASE)); hpa 290 arch/ia64/hp/common/sba_iommu.c DBG_INIT("IOC_IMASK : %016lx\n", READ_REG(hpa+IOC_IMASK)); hpa 291 arch/ia64/hp/common/sba_iommu.c DBG_INIT("IOC_TCNFG : %016lx\n", READ_REG(hpa+IOC_TCNFG)); hpa 292 arch/ia64/hp/common/sba_iommu.c DBG_INIT("IOC_PDIR_BASE: %016lx\n", READ_REG(hpa+IOC_PDIR_BASE)); hpa 1783 arch/ia64/hp/common/sba_iommu.c static void __init ioc_init(unsigned long hpa, struct ioc *ioc) hpa 1790 arch/ia64/hp/common/sba_iommu.c ioc->ioc_hpa = ioremap(hpa, 0x1000); hpa 1826 arch/ia64/hp/common/sba_iommu.c hpa, ioc->iov_size >> 20, ioc->ibase); hpa 1984 arch/ia64/hp/common/sba_iommu.c u64 hpa, length; hpa 1988 arch/ia64/hp/common/sba_iommu.c status = hp_acpi_csr_space(handle, &hpa, &length); hpa 2001 arch/ia64/hp/common/sba_iommu.c hpa += ZX1_IOC_OFFSET; hpa 2015 arch/ia64/hp/common/sba_iommu.c ioc_init(hpa, ioc); hpa 108 arch/parisc/include/asm/hardware.h extern struct parisc_device *alloc_pa_dev(unsigned long hpa, hpa 8 arch/parisc/include/asm/parisc-device.h struct resource hpa; /* Hard Physical Address */ hpa 36 arch/parisc/include/asm/pdc.h int pdc_iodc_read(unsigned long *actcnt, unsigned long hpa, unsigned int index, hpa 62 arch/parisc/include/asm/pdc.h int pdc_pci_irt_size(unsigned long *num_entries, unsigned long hpa); hpa 63 arch/parisc/include/asm/pdc.h int pdc_pci_irt(unsigned long num_entries, unsigned long hpa, void *tbl); hpa 369 arch/parisc/include/asm/pdcpat.h extern int pdc_pat_cpu_get_number(struct pdc_pat_cpu_num *cpu_info, unsigned long hpa); hpa 88 arch/parisc/include/asm/processor.h unsigned long hpa; /* Host Physical address */ hpa 237 arch/parisc/include/asm/ropes.h static inline int agp_mode_mercury(void __iomem *hpa) { hpa 240 arch/parisc/include/asm/ropes.h bus_mode = readl(hpa + 0x0620); hpa 255 arch/parisc/include/asm/ropes.h extern void *iosapic_register(unsigned long hpa); hpa 381 arch/parisc/include/uapi/asm/pdc.h unsigned int hpa; /* HPA base address */ hpa 629 arch/parisc/include/uapi/asm/pdc.h unsigned long hpa; /* mod's register set address */ hpa 228 arch/parisc/kernel/drivers.c unsigned long hpa; hpa 238 arch/parisc/kernel/drivers.c if (pdev->hpa.start == d->hpa) { hpa 246 arch/parisc/kernel/drivers.c static struct parisc_device *find_device_by_addr(unsigned long hpa) hpa 249 arch/parisc/kernel/drivers.c .hpa = hpa, hpa 494 arch/parisc/kernel/drivers.c alloc_pa_dev(unsigned long hpa, struct hardware_path *mod_path) hpa 503 arch/parisc/kernel/drivers.c if (find_device_by_addr(hpa) != NULL) hpa 506 arch/parisc/kernel/drivers.c status = pdc_iodc_read(&bytecnt, hpa, 0, &iodc_data, 32); hpa 523 arch/parisc/kernel/drivers.c dev->hpa.name = parisc_pathname(dev); hpa 524 arch/parisc/kernel/drivers.c dev->hpa.start = hpa; hpa 529 arch/parisc/kernel/drivers.c if (hpa == 0xf4000000 || hpa == 0xf8000000) { hpa 530 arch/parisc/kernel/drivers.c dev->hpa.end = hpa + 0x03ffffff; hpa 531 arch/parisc/kernel/drivers.c } else if (hpa == 0xf6000000 || hpa == 0xfa000000) { hpa 532 arch/parisc/kernel/drivers.c dev->hpa.end = hpa + 0x01ffffff; hpa 534 arch/parisc/kernel/drivers.c dev->hpa.end = hpa + 0xfff; hpa 536 arch/parisc/kernel/drivers.c dev->hpa.flags = IORESOURCE_MEM; hpa 545 arch/parisc/kernel/drivers.c if ((hpa & 0xfff) == 0 && insert_resource(&iomem_resource, &dev->hpa)) hpa 546 arch/parisc/kernel/drivers.c pr_warn("Unable to claim HPA %lx for device %s\n", hpa, name); hpa 798 arch/parisc/kernel/drivers.c ((gsc_readl(dev->hpa.start + offsetof(struct bc_module, io_status)) \ hpa 807 arch/parisc/kernel/drivers.c #define READ_IO_IO_LOW(dev) (unsigned long)(signed int)gsc_readl(dev->hpa.start + IO_IO_LOW) hpa 808 arch/parisc/kernel/drivers.c #define READ_IO_IO_HIGH(dev) (unsigned long)(signed int)gsc_readl(dev->hpa.start + IO_IO_HIGH) hpa 847 arch/parisc/kernel/drivers.c unsigned long hpa = io_io_low; hpa 852 arch/parisc/kernel/drivers.c for(i = 0; i < MAX_NATIVE_DEVICES; i++, hpa += NATIVE_DEVICE_OFFSET) { hpa 856 arch/parisc/kernel/drivers.c dev = find_device_by_addr(hpa); hpa 859 arch/parisc/kernel/drivers.c dev = alloc_pa_dev(hpa, &path); hpa 868 arch/parisc/kernel/drivers.c } while(!devices_found && hpa < io_io_high); hpa 893 arch/parisc/kernel/drivers.c ++count, dev->name, &(dev->hpa.start), hw_path, dev->id.hw_type, hpa 967 arch/parisc/kernel/drivers.c unsigned long hpa = dev->hpa.start; hpa 969 arch/parisc/kernel/drivers.c pr_cont("\t{\t.hpa = 0x%08lx,\\\n", hpa); hpa 970 arch/parisc/kernel/drivers.c pr_cont("\t\t.iodc = &iodc_data_hpa_%08lx,\\\n", hpa); hpa 971 arch/parisc/kernel/drivers.c pr_cont("\t\t.mod_info = &mod_info_hpa_%08lx,\\\n", hpa); hpa 972 arch/parisc/kernel/drivers.c pr_cont("\t\t.mod_path = &mod_path_hpa_%08lx,\\\n", hpa); hpa 973 arch/parisc/kernel/drivers.c pr_cont("\t\t.num_addr = HPA_%08lx_num_addr,\\\n", hpa); hpa 974 arch/parisc/kernel/drivers.c pr_cont("\t\t.add_addr = { HPA_%08lx_add_addr } },\\\n", hpa); hpa 992 arch/parisc/kernel/drivers.c unsigned long hpa = dev->hpa.start; hpa 1000 arch/parisc/kernel/drivers.c status = pdc_iodc_read(&count, hpa, 0, hpa 1003 arch/parisc/kernel/drivers.c pr_info("No IODC data for hpa 0x%08lx\n", hpa); hpa 1010 arch/parisc/kernel/drivers.c hpa, parisc_hardware_description(&dev->id)); hpa 1016 arch/parisc/kernel/drivers.c } while (status == PDC_OK && pdc_mod_info.mod_addr != hpa); hpa 1019 arch/parisc/kernel/drivers.c " mod_info_hpa_%08lx = {\n", hpa); hpa 1029 arch/parisc/kernel/drivers.c "mod_path_hpa_%08lx = {\n", hpa); hpa 1046 arch/parisc/kernel/drivers.c pr_info("static struct pdc_iodc iodc_data_hpa_%08lx = {\n", hpa); hpa 1067 arch/parisc/kernel/drivers.c pr_info("#define HPA_%08lx_num_addr %d\n", hpa, dev->num_addrs); hpa 1068 arch/parisc/kernel/drivers.c pr_info("#define HPA_%08lx_add_addr ", hpa); hpa 388 arch/parisc/kernel/firmware.c int pdc_iodc_read(unsigned long *actcnt, unsigned long hpa, unsigned int index, hpa 395 arch/parisc/kernel/firmware.c retval = mem_pdc_call(PDC_IODC, PDC_IODC_READ, __pa(pdc_result), hpa, hpa 703 arch/parisc/kernel/firmware.c int pdc_lan_station_id(char *lan_addr, unsigned long hpa) hpa 710 arch/parisc/kernel/firmware.c __pa(pdc_result), hpa); hpa 911 arch/parisc/kernel/firmware.c int pdc_pci_irt_size(unsigned long *num_entries, unsigned long hpa) hpa 918 arch/parisc/kernel/firmware.c __pa(pdc_result), hpa); hpa 935 arch/parisc/kernel/firmware.c int pdc_pci_irt(unsigned long num_entries, unsigned long hpa, void *tbl) hpa 945 arch/parisc/kernel/firmware.c __pa(pdc_result), hpa, __pa(tbl)); hpa 961 arch/parisc/kernel/firmware.c unsigned int pdc_pci_config_read(void *hpa, unsigned long cfg_addr) hpa 970 arch/parisc/kernel/firmware.c __pa(pdc_result), hpa, cfg_addr&~3UL, 4UL); hpa 985 arch/parisc/kernel/firmware.c void pdc_pci_config_write(void *hpa, unsigned long cfg_addr, unsigned int val) hpa 993 arch/parisc/kernel/firmware.c __pa(pdc_result), hpa, hpa 1252 arch/parisc/kernel/firmware.c (unsigned long)PAGE0->mem_cons.hpa, ENTRY_IO_COUT, hpa 1280 arch/parisc/kernel/firmware.c (unsigned long)PAGE0->mem_kbd.hpa, ENTRY_IO_CIN, hpa 1397 arch/parisc/kernel/firmware.c int pdc_pat_cpu_get_number(struct pdc_pat_cpu_num *cpu_info, unsigned long hpa) hpa 1404 arch/parisc/kernel/firmware.c __pa(&pdc_result), hpa); hpa 459 arch/parisc/kernel/inventory.c dev = alloc_pa_dev(r_addr->hpa, &module_path->path); hpa 541 arch/parisc/kernel/irq.c per_cpu(cpu_data, cpu).hpa); hpa 33 arch/parisc/kernel/pa7300lc.c u32 hpa; hpa 38 arch/parisc/kernel/pa7300lc.c hpa = cpu_hpa(); hpa 43 arch/parisc/kernel/pa7300lc.c gsc_readl(hpa+MIOC_CONTROL), gsc_readl(hpa+MIOC_STATUS), hpa 44 arch/parisc/kernel/pa7300lc.c gsc_readl(hpa+MDERRADD), gsc_readl(hpa+DMAERR), hpa 45 arch/parisc/kernel/pa7300lc.c gsc_readl(hpa+DIOERR), gsc_readl(hpa+HIDMAMEM)); hpa 795 arch/parisc/kernel/perf.c runway = ioremap_nocache(cpu_device->hpa.start, 4096); hpa 102 arch/parisc/kernel/processor.c txn_addr = dev->hpa.start; /* for legacy PDC */ hpa 129 arch/parisc/kernel/processor.c status = pdc_pat_cpu_get_number(&cpu_info, dev->hpa.start); hpa 135 arch/parisc/kernel/processor.c &dev->hpa.start); hpa 148 arch/parisc/kernel/processor.c &dev->hpa.start, cpu_info.cpu_num, NR_CPUS); hpa 168 arch/parisc/kernel/processor.c p->hpa = dev->hpa.start; /* save CPU hpa */ hpa 380 arch/parisc/kernel/processor.c if (0 == cpuinfo->hpa) hpa 196 arch/parisc/kernel/smp.c gsc_writel(IPI_IRQ - CPU_IRQ_BASE, p->hpa); hpa 339 arch/parisc/kernel/smp.c printk(KERN_INFO "Releasing cpu %d now, hpa=%lx\n", cpuid, p->hpa); hpa 350 arch/parisc/kernel/smp.c gsc_writel(TIMER_IRQ - CPU_IRQ_BASE, p->hpa); hpa 53 arch/powerpc/include/asm/iommu.h unsigned long *hpa, hpa 211 arch/powerpc/include/asm/iommu.h unsigned long entry, unsigned long *hpa, hpa 215 arch/powerpc/include/asm/iommu.h unsigned long entry, unsigned long *hpa, hpa 631 arch/powerpc/include/asm/kvm_book3s_64.h unsigned long hpa, unsigned long nbytes); hpa 634 arch/powerpc/include/asm/kvm_book3s_64.h unsigned long gpa, unsigned long hpa, hpa 41 arch/powerpc/include/asm/mmu_context.h unsigned long ua, unsigned int pageshift, unsigned long *hpa); hpa 43 arch/powerpc/include/asm/mmu_context.h unsigned long ua, unsigned int pageshift, unsigned long *hpa); hpa 45 arch/powerpc/include/asm/mmu_context.h extern bool mm_iommu_is_devmem(struct mm_struct *mm, unsigned long hpa, hpa 50 arch/powerpc/include/asm/mmu_context.h static inline bool mm_iommu_is_devmem(struct mm_struct *mm, unsigned long hpa, hpa 1018 arch/powerpc/kernel/iommu.c unsigned long entry, unsigned long *hpa, hpa 1024 arch/powerpc/kernel/iommu.c ret = tbl->it_ops->xchg_no_kill(tbl, entry, hpa, direction, false); hpa 1027 arch/powerpc/kernel/iommu.c !mm_iommu_is_devmem(mm, *hpa, tbl->it_page_shift, hpa 1029 arch/powerpc/kernel/iommu.c SetPageDirty(pfn_to_page(*hpa >> PAGE_SHIFT)); hpa 384 arch/powerpc/kvm/book3s_64_mmu_radix.c unsigned long hpa; hpa 407 arch/powerpc/kvm/book3s_64_mmu_radix.c hpa = old & PTE_RPN_MASK; hpa 408 arch/powerpc/kvm/book3s_64_mmu_radix.c kvmhv_remove_nest_rmap_range(kvm, memslot, gpa, hpa, page_size); hpa 368 arch/powerpc/kvm/book3s_64_vio.c unsigned long hpa = 0; hpa 376 arch/powerpc/kvm/book3s_64_vio.c if (mm_iommu_ua_to_hpa(mem, ua, shift, &hpa)) hpa 416 arch/powerpc/kvm/book3s_64_vio.c unsigned long hpa = 0; hpa 419 arch/powerpc/kvm/book3s_64_vio.c iommu_tce_xchg_no_kill(mm, tbl, entry, &hpa, &dir); hpa 447 arch/powerpc/kvm/book3s_64_vio.c unsigned long hpa = 0; hpa 450 arch/powerpc/kvm/book3s_64_vio.c if (WARN_ON_ONCE(iommu_tce_xchg_no_kill(kvm->mm, tbl, entry, &hpa, hpa 459 arch/powerpc/kvm/book3s_64_vio.c iommu_tce_xchg_no_kill(kvm->mm, tbl, entry, &hpa, &dir); hpa 486 arch/powerpc/kvm/book3s_64_vio.c unsigned long hpa; hpa 499 arch/powerpc/kvm/book3s_64_vio.c if (WARN_ON_ONCE(mm_iommu_ua_to_hpa(mem, ua, tbl->it_page_shift, &hpa))) hpa 505 arch/powerpc/kvm/book3s_64_vio.c ret = iommu_tce_xchg_no_kill(kvm->mm, tbl, entry, &hpa, &dir); hpa 124 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long hpa = 0; hpa 132 arch/powerpc/kvm/book3s_64_vio_hv.c if (mm_iommu_ua_to_hpa_rm(mem, ua, shift, &hpa)) hpa 223 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long entry, unsigned long *hpa, hpa 228 arch/powerpc/kvm/book3s_64_vio_hv.c ret = tbl->it_ops->xchg_no_kill(tbl, entry, hpa, direction, true); hpa 254 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long hpa = 0; hpa 257 arch/powerpc/kvm/book3s_64_vio_hv.c iommu_tce_xchg_no_kill_rm(kvm->mm, tbl, entry, &hpa, &dir); hpa 286 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long hpa = 0; hpa 289 arch/powerpc/kvm/book3s_64_vio_hv.c if (iommu_tce_xchg_no_kill_rm(kvm->mm, tbl, entry, &hpa, &dir)) hpa 301 arch/powerpc/kvm/book3s_64_vio_hv.c iommu_tce_xchg_no_kill_rm(kvm->mm, tbl, entry, &hpa, &dir); hpa 328 arch/powerpc/kvm/book3s_64_vio_hv.c unsigned long hpa = 0; hpa 341 arch/powerpc/kvm/book3s_64_vio_hv.c &hpa))) hpa 347 arch/powerpc/kvm/book3s_64_vio_hv.c ret = iommu_tce_xchg_no_kill_rm(kvm->mm, tbl, entry, &hpa, &dir); hpa 793 arch/powerpc/kvm/book3s_hv_nested.c unsigned long hpa, unsigned long mask) hpa 814 arch/powerpc/kvm/book3s_hv_nested.c if (ptep && pte_present(*ptep) && ((pte_val(*ptep) & mask) == hpa)) { hpa 826 arch/powerpc/kvm/book3s_hv_nested.c unsigned long hpa, unsigned long nbytes) hpa 836 arch/powerpc/kvm/book3s_hv_nested.c hpa &= mask; hpa 839 arch/powerpc/kvm/book3s_hv_nested.c kvmhv_update_nest_rmap_rc(kvm, rmap, clr, set, hpa, mask); hpa 843 arch/powerpc/kvm/book3s_hv_nested.c unsigned long hpa, unsigned long mask) hpa 859 arch/powerpc/kvm/book3s_hv_nested.c if (ptep && pte_present(*ptep) && ((pte_val(*ptep) & mask) == hpa)) hpa 864 arch/powerpc/kvm/book3s_hv_nested.c unsigned long hpa, unsigned long mask) hpa 871 arch/powerpc/kvm/book3s_hv_nested.c kvmhv_remove_nest_rmap(kvm, rmap, hpa, mask); hpa 879 arch/powerpc/kvm/book3s_hv_nested.c unsigned long gpa, unsigned long hpa, hpa 891 arch/powerpc/kvm/book3s_hv_nested.c hpa &= addr_mask; hpa 895 arch/powerpc/kvm/book3s_hv_nested.c kvmhv_remove_nest_rmap_list(kvm, rmap, hpa, addr_mask); hpa 892 arch/powerpc/kvm/book3s_hv_rm_mmu.c int writing, unsigned long *hpa, hpa 925 arch/powerpc/kvm/book3s_hv_rm_mmu.c if (hpa) hpa 926 arch/powerpc/kvm/book3s_hv_rm_mmu.c *hpa = pa; hpa 342 arch/powerpc/mm/book3s64/iommu_api.c unsigned long ua, unsigned int pageshift, unsigned long *hpa) hpa 354 arch/powerpc/mm/book3s64/iommu_api.c *hpa = mem->dev_hpa + (ua - mem->ua); hpa 359 arch/powerpc/mm/book3s64/iommu_api.c *hpa = (*va & MM_IOMMU_TABLE_GROUP_PAGE_MASK) | (ua & ~PAGE_MASK); hpa 366 arch/powerpc/mm/book3s64/iommu_api.c unsigned long ua, unsigned int pageshift, unsigned long *hpa) hpa 378 arch/powerpc/mm/book3s64/iommu_api.c *hpa = mem->dev_hpa + (ua - mem->ua); hpa 386 arch/powerpc/mm/book3s64/iommu_api.c *hpa = (*pa & MM_IOMMU_TABLE_GROUP_PAGE_MASK) | (ua & ~PAGE_MASK); hpa 415 arch/powerpc/mm/book3s64/iommu_api.c bool mm_iommu_is_devmem(struct mm_struct *mm, unsigned long hpa, hpa 426 arch/powerpc/mm/book3s64/iommu_api.c if ((mem->dev_hpa <= hpa) && (hpa < end)) { hpa 433 arch/powerpc/mm/book3s64/iommu_api.c *size = min(1UL << pageshift, end - hpa); hpa 120 arch/powerpc/platforms/powernv/pci-ioda-tce.c unsigned long *hpa, enum dma_data_direction *direction, hpa 124 arch/powerpc/platforms/powernv/pci-ioda-tce.c unsigned long newtce = *hpa | proto_tce, oldtce; hpa 128 arch/powerpc/platforms/powernv/pci-ioda-tce.c BUG_ON(*hpa & ~IOMMU_PAGE_MASK(tbl)); hpa 133 arch/powerpc/platforms/powernv/pci-ioda-tce.c *hpa = 0; hpa 148 arch/powerpc/platforms/powernv/pci-ioda-tce.c *hpa = oldtce & ~(TCE_PCI_READ | TCE_PCI_WRITE); hpa 200 arch/powerpc/platforms/powernv/pci-ioda-tce.c unsigned long hpa = be64_to_cpu(tmp[i]); hpa 202 arch/powerpc/platforms/powernv/pci-ioda-tce.c if (!(hpa & (TCE_PCI_READ | TCE_PCI_WRITE))) hpa 205 arch/powerpc/platforms/powernv/pci-ioda-tce.c pnv_pci_ioda2_table_do_free_pages(__va(hpa), size, hpa 1955 arch/powerpc/platforms/powernv/pci-ioda.c unsigned long *hpa, enum dma_data_direction *direction, hpa 1958 arch/powerpc/platforms/powernv/pci-ioda.c return pnv_tce_xchg(tbl, index, hpa, direction, !realmode); hpa 230 arch/powerpc/platforms/powernv/pci.h unsigned long *hpa, enum dma_data_direction *direction, hpa 643 arch/s390/kvm/vsie.c static int pin_guest_page(struct kvm *kvm, gpa_t gpa, hpa_t *hpa) hpa 650 arch/s390/kvm/vsie.c *hpa = (hpa_t) page_to_virt(page) + (gpa & ~PAGE_MASK); hpa 655 arch/s390/kvm/vsie.c static void unpin_guest_page(struct kvm *kvm, gpa_t gpa, hpa_t hpa) hpa 657 arch/s390/kvm/vsie.c kvm_release_pfn_dirty(hpa >> PAGE_SHIFT); hpa 666 arch/s390/kvm/vsie.c hpa_t hpa; hpa 668 arch/s390/kvm/vsie.c hpa = (u64) scb_s->scaoh << 32 | scb_s->scaol; hpa 669 arch/s390/kvm/vsie.c if (hpa) { hpa 670 arch/s390/kvm/vsie.c unpin_guest_page(vcpu->kvm, vsie_page->sca_gpa, hpa); hpa 676 arch/s390/kvm/vsie.c hpa = scb_s->itdba; hpa 677 arch/s390/kvm/vsie.c if (hpa) { hpa 678 arch/s390/kvm/vsie.c unpin_guest_page(vcpu->kvm, vsie_page->itdba_gpa, hpa); hpa 683 arch/s390/kvm/vsie.c hpa = scb_s->gvrd; hpa 684 arch/s390/kvm/vsie.c if (hpa) { hpa 685 arch/s390/kvm/vsie.c unpin_guest_page(vcpu->kvm, vsie_page->gvrd_gpa, hpa); hpa 690 arch/s390/kvm/vsie.c hpa = scb_s->riccbd; hpa 691 arch/s390/kvm/vsie.c if (hpa) { hpa 692 arch/s390/kvm/vsie.c unpin_guest_page(vcpu->kvm, vsie_page->riccbd_gpa, hpa); hpa 697 arch/s390/kvm/vsie.c hpa = scb_s->sdnxo; hpa 698 arch/s390/kvm/vsie.c if (hpa) { hpa 699 arch/s390/kvm/vsie.c unpin_guest_page(vcpu->kvm, vsie_page->sdnx_gpa, hpa); hpa 723 arch/s390/kvm/vsie.c hpa_t hpa; hpa 739 arch/s390/kvm/vsie.c rc = pin_guest_page(vcpu->kvm, gpa, &hpa); hpa 746 arch/s390/kvm/vsie.c scb_s->scaoh = (u32)((u64)hpa >> 32); hpa 747 arch/s390/kvm/vsie.c scb_s->scaol = (u32)(u64)hpa; hpa 757 arch/s390/kvm/vsie.c rc = pin_guest_page(vcpu->kvm, gpa, &hpa); hpa 763 arch/s390/kvm/vsie.c scb_s->itdba = hpa; hpa 776 arch/s390/kvm/vsie.c rc = pin_guest_page(vcpu->kvm, gpa, &hpa); hpa 782 arch/s390/kvm/vsie.c scb_s->gvrd = hpa; hpa 792 arch/s390/kvm/vsie.c rc = pin_guest_page(vcpu->kvm, gpa, &hpa); hpa 799 arch/s390/kvm/vsie.c scb_s->riccbd = hpa; hpa 822 arch/s390/kvm/vsie.c rc = pin_guest_page(vcpu->kvm, gpa, &hpa); hpa 828 arch/s390/kvm/vsie.c scb_s->sdnxo = hpa | sdnxc; hpa 840 arch/s390/kvm/vsie.c hpa_t hpa = (hpa_t) vsie_page->scb_o; hpa 842 arch/s390/kvm/vsie.c if (hpa) hpa 843 arch/s390/kvm/vsie.c unpin_guest_page(vcpu->kvm, gpa, hpa); hpa 856 arch/s390/kvm/vsie.c hpa_t hpa; hpa 859 arch/s390/kvm/vsie.c rc = pin_guest_page(vcpu->kvm, gpa, &hpa); hpa 865 arch/s390/kvm/vsie.c vsie_page->scb_o = (struct kvm_s390_sie_block *) hpa; hpa 366 arch/x86/include/asm/kvm_host.h hpa_t hpa; hpa 370 arch/x86/include/asm/kvm_host.h ((struct kvm_mmu_root_info) { .cr3 = INVALID_PAGE, .hpa = INVALID_PAGE }) hpa 1101 arch/x86/include/asm/kvm_host.h void (*set_apic_access_page_addr)(struct kvm_vcpu *vcpu, hpa_t hpa); hpa 3730 arch/x86/kvm/mmu.c VALID_PAGE(mmu->prev_roots[i].hpa)) hpa 3741 arch/x86/kvm/mmu.c mmu_free_root_page(vcpu->kvm, &mmu->prev_roots[i].hpa, hpa 4370 arch/x86/kvm/mmu.c root.hpa = mmu->root_hpa; hpa 4375 arch/x86/kvm/mmu.c if (new_cr3 == root.cr3 && VALID_PAGE(root.hpa) && hpa 4376 arch/x86/kvm/mmu.c page_header(root.hpa) != NULL && hpa 4377 arch/x86/kvm/mmu.c new_role.word == page_header(root.hpa)->role.word) hpa 4381 arch/x86/kvm/mmu.c mmu->root_hpa = root.hpa; hpa 5633 arch/x86/kvm/mmu.c if (VALID_PAGE(mmu->prev_roots[i].hpa)) hpa 5634 arch/x86/kvm/mmu.c mmu->invlpg(vcpu, gva, mmu->prev_roots[i].hpa); hpa 5653 arch/x86/kvm/mmu.c if (VALID_PAGE(mmu->prev_roots[i].hpa) && hpa 5655 arch/x86/kvm/mmu.c mmu->invlpg(vcpu, gva, mmu->prev_roots[i].hpa); hpa 98 arch/x86/kvm/mmu_audit.c hpa_t hpa; hpa 119 arch/x86/kvm/mmu_audit.c hpa = pfn << PAGE_SHIFT; hpa 120 arch/x86/kvm/mmu_audit.c if ((*sptep & PT64_BASE_ADDR_MASK) != hpa) hpa 123 arch/x86/kvm/mmu_audit.c hpa, *sptep); hpa 2937 arch/x86/kvm/vmx/nested.c u64 hpa; hpa 2953 arch/x86/kvm/vmx/nested.c hpa = page_to_phys(vmx->nested.apic_access_page); hpa 2954 arch/x86/kvm/vmx/nested.c vmcs_write64(APIC_ACCESS_ADDR, hpa); hpa 6077 arch/x86/kvm/vmx/vmx.c static void vmx_set_apic_access_page_addr(struct kvm_vcpu *vcpu, hpa_t hpa) hpa 6080 arch/x86/kvm/vmx/vmx.c vmcs_write64(APIC_ACCESS_ADDR, hpa); hpa 160 drivers/char/agp/hp-agp.c hp_zx1_ioc_init (u64 hpa) hpa 164 drivers/char/agp/hp-agp.c hp->ioc_regs = ioremap(hpa, 1024); hpa 181 drivers/char/agp/hp-agp.c hp_zx1_lba_find_capability (volatile u8 __iomem *hpa, int cap) hpa 187 drivers/char/agp/hp-agp.c status = readw(hpa+PCI_STATUS); hpa 190 drivers/char/agp/hp-agp.c pos = readb(hpa+PCI_CAPABILITY_LIST); hpa 193 drivers/char/agp/hp-agp.c id = readb(hpa+pos+PCI_CAP_LIST_ID); hpa 198 drivers/char/agp/hp-agp.c pos = readb(hpa+pos+PCI_CAP_LIST_NEXT); hpa 204 drivers/char/agp/hp-agp.c hp_zx1_lba_init (u64 hpa) hpa 209 drivers/char/agp/hp-agp.c hp->lba_regs = ioremap(hpa, 256); hpa 19 drivers/char/ipmi/ipmi_si_parisc.c io.addr_data = dev->hpa.start; hpa 306 drivers/input/keyboard/hilkbd.c (void *)dev->hpa.start); hpa 310 drivers/input/keyboard/hilkbd.c hil_base = dev->hpa.start; hpa 331 drivers/input/serio/gscps2.c unsigned long hpa = dev->hpa.start; hpa 339 drivers/input/serio/gscps2.c hpa += GSC_DINO_OFFSET; hpa 352 drivers/input/serio/gscps2.c ps2port->addr = ioremap_nocache(hpa, GSC_STATUS + 4); hpa 374 drivers/input/serio/gscps2.c hpa, ps2port->id); hpa 380 drivers/input/serio/gscps2.c if (!request_mem_region(hpa, GSC_STATUS + 4, ps2port->port.name)) hpa 386 drivers/input/serio/gscps2.c hpa, hpa 401 drivers/input/serio/gscps2.c release_mem_region(dev->hpa.start, GSC_STATUS + 4); hpa 424 drivers/input/serio/gscps2.c release_mem_region(dev->hpa, GSC_STATUS + 4); hpa 946 drivers/input/serio/hp_sdc.c hp_sdc.base_io = d->hpa.start; hpa 947 drivers/input/serio/hp_sdc.c hp_sdc.data_io = d->hpa.start + 0x800; hpa 948 drivers/input/serio/hp_sdc.c hp_sdc.status_io = d->hpa.start + 0x801; hpa 5440 drivers/iommu/intel-iommu.c unsigned long iova, phys_addr_t hpa, hpa 5471 drivers/iommu/intel-iommu.c size = aligned_nrpages(hpa, size); hpa 5473 drivers/iommu/intel-iommu.c hpa >> VTD_PAGE_SHIFT, size, prot); hpa 163 drivers/net/ethernet/i825xx/lasi_82596.c __FILE__, (unsigned long)dev->hpa.start); hpa 168 drivers/net/ethernet/i825xx/lasi_82596.c (unsigned long)dev->hpa.start, dev->irq); hpa 176 drivers/net/ethernet/i825xx/lasi_82596.c netdevice->base_addr = dev->hpa.start; hpa 74 drivers/parisc/asp.c asp.version = gsc_readb(dev->hpa.start + ASP_VER_OFFSET) & 0xf; hpa 76 drivers/parisc/asp.c asp.hpa = ASP_INTERRUPT_ADDR; hpa 79 drivers/parisc/asp.c asp.name, asp.version, (unsigned long)dev->hpa.start); hpa 1529 drivers/parisc/ccio-dma.c (unsigned long)dev->hpa.start); hpa 1537 drivers/parisc/ccio-dma.c ioc->ioc_regs = ioremap_nocache(dev->hpa.start, 4096); hpa 917 drivers/parisc/dino.c unsigned long hpa = dev->hpa.start; hpa 937 drivers/parisc/dino.c printk("%s version %s found at 0x%lx\n", name, version, hpa); hpa 939 drivers/parisc/dino.c if (!request_mem_region(hpa, PAGE_SIZE, name)) { hpa 941 drivers/parisc/dino.c hpa); hpa 949 drivers/parisc/dino.c if (hpa == (unsigned long)CUJO_RAVEN_ADDR) { hpa 951 drivers/parisc/dino.c } else if (hpa == (unsigned long)CUJO_FIREHAWK_ADDR) { hpa 954 drivers/parisc/dino.c printk("Don't recognise Cujo at address 0x%lx, not enabling workaround\n", hpa); hpa 977 drivers/parisc/dino.c dino_dev->hba.base_addr = ioremap_nocache(hpa, 4096); hpa 306 drivers/parisc/eisa.c name, (unsigned long)dev->hpa.start); hpa 76 drivers/parisc/gsc.c irr = gsc_readl(gsc_asic->hpa + OFFSET_IRR); hpa 114 drivers/parisc/gsc.c imr = gsc_readl(irq_dev->hpa + OFFSET_IMR); hpa 116 drivers/parisc/gsc.c gsc_writel(imr, irq_dev->hpa + OFFSET_IMR); hpa 129 drivers/parisc/gsc.c imr = gsc_readl(irq_dev->hpa + OFFSET_IMR); hpa 131 drivers/parisc/gsc.c gsc_writel(imr, irq_dev->hpa + OFFSET_IMR); hpa 214 drivers/parisc/gsc.c res = request_mem_region(gsc_asic->hpa, 0x100000, gsc_asic->name); hpa 222 drivers/parisc/gsc.c if (gsc_readl(gsc_asic->hpa + OFFSET_IMR)) hpa 224 drivers/parisc/gsc.c gsc_readl(gsc_asic->hpa + OFFSET_IMR)); hpa 29 drivers/parisc/gsc.h unsigned long hpa; hpa 26 drivers/parisc/hppb.c unsigned long hpa; hpa 32 drivers/parisc/hppb.c .hpa = 0, hpa 56 drivers/parisc/hppb.c if(card->hpa) { hpa 65 drivers/parisc/hppb.c card->hpa = dev->hpa.start; hpa 69 drivers/parisc/hppb.c card->mmio_region.start = gsc_readl(dev->hpa.start + IO_IO_LOW); hpa 70 drivers/parisc/hppb.c card->mmio_region.end = gsc_readl(dev->hpa.start + IO_IO_HIGH) - 1; hpa 75 drivers/parisc/hppb.c &dev->hpa.start, hpa 158 drivers/parisc/iosapic.c #define COMPARE_IRTE_ADDR(irte, hpa) ((irte)->dest_iosapic_addr == (hpa)) hpa 160 drivers/parisc/iosapic.c #define COMPARE_IRTE_ADDR(irte, hpa) \ hpa 161 drivers/parisc/iosapic.c ((irte)->dest_iosapic_addr == ((hpa) | 0xffffffff00000000ULL)) hpa 901 drivers/parisc/iosapic.c void *iosapic_register(unsigned long hpa) hpa 915 drivers/parisc/iosapic.c if (COMPARE_IRTE_ADDR(irte, hpa)) hpa 920 drivers/parisc/iosapic.c DBG("iosapic_register() ignoring 0x%lx (NOT FOUND)\n", hpa); hpa 930 drivers/parisc/iosapic.c isi->addr = ioremap_nocache(hpa, 4096); hpa 931 drivers/parisc/iosapic.c isi->isi_hpa = hpa; hpa 58 drivers/parisc/lasi.c unsigned long lasi_base = this_lasi->hpa; hpa 174 drivers/parisc/lasi.c lasi->hpa = dev->hpa.start; hpa 177 drivers/parisc/lasi.c lasi->version = gsc_readl(lasi->hpa + LASI_VER) & 0xf; hpa 179 drivers/parisc/lasi.c lasi->name, lasi->version, lasi->hpa); hpa 182 drivers/parisc/lasi.c lasi_led_init(lasi->hpa); hpa 205 drivers/parisc/lasi.c gsc_writel(lasi->eim, lasi->hpa + OFFSET_IAR); hpa 220 drivers/parisc/lasi.c lasi_power_off_hpa = lasi->hpa; hpa 1283 drivers/parisc/lba_pci.c r->start += (rsize + 1) * LBA_NUM(pa_dev->hpa.start); hpa 1479 drivers/parisc/lba_pci.c void __iomem *addr = ioremap_nocache(dev->hpa.start, 4096); hpa 1498 drivers/parisc/lba_pci.c version, func_class & 0xf, (long)dev->hpa.start); hpa 1528 drivers/parisc/lba_pci.c minor, func_class, (long)dev->hpa.start); hpa 1533 drivers/parisc/lba_pci.c (long)dev->hpa.start); hpa 1538 drivers/parisc/lba_pci.c tmp_obj = iosapic_register(dev->hpa.start + LBA_IOSAPIC_BASE); hpa 1696 drivers/parisc/lba_pci.c void __iomem * base_addr = ioremap_nocache(lba->hpa.start, 4096); hpa 153 drivers/parisc/sba_iommu.c sba_dump_ranges(void __iomem *hpa) hpa 155 drivers/parisc/sba_iommu.c DBG_INIT("SBA at 0x%p\n", hpa); hpa 156 drivers/parisc/sba_iommu.c DBG_INIT("IOS_DIST_BASE : %Lx\n", READ_REG64(hpa+IOS_DIST_BASE)); hpa 157 drivers/parisc/sba_iommu.c DBG_INIT("IOS_DIST_MASK : %Lx\n", READ_REG64(hpa+IOS_DIST_MASK)); hpa 158 drivers/parisc/sba_iommu.c DBG_INIT("IOS_DIST_ROUTE : %Lx\n", READ_REG64(hpa+IOS_DIST_ROUTE)); hpa 160 drivers/parisc/sba_iommu.c DBG_INIT("IOS_DIRECT_BASE : %Lx\n", READ_REG64(hpa+IOS_DIRECT_BASE)); hpa 161 drivers/parisc/sba_iommu.c DBG_INIT("IOS_DIRECT_MASK : %Lx\n", READ_REG64(hpa+IOS_DIRECT_MASK)); hpa 162 drivers/parisc/sba_iommu.c DBG_INIT("IOS_DIRECT_ROUTE: %Lx\n", READ_REG64(hpa+IOS_DIRECT_ROUTE)); hpa 171 drivers/parisc/sba_iommu.c static void sba_dump_tlb(void __iomem *hpa) hpa 173 drivers/parisc/sba_iommu.c DBG_INIT("IO TLB at 0x%p\n", hpa); hpa 174 drivers/parisc/sba_iommu.c DBG_INIT("IOC_IBASE : 0x%Lx\n", READ_REG64(hpa+IOC_IBASE)); hpa 175 drivers/parisc/sba_iommu.c DBG_INIT("IOC_IMASK : 0x%Lx\n", READ_REG64(hpa+IOC_IMASK)); hpa 176 drivers/parisc/sba_iommu.c DBG_INIT("IOC_TCNFG : 0x%Lx\n", READ_REG64(hpa+IOC_TCNFG)); hpa 177 drivers/parisc/sba_iommu.c DBG_INIT("IOC_PDIR_BASE: 0x%Lx\n", READ_REG64(hpa+IOC_PDIR_BASE)); hpa 1227 drivers/parisc/sba_iommu.c int rope_num = (lba->hpa.start >> 13) & 0xf; hpa 1516 drivers/parisc/sba_iommu.c return ioremap_nocache(sba_dev->dev->hpa.start + offset, SBA_FUNC_SIZE); hpa 1551 drivers/parisc/sba_iommu.c printk("sba_hw_init(): mem_boot 0x%x 0x%x 0x%x 0x%x\n", PAGE0->mem_boot.hpa, hpa 1886 drivers/parisc/sba_iommu.c void __iomem *sba_addr = ioremap_nocache(dev->hpa.start, SBA_FUNC_SIZE); hpa 1931 drivers/parisc/sba_iommu.c MODULE_NAME, version, (unsigned long long)dev->hpa.start); hpa 50 drivers/parisc/wax.c unsigned long base = wax->hpa; hpa 79 drivers/parisc/wax.c wax->hpa = dev->hpa.start; hpa 82 drivers/parisc/wax.c printk(KERN_INFO "%s at 0x%lx found.\n", wax->name, wax->hpa); hpa 105 drivers/parisc/wax.c gsc_writel(wax->eim, wax->hpa + OFFSET_IAR); hpa 351 drivers/parport/parport_gsc.c (unsigned long long)dev->hpa.start); hpa 355 drivers/parport/parport_gsc.c port = dev->hpa.start + PARPORT_GSC_OFFSET; hpa 89 drivers/scsi/lasi700.c unsigned long base = dev->hpa.start + LASI_SCSI_CORE_OFFSET; hpa 64 drivers/scsi/zalon.c lasi_scsi_clock(void * hpa, int defaultclock) hpa 68 drivers/scsi/zalon.c status = pdc_iodc_read(&pdc_result, hpa, 0, &iodc_data, 32 ); hpa 92 drivers/scsi/zalon.c void __iomem *zalon = ioremap_nocache(dev->hpa.start, 4096); hpa 131 drivers/scsi/zalon.c device.slot.base = dev->hpa.start + GSC_SCSI_ZALON_OFFSET; hpa 44 drivers/tty/serial/8250/8250_gsc.c (unsigned long long)dev->hpa.start); hpa 48 drivers/tty/serial/8250/8250_gsc.c address = dev->hpa.start; hpa 93 drivers/tty/serial/mux.c status = pdc_iodc_read(&bytecnt, dev->hpa.start, 0, iodc_data, 32); hpa 459 drivers/tty/serial/mux.c request_mem_region(dev->hpa.start + MUX_OFFSET, hpa 475 drivers/tty/serial/mux.c port->mapbase = dev->hpa.start + MUX_OFFSET + hpa 509 drivers/tty/serial/mux.c if(mux_ports[i].port.mapbase == dev->hpa.start + MUX_OFFSET) hpa 523 drivers/tty/serial/mux.c release_mem_region(dev->hpa.start + MUX_OFFSET, port_count * MUX_LINE_OFFSET); hpa 18 drivers/vfio/pci/trace.h TP_PROTO(struct pci_dev *pdev, unsigned long hpa, unsigned long ua, hpa 20 drivers/vfio/pci/trace.h TP_ARGS(pdev, hpa, ua, ret), hpa 24 drivers/vfio/pci/trace.h __field(unsigned long, hpa) hpa 31 drivers/vfio/pci/trace.h __entry->hpa = hpa; hpa 36 drivers/vfio/pci/trace.h TP_printk("%s: %lx -> %lx ret=%d", __entry->name, __entry->hpa, hpa 41 drivers/vfio/pci/trace.h TP_PROTO(struct pci_dev *pdev, unsigned long hpa, unsigned long ua, hpa 43 drivers/vfio/pci/trace.h TP_ARGS(pdev, hpa, ua, size, ret), hpa 47 drivers/vfio/pci/trace.h __field(unsigned long, hpa) hpa 55 drivers/vfio/pci/trace.h __entry->hpa = hpa; hpa 61 drivers/vfio/pci/trace.h TP_printk("%s: %lx -> %lx size=%lx ret=%d", __entry->name, __entry->hpa, hpa 66 drivers/vfio/pci/trace.h TP_PROTO(struct pci_dev *pdev, unsigned long hpa, unsigned long ua, hpa 68 drivers/vfio/pci/trace.h TP_ARGS(pdev, hpa, ua, size, ret), hpa 72 drivers/vfio/pci/trace.h __field(unsigned long, hpa) hpa 80 drivers/vfio/pci/trace.h __entry->hpa = hpa; hpa 86 drivers/vfio/pci/trace.h TP_printk("%s: %lx -> %lx size=%lx ret=%d", __entry->name, __entry->hpa, hpa 178 drivers/vfio/vfio_iommu_spapr_tce.c static bool tce_page_is_contained(struct mm_struct *mm, unsigned long hpa, hpa 184 drivers/vfio/vfio_iommu_spapr_tce.c if (mm_iommu_is_devmem(mm, hpa, it_page_shift, &size)) hpa 187 drivers/vfio/vfio_iommu_spapr_tce.c page = pfn_to_page(hpa >> PAGE_SHIFT); hpa 381 drivers/vfio/vfio_iommu_spapr_tce.c unsigned long hpa) hpa 385 drivers/vfio/vfio_iommu_spapr_tce.c page = pfn_to_page(hpa >> PAGE_SHIFT); hpa 414 drivers/vfio/vfio_iommu_spapr_tce.c unsigned long hpa = 0; hpa 421 drivers/vfio/vfio_iommu_spapr_tce.c tbl->it_page_shift, &hpa, &mem); hpa 484 drivers/vfio/vfio_iommu_spapr_tce.c static int tce_iommu_use_page(unsigned long tce, unsigned long *hpa) hpa 494 drivers/vfio/vfio_iommu_spapr_tce.c *hpa = __pa((unsigned long) page_address(page)); hpa 505 drivers/vfio/vfio_iommu_spapr_tce.c unsigned long hpa; hpa 511 drivers/vfio/vfio_iommu_spapr_tce.c ret = tce_iommu_use_page(tce, &hpa); hpa 515 drivers/vfio/vfio_iommu_spapr_tce.c if (!tce_page_is_contained(container->mm, hpa, hpa 521 drivers/vfio/vfio_iommu_spapr_tce.c hpa |= offset; hpa 524 drivers/vfio/vfio_iommu_spapr_tce.c &hpa, &dirtmp); hpa 526 drivers/vfio/vfio_iommu_spapr_tce.c tce_iommu_unuse_page(container, hpa); hpa 534 drivers/vfio/vfio_iommu_spapr_tce.c tce_iommu_unuse_page(container, hpa); hpa 553 drivers/vfio/vfio_iommu_spapr_tce.c unsigned long hpa; hpa 561 drivers/vfio/vfio_iommu_spapr_tce.c tce, tbl->it_page_shift, &hpa, &mem); hpa 565 drivers/vfio/vfio_iommu_spapr_tce.c if (!tce_page_is_contained(container->mm, hpa, hpa 572 drivers/vfio/vfio_iommu_spapr_tce.c hpa |= tce & IOMMU_PAGE_MASK(tbl) & ~PAGE_MASK; hpa 580 drivers/vfio/vfio_iommu_spapr_tce.c &hpa, &dirtmp); hpa 416 drivers/video/console/sticore.c unsigned long hpa) hpa 458 drivers/video/console/sticore.c newhpa = (i == 0) ? rom_address : hpa; hpa 818 drivers/video/console/sticore.c unsigned long hpa, hpa 886 drivers/video/console/sticore.c address, hpa)); hpa 892 drivers/video/console/sticore.c address, hpa)); hpa 899 drivers/video/console/sticore.c if (sti_init_glob_cfg(sti, address, hpa)) hpa 947 drivers/video/console/sticore.c int hpa = dev->hpa.start; hpa 950 drivers/video/console/sticore.c sti = sti_try_rom_generic(dev->addr[0], hpa, NULL); hpa 952 drivers/video/console/sticore.c sti = sti_try_rom_generic(hpa, hpa, NULL); hpa 954 drivers/video/console/sticore.c sti = sti_try_rom_generic(PAGE0->proc_sti, hpa, NULL); hpa 77 drivers/video/fbdev/sticore.h #define REGION_OFFSET_TO_PHYS( rt, hpa ) \ hpa 78 drivers/video/fbdev/sticore.h (((rt).region_desc.offset << 12) + (hpa)) hpa 906 sound/parisc/harmony.c h->hpa = padev->hpa.start; hpa 910 sound/parisc/harmony.c h->iobase = ioremap_nocache(padev->hpa.start, HARMONY_SIZE); hpa 913 sound/parisc/harmony.c (unsigned long)padev->hpa.start); hpa 970 sound/parisc/harmony.c card->shortname, h->hpa, h->irq); hpa 20 sound/parisc/harmony.h unsigned long hpa; /* hard physical address */