Lines Matching refs:pa

149 	u32 l, pa;  in omap2_iommu_enable()  local
154 pa = virt_to_phys(obj->iopgd); in omap2_iommu_enable()
155 if (!IS_ALIGNED(pa, SZ_16K)) in omap2_iommu_enable()
162 iommu_write_reg(obj, pa, MMU_TTB); in omap2_iommu_enable()
330 cr->ram = e->pa | e->endian | e->elsz | e->mixed; in iotlb_alloc_cr()
536 static int iopgd_alloc_section(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) in iopgd_alloc_section() argument
540 if ((da | pa) & ~IOSECTION_MASK) { in iopgd_alloc_section()
542 __func__, da, pa, IOSECTION_SIZE); in iopgd_alloc_section()
546 *iopgd = (pa & IOSECTION_MASK) | prot | IOPGD_SECTION; in iopgd_alloc_section()
551 static int iopgd_alloc_super(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) in iopgd_alloc_super() argument
556 if ((da | pa) & ~IOSUPER_MASK) { in iopgd_alloc_super()
558 __func__, da, pa, IOSUPER_SIZE); in iopgd_alloc_super()
563 *(iopgd + i) = (pa & IOSUPER_MASK) | prot | IOPGD_SUPER; in iopgd_alloc_super()
568 static int iopte_alloc_page(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) in iopte_alloc_page() argument
576 *iopte = (pa & IOPAGE_MASK) | prot | IOPTE_SMALL; in iopte_alloc_page()
580 __func__, da, pa, iopte, *iopte); in iopte_alloc_page()
585 static int iopte_alloc_large(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) in iopte_alloc_large() argument
591 if ((da | pa) & ~IOLARGE_MASK) { in iopte_alloc_large()
593 __func__, da, pa, IOLARGE_SIZE); in iopte_alloc_large()
601 *(iopte + i) = (pa & IOLARGE_MASK) | prot | IOPTE_LARGE; in iopte_alloc_large()
638 err = fn(obj, e->da, e->pa, prot); in iopgtable_store_entry_core()
1021 static u32 iotlb_init_entry(struct iotlb_entry *e, u32 da, u32 pa, int pgsz) in iotlb_init_entry() argument
1026 e->pa = pa; in iotlb_init_entry()
1037 phys_addr_t pa, size_t bytes, int prot) in omap_iommu_map() argument
1052 dev_dbg(dev, "mapping da 0x%lx to pa %pa size 0x%x\n", da, &pa, bytes); in omap_iommu_map()
1054 iotlb_init_entry(&e, da, pa, omap_pgsz); in omap_iommu_map()