Lines Matching refs:pa

146 	u32 l, pa;  in omap2_iommu_enable()  local
151 pa = virt_to_phys(obj->iopgd); in omap2_iommu_enable()
152 if (!IS_ALIGNED(pa, SZ_16K)) in omap2_iommu_enable()
159 iommu_write_reg(obj, pa, MMU_TTB); in omap2_iommu_enable()
333 cr->ram = e->pa | e->endian | e->elsz | e->mixed; in iotlb_alloc_cr()
662 static int iopgd_alloc_section(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) in iopgd_alloc_section() argument
666 if ((da | pa) & ~IOSECTION_MASK) { in iopgd_alloc_section()
668 __func__, da, pa, IOSECTION_SIZE); in iopgd_alloc_section()
672 *iopgd = (pa & IOSECTION_MASK) | prot | IOPGD_SECTION; in iopgd_alloc_section()
677 static int iopgd_alloc_super(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) in iopgd_alloc_super() argument
682 if ((da | pa) & ~IOSUPER_MASK) { in iopgd_alloc_super()
684 __func__, da, pa, IOSUPER_SIZE); in iopgd_alloc_super()
689 *(iopgd + i) = (pa & IOSUPER_MASK) | prot | IOPGD_SUPER; in iopgd_alloc_super()
694 static int iopte_alloc_page(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) in iopte_alloc_page() argument
702 *iopte = (pa & IOPAGE_MASK) | prot | IOPTE_SMALL; in iopte_alloc_page()
706 __func__, da, pa, iopte, *iopte); in iopte_alloc_page()
711 static int iopte_alloc_large(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) in iopte_alloc_large() argument
717 if ((da | pa) & ~IOLARGE_MASK) { in iopte_alloc_large()
719 __func__, da, pa, IOLARGE_SIZE); in iopte_alloc_large()
727 *(iopte + i) = (pa & IOLARGE_MASK) | prot | IOPTE_LARGE; in iopte_alloc_large()
764 err = fn(obj, e->da, e->pa, prot); in iopgtable_store_entry_core()
1108 static u32 iotlb_init_entry(struct iotlb_entry *e, u32 da, u32 pa, int pgsz) in iotlb_init_entry() argument
1113 e->pa = pa; in iotlb_init_entry()
1124 phys_addr_t pa, size_t bytes, int prot) in omap_iommu_map() argument
1139 dev_dbg(dev, "mapping da 0x%lx to pa %pa size 0x%x\n", da, &pa, bytes); in omap_iommu_map()
1141 iotlb_init_entry(&e, da, pa, omap_pgsz); in omap_iommu_map()