pa 126 arch/alpha/include/asm/hwrpb.h unsigned long pa; pa 19 arch/alpha/include/asm/mmzone.h #define alpha_pa_to_nid(pa) \ pa 21 arch/alpha/include/asm/mmzone.h ? alpha_mv.pa_to_nid(pa) \ pa 32 arch/alpha/include/asm/mmzone.h #define pa_to_nid(pa) alpha_pa_to_nid(pa) pa 857 arch/alpha/kernel/core_marvel.c marvel_pa_to_nid(unsigned long pa) pa 861 arch/alpha/kernel/core_marvel.c if ((pa >> 43) & 1) /* I/O */ pa 862 arch/alpha/kernel/core_marvel.c cpuid = (~(pa >> 35) & 0xff); pa 864 arch/alpha/kernel/core_marvel.c cpuid = ((pa >> 34) & 0x3) | ((pa >> (37 - 2)) & (0x1f << 2)); pa 878 arch/alpha/kernel/core_marvel.c unsigned long pa; pa 880 arch/alpha/kernel/core_marvel.c pa = (nid & 0x3) | ((nid & (0x1f << 2)) << 1); pa 881 arch/alpha/kernel/core_marvel.c pa <<= 34; pa 883 arch/alpha/kernel/core_marvel.c return pa; pa 447 arch/alpha/kernel/core_wildfire.c int wildfire_pa_to_nid(unsigned long pa) pa 449 arch/alpha/kernel/core_wildfire.c return pa >> 36; pa 211 arch/alpha/mm/init.c unsigned long pfn = crb->map[i].pa >> PAGE_SHIFT; pa 58 arch/arm/include/asm/kvm_mmu.h phys_addr_t pa, unsigned long size, bool writable); pa 21 arch/arm/include/asm/smp_scu.h unsigned long pa; pa 23 arch/arm/include/asm/smp_scu.h asm("mrc p15, 4, %0, c15, c0, 0" : "=r" (pa)); pa 25 arch/arm/include/asm/smp_scu.h return pa; pa 31 arch/arm/mach-davinci/include/mach/hardware.h #define IO_ADDRESS(pa) IOMEM(__IO_ADDRESS(pa)) pa 971 arch/arm/mach-omap1/clock.c struct clk *pa; pa 978 arch/arm/mach-omap1/clock.c pa = c->parent; pa 980 arch/arm/mach-omap1/clock.c c->name, pa ? pa->name : "none", c->rate, pa 993 arch/arm/mach-omap1/clock.c struct clk *pa = c->parent; pa 995 arch/arm/mach-omap1/clock.c d = debugfs_create_dir(c->name, pa ? pa->dent : clk_debugfs_root); pa 1005 arch/arm/mach-omap1/clock.c struct clk *pa = c->parent; pa 1007 arch/arm/mach-omap1/clock.c if (pa && !pa->dent) pa 1008 arch/arm/mach-omap1/clock.c clk_debugfs_register(pa); pa 47 arch/arm/mach-omap1/include/mach/hardware.h extern u8 omap_readb(u32 pa); pa 48 arch/arm/mach-omap1/include/mach/hardware.h extern u16 omap_readw(u32 pa); pa 49 arch/arm/mach-omap1/include/mach/hardware.h extern u32 omap_readl(u32 pa); pa 50 arch/arm/mach-omap1/include/mach/hardware.h extern void omap_writeb(u8 v, u32 pa); pa 51 arch/arm/mach-omap1/include/mach/hardware.h extern void omap_writew(u16 v, u32 pa); pa 52 arch/arm/mach-omap1/include/mach/hardware.h extern void omap_writel(u32 v, u32 pa); pa 76 arch/arm/mach-omap1/include/mach/hardware.h #define OMAP1_IO_ADDRESS(pa) IOMEM((pa) - OMAP1_IO_OFFSET) pa 145 arch/arm/mach-omap1/io.c u8 omap_readb(u32 pa) pa 147 arch/arm/mach-omap1/io.c return __raw_readb(OMAP1_IO_ADDRESS(pa)); pa 151 arch/arm/mach-omap1/io.c u16 omap_readw(u32 pa) pa 153 arch/arm/mach-omap1/io.c return __raw_readw(OMAP1_IO_ADDRESS(pa)); pa 157 arch/arm/mach-omap1/io.c u32 omap_readl(u32 pa) pa 159 arch/arm/mach-omap1/io.c return __raw_readl(OMAP1_IO_ADDRESS(pa)); pa 163 arch/arm/mach-omap1/io.c void omap_writeb(u8 v, u32 pa) pa 165 arch/arm/mach-omap1/io.c __raw_writeb(v, OMAP1_IO_ADDRESS(pa)); pa 169 arch/arm/mach-omap1/io.c void omap_writew(u16 v, u32 pa) pa 171 arch/arm/mach-omap1/io.c __raw_writew(v, OMAP1_IO_ADDRESS(pa)); pa 175 arch/arm/mach-omap1/io.c void omap_writel(u32 v, u32 pa) pa 177 arch/arm/mach-omap1/io.c __raw_writel(v, OMAP1_IO_ADDRESS(pa)); pa 349 arch/arm/mach-omap2/cm33xx.c return cm_base.pa + inst + offset; pa 348 arch/arm/mach-omap2/cm_common.c mem->pa = res.start + data->offset; pa 478 arch/arm/mach-omap2/cminst44xx.c return _cm_bases[part].pa + inst + offset; pa 34 arch/arm/mach-omap2/iomap.h #define OMAP2_L3_IO_ADDRESS(pa) IOMEM((pa) + OMAP2_L3_IO_OFFSET) /* L3 */ pa 37 arch/arm/mach-omap2/iomap.h #define OMAP2_L4_IO_ADDRESS(pa) IOMEM((pa) + OMAP2_L4_IO_OFFSET) /* L4 */ pa 40 arch/arm/mach-omap2/iomap.h #define OMAP4_L3_IO_ADDRESS(pa) IOMEM((pa) + OMAP4_L3_IO_OFFSET) /* L3 */ pa 43 arch/arm/mach-omap2/iomap.h #define AM33XX_L4_WK_IO_ADDRESS(pa) IOMEM((pa) + AM33XX_L4_WK_IO_OFFSET) pa 46 arch/arm/mach-omap2/iomap.h #define OMAP4_L3_PER_IO_ADDRESS(pa) IOMEM((pa) + OMAP4_L3_PER_IO_OFFSET) pa 49 arch/arm/mach-omap2/iomap.h #define OMAP2_EMU_IO_ADDRESS(pa) IOMEM((pa) + OMAP2_EMU_IO_OFFSET) pa 526 arch/arm/mach-omap2/prcm-common.h u32 pa; pa 762 arch/arm/mach-omap2/prm_common.c prm_base.pa = res.start + data->offset; pa 66 arch/arm/mm/cache-xsc3l2.c static inline unsigned long l2_map_va(unsigned long pa, unsigned long prev_va) pa 70 arch/arm/mm/cache-xsc3l2.c unsigned long pa_offset = pa << (32 - PAGE_SHIFT); pa 78 arch/arm/mm/cache-xsc3l2.c va = (unsigned long)kmap_atomic_pfn(pa >> PAGE_SHIFT); pa 82 arch/arm/mm/cache-xsc3l2.c return __phys_to_virt(pa); pa 159 arch/arm64/include/asm/kvm_mmu.h phys_addr_t pa, unsigned long size, bool writable); pa 51 arch/arm64/kernel/acpi_numa.c struct acpi_srat_gicc_affinity *pa; pa 57 arch/arm64/kernel/acpi_numa.c pa = (struct acpi_srat_gicc_affinity *)header; pa 58 arch/arm64/kernel/acpi_numa.c if (!pa) pa 61 arch/arm64/kernel/acpi_numa.c if (!(pa->flags & ACPI_SRAT_GICC_ENABLED)) pa 64 arch/arm64/kernel/acpi_numa.c pxm = pa->proximity_domain; pa 73 arch/arm64/kernel/acpi_numa.c cpu = get_cpu_for_acpi_id(pa->acpi_processor_uid); pa 92 arch/arm64/kernel/acpi_numa.c void __init acpi_numa_gicc_affinity_init(struct acpi_srat_gicc_affinity *pa) pa 99 arch/arm64/kernel/acpi_numa.c if (pa->header.length < sizeof(struct acpi_srat_gicc_affinity)) { pa 101 arch/arm64/kernel/acpi_numa.c pa->header.length); pa 106 arch/arm64/kernel/acpi_numa.c if (!(pa->flags & ACPI_SRAT_GICC_ENABLED)) pa 109 arch/arm64/kernel/acpi_numa.c pxm = pa->proximity_domain; pa 376 arch/arm64/mm/mmu.c phys_addr_t pa = __pgd_pgtable_alloc(shift); pa 387 arch/arm64/mm/mmu.c BUG_ON(!pgtable_pte_page_ctor(phys_to_page(pa))); pa 389 arch/arm64/mm/mmu.c BUG_ON(!pgtable_pmd_page_ctor(phys_to_page(pa))); pa 391 arch/arm64/mm/mmu.c return pa; pa 333 arch/ia64/kernel/acpi.c get_processor_proximity_domain(struct acpi_srat_cpu_affinity *pa) pa 337 arch/ia64/kernel/acpi.c pxm = pa->proximity_domain_lo; pa 339 arch/ia64/kernel/acpi.c pxm += pa->proximity_domain_hi[0] << 8; pa 375 arch/ia64/kernel/acpi.c acpi_numa_processor_affinity_init(struct acpi_srat_cpu_affinity *pa) pa 379 arch/ia64/kernel/acpi.c if (!(pa->flags & ACPI_SRAT_CPU_ENABLED)) pa 388 arch/ia64/kernel/acpi.c pxm = get_processor_proximity_domain(pa); pa 394 arch/ia64/kernel/acpi.c (pa->apic_id << 8) | (pa->local_sapic_eid); pa 41 arch/ia64/kernel/brl_emu.c #define unimplemented_physical_address(pa) ( \ pa 42 arch/ia64/kernel/brl_emu.c ((pa) & local_cpu_data->unimpl_pa_mask) != 0 \ pa 148 arch/m68k/coldfire/intc-2.c u16 pa, tb; pa 170 arch/m68k/coldfire/intc-2.c pa = __raw_readw(MCFEPORT_EPPAR); pa 171 arch/m68k/coldfire/intc-2.c pa = (pa & ~(0x3 << (irq * 2))) | (tb << (irq * 2)); pa 172 arch/m68k/coldfire/intc-2.c __raw_writew(pa, MCFEPORT_EPPAR); pa 132 arch/m68k/coldfire/intc-simr.c u16 pa, tb; pa 154 arch/m68k/coldfire/intc-simr.c pa = __raw_readw(MCFEPORT_EPPAR); pa 155 arch/m68k/coldfire/intc-simr.c pa = (pa & ~(0x3 << ebit)) | (tb << ebit); pa 156 arch/m68k/coldfire/intc-simr.c __raw_writew(pa, MCFEPORT_EPPAR); pa 15 arch/m68k/include/asm/sun3xprom.h unsigned long sun3x_prom_ptov(unsigned long pa, unsigned long size); pa 531 arch/microblaze/include/asm/pgtable.h int map_page(unsigned long va, phys_addr_t pa, int flags); pa 135 arch/microblaze/mm/pgtable.c int map_page(unsigned long va, phys_addr_t pa, int flags) pa 148 arch/microblaze/mm/pgtable.c set_pte_at(&init_mm, va, pg, pfn_pte(pa >> PAGE_SHIFT, pa 214 arch/microblaze/mm/pgtable.c unsigned long pa; pa 227 arch/microblaze/mm/pgtable.c pa = 0; pa 229 arch/microblaze/mm/pgtable.c pa = (pte_val(*pte) & PAGE_MASK) | (addr & ~PAGE_MASK); pa 231 arch/microblaze/mm/pgtable.c return pa; pa 43 arch/mips/bmips/dma.c dma_addr_t __phys_to_dma(struct device *dev, phys_addr_t pa) pa 48 arch/mips/bmips/dma.c if (pa >= r->child_addr && pa 49 arch/mips/bmips/dma.c pa < (r->child_addr + r->size)) pa 50 arch/mips/bmips/dma.c return pa - r->child_addr + r->parent_addr; pa 52 arch/mips/bmips/dma.c return pa; pa 135 arch/mips/include/asm/octeon/cvmx-address.h uint64_t pa:49; /* physical address */ pa 224 arch/mips/include/asm/octeon/cvmx-address.h uint64_t pa:49; pa 119 arch/mips/kernel/spram.c unsigned int firstpa = 0, lastpa = 0, pa = 0; pa 143 arch/mips/kernel/spram.c if ((pa == firstpa && size == firstsize) || pa 144 arch/mips/kernel/spram.c (pa == lastpa && size == lastsize)) pa 159 arch/mips/kernel/spram.c pa = tag0 & SPRAM_TAG0_PA_MASK; pa 163 arch/mips/kernel/spram.c firstpa = pa; pa 167 arch/mips/kernel/spram.c lastpa = pa; pa 171 arch/mips/kernel/spram.c unsigned int *vp = (unsigned int *)(CKSEG1 | pa); pa 190 arch/mips/kernel/spram.c type, i, pa, size, enabled ? ",enabled" : ""); pa 309 arch/mips/kvm/tlb.c unsigned long entrylo[2], pagemask, pagemaskbit, pa; pa 364 arch/mips/kvm/tlb.c pa = entrylo[!!(gva & pagemaskbit)]; pa 370 arch/mips/kvm/tlb.c if (!(pa & ENTRYLO_V)) pa 377 arch/mips/kvm/tlb.c pa = (pa << 6) & ~0xfffl; pa 378 arch/mips/kvm/tlb.c pa |= gva & ~(pagemask | pagemaskbit); pa 380 arch/mips/kvm/tlb.c *gpa = pa; pa 723 arch/mips/kvm/vz.c unsigned long mask, pa; pa 765 arch/mips/kvm/vz.c pa = (segctl << 20) & mask; pa 766 arch/mips/kvm/vz.c pa |= gva32 & ~mask; pa 767 arch/mips/kvm/vz.c *gpa = pa; pa 78 arch/mips/lib/dump_tlb.c unsigned long long entrylo0, entrylo1, pa; pa 159 arch/mips/lib/dump_tlb.c pa = entrylo0 & ~(MIPS_ENTRYLO_RI | MIPS_ENTRYLO_XI); pa 161 arch/mips/lib/dump_tlb.c pa |= (unsigned long long)readx_c0_entrylo0() << 30; pa 162 arch/mips/lib/dump_tlb.c pa = (pa << 6) & PAGE_MASK; pa 169 arch/mips/lib/dump_tlb.c pwidth, pa, c0, pa 174 arch/mips/lib/dump_tlb.c pa = entrylo1 & ~(MIPS_ENTRYLO_RI | MIPS_ENTRYLO_XI); pa 176 arch/mips/lib/dump_tlb.c pa |= (unsigned long long)readx_c0_entrylo1() << 30; pa 177 arch/mips/lib/dump_tlb.c pa = (pa << 6) & PAGE_MASK; pa 183 arch/mips/lib/dump_tlb.c pwidth, pa, c1, pa 480 arch/mips/mm/cerr-sb1.c unsigned long long taglo, pa; pa 501 arch/mips/mm/cerr-sb1.c pa = (taglo & 0xFFFFFFE000ULL) | addr; pa 515 arch/mips/mm/cerr-sb1.c way, pa, dc_state_str(state), state, taghi, taglo); pa 86 arch/nds32/mm/cacheflush.c static inline unsigned long kremap0(unsigned long uaddr, unsigned long pa) pa 92 arch/nds32/mm/cacheflush.c pte = (pa | PAGE_KERNEL); pa 106 arch/nds32/mm/cacheflush.c static inline unsigned long kremap1(unsigned long uaddr, unsigned long pa) pa 112 arch/nds32/mm/cacheflush.c pte = (pa | PAGE_KERNEL); pa 6 arch/parisc/include/asm/special_insns.h unsigned long pa; \ pa 10 arch/parisc/include/asm/special_insns.h : "=r" (pa) \ pa 14 arch/parisc/include/asm/special_insns.h pa; \ pa 18 arch/parisc/include/asm/special_insns.h unsigned long pa; \ pa 22 arch/parisc/include/asm/special_insns.h : "=r" (pa) \ pa 26 arch/parisc/include/asm/special_insns.h pa; \ pa 39 arch/powerpc/boot/wii.c static int mipc_check_address(u32 pa) pa 42 arch/powerpc/boot/wii.c if (pa < 0x10000000 || pa > 0x14000000) pa 140 arch/powerpc/include/asm/book3s/32/pgtable.h int map_kernel_page(unsigned long va, phys_addr_t pa, pgprot_t prot); pa 247 arch/powerpc/include/asm/book3s/64/hash.h int hash__map_kernel_page(unsigned long ea, unsigned long pa, pgprot_t prot); pa 404 arch/powerpc/include/asm/book3s/64/mmu-hash.h static inline unsigned long hpte_encode_r(unsigned long pa, int base_psize, pa 409 arch/powerpc/include/asm/book3s/64/mmu-hash.h return pa & HPTE_R_RPN; pa 413 arch/powerpc/include/asm/book3s/64/mmu-hash.h return (pa & ~((1ul << shift) - 1)) | (penc << LP_SHIFT); pa 1037 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int map_kernel_page(unsigned long ea, unsigned long pa, pgprot_t prot) pa 1044 arch/powerpc/include/asm/book3s/64/pgtable.h return radix__map_kernel_page(ea, pa, prot, PAGE_SIZE); pa 1046 arch/powerpc/include/asm/book3s/64/pgtable.h return hash__map_kernel_page(ea, pa, prot); pa 277 arch/powerpc/include/asm/book3s/64/radix.h extern int radix__map_kernel_page(unsigned long ea, unsigned long pa, pa 725 arch/powerpc/include/asm/io.h int early_ioremap_range(unsigned long ea, phys_addr_t pa, pa 727 arch/powerpc/include/asm/io.h void __iomem *do_ioremap(phys_addr_t pa, phys_addr_t offset, unsigned long size, pa 733 arch/powerpc/include/asm/io.h extern void __iomem * __ioremap_at(phys_addr_t pa, void *ea, pa 67 arch/powerpc/include/asm/nohash/32/pgtable.h int map_kernel_page(unsigned long va, phys_addr_t pa, pgprot_t prot); pa 358 arch/powerpc/include/asm/nohash/64/pgtable.h int map_kernel_page(unsigned long ea, unsigned long pa, pgprot_t prot); pa 356 arch/powerpc/kernel/eeh.c unsigned long pa; pa 368 arch/powerpc/kernel/eeh.c pa = pte_pfn(*ptep); pa 372 arch/powerpc/kernel/eeh.c pa <<= hugepage_shift; pa 373 arch/powerpc/kernel/eeh.c pa |= token & ((1ul << hugepage_shift) - 1); pa 375 arch/powerpc/kernel/eeh.c pa <<= PAGE_SHIFT; pa 376 arch/powerpc/kernel/eeh.c pa |= token & (PAGE_SIZE - 1); pa 379 arch/powerpc/kernel/eeh.c return pa; pa 329 arch/powerpc/kernel/mce.c uint64_t ea = 0, pa = 0; pa 445 arch/powerpc/kernel/mce.c pa = evt->u.ue_error.physical_address; pa 520 arch/powerpc/kernel/mce.c if (pa) pa 521 arch/powerpc/kernel/mce.c sprintf(dar_str + n, "paddr: %016llx ", pa); pa 522 arch/powerpc/kernel/mce.c } else if (pa) { pa 523 arch/powerpc/kernel/mce.c sprintf(pa_str, " paddr: %016llx", pa); pa 201 arch/powerpc/kvm/book3s_hv_rm_mmu.c unsigned long i, pa, gpa, gfn, psize; pa 233 arch/powerpc/kvm/book3s_hv_rm_mmu.c pa = 0; pa 285 arch/powerpc/kvm/book3s_hv_rm_mmu.c pa = pte_pfn(pte) << PAGE_SHIFT; pa 286 arch/powerpc/kvm/book3s_hv_rm_mmu.c pa |= hva & (host_pte_size - 1); pa 287 arch/powerpc/kvm/book3s_hv_rm_mmu.c pa |= gpa & ~PAGE_MASK; pa 294 arch/powerpc/kvm/book3s_hv_rm_mmu.c ptel |= pa; pa 296 arch/powerpc/kvm/book3s_hv_rm_mmu.c if (pa) pa 897 arch/powerpc/kvm/book3s_hv_rm_mmu.c unsigned long gfn, hva, pa, psize = PAGE_SHIFT; pa 921 arch/powerpc/kvm/book3s_hv_rm_mmu.c pa = pte_pfn(pte) << PAGE_SHIFT; pa 922 arch/powerpc/kvm/book3s_hv_rm_mmu.c pa |= hva & (psize - 1); pa 923 arch/powerpc/kvm/book3s_hv_rm_mmu.c pa |= gpa & ~PAGE_MASK; pa 926 arch/powerpc/kvm/book3s_hv_rm_mmu.c *hpa = pa; pa 938 arch/powerpc/kvm/book3s_hv_rm_mmu.c unsigned long pa, mmu_seq; pa 946 arch/powerpc/kvm/book3s_hv_rm_mmu.c ret = kvmppc_get_hpa(vcpu, dest, 1, &pa, &memslot); pa 958 arch/powerpc/kvm/book3s_hv_rm_mmu.c for (i = 0; i < SZ_4K; i += L1_CACHE_BYTES, pa += L1_CACHE_BYTES) pa 959 arch/powerpc/kvm/book3s_hv_rm_mmu.c dcbz((void *)pa); pa 62 arch/powerpc/mm/book3s32/mmu.c unsigned long p_block_mapped(phys_addr_t pa) pa 66 arch/powerpc/mm/book3s32/mmu.c if (pa >= bat_addrs[b].phys pa 67 arch/powerpc/mm/book3s32/mmu.c && pa < (bat_addrs[b].limit-bat_addrs[b].start) pa 69 arch/powerpc/mm/book3s32/mmu.c return bat_addrs[b].start+(pa-bat_addrs[b].phys); pa 25 arch/powerpc/mm/book3s64/hash_4k.c unsigned long rflags, pa; pa 79 arch/powerpc/mm/book3s64/hash_4k.c pa = pte_pfn(__pte(old_pte)) << PAGE_SHIFT; pa 86 arch/powerpc/mm/book3s64/hash_4k.c slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa, rflags, 0, pa 93 arch/powerpc/mm/book3s64/hash_4k.c slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa, pa 43 arch/powerpc/mm/book3s64/hash_64k.c unsigned long rflags, pa; pa 150 arch/powerpc/mm/book3s64/hash_64k.c pa = pte_pfn(__pte(old_pte)) << HW_PAGE_SHIFT; pa 152 arch/powerpc/mm/book3s64/hash_64k.c pa = pte_pfn(__pte(old_pte)) << PAGE_SHIFT; pa 153 arch/powerpc/mm/book3s64/hash_64k.c pa += (subpg_index << shift); pa 160 arch/powerpc/mm/book3s64/hash_64k.c slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa, rflags, 0, pa 169 arch/powerpc/mm/book3s64/hash_64k.c slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa, pa 229 arch/powerpc/mm/book3s64/hash_64k.c unsigned long rflags, pa; pa 286 arch/powerpc/mm/book3s64/hash_64k.c pa = pte_pfn(__pte(old_pte)) << PAGE_SHIFT; pa 293 arch/powerpc/mm/book3s64/hash_64k.c slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa, rflags, 0, pa 301 arch/powerpc/mm/book3s64/hash_64k.c slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa, pa 27 arch/powerpc/mm/book3s64/hash_hugepage.c unsigned long rflags, pa, hidx; pa 133 arch/powerpc/mm/book3s64/hash_hugepage.c pa = pmd_pfn(__pmd(old_pmd)) << PAGE_SHIFT; pa 140 arch/powerpc/mm/book3s64/hash_hugepage.c slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa, rflags, 0, pa 147 arch/powerpc/mm/book3s64/hash_hugepage.c slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa, pa 22 arch/powerpc/mm/book3s64/hash_hugetlbpage.c unsigned long pa, unsigned long rlags, pa 32 arch/powerpc/mm/book3s64/hash_hugetlbpage.c unsigned long rflags, pa; pa 103 arch/powerpc/mm/book3s64/hash_hugetlbpage.c pa = pte_pfn(__pte(old_pte)) << PAGE_SHIFT; pa 108 arch/powerpc/mm/book3s64/hash_hugetlbpage.c slot = hpte_insert_repeating(hash, vpn, pa, rflags, 0, pa 338 arch/powerpc/mm/book3s64/hash_native.c unsigned long pa, unsigned long rflags, pa 348 arch/powerpc/mm/book3s64/hash_native.c hpte_group, vpn, pa, rflags, vflags, psize); pa 367 arch/powerpc/mm/book3s64/hash_native.c hpte_r = hpte_encode_r(pa, psize, apsize) | rflags; pa 148 arch/powerpc/mm/book3s64/hash_pgtable.c int hash__map_kernel_page(unsigned long ea, unsigned long pa, pgprot_t prot) pa 167 arch/powerpc/mm/book3s64/hash_pgtable.c set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, prot)); pa 175 arch/powerpc/mm/book3s64/hash_pgtable.c if (htab_bolt_mapping(ea, ea + PAGE_SIZE, pa, pgprot_val(prot), pa 178 arch/powerpc/mm/book3s64/hash_pgtable.c "memory at %016lx !\n", pa); pa 1838 arch/powerpc/mm/book3s64/hash_utils.c unsigned long pa, unsigned long rflags, pa 1848 arch/powerpc/mm/book3s64/hash_utils.c slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa, rflags, vflags, pa 1854 arch/powerpc/mm/book3s64/hash_utils.c slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa, rflags, pa 369 arch/powerpc/mm/book3s64/iommu_api.c unsigned long *pa; pa 382 arch/powerpc/mm/book3s64/iommu_api.c pa = (void *) vmalloc_to_phys(&mem->hpas[entry]); pa 383 arch/powerpc/mm/book3s64/iommu_api.c if (!pa) pa 386 arch/powerpc/mm/book3s64/iommu_api.c *hpa = (*pa & MM_IOMMU_TABLE_GROUP_PAGE_MASK) | (ua & ~PAGE_MASK); pa 396 arch/powerpc/mm/book3s64/iommu_api.c unsigned long *pa; pa 408 arch/powerpc/mm/book3s64/iommu_api.c pa = (void *) vmalloc_to_phys(va); pa 409 arch/powerpc/mm/book3s64/iommu_api.c if (!pa) pa 412 arch/powerpc/mm/book3s64/iommu_api.c *pa |= MM_IOMMU_TABLE_GROUP_PAGE_DIRTY; pa 58 arch/powerpc/mm/book3s64/radix_pgtable.c static int early_map_kernel_page(unsigned long ea, unsigned long pa, pa 64 arch/powerpc/mm/book3s64/radix_pgtable.c unsigned long pfn = pa >> PAGE_SHIFT; pa 108 arch/powerpc/mm/book3s64/radix_pgtable.c static int __map_kernel_page(unsigned long ea, unsigned long pa, pa 114 arch/powerpc/mm/book3s64/radix_pgtable.c unsigned long pfn = pa >> PAGE_SHIFT; pa 129 arch/powerpc/mm/book3s64/radix_pgtable.c return early_map_kernel_page(ea, pa, flags, map_page_size, pa 162 arch/powerpc/mm/book3s64/radix_pgtable.c int radix__map_kernel_page(unsigned long ea, unsigned long pa, pa 166 arch/powerpc/mm/book3s64/radix_pgtable.c return __map_kernel_page(ea, pa, flags, map_page_size, -1, 0, 0); pa 886 arch/powerpc/mm/book3s64/radix_pgtable.c static int __map_kernel_page_nid(unsigned long ea, unsigned long pa, pa 890 arch/powerpc/mm/book3s64/radix_pgtable.c return __map_kernel_page(ea, pa, flags, map_page_size, nid, 0, 0); pa 62 arch/powerpc/mm/ioremap.c int early_ioremap_range(unsigned long ea, phys_addr_t pa, pa 68 arch/powerpc/mm/ioremap.c int err = map_kernel_page(ea + i, pa + i, prot); pa 77 arch/powerpc/mm/ioremap.c void __iomem *do_ioremap(phys_addr_t pa, phys_addr_t offset, unsigned long size, pa 88 arch/powerpc/mm/ioremap.c area->phys_addr = pa; pa 91 arch/powerpc/mm/ioremap.c ret = ioremap_page_range(va, va + size, pa, prot); pa 10 arch/powerpc/mm/ioremap_64.c void __iomem *__ioremap_at(phys_addr_t pa, void *ea, unsigned long size, pgprot_t prot) pa 24 arch/powerpc/mm/ioremap_64.c WARN_ON(pa & ~PAGE_MASK); pa 29 arch/powerpc/mm/ioremap_64.c ret = ioremap_page_range(va, va + size, pa, prot); pa 33 arch/powerpc/mm/ioremap_64.c ret = early_ioremap_range(va, pa, size, prot); pa 26 arch/powerpc/mm/kasan/kasan_init_32.c phys_addr_t pa = __pa(kasan_early_shadow_page); pa 30 arch/powerpc/mm/kasan/kasan_init_32.c __set_pte_at(&init_mm, va, ptep, pfn_pte(PHYS_PFN(pa), prot), 0); pa 116 arch/powerpc/mm/kasan/kasan_init_32.c phys_addr_t pa = __pa(kasan_early_shadow_page); pa 124 arch/powerpc/mm/kasan/kasan_init_32.c if ((pte_val(*ptep) & PTE_RPN_MASK) != pa) pa 127 arch/powerpc/mm/kasan/kasan_init_32.c __set_pte_at(&init_mm, k_cur, ptep, pfn_pte(PHYS_PFN(pa), prot), 0); pa 160 arch/powerpc/mm/mmu_decl.h unsigned long p_block_mapped(phys_addr_t pa); pa 163 arch/powerpc/mm/mmu_decl.h static inline unsigned long p_block_mapped(phys_addr_t pa) { return 0; } pa 43 arch/powerpc/mm/nohash/8xx.c unsigned long p_block_mapped(phys_addr_t pa) pa 49 arch/powerpc/mm/nohash/8xx.c if (pa >= p && pa < p + IMMR_SIZE) pa 50 arch/powerpc/mm/nohash/8xx.c return VIRT_IMMR_BASE + pa - p; pa 51 arch/powerpc/mm/nohash/8xx.c if (pa < block_mapped_ram) pa 52 arch/powerpc/mm/nohash/8xx.c return (unsigned long)__va(pa); pa 73 arch/powerpc/mm/nohash/book3e_pgtable.c int __ref map_kernel_page(unsigned long ea, unsigned long pa, pgprot_t prot) pa 112 arch/powerpc/mm/nohash/book3e_pgtable.c set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, prot)); pa 86 arch/powerpc/mm/nohash/fsl_booke.c unsigned long p_block_mapped(phys_addr_t pa) pa 90 arch/powerpc/mm/nohash/fsl_booke.c if (pa >= tlbcam_addrs[b].phys pa 91 arch/powerpc/mm/nohash/fsl_booke.c && pa < (tlbcam_addrs[b].limit-tlbcam_addrs[b].start) pa 93 arch/powerpc/mm/nohash/fsl_booke.c return tlbcam_addrs[b].start+(pa-tlbcam_addrs[b].phys); pa 59 arch/powerpc/mm/pgtable_32.c int __ref map_kernel_page(unsigned long va, phys_addr_t pa, pgprot_t prot) pa 78 arch/powerpc/mm/pgtable_32.c set_pte_at(&init_mm, va, pg, pfn_pte(pa >> PAGE_SHIFT, prot)); pa 195 arch/powerpc/mm/ptdump/ptdump.c u64 pa = val & PTE_RPN_MASK; pa 202 arch/powerpc/mm/ptdump/ptdump.c st->start_pa = pa; pa 203 arch/powerpc/mm/ptdump/ptdump.c st->last_pa = pa; pa 216 arch/powerpc/mm/ptdump/ptdump.c (pa != st->last_pa + st->page_size && pa 217 arch/powerpc/mm/ptdump/ptdump.c (pa != st->start_pa || st->start_pa != st->last_pa))) { pa 242 arch/powerpc/mm/ptdump/ptdump.c st->start_pa = pa; pa 243 arch/powerpc/mm/ptdump/ptdump.c st->last_pa = pa; pa 248 arch/powerpc/mm/ptdump/ptdump.c st->last_pa = pa; pa 36 arch/powerpc/platforms/ps3/htab.c unsigned long pa, unsigned long rflags, unsigned long vflags, pa 54 arch/powerpc/platforms/ps3/htab.c hpte_r = hpte_encode_r(ps3_mm_phys_to_lpar(pa), psize, apsize) | rflags; pa 68 arch/powerpc/platforms/ps3/htab.c __func__, ps3_result(result), vpn, pa, hpte_group, pa 141 arch/powerpc/platforms/pseries/cmm.c struct cmm_page_array *pa, *npa; pa 164 arch/powerpc/platforms/pseries/cmm.c pa = cmm_page_list; pa 165 arch/powerpc/platforms/pseries/cmm.c if (!pa || pa->index >= CMM_NR_PAGES) { pa 177 arch/powerpc/platforms/pseries/cmm.c pa = cmm_page_list; pa 179 arch/powerpc/platforms/pseries/cmm.c if (!pa || pa->index >= CMM_NR_PAGES) { pa 180 arch/powerpc/platforms/pseries/cmm.c npa->next = pa; pa 182 arch/powerpc/platforms/pseries/cmm.c pa = npa; pa 183 arch/powerpc/platforms/pseries/cmm.c cmm_page_list = pa; pa 195 arch/powerpc/platforms/pseries/cmm.c pa->page[pa->index++] = addr; pa 215 arch/powerpc/platforms/pseries/cmm.c struct cmm_page_array *pa; pa 220 arch/powerpc/platforms/pseries/cmm.c pa = cmm_page_list; pa 222 arch/powerpc/platforms/pseries/cmm.c if (!pa || pa->index <= 0) pa 224 arch/powerpc/platforms/pseries/cmm.c addr = pa->page[--pa->index]; pa 226 arch/powerpc/platforms/pseries/cmm.c if (pa->index == 0) { pa 227 arch/powerpc/platforms/pseries/cmm.c pa = pa->next; pa 229 arch/powerpc/platforms/pseries/cmm.c cmm_page_list = pa; pa 499 arch/powerpc/platforms/pseries/cmm.c struct cmm_page_array *pa; pa 505 arch/powerpc/platforms/pseries/cmm.c pa = cmm_page_list; pa 506 arch/powerpc/platforms/pseries/cmm.c while (pa) { pa 507 arch/powerpc/platforms/pseries/cmm.c if ((unsigned long)pa >= start && (unsigned long)pa < end) pa 509 arch/powerpc/platforms/pseries/cmm.c for (idx = 0; idx < pa->index; idx++) pa 510 arch/powerpc/platforms/pseries/cmm.c if (pa->page[idx] >= start && pa->page[idx] < end) pa 512 arch/powerpc/platforms/pseries/cmm.c pa = pa->next; pa 706 arch/powerpc/platforms/pseries/lpar.c unsigned long vpn, unsigned long pa, pa 718 arch/powerpc/platforms/pseries/lpar.c hpte_group, vpn, pa, rflags, vflags, psize); pa 721 arch/powerpc/platforms/pseries/lpar.c hpte_r = hpte_encode_r(pa, psize, apsize) | rflags; pa 178 arch/riscv/mm/init.c static pte_t *__init get_pte_virt(phys_addr_t pa) pa 182 arch/riscv/mm/init.c return (pte_t *)set_fixmap_offset(FIX_PTE, pa); pa 184 arch/riscv/mm/init.c return (pte_t *)((uintptr_t)pa); pa 200 arch/riscv/mm/init.c uintptr_t va, phys_addr_t pa, pa 208 arch/riscv/mm/init.c ptep[pte_index] = pfn_pte(PFN_DOWN(pa), prot); pa 223 arch/riscv/mm/init.c static pmd_t *__init get_pmd_virt(phys_addr_t pa) pa 227 arch/riscv/mm/init.c return (pmd_t *)set_fixmap_offset(FIX_PMD, pa); pa 229 arch/riscv/mm/init.c return (pmd_t *)((uintptr_t)pa); pa 246 arch/riscv/mm/init.c uintptr_t va, phys_addr_t pa, pa 255 arch/riscv/mm/init.c pmdp[pmd_index] = pfn_pmd(PFN_DOWN(pa), prot); pa 269 arch/riscv/mm/init.c create_pte_mapping(ptep, va, pa, sz, prot); pa 290 arch/riscv/mm/init.c uintptr_t va, phys_addr_t pa, pa 299 arch/riscv/mm/init.c pgdp[pgd_index] = pfn_pgd(PFN_DOWN(pa), prot); pa 313 arch/riscv/mm/init.c create_pgd_next_mapping(nextp, va, pa, sz, prot); pa 413 arch/riscv/mm/init.c phys_addr_t pa, start, end; pa 438 arch/riscv/mm/init.c for (pa = start; pa < end; pa += map_size) { pa 439 arch/riscv/mm/init.c va = (uintptr_t)__va(pa); pa 440 arch/riscv/mm/init.c create_pgd_mapping(swapper_pg_dir, va, pa, pa 584 arch/s390/kernel/smp.c unsigned long pa; pa 586 arch/s390/kernel/smp.c pa = __pa(&pcpu->lowcore->floating_pt_save_area); pa 588 arch/s390/kernel/smp.c pa) != SIGP_CC_ORDER_CODE_ACCEPTED) pa 592 arch/s390/kernel/smp.c pa = __pa(pcpu->lowcore->mcesad & MCESA_ORIGIN_MASK); pa 594 arch/s390/kernel/smp.c pa |= pcpu->lowcore->mcesad & MCESA_LC_MASK; pa 596 arch/s390/kernel/smp.c pa) != SIGP_CC_ORDER_CODE_ACCEPTED) pa 68 arch/s390/mm/cmm.c struct cmm_page_array *pa, *npa; pa 76 arch/s390/mm/cmm.c pa = *list; pa 77 arch/s390/mm/cmm.c if (!pa || pa->index >= CMM_NR_PAGES) { pa 87 arch/s390/mm/cmm.c pa = *list; pa 88 arch/s390/mm/cmm.c if (!pa || pa->index >= CMM_NR_PAGES) { pa 89 arch/s390/mm/cmm.c npa->next = pa; pa 91 arch/s390/mm/cmm.c pa = npa; pa 92 arch/s390/mm/cmm.c *list = pa; pa 97 arch/s390/mm/cmm.c pa->pages[pa->index++] = addr; pa 107 arch/s390/mm/cmm.c struct cmm_page_array *pa; pa 111 arch/s390/mm/cmm.c pa = *list; pa 113 arch/s390/mm/cmm.c if (!pa || pa->index <= 0) pa 115 arch/s390/mm/cmm.c addr = pa->pages[--pa->index]; pa 116 arch/s390/mm/cmm.c if (pa->index == 0) { pa 117 arch/s390/mm/cmm.c pa = pa->next; pa 119 arch/s390/mm/cmm.c *list = pa; pa 133 arch/s390/pci/pci_dma.c static int __dma_update_trans(struct zpci_dev *zdev, unsigned long pa, pa 137 arch/s390/pci/pci_dma.c u8 *page_addr = (u8 *) (pa & PAGE_MASK); pa 218 arch/s390/pci/pci_dma.c static int dma_update_trans(struct zpci_dev *zdev, unsigned long pa, pa 223 arch/s390/pci/pci_dma.c rc = __dma_update_trans(zdev, pa, dma_addr, size, flags); pa 229 arch/s390/pci/pci_dma.c __dma_update_trans(zdev, pa, dma_addr, size, ZPCI_PTE_INVALID); pa 341 arch/s390/pci/pci_dma.c unsigned long pa = page_to_phys(page) + offset; pa 348 arch/s390/pci/pci_dma.c nr_pages = iommu_num_pages(pa, size, PAGE_SIZE); pa 361 arch/s390/pci/pci_dma.c ret = dma_update_trans(zdev, pa, dma_addr, size, flags); pa 372 arch/s390/pci/pci_dma.c zpci_err_dma(ret, pa); pa 403 arch/s390/pci/pci_dma.c unsigned long pa; pa 411 arch/s390/pci/pci_dma.c pa = page_to_phys(page); pa 414 arch/s390/pci/pci_dma.c free_pages(pa, get_order(size)); pa 421 arch/s390/pci/pci_dma.c return (void *) pa; pa 425 arch/s390/pci/pci_dma.c void *pa, dma_addr_t dma_handle, pa 433 arch/s390/pci/pci_dma.c free_pages((unsigned long) pa, get_order(size)); pa 446 arch/s390/pci/pci_dma.c unsigned long pa = 0; pa 458 arch/s390/pci/pci_dma.c pa = page_to_phys(sg_page(s)); pa 459 arch/s390/pci/pci_dma.c ret = __dma_update_trans(zdev, pa, dma_addr, pa 480 arch/s390/pci/pci_dma.c zpci_err_dma(ret, pa); pa 31 arch/sparc/include/asm/pgalloc_32.h unsigned long pa = __nocache_pa(pmdp); pa 33 arch/sparc/include/asm/pgalloc_32.h set_pte((pte_t *)pgdp, __pte((SRMMU_ET_PTD | (pa >> 4)))); pa 58 arch/sparc/kernel/ioport.c static inline void dma_make_coherent(unsigned long pa, unsigned long len) pa 66 arch/sparc/kernel/ioport.c static void __iomem *_sparc_ioremap(struct resource *res, u32 bus, u32 pa, int sz); pa 216 arch/sparc/kernel/ioport.c _sparc_ioremap(struct resource *res, u32 bus, u32 pa, int sz) pa 218 arch/sparc/kernel/ioport.c unsigned long offset = ((unsigned long) pa) & (~PAGE_MASK); pa 229 arch/sparc/kernel/ioport.c pa &= PAGE_MASK; pa 230 arch/sparc/kernel/ioport.c srmmu_mapiorange(bus, pa, res->start, resource_size(res)); pa 2084 arch/sparc/kernel/ldc.c static void fill_cookies(struct cookie_state *sp, unsigned long pa, pa 2088 arch/sparc/kernel/ldc.c unsigned long tlen, new = pa + PAGE_SIZE; pa 2091 arch/sparc/kernel/ldc.c sp->page_table[sp->pte_idx].mte = sp->mte_base | pa; pa 2116 arch/sparc/kernel/ldc.c pa = new; pa 2198 arch/sparc/kernel/ldc.c unsigned long npages, pa; pa 2206 arch/sparc/kernel/ldc.c pa = __pa(buf); pa 2207 arch/sparc/kernel/ldc.c if ((pa | len) & (8UL - 1)) pa 2210 arch/sparc/kernel/ldc.c npages = pages_in_region(pa, len); pa 2225 arch/sparc/kernel/ldc.c fill_cookies(&state, (pa & PAGE_MASK), (pa & ~PAGE_MASK), len); pa 86 arch/sparc/mm/fault_64.c unsigned long pa; pa 104 arch/sparc/mm/fault_64.c pa = pmd_pfn(*pmdp) << PAGE_SHIFT; pa 105 arch/sparc/mm/fault_64.c pa += tpc & ~HPAGE_MASK; pa 110 arch/sparc/mm/fault_64.c : "r" (pa), "i" (ASI_PHYS_USE_EC)); pa 117 arch/sparc/mm/fault_64.c pa = (pte_pfn(pte) << PAGE_SHIFT); pa 118 arch/sparc/mm/fault_64.c pa += (tpc & ~PAGE_MASK); pa 123 arch/sparc/mm/fault_64.c : "r" (pa), "i" (ASI_PHYS_USE_EC)); pa 1661 arch/sparc/mm/init_64.c unsigned long pa = __pa(addr); pa 1663 arch/sparc/mm/init_64.c if ((pa >> max_phys_bits) != 0UL) pa 1666 arch/sparc/mm/init_64.c return pfn_valid(pa >> PAGE_SHIFT); pa 2093 arch/sparc/mm/init_64.c static void patch_one_ktsb_phys(unsigned int *start, unsigned int *end, unsigned long pa) pa 2097 arch/sparc/mm/init_64.c high_bits = (pa >> 32) & 0xffffffff; pa 2098 arch/sparc/mm/init_64.c low_bits = (pa >> 0) & 0xffffffff; pa 2197 arch/sparc/mm/init_64.c unsigned long pa, ret; pa 2199 arch/sparc/mm/init_64.c pa = kern_base + ((unsigned long)&ktsb_descr[0] - KERNBASE); pa 2201 arch/sparc/mm/init_64.c ret = sun4v_mmu_tsb_ctx0(NUM_KTSB_DESCR, pa); pa 2204 arch/sparc/mm/init_64.c "errors with %lx\n", pa, ret); pa 255 arch/x86/boot/compressed/acpi.c acpi_physical_address pa; pa 257 arch/x86/boot/compressed/acpi.c pa = boot_params->acpi_rsdp_addr; pa 264 arch/x86/boot/compressed/acpi.c if (!pa) pa 265 arch/x86/boot/compressed/acpi.c pa = kexec_get_rsdp_addr(); pa 267 arch/x86/boot/compressed/acpi.c if (!pa) pa 268 arch/x86/boot/compressed/acpi.c pa = efi_get_rsdp_addr(); pa 270 arch/x86/boot/compressed/acpi.c if (!pa) pa 271 arch/x86/boot/compressed/acpi.c pa = bios_get_rsdp_addr(); pa 273 arch/x86/boot/compressed/acpi.c return pa; pa 288 arch/x86/events/intel/ds.c phys_addr_t pa; pa 291 arch/x86/events/intel/ds.c pa = virt_to_phys(addr); pa 294 arch/x86/events/intel/ds.c for (; msz < size; msz += PAGE_SIZE, pa += PAGE_SIZE, cea += PAGE_SIZE) pa 295 arch/x86/events/intel/ds.c cea_set_pte(cea, pa, prot); pa 395 arch/x86/hyperv/hv_init.c void hyperv_report_panic_msg(phys_addr_t pa, size_t size) pa 405 arch/x86/hyperv/hv_init.c wrmsrl(HV_X64_MSR_CRASH_P3, pa); pa 123 arch/x86/include/asm/cpu_entry_area.h extern void cea_set_pte(void *cea_vaddr, phys_addr_t pa, pgprot_t flags); pa 454 arch/x86/include/asm/uv/uv_hub.h static inline struct uv_gam_range_s *uv_gam_range(unsigned long pa) pa 457 arch/x86/include/asm/uv/uv_hub.h unsigned long pal = (pa & uv_hub_info->gpa_mask) >> UV_GAM_RANGE_SHFT; pa 466 arch/x86/include/asm/uv/uv_hub.h pr_crit("UV: GAM Range for 0x%lx not found at %p!\n", pa, gr); pa 471 arch/x86/include/asm/uv/uv_hub.h static inline unsigned long uv_gam_range_base(unsigned long pa) pa 473 arch/x86/include/asm/uv/uv_hub.h struct uv_gam_range_s *gr = uv_gam_range(pa); pa 41 arch/x86/kernel/irq_64.c phys_addr_t pa = per_cpu_ptr_to_phys(stack + (i << PAGE_SHIFT)); pa 43 arch/x86/kernel/irq_64.c pages[i] = pfn_to_page(pa >> PAGE_SHIFT); pa 36 arch/x86/kernel/kdebugfs.c u64 pa; pa 47 arch/x86/kernel/kdebugfs.c pa = node->paddr + sizeof(struct setup_data) + pos; pa 48 arch/x86/kernel/kdebugfs.c p = memremap(pa, count, MEMREMAP_WB); pa 312 arch/x86/kernel/kvm.c u64 pa = slow_virt_to_phys(this_cpu_ptr(&apf_reason)); pa 315 arch/x86/kernel/kvm.c pa |= KVM_ASYNC_PF_SEND_ALWAYS; pa 317 arch/x86/kernel/kvm.c pa |= KVM_ASYNC_PF_ENABLED; pa 320 arch/x86/kernel/kvm.c pa |= KVM_ASYNC_PF_DELIVERY_AS_PF_VMEXIT; pa 322 arch/x86/kernel/kvm.c wrmsrl(MSR_KVM_ASYNC_PF_EN, pa); pa 329 arch/x86/kernel/kvm.c unsigned long pa; pa 333 arch/x86/kernel/kvm.c pa = slow_virt_to_phys(this_cpu_ptr(&kvm_apic_eoi)) pa 335 arch/x86/kernel/kvm.c wrmsrl(MSR_KVM_PV_EOI_EN, pa); pa 174 arch/x86/kernel/kvmclock.c u64 pa; pa 179 arch/x86/kernel/kvmclock.c pa = slow_virt_to_phys(&src->pvti) | 0x01ULL; pa 180 arch/x86/kernel/kvmclock.c wrmsrl(msr_kvm_system_time, pa); pa 181 arch/x86/kernel/kvmclock.c pr_info("kvm-clock: cpu %d, msr %llx, %s", smp_processor_id(), pa, txt); pa 29 arch/x86/mm/cpu_entry_area.c void cea_set_pte(void *cea_vaddr, phys_addr_t pa, pgprot_t flags) pa 32 arch/x86/mm/cpu_entry_area.c pte_t pte = pfn_pte(pa >> PAGE_SHIFT, flags); pa 202 arch/x86/mm/mem_encrypt.c unsigned long pfn, pa, size; pa 232 arch/x86/mm/mem_encrypt.c pa = pfn << page_level_shift(level); pa 240 arch/x86/mm/mem_encrypt.c clflush_cache_range(__va(pa), size); pa 244 arch/x86/mm/mem_encrypt.c sme_early_encrypt(pa, size); pa 246 arch/x86/mm/mem_encrypt.c sme_early_decrypt(pa, size); pa 460 arch/x86/mm/pti.c phys_addr_t pa = per_cpu_ptr_to_phys((void *)va); pa 467 arch/x86/mm/pti.c *target_pte = pfn_pte(pa >> PAGE_SHIFT, PAGE_KERNEL); pa 28 arch/x86/mm/srat.c acpi_numa_x2apic_affinity_init(struct acpi_srat_x2apic_cpu_affinity *pa) pa 35 arch/x86/mm/srat.c if (pa->header.length < sizeof(struct acpi_srat_x2apic_cpu_affinity)) { pa 39 arch/x86/mm/srat.c if ((pa->flags & ACPI_SRAT_CPU_ENABLED) == 0) pa 41 arch/x86/mm/srat.c pxm = pa->proximity_domain; pa 42 arch/x86/mm/srat.c apic_id = pa->apic_id; pa 67 arch/x86/mm/srat.c acpi_numa_processor_affinity_init(struct acpi_srat_cpu_affinity *pa) pa 74 arch/x86/mm/srat.c if (pa->header.length != sizeof(struct acpi_srat_cpu_affinity)) { pa 78 arch/x86/mm/srat.c if ((pa->flags & ACPI_SRAT_CPU_ENABLED) == 0) pa 80 arch/x86/mm/srat.c pxm = pa->proximity_domain_lo; pa 82 arch/x86/mm/srat.c pxm |= *((unsigned int*)pa->proximity_domain_hi) << 8; pa 91 arch/x86/mm/srat.c apic_id = (pa->apic_id << 8) | pa->local_sapic_eid; pa 93 arch/x86/mm/srat.c apic_id = pa->apic_id; pa 953 arch/x86/platform/efi/efi.c unsigned long pa; pa 969 arch/x86/platform/efi/efi.c pa = __pa(new_memmap); pa 978 arch/x86/platform/efi/efi.c if (efi_memmap_init_late(pa, efi.memmap.desc_size * count)) { pa 991 arch/x86/platform/efi/efi.c if (efi_setup_page_tables(pa, 1 << pg_shift)) pa 1001 arch/x86/platform/efi/efi.c (efi_memory_desc_t *)pa); pa 1008 arch/x86/platform/efi/efi.c (efi_memory_desc_t *)pa); pa 319 arch/x86/platform/efi/efi_64.c phys_addr_t pa; pa 327 arch/x86/platform/efi/efi_64.c pa = slow_virt_to_phys(va); pa 330 arch/x86/platform/efi/efi_64.c if (WARN_ON((pa ^ (pa + size - 1)) & PAGE_MASK)) pa 333 arch/x86/platform/efi/efi_64.c return pa; pa 433 arch/x86/platform/efi/efi_64.c u64 pa = md->phys_addr; pa 458 arch/x86/platform/efi/efi_64.c if (!(pa & (PMD_SIZE - 1))) { pa 461 arch/x86/platform/efi/efi_64.c u64 pa_offset = pa & (PMD_SIZE - 1); pa 379 arch/x86/platform/efi/quirks.c u64 pa = md->phys_addr; pa 398 arch/x86/platform/efi/quirks.c if (kernel_unmap_pages_in_pgd(pgd, pa, md->num_pages)) pa 399 arch/x86/platform/efi/quirks.c pr_err("Failed to unmap 1:1 mapping for 0x%llx\n", pa); pa 44 arch/x86/xen/enlighten_hvm.c u64 pa; pa 54 arch/x86/xen/enlighten_hvm.c for (pa = PAGE_SIZE; pa 55 arch/x86/xen/enlighten_hvm.c !e820__mapped_all(pa, pa + PAGE_SIZE, E820_TYPE_RAM) || pa 56 arch/x86/xen/enlighten_hvm.c memblock_is_reserved(pa); pa 57 arch/x86/xen/enlighten_hvm.c pa += PAGE_SIZE) pa 60 arch/x86/xen/enlighten_hvm.c shared_info_pfn = PHYS_PFN(pa); pa 62 arch/x86/xen/enlighten_hvm.c memblock_reserve(pa, PAGE_SIZE); pa 63 arch/x86/xen/enlighten_hvm.c HYPERVISOR_shared_info = early_memremap(pa, PAGE_SIZE); pa 1129 arch/x86/xen/mmu_pv.c unsigned long pa = __pa(pgtbl) & PHYSICAL_PAGE_MASK; pa 1132 arch/x86/xen/mmu_pv.c pin_pagetable_pfn(MMUEXT_UNPIN_TABLE, PFN_DOWN(pa)); pa 1133 arch/x86/xen/mmu_pv.c ClearPagePinned(virt_to_page(__va(pa))); pa 1134 arch/x86/xen/mmu_pv.c xen_free_ro_pages(pa, PAGE_SIZE); pa 1139 arch/x86/xen/mmu_pv.c unsigned long pa; pa 1144 arch/x86/xen/mmu_pv.c pa = pmd_val(*pmd) & PHYSICAL_PAGE_MASK; pa 1145 arch/x86/xen/mmu_pv.c xen_free_ro_pages(pa, PMD_SIZE); pa 1153 arch/x86/xen/mmu_pv.c pa = pte_pfn(pte_tbl[i]) << PAGE_SHIFT; pa 1154 arch/x86/xen/mmu_pv.c xen_free_ro_pages(pa, PAGE_SIZE); pa 1162 arch/x86/xen/mmu_pv.c unsigned long pa; pa 1167 arch/x86/xen/mmu_pv.c pa = pud_val(*pud) & PHYSICAL_PAGE_MASK; pa 1168 arch/x86/xen/mmu_pv.c xen_free_ro_pages(pa, PUD_SIZE); pa 1184 arch/x86/xen/mmu_pv.c unsigned long pa; pa 1189 arch/x86/xen/mmu_pv.c pa = p4d_val(*p4d) & PHYSICAL_PAGE_MASK; pa 1190 arch/x86/xen/mmu_pv.c xen_free_ro_pages(pa, P4D_SIZE); pa 2014 arch/x86/xen/mmu_pv.c phys_addr_t pa; pa 2020 arch/x86/xen/mmu_pv.c pa = read_cr3_pa(); pa 2021 arch/x86/xen/mmu_pv.c pgd = native_make_pgd(xen_read_phys_ulong(pa + pgd_index(vaddr) * pa 2026 arch/x86/xen/mmu_pv.c pa = pgd_val(pgd) & PTE_PFN_MASK; pa 2027 arch/x86/xen/mmu_pv.c pud = native_make_pud(xen_read_phys_ulong(pa + pud_index(vaddr) * pa 2031 arch/x86/xen/mmu_pv.c pa = pud_val(pud) & PTE_PFN_MASK; pa 2033 arch/x86/xen/mmu_pv.c return pa + (vaddr & ~PUD_MASK); pa 2035 arch/x86/xen/mmu_pv.c pmd = native_make_pmd(xen_read_phys_ulong(pa + pmd_index(vaddr) * pa 2039 arch/x86/xen/mmu_pv.c pa = pmd_val(pmd) & PTE_PFN_MASK; pa 2041 arch/x86/xen/mmu_pv.c return pa + (vaddr & ~PMD_MASK); pa 2043 arch/x86/xen/mmu_pv.c pte = native_make_pte(xen_read_phys_ulong(pa + pte_index(vaddr) * pa 2047 arch/x86/xen/mmu_pv.c pa = pte_pfn(pte) << PAGE_SHIFT; pa 2049 arch/x86/xen/mmu_pv.c return pa | (vaddr & ~PAGE_MASK); pa 322 drivers/acpi/numa.c acpi_numa_x2apic_affinity_init(struct acpi_srat_x2apic_cpu_affinity *pa) pa 324 drivers/acpi/numa.c pr_warn("Found unsupported x2apic [0x%08x] SRAT entry\n", pa->apic_id); pa 183 drivers/acpi/osl.c acpi_physical_address pa; pa 200 drivers/acpi/osl.c pa = acpi_arch_get_root_pointer(); pa 201 drivers/acpi/osl.c if (pa) pa 202 drivers/acpi/osl.c return pa; pa 211 drivers/acpi/osl.c acpi_find_root_pointer(&pa); pa 214 drivers/acpi/osl.c return pa; pa 360 drivers/base/platform.c struct platform_object *pa = container_of(dev, struct platform_object, pa 363 drivers/base/platform.c of_device_node_put(&pa->pdev.dev); pa 364 drivers/base/platform.c kfree(pa->pdev.dev.platform_data); pa 365 drivers/base/platform.c kfree(pa->pdev.mfd_cell); pa 366 drivers/base/platform.c kfree(pa->pdev.resource); pa 367 drivers/base/platform.c kfree(pa->pdev.driver_override); pa 368 drivers/base/platform.c kfree(pa); pa 381 drivers/base/platform.c struct platform_object *pa; pa 383 drivers/base/platform.c pa = kzalloc(sizeof(*pa) + strlen(name) + 1, GFP_KERNEL); pa 384 drivers/base/platform.c if (pa) { pa 385 drivers/base/platform.c strcpy(pa->name, name); pa 386 drivers/base/platform.c pa->pdev.name = pa->name; pa 387 drivers/base/platform.c pa->pdev.id = id; pa 388 drivers/base/platform.c device_initialize(&pa->pdev.dev); pa 389 drivers/base/platform.c pa->pdev.dev.release = platform_device_release; pa 390 drivers/base/platform.c setup_pdev_dma_masks(&pa->pdev); pa 393 drivers/base/platform.c return pa ? &pa->pdev : NULL; pa 18 drivers/char/agp/alpha-agp.c unsigned long pa; pa 22 drivers/char/agp/alpha-agp.c pa = agp->ops->translate(agp, dma_addr); pa 24 drivers/char/agp/alpha-agp.c if (pa == (unsigned long)-EINVAL) pa 30 drivers/char/agp/alpha-agp.c page = virt_to_page(__va(pa)); pa 171 drivers/clk/ti/adpll.c unsigned long pa; pa 204 drivers/clk/ti/adpll.c sprintf(buf, "%08lx.%s.%s", d->pa, base_name, postfix); pa 230 drivers/clk/ti/adpll.c snprintf(con_id, 16, "pll%03lx%s", d->pa & 0xfff, postfix + 1); pa 903 drivers/clk/ti/adpll.c d->pa = res->start; pa 1077 drivers/crypto/n2_core.c unsigned long pa; pa 1079 drivers/crypto/n2_core.c pa = (c->arr[c->arr_len-1].src_paddr + pa 1083 drivers/crypto/n2_core.c memcpy(rctx->temp_iv, __va(pa), pa 94 drivers/crypto/nx/nx-842-powernv.c unsigned long pa, unsigned int len) pa 96 drivers/crypto/nx/nx-842-powernv.c unsigned int l = min_t(unsigned int, len, LEN_ON_PAGE(pa)); pa 102 drivers/crypto/nx/nx-842-powernv.c dde->address = cpu_to_be64(pa); pa 118 drivers/crypto/nx/nx-842-powernv.c unsigned long pa = nx842_get_pa(buf); pa 121 drivers/crypto/nx/nx-842-powernv.c if (!IS_ALIGNED(pa, DDE_BUFFER_ALIGN)) { pa 123 drivers/crypto/nx/nx-842-powernv.c in ? "input" : "output", pa, DDE_BUFFER_ALIGN); pa 141 drivers/crypto/nx/nx-842-powernv.c if (len <= LEN_ON_PAGE(pa)) { pa 142 drivers/crypto/nx/nx-842-powernv.c ret = setup_direct_dde(dde, pa, len); pa 149 drivers/crypto/nx/nx-842-powernv.c ret = setup_direct_dde(&ddl[i], pa, len); pa 152 drivers/crypto/nx/nx-842-powernv.c pa = nx842_get_pa(buf); pa 101 drivers/crypto/nx/nx-842.h #define LEN_ON_SIZE(pa, size) ((size) - ((pa) & ((size) - 1))) pa 102 drivers/crypto/nx/nx-842.h #define LEN_ON_PAGE(pa) LEN_ON_SIZE(pa, PAGE_SIZE) pa 123 drivers/dio/dio.c unsigned long pa; pa 128 drivers/dio/dio.c pa = dio_scodetophysaddr(scode); pa 130 drivers/dio/dio.c if (!pa) pa 134 drivers/dio/dio.c va = (void *)(pa + DIO_VIRADDRBASE); pa 136 drivers/dio/dio.c va = ioremap(pa, PAGE_SIZE); pa 196 drivers/dio/dio.c unsigned long pa; pa 201 drivers/dio/dio.c pa = dio_scodetophysaddr(scode); pa 203 drivers/dio/dio.c if (!pa) pa 207 drivers/dio/dio.c va = (void *)(pa + DIO_VIRADDRBASE); pa 209 drivers/dio/dio.c va = ioremap(pa, PAGE_SIZE); pa 226 drivers/dio/dio.c dev->resource.start = pa; pa 227 drivers/dio/dio.c dev->resource.end = pa + DIO_SIZE(scode, va); pa 440 drivers/edac/cpc925_edac.c unsigned long pa; pa 460 drivers/edac/cpc925_edac.c pa = mci->csrows[rank]->first_page << PAGE_SHIFT; pa 467 drivers/edac/cpc925_edac.c pa |= c << (14 - i); pa 471 drivers/edac/cpc925_edac.c pa |= bank << 19; pa 477 drivers/edac/cpc925_edac.c pa |= c << (26 - i); pa 483 drivers/edac/cpc925_edac.c pa |= c << (21 + i); pa 489 drivers/edac/cpc925_edac.c pa |= c << (18 - i); pa 495 drivers/edac/cpc925_edac.c pa |= c << (29 - i); pa 498 drivers/edac/cpc925_edac.c *offset = pa & (PAGE_SIZE - 1); pa 499 drivers/edac/cpc925_edac.c *pfn = pa >> PAGE_SHIFT; pa 501 drivers/edac/cpc925_edac.c edac_dbg(0, "ECC physical address 0x%lx\n", pa); pa 901 drivers/firmware/stratix10-svc.c phys_addr_t pa; pa 914 drivers/firmware/stratix10-svc.c pa = gen_pool_virt_to_phys(genpool, va); pa 917 drivers/firmware/stratix10-svc.c pmem->paddr = pa; pa 911 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c struct kfd_process_device_apertures *pa; pa 941 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c pa = kzalloc((sizeof(struct kfd_process_device_apertures) * pa 943 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c if (!pa) pa 950 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c kfree(pa); pa 957 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c pa[nodes].gpu_id = pdd->dev->id; pa 958 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c pa[nodes].lds_base = pdd->lds_base; pa 959 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c pa[nodes].lds_limit = pdd->lds_limit; pa 960 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c pa[nodes].gpuvm_base = pdd->gpuvm_base; pa 961 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c pa[nodes].gpuvm_limit = pdd->gpuvm_limit; pa 962 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c pa[nodes].scratch_base = pdd->scratch_base; pa 963 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c pa[nodes].scratch_limit = pdd->scratch_limit; pa 988 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c pa, pa 990 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c kfree(pa); pa 83 drivers/gpu/drm/etnaviv/etnaviv_mmu.c u32 pa = sg_dma_address(sg) - sg->offset; pa 86 drivers/gpu/drm/etnaviv/etnaviv_mmu.c VERB("map[%d]: %08x %08x(%zx)", i, iova, pa, bytes); pa 88 drivers/gpu/drm/etnaviv/etnaviv_mmu.c ret = etnaviv_context_map(context, da, pa, bytes, prot); pa 777 drivers/gpu/drm/i915/gvt/gtt.c u64 pa, void *p_data, int bytes); pa 1711 drivers/gpu/drm/i915/gvt/gtt.c u64 pa, void *p_data, int bytes) pa 1720 drivers/gpu/drm/i915/gvt/gtt.c index = (pa & (PAGE_SIZE - 1)) >> info->gtt_entry_size_shift; pa 60 drivers/gpu/drm/i915/gvt/mmio.c static void failsafe_emulate_mmio_rw(struct intel_vgpu *vgpu, u64 pa, pa 72 drivers/gpu/drm/i915/gvt/mmio.c offset = intel_vgpu_gpa_to_mmio_offset(vgpu, pa); pa 102 drivers/gpu/drm/i915/gvt/mmio.c int intel_vgpu_emulate_mmio_read(struct intel_vgpu *vgpu, u64 pa, pa 110 drivers/gpu/drm/i915/gvt/mmio.c failsafe_emulate_mmio_rw(vgpu, pa, p_data, bytes, true); pa 115 drivers/gpu/drm/i915/gvt/mmio.c offset = intel_vgpu_gpa_to_mmio_offset(vgpu, pa); pa 136 drivers/gpu/drm/i915/gvt/mmio.c ret = intel_gvt_hypervisor_read_gpa(vgpu, pa, p_data, bytes); pa 174 drivers/gpu/drm/i915/gvt/mmio.c int intel_vgpu_emulate_mmio_write(struct intel_vgpu *vgpu, u64 pa, pa 182 drivers/gpu/drm/i915/gvt/mmio.c failsafe_emulate_mmio_rw(vgpu, pa, p_data, bytes, false); pa 188 drivers/gpu/drm/i915/gvt/mmio.c offset = intel_vgpu_gpa_to_mmio_offset(vgpu, pa); pa 209 drivers/gpu/drm/i915/gvt/mmio.c ret = intel_gvt_hypervisor_write_gpa(vgpu, pa, p_data, bytes); pa 87 drivers/gpu/drm/i915/gvt/mmio.h int intel_vgpu_emulate_mmio_read(struct intel_vgpu *vgpu, u64 pa, pa 89 drivers/gpu/drm/i915/gvt/mmio.h int intel_vgpu_emulate_mmio_write(struct intel_vgpu *vgpu, u64 pa, pa 573 drivers/gpu/drm/msm/disp/mdp5/mdp5_crtc.c struct plane_state *pa = (struct plane_state *)a; pa 575 drivers/gpu/drm/msm/disp/mdp5/mdp5_crtc.c return pa->state->zpos - pb->state->zpos; pa 332 drivers/gpu/drm/omapdrm/dss/hdmi4_cec.c void hdmi4_cec_set_phys_addr(struct hdmi_core_data *core, u16 pa) pa 334 drivers/gpu/drm/omapdrm/dss/hdmi4_cec.c cec_s_phys_addr(core->adap, pa, false); pa 29 drivers/gpu/drm/omapdrm/dss/hdmi4_cec.h void hdmi4_cec_set_phys_addr(struct hdmi_core_data *core, u16 pa); pa 35 drivers/gpu/drm/omapdrm/dss/hdmi4_cec.h static inline void hdmi4_cec_set_phys_addr(struct hdmi_core_data *core, u16 pa) pa 218 drivers/gpu/drm/omapdrm/omap_dmm_tiler.c static void *alloc_dma(struct dmm_txn *txn, size_t sz, dma_addr_t *pa) pa 228 drivers/gpu/drm/omapdrm/omap_dmm_tiler.c *pa = txn->current_pa; pa 506 drivers/gpu/drm/vmwgfx/device_include/svga_reg.h PA pa; pa 566 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c cb_hdr->ptr.pa += (u64) new_start_offset; pa 893 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c cb_hdr->ptr.pa = (u64)man->handle + (u64)offset; pa 936 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c cb_hdr->ptr.pa = (u64)header->handle + pa 9894 drivers/infiniband/hw/hfi1/chip.c u32 type, unsigned long pa, u16 order) pa 9902 drivers/infiniband/hw/hfi1/chip.c pa = 0; pa 9910 drivers/infiniband/hw/hfi1/chip.c trace_hfi1_put_tid(dd, index, type, pa, order); pa 9915 drivers/infiniband/hw/hfi1/chip.c | ((pa >> RT_ADDR_SHIFT) & RCV_ARRAY_RT_ADDR_MASK) pa 1422 drivers/infiniband/hw/hfi1/chip.h u32 type, unsigned long pa, u16 order); pa 89 drivers/infiniband/hw/hfi1/trace_tid.h unsigned long va, unsigned long pa, dma_addr_t dma), pa 90 drivers/infiniband/hw/hfi1/trace_tid.h TP_ARGS(ctxt, subctxt, rarr, npages, va, pa, dma), pa 97 drivers/infiniband/hw/hfi1/trace_tid.h __field(unsigned long, pa) pa 106 drivers/infiniband/hw/hfi1/trace_tid.h __entry->pa = pa; pa 114 drivers/infiniband/hw/hfi1/trace_tid.h __entry->pa, pa 123 drivers/infiniband/hw/hfi1/trace_tid.h unsigned long va, unsigned long pa, dma_addr_t dma), pa 124 drivers/infiniband/hw/hfi1/trace_tid.h TP_ARGS(ctxt, subctxt, rarr, npages, va, pa, dma) pa 130 drivers/infiniband/hw/hfi1/trace_tid.h unsigned long va, unsigned long pa, dma_addr_t dma), pa 131 drivers/infiniband/hw/hfi1/trace_tid.h TP_ARGS(ctxt, subctxt, rarr, npages, va, pa, dma) pa 137 drivers/infiniband/hw/hfi1/trace_tid.h u32 index, u32 type, unsigned long pa, u16 order), pa 138 drivers/infiniband/hw/hfi1/trace_tid.h TP_ARGS(dd, index, type, pa, order), pa 141 drivers/infiniband/hw/hfi1/trace_tid.h __field(unsigned long, pa); pa 148 drivers/infiniband/hw/hfi1/trace_tid.h __entry->pa = pa; pa 156 drivers/infiniband/hw/hfi1/trace_tid.h __entry->pa, pa 3711 drivers/infiniband/hw/i40iw/i40iw_cm.c iwqp->ietf_mem.pa, pa 998 drivers/infiniband/hw/i40iw/i40iw_ctrl.c set_64bit_val(wqe, 32, commit_fpm_mem->pa); pa 1060 drivers/infiniband/hw/i40iw/i40iw_ctrl.c set_64bit_val(wqe, 32, query_fpm_mem->pa); pa 3484 drivers/infiniband/hw/i40iw/i40iw_ctrl.c query_fpm_mem.pa = dev->fpm_query_buf_pa; pa 3503 drivers/infiniband/hw/i40iw/i40iw_ctrl.c vf_dev->fpm_query_buf_pa = dev->vf_fpm_query_buf[iw_vf_idx].pa; pa 3505 drivers/infiniband/hw/i40iw/i40iw_ctrl.c query_fpm_mem.pa = vf_dev->fpm_query_buf_pa; pa 3603 drivers/infiniband/hw/i40iw/i40iw_ctrl.c commit_fpm_mem.pa = dev->fpm_commit_buf_pa; pa 3661 drivers/infiniband/hw/i40iw/i40iw_ctrl.c data = (u64)sdbuf->pa + offset; pa 4249 drivers/infiniband/hw/i40iw/i40iw_ctrl.c values_mem.pa = pcmdinfo->in.u.query_fpm_values.fpm_values_pa; pa 4258 drivers/infiniband/hw/i40iw/i40iw_ctrl.c values_mem.pa = pcmdinfo->in.u.commit_fpm_values.fpm_values_pa; pa 112 drivers/infiniband/hw/i40iw/i40iw_hmc.c static inline void i40iw_set_sd_entry(u64 pa, pa 117 drivers/infiniband/hw/i40iw/i40iw_hmc.c entry->data = pa | (I40IW_HMC_MAX_BP_COUNT << I40E_PFHMC_SDDATALOW_PMSDBPCOUNT_SHIFT) | pa 151 drivers/infiniband/hw/i40iw/i40iw_hmc.c u64 pa, u32 sd_idx, pa 160 drivers/infiniband/hw/i40iw/i40iw_hmc.c i40iw_set_sd_entry(pa, sd_idx, type, sdinfo.entry); pa 183 drivers/infiniband/hw/i40iw/i40iw_hmc.c u64 pa; pa 196 drivers/infiniband/hw/i40iw/i40iw_hmc.c pa = (sd_entry->entry_type == I40IW_SD_TYPE_PAGED) ? pa 197 drivers/infiniband/hw/i40iw/i40iw_hmc.c sd_entry->u.pd_table.pd_page_addr.pa : pa 198 drivers/infiniband/hw/i40iw/i40iw_hmc.c sd_entry->u.bp.addr.pa; pa 199 drivers/infiniband/hw/i40iw/i40iw_hmc.c i40iw_set_sd_entry(pa, i, sd_entry->entry_type, pa 655 drivers/infiniband/hw/i40iw/i40iw_hmc.c page_desc = page->pa | 0x1; pa 220 drivers/infiniband/hw/i40iw/i40iw_hmc.h u64 pa, u32 sd_idx, enum i40iw_sd_entry_type type, pa 540 drivers/infiniband/hw/i40iw/i40iw_main.c memptr->pa = iwdev->obj_next.pa + extra; pa 546 drivers/infiniband/hw/i40iw/i40iw_main.c iwdev->obj_next.pa = memptr->pa + size; pa 588 drivers/infiniband/hw/i40iw/i40iw_main.c dev->cqp->host_ctx_pa = mem.pa; pa 594 drivers/infiniband/hw/i40iw/i40iw_main.c cqp_init_info.sq_pa = cqp->sq.pa; pa 595 drivers/infiniband/hw/i40iw/i40iw_main.c cqp_init_info.host_ctx_pa = mem.pa; pa 658 drivers/infiniband/hw/i40iw/i40iw_main.c info.cq_pa = ccq->mem_cq.pa; pa 661 drivers/infiniband/hw/i40iw/i40iw_main.c info.shadow_area_pa = mem.pa; pa 741 drivers/infiniband/hw/i40iw/i40iw_main.c info.ceqe_pa = iwceq->mem.pa; pa 881 drivers/infiniband/hw/i40iw/i40iw_main.c info.aeq_elem_pa = aeq->mem.pa; pa 1341 drivers/infiniband/hw/i40iw/i40iw_main.c info.fpm_query_buf_pa = mem.pa; pa 1347 drivers/infiniband/hw/i40iw/i40iw_main.c info.fpm_commit_buf_pa = mem.pa; pa 88 drivers/infiniband/hw/i40iw/i40iw_osdep.h dma_addr_t pa; pa 286 drivers/infiniband/hw/i40iw/i40iw_pble.c mem.pa = chunk->dmaaddrs[i]; pa 303 drivers/infiniband/hw/i40iw/i40iw_pble.c vf_pble_info.pd_pl_pba = sd_entry->u.pd_table.pd_page_addr.pa; pa 397 drivers/infiniband/hw/i40iw/i40iw_pble.c sd_entry->u.pd_table.pd_page_addr.pa : sd_entry->u.bp.addr.pa; pa 126 drivers/infiniband/hw/i40iw/i40iw_puda.c set_64bit_val(wqe, 0, buf->mem.pa); pa 452 drivers/infiniband/hw/i40iw/i40iw_puda.c info.paddr = buf->mem.pa; pa 581 drivers/infiniband/hw/i40iw/i40iw_puda.c qp->sq_pa = mem->pa; pa 697 drivers/infiniband/hw/i40iw/i40iw_puda.c info.cq_base_pa = mem->pa; pa 698 drivers/infiniband/hw/i40iw/i40iw_puda.c info.shadow_area_pa = mem->pa + cqsize; pa 762 drivers/infiniband/hw/i40iw/i40iw_utils.c (dma_addr_t *)&mem->pa, GFP_KERNEL); pa 781 drivers/infiniband/hw/i40iw/i40iw_utils.c mem->va, (dma_addr_t)mem->pa); pa 1111 drivers/infiniband/hw/i40iw/i40iw_utils.c cqp_info->in.u.query_fpm_values.fpm_values_pa = values_mem->pa; pa 1144 drivers/infiniband/hw/i40iw/i40iw_utils.c cqp_info->in.u.commit_fpm_values.fpm_values_pa = values_mem->pa; pa 497 drivers/infiniband/hw/i40iw/i40iw_verbs.c info->sq_pa = mem->pa; pa 594 drivers/infiniband/hw/i40iw/i40iw_verbs.c init_info.q2_pa = iwqp->q2_ctx_mem.pa; pa 614 drivers/infiniband/hw/i40iw/i40iw_verbs.c iwqp->host_ctx.pa = init_info.host_ctx_pa; pa 1172 drivers/infiniband/hw/i40iw/i40iw_verbs.c info.cq_base_pa = iwcq->kmem.pa; pa 146 drivers/infiniband/hw/mlx5/mem.c dma_addr_t pa = pa 149 drivers/infiniband/hw/mlx5/mem.c pas[i] = cpu_to_be64(umem_dma_to_mtt(pa)); pa 122 drivers/infiniband/hw/ocrdma/ocrdma.h dma_addr_t pa; pa 128 drivers/infiniband/hw/ocrdma/ocrdma.h dma_addr_t pa; pa 224 drivers/infiniband/hw/ocrdma/ocrdma.h dma_addr_t pa; pa 267 drivers/infiniband/hw/ocrdma/ocrdma.h dma_addr_t pa; pa 336 drivers/infiniband/hw/ocrdma/ocrdma.h dma_addr_t pa; pa 372 drivers/infiniband/hw/ocrdma/ocrdma.h dma_addr_t pa; pa 444 drivers/infiniband/hw/ocrdma/ocrdma.h dma_addr_t pa; pa 448 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ocrdma_build_q_pages(&cmd->pa[0], cmd->num_pages, eq->q.dma, pa 544 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ocrdma_build_q_pages(&cmd->pa[0], cq->size / OCRDMA_MIN_Q_PAGE_SIZE, pa 571 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_pa *pa; pa 592 drivers/infiniband/hw/ocrdma/ocrdma_hw.c pa = &cmd->pa[0]; pa 594 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ocrdma_build_q_pages(pa, num_pages, mq->dma, PAGE_SIZE_4K); pa 1302 drivers/infiniband/hw/ocrdma/ocrdma_hw.c mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dev->stats_mem.pa & 0xffffffff); pa 1303 drivers/infiniband/hw/ocrdma/ocrdma_hw.c mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dev->stats_mem.pa); pa 1342 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dma.size, &dma.pa, GFP_KERNEL); pa 1350 drivers/infiniband/hw/ocrdma/ocrdma_hw.c mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dma.pa & 0xffffffff); pa 1351 drivers/infiniband/hw/ocrdma/ocrdma_hw.c mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dma.pa); pa 1371 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dma_free_coherent(&dev->nic_info.pdev->dev, dma.size, dma.va, dma.pa); pa 1652 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dma_addr_t pa; pa 1681 drivers/infiniband/hw/ocrdma/ocrdma_hw.c &dev->av_tbl.pbl.pa, pa 1687 drivers/infiniband/hw/ocrdma/ocrdma_hw.c &pa, GFP_KERNEL); pa 1690 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dev->av_tbl.pa = pa; pa 1695 drivers/infiniband/hw/ocrdma/ocrdma_hw.c pbes[i].pa_lo = (u32)cpu_to_le32(pa & 0xffffffff); pa 1696 drivers/infiniband/hw/ocrdma/ocrdma_hw.c pbes[i].pa_hi = (u32)cpu_to_le32(upper_32_bits(pa)); pa 1697 drivers/infiniband/hw/ocrdma/ocrdma_hw.c pa += PAGE_SIZE; pa 1699 drivers/infiniband/hw/ocrdma/ocrdma_hw.c cmd->tbl_addr[0].lo = (u32)(dev->av_tbl.pbl.pa & 0xFFFFFFFF); pa 1700 drivers/infiniband/hw/ocrdma/ocrdma_hw.c cmd->tbl_addr[0].hi = (u32)upper_32_bits(dev->av_tbl.pbl.pa); pa 1711 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dev->av_tbl.pa); pa 1715 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dev->av_tbl.pbl.pa); pa 1738 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dev->av_tbl.pa); pa 1741 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dev->av_tbl.pbl.pa); pa 1820 drivers/infiniband/hw/ocrdma/ocrdma_hw.c cq->va = dma_alloc_coherent(&pdev->dev, cq->len, &cq->pa, GFP_KERNEL); pa 1872 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ocrdma_build_q_pages(&cmd->cmd.pa[0], hw_pages, cq->pa, page_size); pa 1883 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dma_free_coherent(&pdev->dev, cq->len, cq->va, cq->pa); pa 1905 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dma_free_coherent(&dev->nic_info.pdev->dev, cq->len, cq->va, cq->pa); pa 1997 drivers/infiniband/hw/ocrdma/ocrdma_hw.c cmd->pbl[i].lo = (u32) (hwmr->pbl_table[i].pa & 0xffffffff); pa 1998 drivers/infiniband/hw/ocrdma/ocrdma_hw.c cmd->pbl[i].hi = upper_32_bits(hwmr->pbl_table[i].pa); pa 2028 drivers/infiniband/hw/ocrdma/ocrdma_hw.c (u32) (hwmr->pbl_table[i + pbl_offset].pa & 0xffffffff); pa 2030 drivers/infiniband/hw/ocrdma/ocrdma_hw.c upper_32_bits(hwmr->pbl_table[i + pbl_offset].pa); pa 2187 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dma_addr_t pa; pa 2208 drivers/infiniband/hw/ocrdma/ocrdma_hw.c qp->sq.va = dma_alloc_coherent(&pdev->dev, len, &pa, GFP_KERNEL); pa 2212 drivers/infiniband/hw/ocrdma/ocrdma_hw.c qp->sq.pa = pa; pa 2214 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ocrdma_build_q_pages(&cmd->wq_addr[0], hw_pages, pa, hw_page_size); pa 2242 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dma_addr_t pa = 0; pa 2258 drivers/infiniband/hw/ocrdma/ocrdma_hw.c qp->rq.va = dma_alloc_coherent(&pdev->dev, len, &pa, GFP_KERNEL); pa 2261 drivers/infiniband/hw/ocrdma/ocrdma_hw.c qp->rq.pa = pa; pa 2265 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ocrdma_build_q_pages(&cmd->rq_addr[0], hw_pages, pa, hw_page_size); pa 2305 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dma_addr_t pa = 0; pa 2314 drivers/infiniband/hw/ocrdma/ocrdma_hw.c qp->ird_q_va = dma_alloc_coherent(&pdev->dev, ird_q_len, &pa, pa 2319 drivers/infiniband/hw/ocrdma/ocrdma_hw.c pa, ird_page_size); pa 2456 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dma_free_coherent(&pdev->dev, qp->rq.len, qp->rq.va, qp->rq.pa); pa 2459 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dma_free_coherent(&pdev->dev, qp->sq.len, qp->sq.va, qp->sq.pa); pa 2735 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dma_free_coherent(&pdev->dev, qp->sq.len, qp->sq.va, qp->sq.pa); pa 2737 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dma_free_coherent(&pdev->dev, qp->rq.len, qp->rq.va, qp->rq.pa); pa 2752 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dma_addr_t pa; pa 2772 drivers/infiniband/hw/ocrdma/ocrdma_hw.c srq->rq.va = dma_alloc_coherent(&pdev->dev, len, &pa, GFP_KERNEL); pa 2777 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ocrdma_build_q_pages(&cmd->rq_addr[0], hw_pages, pa, hw_page_size); pa 2780 drivers/infiniband/hw/ocrdma/ocrdma_hw.c srq->rq.pa = pa; pa 2812 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dma_free_coherent(&pdev->dev, srq->rq.len, srq->rq.va, pa); pa 2873 drivers/infiniband/hw/ocrdma/ocrdma_hw.c srq->rq.va, srq->rq.pa); pa 2881 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dma_addr_t pa; pa 2892 drivers/infiniband/hw/ocrdma/ocrdma_hw.c req = dma_alloc_coherent(&pdev->dev, cmd.hdr.pyld_len, &pa, GFP_KERNEL); pa 2900 drivers/infiniband/hw/ocrdma/ocrdma_hw.c mqe_sge->pa_lo = (u32) (pa & 0xFFFFFFFFUL); pa 2901 drivers/infiniband/hw/ocrdma/ocrdma_hw.c mqe_sge->pa_hi = (u32) upper_32_bits(pa); pa 2917 drivers/infiniband/hw/ocrdma/ocrdma_hw.c dma_free_coherent(&pdev->dev, cmd.hdr.pyld_len, req, pa); pa 325 drivers/infiniband/hw/ocrdma/ocrdma_sli.h struct ocrdma_pa pa[MAX_OCRDMA_EQ_PAGES]; pa 797 drivers/infiniband/hw/ocrdma/ocrdma_sli.h struct ocrdma_pa pa[OCRDMA_CREATE_CQ_MAX_PAGES]; pa 839 drivers/infiniband/hw/ocrdma/ocrdma_sli.h struct ocrdma_pa pa[8]; pa 77 drivers/infiniband/hw/ocrdma/ocrdma_stats.c &mem->pa, GFP_KERNEL); pa 97 drivers/infiniband/hw/ocrdma/ocrdma_stats.c mem->va, mem->pa); pa 487 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c &ctx->ah_tbl.pa, GFP_KERNEL); pa 519 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_del_mmap(ctx, ctx->ah_tbl.pa, ctx->ah_tbl.len); pa 522 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ctx->ah_tbl.pa); pa 535 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_del_mmap(uctx, uctx->ah_tbl.pa, uctx->ah_tbl.len); pa 537 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c uctx->ah_tbl.pa); pa 761 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c mr->pbl_table[i].pa); pa 800 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c dma_addr_t pa; pa 809 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c va = dma_alloc_coherent(&pdev->dev, dma_len, &pa, GFP_KERNEL); pa 816 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c mr->pbl_table[i].pa = pa; pa 1086 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_del_mmap(cq->ucontext, (u64) cq->pa, pa 1732 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_del_mmap(pd->uctx, (u64) qp->sq.pa, pa 1735 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_del_mmap(pd->uctx, (u64) qp->rq.pa, pa 1878 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_del_mmap(srq->pd->uctx, (u64) srq->rq.pa, pa 226 drivers/infiniband/hw/qedr/qedr.h dma_addr_t pa; pa 496 drivers/infiniband/hw/qedr/verbs.c pbl[i].va, pbl[i].pa); pa 516 drivers/infiniband/hw/qedr/verbs.c dma_addr_t pa; pa 525 drivers/infiniband/hw/qedr/verbs.c va = dma_alloc_coherent(&pdev->dev, pbl_info->pbl_size, &pa, pa 531 drivers/infiniband/hw/qedr/verbs.c pbl_table[i].pa = pa; pa 539 drivers/infiniband/hw/qedr/verbs.c pbl_main_tbl[i] = pbl_table[i + 1].pa; pa 546 drivers/infiniband/hw/qedr/verbs.c pbl_table[i].va, pbl_table[i].pa); pa 865 drivers/infiniband/hw/qedr/verbs.c pbl_ptr = cq->q.pbl_tbl->pa; pa 1399 drivers/infiniband/hw/qedr/verbs.c pbl_base_addr = srq->usrq.pbl_tbl->pa; pa 1564 drivers/infiniband/hw/qedr/verbs.c qp->usq.pbl_tbl->pa = out_params->sq_pbl_phys; pa 1570 drivers/infiniband/hw/qedr/verbs.c qp->urq.pbl_tbl->pa = out_params->rq_pbl_phys; pa 1634 drivers/infiniband/hw/qedr/verbs.c in_params.sq_pbl_ptr = qp->usq.pbl_tbl->pa; pa 1637 drivers/infiniband/hw/qedr/verbs.c in_params.rq_pbl_ptr = qp->urq.pbl_tbl->pa; pa 2580 drivers/infiniband/hw/qedr/verbs.c &info->pbl_table->pa); pa 2593 drivers/infiniband/hw/qedr/verbs.c DP_DEBUG(dev, QEDR_MSG_MR, "extra pbl_table_pa = %pa\n", &tmp->pa); pa 2653 drivers/infiniband/hw/qedr/verbs.c mr->hw_mr.pbl_ptr = mr->info.pbl_table[0].pa; pa 3105 drivers/infiniband/hw/qedr/verbs.c fwqe2->pbl_addr.hi = upper_32_bits(mr->info.pbl_table->pa); pa 3106 drivers/infiniband/hw/qedr/verbs.c fwqe2->pbl_addr.lo = lower_32_bits(mr->info.pbl_table->pa); pa 1836 drivers/infiniband/hw/qib/qib_iba6120.c u32 type, unsigned long pa) pa 1846 drivers/infiniband/hw/qib/qib_iba6120.c if (pa != dd->tidinvalid) { pa 1847 drivers/infiniband/hw/qib/qib_iba6120.c if (pa & ((1U << 11) - 1)) { pa 1849 drivers/infiniband/hw/qib/qib_iba6120.c pa); pa 1852 drivers/infiniband/hw/qib/qib_iba6120.c pa >>= 11; pa 1853 drivers/infiniband/hw/qib/qib_iba6120.c if (pa & ~QLOGIC_IB_RT_ADDR_MASK) { pa 1856 drivers/infiniband/hw/qib/qib_iba6120.c pa); pa 1861 drivers/infiniband/hw/qib/qib_iba6120.c pa |= dd->tidtemplate; pa 1863 drivers/infiniband/hw/qib/qib_iba6120.c pa |= 2 << 29; pa 1885 drivers/infiniband/hw/qib/qib_iba6120.c writel(pa, tidp32); pa 1903 drivers/infiniband/hw/qib/qib_iba6120.c u32 type, unsigned long pa) pa 1910 drivers/infiniband/hw/qib/qib_iba6120.c if (pa != dd->tidinvalid) { pa 1911 drivers/infiniband/hw/qib/qib_iba6120.c if (pa & ((1U << 11) - 1)) { pa 1913 drivers/infiniband/hw/qib/qib_iba6120.c pa); pa 1916 drivers/infiniband/hw/qib/qib_iba6120.c pa >>= 11; pa 1917 drivers/infiniband/hw/qib/qib_iba6120.c if (pa & ~QLOGIC_IB_RT_ADDR_MASK) { pa 1920 drivers/infiniband/hw/qib/qib_iba6120.c pa); pa 1925 drivers/infiniband/hw/qib/qib_iba6120.c pa |= dd->tidtemplate; pa 1927 drivers/infiniband/hw/qib/qib_iba6120.c pa |= 2 << 29; pa 1929 drivers/infiniband/hw/qib/qib_iba6120.c writel(pa, tidp32); pa 2153 drivers/infiniband/hw/qib/qib_iba7220.c u32 type, unsigned long pa) pa 2155 drivers/infiniband/hw/qib/qib_iba7220.c if (pa != dd->tidinvalid) { pa 2156 drivers/infiniband/hw/qib/qib_iba7220.c u64 chippa = pa >> IBA7220_TID_PA_SHIFT; pa 2159 drivers/infiniband/hw/qib/qib_iba7220.c if (pa != (chippa << IBA7220_TID_PA_SHIFT)) { pa 2161 drivers/infiniband/hw/qib/qib_iba7220.c pa); pa 2167 drivers/infiniband/hw/qib/qib_iba7220.c pa); pa 2175 drivers/infiniband/hw/qib/qib_iba7220.c pa = chippa; pa 2177 drivers/infiniband/hw/qib/qib_iba7220.c writeq(pa, tidptr); pa 3769 drivers/infiniband/hw/qib/qib_iba7322.c u32 type, unsigned long pa) pa 3773 drivers/infiniband/hw/qib/qib_iba7322.c if (pa != dd->tidinvalid) { pa 3774 drivers/infiniband/hw/qib/qib_iba7322.c u64 chippa = pa >> IBA7322_TID_PA_SHIFT; pa 3777 drivers/infiniband/hw/qib/qib_iba7322.c if (pa != (chippa << IBA7322_TID_PA_SHIFT)) { pa 3779 drivers/infiniband/hw/qib/qib_iba7322.c pa); pa 3785 drivers/infiniband/hw/qib/qib_iba7322.c pa); pa 3793 drivers/infiniband/hw/qib/qib_iba7322.c pa = chippa; pa 3795 drivers/infiniband/hw/qib/qib_iba7322.c writeq(pa, tidptr); pa 1676 drivers/infiniband/hw/qib/qib_init.c dma_addr_t pa = rcd->rcvegrbuf_phys[chunk]; pa 1688 drivers/infiniband/hw/qib/qib_init.c RCVHQ_RCV_TYPE_EAGER, pa); pa 1689 drivers/infiniband/hw/qib/qib_init.c pa += egrsize; pa 72 drivers/infiniband/hw/usnic/usnic_uiom.c dma_addr_t pa; pa 77 drivers/infiniband/hw/usnic/usnic_uiom.c pa = sg_phys(sg); pa 79 drivers/infiniband/hw/usnic/usnic_uiom.c usnic_dbg("pa: %pa\n", &pa); pa 100 drivers/infiniband/hw/usnic/usnic_uiom.c dma_addr_t pa; pa 170 drivers/infiniband/hw/usnic/usnic_uiom.c pa = sg_phys(sg); pa 172 drivers/infiniband/hw/usnic/usnic_uiom.c cur_base + i*PAGE_SIZE, &pa); pa 251 drivers/infiniband/hw/usnic/usnic_uiom.c dma_addr_t pa; pa 265 drivers/infiniband/hw/usnic/usnic_uiom.c pa = sg_phys(&chunk->page_list[i]); pa 272 drivers/infiniband/hw/usnic/usnic_uiom.c pa_start = pa; pa 273 drivers/infiniband/hw/usnic/usnic_uiom.c pa_end = pa; pa 278 drivers/infiniband/hw/usnic/usnic_uiom.c if ((pa_end + PAGE_SIZE != pa) && pa 279 drivers/infiniband/hw/usnic/usnic_uiom.c (pa != pa_start)) { pa 292 drivers/infiniband/hw/usnic/usnic_uiom.c pa_start = pa; pa 293 drivers/infiniband/hw/usnic/usnic_uiom.c pa_end = pa; pa 298 drivers/infiniband/hw/usnic/usnic_uiom.c size = pa - pa_start + PAGE_SIZE; pa 311 drivers/infiniband/hw/usnic/usnic_uiom.c if (pa != pa_start) pa 124 drivers/iommu/exynos-iommu.c #define mk_lv1ent_sect(pa, prot) ((pa >> PG_ENT_SHIFT) | LV1_PROT[prot] | 2) pa 125 drivers/iommu/exynos-iommu.c #define mk_lv1ent_page(pa) ((pa >> PG_ENT_SHIFT) | 1) pa 126 drivers/iommu/exynos-iommu.c #define mk_lv2ent_lpage(pa, prot) ((pa >> PG_ENT_SHIFT) | LV2_PROT[prot] | 1) pa 127 drivers/iommu/exynos-iommu.c #define mk_lv2ent_spage(pa, prot) ((pa >> PG_ENT_SHIFT) | LV2_PROT[prot] | 2) pa 507 drivers/iommu/msm_iommu.c phys_addr_t pa, size_t len, int prot) pa 514 drivers/iommu/msm_iommu.c ret = priv->iop->map(priv->iop, iova, pa, len, prot); pa 485 drivers/iommu/mtk_iommu.c phys_addr_t pa; pa 488 drivers/iommu/mtk_iommu.c pa = dom->iop->iova_to_phys(dom->iop, iova); pa 491 drivers/iommu/mtk_iommu.c if (data->enable_4GB && pa >= MTK_IOMMU_4GB_MODE_REMAP_BASE) pa 492 drivers/iommu/mtk_iommu.c pa &= ~BIT_ULL(32); pa 494 drivers/iommu/mtk_iommu.c return pa; pa 349 drivers/iommu/mtk_iommu_v1.c phys_addr_t pa; pa 352 drivers/iommu/mtk_iommu_v1.c pa = *(dom->pgt_va + (iova >> MT2701_IOMMU_PAGE_SHIFT)); pa 353 drivers/iommu/mtk_iommu_v1.c pa = pa & (~(MT2701_IOMMU_PAGE_SIZE - 1)); pa 356 drivers/iommu/mtk_iommu_v1.c return pa; pa 168 drivers/iommu/omap-iommu.c u32 l, pa; pa 173 drivers/iommu/omap-iommu.c pa = virt_to_phys(obj->iopgd); pa 174 drivers/iommu/omap-iommu.c if (!IS_ALIGNED(pa, SZ_16K)) pa 181 drivers/iommu/omap-iommu.c iommu_write_reg(obj, pa, MMU_TTB); pa 331 drivers/iommu/omap-iommu.c cr->ram = e->pa | e->endian | e->elsz | e->mixed; pa 556 drivers/iommu/omap-iommu.c static int iopgd_alloc_section(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) pa 561 drivers/iommu/omap-iommu.c if ((da | pa) & ~IOSECTION_MASK) { pa 563 drivers/iommu/omap-iommu.c __func__, da, pa, IOSECTION_SIZE); pa 567 drivers/iommu/omap-iommu.c *iopgd = (pa & IOSECTION_MASK) | prot | IOPGD_SECTION; pa 572 drivers/iommu/omap-iommu.c static int iopgd_alloc_super(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) pa 578 drivers/iommu/omap-iommu.c if ((da | pa) & ~IOSUPER_MASK) { pa 580 drivers/iommu/omap-iommu.c __func__, da, pa, IOSUPER_SIZE); pa 585 drivers/iommu/omap-iommu.c *(iopgd + i) = (pa & IOSUPER_MASK) | prot | IOPGD_SUPER; pa 590 drivers/iommu/omap-iommu.c static int iopte_alloc_page(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) pa 600 drivers/iommu/omap-iommu.c *iopte = (pa & IOPAGE_MASK) | prot | IOPTE_SMALL; pa 604 drivers/iommu/omap-iommu.c __func__, da, pa, iopte, *iopte); pa 609 drivers/iommu/omap-iommu.c static int iopte_alloc_large(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) pa 617 drivers/iommu/omap-iommu.c if ((da | pa) & ~IOLARGE_MASK) { pa 619 drivers/iommu/omap-iommu.c __func__, da, pa, IOLARGE_SIZE); pa 627 drivers/iommu/omap-iommu.c *(iopte + i) = (pa & IOLARGE_MASK) | prot | IOPTE_LARGE; pa 666 drivers/iommu/omap-iommu.c err = fn(obj, e->da, e->pa, prot); pa 1326 drivers/iommu/omap-iommu.c static u32 iotlb_init_entry(struct iotlb_entry *e, u32 da, u32 pa, int pgsz) pa 1331 drivers/iommu/omap-iommu.c e->pa = pa; pa 1342 drivers/iommu/omap-iommu.c phys_addr_t pa, size_t bytes, int prot) pa 1359 drivers/iommu/omap-iommu.c dev_dbg(dev, "mapping da 0x%lx to pa %pa size 0x%x\n", da, &pa, bytes); pa 1361 drivers/iommu/omap-iommu.c iotlb_init_entry(&e, da, pa, omap_pgsz); pa 23 drivers/iommu/omap-iommu.h u32 pa; pa 210 drivers/iommu/s390-iommu.c unsigned long pa, dma_addr_t dma_addr, pa 214 drivers/iommu/s390-iommu.c u8 *page_addr = (u8 *) (pa & PAGE_MASK); pa 168 drivers/iommu/tegra-gart.c unsigned long pa) pa 175 drivers/iommu/tegra-gart.c gart_set_pte(gart, iova, GART_ENTRY_PHYS_ADDR_VALID | pa); pa 181 drivers/iommu/tegra-gart.c phys_addr_t pa, size_t bytes, int prot) pa 190 drivers/iommu/tegra-gart.c ret = __gart_iommu_map(gart, iova, (unsigned long)pa); pa 101 drivers/mailbox/bcm-flexrm-mailbox.c #define BD_START_ADDR_VALUE(pa) \ pa 102 drivers/mailbox/bcm-flexrm-mailbox.c ((u32)((((dma_addr_t)(pa)) >> RING_BD_ALIGN_ORDER) & 0x0fffffff)) pa 107 drivers/mailbox/bcm-flexrm-mailbox.c #define CMPL_START_ADDR_VALUE(pa) \ pa 108 drivers/mailbox/bcm-flexrm-mailbox.c ((u32)((((u64)(pa)) >> RING_CMPL_ALIGN_ORDER) & 0x07ffffff)) pa 1456 drivers/mailbox/bcm-flexrm-mailbox.c const struct of_phandle_args *pa) pa 1461 drivers/mailbox/bcm-flexrm-mailbox.c if (pa->args_count < 3) pa 1464 drivers/mailbox/bcm-flexrm-mailbox.c if (pa->args[0] >= cntlr->num_chans) pa 1467 drivers/mailbox/bcm-flexrm-mailbox.c if (pa->args[1] > MSI_COUNT_MASK) pa 1470 drivers/mailbox/bcm-flexrm-mailbox.c if (pa->args[2] > MSI_TIMER_VAL_MASK) pa 1473 drivers/mailbox/bcm-flexrm-mailbox.c chan = &cntlr->chans[pa->args[0]]; pa 1475 drivers/mailbox/bcm-flexrm-mailbox.c ring->msi_count_threshold = pa->args[1]; pa 1476 drivers/mailbox/bcm-flexrm-mailbox.c ring->msi_timer_val = pa->args[2]; pa 1619 drivers/media/cec/cec-adap.c u16 pa = CEC_PHYS_ADDR_INVALID; pa 1622 drivers/media/cec/cec-adap.c pa = cec_get_edid_phys_addr((const u8 *)edid, pa 1624 drivers/media/cec/cec-adap.c cec_s_phys_addr(adap, pa, false); pa 1973 drivers/media/cec/cec-adap.c u16 pa = (msg->msg[2] << 8) | msg->msg[3]; pa 1976 drivers/media/cec/cec-adap.c adap->phys_addrs[init_laddr] = pa; pa 1978 drivers/media/cec/cec-adap.c cec_phys_addr_exp(pa), init_laddr); pa 187 drivers/media/cec/cec-core.c static void cec_cec_notify(struct cec_adapter *adap, u16 pa) pa 189 drivers/media/cec/cec-core.c cec_s_phys_addr(adap, pa, false); pa 28 drivers/media/cec/cec-notifier.c void (*callback)(struct cec_adapter *adap, u16 pa); pa 170 drivers/media/cec/cec-notifier.c void cec_notifier_set_phys_addr(struct cec_notifier *n, u16 pa) pa 176 drivers/media/cec/cec-notifier.c n->phys_addr = pa; pa 188 drivers/media/cec/cec-notifier.c u16 pa = CEC_PHYS_ADDR_INVALID; pa 194 drivers/media/cec/cec-notifier.c pa = cec_get_edid_phys_addr((const u8 *)edid, pa 196 drivers/media/cec/cec-notifier.c cec_notifier_set_phys_addr(n, pa); pa 202 drivers/media/cec/cec-notifier.c void (*callback)(struct cec_adapter *adap, u16 pa)) pa 2291 drivers/media/i2c/adv7604.c u16 pa; pa 2324 drivers/media/i2c/adv7604.c pa = v4l2_get_edid_phys_addr(edid->edid, edid->blocks * 128, &spa_loc); pa 2325 drivers/media/i2c/adv7604.c err = v4l2_phys_addr_validate(pa, &pa, NULL); pa 2402 drivers/media/i2c/adv7604.c cec_s_phys_addr(state->cec_adap, pa, false); pa 776 drivers/media/i2c/adv7842.c u16 pa; pa 794 drivers/media/i2c/adv7842.c pa = v4l2_get_edid_phys_addr(edid, 256, &spa_loc); pa 795 drivers/media/i2c/adv7842.c err = v4l2_phys_addr_validate(pa, &pa, NULL); pa 840 drivers/media/i2c/adv7842.c cec_s_phys_addr(state->cec_adap, pa, false); pa 1772 drivers/media/i2c/tc358743.c u16 pa; pa 1791 drivers/media/i2c/tc358743.c pa = cec_get_edid_phys_addr(edid->edid, edid->blocks * 128, NULL); pa 1792 drivers/media/i2c/tc358743.c err = v4l2_phys_addr_validate(pa, &pa, NULL); pa 1813 drivers/media/i2c/tc358743.c cec_s_phys_addr(state->cec_adap, pa, false); pa 37 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c unsigned long pa; pa 402 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c vsi->mv_buf.pa = (unsigned long)mem->dma_addr; pa 419 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c vsi->seg_id_buf.pa = (unsigned long)mem->dma_addr; pa 668 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c inst->vsi->mv_buf.pa = (unsigned long)inst->mv_buf.dma_addr; pa 673 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c inst->vsi->seg_id_buf.pa = (unsigned long)inst->seg_id_buf.dma_addr; pa 85 drivers/media/platform/mtk-vpu/mtk_vpu.c dma_addr_t pa; pa 668 drivers/media/platform/mtk-vpu/mtk_vpu.c vpu->extmem[fw_type].pa); pa 681 drivers/media/platform/mtk-vpu/mtk_vpu.c &vpu->extmem[fw_type].pa, pa 690 drivers/media/platform/mtk-vpu/mtk_vpu.c vpu_cfg_writel(vpu, (vpu->extmem[fw_type].pa & 0xFFFFF000) + offset_4gb, pa 695 drivers/media/platform/mtk-vpu/mtk_vpu.c (unsigned long long)vpu->extmem[fw_type].pa, pa 58 drivers/media/radio/radio-keene.c u8 pa; pa 79 drivers/media/radio/radio-keene.c radio->buffer[4] = radio->pa; pa 245 drivers/media/radio/radio-keene.c radio->pa = (ctrl->val - 71) * 100 / 62; pa 335 drivers/media/radio/radio-keene.c radio->pa = 118; pa 330 drivers/media/usb/pulse8-cec/pulse8-cec.c struct cec_log_addrs *log_addrs, u16 *pa) pa 347 drivers/media/usb/pulse8-cec/pulse8-cec.c *pa = CEC_PHYS_ADDR_INVALID; pa 427 drivers/media/usb/pulse8-cec/pulse8-cec.c *pa = (data[0] << 8) | data[1]; pa 429 drivers/media/usb/pulse8-cec/pulse8-cec.c cec_phys_addr_exp(*pa)); pa 450 drivers/media/usb/pulse8-cec/pulse8-cec.c u16 pa) pa 458 drivers/media/usb/pulse8-cec/pulse8-cec.c cec_s_phys_addr(pulse8->adap, pa, false); pa 480 drivers/media/usb/pulse8-cec/pulse8-cec.c u16 pa = adap->phys_addr; pa 554 drivers/media/usb/pulse8-cec/pulse8-cec.c cmd[1] = pa >> 8; pa 555 drivers/media/usb/pulse8-cec/pulse8-cec.c cmd[2] = pa & 0xff; pa 654 drivers/media/usb/pulse8-cec/pulse8-cec.c u16 pa = CEC_PHYS_ADDR_INVALID; pa 679 drivers/media/usb/pulse8-cec/pulse8-cec.c err = pulse8_setup(pulse8, serio, &log_addrs, &pa); pa 690 drivers/media/usb/pulse8-cec/pulse8-cec.c err = pulse8_apply_persistent_config(pulse8, &log_addrs, pa); pa 240 drivers/media/usb/rainshadow-cec/rainshadow-cec.c struct cec_log_addrs *log_addrs, u16 *pa) pa 312 drivers/media/usb/rainshadow-cec/rainshadow-cec.c u16 pa = CEC_PHYS_ADDR_INVALID; pa 336 drivers/media/usb/rainshadow-cec/rainshadow-cec.c err = rain_setup(rain, serio, &log_addrs, &pa); pa 443 drivers/message/fusion/mptbase.c mpt_turbo_reply(MPT_ADAPTER *ioc, u32 pa) pa 451 drivers/message/fusion/mptbase.c ioc->name, pa)); pa 453 drivers/message/fusion/mptbase.c switch (pa >> MPI_CONTEXT_REPLY_TYPE_SHIFT) { pa 455 drivers/message/fusion/mptbase.c req_idx = pa & 0x0000FFFF; pa 456 drivers/message/fusion/mptbase.c cb_idx = (pa & 0x00FF0000) >> 16; pa 470 drivers/message/fusion/mptbase.c if ((pa & 0x58000000) == 0x58000000) { pa 471 drivers/message/fusion/mptbase.c req_idx = pa & 0x0000FFFF; pa 478 drivers/message/fusion/mptbase.c mr = (MPT_FRAME_HDR *) CAST_U32_TO_PTR(pa); pa 482 drivers/message/fusion/mptbase.c mr = (MPT_FRAME_HDR *) CAST_U32_TO_PTR(pa); pa 504 drivers/message/fusion/mptbase.c mpt_reply(MPT_ADAPTER *ioc, u32 pa) pa 525 drivers/message/fusion/mptbase.c reply_dma_low = (pa <<= 1); pa 566 drivers/message/fusion/mptbase.c CHIPREG_WRITE32(&ioc->chip->ReplyFifo, pa); pa 594 drivers/message/fusion/mptbase.c u32 pa = CHIPREG_READ32_dmasync(&ioc->chip->ReplyFifo); pa 596 drivers/message/fusion/mptbase.c if (pa == 0xFFFFFFFF) pa 603 drivers/message/fusion/mptbase.c if (pa & MPI_ADDRESS_REPLY_A_BIT) pa 604 drivers/message/fusion/mptbase.c mpt_reply(ioc, pa); pa 606 drivers/message/fusion/mptbase.c mpt_turbo_reply(ioc, pa); pa 607 drivers/message/fusion/mptbase.c pa = CHIPREG_READ32_dmasync(&ioc->chip->ReplyFifo); pa 608 drivers/message/fusion/mptbase.c } while (pa != 0xFFFFFFFF); pa 1020 drivers/message/fusion/mptctl.c dma_addr_t pa; // phys addr pa 1069 drivers/message/fusion/mptctl.c &pa); pa 363 drivers/misc/fastrpc.c struct fastrpc_buf_overlap *pa = (struct fastrpc_buf_overlap *)a; pa 366 drivers/misc/fastrpc.c int st = CMP(pa->start, pb->start); pa 368 drivers/misc/fastrpc.c int ed = CMP(pb->end, pa->end); pa 100 drivers/misc/mic/bus/scif_bus.h phys_addr_t pa, size_t len); pa 97 drivers/misc/mic/bus/vop_bus.h dma_addr_t pa, size_t len); pa 215 drivers/misc/mic/card/mic_device.c phys_addr_t pa, size_t len) pa 219 drivers/misc/mic/card/mic_device.c return mic_card_map(&mdrv->mdev, pa, len); pa 286 drivers/misc/mic/card/mic_device.c dma_addr_t pa, size_t len) pa 290 drivers/misc/mic/card/mic_device.c return mic_card_map(&mdrv->mdev, pa, len); pa 243 drivers/misc/mic/card/mic_x100.c mdev->mmio.pa = MIC_X100_MMIO_BASE; pa 33 drivers/misc/mic/common/mic_dev.h phys_addr_t pa; pa 104 drivers/misc/mic/host/mic_boot.c dma_addr_t pa, size_t len) pa 108 drivers/misc/mic/host/mic_boot.c return mdev->aper.va + pa; pa 288 drivers/misc/mic/host/mic_boot.c phys_addr_t pa, size_t len) pa 292 drivers/misc/mic/host/mic_boot.c return mdev->aper.va + pa; pa 199 drivers/misc/mic/host/mic_main.c mdev->mmio.pa = pci_resource_start(pdev, mdev->ops->mmio_bar); pa 208 drivers/misc/mic/host/mic_main.c mdev->aper.pa = pci_resource_start(pdev, mdev->ops->aper_bar); pa 210 drivers/misc/mic/host/mic_main.c mdev->aper.va = ioremap_wc(mdev->aper.pa, mdev->aper.len); pa 20 drivers/misc/mic/host/mic_smpt.c static inline u8 mic_sys_addr_to_smpt(struct mic_device *mdev, dma_addr_t pa) pa 22 drivers/misc/mic/host/mic_smpt.c return (pa - mdev->smpt->info.base) >> mdev->smpt->info.page_shift; pa 30 drivers/misc/mic/host/mic_smpt.c static inline u64 mic_smpt_offset(struct mic_device *mdev, dma_addr_t pa) pa 32 drivers/misc/mic/host/mic_smpt.c return pa & mic_system_page_mask(mdev); pa 35 drivers/misc/mic/host/mic_smpt.c static inline u64 mic_smpt_align_low(struct mic_device *mdev, dma_addr_t pa) pa 37 drivers/misc/mic/host/mic_smpt.c return ALIGN(pa - mic_system_page_mask(mdev), pa 41 drivers/misc/mic/host/mic_smpt.c static inline u64 mic_smpt_align_high(struct mic_device *mdev, dma_addr_t pa) pa 43 drivers/misc/mic/host/mic_smpt.c return ALIGN(pa, mdev->smpt->info.page_size); pa 60 drivers/misc/mic/host/mic_smpt.c mic_is_system_addr(struct mic_device *mdev, dma_addr_t pa) pa 62 drivers/misc/mic/host/mic_smpt.c return pa >= mdev->smpt->info.base && pa <= mic_max_system_addr(mdev); pa 674 drivers/misc/mic/scif/scif_dma.c phys = phys - dev->sdev->aper->pa; pa 207 drivers/misc/mic/scif/scif_mmap.c apt_base = sdev->aper->pa; pa 288 drivers/misc/mic/scif/scif_mmap.c ep->remote_dev->sdev->aper->pa; pa 360 drivers/misc/mic/scif/scif_nodeqp.c scif_p2p_setsg(phys_addr_t pa, int page_size, int page_cnt) pa 371 drivers/misc/mic/scif/scif_nodeqp.c page = pfn_to_page(pa >> PAGE_SHIFT); pa 373 drivers/misc/mic/scif/scif_nodeqp.c pa += page_size; pa 393 drivers/misc/mic/scif/scif_nodeqp.c p2p->ppi_sg[SCIF_PPI_MMIO] = scif_p2p_setsg(psdev->mmio->pa, pa 400 drivers/misc/mic/scif/scif_nodeqp.c p2p->ppi_sg[SCIF_PPI_APER] = scif_p2p_setsg(psdev->aper->pa, pa 1120 drivers/misc/mic/scif/scif_rma.c apt_base = sdev->aper->pa; pa 552 drivers/misc/mic/vop/vop_vringh.c daddr += vpdev->aper->pa; pa 629 drivers/misc/mic/vop/vop_vringh.c daddr += vpdev->aper->pa; pa 1039 drivers/misc/mic/vop/vop_vringh.c unsigned long *size, unsigned long *pa) pa 1054 drivers/misc/mic/vop/vop_vringh.c *pa = virt_to_phys(vpdev->hw_ops->get_dp(vpdev)); pa 1063 drivers/misc/mic/vop/vop_vringh.c *pa = virt_to_phys(vvr->vring.va); pa 1079 drivers/misc/mic/vop/vop_vringh.c unsigned long pa, size = vma->vm_end - vma->vm_start, size_rem = size; pa 1090 drivers/misc/mic/vop/vop_vringh.c i = vop_query_offset(vdev, offset, &size, &pa); pa 1096 drivers/misc/mic/vop/vop_vringh.c pa >> PAGE_SHIFT, size, pa 681 drivers/net/can/usb/peak_usb/pcan_usb_core.c const struct peak_usb_adapter *pa = dev->adapter; pa 683 drivers/net/can/usb/peak_usb/pcan_usb_core.c if (pa->dev_set_bittiming) { pa 685 drivers/net/can/usb/peak_usb/pcan_usb_core.c int err = pa->dev_set_bittiming(dev, bt); pa 702 drivers/net/can/usb/peak_usb/pcan_usb_core.c const struct peak_usb_adapter *pa = dev->adapter; pa 704 drivers/net/can/usb/peak_usb/pcan_usb_core.c if (pa->dev_set_data_bittiming) { pa 706 drivers/net/can/usb/peak_usb/pcan_usb_core.c int err = pa->dev_set_data_bittiming(dev, bt); pa 286 drivers/net/ethernet/amd/xgbe/xgbe-desc.c struct xgbe_page_alloc *pa, int alloc_order, pa 324 drivers/net/ethernet/amd/xgbe/xgbe-desc.c pa->pages = pages; pa 325 drivers/net/ethernet/amd/xgbe/xgbe-desc.c pa->pages_len = PAGE_SIZE << order; pa 326 drivers/net/ethernet/amd/xgbe/xgbe-desc.c pa->pages_offset = 0; pa 327 drivers/net/ethernet/amd/xgbe/xgbe-desc.c pa->pages_dma = pages_dma; pa 333 drivers/net/ethernet/amd/xgbe/xgbe-desc.c struct xgbe_page_alloc *pa, pa 336 drivers/net/ethernet/amd/xgbe/xgbe-desc.c get_page(pa->pages); pa 337 drivers/net/ethernet/amd/xgbe/xgbe-desc.c bd->pa = *pa; pa 339 drivers/net/ethernet/amd/xgbe/xgbe-desc.c bd->dma_base = pa->pages_dma; pa 340 drivers/net/ethernet/amd/xgbe/xgbe-desc.c bd->dma_off = pa->pages_offset; pa 343 drivers/net/ethernet/amd/xgbe/xgbe-desc.c pa->pages_offset += len; pa 344 drivers/net/ethernet/amd/xgbe/xgbe-desc.c if ((pa->pages_offset + len) > pa->pages_len) { pa 346 drivers/net/ethernet/amd/xgbe/xgbe-desc.c bd->pa_unmap = *pa; pa 349 drivers/net/ethernet/amd/xgbe/xgbe-desc.c pa->pages = NULL; pa 350 drivers/net/ethernet/amd/xgbe/xgbe-desc.c pa->pages_len = 0; pa 351 drivers/net/ethernet/amd/xgbe/xgbe-desc.c pa->pages_offset = 0; pa 352 drivers/net/ethernet/amd/xgbe/xgbe-desc.c pa->pages_dma = 0; pa 490 drivers/net/ethernet/amd/xgbe/xgbe-desc.c if (rdata->rx.hdr.pa.pages) pa 491 drivers/net/ethernet/amd/xgbe/xgbe-desc.c put_page(rdata->rx.hdr.pa.pages); pa 500 drivers/net/ethernet/amd/xgbe/xgbe-desc.c if (rdata->rx.buf.pa.pages) pa 501 drivers/net/ethernet/amd/xgbe/xgbe-desc.c put_page(rdata->rx.buf.pa.pages); pa 2567 drivers/net/ethernet/amd/xgbe/xgbe-drv.c packet = page_address(rdata->rx.hdr.pa.pages) + pa 2568 drivers/net/ethernet/amd/xgbe/xgbe-drv.c rdata->rx.hdr.pa.pages_offset; pa 2784 drivers/net/ethernet/amd/xgbe/xgbe-drv.c rdata->rx.buf.pa.pages, pa 2785 drivers/net/ethernet/amd/xgbe/xgbe-drv.c rdata->rx.buf.pa.pages_offset, pa 2788 drivers/net/ethernet/amd/xgbe/xgbe-drv.c rdata->rx.buf.pa.pages = NULL; pa 395 drivers/net/ethernet/amd/xgbe/xgbe.h struct xgbe_page_alloc pa; pa 465 drivers/net/ethernet/aquantia/atlantic/aq_nic.c dx_buff->pa = dma_map_single(aq_nic_get_dev(self), pa 470 drivers/net/ethernet/aquantia/atlantic/aq_nic.c if (unlikely(dma_mapping_error(aq_nic_get_dev(self), dx_buff->pa))) { pa 532 drivers/net/ethernet/aquantia/atlantic/aq_nic.c dx_buff->pa = frag_pa; pa 554 drivers/net/ethernet/aquantia/atlantic/aq_nic.c if (!dx_buff->is_gso && !dx_buff->is_vlan && dx_buff->pa) { pa 557 drivers/net/ethernet/aquantia/atlantic/aq_nic.c dx_buff->pa, pa 562 drivers/net/ethernet/aquantia/atlantic/aq_nic.c dx_buff->pa, pa 238 drivers/net/ethernet/aquantia/atlantic/aq_ring.c dma_unmap_single(dev, buff->pa, buff->len, pa 241 drivers/net/ethernet/aquantia/atlantic/aq_ring.c dma_unmap_page(dev, buff->pa, buff->len, pa 252 drivers/net/ethernet/aquantia/atlantic/aq_ring.c buff->pa = 0U; pa 462 drivers/net/ethernet/aquantia/atlantic/aq_ring.c buff->pa = aq_buf_daddr(&buff->rxdata); pa 39 drivers/net/ethernet/aquantia/atlantic/aq_ring.h dma_addr_t pa; pa 473 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c txd->buf_addr = buff->pa; pa 590 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c rxd->buf_addr = buff->pa; pa 534 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c txd->buf_addr = buff->pa; pa 658 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c rxd->buf_addr = buff->pa; pa 142 drivers/net/ethernet/brocade/bna/bfa_cee.c cee->attr_dma.pa = dma_pa; pa 144 drivers/net/ethernet/brocade/bna/bfa_cee.c cee->stats_dma.pa = dma_pa + bfa_cee_attr_meminfo(); pa 177 drivers/net/ethernet/brocade/bna/bfa_cee.c bfa_dma_be_addr_set(cmd->dma_addr, cee->attr_dma.pa); pa 1953 drivers/net/ethernet/brocade/bna/bfa_ioc.c bfa_dma_be_addr_set(attr_req.attr_addr, ioc->attr_dma.pa); pa 2584 drivers/net/ethernet/brocade/bna/bfa_ioc.c ioc->attr_dma.pa = dm_pa; pa 40 drivers/net/ethernet/brocade/bna/bfa_ioc.h u64 pa; /* ! Physical address */ pa 50 drivers/net/ethernet/brocade/bna/bfa_ioc.h #define bfa_dma_be_addr_set(dma_addr, pa) \ pa 51 drivers/net/ethernet/brocade/bna/bfa_ioc.h __bfa_dma_be_addr_set(&dma_addr, (u64)pa) pa 53 drivers/net/ethernet/brocade/bna/bfa_ioc.h __bfa_dma_be_addr_set(union bfi_addr_u *dma_addr, u64 pa) pa 55 drivers/net/ethernet/brocade/bna/bfa_ioc.h dma_addr->a32.addr_lo = (u32) htonl(pa); pa 56 drivers/net/ethernet/brocade/bna/bfa_ioc.h dma_addr->a32.addr_hi = (u32) htonl(upper_32_bits(pa)); pa 63 drivers/net/ethernet/brocade/bna/bfa_ioc.h __bfa_alen_set(struct bfi_alen *alen, u32 len, u64 pa) pa 66 drivers/net/ethernet/brocade/bna/bfa_ioc.h bfa_dma_be_addr_set(alen->al_addr, pa); pa 517 drivers/net/ethernet/brocade/bna/bfa_msgq.c bfa_dma_be_addr_set(msgq_cfg->cmdq.addr, msgq->cmdq.addr.pa); pa 519 drivers/net/ethernet/brocade/bna/bfa_msgq.c bfa_dma_be_addr_set(msgq_cfg->rspq.addr, msgq->rspq.addr.pa); pa 590 drivers/net/ethernet/brocade/bna/bfa_msgq.c bfa_msgq_memclaim(struct bfa_msgq *msgq, u8 *kva, u64 pa) pa 593 drivers/net/ethernet/brocade/bna/bfa_msgq.c msgq->cmdq.addr.pa = pa; pa 596 drivers/net/ethernet/brocade/bna/bfa_msgq.c pa += roundup(BFA_MSGQ_CMDQ_SIZE, BFA_DMA_ALIGN_SZ); pa 599 drivers/net/ethernet/brocade/bna/bfa_msgq.c msgq->rspq.addr.pa = pa; pa 115 drivers/net/ethernet/brocade/bna/bfa_msgq.h void bfa_msgq_memclaim(struct bfa_msgq *msgq, u8 *kva, u64 pa); pa 1955 drivers/net/ethernet/dec/tulip/de4x5.c char *pa; pa 1960 drivers/net/ethernet/dec/tulip/de4x5.c pa = build_setup_frame(dev, ALL); /* Build the basic frame */ pa 1982 drivers/net/ethernet/dec/tulip/de4x5.c *(pa + (i&1)) = *addrs++; pa 1983 drivers/net/ethernet/dec/tulip/de4x5.c if (i & 0x01) pa += 4; pa 5037 drivers/net/ethernet/dec/tulip/de4x5.c char *pa = lp->setup_frame; pa 5045 drivers/net/ethernet/dec/tulip/de4x5.c for (pa=lp->setup_frame+IMPERF_PA_OFFSET, i=0; i<ETH_ALEN; i++) { pa 5046 drivers/net/ethernet/dec/tulip/de4x5.c *(pa + i) = dev->dev_addr[i]; /* Host address */ pa 5047 drivers/net/ethernet/dec/tulip/de4x5.c if (i & 0x01) pa += 2; pa 5052 drivers/net/ethernet/dec/tulip/de4x5.c *(pa + (i&1)) = dev->dev_addr[i]; pa 5053 drivers/net/ethernet/dec/tulip/de4x5.c if (i & 0x01) pa += 4; pa 5056 drivers/net/ethernet/dec/tulip/de4x5.c *(pa + (i&1)) = (char) 0xff; pa 5057 drivers/net/ethernet/dec/tulip/de4x5.c if (i & 0x01) pa += 4; pa 5061 drivers/net/ethernet/dec/tulip/de4x5.c return pa; /* Points to the next entry */ pa 161 drivers/net/ethernet/intel/i40e/i40e_adminq.c cpu_to_le32(upper_32_bits(bi->pa)); pa 163 drivers/net/ethernet/intel/i40e/i40e_adminq.c cpu_to_le32(lower_32_bits(bi->pa)); pa 250 drivers/net/ethernet/intel/i40e/i40e_adminq.c if (hw->aq.asq.r.asq_bi[i].pa) pa 281 drivers/net/ethernet/intel/i40e/i40e_adminq.c wr32(hw, hw->aq.asq.bal, lower_32_bits(hw->aq.asq.desc_buf.pa)); pa 282 drivers/net/ethernet/intel/i40e/i40e_adminq.c wr32(hw, hw->aq.asq.bah, upper_32_bits(hw->aq.asq.desc_buf.pa)); pa 286 drivers/net/ethernet/intel/i40e/i40e_adminq.c if (reg != lower_32_bits(hw->aq.asq.desc_buf.pa)) pa 310 drivers/net/ethernet/intel/i40e/i40e_adminq.c wr32(hw, hw->aq.arq.bal, lower_32_bits(hw->aq.arq.desc_buf.pa)); pa 311 drivers/net/ethernet/intel/i40e/i40e_adminq.c wr32(hw, hw->aq.arq.bah, upper_32_bits(hw->aq.arq.desc_buf.pa)); pa 318 drivers/net/ethernet/intel/i40e/i40e_adminq.c if (reg != lower_32_bits(hw->aq.arq.desc_buf.pa)) pa 834 drivers/net/ethernet/intel/i40e/i40e_adminq.c cpu_to_le32(upper_32_bits(dma_buff->pa)); pa 836 drivers/net/ethernet/intel/i40e/i40e_adminq.c cpu_to_le32(lower_32_bits(dma_buff->pa)); pa 1015 drivers/net/ethernet/intel/i40e/i40e_adminq.c desc->params.external.addr_high = cpu_to_le32(upper_32_bits(bi->pa)); pa 1016 drivers/net/ethernet/intel/i40e/i40e_adminq.c desc->params.external.addr_low = cpu_to_le32(lower_32_bits(bi->pa)); pa 156 drivers/net/ethernet/intel/i40e/i40e_hmc.c page_desc = page->pa | 0x1; pa 99 drivers/net/ethernet/intel/i40e/i40e_hmc.h #define I40E_SET_PF_SD_ENTRY(hw, pa, sd_index, type) \ pa 102 drivers/net/ethernet/intel/i40e/i40e_hmc.h val1 = (u32)(upper_32_bits(pa)); \ pa 103 drivers/net/ethernet/intel/i40e/i40e_hmc.h val2 = (u32)(pa) | (I40E_HMC_MAX_BP_COUNT << \ pa 388 drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c sd_entry->u.pd_table.pd_page_addr.pa, pa 392 drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c I40E_SET_PF_SD_ENTRY(hw, sd_entry->u.bp.addr.pa, pa 121 drivers/net/ethernet/intel/i40e/i40e_main.c mem->va = dma_alloc_coherent(&pf->pdev->dev, mem->size, &mem->pa, pa 138 drivers/net/ethernet/intel/i40e/i40e_main.c dma_free_coherent(&pf->pdev->dev, mem->size, mem->va, mem->pa); pa 140 drivers/net/ethernet/intel/i40e/i40e_main.c mem->pa = 0; pa 36 drivers/net/ethernet/intel/i40e/i40e_osdep.h dma_addr_t pa; pa 147 drivers/net/ethernet/intel/iavf/iavf_adminq.c cpu_to_le32(upper_32_bits(bi->pa)); pa 149 drivers/net/ethernet/intel/iavf/iavf_adminq.c cpu_to_le32(lower_32_bits(bi->pa)); pa 237 drivers/net/ethernet/intel/iavf/iavf_adminq.c if (hw->aq.asq.r.asq_bi[i].pa) pa 268 drivers/net/ethernet/intel/iavf/iavf_adminq.c wr32(hw, hw->aq.asq.bal, lower_32_bits(hw->aq.asq.desc_buf.pa)); pa 269 drivers/net/ethernet/intel/iavf/iavf_adminq.c wr32(hw, hw->aq.asq.bah, upper_32_bits(hw->aq.asq.desc_buf.pa)); pa 273 drivers/net/ethernet/intel/iavf/iavf_adminq.c if (reg != lower_32_bits(hw->aq.asq.desc_buf.pa)) pa 297 drivers/net/ethernet/intel/iavf/iavf_adminq.c wr32(hw, hw->aq.arq.bal, lower_32_bits(hw->aq.arq.desc_buf.pa)); pa 298 drivers/net/ethernet/intel/iavf/iavf_adminq.c wr32(hw, hw->aq.arq.bah, upper_32_bits(hw->aq.arq.desc_buf.pa)); pa 305 drivers/net/ethernet/intel/iavf/iavf_adminq.c if (reg != lower_32_bits(hw->aq.arq.desc_buf.pa)) pa 736 drivers/net/ethernet/intel/iavf/iavf_adminq.c cpu_to_le32(upper_32_bits(dma_buff->pa)); pa 738 drivers/net/ethernet/intel/iavf/iavf_adminq.c cpu_to_le32(lower_32_bits(dma_buff->pa)); pa 916 drivers/net/ethernet/intel/iavf/iavf_adminq.c desc->params.external.addr_high = cpu_to_le32(upper_32_bits(bi->pa)); pa 917 drivers/net/ethernet/intel/iavf/iavf_adminq.c desc->params.external.addr_low = cpu_to_le32(lower_32_bits(bi->pa)); pa 83 drivers/net/ethernet/intel/iavf/iavf_main.c (dma_addr_t *)&mem->pa, GFP_KERNEL); pa 103 drivers/net/ethernet/intel/iavf/iavf_main.c mem->va, (dma_addr_t)mem->pa); pa 32 drivers/net/ethernet/intel/iavf/iavf_osdep.h dma_addr_t pa; pa 81 drivers/net/ethernet/intel/ice/ice_controlq.c &cq->sq.desc_buf.pa, pa 91 drivers/net/ethernet/intel/ice/ice_controlq.c cq->sq.desc_buf.va, cq->sq.desc_buf.pa); pa 93 drivers/net/ethernet/intel/ice/ice_controlq.c cq->sq.desc_buf.pa = 0; pa 112 drivers/net/ethernet/intel/ice/ice_controlq.c &cq->rq.desc_buf.pa, pa 131 drivers/net/ethernet/intel/ice/ice_controlq.c ring->desc_buf.va, ring->desc_buf.pa); pa 133 drivers/net/ethernet/intel/ice/ice_controlq.c ring->desc_buf.pa = 0; pa 163 drivers/net/ethernet/intel/ice/ice_controlq.c cq->rq_buf_size, &bi->pa, pa 184 drivers/net/ethernet/intel/ice/ice_controlq.c cpu_to_le32(upper_32_bits(bi->pa)); pa 186 drivers/net/ethernet/intel/ice/ice_controlq.c cpu_to_le32(lower_32_bits(bi->pa)); pa 197 drivers/net/ethernet/intel/ice/ice_controlq.c cq->rq.r.rq_bi[i].va, cq->rq.r.rq_bi[i].pa); pa 199 drivers/net/ethernet/intel/ice/ice_controlq.c cq->rq.r.rq_bi[i].pa = 0; pa 230 drivers/net/ethernet/intel/ice/ice_controlq.c cq->sq_buf_size, &bi->pa, pa 243 drivers/net/ethernet/intel/ice/ice_controlq.c cq->sq.r.sq_bi[i].va, cq->sq.r.sq_bi[i].pa); pa 245 drivers/net/ethernet/intel/ice/ice_controlq.c cq->sq.r.sq_bi[i].pa = 0; pa 262 drivers/net/ethernet/intel/ice/ice_controlq.c wr32(hw, ring->bal, lower_32_bits(ring->desc_buf.pa)); pa 263 drivers/net/ethernet/intel/ice/ice_controlq.c wr32(hw, ring->bah, upper_32_bits(ring->desc_buf.pa)); pa 266 drivers/net/ethernet/intel/ice/ice_controlq.c if (rd32(hw, ring->bal) != lower_32_bits(ring->desc_buf.pa)) pa 430 drivers/net/ethernet/intel/ice/ice_controlq.c if ((qi)->ring.r.ring##_bi[i].pa) { \ pa 434 drivers/net/ethernet/intel/ice/ice_controlq.c (qi)->ring.r.ring##_bi[i].pa);\ pa 436 drivers/net/ethernet/intel/ice/ice_controlq.c (qi)->ring.r.ring##_bi[i].pa = 0; \ pa 931 drivers/net/ethernet/intel/ice/ice_controlq.c cpu_to_le32(upper_32_bits(dma_buf->pa)); pa 933 drivers/net/ethernet/intel/ice/ice_controlq.c cpu_to_le32(lower_32_bits(dma_buf->pa)); pa 1102 drivers/net/ethernet/intel/ice/ice_controlq.c desc->params.generic.addr_high = cpu_to_le32(upper_32_bits(bi->pa)); pa 1103 drivers/net/ethernet/intel/ice/ice_controlq.c desc->params.generic.addr_low = cpu_to_le32(lower_32_bits(bi->pa)); pa 23 drivers/net/ethernet/intel/ice/ice_osdep.h dma_addr_t pa; pa 472 drivers/net/ethernet/mellanox/mlxsw/cmd.h MLXSW_ITEM64_INDEXED(cmd_mbox, map_fa, pa, 0x00, 12, 52, 0x08, 0x00, true); pa 922 drivers/net/ethernet/mellanox/mlxsw/cmd.h MLXSW_ITEM64_INDEXED(cmd_mbox, sw2hw_dq, pa, 0x10, 12, 52, 0x08, 0x00, true); pa 1077 drivers/net/ethernet/mellanox/mlxsw/cmd.h MLXSW_ITEM64_INDEXED(cmd_mbox, sw2hw_cq, pa, 0x10, 11, 53, 0x08, 0x00, true); pa 1161 drivers/net/ethernet/mellanox/mlxsw/cmd.h MLXSW_ITEM64_INDEXED(cmd_mbox, sw2hw_eq, pa, 0x10, 11, 53, 0x08, 0x00, true); pa 46 drivers/net/ethernet/realtek/r8169_firmware.c struct rtl_fw_phy_action *pa = &rtl_fw->phy_action; pa 73 drivers/net/ethernet/realtek/r8169_firmware.c pa->code = (__le32 *)(fw->data + start); pa 74 drivers/net/ethernet/realtek/r8169_firmware.c pa->size = size; pa 81 drivers/net/ethernet/realtek/r8169_firmware.c pa->code = (__le32 *)fw->data; pa 82 drivers/net/ethernet/realtek/r8169_firmware.c pa->size = fw->size / FW_OPCODE_SIZE; pa 90 drivers/net/ethernet/realtek/r8169_firmware.c struct rtl_fw_phy_action *pa = &rtl_fw->phy_action; pa 93 drivers/net/ethernet/realtek/r8169_firmware.c for (index = 0; index < pa->size; index++) { pa 94 drivers/net/ethernet/realtek/r8169_firmware.c u32 action = le32_to_cpu(pa->code[index]); pa 113 drivers/net/ethernet/realtek/r8169_firmware.c if (index + 2 >= pa->size) pa 119 drivers/net/ethernet/realtek/r8169_firmware.c if (index + 1 + regno >= pa->size) pa 137 drivers/net/ethernet/realtek/r8169_firmware.c struct rtl_fw_phy_action *pa = &rtl_fw->phy_action; pa 143 drivers/net/ethernet/realtek/r8169_firmware.c for (index = 0; index < pa->size; index++) { pa 144 drivers/net/ethernet/realtek/r8169_firmware.c u32 action = le32_to_cpu(pa->code[index]); pa 41 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c if (desc_data->rx.hdr.pa.pages) pa 42 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c put_page(desc_data->rx.hdr.pa.pages); pa 51 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c if (desc_data->rx.buf.pa.pages) pa 52 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c put_page(desc_data->rx.buf.pa.pages); pa 331 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c struct xlgmac_page_alloc *pa, pa 357 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c pa->pages = pages; pa 358 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c pa->pages_len = PAGE_SIZE << order; pa 359 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c pa->pages_offset = 0; pa 360 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c pa->pages_dma = pages_dma; pa 366 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c struct xlgmac_page_alloc *pa, pa 369 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c get_page(pa->pages); pa 370 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c bd->pa = *pa; pa 372 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c bd->dma_base = pa->pages_dma; pa 373 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c bd->dma_off = pa->pages_offset; pa 376 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c pa->pages_offset += len; pa 377 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c if ((pa->pages_offset + len) > pa->pages_len) { pa 379 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c bd->pa_unmap = *pa; pa 382 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c pa->pages = NULL; pa 383 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c pa->pages_len = 0; pa 384 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c pa->pages_offset = 0; pa 385 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c pa->pages_dma = 0; pa 1008 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c packet = page_address(desc_data->rx.hdr.pa.pages) + pa 1009 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c desc_data->rx.hdr.pa.pages_offset; pa 1025 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c desc_data->rx.buf.pa.pages, pa 1026 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c desc_data->rx.buf.pa.pages_offset, pa 1028 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c desc_data->rx.buf.pa.pages = NULL; pa 1206 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c desc_data->rx.buf.pa.pages, pa 1207 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c desc_data->rx.buf.pa.pages_offset, pa 1210 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c desc_data->rx.buf.pa.pages = NULL; pa 231 drivers/net/ethernet/synopsys/dwc-xlgmac.h struct xlgmac_page_alloc pa; pa 675 drivers/net/ethernet/ti/cpsw.c void *pa = page_address(page); pa 676 drivers/net/ethernet/ti/cpsw.c struct cpsw_meta_xdp *xmeta = pa + CPSW_XMETA_OFFSET; pa 724 drivers/net/ethernet/ti/cpsw.c xdp.data = pa + CPSW_HEADROOM + pa 729 drivers/net/ethernet/ti/cpsw.c xdp.data = pa + CPSW_HEADROOM; pa 735 drivers/net/ethernet/ti/cpsw.c xdp.data_hard_start = pa; pa 751 drivers/net/ethernet/ti/cpsw.c skb = build_skb(pa, cpsw_rxbuf_total_len(pkt_size)); pa 789 drivers/net/fddi/defza.h u8 pa[2]; /* preamble */ pa 35 drivers/net/fddi/skfp/pmf.c static int smt_set_para(struct s_smc *smc, struct smt_para *pa, int index, pa 312 drivers/net/fddi/skfp/pmf.c struct smt_para *pa ; pa 369 drivers/net/fddi/skfp/pmf.c pa = (struct smt_para *) (req + 1) ; pa 374 drivers/net/fddi/skfp/pmf.c if (((u_short)len < pa->p_len + PARA_LEN) || (pa->p_len & 3)) { pa 379 drivers/net/fddi/skfp/pmf.c if (((range = (pa->p_type & 0xf000)) == 0x2000) || pa 384 drivers/net/fddi/skfp/pmf.c index = *((u_char *)pa + PARA_LEN + 3) ;/* index */ pa 386 drivers/net/fddi/skfp/pmf.c if (!set && (pa->p_len != 4)) { pa 415 drivers/net/fddi/skfp/pmf.c if (!set && (pa->p_len != 0)) { pa 427 drivers/net/fddi/skfp/pmf.c pt = smt_get_ptab(pa->p_type) ; pa 442 drivers/net/fddi/skfp/pmf.c else if (pa->p_type != SMT_P_AUTHOR && pa 443 drivers/net/fddi/skfp/pmf.c (!set || (pa->p_type != SMT_P1035))) { pa 446 drivers/net/fddi/skfp/pmf.c smt_add_para(smc,&pcon,pa->p_type, pa 450 drivers/net/fddi/skfp/pmf.c st = smt_set_para(smc,pa,index,local,1); pa 454 drivers/net/fddi/skfp/pmf.c smt_add_para(smc,&pcon,pa->p_type, pa 463 drivers/net/fddi/skfp/pmf.c smt_add_para(smc,&pcon,pa->p_type, pa 471 drivers/net/fddi/skfp/pmf.c len -= pa->p_len + PARA_LEN ; pa 472 drivers/net/fddi/skfp/pmf.c pa = (struct smt_para *) ((char *)pa + pa->p_len + PARA_LEN) ; pa 497 drivers/net/fddi/skfp/pmf.c struct smt_para *pa ; pa 519 drivers/net/fddi/skfp/pmf.c pa = (struct smt_para *) sm_to_para(smc,sm,SMT_P_AUTHOR) ; pa 520 drivers/net/fddi/skfp/pmf.c if (!pa) pa 522 drivers/net/fddi/skfp/pmf.c if (pa->p_len != 8) pa 524 drivers/net/fddi/skfp/pmf.c if (memcmp((char *)(pa+1),(char *)smc->mib.fddiPRPMFPasswd,8)) pa 532 drivers/net/fddi/skfp/pmf.c struct smt_para *pa ; pa 535 drivers/net/fddi/skfp/pmf.c pa = (struct smt_para *) sm_to_para(smc,sm,SMT_P1035) ; pa 536 drivers/net/fddi/skfp/pmf.c if (pa) { pa 537 drivers/net/fddi/skfp/pmf.c sc = (struct smt_p_setcount *) pa ; pa 549 drivers/net/fddi/skfp/pmf.c struct smt_para *pa ; pa 582 drivers/net/fddi/skfp/pmf.c pa = (struct smt_para *) to ; /* type/length pointer */ pa 1047 drivers/net/fddi/skfp/pmf.c pa->p_type = para ; pa 1048 drivers/net/fddi/skfp/pmf.c pa->p_len = plen - len - PARA_LEN ; pa 1071 drivers/net/fddi/skfp/pmf.c static int smt_set_para(struct s_smc *smc, struct smt_para *pa, int index, pa 1097 drivers/net/fddi/skfp/pmf.c len = pa->p_len ; pa 1098 drivers/net/fddi/skfp/pmf.c from = (char *) (pa + 1 ) ; pa 1101 drivers/net/fddi/skfp/pmf.c switch (pa->p_type & 0xf000) { pa 1134 drivers/net/fddi/skfp/pmf.c switch (pa->p_type) { pa 1154 drivers/net/fddi/skfp/pmf.c pt = smt_get_ptab(pa->p_type) ; pa 1156 drivers/net/fddi/skfp/pmf.c return (pa->p_type & 0xff00) ? SMT_RDF_NOPARAM : pa 1295 drivers/net/fddi/skfp/pmf.c switch (pa->p_type) { pa 1559 drivers/net/fddi/skfp/pmf.c struct smt_para *pa ; pa 1589 drivers/net/fddi/skfp/pmf.c pa = (struct smt_para *) (sm + 1) ; pa 1593 drivers/net/fddi/skfp/pmf.c printf("TYPE %x LEN %x VALUE\t",pa->p_type,pa->p_len) ; pa 1595 drivers/net/fddi/skfp/pmf.c printf("TYPE %04x LEN %2x VALUE\t",pa->p_type,pa->p_len) ; pa 1597 drivers/net/fddi/skfp/pmf.c n = pa->p_len ; pa 1607 drivers/net/fddi/skfp/pmf.c dump_hex((char *)(pa+1),(int) n) ; pa 1612 drivers/net/fddi/skfp/pmf.c c = (char *)(pa+1) ; pa 1636 drivers/net/fddi/skfp/pmf.c plen = (pa->p_len + PARA_LEN + 3) & ~3 ; pa 1638 drivers/net/fddi/skfp/pmf.c pa = (struct smt_para *)((char *)pa + plen) ; pa 1824 drivers/net/fddi/skfp/smt.c struct smt_para *pa ; pa 1841 drivers/net/fddi/skfp/smt.c pa = (struct smt_para *) p ; pa 1842 drivers/net/fddi/skfp/smt.c plen = pa->p_len ; pa 1843 drivers/net/fddi/skfp/smt.c type = pa->p_type ; pa 1844 drivers/net/fddi/skfp/smt.c pa->p_type = smt_swap_short(pa->p_type) ; pa 1845 drivers/net/fddi/skfp/smt.c pa->p_len = smt_swap_short(pa->p_len) ; pa 1847 drivers/net/fddi/skfp/smt.c plen = pa->p_len ; pa 1848 drivers/net/fddi/skfp/smt.c type = pa->p_type ; pa 710 drivers/net/hyperv/hyperv_net.h u64 pa; /* Physical Address */ pa 750 drivers/net/hyperv/hyperv_net.h u64 pa; /* Physical Address */ pa 93 drivers/net/wireless/ath/wil6210/debugfs.c seq_printf(s, " pa = %pad\n", &ring->pa); pa 213 drivers/net/wireless/ath/wil6210/debugfs.c seq_printf(s, " pa = %pad\n", &sring->pa); pa 28 drivers/net/wireless/ath/wil6210/pmc.c dma_addr_t pa; pa 159 drivers/net/wireless/ath/wil6210/pmc.c &pmc->descriptors[i].pa, pa 174 drivers/net/wireless/ath/wil6210/pmc.c cpu_to_le32(lower_32_bits(pmc->descriptors[i].pa)); pa 176 drivers/net/wireless/ath/wil6210/pmc.c cpu_to_le16((u16)upper_32_bits(pmc->descriptors[i].pa)); pa 212 drivers/net/wireless/ath/wil6210/pmc.c pmc->descriptors[i].pa); pa 296 drivers/net/wireless/ath/wil6210/pmc.c pmc->descriptors[i].pa); pa 153 drivers/net/wireless/ath/wil6210/txrx.c vring->va = dma_alloc_coherent(dev, sz, &vring->pa, GFP_KERNEL); pa 176 drivers/net/wireless/ath/wil6210/txrx.c vring->va, &vring->pa, vring->ctx); pa 185 drivers/net/wireless/ath/wil6210/txrx.c dma_addr_t pa = wil_desc_addr(&d->dma.addr); pa 190 drivers/net/wireless/ath/wil6210/txrx.c dma_unmap_single(dev, pa, dmalen, DMA_TO_DEVICE); pa 193 drivers/net/wireless/ath/wil6210/txrx.c dma_unmap_page(dev, pa, dmalen, DMA_TO_DEVICE); pa 211 drivers/net/wireless/ath/wil6210/txrx.c &vring->pa, vring->ctx); pa 215 drivers/net/wireless/ath/wil6210/txrx.c &vring->pa, vring->ctx); pa 219 drivers/net/wireless/ath/wil6210/txrx.c dma_addr_t pa; pa 248 drivers/net/wireless/ath/wil6210/txrx.c pa = wil_desc_addr(&d->dma.addr); pa 250 drivers/net/wireless/ath/wil6210/txrx.c dma_unmap_single(dev, pa, dmalen, DMA_FROM_DEVICE); pa 255 drivers/net/wireless/ath/wil6210/txrx.c dma_free_coherent(dev, sz, (void *)vring->va, vring->pa); pa 257 drivers/net/wireless/ath/wil6210/txrx.c vring->pa = 0; pa 274 drivers/net/wireless/ath/wil6210/txrx.c dma_addr_t pa; pa 289 drivers/net/wireless/ath/wil6210/txrx.c pa = dma_map_single(dev, skb->data, skb->len, DMA_FROM_DEVICE); pa 290 drivers/net/wireless/ath/wil6210/txrx.c if (unlikely(dma_mapping_error(dev, pa))) { pa 296 drivers/net/wireless/ath/wil6210/txrx.c wil_desc_addr_set(&d->dma.addr, pa); pa 461 drivers/net/wireless/ath/wil6210/txrx.c dma_addr_t pa; pa 492 drivers/net/wireless/ath/wil6210/txrx.c pa = wil_desc_addr(&d->dma.addr); pa 494 drivers/net/wireless/ath/wil6210/txrx.c dma_unmap_single(dev, pa, sz, DMA_FROM_DEVICE); pa 1126 drivers/net/wireless/ath/wil6210/txrx.c static int wil_tx_desc_map(union wil_tx_desc *desc, dma_addr_t pa, pa 1131 drivers/net/wireless/ath/wil6210/txrx.c wil_desc_addr_set(&d->dma.addr, pa); pa 1227 drivers/net/wireless/ath/wil6210/txrx.c cmd.vring_cfg.tx_sw_ring.ring_mem_base = cpu_to_le64(vring->pa); pa 1318 drivers/net/wireless/ath/wil6210/txrx.c cmd.vring_cfg.tx_sw_ring.ring_mem_base = cpu_to_le64(vring->pa); pa 1396 drivers/net/wireless/ath/wil6210/txrx.c cmd.vring_cfg.tx_sw_ring.ring_mem_base = cpu_to_le64(vring->pa); pa 1759 drivers/net/wireless/ath/wil6210/txrx.c dma_addr_t pa; pa 1827 drivers/net/wireless/ath/wil6210/txrx.c pa = dma_map_single(dev, skb->data, hdrlen, DMA_TO_DEVICE); pa 1828 drivers/net/wireless/ath/wil6210/txrx.c if (unlikely(dma_mapping_error(dev, pa))) { pa 1833 drivers/net/wireless/ath/wil6210/txrx.c wil->txrx_ops.tx_desc_map((union wil_tx_desc *)hdr_desc, pa, pa 1872 drivers/net/wireless/ath/wil6210/txrx.c pa = skb_frag_dma_map(dev, frag, pa 1877 drivers/net/wireless/ath/wil6210/txrx.c pa = dma_map_single(dev, pa 1886 drivers/net/wireless/ath/wil6210/txrx.c if (unlikely(dma_mapping_error(dev, pa))) { pa 1902 drivers/net/wireless/ath/wil6210/txrx.c pa, lenmss, vring_index); pa 2046 drivers/net/wireless/ath/wil6210/txrx.c dma_addr_t pa; pa 2065 drivers/net/wireless/ath/wil6210/txrx.c pa = dma_map_single(dev, skb->data, skb_headlen(skb), DMA_TO_DEVICE); pa 2068 drivers/net/wireless/ath/wil6210/txrx.c skb_headlen(skb), skb->data, &pa); pa 2072 drivers/net/wireless/ath/wil6210/txrx.c if (unlikely(dma_mapping_error(dev, pa))) pa 2076 drivers/net/wireless/ath/wil6210/txrx.c wil->txrx_ops.tx_desc_map((union wil_tx_desc *)d, pa, len, pa 2104 drivers/net/wireless/ath/wil6210/txrx.c pa = skb_frag_dma_map(dev, frag, 0, skb_frag_size(frag), pa 2106 drivers/net/wireless/ath/wil6210/txrx.c if (unlikely(dma_mapping_error(dev, pa))) { pa 2113 drivers/net/wireless/ath/wil6210/txrx.c pa, len, ring_index); pa 42 drivers/net/wireless/ath/wil6210/txrx.h dma_addr_t pa) pa 44 drivers/net/wireless/ath/wil6210/txrx.h addr->addr_low = cpu_to_le32(lower_32_bits(pa)); pa 45 drivers/net/wireless/ath/wil6210/txrx.h addr->addr_high = cpu_to_le16((u16)upper_32_bits(pa)); pa 43 drivers/net/wireless/ath/wil6210/txrx_edma.c dma_addr_t pa = wil_tx_desc_get_addr_edma(&d->dma); pa 48 drivers/net/wireless/ath/wil6210/txrx_edma.c dma_unmap_single(dev, pa, dmalen, DMA_TO_DEVICE); pa 51 drivers/net/wireless/ath/wil6210/txrx_edma.c dma_unmap_page(dev, pa, dmalen, DMA_TO_DEVICE); pa 82 drivers/net/wireless/ath/wil6210/txrx_edma.c sz, sring->va, &sring->pa); pa 84 drivers/net/wireless/ath/wil6210/txrx_edma.c dma_free_coherent(dev, sz, (void *)sring->va, sring->pa); pa 85 drivers/net/wireless/ath/wil6210/txrx_edma.c sring->pa = 0; pa 107 drivers/net/wireless/ath/wil6210/txrx_edma.c sring->va = dma_alloc_coherent(dev, sz, &sring->pa, GFP_KERNEL); pa 112 drivers/net/wireless/ath/wil6210/txrx_edma.c &sring->pa); pa 169 drivers/net/wireless/ath/wil6210/txrx_edma.c dma_addr_t pa; pa 197 drivers/net/wireless/ath/wil6210/txrx_edma.c pa = dma_map_single(dev, skb->data, skb->len, DMA_FROM_DEVICE); pa 198 drivers/net/wireless/ath/wil6210/txrx_edma.c if (unlikely(dma_mapping_error(dev, pa))) { pa 212 drivers/net/wireless/ath/wil6210/txrx_edma.c wil_desc_set_addr_edma(&d->dma.addr, &d->dma.addr_high_high, pa); pa 218 drivers/net/wireless/ath/wil6210/txrx_edma.c memcpy(skb->cb, &pa, sizeof(pa)); pa 281 drivers/net/wireless/ath/wil6210/txrx_edma.c dma_addr_t pa; pa 295 drivers/net/wireless/ath/wil6210/txrx_edma.c memcpy(&pa, skb->cb, sizeof(pa)); pa 296 drivers/net/wireless/ath/wil6210/txrx_edma.c dma_unmap_single(dev, pa, wil->rx_buf_len, pa 402 drivers/net/wireless/ath/wil6210/txrx_edma.c ring->va = dma_alloc_coherent(dev, sz, &ring->pa, GFP_KERNEL); pa 409 drivers/net/wireless/ath/wil6210/txrx_edma.c dma_alloc_coherent(dev, sz, &ring->edma_rx_swtail.pa, pa 417 drivers/net/wireless/ath/wil6210/txrx_edma.c ring->size, ring->va, &ring->pa, ring->ctx); pa 422 drivers/net/wireless/ath/wil6210/txrx_edma.c (void *)ring->va, ring->pa); pa 446 drivers/net/wireless/ath/wil6210/txrx_edma.c &ring->pa, ring->ctx); pa 451 drivers/net/wireless/ath/wil6210/txrx_edma.c ring->edma_rx_swtail.pa); pa 460 drivers/net/wireless/ath/wil6210/txrx_edma.c &ring->pa, ring->ctx); pa 486 drivers/net/wireless/ath/wil6210/txrx_edma.c dma_free_coherent(dev, sz, (void *)ring->va, ring->pa); pa 488 drivers/net/wireless/ath/wil6210/txrx_edma.c ring->pa = 0; pa 871 drivers/net/wireless/ath/wil6210/txrx_edma.c dma_addr_t pa; pa 940 drivers/net/wireless/ath/wil6210/txrx_edma.c memcpy(&pa, skb->cb, sizeof(pa)); pa 941 drivers/net/wireless/ath/wil6210/txrx_edma.c dma_unmap_single(dev, pa, sz, DMA_FROM_DEVICE); pa 1120 drivers/net/wireless/ath/wil6210/txrx_edma.c dma_addr_t pa, pa 1129 drivers/net/wireless/ath/wil6210/txrx_edma.c wil_desc_set_addr_edma(&d->dma.addr, &d->dma.addr_high_high, pa); pa 1366 drivers/net/wireless/ath/wil6210/txrx_edma.c dma_addr_t pa; pa 1372 drivers/net/wireless/ath/wil6210/txrx_edma.c pa = dma_map_single(dev, buff_addr, len, DMA_TO_DEVICE); pa 1375 drivers/net/wireless/ath/wil6210/txrx_edma.c pa = skb_frag_dma_map(dev, frag, 0, len, DMA_TO_DEVICE); pa 1378 drivers/net/wireless/ath/wil6210/txrx_edma.c if (unlikely(dma_mapping_error(dev, pa))) { pa 1383 drivers/net/wireless/ath/wil6210/txrx_edma.c wil->txrx_ops.tx_desc_map((union wil_tx_desc *)d, pa, pa 583 drivers/net/wireless/ath/wil6210/txrx_edma.h dma_addr_t pa) pa 585 drivers/net/wireless/ath/wil6210/txrx_edma.h addr->addr_low = cpu_to_le32(lower_32_bits(pa)); pa 586 drivers/net/wireless/ath/wil6210/txrx_edma.h addr->addr_high = cpu_to_le16((u16)upper_32_bits(pa)); pa 587 drivers/net/wireless/ath/wil6210/txrx_edma.h *addr_high_high = cpu_to_le16((u16)(upper_32_bits(pa) >> 16)); pa 523 drivers/net/wireless/ath/wil6210/wil6210.h dma_addr_t pa; pa 532 drivers/net/wireless/ath/wil6210/wil6210.h dma_addr_t pa; pa 559 drivers/net/wireless/ath/wil6210/wil6210.h dma_addr_t pa; pa 611 drivers/net/wireless/ath/wil6210/wil6210.h int (*tx_desc_map)(union wil_tx_desc *desc, dma_addr_t pa, pa 2604 drivers/net/wireless/ath/wil6210/wmi.c .ring_mem_base = cpu_to_le64(vring->pa), pa 3721 drivers/net/wireless/ath/wil6210/wmi.c cmd.ring_cfg.ring_mem_base = cpu_to_le64(sring->pa); pa 3800 drivers/net/wireless/ath/wil6210/wmi.c cmd.ring_cfg.ring_mem_base = cpu_to_le64(sring->pa); pa 3841 drivers/net/wireless/ath/wil6210/wmi.c cmd.ring_cfg.ring_mem_base = cpu_to_le64(ring->pa); pa 3842 drivers/net/wireless/ath/wil6210/wmi.c cmd.sw_tail_host_addr = cpu_to_le64(ring->edma_rx_swtail.pa); pa 3892 drivers/net/wireless/ath/wil6210/wmi.c cmd.ring_cfg.ring_mem_base = cpu_to_le64(ring->pa); pa 3938 drivers/net/wireless/ath/wil6210/wmi.c cmd.ring_cfg.ring_mem_base = cpu_to_le64(ring->pa); pa 1840 drivers/net/wireless/broadcom/b43/phy_lp.c bool rx, bool pa, struct lpphy_tx_gains *gains) pa 1877 drivers/net/wireless/broadcom/b43/phy_lp.c pa = false; pa 1885 drivers/net/wireless/broadcom/b43/phy_lp.c 0xFFF7, pa << 3); pa 1889 drivers/net/wireless/broadcom/b43/phy_lp.c 0xFFDF, pa << 5); pa 716 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c dma_addr_t pa, uint outidx, u32 *flags, u32 bufcount) pa 721 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c if ((di->dataoffsetlow == 0) || !(pa & PCI32ADDR_HIGH)) { pa 722 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c ddring[outidx].addrlow = cpu_to_le32(pa + di->dataoffsetlow); pa 730 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c ae = (pa & PCI32ADDR_HIGH) >> PCI32ADDR_HIGH_SHIFT; pa 731 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c pa &= ~PCI32ADDR_HIGH; pa 734 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c ddring[outidx].addrlow = cpu_to_le32(pa + di->dataoffsetlow); pa 774 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c _dma_ddtable_init(struct dma_info *di, uint direction, dma_addr_t pa) pa 778 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c di->xmtptrbase = pa; pa 780 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c di->rcvptrbase = pa; pa 784 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c || !(pa & PCI32ADDR_HIGH)) { pa 787 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c pa + di->ddoffsetlow); pa 792 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c pa + di->ddoffsetlow); pa 801 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c ae = (pa & PCI32ADDR_HIGH) >> PCI32ADDR_HIGH_SHIFT; pa 802 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c pa &= ~PCI32ADDR_HIGH; pa 806 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c pa + di->ddoffsetlow); pa 813 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c pa + di->ddoffsetlow); pa 873 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c dma_addr_t pa; pa 894 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c pa = le32_to_cpu(di->rxd64[i].addrlow) - di->dataoffsetlow; pa 897 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c dma_unmap_single(di->dmadev, pa, di->rxbufsize, DMA_FROM_DEVICE); pa 1031 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c dma_addr_t pa; pa 1080 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c pa = dma_map_single(di->dmadev, p->data, di->rxbufsize, pa 1082 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c if (dma_mapping_error(di->dmadev, pa)) { pa 1095 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c dma64_dd_upd(di, di->rxd64, pa, rxout, &flags, pa 1277 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c dma_addr_t pa; pa 1291 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c pa = dma_map_single(di->dmadev, data, len, DMA_TO_DEVICE); pa 1293 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c if (dma_mapping_error(di->dmadev, pa)) { pa 1306 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c dma64_dd_upd(di, di->txd64, pa, txout, &flags, len); pa 1512 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c dma_addr_t pa; pa 1515 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c pa = le32_to_cpu(di->txd64[i].addrlow) - di->dataoffsetlow; pa 1527 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c dma_unmap_single(di->dmadev, pa, size, DMA_TO_DEVICE); pa 175 drivers/of/fdt.c const char *p = nodename, *ps = p, *pa = NULL; pa 180 drivers/of/fdt.c pa = p; pa 186 drivers/of/fdt.c if (pa < ps) pa 187 drivers/of/fdt.c pa = p; pa 188 drivers/of/fdt.c len = (pa - ps) + 1; pa 559 drivers/parisc/ccio-dma.c register unsigned long pa; pa 570 drivers/parisc/ccio-dma.c pa = lpa(vba); pa 571 drivers/parisc/ccio-dma.c asm volatile("depw %1,31,12,%0" : "+r" (pa) : "r" (hints)); pa 572 drivers/parisc/ccio-dma.c ((u32 *)pdir_ptr)[1] = (u32) pa; pa 584 drivers/parisc/ccio-dma.c asm volatile ("extrd,u %1,15,4,%0" : "=r" (ci) : "r" (pa)); pa 585 drivers/parisc/ccio-dma.c asm volatile ("extrd,u %1,31,16,%0" : "+r" (pa) : "r" (pa)); pa 586 drivers/parisc/ccio-dma.c asm volatile ("depd %1,35,4,%0" : "+r" (pa) : "r" (ci)); pa 588 drivers/parisc/ccio-dma.c pa = 0; pa 597 drivers/parisc/ccio-dma.c asm volatile ("depw %1,15,12,%0" : "+r" (pa) : "r" (ci)); pa 599 drivers/parisc/ccio-dma.c ((u32 *)pdir_ptr)[0] = (u32) pa; pa 569 drivers/parisc/sba_iommu.c u64 pa; /* physical address */ pa 572 drivers/parisc/sba_iommu.c pa = lpa(vba); pa 573 drivers/parisc/sba_iommu.c pa &= IOVP_MASK; pa 576 drivers/parisc/sba_iommu.c pa |= (ci >> PAGE_SHIFT) & 0xff; /* move CI (8 bits) into lowest byte */ pa 578 drivers/parisc/sba_iommu.c pa |= SBA_PDIR_VALID_BIT; /* set "valid" bit */ pa 579 drivers/parisc/sba_iommu.c *pdir_ptr = cpu_to_le64(pa); /* swap and store into I/O Pdir */ pa 1527 drivers/pci/msi.c u32 *pa = data; pa 1528 drivers/pci/msi.c u8 bus = PCI_BUS_NUM(*pa); pa 1531 drivers/pci/msi.c *pa = alias; pa 436 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-req-buf-alloc-all, 0x01), pa 437 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-req-buf-alloc-rd, 0x02), pa 438 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-req-buf-alloc-wr, 0x03), pa 439 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-all-cp-req, 0x04), pa 440 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-cp-blk-req, 0x05), pa 441 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-cp-ptl-req, 0x06), pa 442 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-cp-rd-req, 0x07), pa 443 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-cp-wr-req, 0x08), pa 447 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-rd-shared-req-issued, 0x10), pa 448 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-rd-exclusive-req-issued, 0x11), pa 449 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-wr-invalidate-req-issued-stashable, 0x12), pa 450 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-wr-invalidate-req-issued-nonstashable, 0x13), pa 451 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-wr-back-req-issued-stashable, 0x14), pa 452 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-wr-back-req-issued-nonstashable, 0x15), pa 453 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-ptl-wr-req, 0x16), pa 454 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-ptl-rd-req, 0x17), pa 455 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-wr-back-clean-data, 0x18), pa 456 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-wr-back-cancelled-on-SS, 0x1b), pa 457 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-barrier-occurrence, 0x1c), pa 458 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-barrier-cycles, 0x1d), pa 459 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-total-cp-snoops, 0x20), pa 460 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-rd-shared-snoop, 0x21), pa 461 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-rd-shared-snoop-hit, 0x22), pa 462 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-rd-exclusive-snoop, 0x23), pa 463 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-rd-exclusive-snoop-hit, 0x24), pa 464 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-rd-wr-invalid-snoop, 0x25), pa 465 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-rd-wr-invalid-snoop-hit, 0x26), pa 466 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-req-buffer-full, 0x28), pa 479 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-axi0-rd-req, 0x01), pa 480 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-axi0-wr-req, 0x02), pa 481 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-axi1-rd-req, 0x03), pa 482 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-axi1-wr-req, 0x04), pa 123 drivers/pinctrl/pinctrl-single.c struct pinctrl_pin_desc *pa; pa 277 drivers/pinctrl/pinctrl-single.c size_t pa; pa 284 drivers/pinctrl/pinctrl-single.c pa = pcs->res->start + offset; pa 286 drivers/pinctrl/pinctrl-single.c seq_printf(s, "%zx %08x %s ", pa, val, DRIVER_NAME); pa 683 drivers/pinctrl/pinctrl-single.c pin = &pcs->pins.pa[i]; pa 715 drivers/pinctrl/pinctrl-single.c pcs->pins.pa = devm_kcalloc(pcs->dev, pa 716 drivers/pinctrl/pinctrl-single.c nr_pins, sizeof(*pcs->pins.pa), pa 718 drivers/pinctrl/pinctrl-single.c if (!pcs->pins.pa) pa 721 drivers/pinctrl/pinctrl-single.c pcs->desc.pins = pcs->pins.pa; pa 152 drivers/pinctrl/ti/pinctrl-ti-iodelay.c struct pinctrl_pin_desc *pa; pa 461 drivers/pinctrl/ti/pinctrl-ti-iodelay.c pd = &iod->pa[pin]; pa 666 drivers/pinctrl/ti/pinctrl-ti-iodelay.c pd = &iod->pa[pin]; pa 752 drivers/pinctrl/ti/pinctrl-ti-iodelay.c iod->pa = devm_kcalloc(dev, nr_pins, sizeof(*iod->pa), GFP_KERNEL); pa 753 drivers/pinctrl/ti/pinctrl-ti-iodelay.c if (!iod->pa) pa 756 drivers/pinctrl/ti/pinctrl-ti-iodelay.c iod->desc.pins = iod->pa; pa 762 drivers/pinctrl/ti/pinctrl-ti-iodelay.c pin = &iod->pa[i]; pa 703 drivers/remoteproc/remoteproc_core.c ret = iommu_map(rproc->domain, rsc->da, rsc->pa, rsc->len, rsc->flags); pa 721 drivers/remoteproc/remoteproc_core.c rsc->pa, rsc->da, rsc->len); pa 890 drivers/remoteproc/remoteproc_core.c rsc->name, rsc->da, rsc->pa, rsc->len, rsc->flags); pa 1177 drivers/remoteproc/remoteproc_core.c u64 pa; pa 1214 drivers/remoteproc/remoteproc_core.c pa = (u64)rproc_va_to_pa(entry->va); pa 1216 drivers/remoteproc/remoteproc_core.c pa = (u64)entry->dma; pa 1218 drivers/remoteproc/remoteproc_core.c if (((u64)pa) & HIGH_BITS_MASK) pa 1222 drivers/remoteproc/remoteproc_core.c rsc->pa = (u32)pa; pa 214 drivers/remoteproc/remoteproc_debugfs.c seq_printf(seq, " Physical Address 0x%x\n", c->pa); pa 224 drivers/remoteproc/remoteproc_debugfs.c seq_printf(seq, " Physical Address 0x%x\n", d->pa); pa 259 drivers/remoteproc/remoteproc_debugfs.c v->vring[j].pa); pa 347 drivers/remoteproc/remoteproc_virtio.c phys_addr_t pa; pa 363 drivers/remoteproc/remoteproc_virtio.c pa = rproc_va_to_pa(mem->va); pa 366 drivers/remoteproc/remoteproc_virtio.c pa = (phys_addr_t)mem->dma; pa 370 drivers/remoteproc/remoteproc_virtio.c ret = dma_declare_coherent_memory(dev, pa, pa 74 drivers/remoteproc/stm32_rproc.c static int stm32_rproc_pa_to_da(struct rproc *rproc, phys_addr_t pa, u64 *da) pa 83 drivers/remoteproc/stm32_rproc.c if (pa < p_mem->bus_addr || pa 84 drivers/remoteproc/stm32_rproc.c pa >= p_mem->bus_addr + p_mem->size) pa 86 drivers/remoteproc/stm32_rproc.c *da = pa - p_mem->bus_addr + p_mem->dev_addr; pa 87 drivers/remoteproc/stm32_rproc.c dev_dbg(rproc->dev.parent, "pa %pa to da %llx\n", &pa, *da); pa 58 drivers/s390/cio/vfio_ccw_cp.c static int pfn_array_alloc(struct pfn_array *pa, u64 iova, unsigned int len) pa 62 drivers/s390/cio/vfio_ccw_cp.c if (pa->pa_nr || pa->pa_iova_pfn) pa 65 drivers/s390/cio/vfio_ccw_cp.c pa->pa_iova = iova; pa 67 drivers/s390/cio/vfio_ccw_cp.c pa->pa_nr = ((iova & ~PAGE_MASK) + len + (PAGE_SIZE - 1)) >> PAGE_SHIFT; pa 68 drivers/s390/cio/vfio_ccw_cp.c if (!pa->pa_nr) pa 71 drivers/s390/cio/vfio_ccw_cp.c pa->pa_iova_pfn = kcalloc(pa->pa_nr, pa 72 drivers/s390/cio/vfio_ccw_cp.c sizeof(*pa->pa_iova_pfn) + pa 73 drivers/s390/cio/vfio_ccw_cp.c sizeof(*pa->pa_pfn), pa 75 drivers/s390/cio/vfio_ccw_cp.c if (unlikely(!pa->pa_iova_pfn)) { pa 76 drivers/s390/cio/vfio_ccw_cp.c pa->pa_nr = 0; pa 79 drivers/s390/cio/vfio_ccw_cp.c pa->pa_pfn = pa->pa_iova_pfn + pa->pa_nr; pa 81 drivers/s390/cio/vfio_ccw_cp.c pa->pa_iova_pfn[0] = pa->pa_iova >> PAGE_SHIFT; pa 82 drivers/s390/cio/vfio_ccw_cp.c pa->pa_pfn[0] = -1ULL; pa 83 drivers/s390/cio/vfio_ccw_cp.c for (i = 1; i < pa->pa_nr; i++) { pa 84 drivers/s390/cio/vfio_ccw_cp.c pa->pa_iova_pfn[i] = pa->pa_iova_pfn[i - 1] + 1; pa 85 drivers/s390/cio/vfio_ccw_cp.c pa->pa_pfn[i] = -1ULL; pa 100 drivers/s390/cio/vfio_ccw_cp.c static int pfn_array_pin(struct pfn_array *pa, struct device *mdev) pa 104 drivers/s390/cio/vfio_ccw_cp.c ret = vfio_pin_pages(mdev, pa->pa_iova_pfn, pa->pa_nr, pa 105 drivers/s390/cio/vfio_ccw_cp.c IOMMU_READ | IOMMU_WRITE, pa->pa_pfn); pa 109 drivers/s390/cio/vfio_ccw_cp.c } else if (ret > 0 && ret != pa->pa_nr) { pa 110 drivers/s390/cio/vfio_ccw_cp.c vfio_unpin_pages(mdev, pa->pa_iova_pfn, ret); pa 118 drivers/s390/cio/vfio_ccw_cp.c pa->pa_nr = 0; pa 124 drivers/s390/cio/vfio_ccw_cp.c static void pfn_array_unpin_free(struct pfn_array *pa, struct device *mdev) pa 127 drivers/s390/cio/vfio_ccw_cp.c if (pa->pa_nr) pa 128 drivers/s390/cio/vfio_ccw_cp.c vfio_unpin_pages(mdev, pa->pa_iova_pfn, pa->pa_nr); pa 129 drivers/s390/cio/vfio_ccw_cp.c pa->pa_nr = 0; pa 130 drivers/s390/cio/vfio_ccw_cp.c kfree(pa->pa_iova_pfn); pa 133 drivers/s390/cio/vfio_ccw_cp.c static bool pfn_array_iova_pinned(struct pfn_array *pa, unsigned long iova) pa 138 drivers/s390/cio/vfio_ccw_cp.c for (i = 0; i < pa->pa_nr; i++) pa 139 drivers/s390/cio/vfio_ccw_cp.c if (pa->pa_iova_pfn[i] == iova_pfn) pa 146 drivers/s390/cio/vfio_ccw_cp.c struct pfn_array *pa, pa 159 drivers/s390/cio/vfio_ccw_cp.c for (i = 0; i < pa->pa_nr; i++) pa 160 drivers/s390/cio/vfio_ccw_cp.c idaws[i] = pa->pa_pfn[i] << PAGE_SHIFT; pa 163 drivers/s390/cio/vfio_ccw_cp.c idaws[0] += pa->pa_iova & (PAGE_SIZE - 1); pa 196 drivers/s390/cio/vfio_ccw_cp.c struct pfn_array pa = {0}; pa 201 drivers/s390/cio/vfio_ccw_cp.c ret = pfn_array_alloc(&pa, iova, n); pa 205 drivers/s390/cio/vfio_ccw_cp.c ret = pfn_array_pin(&pa, mdev); pa 207 drivers/s390/cio/vfio_ccw_cp.c pfn_array_unpin_free(&pa, mdev); pa 212 drivers/s390/cio/vfio_ccw_cp.c for (i = 0; i < pa.pa_nr; i++) { pa 213 drivers/s390/cio/vfio_ccw_cp.c from = pa.pa_pfn[i] << PAGE_SHIFT; pa 228 drivers/s390/cio/vfio_ccw_cp.c pfn_array_unpin_free(&pa, mdev); pa 511 drivers/s390/cio/vfio_ccw_cp.c struct pfn_array *pa; pa 550 drivers/s390/cio/vfio_ccw_cp.c pa = chain->ch_pa + idx; pa 551 drivers/s390/cio/vfio_ccw_cp.c ret = pfn_array_alloc(pa, iova, bytes); pa 566 drivers/s390/cio/vfio_ccw_cp.c pa->pa_iova_pfn[i] = idaws[i] >> PAGE_SHIFT; pa 576 drivers/s390/cio/vfio_ccw_cp.c ret = pfn_array_pin(pa, cp->mdev); pa 580 drivers/s390/cio/vfio_ccw_cp.c pa->pa_nr = 0; pa 587 drivers/s390/cio/vfio_ccw_cp.c pfn_array_idal_create_words(pa, idaws); pa 592 drivers/s390/cio/vfio_ccw_cp.c pfn_array_unpin_free(pa, cp->mdev); pa 1544 drivers/scsi/be2iscsi/be_main.c if (pasync_handle->pa.u.a64.address != phys_addr.u.a64.address || pa 1551 drivers/scsi/be2iscsi/be_main.c pasync_handle->pa.u.a64.address, pa 1744 drivers/scsi/be2iscsi/be_main.c pasync_sge[pi].hi = pasync_handle->pa.u.a32.address_lo; pa 1745 drivers/scsi/be2iscsi/be_main.c pasync_sge[pi].lo = pasync_handle->pa.u.a32.address_hi; pa 2880 drivers/scsi/be2iscsi/be_main.c pasync_header_h->pa.u.a64.address = pa 2915 drivers/scsi/be2iscsi/be_main.c pasync_data_h->pa.u.a64.address = pa 545 drivers/scsi/be2iscsi/be_main.h struct be_bus_address pa; pa 143 drivers/scsi/be2iscsi/be_mgmt.h #define ISCSI_GET_PDU_TEMPLATE_ADDRESS(pc, pa) {\ pa 144 drivers/scsi/be2iscsi/be_mgmt.h pa->lo = phba->init_mem[ISCSI_MEM_GLOBAL_HEADER].mem_array[0].\ pa 146 drivers/scsi/be2iscsi/be_mgmt.h pa->hi = phba->init_mem[ISCSI_MEM_GLOBAL_HEADER].mem_array[0].\ pa 997 drivers/scsi/bfa/bfa_core.c bfa_dma_be_addr_set(cfg_info->cfgrsp_addr, iocfc->cfgrsp_dma.pa); pa 1003 drivers/scsi/bfa/bfa_core.c iocfc->req_cq_ba[i].pa); pa 1005 drivers/scsi/bfa/bfa_core.c iocfc->req_cq_shadow_ci[i].pa); pa 1010 drivers/scsi/bfa/bfa_core.c iocfc->rsp_cq_ba[i].pa); pa 1012 drivers/scsi/bfa/bfa_core.c iocfc->rsp_cq_shadow_pi[i].pa); pa 1029 drivers/scsi/bfa/bfa_core.c bfa_dma_be_addr_set(cfg_req.ioc_cfg_dma_addr, iocfc->cfg_info.pa); pa 1112 drivers/scsi/bfa/bfa_core.c iocfc->req_cq_ba[i].pa = bfa_mem_dma_phys(reqq_dma); pa 1117 drivers/scsi/bfa/bfa_core.c iocfc->rsp_cq_ba[i].pa = bfa_mem_dma_phys(rspq_dma); pa 1127 drivers/scsi/bfa/bfa_core.c iocfc->req_cq_shadow_ci[i].pa = dm_pa; pa 1132 drivers/scsi/bfa/bfa_core.c iocfc->rsp_cq_shadow_pi[i].pa = dm_pa; pa 1139 drivers/scsi/bfa/bfa_core.c bfa->iocfc.cfg_info.pa = dm_pa; pa 1146 drivers/scsi/bfa/bfa_core.c bfa->iocfc.cfgrsp_dma.pa = dm_pa; pa 3713 drivers/scsi/bfa/bfa_fcpim.c fcp->snsbase[idx].pa = bfa_mem_dma_phys(seg_ptr); pa 3715 drivers/scsi/bfa/bfa_fcpim.c bfa_iocfc_set_snsbase(bfa, idx, fcp->snsbase[idx].pa); pa 1829 drivers/scsi/bfa/bfa_ioc.c bfa_dma_be_addr_set(attr_req.attr_addr, ioc->attr_dma.pa); pa 2490 drivers/scsi/bfa/bfa_ioc.c ioc->attr_dma.pa = dm_pa; pa 3372 drivers/scsi/bfa/bfa_ioc.c ablk->dma_addr.pa = dma_pa; pa 3412 drivers/scsi/bfa/bfa_ioc.c bfa_dma_be_addr_set(m->addr, ablk->dma_addr.pa); pa 163 drivers/scsi/bfa/bfa_ioc.h u64 pa; /* ! Physical address */ pa 175 drivers/scsi/bfa/bfa_ioc.h #define bfa_dma_be_addr_set(dma_addr, pa) \ pa 176 drivers/scsi/bfa/bfa_ioc.h __bfa_dma_be_addr_set(&dma_addr, (u64)pa) pa 178 drivers/scsi/bfa/bfa_ioc.h __bfa_dma_be_addr_set(union bfi_addr_u *dma_addr, u64 pa) pa 180 drivers/scsi/bfa/bfa_ioc.h dma_addr->a32.addr_lo = cpu_to_be32(pa); pa 181 drivers/scsi/bfa/bfa_ioc.h dma_addr->a32.addr_hi = cpu_to_be32(pa >> 32); pa 188 drivers/scsi/bfa/bfa_ioc.h __bfa_alen_set(struct bfi_alen_s *alen, u32 len, u64 pa) pa 191 drivers/scsi/bfa/bfa_ioc.h bfa_dma_be_addr_set(alen->al_addr, pa); pa 206 drivers/scsi/bfa/bfa_port.c port->stats_dma.pa = dma_pa; pa 352 drivers/scsi/bfa/bfa_port.c bfa_dma_be_addr_set(m->dma_addr, port->stats_dma.pa); pa 623 drivers/scsi/bfa/bfa_port.c cee->attr_dma.pa = dma_pa; pa 626 drivers/scsi/bfa/bfa_port.c cee->stats_dma.pa = dma_pa + BFA_ROUNDUP( pa 667 drivers/scsi/bfa/bfa_port.c bfa_dma_be_addr_set(cmd->dma_addr, cee->attr_dma.pa); pa 707 drivers/scsi/bfa/bfa_port.c bfa_dma_be_addr_set(cmd->dma_addr, cee->stats_dma.pa); pa 5191 drivers/scsi/bfa/bfa_svc.c u64 pa; pa 5218 drivers/scsi/bfa/bfa_svc.c sgpg_pa.pa = bfa_mem_dma_phys(seg_ptr) + align_len; pa 5219 drivers/scsi/bfa/bfa_svc.c WARN_ON(sgpg_pa.pa & (sgpg_sz - 1)); pa 5228 drivers/scsi/bfa/bfa_svc.c sgpg_pa_tmp.pa = bfa_sgaddr_le(sgpg_pa.pa); pa 5234 drivers/scsi/bfa/bfa_svc.c sgpg_pa.pa += sgpg_sz; pa 982 drivers/scsi/fnic/fnic_fcs.c dma_addr_t pa; pa 996 drivers/scsi/fnic/fnic_fcs.c pa = dma_map_single(&fnic->pdev->dev, skb->data, len, DMA_FROM_DEVICE); pa 997 drivers/scsi/fnic/fnic_fcs.c if (dma_mapping_error(&fnic->pdev->dev, pa)) { pa 1003 drivers/scsi/fnic/fnic_fcs.c fnic_queue_rq_desc(rq, skb, pa, len); pa 1032 drivers/scsi/fnic/fnic_fcs.c dma_addr_t pa; pa 1055 drivers/scsi/fnic/fnic_fcs.c pa = dma_map_single(&fnic->pdev->dev, skb->data, skb->len, pa 1057 drivers/scsi/fnic/fnic_fcs.c if (dma_mapping_error(&fnic->pdev->dev, pa)) { pa 1066 drivers/scsi/fnic/fnic_fcs.c fnic_queue_wq_eth_desc(wq, skb, pa, skb->len, pa 1074 drivers/scsi/fnic/fnic_fcs.c dma_unmap_single(&fnic->pdev->dev, pa, skb->len, DMA_TO_DEVICE); pa 1086 drivers/scsi/fnic/fnic_fcs.c dma_addr_t pa; pa 1131 drivers/scsi/fnic/fnic_fcs.c pa = dma_map_single(&fnic->pdev->dev, eth_hdr, tot_len, DMA_TO_DEVICE); pa 1132 drivers/scsi/fnic/fnic_fcs.c if (dma_mapping_error(&fnic->pdev->dev, pa)) { pa 1146 drivers/scsi/fnic/fnic_fcs.c dma_unmap_single(&fnic->pdev->dev, pa, tot_len, DMA_TO_DEVICE); pa 1151 drivers/scsi/fnic/fnic_fcs.c fnic_queue_wq_desc(wq, skb, pa, tot_len, fr_eof(fp), pa 3195 drivers/scsi/lpfc/lpfc_hbadisc.c } pa; pa 3215 drivers/scsi/lpfc/lpfc_hbadisc.c un.pa.wd1, un.pa.wd2, pa 3216 drivers/scsi/lpfc/lpfc_hbadisc.c un.pa.wd3, un.pa.wd4); pa 85 drivers/scsi/snic/snic_disc.c dma_addr_t pa = 0; pa 114 drivers/scsi/snic/snic_disc.c pa = dma_map_single(&snic->pdev->dev, buf, buf_len, DMA_FROM_DEVICE); pa 115 drivers/scsi/snic/snic_disc.c if (dma_mapping_error(&snic->pdev->dev, pa)) { pa 127 drivers/scsi/snic/snic_disc.c SNIC_BUG_ON(pa == 0); pa 134 drivers/scsi/snic/snic_disc.c pa, pa 141 drivers/scsi/snic/snic_disc.c dma_unmap_single(&snic->pdev->dev, pa, buf_len, pa 165 drivers/scsi/snic/snic_io.c dma_addr_t pa = 0; pa 176 drivers/scsi/snic/snic_io.c pa = dma_map_single(&snic->pdev->dev, os_buf, len, DMA_TO_DEVICE); pa 177 drivers/scsi/snic/snic_io.c if (dma_mapping_error(&snic->pdev->dev, pa)) { pa 183 drivers/scsi/snic/snic_io.c req->req_pa = (ulong)pa; pa 190 drivers/scsi/snic/snic_io.c dma_unmap_single(&snic->pdev->dev, pa, len, DMA_TO_DEVICE); pa 199 drivers/scsi/snic/snic_io.c snic_queue_wq_eth_desc(&snic->wq[q_num], os_buf, pa, len, 0, 0, 1); pa 170 drivers/scsi/snic/snic_scsi.c dma_addr_t pa = 0; pa 188 drivers/scsi/snic/snic_scsi.c pa = dma_map_single(&snic->pdev->dev, pa 192 drivers/scsi/snic/snic_scsi.c if (dma_mapping_error(&snic->pdev->dev, pa)) { pa 220 drivers/scsi/snic/snic_scsi.c pa, /* sense buffer pa */ pa 247 drivers/scsi/sym53c8xx_2/sym_fw.c u32 *pa; pa 256 drivers/scsi/sym53c8xx_2/sym_fw.c pa = (u32 *) &np->fwa_bas; pa 258 drivers/scsi/sym53c8xx_2/sym_fw.c pa[i] = np->scripta_ba + po[i]; pa 264 drivers/scsi/sym53c8xx_2/sym_fw.c pa = (u32 *) &np->fwb_bas; pa 266 drivers/scsi/sym53c8xx_2/sym_fw.c pa[i] = np->scriptb_ba + po[i]; pa 272 drivers/scsi/sym53c8xx_2/sym_fw.c pa = (u32 *) &np->fwz_bas; pa 274 drivers/scsi/sym53c8xx_2/sym_fw.c pa[i] = np->scriptz_ba + po[i]; pa 121 drivers/sfi/sfi_core.c static void sfi_print_table_header(unsigned long long pa, pa 125 drivers/sfi/sfi_core.c header->sig, pa, pa 165 drivers/sfi/sfi_core.c static struct sfi_table_header *sfi_map_table(u64 pa) pa 170 drivers/sfi/sfi_core.c if (!TABLE_ON_PAGE(syst_pa, pa, sizeof(struct sfi_table_header))) pa 171 drivers/sfi/sfi_core.c th = sfi_map_memory(pa, sizeof(struct sfi_table_header)); pa 173 drivers/sfi/sfi_core.c th = (void *)syst_va + (pa - syst_pa); pa 181 drivers/sfi/sfi_core.c if (!TABLE_ON_PAGE(syst_pa, pa, sizeof(struct sfi_table_header))) pa 184 drivers/sfi/sfi_core.c return sfi_map_memory(pa, length); pa 232 drivers/sfi/sfi_core.c __ref sfi_check_table(u64 pa, struct sfi_table_key *key) pa 237 drivers/sfi/sfi_core.c th = sfi_map_table(pa); pa 242 drivers/sfi/sfi_core.c sfi_print_table_header(pa, th); pa 419 drivers/sfi/sfi_core.c struct sfi_table_attr __init *sfi_sysfs_install_table(u64 pa) pa 429 drivers/sfi/sfi_core.c th = sfi_map_table(pa); pa 80 drivers/sfi/sfi_core.h extern struct sfi_table_attr __init *sfi_sysfs_install_table(u64 pa); pa 178 drivers/staging/gdm724x/gdm_lte.c u16 pa; pa 188 drivers/staging/gdm724x/gdm_lte.c u16 pa[20]; pa 198 drivers/staging/gdm724x/gdm_lte.c for (i = 0; i < ARRAY_SIZE(pseudo_header.pa); i++) { pa 199 drivers/staging/gdm724x/gdm_lte.c pa = pseudo_header.pa[i]; pa 200 drivers/staging/gdm724x/gdm_lte.c sum = csum_add(sum, csum_unfold((__force __sum16)pa)); pa 42 drivers/tee/optee/core.c phys_addr_t pa; pa 76 drivers/tee/optee/core.c rc = tee_shm_get_pa(shm, 0, &pa); pa 79 drivers/tee/optee/core.c p->u.memref.shm_offs = mp->u.tmem.buf_ptr - pa; pa 122 drivers/tee/optee/core.c phys_addr_t pa; pa 135 drivers/tee/optee/core.c rc = tee_shm_get_pa(p->u.memref.shm, p->u.memref.shm_offs, &pa); pa 139 drivers/tee/optee/core.c mp->u.tmem.buf_ptr = pa; pa 197 drivers/tee/optee/rpc.c phys_addr_t pa; pa 235 drivers/tee/optee/rpc.c if (tee_shm_get_pa(shm, 0, &pa)) { pa 278 drivers/tee/optee/rpc.c arg->params[0].u.tmem.buf_ptr = pa; pa 404 drivers/tee/optee/rpc.c phys_addr_t pa; pa 409 drivers/tee/optee/rpc.c if (!IS_ERR(shm) && !tee_shm_get_pa(shm, 0, &pa)) { pa 410 drivers/tee/optee/rpc.c reg_pair_from_64(¶m->a1, ¶m->a2, pa); pa 383 drivers/tee/tee_shm.c int tee_shm_va2pa(struct tee_shm *shm, void *va, phys_addr_t *pa) pa 394 drivers/tee/tee_shm.c shm, (unsigned long)va - (unsigned long)shm->kaddr, pa); pa 405 drivers/tee/tee_shm.c int tee_shm_pa2va(struct tee_shm *shm, phys_addr_t pa, void **va) pa 410 drivers/tee/tee_shm.c if (pa < shm->paddr) pa 412 drivers/tee/tee_shm.c if (pa >= (shm->paddr + shm->size)) pa 416 drivers/tee/tee_shm.c void *v = tee_shm_get_va(shm, pa - shm->paddr); pa 451 drivers/tee/tee_shm.c int tee_shm_get_pa(struct tee_shm *shm, size_t offs, phys_addr_t *pa) pa 455 drivers/tee/tee_shm.c if (pa) pa 456 drivers/tee/tee_shm.c *pa = shm->paddr + offs; pa 127 drivers/tty/serial/8250/8250_hp300.c unsigned long pa = dio_scodetophysaddr(scode); pa 128 drivers/tty/serial/8250/8250_hp300.c if (!pa) pa 134 drivers/tty/serial/8250/8250_hp300.c port.mapbase = (pa + UART_OFFSET); pa 137 drivers/tty/serial/8250/8250_hp300.c port.irq = DIO_IPL(pa + DIO_VIRADDRBASE); pa 140 drivers/tty/serial/8250/8250_hp300.c out_8(pa + DIO_VIRADDRBASE + DCA_IC, DCA_IC_IE); pa 142 drivers/tty/serial/8250/8250_hp300.c if (DIO_ID(pa + DIO_VIRADDRBASE) & 0x80) pa 631 fs/ext4/mballoc.c struct ext4_prealloc_space *pa; pa 632 fs/ext4/mballoc.c pa = list_entry(cur, struct ext4_prealloc_space, pa_group_list); pa 633 fs/ext4/mballoc.c ext4_get_group_no_and_offset(sb, pa->pa_pstart, &groupnr, &k); pa 635 fs/ext4/mballoc.c for (i = 0; i < pa->pa_len; i++) pa 2700 fs/ext4/mballoc.c struct ext4_prealloc_space *pa; pa 2705 fs/ext4/mballoc.c pa = list_entry(cur, struct ext4_prealloc_space, pa_group_list); pa 2706 fs/ext4/mballoc.c list_del(&pa->pa_group_list); pa 2708 fs/ext4/mballoc.c kmem_cache_free(ext4_pspace_cachep, pa); pa 3092 fs/ext4/mballoc.c struct ext4_prealloc_space *pa; pa 3186 fs/ext4/mballoc.c list_for_each_entry_rcu(pa, &ei->i_prealloc_list, pa_inode_list) { pa 3189 fs/ext4/mballoc.c if (pa->pa_deleted) pa 3191 fs/ext4/mballoc.c spin_lock(&pa->pa_lock); pa 3192 fs/ext4/mballoc.c if (pa->pa_deleted) { pa 3193 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 3197 fs/ext4/mballoc.c pa_end = pa->pa_lstart + EXT4_C2B(EXT4_SB(ac->ac_sb), pa 3198 fs/ext4/mballoc.c pa->pa_len); pa 3202 fs/ext4/mballoc.c ac->ac_o_ex.fe_logical < pa->pa_lstart)); pa 3205 fs/ext4/mballoc.c if (pa->pa_lstart >= end || pa_end <= start) { pa 3206 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 3209 fs/ext4/mballoc.c BUG_ON(pa->pa_lstart <= start && pa_end >= end); pa 3215 fs/ext4/mballoc.c } else if (pa->pa_lstart > ac->ac_o_ex.fe_logical) { pa 3216 fs/ext4/mballoc.c BUG_ON(pa->pa_lstart > end); pa 3217 fs/ext4/mballoc.c end = pa->pa_lstart; pa 3219 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 3226 fs/ext4/mballoc.c list_for_each_entry_rcu(pa, &ei->i_prealloc_list, pa_inode_list) { pa 3229 fs/ext4/mballoc.c spin_lock(&pa->pa_lock); pa 3230 fs/ext4/mballoc.c if (pa->pa_deleted == 0) { pa 3231 fs/ext4/mballoc.c pa_end = pa->pa_lstart + EXT4_C2B(EXT4_SB(ac->ac_sb), pa 3232 fs/ext4/mballoc.c pa->pa_len); pa 3233 fs/ext4/mballoc.c BUG_ON(!(start >= pa_end || end <= pa->pa_lstart)); pa 3235 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 3307 fs/ext4/mballoc.c struct ext4_prealloc_space *pa = ac->ac_pa; pa 3311 fs/ext4/mballoc.c if (pa == NULL) { pa 3331 fs/ext4/mballoc.c if (pa->pa_type == MB_INODE_PA) pa 3332 fs/ext4/mballoc.c pa->pa_free += ac->ac_b_ex.fe_len; pa 3339 fs/ext4/mballoc.c struct ext4_prealloc_space *pa) pa 3347 fs/ext4/mballoc.c start = pa->pa_pstart + (ac->ac_o_ex.fe_logical - pa->pa_lstart); pa 3348 fs/ext4/mballoc.c end = min(pa->pa_pstart + EXT4_C2B(sbi, pa->pa_len), pa 3355 fs/ext4/mballoc.c ac->ac_pa = pa; pa 3357 fs/ext4/mballoc.c BUG_ON(start < pa->pa_pstart); pa 3358 fs/ext4/mballoc.c BUG_ON(end > pa->pa_pstart + EXT4_C2B(sbi, pa->pa_len)); pa 3359 fs/ext4/mballoc.c BUG_ON(pa->pa_free < len); pa 3360 fs/ext4/mballoc.c pa->pa_free -= len; pa 3362 fs/ext4/mballoc.c mb_debug(1, "use %llu/%u from inode pa %p\n", start, len, pa); pa 3369 fs/ext4/mballoc.c struct ext4_prealloc_space *pa) pa 3373 fs/ext4/mballoc.c ext4_get_group_no_and_offset(ac->ac_sb, pa->pa_pstart, pa 3378 fs/ext4/mballoc.c ac->ac_pa = pa; pa 3386 fs/ext4/mballoc.c mb_debug(1, "use %u/%u from group pa %p\n", pa->pa_lstart-len, len, pa); pa 3397 fs/ext4/mballoc.c struct ext4_prealloc_space *pa, pa 3403 fs/ext4/mballoc.c atomic_inc(&pa->pa_count); pa 3404 fs/ext4/mballoc.c return pa; pa 3407 fs/ext4/mballoc.c new_distance = abs(goal_block - pa->pa_pstart); pa 3414 fs/ext4/mballoc.c atomic_inc(&pa->pa_count); pa 3415 fs/ext4/mballoc.c return pa; pa 3428 fs/ext4/mballoc.c struct ext4_prealloc_space *pa, *cpa = NULL; pa 3437 fs/ext4/mballoc.c list_for_each_entry_rcu(pa, &ei->i_prealloc_list, pa_inode_list) { pa 3441 fs/ext4/mballoc.c if (ac->ac_o_ex.fe_logical < pa->pa_lstart || pa 3442 fs/ext4/mballoc.c ac->ac_o_ex.fe_logical >= (pa->pa_lstart + pa 3443 fs/ext4/mballoc.c EXT4_C2B(sbi, pa->pa_len))) pa 3448 fs/ext4/mballoc.c (pa->pa_pstart + EXT4_C2B(sbi, pa->pa_len) > pa 3453 fs/ext4/mballoc.c spin_lock(&pa->pa_lock); pa 3454 fs/ext4/mballoc.c if (pa->pa_deleted == 0 && pa->pa_free) { pa 3455 fs/ext4/mballoc.c atomic_inc(&pa->pa_count); pa 3456 fs/ext4/mballoc.c ext4_mb_use_inode_pa(ac, pa); pa 3457 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 3462 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 3486 fs/ext4/mballoc.c list_for_each_entry_rcu(pa, &lg->lg_prealloc_list[i], pa 3488 fs/ext4/mballoc.c spin_lock(&pa->pa_lock); pa 3489 fs/ext4/mballoc.c if (pa->pa_deleted == 0 && pa 3490 fs/ext4/mballoc.c pa->pa_free >= ac->ac_o_ex.fe_len) { pa 3493 fs/ext4/mballoc.c pa, cpa); pa 3495 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 3541 fs/ext4/mballoc.c struct ext4_prealloc_space *pa; pa 3557 fs/ext4/mballoc.c pa = list_entry(cur, struct ext4_prealloc_space, pa_group_list); pa 3558 fs/ext4/mballoc.c spin_lock(&pa->pa_lock); pa 3559 fs/ext4/mballoc.c ext4_get_group_no_and_offset(sb, pa->pa_pstart, pa 3561 fs/ext4/mballoc.c len = pa->pa_len; pa 3562 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 3574 fs/ext4/mballoc.c struct ext4_prealloc_space *pa; pa 3575 fs/ext4/mballoc.c pa = container_of(head, struct ext4_prealloc_space, u.pa_rcu); pa 3577 fs/ext4/mballoc.c BUG_ON(atomic_read(&pa->pa_count)); pa 3578 fs/ext4/mballoc.c BUG_ON(pa->pa_deleted == 0); pa 3579 fs/ext4/mballoc.c kmem_cache_free(ext4_pspace_cachep, pa); pa 3587 fs/ext4/mballoc.c struct super_block *sb, struct ext4_prealloc_space *pa) pa 3593 fs/ext4/mballoc.c spin_lock(&pa->pa_lock); pa 3594 fs/ext4/mballoc.c if (!atomic_dec_and_test(&pa->pa_count) || pa->pa_free != 0) { pa 3595 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 3599 fs/ext4/mballoc.c if (pa->pa_deleted == 1) { pa 3600 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 3604 fs/ext4/mballoc.c pa->pa_deleted = 1; pa 3605 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 3607 fs/ext4/mballoc.c grp_blk = pa->pa_pstart; pa 3612 fs/ext4/mballoc.c if (pa->pa_type == MB_GROUP_PA) pa 3632 fs/ext4/mballoc.c list_del(&pa->pa_group_list); pa 3635 fs/ext4/mballoc.c spin_lock(pa->pa_obj_lock); pa 3636 fs/ext4/mballoc.c list_del_rcu(&pa->pa_inode_list); pa 3637 fs/ext4/mballoc.c spin_unlock(pa->pa_obj_lock); pa 3639 fs/ext4/mballoc.c call_rcu(&(pa)->u.pa_rcu, ext4_mb_pa_callback); pa 3650 fs/ext4/mballoc.c struct ext4_prealloc_space *pa; pa 3659 fs/ext4/mballoc.c pa = kmem_cache_alloc(ext4_pspace_cachep, GFP_NOFS); pa 3660 fs/ext4/mballoc.c if (pa == NULL) pa 3701 fs/ext4/mballoc.c pa->pa_lstart = ac->ac_b_ex.fe_logical; pa 3702 fs/ext4/mballoc.c pa->pa_pstart = ext4_grp_offs_to_block(sb, &ac->ac_b_ex); pa 3703 fs/ext4/mballoc.c pa->pa_len = ac->ac_b_ex.fe_len; pa 3704 fs/ext4/mballoc.c pa->pa_free = pa->pa_len; pa 3705 fs/ext4/mballoc.c atomic_set(&pa->pa_count, 1); pa 3706 fs/ext4/mballoc.c spin_lock_init(&pa->pa_lock); pa 3707 fs/ext4/mballoc.c INIT_LIST_HEAD(&pa->pa_inode_list); pa 3708 fs/ext4/mballoc.c INIT_LIST_HEAD(&pa->pa_group_list); pa 3709 fs/ext4/mballoc.c pa->pa_deleted = 0; pa 3710 fs/ext4/mballoc.c pa->pa_type = MB_INODE_PA; pa 3712 fs/ext4/mballoc.c mb_debug(1, "new inode pa %p: %llu/%u for %u\n", pa, pa 3713 fs/ext4/mballoc.c pa->pa_pstart, pa->pa_len, pa->pa_lstart); pa 3714 fs/ext4/mballoc.c trace_ext4_mb_new_inode_pa(ac, pa); pa 3716 fs/ext4/mballoc.c ext4_mb_use_inode_pa(ac, pa); pa 3717 fs/ext4/mballoc.c atomic_add(pa->pa_free, &sbi->s_mb_preallocated); pa 3722 fs/ext4/mballoc.c pa->pa_obj_lock = &ei->i_prealloc_lock; pa 3723 fs/ext4/mballoc.c pa->pa_inode = ac->ac_inode; pa 3726 fs/ext4/mballoc.c list_add(&pa->pa_group_list, &grp->bb_prealloc_list); pa 3729 fs/ext4/mballoc.c spin_lock(pa->pa_obj_lock); pa 3730 fs/ext4/mballoc.c list_add_rcu(&pa->pa_inode_list, &ei->i_prealloc_list); pa 3731 fs/ext4/mballoc.c spin_unlock(pa->pa_obj_lock); pa 3744 fs/ext4/mballoc.c struct ext4_prealloc_space *pa; pa 3753 fs/ext4/mballoc.c pa = kmem_cache_alloc(ext4_pspace_cachep, GFP_NOFS); pa 3754 fs/ext4/mballoc.c if (pa == NULL) pa 3761 fs/ext4/mballoc.c pa->pa_pstart = ext4_grp_offs_to_block(sb, &ac->ac_b_ex); pa 3762 fs/ext4/mballoc.c pa->pa_lstart = pa->pa_pstart; pa 3763 fs/ext4/mballoc.c pa->pa_len = ac->ac_b_ex.fe_len; pa 3764 fs/ext4/mballoc.c pa->pa_free = pa->pa_len; pa 3765 fs/ext4/mballoc.c atomic_set(&pa->pa_count, 1); pa 3766 fs/ext4/mballoc.c spin_lock_init(&pa->pa_lock); pa 3767 fs/ext4/mballoc.c INIT_LIST_HEAD(&pa->pa_inode_list); pa 3768 fs/ext4/mballoc.c INIT_LIST_HEAD(&pa->pa_group_list); pa 3769 fs/ext4/mballoc.c pa->pa_deleted = 0; pa 3770 fs/ext4/mballoc.c pa->pa_type = MB_GROUP_PA; pa 3772 fs/ext4/mballoc.c mb_debug(1, "new group pa %p: %llu/%u for %u\n", pa, pa 3773 fs/ext4/mballoc.c pa->pa_pstart, pa->pa_len, pa->pa_lstart); pa 3774 fs/ext4/mballoc.c trace_ext4_mb_new_group_pa(ac, pa); pa 3776 fs/ext4/mballoc.c ext4_mb_use_group_pa(ac, pa); pa 3777 fs/ext4/mballoc.c atomic_add(pa->pa_free, &EXT4_SB(sb)->s_mb_preallocated); pa 3783 fs/ext4/mballoc.c pa->pa_obj_lock = &lg->lg_prealloc_lock; pa 3784 fs/ext4/mballoc.c pa->pa_inode = NULL; pa 3787 fs/ext4/mballoc.c list_add(&pa->pa_group_list, &grp->bb_prealloc_list); pa 3818 fs/ext4/mballoc.c struct ext4_prealloc_space *pa) pa 3829 fs/ext4/mballoc.c BUG_ON(pa->pa_deleted == 0); pa 3830 fs/ext4/mballoc.c ext4_get_group_no_and_offset(sb, pa->pa_pstart, &group, &bit); pa 3831 fs/ext4/mballoc.c grp_blk_start = pa->pa_pstart - EXT4_C2B(sbi, bit); pa 3832 fs/ext4/mballoc.c BUG_ON(group != e4b->bd_group && pa->pa_len != 0); pa 3833 fs/ext4/mballoc.c end = bit + pa->pa_len; pa 3846 fs/ext4/mballoc.c trace_ext4_mb_release_inode_pa(pa, (grp_blk_start + pa 3849 fs/ext4/mballoc.c mb_free_blocks(pa->pa_inode, e4b, bit, next - bit); pa 3852 fs/ext4/mballoc.c if (free != pa->pa_free) { pa 3855 fs/ext4/mballoc.c pa, (unsigned long) pa->pa_lstart, pa 3856 fs/ext4/mballoc.c (unsigned long) pa->pa_pstart, pa 3857 fs/ext4/mballoc.c (unsigned long) pa->pa_len); pa 3859 fs/ext4/mballoc.c free, pa->pa_free); pa 3872 fs/ext4/mballoc.c struct ext4_prealloc_space *pa) pa 3878 fs/ext4/mballoc.c trace_ext4_mb_release_group_pa(sb, pa); pa 3879 fs/ext4/mballoc.c BUG_ON(pa->pa_deleted == 0); pa 3880 fs/ext4/mballoc.c ext4_get_group_no_and_offset(sb, pa->pa_pstart, &group, &bit); pa 3881 fs/ext4/mballoc.c BUG_ON(group != e4b->bd_group && pa->pa_len != 0); pa 3882 fs/ext4/mballoc.c mb_free_blocks(pa->pa_inode, e4b, bit, pa->pa_len); pa 3883 fs/ext4/mballoc.c atomic_add(pa->pa_len, &EXT4_SB(sb)->s_mb_discarded); pa 3884 fs/ext4/mballoc.c trace_ext4_mballoc_discard(sb, NULL, group, bit, pa->pa_len); pa 3904 fs/ext4/mballoc.c struct ext4_prealloc_space *pa, *tmp; pa 3938 fs/ext4/mballoc.c list_for_each_entry_safe(pa, tmp, pa 3940 fs/ext4/mballoc.c spin_lock(&pa->pa_lock); pa 3941 fs/ext4/mballoc.c if (atomic_read(&pa->pa_count)) { pa 3942 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 3946 fs/ext4/mballoc.c if (pa->pa_deleted) { pa 3947 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 3952 fs/ext4/mballoc.c pa->pa_deleted = 1; pa 3955 fs/ext4/mballoc.c free += pa->pa_free; pa 3957 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 3959 fs/ext4/mballoc.c list_del(&pa->pa_group_list); pa 3960 fs/ext4/mballoc.c list_add(&pa->u.pa_tmp_list, &list); pa 3978 fs/ext4/mballoc.c list_for_each_entry_safe(pa, tmp, &list, u.pa_tmp_list) { pa 3981 fs/ext4/mballoc.c spin_lock(pa->pa_obj_lock); pa 3982 fs/ext4/mballoc.c list_del_rcu(&pa->pa_inode_list); pa 3983 fs/ext4/mballoc.c spin_unlock(pa->pa_obj_lock); pa 3985 fs/ext4/mballoc.c if (pa->pa_type == MB_GROUP_PA) pa 3986 fs/ext4/mballoc.c ext4_mb_release_group_pa(&e4b, pa); pa 3988 fs/ext4/mballoc.c ext4_mb_release_inode_pa(&e4b, bitmap_bh, pa); pa 3990 fs/ext4/mballoc.c list_del(&pa->u.pa_tmp_list); pa 3991 fs/ext4/mballoc.c call_rcu(&(pa)->u.pa_rcu, ext4_mb_pa_callback); pa 4015 fs/ext4/mballoc.c struct ext4_prealloc_space *pa, *tmp; pa 4035 fs/ext4/mballoc.c pa = list_entry(ei->i_prealloc_list.next, pa 4037 fs/ext4/mballoc.c BUG_ON(pa->pa_obj_lock != &ei->i_prealloc_lock); pa 4038 fs/ext4/mballoc.c spin_lock(&pa->pa_lock); pa 4039 fs/ext4/mballoc.c if (atomic_read(&pa->pa_count)) { pa 4042 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 4051 fs/ext4/mballoc.c if (pa->pa_deleted == 0) { pa 4052 fs/ext4/mballoc.c pa->pa_deleted = 1; pa 4053 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 4054 fs/ext4/mballoc.c list_del_rcu(&pa->pa_inode_list); pa 4055 fs/ext4/mballoc.c list_add(&pa->u.pa_tmp_list, &list); pa 4060 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 4080 fs/ext4/mballoc.c list_for_each_entry_safe(pa, tmp, &list, u.pa_tmp_list) { pa 4081 fs/ext4/mballoc.c BUG_ON(pa->pa_type != MB_INODE_PA); pa 4082 fs/ext4/mballoc.c group = ext4_get_group_number(sb, pa->pa_pstart); pa 4102 fs/ext4/mballoc.c list_del(&pa->pa_group_list); pa 4103 fs/ext4/mballoc.c ext4_mb_release_inode_pa(&e4b, bitmap_bh, pa); pa 4109 fs/ext4/mballoc.c list_del(&pa->u.pa_tmp_list); pa 4110 fs/ext4/mballoc.c call_rcu(&(pa)->u.pa_rcu, ext4_mb_pa_callback); pa 4149 fs/ext4/mballoc.c struct ext4_prealloc_space *pa; pa 4154 fs/ext4/mballoc.c pa = list_entry(cur, struct ext4_prealloc_space, pa 4156 fs/ext4/mballoc.c spin_lock(&pa->pa_lock); pa 4157 fs/ext4/mballoc.c ext4_get_group_no_and_offset(sb, pa->pa_pstart, pa 4159 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 4161 fs/ext4/mballoc.c start, pa->pa_len); pa 4296 fs/ext4/mballoc.c struct ext4_prealloc_space *pa, *tmp; pa 4303 fs/ext4/mballoc.c list_for_each_entry_rcu(pa, &lg->lg_prealloc_list[order], pa 4305 fs/ext4/mballoc.c spin_lock(&pa->pa_lock); pa 4306 fs/ext4/mballoc.c if (atomic_read(&pa->pa_count)) { pa 4312 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 4315 fs/ext4/mballoc.c if (pa->pa_deleted) { pa 4316 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 4320 fs/ext4/mballoc.c BUG_ON(pa->pa_type != MB_GROUP_PA); pa 4323 fs/ext4/mballoc.c pa->pa_deleted = 1; pa 4324 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 4326 fs/ext4/mballoc.c list_del_rcu(&pa->pa_inode_list); pa 4327 fs/ext4/mballoc.c list_add(&pa->u.pa_tmp_list, &discard_list); pa 4342 fs/ext4/mballoc.c list_for_each_entry_safe(pa, tmp, &discard_list, u.pa_tmp_list) { pa 4345 fs/ext4/mballoc.c group = ext4_get_group_number(sb, pa->pa_pstart); pa 4354 fs/ext4/mballoc.c list_del(&pa->pa_group_list); pa 4355 fs/ext4/mballoc.c ext4_mb_release_group_pa(&e4b, pa); pa 4359 fs/ext4/mballoc.c list_del(&pa->u.pa_tmp_list); pa 4360 fs/ext4/mballoc.c call_rcu(&(pa)->u.pa_rcu, ext4_mb_pa_callback); pa 4378 fs/ext4/mballoc.c struct ext4_prealloc_space *tmp_pa, *pa = ac->ac_pa; pa 4380 fs/ext4/mballoc.c order = fls(pa->pa_free) - 1; pa 4393 fs/ext4/mballoc.c if (!added && pa->pa_free < tmp_pa->pa_free) { pa 4395 fs/ext4/mballoc.c list_add_tail_rcu(&pa->pa_inode_list, pa 4407 fs/ext4/mballoc.c list_add_tail_rcu(&pa->pa_inode_list, pa 4426 fs/ext4/mballoc.c struct ext4_prealloc_space *pa = ac->ac_pa; pa 4427 fs/ext4/mballoc.c if (pa) { pa 4428 fs/ext4/mballoc.c if (pa->pa_type == MB_GROUP_PA) { pa 4430 fs/ext4/mballoc.c spin_lock(&pa->pa_lock); pa 4431 fs/ext4/mballoc.c pa->pa_pstart += EXT4_C2B(sbi, ac->ac_b_ex.fe_len); pa 4432 fs/ext4/mballoc.c pa->pa_lstart += EXT4_C2B(sbi, ac->ac_b_ex.fe_len); pa 4433 fs/ext4/mballoc.c pa->pa_free -= ac->ac_b_ex.fe_len; pa 4434 fs/ext4/mballoc.c pa->pa_len -= ac->ac_b_ex.fe_len; pa 4435 fs/ext4/mballoc.c spin_unlock(&pa->pa_lock); pa 4438 fs/ext4/mballoc.c if (pa) { pa 4445 fs/ext4/mballoc.c if ((pa->pa_type == MB_GROUP_PA) && likely(pa->pa_free)) { pa 4446 fs/ext4/mballoc.c spin_lock(pa->pa_obj_lock); pa 4447 fs/ext4/mballoc.c list_del_rcu(&pa->pa_inode_list); pa 4448 fs/ext4/mballoc.c spin_unlock(pa->pa_obj_lock); pa 4451 fs/ext4/mballoc.c ext4_mb_put_pa(ac, ac->ac_sb, pa); pa 280 fs/f2fs/acl.c struct posix_acl_entry *pa, *pe; pa 287 fs/f2fs/acl.c FOREACH_ACL_ENTRY(pa, acl, pe) { pa 288 fs/f2fs/acl.c switch (pa->e_tag) { pa 290 fs/f2fs/acl.c pa->e_perm &= (mode >> 6) | ~S_IRWXO; pa 291 fs/f2fs/acl.c mode &= (pa->e_perm << 6) | ~S_IRWXU; pa 300 fs/f2fs/acl.c group_obj = pa; pa 304 fs/f2fs/acl.c pa->e_perm &= mode | ~S_IRWXO; pa 305 fs/f2fs/acl.c mode &= pa->e_perm | ~S_IRWXO; pa 309 fs/f2fs/acl.c mask_obj = pa; pa 221 fs/nfs_common/nfsacl.c struct posix_acl_entry *pa, *pe, pa 231 fs/nfs_common/nfsacl.c FOREACH_ACL_ENTRY(pa, acl, pe) { pa 232 fs/nfs_common/nfsacl.c switch(pa->e_tag) { pa 236 fs/nfs_common/nfsacl.c group_obj = pa; pa 239 fs/nfs_common/nfsacl.c mask = pa; pa 191 fs/nfsd/nfs4acl.c struct posix_acl_entry *pa, *pe; pa 203 fs/nfsd/nfs4acl.c FOREACH_ACL_ENTRY(pa, acl, pe) { pa 204 fs/nfsd/nfs4acl.c switch (pa->e_tag) { pa 206 fs/nfsd/nfs4acl.c pas->owner = pa->e_perm; pa 209 fs/nfsd/nfs4acl.c pas->group = pa->e_perm; pa 212 fs/nfsd/nfs4acl.c pas->users |= pa->e_perm; pa 215 fs/nfsd/nfs4acl.c pas->groups |= pa->e_perm; pa 218 fs/nfsd/nfs4acl.c pas->other = pa->e_perm; pa 221 fs/nfsd/nfs4acl.c pas->mask = pa->e_perm; pa 236 fs/nfsd/nfs4acl.c struct posix_acl_entry *pa, *group_owner_entry; pa 246 fs/nfsd/nfs4acl.c pa = pacl->a_entries; pa 267 fs/nfsd/nfs4acl.c ace->access_mask = mask_from_posix(pa->e_perm, flags | NFS4_ACL_OWNER); pa 271 fs/nfsd/nfs4acl.c pa++; pa 273 fs/nfsd/nfs4acl.c while (pa->e_tag == ACL_USER) { pa 274 fs/nfsd/nfs4acl.c deny = ~(pa->e_perm & pas.mask); pa 281 fs/nfsd/nfs4acl.c ace->who_uid = pa->e_uid; pa 287 fs/nfsd/nfs4acl.c ace->access_mask = mask_from_posix(pa->e_perm & pas.mask, pa 290 fs/nfsd/nfs4acl.c ace->who_uid = pa->e_uid; pa 293 fs/nfsd/nfs4acl.c pa++; pa 301 fs/nfsd/nfs4acl.c group_owner_entry = pa; pa 309 fs/nfsd/nfs4acl.c pa++; pa 311 fs/nfsd/nfs4acl.c while (pa->e_tag == ACL_GROUP) { pa 314 fs/nfsd/nfs4acl.c ace->access_mask = mask_from_posix(pa->e_perm & pas.mask, pa 317 fs/nfsd/nfs4acl.c ace->who_gid = pa->e_gid; pa 320 fs/nfsd/nfs4acl.c pa++; pa 325 fs/nfsd/nfs4acl.c pa = group_owner_entry; pa 336 fs/nfsd/nfs4acl.c pa++; pa 338 fs/nfsd/nfs4acl.c while (pa->e_tag == ACL_GROUP) { pa 339 fs/nfsd/nfs4acl.c deny = ~(pa->e_perm & pas.mask); pa 346 fs/nfsd/nfs4acl.c ace->who_gid = pa->e_gid; pa 350 fs/nfsd/nfs4acl.c pa++; pa 353 fs/nfsd/nfs4acl.c if (pa->e_tag == ACL_MASK) pa 354 fs/nfsd/nfs4acl.c pa++; pa 357 fs/nfsd/nfs4acl.c ace->access_mask = mask_from_posix(pa->e_perm, flags); pa 212 fs/posix_acl.c const struct posix_acl_entry *pa, *pe; pa 216 fs/posix_acl.c FOREACH_ACL_ENTRY(pa, acl, pe) { pa 217 fs/posix_acl.c if (pa->e_perm & ~(ACL_READ|ACL_WRITE|ACL_EXECUTE)) pa 219 fs/posix_acl.c switch (pa->e_tag) { pa 230 fs/posix_acl.c if (!kuid_has_mapping(user_ns, pa->e_uid)) pa 245 fs/posix_acl.c if (!kgid_has_mapping(user_ns, pa->e_gid)) pa 281 fs/posix_acl.c const struct posix_acl_entry *pa, *pe; pa 291 fs/posix_acl.c FOREACH_ACL_ENTRY(pa, acl, pe) { pa 292 fs/posix_acl.c switch (pa->e_tag) { pa 294 fs/posix_acl.c mode |= (pa->e_perm & S_IRWXO) << 6; pa 297 fs/posix_acl.c mode |= (pa->e_perm & S_IRWXO) << 3; pa 300 fs/posix_acl.c mode |= pa->e_perm & S_IRWXO; pa 304 fs/posix_acl.c ((pa->e_perm & S_IRWXO) << 3); pa 350 fs/posix_acl.c const struct posix_acl_entry *pa, *pe, *mask_obj; pa 355 fs/posix_acl.c FOREACH_ACL_ENTRY(pa, acl, pe) { pa 356 fs/posix_acl.c switch(pa->e_tag) { pa 363 fs/posix_acl.c if (uid_eq(pa->e_uid, current_fsuid())) pa 369 fs/posix_acl.c if ((pa->e_perm & want) == want) pa 374 fs/posix_acl.c if (in_group_p(pa->e_gid)) { pa 376 fs/posix_acl.c if ((pa->e_perm & want) == want) pa 394 fs/posix_acl.c for (mask_obj = pa+1; mask_obj != pe; mask_obj++) { pa 396 fs/posix_acl.c if ((pa->e_perm & mask_obj->e_perm & want) == want) pa 403 fs/posix_acl.c if ((pa->e_perm & want) == want) pa 418 fs/posix_acl.c struct posix_acl_entry *pa, *pe; pa 425 fs/posix_acl.c FOREACH_ACL_ENTRY(pa, acl, pe) { pa 426 fs/posix_acl.c switch(pa->e_tag) { pa 428 fs/posix_acl.c pa->e_perm &= (mode >> 6) | ~S_IRWXO; pa 429 fs/posix_acl.c mode &= (pa->e_perm << 6) | ~S_IRWXU; pa 438 fs/posix_acl.c group_obj = pa; pa 442 fs/posix_acl.c pa->e_perm &= mode | ~S_IRWXO; pa 443 fs/posix_acl.c mode &= pa->e_perm | ~S_IRWXO; pa 447 fs/posix_acl.c mask_obj = pa; pa 476 fs/posix_acl.c struct posix_acl_entry *pa, *pe; pa 480 fs/posix_acl.c FOREACH_ACL_ENTRY(pa, acl, pe) { pa 481 fs/posix_acl.c switch(pa->e_tag) { pa 483 fs/posix_acl.c pa->e_perm = (mode & S_IRWXU) >> 6; pa 491 fs/posix_acl.c group_obj = pa; pa 495 fs/posix_acl.c mask_obj = pa; pa 499 fs/posix_acl.c pa->e_perm = (mode & S_IRWXO); pa 649 fs/xfs/scrub/agheader.c const void *pa, pa 652 fs/xfs/scrub/agheader.c const xfs_agblock_t *a = pa; pa 167 include/asm-generic/mshyperv.h void hyperv_report_panic_msg(phys_addr_t pa, size_t size); pa 52 include/dt-bindings/pinctrl/am43xx.h #define AM4372_IOPAD(pa, val) (((pa) & 0xffff) - 0x0800) (val) pa 71 include/dt-bindings/pinctrl/dra.h #define DRA7XX_CORE_IOPAD(pa, val) (((pa) & 0xffff) - 0x3400) (val) pa 32 include/dt-bindings/pinctrl/k3.h #define AM65X_IOPAD(pa, val, muxmode) (((pa) & 0x1fff)) ((val) | (muxmode)) pa 33 include/dt-bindings/pinctrl/k3.h #define AM65X_WKUP_IOPAD(pa, val, muxmode) (((pa) & 0x1fff)) ((val) | (muxmode)) pa 35 include/dt-bindings/pinctrl/k3.h #define J721E_IOPAD(pa, val, muxmode) (((pa) & 0x1fff)) ((val) | (muxmode)) pa 36 include/dt-bindings/pinctrl/k3.h #define J721E_WKUP_IOPAD(pa, val, muxmode) (((pa) & 0x1fff)) ((val) | (muxmode)) pa 35 include/dt-bindings/pinctrl/keystone.h #define KEYSTONE_IOPAD_OFFSET(pa, offset) (((pa) & 0xffff) - (offset)) pa 37 include/dt-bindings/pinctrl/keystone.h #define K2G_CORE_IOPAD(pa) KEYSTONE_IOPAD_OFFSET((pa), 0x1000) pa 57 include/dt-bindings/pinctrl/omap.h #define OMAP_IOPAD_OFFSET(pa, offset) (((pa) & 0xffff) - (offset)) pa 59 include/dt-bindings/pinctrl/omap.h #define OMAP2420_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0030) (val) pa 60 include/dt-bindings/pinctrl/omap.h #define OMAP2430_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val) pa 61 include/dt-bindings/pinctrl/omap.h #define OMAP3_CORE1_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val) pa 62 include/dt-bindings/pinctrl/omap.h #define OMAP3430_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25d8) (val) pa 63 include/dt-bindings/pinctrl/omap.h #define OMAP3630_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25a0) (val) pa 64 include/dt-bindings/pinctrl/omap.h #define OMAP3_WKUP_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2a00) (val) pa 65 include/dt-bindings/pinctrl/omap.h #define DM814X_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) pa 66 include/dt-bindings/pinctrl/omap.h #define DM816X_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) pa 67 include/dt-bindings/pinctrl/omap.h #define AM33XX_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) pa 68 include/dt-bindings/pinctrl/omap.h #define AM33XX_PADCONF(pa, dir, mux) OMAP_IOPAD_OFFSET((pa), 0x0800) ((dir) | (mux)) pa 248 include/linux/acpi.h void acpi_numa_processor_affinity_init (struct acpi_srat_cpu_affinity *pa); pa 251 include/linux/acpi.h acpi_numa_processor_affinity_init(struct acpi_srat_cpu_affinity *pa) { } pa 254 include/linux/acpi.h void acpi_numa_x2apic_affinity_init(struct acpi_srat_x2apic_cpu_affinity *pa); pa 257 include/linux/acpi.h void acpi_numa_gicc_affinity_init(struct acpi_srat_gicc_affinity *pa); pa 260 include/linux/acpi.h acpi_numa_gicc_affinity_init(struct acpi_srat_gicc_affinity *pa) { } pa 47 include/linux/fsl/bestcomm/sram.h static inline void *bcom_sram_pa2va(phys_addr_t pa) { pa 49 include/linux/fsl/bestcomm/sram.h (unsigned long)(pa - bcom_sram->base_phys); pa 29 include/linux/if_pppox.h struct pppoe_addr pa; /* what this socket is bound to*/ pa 57 include/linux/if_pppox.h #define pppoe_pa proto.pppoe.pa pa 230 include/linux/irqdomain.h const char *name, phys_addr_t *pa); pa 251 include/linux/irqdomain.h static inline struct fwnode_handle *irq_domain_alloc_fwnode(phys_addr_t *pa) pa 253 include/linux/irqdomain.h return __irq_domain_alloc_fwnode(IRQCHIP_FWNODE_REAL, 0, NULL, pa); pa 775 include/linux/mlx5/driver.h u64 pa; pa 34 include/linux/posix_acl.h #define FOREACH_ACL_ENTRY(pa, acl, pe) \ pa 35 include/linux/posix_acl.h for(pa=(acl)->a_entries, pe=pa+(acl)->a_count; pa<pe; pa++) pa 172 include/linux/remoteproc.h u32 pa; pa 210 include/linux/remoteproc.h u32 pa; pa 262 include/linux/remoteproc.h u32 pa; pa 374 include/linux/tee_drv.h int tee_shm_va2pa(struct tee_shm *shm, void *va, phys_addr_t *pa); pa 383 include/linux/tee_drv.h int tee_shm_pa2va(struct tee_shm *shm, phys_addr_t pa, void **va); pa 402 include/linux/tee_drv.h int tee_shm_get_pa(struct tee_shm *shm, size_t offs, phys_addr_t *pa); pa 107 include/media/cec-notifier.h void cec_notifier_set_phys_addr(struct cec_notifier *n, u16 pa); pa 167 include/media/cec-notifier.h static inline void cec_notifier_set_phys_addr(struct cec_notifier *n, u16 pa) pa 264 include/media/cec.h #define cec_phys_addr_exp(pa) \ pa 265 include/media/cec.h ((pa) >> 12), ((pa) >> 8) & 0xf, ((pa) >> 4) & 0xf, (pa) & 0xf pa 430 include/media/cec.h void (*callback)(struct cec_adapter *adap, u16 pa)); pa 451 include/media/cec.h void (*callback)(struct cec_adapter *adap, u16 pa)) pa 40 include/ras/ras_event.h __field(u64, pa) pa 55 include/ras/ras_event.h __entry->pa = mem->physical_addr; pa 57 include/ras/ras_event.h __entry->pa = ~0ull; pa 72 include/ras/ras_event.h __entry->pa, pa 197 include/trace/events/erofs.h __field( erofs_off_t, pa ) pa 209 include/trace/events/erofs.h __entry->pa = map->m_pa; pa 220 include/trace/events/erofs.h __entry->la, __entry->pa, __entry->llen, __entry->plen, pa 635 include/trace/events/ext4.h struct ext4_prealloc_space *pa), pa 637 include/trace/events/ext4.h TP_ARGS(ac, pa), pa 651 include/trace/events/ext4.h __entry->pa_pstart = pa->pa_pstart; pa 652 include/trace/events/ext4.h __entry->pa_lstart = pa->pa_lstart; pa 653 include/trace/events/ext4.h __entry->pa_len = pa->pa_len; pa 665 include/trace/events/ext4.h struct ext4_prealloc_space *pa), pa 667 include/trace/events/ext4.h TP_ARGS(ac, pa) pa 673 include/trace/events/ext4.h struct ext4_prealloc_space *pa), pa 675 include/trace/events/ext4.h TP_ARGS(ac, pa) pa 679 include/trace/events/ext4.h TP_PROTO(struct ext4_prealloc_space *pa, pa 682 include/trace/events/ext4.h TP_ARGS(pa, block, count), pa 693 include/trace/events/ext4.h __entry->dev = pa->pa_inode->i_sb->s_dev; pa 694 include/trace/events/ext4.h __entry->ino = pa->pa_inode->i_ino; pa 706 include/trace/events/ext4.h TP_PROTO(struct super_block *sb, struct ext4_prealloc_space *pa), pa 708 include/trace/events/ext4.h TP_ARGS(sb, pa), pa 719 include/trace/events/ext4.h __entry->pa_pstart = pa->pa_pstart; pa 720 include/trace/events/ext4.h __entry->pa_len = pa->pa_len; pa 34 kernel/irq/irqdomain.c phys_addr_t *pa; pa 66 kernel/irq/irqdomain.c phys_addr_t *pa) pa 81 kernel/irq/irqdomain.c n = kasprintf(GFP_KERNEL, "irqchip@%pa", pa); pa 93 kernel/irq/irqdomain.c fwid->pa = pa; pa 24 lib/test_debug_virtual.c phys_addr_t pa; pa 28 lib/test_debug_virtual.c pa = virt_to_phys(va); pa 30 lib/test_debug_virtual.c pr_info("PA: %pa for VA: 0x%lx\n", &pa, (unsigned long)va); pa 36 lib/test_debug_virtual.c pa = virt_to_phys(foo); pa 38 lib/test_debug_virtual.c pr_info("PA: %pa for VA: 0x%lx\n", &pa, (unsigned long)va); pa 82 mm/process_vm_access.c unsigned long pa = addr & PAGE_MASK; pa 83 mm/process_vm_access.c unsigned long start_offset = addr - pa; pa 109 mm/process_vm_access.c pages = get_user_pages_remote(task, mm, pa, pages, flags, pa 126 mm/process_vm_access.c pa += pages * PAGE_SIZE; pa 387 net/8021q/vlan_dev.c static int vlan_dev_neigh_setup(struct net_device *dev, struct neigh_parms *pa) pa 394 net/8021q/vlan_dev.c err = ops->ndo_neigh_setup(real_dev, pa); pa 268 net/wireless/core.h } pa; pa 1062 net/wireless/sme.c memcpy(ev->pa.bssid, bssid, ETH_ALEN); pa 892 net/wireless/util.c __cfg80211_port_authorized(wdev, ev->pa.bssid); pa 52 scripts/dtc/include-prefixes/dt-bindings/pinctrl/am43xx.h #define AM4372_IOPAD(pa, val) (((pa) & 0xffff) - 0x0800) (val) pa 71 scripts/dtc/include-prefixes/dt-bindings/pinctrl/dra.h #define DRA7XX_CORE_IOPAD(pa, val) (((pa) & 0xffff) - 0x3400) (val) pa 32 scripts/dtc/include-prefixes/dt-bindings/pinctrl/k3.h #define AM65X_IOPAD(pa, val, muxmode) (((pa) & 0x1fff)) ((val) | (muxmode)) pa 33 scripts/dtc/include-prefixes/dt-bindings/pinctrl/k3.h #define AM65X_WKUP_IOPAD(pa, val, muxmode) (((pa) & 0x1fff)) ((val) | (muxmode)) pa 35 scripts/dtc/include-prefixes/dt-bindings/pinctrl/k3.h #define J721E_IOPAD(pa, val, muxmode) (((pa) & 0x1fff)) ((val) | (muxmode)) pa 36 scripts/dtc/include-prefixes/dt-bindings/pinctrl/k3.h #define J721E_WKUP_IOPAD(pa, val, muxmode) (((pa) & 0x1fff)) ((val) | (muxmode)) pa 35 scripts/dtc/include-prefixes/dt-bindings/pinctrl/keystone.h #define KEYSTONE_IOPAD_OFFSET(pa, offset) (((pa) & 0xffff) - (offset)) pa 37 scripts/dtc/include-prefixes/dt-bindings/pinctrl/keystone.h #define K2G_CORE_IOPAD(pa) KEYSTONE_IOPAD_OFFSET((pa), 0x1000) pa 57 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define OMAP_IOPAD_OFFSET(pa, offset) (((pa) & 0xffff) - (offset)) pa 59 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define OMAP2420_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0030) (val) pa 60 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define OMAP2430_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val) pa 61 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define OMAP3_CORE1_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val) pa 62 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define OMAP3430_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25d8) (val) pa 63 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define OMAP3630_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25a0) (val) pa 64 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define OMAP3_WKUP_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2a00) (val) pa 65 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define DM814X_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) pa 66 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define DM816X_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) pa 67 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define AM33XX_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) pa 68 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define AM33XX_PADCONF(pa, dir, mux) OMAP_IOPAD_OFFSET((pa), 0x0800) ((dir) | (mux)) pa 180 sound/pci/asihpi/hpioctl.c struct hpi_adapter *pa = NULL; pa 183 sound/pci/asihpi/hpioctl.c pa = &adapters[array_index_nospec(hm->h.adapter_index, pa 186 sound/pci/asihpi/hpioctl.c if (!pa || !pa->adapter || !pa->adapter->type) { pa 199 sound/pci/asihpi/hpioctl.c if (mutex_lock_interruptible(&pa->mutex)) { pa 216 sound/pci/asihpi/hpioctl.c if (pa->buffer_size < size) { pa 221 sound/pci/asihpi/hpioctl.c pa->buffer_size, size); pa 222 sound/pci/asihpi/hpioctl.c if (pa->p_buffer) { pa 223 sound/pci/asihpi/hpioctl.c pa->buffer_size = 0; pa 224 sound/pci/asihpi/hpioctl.c vfree(pa->p_buffer); pa 226 sound/pci/asihpi/hpioctl.c pa->p_buffer = vmalloc(size); pa 227 sound/pci/asihpi/hpioctl.c if (pa->p_buffer) pa 228 sound/pci/asihpi/hpioctl.c pa->buffer_size = size; pa 235 sound/pci/asihpi/hpioctl.c mutex_unlock(&pa->mutex); pa 241 sound/pci/asihpi/hpioctl.c hm->m0.u.d.u.data.pb_data = pa->p_buffer; pa 257 sound/pci/asihpi/hpioctl.c copy_from_user(pa->p_buffer, ptr, size); pa 269 sound/pci/asihpi/hpioctl.c copy_to_user(ptr, pa->p_buffer, size); pa 276 sound/pci/asihpi/hpioctl.c mutex_unlock(&pa->mutex); pa 523 sound/pci/asihpi/hpioctl.c struct hpi_adapter *pa; pa 526 sound/pci/asihpi/hpioctl.c pa = pci_get_drvdata(pci_dev); pa 527 sound/pci/asihpi/hpioctl.c pci = pa->adapter->pci; pa 532 sound/pci/asihpi/hpioctl.c hm.adapter_index = pa->adapter->index; pa 540 sound/pci/asihpi/hpioctl.c hm.adapter_index = pa->adapter->index; pa 547 sound/pci/asihpi/hpioctl.c if (pa->irq) pa 548 sound/pci/asihpi/hpioctl.c free_irq(pa->irq, pa); pa 550 sound/pci/asihpi/hpioctl.c vfree(pa->p_buffer); pa 557 sound/pci/asihpi/hpioctl.c pci_dev->devfn, pa->adapter->index); pa 559 sound/pci/asihpi/hpioctl.c memset(pa, 0, sizeof(*pa)); pa 992 sound/pci/es1968.c u32 pa; pa 1010 sound/pci/es1968.c pa = es->memory->buf.addr; pa 1011 sound/pci/es1968.c pa -= chip->dma.addr; pa 1012 sound/pci/es1968.c pa >>= 1; /* words */ pa 1014 sound/pci/es1968.c pa |= 0x00400000; /* System RAM (Bit 22) */ pa 1019 sound/pci/es1968.c pa |= 0x00800000; /* (Bit 23) */ pa 1021 sound/pci/es1968.c pa >>= 1; pa 1026 sound/pci/es1968.c es->base[channel] = pa & 0xFFFF; pa 1032 sound/pci/es1968.c apu_set_register(chip, apu, 4, ((pa >> 16) & 0xFF) << 8); pa 1033 sound/pci/es1968.c apu_set_register(chip, apu, 5, pa & 0xFFFF); pa 1034 sound/pci/es1968.c apu_set_register(chip, apu, 6, (pa + size) & 0xFFFF); pa 1093 sound/pci/es1968.c unsigned int pa, unsigned int bsize, pa 1101 sound/pci/es1968.c snd_es1968_program_wavecache(chip, es, channel, pa, 1); pa 1104 sound/pci/es1968.c pa -= chip->dma.addr; pa 1105 sound/pci/es1968.c pa >>= 1; /* words */ pa 1109 sound/pci/es1968.c es->base[channel] = pa & 0xFFFF; pa 1110 sound/pci/es1968.c pa |= 0x00400000; /* bit 22 -> System RAM */ pa 1121 sound/pci/es1968.c apu_set_register(chip, apu, 4, ((pa >> 16) & 0xFF) << 8); pa 1122 sound/pci/es1968.c apu_set_register(chip, apu, 5, pa & 0xFFFF); pa 1123 sound/pci/es1968.c apu_set_register(chip, apu, 6, (pa + bsize) & 0xFFFF); pa 1697 sound/pci/es1968.c unsigned int pa, offset, t; pa 1722 sound/pci/es1968.c pa = (unsigned int)((memory->buf.addr - chip->dma.addr) >> 1); pa 1723 sound/pci/es1968.c pa |= 0x00400000; /* System RAM (Bit 22) */ pa 1730 sound/pci/es1968.c apu_set_register(chip, apu, 4, ((pa >> 16) & 0xff) << 8); pa 1731 sound/pci/es1968.c apu_set_register(chip, apu, 5, pa & 0xffff); pa 1732 sound/pci/es1968.c apu_set_register(chip, apu, 6, (pa + CLOCK_MEASURE_BUFSIZE/2) & 0xffff); pa 1749 sound/pci/es1968.c __apu_set_register(chip, apu, 5, pa & 0xffff); pa 1763 sound/pci/es1968.c offset -= (pa & 0xffff); pa 686 tools/lib/traceevent/event-parse.c const struct printk_map *pa = a; pa 689 tools/lib/traceevent/event-parse.c if (pa->addr < pb->addr) pa 691 tools/lib/traceevent/event-parse.c if (pa->addr > pb->addr) pa 1861 tools/perf/util/probe-event.c char *synthesize_perf_probe_arg(struct perf_probe_arg *pa) pa 1863 tools/perf/util/probe-event.c struct perf_probe_arg_field *field = pa->field; pa 1871 tools/perf/util/probe-event.c if (pa->name && pa->var) pa 1872 tools/perf/util/probe-event.c err = strbuf_addf(&buf, "%s=%s", pa->name, pa->var); pa 1874 tools/perf/util/probe-event.c err = strbuf_addstr(&buf, pa->name ?: pa->var); pa 1889 tools/perf/util/probe-event.c if (pa->type) pa 1890 tools/perf/util/probe-event.c if (strbuf_addf(&buf, ":%s", pa->type) < 0) pa 135 tools/perf/util/probe-event.h char *synthesize_perf_probe_arg(struct perf_probe_arg *pa); pa 348 tools/testing/selftests/kvm/include/x86_64/vmx.h : [pa]"m"(phys) pa 365 tools/testing/selftests/kvm/include/x86_64/vmx.h : [pa]"m"(vmcs_pa) pa 380 tools/testing/selftests/kvm/include/x86_64/vmx.h : [pa]"m"(vmcs_pa) pa 777 tools/testing/selftests/net/nettest.c const uint32_t *pa = (uint32_t *) &in6->s6_addr; pa 782 tools/testing/selftests/net/nettest.c pa += 3; pa 783 tools/testing/selftests/net/nettest.c in4.s_addr = *pa; pa 1340 virt/kvm/arm/mmu.c phys_addr_t pa, unsigned long size, bool writable) pa 1348 virt/kvm/arm/mmu.c pfn = __phys_to_pfn(pa); pa 2336 virt/kvm/arm/mmu.c phys_addr_t pa; pa 2338 virt/kvm/arm/mmu.c pa = (phys_addr_t)vma->vm_pgoff << PAGE_SHIFT; pa 2339 virt/kvm/arm/mmu.c pa += vm_start - vma->vm_start; pa 2347 virt/kvm/arm/mmu.c ret = kvm_phys_addr_ioremap(kvm, gpa, pa,