pa                126 arch/alpha/include/asm/hwrpb.h 	unsigned long pa;
pa                 19 arch/alpha/include/asm/mmzone.h #define alpha_pa_to_nid(pa)		\
pa                 21 arch/alpha/include/asm/mmzone.h 	 ? alpha_mv.pa_to_nid(pa)	\
pa                 32 arch/alpha/include/asm/mmzone.h #define pa_to_nid(pa)		alpha_pa_to_nid(pa)
pa                857 arch/alpha/kernel/core_marvel.c marvel_pa_to_nid(unsigned long pa)
pa                861 arch/alpha/kernel/core_marvel.c 	if ((pa >> 43) & 1) 	/* I/O */ 
pa                862 arch/alpha/kernel/core_marvel.c 		cpuid = (~(pa >> 35) & 0xff);
pa                864 arch/alpha/kernel/core_marvel.c 		cpuid = ((pa >> 34) & 0x3) | ((pa >> (37 - 2)) & (0x1f << 2));
pa                878 arch/alpha/kernel/core_marvel.c 	unsigned long pa;
pa                880 arch/alpha/kernel/core_marvel.c 	pa = (nid & 0x3) | ((nid & (0x1f << 2)) << 1);
pa                881 arch/alpha/kernel/core_marvel.c 	pa <<= 34;
pa                883 arch/alpha/kernel/core_marvel.c 	return pa;
pa                447 arch/alpha/kernel/core_wildfire.c int wildfire_pa_to_nid(unsigned long pa)
pa                449 arch/alpha/kernel/core_wildfire.c 	return pa >> 36;
pa                211 arch/alpha/mm/init.c 			unsigned long pfn = crb->map[i].pa >> PAGE_SHIFT;
pa                 58 arch/arm/include/asm/kvm_mmu.h 			  phys_addr_t pa, unsigned long size, bool writable);
pa                 21 arch/arm/include/asm/smp_scu.h 	unsigned long pa;
pa                 23 arch/arm/include/asm/smp_scu.h 	asm("mrc p15, 4, %0, c15, c0, 0" : "=r" (pa));
pa                 25 arch/arm/include/asm/smp_scu.h 	return pa;
pa                 31 arch/arm/mach-davinci/include/mach/hardware.h #define IO_ADDRESS(pa)			IOMEM(__IO_ADDRESS(pa))
pa                971 arch/arm/mach-omap1/clock.c 	struct clk *pa;
pa                978 arch/arm/mach-omap1/clock.c 		pa = c->parent;
pa                980 arch/arm/mach-omap1/clock.c 			   c->name, pa ? pa->name : "none", c->rate,
pa                993 arch/arm/mach-omap1/clock.c 	struct clk *pa = c->parent;
pa                995 arch/arm/mach-omap1/clock.c 	d = debugfs_create_dir(c->name, pa ? pa->dent : clk_debugfs_root);
pa               1005 arch/arm/mach-omap1/clock.c 	struct clk *pa = c->parent;
pa               1007 arch/arm/mach-omap1/clock.c 	if (pa && !pa->dent)
pa               1008 arch/arm/mach-omap1/clock.c 		clk_debugfs_register(pa);
pa                 47 arch/arm/mach-omap1/include/mach/hardware.h extern u8 omap_readb(u32 pa);
pa                 48 arch/arm/mach-omap1/include/mach/hardware.h extern u16 omap_readw(u32 pa);
pa                 49 arch/arm/mach-omap1/include/mach/hardware.h extern u32 omap_readl(u32 pa);
pa                 50 arch/arm/mach-omap1/include/mach/hardware.h extern void omap_writeb(u8 v, u32 pa);
pa                 51 arch/arm/mach-omap1/include/mach/hardware.h extern void omap_writew(u16 v, u32 pa);
pa                 52 arch/arm/mach-omap1/include/mach/hardware.h extern void omap_writel(u32 v, u32 pa);
pa                 76 arch/arm/mach-omap1/include/mach/hardware.h #define OMAP1_IO_ADDRESS(pa)	IOMEM((pa) - OMAP1_IO_OFFSET)
pa                145 arch/arm/mach-omap1/io.c u8 omap_readb(u32 pa)
pa                147 arch/arm/mach-omap1/io.c 	return __raw_readb(OMAP1_IO_ADDRESS(pa));
pa                151 arch/arm/mach-omap1/io.c u16 omap_readw(u32 pa)
pa                153 arch/arm/mach-omap1/io.c 	return __raw_readw(OMAP1_IO_ADDRESS(pa));
pa                157 arch/arm/mach-omap1/io.c u32 omap_readl(u32 pa)
pa                159 arch/arm/mach-omap1/io.c 	return __raw_readl(OMAP1_IO_ADDRESS(pa));
pa                163 arch/arm/mach-omap1/io.c void omap_writeb(u8 v, u32 pa)
pa                165 arch/arm/mach-omap1/io.c 	__raw_writeb(v, OMAP1_IO_ADDRESS(pa));
pa                169 arch/arm/mach-omap1/io.c void omap_writew(u16 v, u32 pa)
pa                171 arch/arm/mach-omap1/io.c 	__raw_writew(v, OMAP1_IO_ADDRESS(pa));
pa                175 arch/arm/mach-omap1/io.c void omap_writel(u32 v, u32 pa)
pa                177 arch/arm/mach-omap1/io.c 	__raw_writel(v, OMAP1_IO_ADDRESS(pa));
pa                349 arch/arm/mach-omap2/cm33xx.c 	return cm_base.pa + inst + offset;
pa                348 arch/arm/mach-omap2/cm_common.c 			mem->pa = res.start + data->offset;
pa                478 arch/arm/mach-omap2/cminst44xx.c 	return _cm_bases[part].pa + inst + offset;
pa                 34 arch/arm/mach-omap2/iomap.h #define OMAP2_L3_IO_ADDRESS(pa)	IOMEM((pa) + OMAP2_L3_IO_OFFSET) /* L3 */
pa                 37 arch/arm/mach-omap2/iomap.h #define OMAP2_L4_IO_ADDRESS(pa)	IOMEM((pa) + OMAP2_L4_IO_OFFSET) /* L4 */
pa                 40 arch/arm/mach-omap2/iomap.h #define OMAP4_L3_IO_ADDRESS(pa)	IOMEM((pa) + OMAP4_L3_IO_OFFSET) /* L3 */
pa                 43 arch/arm/mach-omap2/iomap.h #define AM33XX_L4_WK_IO_ADDRESS(pa)	IOMEM((pa) + AM33XX_L4_WK_IO_OFFSET)
pa                 46 arch/arm/mach-omap2/iomap.h #define OMAP4_L3_PER_IO_ADDRESS(pa)	IOMEM((pa) + OMAP4_L3_PER_IO_OFFSET)
pa                 49 arch/arm/mach-omap2/iomap.h #define OMAP2_EMU_IO_ADDRESS(pa)	IOMEM((pa) + OMAP2_EMU_IO_OFFSET)
pa                526 arch/arm/mach-omap2/prcm-common.h 	u32 pa;
pa                762 arch/arm/mach-omap2/prm_common.c 			prm_base.pa = res.start + data->offset;
pa                 66 arch/arm/mm/cache-xsc3l2.c static inline unsigned long l2_map_va(unsigned long pa, unsigned long prev_va)
pa                 70 arch/arm/mm/cache-xsc3l2.c 	unsigned long pa_offset = pa << (32 - PAGE_SHIFT);
pa                 78 arch/arm/mm/cache-xsc3l2.c 		va = (unsigned long)kmap_atomic_pfn(pa >> PAGE_SHIFT);
pa                 82 arch/arm/mm/cache-xsc3l2.c 	return __phys_to_virt(pa);
pa                159 arch/arm64/include/asm/kvm_mmu.h 			  phys_addr_t pa, unsigned long size, bool writable);
pa                 51 arch/arm64/kernel/acpi_numa.c 	struct acpi_srat_gicc_affinity *pa;
pa                 57 arch/arm64/kernel/acpi_numa.c 	pa = (struct acpi_srat_gicc_affinity *)header;
pa                 58 arch/arm64/kernel/acpi_numa.c 	if (!pa)
pa                 61 arch/arm64/kernel/acpi_numa.c 	if (!(pa->flags & ACPI_SRAT_GICC_ENABLED))
pa                 64 arch/arm64/kernel/acpi_numa.c 	pxm = pa->proximity_domain;
pa                 73 arch/arm64/kernel/acpi_numa.c 	cpu = get_cpu_for_acpi_id(pa->acpi_processor_uid);
pa                 92 arch/arm64/kernel/acpi_numa.c void __init acpi_numa_gicc_affinity_init(struct acpi_srat_gicc_affinity *pa)
pa                 99 arch/arm64/kernel/acpi_numa.c 	if (pa->header.length < sizeof(struct acpi_srat_gicc_affinity)) {
pa                101 arch/arm64/kernel/acpi_numa.c 			pa->header.length);
pa                106 arch/arm64/kernel/acpi_numa.c 	if (!(pa->flags & ACPI_SRAT_GICC_ENABLED))
pa                109 arch/arm64/kernel/acpi_numa.c 	pxm = pa->proximity_domain;
pa                376 arch/arm64/mm/mmu.c 	phys_addr_t pa = __pgd_pgtable_alloc(shift);
pa                387 arch/arm64/mm/mmu.c 		BUG_ON(!pgtable_pte_page_ctor(phys_to_page(pa)));
pa                389 arch/arm64/mm/mmu.c 		BUG_ON(!pgtable_pmd_page_ctor(phys_to_page(pa)));
pa                391 arch/arm64/mm/mmu.c 	return pa;
pa                333 arch/ia64/kernel/acpi.c get_processor_proximity_domain(struct acpi_srat_cpu_affinity *pa)
pa                337 arch/ia64/kernel/acpi.c 	pxm = pa->proximity_domain_lo;
pa                339 arch/ia64/kernel/acpi.c 		pxm += pa->proximity_domain_hi[0] << 8;
pa                375 arch/ia64/kernel/acpi.c acpi_numa_processor_affinity_init(struct acpi_srat_cpu_affinity *pa)
pa                379 arch/ia64/kernel/acpi.c 	if (!(pa->flags & ACPI_SRAT_CPU_ENABLED))
pa                388 arch/ia64/kernel/acpi.c 	pxm = get_processor_proximity_domain(pa);
pa                394 arch/ia64/kernel/acpi.c 	    (pa->apic_id << 8) | (pa->local_sapic_eid);
pa                 41 arch/ia64/kernel/brl_emu.c #define unimplemented_physical_address(pa) (		\
pa                 42 arch/ia64/kernel/brl_emu.c 	((pa) & local_cpu_data->unimpl_pa_mask) != 0	\
pa                148 arch/m68k/coldfire/intc-2.c 	u16 pa, tb;
pa                170 arch/m68k/coldfire/intc-2.c 	pa = __raw_readw(MCFEPORT_EPPAR);
pa                171 arch/m68k/coldfire/intc-2.c 	pa = (pa & ~(0x3 << (irq * 2))) | (tb << (irq * 2));
pa                172 arch/m68k/coldfire/intc-2.c 	__raw_writew(pa, MCFEPORT_EPPAR);
pa                132 arch/m68k/coldfire/intc-simr.c 	u16 pa, tb;
pa                154 arch/m68k/coldfire/intc-simr.c 	pa = __raw_readw(MCFEPORT_EPPAR);
pa                155 arch/m68k/coldfire/intc-simr.c 	pa = (pa & ~(0x3 << ebit)) | (tb << ebit);
pa                156 arch/m68k/coldfire/intc-simr.c 	__raw_writew(pa, MCFEPORT_EPPAR);
pa                 15 arch/m68k/include/asm/sun3xprom.h unsigned long sun3x_prom_ptov(unsigned long pa, unsigned long size);
pa                531 arch/microblaze/include/asm/pgtable.h int map_page(unsigned long va, phys_addr_t pa, int flags);
pa                135 arch/microblaze/mm/pgtable.c int map_page(unsigned long va, phys_addr_t pa, int flags)
pa                148 arch/microblaze/mm/pgtable.c 		set_pte_at(&init_mm, va, pg, pfn_pte(pa >> PAGE_SHIFT,
pa                214 arch/microblaze/mm/pgtable.c 	unsigned long pa;
pa                227 arch/microblaze/mm/pgtable.c 	pa = 0;
pa                229 arch/microblaze/mm/pgtable.c 		pa = (pte_val(*pte) & PAGE_MASK) | (addr & ~PAGE_MASK);
pa                231 arch/microblaze/mm/pgtable.c 	return pa;
pa                 43 arch/mips/bmips/dma.c dma_addr_t __phys_to_dma(struct device *dev, phys_addr_t pa)
pa                 48 arch/mips/bmips/dma.c 		if (pa >= r->child_addr &&
pa                 49 arch/mips/bmips/dma.c 		    pa < (r->child_addr + r->size))
pa                 50 arch/mips/bmips/dma.c 			return pa - r->child_addr + r->parent_addr;
pa                 52 arch/mips/bmips/dma.c 	return pa;
pa                135 arch/mips/include/asm/octeon/cvmx-address.h 		uint64_t pa:49; /* physical address */
pa                224 arch/mips/include/asm/octeon/cvmx-address.h 		uint64_t pa:49;
pa                119 arch/mips/kernel/spram.c 	unsigned int firstpa = 0, lastpa = 0, pa = 0;
pa                143 arch/mips/kernel/spram.c 			if ((pa == firstpa && size == firstsize) ||
pa                144 arch/mips/kernel/spram.c 			    (pa == lastpa && size == lastsize))
pa                159 arch/mips/kernel/spram.c 		pa = tag0 & SPRAM_TAG0_PA_MASK;
pa                163 arch/mips/kernel/spram.c 			firstpa = pa;
pa                167 arch/mips/kernel/spram.c 		lastpa = pa;
pa                171 arch/mips/kernel/spram.c 			unsigned int *vp = (unsigned int *)(CKSEG1 | pa);
pa                190 arch/mips/kernel/spram.c 			type, i, pa, size, enabled ? ",enabled" : "");
pa                309 arch/mips/kvm/tlb.c 	unsigned long entrylo[2], pagemask, pagemaskbit, pa;
pa                364 arch/mips/kvm/tlb.c 	pa = entrylo[!!(gva & pagemaskbit)];
pa                370 arch/mips/kvm/tlb.c 	if (!(pa & ENTRYLO_V))
pa                377 arch/mips/kvm/tlb.c 	pa = (pa << 6) & ~0xfffl;
pa                378 arch/mips/kvm/tlb.c 	pa |= gva & ~(pagemask | pagemaskbit);
pa                380 arch/mips/kvm/tlb.c 	*gpa = pa;
pa                723 arch/mips/kvm/vz.c 			unsigned long mask, pa;
pa                765 arch/mips/kvm/vz.c 			pa = (segctl << 20) & mask;
pa                766 arch/mips/kvm/vz.c 			pa |= gva32 & ~mask;
pa                767 arch/mips/kvm/vz.c 			*gpa = pa;
pa                 78 arch/mips/lib/dump_tlb.c 	unsigned long long entrylo0, entrylo1, pa;
pa                159 arch/mips/lib/dump_tlb.c 		pa = entrylo0 & ~(MIPS_ENTRYLO_RI | MIPS_ENTRYLO_XI);
pa                161 arch/mips/lib/dump_tlb.c 			pa |= (unsigned long long)readx_c0_entrylo0() << 30;
pa                162 arch/mips/lib/dump_tlb.c 		pa = (pa << 6) & PAGE_MASK;
pa                169 arch/mips/lib/dump_tlb.c 			pwidth, pa, c0,
pa                174 arch/mips/lib/dump_tlb.c 		pa = entrylo1 & ~(MIPS_ENTRYLO_RI | MIPS_ENTRYLO_XI);
pa                176 arch/mips/lib/dump_tlb.c 			pa |= (unsigned long long)readx_c0_entrylo1() << 30;
pa                177 arch/mips/lib/dump_tlb.c 		pa = (pa << 6) & PAGE_MASK;
pa                183 arch/mips/lib/dump_tlb.c 			pwidth, pa, c1,
pa                480 arch/mips/mm/cerr-sb1.c 	unsigned long long taglo, pa;
pa                501 arch/mips/mm/cerr-sb1.c 		pa = (taglo & 0xFFFFFFE000ULL) | addr;
pa                515 arch/mips/mm/cerr-sb1.c 			    way, pa, dc_state_str(state), state, taghi, taglo);
pa                 86 arch/nds32/mm/cacheflush.c static inline unsigned long kremap0(unsigned long uaddr, unsigned long pa)
pa                 92 arch/nds32/mm/cacheflush.c 	pte = (pa | PAGE_KERNEL);
pa                106 arch/nds32/mm/cacheflush.c static inline unsigned long kremap1(unsigned long uaddr, unsigned long pa)
pa                112 arch/nds32/mm/cacheflush.c 	pte = (pa | PAGE_KERNEL);
pa                  6 arch/parisc/include/asm/special_insns.h 	unsigned long pa;		\
pa                 10 arch/parisc/include/asm/special_insns.h 		: "=r" (pa)		\
pa                 14 arch/parisc/include/asm/special_insns.h 	pa;				\
pa                 18 arch/parisc/include/asm/special_insns.h 	unsigned long pa;		\
pa                 22 arch/parisc/include/asm/special_insns.h 		: "=r" (pa)		\
pa                 26 arch/parisc/include/asm/special_insns.h 	pa;				\
pa                 39 arch/powerpc/boot/wii.c static int mipc_check_address(u32 pa)
pa                 42 arch/powerpc/boot/wii.c 	if (pa < 0x10000000 || pa > 0x14000000)
pa                140 arch/powerpc/include/asm/book3s/32/pgtable.h int map_kernel_page(unsigned long va, phys_addr_t pa, pgprot_t prot);
pa                247 arch/powerpc/include/asm/book3s/64/hash.h int hash__map_kernel_page(unsigned long ea, unsigned long pa, pgprot_t prot);
pa                404 arch/powerpc/include/asm/book3s/64/mmu-hash.h static inline unsigned long hpte_encode_r(unsigned long pa, int base_psize,
pa                409 arch/powerpc/include/asm/book3s/64/mmu-hash.h 		return pa & HPTE_R_RPN;
pa                413 arch/powerpc/include/asm/book3s/64/mmu-hash.h 		return (pa & ~((1ul << shift) - 1)) | (penc << LP_SHIFT);
pa               1037 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int map_kernel_page(unsigned long ea, unsigned long pa, pgprot_t prot)
pa               1044 arch/powerpc/include/asm/book3s/64/pgtable.h 		return radix__map_kernel_page(ea, pa, prot, PAGE_SIZE);
pa               1046 arch/powerpc/include/asm/book3s/64/pgtable.h 	return hash__map_kernel_page(ea, pa, prot);
pa                277 arch/powerpc/include/asm/book3s/64/radix.h extern int radix__map_kernel_page(unsigned long ea, unsigned long pa,
pa                725 arch/powerpc/include/asm/io.h int early_ioremap_range(unsigned long ea, phys_addr_t pa,
pa                727 arch/powerpc/include/asm/io.h void __iomem *do_ioremap(phys_addr_t pa, phys_addr_t offset, unsigned long size,
pa                733 arch/powerpc/include/asm/io.h extern void __iomem * __ioremap_at(phys_addr_t pa, void *ea,
pa                 67 arch/powerpc/include/asm/nohash/32/pgtable.h int map_kernel_page(unsigned long va, phys_addr_t pa, pgprot_t prot);
pa                358 arch/powerpc/include/asm/nohash/64/pgtable.h int map_kernel_page(unsigned long ea, unsigned long pa, pgprot_t prot);
pa                356 arch/powerpc/kernel/eeh.c 	unsigned long pa;
pa                368 arch/powerpc/kernel/eeh.c 	pa = pte_pfn(*ptep);
pa                372 arch/powerpc/kernel/eeh.c 		pa <<= hugepage_shift;
pa                373 arch/powerpc/kernel/eeh.c 		pa |= token & ((1ul << hugepage_shift) - 1);
pa                375 arch/powerpc/kernel/eeh.c 		pa <<= PAGE_SHIFT;
pa                376 arch/powerpc/kernel/eeh.c 		pa |= token & (PAGE_SIZE - 1);
pa                379 arch/powerpc/kernel/eeh.c 	return pa;
pa                329 arch/powerpc/kernel/mce.c 	uint64_t ea = 0, pa = 0;
pa                445 arch/powerpc/kernel/mce.c 			pa = evt->u.ue_error.physical_address;
pa                520 arch/powerpc/kernel/mce.c 		if (pa)
pa                521 arch/powerpc/kernel/mce.c 			sprintf(dar_str + n, "paddr: %016llx ", pa);
pa                522 arch/powerpc/kernel/mce.c 	} else if (pa) {
pa                523 arch/powerpc/kernel/mce.c 		sprintf(pa_str, " paddr: %016llx", pa);
pa                201 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	unsigned long i, pa, gpa, gfn, psize;
pa                233 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	pa = 0;
pa                285 arch/powerpc/kvm/book3s_hv_rm_mmu.c 			pa = pte_pfn(pte) << PAGE_SHIFT;
pa                286 arch/powerpc/kvm/book3s_hv_rm_mmu.c 			pa |= hva & (host_pte_size - 1);
pa                287 arch/powerpc/kvm/book3s_hv_rm_mmu.c 			pa |= gpa & ~PAGE_MASK;
pa                294 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	ptel |= pa;
pa                296 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	if (pa)
pa                897 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	unsigned long gfn, hva, pa, psize = PAGE_SHIFT;
pa                921 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	pa = pte_pfn(pte) << PAGE_SHIFT;
pa                922 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	pa |= hva & (psize - 1);
pa                923 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	pa |= gpa & ~PAGE_MASK;
pa                926 arch/powerpc/kvm/book3s_hv_rm_mmu.c 		*hpa = pa;
pa                938 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	unsigned long pa, mmu_seq;
pa                946 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	ret = kvmppc_get_hpa(vcpu, dest, 1, &pa, &memslot);
pa                958 arch/powerpc/kvm/book3s_hv_rm_mmu.c 	for (i = 0; i < SZ_4K; i += L1_CACHE_BYTES, pa += L1_CACHE_BYTES)
pa                959 arch/powerpc/kvm/book3s_hv_rm_mmu.c 		dcbz((void *)pa);
pa                 62 arch/powerpc/mm/book3s32/mmu.c unsigned long p_block_mapped(phys_addr_t pa)
pa                 66 arch/powerpc/mm/book3s32/mmu.c 		if (pa >= bat_addrs[b].phys
pa                 67 arch/powerpc/mm/book3s32/mmu.c 	    	    && pa < (bat_addrs[b].limit-bat_addrs[b].start)
pa                 69 arch/powerpc/mm/book3s32/mmu.c 			return bat_addrs[b].start+(pa-bat_addrs[b].phys);
pa                 25 arch/powerpc/mm/book3s64/hash_4k.c 	unsigned long rflags, pa;
pa                 79 arch/powerpc/mm/book3s64/hash_4k.c 		pa = pte_pfn(__pte(old_pte)) << PAGE_SHIFT;
pa                 86 arch/powerpc/mm/book3s64/hash_4k.c 		slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa, rflags, 0,
pa                 93 arch/powerpc/mm/book3s64/hash_4k.c 			slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa,
pa                 43 arch/powerpc/mm/book3s64/hash_64k.c 	unsigned long rflags, pa;
pa                150 arch/powerpc/mm/book3s64/hash_64k.c 		pa = pte_pfn(__pte(old_pte)) << HW_PAGE_SHIFT;
pa                152 arch/powerpc/mm/book3s64/hash_64k.c 		pa = pte_pfn(__pte(old_pte)) << PAGE_SHIFT;
pa                153 arch/powerpc/mm/book3s64/hash_64k.c 		pa += (subpg_index << shift);
pa                160 arch/powerpc/mm/book3s64/hash_64k.c 	slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa, rflags, 0,
pa                169 arch/powerpc/mm/book3s64/hash_64k.c 		slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa,
pa                229 arch/powerpc/mm/book3s64/hash_64k.c 	unsigned long rflags, pa;
pa                286 arch/powerpc/mm/book3s64/hash_64k.c 		pa = pte_pfn(__pte(old_pte)) << PAGE_SHIFT;
pa                293 arch/powerpc/mm/book3s64/hash_64k.c 		slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa, rflags, 0,
pa                301 arch/powerpc/mm/book3s64/hash_64k.c 			slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa,
pa                 27 arch/powerpc/mm/book3s64/hash_hugepage.c 	unsigned long rflags, pa, hidx;
pa                133 arch/powerpc/mm/book3s64/hash_hugepage.c 		pa = pmd_pfn(__pmd(old_pmd)) << PAGE_SHIFT;
pa                140 arch/powerpc/mm/book3s64/hash_hugepage.c 		slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa, rflags, 0,
pa                147 arch/powerpc/mm/book3s64/hash_hugepage.c 			slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa,
pa                 22 arch/powerpc/mm/book3s64/hash_hugetlbpage.c 				  unsigned long pa, unsigned long rlags,
pa                 32 arch/powerpc/mm/book3s64/hash_hugetlbpage.c 	unsigned long rflags, pa;
pa                103 arch/powerpc/mm/book3s64/hash_hugetlbpage.c 		pa = pte_pfn(__pte(old_pte)) << PAGE_SHIFT;
pa                108 arch/powerpc/mm/book3s64/hash_hugetlbpage.c 		slot = hpte_insert_repeating(hash, vpn, pa, rflags, 0,
pa                338 arch/powerpc/mm/book3s64/hash_native.c 			unsigned long pa, unsigned long rflags,
pa                348 arch/powerpc/mm/book3s64/hash_native.c 			hpte_group, vpn, pa, rflags, vflags, psize);
pa                367 arch/powerpc/mm/book3s64/hash_native.c 	hpte_r = hpte_encode_r(pa, psize, apsize) | rflags;
pa                148 arch/powerpc/mm/book3s64/hash_pgtable.c int hash__map_kernel_page(unsigned long ea, unsigned long pa, pgprot_t prot)
pa                167 arch/powerpc/mm/book3s64/hash_pgtable.c 		set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, prot));
pa                175 arch/powerpc/mm/book3s64/hash_pgtable.c 		if (htab_bolt_mapping(ea, ea + PAGE_SIZE, pa, pgprot_val(prot),
pa                178 arch/powerpc/mm/book3s64/hash_pgtable.c 			       "memory at %016lx !\n", pa);
pa               1838 arch/powerpc/mm/book3s64/hash_utils.c 			   unsigned long pa, unsigned long rflags,
pa               1848 arch/powerpc/mm/book3s64/hash_utils.c 	slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa, rflags, vflags,
pa               1854 arch/powerpc/mm/book3s64/hash_utils.c 		slot = mmu_hash_ops.hpte_insert(hpte_group, vpn, pa, rflags,
pa                369 arch/powerpc/mm/book3s64/iommu_api.c 	unsigned long *pa;
pa                382 arch/powerpc/mm/book3s64/iommu_api.c 	pa = (void *) vmalloc_to_phys(&mem->hpas[entry]);
pa                383 arch/powerpc/mm/book3s64/iommu_api.c 	if (!pa)
pa                386 arch/powerpc/mm/book3s64/iommu_api.c 	*hpa = (*pa & MM_IOMMU_TABLE_GROUP_PAGE_MASK) | (ua & ~PAGE_MASK);
pa                396 arch/powerpc/mm/book3s64/iommu_api.c 	unsigned long *pa;
pa                408 arch/powerpc/mm/book3s64/iommu_api.c 	pa = (void *) vmalloc_to_phys(va);
pa                409 arch/powerpc/mm/book3s64/iommu_api.c 	if (!pa)
pa                412 arch/powerpc/mm/book3s64/iommu_api.c 	*pa |= MM_IOMMU_TABLE_GROUP_PAGE_DIRTY;
pa                 58 arch/powerpc/mm/book3s64/radix_pgtable.c static int early_map_kernel_page(unsigned long ea, unsigned long pa,
pa                 64 arch/powerpc/mm/book3s64/radix_pgtable.c 	unsigned long pfn = pa >> PAGE_SHIFT;
pa                108 arch/powerpc/mm/book3s64/radix_pgtable.c static int __map_kernel_page(unsigned long ea, unsigned long pa,
pa                114 arch/powerpc/mm/book3s64/radix_pgtable.c 	unsigned long pfn = pa >> PAGE_SHIFT;
pa                129 arch/powerpc/mm/book3s64/radix_pgtable.c 		return early_map_kernel_page(ea, pa, flags, map_page_size,
pa                162 arch/powerpc/mm/book3s64/radix_pgtable.c int radix__map_kernel_page(unsigned long ea, unsigned long pa,
pa                166 arch/powerpc/mm/book3s64/radix_pgtable.c 	return __map_kernel_page(ea, pa, flags, map_page_size, -1, 0, 0);
pa                886 arch/powerpc/mm/book3s64/radix_pgtable.c static int __map_kernel_page_nid(unsigned long ea, unsigned long pa,
pa                890 arch/powerpc/mm/book3s64/radix_pgtable.c 	return __map_kernel_page(ea, pa, flags, map_page_size, nid, 0, 0);
pa                 62 arch/powerpc/mm/ioremap.c int early_ioremap_range(unsigned long ea, phys_addr_t pa,
pa                 68 arch/powerpc/mm/ioremap.c 		int err = map_kernel_page(ea + i, pa + i, prot);
pa                 77 arch/powerpc/mm/ioremap.c void __iomem *do_ioremap(phys_addr_t pa, phys_addr_t offset, unsigned long size,
pa                 88 arch/powerpc/mm/ioremap.c 	area->phys_addr = pa;
pa                 91 arch/powerpc/mm/ioremap.c 	ret = ioremap_page_range(va, va + size, pa, prot);
pa                 10 arch/powerpc/mm/ioremap_64.c void __iomem *__ioremap_at(phys_addr_t pa, void *ea, unsigned long size, pgprot_t prot)
pa                 24 arch/powerpc/mm/ioremap_64.c 	WARN_ON(pa & ~PAGE_MASK);
pa                 29 arch/powerpc/mm/ioremap_64.c 		ret = ioremap_page_range(va, va + size, pa, prot);
pa                 33 arch/powerpc/mm/ioremap_64.c 		ret = early_ioremap_range(va, pa, size, prot);
pa                 26 arch/powerpc/mm/kasan/kasan_init_32.c 	phys_addr_t pa = __pa(kasan_early_shadow_page);
pa                 30 arch/powerpc/mm/kasan/kasan_init_32.c 		__set_pte_at(&init_mm, va, ptep, pfn_pte(PHYS_PFN(pa), prot), 0);
pa                116 arch/powerpc/mm/kasan/kasan_init_32.c 	phys_addr_t pa = __pa(kasan_early_shadow_page);
pa                124 arch/powerpc/mm/kasan/kasan_init_32.c 		if ((pte_val(*ptep) & PTE_RPN_MASK) != pa)
pa                127 arch/powerpc/mm/kasan/kasan_init_32.c 		__set_pte_at(&init_mm, k_cur, ptep, pfn_pte(PHYS_PFN(pa), prot), 0);
pa                160 arch/powerpc/mm/mmu_decl.h unsigned long p_block_mapped(phys_addr_t pa);
pa                163 arch/powerpc/mm/mmu_decl.h static inline unsigned long p_block_mapped(phys_addr_t pa) { return 0; }
pa                 43 arch/powerpc/mm/nohash/8xx.c unsigned long p_block_mapped(phys_addr_t pa)
pa                 49 arch/powerpc/mm/nohash/8xx.c 	if (pa >= p && pa < p + IMMR_SIZE)
pa                 50 arch/powerpc/mm/nohash/8xx.c 		return VIRT_IMMR_BASE + pa - p;
pa                 51 arch/powerpc/mm/nohash/8xx.c 	if (pa < block_mapped_ram)
pa                 52 arch/powerpc/mm/nohash/8xx.c 		return (unsigned long)__va(pa);
pa                 73 arch/powerpc/mm/nohash/book3e_pgtable.c int __ref map_kernel_page(unsigned long ea, unsigned long pa, pgprot_t prot)
pa                112 arch/powerpc/mm/nohash/book3e_pgtable.c 	set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, prot));
pa                 86 arch/powerpc/mm/nohash/fsl_booke.c unsigned long p_block_mapped(phys_addr_t pa)
pa                 90 arch/powerpc/mm/nohash/fsl_booke.c 		if (pa >= tlbcam_addrs[b].phys
pa                 91 arch/powerpc/mm/nohash/fsl_booke.c 			&& pa < (tlbcam_addrs[b].limit-tlbcam_addrs[b].start)
pa                 93 arch/powerpc/mm/nohash/fsl_booke.c 			return tlbcam_addrs[b].start+(pa-tlbcam_addrs[b].phys);
pa                 59 arch/powerpc/mm/pgtable_32.c int __ref map_kernel_page(unsigned long va, phys_addr_t pa, pgprot_t prot)
pa                 78 arch/powerpc/mm/pgtable_32.c 		set_pte_at(&init_mm, va, pg, pfn_pte(pa >> PAGE_SHIFT, prot));
pa                195 arch/powerpc/mm/ptdump/ptdump.c 	u64 pa = val & PTE_RPN_MASK;
pa                202 arch/powerpc/mm/ptdump/ptdump.c 		st->start_pa = pa;
pa                203 arch/powerpc/mm/ptdump/ptdump.c 		st->last_pa = pa;
pa                216 arch/powerpc/mm/ptdump/ptdump.c 		   (pa != st->last_pa + st->page_size &&
pa                217 arch/powerpc/mm/ptdump/ptdump.c 		    (pa != st->start_pa || st->start_pa != st->last_pa))) {
pa                242 arch/powerpc/mm/ptdump/ptdump.c 		st->start_pa = pa;
pa                243 arch/powerpc/mm/ptdump/ptdump.c 		st->last_pa = pa;
pa                248 arch/powerpc/mm/ptdump/ptdump.c 		st->last_pa = pa;
pa                 36 arch/powerpc/platforms/ps3/htab.c 	unsigned long pa, unsigned long rflags, unsigned long vflags,
pa                 54 arch/powerpc/platforms/ps3/htab.c 	hpte_r = hpte_encode_r(ps3_mm_phys_to_lpar(pa), psize, apsize) | rflags;
pa                 68 arch/powerpc/platforms/ps3/htab.c 			__func__, ps3_result(result), vpn, pa, hpte_group,
pa                141 arch/powerpc/platforms/pseries/cmm.c 	struct cmm_page_array *pa, *npa;
pa                164 arch/powerpc/platforms/pseries/cmm.c 		pa = cmm_page_list;
pa                165 arch/powerpc/platforms/pseries/cmm.c 		if (!pa || pa->index >= CMM_NR_PAGES) {
pa                177 arch/powerpc/platforms/pseries/cmm.c 			pa = cmm_page_list;
pa                179 arch/powerpc/platforms/pseries/cmm.c 			if (!pa || pa->index >= CMM_NR_PAGES) {
pa                180 arch/powerpc/platforms/pseries/cmm.c 				npa->next = pa;
pa                182 arch/powerpc/platforms/pseries/cmm.c 				pa = npa;
pa                183 arch/powerpc/platforms/pseries/cmm.c 				cmm_page_list = pa;
pa                195 arch/powerpc/platforms/pseries/cmm.c 		pa->page[pa->index++] = addr;
pa                215 arch/powerpc/platforms/pseries/cmm.c 	struct cmm_page_array *pa;
pa                220 arch/powerpc/platforms/pseries/cmm.c 	pa = cmm_page_list;
pa                222 arch/powerpc/platforms/pseries/cmm.c 		if (!pa || pa->index <= 0)
pa                224 arch/powerpc/platforms/pseries/cmm.c 		addr = pa->page[--pa->index];
pa                226 arch/powerpc/platforms/pseries/cmm.c 		if (pa->index == 0) {
pa                227 arch/powerpc/platforms/pseries/cmm.c 			pa = pa->next;
pa                229 arch/powerpc/platforms/pseries/cmm.c 			cmm_page_list = pa;
pa                499 arch/powerpc/platforms/pseries/cmm.c 	struct cmm_page_array *pa;
pa                505 arch/powerpc/platforms/pseries/cmm.c 	pa = cmm_page_list;
pa                506 arch/powerpc/platforms/pseries/cmm.c 	while (pa) {
pa                507 arch/powerpc/platforms/pseries/cmm.c 		if ((unsigned long)pa >= start && (unsigned long)pa < end)
pa                509 arch/powerpc/platforms/pseries/cmm.c 		for (idx = 0; idx < pa->index; idx++)
pa                510 arch/powerpc/platforms/pseries/cmm.c 			if (pa->page[idx] >= start && pa->page[idx] < end)
pa                512 arch/powerpc/platforms/pseries/cmm.c 		pa = pa->next;
pa                706 arch/powerpc/platforms/pseries/lpar.c 				     unsigned long vpn, unsigned long pa,
pa                718 arch/powerpc/platforms/pseries/lpar.c 			 hpte_group, vpn,  pa, rflags, vflags, psize);
pa                721 arch/powerpc/platforms/pseries/lpar.c 	hpte_r = hpte_encode_r(pa, psize, apsize) | rflags;
pa                178 arch/riscv/mm/init.c static pte_t *__init get_pte_virt(phys_addr_t pa)
pa                182 arch/riscv/mm/init.c 		return (pte_t *)set_fixmap_offset(FIX_PTE, pa);
pa                184 arch/riscv/mm/init.c 		return (pte_t *)((uintptr_t)pa);
pa                200 arch/riscv/mm/init.c 				      uintptr_t va, phys_addr_t pa,
pa                208 arch/riscv/mm/init.c 		ptep[pte_index] = pfn_pte(PFN_DOWN(pa), prot);
pa                223 arch/riscv/mm/init.c static pmd_t *__init get_pmd_virt(phys_addr_t pa)
pa                227 arch/riscv/mm/init.c 		return (pmd_t *)set_fixmap_offset(FIX_PMD, pa);
pa                229 arch/riscv/mm/init.c 		return (pmd_t *)((uintptr_t)pa);
pa                246 arch/riscv/mm/init.c 				      uintptr_t va, phys_addr_t pa,
pa                255 arch/riscv/mm/init.c 			pmdp[pmd_index] = pfn_pmd(PFN_DOWN(pa), prot);
pa                269 arch/riscv/mm/init.c 	create_pte_mapping(ptep, va, pa, sz, prot);
pa                290 arch/riscv/mm/init.c 				      uintptr_t va, phys_addr_t pa,
pa                299 arch/riscv/mm/init.c 			pgdp[pgd_index] = pfn_pgd(PFN_DOWN(pa), prot);
pa                313 arch/riscv/mm/init.c 	create_pgd_next_mapping(nextp, va, pa, sz, prot);
pa                413 arch/riscv/mm/init.c 	phys_addr_t pa, start, end;
pa                438 arch/riscv/mm/init.c 		for (pa = start; pa < end; pa += map_size) {
pa                439 arch/riscv/mm/init.c 			va = (uintptr_t)__va(pa);
pa                440 arch/riscv/mm/init.c 			create_pgd_mapping(swapper_pg_dir, va, pa,
pa                584 arch/s390/kernel/smp.c 	unsigned long pa;
pa                586 arch/s390/kernel/smp.c 	pa = __pa(&pcpu->lowcore->floating_pt_save_area);
pa                588 arch/s390/kernel/smp.c 			      pa) != SIGP_CC_ORDER_CODE_ACCEPTED)
pa                592 arch/s390/kernel/smp.c 	pa = __pa(pcpu->lowcore->mcesad & MCESA_ORIGIN_MASK);
pa                594 arch/s390/kernel/smp.c 		pa |= pcpu->lowcore->mcesad & MCESA_LC_MASK;
pa                596 arch/s390/kernel/smp.c 			      pa) != SIGP_CC_ORDER_CODE_ACCEPTED)
pa                 68 arch/s390/mm/cmm.c 	struct cmm_page_array *pa, *npa;
pa                 76 arch/s390/mm/cmm.c 		pa = *list;
pa                 77 arch/s390/mm/cmm.c 		if (!pa || pa->index >= CMM_NR_PAGES) {
pa                 87 arch/s390/mm/cmm.c 			pa = *list;
pa                 88 arch/s390/mm/cmm.c 			if (!pa || pa->index >= CMM_NR_PAGES) {
pa                 89 arch/s390/mm/cmm.c 				npa->next = pa;
pa                 91 arch/s390/mm/cmm.c 				pa = npa;
pa                 92 arch/s390/mm/cmm.c 				*list = pa;
pa                 97 arch/s390/mm/cmm.c 		pa->pages[pa->index++] = addr;
pa                107 arch/s390/mm/cmm.c 	struct cmm_page_array *pa;
pa                111 arch/s390/mm/cmm.c 	pa = *list;
pa                113 arch/s390/mm/cmm.c 		if (!pa || pa->index <= 0)
pa                115 arch/s390/mm/cmm.c 		addr = pa->pages[--pa->index];
pa                116 arch/s390/mm/cmm.c 		if (pa->index == 0) {
pa                117 arch/s390/mm/cmm.c 			pa = pa->next;
pa                119 arch/s390/mm/cmm.c 			*list = pa;
pa                133 arch/s390/pci/pci_dma.c static int __dma_update_trans(struct zpci_dev *zdev, unsigned long pa,
pa                137 arch/s390/pci/pci_dma.c 	u8 *page_addr = (u8 *) (pa & PAGE_MASK);
pa                218 arch/s390/pci/pci_dma.c static int dma_update_trans(struct zpci_dev *zdev, unsigned long pa,
pa                223 arch/s390/pci/pci_dma.c 	rc = __dma_update_trans(zdev, pa, dma_addr, size, flags);
pa                229 arch/s390/pci/pci_dma.c 		__dma_update_trans(zdev, pa, dma_addr, size, ZPCI_PTE_INVALID);
pa                341 arch/s390/pci/pci_dma.c 	unsigned long pa = page_to_phys(page) + offset;
pa                348 arch/s390/pci/pci_dma.c 	nr_pages = iommu_num_pages(pa, size, PAGE_SIZE);
pa                361 arch/s390/pci/pci_dma.c 	ret = dma_update_trans(zdev, pa, dma_addr, size, flags);
pa                372 arch/s390/pci/pci_dma.c 	zpci_err_dma(ret, pa);
pa                403 arch/s390/pci/pci_dma.c 	unsigned long pa;
pa                411 arch/s390/pci/pci_dma.c 	pa = page_to_phys(page);
pa                414 arch/s390/pci/pci_dma.c 		free_pages(pa, get_order(size));
pa                421 arch/s390/pci/pci_dma.c 	return (void *) pa;
pa                425 arch/s390/pci/pci_dma.c 			  void *pa, dma_addr_t dma_handle,
pa                433 arch/s390/pci/pci_dma.c 	free_pages((unsigned long) pa, get_order(size));
pa                446 arch/s390/pci/pci_dma.c 	unsigned long pa = 0;
pa                458 arch/s390/pci/pci_dma.c 		pa = page_to_phys(sg_page(s));
pa                459 arch/s390/pci/pci_dma.c 		ret = __dma_update_trans(zdev, pa, dma_addr,
pa                480 arch/s390/pci/pci_dma.c 	zpci_err_dma(ret, pa);
pa                 31 arch/sparc/include/asm/pgalloc_32.h 	unsigned long pa = __nocache_pa(pmdp);
pa                 33 arch/sparc/include/asm/pgalloc_32.h 	set_pte((pte_t *)pgdp, __pte((SRMMU_ET_PTD | (pa >> 4))));
pa                 58 arch/sparc/kernel/ioport.c static inline void dma_make_coherent(unsigned long pa, unsigned long len)
pa                 66 arch/sparc/kernel/ioport.c static void __iomem *_sparc_ioremap(struct resource *res, u32 bus, u32 pa, int sz);
pa                216 arch/sparc/kernel/ioport.c _sparc_ioremap(struct resource *res, u32 bus, u32 pa, int sz)
pa                218 arch/sparc/kernel/ioport.c 	unsigned long offset = ((unsigned long) pa) & (~PAGE_MASK);
pa                229 arch/sparc/kernel/ioport.c 	pa &= PAGE_MASK;
pa                230 arch/sparc/kernel/ioport.c 	srmmu_mapiorange(bus, pa, res->start, resource_size(res));
pa               2084 arch/sparc/kernel/ldc.c static void fill_cookies(struct cookie_state *sp, unsigned long pa,
pa               2088 arch/sparc/kernel/ldc.c 		unsigned long tlen, new = pa + PAGE_SIZE;
pa               2091 arch/sparc/kernel/ldc.c 		sp->page_table[sp->pte_idx].mte = sp->mte_base | pa;
pa               2116 arch/sparc/kernel/ldc.c 		pa = new;
pa               2198 arch/sparc/kernel/ldc.c 	unsigned long npages, pa;
pa               2206 arch/sparc/kernel/ldc.c 	pa = __pa(buf);
pa               2207 arch/sparc/kernel/ldc.c 	if ((pa | len) & (8UL - 1))
pa               2210 arch/sparc/kernel/ldc.c 	npages = pages_in_region(pa, len);
pa               2225 arch/sparc/kernel/ldc.c 	fill_cookies(&state, (pa & PAGE_MASK), (pa & ~PAGE_MASK), len);
pa                 86 arch/sparc/mm/fault_64.c 	unsigned long pa;
pa                104 arch/sparc/mm/fault_64.c 		pa  = pmd_pfn(*pmdp) << PAGE_SHIFT;
pa                105 arch/sparc/mm/fault_64.c 		pa += tpc & ~HPAGE_MASK;
pa                110 arch/sparc/mm/fault_64.c 				     : "r" (pa), "i" (ASI_PHYS_USE_EC));
pa                117 arch/sparc/mm/fault_64.c 			pa  = (pte_pfn(pte) << PAGE_SHIFT);
pa                118 arch/sparc/mm/fault_64.c 			pa += (tpc & ~PAGE_MASK);
pa                123 arch/sparc/mm/fault_64.c 					     : "r" (pa), "i" (ASI_PHYS_USE_EC));
pa               1661 arch/sparc/mm/init_64.c 		unsigned long pa = __pa(addr);
pa               1663 arch/sparc/mm/init_64.c 		if ((pa >> max_phys_bits) != 0UL)
pa               1666 arch/sparc/mm/init_64.c 		return pfn_valid(pa >> PAGE_SHIFT);
pa               2093 arch/sparc/mm/init_64.c static void patch_one_ktsb_phys(unsigned int *start, unsigned int *end, unsigned long pa)
pa               2097 arch/sparc/mm/init_64.c 	high_bits = (pa >> 32) & 0xffffffff;
pa               2098 arch/sparc/mm/init_64.c 	low_bits = (pa >> 0) & 0xffffffff;
pa               2197 arch/sparc/mm/init_64.c 	unsigned long pa, ret;
pa               2199 arch/sparc/mm/init_64.c 	pa = kern_base + ((unsigned long)&ktsb_descr[0] - KERNBASE);
pa               2201 arch/sparc/mm/init_64.c 	ret = sun4v_mmu_tsb_ctx0(NUM_KTSB_DESCR, pa);
pa               2204 arch/sparc/mm/init_64.c 			    "errors with %lx\n", pa, ret);
pa                255 arch/x86/boot/compressed/acpi.c 	acpi_physical_address pa;
pa                257 arch/x86/boot/compressed/acpi.c 	pa = boot_params->acpi_rsdp_addr;
pa                264 arch/x86/boot/compressed/acpi.c 	if (!pa)
pa                265 arch/x86/boot/compressed/acpi.c 		pa = kexec_get_rsdp_addr();
pa                267 arch/x86/boot/compressed/acpi.c 	if (!pa)
pa                268 arch/x86/boot/compressed/acpi.c 		pa = efi_get_rsdp_addr();
pa                270 arch/x86/boot/compressed/acpi.c 	if (!pa)
pa                271 arch/x86/boot/compressed/acpi.c 		pa = bios_get_rsdp_addr();
pa                273 arch/x86/boot/compressed/acpi.c 	return pa;
pa                288 arch/x86/events/intel/ds.c 	phys_addr_t pa;
pa                291 arch/x86/events/intel/ds.c 	pa = virt_to_phys(addr);
pa                294 arch/x86/events/intel/ds.c 	for (; msz < size; msz += PAGE_SIZE, pa += PAGE_SIZE, cea += PAGE_SIZE)
pa                295 arch/x86/events/intel/ds.c 		cea_set_pte(cea, pa, prot);
pa                395 arch/x86/hyperv/hv_init.c void hyperv_report_panic_msg(phys_addr_t pa, size_t size)
pa                405 arch/x86/hyperv/hv_init.c 	wrmsrl(HV_X64_MSR_CRASH_P3, pa);
pa                123 arch/x86/include/asm/cpu_entry_area.h extern void cea_set_pte(void *cea_vaddr, phys_addr_t pa, pgprot_t flags);
pa                454 arch/x86/include/asm/uv/uv_hub.h static inline struct uv_gam_range_s *uv_gam_range(unsigned long pa)
pa                457 arch/x86/include/asm/uv/uv_hub.h 	unsigned long pal = (pa & uv_hub_info->gpa_mask) >> UV_GAM_RANGE_SHFT;
pa                466 arch/x86/include/asm/uv/uv_hub.h 	pr_crit("UV: GAM Range for 0x%lx not found at %p!\n", pa, gr);
pa                471 arch/x86/include/asm/uv/uv_hub.h static inline unsigned long uv_gam_range_base(unsigned long pa)
pa                473 arch/x86/include/asm/uv/uv_hub.h 	struct uv_gam_range_s *gr = uv_gam_range(pa);
pa                 41 arch/x86/kernel/irq_64.c 		phys_addr_t pa = per_cpu_ptr_to_phys(stack + (i << PAGE_SHIFT));
pa                 43 arch/x86/kernel/irq_64.c 		pages[i] = pfn_to_page(pa >> PAGE_SHIFT);
pa                 36 arch/x86/kernel/kdebugfs.c 	u64 pa;
pa                 47 arch/x86/kernel/kdebugfs.c 	pa = node->paddr + sizeof(struct setup_data) + pos;
pa                 48 arch/x86/kernel/kdebugfs.c 	p = memremap(pa, count, MEMREMAP_WB);
pa                312 arch/x86/kernel/kvm.c 		u64 pa = slow_virt_to_phys(this_cpu_ptr(&apf_reason));
pa                315 arch/x86/kernel/kvm.c 		pa |= KVM_ASYNC_PF_SEND_ALWAYS;
pa                317 arch/x86/kernel/kvm.c 		pa |= KVM_ASYNC_PF_ENABLED;
pa                320 arch/x86/kernel/kvm.c 			pa |= KVM_ASYNC_PF_DELIVERY_AS_PF_VMEXIT;
pa                322 arch/x86/kernel/kvm.c 		wrmsrl(MSR_KVM_ASYNC_PF_EN, pa);
pa                329 arch/x86/kernel/kvm.c 		unsigned long pa;
pa                333 arch/x86/kernel/kvm.c 		pa = slow_virt_to_phys(this_cpu_ptr(&kvm_apic_eoi))
pa                335 arch/x86/kernel/kvm.c 		wrmsrl(MSR_KVM_PV_EOI_EN, pa);
pa                174 arch/x86/kernel/kvmclock.c 	u64 pa;
pa                179 arch/x86/kernel/kvmclock.c 	pa = slow_virt_to_phys(&src->pvti) | 0x01ULL;
pa                180 arch/x86/kernel/kvmclock.c 	wrmsrl(msr_kvm_system_time, pa);
pa                181 arch/x86/kernel/kvmclock.c 	pr_info("kvm-clock: cpu %d, msr %llx, %s", smp_processor_id(), pa, txt);
pa                 29 arch/x86/mm/cpu_entry_area.c void cea_set_pte(void *cea_vaddr, phys_addr_t pa, pgprot_t flags)
pa                 32 arch/x86/mm/cpu_entry_area.c 	pte_t pte = pfn_pte(pa >> PAGE_SHIFT, flags);
pa                202 arch/x86/mm/mem_encrypt.c 	unsigned long pfn, pa, size;
pa                232 arch/x86/mm/mem_encrypt.c 	pa = pfn << page_level_shift(level);
pa                240 arch/x86/mm/mem_encrypt.c 	clflush_cache_range(__va(pa), size);
pa                244 arch/x86/mm/mem_encrypt.c 		sme_early_encrypt(pa, size);
pa                246 arch/x86/mm/mem_encrypt.c 		sme_early_decrypt(pa, size);
pa                460 arch/x86/mm/pti.c 		phys_addr_t pa = per_cpu_ptr_to_phys((void *)va);
pa                467 arch/x86/mm/pti.c 		*target_pte = pfn_pte(pa >> PAGE_SHIFT, PAGE_KERNEL);
pa                 28 arch/x86/mm/srat.c acpi_numa_x2apic_affinity_init(struct acpi_srat_x2apic_cpu_affinity *pa)
pa                 35 arch/x86/mm/srat.c 	if (pa->header.length < sizeof(struct acpi_srat_x2apic_cpu_affinity)) {
pa                 39 arch/x86/mm/srat.c 	if ((pa->flags & ACPI_SRAT_CPU_ENABLED) == 0)
pa                 41 arch/x86/mm/srat.c 	pxm = pa->proximity_domain;
pa                 42 arch/x86/mm/srat.c 	apic_id = pa->apic_id;
pa                 67 arch/x86/mm/srat.c acpi_numa_processor_affinity_init(struct acpi_srat_cpu_affinity *pa)
pa                 74 arch/x86/mm/srat.c 	if (pa->header.length != sizeof(struct acpi_srat_cpu_affinity)) {
pa                 78 arch/x86/mm/srat.c 	if ((pa->flags & ACPI_SRAT_CPU_ENABLED) == 0)
pa                 80 arch/x86/mm/srat.c 	pxm = pa->proximity_domain_lo;
pa                 82 arch/x86/mm/srat.c 		pxm |= *((unsigned int*)pa->proximity_domain_hi) << 8;
pa                 91 arch/x86/mm/srat.c 		apic_id = (pa->apic_id << 8) | pa->local_sapic_eid;
pa                 93 arch/x86/mm/srat.c 		apic_id = pa->apic_id;
pa                953 arch/x86/platform/efi/efi.c 	unsigned long pa;
pa                969 arch/x86/platform/efi/efi.c 	pa = __pa(new_memmap);
pa                978 arch/x86/platform/efi/efi.c 	if (efi_memmap_init_late(pa, efi.memmap.desc_size * count)) {
pa                991 arch/x86/platform/efi/efi.c 	if (efi_setup_page_tables(pa, 1 << pg_shift))
pa               1001 arch/x86/platform/efi/efi.c 				(efi_memory_desc_t *)pa);
pa               1008 arch/x86/platform/efi/efi.c 				(efi_memory_desc_t *)pa);
pa                319 arch/x86/platform/efi/efi_64.c 	phys_addr_t pa;
pa                327 arch/x86/platform/efi/efi_64.c 	pa = slow_virt_to_phys(va);
pa                330 arch/x86/platform/efi/efi_64.c 	if (WARN_ON((pa ^ (pa + size - 1)) & PAGE_MASK))
pa                333 arch/x86/platform/efi/efi_64.c 	return pa;
pa                433 arch/x86/platform/efi/efi_64.c 	u64 pa = md->phys_addr;
pa                458 arch/x86/platform/efi/efi_64.c 	if (!(pa & (PMD_SIZE - 1))) {
pa                461 arch/x86/platform/efi/efi_64.c 		u64 pa_offset = pa & (PMD_SIZE - 1);
pa                379 arch/x86/platform/efi/quirks.c 	u64 pa = md->phys_addr;
pa                398 arch/x86/platform/efi/quirks.c 	if (kernel_unmap_pages_in_pgd(pgd, pa, md->num_pages))
pa                399 arch/x86/platform/efi/quirks.c 		pr_err("Failed to unmap 1:1 mapping for 0x%llx\n", pa);
pa                 44 arch/x86/xen/enlighten_hvm.c 	u64 pa;
pa                 54 arch/x86/xen/enlighten_hvm.c 	for (pa = PAGE_SIZE;
pa                 55 arch/x86/xen/enlighten_hvm.c 	     !e820__mapped_all(pa, pa + PAGE_SIZE, E820_TYPE_RAM) ||
pa                 56 arch/x86/xen/enlighten_hvm.c 	     memblock_is_reserved(pa);
pa                 57 arch/x86/xen/enlighten_hvm.c 	     pa += PAGE_SIZE)
pa                 60 arch/x86/xen/enlighten_hvm.c 	shared_info_pfn = PHYS_PFN(pa);
pa                 62 arch/x86/xen/enlighten_hvm.c 	memblock_reserve(pa, PAGE_SIZE);
pa                 63 arch/x86/xen/enlighten_hvm.c 	HYPERVISOR_shared_info = early_memremap(pa, PAGE_SIZE);
pa               1129 arch/x86/xen/mmu_pv.c 	unsigned long pa = __pa(pgtbl) & PHYSICAL_PAGE_MASK;
pa               1132 arch/x86/xen/mmu_pv.c 		pin_pagetable_pfn(MMUEXT_UNPIN_TABLE, PFN_DOWN(pa));
pa               1133 arch/x86/xen/mmu_pv.c 	ClearPagePinned(virt_to_page(__va(pa)));
pa               1134 arch/x86/xen/mmu_pv.c 	xen_free_ro_pages(pa, PAGE_SIZE);
pa               1139 arch/x86/xen/mmu_pv.c 	unsigned long pa;
pa               1144 arch/x86/xen/mmu_pv.c 		pa = pmd_val(*pmd) & PHYSICAL_PAGE_MASK;
pa               1145 arch/x86/xen/mmu_pv.c 		xen_free_ro_pages(pa, PMD_SIZE);
pa               1153 arch/x86/xen/mmu_pv.c 		pa = pte_pfn(pte_tbl[i]) << PAGE_SHIFT;
pa               1154 arch/x86/xen/mmu_pv.c 		xen_free_ro_pages(pa, PAGE_SIZE);
pa               1162 arch/x86/xen/mmu_pv.c 	unsigned long pa;
pa               1167 arch/x86/xen/mmu_pv.c 		pa = pud_val(*pud) & PHYSICAL_PAGE_MASK;
pa               1168 arch/x86/xen/mmu_pv.c 		xen_free_ro_pages(pa, PUD_SIZE);
pa               1184 arch/x86/xen/mmu_pv.c 	unsigned long pa;
pa               1189 arch/x86/xen/mmu_pv.c 		pa = p4d_val(*p4d) & PHYSICAL_PAGE_MASK;
pa               1190 arch/x86/xen/mmu_pv.c 		xen_free_ro_pages(pa, P4D_SIZE);
pa               2014 arch/x86/xen/mmu_pv.c 	phys_addr_t pa;
pa               2020 arch/x86/xen/mmu_pv.c 	pa = read_cr3_pa();
pa               2021 arch/x86/xen/mmu_pv.c 	pgd = native_make_pgd(xen_read_phys_ulong(pa + pgd_index(vaddr) *
pa               2026 arch/x86/xen/mmu_pv.c 	pa = pgd_val(pgd) & PTE_PFN_MASK;
pa               2027 arch/x86/xen/mmu_pv.c 	pud = native_make_pud(xen_read_phys_ulong(pa + pud_index(vaddr) *
pa               2031 arch/x86/xen/mmu_pv.c 	pa = pud_val(pud) & PTE_PFN_MASK;
pa               2033 arch/x86/xen/mmu_pv.c 		return pa + (vaddr & ~PUD_MASK);
pa               2035 arch/x86/xen/mmu_pv.c 	pmd = native_make_pmd(xen_read_phys_ulong(pa + pmd_index(vaddr) *
pa               2039 arch/x86/xen/mmu_pv.c 	pa = pmd_val(pmd) & PTE_PFN_MASK;
pa               2041 arch/x86/xen/mmu_pv.c 		return pa + (vaddr & ~PMD_MASK);
pa               2043 arch/x86/xen/mmu_pv.c 	pte = native_make_pte(xen_read_phys_ulong(pa + pte_index(vaddr) *
pa               2047 arch/x86/xen/mmu_pv.c 	pa = pte_pfn(pte) << PAGE_SHIFT;
pa               2049 arch/x86/xen/mmu_pv.c 	return pa | (vaddr & ~PAGE_MASK);
pa                322 drivers/acpi/numa.c acpi_numa_x2apic_affinity_init(struct acpi_srat_x2apic_cpu_affinity *pa)
pa                324 drivers/acpi/numa.c 	pr_warn("Found unsupported x2apic [0x%08x] SRAT entry\n", pa->apic_id);
pa                183 drivers/acpi/osl.c 	acpi_physical_address pa;
pa                200 drivers/acpi/osl.c 	pa = acpi_arch_get_root_pointer();
pa                201 drivers/acpi/osl.c 	if (pa)
pa                202 drivers/acpi/osl.c 		return pa;
pa                211 drivers/acpi/osl.c 		acpi_find_root_pointer(&pa);
pa                214 drivers/acpi/osl.c 	return pa;
pa                360 drivers/base/platform.c 	struct platform_object *pa = container_of(dev, struct platform_object,
pa                363 drivers/base/platform.c 	of_device_node_put(&pa->pdev.dev);
pa                364 drivers/base/platform.c 	kfree(pa->pdev.dev.platform_data);
pa                365 drivers/base/platform.c 	kfree(pa->pdev.mfd_cell);
pa                366 drivers/base/platform.c 	kfree(pa->pdev.resource);
pa                367 drivers/base/platform.c 	kfree(pa->pdev.driver_override);
pa                368 drivers/base/platform.c 	kfree(pa);
pa                381 drivers/base/platform.c 	struct platform_object *pa;
pa                383 drivers/base/platform.c 	pa = kzalloc(sizeof(*pa) + strlen(name) + 1, GFP_KERNEL);
pa                384 drivers/base/platform.c 	if (pa) {
pa                385 drivers/base/platform.c 		strcpy(pa->name, name);
pa                386 drivers/base/platform.c 		pa->pdev.name = pa->name;
pa                387 drivers/base/platform.c 		pa->pdev.id = id;
pa                388 drivers/base/platform.c 		device_initialize(&pa->pdev.dev);
pa                389 drivers/base/platform.c 		pa->pdev.dev.release = platform_device_release;
pa                390 drivers/base/platform.c 		setup_pdev_dma_masks(&pa->pdev);
pa                393 drivers/base/platform.c 	return pa ? &pa->pdev : NULL;
pa                 18 drivers/char/agp/alpha-agp.c 	unsigned long pa;
pa                 22 drivers/char/agp/alpha-agp.c 	pa = agp->ops->translate(agp, dma_addr);
pa                 24 drivers/char/agp/alpha-agp.c 	if (pa == (unsigned long)-EINVAL)
pa                 30 drivers/char/agp/alpha-agp.c 	page = virt_to_page(__va(pa));
pa                171 drivers/clk/ti/adpll.c 	unsigned long pa;
pa                204 drivers/clk/ti/adpll.c 		sprintf(buf, "%08lx.%s.%s", d->pa, base_name, postfix);
pa                230 drivers/clk/ti/adpll.c 		snprintf(con_id, 16, "pll%03lx%s", d->pa & 0xfff, postfix + 1);
pa                903 drivers/clk/ti/adpll.c 	d->pa = res->start;
pa               1077 drivers/crypto/n2_core.c 				unsigned long pa;
pa               1079 drivers/crypto/n2_core.c 				pa = (c->arr[c->arr_len-1].src_paddr +
pa               1083 drivers/crypto/n2_core.c 				memcpy(rctx->temp_iv, __va(pa),
pa                 94 drivers/crypto/nx/nx-842-powernv.c 				     unsigned long pa, unsigned int len)
pa                 96 drivers/crypto/nx/nx-842-powernv.c 	unsigned int l = min_t(unsigned int, len, LEN_ON_PAGE(pa));
pa                102 drivers/crypto/nx/nx-842-powernv.c 	dde->address = cpu_to_be64(pa);
pa                118 drivers/crypto/nx/nx-842-powernv.c 	unsigned long pa = nx842_get_pa(buf);
pa                121 drivers/crypto/nx/nx-842-powernv.c 	if (!IS_ALIGNED(pa, DDE_BUFFER_ALIGN)) {
pa                123 drivers/crypto/nx/nx-842-powernv.c 			 in ? "input" : "output", pa, DDE_BUFFER_ALIGN);
pa                141 drivers/crypto/nx/nx-842-powernv.c 	if (len <= LEN_ON_PAGE(pa)) {
pa                142 drivers/crypto/nx/nx-842-powernv.c 		ret = setup_direct_dde(dde, pa, len);
pa                149 drivers/crypto/nx/nx-842-powernv.c 		ret = setup_direct_dde(&ddl[i], pa, len);
pa                152 drivers/crypto/nx/nx-842-powernv.c 		pa = nx842_get_pa(buf);
pa                101 drivers/crypto/nx/nx-842.h #define LEN_ON_SIZE(pa, size)	((size) - ((pa) & ((size) - 1)))
pa                102 drivers/crypto/nx/nx-842.h #define LEN_ON_PAGE(pa)		LEN_ON_SIZE(pa, PAGE_SIZE)
pa                123 drivers/dio/dio.c 		unsigned long pa;
pa                128 drivers/dio/dio.c                 pa = dio_scodetophysaddr(scode);
pa                130 drivers/dio/dio.c 		if (!pa)
pa                134 drivers/dio/dio.c 			va = (void *)(pa + DIO_VIRADDRBASE);
pa                136 drivers/dio/dio.c 			va = ioremap(pa, PAGE_SIZE);
pa                196 drivers/dio/dio.c 		unsigned long pa;
pa                201 drivers/dio/dio.c 		pa = dio_scodetophysaddr(scode);
pa                203 drivers/dio/dio.c 		if (!pa)
pa                207 drivers/dio/dio.c 			va = (void *)(pa + DIO_VIRADDRBASE);
pa                209 drivers/dio/dio.c 			va = ioremap(pa, PAGE_SIZE);
pa                226 drivers/dio/dio.c 		dev->resource.start = pa;
pa                227 drivers/dio/dio.c 		dev->resource.end = pa + DIO_SIZE(scode, va);
pa                440 drivers/edac/cpc925_edac.c 	unsigned long pa;
pa                460 drivers/edac/cpc925_edac.c 	pa = mci->csrows[rank]->first_page << PAGE_SHIFT;
pa                467 drivers/edac/cpc925_edac.c 		pa |= c << (14 - i);
pa                471 drivers/edac/cpc925_edac.c 	pa |= bank << 19;
pa                477 drivers/edac/cpc925_edac.c 		pa |= c << (26 - i);
pa                483 drivers/edac/cpc925_edac.c 		pa |= c << (21 + i);
pa                489 drivers/edac/cpc925_edac.c 		pa |= c << (18 - i);
pa                495 drivers/edac/cpc925_edac.c 		pa |= c << (29 - i);
pa                498 drivers/edac/cpc925_edac.c 	*offset = pa & (PAGE_SIZE - 1);
pa                499 drivers/edac/cpc925_edac.c 	*pfn = pa >> PAGE_SHIFT;
pa                501 drivers/edac/cpc925_edac.c 	edac_dbg(0, "ECC physical address 0x%lx\n", pa);
pa                901 drivers/firmware/stratix10-svc.c 	phys_addr_t pa;
pa                914 drivers/firmware/stratix10-svc.c 	pa = gen_pool_virt_to_phys(genpool, va);
pa                917 drivers/firmware/stratix10-svc.c 	pmem->paddr = pa;
pa                911 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c 	struct kfd_process_device_apertures *pa;
pa                941 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c 	pa = kzalloc((sizeof(struct kfd_process_device_apertures) *
pa                943 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c 	if (!pa)
pa                950 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c 		kfree(pa);
pa                957 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c 		pa[nodes].gpu_id = pdd->dev->id;
pa                958 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c 		pa[nodes].lds_base = pdd->lds_base;
pa                959 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c 		pa[nodes].lds_limit = pdd->lds_limit;
pa                960 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c 		pa[nodes].gpuvm_base = pdd->gpuvm_base;
pa                961 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c 		pa[nodes].gpuvm_limit = pdd->gpuvm_limit;
pa                962 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c 		pa[nodes].scratch_base = pdd->scratch_base;
pa                963 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c 		pa[nodes].scratch_limit = pdd->scratch_limit;
pa                988 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c 			pa,
pa                990 drivers/gpu/drm/amd/amdkfd/kfd_chardev.c 	kfree(pa);
pa                 83 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 		u32 pa = sg_dma_address(sg) - sg->offset;
pa                 86 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 		VERB("map[%d]: %08x %08x(%zx)", i, iova, pa, bytes);
pa                 88 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 		ret = etnaviv_context_map(context, da, pa, bytes, prot);
pa                777 drivers/gpu/drm/i915/gvt/gtt.c 		u64 pa, void *p_data, int bytes);
pa               1711 drivers/gpu/drm/i915/gvt/gtt.c 		u64 pa, void *p_data, int bytes)
pa               1720 drivers/gpu/drm/i915/gvt/gtt.c 	index = (pa & (PAGE_SIZE - 1)) >> info->gtt_entry_size_shift;
pa                 60 drivers/gpu/drm/i915/gvt/mmio.c static void failsafe_emulate_mmio_rw(struct intel_vgpu *vgpu, u64 pa,
pa                 72 drivers/gpu/drm/i915/gvt/mmio.c 	offset = intel_vgpu_gpa_to_mmio_offset(vgpu, pa);
pa                102 drivers/gpu/drm/i915/gvt/mmio.c int intel_vgpu_emulate_mmio_read(struct intel_vgpu *vgpu, u64 pa,
pa                110 drivers/gpu/drm/i915/gvt/mmio.c 		failsafe_emulate_mmio_rw(vgpu, pa, p_data, bytes, true);
pa                115 drivers/gpu/drm/i915/gvt/mmio.c 	offset = intel_vgpu_gpa_to_mmio_offset(vgpu, pa);
pa                136 drivers/gpu/drm/i915/gvt/mmio.c 		ret = intel_gvt_hypervisor_read_gpa(vgpu, pa, p_data, bytes);
pa                174 drivers/gpu/drm/i915/gvt/mmio.c int intel_vgpu_emulate_mmio_write(struct intel_vgpu *vgpu, u64 pa,
pa                182 drivers/gpu/drm/i915/gvt/mmio.c 		failsafe_emulate_mmio_rw(vgpu, pa, p_data, bytes, false);
pa                188 drivers/gpu/drm/i915/gvt/mmio.c 	offset = intel_vgpu_gpa_to_mmio_offset(vgpu, pa);
pa                209 drivers/gpu/drm/i915/gvt/mmio.c 		ret = intel_gvt_hypervisor_write_gpa(vgpu, pa, p_data, bytes);
pa                 87 drivers/gpu/drm/i915/gvt/mmio.h int intel_vgpu_emulate_mmio_read(struct intel_vgpu *vgpu, u64 pa,
pa                 89 drivers/gpu/drm/i915/gvt/mmio.h int intel_vgpu_emulate_mmio_write(struct intel_vgpu *vgpu, u64 pa,
pa                573 drivers/gpu/drm/msm/disp/mdp5/mdp5_crtc.c 	struct plane_state *pa = (struct plane_state *)a;
pa                575 drivers/gpu/drm/msm/disp/mdp5/mdp5_crtc.c 	return pa->state->zpos - pb->state->zpos;
pa                332 drivers/gpu/drm/omapdrm/dss/hdmi4_cec.c void hdmi4_cec_set_phys_addr(struct hdmi_core_data *core, u16 pa)
pa                334 drivers/gpu/drm/omapdrm/dss/hdmi4_cec.c 	cec_s_phys_addr(core->adap, pa, false);
pa                 29 drivers/gpu/drm/omapdrm/dss/hdmi4_cec.h void hdmi4_cec_set_phys_addr(struct hdmi_core_data *core, u16 pa);
pa                 35 drivers/gpu/drm/omapdrm/dss/hdmi4_cec.h static inline void hdmi4_cec_set_phys_addr(struct hdmi_core_data *core, u16 pa)
pa                218 drivers/gpu/drm/omapdrm/omap_dmm_tiler.c static void *alloc_dma(struct dmm_txn *txn, size_t sz, dma_addr_t *pa)
pa                228 drivers/gpu/drm/omapdrm/omap_dmm_tiler.c 	*pa = txn->current_pa;
pa                506 drivers/gpu/drm/vmwgfx/device_include/svga_reg.h       PA pa;
pa                566 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 			cb_hdr->ptr.pa += (u64) new_start_offset;
pa                893 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 		cb_hdr->ptr.pa = (u64)man->handle + (u64)offset;
pa                936 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	cb_hdr->ptr.pa = (u64)header->handle +
pa               9894 drivers/infiniband/hw/hfi1/chip.c 		  u32 type, unsigned long pa, u16 order)
pa               9902 drivers/infiniband/hw/hfi1/chip.c 		pa = 0;
pa               9910 drivers/infiniband/hw/hfi1/chip.c 	trace_hfi1_put_tid(dd, index, type, pa, order);
pa               9915 drivers/infiniband/hw/hfi1/chip.c 		| ((pa >> RT_ADDR_SHIFT) & RCV_ARRAY_RT_ADDR_MASK)
pa               1422 drivers/infiniband/hw/hfi1/chip.h 		  u32 type, unsigned long pa, u16 order);
pa                 89 drivers/infiniband/hw/hfi1/trace_tid.h 		 unsigned long va, unsigned long pa, dma_addr_t dma),
pa                 90 drivers/infiniband/hw/hfi1/trace_tid.h 	TP_ARGS(ctxt, subctxt, rarr, npages, va, pa, dma),
pa                 97 drivers/infiniband/hw/hfi1/trace_tid.h 		__field(unsigned long, pa)
pa                106 drivers/infiniband/hw/hfi1/trace_tid.h 		__entry->pa = pa;
pa                114 drivers/infiniband/hw/hfi1/trace_tid.h 		  __entry->pa,
pa                123 drivers/infiniband/hw/hfi1/trace_tid.h 		 unsigned long va, unsigned long pa, dma_addr_t dma),
pa                124 drivers/infiniband/hw/hfi1/trace_tid.h 	TP_ARGS(ctxt, subctxt, rarr, npages, va, pa, dma)
pa                130 drivers/infiniband/hw/hfi1/trace_tid.h 		 unsigned long va, unsigned long pa, dma_addr_t dma),
pa                131 drivers/infiniband/hw/hfi1/trace_tid.h 	TP_ARGS(ctxt, subctxt, rarr, npages, va, pa, dma)
pa                137 drivers/infiniband/hw/hfi1/trace_tid.h 		 u32 index, u32 type, unsigned long pa, u16 order),
pa                138 drivers/infiniband/hw/hfi1/trace_tid.h 	TP_ARGS(dd, index, type, pa, order),
pa                141 drivers/infiniband/hw/hfi1/trace_tid.h 		__field(unsigned long, pa);
pa                148 drivers/infiniband/hw/hfi1/trace_tid.h 		__entry->pa = pa;
pa                156 drivers/infiniband/hw/hfi1/trace_tid.h 		  __entry->pa,
pa               3711 drivers/infiniband/hw/i40iw/i40iw_cm.c 					 iwqp->ietf_mem.pa,
pa                998 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	set_64bit_val(wqe, 32, commit_fpm_mem->pa);
pa               1060 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	set_64bit_val(wqe, 32, query_fpm_mem->pa);
pa               3484 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 		query_fpm_mem.pa = dev->fpm_query_buf_pa;
pa               3503 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 			vf_dev->fpm_query_buf_pa = dev->vf_fpm_query_buf[iw_vf_idx].pa;
pa               3505 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 		query_fpm_mem.pa = vf_dev->fpm_query_buf_pa;
pa               3603 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	commit_fpm_mem.pa = dev->fpm_commit_buf_pa;
pa               3661 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 		data = (u64)sdbuf->pa + offset;
pa               4249 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 		values_mem.pa = pcmdinfo->in.u.query_fpm_values.fpm_values_pa;
pa               4258 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 		values_mem.pa = pcmdinfo->in.u.commit_fpm_values.fpm_values_pa;
pa                112 drivers/infiniband/hw/i40iw/i40iw_hmc.c static inline void i40iw_set_sd_entry(u64 pa,
pa                117 drivers/infiniband/hw/i40iw/i40iw_hmc.c 	entry->data = pa | (I40IW_HMC_MAX_BP_COUNT << I40E_PFHMC_SDDATALOW_PMSDBPCOUNT_SHIFT) |
pa                151 drivers/infiniband/hw/i40iw/i40iw_hmc.c 					u64 pa, u32 sd_idx,
pa                160 drivers/infiniband/hw/i40iw/i40iw_hmc.c 		i40iw_set_sd_entry(pa, sd_idx, type, sdinfo.entry);
pa                183 drivers/infiniband/hw/i40iw/i40iw_hmc.c 	u64 pa;
pa                196 drivers/infiniband/hw/i40iw/i40iw_hmc.c 			pa = (sd_entry->entry_type == I40IW_SD_TYPE_PAGED) ?
pa                197 drivers/infiniband/hw/i40iw/i40iw_hmc.c 			    sd_entry->u.pd_table.pd_page_addr.pa :
pa                198 drivers/infiniband/hw/i40iw/i40iw_hmc.c 			    sd_entry->u.bp.addr.pa;
pa                199 drivers/infiniband/hw/i40iw/i40iw_hmc.c 			i40iw_set_sd_entry(pa, i, sd_entry->entry_type,
pa                655 drivers/infiniband/hw/i40iw/i40iw_hmc.c 		page_desc = page->pa | 0x1;
pa                220 drivers/infiniband/hw/i40iw/i40iw_hmc.h 					u64 pa, u32 sd_idx, enum i40iw_sd_entry_type type,
pa                540 drivers/infiniband/hw/i40iw/i40iw_main.c 	memptr->pa = iwdev->obj_next.pa + extra;
pa                546 drivers/infiniband/hw/i40iw/i40iw_main.c 	iwdev->obj_next.pa = memptr->pa + size;
pa                588 drivers/infiniband/hw/i40iw/i40iw_main.c 	dev->cqp->host_ctx_pa = mem.pa;
pa                594 drivers/infiniband/hw/i40iw/i40iw_main.c 	cqp_init_info.sq_pa = cqp->sq.pa;
pa                595 drivers/infiniband/hw/i40iw/i40iw_main.c 	cqp_init_info.host_ctx_pa = mem.pa;
pa                658 drivers/infiniband/hw/i40iw/i40iw_main.c 	info.cq_pa = ccq->mem_cq.pa;
pa                661 drivers/infiniband/hw/i40iw/i40iw_main.c 	info.shadow_area_pa = mem.pa;
pa                741 drivers/infiniband/hw/i40iw/i40iw_main.c 	info.ceqe_pa = iwceq->mem.pa;
pa                881 drivers/infiniband/hw/i40iw/i40iw_main.c 	info.aeq_elem_pa = aeq->mem.pa;
pa               1341 drivers/infiniband/hw/i40iw/i40iw_main.c 	info.fpm_query_buf_pa = mem.pa;
pa               1347 drivers/infiniband/hw/i40iw/i40iw_main.c 	info.fpm_commit_buf_pa = mem.pa;
pa                 88 drivers/infiniband/hw/i40iw/i40iw_osdep.h 	dma_addr_t pa;
pa                286 drivers/infiniband/hw/i40iw/i40iw_pble.c 		mem.pa = chunk->dmaaddrs[i];
pa                303 drivers/infiniband/hw/i40iw/i40iw_pble.c 		vf_pble_info.pd_pl_pba = sd_entry->u.pd_table.pd_page_addr.pa;
pa                397 drivers/infiniband/hw/i40iw/i40iw_pble.c 			sd_entry->u.pd_table.pd_page_addr.pa : sd_entry->u.bp.addr.pa;
pa                126 drivers/infiniband/hw/i40iw/i40iw_puda.c 	set_64bit_val(wqe, 0, buf->mem.pa);
pa                452 drivers/infiniband/hw/i40iw/i40iw_puda.c 	info.paddr = buf->mem.pa;
pa                581 drivers/infiniband/hw/i40iw/i40iw_puda.c 	qp->sq_pa = mem->pa;
pa                697 drivers/infiniband/hw/i40iw/i40iw_puda.c 	info.cq_base_pa = mem->pa;
pa                698 drivers/infiniband/hw/i40iw/i40iw_puda.c 	info.shadow_area_pa = mem->pa + cqsize;
pa                762 drivers/infiniband/hw/i40iw/i40iw_utils.c 				     (dma_addr_t *)&mem->pa, GFP_KERNEL);
pa                781 drivers/infiniband/hw/i40iw/i40iw_utils.c 			  mem->va, (dma_addr_t)mem->pa);
pa               1111 drivers/infiniband/hw/i40iw/i40iw_utils.c 	cqp_info->in.u.query_fpm_values.fpm_values_pa = values_mem->pa;
pa               1144 drivers/infiniband/hw/i40iw/i40iw_utils.c 	cqp_info->in.u.commit_fpm_values.fpm_values_pa = values_mem->pa;
pa                497 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	info->sq_pa = mem->pa;
pa                594 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	init_info.q2_pa = iwqp->q2_ctx_mem.pa;
pa                614 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	iwqp->host_ctx.pa = init_info.host_ctx_pa;
pa               1172 drivers/infiniband/hw/i40iw/i40iw_verbs.c 		info.cq_base_pa = iwcq->kmem.pa;
pa                146 drivers/infiniband/hw/mlx5/mem.c 			dma_addr_t pa =
pa                149 drivers/infiniband/hw/mlx5/mem.c 			pas[i] = cpu_to_be64(umem_dma_to_mtt(pa));
pa                122 drivers/infiniband/hw/ocrdma/ocrdma.h 	dma_addr_t pa;
pa                128 drivers/infiniband/hw/ocrdma/ocrdma.h 	dma_addr_t pa;
pa                224 drivers/infiniband/hw/ocrdma/ocrdma.h 	dma_addr_t pa;
pa                267 drivers/infiniband/hw/ocrdma/ocrdma.h 		dma_addr_t pa;
pa                336 drivers/infiniband/hw/ocrdma/ocrdma.h 	dma_addr_t pa;
pa                372 drivers/infiniband/hw/ocrdma/ocrdma.h 	dma_addr_t pa;
pa                444 drivers/infiniband/hw/ocrdma/ocrdma.h 		dma_addr_t pa;
pa                448 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	ocrdma_build_q_pages(&cmd->pa[0], cmd->num_pages, eq->q.dma,
pa                544 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	ocrdma_build_q_pages(&cmd->pa[0], cq->size / OCRDMA_MIN_Q_PAGE_SIZE,
pa                571 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	struct ocrdma_pa *pa;
pa                592 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	pa = &cmd->pa[0];
pa                594 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	ocrdma_build_q_pages(pa, num_pages, mq->dma, PAGE_SIZE_4K);
pa               1302 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dev->stats_mem.pa & 0xffffffff);
pa               1303 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dev->stats_mem.pa);
pa               1342 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 					dma.size, &dma.pa, GFP_KERNEL);
pa               1350 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dma.pa & 0xffffffff);
pa               1351 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dma.pa);
pa               1371 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	dma_free_coherent(&dev->nic_info.pdev->dev, dma.size, dma.va, dma.pa);
pa               1652 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	dma_addr_t pa;
pa               1681 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 						&dev->av_tbl.pbl.pa,
pa               1687 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 					    &pa, GFP_KERNEL);
pa               1690 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	dev->av_tbl.pa = pa;
pa               1695 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		pbes[i].pa_lo = (u32)cpu_to_le32(pa & 0xffffffff);
pa               1696 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		pbes[i].pa_hi = (u32)cpu_to_le32(upper_32_bits(pa));
pa               1697 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		pa += PAGE_SIZE;
pa               1699 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	cmd->tbl_addr[0].lo = (u32)(dev->av_tbl.pbl.pa & 0xFFFFFFFF);
pa               1700 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	cmd->tbl_addr[0].hi = (u32)upper_32_bits(dev->av_tbl.pbl.pa);
pa               1711 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 			  dev->av_tbl.pa);
pa               1715 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 			  dev->av_tbl.pbl.pa);
pa               1738 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 			  dev->av_tbl.pa);
pa               1741 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 			  dev->av_tbl.pbl.pa);
pa               1820 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	cq->va = dma_alloc_coherent(&pdev->dev, cq->len, &cq->pa, GFP_KERNEL);
pa               1872 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	ocrdma_build_q_pages(&cmd->cmd.pa[0], hw_pages, cq->pa, page_size);
pa               1883 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	dma_free_coherent(&pdev->dev, cq->len, cq->va, cq->pa);
pa               1905 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	dma_free_coherent(&dev->nic_info.pdev->dev, cq->len, cq->va, cq->pa);
pa               1997 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		cmd->pbl[i].lo = (u32) (hwmr->pbl_table[i].pa & 0xffffffff);
pa               1998 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		cmd->pbl[i].hi = upper_32_bits(hwmr->pbl_table[i].pa);
pa               2028 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		    (u32) (hwmr->pbl_table[i + pbl_offset].pa & 0xffffffff);
pa               2030 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		    upper_32_bits(hwmr->pbl_table[i + pbl_offset].pa);
pa               2187 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	dma_addr_t pa;
pa               2208 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	qp->sq.va = dma_alloc_coherent(&pdev->dev, len, &pa, GFP_KERNEL);
pa               2212 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	qp->sq.pa = pa;
pa               2214 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	ocrdma_build_q_pages(&cmd->wq_addr[0], hw_pages, pa, hw_page_size);
pa               2242 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	dma_addr_t pa = 0;
pa               2258 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	qp->rq.va = dma_alloc_coherent(&pdev->dev, len, &pa, GFP_KERNEL);
pa               2261 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	qp->rq.pa = pa;
pa               2265 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	ocrdma_build_q_pages(&cmd->rq_addr[0], hw_pages, pa, hw_page_size);
pa               2305 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	dma_addr_t pa = 0;
pa               2314 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	qp->ird_q_va = dma_alloc_coherent(&pdev->dev, ird_q_len, &pa,
pa               2319 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 			     pa, ird_page_size);
pa               2456 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		dma_free_coherent(&pdev->dev, qp->rq.len, qp->rq.va, qp->rq.pa);
pa               2459 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	dma_free_coherent(&pdev->dev, qp->sq.len, qp->sq.va, qp->sq.pa);
pa               2735 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		dma_free_coherent(&pdev->dev, qp->sq.len, qp->sq.va, qp->sq.pa);
pa               2737 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		dma_free_coherent(&pdev->dev, qp->rq.len, qp->rq.va, qp->rq.pa);
pa               2752 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	dma_addr_t pa;
pa               2772 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	srq->rq.va = dma_alloc_coherent(&pdev->dev, len, &pa, GFP_KERNEL);
pa               2777 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	ocrdma_build_q_pages(&cmd->rq_addr[0], hw_pages, pa, hw_page_size);
pa               2780 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	srq->rq.pa = pa;
pa               2812 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	dma_free_coherent(&pdev->dev, srq->rq.len, srq->rq.va, pa);
pa               2873 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 				  srq->rq.va, srq->rq.pa);
pa               2881 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	dma_addr_t pa;
pa               2892 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	req = dma_alloc_coherent(&pdev->dev, cmd.hdr.pyld_len, &pa, GFP_KERNEL);
pa               2900 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	mqe_sge->pa_lo = (u32) (pa & 0xFFFFFFFFUL);
pa               2901 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	mqe_sge->pa_hi = (u32) upper_32_bits(pa);
pa               2917 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	dma_free_coherent(&pdev->dev, cmd.hdr.pyld_len, req, pa);
pa                325 drivers/infiniband/hw/ocrdma/ocrdma_sli.h 	struct ocrdma_pa pa[MAX_OCRDMA_EQ_PAGES];
pa                797 drivers/infiniband/hw/ocrdma/ocrdma_sli.h 	struct ocrdma_pa pa[OCRDMA_CREATE_CQ_MAX_PAGES];
pa                839 drivers/infiniband/hw/ocrdma/ocrdma_sli.h 	struct ocrdma_pa pa[8];
pa                 77 drivers/infiniband/hw/ocrdma/ocrdma_stats.c 				     &mem->pa, GFP_KERNEL);
pa                 97 drivers/infiniband/hw/ocrdma/ocrdma_stats.c 				  mem->va, mem->pa);
pa                487 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 					    &ctx->ah_tbl.pa, GFP_KERNEL);
pa                519 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	ocrdma_del_mmap(ctx, ctx->ah_tbl.pa, ctx->ah_tbl.len);
pa                522 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			  ctx->ah_tbl.pa);
pa                535 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	ocrdma_del_mmap(uctx, uctx->ah_tbl.pa, uctx->ah_tbl.len);
pa                537 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			  uctx->ah_tbl.pa);
pa                761 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 					  mr->pbl_table[i].pa);
pa                800 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	dma_addr_t pa;
pa                809 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		va = dma_alloc_coherent(&pdev->dev, dma_len, &pa, GFP_KERNEL);
pa                816 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		mr->pbl_table[i].pa = pa;
pa               1086 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		ocrdma_del_mmap(cq->ucontext, (u64) cq->pa,
pa               1732 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		ocrdma_del_mmap(pd->uctx, (u64) qp->sq.pa,
pa               1735 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			ocrdma_del_mmap(pd->uctx, (u64) qp->rq.pa,
pa               1878 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		ocrdma_del_mmap(srq->pd->uctx, (u64) srq->rq.pa,
pa                226 drivers/infiniband/hw/qedr/qedr.h 	dma_addr_t pa;
pa                496 drivers/infiniband/hw/qedr/verbs.c 				  pbl[i].va, pbl[i].pa);
pa                516 drivers/infiniband/hw/qedr/verbs.c 	dma_addr_t pa;
pa                525 drivers/infiniband/hw/qedr/verbs.c 		va = dma_alloc_coherent(&pdev->dev, pbl_info->pbl_size, &pa,
pa                531 drivers/infiniband/hw/qedr/verbs.c 		pbl_table[i].pa = pa;
pa                539 drivers/infiniband/hw/qedr/verbs.c 		pbl_main_tbl[i] = pbl_table[i + 1].pa;
pa                546 drivers/infiniband/hw/qedr/verbs.c 				  pbl_table[i].va, pbl_table[i].pa);
pa                865 drivers/infiniband/hw/qedr/verbs.c 		pbl_ptr = cq->q.pbl_tbl->pa;
pa               1399 drivers/infiniband/hw/qedr/verbs.c 		pbl_base_addr = srq->usrq.pbl_tbl->pa;
pa               1564 drivers/infiniband/hw/qedr/verbs.c 	qp->usq.pbl_tbl->pa = out_params->sq_pbl_phys;
pa               1570 drivers/infiniband/hw/qedr/verbs.c 		qp->urq.pbl_tbl->pa = out_params->rq_pbl_phys;
pa               1634 drivers/infiniband/hw/qedr/verbs.c 	in_params.sq_pbl_ptr = qp->usq.pbl_tbl->pa;
pa               1637 drivers/infiniband/hw/qedr/verbs.c 		in_params.rq_pbl_ptr = qp->urq.pbl_tbl->pa;
pa               2580 drivers/infiniband/hw/qedr/verbs.c 		 &info->pbl_table->pa);
pa               2593 drivers/infiniband/hw/qedr/verbs.c 	DP_DEBUG(dev, QEDR_MSG_MR, "extra pbl_table_pa = %pa\n", &tmp->pa);
pa               2653 drivers/infiniband/hw/qedr/verbs.c 	mr->hw_mr.pbl_ptr = mr->info.pbl_table[0].pa;
pa               3105 drivers/infiniband/hw/qedr/verbs.c 	fwqe2->pbl_addr.hi = upper_32_bits(mr->info.pbl_table->pa);
pa               3106 drivers/infiniband/hw/qedr/verbs.c 	fwqe2->pbl_addr.lo = lower_32_bits(mr->info.pbl_table->pa);
pa               1836 drivers/infiniband/hw/qib/qib_iba6120.c 			     u32 type, unsigned long pa)
pa               1846 drivers/infiniband/hw/qib/qib_iba6120.c 	if (pa != dd->tidinvalid) {
pa               1847 drivers/infiniband/hw/qib/qib_iba6120.c 		if (pa & ((1U << 11) - 1)) {
pa               1849 drivers/infiniband/hw/qib/qib_iba6120.c 				    pa);
pa               1852 drivers/infiniband/hw/qib/qib_iba6120.c 		pa >>= 11;
pa               1853 drivers/infiniband/hw/qib/qib_iba6120.c 		if (pa & ~QLOGIC_IB_RT_ADDR_MASK) {
pa               1856 drivers/infiniband/hw/qib/qib_iba6120.c 				pa);
pa               1861 drivers/infiniband/hw/qib/qib_iba6120.c 			pa |= dd->tidtemplate;
pa               1863 drivers/infiniband/hw/qib/qib_iba6120.c 			pa |= 2 << 29;
pa               1885 drivers/infiniband/hw/qib/qib_iba6120.c 	writel(pa, tidp32);
pa               1903 drivers/infiniband/hw/qib/qib_iba6120.c 			       u32 type, unsigned long pa)
pa               1910 drivers/infiniband/hw/qib/qib_iba6120.c 	if (pa != dd->tidinvalid) {
pa               1911 drivers/infiniband/hw/qib/qib_iba6120.c 		if (pa & ((1U << 11) - 1)) {
pa               1913 drivers/infiniband/hw/qib/qib_iba6120.c 				    pa);
pa               1916 drivers/infiniband/hw/qib/qib_iba6120.c 		pa >>= 11;
pa               1917 drivers/infiniband/hw/qib/qib_iba6120.c 		if (pa & ~QLOGIC_IB_RT_ADDR_MASK) {
pa               1920 drivers/infiniband/hw/qib/qib_iba6120.c 				pa);
pa               1925 drivers/infiniband/hw/qib/qib_iba6120.c 			pa |= dd->tidtemplate;
pa               1927 drivers/infiniband/hw/qib/qib_iba6120.c 			pa |= 2 << 29;
pa               1929 drivers/infiniband/hw/qib/qib_iba6120.c 	writel(pa, tidp32);
pa               2153 drivers/infiniband/hw/qib/qib_iba7220.c 			     u32 type, unsigned long pa)
pa               2155 drivers/infiniband/hw/qib/qib_iba7220.c 	if (pa != dd->tidinvalid) {
pa               2156 drivers/infiniband/hw/qib/qib_iba7220.c 		u64 chippa = pa >> IBA7220_TID_PA_SHIFT;
pa               2159 drivers/infiniband/hw/qib/qib_iba7220.c 		if (pa != (chippa << IBA7220_TID_PA_SHIFT)) {
pa               2161 drivers/infiniband/hw/qib/qib_iba7220.c 				    pa);
pa               2167 drivers/infiniband/hw/qib/qib_iba7220.c 				pa);
pa               2175 drivers/infiniband/hw/qib/qib_iba7220.c 		pa = chippa;
pa               2177 drivers/infiniband/hw/qib/qib_iba7220.c 	writeq(pa, tidptr);
pa               3769 drivers/infiniband/hw/qib/qib_iba7322.c 			     u32 type, unsigned long pa)
pa               3773 drivers/infiniband/hw/qib/qib_iba7322.c 	if (pa != dd->tidinvalid) {
pa               3774 drivers/infiniband/hw/qib/qib_iba7322.c 		u64 chippa = pa >> IBA7322_TID_PA_SHIFT;
pa               3777 drivers/infiniband/hw/qib/qib_iba7322.c 		if (pa != (chippa << IBA7322_TID_PA_SHIFT)) {
pa               3779 drivers/infiniband/hw/qib/qib_iba7322.c 				    pa);
pa               3785 drivers/infiniband/hw/qib/qib_iba7322.c 				pa);
pa               3793 drivers/infiniband/hw/qib/qib_iba7322.c 		pa = chippa;
pa               3795 drivers/infiniband/hw/qib/qib_iba7322.c 	writeq(pa, tidptr);
pa               1676 drivers/infiniband/hw/qib/qib_init.c 		dma_addr_t pa = rcd->rcvegrbuf_phys[chunk];
pa               1688 drivers/infiniband/hw/qib/qib_init.c 					  RCVHQ_RCV_TYPE_EAGER, pa);
pa               1689 drivers/infiniband/hw/qib/qib_init.c 			pa += egrsize;
pa                 72 drivers/infiniband/hw/usnic/usnic_uiom.c 	dma_addr_t pa;
pa                 77 drivers/infiniband/hw/usnic/usnic_uiom.c 			pa = sg_phys(sg);
pa                 79 drivers/infiniband/hw/usnic/usnic_uiom.c 			usnic_dbg("pa: %pa\n", &pa);
pa                100 drivers/infiniband/hw/usnic/usnic_uiom.c 	dma_addr_t pa;
pa                170 drivers/infiniband/hw/usnic/usnic_uiom.c 				pa = sg_phys(sg);
pa                172 drivers/infiniband/hw/usnic/usnic_uiom.c 						cur_base + i*PAGE_SIZE, &pa);
pa                251 drivers/infiniband/hw/usnic/usnic_uiom.c 	dma_addr_t pa;
pa                265 drivers/infiniband/hw/usnic/usnic_uiom.c 			pa = sg_phys(&chunk->page_list[i]);
pa                272 drivers/infiniband/hw/usnic/usnic_uiom.c 				pa_start = pa;
pa                273 drivers/infiniband/hw/usnic/usnic_uiom.c 				pa_end = pa;
pa                278 drivers/infiniband/hw/usnic/usnic_uiom.c 			if ((pa_end + PAGE_SIZE != pa) &&
pa                279 drivers/infiniband/hw/usnic/usnic_uiom.c 					(pa != pa_start)) {
pa                292 drivers/infiniband/hw/usnic/usnic_uiom.c 				pa_start = pa;
pa                293 drivers/infiniband/hw/usnic/usnic_uiom.c 				pa_end = pa;
pa                298 drivers/infiniband/hw/usnic/usnic_uiom.c 				size = pa - pa_start + PAGE_SIZE;
pa                311 drivers/infiniband/hw/usnic/usnic_uiom.c 			if (pa != pa_start)
pa                124 drivers/iommu/exynos-iommu.c #define mk_lv1ent_sect(pa, prot) ((pa >> PG_ENT_SHIFT) | LV1_PROT[prot] | 2)
pa                125 drivers/iommu/exynos-iommu.c #define mk_lv1ent_page(pa) ((pa >> PG_ENT_SHIFT) | 1)
pa                126 drivers/iommu/exynos-iommu.c #define mk_lv2ent_lpage(pa, prot) ((pa >> PG_ENT_SHIFT) | LV2_PROT[prot] | 1)
pa                127 drivers/iommu/exynos-iommu.c #define mk_lv2ent_spage(pa, prot) ((pa >> PG_ENT_SHIFT) | LV2_PROT[prot] | 2)
pa                507 drivers/iommu/msm_iommu.c 			 phys_addr_t pa, size_t len, int prot)
pa                514 drivers/iommu/msm_iommu.c 	ret = priv->iop->map(priv->iop, iova, pa, len, prot);
pa                485 drivers/iommu/mtk_iommu.c 	phys_addr_t pa;
pa                488 drivers/iommu/mtk_iommu.c 	pa = dom->iop->iova_to_phys(dom->iop, iova);
pa                491 drivers/iommu/mtk_iommu.c 	if (data->enable_4GB && pa >= MTK_IOMMU_4GB_MODE_REMAP_BASE)
pa                492 drivers/iommu/mtk_iommu.c 		pa &= ~BIT_ULL(32);
pa                494 drivers/iommu/mtk_iommu.c 	return pa;
pa                349 drivers/iommu/mtk_iommu_v1.c 	phys_addr_t pa;
pa                352 drivers/iommu/mtk_iommu_v1.c 	pa = *(dom->pgt_va + (iova >> MT2701_IOMMU_PAGE_SHIFT));
pa                353 drivers/iommu/mtk_iommu_v1.c 	pa = pa & (~(MT2701_IOMMU_PAGE_SIZE - 1));
pa                356 drivers/iommu/mtk_iommu_v1.c 	return pa;
pa                168 drivers/iommu/omap-iommu.c 	u32 l, pa;
pa                173 drivers/iommu/omap-iommu.c 	pa = virt_to_phys(obj->iopgd);
pa                174 drivers/iommu/omap-iommu.c 	if (!IS_ALIGNED(pa, SZ_16K))
pa                181 drivers/iommu/omap-iommu.c 	iommu_write_reg(obj, pa, MMU_TTB);
pa                331 drivers/iommu/omap-iommu.c 	cr->ram = e->pa | e->endian | e->elsz | e->mixed;
pa                556 drivers/iommu/omap-iommu.c static int iopgd_alloc_section(struct omap_iommu *obj, u32 da, u32 pa, u32 prot)
pa                561 drivers/iommu/omap-iommu.c 	if ((da | pa) & ~IOSECTION_MASK) {
pa                563 drivers/iommu/omap-iommu.c 			__func__, da, pa, IOSECTION_SIZE);
pa                567 drivers/iommu/omap-iommu.c 	*iopgd = (pa & IOSECTION_MASK) | prot | IOPGD_SECTION;
pa                572 drivers/iommu/omap-iommu.c static int iopgd_alloc_super(struct omap_iommu *obj, u32 da, u32 pa, u32 prot)
pa                578 drivers/iommu/omap-iommu.c 	if ((da | pa) & ~IOSUPER_MASK) {
pa                580 drivers/iommu/omap-iommu.c 			__func__, da, pa, IOSUPER_SIZE);
pa                585 drivers/iommu/omap-iommu.c 		*(iopgd + i) = (pa & IOSUPER_MASK) | prot | IOPGD_SUPER;
pa                590 drivers/iommu/omap-iommu.c static int iopte_alloc_page(struct omap_iommu *obj, u32 da, u32 pa, u32 prot)
pa                600 drivers/iommu/omap-iommu.c 	*iopte = (pa & IOPAGE_MASK) | prot | IOPTE_SMALL;
pa                604 drivers/iommu/omap-iommu.c 		 __func__, da, pa, iopte, *iopte);
pa                609 drivers/iommu/omap-iommu.c static int iopte_alloc_large(struct omap_iommu *obj, u32 da, u32 pa, u32 prot)
pa                617 drivers/iommu/omap-iommu.c 	if ((da | pa) & ~IOLARGE_MASK) {
pa                619 drivers/iommu/omap-iommu.c 			__func__, da, pa, IOLARGE_SIZE);
pa                627 drivers/iommu/omap-iommu.c 		*(iopte + i) = (pa & IOLARGE_MASK) | prot | IOPTE_LARGE;
pa                666 drivers/iommu/omap-iommu.c 	err = fn(obj, e->da, e->pa, prot);
pa               1326 drivers/iommu/omap-iommu.c static u32 iotlb_init_entry(struct iotlb_entry *e, u32 da, u32 pa, int pgsz)
pa               1331 drivers/iommu/omap-iommu.c 	e->pa		= pa;
pa               1342 drivers/iommu/omap-iommu.c 			  phys_addr_t pa, size_t bytes, int prot)
pa               1359 drivers/iommu/omap-iommu.c 	dev_dbg(dev, "mapping da 0x%lx to pa %pa size 0x%x\n", da, &pa, bytes);
pa               1361 drivers/iommu/omap-iommu.c 	iotlb_init_entry(&e, da, pa, omap_pgsz);
pa                 23 drivers/iommu/omap-iommu.h 	u32 pa;
pa                210 drivers/iommu/s390-iommu.c 				   unsigned long pa, dma_addr_t dma_addr,
pa                214 drivers/iommu/s390-iommu.c 	u8 *page_addr = (u8 *) (pa & PAGE_MASK);
pa                168 drivers/iommu/tegra-gart.c 				   unsigned long pa)
pa                175 drivers/iommu/tegra-gart.c 	gart_set_pte(gart, iova, GART_ENTRY_PHYS_ADDR_VALID | pa);
pa                181 drivers/iommu/tegra-gart.c 			  phys_addr_t pa, size_t bytes, int prot)
pa                190 drivers/iommu/tegra-gart.c 	ret = __gart_iommu_map(gart, iova, (unsigned long)pa);
pa                101 drivers/mailbox/bcm-flexrm-mailbox.c #define BD_START_ADDR_VALUE(pa)				\
pa                102 drivers/mailbox/bcm-flexrm-mailbox.c 	((u32)((((dma_addr_t)(pa)) >> RING_BD_ALIGN_ORDER) & 0x0fffffff))
pa                107 drivers/mailbox/bcm-flexrm-mailbox.c #define CMPL_START_ADDR_VALUE(pa)			\
pa                108 drivers/mailbox/bcm-flexrm-mailbox.c 	((u32)((((u64)(pa)) >> RING_CMPL_ALIGN_ORDER) & 0x07ffffff))
pa               1456 drivers/mailbox/bcm-flexrm-mailbox.c 					const struct of_phandle_args *pa)
pa               1461 drivers/mailbox/bcm-flexrm-mailbox.c 	if (pa->args_count < 3)
pa               1464 drivers/mailbox/bcm-flexrm-mailbox.c 	if (pa->args[0] >= cntlr->num_chans)
pa               1467 drivers/mailbox/bcm-flexrm-mailbox.c 	if (pa->args[1] > MSI_COUNT_MASK)
pa               1470 drivers/mailbox/bcm-flexrm-mailbox.c 	if (pa->args[2] > MSI_TIMER_VAL_MASK)
pa               1473 drivers/mailbox/bcm-flexrm-mailbox.c 	chan = &cntlr->chans[pa->args[0]];
pa               1475 drivers/mailbox/bcm-flexrm-mailbox.c 	ring->msi_count_threshold = pa->args[1];
pa               1476 drivers/mailbox/bcm-flexrm-mailbox.c 	ring->msi_timer_val = pa->args[2];
pa               1619 drivers/media/cec/cec-adap.c 	u16 pa = CEC_PHYS_ADDR_INVALID;
pa               1622 drivers/media/cec/cec-adap.c 		pa = cec_get_edid_phys_addr((const u8 *)edid,
pa               1624 drivers/media/cec/cec-adap.c 	cec_s_phys_addr(adap, pa, false);
pa               1973 drivers/media/cec/cec-adap.c 		u16 pa = (msg->msg[2] << 8) | msg->msg[3];
pa               1976 drivers/media/cec/cec-adap.c 			adap->phys_addrs[init_laddr] = pa;
pa               1978 drivers/media/cec/cec-adap.c 			cec_phys_addr_exp(pa), init_laddr);
pa                187 drivers/media/cec/cec-core.c static void cec_cec_notify(struct cec_adapter *adap, u16 pa)
pa                189 drivers/media/cec/cec-core.c 	cec_s_phys_addr(adap, pa, false);
pa                 28 drivers/media/cec/cec-notifier.c 	void (*callback)(struct cec_adapter *adap, u16 pa);
pa                170 drivers/media/cec/cec-notifier.c void cec_notifier_set_phys_addr(struct cec_notifier *n, u16 pa)
pa                176 drivers/media/cec/cec-notifier.c 	n->phys_addr = pa;
pa                188 drivers/media/cec/cec-notifier.c 	u16 pa = CEC_PHYS_ADDR_INVALID;
pa                194 drivers/media/cec/cec-notifier.c 		pa = cec_get_edid_phys_addr((const u8 *)edid,
pa                196 drivers/media/cec/cec-notifier.c 	cec_notifier_set_phys_addr(n, pa);
pa                202 drivers/media/cec/cec-notifier.c 			   void (*callback)(struct cec_adapter *adap, u16 pa))
pa               2291 drivers/media/i2c/adv7604.c 	u16 pa;
pa               2324 drivers/media/i2c/adv7604.c 	pa = v4l2_get_edid_phys_addr(edid->edid, edid->blocks * 128, &spa_loc);
pa               2325 drivers/media/i2c/adv7604.c 	err = v4l2_phys_addr_validate(pa, &pa, NULL);
pa               2402 drivers/media/i2c/adv7604.c 	cec_s_phys_addr(state->cec_adap, pa, false);
pa                776 drivers/media/i2c/adv7842.c 	u16 pa;
pa                794 drivers/media/i2c/adv7842.c 	pa = v4l2_get_edid_phys_addr(edid, 256, &spa_loc);
pa                795 drivers/media/i2c/adv7842.c 	err = v4l2_phys_addr_validate(pa, &pa, NULL);
pa                840 drivers/media/i2c/adv7842.c 	cec_s_phys_addr(state->cec_adap, pa, false);
pa               1772 drivers/media/i2c/tc358743.c 	u16 pa;
pa               1791 drivers/media/i2c/tc358743.c 	pa = cec_get_edid_phys_addr(edid->edid, edid->blocks * 128, NULL);
pa               1792 drivers/media/i2c/tc358743.c 	err = v4l2_phys_addr_validate(pa, &pa, NULL);
pa               1813 drivers/media/i2c/tc358743.c 	cec_s_phys_addr(state->cec_adap, pa, false);
pa                 37 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	unsigned long pa;
pa                402 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	vsi->mv_buf.pa = (unsigned long)mem->dma_addr;
pa                419 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	vsi->seg_id_buf.pa = (unsigned long)mem->dma_addr;
pa                668 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	inst->vsi->mv_buf.pa = (unsigned long)inst->mv_buf.dma_addr;
pa                673 drivers/media/platform/mtk-vcodec/vdec/vdec_vp9_if.c 	inst->vsi->seg_id_buf.pa = (unsigned long)inst->seg_id_buf.dma_addr;
pa                 85 drivers/media/platform/mtk-vpu/mtk_vpu.c 	dma_addr_t pa;
pa                668 drivers/media/platform/mtk-vpu/mtk_vpu.c 			  vpu->extmem[fw_type].pa);
pa                681 drivers/media/platform/mtk-vpu/mtk_vpu.c 					       &vpu->extmem[fw_type].pa,
pa                690 drivers/media/platform/mtk-vpu/mtk_vpu.c 	vpu_cfg_writel(vpu, (vpu->extmem[fw_type].pa & 0xFFFFF000) + offset_4gb,
pa                695 drivers/media/platform/mtk-vpu/mtk_vpu.c 		 (unsigned long long)vpu->extmem[fw_type].pa,
pa                 58 drivers/media/radio/radio-keene.c 	u8 pa;
pa                 79 drivers/media/radio/radio-keene.c 	radio->buffer[4] = radio->pa;
pa                245 drivers/media/radio/radio-keene.c 		radio->pa = (ctrl->val - 71) * 100 / 62;
pa                335 drivers/media/radio/radio-keene.c 	radio->pa = 118;
pa                330 drivers/media/usb/pulse8-cec/pulse8-cec.c 			struct cec_log_addrs *log_addrs, u16 *pa)
pa                347 drivers/media/usb/pulse8-cec/pulse8-cec.c 		*pa = CEC_PHYS_ADDR_INVALID;
pa                427 drivers/media/usb/pulse8-cec/pulse8-cec.c 	*pa = (data[0] << 8) | data[1];
pa                429 drivers/media/usb/pulse8-cec/pulse8-cec.c 		cec_phys_addr_exp(*pa));
pa                450 drivers/media/usb/pulse8-cec/pulse8-cec.c 					  u16 pa)
pa                458 drivers/media/usb/pulse8-cec/pulse8-cec.c 	cec_s_phys_addr(pulse8->adap, pa, false);
pa                480 drivers/media/usb/pulse8-cec/pulse8-cec.c 	u16 pa = adap->phys_addr;
pa                554 drivers/media/usb/pulse8-cec/pulse8-cec.c 	cmd[1] = pa >> 8;
pa                555 drivers/media/usb/pulse8-cec/pulse8-cec.c 	cmd[2] = pa & 0xff;
pa                654 drivers/media/usb/pulse8-cec/pulse8-cec.c 	u16 pa = CEC_PHYS_ADDR_INVALID;
pa                679 drivers/media/usb/pulse8-cec/pulse8-cec.c 	err = pulse8_setup(pulse8, serio, &log_addrs, &pa);
pa                690 drivers/media/usb/pulse8-cec/pulse8-cec.c 		err = pulse8_apply_persistent_config(pulse8, &log_addrs, pa);
pa                240 drivers/media/usb/rainshadow-cec/rainshadow-cec.c 			struct cec_log_addrs *log_addrs, u16 *pa)
pa                312 drivers/media/usb/rainshadow-cec/rainshadow-cec.c 	u16 pa = CEC_PHYS_ADDR_INVALID;
pa                336 drivers/media/usb/rainshadow-cec/rainshadow-cec.c 	err = rain_setup(rain, serio, &log_addrs, &pa);
pa                443 drivers/message/fusion/mptbase.c mpt_turbo_reply(MPT_ADAPTER *ioc, u32 pa)
pa                451 drivers/message/fusion/mptbase.c 				ioc->name, pa));
pa                453 drivers/message/fusion/mptbase.c 	switch (pa >> MPI_CONTEXT_REPLY_TYPE_SHIFT) {
pa                455 drivers/message/fusion/mptbase.c 		req_idx = pa & 0x0000FFFF;
pa                456 drivers/message/fusion/mptbase.c 		cb_idx = (pa & 0x00FF0000) >> 16;
pa                470 drivers/message/fusion/mptbase.c 		if ((pa & 0x58000000) == 0x58000000) {
pa                471 drivers/message/fusion/mptbase.c 			req_idx = pa & 0x0000FFFF;
pa                478 drivers/message/fusion/mptbase.c 		mr = (MPT_FRAME_HDR *) CAST_U32_TO_PTR(pa);
pa                482 drivers/message/fusion/mptbase.c 		mr = (MPT_FRAME_HDR *) CAST_U32_TO_PTR(pa);
pa                504 drivers/message/fusion/mptbase.c mpt_reply(MPT_ADAPTER *ioc, u32 pa)
pa                525 drivers/message/fusion/mptbase.c 	reply_dma_low = (pa <<= 1);
pa                566 drivers/message/fusion/mptbase.c 	CHIPREG_WRITE32(&ioc->chip->ReplyFifo, pa);
pa                594 drivers/message/fusion/mptbase.c 	u32 pa = CHIPREG_READ32_dmasync(&ioc->chip->ReplyFifo);
pa                596 drivers/message/fusion/mptbase.c 	if (pa == 0xFFFFFFFF)
pa                603 drivers/message/fusion/mptbase.c 		if (pa & MPI_ADDRESS_REPLY_A_BIT)
pa                604 drivers/message/fusion/mptbase.c 			mpt_reply(ioc, pa);
pa                606 drivers/message/fusion/mptbase.c 			mpt_turbo_reply(ioc, pa);
pa                607 drivers/message/fusion/mptbase.c 		pa = CHIPREG_READ32_dmasync(&ioc->chip->ReplyFifo);
pa                608 drivers/message/fusion/mptbase.c 	} while (pa != 0xFFFFFFFF);
pa               1020 drivers/message/fusion/mptctl.c 	dma_addr_t	 pa;					// phys addr
pa               1069 drivers/message/fusion/mptctl.c 								 &pa);
pa                363 drivers/misc/fastrpc.c 	struct fastrpc_buf_overlap *pa = (struct fastrpc_buf_overlap *)a;
pa                366 drivers/misc/fastrpc.c 	int st = CMP(pa->start, pb->start);
pa                368 drivers/misc/fastrpc.c 	int ed = CMP(pb->end, pa->end);
pa                100 drivers/misc/mic/bus/scif_bus.h 				  phys_addr_t pa, size_t len);
pa                 97 drivers/misc/mic/bus/vop_bus.h 				  dma_addr_t pa, size_t len);
pa                215 drivers/misc/mic/card/mic_device.c 	       phys_addr_t pa, size_t len)
pa                219 drivers/misc/mic/card/mic_device.c 	return mic_card_map(&mdrv->mdev, pa, len);
pa                286 drivers/misc/mic/card/mic_device.c 				   dma_addr_t pa, size_t len)
pa                290 drivers/misc/mic/card/mic_device.c 	return mic_card_map(&mdrv->mdev, pa, len);
pa                243 drivers/misc/mic/card/mic_x100.c 	mdev->mmio.pa = MIC_X100_MMIO_BASE;
pa                 33 drivers/misc/mic/common/mic_dev.h 	phys_addr_t pa;
pa                104 drivers/misc/mic/host/mic_boot.c 				   dma_addr_t pa, size_t len)
pa                108 drivers/misc/mic/host/mic_boot.c 	return mdev->aper.va + pa;
pa                288 drivers/misc/mic/host/mic_boot.c 				    phys_addr_t pa, size_t len)
pa                292 drivers/misc/mic/host/mic_boot.c 	return mdev->aper.va + pa;
pa                199 drivers/misc/mic/host/mic_main.c 	mdev->mmio.pa = pci_resource_start(pdev, mdev->ops->mmio_bar);
pa                208 drivers/misc/mic/host/mic_main.c 	mdev->aper.pa = pci_resource_start(pdev, mdev->ops->aper_bar);
pa                210 drivers/misc/mic/host/mic_main.c 	mdev->aper.va = ioremap_wc(mdev->aper.pa, mdev->aper.len);
pa                 20 drivers/misc/mic/host/mic_smpt.c static inline u8 mic_sys_addr_to_smpt(struct mic_device *mdev, dma_addr_t pa)
pa                 22 drivers/misc/mic/host/mic_smpt.c 	return (pa - mdev->smpt->info.base) >> mdev->smpt->info.page_shift;
pa                 30 drivers/misc/mic/host/mic_smpt.c static inline u64 mic_smpt_offset(struct mic_device *mdev, dma_addr_t pa)
pa                 32 drivers/misc/mic/host/mic_smpt.c 	return pa & mic_system_page_mask(mdev);
pa                 35 drivers/misc/mic/host/mic_smpt.c static inline u64 mic_smpt_align_low(struct mic_device *mdev, dma_addr_t pa)
pa                 37 drivers/misc/mic/host/mic_smpt.c 	return ALIGN(pa - mic_system_page_mask(mdev),
pa                 41 drivers/misc/mic/host/mic_smpt.c static inline u64 mic_smpt_align_high(struct mic_device *mdev, dma_addr_t pa)
pa                 43 drivers/misc/mic/host/mic_smpt.c 	return ALIGN(pa, mdev->smpt->info.page_size);
pa                 60 drivers/misc/mic/host/mic_smpt.c mic_is_system_addr(struct mic_device *mdev, dma_addr_t pa)
pa                 62 drivers/misc/mic/host/mic_smpt.c 	return pa >= mdev->smpt->info.base && pa <= mic_max_system_addr(mdev);
pa                674 drivers/misc/mic/scif/scif_dma.c 		phys = phys - dev->sdev->aper->pa;
pa                207 drivers/misc/mic/scif/scif_mmap.c 		apt_base = sdev->aper->pa;
pa                288 drivers/misc/mic/scif/scif_mmap.c 				ep->remote_dev->sdev->aper->pa;
pa                360 drivers/misc/mic/scif/scif_nodeqp.c scif_p2p_setsg(phys_addr_t pa, int page_size, int page_cnt)
pa                371 drivers/misc/mic/scif/scif_nodeqp.c 		page = pfn_to_page(pa >> PAGE_SHIFT);
pa                373 drivers/misc/mic/scif/scif_nodeqp.c 		pa += page_size;
pa                393 drivers/misc/mic/scif/scif_nodeqp.c 	p2p->ppi_sg[SCIF_PPI_MMIO] = scif_p2p_setsg(psdev->mmio->pa,
pa                400 drivers/misc/mic/scif/scif_nodeqp.c 	p2p->ppi_sg[SCIF_PPI_APER] = scif_p2p_setsg(psdev->aper->pa,
pa               1120 drivers/misc/mic/scif/scif_rma.c 		apt_base = sdev->aper->pa;
pa                552 drivers/misc/mic/vop/vop_vringh.c 		daddr += vpdev->aper->pa;
pa                629 drivers/misc/mic/vop/vop_vringh.c 		daddr += vpdev->aper->pa;
pa               1039 drivers/misc/mic/vop/vop_vringh.c 		 unsigned long *size, unsigned long *pa)
pa               1054 drivers/misc/mic/vop/vop_vringh.c 		*pa = virt_to_phys(vpdev->hw_ops->get_dp(vpdev));
pa               1063 drivers/misc/mic/vop/vop_vringh.c 			*pa = virt_to_phys(vvr->vring.va);
pa               1079 drivers/misc/mic/vop/vop_vringh.c 	unsigned long pa, size = vma->vm_end - vma->vm_start, size_rem = size;
pa               1090 drivers/misc/mic/vop/vop_vringh.c 		i = vop_query_offset(vdev, offset, &size, &pa);
pa               1096 drivers/misc/mic/vop/vop_vringh.c 				      pa >> PAGE_SHIFT, size,
pa                681 drivers/net/can/usb/peak_usb/pcan_usb_core.c 	const struct peak_usb_adapter *pa = dev->adapter;
pa                683 drivers/net/can/usb/peak_usb/pcan_usb_core.c 	if (pa->dev_set_bittiming) {
pa                685 drivers/net/can/usb/peak_usb/pcan_usb_core.c 		int err = pa->dev_set_bittiming(dev, bt);
pa                702 drivers/net/can/usb/peak_usb/pcan_usb_core.c 	const struct peak_usb_adapter *pa = dev->adapter;
pa                704 drivers/net/can/usb/peak_usb/pcan_usb_core.c 	if (pa->dev_set_data_bittiming) {
pa                706 drivers/net/can/usb/peak_usb/pcan_usb_core.c 		int err = pa->dev_set_data_bittiming(dev, bt);
pa                286 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 			    struct xgbe_page_alloc *pa, int alloc_order,
pa                324 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	pa->pages = pages;
pa                325 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	pa->pages_len = PAGE_SIZE << order;
pa                326 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	pa->pages_offset = 0;
pa                327 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	pa->pages_dma = pages_dma;
pa                333 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 				 struct xgbe_page_alloc *pa,
pa                336 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	get_page(pa->pages);
pa                337 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	bd->pa = *pa;
pa                339 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	bd->dma_base = pa->pages_dma;
pa                340 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	bd->dma_off = pa->pages_offset;
pa                343 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	pa->pages_offset += len;
pa                344 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	if ((pa->pages_offset + len) > pa->pages_len) {
pa                346 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 		bd->pa_unmap = *pa;
pa                349 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 		pa->pages = NULL;
pa                350 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 		pa->pages_len = 0;
pa                351 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 		pa->pages_offset = 0;
pa                352 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 		pa->pages_dma = 0;
pa                490 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	if (rdata->rx.hdr.pa.pages)
pa                491 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 		put_page(rdata->rx.hdr.pa.pages);
pa                500 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	if (rdata->rx.buf.pa.pages)
pa                501 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 		put_page(rdata->rx.buf.pa.pages);
pa               2567 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 	packet = page_address(rdata->rx.hdr.pa.pages) +
pa               2568 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 		 rdata->rx.hdr.pa.pages_offset;
pa               2784 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 						rdata->rx.buf.pa.pages,
pa               2785 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 						rdata->rx.buf.pa.pages_offset,
pa               2788 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 				rdata->rx.buf.pa.pages = NULL;
pa                395 drivers/net/ethernet/amd/xgbe/xgbe.h 	struct xgbe_page_alloc pa;
pa                465 drivers/net/ethernet/aquantia/atlantic/aq_nic.c 	dx_buff->pa = dma_map_single(aq_nic_get_dev(self),
pa                470 drivers/net/ethernet/aquantia/atlantic/aq_nic.c 	if (unlikely(dma_mapping_error(aq_nic_get_dev(self), dx_buff->pa))) {
pa                532 drivers/net/ethernet/aquantia/atlantic/aq_nic.c 			dx_buff->pa = frag_pa;
pa                554 drivers/net/ethernet/aquantia/atlantic/aq_nic.c 		if (!dx_buff->is_gso && !dx_buff->is_vlan && dx_buff->pa) {
pa                557 drivers/net/ethernet/aquantia/atlantic/aq_nic.c 						 dx_buff->pa,
pa                562 drivers/net/ethernet/aquantia/atlantic/aq_nic.c 					       dx_buff->pa,
pa                238 drivers/net/ethernet/aquantia/atlantic/aq_ring.c 				dma_unmap_single(dev, buff->pa, buff->len,
pa                241 drivers/net/ethernet/aquantia/atlantic/aq_ring.c 				dma_unmap_page(dev, buff->pa, buff->len,
pa                252 drivers/net/ethernet/aquantia/atlantic/aq_ring.c 		buff->pa = 0U;
pa                462 drivers/net/ethernet/aquantia/atlantic/aq_ring.c 		buff->pa = aq_buf_daddr(&buff->rxdata);
pa                 39 drivers/net/ethernet/aquantia/atlantic/aq_ring.h 		dma_addr_t pa;
pa                473 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 			txd->buf_addr = buff->pa;
pa                590 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c 		rxd->buf_addr = buff->pa;
pa                534 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 			txd->buf_addr = buff->pa;
pa                658 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c 		rxd->buf_addr = buff->pa;
pa                142 drivers/net/ethernet/brocade/bna/bfa_cee.c 	cee->attr_dma.pa = dma_pa;
pa                144 drivers/net/ethernet/brocade/bna/bfa_cee.c 	cee->stats_dma.pa = dma_pa + bfa_cee_attr_meminfo();
pa                177 drivers/net/ethernet/brocade/bna/bfa_cee.c 	bfa_dma_be_addr_set(cmd->dma_addr, cee->attr_dma.pa);
pa               1953 drivers/net/ethernet/brocade/bna/bfa_ioc.c 	bfa_dma_be_addr_set(attr_req.attr_addr, ioc->attr_dma.pa);
pa               2584 drivers/net/ethernet/brocade/bna/bfa_ioc.c 	ioc->attr_dma.pa = dm_pa;
pa                 40 drivers/net/ethernet/brocade/bna/bfa_ioc.h 	u64	pa;	/* ! Physical address		*/
pa                 50 drivers/net/ethernet/brocade/bna/bfa_ioc.h #define bfa_dma_be_addr_set(dma_addr, pa)	\
pa                 51 drivers/net/ethernet/brocade/bna/bfa_ioc.h 		__bfa_dma_be_addr_set(&dma_addr, (u64)pa)
pa                 53 drivers/net/ethernet/brocade/bna/bfa_ioc.h __bfa_dma_be_addr_set(union bfi_addr_u *dma_addr, u64 pa)
pa                 55 drivers/net/ethernet/brocade/bna/bfa_ioc.h 	dma_addr->a32.addr_lo = (u32) htonl(pa);
pa                 56 drivers/net/ethernet/brocade/bna/bfa_ioc.h 	dma_addr->a32.addr_hi = (u32) htonl(upper_32_bits(pa));
pa                 63 drivers/net/ethernet/brocade/bna/bfa_ioc.h __bfa_alen_set(struct bfi_alen *alen, u32 len, u64 pa)
pa                 66 drivers/net/ethernet/brocade/bna/bfa_ioc.h 	bfa_dma_be_addr_set(alen->al_addr, pa);
pa                517 drivers/net/ethernet/brocade/bna/bfa_msgq.c 	bfa_dma_be_addr_set(msgq_cfg->cmdq.addr, msgq->cmdq.addr.pa);
pa                519 drivers/net/ethernet/brocade/bna/bfa_msgq.c 	bfa_dma_be_addr_set(msgq_cfg->rspq.addr, msgq->rspq.addr.pa);
pa                590 drivers/net/ethernet/brocade/bna/bfa_msgq.c bfa_msgq_memclaim(struct bfa_msgq *msgq, u8 *kva, u64 pa)
pa                593 drivers/net/ethernet/brocade/bna/bfa_msgq.c 	msgq->cmdq.addr.pa  = pa;
pa                596 drivers/net/ethernet/brocade/bna/bfa_msgq.c 	pa += roundup(BFA_MSGQ_CMDQ_SIZE, BFA_DMA_ALIGN_SZ);
pa                599 drivers/net/ethernet/brocade/bna/bfa_msgq.c 	msgq->rspq.addr.pa = pa;
pa                115 drivers/net/ethernet/brocade/bna/bfa_msgq.h void bfa_msgq_memclaim(struct bfa_msgq *msgq, u8 *kva, u64 pa);
pa               1955 drivers/net/ethernet/dec/tulip/de4x5.c     char *pa;
pa               1960 drivers/net/ethernet/dec/tulip/de4x5.c     pa = build_setup_frame(dev, ALL);        /* Build the basic frame */
pa               1982 drivers/net/ethernet/dec/tulip/de4x5.c 		*(pa + (i&1)) = *addrs++;
pa               1983 drivers/net/ethernet/dec/tulip/de4x5.c 		if (i & 0x01) pa += 4;
pa               5037 drivers/net/ethernet/dec/tulip/de4x5.c     char *pa = lp->setup_frame;
pa               5045 drivers/net/ethernet/dec/tulip/de4x5.c 	for (pa=lp->setup_frame+IMPERF_PA_OFFSET, i=0; i<ETH_ALEN; i++) {
pa               5046 drivers/net/ethernet/dec/tulip/de4x5.c 	    *(pa + i) = dev->dev_addr[i];                 /* Host address */
pa               5047 drivers/net/ethernet/dec/tulip/de4x5.c 	    if (i & 0x01) pa += 2;
pa               5052 drivers/net/ethernet/dec/tulip/de4x5.c 	    *(pa + (i&1)) = dev->dev_addr[i];
pa               5053 drivers/net/ethernet/dec/tulip/de4x5.c 	    if (i & 0x01) pa += 4;
pa               5056 drivers/net/ethernet/dec/tulip/de4x5.c 	    *(pa + (i&1)) = (char) 0xff;
pa               5057 drivers/net/ethernet/dec/tulip/de4x5.c 	    if (i & 0x01) pa += 4;
pa               5061 drivers/net/ethernet/dec/tulip/de4x5.c     return pa;                     /* Points to the next entry */
pa                161 drivers/net/ethernet/intel/i40e/i40e_adminq.c 			cpu_to_le32(upper_32_bits(bi->pa));
pa                163 drivers/net/ethernet/intel/i40e/i40e_adminq.c 			cpu_to_le32(lower_32_bits(bi->pa));
pa                250 drivers/net/ethernet/intel/i40e/i40e_adminq.c 		if (hw->aq.asq.r.asq_bi[i].pa)
pa                281 drivers/net/ethernet/intel/i40e/i40e_adminq.c 	wr32(hw, hw->aq.asq.bal, lower_32_bits(hw->aq.asq.desc_buf.pa));
pa                282 drivers/net/ethernet/intel/i40e/i40e_adminq.c 	wr32(hw, hw->aq.asq.bah, upper_32_bits(hw->aq.asq.desc_buf.pa));
pa                286 drivers/net/ethernet/intel/i40e/i40e_adminq.c 	if (reg != lower_32_bits(hw->aq.asq.desc_buf.pa))
pa                310 drivers/net/ethernet/intel/i40e/i40e_adminq.c 	wr32(hw, hw->aq.arq.bal, lower_32_bits(hw->aq.arq.desc_buf.pa));
pa                311 drivers/net/ethernet/intel/i40e/i40e_adminq.c 	wr32(hw, hw->aq.arq.bah, upper_32_bits(hw->aq.arq.desc_buf.pa));
pa                318 drivers/net/ethernet/intel/i40e/i40e_adminq.c 	if (reg != lower_32_bits(hw->aq.arq.desc_buf.pa))
pa                834 drivers/net/ethernet/intel/i40e/i40e_adminq.c 				cpu_to_le32(upper_32_bits(dma_buff->pa));
pa                836 drivers/net/ethernet/intel/i40e/i40e_adminq.c 				cpu_to_le32(lower_32_bits(dma_buff->pa));
pa               1015 drivers/net/ethernet/intel/i40e/i40e_adminq.c 	desc->params.external.addr_high = cpu_to_le32(upper_32_bits(bi->pa));
pa               1016 drivers/net/ethernet/intel/i40e/i40e_adminq.c 	desc->params.external.addr_low = cpu_to_le32(lower_32_bits(bi->pa));
pa                156 drivers/net/ethernet/intel/i40e/i40e_hmc.c 		page_desc = page->pa | 0x1;
pa                 99 drivers/net/ethernet/intel/i40e/i40e_hmc.h #define I40E_SET_PF_SD_ENTRY(hw, pa, sd_index, type)			\
pa                102 drivers/net/ethernet/intel/i40e/i40e_hmc.h 	val1 = (u32)(upper_32_bits(pa));				\
pa                103 drivers/net/ethernet/intel/i40e/i40e_hmc.h 	val2 = (u32)(pa) | (I40E_HMC_MAX_BP_COUNT <<			\
pa                388 drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c 					sd_entry->u.pd_table.pd_page_addr.pa,
pa                392 drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c 				I40E_SET_PF_SD_ENTRY(hw, sd_entry->u.bp.addr.pa,
pa                121 drivers/net/ethernet/intel/i40e/i40e_main.c 	mem->va = dma_alloc_coherent(&pf->pdev->dev, mem->size, &mem->pa,
pa                138 drivers/net/ethernet/intel/i40e/i40e_main.c 	dma_free_coherent(&pf->pdev->dev, mem->size, mem->va, mem->pa);
pa                140 drivers/net/ethernet/intel/i40e/i40e_main.c 	mem->pa = 0;
pa                 36 drivers/net/ethernet/intel/i40e/i40e_osdep.h 	dma_addr_t pa;
pa                147 drivers/net/ethernet/intel/iavf/iavf_adminq.c 			cpu_to_le32(upper_32_bits(bi->pa));
pa                149 drivers/net/ethernet/intel/iavf/iavf_adminq.c 			cpu_to_le32(lower_32_bits(bi->pa));
pa                237 drivers/net/ethernet/intel/iavf/iavf_adminq.c 		if (hw->aq.asq.r.asq_bi[i].pa)
pa                268 drivers/net/ethernet/intel/iavf/iavf_adminq.c 	wr32(hw, hw->aq.asq.bal, lower_32_bits(hw->aq.asq.desc_buf.pa));
pa                269 drivers/net/ethernet/intel/iavf/iavf_adminq.c 	wr32(hw, hw->aq.asq.bah, upper_32_bits(hw->aq.asq.desc_buf.pa));
pa                273 drivers/net/ethernet/intel/iavf/iavf_adminq.c 	if (reg != lower_32_bits(hw->aq.asq.desc_buf.pa))
pa                297 drivers/net/ethernet/intel/iavf/iavf_adminq.c 	wr32(hw, hw->aq.arq.bal, lower_32_bits(hw->aq.arq.desc_buf.pa));
pa                298 drivers/net/ethernet/intel/iavf/iavf_adminq.c 	wr32(hw, hw->aq.arq.bah, upper_32_bits(hw->aq.arq.desc_buf.pa));
pa                305 drivers/net/ethernet/intel/iavf/iavf_adminq.c 	if (reg != lower_32_bits(hw->aq.arq.desc_buf.pa))
pa                736 drivers/net/ethernet/intel/iavf/iavf_adminq.c 				cpu_to_le32(upper_32_bits(dma_buff->pa));
pa                738 drivers/net/ethernet/intel/iavf/iavf_adminq.c 				cpu_to_le32(lower_32_bits(dma_buff->pa));
pa                916 drivers/net/ethernet/intel/iavf/iavf_adminq.c 	desc->params.external.addr_high = cpu_to_le32(upper_32_bits(bi->pa));
pa                917 drivers/net/ethernet/intel/iavf/iavf_adminq.c 	desc->params.external.addr_low = cpu_to_le32(lower_32_bits(bi->pa));
pa                 83 drivers/net/ethernet/intel/iavf/iavf_main.c 				     (dma_addr_t *)&mem->pa, GFP_KERNEL);
pa                103 drivers/net/ethernet/intel/iavf/iavf_main.c 			  mem->va, (dma_addr_t)mem->pa);
pa                 32 drivers/net/ethernet/intel/iavf/iavf_osdep.h 	dma_addr_t pa;
pa                 81 drivers/net/ethernet/intel/ice/ice_controlq.c 						 &cq->sq.desc_buf.pa,
pa                 91 drivers/net/ethernet/intel/ice/ice_controlq.c 				   cq->sq.desc_buf.va, cq->sq.desc_buf.pa);
pa                 93 drivers/net/ethernet/intel/ice/ice_controlq.c 		cq->sq.desc_buf.pa = 0;
pa                112 drivers/net/ethernet/intel/ice/ice_controlq.c 						 &cq->rq.desc_buf.pa,
pa                131 drivers/net/ethernet/intel/ice/ice_controlq.c 			   ring->desc_buf.va, ring->desc_buf.pa);
pa                133 drivers/net/ethernet/intel/ice/ice_controlq.c 	ring->desc_buf.pa = 0;
pa                163 drivers/net/ethernet/intel/ice/ice_controlq.c 					     cq->rq_buf_size, &bi->pa,
pa                184 drivers/net/ethernet/intel/ice/ice_controlq.c 			cpu_to_le32(upper_32_bits(bi->pa));
pa                186 drivers/net/ethernet/intel/ice/ice_controlq.c 			cpu_to_le32(lower_32_bits(bi->pa));
pa                197 drivers/net/ethernet/intel/ice/ice_controlq.c 				   cq->rq.r.rq_bi[i].va, cq->rq.r.rq_bi[i].pa);
pa                199 drivers/net/ethernet/intel/ice/ice_controlq.c 		cq->rq.r.rq_bi[i].pa = 0;
pa                230 drivers/net/ethernet/intel/ice/ice_controlq.c 					     cq->sq_buf_size, &bi->pa,
pa                243 drivers/net/ethernet/intel/ice/ice_controlq.c 				   cq->sq.r.sq_bi[i].va, cq->sq.r.sq_bi[i].pa);
pa                245 drivers/net/ethernet/intel/ice/ice_controlq.c 		cq->sq.r.sq_bi[i].pa = 0;
pa                262 drivers/net/ethernet/intel/ice/ice_controlq.c 	wr32(hw, ring->bal, lower_32_bits(ring->desc_buf.pa));
pa                263 drivers/net/ethernet/intel/ice/ice_controlq.c 	wr32(hw, ring->bah, upper_32_bits(ring->desc_buf.pa));
pa                266 drivers/net/ethernet/intel/ice/ice_controlq.c 	if (rd32(hw, ring->bal) != lower_32_bits(ring->desc_buf.pa))
pa                430 drivers/net/ethernet/intel/ice/ice_controlq.c 		if ((qi)->ring.r.ring##_bi[i].pa) {			\
pa                434 drivers/net/ethernet/intel/ice/ice_controlq.c 					   (qi)->ring.r.ring##_bi[i].pa);\
pa                436 drivers/net/ethernet/intel/ice/ice_controlq.c 			(qi)->ring.r.ring##_bi[i].pa = 0;		\
pa                931 drivers/net/ethernet/intel/ice/ice_controlq.c 			cpu_to_le32(upper_32_bits(dma_buf->pa));
pa                933 drivers/net/ethernet/intel/ice/ice_controlq.c 			cpu_to_le32(lower_32_bits(dma_buf->pa));
pa               1102 drivers/net/ethernet/intel/ice/ice_controlq.c 	desc->params.generic.addr_high = cpu_to_le32(upper_32_bits(bi->pa));
pa               1103 drivers/net/ethernet/intel/ice/ice_controlq.c 	desc->params.generic.addr_low = cpu_to_le32(lower_32_bits(bi->pa));
pa                 23 drivers/net/ethernet/intel/ice/ice_osdep.h 	dma_addr_t pa;
pa                472 drivers/net/ethernet/mellanox/mlxsw/cmd.h MLXSW_ITEM64_INDEXED(cmd_mbox, map_fa, pa, 0x00, 12, 52, 0x08, 0x00, true);
pa                922 drivers/net/ethernet/mellanox/mlxsw/cmd.h MLXSW_ITEM64_INDEXED(cmd_mbox, sw2hw_dq, pa, 0x10, 12, 52, 0x08, 0x00, true);
pa               1077 drivers/net/ethernet/mellanox/mlxsw/cmd.h MLXSW_ITEM64_INDEXED(cmd_mbox, sw2hw_cq, pa, 0x10, 11, 53, 0x08, 0x00, true);
pa               1161 drivers/net/ethernet/mellanox/mlxsw/cmd.h MLXSW_ITEM64_INDEXED(cmd_mbox, sw2hw_eq, pa, 0x10, 11, 53, 0x08, 0x00, true);
pa                 46 drivers/net/ethernet/realtek/r8169_firmware.c 	struct rtl_fw_phy_action *pa = &rtl_fw->phy_action;
pa                 73 drivers/net/ethernet/realtek/r8169_firmware.c 		pa->code = (__le32 *)(fw->data + start);
pa                 74 drivers/net/ethernet/realtek/r8169_firmware.c 		pa->size = size;
pa                 81 drivers/net/ethernet/realtek/r8169_firmware.c 		pa->code = (__le32 *)fw->data;
pa                 82 drivers/net/ethernet/realtek/r8169_firmware.c 		pa->size = fw->size / FW_OPCODE_SIZE;
pa                 90 drivers/net/ethernet/realtek/r8169_firmware.c 	struct rtl_fw_phy_action *pa = &rtl_fw->phy_action;
pa                 93 drivers/net/ethernet/realtek/r8169_firmware.c 	for (index = 0; index < pa->size; index++) {
pa                 94 drivers/net/ethernet/realtek/r8169_firmware.c 		u32 action = le32_to_cpu(pa->code[index]);
pa                113 drivers/net/ethernet/realtek/r8169_firmware.c 			if (index + 2 >= pa->size)
pa                119 drivers/net/ethernet/realtek/r8169_firmware.c 			if (index + 1 + regno >= pa->size)
pa                137 drivers/net/ethernet/realtek/r8169_firmware.c 	struct rtl_fw_phy_action *pa = &rtl_fw->phy_action;
pa                143 drivers/net/ethernet/realtek/r8169_firmware.c 	for (index = 0; index < pa->size; index++) {
pa                144 drivers/net/ethernet/realtek/r8169_firmware.c 		u32 action = le32_to_cpu(pa->code[index]);
pa                 41 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 	if (desc_data->rx.hdr.pa.pages)
pa                 42 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 		put_page(desc_data->rx.hdr.pa.pages);
pa                 51 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 	if (desc_data->rx.buf.pa.pages)
pa                 52 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 		put_page(desc_data->rx.buf.pa.pages);
pa                331 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 			      struct xlgmac_page_alloc *pa,
pa                357 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 	pa->pages = pages;
pa                358 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 	pa->pages_len = PAGE_SIZE << order;
pa                359 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 	pa->pages_offset = 0;
pa                360 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 	pa->pages_dma = pages_dma;
pa                366 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 				   struct xlgmac_page_alloc *pa,
pa                369 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 	get_page(pa->pages);
pa                370 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 	bd->pa = *pa;
pa                372 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 	bd->dma_base = pa->pages_dma;
pa                373 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 	bd->dma_off = pa->pages_offset;
pa                376 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 	pa->pages_offset += len;
pa                377 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 	if ((pa->pages_offset + len) > pa->pages_len) {
pa                379 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 		bd->pa_unmap = *pa;
pa                382 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 		pa->pages = NULL;
pa                383 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 		pa->pages_len = 0;
pa                384 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 		pa->pages_offset = 0;
pa                385 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 		pa->pages_dma = 0;
pa               1008 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 	packet = page_address(desc_data->rx.hdr.pa.pages) +
pa               1009 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 		 desc_data->rx.hdr.pa.pages_offset;
pa               1025 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 				desc_data->rx.buf.pa.pages,
pa               1026 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 				desc_data->rx.buf.pa.pages_offset,
pa               1028 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 		desc_data->rx.buf.pa.pages = NULL;
pa               1206 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 					desc_data->rx.buf.pa.pages,
pa               1207 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 					desc_data->rx.buf.pa.pages_offset,
pa               1210 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 				desc_data->rx.buf.pa.pages = NULL;
pa                231 drivers/net/ethernet/synopsys/dwc-xlgmac.h 	struct xlgmac_page_alloc pa;
pa                675 drivers/net/ethernet/ti/cpsw.c 	void			*pa = page_address(page);
pa                676 drivers/net/ethernet/ti/cpsw.c 	struct cpsw_meta_xdp	*xmeta = pa + CPSW_XMETA_OFFSET;
pa                724 drivers/net/ethernet/ti/cpsw.c 			xdp.data = pa + CPSW_HEADROOM +
pa                729 drivers/net/ethernet/ti/cpsw.c 			xdp.data = pa + CPSW_HEADROOM;
pa                735 drivers/net/ethernet/ti/cpsw.c 		xdp.data_hard_start = pa;
pa                751 drivers/net/ethernet/ti/cpsw.c 	skb = build_skb(pa, cpsw_rxbuf_total_len(pkt_size));
pa                789 drivers/net/fddi/defza.h 	u8 pa[2];			/* preamble */
pa                 35 drivers/net/fddi/skfp/pmf.c static int smt_set_para(struct s_smc *smc, struct smt_para *pa, int index,
pa                312 drivers/net/fddi/skfp/pmf.c 	struct smt_para		*pa ;
pa                369 drivers/net/fddi/skfp/pmf.c 	pa = (struct smt_para *) (req + 1) ;
pa                374 drivers/net/fddi/skfp/pmf.c 		if (((u_short)len < pa->p_len + PARA_LEN) || (pa->p_len & 3)) {
pa                379 drivers/net/fddi/skfp/pmf.c 		if (((range = (pa->p_type & 0xf000)) == 0x2000) ||
pa                384 drivers/net/fddi/skfp/pmf.c 			index = *((u_char *)pa + PARA_LEN + 3) ;/* index */
pa                386 drivers/net/fddi/skfp/pmf.c 			if (!set && (pa->p_len != 4)) {
pa                415 drivers/net/fddi/skfp/pmf.c 			if (!set && (pa->p_len != 0)) {
pa                427 drivers/net/fddi/skfp/pmf.c 			pt = smt_get_ptab(pa->p_type) ;
pa                442 drivers/net/fddi/skfp/pmf.c 			else if (pa->p_type != SMT_P_AUTHOR &&
pa                443 drivers/net/fddi/skfp/pmf.c 				 (!set || (pa->p_type != SMT_P1035))) {
pa                446 drivers/net/fddi/skfp/pmf.c 					smt_add_para(smc,&pcon,pa->p_type,
pa                450 drivers/net/fddi/skfp/pmf.c 					st = smt_set_para(smc,pa,index,local,1);
pa                454 drivers/net/fddi/skfp/pmf.c 					smt_add_para(smc,&pcon,pa->p_type,
pa                463 drivers/net/fddi/skfp/pmf.c 					smt_add_para(smc,&pcon,pa->p_type,
pa                471 drivers/net/fddi/skfp/pmf.c 		len -= pa->p_len + PARA_LEN ;
pa                472 drivers/net/fddi/skfp/pmf.c 		pa = (struct smt_para *) ((char *)pa + pa->p_len + PARA_LEN) ;
pa                497 drivers/net/fddi/skfp/pmf.c 	struct smt_para	*pa ;
pa                519 drivers/net/fddi/skfp/pmf.c 		pa = (struct smt_para *) sm_to_para(smc,sm,SMT_P_AUTHOR) ;
pa                520 drivers/net/fddi/skfp/pmf.c 		if (!pa)
pa                522 drivers/net/fddi/skfp/pmf.c 		if (pa->p_len != 8)
pa                524 drivers/net/fddi/skfp/pmf.c 		if (memcmp((char *)(pa+1),(char *)smc->mib.fddiPRPMFPasswd,8))
pa                532 drivers/net/fddi/skfp/pmf.c 	struct smt_para	*pa ;
pa                535 drivers/net/fddi/skfp/pmf.c 	pa = (struct smt_para *) sm_to_para(smc,sm,SMT_P1035) ;
pa                536 drivers/net/fddi/skfp/pmf.c 	if (pa) {
pa                537 drivers/net/fddi/skfp/pmf.c 		sc = (struct smt_p_setcount *) pa ;
pa                549 drivers/net/fddi/skfp/pmf.c 	struct smt_para	*pa ;
pa                582 drivers/net/fddi/skfp/pmf.c 	pa = (struct smt_para *) to ;	/* type/length pointer */
pa               1047 drivers/net/fddi/skfp/pmf.c 	pa->p_type = para ;
pa               1048 drivers/net/fddi/skfp/pmf.c 	pa->p_len = plen - len - PARA_LEN ;
pa               1071 drivers/net/fddi/skfp/pmf.c static int smt_set_para(struct s_smc *smc, struct smt_para *pa, int index,
pa               1097 drivers/net/fddi/skfp/pmf.c 	len = pa->p_len ;
pa               1098 drivers/net/fddi/skfp/pmf.c 	from = (char *) (pa + 1 ) ;
pa               1101 drivers/net/fddi/skfp/pmf.c 	switch (pa->p_type & 0xf000) {
pa               1134 drivers/net/fddi/skfp/pmf.c 	switch (pa->p_type) {
pa               1154 drivers/net/fddi/skfp/pmf.c 	pt = smt_get_ptab(pa->p_type) ;
pa               1156 drivers/net/fddi/skfp/pmf.c 		return (pa->p_type & 0xff00) ? SMT_RDF_NOPARAM :
pa               1295 drivers/net/fddi/skfp/pmf.c 	switch (pa->p_type) {
pa               1559 drivers/net/fddi/skfp/pmf.c 	struct smt_para	*pa ;
pa               1589 drivers/net/fddi/skfp/pmf.c 	pa = (struct smt_para *) (sm + 1) ;
pa               1593 drivers/net/fddi/skfp/pmf.c 		printf("TYPE %x LEN %x VALUE\t",pa->p_type,pa->p_len) ;
pa               1595 drivers/net/fddi/skfp/pmf.c 		printf("TYPE %04x LEN %2x VALUE\t",pa->p_type,pa->p_len) ;
pa               1597 drivers/net/fddi/skfp/pmf.c 		n = pa->p_len ;
pa               1607 drivers/net/fddi/skfp/pmf.c 			dump_hex((char *)(pa+1),(int) n) ;
pa               1612 drivers/net/fddi/skfp/pmf.c 			c = (char *)(pa+1) ;
pa               1636 drivers/net/fddi/skfp/pmf.c 		plen = (pa->p_len + PARA_LEN + 3) & ~3 ;
pa               1638 drivers/net/fddi/skfp/pmf.c 		pa = (struct smt_para *)((char *)pa + plen) ;
pa               1824 drivers/net/fddi/skfp/smt.c 	struct smt_para	*pa ;
pa               1841 drivers/net/fddi/skfp/smt.c 		pa = (struct smt_para *) p ;
pa               1842 drivers/net/fddi/skfp/smt.c 		plen = pa->p_len ;
pa               1843 drivers/net/fddi/skfp/smt.c 		type = pa->p_type ;
pa               1844 drivers/net/fddi/skfp/smt.c 		pa->p_type = smt_swap_short(pa->p_type) ;
pa               1845 drivers/net/fddi/skfp/smt.c 		pa->p_len = smt_swap_short(pa->p_len) ;
pa               1847 drivers/net/fddi/skfp/smt.c 			plen = pa->p_len ;
pa               1848 drivers/net/fddi/skfp/smt.c 			type = pa->p_type ;
pa                710 drivers/net/hyperv/hyperv_net.h 			u64 pa; /* Physical Address */
pa                750 drivers/net/hyperv/hyperv_net.h 			u64 pa; /* Physical Address */
pa                 93 drivers/net/wireless/ath/wil6210/debugfs.c 	seq_printf(s, "  pa     = %pad\n", &ring->pa);
pa                213 drivers/net/wireless/ath/wil6210/debugfs.c 	seq_printf(s, "  pa     = %pad\n", &sring->pa);
pa                 28 drivers/net/wireless/ath/wil6210/pmc.c 	dma_addr_t pa;
pa                159 drivers/net/wireless/ath/wil6210/pmc.c 			&pmc->descriptors[i].pa,
pa                174 drivers/net/wireless/ath/wil6210/pmc.c 			cpu_to_le32(lower_32_bits(pmc->descriptors[i].pa));
pa                176 drivers/net/wireless/ath/wil6210/pmc.c 			cpu_to_le16((u16)upper_32_bits(pmc->descriptors[i].pa));
pa                212 drivers/net/wireless/ath/wil6210/pmc.c 				  pmc->descriptors[i].pa);
pa                296 drivers/net/wireless/ath/wil6210/pmc.c 					  pmc->descriptors[i].pa);
pa                153 drivers/net/wireless/ath/wil6210/txrx.c 	vring->va = dma_alloc_coherent(dev, sz, &vring->pa, GFP_KERNEL);
pa                176 drivers/net/wireless/ath/wil6210/txrx.c 		     vring->va, &vring->pa, vring->ctx);
pa                185 drivers/net/wireless/ath/wil6210/txrx.c 	dma_addr_t pa = wil_desc_addr(&d->dma.addr);
pa                190 drivers/net/wireless/ath/wil6210/txrx.c 		dma_unmap_single(dev, pa, dmalen, DMA_TO_DEVICE);
pa                193 drivers/net/wireless/ath/wil6210/txrx.c 		dma_unmap_page(dev, pa, dmalen, DMA_TO_DEVICE);
pa                211 drivers/net/wireless/ath/wil6210/txrx.c 			     &vring->pa, vring->ctx);
pa                215 drivers/net/wireless/ath/wil6210/txrx.c 			     &vring->pa, vring->ctx);
pa                219 drivers/net/wireless/ath/wil6210/txrx.c 		dma_addr_t pa;
pa                248 drivers/net/wireless/ath/wil6210/txrx.c 			pa = wil_desc_addr(&d->dma.addr);
pa                250 drivers/net/wireless/ath/wil6210/txrx.c 			dma_unmap_single(dev, pa, dmalen, DMA_FROM_DEVICE);
pa                255 drivers/net/wireless/ath/wil6210/txrx.c 	dma_free_coherent(dev, sz, (void *)vring->va, vring->pa);
pa                257 drivers/net/wireless/ath/wil6210/txrx.c 	vring->pa = 0;
pa                274 drivers/net/wireless/ath/wil6210/txrx.c 	dma_addr_t pa;
pa                289 drivers/net/wireless/ath/wil6210/txrx.c 	pa = dma_map_single(dev, skb->data, skb->len, DMA_FROM_DEVICE);
pa                290 drivers/net/wireless/ath/wil6210/txrx.c 	if (unlikely(dma_mapping_error(dev, pa))) {
pa                296 drivers/net/wireless/ath/wil6210/txrx.c 	wil_desc_addr_set(&d->dma.addr, pa);
pa                461 drivers/net/wireless/ath/wil6210/txrx.c 	dma_addr_t pa;
pa                492 drivers/net/wireless/ath/wil6210/txrx.c 	pa = wil_desc_addr(&d->dma.addr);
pa                494 drivers/net/wireless/ath/wil6210/txrx.c 	dma_unmap_single(dev, pa, sz, DMA_FROM_DEVICE);
pa               1126 drivers/net/wireless/ath/wil6210/txrx.c static int wil_tx_desc_map(union wil_tx_desc *desc, dma_addr_t pa,
pa               1131 drivers/net/wireless/ath/wil6210/txrx.c 	wil_desc_addr_set(&d->dma.addr, pa);
pa               1227 drivers/net/wireless/ath/wil6210/txrx.c 	cmd.vring_cfg.tx_sw_ring.ring_mem_base = cpu_to_le64(vring->pa);
pa               1318 drivers/net/wireless/ath/wil6210/txrx.c 	cmd.vring_cfg.tx_sw_ring.ring_mem_base = cpu_to_le64(vring->pa);
pa               1396 drivers/net/wireless/ath/wil6210/txrx.c 	cmd.vring_cfg.tx_sw_ring.ring_mem_base = cpu_to_le64(vring->pa);
pa               1759 drivers/net/wireless/ath/wil6210/txrx.c 	dma_addr_t pa;
pa               1827 drivers/net/wireless/ath/wil6210/txrx.c 	pa = dma_map_single(dev, skb->data, hdrlen, DMA_TO_DEVICE);
pa               1828 drivers/net/wireless/ath/wil6210/txrx.c 	if (unlikely(dma_mapping_error(dev, pa))) {
pa               1833 drivers/net/wireless/ath/wil6210/txrx.c 	wil->txrx_ops.tx_desc_map((union wil_tx_desc *)hdr_desc, pa,
pa               1872 drivers/net/wireless/ath/wil6210/txrx.c 				pa = skb_frag_dma_map(dev, frag,
pa               1877 drivers/net/wireless/ath/wil6210/txrx.c 				pa = dma_map_single(dev,
pa               1886 drivers/net/wireless/ath/wil6210/txrx.c 			if (unlikely(dma_mapping_error(dev, pa))) {
pa               1902 drivers/net/wireless/ath/wil6210/txrx.c 						  pa, lenmss, vring_index);
pa               2046 drivers/net/wireless/ath/wil6210/txrx.c 	dma_addr_t pa;
pa               2065 drivers/net/wireless/ath/wil6210/txrx.c 	pa = dma_map_single(dev, skb->data, skb_headlen(skb), DMA_TO_DEVICE);
pa               2068 drivers/net/wireless/ath/wil6210/txrx.c 		     skb_headlen(skb), skb->data, &pa);
pa               2072 drivers/net/wireless/ath/wil6210/txrx.c 	if (unlikely(dma_mapping_error(dev, pa)))
pa               2076 drivers/net/wireless/ath/wil6210/txrx.c 	wil->txrx_ops.tx_desc_map((union wil_tx_desc *)d, pa, len,
pa               2104 drivers/net/wireless/ath/wil6210/txrx.c 		pa = skb_frag_dma_map(dev, frag, 0, skb_frag_size(frag),
pa               2106 drivers/net/wireless/ath/wil6210/txrx.c 		if (unlikely(dma_mapping_error(dev, pa))) {
pa               2113 drivers/net/wireless/ath/wil6210/txrx.c 					   pa, len, ring_index);
pa                 42 drivers/net/wireless/ath/wil6210/txrx.h 				     dma_addr_t pa)
pa                 44 drivers/net/wireless/ath/wil6210/txrx.h 	addr->addr_low = cpu_to_le32(lower_32_bits(pa));
pa                 45 drivers/net/wireless/ath/wil6210/txrx.h 	addr->addr_high = cpu_to_le16((u16)upper_32_bits(pa));
pa                 43 drivers/net/wireless/ath/wil6210/txrx_edma.c 	dma_addr_t pa = wil_tx_desc_get_addr_edma(&d->dma);
pa                 48 drivers/net/wireless/ath/wil6210/txrx_edma.c 		dma_unmap_single(dev, pa, dmalen, DMA_TO_DEVICE);
pa                 51 drivers/net/wireless/ath/wil6210/txrx_edma.c 		dma_unmap_page(dev, pa, dmalen, DMA_TO_DEVICE);
pa                 82 drivers/net/wireless/ath/wil6210/txrx_edma.c 		     sz, sring->va, &sring->pa);
pa                 84 drivers/net/wireless/ath/wil6210/txrx_edma.c 	dma_free_coherent(dev, sz, (void *)sring->va, sring->pa);
pa                 85 drivers/net/wireless/ath/wil6210/txrx_edma.c 	sring->pa = 0;
pa                107 drivers/net/wireless/ath/wil6210/txrx_edma.c 	sring->va = dma_alloc_coherent(dev, sz, &sring->pa, GFP_KERNEL);
pa                112 drivers/net/wireless/ath/wil6210/txrx_edma.c 		     &sring->pa);
pa                169 drivers/net/wireless/ath/wil6210/txrx_edma.c 	dma_addr_t pa;
pa                197 drivers/net/wireless/ath/wil6210/txrx_edma.c 	pa = dma_map_single(dev, skb->data, skb->len, DMA_FROM_DEVICE);
pa                198 drivers/net/wireless/ath/wil6210/txrx_edma.c 	if (unlikely(dma_mapping_error(dev, pa))) {
pa                212 drivers/net/wireless/ath/wil6210/txrx_edma.c 	wil_desc_set_addr_edma(&d->dma.addr, &d->dma.addr_high_high, pa);
pa                218 drivers/net/wireless/ath/wil6210/txrx_edma.c 	memcpy(skb->cb, &pa, sizeof(pa));
pa                281 drivers/net/wireless/ath/wil6210/txrx_edma.c 	dma_addr_t pa;
pa                295 drivers/net/wireless/ath/wil6210/txrx_edma.c 			memcpy(&pa, skb->cb, sizeof(pa));
pa                296 drivers/net/wireless/ath/wil6210/txrx_edma.c 			dma_unmap_single(dev, pa, wil->rx_buf_len,
pa                402 drivers/net/wireless/ath/wil6210/txrx_edma.c 	ring->va = dma_alloc_coherent(dev, sz, &ring->pa, GFP_KERNEL);
pa                409 drivers/net/wireless/ath/wil6210/txrx_edma.c 			dma_alloc_coherent(dev, sz, &ring->edma_rx_swtail.pa,
pa                417 drivers/net/wireless/ath/wil6210/txrx_edma.c 		     ring->size, ring->va, &ring->pa, ring->ctx);
pa                422 drivers/net/wireless/ath/wil6210/txrx_edma.c 			  (void *)ring->va, ring->pa);
pa                446 drivers/net/wireless/ath/wil6210/txrx_edma.c 			     &ring->pa, ring->ctx);
pa                451 drivers/net/wireless/ath/wil6210/txrx_edma.c 				  ring->edma_rx_swtail.pa);
pa                460 drivers/net/wireless/ath/wil6210/txrx_edma.c 		     &ring->pa, ring->ctx);
pa                486 drivers/net/wireless/ath/wil6210/txrx_edma.c 	dma_free_coherent(dev, sz, (void *)ring->va, ring->pa);
pa                488 drivers/net/wireless/ath/wil6210/txrx_edma.c 	ring->pa = 0;
pa                871 drivers/net/wireless/ath/wil6210/txrx_edma.c 	dma_addr_t pa;
pa                940 drivers/net/wireless/ath/wil6210/txrx_edma.c 	memcpy(&pa, skb->cb, sizeof(pa));
pa                941 drivers/net/wireless/ath/wil6210/txrx_edma.c 	dma_unmap_single(dev, pa, sz, DMA_FROM_DEVICE);
pa               1120 drivers/net/wireless/ath/wil6210/txrx_edma.c 				dma_addr_t pa,
pa               1129 drivers/net/wireless/ath/wil6210/txrx_edma.c 	wil_desc_set_addr_edma(&d->dma.addr, &d->dma.addr_high_high, pa);
pa               1366 drivers/net/wireless/ath/wil6210/txrx_edma.c 	dma_addr_t pa;
pa               1372 drivers/net/wireless/ath/wil6210/txrx_edma.c 		pa = dma_map_single(dev, buff_addr, len, DMA_TO_DEVICE);
pa               1375 drivers/net/wireless/ath/wil6210/txrx_edma.c 		pa = skb_frag_dma_map(dev, frag, 0, len, DMA_TO_DEVICE);
pa               1378 drivers/net/wireless/ath/wil6210/txrx_edma.c 	if (unlikely(dma_mapping_error(dev, pa))) {
pa               1383 drivers/net/wireless/ath/wil6210/txrx_edma.c 	wil->txrx_ops.tx_desc_map((union wil_tx_desc *)d, pa,
pa                583 drivers/net/wireless/ath/wil6210/txrx_edma.h 					  dma_addr_t pa)
pa                585 drivers/net/wireless/ath/wil6210/txrx_edma.h 	addr->addr_low = cpu_to_le32(lower_32_bits(pa));
pa                586 drivers/net/wireless/ath/wil6210/txrx_edma.h 	addr->addr_high = cpu_to_le16((u16)upper_32_bits(pa));
pa                587 drivers/net/wireless/ath/wil6210/txrx_edma.h 	*addr_high_high = cpu_to_le16((u16)(upper_32_bits(pa) >> 16));
pa                523 drivers/net/wireless/ath/wil6210/wil6210.h 	dma_addr_t pa;
pa                532 drivers/net/wireless/ath/wil6210/wil6210.h 	dma_addr_t pa;
pa                559 drivers/net/wireless/ath/wil6210/wil6210.h 	dma_addr_t pa;
pa                611 drivers/net/wireless/ath/wil6210/wil6210.h 	int (*tx_desc_map)(union wil_tx_desc *desc, dma_addr_t pa,
pa               2604 drivers/net/wireless/ath/wil6210/wmi.c 			.ring_mem_base = cpu_to_le64(vring->pa),
pa               3721 drivers/net/wireless/ath/wil6210/wmi.c 	cmd.ring_cfg.ring_mem_base = cpu_to_le64(sring->pa);
pa               3800 drivers/net/wireless/ath/wil6210/wmi.c 	cmd.ring_cfg.ring_mem_base = cpu_to_le64(sring->pa);
pa               3841 drivers/net/wireless/ath/wil6210/wmi.c 	cmd.ring_cfg.ring_mem_base = cpu_to_le64(ring->pa);
pa               3842 drivers/net/wireless/ath/wil6210/wmi.c 	cmd.sw_tail_host_addr = cpu_to_le64(ring->edma_rx_swtail.pa);
pa               3892 drivers/net/wireless/ath/wil6210/wmi.c 	cmd.ring_cfg.ring_mem_base = cpu_to_le64(ring->pa);
pa               3938 drivers/net/wireless/ath/wil6210/wmi.c 	cmd.ring_cfg.ring_mem_base = cpu_to_le64(ring->pa);
pa               1840 drivers/net/wireless/broadcom/b43/phy_lp.c 			    bool rx, bool pa, struct lpphy_tx_gains *gains)
pa               1877 drivers/net/wireless/broadcom/b43/phy_lp.c 		pa = false;
pa               1885 drivers/net/wireless/broadcom/b43/phy_lp.c 				0xFFF7, pa << 3);
pa               1889 drivers/net/wireless/broadcom/b43/phy_lp.c 				0xFFDF, pa << 5);
pa                716 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 	     dma_addr_t pa, uint outidx, u32 *flags, u32 bufcount)
pa                721 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 	if ((di->dataoffsetlow == 0) || !(pa & PCI32ADDR_HIGH)) {
pa                722 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 		ddring[outidx].addrlow = cpu_to_le32(pa + di->dataoffsetlow);
pa                730 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 		ae = (pa & PCI32ADDR_HIGH) >> PCI32ADDR_HIGH_SHIFT;
pa                731 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 		pa &= ~PCI32ADDR_HIGH;
pa                734 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 		ddring[outidx].addrlow = cpu_to_le32(pa + di->dataoffsetlow);
pa                774 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c _dma_ddtable_init(struct dma_info *di, uint direction, dma_addr_t pa)
pa                778 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 			di->xmtptrbase = pa;
pa                780 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 			di->rcvptrbase = pa;
pa                784 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 	    || !(pa & PCI32ADDR_HIGH)) {
pa                787 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 				     pa + di->ddoffsetlow);
pa                792 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 				     pa + di->ddoffsetlow);
pa                801 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 		ae = (pa & PCI32ADDR_HIGH) >> PCI32ADDR_HIGH_SHIFT;
pa                802 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 		pa &= ~PCI32ADDR_HIGH;
pa                806 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 				     pa + di->ddoffsetlow);
pa                813 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 				     pa + di->ddoffsetlow);
pa                873 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 	dma_addr_t pa;
pa                894 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 	pa = le32_to_cpu(di->rxd64[i].addrlow) - di->dataoffsetlow;
pa                897 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 	dma_unmap_single(di->dmadev, pa, di->rxbufsize, DMA_FROM_DEVICE);
pa               1031 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 	dma_addr_t pa;
pa               1080 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 		pa = dma_map_single(di->dmadev, p->data, di->rxbufsize,
pa               1082 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 		if (dma_mapping_error(di->dmadev, pa)) {
pa               1095 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 		dma64_dd_upd(di, di->rxd64, pa, rxout, &flags,
pa               1277 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 	dma_addr_t pa;
pa               1291 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 	pa = dma_map_single(di->dmadev, data, len, DMA_TO_DEVICE);
pa               1293 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 	if (dma_mapping_error(di->dmadev, pa)) {
pa               1306 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 	dma64_dd_upd(di, di->txd64, pa, txout, &flags, len);
pa               1512 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 		dma_addr_t pa;
pa               1515 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 		pa = le32_to_cpu(di->txd64[i].addrlow) - di->dataoffsetlow;
pa               1527 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 		dma_unmap_single(di->dmadev, pa, size, DMA_TO_DEVICE);
pa                175 drivers/of/fdt.c 		const char *p = nodename, *ps = p, *pa = NULL;
pa                180 drivers/of/fdt.c 				pa = p;
pa                186 drivers/of/fdt.c 		if (pa < ps)
pa                187 drivers/of/fdt.c 			pa = p;
pa                188 drivers/of/fdt.c 		len = (pa - ps) + 1;
pa                559 drivers/parisc/ccio-dma.c 	register unsigned long pa;
pa                570 drivers/parisc/ccio-dma.c 	pa = lpa(vba);
pa                571 drivers/parisc/ccio-dma.c 	asm volatile("depw  %1,31,12,%0" : "+r" (pa) : "r" (hints));
pa                572 drivers/parisc/ccio-dma.c 	((u32 *)pdir_ptr)[1] = (u32) pa;
pa                584 drivers/parisc/ccio-dma.c 	asm volatile ("extrd,u %1,15,4,%0" : "=r" (ci) : "r" (pa));
pa                585 drivers/parisc/ccio-dma.c 	asm volatile ("extrd,u %1,31,16,%0" : "+r" (pa) : "r" (pa));
pa                586 drivers/parisc/ccio-dma.c 	asm volatile ("depd  %1,35,4,%0" : "+r" (pa) : "r" (ci));
pa                588 drivers/parisc/ccio-dma.c 	pa = 0;
pa                597 drivers/parisc/ccio-dma.c 	asm volatile ("depw  %1,15,12,%0" : "+r" (pa) : "r" (ci));
pa                599 drivers/parisc/ccio-dma.c 	((u32 *)pdir_ptr)[0] = (u32) pa;
pa                569 drivers/parisc/sba_iommu.c 	u64 pa; /* physical address */
pa                572 drivers/parisc/sba_iommu.c 	pa = lpa(vba);
pa                573 drivers/parisc/sba_iommu.c 	pa &= IOVP_MASK;
pa                576 drivers/parisc/sba_iommu.c 	pa |= (ci >> PAGE_SHIFT) & 0xff;  /* move CI (8 bits) into lowest byte */
pa                578 drivers/parisc/sba_iommu.c 	pa |= SBA_PDIR_VALID_BIT;	/* set "valid" bit */
pa                579 drivers/parisc/sba_iommu.c 	*pdir_ptr = cpu_to_le64(pa);	/* swap and store into I/O Pdir */
pa               1527 drivers/pci/msi.c 	u32 *pa = data;
pa               1528 drivers/pci/msi.c 	u8 bus = PCI_BUS_NUM(*pa);
pa               1531 drivers/pci/msi.c 		*pa = alias;
pa                436 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-req-buf-alloc-all,		0x01),
pa                437 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-req-buf-alloc-rd,		0x02),
pa                438 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-req-buf-alloc-wr,		0x03),
pa                439 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-all-cp-req,			0x04),
pa                440 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-cp-blk-req,			0x05),
pa                441 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-cp-ptl-req,			0x06),
pa                442 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-cp-rd-req,			0x07),
pa                443 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-cp-wr-req,			0x08),
pa                447 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-rd-shared-req-issued,		0x10),
pa                448 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-rd-exclusive-req-issued,	0x11),
pa                449 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-wr-invalidate-req-issued-stashable, 0x12),
pa                450 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-wr-invalidate-req-issued-nonstashable, 0x13),
pa                451 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-wr-back-req-issued-stashable,	0x14),
pa                452 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-wr-back-req-issued-nonstashable, 0x15),
pa                453 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-ptl-wr-req,			0x16),
pa                454 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-ptl-rd-req,			0x17),
pa                455 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-wr-back-clean-data,		0x18),
pa                456 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-wr-back-cancelled-on-SS,	0x1b),
pa                457 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-barrier-occurrence,		0x1c),
pa                458 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-barrier-cycles,			0x1d),
pa                459 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-total-cp-snoops,		0x20),
pa                460 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-rd-shared-snoop,		0x21),
pa                461 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-rd-shared-snoop-hit,		0x22),
pa                462 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-rd-exclusive-snoop,		0x23),
pa                463 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-rd-exclusive-snoop-hit,		0x24),
pa                464 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-rd-wr-invalid-snoop,		0x25),
pa                465 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-rd-wr-invalid-snoop-hit,	0x26),
pa                466 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-req-buffer-full,		0x28),
pa                479 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-axi0-rd-req,			0x01),
pa                480 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-axi0-wr-req,			0x02),
pa                481 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-axi1-rd-req,			0x03),
pa                482 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pa-axi1-wr-req,			0x04),
pa                123 drivers/pinctrl/pinctrl-single.c 	struct pinctrl_pin_desc *pa;
pa                277 drivers/pinctrl/pinctrl-single.c 	size_t pa;
pa                284 drivers/pinctrl/pinctrl-single.c 	pa = pcs->res->start + offset;
pa                286 drivers/pinctrl/pinctrl-single.c 	seq_printf(s, "%zx %08x %s ", pa, val, DRIVER_NAME);
pa                683 drivers/pinctrl/pinctrl-single.c 	pin = &pcs->pins.pa[i];
pa                715 drivers/pinctrl/pinctrl-single.c 	pcs->pins.pa = devm_kcalloc(pcs->dev,
pa                716 drivers/pinctrl/pinctrl-single.c 				nr_pins, sizeof(*pcs->pins.pa),
pa                718 drivers/pinctrl/pinctrl-single.c 	if (!pcs->pins.pa)
pa                721 drivers/pinctrl/pinctrl-single.c 	pcs->desc.pins = pcs->pins.pa;
pa                152 drivers/pinctrl/ti/pinctrl-ti-iodelay.c 	struct pinctrl_pin_desc *pa;
pa                461 drivers/pinctrl/ti/pinctrl-ti-iodelay.c 	pd = &iod->pa[pin];
pa                666 drivers/pinctrl/ti/pinctrl-ti-iodelay.c 	pd = &iod->pa[pin];
pa                752 drivers/pinctrl/ti/pinctrl-ti-iodelay.c 	iod->pa = devm_kcalloc(dev, nr_pins, sizeof(*iod->pa), GFP_KERNEL);
pa                753 drivers/pinctrl/ti/pinctrl-ti-iodelay.c 	if (!iod->pa)
pa                756 drivers/pinctrl/ti/pinctrl-ti-iodelay.c 	iod->desc.pins = iod->pa;
pa                762 drivers/pinctrl/ti/pinctrl-ti-iodelay.c 		pin = &iod->pa[i];
pa                703 drivers/remoteproc/remoteproc_core.c 	ret = iommu_map(rproc->domain, rsc->da, rsc->pa, rsc->len, rsc->flags);
pa                721 drivers/remoteproc/remoteproc_core.c 		rsc->pa, rsc->da, rsc->len);
pa                890 drivers/remoteproc/remoteproc_core.c 		rsc->name, rsc->da, rsc->pa, rsc->len, rsc->flags);
pa               1177 drivers/remoteproc/remoteproc_core.c 	u64 pa;
pa               1214 drivers/remoteproc/remoteproc_core.c 				pa = (u64)rproc_va_to_pa(entry->va);
pa               1216 drivers/remoteproc/remoteproc_core.c 				pa = (u64)entry->dma;
pa               1218 drivers/remoteproc/remoteproc_core.c 			if (((u64)pa) & HIGH_BITS_MASK)
pa               1222 drivers/remoteproc/remoteproc_core.c 			rsc->pa = (u32)pa;
pa                214 drivers/remoteproc/remoteproc_debugfs.c 			seq_printf(seq, "  Physical Address 0x%x\n", c->pa);
pa                224 drivers/remoteproc/remoteproc_debugfs.c 			seq_printf(seq, "  Physical Address 0x%x\n", d->pa);
pa                259 drivers/remoteproc/remoteproc_debugfs.c 					   v->vring[j].pa);
pa                347 drivers/remoteproc/remoteproc_virtio.c 		phys_addr_t pa;
pa                363 drivers/remoteproc/remoteproc_virtio.c 				pa = rproc_va_to_pa(mem->va);
pa                366 drivers/remoteproc/remoteproc_virtio.c 				pa = (phys_addr_t)mem->dma;
pa                370 drivers/remoteproc/remoteproc_virtio.c 			ret = dma_declare_coherent_memory(dev, pa,
pa                 74 drivers/remoteproc/stm32_rproc.c static int stm32_rproc_pa_to_da(struct rproc *rproc, phys_addr_t pa, u64 *da)
pa                 83 drivers/remoteproc/stm32_rproc.c 		if (pa < p_mem->bus_addr ||
pa                 84 drivers/remoteproc/stm32_rproc.c 		    pa >= p_mem->bus_addr + p_mem->size)
pa                 86 drivers/remoteproc/stm32_rproc.c 		*da = pa - p_mem->bus_addr + p_mem->dev_addr;
pa                 87 drivers/remoteproc/stm32_rproc.c 		dev_dbg(rproc->dev.parent, "pa %pa to da %llx\n", &pa, *da);
pa                 58 drivers/s390/cio/vfio_ccw_cp.c static int pfn_array_alloc(struct pfn_array *pa, u64 iova, unsigned int len)
pa                 62 drivers/s390/cio/vfio_ccw_cp.c 	if (pa->pa_nr || pa->pa_iova_pfn)
pa                 65 drivers/s390/cio/vfio_ccw_cp.c 	pa->pa_iova = iova;
pa                 67 drivers/s390/cio/vfio_ccw_cp.c 	pa->pa_nr = ((iova & ~PAGE_MASK) + len + (PAGE_SIZE - 1)) >> PAGE_SHIFT;
pa                 68 drivers/s390/cio/vfio_ccw_cp.c 	if (!pa->pa_nr)
pa                 71 drivers/s390/cio/vfio_ccw_cp.c 	pa->pa_iova_pfn = kcalloc(pa->pa_nr,
pa                 72 drivers/s390/cio/vfio_ccw_cp.c 				  sizeof(*pa->pa_iova_pfn) +
pa                 73 drivers/s390/cio/vfio_ccw_cp.c 				  sizeof(*pa->pa_pfn),
pa                 75 drivers/s390/cio/vfio_ccw_cp.c 	if (unlikely(!pa->pa_iova_pfn)) {
pa                 76 drivers/s390/cio/vfio_ccw_cp.c 		pa->pa_nr = 0;
pa                 79 drivers/s390/cio/vfio_ccw_cp.c 	pa->pa_pfn = pa->pa_iova_pfn + pa->pa_nr;
pa                 81 drivers/s390/cio/vfio_ccw_cp.c 	pa->pa_iova_pfn[0] = pa->pa_iova >> PAGE_SHIFT;
pa                 82 drivers/s390/cio/vfio_ccw_cp.c 	pa->pa_pfn[0] = -1ULL;
pa                 83 drivers/s390/cio/vfio_ccw_cp.c 	for (i = 1; i < pa->pa_nr; i++) {
pa                 84 drivers/s390/cio/vfio_ccw_cp.c 		pa->pa_iova_pfn[i] = pa->pa_iova_pfn[i - 1] + 1;
pa                 85 drivers/s390/cio/vfio_ccw_cp.c 		pa->pa_pfn[i] = -1ULL;
pa                100 drivers/s390/cio/vfio_ccw_cp.c static int pfn_array_pin(struct pfn_array *pa, struct device *mdev)
pa                104 drivers/s390/cio/vfio_ccw_cp.c 	ret = vfio_pin_pages(mdev, pa->pa_iova_pfn, pa->pa_nr,
pa                105 drivers/s390/cio/vfio_ccw_cp.c 			     IOMMU_READ | IOMMU_WRITE, pa->pa_pfn);
pa                109 drivers/s390/cio/vfio_ccw_cp.c 	} else if (ret > 0 && ret != pa->pa_nr) {
pa                110 drivers/s390/cio/vfio_ccw_cp.c 		vfio_unpin_pages(mdev, pa->pa_iova_pfn, ret);
pa                118 drivers/s390/cio/vfio_ccw_cp.c 	pa->pa_nr = 0;
pa                124 drivers/s390/cio/vfio_ccw_cp.c static void pfn_array_unpin_free(struct pfn_array *pa, struct device *mdev)
pa                127 drivers/s390/cio/vfio_ccw_cp.c 	if (pa->pa_nr)
pa                128 drivers/s390/cio/vfio_ccw_cp.c 		vfio_unpin_pages(mdev, pa->pa_iova_pfn, pa->pa_nr);
pa                129 drivers/s390/cio/vfio_ccw_cp.c 	pa->pa_nr = 0;
pa                130 drivers/s390/cio/vfio_ccw_cp.c 	kfree(pa->pa_iova_pfn);
pa                133 drivers/s390/cio/vfio_ccw_cp.c static bool pfn_array_iova_pinned(struct pfn_array *pa, unsigned long iova)
pa                138 drivers/s390/cio/vfio_ccw_cp.c 	for (i = 0; i < pa->pa_nr; i++)
pa                139 drivers/s390/cio/vfio_ccw_cp.c 		if (pa->pa_iova_pfn[i] == iova_pfn)
pa                146 drivers/s390/cio/vfio_ccw_cp.c 	struct pfn_array *pa,
pa                159 drivers/s390/cio/vfio_ccw_cp.c 	for (i = 0; i < pa->pa_nr; i++)
pa                160 drivers/s390/cio/vfio_ccw_cp.c 		idaws[i] = pa->pa_pfn[i] << PAGE_SHIFT;
pa                163 drivers/s390/cio/vfio_ccw_cp.c 	idaws[0] += pa->pa_iova & (PAGE_SIZE - 1);
pa                196 drivers/s390/cio/vfio_ccw_cp.c 	struct pfn_array pa = {0};
pa                201 drivers/s390/cio/vfio_ccw_cp.c 	ret = pfn_array_alloc(&pa, iova, n);
pa                205 drivers/s390/cio/vfio_ccw_cp.c 	ret = pfn_array_pin(&pa, mdev);
pa                207 drivers/s390/cio/vfio_ccw_cp.c 		pfn_array_unpin_free(&pa, mdev);
pa                212 drivers/s390/cio/vfio_ccw_cp.c 	for (i = 0; i < pa.pa_nr; i++) {
pa                213 drivers/s390/cio/vfio_ccw_cp.c 		from = pa.pa_pfn[i] << PAGE_SHIFT;
pa                228 drivers/s390/cio/vfio_ccw_cp.c 	pfn_array_unpin_free(&pa, mdev);
pa                511 drivers/s390/cio/vfio_ccw_cp.c 	struct pfn_array *pa;
pa                550 drivers/s390/cio/vfio_ccw_cp.c 	pa = chain->ch_pa + idx;
pa                551 drivers/s390/cio/vfio_ccw_cp.c 	ret = pfn_array_alloc(pa, iova, bytes);
pa                566 drivers/s390/cio/vfio_ccw_cp.c 			pa->pa_iova_pfn[i] = idaws[i] >> PAGE_SHIFT;
pa                576 drivers/s390/cio/vfio_ccw_cp.c 		ret = pfn_array_pin(pa, cp->mdev);
pa                580 drivers/s390/cio/vfio_ccw_cp.c 		pa->pa_nr = 0;
pa                587 drivers/s390/cio/vfio_ccw_cp.c 	pfn_array_idal_create_words(pa, idaws);
pa                592 drivers/s390/cio/vfio_ccw_cp.c 	pfn_array_unpin_free(pa, cp->mdev);
pa               1544 drivers/scsi/be2iscsi/be_main.c 	if (pasync_handle->pa.u.a64.address != phys_addr.u.a64.address ||
pa               1551 drivers/scsi/be2iscsi/be_main.c 			    pasync_handle->pa.u.a64.address,
pa               1744 drivers/scsi/be2iscsi/be_main.c 			pasync_sge[pi].hi = pasync_handle->pa.u.a32.address_lo;
pa               1745 drivers/scsi/be2iscsi/be_main.c 			pasync_sge[pi].lo = pasync_handle->pa.u.a32.address_hi;
pa               2880 drivers/scsi/be2iscsi/be_main.c 				pasync_header_h->pa.u.a64.address =
pa               2915 drivers/scsi/be2iscsi/be_main.c 				pasync_data_h->pa.u.a64.address =
pa                545 drivers/scsi/be2iscsi/be_main.h 	struct be_bus_address pa;
pa                143 drivers/scsi/be2iscsi/be_mgmt.h #define ISCSI_GET_PDU_TEMPLATE_ADDRESS(pc, pa) {\
pa                144 drivers/scsi/be2iscsi/be_mgmt.h 	pa->lo = phba->init_mem[ISCSI_MEM_GLOBAL_HEADER].mem_array[0].\
pa                146 drivers/scsi/be2iscsi/be_mgmt.h 	pa->hi = phba->init_mem[ISCSI_MEM_GLOBAL_HEADER].mem_array[0].\
pa                997 drivers/scsi/bfa/bfa_core.c 	bfa_dma_be_addr_set(cfg_info->cfgrsp_addr, iocfc->cfgrsp_dma.pa);
pa               1003 drivers/scsi/bfa/bfa_core.c 				    iocfc->req_cq_ba[i].pa);
pa               1005 drivers/scsi/bfa/bfa_core.c 				    iocfc->req_cq_shadow_ci[i].pa);
pa               1010 drivers/scsi/bfa/bfa_core.c 				    iocfc->rsp_cq_ba[i].pa);
pa               1012 drivers/scsi/bfa/bfa_core.c 				    iocfc->rsp_cq_shadow_pi[i].pa);
pa               1029 drivers/scsi/bfa/bfa_core.c 	bfa_dma_be_addr_set(cfg_req.ioc_cfg_dma_addr, iocfc->cfg_info.pa);
pa               1112 drivers/scsi/bfa/bfa_core.c 		iocfc->req_cq_ba[i].pa = bfa_mem_dma_phys(reqq_dma);
pa               1117 drivers/scsi/bfa/bfa_core.c 		iocfc->rsp_cq_ba[i].pa = bfa_mem_dma_phys(rspq_dma);
pa               1127 drivers/scsi/bfa/bfa_core.c 		iocfc->req_cq_shadow_ci[i].pa = dm_pa;
pa               1132 drivers/scsi/bfa/bfa_core.c 		iocfc->rsp_cq_shadow_pi[i].pa = dm_pa;
pa               1139 drivers/scsi/bfa/bfa_core.c 	bfa->iocfc.cfg_info.pa = dm_pa;
pa               1146 drivers/scsi/bfa/bfa_core.c 	bfa->iocfc.cfgrsp_dma.pa = dm_pa;
pa               3713 drivers/scsi/bfa/bfa_fcpim.c 		fcp->snsbase[idx].pa = bfa_mem_dma_phys(seg_ptr);
pa               3715 drivers/scsi/bfa/bfa_fcpim.c 		bfa_iocfc_set_snsbase(bfa, idx, fcp->snsbase[idx].pa);
pa               1829 drivers/scsi/bfa/bfa_ioc.c 	bfa_dma_be_addr_set(attr_req.attr_addr, ioc->attr_dma.pa);
pa               2490 drivers/scsi/bfa/bfa_ioc.c 	ioc->attr_dma.pa = dm_pa;
pa               3372 drivers/scsi/bfa/bfa_ioc.c 	ablk->dma_addr.pa  = dma_pa;
pa               3412 drivers/scsi/bfa/bfa_ioc.c 	bfa_dma_be_addr_set(m->addr, ablk->dma_addr.pa);
pa                163 drivers/scsi/bfa/bfa_ioc.h 	u64	pa;	/* ! Physical address		*/
pa                175 drivers/scsi/bfa/bfa_ioc.h #define bfa_dma_be_addr_set(dma_addr, pa)	\
pa                176 drivers/scsi/bfa/bfa_ioc.h 		__bfa_dma_be_addr_set(&dma_addr, (u64)pa)
pa                178 drivers/scsi/bfa/bfa_ioc.h __bfa_dma_be_addr_set(union bfi_addr_u *dma_addr, u64 pa)
pa                180 drivers/scsi/bfa/bfa_ioc.h 	dma_addr->a32.addr_lo = cpu_to_be32(pa);
pa                181 drivers/scsi/bfa/bfa_ioc.h 	dma_addr->a32.addr_hi = cpu_to_be32(pa >> 32);
pa                188 drivers/scsi/bfa/bfa_ioc.h __bfa_alen_set(struct bfi_alen_s *alen, u32 len, u64 pa)
pa                191 drivers/scsi/bfa/bfa_ioc.h 	bfa_dma_be_addr_set(alen->al_addr, pa);
pa                206 drivers/scsi/bfa/bfa_port.c 	port->stats_dma.pa  = dma_pa;
pa                352 drivers/scsi/bfa/bfa_port.c 	bfa_dma_be_addr_set(m->dma_addr, port->stats_dma.pa);
pa                623 drivers/scsi/bfa/bfa_port.c 	cee->attr_dma.pa = dma_pa;
pa                626 drivers/scsi/bfa/bfa_port.c 	cee->stats_dma.pa = dma_pa + BFA_ROUNDUP(
pa                667 drivers/scsi/bfa/bfa_port.c 	bfa_dma_be_addr_set(cmd->dma_addr, cee->attr_dma.pa);
pa                707 drivers/scsi/bfa/bfa_port.c 	bfa_dma_be_addr_set(cmd->dma_addr, cee->stats_dma.pa);
pa               5191 drivers/scsi/bfa/bfa_svc.c 		u64 pa;
pa               5218 drivers/scsi/bfa/bfa_svc.c 		sgpg_pa.pa = bfa_mem_dma_phys(seg_ptr) + align_len;
pa               5219 drivers/scsi/bfa/bfa_svc.c 		WARN_ON(sgpg_pa.pa & (sgpg_sz - 1));
pa               5228 drivers/scsi/bfa/bfa_svc.c 			sgpg_pa_tmp.pa = bfa_sgaddr_le(sgpg_pa.pa);
pa               5234 drivers/scsi/bfa/bfa_svc.c 			sgpg_pa.pa += sgpg_sz;
pa                982 drivers/scsi/fnic/fnic_fcs.c 	dma_addr_t pa;
pa                996 drivers/scsi/fnic/fnic_fcs.c 	pa = dma_map_single(&fnic->pdev->dev, skb->data, len, DMA_FROM_DEVICE);
pa                997 drivers/scsi/fnic/fnic_fcs.c 	if (dma_mapping_error(&fnic->pdev->dev, pa)) {
pa               1003 drivers/scsi/fnic/fnic_fcs.c 	fnic_queue_rq_desc(rq, skb, pa, len);
pa               1032 drivers/scsi/fnic/fnic_fcs.c 	dma_addr_t pa;
pa               1055 drivers/scsi/fnic/fnic_fcs.c 	pa = dma_map_single(&fnic->pdev->dev, skb->data, skb->len,
pa               1057 drivers/scsi/fnic/fnic_fcs.c 	if (dma_mapping_error(&fnic->pdev->dev, pa)) {
pa               1066 drivers/scsi/fnic/fnic_fcs.c 	fnic_queue_wq_eth_desc(wq, skb, pa, skb->len,
pa               1074 drivers/scsi/fnic/fnic_fcs.c 	dma_unmap_single(&fnic->pdev->dev, pa, skb->len, DMA_TO_DEVICE);
pa               1086 drivers/scsi/fnic/fnic_fcs.c 	dma_addr_t pa;
pa               1131 drivers/scsi/fnic/fnic_fcs.c 	pa = dma_map_single(&fnic->pdev->dev, eth_hdr, tot_len, DMA_TO_DEVICE);
pa               1132 drivers/scsi/fnic/fnic_fcs.c 	if (dma_mapping_error(&fnic->pdev->dev, pa)) {
pa               1146 drivers/scsi/fnic/fnic_fcs.c 		dma_unmap_single(&fnic->pdev->dev, pa, tot_len, DMA_TO_DEVICE);
pa               1151 drivers/scsi/fnic/fnic_fcs.c 	fnic_queue_wq_desc(wq, skb, pa, tot_len, fr_eof(fp),
pa               3195 drivers/scsi/lpfc/lpfc_hbadisc.c 					} pa;
pa               3215 drivers/scsi/lpfc/lpfc_hbadisc.c 							un.pa.wd1, un.pa.wd2,
pa               3216 drivers/scsi/lpfc/lpfc_hbadisc.c 							un.pa.wd3, un.pa.wd4);
pa                 85 drivers/scsi/snic/snic_disc.c 	dma_addr_t pa = 0;
pa                114 drivers/scsi/snic/snic_disc.c 	pa = dma_map_single(&snic->pdev->dev, buf, buf_len, DMA_FROM_DEVICE);
pa                115 drivers/scsi/snic/snic_disc.c 	if (dma_mapping_error(&snic->pdev->dev, pa)) {
pa                127 drivers/scsi/snic/snic_disc.c 	SNIC_BUG_ON(pa == 0);
pa                134 drivers/scsi/snic/snic_disc.c 			     pa,
pa                141 drivers/scsi/snic/snic_disc.c 		dma_unmap_single(&snic->pdev->dev, pa, buf_len,
pa                165 drivers/scsi/snic/snic_io.c 	dma_addr_t pa = 0;
pa                176 drivers/scsi/snic/snic_io.c 	pa = dma_map_single(&snic->pdev->dev, os_buf, len, DMA_TO_DEVICE);
pa                177 drivers/scsi/snic/snic_io.c 	if (dma_mapping_error(&snic->pdev->dev, pa)) {
pa                183 drivers/scsi/snic/snic_io.c 	req->req_pa = (ulong)pa;
pa                190 drivers/scsi/snic/snic_io.c 		dma_unmap_single(&snic->pdev->dev, pa, len, DMA_TO_DEVICE);
pa                199 drivers/scsi/snic/snic_io.c 	snic_queue_wq_eth_desc(&snic->wq[q_num], os_buf, pa, len, 0, 0, 1);
pa                170 drivers/scsi/snic/snic_scsi.c 	dma_addr_t pa = 0;
pa                188 drivers/scsi/snic/snic_scsi.c 	pa = dma_map_single(&snic->pdev->dev,
pa                192 drivers/scsi/snic/snic_scsi.c 	if (dma_mapping_error(&snic->pdev->dev, pa)) {
pa                220 drivers/scsi/snic/snic_scsi.c 			pa, /* sense buffer pa */
pa                247 drivers/scsi/sym53c8xx_2/sym_fw.c 	u32 *pa;
pa                256 drivers/scsi/sym53c8xx_2/sym_fw.c 	pa = (u32 *) &np->fwa_bas;
pa                258 drivers/scsi/sym53c8xx_2/sym_fw.c 		pa[i] = np->scripta_ba + po[i];
pa                264 drivers/scsi/sym53c8xx_2/sym_fw.c 	pa = (u32 *) &np->fwb_bas;
pa                266 drivers/scsi/sym53c8xx_2/sym_fw.c 		pa[i] = np->scriptb_ba + po[i];
pa                272 drivers/scsi/sym53c8xx_2/sym_fw.c 	pa = (u32 *) &np->fwz_bas;
pa                274 drivers/scsi/sym53c8xx_2/sym_fw.c 		pa[i] = np->scriptz_ba + po[i];
pa                121 drivers/sfi/sfi_core.c static void sfi_print_table_header(unsigned long long pa,
pa                125 drivers/sfi/sfi_core.c 		header->sig, pa,
pa                165 drivers/sfi/sfi_core.c static struct sfi_table_header *sfi_map_table(u64 pa)
pa                170 drivers/sfi/sfi_core.c 	if (!TABLE_ON_PAGE(syst_pa, pa, sizeof(struct sfi_table_header)))
pa                171 drivers/sfi/sfi_core.c 		th = sfi_map_memory(pa, sizeof(struct sfi_table_header));
pa                173 drivers/sfi/sfi_core.c 		th = (void *)syst_va + (pa - syst_pa);
pa                181 drivers/sfi/sfi_core.c 	if (!TABLE_ON_PAGE(syst_pa, pa, sizeof(struct sfi_table_header)))
pa                184 drivers/sfi/sfi_core.c 	return sfi_map_memory(pa, length);
pa                232 drivers/sfi/sfi_core.c  __ref sfi_check_table(u64 pa, struct sfi_table_key *key)
pa                237 drivers/sfi/sfi_core.c 	th = sfi_map_table(pa);
pa                242 drivers/sfi/sfi_core.c 		sfi_print_table_header(pa, th);
pa                419 drivers/sfi/sfi_core.c struct sfi_table_attr __init *sfi_sysfs_install_table(u64 pa)
pa                429 drivers/sfi/sfi_core.c 	th = sfi_map_table(pa);
pa                 80 drivers/sfi/sfi_core.h extern struct sfi_table_attr __init *sfi_sysfs_install_table(u64 pa);
pa                178 drivers/staging/gdm724x/gdm_lte.c 	u16 pa;
pa                188 drivers/staging/gdm724x/gdm_lte.c 		u16 pa[20];
pa                198 drivers/staging/gdm724x/gdm_lte.c 	for (i = 0; i < ARRAY_SIZE(pseudo_header.pa); i++) {
pa                199 drivers/staging/gdm724x/gdm_lte.c 		pa = pseudo_header.pa[i];
pa                200 drivers/staging/gdm724x/gdm_lte.c 		sum = csum_add(sum, csum_unfold((__force __sum16)pa));
pa                 42 drivers/tee/optee/core.c 	phys_addr_t pa;
pa                 76 drivers/tee/optee/core.c 			rc = tee_shm_get_pa(shm, 0, &pa);
pa                 79 drivers/tee/optee/core.c 			p->u.memref.shm_offs = mp->u.tmem.buf_ptr - pa;
pa                122 drivers/tee/optee/core.c 	phys_addr_t pa;
pa                135 drivers/tee/optee/core.c 	rc = tee_shm_get_pa(p->u.memref.shm, p->u.memref.shm_offs, &pa);
pa                139 drivers/tee/optee/core.c 	mp->u.tmem.buf_ptr = pa;
pa                197 drivers/tee/optee/rpc.c 	phys_addr_t pa;
pa                235 drivers/tee/optee/rpc.c 	if (tee_shm_get_pa(shm, 0, &pa)) {
pa                278 drivers/tee/optee/rpc.c 		arg->params[0].u.tmem.buf_ptr = pa;
pa                404 drivers/tee/optee/rpc.c 	phys_addr_t pa;
pa                409 drivers/tee/optee/rpc.c 		if (!IS_ERR(shm) && !tee_shm_get_pa(shm, 0, &pa)) {
pa                410 drivers/tee/optee/rpc.c 			reg_pair_from_64(&param->a1, &param->a2, pa);
pa                383 drivers/tee/tee_shm.c int tee_shm_va2pa(struct tee_shm *shm, void *va, phys_addr_t *pa)
pa                394 drivers/tee/tee_shm.c 			shm, (unsigned long)va - (unsigned long)shm->kaddr, pa);
pa                405 drivers/tee/tee_shm.c int tee_shm_pa2va(struct tee_shm *shm, phys_addr_t pa, void **va)
pa                410 drivers/tee/tee_shm.c 	if (pa < shm->paddr)
pa                412 drivers/tee/tee_shm.c 	if (pa >= (shm->paddr + shm->size))
pa                416 drivers/tee/tee_shm.c 		void *v = tee_shm_get_va(shm, pa - shm->paddr);
pa                451 drivers/tee/tee_shm.c int tee_shm_get_pa(struct tee_shm *shm, size_t offs, phys_addr_t *pa)
pa                455 drivers/tee/tee_shm.c 	if (pa)
pa                456 drivers/tee/tee_shm.c 		*pa = shm->paddr + offs;
pa                127 drivers/tty/serial/8250/8250_hp300.c 		unsigned long pa = dio_scodetophysaddr(scode);
pa                128 drivers/tty/serial/8250/8250_hp300.c 		if (!pa)
pa                134 drivers/tty/serial/8250/8250_hp300.c 		port.mapbase = (pa + UART_OFFSET);
pa                137 drivers/tty/serial/8250/8250_hp300.c 		port.irq = DIO_IPL(pa + DIO_VIRADDRBASE);
pa                140 drivers/tty/serial/8250/8250_hp300.c 		out_8(pa + DIO_VIRADDRBASE + DCA_IC, DCA_IC_IE);
pa                142 drivers/tty/serial/8250/8250_hp300.c 		if (DIO_ID(pa + DIO_VIRADDRBASE) & 0x80)
pa                631 fs/ext4/mballoc.c 		struct ext4_prealloc_space *pa;
pa                632 fs/ext4/mballoc.c 		pa = list_entry(cur, struct ext4_prealloc_space, pa_group_list);
pa                633 fs/ext4/mballoc.c 		ext4_get_group_no_and_offset(sb, pa->pa_pstart, &groupnr, &k);
pa                635 fs/ext4/mballoc.c 		for (i = 0; i < pa->pa_len; i++)
pa               2700 fs/ext4/mballoc.c 	struct ext4_prealloc_space *pa;
pa               2705 fs/ext4/mballoc.c 		pa = list_entry(cur, struct ext4_prealloc_space, pa_group_list);
pa               2706 fs/ext4/mballoc.c 		list_del(&pa->pa_group_list);
pa               2708 fs/ext4/mballoc.c 		kmem_cache_free(ext4_pspace_cachep, pa);
pa               3092 fs/ext4/mballoc.c 	struct ext4_prealloc_space *pa;
pa               3186 fs/ext4/mballoc.c 	list_for_each_entry_rcu(pa, &ei->i_prealloc_list, pa_inode_list) {
pa               3189 fs/ext4/mballoc.c 		if (pa->pa_deleted)
pa               3191 fs/ext4/mballoc.c 		spin_lock(&pa->pa_lock);
pa               3192 fs/ext4/mballoc.c 		if (pa->pa_deleted) {
pa               3193 fs/ext4/mballoc.c 			spin_unlock(&pa->pa_lock);
pa               3197 fs/ext4/mballoc.c 		pa_end = pa->pa_lstart + EXT4_C2B(EXT4_SB(ac->ac_sb),
pa               3198 fs/ext4/mballoc.c 						  pa->pa_len);
pa               3202 fs/ext4/mballoc.c 			ac->ac_o_ex.fe_logical < pa->pa_lstart));
pa               3205 fs/ext4/mballoc.c 		if (pa->pa_lstart >= end || pa_end <= start) {
pa               3206 fs/ext4/mballoc.c 			spin_unlock(&pa->pa_lock);
pa               3209 fs/ext4/mballoc.c 		BUG_ON(pa->pa_lstart <= start && pa_end >= end);
pa               3215 fs/ext4/mballoc.c 		} else if (pa->pa_lstart > ac->ac_o_ex.fe_logical) {
pa               3216 fs/ext4/mballoc.c 			BUG_ON(pa->pa_lstart > end);
pa               3217 fs/ext4/mballoc.c 			end = pa->pa_lstart;
pa               3219 fs/ext4/mballoc.c 		spin_unlock(&pa->pa_lock);
pa               3226 fs/ext4/mballoc.c 	list_for_each_entry_rcu(pa, &ei->i_prealloc_list, pa_inode_list) {
pa               3229 fs/ext4/mballoc.c 		spin_lock(&pa->pa_lock);
pa               3230 fs/ext4/mballoc.c 		if (pa->pa_deleted == 0) {
pa               3231 fs/ext4/mballoc.c 			pa_end = pa->pa_lstart + EXT4_C2B(EXT4_SB(ac->ac_sb),
pa               3232 fs/ext4/mballoc.c 							  pa->pa_len);
pa               3233 fs/ext4/mballoc.c 			BUG_ON(!(start >= pa_end || end <= pa->pa_lstart));
pa               3235 fs/ext4/mballoc.c 		spin_unlock(&pa->pa_lock);
pa               3307 fs/ext4/mballoc.c 	struct ext4_prealloc_space *pa = ac->ac_pa;
pa               3311 fs/ext4/mballoc.c 	if (pa == NULL) {
pa               3331 fs/ext4/mballoc.c 	if (pa->pa_type == MB_INODE_PA)
pa               3332 fs/ext4/mballoc.c 		pa->pa_free += ac->ac_b_ex.fe_len;
pa               3339 fs/ext4/mballoc.c 				struct ext4_prealloc_space *pa)
pa               3347 fs/ext4/mballoc.c 	start = pa->pa_pstart + (ac->ac_o_ex.fe_logical - pa->pa_lstart);
pa               3348 fs/ext4/mballoc.c 	end = min(pa->pa_pstart + EXT4_C2B(sbi, pa->pa_len),
pa               3355 fs/ext4/mballoc.c 	ac->ac_pa = pa;
pa               3357 fs/ext4/mballoc.c 	BUG_ON(start < pa->pa_pstart);
pa               3358 fs/ext4/mballoc.c 	BUG_ON(end > pa->pa_pstart + EXT4_C2B(sbi, pa->pa_len));
pa               3359 fs/ext4/mballoc.c 	BUG_ON(pa->pa_free < len);
pa               3360 fs/ext4/mballoc.c 	pa->pa_free -= len;
pa               3362 fs/ext4/mballoc.c 	mb_debug(1, "use %llu/%u from inode pa %p\n", start, len, pa);
pa               3369 fs/ext4/mballoc.c 				struct ext4_prealloc_space *pa)
pa               3373 fs/ext4/mballoc.c 	ext4_get_group_no_and_offset(ac->ac_sb, pa->pa_pstart,
pa               3378 fs/ext4/mballoc.c 	ac->ac_pa = pa;
pa               3386 fs/ext4/mballoc.c 	mb_debug(1, "use %u/%u from group pa %p\n", pa->pa_lstart-len, len, pa);
pa               3397 fs/ext4/mballoc.c 			struct ext4_prealloc_space *pa,
pa               3403 fs/ext4/mballoc.c 		atomic_inc(&pa->pa_count);
pa               3404 fs/ext4/mballoc.c 		return pa;
pa               3407 fs/ext4/mballoc.c 	new_distance = abs(goal_block - pa->pa_pstart);
pa               3414 fs/ext4/mballoc.c 	atomic_inc(&pa->pa_count);
pa               3415 fs/ext4/mballoc.c 	return pa;
pa               3428 fs/ext4/mballoc.c 	struct ext4_prealloc_space *pa, *cpa = NULL;
pa               3437 fs/ext4/mballoc.c 	list_for_each_entry_rcu(pa, &ei->i_prealloc_list, pa_inode_list) {
pa               3441 fs/ext4/mballoc.c 		if (ac->ac_o_ex.fe_logical < pa->pa_lstart ||
pa               3442 fs/ext4/mballoc.c 		    ac->ac_o_ex.fe_logical >= (pa->pa_lstart +
pa               3443 fs/ext4/mballoc.c 					       EXT4_C2B(sbi, pa->pa_len)))
pa               3448 fs/ext4/mballoc.c 		    (pa->pa_pstart + EXT4_C2B(sbi, pa->pa_len) >
pa               3453 fs/ext4/mballoc.c 		spin_lock(&pa->pa_lock);
pa               3454 fs/ext4/mballoc.c 		if (pa->pa_deleted == 0 && pa->pa_free) {
pa               3455 fs/ext4/mballoc.c 			atomic_inc(&pa->pa_count);
pa               3456 fs/ext4/mballoc.c 			ext4_mb_use_inode_pa(ac, pa);
pa               3457 fs/ext4/mballoc.c 			spin_unlock(&pa->pa_lock);
pa               3462 fs/ext4/mballoc.c 		spin_unlock(&pa->pa_lock);
pa               3486 fs/ext4/mballoc.c 		list_for_each_entry_rcu(pa, &lg->lg_prealloc_list[i],
pa               3488 fs/ext4/mballoc.c 			spin_lock(&pa->pa_lock);
pa               3489 fs/ext4/mballoc.c 			if (pa->pa_deleted == 0 &&
pa               3490 fs/ext4/mballoc.c 					pa->pa_free >= ac->ac_o_ex.fe_len) {
pa               3493 fs/ext4/mballoc.c 								pa, cpa);
pa               3495 fs/ext4/mballoc.c 			spin_unlock(&pa->pa_lock);
pa               3541 fs/ext4/mballoc.c 	struct ext4_prealloc_space *pa;
pa               3557 fs/ext4/mballoc.c 		pa = list_entry(cur, struct ext4_prealloc_space, pa_group_list);
pa               3558 fs/ext4/mballoc.c 		spin_lock(&pa->pa_lock);
pa               3559 fs/ext4/mballoc.c 		ext4_get_group_no_and_offset(sb, pa->pa_pstart,
pa               3561 fs/ext4/mballoc.c 		len = pa->pa_len;
pa               3562 fs/ext4/mballoc.c 		spin_unlock(&pa->pa_lock);
pa               3574 fs/ext4/mballoc.c 	struct ext4_prealloc_space *pa;
pa               3575 fs/ext4/mballoc.c 	pa = container_of(head, struct ext4_prealloc_space, u.pa_rcu);
pa               3577 fs/ext4/mballoc.c 	BUG_ON(atomic_read(&pa->pa_count));
pa               3578 fs/ext4/mballoc.c 	BUG_ON(pa->pa_deleted == 0);
pa               3579 fs/ext4/mballoc.c 	kmem_cache_free(ext4_pspace_cachep, pa);
pa               3587 fs/ext4/mballoc.c 			struct super_block *sb, struct ext4_prealloc_space *pa)
pa               3593 fs/ext4/mballoc.c 	spin_lock(&pa->pa_lock);
pa               3594 fs/ext4/mballoc.c 	if (!atomic_dec_and_test(&pa->pa_count) || pa->pa_free != 0) {
pa               3595 fs/ext4/mballoc.c 		spin_unlock(&pa->pa_lock);
pa               3599 fs/ext4/mballoc.c 	if (pa->pa_deleted == 1) {
pa               3600 fs/ext4/mballoc.c 		spin_unlock(&pa->pa_lock);
pa               3604 fs/ext4/mballoc.c 	pa->pa_deleted = 1;
pa               3605 fs/ext4/mballoc.c 	spin_unlock(&pa->pa_lock);
pa               3607 fs/ext4/mballoc.c 	grp_blk = pa->pa_pstart;
pa               3612 fs/ext4/mballoc.c 	if (pa->pa_type == MB_GROUP_PA)
pa               3632 fs/ext4/mballoc.c 	list_del(&pa->pa_group_list);
pa               3635 fs/ext4/mballoc.c 	spin_lock(pa->pa_obj_lock);
pa               3636 fs/ext4/mballoc.c 	list_del_rcu(&pa->pa_inode_list);
pa               3637 fs/ext4/mballoc.c 	spin_unlock(pa->pa_obj_lock);
pa               3639 fs/ext4/mballoc.c 	call_rcu(&(pa)->u.pa_rcu, ext4_mb_pa_callback);
pa               3650 fs/ext4/mballoc.c 	struct ext4_prealloc_space *pa;
pa               3659 fs/ext4/mballoc.c 	pa = kmem_cache_alloc(ext4_pspace_cachep, GFP_NOFS);
pa               3660 fs/ext4/mballoc.c 	if (pa == NULL)
pa               3701 fs/ext4/mballoc.c 	pa->pa_lstart = ac->ac_b_ex.fe_logical;
pa               3702 fs/ext4/mballoc.c 	pa->pa_pstart = ext4_grp_offs_to_block(sb, &ac->ac_b_ex);
pa               3703 fs/ext4/mballoc.c 	pa->pa_len = ac->ac_b_ex.fe_len;
pa               3704 fs/ext4/mballoc.c 	pa->pa_free = pa->pa_len;
pa               3705 fs/ext4/mballoc.c 	atomic_set(&pa->pa_count, 1);
pa               3706 fs/ext4/mballoc.c 	spin_lock_init(&pa->pa_lock);
pa               3707 fs/ext4/mballoc.c 	INIT_LIST_HEAD(&pa->pa_inode_list);
pa               3708 fs/ext4/mballoc.c 	INIT_LIST_HEAD(&pa->pa_group_list);
pa               3709 fs/ext4/mballoc.c 	pa->pa_deleted = 0;
pa               3710 fs/ext4/mballoc.c 	pa->pa_type = MB_INODE_PA;
pa               3712 fs/ext4/mballoc.c 	mb_debug(1, "new inode pa %p: %llu/%u for %u\n", pa,
pa               3713 fs/ext4/mballoc.c 			pa->pa_pstart, pa->pa_len, pa->pa_lstart);
pa               3714 fs/ext4/mballoc.c 	trace_ext4_mb_new_inode_pa(ac, pa);
pa               3716 fs/ext4/mballoc.c 	ext4_mb_use_inode_pa(ac, pa);
pa               3717 fs/ext4/mballoc.c 	atomic_add(pa->pa_free, &sbi->s_mb_preallocated);
pa               3722 fs/ext4/mballoc.c 	pa->pa_obj_lock = &ei->i_prealloc_lock;
pa               3723 fs/ext4/mballoc.c 	pa->pa_inode = ac->ac_inode;
pa               3726 fs/ext4/mballoc.c 	list_add(&pa->pa_group_list, &grp->bb_prealloc_list);
pa               3729 fs/ext4/mballoc.c 	spin_lock(pa->pa_obj_lock);
pa               3730 fs/ext4/mballoc.c 	list_add_rcu(&pa->pa_inode_list, &ei->i_prealloc_list);
pa               3731 fs/ext4/mballoc.c 	spin_unlock(pa->pa_obj_lock);
pa               3744 fs/ext4/mballoc.c 	struct ext4_prealloc_space *pa;
pa               3753 fs/ext4/mballoc.c 	pa = kmem_cache_alloc(ext4_pspace_cachep, GFP_NOFS);
pa               3754 fs/ext4/mballoc.c 	if (pa == NULL)
pa               3761 fs/ext4/mballoc.c 	pa->pa_pstart = ext4_grp_offs_to_block(sb, &ac->ac_b_ex);
pa               3762 fs/ext4/mballoc.c 	pa->pa_lstart = pa->pa_pstart;
pa               3763 fs/ext4/mballoc.c 	pa->pa_len = ac->ac_b_ex.fe_len;
pa               3764 fs/ext4/mballoc.c 	pa->pa_free = pa->pa_len;
pa               3765 fs/ext4/mballoc.c 	atomic_set(&pa->pa_count, 1);
pa               3766 fs/ext4/mballoc.c 	spin_lock_init(&pa->pa_lock);
pa               3767 fs/ext4/mballoc.c 	INIT_LIST_HEAD(&pa->pa_inode_list);
pa               3768 fs/ext4/mballoc.c 	INIT_LIST_HEAD(&pa->pa_group_list);
pa               3769 fs/ext4/mballoc.c 	pa->pa_deleted = 0;
pa               3770 fs/ext4/mballoc.c 	pa->pa_type = MB_GROUP_PA;
pa               3772 fs/ext4/mballoc.c 	mb_debug(1, "new group pa %p: %llu/%u for %u\n", pa,
pa               3773 fs/ext4/mballoc.c 			pa->pa_pstart, pa->pa_len, pa->pa_lstart);
pa               3774 fs/ext4/mballoc.c 	trace_ext4_mb_new_group_pa(ac, pa);
pa               3776 fs/ext4/mballoc.c 	ext4_mb_use_group_pa(ac, pa);
pa               3777 fs/ext4/mballoc.c 	atomic_add(pa->pa_free, &EXT4_SB(sb)->s_mb_preallocated);
pa               3783 fs/ext4/mballoc.c 	pa->pa_obj_lock = &lg->lg_prealloc_lock;
pa               3784 fs/ext4/mballoc.c 	pa->pa_inode = NULL;
pa               3787 fs/ext4/mballoc.c 	list_add(&pa->pa_group_list, &grp->bb_prealloc_list);
pa               3818 fs/ext4/mballoc.c 			struct ext4_prealloc_space *pa)
pa               3829 fs/ext4/mballoc.c 	BUG_ON(pa->pa_deleted == 0);
pa               3830 fs/ext4/mballoc.c 	ext4_get_group_no_and_offset(sb, pa->pa_pstart, &group, &bit);
pa               3831 fs/ext4/mballoc.c 	grp_blk_start = pa->pa_pstart - EXT4_C2B(sbi, bit);
pa               3832 fs/ext4/mballoc.c 	BUG_ON(group != e4b->bd_group && pa->pa_len != 0);
pa               3833 fs/ext4/mballoc.c 	end = bit + pa->pa_len;
pa               3846 fs/ext4/mballoc.c 		trace_ext4_mb_release_inode_pa(pa, (grp_blk_start +
pa               3849 fs/ext4/mballoc.c 		mb_free_blocks(pa->pa_inode, e4b, bit, next - bit);
pa               3852 fs/ext4/mballoc.c 	if (free != pa->pa_free) {
pa               3855 fs/ext4/mballoc.c 			 pa, (unsigned long) pa->pa_lstart,
pa               3856 fs/ext4/mballoc.c 			 (unsigned long) pa->pa_pstart,
pa               3857 fs/ext4/mballoc.c 			 (unsigned long) pa->pa_len);
pa               3859 fs/ext4/mballoc.c 					free, pa->pa_free);
pa               3872 fs/ext4/mballoc.c 				struct ext4_prealloc_space *pa)
pa               3878 fs/ext4/mballoc.c 	trace_ext4_mb_release_group_pa(sb, pa);
pa               3879 fs/ext4/mballoc.c 	BUG_ON(pa->pa_deleted == 0);
pa               3880 fs/ext4/mballoc.c 	ext4_get_group_no_and_offset(sb, pa->pa_pstart, &group, &bit);
pa               3881 fs/ext4/mballoc.c 	BUG_ON(group != e4b->bd_group && pa->pa_len != 0);
pa               3882 fs/ext4/mballoc.c 	mb_free_blocks(pa->pa_inode, e4b, bit, pa->pa_len);
pa               3883 fs/ext4/mballoc.c 	atomic_add(pa->pa_len, &EXT4_SB(sb)->s_mb_discarded);
pa               3884 fs/ext4/mballoc.c 	trace_ext4_mballoc_discard(sb, NULL, group, bit, pa->pa_len);
pa               3904 fs/ext4/mballoc.c 	struct ext4_prealloc_space *pa, *tmp;
pa               3938 fs/ext4/mballoc.c 	list_for_each_entry_safe(pa, tmp,
pa               3940 fs/ext4/mballoc.c 		spin_lock(&pa->pa_lock);
pa               3941 fs/ext4/mballoc.c 		if (atomic_read(&pa->pa_count)) {
pa               3942 fs/ext4/mballoc.c 			spin_unlock(&pa->pa_lock);
pa               3946 fs/ext4/mballoc.c 		if (pa->pa_deleted) {
pa               3947 fs/ext4/mballoc.c 			spin_unlock(&pa->pa_lock);
pa               3952 fs/ext4/mballoc.c 		pa->pa_deleted = 1;
pa               3955 fs/ext4/mballoc.c 		free += pa->pa_free;
pa               3957 fs/ext4/mballoc.c 		spin_unlock(&pa->pa_lock);
pa               3959 fs/ext4/mballoc.c 		list_del(&pa->pa_group_list);
pa               3960 fs/ext4/mballoc.c 		list_add(&pa->u.pa_tmp_list, &list);
pa               3978 fs/ext4/mballoc.c 	list_for_each_entry_safe(pa, tmp, &list, u.pa_tmp_list) {
pa               3981 fs/ext4/mballoc.c 		spin_lock(pa->pa_obj_lock);
pa               3982 fs/ext4/mballoc.c 		list_del_rcu(&pa->pa_inode_list);
pa               3983 fs/ext4/mballoc.c 		spin_unlock(pa->pa_obj_lock);
pa               3985 fs/ext4/mballoc.c 		if (pa->pa_type == MB_GROUP_PA)
pa               3986 fs/ext4/mballoc.c 			ext4_mb_release_group_pa(&e4b, pa);
pa               3988 fs/ext4/mballoc.c 			ext4_mb_release_inode_pa(&e4b, bitmap_bh, pa);
pa               3990 fs/ext4/mballoc.c 		list_del(&pa->u.pa_tmp_list);
pa               3991 fs/ext4/mballoc.c 		call_rcu(&(pa)->u.pa_rcu, ext4_mb_pa_callback);
pa               4015 fs/ext4/mballoc.c 	struct ext4_prealloc_space *pa, *tmp;
pa               4035 fs/ext4/mballoc.c 		pa = list_entry(ei->i_prealloc_list.next,
pa               4037 fs/ext4/mballoc.c 		BUG_ON(pa->pa_obj_lock != &ei->i_prealloc_lock);
pa               4038 fs/ext4/mballoc.c 		spin_lock(&pa->pa_lock);
pa               4039 fs/ext4/mballoc.c 		if (atomic_read(&pa->pa_count)) {
pa               4042 fs/ext4/mballoc.c 			spin_unlock(&pa->pa_lock);
pa               4051 fs/ext4/mballoc.c 		if (pa->pa_deleted == 0) {
pa               4052 fs/ext4/mballoc.c 			pa->pa_deleted = 1;
pa               4053 fs/ext4/mballoc.c 			spin_unlock(&pa->pa_lock);
pa               4054 fs/ext4/mballoc.c 			list_del_rcu(&pa->pa_inode_list);
pa               4055 fs/ext4/mballoc.c 			list_add(&pa->u.pa_tmp_list, &list);
pa               4060 fs/ext4/mballoc.c 		spin_unlock(&pa->pa_lock);
pa               4080 fs/ext4/mballoc.c 	list_for_each_entry_safe(pa, tmp, &list, u.pa_tmp_list) {
pa               4081 fs/ext4/mballoc.c 		BUG_ON(pa->pa_type != MB_INODE_PA);
pa               4082 fs/ext4/mballoc.c 		group = ext4_get_group_number(sb, pa->pa_pstart);
pa               4102 fs/ext4/mballoc.c 		list_del(&pa->pa_group_list);
pa               4103 fs/ext4/mballoc.c 		ext4_mb_release_inode_pa(&e4b, bitmap_bh, pa);
pa               4109 fs/ext4/mballoc.c 		list_del(&pa->u.pa_tmp_list);
pa               4110 fs/ext4/mballoc.c 		call_rcu(&(pa)->u.pa_rcu, ext4_mb_pa_callback);
pa               4149 fs/ext4/mballoc.c 		struct ext4_prealloc_space *pa;
pa               4154 fs/ext4/mballoc.c 			pa = list_entry(cur, struct ext4_prealloc_space,
pa               4156 fs/ext4/mballoc.c 			spin_lock(&pa->pa_lock);
pa               4157 fs/ext4/mballoc.c 			ext4_get_group_no_and_offset(sb, pa->pa_pstart,
pa               4159 fs/ext4/mballoc.c 			spin_unlock(&pa->pa_lock);
pa               4161 fs/ext4/mballoc.c 			       start, pa->pa_len);
pa               4296 fs/ext4/mballoc.c 	struct ext4_prealloc_space *pa, *tmp;
pa               4303 fs/ext4/mballoc.c 	list_for_each_entry_rcu(pa, &lg->lg_prealloc_list[order],
pa               4305 fs/ext4/mballoc.c 		spin_lock(&pa->pa_lock);
pa               4306 fs/ext4/mballoc.c 		if (atomic_read(&pa->pa_count)) {
pa               4312 fs/ext4/mballoc.c 			spin_unlock(&pa->pa_lock);
pa               4315 fs/ext4/mballoc.c 		if (pa->pa_deleted) {
pa               4316 fs/ext4/mballoc.c 			spin_unlock(&pa->pa_lock);
pa               4320 fs/ext4/mballoc.c 		BUG_ON(pa->pa_type != MB_GROUP_PA);
pa               4323 fs/ext4/mballoc.c 		pa->pa_deleted = 1;
pa               4324 fs/ext4/mballoc.c 		spin_unlock(&pa->pa_lock);
pa               4326 fs/ext4/mballoc.c 		list_del_rcu(&pa->pa_inode_list);
pa               4327 fs/ext4/mballoc.c 		list_add(&pa->u.pa_tmp_list, &discard_list);
pa               4342 fs/ext4/mballoc.c 	list_for_each_entry_safe(pa, tmp, &discard_list, u.pa_tmp_list) {
pa               4345 fs/ext4/mballoc.c 		group = ext4_get_group_number(sb, pa->pa_pstart);
pa               4354 fs/ext4/mballoc.c 		list_del(&pa->pa_group_list);
pa               4355 fs/ext4/mballoc.c 		ext4_mb_release_group_pa(&e4b, pa);
pa               4359 fs/ext4/mballoc.c 		list_del(&pa->u.pa_tmp_list);
pa               4360 fs/ext4/mballoc.c 		call_rcu(&(pa)->u.pa_rcu, ext4_mb_pa_callback);
pa               4378 fs/ext4/mballoc.c 	struct ext4_prealloc_space *tmp_pa, *pa = ac->ac_pa;
pa               4380 fs/ext4/mballoc.c 	order = fls(pa->pa_free) - 1;
pa               4393 fs/ext4/mballoc.c 		if (!added && pa->pa_free < tmp_pa->pa_free) {
pa               4395 fs/ext4/mballoc.c 			list_add_tail_rcu(&pa->pa_inode_list,
pa               4407 fs/ext4/mballoc.c 		list_add_tail_rcu(&pa->pa_inode_list,
pa               4426 fs/ext4/mballoc.c 	struct ext4_prealloc_space *pa = ac->ac_pa;
pa               4427 fs/ext4/mballoc.c 	if (pa) {
pa               4428 fs/ext4/mballoc.c 		if (pa->pa_type == MB_GROUP_PA) {
pa               4430 fs/ext4/mballoc.c 			spin_lock(&pa->pa_lock);
pa               4431 fs/ext4/mballoc.c 			pa->pa_pstart += EXT4_C2B(sbi, ac->ac_b_ex.fe_len);
pa               4432 fs/ext4/mballoc.c 			pa->pa_lstart += EXT4_C2B(sbi, ac->ac_b_ex.fe_len);
pa               4433 fs/ext4/mballoc.c 			pa->pa_free -= ac->ac_b_ex.fe_len;
pa               4434 fs/ext4/mballoc.c 			pa->pa_len -= ac->ac_b_ex.fe_len;
pa               4435 fs/ext4/mballoc.c 			spin_unlock(&pa->pa_lock);
pa               4438 fs/ext4/mballoc.c 	if (pa) {
pa               4445 fs/ext4/mballoc.c 		if ((pa->pa_type == MB_GROUP_PA) && likely(pa->pa_free)) {
pa               4446 fs/ext4/mballoc.c 			spin_lock(pa->pa_obj_lock);
pa               4447 fs/ext4/mballoc.c 			list_del_rcu(&pa->pa_inode_list);
pa               4448 fs/ext4/mballoc.c 			spin_unlock(pa->pa_obj_lock);
pa               4451 fs/ext4/mballoc.c 		ext4_mb_put_pa(ac, ac->ac_sb, pa);
pa                280 fs/f2fs/acl.c  	struct posix_acl_entry *pa, *pe;
pa                287 fs/f2fs/acl.c  	FOREACH_ACL_ENTRY(pa, acl, pe) {
pa                288 fs/f2fs/acl.c  		switch (pa->e_tag) {
pa                290 fs/f2fs/acl.c  			pa->e_perm &= (mode >> 6) | ~S_IRWXO;
pa                291 fs/f2fs/acl.c  			mode &= (pa->e_perm << 6) | ~S_IRWXU;
pa                300 fs/f2fs/acl.c  			group_obj = pa;
pa                304 fs/f2fs/acl.c  			pa->e_perm &= mode | ~S_IRWXO;
pa                305 fs/f2fs/acl.c  			mode &= pa->e_perm | ~S_IRWXO;
pa                309 fs/f2fs/acl.c  			mask_obj = pa;
pa                221 fs/nfs_common/nfsacl.c 	struct posix_acl_entry *pa, *pe,
pa                231 fs/nfs_common/nfsacl.c 	FOREACH_ACL_ENTRY(pa, acl, pe) {
pa                232 fs/nfs_common/nfsacl.c 		switch(pa->e_tag) {
pa                236 fs/nfs_common/nfsacl.c 				group_obj = pa;
pa                239 fs/nfs_common/nfsacl.c 				mask = pa;
pa                191 fs/nfsd/nfs4acl.c 	struct posix_acl_entry *pa, *pe;
pa                203 fs/nfsd/nfs4acl.c 	FOREACH_ACL_ENTRY(pa, acl, pe) {
pa                204 fs/nfsd/nfs4acl.c 		switch (pa->e_tag) {
pa                206 fs/nfsd/nfs4acl.c 				pas->owner = pa->e_perm;
pa                209 fs/nfsd/nfs4acl.c 				pas->group = pa->e_perm;
pa                212 fs/nfsd/nfs4acl.c 				pas->users |= pa->e_perm;
pa                215 fs/nfsd/nfs4acl.c 				pas->groups |= pa->e_perm;
pa                218 fs/nfsd/nfs4acl.c 				pas->other = pa->e_perm;
pa                221 fs/nfsd/nfs4acl.c 				pas->mask = pa->e_perm;
pa                236 fs/nfsd/nfs4acl.c 	struct posix_acl_entry *pa, *group_owner_entry;
pa                246 fs/nfsd/nfs4acl.c 	pa = pacl->a_entries;
pa                267 fs/nfsd/nfs4acl.c 	ace->access_mask = mask_from_posix(pa->e_perm, flags | NFS4_ACL_OWNER);
pa                271 fs/nfsd/nfs4acl.c 	pa++;
pa                273 fs/nfsd/nfs4acl.c 	while (pa->e_tag == ACL_USER) {
pa                274 fs/nfsd/nfs4acl.c 		deny = ~(pa->e_perm & pas.mask);
pa                281 fs/nfsd/nfs4acl.c 			ace->who_uid = pa->e_uid;
pa                287 fs/nfsd/nfs4acl.c 		ace->access_mask = mask_from_posix(pa->e_perm & pas.mask,
pa                290 fs/nfsd/nfs4acl.c 		ace->who_uid = pa->e_uid;
pa                293 fs/nfsd/nfs4acl.c 		pa++;
pa                301 fs/nfsd/nfs4acl.c 	group_owner_entry = pa;
pa                309 fs/nfsd/nfs4acl.c 	pa++;
pa                311 fs/nfsd/nfs4acl.c 	while (pa->e_tag == ACL_GROUP) {
pa                314 fs/nfsd/nfs4acl.c 		ace->access_mask = mask_from_posix(pa->e_perm & pas.mask,
pa                317 fs/nfsd/nfs4acl.c 		ace->who_gid = pa->e_gid;
pa                320 fs/nfsd/nfs4acl.c 		pa++;
pa                325 fs/nfsd/nfs4acl.c 	pa = group_owner_entry;
pa                336 fs/nfsd/nfs4acl.c 	pa++;
pa                338 fs/nfsd/nfs4acl.c 	while (pa->e_tag == ACL_GROUP) {
pa                339 fs/nfsd/nfs4acl.c 		deny = ~(pa->e_perm & pas.mask);
pa                346 fs/nfsd/nfs4acl.c 			ace->who_gid = pa->e_gid;
pa                350 fs/nfsd/nfs4acl.c 		pa++;
pa                353 fs/nfsd/nfs4acl.c 	if (pa->e_tag == ACL_MASK)
pa                354 fs/nfsd/nfs4acl.c 		pa++;
pa                357 fs/nfsd/nfs4acl.c 	ace->access_mask = mask_from_posix(pa->e_perm, flags);
pa                212 fs/posix_acl.c 	const struct posix_acl_entry *pa, *pe;
pa                216 fs/posix_acl.c 	FOREACH_ACL_ENTRY(pa, acl, pe) {
pa                217 fs/posix_acl.c 		if (pa->e_perm & ~(ACL_READ|ACL_WRITE|ACL_EXECUTE))
pa                219 fs/posix_acl.c 		switch (pa->e_tag) {
pa                230 fs/posix_acl.c 				if (!kuid_has_mapping(user_ns, pa->e_uid))
pa                245 fs/posix_acl.c 				if (!kgid_has_mapping(user_ns, pa->e_gid))
pa                281 fs/posix_acl.c 	const struct posix_acl_entry *pa, *pe;
pa                291 fs/posix_acl.c 	FOREACH_ACL_ENTRY(pa, acl, pe) {
pa                292 fs/posix_acl.c 		switch (pa->e_tag) {
pa                294 fs/posix_acl.c 				mode |= (pa->e_perm & S_IRWXO) << 6;
pa                297 fs/posix_acl.c 				mode |= (pa->e_perm & S_IRWXO) << 3;
pa                300 fs/posix_acl.c 				mode |= pa->e_perm & S_IRWXO;
pa                304 fs/posix_acl.c 				       ((pa->e_perm & S_IRWXO) << 3);
pa                350 fs/posix_acl.c 	const struct posix_acl_entry *pa, *pe, *mask_obj;
pa                355 fs/posix_acl.c 	FOREACH_ACL_ENTRY(pa, acl, pe) {
pa                356 fs/posix_acl.c                 switch(pa->e_tag) {
pa                363 fs/posix_acl.c 				if (uid_eq(pa->e_uid, current_fsuid()))
pa                369 fs/posix_acl.c 					if ((pa->e_perm & want) == want)
pa                374 fs/posix_acl.c 				if (in_group_p(pa->e_gid)) {
pa                376 fs/posix_acl.c 					if ((pa->e_perm & want) == want)
pa                394 fs/posix_acl.c 	for (mask_obj = pa+1; mask_obj != pe; mask_obj++) {
pa                396 fs/posix_acl.c 			if ((pa->e_perm & mask_obj->e_perm & want) == want)
pa                403 fs/posix_acl.c 	if ((pa->e_perm & want) == want)
pa                418 fs/posix_acl.c 	struct posix_acl_entry *pa, *pe;
pa                425 fs/posix_acl.c 	FOREACH_ACL_ENTRY(pa, acl, pe) {
pa                426 fs/posix_acl.c                 switch(pa->e_tag) {
pa                428 fs/posix_acl.c 				pa->e_perm &= (mode >> 6) | ~S_IRWXO;
pa                429 fs/posix_acl.c 				mode &= (pa->e_perm << 6) | ~S_IRWXU;
pa                438 fs/posix_acl.c 				group_obj = pa;
pa                442 fs/posix_acl.c 				pa->e_perm &= mode | ~S_IRWXO;
pa                443 fs/posix_acl.c 				mode &= pa->e_perm | ~S_IRWXO;
pa                447 fs/posix_acl.c 				mask_obj = pa;
pa                476 fs/posix_acl.c 	struct posix_acl_entry *pa, *pe;
pa                480 fs/posix_acl.c 	FOREACH_ACL_ENTRY(pa, acl, pe) {
pa                481 fs/posix_acl.c 		switch(pa->e_tag) {
pa                483 fs/posix_acl.c 				pa->e_perm = (mode & S_IRWXU) >> 6;
pa                491 fs/posix_acl.c 				group_obj = pa;
pa                495 fs/posix_acl.c 				mask_obj = pa;
pa                499 fs/posix_acl.c 				pa->e_perm = (mode & S_IRWXO);
pa                649 fs/xfs/scrub/agheader.c 	const void		*pa,
pa                652 fs/xfs/scrub/agheader.c 	const xfs_agblock_t	*a = pa;
pa                167 include/asm-generic/mshyperv.h void hyperv_report_panic_msg(phys_addr_t pa, size_t size);
pa                 52 include/dt-bindings/pinctrl/am43xx.h #define AM4372_IOPAD(pa, val)	(((pa) & 0xffff) - 0x0800) (val)
pa                 71 include/dt-bindings/pinctrl/dra.h #define DRA7XX_CORE_IOPAD(pa, val)	(((pa) & 0xffff) - 0x3400) (val)
pa                 32 include/dt-bindings/pinctrl/k3.h #define AM65X_IOPAD(pa, val, muxmode)		(((pa) & 0x1fff)) ((val) | (muxmode))
pa                 33 include/dt-bindings/pinctrl/k3.h #define AM65X_WKUP_IOPAD(pa, val, muxmode)	(((pa) & 0x1fff)) ((val) | (muxmode))
pa                 35 include/dt-bindings/pinctrl/k3.h #define J721E_IOPAD(pa, val, muxmode)		(((pa) & 0x1fff)) ((val) | (muxmode))
pa                 36 include/dt-bindings/pinctrl/k3.h #define J721E_WKUP_IOPAD(pa, val, muxmode)	(((pa) & 0x1fff)) ((val) | (muxmode))
pa                 35 include/dt-bindings/pinctrl/keystone.h #define KEYSTONE_IOPAD_OFFSET(pa, offset) (((pa) & 0xffff) - (offset))
pa                 37 include/dt-bindings/pinctrl/keystone.h #define K2G_CORE_IOPAD(pa) KEYSTONE_IOPAD_OFFSET((pa), 0x1000)
pa                 57 include/dt-bindings/pinctrl/omap.h #define OMAP_IOPAD_OFFSET(pa, offset)	(((pa) & 0xffff) - (offset))
pa                 59 include/dt-bindings/pinctrl/omap.h #define OMAP2420_CORE_IOPAD(pa, val)	OMAP_IOPAD_OFFSET((pa), 0x0030) (val)
pa                 60 include/dt-bindings/pinctrl/omap.h #define OMAP2430_CORE_IOPAD(pa, val)	OMAP_IOPAD_OFFSET((pa), 0x2030) (val)
pa                 61 include/dt-bindings/pinctrl/omap.h #define OMAP3_CORE1_IOPAD(pa, val)	OMAP_IOPAD_OFFSET((pa), 0x2030) (val)
pa                 62 include/dt-bindings/pinctrl/omap.h #define OMAP3430_CORE2_IOPAD(pa, val)	OMAP_IOPAD_OFFSET((pa), 0x25d8) (val)
pa                 63 include/dt-bindings/pinctrl/omap.h #define OMAP3630_CORE2_IOPAD(pa, val)	OMAP_IOPAD_OFFSET((pa), 0x25a0) (val)
pa                 64 include/dt-bindings/pinctrl/omap.h #define OMAP3_WKUP_IOPAD(pa, val)	OMAP_IOPAD_OFFSET((pa), 0x2a00) (val)
pa                 65 include/dt-bindings/pinctrl/omap.h #define DM814X_IOPAD(pa, val)		OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
pa                 66 include/dt-bindings/pinctrl/omap.h #define DM816X_IOPAD(pa, val)		OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
pa                 67 include/dt-bindings/pinctrl/omap.h #define AM33XX_IOPAD(pa, val)		OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
pa                 68 include/dt-bindings/pinctrl/omap.h #define AM33XX_PADCONF(pa, dir, mux)	OMAP_IOPAD_OFFSET((pa), 0x0800) ((dir) | (mux))
pa                248 include/linux/acpi.h void acpi_numa_processor_affinity_init (struct acpi_srat_cpu_affinity *pa);
pa                251 include/linux/acpi.h acpi_numa_processor_affinity_init(struct acpi_srat_cpu_affinity *pa) { }
pa                254 include/linux/acpi.h void acpi_numa_x2apic_affinity_init(struct acpi_srat_x2apic_cpu_affinity *pa);
pa                257 include/linux/acpi.h void acpi_numa_gicc_affinity_init(struct acpi_srat_gicc_affinity *pa);
pa                260 include/linux/acpi.h acpi_numa_gicc_affinity_init(struct acpi_srat_gicc_affinity *pa) { }
pa                 47 include/linux/fsl/bestcomm/sram.h static inline void *bcom_sram_pa2va(phys_addr_t pa) {
pa                 49 include/linux/fsl/bestcomm/sram.h 		(unsigned long)(pa - bcom_sram->base_phys);
pa                 29 include/linux/if_pppox.h 	struct pppoe_addr	pa;	  /* what this socket is bound to*/
pa                 57 include/linux/if_pppox.h #define pppoe_pa	proto.pppoe.pa
pa                230 include/linux/irqdomain.h 						const char *name, phys_addr_t *pa);
pa                251 include/linux/irqdomain.h static inline struct fwnode_handle *irq_domain_alloc_fwnode(phys_addr_t *pa)
pa                253 include/linux/irqdomain.h 	return __irq_domain_alloc_fwnode(IRQCHIP_FWNODE_REAL, 0, NULL, pa);
pa                775 include/linux/mlx5/driver.h 	u64	pa;
pa                 34 include/linux/posix_acl.h #define FOREACH_ACL_ENTRY(pa, acl, pe) \
pa                 35 include/linux/posix_acl.h 	for(pa=(acl)->a_entries, pe=pa+(acl)->a_count; pa<pe; pa++)
pa                172 include/linux/remoteproc.h 	u32 pa;
pa                210 include/linux/remoteproc.h 	u32 pa;
pa                262 include/linux/remoteproc.h 	u32 pa;
pa                374 include/linux/tee_drv.h int tee_shm_va2pa(struct tee_shm *shm, void *va, phys_addr_t *pa);
pa                383 include/linux/tee_drv.h int tee_shm_pa2va(struct tee_shm *shm, phys_addr_t pa, void **va);
pa                402 include/linux/tee_drv.h int tee_shm_get_pa(struct tee_shm *shm, size_t offs, phys_addr_t *pa);
pa                107 include/media/cec-notifier.h void cec_notifier_set_phys_addr(struct cec_notifier *n, u16 pa);
pa                167 include/media/cec-notifier.h static inline void cec_notifier_set_phys_addr(struct cec_notifier *n, u16 pa)
pa                264 include/media/cec.h #define cec_phys_addr_exp(pa) \
pa                265 include/media/cec.h 	((pa) >> 12), ((pa) >> 8) & 0xf, ((pa) >> 4) & 0xf, (pa) & 0xf
pa                430 include/media/cec.h 			   void (*callback)(struct cec_adapter *adap, u16 pa));
pa                451 include/media/cec.h 		      void (*callback)(struct cec_adapter *adap, u16 pa))
pa                 40 include/ras/ras_event.h 		__field(u64, pa)
pa                 55 include/ras/ras_event.h 			__entry->pa = mem->physical_addr;
pa                 57 include/ras/ras_event.h 			__entry->pa = ~0ull;
pa                 72 include/ras/ras_event.h 		  __entry->pa,
pa                197 include/trace/events/erofs.h 		__field(	erofs_off_t,	pa		)
pa                209 include/trace/events/erofs.h 		__entry->pa	= map->m_pa;
pa                220 include/trace/events/erofs.h 		  __entry->la, __entry->pa, __entry->llen, __entry->plen,
pa                635 include/trace/events/ext4.h 		 struct ext4_prealloc_space *pa),
pa                637 include/trace/events/ext4.h 	TP_ARGS(ac, pa),
pa                651 include/trace/events/ext4.h 		__entry->pa_pstart	= pa->pa_pstart;
pa                652 include/trace/events/ext4.h 		__entry->pa_lstart	= pa->pa_lstart;
pa                653 include/trace/events/ext4.h 		__entry->pa_len		= pa->pa_len;
pa                665 include/trace/events/ext4.h 		 struct ext4_prealloc_space *pa),
pa                667 include/trace/events/ext4.h 	TP_ARGS(ac, pa)
pa                673 include/trace/events/ext4.h 		 struct ext4_prealloc_space *pa),
pa                675 include/trace/events/ext4.h 	TP_ARGS(ac, pa)
pa                679 include/trace/events/ext4.h 	TP_PROTO(struct ext4_prealloc_space *pa,
pa                682 include/trace/events/ext4.h 	TP_ARGS(pa, block, count),
pa                693 include/trace/events/ext4.h 		__entry->dev		= pa->pa_inode->i_sb->s_dev;
pa                694 include/trace/events/ext4.h 		__entry->ino		= pa->pa_inode->i_ino;
pa                706 include/trace/events/ext4.h 	TP_PROTO(struct super_block *sb, struct ext4_prealloc_space *pa),
pa                708 include/trace/events/ext4.h 	TP_ARGS(sb, pa),
pa                719 include/trace/events/ext4.h 		__entry->pa_pstart	= pa->pa_pstart;
pa                720 include/trace/events/ext4.h 		__entry->pa_len		= pa->pa_len;
pa                 34 kernel/irq/irqdomain.c 	phys_addr_t		*pa;
pa                 66 kernel/irq/irqdomain.c 						phys_addr_t *pa)
pa                 81 kernel/irq/irqdomain.c 		n = kasprintf(GFP_KERNEL, "irqchip@%pa", pa);
pa                 93 kernel/irq/irqdomain.c 	fwid->pa = pa;
pa                 24 lib/test_debug_virtual.c 	phys_addr_t pa;
pa                 28 lib/test_debug_virtual.c 	pa = virt_to_phys(va);
pa                 30 lib/test_debug_virtual.c 	pr_info("PA: %pa for VA: 0x%lx\n", &pa, (unsigned long)va);
pa                 36 lib/test_debug_virtual.c 	pa = virt_to_phys(foo);
pa                 38 lib/test_debug_virtual.c 	pr_info("PA: %pa for VA: 0x%lx\n", &pa, (unsigned long)va);
pa                 82 mm/process_vm_access.c 	unsigned long pa = addr & PAGE_MASK;
pa                 83 mm/process_vm_access.c 	unsigned long start_offset = addr - pa;
pa                109 mm/process_vm_access.c 		pages = get_user_pages_remote(task, mm, pa, pages, flags,
pa                126 mm/process_vm_access.c 		pa += pages * PAGE_SIZE;
pa                387 net/8021q/vlan_dev.c static int vlan_dev_neigh_setup(struct net_device *dev, struct neigh_parms *pa)
pa                394 net/8021q/vlan_dev.c 		err = ops->ndo_neigh_setup(real_dev, pa);
pa                268 net/wireless/core.h 		} pa;
pa               1062 net/wireless/sme.c 	memcpy(ev->pa.bssid, bssid, ETH_ALEN);
pa                892 net/wireless/util.c 			__cfg80211_port_authorized(wdev, ev->pa.bssid);
pa                 52 scripts/dtc/include-prefixes/dt-bindings/pinctrl/am43xx.h #define AM4372_IOPAD(pa, val)	(((pa) & 0xffff) - 0x0800) (val)
pa                 71 scripts/dtc/include-prefixes/dt-bindings/pinctrl/dra.h #define DRA7XX_CORE_IOPAD(pa, val)	(((pa) & 0xffff) - 0x3400) (val)
pa                 32 scripts/dtc/include-prefixes/dt-bindings/pinctrl/k3.h #define AM65X_IOPAD(pa, val, muxmode)		(((pa) & 0x1fff)) ((val) | (muxmode))
pa                 33 scripts/dtc/include-prefixes/dt-bindings/pinctrl/k3.h #define AM65X_WKUP_IOPAD(pa, val, muxmode)	(((pa) & 0x1fff)) ((val) | (muxmode))
pa                 35 scripts/dtc/include-prefixes/dt-bindings/pinctrl/k3.h #define J721E_IOPAD(pa, val, muxmode)		(((pa) & 0x1fff)) ((val) | (muxmode))
pa                 36 scripts/dtc/include-prefixes/dt-bindings/pinctrl/k3.h #define J721E_WKUP_IOPAD(pa, val, muxmode)	(((pa) & 0x1fff)) ((val) | (muxmode))
pa                 35 scripts/dtc/include-prefixes/dt-bindings/pinctrl/keystone.h #define KEYSTONE_IOPAD_OFFSET(pa, offset) (((pa) & 0xffff) - (offset))
pa                 37 scripts/dtc/include-prefixes/dt-bindings/pinctrl/keystone.h #define K2G_CORE_IOPAD(pa) KEYSTONE_IOPAD_OFFSET((pa), 0x1000)
pa                 57 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define OMAP_IOPAD_OFFSET(pa, offset)	(((pa) & 0xffff) - (offset))
pa                 59 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define OMAP2420_CORE_IOPAD(pa, val)	OMAP_IOPAD_OFFSET((pa), 0x0030) (val)
pa                 60 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define OMAP2430_CORE_IOPAD(pa, val)	OMAP_IOPAD_OFFSET((pa), 0x2030) (val)
pa                 61 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define OMAP3_CORE1_IOPAD(pa, val)	OMAP_IOPAD_OFFSET((pa), 0x2030) (val)
pa                 62 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define OMAP3430_CORE2_IOPAD(pa, val)	OMAP_IOPAD_OFFSET((pa), 0x25d8) (val)
pa                 63 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define OMAP3630_CORE2_IOPAD(pa, val)	OMAP_IOPAD_OFFSET((pa), 0x25a0) (val)
pa                 64 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define OMAP3_WKUP_IOPAD(pa, val)	OMAP_IOPAD_OFFSET((pa), 0x2a00) (val)
pa                 65 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define DM814X_IOPAD(pa, val)		OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
pa                 66 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define DM816X_IOPAD(pa, val)		OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
pa                 67 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define AM33XX_IOPAD(pa, val)		OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
pa                 68 scripts/dtc/include-prefixes/dt-bindings/pinctrl/omap.h #define AM33XX_PADCONF(pa, dir, mux)	OMAP_IOPAD_OFFSET((pa), 0x0800) ((dir) | (mux))
pa                180 sound/pci/asihpi/hpioctl.c 		struct hpi_adapter *pa = NULL;
pa                183 sound/pci/asihpi/hpioctl.c 			pa = &adapters[array_index_nospec(hm->h.adapter_index,
pa                186 sound/pci/asihpi/hpioctl.c 		if (!pa || !pa->adapter || !pa->adapter->type) {
pa                199 sound/pci/asihpi/hpioctl.c 		if (mutex_lock_interruptible(&pa->mutex)) {
pa                216 sound/pci/asihpi/hpioctl.c 				if (pa->buffer_size < size) {
pa                221 sound/pci/asihpi/hpioctl.c 						pa->buffer_size, size);
pa                222 sound/pci/asihpi/hpioctl.c 					if (pa->p_buffer) {
pa                223 sound/pci/asihpi/hpioctl.c 						pa->buffer_size = 0;
pa                224 sound/pci/asihpi/hpioctl.c 						vfree(pa->p_buffer);
pa                226 sound/pci/asihpi/hpioctl.c 					pa->p_buffer = vmalloc(size);
pa                227 sound/pci/asihpi/hpioctl.c 					if (pa->p_buffer)
pa                228 sound/pci/asihpi/hpioctl.c 						pa->buffer_size = size;
pa                235 sound/pci/asihpi/hpioctl.c 						mutex_unlock(&pa->mutex);
pa                241 sound/pci/asihpi/hpioctl.c 				hm->m0.u.d.u.data.pb_data = pa->p_buffer;
pa                257 sound/pci/asihpi/hpioctl.c 				copy_from_user(pa->p_buffer, ptr, size);
pa                269 sound/pci/asihpi/hpioctl.c 				copy_to_user(ptr, pa->p_buffer, size);
pa                276 sound/pci/asihpi/hpioctl.c 		mutex_unlock(&pa->mutex);
pa                523 sound/pci/asihpi/hpioctl.c 	struct hpi_adapter *pa;
pa                526 sound/pci/asihpi/hpioctl.c 	pa = pci_get_drvdata(pci_dev);
pa                527 sound/pci/asihpi/hpioctl.c 	pci = pa->adapter->pci;
pa                532 sound/pci/asihpi/hpioctl.c 	hm.adapter_index = pa->adapter->index;
pa                540 sound/pci/asihpi/hpioctl.c 	hm.adapter_index = pa->adapter->index;
pa                547 sound/pci/asihpi/hpioctl.c 	if (pa->irq)
pa                548 sound/pci/asihpi/hpioctl.c 		free_irq(pa->irq, pa);
pa                550 sound/pci/asihpi/hpioctl.c 	vfree(pa->p_buffer);
pa                557 sound/pci/asihpi/hpioctl.c 			 pci_dev->devfn, pa->adapter->index);
pa                559 sound/pci/asihpi/hpioctl.c 	memset(pa, 0, sizeof(*pa));
pa                992 sound/pci/es1968.c 	u32 pa;
pa               1010 sound/pci/es1968.c 		pa = es->memory->buf.addr;
pa               1011 sound/pci/es1968.c 		pa -= chip->dma.addr;
pa               1012 sound/pci/es1968.c 		pa >>= 1;	/* words */
pa               1014 sound/pci/es1968.c 		pa |= 0x00400000;	/* System RAM (Bit 22) */
pa               1019 sound/pci/es1968.c 				pa |= 0x00800000;	/* (Bit 23) */
pa               1021 sound/pci/es1968.c 				pa >>= 1;
pa               1026 sound/pci/es1968.c 		es->base[channel] = pa & 0xFFFF;
pa               1032 sound/pci/es1968.c 		apu_set_register(chip, apu, 4, ((pa >> 16) & 0xFF) << 8);
pa               1033 sound/pci/es1968.c 		apu_set_register(chip, apu, 5, pa & 0xFFFF);
pa               1034 sound/pci/es1968.c 		apu_set_register(chip, apu, 6, (pa + size) & 0xFFFF);
pa               1093 sound/pci/es1968.c 			     unsigned int pa, unsigned int bsize,
pa               1101 sound/pci/es1968.c 	snd_es1968_program_wavecache(chip, es, channel, pa, 1);
pa               1104 sound/pci/es1968.c 	pa -= chip->dma.addr;
pa               1105 sound/pci/es1968.c 	pa >>= 1;	/* words */
pa               1109 sound/pci/es1968.c 	es->base[channel] = pa & 0xFFFF;
pa               1110 sound/pci/es1968.c 	pa |= 0x00400000;	/* bit 22 -> System RAM */
pa               1121 sound/pci/es1968.c 	apu_set_register(chip, apu, 4, ((pa >> 16) & 0xFF) << 8);
pa               1122 sound/pci/es1968.c 	apu_set_register(chip, apu, 5, pa & 0xFFFF);
pa               1123 sound/pci/es1968.c 	apu_set_register(chip, apu, 6, (pa + bsize) & 0xFFFF);
pa               1697 sound/pci/es1968.c 	unsigned int pa, offset, t;
pa               1722 sound/pci/es1968.c 	pa = (unsigned int)((memory->buf.addr - chip->dma.addr) >> 1);
pa               1723 sound/pci/es1968.c 	pa |= 0x00400000;	/* System RAM (Bit 22) */
pa               1730 sound/pci/es1968.c 	apu_set_register(chip, apu, 4, ((pa >> 16) & 0xff) << 8);
pa               1731 sound/pci/es1968.c 	apu_set_register(chip, apu, 5, pa & 0xffff);
pa               1732 sound/pci/es1968.c 	apu_set_register(chip, apu, 6, (pa + CLOCK_MEASURE_BUFSIZE/2) & 0xffff);
pa               1749 sound/pci/es1968.c 	__apu_set_register(chip, apu, 5, pa & 0xffff);
pa               1763 sound/pci/es1968.c 	offset -= (pa & 0xffff);
pa                686 tools/lib/traceevent/event-parse.c 	const struct printk_map *pa = a;
pa                689 tools/lib/traceevent/event-parse.c 	if (pa->addr < pb->addr)
pa                691 tools/lib/traceevent/event-parse.c 	if (pa->addr > pb->addr)
pa               1861 tools/perf/util/probe-event.c char *synthesize_perf_probe_arg(struct perf_probe_arg *pa)
pa               1863 tools/perf/util/probe-event.c 	struct perf_probe_arg_field *field = pa->field;
pa               1871 tools/perf/util/probe-event.c 	if (pa->name && pa->var)
pa               1872 tools/perf/util/probe-event.c 		err = strbuf_addf(&buf, "%s=%s", pa->name, pa->var);
pa               1874 tools/perf/util/probe-event.c 		err = strbuf_addstr(&buf, pa->name ?: pa->var);
pa               1889 tools/perf/util/probe-event.c 	if (pa->type)
pa               1890 tools/perf/util/probe-event.c 		if (strbuf_addf(&buf, ":%s", pa->type) < 0)
pa                135 tools/perf/util/probe-event.h char *synthesize_perf_probe_arg(struct perf_probe_arg *pa);
pa                348 tools/testing/selftests/kvm/include/x86_64/vmx.h 		: [pa]"m"(phys)
pa                365 tools/testing/selftests/kvm/include/x86_64/vmx.h 		: [pa]"m"(vmcs_pa)
pa                380 tools/testing/selftests/kvm/include/x86_64/vmx.h 		: [pa]"m"(vmcs_pa)
pa                777 tools/testing/selftests/net/nettest.c 				const uint32_t *pa = (uint32_t *) &in6->s6_addr;
pa                782 tools/testing/selftests/net/nettest.c 				pa += 3;
pa                783 tools/testing/selftests/net/nettest.c 				in4.s_addr = *pa;
pa               1340 virt/kvm/arm/mmu.c 			  phys_addr_t pa, unsigned long size, bool writable)
pa               1348 virt/kvm/arm/mmu.c 	pfn = __phys_to_pfn(pa);
pa               2336 virt/kvm/arm/mmu.c 			phys_addr_t pa;
pa               2338 virt/kvm/arm/mmu.c 			pa = (phys_addr_t)vma->vm_pgoff << PAGE_SHIFT;
pa               2339 virt/kvm/arm/mmu.c 			pa += vm_start - vma->vm_start;
pa               2347 virt/kvm/arm/mmu.c 			ret = kvm_phys_addr_ioremap(kvm, gpa, pa,