Searched refs:pa (Results 1 - 200 of 322) sorted by relevance

12

/linux-4.1.27/arch/mips/boot/dts/include/dt-bindings/pinctrl/
H A Domap.h56 #define OMAP_IOPAD_OFFSET(pa, offset) (((pa) & 0xffff) - (offset))
58 #define OMAP2420_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0030) (val)
59 #define OMAP2430_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val)
60 #define OMAP3_CORE1_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val)
61 #define OMAP3430_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25d8) (val)
62 #define OMAP3630_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25a0) (val)
63 #define OMAP3_WKUP_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2a00) (val)
64 #define DM816X_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
65 #define AM33XX_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
66 #define AM4372_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
67 #define DRA7XX_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x3400) (val)
/linux-4.1.27/arch/powerpc/boot/dts/include/dt-bindings/pinctrl/
H A Domap.h56 #define OMAP_IOPAD_OFFSET(pa, offset) (((pa) & 0xffff) - (offset))
58 #define OMAP2420_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0030) (val)
59 #define OMAP2430_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val)
60 #define OMAP3_CORE1_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val)
61 #define OMAP3430_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25d8) (val)
62 #define OMAP3630_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25a0) (val)
63 #define OMAP3_WKUP_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2a00) (val)
64 #define DM816X_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
65 #define AM33XX_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
66 #define AM4372_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
67 #define DRA7XX_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x3400) (val)
/linux-4.1.27/arch/arm64/boot/dts/include/dt-bindings/pinctrl/
H A Domap.h56 #define OMAP_IOPAD_OFFSET(pa, offset) (((pa) & 0xffff) - (offset))
58 #define OMAP2420_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0030) (val)
59 #define OMAP2430_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val)
60 #define OMAP3_CORE1_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val)
61 #define OMAP3430_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25d8) (val)
62 #define OMAP3630_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25a0) (val)
63 #define OMAP3_WKUP_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2a00) (val)
64 #define DM816X_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
65 #define AM33XX_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
66 #define AM4372_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
67 #define DRA7XX_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x3400) (val)
/linux-4.1.27/arch/metag/boot/dts/include/dt-bindings/pinctrl/
H A Domap.h56 #define OMAP_IOPAD_OFFSET(pa, offset) (((pa) & 0xffff) - (offset))
58 #define OMAP2420_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0030) (val)
59 #define OMAP2430_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val)
60 #define OMAP3_CORE1_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val)
61 #define OMAP3430_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25d8) (val)
62 #define OMAP3630_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25a0) (val)
63 #define OMAP3_WKUP_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2a00) (val)
64 #define DM816X_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
65 #define AM33XX_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
66 #define AM4372_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
67 #define DRA7XX_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x3400) (val)
/linux-4.1.27/arch/arm/boot/dts/include/dt-bindings/pinctrl/
H A Domap.h56 #define OMAP_IOPAD_OFFSET(pa, offset) (((pa) & 0xffff) - (offset))
58 #define OMAP2420_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0030) (val)
59 #define OMAP2430_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val)
60 #define OMAP3_CORE1_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val)
61 #define OMAP3430_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25d8) (val)
62 #define OMAP3630_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25a0) (val)
63 #define OMAP3_WKUP_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2a00) (val)
64 #define DM816X_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
65 #define AM33XX_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
66 #define AM4372_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
67 #define DRA7XX_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x3400) (val)
/linux-4.1.27/include/dt-bindings/pinctrl/
H A Domap.h56 #define OMAP_IOPAD_OFFSET(pa, offset) (((pa) & 0xffff) - (offset))
58 #define OMAP2420_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0030) (val)
59 #define OMAP2430_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val)
60 #define OMAP3_CORE1_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val)
61 #define OMAP3430_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25d8) (val)
62 #define OMAP3630_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25a0) (val)
63 #define OMAP3_WKUP_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2a00) (val)
64 #define DM816X_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
65 #define AM33XX_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
66 #define AM4372_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val)
67 #define DRA7XX_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x3400) (val)
/linux-4.1.27/drivers/spmi/
H A Dspmi-pmic-arb.c412 struct spmi_pmic_arb_dev *pa = irq_data_get_irq_chip_data(d); qpnpint_spmi_write() local
416 if (pmic_arb_write_cmd(pa->spmic, SPMI_CMD_EXT_WRITEL, sid, qpnpint_spmi_write()
418 dev_err_ratelimited(&pa->spmic->dev, qpnpint_spmi_write()
425 struct spmi_pmic_arb_dev *pa = irq_data_get_irq_chip_data(d); qpnpint_spmi_read() local
429 if (pmic_arb_read_cmd(pa->spmic, SPMI_CMD_EXT_READL, sid, qpnpint_spmi_read()
431 dev_err_ratelimited(&pa->spmic->dev, qpnpint_spmi_read()
436 static void periph_interrupt(struct spmi_pmic_arb_dev *pa, u8 apid) periph_interrupt() argument
442 status = readl_relaxed(pa->intr + pa->ver_ops->irq_status(apid)); periph_interrupt()
446 irq = irq_find_mapping(pa->domain, periph_interrupt()
447 pa->apid_to_ppid[apid] << 16 periph_interrupt()
456 struct spmi_pmic_arb_dev *pa = irq_get_handler_data(irq); pmic_arb_chained_irq() local
458 void __iomem *intr = pa->intr; pmic_arb_chained_irq()
459 int first = pa->min_apid >> 5; pmic_arb_chained_irq()
460 int last = pa->max_apid >> 5; pmic_arb_chained_irq()
468 pa->ver_ops->owner_acc_status(pa->ee, i)); pmic_arb_chained_irq()
472 periph_interrupt(pa, id + i * 32); pmic_arb_chained_irq()
481 struct spmi_pmic_arb_dev *pa = irq_data_get_irq_chip_data(d); qpnpint_irq_ack() local
487 raw_spin_lock_irqsave(&pa->lock, flags); qpnpint_irq_ack()
488 writel_relaxed(1 << irq, pa->intr + pa->ver_ops->irq_clear(apid)); qpnpint_irq_ack()
489 raw_spin_unlock_irqrestore(&pa->lock, flags); qpnpint_irq_ack()
497 struct spmi_pmic_arb_dev *pa = irq_data_get_irq_chip_data(d); qpnpint_irq_mask() local
504 raw_spin_lock_irqsave(&pa->lock, flags); qpnpint_irq_mask()
505 status = readl_relaxed(pa->intr + pa->ver_ops->acc_enable(apid)); qpnpint_irq_mask()
508 writel_relaxed(status, pa->intr + qpnpint_irq_mask()
509 pa->ver_ops->acc_enable(apid)); qpnpint_irq_mask()
511 raw_spin_unlock_irqrestore(&pa->lock, flags); qpnpint_irq_mask()
519 struct spmi_pmic_arb_dev *pa = irq_data_get_irq_chip_data(d); qpnpint_irq_unmask() local
526 raw_spin_lock_irqsave(&pa->lock, flags); qpnpint_irq_unmask()
527 status = readl_relaxed(pa->intr + pa->ver_ops->acc_enable(apid)); qpnpint_irq_unmask()
530 pa->intr + pa->ver_ops->acc_enable(apid)); qpnpint_irq_unmask()
532 raw_spin_unlock_irqrestore(&pa->lock, flags); qpnpint_irq_unmask()
595 static int search_mapping_table(struct spmi_pmic_arb_dev *pa, search_mapping_table() argument
600 u32 *mapping_table = pa->mapping_table; search_mapping_table()
634 struct spmi_pmic_arb_dev *pa = d->host_data; qpnpint_irq_domain_dt_translate() local
639 dev_dbg(&pa->spmic->dev, qpnpint_irq_domain_dt_translate()
654 err = search_mapping_table(pa, &spec, &apid); qpnpint_irq_domain_dt_translate()
658 pa->apid_to_ppid[apid] = spec.slave << 8 | spec.per; qpnpint_irq_domain_dt_translate()
661 if (apid > pa->max_apid) qpnpint_irq_domain_dt_translate()
662 pa->max_apid = apid; qpnpint_irq_domain_dt_translate()
663 if (apid < pa->min_apid) qpnpint_irq_domain_dt_translate()
664 pa->min_apid = apid; qpnpint_irq_domain_dt_translate()
672 dev_dbg(&pa->spmic->dev, "out_hwirq = %lu\n", *out_hwirq); qpnpint_irq_domain_dt_translate()
681 struct spmi_pmic_arb_dev *pa = d->host_data; qpnpint_irq_domain_map() local
683 dev_dbg(&pa->spmic->dev, "virq = %u, hwirq = %lu\n", virq, hwirq); qpnpint_irq_domain_map()
692 static u32 pmic_arb_offset_v1(struct spmi_pmic_arb_dev *pa, u8 sid, u16 addr) pmic_arb_offset_v1() argument
694 return 0x800 + 0x80 * pa->channel; pmic_arb_offset_v1()
698 static u32 pmic_arb_offset_v2(struct spmi_pmic_arb_dev *pa, u8 sid, u16 addr) pmic_arb_offset_v2() argument
701 u8 chan = pa->ppid_to_chan[ppid]; pmic_arb_offset_v2()
703 return 0x1000 * pa->ee + 0x8000 * chan; pmic_arb_offset_v2()
783 struct spmi_pmic_arb_dev *pa; spmi_pmic_arb_probe() local
791 ctrl = spmi_controller_alloc(&pdev->dev, sizeof(*pa)); spmi_pmic_arb_probe()
795 pa = spmi_controller_get_drvdata(ctrl); spmi_pmic_arb_probe()
796 pa->spmic = ctrl; spmi_pmic_arb_probe()
812 pa->ver_ops = &pmic_arb_v1; spmi_pmic_arb_probe()
813 pa->wr_base = core; spmi_pmic_arb_probe()
814 pa->rd_base = core; spmi_pmic_arb_probe()
820 pa->ver_ops = &pmic_arb_v2; spmi_pmic_arb_probe()
824 pa->rd_base = devm_ioremap_resource(&ctrl->dev, res); spmi_pmic_arb_probe()
825 if (IS_ERR(pa->rd_base)) { spmi_pmic_arb_probe()
826 err = PTR_ERR(pa->rd_base); spmi_pmic_arb_probe()
832 pa->wr_base = devm_ioremap_resource(&ctrl->dev, res); spmi_pmic_arb_probe()
833 if (IS_ERR(pa->wr_base)) { spmi_pmic_arb_probe()
834 err = PTR_ERR(pa->wr_base); spmi_pmic_arb_probe()
838 pa->ppid_to_chan = devm_kzalloc(&ctrl->dev, spmi_pmic_arb_probe()
840 if (!pa->ppid_to_chan) { spmi_pmic_arb_probe()
854 pa->ppid_to_chan[ppid] = chan; spmi_pmic_arb_probe()
859 pa->intr = devm_ioremap_resource(&ctrl->dev, res); spmi_pmic_arb_probe()
860 if (IS_ERR(pa->intr)) { spmi_pmic_arb_probe()
861 err = PTR_ERR(pa->intr); spmi_pmic_arb_probe()
866 pa->cnfg = devm_ioremap_resource(&ctrl->dev, res); spmi_pmic_arb_probe()
867 if (IS_ERR(pa->cnfg)) { spmi_pmic_arb_probe()
868 err = PTR_ERR(pa->cnfg); spmi_pmic_arb_probe()
872 pa->irq = platform_get_irq_byname(pdev, "periph_irq"); spmi_pmic_arb_probe()
873 if (pa->irq < 0) { spmi_pmic_arb_probe()
874 err = pa->irq; spmi_pmic_arb_probe()
890 pa->channel = channel; spmi_pmic_arb_probe()
904 pa->ee = ee; spmi_pmic_arb_probe()
906 for (i = 0; i < ARRAY_SIZE(pa->mapping_table); ++i) spmi_pmic_arb_probe()
907 pa->mapping_table[i] = readl_relaxed( spmi_pmic_arb_probe()
908 pa->cnfg + SPMI_MAPPING_TABLE_REG(i)); spmi_pmic_arb_probe()
912 pa->max_apid = 0; spmi_pmic_arb_probe()
913 pa->min_apid = PMIC_ARB_MAX_PERIPHS - 1; spmi_pmic_arb_probe()
916 raw_spin_lock_init(&pa->lock); spmi_pmic_arb_probe()
923 pa->domain = irq_domain_add_tree(pdev->dev.of_node, spmi_pmic_arb_probe()
924 &pmic_arb_irq_domain_ops, pa); spmi_pmic_arb_probe()
925 if (!pa->domain) { spmi_pmic_arb_probe()
931 irq_set_handler_data(pa->irq, pa); spmi_pmic_arb_probe()
932 irq_set_chained_handler(pa->irq, pmic_arb_chained_irq); spmi_pmic_arb_probe()
941 irq_set_chained_handler(pa->irq, NULL); spmi_pmic_arb_probe()
942 irq_set_handler_data(pa->irq, NULL); spmi_pmic_arb_probe()
943 irq_domain_remove(pa->domain); spmi_pmic_arb_probe()
952 struct spmi_pmic_arb_dev *pa = spmi_controller_get_drvdata(ctrl); spmi_pmic_arb_remove() local
954 irq_set_chained_handler(pa->irq, NULL); spmi_pmic_arb_remove()
955 irq_set_handler_data(pa->irq, NULL); spmi_pmic_arb_remove()
956 irq_domain_remove(pa->domain); spmi_pmic_arb_remove()
/linux-4.1.27/arch/unicore32/mm/
H A Dproc-macros.S88 * va2pa va, pa, tbl, msk, off, err
92 * pa: physical address, result is stored in this register
97 .macro va2pa, va, pa, tbl, msk, off, err=990f
98 movc \pa, p0.c2, #0
102 add \pa, \pa, #PAGE_OFFSET @ pa <- virt addr of page table
103 ldw \pa, [\pa+], \off << #2 @ pa <- the content of pt
104 cand.a \pa, #4 @ test exist bit
106 and \off, \pa, #3 @ off <- the last 2 bits
111 and \pa, \pa, \msk @ pa <- phys addr of 2nd pt
131 and \pa, \pa, \msk
132 or \pa, \pa, \tbl
/linux-4.1.27/arch/parisc/include/asm/
H A Dhardware.h22 pcx = 0, /* pa7000 pa 1.0 */
23 pcxs = 1, /* pa7000 pa 1.1a */
24 pcxt = 2, /* pa7100 pa 1.1b */
25 pcxt_ = 3, /* pa7200 (t') pa 1.1c */
26 pcxl = 4, /* pa7100lc pa 1.1d */
27 pcxl2 = 5, /* pa7300lc pa 1.1e */
28 pcxu = 6, /* pa8000 pa 2.0 */
29 pcxu_ = 7, /* pa8200 (u+) pa 2.0 */
30 pcxw = 8, /* pa8500 pa 2.0 */
31 pcxw_ = 9, /* pa8600 (w+) pa 2.0 */
32 pcxw2 = 10, /* pa8700 pa 2.0 */
33 mako = 11, /* pa8800 pa 2.0 */
34 mako2 = 12 /* pa8900 pa 2.0 */
/linux-4.1.27/arch/arm/include/asm/
H A Dsmp_scu.h19 unsigned long pa; scu_a9_get_base() local
21 asm("mrc p15, 4, %0, c15, c0, 0" : "=r" (pa)); scu_a9_get_base()
23 return pa; scu_a9_get_base()
H A Dkvm_mmu.h60 phys_addr_t pa, unsigned long size, bool writable);
/linux-4.1.27/arch/arm/mach-omap1/
H A Dio.c148 u8 omap_readb(u32 pa) omap_readb() argument
150 return __raw_readb(OMAP1_IO_ADDRESS(pa)); omap_readb()
154 u16 omap_readw(u32 pa) omap_readw() argument
156 return __raw_readw(OMAP1_IO_ADDRESS(pa)); omap_readw()
160 u32 omap_readl(u32 pa) omap_readl() argument
162 return __raw_readl(OMAP1_IO_ADDRESS(pa)); omap_readl()
166 void omap_writeb(u8 v, u32 pa) omap_writeb() argument
168 __raw_writeb(v, OMAP1_IO_ADDRESS(pa)); omap_writeb()
172 void omap_writew(u16 v, u32 pa) omap_writew() argument
174 __raw_writew(v, OMAP1_IO_ADDRESS(pa)); omap_writew()
178 void omap_writel(u32 v, u32 pa) omap_writel() argument
180 __raw_writel(v, OMAP1_IO_ADDRESS(pa)); omap_writel()
H A Dclock.c994 struct clk *pa; clk_dbg_show_summary() local
1001 pa = c->parent; clk_dbg_show_summary()
1003 c->name, pa ? pa->name : "none", c->rate, clk_dbg_show_summary()
1027 struct clk *pa = c->parent; clk_debugfs_register_one() local
1029 d = debugfs_create_dir(c->name, pa ? pa->dent : clk_debugfs_root); clk_debugfs_register_one()
1059 struct clk *pa = c->parent; clk_debugfs_register() local
1061 if (pa && !pa->dent) { clk_debugfs_register()
1062 err = clk_debugfs_register(pa); clk_debugfs_register()
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/engine/dmaobj/
H A Dpriv.h5 #define nvkm_dmaobj_create(p,e,c,pa,sa,d) \
6 nvkm_dmaobj_create_((p), (e), (c), (pa), (sa), sizeof(**d), (void **)d)
/linux-4.1.27/arch/s390/mm/
H A Dcmm.c66 struct cmm_page_array *pa, *npa; cmm_alloc_pages() local
74 pa = *list; cmm_alloc_pages()
75 if (!pa || pa->index >= CMM_NR_PAGES) { cmm_alloc_pages()
85 pa = *list; cmm_alloc_pages()
86 if (!pa || pa->index >= CMM_NR_PAGES) { cmm_alloc_pages()
87 npa->next = pa; cmm_alloc_pages()
89 pa = npa; cmm_alloc_pages()
90 *list = pa; cmm_alloc_pages()
95 pa->pages[pa->index++] = addr; cmm_alloc_pages()
105 struct cmm_page_array *pa; cmm_free_pages() local
109 pa = *list; cmm_free_pages()
111 if (!pa || pa->index <= 0) cmm_free_pages()
113 addr = pa->pages[--pa->index]; cmm_free_pages()
114 if (pa->index == 0) { cmm_free_pages()
115 pa = pa->next; cmm_free_pages()
117 *list = pa; cmm_free_pages()
/linux-4.1.27/arch/mips/include/asm/mach-ip32/
H A Ddma-coherence.h32 dma_addr_t pa = virt_to_phys(addr) & RAM_OFFSET_MASK; plat_map_dma_mem() local
35 pa += CRIME_HI_MEM_BASE; plat_map_dma_mem()
37 return pa; plat_map_dma_mem()
43 dma_addr_t pa; plat_map_dma_mem_page() local
45 pa = page_to_phys(page) & RAM_OFFSET_MASK; plat_map_dma_mem_page()
48 pa += CRIME_HI_MEM_BASE; plat_map_dma_mem_page()
50 return pa; plat_map_dma_mem_page()
/linux-4.1.27/arch/x86/mm/
H A Dsrat.c74 acpi_numa_x2apic_affinity_init(struct acpi_srat_x2apic_cpu_affinity *pa) acpi_numa_x2apic_affinity_init() argument
81 if (pa->header.length < sizeof(struct acpi_srat_x2apic_cpu_affinity)) { acpi_numa_x2apic_affinity_init()
85 if ((pa->flags & ACPI_SRAT_CPU_ENABLED) == 0) acpi_numa_x2apic_affinity_init()
87 pxm = pa->proximity_domain; acpi_numa_x2apic_affinity_init()
88 apic_id = pa->apic_id; acpi_numa_x2apic_affinity_init()
114 acpi_numa_processor_affinity_init(struct acpi_srat_cpu_affinity *pa) acpi_numa_processor_affinity_init() argument
121 if (pa->header.length != sizeof(struct acpi_srat_cpu_affinity)) { acpi_numa_processor_affinity_init()
125 if ((pa->flags & ACPI_SRAT_CPU_ENABLED) == 0) acpi_numa_processor_affinity_init()
127 pxm = pa->proximity_domain_lo; acpi_numa_processor_affinity_init()
129 pxm |= *((unsigned int*)pa->proximity_domain_hi) << 8; acpi_numa_processor_affinity_init()
138 apic_id = (pa->apic_id << 8) | pa->local_sapic_eid; acpi_numa_processor_affinity_init()
140 apic_id = pa->apic_id; acpi_numa_processor_affinity_init()
/linux-4.1.27/fs/ext4/
H A Dmballoc.c318 * - per-pa lock (pa)
321 * - new pa
325 * - find and use pa:
326 * pa
328 * - release consumed pa:
329 * pa
335 * pa
339 * pa
344 * pa
641 struct ext4_prealloc_space *pa; __mb_check_buddy() local
642 pa = list_entry(cur, struct ext4_prealloc_space, pa_group_list); __mb_check_buddy()
643 ext4_get_group_no_and_offset(sb, pa->pa_pstart, &groupnr, &k); __mb_check_buddy()
645 for (i = 0; i < pa->pa_len; i++) __mb_check_buddy()
2672 struct ext4_prealloc_space *pa; ext4_mb_cleanup_pa() local
2677 pa = list_entry(cur, struct ext4_prealloc_space, pa_group_list); ext4_mb_cleanup_pa()
2678 list_del(&pa->pa_group_list); ext4_mb_cleanup_pa()
2680 kmem_cache_free(ext4_pspace_cachep, pa); ext4_mb_cleanup_pa()
3015 struct ext4_prealloc_space *pa; ext4_mb_normalize_request() local
3102 list_for_each_entry_rcu(pa, &ei->i_prealloc_list, pa_inode_list) { ext4_mb_normalize_request()
3105 if (pa->pa_deleted) ext4_mb_normalize_request()
3107 spin_lock(&pa->pa_lock); ext4_mb_normalize_request()
3108 if (pa->pa_deleted) { ext4_mb_normalize_request()
3109 spin_unlock(&pa->pa_lock); ext4_mb_normalize_request()
3113 pa_end = pa->pa_lstart + EXT4_C2B(EXT4_SB(ac->ac_sb), ext4_mb_normalize_request()
3114 pa->pa_len); ext4_mb_normalize_request()
3118 ac->ac_o_ex.fe_logical < pa->pa_lstart)); ext4_mb_normalize_request()
3121 if (pa->pa_lstart >= end || pa_end <= start) { ext4_mb_normalize_request()
3122 spin_unlock(&pa->pa_lock); ext4_mb_normalize_request()
3125 BUG_ON(pa->pa_lstart <= start && pa_end >= end); ext4_mb_normalize_request()
3127 /* adjust start or end to be adjacent to this pa */ ext4_mb_normalize_request()
3131 } else if (pa->pa_lstart > ac->ac_o_ex.fe_logical) { ext4_mb_normalize_request()
3132 BUG_ON(pa->pa_lstart > end); ext4_mb_normalize_request()
3133 end = pa->pa_lstart; ext4_mb_normalize_request()
3135 spin_unlock(&pa->pa_lock); ext4_mb_normalize_request()
3142 list_for_each_entry_rcu(pa, &ei->i_prealloc_list, pa_inode_list) { ext4_mb_normalize_request()
3145 spin_lock(&pa->pa_lock); ext4_mb_normalize_request()
3146 if (pa->pa_deleted == 0) { ext4_mb_normalize_request()
3147 pa_end = pa->pa_lstart + EXT4_C2B(EXT4_SB(ac->ac_sb), ext4_mb_normalize_request()
3148 pa->pa_len); ext4_mb_normalize_request()
3149 BUG_ON(!(start >= pa_end || end <= pa->pa_lstart)); ext4_mb_normalize_request()
3151 spin_unlock(&pa->pa_lock); ext4_mb_normalize_request()
3223 struct ext4_prealloc_space *pa = ac->ac_pa; ext4_discard_allocated_blocks() local
3227 if (pa == NULL) { ext4_discard_allocated_blocks()
3247 if (pa->pa_type == MB_INODE_PA) ext4_discard_allocated_blocks()
3248 pa->pa_free += ac->ac_b_ex.fe_len; ext4_discard_allocated_blocks()
3255 struct ext4_prealloc_space *pa) ext4_mb_use_inode_pa()
3263 start = pa->pa_pstart + (ac->ac_o_ex.fe_logical - pa->pa_lstart); ext4_mb_use_inode_pa()
3264 end = min(pa->pa_pstart + EXT4_C2B(sbi, pa->pa_len), ext4_mb_use_inode_pa()
3271 ac->ac_pa = pa; ext4_mb_use_inode_pa()
3273 BUG_ON(start < pa->pa_pstart); ext4_mb_use_inode_pa()
3274 BUG_ON(end > pa->pa_pstart + EXT4_C2B(sbi, pa->pa_len)); ext4_mb_use_inode_pa()
3275 BUG_ON(pa->pa_free < len); ext4_mb_use_inode_pa()
3276 pa->pa_free -= len; ext4_mb_use_inode_pa()
3278 mb_debug(1, "use %llu/%u from inode pa %p\n", start, len, pa); ext4_mb_use_inode_pa()
3285 struct ext4_prealloc_space *pa) ext4_mb_use_group_pa()
3289 ext4_get_group_no_and_offset(ac->ac_sb, pa->pa_pstart, ext4_mb_use_group_pa()
3294 ac->ac_pa = pa; ext4_mb_use_group_pa()
3298 * instead we correct pa later, after blocks are marked ext4_mb_use_group_pa()
3300 * Other CPUs are prevented from allocating from this pa by lg_mutex ext4_mb_use_group_pa()
3302 mb_debug(1, "use %u/%u from group pa %p\n", pa->pa_lstart-len, len, pa); ext4_mb_use_group_pa()
3313 struct ext4_prealloc_space *pa, ext4_mb_check_group_pa()
3319 atomic_inc(&pa->pa_count); ext4_mb_check_group_pa()
3320 return pa; ext4_mb_check_group_pa()
3323 new_distance = abs(goal_block - pa->pa_pstart); ext4_mb_check_group_pa()
3330 atomic_inc(&pa->pa_count); ext4_mb_check_group_pa()
3331 return pa; ext4_mb_check_group_pa()
3344 struct ext4_prealloc_space *pa, *cpa = NULL; ext4_mb_use_preallocated() local
3353 list_for_each_entry_rcu(pa, &ei->i_prealloc_list, pa_inode_list) { ext4_mb_use_preallocated()
3357 if (ac->ac_o_ex.fe_logical < pa->pa_lstart || ext4_mb_use_preallocated()
3358 ac->ac_o_ex.fe_logical >= (pa->pa_lstart + ext4_mb_use_preallocated()
3359 EXT4_C2B(sbi, pa->pa_len))) ext4_mb_use_preallocated()
3364 (pa->pa_pstart + EXT4_C2B(sbi, pa->pa_len) > ext4_mb_use_preallocated()
3369 spin_lock(&pa->pa_lock); ext4_mb_use_preallocated()
3370 if (pa->pa_deleted == 0 && pa->pa_free) { ext4_mb_use_preallocated()
3371 atomic_inc(&pa->pa_count); ext4_mb_use_preallocated()
3372 ext4_mb_use_inode_pa(ac, pa); ext4_mb_use_preallocated()
3373 spin_unlock(&pa->pa_lock); ext4_mb_use_preallocated()
3378 spin_unlock(&pa->pa_lock); ext4_mb_use_preallocated()
3402 list_for_each_entry_rcu(pa, &lg->lg_prealloc_list[i], ext4_mb_use_preallocated()
3404 spin_lock(&pa->pa_lock); ext4_mb_use_preallocated()
3405 if (pa->pa_deleted == 0 && ext4_mb_use_preallocated()
3406 pa->pa_free >= ac->ac_o_ex.fe_len) { ext4_mb_use_preallocated()
3409 pa, cpa); ext4_mb_use_preallocated()
3411 spin_unlock(&pa->pa_lock); ext4_mb_use_preallocated()
3457 struct ext4_prealloc_space *pa; ext4_mb_generate_from_pa() local
3473 pa = list_entry(cur, struct ext4_prealloc_space, pa_group_list); ext4_mb_generate_from_pa()
3474 spin_lock(&pa->pa_lock); ext4_mb_generate_from_pa()
3475 ext4_get_group_no_and_offset(sb, pa->pa_pstart, ext4_mb_generate_from_pa()
3477 len = pa->pa_len; ext4_mb_generate_from_pa()
3478 spin_unlock(&pa->pa_lock); ext4_mb_generate_from_pa()
3490 struct ext4_prealloc_space *pa; ext4_mb_pa_callback() local
3491 pa = container_of(head, struct ext4_prealloc_space, u.pa_rcu); ext4_mb_pa_callback()
3493 BUG_ON(atomic_read(&pa->pa_count)); ext4_mb_pa_callback()
3494 BUG_ON(pa->pa_deleted == 0); ext4_mb_pa_callback()
3495 kmem_cache_free(ext4_pspace_cachep, pa); ext4_mb_pa_callback()
3503 struct super_block *sb, struct ext4_prealloc_space *pa) ext4_mb_put_pa()
3509 spin_lock(&pa->pa_lock); ext4_mb_put_pa()
3510 if (!atomic_dec_and_test(&pa->pa_count) || pa->pa_free != 0) { ext4_mb_put_pa()
3511 spin_unlock(&pa->pa_lock); ext4_mb_put_pa()
3515 if (pa->pa_deleted == 1) { ext4_mb_put_pa()
3516 spin_unlock(&pa->pa_lock); ext4_mb_put_pa()
3520 pa->pa_deleted = 1; ext4_mb_put_pa()
3521 spin_unlock(&pa->pa_lock); ext4_mb_put_pa()
3523 grp_blk = pa->pa_pstart; ext4_mb_put_pa()
3526 * next group when pa is used up ext4_mb_put_pa()
3528 if (pa->pa_type == MB_GROUP_PA) ext4_mb_put_pa()
3548 list_del(&pa->pa_group_list); ext4_mb_put_pa()
3551 spin_lock(pa->pa_obj_lock); ext4_mb_put_pa()
3552 list_del_rcu(&pa->pa_inode_list); ext4_mb_put_pa()
3553 spin_unlock(pa->pa_obj_lock); ext4_mb_put_pa()
3555 call_rcu(&(pa)->u.pa_rcu, ext4_mb_pa_callback); ext4_mb_put_pa()
3566 struct ext4_prealloc_space *pa; ext4_mb_new_inode_pa() local
3575 pa = kmem_cache_alloc(ext4_pspace_cachep, GFP_NOFS); ext4_mb_new_inode_pa()
3576 if (pa == NULL) ext4_mb_new_inode_pa()
3617 pa->pa_lstart = ac->ac_b_ex.fe_logical; ext4_mb_new_inode_pa()
3618 pa->pa_pstart = ext4_grp_offs_to_block(sb, &ac->ac_b_ex); ext4_mb_new_inode_pa()
3619 pa->pa_len = ac->ac_b_ex.fe_len; ext4_mb_new_inode_pa()
3620 pa->pa_free = pa->pa_len; ext4_mb_new_inode_pa()
3621 atomic_set(&pa->pa_count, 1); ext4_mb_new_inode_pa()
3622 spin_lock_init(&pa->pa_lock); ext4_mb_new_inode_pa()
3623 INIT_LIST_HEAD(&pa->pa_inode_list); ext4_mb_new_inode_pa()
3624 INIT_LIST_HEAD(&pa->pa_group_list); ext4_mb_new_inode_pa()
3625 pa->pa_deleted = 0; ext4_mb_new_inode_pa()
3626 pa->pa_type = MB_INODE_PA; ext4_mb_new_inode_pa()
3628 mb_debug(1, "new inode pa %p: %llu/%u for %u\n", pa, ext4_mb_new_inode_pa()
3629 pa->pa_pstart, pa->pa_len, pa->pa_lstart); ext4_mb_new_inode_pa()
3630 trace_ext4_mb_new_inode_pa(ac, pa); ext4_mb_new_inode_pa()
3632 ext4_mb_use_inode_pa(ac, pa); ext4_mb_new_inode_pa()
3633 atomic_add(pa->pa_free, &sbi->s_mb_preallocated); ext4_mb_new_inode_pa()
3638 pa->pa_obj_lock = &ei->i_prealloc_lock; ext4_mb_new_inode_pa()
3639 pa->pa_inode = ac->ac_inode; ext4_mb_new_inode_pa()
3642 list_add(&pa->pa_group_list, &grp->bb_prealloc_list); ext4_mb_new_inode_pa()
3645 spin_lock(pa->pa_obj_lock); ext4_mb_new_inode_pa()
3646 list_add_rcu(&pa->pa_inode_list, &ei->i_prealloc_list); ext4_mb_new_inode_pa()
3647 spin_unlock(pa->pa_obj_lock); ext4_mb_new_inode_pa()
3660 struct ext4_prealloc_space *pa; ext4_mb_new_group_pa() local
3669 pa = kmem_cache_alloc(ext4_pspace_cachep, GFP_NOFS); ext4_mb_new_group_pa()
3670 if (pa == NULL) ext4_mb_new_group_pa()
3677 pa->pa_pstart = ext4_grp_offs_to_block(sb, &ac->ac_b_ex); ext4_mb_new_group_pa()
3678 pa->pa_lstart = pa->pa_pstart; ext4_mb_new_group_pa()
3679 pa->pa_len = ac->ac_b_ex.fe_len; ext4_mb_new_group_pa()
3680 pa->pa_free = pa->pa_len; ext4_mb_new_group_pa()
3681 atomic_set(&pa->pa_count, 1); ext4_mb_new_group_pa()
3682 spin_lock_init(&pa->pa_lock); ext4_mb_new_group_pa()
3683 INIT_LIST_HEAD(&pa->pa_inode_list); ext4_mb_new_group_pa()
3684 INIT_LIST_HEAD(&pa->pa_group_list); ext4_mb_new_group_pa()
3685 pa->pa_deleted = 0; ext4_mb_new_group_pa()
3686 pa->pa_type = MB_GROUP_PA; ext4_mb_new_group_pa()
3688 mb_debug(1, "new group pa %p: %llu/%u for %u\n", pa, ext4_mb_new_group_pa()
3689 pa->pa_pstart, pa->pa_len, pa->pa_lstart); ext4_mb_new_group_pa()
3690 trace_ext4_mb_new_group_pa(ac, pa); ext4_mb_new_group_pa()
3692 ext4_mb_use_group_pa(ac, pa); ext4_mb_new_group_pa()
3693 atomic_add(pa->pa_free, &EXT4_SB(sb)->s_mb_preallocated); ext4_mb_new_group_pa()
3699 pa->pa_obj_lock = &lg->lg_prealloc_lock; ext4_mb_new_group_pa()
3700 pa->pa_inode = NULL; ext4_mb_new_group_pa()
3703 list_add(&pa->pa_group_list, &grp->bb_prealloc_list); ext4_mb_new_group_pa()
3707 * We will later add the new pa to the right bucket ext4_mb_new_group_pa()
3727 * @pa must be unlinked from inode and group lists, so that
3734 struct ext4_prealloc_space *pa) ext4_mb_release_inode_pa()
3746 BUG_ON(pa->pa_deleted == 0); ext4_mb_release_inode_pa()
3747 ext4_get_group_no_and_offset(sb, pa->pa_pstart, &group, &bit); ext4_mb_release_inode_pa()
3748 grp_blk_start = pa->pa_pstart - EXT4_C2B(sbi, bit); ext4_mb_release_inode_pa()
3749 BUG_ON(group != e4b->bd_group && pa->pa_len != 0); ext4_mb_release_inode_pa()
3750 end = bit + pa->pa_len; ext4_mb_release_inode_pa()
3763 trace_ext4_mb_release_inode_pa(pa, (grp_blk_start + ext4_mb_release_inode_pa()
3766 mb_free_blocks(pa->pa_inode, e4b, bit, next - bit); ext4_mb_release_inode_pa()
3769 if (free != pa->pa_free) { ext4_mb_release_inode_pa()
3771 "pa %p: logic %lu, phys. %lu, len %lu", ext4_mb_release_inode_pa()
3772 pa, (unsigned long) pa->pa_lstart, ext4_mb_release_inode_pa()
3773 (unsigned long) pa->pa_pstart, ext4_mb_release_inode_pa()
3774 (unsigned long) pa->pa_len); ext4_mb_release_inode_pa()
3776 free, pa->pa_free); ext4_mb_release_inode_pa()
3778 * pa is already deleted so we use the value obtained ext4_mb_release_inode_pa()
3789 struct ext4_prealloc_space *pa) ext4_mb_release_group_pa()
3795 trace_ext4_mb_release_group_pa(sb, pa); ext4_mb_release_group_pa()
3796 BUG_ON(pa->pa_deleted == 0); ext4_mb_release_group_pa()
3797 ext4_get_group_no_and_offset(sb, pa->pa_pstart, &group, &bit); ext4_mb_release_group_pa()
3798 BUG_ON(group != e4b->bd_group && pa->pa_len != 0); ext4_mb_release_group_pa()
3799 mb_free_blocks(pa->pa_inode, e4b, bit, pa->pa_len); ext4_mb_release_group_pa()
3800 atomic_add(pa->pa_len, &EXT4_SB(sb)->s_mb_discarded); ext4_mb_release_group_pa()
3801 trace_ext4_mballoc_discard(sb, NULL, group, bit, pa->pa_len); ext4_mb_release_group_pa()
3821 struct ext4_prealloc_space *pa, *tmp; ext4_mb_discard_group_preallocations() local
3852 list_for_each_entry_safe(pa, tmp, ext4_mb_discard_group_preallocations()
3854 spin_lock(&pa->pa_lock); ext4_mb_discard_group_preallocations()
3855 if (atomic_read(&pa->pa_count)) { ext4_mb_discard_group_preallocations()
3856 spin_unlock(&pa->pa_lock); ext4_mb_discard_group_preallocations()
3860 if (pa->pa_deleted) { ext4_mb_discard_group_preallocations()
3861 spin_unlock(&pa->pa_lock); ext4_mb_discard_group_preallocations()
3866 pa->pa_deleted = 1; ext4_mb_discard_group_preallocations()
3869 free += pa->pa_free; ext4_mb_discard_group_preallocations()
3871 spin_unlock(&pa->pa_lock); ext4_mb_discard_group_preallocations()
3873 list_del(&pa->pa_group_list); ext4_mb_discard_group_preallocations()
3874 list_add(&pa->u.pa_tmp_list, &list); ext4_mb_discard_group_preallocations()
3892 list_for_each_entry_safe(pa, tmp, &list, u.pa_tmp_list) { ext4_mb_discard_group_preallocations()
3895 spin_lock(pa->pa_obj_lock); ext4_mb_discard_group_preallocations()
3896 list_del_rcu(&pa->pa_inode_list); ext4_mb_discard_group_preallocations()
3897 spin_unlock(pa->pa_obj_lock); ext4_mb_discard_group_preallocations()
3899 if (pa->pa_type == MB_GROUP_PA) ext4_mb_discard_group_preallocations()
3900 ext4_mb_release_group_pa(&e4b, pa); ext4_mb_discard_group_preallocations()
3902 ext4_mb_release_inode_pa(&e4b, bitmap_bh, pa); ext4_mb_discard_group_preallocations()
3904 list_del(&pa->u.pa_tmp_list); ext4_mb_discard_group_preallocations()
3905 call_rcu(&(pa)->u.pa_rcu, ext4_mb_pa_callback); ext4_mb_discard_group_preallocations()
3929 struct ext4_prealloc_space *pa, *tmp; ext4_discard_preallocations() local
3946 /* first, collect all pa's in the inode */ ext4_discard_preallocations()
3949 pa = list_entry(ei->i_prealloc_list.next, ext4_discard_preallocations()
3951 BUG_ON(pa->pa_obj_lock != &ei->i_prealloc_lock); ext4_discard_preallocations()
3952 spin_lock(&pa->pa_lock); ext4_discard_preallocations()
3953 if (atomic_read(&pa->pa_count)) { ext4_discard_preallocations()
3956 spin_unlock(&pa->pa_lock); ext4_discard_preallocations()
3959 "uh-oh! used pa while discarding"); ext4_discard_preallocations()
3965 if (pa->pa_deleted == 0) { ext4_discard_preallocations()
3966 pa->pa_deleted = 1; ext4_discard_preallocations()
3967 spin_unlock(&pa->pa_lock); ext4_discard_preallocations()
3968 list_del_rcu(&pa->pa_inode_list); ext4_discard_preallocations()
3969 list_add(&pa->u.pa_tmp_list, &list); ext4_discard_preallocations()
3973 /* someone is deleting pa right now */ ext4_discard_preallocations()
3974 spin_unlock(&pa->pa_lock); ext4_discard_preallocations()
3978 * doesn't mean pa is already unlinked from ext4_discard_preallocations()
3982 * pa from inode's list may access already ext4_discard_preallocations()
3994 list_for_each_entry_safe(pa, tmp, &list, u.pa_tmp_list) { ext4_discard_preallocations()
3995 BUG_ON(pa->pa_type != MB_INODE_PA); ext4_discard_preallocations()
3996 group = ext4_get_group_number(sb, pa->pa_pstart); ext4_discard_preallocations()
4014 list_del(&pa->pa_group_list); ext4_discard_preallocations()
4015 ext4_mb_release_inode_pa(&e4b, bitmap_bh, pa); ext4_discard_preallocations()
4021 list_del(&pa->u.pa_tmp_list); ext4_discard_preallocations()
4022 call_rcu(&(pa)->u.pa_rcu, ext4_mb_pa_callback); ext4_discard_preallocations()
4061 struct ext4_prealloc_space *pa; ext4_mb_show_ac() local
4066 pa = list_entry(cur, struct ext4_prealloc_space, ext4_mb_show_ac()
4068 spin_lock(&pa->pa_lock); ext4_mb_show_ac()
4069 ext4_get_group_no_and_offset(sb, pa->pa_pstart, ext4_mb_show_ac()
4071 spin_unlock(&pa->pa_lock); ext4_mb_show_ac()
4073 start, pa->pa_len); ext4_mb_show_ac()
4209 struct ext4_prealloc_space *pa, *tmp; ext4_mb_discard_lg_preallocations() local
4216 list_for_each_entry_rcu(pa, &lg->lg_prealloc_list[order], ext4_mb_discard_lg_preallocations()
4218 spin_lock(&pa->pa_lock); ext4_mb_discard_lg_preallocations()
4219 if (atomic_read(&pa->pa_count)) { ext4_mb_discard_lg_preallocations()
4221 * This is the pa that we just used ext4_mb_discard_lg_preallocations()
4225 spin_unlock(&pa->pa_lock); ext4_mb_discard_lg_preallocations()
4228 if (pa->pa_deleted) { ext4_mb_discard_lg_preallocations()
4229 spin_unlock(&pa->pa_lock); ext4_mb_discard_lg_preallocations()
4233 BUG_ON(pa->pa_type != MB_GROUP_PA); ext4_mb_discard_lg_preallocations()
4236 pa->pa_deleted = 1; ext4_mb_discard_lg_preallocations()
4237 spin_unlock(&pa->pa_lock); ext4_mb_discard_lg_preallocations()
4239 list_del_rcu(&pa->pa_inode_list); ext4_mb_discard_lg_preallocations()
4240 list_add(&pa->u.pa_tmp_list, &discard_list); ext4_mb_discard_lg_preallocations()
4255 list_for_each_entry_safe(pa, tmp, &discard_list, u.pa_tmp_list) { ext4_mb_discard_lg_preallocations()
4257 group = ext4_get_group_number(sb, pa->pa_pstart); ext4_mb_discard_lg_preallocations()
4264 list_del(&pa->pa_group_list); ext4_mb_discard_lg_preallocations()
4265 ext4_mb_release_group_pa(&e4b, pa); ext4_mb_discard_lg_preallocations()
4269 list_del(&pa->u.pa_tmp_list); ext4_mb_discard_lg_preallocations()
4270 call_rcu(&(pa)->u.pa_rcu, ext4_mb_pa_callback); ext4_mb_discard_lg_preallocations()
4288 struct ext4_prealloc_space *tmp_pa, *pa = ac->ac_pa; ext4_mb_add_n_trim() local
4290 order = fls(pa->pa_free) - 1; ext4_mb_add_n_trim()
4303 if (!added && pa->pa_free < tmp_pa->pa_free) { ext4_mb_add_n_trim()
4305 list_add_tail_rcu(&pa->pa_inode_list, ext4_mb_add_n_trim()
4317 list_add_tail_rcu(&pa->pa_inode_list, ext4_mb_add_n_trim()
4336 struct ext4_prealloc_space *pa = ac->ac_pa; ext4_mb_release_context() local
4337 if (pa) { ext4_mb_release_context()
4338 if (pa->pa_type == MB_GROUP_PA) { ext4_mb_release_context()
4340 spin_lock(&pa->pa_lock); ext4_mb_release_context()
4341 pa->pa_pstart += EXT4_C2B(sbi, ac->ac_b_ex.fe_len); ext4_mb_release_context()
4342 pa->pa_lstart += EXT4_C2B(sbi, ac->ac_b_ex.fe_len); ext4_mb_release_context()
4343 pa->pa_free -= ac->ac_b_ex.fe_len; ext4_mb_release_context()
4344 pa->pa_len -= ac->ac_b_ex.fe_len; ext4_mb_release_context()
4345 spin_unlock(&pa->pa_lock); ext4_mb_release_context()
4348 if (pa) { ext4_mb_release_context()
4350 * We want to add the pa to the right bucket. ext4_mb_release_context()
4355 if ((pa->pa_type == MB_GROUP_PA) && likely(pa->pa_free)) { ext4_mb_release_context()
4356 spin_lock(pa->pa_obj_lock); ext4_mb_release_context()
4357 list_del_rcu(&pa->pa_inode_list); ext4_mb_release_context()
4358 spin_unlock(pa->pa_obj_lock); ext4_mb_release_context()
4361 ext4_mb_put_pa(ac, ac->ac_sb, pa); ext4_mb_release_context()
3254 ext4_mb_use_inode_pa(struct ext4_allocation_context *ac, struct ext4_prealloc_space *pa) ext4_mb_use_inode_pa() argument
3284 ext4_mb_use_group_pa(struct ext4_allocation_context *ac, struct ext4_prealloc_space *pa) ext4_mb_use_group_pa() argument
3312 ext4_mb_check_group_pa(ext4_fsblk_t goal_block, struct ext4_prealloc_space *pa, struct ext4_prealloc_space *cpa) ext4_mb_check_group_pa() argument
3502 ext4_mb_put_pa(struct ext4_allocation_context *ac, struct super_block *sb, struct ext4_prealloc_space *pa) ext4_mb_put_pa() argument
3733 ext4_mb_release_inode_pa(struct ext4_buddy *e4b, struct buffer_head *bitmap_bh, struct ext4_prealloc_space *pa) ext4_mb_release_inode_pa() argument
3788 ext4_mb_release_group_pa(struct ext4_buddy *e4b, struct ext4_prealloc_space *pa) ext4_mb_release_group_pa() argument
H A Dmballoc.h127 unsigned short pa_type; /* pa type. inode or group */
/linux-4.1.27/sound/pci/asihpi/
H A Dhpioctl.c181 struct hpi_adapter *pa = NULL; asihpi_hpi_ioctl() local
184 pa = &adapters[hm->h.adapter_index]; asihpi_hpi_ioctl()
186 if (!pa || !pa->adapter || !pa->adapter->type) { asihpi_hpi_ioctl()
199 if (mutex_lock_interruptible(&pa->mutex)) { asihpi_hpi_ioctl()
216 if (pa->buffer_size < size) { asihpi_hpi_ioctl()
221 pa->buffer_size, size); asihpi_hpi_ioctl()
222 if (pa->p_buffer) { asihpi_hpi_ioctl()
223 pa->buffer_size = 0; asihpi_hpi_ioctl()
224 vfree(pa->p_buffer); asihpi_hpi_ioctl()
226 pa->p_buffer = vmalloc(size); asihpi_hpi_ioctl()
227 if (pa->p_buffer) asihpi_hpi_ioctl()
228 pa->buffer_size = size; asihpi_hpi_ioctl()
235 mutex_unlock(&pa->mutex); asihpi_hpi_ioctl()
241 hm->m0.u.d.u.data.pb_data = pa->p_buffer; asihpi_hpi_ioctl()
257 copy_from_user(pa->p_buffer, ptr, size); asihpi_hpi_ioctl()
269 copy_to_user(ptr, pa->p_buffer, size); asihpi_hpi_ioctl()
276 mutex_unlock(&pa->mutex); asihpi_hpi_ioctl()
523 struct hpi_adapter *pa; asihpi_adapter_remove() local
526 pa = pci_get_drvdata(pci_dev); asihpi_adapter_remove()
527 pci = pa->adapter->pci; asihpi_adapter_remove()
532 hm.adapter_index = pa->adapter->index; asihpi_adapter_remove()
540 hm.adapter_index = pa->adapter->index; asihpi_adapter_remove()
547 if (pa->irq) asihpi_adapter_remove()
548 free_irq(pa->irq, pa); asihpi_adapter_remove()
550 vfree(pa->p_buffer); asihpi_adapter_remove()
557 pci_dev->devfn, pa->adapter->index); asihpi_adapter_remove()
559 memset(pa, 0, sizeof(*pa)); asihpi_adapter_remove()
/linux-4.1.27/arch/mips/include/asm/mach-ip27/
H A Ddma-coherence.h24 dma_addr_t pa = dev_to_baddr(dev, virt_to_phys(addr)); plat_map_dma_mem() local
26 return pa; plat_map_dma_mem()
32 dma_addr_t pa = dev_to_baddr(dev, page_to_phys(page)); plat_map_dma_mem_page() local
34 return pa; plat_map_dma_mem_page()
/linux-4.1.27/drivers/infiniband/hw/usnic/
H A Dusnic_uiom.c85 dma_addr_t pa; usnic_uiom_put_pages() local
90 pa = sg_phys(sg); list_for_each_entry_safe()
94 usnic_dbg("pa: %pa\n", &pa); list_for_each_entry_safe()
114 dma_addr_t pa; usnic_uiom_get_pages() local
173 pa = sg_phys(sg); usnic_uiom_get_pages()
174 usnic_dbg("va: 0x%lx pa: %pa\n", usnic_uiom_get_pages()
175 cur_base + i*PAGE_SIZE, &pa); usnic_uiom_get_pages()
253 dma_addr_t pa; usnic_uiom_map_sorted_intervals() local
267 pa = sg_phys(&chunk->page_list[i]); list_for_each_entry()
274 pa_start = pa; list_for_each_entry()
275 pa_end = pa; list_for_each_entry()
280 if ((pa_end + PAGE_SIZE != pa) && list_for_each_entry()
281 (pa != pa_start)) { list_for_each_entry()
284 usnic_dbg("va 0x%lx pa %pa size 0x%zx flags 0x%x", list_for_each_entry()
289 usnic_err("Failed to map va 0x%lx pa %pa size 0x%zx with err %d\n", list_for_each_entry()
294 pa_start = pa; list_for_each_entry()
295 pa_end = pa; list_for_each_entry()
300 size = pa - pa_start + PAGE_SIZE; list_for_each_entry()
301 usnic_dbg("va 0x%lx pa %pa size 0x%zx flags 0x%x\n", list_for_each_entry()
306 usnic_err("Failed to map va 0x%lx pa %pa size 0x%zx with err %d\n", list_for_each_entry()
313 if (pa != pa_start) list_for_each_entry()
/linux-4.1.27/arch/mips/bmips/
H A Ddma.c43 static dma_addr_t bmips_phys_to_dma(struct device *dev, phys_addr_t pa) bmips_phys_to_dma() argument
48 if (pa >= r->child_addr && bmips_phys_to_dma()
49 pa < (r->child_addr + r->size)) bmips_phys_to_dma()
50 return pa - r->child_addr + r->parent_addr; bmips_phys_to_dma()
52 return pa; bmips_phys_to_dma()
/linux-4.1.27/arch/x86/kernel/
H A Dhead_32.S28 #define pa(X) ((X) - __PAGE_OFFSET) define
55 * KERNEL_IMAGE_SIZE should be greater than pa(_end)
97 movl pa(stack_start),%ecx
107 lgdt pa(boot_gdt_descr)
122 movl $pa(__bss_start),%edi
123 movl $pa(__bss_stop),%ecx
135 movl $pa(boot_params),%edi
140 movl pa(boot_params) + NEW_CL_POINTER,%esi
143 movl $pa(boot_command_line),%edi
152 movl %eax, pa(olpc_ofw_pgd)
182 movl $pa(__brk_base), %edi
183 movl $pa(initial_pg_pmd), %edx
202 movl $pa(_end) + MAPPING_BEYOND_END + PTE_IDENT_ATTR, %ebp
207 movl %edi, pa(_brk_end)
209 movl %eax, pa(max_pfn_mapped)
212 movl $pa(initial_pg_fixmap)+PDE_IDENT_ATTR,%eax
213 movl %eax,pa(initial_pg_pmd+0x1000*KPMDS-8)
218 movl $pa(__brk_base), %edi
219 movl $pa(initial_page_table), %edx
234 movl $pa(_end) + MAPPING_BEYOND_END + PTE_IDENT_ATTR, %ebp
238 movl %edi, pa(_brk_end)
240 movl %eax, pa(max_pfn_mapped)
243 movl $pa(initial_pg_fixmap)+PDE_IDENT_ATTR,%eax
244 movl %eax,pa(initial_page_table+0xffc)
249 cmpw $0x207, pa(boot_params + BP_version)
254 movl pa(boot_params + BP_hardware_subarch), %eax
258 movl pa(subarch_entries)(,%eax,4), %eax
310 movl pa(stack_start),%ecx
343 movl $-1,pa(X86_CPUID) # preset CPUID level
355 movl %eax,pa(X86_CPUID) # save largest std CPUID function
362 movl pa(mmu_cr4_features),%eax
398 movl $pa(initial_page_table), %eax
691 .long pa(initial_pg_pmd+PGD_IDENT_ATTR),0 /* low identity map */
693 .long pa(initial_pg_pmd+PGD_IDENT_ATTR),0
694 .long pa(initial_pg_pmd+PGD_IDENT_ATTR+0x1000),0
695 .long pa(initial_pg_pmd+PGD_IDENT_ATTR+0x2000),0
698 .long pa(initial_pg_pmd+PGD_IDENT_ATTR),0
699 .long pa(initial_pg_pmd+PGD_IDENT_ATTR+0x1000),0
703 .long pa(initial_pg_pmd+PGD_IDENT_ATTR),0
H A Dkdebugfs.c38 u64 pa; setup_data_read() local
49 pa = node->paddr + sizeof(struct setup_data) + pos; setup_data_read()
50 pg = pfn_to_page((pa + count - 1) >> PAGE_SHIFT); setup_data_read()
52 p = ioremap_cache(pa, count); setup_data_read()
56 p = __va(pa); setup_data_read()
H A Dkvm.c340 u64 pa = slow_virt_to_phys(this_cpu_ptr(&apf_reason)); kvm_guest_cpu_init() local
343 pa |= KVM_ASYNC_PF_SEND_ALWAYS; kvm_guest_cpu_init()
345 wrmsrl(MSR_KVM_ASYNC_PF_EN, pa | KVM_ASYNC_PF_ENABLED); kvm_guest_cpu_init()
352 unsigned long pa; kvm_guest_cpu_init() local
356 pa = slow_virt_to_phys(this_cpu_ptr(&kvm_apic_eoi)) kvm_guest_cpu_init()
358 wrmsrl(MSR_KVM_PV_EOI_EN, pa); kvm_guest_cpu_init()
/linux-4.1.27/fs/
H A Dposix_acl.c176 const struct posix_acl_entry *pa, *pe; posix_acl_valid() local
180 FOREACH_ACL_ENTRY(pa, acl, pe) { FOREACH_ACL_ENTRY()
181 if (pa->e_perm & ~(ACL_READ|ACL_WRITE|ACL_EXECUTE)) FOREACH_ACL_ENTRY()
183 switch (pa->e_tag) { FOREACH_ACL_ENTRY()
194 if (!uid_valid(pa->e_uid)) FOREACH_ACL_ENTRY()
209 if (!gid_valid(pa->e_gid)) FOREACH_ACL_ENTRY()
245 const struct posix_acl_entry *pa, *pe; posix_acl_equiv_mode() local
255 FOREACH_ACL_ENTRY(pa, acl, pe) { FOREACH_ACL_ENTRY()
256 switch (pa->e_tag) { FOREACH_ACL_ENTRY()
258 mode |= (pa->e_perm & S_IRWXO) << 6; FOREACH_ACL_ENTRY()
261 mode |= (pa->e_perm & S_IRWXO) << 3; FOREACH_ACL_ENTRY()
264 mode |= pa->e_perm & S_IRWXO; FOREACH_ACL_ENTRY()
268 ((pa->e_perm & S_IRWXO) << 3); FOREACH_ACL_ENTRY()
314 const struct posix_acl_entry *pa, *pe, *mask_obj; posix_acl_permission() local
319 FOREACH_ACL_ENTRY(pa, acl, pe) { FOREACH_ACL_ENTRY()
320 switch(pa->e_tag) { FOREACH_ACL_ENTRY()
327 if (uid_eq(pa->e_uid, current_fsuid())) FOREACH_ACL_ENTRY()
333 if ((pa->e_perm & want) == want) FOREACH_ACL_ENTRY()
338 if (in_group_p(pa->e_gid)) { FOREACH_ACL_ENTRY()
340 if ((pa->e_perm & want) == want) FOREACH_ACL_ENTRY()
358 for (mask_obj = pa+1; mask_obj != pe; mask_obj++) {
360 if ((pa->e_perm & mask_obj->e_perm & want) == want)
367 if ((pa->e_perm & want) == want)
382 struct posix_acl_entry *pa, *pe; posix_acl_create_masq() local
389 FOREACH_ACL_ENTRY(pa, acl, pe) { FOREACH_ACL_ENTRY()
390 switch(pa->e_tag) { FOREACH_ACL_ENTRY()
392 pa->e_perm &= (mode >> 6) | ~S_IRWXO; FOREACH_ACL_ENTRY()
393 mode &= (pa->e_perm << 6) | ~S_IRWXU; FOREACH_ACL_ENTRY()
402 group_obj = pa; FOREACH_ACL_ENTRY()
406 pa->e_perm &= mode | ~S_IRWXO; FOREACH_ACL_ENTRY()
407 mode &= pa->e_perm | ~S_IRWXO; FOREACH_ACL_ENTRY()
411 mask_obj = pa; FOREACH_ACL_ENTRY()
440 struct posix_acl_entry *pa, *pe; __posix_acl_chmod_masq() local
444 FOREACH_ACL_ENTRY(pa, acl, pe) { FOREACH_ACL_ENTRY()
445 switch(pa->e_tag) { FOREACH_ACL_ENTRY()
447 pa->e_perm = (mode & S_IRWXU) >> 6; FOREACH_ACL_ENTRY()
455 group_obj = pa; FOREACH_ACL_ENTRY()
459 mask_obj = pa; FOREACH_ACL_ENTRY()
463 pa->e_perm = (mode & S_IRWXO); FOREACH_ACL_ENTRY()
/linux-4.1.27/drivers/isdn/hardware/eicon/
H A Ddiva.c169 diva_os_xdi_adapter_t *pdiva, *pa; diva_driver_add_card() local
207 pa = pdiva; diva_driver_add_card()
209 pa = diva_q_get_next(&pa->link); diva_driver_add_card()
210 if (pa && !pa->interface.cleanup_adapter_proc) { diva_driver_add_card()
211 pa->controller = i + 1 + j; diva_driver_add_card()
212 pa->xdi_adapter.ANum = pa->controller; diva_driver_add_card()
213 IoAdapters[i + j] = &pa->xdi_adapter; diva_driver_add_card()
216 pa->controller)) diva_driver_add_card()
217 create_adapter_proc(pa); /* add adapter to proc file system */ diva_driver_add_card()
283 diva_os_xdi_adapter_t *pa; diva_driver_remove_card() local
286 pa = a[0] = (diva_os_xdi_adapter_t *) pdiva; diva_driver_remove_card()
292 if ((pa = diva_q_get_next(&pa->link)) diva_driver_remove_card()
293 && !pa->interface.cleanup_adapter_proc) { diva_driver_remove_card()
294 a[i] = pa; diva_driver_remove_card()
/linux-4.1.27/fs/nfsd/
H A Dnfs4acl.c201 struct posix_acl_entry *pa, *pe; summarize_posix_acl() local
213 FOREACH_ACL_ENTRY(pa, acl, pe) { FOREACH_ACL_ENTRY()
214 switch (pa->e_tag) { FOREACH_ACL_ENTRY()
216 pas->owner = pa->e_perm; FOREACH_ACL_ENTRY()
219 pas->group = pa->e_perm; FOREACH_ACL_ENTRY()
222 pas->users |= pa->e_perm; FOREACH_ACL_ENTRY()
225 pas->groups |= pa->e_perm; FOREACH_ACL_ENTRY()
228 pas->other = pa->e_perm; FOREACH_ACL_ENTRY()
231 pas->mask = pa->e_perm; FOREACH_ACL_ENTRY()
246 struct posix_acl_entry *pa, *group_owner_entry; _posix_to_nfsv4_one() local
256 pa = pacl->a_entries; _posix_to_nfsv4_one()
277 ace->access_mask = mask_from_posix(pa->e_perm, flags | NFS4_ACL_OWNER); _posix_to_nfsv4_one()
281 pa++; _posix_to_nfsv4_one()
283 while (pa->e_tag == ACL_USER) { _posix_to_nfsv4_one()
284 deny = ~(pa->e_perm & pas.mask); _posix_to_nfsv4_one()
291 ace->who_uid = pa->e_uid; _posix_to_nfsv4_one()
297 ace->access_mask = mask_from_posix(pa->e_perm & pas.mask, _posix_to_nfsv4_one()
300 ace->who_uid = pa->e_uid; _posix_to_nfsv4_one()
303 pa++; _posix_to_nfsv4_one()
311 group_owner_entry = pa; _posix_to_nfsv4_one()
319 pa++; _posix_to_nfsv4_one()
321 while (pa->e_tag == ACL_GROUP) { _posix_to_nfsv4_one()
324 ace->access_mask = mask_from_posix(pa->e_perm & pas.mask, _posix_to_nfsv4_one()
327 ace->who_gid = pa->e_gid; _posix_to_nfsv4_one()
330 pa++; _posix_to_nfsv4_one()
335 pa = group_owner_entry; _posix_to_nfsv4_one()
346 pa++; _posix_to_nfsv4_one()
348 while (pa->e_tag == ACL_GROUP) { _posix_to_nfsv4_one()
349 deny = ~(pa->e_perm & pas.mask); _posix_to_nfsv4_one()
356 ace->who_gid = pa->e_gid; _posix_to_nfsv4_one()
360 pa++; _posix_to_nfsv4_one()
363 if (pa->e_tag == ACL_MASK) _posix_to_nfsv4_one()
364 pa++; _posix_to_nfsv4_one()
367 ace->access_mask = mask_from_posix(pa->e_perm, flags); _posix_to_nfsv4_one()
/linux-4.1.27/drivers/net/fddi/skfp/
H A Dpmf.c39 static int smt_set_para(struct s_smc *smc, struct smt_para *pa, int index,
316 struct smt_para *pa ; smt_build_pmf_response() local
373 pa = (struct smt_para *) (req + 1) ; smt_build_pmf_response()
378 if (((u_short)len < pa->p_len + PARA_LEN) || (pa->p_len & 3)) { smt_build_pmf_response()
383 if (((range = (pa->p_type & 0xf000)) == 0x2000) || smt_build_pmf_response()
388 index = *((u_char *)pa + PARA_LEN + 3) ;/* index */ smt_build_pmf_response()
390 if (!set && (pa->p_len != 4)) { smt_build_pmf_response()
419 if (!set && (pa->p_len != 0)) { smt_build_pmf_response()
431 pt = smt_get_ptab(pa->p_type) ; smt_build_pmf_response()
446 else if (pa->p_type != SMT_P_AUTHOR && smt_build_pmf_response()
447 (!set || (pa->p_type != SMT_P1035))) { smt_build_pmf_response()
450 smt_add_para(smc,&pcon,pa->p_type, smt_build_pmf_response()
454 st = smt_set_para(smc,pa,index,local,1); smt_build_pmf_response()
458 smt_add_para(smc,&pcon,pa->p_type, smt_build_pmf_response()
467 smt_add_para(smc,&pcon,pa->p_type, smt_build_pmf_response()
475 len -= pa->p_len + PARA_LEN ; smt_build_pmf_response()
476 pa = (struct smt_para *) ((char *)pa + pa->p_len + PARA_LEN) ; smt_build_pmf_response()
501 struct smt_para *pa ; smt_authorize() local
523 pa = (struct smt_para *) sm_to_para(smc,sm,SMT_P_AUTHOR) ; smt_authorize()
524 if (!pa) smt_authorize()
526 if (pa->p_len != 8) smt_authorize()
528 if (memcmp((char *)(pa+1),(char *)smc->mib.fddiPRPMFPasswd,8)) smt_authorize()
536 struct smt_para *pa ; smt_check_set_count() local
539 pa = (struct smt_para *) sm_to_para(smc,sm,SMT_P1035) ; smt_check_set_count()
540 if (pa) { smt_check_set_count()
541 sc = (struct smt_p_setcount *) pa ; smt_check_set_count()
553 struct smt_para *pa ; smt_add_para() local
586 pa = (struct smt_para *) to ; /* type/length pointer */ smt_add_para()
1051 pa->p_type = para ; smt_add_para()
1052 pa->p_len = plen - len - PARA_LEN ; smt_add_para()
1075 static int smt_set_para(struct s_smc *smc, struct smt_para *pa, int index, smt_set_para() argument
1101 len = pa->p_len ; smt_set_para()
1102 from = (char *) (pa + 1 ) ; smt_set_para()
1105 switch (pa->p_type & 0xf000) { smt_set_para()
1138 switch (pa->p_type) { smt_set_para()
1158 pt = smt_get_ptab(pa->p_type) ; smt_set_para()
1160 return (pa->p_type & 0xff00) ? SMT_RDF_NOPARAM : smt_set_para()
1299 switch (pa->p_type) { smt_set_para()
1563 struct smt_para *pa ; dump_smt() local
1593 pa = (struct smt_para *) (sm + 1) ; dump_smt()
1597 printf("TYPE %x LEN %x VALUE\t",pa->p_type,pa->p_len) ; dump_smt()
1599 printf("TYPE %04x LEN %2x VALUE\t",pa->p_type,pa->p_len) ; dump_smt()
1601 n = pa->p_len ; dump_smt()
1611 dump_hex((char *)(pa+1),(int) n) ; dump_smt()
1616 c = (char *)(pa+1) ; dump_smt()
1640 plen = (pa->p_len + PARA_LEN + 3) & ~3 ; dump_smt()
1642 pa = (struct smt_para *)((char *)pa + plen) ; dump_smt()
H A Dsmt.c1829 struct smt_para *pa ; smt_swap_para() local
1846 pa = (struct smt_para *) p ; smt_swap_para()
1847 plen = pa->p_len ; smt_swap_para()
1848 type = pa->p_type ; smt_swap_para()
1849 pa->p_type = smt_swap_short(pa->p_type) ; smt_swap_para()
1850 pa->p_len = smt_swap_short(pa->p_len) ; smt_swap_para()
1852 plen = pa->p_len ; smt_swap_para()
1853 type = pa->p_type ; smt_swap_para()
/linux-4.1.27/drivers/misc/mic/host/
H A Dmic_fops.c156 unsigned long *size, unsigned long *pa) mic_query_offset()
171 *pa = virt_to_phys(mdev->dp); mic_query_offset()
179 *pa = virt_to_phys(mvr->vring.va); mic_query_offset()
196 unsigned long pa, size = vma->vm_end - vma->vm_start, size_rem = size; mic_mmap() local
207 i = mic_query_offset(mvdev, offset, &size, &pa); mic_mmap()
211 pa >> PAGE_SHIFT, size, vma->vm_page_prot); mic_mmap()
215 "%s %d type %d size 0x%lx off 0x%lx pa 0x%lx vma 0x%lx\n", mic_mmap()
217 pa, vma->vm_start + offset); mic_mmap()
155 mic_query_offset(struct mic_vdev *mvdev, unsigned long offset, unsigned long *size, unsigned long *pa) mic_query_offset() argument
H A Dmic_smpt.c32 static inline u8 mic_sys_addr_to_smpt(struct mic_device *mdev, dma_addr_t pa) mic_sys_addr_to_smpt() argument
34 return (pa - mdev->smpt->info.base) >> mdev->smpt->info.page_shift; mic_sys_addr_to_smpt()
42 static inline u64 mic_smpt_offset(struct mic_device *mdev, dma_addr_t pa) mic_smpt_offset() argument
44 return pa & mic_system_page_mask(mdev); mic_smpt_offset()
47 static inline u64 mic_smpt_align_low(struct mic_device *mdev, dma_addr_t pa) mic_smpt_align_low() argument
49 return ALIGN(pa - mic_system_page_mask(mdev), mic_smpt_align_low()
53 static inline u64 mic_smpt_align_high(struct mic_device *mdev, dma_addr_t pa) mic_smpt_align_high() argument
55 return ALIGN(pa, mdev->smpt->info.page_size); mic_smpt_align_high()
72 mic_is_system_addr(struct mic_device *mdev, dma_addr_t pa) mic_is_system_addr() argument
74 return pa >= mdev->smpt->info.base && pa <= mic_max_system_addr(mdev); mic_is_system_addr()
H A Dmic_main.c336 mdev->mmio.pa = pci_resource_start(pdev, mdev->ops->mmio_bar); mic_probe()
345 mdev->aper.pa = pci_resource_start(pdev, mdev->ops->aper_bar); mic_probe()
347 mdev->aper.va = ioremap_wc(mdev->aper.pa, mdev->aper.len); mic_probe()
/linux-4.1.27/arch/powerpc/platforms/pseries/
H A Dcmm.c121 struct cmm_page_array *pa, *npa; cmm_alloc_pages() local
144 pa = cmm_page_list; cmm_alloc_pages()
145 if (!pa || pa->index >= CMM_NR_PAGES) { cmm_alloc_pages()
157 pa = cmm_page_list; cmm_alloc_pages()
159 if (!pa || pa->index >= CMM_NR_PAGES) { cmm_alloc_pages()
160 npa->next = pa; cmm_alloc_pages()
162 pa = npa; cmm_alloc_pages()
163 cmm_page_list = pa; cmm_alloc_pages()
175 pa->page[pa->index++] = addr; cmm_alloc_pages()
195 struct cmm_page_array *pa; cmm_free_pages() local
200 pa = cmm_page_list; cmm_free_pages()
202 if (!pa || pa->index <= 0) cmm_free_pages()
204 addr = pa->page[--pa->index]; cmm_free_pages()
206 if (pa->index == 0) { cmm_free_pages()
207 pa = pa->next; cmm_free_pages()
209 cmm_page_list = pa; cmm_free_pages()
474 struct cmm_page_array *pa; cmm_count_pages() local
480 pa = cmm_page_list; cmm_count_pages()
481 while (pa) { cmm_count_pages()
482 if ((unsigned long)pa >= start && (unsigned long)pa < end) cmm_count_pages()
484 for (idx = 0; idx < pa->index; idx++) cmm_count_pages()
485 if (pa->page[idx] >= start && pa->page[idx] < end) cmm_count_pages()
487 pa = pa->next; cmm_count_pages()
H A Dlpar.c127 unsigned long vpn, unsigned long pa, pSeries_lpar_hpte_insert()
138 "pa=%016lx, rflags=%lx, vflags=%lx, psize=%d)\n", pSeries_lpar_hpte_insert()
139 hpte_group, vpn, pa, rflags, vflags, psize); pSeries_lpar_hpte_insert()
142 hpte_r = hpte_encode_r(pa, psize, apsize) | rflags; pSeries_lpar_hpte_insert()
126 pSeries_lpar_hpte_insert(unsigned long hpte_group, unsigned long vpn, unsigned long pa, unsigned long rflags, unsigned long vflags, int psize, int apsize, int ssize) pSeries_lpar_hpte_insert() argument
/linux-4.1.27/arch/mips/kernel/
H A Dspram.c123 unsigned int firstpa = 0, lastpa = 0, pa = 0; probe_spram() local
147 if ((pa == firstpa && size == firstsize) || probe_spram()
148 (pa == lastpa && size == lastsize)) probe_spram()
163 pa = tag0 & SPRAM_TAG0_PA_MASK; probe_spram()
167 firstpa = pa; probe_spram()
171 lastpa = pa; probe_spram()
175 unsigned int *vp = (unsigned int *)(CKSEG1 | pa); probe_spram()
194 type, i, pa, size, enabled ? ",enabled" : ""); probe_spram()
/linux-4.1.27/arch/arm/mach-davinci/include/mach/
H A Dhardware.h31 #define IO_ADDRESS(pa) IOMEM(__IO_ADDRESS(pa))
/linux-4.1.27/include/linux/fsl/bestcomm/
H A Dsram.h47 static inline void *bcom_sram_pa2va(phys_addr_t pa) { bcom_sram_pa2va() argument
49 (unsigned long)(pa - bcom_sram->base_phys); bcom_sram_pa2va()
/linux-4.1.27/arch/alpha/include/asm/
H A Dmmzone.h18 #define alpha_pa_to_nid(pa) \
20 ? alpha_mv.pa_to_nid(pa) \
31 #define pa_to_nid(pa) alpha_pa_to_nid(pa)
H A Dhwrpb.h125 unsigned long pa; member in struct:vf_map_struct
/linux-4.1.27/drivers/dio/
H A Ddio.c123 unsigned long pa; dio_find() local
128 pa = dio_scodetophysaddr(scode); dio_find()
130 if (!pa) dio_find()
134 va = (void *)(pa + DIO_VIRADDRBASE); dio_find()
136 va = ioremap(pa, PAGE_SIZE); dio_find()
202 unsigned long pa; dio_init() local
207 pa = dio_scodetophysaddr(scode); dio_init()
209 if (!pa) dio_init()
213 va = (void *)(pa + DIO_VIRADDRBASE); dio_init()
215 va = ioremap(pa, PAGE_SIZE); dio_init()
238 dev->resource.start = pa; dio_init()
239 dev->resource.end = pa + DIO_SIZE(scode, va); dio_init()
/linux-4.1.27/arch/arm/mach-omap2/
H A Diomap.h34 #define OMAP2_L3_IO_ADDRESS(pa) IOMEM((pa) + OMAP2_L3_IO_OFFSET) /* L3 */
37 #define OMAP2_L4_IO_ADDRESS(pa) IOMEM((pa) + OMAP2_L4_IO_OFFSET) /* L4 */
40 #define OMAP4_L3_IO_ADDRESS(pa) IOMEM((pa) + OMAP4_L3_IO_OFFSET) /* L3 */
43 #define AM33XX_L4_WK_IO_ADDRESS(pa) IOMEM((pa) + AM33XX_L4_WK_IO_OFFSET)
46 #define OMAP4_L3_PER_IO_ADDRESS(pa) IOMEM((pa) + OMAP4_L3_PER_IO_OFFSET)
49 #define OMAP2_EMU_IO_ADDRESS(pa) IOMEM((pa) + OMAP2_EMU_IO_OFFSET)
/linux-4.1.27/arch/s390/pci/
H A Dpci_dma.c142 static int dma_update_trans(struct zpci_dev *zdev, unsigned long pa, dma_update_trans() argument
146 u8 *page_addr = (u8 *) (pa & PAGE_MASK); dma_update_trans()
282 unsigned long pa = page_to_phys(page) + offset; s390_dma_map_pages() local
287 nr_pages = iommu_num_pages(pa, size, PAGE_SIZE); s390_dma_map_pages()
302 if (!dma_update_trans(zdev, pa, dma_addr, size, flags)) { s390_dma_map_pages()
311 zpci_err_hex(&pa, sizeof(pa)); s390_dma_map_pages()
342 unsigned long pa; s390_dma_alloc() local
350 pa = page_to_phys(page); s390_dma_alloc()
351 memset((void *) pa, 0, size); s390_dma_alloc()
353 map = s390_dma_map_pages(dev, page, pa % PAGE_SIZE, s390_dma_alloc()
356 free_pages(pa, get_order(size)); s390_dma_alloc()
363 return (void *) pa; s390_dma_alloc()
367 void *pa, dma_addr_t dma_handle, s390_dma_free()
375 free_pages((unsigned long) pa, get_order(size)); s390_dma_free()
366 s390_dma_free(struct device *dev, size_t size, void *pa, dma_addr_t dma_handle, struct dma_attrs *attrs) s390_dma_free() argument
/linux-4.1.27/include/linux/
H A Dposix_acl.h54 #define FOREACH_ACL_ENTRY(pa, acl, pe) \
55 for(pa=(acl)->a_entries, pe=pa+(acl)->a_count; pa<pe; pa++)
H A Dif_pppox.h33 struct pppoe_addr pa; /* what this socket is bound to*/ member in struct:pppoe_opt
61 #define pppoe_pa proto.pppoe.pa
H A Dremoteproc.h126 * @pa: physical address
157 * overwrite @pa with the dynamically allocated physical address.
168 u32 pa; member in struct:fw_rsc_carveout
178 * @pa: physical address
192 * @da should specify the required device address, @pa should specify
206 u32 pa; member in struct:fw_rsc_devmem
/linux-4.1.27/arch/cris/arch-v32/mach-fs/
H A Dpinmux.c56 reg_pinmux_rw_pa pa = REG_RD(pinmux, regi_pinmux, rw_pa); crisv32_pinmux_init() local
59 pa.pa0 = pa.pa1 = pa.pa2 = pa.pa3 = crisv32_pinmux_init()
60 pa.pa4 = pa.pa5 = pa.pa6 = pa.pa7 = regk_pinmux_yes; crisv32_pinmux_init()
61 REG_WR(pinmux, regi_pinmux, rw_pa, pa); crisv32_pinmux_init()
/linux-4.1.27/drivers/char/tpm/
H A Dtpm_crb.c214 u64 pa; crb_acpi_add() local
266 memcpy_fromio(&pa, &priv->cca->cmd_pa, 8); crb_acpi_add()
267 pa = le64_to_cpu(pa); crb_acpi_add()
268 priv->cmd = devm_ioremap_nocache(dev, pa, crb_acpi_add()
275 memcpy_fromio(&pa, &priv->cca->rsp_pa, 8); crb_acpi_add()
276 pa = le64_to_cpu(pa); crb_acpi_add()
277 priv->rsp = devm_ioremap_nocache(dev, pa, crb_acpi_add()
/linux-4.1.27/arch/arm/xen/
H A Dp2m.c56 pr_warn("%s: cannot add pfn=%pa -> mfn=%pa: pfn=%pa -> mfn=%pa already exists\n", xen_add_phys_to_mach_entry()
H A Denlighten.c182 pr_info("Xen %s support found, events_irq=%d gnttab_frame=%pa\n", xen_guest_init()
/linux-4.1.27/arch/m68k/include/asm/
H A Dsun3xprom.h14 unsigned long sun3x_prom_ptov(unsigned long pa, unsigned long size);
/linux-4.1.27/fs/f2fs/
H A Dacl.c280 struct posix_acl_entry *pa, *pe; f2fs_acl_create_masq() local
287 FOREACH_ACL_ENTRY(pa, acl, pe) { FOREACH_ACL_ENTRY()
288 switch(pa->e_tag) { FOREACH_ACL_ENTRY()
290 pa->e_perm &= (mode >> 6) | ~S_IRWXO; FOREACH_ACL_ENTRY()
291 mode &= (pa->e_perm << 6) | ~S_IRWXU; FOREACH_ACL_ENTRY()
300 group_obj = pa; FOREACH_ACL_ENTRY()
304 pa->e_perm &= mode | ~S_IRWXO; FOREACH_ACL_ENTRY()
305 mode &= pa->e_perm | ~S_IRWXO; FOREACH_ACL_ENTRY()
309 mask_obj = pa; FOREACH_ACL_ENTRY()
/linux-4.1.27/drivers/net/ethernet/amd/xgbe/
H A Dxgbe-desc.c259 struct xgbe_page_alloc *pa, gfp_t gfp, int order) xgbe_alloc_pages()
286 pa->pages = pages; xgbe_alloc_pages()
287 pa->pages_len = PAGE_SIZE << order; xgbe_alloc_pages()
288 pa->pages_offset = 0; xgbe_alloc_pages()
289 pa->pages_dma = pages_dma; xgbe_alloc_pages()
295 struct xgbe_page_alloc *pa, xgbe_set_buffer_data()
298 get_page(pa->pages); xgbe_set_buffer_data()
299 bd->pa = *pa; xgbe_set_buffer_data()
301 bd->dma = pa->pages_dma + pa->pages_offset; xgbe_set_buffer_data()
304 pa->pages_offset += len; xgbe_set_buffer_data()
305 if ((pa->pages_offset + len) > pa->pages_len) { xgbe_set_buffer_data()
307 bd->pa_unmap = *pa; xgbe_set_buffer_data()
310 pa->pages = NULL; xgbe_set_buffer_data()
311 pa->pages_len = 0; xgbe_set_buffer_data()
312 pa->pages_offset = 0; xgbe_set_buffer_data()
313 pa->pages_dma = 0; xgbe_set_buffer_data()
452 if (rdata->rx.hdr.pa.pages) xgbe_unmap_rdata()
453 put_page(rdata->rx.hdr.pa.pages); xgbe_unmap_rdata()
462 if (rdata->rx.buf.pa.pages) xgbe_unmap_rdata()
463 put_page(rdata->rx.buf.pa.pages); xgbe_unmap_rdata()
258 xgbe_alloc_pages(struct xgbe_prv_data *pdata, struct xgbe_page_alloc *pa, gfp_t gfp, int order) xgbe_alloc_pages() argument
294 xgbe_set_buffer_data(struct xgbe_buffer_data *bd, struct xgbe_page_alloc *pa, unsigned int len) xgbe_set_buffer_data() argument
/linux-4.1.27/lib/
H A Didr.c172 static void idr_mark_full(struct idr_layer **pa, int id) idr_mark_full() argument
174 struct idr_layer *p = pa[0]; idr_mark_full()
185 if (!(p = pa[++l])) idr_mark_full()
208 * @pa: idr_layer[MAX_IDR_LEVEL] used as backtrack buffer
220 static int sub_alloc(struct idr *idp, int *starting_id, struct idr_layer **pa, sub_alloc() argument
231 pa[l--] = NULL; sub_alloc()
249 p = pa[l]; sub_alloc()
281 pa[l--] = p; sub_alloc()
285 pa[l] = p; sub_alloc()
290 struct idr_layer **pa, gfp_t gfp_mask, idr_get_empty_slot()
348 v = sub_alloc(idp, &id, pa, gfp_mask, layer_idr); idr_get_empty_slot()
355 * @id and @pa are from a successful allocation from idr_get_empty_slot().
359 struct idr_layer **pa) idr_fill_slot()
362 rcu_assign_pointer(idr->hint, pa[0]); idr_fill_slot()
364 rcu_assign_pointer(pa[0]->ary[id & IDR_MASK], (struct idr_layer *)ptr); idr_fill_slot()
365 pa[0]->count++; idr_fill_slot()
366 idr_mark_full(pa, id); idr_fill_slot()
453 struct idr_layer *pa[MAX_IDR_LEVEL + 1]; idr_alloc() local
465 id = idr_get_empty_slot(idr, start, pa, gfp_mask, NULL); idr_alloc()
471 idr_fill_slot(idr, ptr, id, pa); idr_alloc()
511 struct idr_layer **pa[MAX_IDR_LEVEL + 1]; sub_remove() local
512 struct idr_layer ***paa = &pa[0]; sub_remove()
588 struct idr_layer *pa[MAX_IDR_LEVEL + 1]; __idr_remove_all() local
589 struct idr_layer **paa = &pa[0]; __idr_remove_all()
691 struct idr_layer *pa[MAX_IDR_LEVEL + 1]; idr_for_each() local
692 struct idr_layer **paa = &pa[0]; idr_for_each()
738 struct idr_layer *p, *pa[MAX_IDR_LEVEL + 1]; idr_get_next() local
739 struct idr_layer **paa = &pa[0]; idr_get_next()
934 struct idr_layer *pa[MAX_IDR_LEVEL + 1]; ida_get_new_above() local
943 t = idr_get_empty_slot(&ida->idr, idr_id, pa, 0, &ida->idr); ida_get_new_above()
955 bitmap = (void *)pa[0]->ary[idr_id & IDR_MASK]; ida_get_new_above()
966 rcu_assign_pointer(pa[0]->ary[idr_id & IDR_MASK], ida_get_new_above()
968 pa[0]->count++; ida_get_new_above()
986 idr_mark_full(pa, idr_id); ida_get_new_above()
289 idr_get_empty_slot(struct idr *idp, int starting_id, struct idr_layer **pa, gfp_t gfp_mask, struct idr *layer_idr) idr_get_empty_slot() argument
358 idr_fill_slot(struct idr *idr, void *ptr, int id, struct idr_layer **pa) idr_fill_slot() argument
/linux-4.1.27/drivers/misc/mic/common/
H A Dmic_dev.h27 * @pa: Base physical address.
32 phys_addr_t pa; member in struct:mic_mw
/linux-4.1.27/arch/sparc/include/asm/
H A Dpgalloc_32.h32 unsigned long pa = __nocache_pa((unsigned long)pmdp); pgd_set() local
34 set_pte((pte_t *)pgdp, (SRMMU_ET_PTD | (pa >> 4))); pgd_set()
/linux-4.1.27/arch/arm/mach-omap1/include/mach/
H A Dhardware.h47 extern u8 omap_readb(u32 pa);
48 extern u16 omap_readw(u32 pa);
49 extern u32 omap_readl(u32 pa);
50 extern void omap_writeb(u8 v, u32 pa);
51 extern void omap_writew(u16 v, u32 pa);
52 extern void omap_writel(u32 v, u32 pa);
76 #define OMAP1_IO_ADDRESS(pa) IOMEM((pa) - OMAP1_IO_OFFSET)
/linux-4.1.27/arch/frv/mm/
H A Ddma-alloc.c50 static int map_page(unsigned long va, unsigned long pa, pgprot_t prot) map_page() argument
67 set_pte(pte, mk_pte_phys(pa & PAGE_MASK, prot)); map_page()
84 unsigned long page, va, pa; consistent_alloc() local
111 *dma_handle = pa = virt_to_bus((void *) page); consistent_alloc()
123 err = map_page(va + i, pa + i, PAGE_KERNEL_NOCACHE); consistent_alloc()
/linux-4.1.27/arch/m68k/coldfire/
H A Dintc-2.c148 u16 pa, tb; intc_irq_set_type() local
170 pa = __raw_readw(MCFEPORT_EPPAR); intc_irq_set_type()
171 pa = (pa & ~(0x3 << (irq * 2))) | (tb << (irq * 2)); intc_irq_set_type()
172 __raw_writew(pa, MCFEPORT_EPPAR); intc_irq_set_type()
H A Dintc-simr.c132 u16 pa, tb; intc_irq_set_type() local
154 pa = __raw_readw(MCFEPORT_EPPAR); intc_irq_set_type()
155 pa = (pa & ~(0x3 << ebit)) | (tb << ebit); intc_irq_set_type()
156 __raw_writew(pa, MCFEPORT_EPPAR); intc_irq_set_type()
/linux-4.1.27/drivers/iommu/
H A Dtegra-gart.c266 phys_addr_t pa, size_t bytes, int prot) gart_iommu_map()
277 pfn = __phys_to_pfn(pa); gart_iommu_map()
279 dev_err(gart->dev, "Invalid page: %pa\n", &pa); gart_iommu_map()
312 phys_addr_t pa; gart_iommu_iova_to_phys() local
322 pa = (pte & GART_PAGE_MASK); gart_iommu_iova_to_phys()
323 if (!pfn_valid(__phys_to_pfn(pa))) { gart_iommu_iova_to_phys()
324 dev_err(gart->dev, "No entry for %08llx:%pa\n", gart_iommu_iova_to_phys()
325 (unsigned long long)iova, &pa); gart_iommu_iova_to_phys()
329 return pa; gart_iommu_iova_to_phys()
265 gart_iommu_map(struct iommu_domain *domain, unsigned long iova, phys_addr_t pa, size_t bytes, int prot) gart_iommu_map() argument
H A Domap-iommu.c146 u32 l, pa; omap2_iommu_enable() local
151 pa = virt_to_phys(obj->iopgd); omap2_iommu_enable()
152 if (!IS_ALIGNED(pa, SZ_16K)) omap2_iommu_enable()
159 iommu_write_reg(obj, pa, MMU_TTB); omap2_iommu_enable()
333 cr->ram = e->pa | e->endian | e->elsz | e->mixed; iotlb_alloc_cr()
662 static int iopgd_alloc_section(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) iopgd_alloc_section() argument
666 if ((da | pa) & ~IOSECTION_MASK) { iopgd_alloc_section()
668 __func__, da, pa, IOSECTION_SIZE); iopgd_alloc_section()
672 *iopgd = (pa & IOSECTION_MASK) | prot | IOPGD_SECTION; iopgd_alloc_section()
677 static int iopgd_alloc_super(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) iopgd_alloc_super() argument
682 if ((da | pa) & ~IOSUPER_MASK) { iopgd_alloc_super()
684 __func__, da, pa, IOSUPER_SIZE); iopgd_alloc_super()
689 *(iopgd + i) = (pa & IOSUPER_MASK) | prot | IOPGD_SUPER; iopgd_alloc_super()
694 static int iopte_alloc_page(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) iopte_alloc_page() argument
702 *iopte = (pa & IOPAGE_MASK) | prot | IOPTE_SMALL; iopte_alloc_page()
705 dev_vdbg(obj->dev, "%s: da:%08x pa:%08x pte:%p *pte:%08x\n", iopte_alloc_page()
706 __func__, da, pa, iopte, *iopte); iopte_alloc_page()
711 static int iopte_alloc_large(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) iopte_alloc_large() argument
717 if ((da | pa) & ~IOLARGE_MASK) { iopte_alloc_large()
719 __func__, da, pa, IOLARGE_SIZE); iopte_alloc_large()
727 *(iopte + i) = (pa & IOLARGE_MASK) | prot | IOPTE_LARGE; iopte_alloc_large()
764 err = fn(obj, e->da, e->pa, prot); iopgtable_store_entry_core()
1108 static u32 iotlb_init_entry(struct iotlb_entry *e, u32 da, u32 pa, int pgsz) iotlb_init_entry() argument
1113 e->pa = pa; iotlb_init_entry()
1124 phys_addr_t pa, size_t bytes, int prot) omap_iommu_map()
1139 dev_dbg(dev, "mapping da 0x%lx to pa %pa size 0x%x\n", da, &pa, bytes); omap_iommu_map()
1141 iotlb_init_entry(&e, da, pa, omap_pgsz); omap_iommu_map()
1123 omap_iommu_map(struct iommu_domain *domain, unsigned long da, phys_addr_t pa, size_t bytes, int prot) omap_iommu_map() argument
H A Domap-iopgtable.h57 * omap_iommu_translate() - va to pa translation
62 * va to pa translation
H A Dexynos-iommu.c91 #define mk_lv1ent_sect(pa) ((pa) | 2)
92 #define mk_lv1ent_page(pa) ((pa) | 1)
93 #define mk_lv2ent_lpage(pa) ((pa) | 1)
94 #define mk_lv2ent_spage(pa) ((pa) | 2)
302 pr_err("%s occurred at %#x by %s(Page table base: %pa)\n", show_fault_information()
807 dev_err(dev, "%s: Failed to attach IOMMU with pgtable %pa\n", exynos_iommu_attach_device()
812 dev_dbg(dev, "%s: Attached IOMMU with pgtable %pa %s\n", exynos_iommu_attach_device()
841 dev_dbg(dev, "%s: Detached IOMMU with pgtable %pa\n", exynos_iommu_detach_device()
H A Dmsm_iommu.c368 phys_addr_t pa, size_t len, int prot) msm_iommu_map()
426 *(fl_pte+i) = (pa & 0xFF000000) | FL_SUPERSECTION | msm_iommu_map()
432 *fl_pte = (pa & 0xFFF00000) | FL_AP_READ | FL_AP_WRITE | FL_NG | msm_iommu_map()
457 *sl_pte = (pa & SL_BASE_MASK_SMALL) | SL_AP0 | SL_AP1 | SL_NG | msm_iommu_map()
464 *(sl_pte+i) = (pa & SL_BASE_MASK_LARGE) | SL_AP0 | msm_iommu_map()
367 msm_iommu_map(struct iommu_domain *domain, unsigned long va, phys_addr_t pa, size_t len, int prot) msm_iommu_map() argument
H A Domap-iommu.h18 u32 pa; member in struct:iotlb_entry
H A Diommu.c1052 pr_err("unaligned: iova 0x%lx pa %pa size 0x%zx min_pagesz 0x%x\n", iommu_map()
1057 pr_debug("map: iova 0x%lx pa %pa size 0x%zx\n", iova, &paddr, size); iommu_map()
1062 pr_debug("mapping: iova 0x%lx pa %pa pgsize 0x%zx\n", iommu_map()
H A Drockchip-iommu.c451 dev_err(iommu->dev, "mmu_dte_addr: %pa dte@%pa: %#08x valid: %u pte@%pa: %#08x valid: %u page@%pa flags: %#03x\n", log_iova()
640 pr_err("iova: %pad already mapped to %pa cannot remap to phys: %pa prot: %#x\n", rk_iommu_map_iova()
/linux-4.1.27/drivers/sfi/
H A Dsfi_core.c120 static void sfi_print_table_header(unsigned long long pa, sfi_print_table_header() argument
124 header->sig, pa, sfi_print_table_header()
164 static struct sfi_table_header *sfi_map_table(u64 pa) sfi_map_table() argument
169 if (!TABLE_ON_PAGE(syst_pa, pa, sizeof(struct sfi_table_header))) sfi_map_table()
170 th = sfi_map_memory(pa, sizeof(struct sfi_table_header)); sfi_map_table()
172 th = (void *)syst_va + (pa - syst_pa); sfi_map_table()
180 if (!TABLE_ON_PAGE(syst_pa, pa, sizeof(struct sfi_table_header))) sfi_map_table()
183 return sfi_map_memory(pa, length); sfi_map_table()
231 __ref sfi_check_table(u64 pa, struct sfi_table_key *key) sfi_check_table() argument
236 th = sfi_map_table(pa); sfi_check_table()
241 sfi_print_table_header(pa, th); sfi_check_table()
418 struct sfi_table_attr __init *sfi_sysfs_install_table(u64 pa) sfi_sysfs_install_table() argument
428 th = sfi_map_table(pa); sfi_sysfs_install_table()
H A Dsfi_core.h80 extern struct sfi_table_attr __init *sfi_sysfs_install_table(u64 pa);
/linux-4.1.27/drivers/gpu/drm/msm/
H A Dmsm_iommu.c60 u32 pa = sg_phys(sg) - sg->offset; msm_iommu_map() local
63 VERB("map[%d]: %08x %08x(%zx)", i, iova, pa, bytes); msm_iommu_map()
65 ret = iommu_map(domain, da, pa, bytes, prot); msm_iommu_map()
/linux-4.1.27/arch/powerpc/boot/
H A Dwii.c44 static int mipc_check_address(u32 pa) mipc_check_address() argument
47 if (pa < 0x10000000 || pa > 0x14000000) mipc_check_address()
/linux-4.1.27/drivers/net/wireless/ath/wil6210/
H A Dtxrx.c117 vring->va = dma_alloc_coherent(dev, sz, &vring->pa, GFP_KERNEL); wil_vring_alloc()
134 vring->va, &vring->pa, vring->ctx); wil_vring_alloc()
142 dma_addr_t pa = wil_desc_addr(&d->dma.addr); wil_txdesc_unmap() local
147 dma_unmap_single(dev, pa, dmalen, DMA_TO_DEVICE); wil_txdesc_unmap()
150 dma_unmap_page(dev, pa, dmalen, DMA_TO_DEVICE); wil_txdesc_unmap()
168 &vring->pa, vring->ctx); wil_vring_free()
172 &vring->pa, vring->ctx); wil_vring_free()
176 dma_addr_t pa; wil_vring_free() local
198 pa = wil_desc_addr(&d->dma.addr); wil_vring_free()
200 dma_unmap_single(dev, pa, dmalen, DMA_FROM_DEVICE); wil_vring_free()
205 dma_free_coherent(dev, sz, (void *)vring->va, vring->pa); wil_vring_free()
207 vring->pa = 0; wil_vring_free()
224 dma_addr_t pa; wil_vring_alloc_skb() local
233 pa = dma_map_single(dev, skb->data, skb->len, DMA_FROM_DEVICE); wil_vring_alloc_skb()
234 if (unlikely(dma_mapping_error(dev, pa))) { wil_vring_alloc_skb()
240 wil_desc_addr_set(&d->dma.addr, pa); wil_vring_alloc_skb()
308 void *pa = PTR_ALIGN(p, 8); wil_rx_add_radiotap_header() local
310 if (skb_tailroom(skb) >= len + (pa - p)) { wil_rx_add_radiotap_header()
312 memcpy(phy_data, pa, phy_length); wil_rx_add_radiotap_header()
376 dma_addr_t pa; wil_vring_reap_rx() local
405 pa = wil_desc_addr(&d->dma.addr); wil_vring_reap_rx()
407 dma_unmap_single(dev, pa, sz, DMA_FROM_DEVICE); wil_vring_reap_rx()
725 cmd.vring_cfg.tx_sw_ring.ring_mem_base = cpu_to_le64(vring->pa); wil_vring_init_tx()
793 cmd.vring_cfg.tx_sw_ring.ring_mem_base = cpu_to_le64(vring->pa); wil_vring_init_bcast()
1025 static int wil_tx_desc_map(struct vring_tx_desc *d, dma_addr_t pa, u32 len, wil_tx_desc_map() argument
1028 wil_desc_addr_set(&d->dma.addr, pa); wil_tx_desc_map()
1115 dma_addr_t pa; __wil_tx_vring() local
1133 pa = dma_map_single(dev, skb->data, skb_headlen(skb), DMA_TO_DEVICE); __wil_tx_vring()
1136 skb_headlen(skb), skb->data, &pa); __wil_tx_vring()
1140 if (unlikely(dma_mapping_error(dev, pa))) __wil_tx_vring()
1144 wil_tx_desc_map(d, pa, len, vring_index); __wil_tx_vring()
1177 pa = skb_frag_dma_map(dev, frag, 0, skb_frag_size(frag), __wil_tx_vring()
1179 if (unlikely(dma_mapping_error(dev, pa))) __wil_tx_vring()
1182 wil_tx_desc_map(d, pa, len, vring_index); __wil_tx_vring()
H A Dtxrx.h44 dma_addr_t pa) wil_desc_addr_set()
46 addr->addr_low = cpu_to_le32(lower_32_bits(pa)); wil_desc_addr_set()
47 addr->addr_high = cpu_to_le16((u16)upper_32_bits(pa)); wil_desc_addr_set()
43 wil_desc_addr_set(struct vring_dma_addr *addr, dma_addr_t pa) wil_desc_addr_set() argument
/linux-4.1.27/drivers/infiniband/hw/ocrdma/
H A Docrdma_hw.c429 ocrdma_build_q_pages(&cmd->pa[0], cmd->num_pages, eq->q.dma, ocrdma_mbx_create_eq()
525 ocrdma_build_q_pages(&cmd->pa[0], cq->size / OCRDMA_MIN_Q_PAGE_SIZE, ocrdma_mbx_mq_cq_create()
552 struct ocrdma_pa *pa; ocrdma_mbx_create_mq() local
571 pa = &cmd->pa[0]; ocrdma_mbx_create_mq()
573 ocrdma_build_q_pages(pa, num_pages, mq->dma, PAGE_SIZE_4K); ocrdma_mbx_create_mq()
1234 mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dev->stats_mem.pa & 0xffffffff); ocrdma_mbx_rdma_stats()
1235 mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dev->stats_mem.pa); ocrdma_mbx_rdma_stats()
1274 dma.size, &dma.pa, GFP_KERNEL); ocrdma_mbx_get_ctrl_attribs()
1282 mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dma.pa & 0xffffffff); ocrdma_mbx_get_ctrl_attribs()
1283 mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dma.pa); ocrdma_mbx_get_ctrl_attribs()
1303 dma_free_coherent(&dev->nic_info.pdev->dev, dma.size, dma.va, dma.pa); ocrdma_mbx_get_ctrl_attribs()
1581 dma_addr_t pa; ocrdma_mbx_create_ah_tbl() local
1610 &dev->av_tbl.pbl.pa, ocrdma_mbx_create_ah_tbl()
1616 &pa, GFP_KERNEL); ocrdma_mbx_create_ah_tbl()
1619 dev->av_tbl.pa = pa; ocrdma_mbx_create_ah_tbl()
1625 pbes[i].pa_lo = (u32)cpu_to_le32(pa & 0xffffffff); ocrdma_mbx_create_ah_tbl()
1626 pbes[i].pa_hi = (u32)cpu_to_le32(upper_32_bits(pa)); ocrdma_mbx_create_ah_tbl()
1627 pa += PAGE_SIZE; ocrdma_mbx_create_ah_tbl()
1629 cmd->tbl_addr[0].lo = (u32)(dev->av_tbl.pbl.pa & 0xFFFFFFFF); ocrdma_mbx_create_ah_tbl()
1630 cmd->tbl_addr[0].hi = (u32)upper_32_bits(dev->av_tbl.pbl.pa); ocrdma_mbx_create_ah_tbl()
1641 dev->av_tbl.pa); ocrdma_mbx_create_ah_tbl()
1645 dev->av_tbl.pbl.pa); ocrdma_mbx_create_ah_tbl()
1668 dev->av_tbl.pa); ocrdma_mbx_delete_ah_tbl()
1671 dev->av_tbl.pbl.pa); ocrdma_mbx_delete_ah_tbl()
1750 cq->va = dma_alloc_coherent(&pdev->dev, cq->len, &cq->pa, GFP_KERNEL); ocrdma_mbx_create_cq()
1803 ocrdma_build_q_pages(&cmd->cmd.pa[0], hw_pages, cq->pa, page_size); ocrdma_mbx_create_cq()
1814 dma_free_coherent(&pdev->dev, cq->len, cq->va, cq->pa); ocrdma_mbx_create_cq()
1837 dma_free_coherent(&dev->nic_info.pdev->dev, cq->len, cq->va, cq->pa); ocrdma_mbx_destroy_cq()
1932 cmd->pbl[i].lo = (u32) (hwmr->pbl_table[i].pa & 0xffffffff); ocrdma_mbx_reg_mr()
1933 cmd->pbl[i].hi = upper_32_bits(hwmr->pbl_table[i].pa); ocrdma_mbx_reg_mr()
1963 (u32) (hwmr->pbl_table[i + pbl_offset].pa & 0xffffffff); ocrdma_mbx_reg_mr_cont()
1965 upper_32_bits(hwmr->pbl_table[i + pbl_offset].pa); ocrdma_mbx_reg_mr_cont()
2125 dma_addr_t pa; ocrdma_set_create_qp_sq_cmd() local
2146 qp->sq.va = dma_alloc_coherent(&pdev->dev, len, &pa, GFP_KERNEL); ocrdma_set_create_qp_sq_cmd()
2151 qp->sq.pa = pa; ocrdma_set_create_qp_sq_cmd()
2153 ocrdma_build_q_pages(&cmd->wq_addr[0], hw_pages, pa, hw_page_size); ocrdma_set_create_qp_sq_cmd()
2181 dma_addr_t pa = 0; ocrdma_set_create_qp_rq_cmd() local
2197 qp->rq.va = dma_alloc_coherent(&pdev->dev, len, &pa, GFP_KERNEL); ocrdma_set_create_qp_rq_cmd()
2201 qp->rq.pa = pa; ocrdma_set_create_qp_rq_cmd()
2205 ocrdma_build_q_pages(&cmd->rq_addr[0], hw_pages, pa, hw_page_size); ocrdma_set_create_qp_rq_cmd()
2245 dma_addr_t pa = 0; ocrdma_set_create_qp_ird_cmd() local
2255 &pa, GFP_KERNEL); ocrdma_set_create_qp_ird_cmd()
2260 pa, ird_page_size); ocrdma_set_create_qp_ird_cmd()
2397 dma_free_coherent(&pdev->dev, qp->rq.len, qp->rq.va, qp->rq.pa); ocrdma_mbx_create_qp()
2400 dma_free_coherent(&pdev->dev, qp->sq.len, qp->sq.va, qp->sq.pa); ocrdma_mbx_create_qp()
2657 dma_free_coherent(&pdev->dev, qp->sq.len, qp->sq.va, qp->sq.pa); ocrdma_mbx_destroy_qp()
2659 dma_free_coherent(&pdev->dev, qp->rq.len, qp->rq.va, qp->rq.pa); ocrdma_mbx_destroy_qp()
2674 dma_addr_t pa; ocrdma_mbx_create_srq() local
2694 srq->rq.va = dma_alloc_coherent(&pdev->dev, len, &pa, GFP_KERNEL); ocrdma_mbx_create_srq()
2699 ocrdma_build_q_pages(&cmd->rq_addr[0], hw_pages, pa, hw_page_size); ocrdma_mbx_create_srq()
2702 srq->rq.pa = pa; ocrdma_mbx_create_srq()
2734 dma_free_coherent(&pdev->dev, srq->rq.len, srq->rq.va, pa); ocrdma_mbx_create_srq()
2796 srq->rq.va, srq->rq.pa); ocrdma_mbx_destroy_srq()
2805 dma_addr_t pa; ocrdma_mbx_get_dcbx_config() local
2816 req = dma_alloc_coherent(&pdev->dev, cmd.hdr.pyld_len, &pa, GFP_KERNEL); ocrdma_mbx_get_dcbx_config()
2824 mqe_sge->pa_lo = (u32) (pa & 0xFFFFFFFFUL); ocrdma_mbx_get_dcbx_config()
2825 mqe_sge->pa_hi = (u32) upper_32_bits(pa); ocrdma_mbx_get_dcbx_config()
2842 dma_free_coherent(&pdev->dev, cmd.hdr.pyld_len, req, pa); ocrdma_mbx_get_dcbx_config()
H A Docrdma.h106 dma_addr_t pa; member in struct:ocrdma_dma_mem
112 dma_addr_t pa; member in struct:ocrdma_pbl
206 dma_addr_t pa; member in struct:stats_mem
246 dma_addr_t pa; member in struct:ocrdma_dev::__anon4960
318 dma_addr_t pa; member in struct:ocrdma_cq
353 dma_addr_t pa; member in struct:ocrdma_qp_hwq_info
425 dma_addr_t pa; member in struct:ocrdma_ucontext::__anon4962
H A Docrdma_verbs.c490 &ctx->ah_tbl.pa, GFP_KERNEL); ocrdma_alloc_ucontext()
524 ocrdma_del_mmap(ctx, ctx->ah_tbl.pa, ctx->ah_tbl.len); ocrdma_alloc_ucontext()
527 ctx->ah_tbl.pa); ocrdma_alloc_ucontext()
542 ocrdma_del_mmap(uctx, uctx->ah_tbl.pa, uctx->ah_tbl.len); ocrdma_dealloc_ucontext()
544 uctx->ah_tbl.pa); ocrdma_dealloc_ucontext()
775 mr->pbl_table[i].pa); ocrdma_free_mr_pbl_tbl()
814 dma_addr_t pa; ocrdma_build_pbl_tbl() local
823 va = dma_alloc_coherent(&pdev->dev, dma_len, &pa, GFP_KERNEL); ocrdma_build_pbl_tbl()
831 mr->pbl_table[i].pa = pa; ocrdma_build_pbl_tbl()
1114 ocrdma_del_mmap(cq->ucontext, (u64) cq->pa, ocrdma_destroy_cq()
1765 ocrdma_del_mmap(pd->uctx, (u64) qp->sq.pa, ocrdma_destroy_qp()
1768 ocrdma_del_mmap(pd->uctx, (u64) qp->rq.pa, ocrdma_destroy_qp()
1914 ocrdma_del_mmap(srq->pd->uctx, (u64) srq->rq.pa, ocrdma_destroy_srq()
/linux-4.1.27/drivers/char/agp/
H A Dalpha-agp.c19 unsigned long pa; alpha_core_agp_vm_fault() local
24 pa = agp->ops->translate(agp, dma_addr); alpha_core_agp_vm_fault()
26 if (pa == (unsigned long)-EINVAL) alpha_core_agp_vm_fault()
32 page = virt_to_page(__va(pa)); alpha_core_agp_vm_fault()
/linux-4.1.27/include/ras/
H A Dras_event.h38 __field(u64, pa)
53 __entry->pa = mem->physical_addr;
55 __entry->pa = ~0ull;
70 __entry->pa,
/linux-4.1.27/arch/powerpc/platforms/ps3/
H A Dhtab.c48 unsigned long pa, unsigned long rflags, unsigned long vflags, ps3_hpte_insert()
66 hpte_r = hpte_encode_r(ps3_mm_phys_to_lpar(pa), psize, apsize) | rflags; ps3_hpte_insert()
79 pr_info("%s:result=%s vpn=%lx pa=%lx ix=%lx v=%llx r=%llx\n", ps3_hpte_insert()
80 __func__, ps3_result(result), vpn, pa, hpte_group, ps3_hpte_insert()
47 ps3_hpte_insert(unsigned long hpte_group, unsigned long vpn, unsigned long pa, unsigned long rflags, unsigned long vflags, int psize, int apsize, int ssize) ps3_hpte_insert() argument
/linux-4.1.27/arch/mips/lib/
H A Dr3k_dump_tlb.c46 " [pa=%06lx n=%d d=%d v=%d g=%d]", dump_tlb()
H A Ddump_tlb.c89 printk("\t[pa=%0*llx c=%d d=%d v=%d g=%d] ", dump_tlb()
95 printk("[pa=%0*llx c=%d d=%d v=%d g=%d]\n", dump_tlb()
/linux-4.1.27/arch/powerpc/mm/
H A Dhugepage-hash64.c27 unsigned long rflags, pa, hidx; __hash_page_thp() local
131 pa = pmd_pfn(__pmd(old_pmd)) << PAGE_SHIFT; __hash_page_thp()
145 slot = ppc_md.hpte_insert(hpte_group, vpn, pa, rflags, 0, __hash_page_thp()
153 slot = ppc_md.hpte_insert(hpte_group, vpn, pa, __hash_page_thp()
H A Dhugetlbpage-hash64.c18 unsigned long pa, unsigned long rlags,
27 unsigned long rflags, pa, sz; __hash_page_huge() local
91 pa = pte_pfn(__pte(old_pte)) << PAGE_SHIFT; __hash_page_huge()
107 slot = hpte_insert_repeating(hash, vpn, pa, rflags, 0, __hash_page_huge()
H A Dfsl_booke_mmu.c90 unsigned long p_mapped_by_tlbcam(phys_addr_t pa) p_mapped_by_tlbcam() argument
94 if (pa >= tlbcam_addrs[b].phys p_mapped_by_tlbcam()
95 && pa < (tlbcam_addrs[b].limit-tlbcam_addrs[b].start) p_mapped_by_tlbcam()
97 return tlbcam_addrs[b].start+(pa-tlbcam_addrs[b].phys); p_mapped_by_tlbcam()
H A Dppc_mmu_32.c64 unsigned long p_mapped_by_bats(phys_addr_t pa) p_mapped_by_bats() argument
68 if (pa >= bat_addrs[b].phys p_mapped_by_bats()
69 && pa < (bat_addrs[b].limit-bat_addrs[b].start) p_mapped_by_bats()
71 return bat_addrs[b].start+(pa-bat_addrs[b].phys); p_mapped_by_bats()
H A Dmmu_decl.h97 extern int map_page(unsigned long va, phys_addr_t pa, int flags);
113 extern int map_kernel_page(unsigned long ea, unsigned long pa, int flags);
H A Dpgtable_32.c56 extern unsigned long p_mapped_by_bats(phys_addr_t pa);
64 extern unsigned long p_mapped_by_tlbcam(phys_addr_t pa);
288 int map_page(unsigned long va, phys_addr_t pa, int flags) map_page() argument
305 set_pte_at(&init_mm, va, pg, pfn_pte(pa >> PAGE_SHIFT, map_page()
H A Dpgtable_64.c91 int map_kernel_page(unsigned long ea, unsigned long pa, int flags) map_kernel_page() argument
109 set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, map_kernel_page()
134 set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, map_kernel_page()
143 if (htab_bolt_mapping(ea, ea + PAGE_SIZE, pa, flags, map_kernel_page()
146 "memory at %016lx !\n", pa); map_kernel_page()
171 void __iomem * __ioremap_at(phys_addr_t pa, void *ea, unsigned long size, __ioremap_at() argument
188 WARN_ON(pa & ~PAGE_MASK); __ioremap_at()
193 if (map_kernel_page((unsigned long)ea+i, pa+i, flags)) __ioremap_at()
H A Dhash_native_64.c195 unsigned long pa, unsigned long rflags, native_hpte_insert()
203 DBG_LOW(" insert(group=%lx, vpn=%016lx, pa=%016lx," native_hpte_insert()
205 hpte_group, vpn, pa, rflags, vflags, psize); native_hpte_insert()
224 hpte_r = hpte_encode_r(pa, psize, apsize) | rflags; native_hpte_insert()
194 native_hpte_insert(unsigned long hpte_group, unsigned long vpn, unsigned long pa, unsigned long rflags, unsigned long vflags, int psize, int apsize, int ssize) native_hpte_insert() argument
/linux-4.1.27/arch/arm/mm/
H A Dcache-xsc3l2.c78 static inline unsigned long l2_map_va(unsigned long pa, unsigned long prev_va) l2_map_va() argument
82 unsigned long pa_offset = pa << (32 - PAGE_SHIFT); l2_map_va()
90 va = (unsigned long)kmap_atomic_pfn(pa >> PAGE_SHIFT); l2_map_va()
94 return __phys_to_virt(pa); l2_map_va()
H A Dnommu.c115 pr_notice("Ignoring RAM after %pa, memory at %pa ignored\n", for_each_memblock()
144 pr_warn("Truncating memory from %pa to %pa (MPU region constraints)",
152 pr_debug("MPU Region from %pa size %pa (end %pa))\n",
H A Dmmu.c1090 pr_notice("Ignoring RAM at %pa-%pa (!CONFIG_HIGHMEM)\n", for_each_memblock()
1099 pr_notice("Truncating RAM at %pa-%pa to -%pa", for_each_memblock()
/linux-4.1.27/arch/arm/mach-alpine/
H A Dplatsmp.c33 pr_err("FAIL: resume address over 32bit (%pa)", &addr); alpine_boot_secondary()
/linux-4.1.27/drivers/net/ethernet/brocade/bna/
H A Dbfa_ioc.h48 u64 pa; /* ! Physical address */ member in struct:bfa_dma
58 #define bfa_dma_be_addr_set(dma_addr, pa) \
59 __bfa_dma_be_addr_set(&dma_addr, (u64)pa)
61 __bfa_dma_be_addr_set(union bfi_addr_u *dma_addr, u64 pa) __bfa_dma_be_addr_set() argument
63 dma_addr->a32.addr_lo = (u32) htonl(pa); __bfa_dma_be_addr_set()
64 dma_addr->a32.addr_hi = (u32) htonl(upper_32_bits(pa)); __bfa_dma_be_addr_set()
71 __bfa_alen_set(struct bfi_alen *alen, u32 len, u64 pa) __bfa_alen_set() argument
74 bfa_dma_be_addr_set(alen->al_addr, pa); __bfa_alen_set()
H A Dbfa_msgq.c524 bfa_dma_be_addr_set(msgq_cfg->cmdq.addr, msgq->cmdq.addr.pa); bfa_msgq_init()
526 bfa_dma_be_addr_set(msgq_cfg->rspq.addr, msgq->rspq.addr.pa); bfa_msgq_init()
597 bfa_msgq_memclaim(struct bfa_msgq *msgq, u8 *kva, u64 pa) bfa_msgq_memclaim() argument
600 msgq->cmdq.addr.pa = pa; bfa_msgq_memclaim()
603 pa += roundup(BFA_MSGQ_CMDQ_SIZE, BFA_DMA_ALIGN_SZ); bfa_msgq_memclaim()
606 msgq->rspq.addr.pa = pa; bfa_msgq_memclaim()
H A Dbfa_cee.c150 cee->attr_dma.pa = dma_pa; bfa_nw_cee_mem_claim()
152 cee->stats_dma.pa = dma_pa + bfa_cee_attr_meminfo(); bfa_nw_cee_mem_claim()
185 bfa_dma_be_addr_set(cmd->dma_addr, cee->attr_dma.pa); bfa_nw_cee_get_attr()
H A Dbfa_msgq.h123 void bfa_msgq_memclaim(struct bfa_msgq *msgq, u8 *kva, u64 pa);
/linux-4.1.27/arch/hexagon/kernel/
H A Dvm_init_segtable.S35 #define BKP(pa) (((pa) & __HVM_PTE_PGMASK_4MB) \
42 #define BKPG_IO(pa) (((pa) & __HVM_PTE_PGMASK_16MB) \
46 #define FOURK_IO(pa) (((pa) & __HVM_PTE_PGMASK_4KB) \
50 #define L2_PTR(pa) (((pa) & __HVM_PTE_PGMASK_4KB) \
/linux-4.1.27/arch/microblaze/mm/
H A Dpgtable.c138 int map_page(unsigned long va, phys_addr_t pa, int flags) map_page() argument
151 set_pte_at(&init_mm, va, pg, pfn_pte(pa >> PAGE_SHIFT, map_page()
217 unsigned long pa; iopa() local
230 pa = 0; iopa()
232 pa = (pte_val(*pte) & PAGE_MASK) | (addr & ~PAGE_MASK); iopa()
234 return pa; iopa()
H A Dconsistent.c70 phys_addr_t pa; consistent_alloc() local
120 *dma_handle = pa = __virt_to_phys(vaddr); consistent_alloc()
137 err = map_page(va + i, pa + i, _PAGE_KERNEL | _PAGE_NO_CACHE); consistent_alloc()
/linux-4.1.27/drivers/tty/serial/8250/
H A D8250_hp300.c127 unsigned long pa = dio_scodetophysaddr(scode); hp300_setup_serial_console() local
128 if (!pa) hp300_setup_serial_console()
134 port.mapbase = (pa + UART_OFFSET); hp300_setup_serial_console()
137 port.irq = DIO_IPL(pa + DIO_VIRADDRBASE); hp300_setup_serial_console()
140 out_8(pa + DIO_VIRADDRBASE + DCA_IC, DCA_IC_IE); hp300_setup_serial_console()
142 if (DIO_ID(pa + DIO_VIRADDRBASE) & 0x80) hp300_setup_serial_console()
/linux-4.1.27/drivers/net/wireless/brcm80211/brcmsmac/
H A Ddma.c716 dma_addr_t pa, uint outidx, u32 *flags, u32 bufcount) dma64_dd_upd()
721 if ((di->dataoffsetlow == 0) || !(pa & PCI32ADDR_HIGH)) { dma64_dd_upd()
722 ddring[outidx].addrlow = cpu_to_le32(pa + di->dataoffsetlow); dma64_dd_upd()
730 ae = (pa & PCI32ADDR_HIGH) >> PCI32ADDR_HIGH_SHIFT; dma64_dd_upd()
731 pa &= ~PCI32ADDR_HIGH; dma64_dd_upd()
734 ddring[outidx].addrlow = cpu_to_le32(pa + di->dataoffsetlow); dma64_dd_upd()
774 _dma_ddtable_init(struct dma_info *di, uint direction, dma_addr_t pa) _dma_ddtable_init() argument
778 di->xmtptrbase = pa; _dma_ddtable_init()
780 di->rcvptrbase = pa; _dma_ddtable_init()
784 || !(pa & PCI32ADDR_HIGH)) { _dma_ddtable_init()
787 pa + di->ddoffsetlow); _dma_ddtable_init()
792 pa + di->ddoffsetlow); _dma_ddtable_init()
800 /* shift the high bit(s) from pa to ae */ _dma_ddtable_init()
801 ae = (pa & PCI32ADDR_HIGH) >> PCI32ADDR_HIGH_SHIFT; _dma_ddtable_init()
802 pa &= ~PCI32ADDR_HIGH; _dma_ddtable_init()
806 pa + di->ddoffsetlow); _dma_ddtable_init()
813 pa + di->ddoffsetlow); _dma_ddtable_init()
873 dma_addr_t pa; dma64_getnextrxp() local
894 pa = le32_to_cpu(di->rxd64[i].addrlow) - di->dataoffsetlow; dma64_getnextrxp()
897 dma_unmap_single(di->dmadev, pa, di->rxbufsize, DMA_FROM_DEVICE); dma64_getnextrxp()
1031 dma_addr_t pa; dma_rxfill() local
1080 pa = dma_map_single(di->dmadev, p->data, di->rxbufsize, dma_rxfill()
1082 if (dma_mapping_error(di->dmadev, pa)) dma_rxfill()
1093 dma64_dd_upd(di, di->rxd64, pa, rxout, &flags, dma_rxfill()
1275 dma_addr_t pa; dma_txenq() local
1289 pa = dma_map_single(di->dmadev, data, len, DMA_TO_DEVICE); dma_txenq()
1291 if (dma_mapping_error(di->dmadev, pa)) { dma_txenq()
1304 dma64_dd_upd(di, di->txd64, pa, txout, &flags, len); dma_txenq()
1510 dma_addr_t pa; dma_getnexttxp() local
1513 pa = le32_to_cpu(di->txd64[i].addrlow) - di->dataoffsetlow; dma_getnexttxp()
1525 dma_unmap_single(di->dmadev, pa, size, DMA_TO_DEVICE); dma_getnexttxp()
715 dma64_dd_upd(struct dma_info *di, struct dma64desc *ddring, dma_addr_t pa, uint outidx, u32 *flags, u32 bufcount) dma64_dd_upd() argument
/linux-4.1.27/drivers/base/
H A Dplatform.c209 struct platform_object *pa = container_of(dev, struct platform_object, platform_device_release() local
212 of_device_node_put(&pa->pdev.dev); platform_device_release()
213 kfree(pa->pdev.dev.platform_data); platform_device_release()
214 kfree(pa->pdev.mfd_cell); platform_device_release()
215 kfree(pa->pdev.resource); platform_device_release()
216 kfree(pa->pdev.driver_override); platform_device_release()
217 kfree(pa); platform_device_release()
230 struct platform_object *pa; platform_device_alloc() local
232 pa = kzalloc(sizeof(*pa) + strlen(name) + 1, GFP_KERNEL); platform_device_alloc()
233 if (pa) { platform_device_alloc()
234 strcpy(pa->name, name); platform_device_alloc()
235 pa->pdev.name = pa->name; platform_device_alloc()
236 pa->pdev.id = id; platform_device_alloc()
237 device_initialize(&pa->pdev.dev); platform_device_alloc()
238 pa->pdev.dev.release = platform_device_release; platform_device_alloc()
239 arch_setup_pdev_archdata(&pa->pdev); platform_device_alloc()
242 return pa ? &pa->pdev : NULL; platform_device_alloc()
H A Ddma-coherent.c287 pr_err("Reserved memory: failed to init DMA memory pool at %pa, size %ld MiB\n", rmem_dma_device_init()
322 pr_info("Reserved memory: created DMA memory pool at %pa, size %ld MiB\n", rmem_dma_setup()
H A Ddma-contiguous.c274 pr_info("Reserved memory: created CMA memory pool at %pa, size %ld MiB\n", rmem_cma_setup()
/linux-4.1.27/fs/nfs_common/
H A Dnfsacl.c220 struct posix_acl_entry *pa, *pe, posix_acl_from_nfsacl() local
230 FOREACH_ACL_ENTRY(pa, acl, pe) { FOREACH_ACL_ENTRY()
231 switch(pa->e_tag) { FOREACH_ACL_ENTRY()
235 group_obj = pa; FOREACH_ACL_ENTRY()
238 mask = pa; FOREACH_ACL_ENTRY()
/linux-4.1.27/drivers/media/v4l2-core/
H A Dvideobuf2-memops.c103 dma_addr_t pa = 0; vb2_get_contig_userptr() local
120 pa = this_pfn << PAGE_SHIFT; vb2_get_contig_userptr()
134 *res_pa = pa + offset; vb2_get_contig_userptr()
/linux-4.1.27/sound/soc/sirf/
H A Dsirf-audio.c101 "spk-pa-gpios", 0); sirf_audio_probe()
103 "hp-pa-gpios", 0); sirf_audio_probe()
/linux-4.1.27/drivers/infiniband/hw/ehca/
H A Dehca_av.c57 struct ib_port_attr pa; ehca_calc_ipd() local
70 ret = ehca_query_port(&shca->ib_device, port, &pa); ehca_calc_ipd()
76 link = ib_width_enum_to_int(pa.active_width) * pa.active_speed; ehca_calc_ipd()
/linux-4.1.27/drivers/net/ethernet/intel/i40e/
H A Di40e_hmc.h117 * @pa: pointer to physical address
121 #define I40E_SET_PF_SD_ENTRY(hw, pa, sd_index, type) \
124 val1 = (u32)(upper_32_bits(pa)); \
125 val2 = (u32)(pa) | (I40E_HMC_MAX_BP_COUNT << \
H A Di40e_adminq.c194 cpu_to_le32(upper_32_bits(bi->pa)); i40e_alloc_arq_bufs()
196 cpu_to_le32(lower_32_bits(bi->pa)); i40e_alloc_arq_bufs()
283 if (hw->aq.asq.r.asq_bi[i].pa) i40e_free_asq_bufs()
314 wr32(hw, hw->aq.asq.bal, lower_32_bits(hw->aq.asq.desc_buf.pa)); i40e_config_asq_regs()
315 wr32(hw, hw->aq.asq.bah, upper_32_bits(hw->aq.asq.desc_buf.pa)); i40e_config_asq_regs()
319 if (reg != lower_32_bits(hw->aq.asq.desc_buf.pa)) i40e_config_asq_regs()
343 wr32(hw, hw->aq.arq.bal, lower_32_bits(hw->aq.arq.desc_buf.pa)); i40e_config_arq_regs()
344 wr32(hw, hw->aq.arq.bah, upper_32_bits(hw->aq.arq.desc_buf.pa)); i40e_config_arq_regs()
351 if (reg != lower_32_bits(hw->aq.arq.desc_buf.pa)) i40e_config_arq_regs()
829 /* Update the address values in the desc with the pa value i40e_asq_send_command()
833 cpu_to_le32(upper_32_bits(dma_buff->pa)); i40e_asq_send_command()
835 cpu_to_le32(lower_32_bits(dma_buff->pa)); i40e_asq_send_command()
994 desc->params.external.addr_high = cpu_to_le32(upper_32_bits(bi->pa)); i40e_clean_arq_element()
995 desc->params.external.addr_low = cpu_to_le32(lower_32_bits(bi->pa)); i40e_clean_arq_element()
H A Di40e_osdep.h56 dma_addr_t pa; member in struct:i40e_dma_mem
/linux-4.1.27/drivers/net/ethernet/intel/i40evf/
H A Di40e_hmc.h117 * @pa: pointer to physical address
121 #define I40E_SET_PF_SD_ENTRY(hw, pa, sd_index, type) \
124 val1 = (u32)(upper_32_bits(pa)); \
125 val2 = (u32)(pa) | (I40E_HMC_MAX_BP_COUNT << \
H A Di40e_adminq.c192 cpu_to_le32(upper_32_bits(bi->pa)); i40e_alloc_arq_bufs()
194 cpu_to_le32(lower_32_bits(bi->pa)); i40e_alloc_arq_bufs()
281 if (hw->aq.asq.r.asq_bi[i].pa) i40e_free_asq_bufs()
312 wr32(hw, hw->aq.asq.bal, lower_32_bits(hw->aq.asq.desc_buf.pa)); i40e_config_asq_regs()
313 wr32(hw, hw->aq.asq.bah, upper_32_bits(hw->aq.asq.desc_buf.pa)); i40e_config_asq_regs()
317 if (reg != lower_32_bits(hw->aq.asq.desc_buf.pa)) i40e_config_asq_regs()
341 wr32(hw, hw->aq.arq.bal, lower_32_bits(hw->aq.arq.desc_buf.pa)); i40e_config_arq_regs()
342 wr32(hw, hw->aq.arq.bah, upper_32_bits(hw->aq.arq.desc_buf.pa)); i40e_config_arq_regs()
349 if (reg != lower_32_bits(hw->aq.arq.desc_buf.pa)) i40e_config_arq_regs()
780 /* Update the address values in the desc with the pa value i40evf_asq_send_command()
784 cpu_to_le32(upper_32_bits(dma_buff->pa)); i40evf_asq_send_command()
786 cpu_to_le32(lower_32_bits(dma_buff->pa)); i40evf_asq_send_command()
946 desc->params.external.addr_high = cpu_to_le32(upper_32_bits(bi->pa)); i40evf_clean_arq_element()
947 desc->params.external.addr_low = cpu_to_le32(lower_32_bits(bi->pa)); i40evf_clean_arq_element()
H A Di40e_osdep.h55 dma_addr_t pa; member in struct:i40e_dma_mem
/linux-4.1.27/net/sunrpc/xprtrdma/
H A Dfrwr_ops.c192 u64 pa; frwr_op_map() local
203 pa = seg->mr_dma; frwr_op_map()
205 frmr->fr_pgl->page_list[page_no++] = pa; frwr_op_map()
206 pa += PAGE_SIZE; frwr_op_map()
/linux-4.1.27/drivers/scsi/sym53c8xx_2/
H A Dsym_fw.c260 u32 *pa; sym_fw_setup_bus_addresses() local
269 pa = (u32 *) &np->fwa_bas; sym_fw_setup_bus_addresses()
271 pa[i] = np->scripta_ba + po[i]; sym_fw_setup_bus_addresses()
277 pa = (u32 *) &np->fwb_bas; sym_fw_setup_bus_addresses()
279 pa[i] = np->scriptb_ba + po[i]; sym_fw_setup_bus_addresses()
285 pa = (u32 *) &np->fwz_bas; sym_fw_setup_bus_addresses()
287 pa[i] = np->scriptz_ba + po[i]; sym_fw_setup_bus_addresses()
/linux-4.1.27/arch/sparc/mm/
H A Dfault_64.c98 unsigned long pa; get_user_insn() local
119 pa = pmd_pfn(*pmdp) << PAGE_SHIFT; get_user_insn()
120 pa += tpc & ~HPAGE_MASK; get_user_insn()
125 : "r" (pa), "i" (ASI_PHYS_USE_EC)); get_user_insn()
132 pa = (pte_pfn(pte) << PAGE_SHIFT); get_user_insn()
133 pa += (tpc & ~PAGE_MASK); get_user_insn()
138 : "r" (pa), "i" (ASI_PHYS_USE_EC)); get_user_insn()
H A Dinit_64.c1379 unsigned long pa = __pa(addr); kern_addr_valid() local
1384 return pfn_valid(pa >> PAGE_SHIFT); kern_addr_valid()
1786 static void patch_one_ktsb_phys(unsigned int *start, unsigned int *end, unsigned long pa) patch_one_ktsb_phys() argument
1790 high_bits = (pa >> 32) & 0xffffffff; patch_one_ktsb_phys()
1791 low_bits = (pa >> 0) & 0xffffffff; patch_one_ktsb_phys()
1890 unsigned long pa, ret; sun4v_ktsb_register() local
1892 pa = kern_base + ((unsigned long)&ktsb_descr[0] - KERNBASE); sun4v_ktsb_register()
1894 ret = sun4v_mmu_tsb_ctx0(NUM_KTSB_DESCR, pa); sun4v_ktsb_register()
1897 "errors with %lx\n", pa, ret); sun4v_ktsb_register()
/linux-4.1.27/mm/
H A Dcma.c249 pr_debug("%s(size %pa, base %pa, limit %pa alignment %pa)\n", cma_declare_contiguous()
288 pr_err("Region at %pa defined on low/high memory boundary (%pa)\n", cma_declare_contiguous()
344 pr_info("Reserved %ld MiB at %pa\n", (unsigned long)size / SZ_1M, cma_declare_contiguous()
H A Dprocess_vm_access.c85 unsigned long pa = addr & PAGE_MASK; process_vm_rw_single_vec() local
86 unsigned long start_offset = addr - pa; process_vm_rw_single_vec()
102 pages = get_user_pages_unlocked(task, mm, pa, pages, process_vm_rw_single_vec()
117 pa += pages * PAGE_SIZE; process_vm_rw_single_vec()
/linux-4.1.27/arch/ia64/kernel/
H A Dacpi.c422 get_processor_proximity_domain(struct acpi_srat_cpu_affinity *pa) get_processor_proximity_domain() argument
426 pxm = pa->proximity_domain_lo; get_processor_proximity_domain()
428 pxm += pa->proximity_domain_hi[0] << 8; get_processor_proximity_domain()
464 acpi_numa_processor_affinity_init(struct acpi_srat_cpu_affinity *pa) acpi_numa_processor_affinity_init() argument
468 if (!(pa->flags & ACPI_SRAT_CPU_ENABLED)) acpi_numa_processor_affinity_init()
477 pxm = get_processor_proximity_domain(pa); acpi_numa_processor_affinity_init()
483 (pa->apic_id << 8) | (pa->local_sapic_eid); acpi_numa_processor_affinity_init()
H A Dbrl_emu.c40 #define unimplemented_physical_address(pa) ( \
41 ((pa) & local_cpu_data->unimpl_pa_mask) != 0 \
H A Dkprobes.c1048 struct param_bsp_cfm pa; setjmp_pre_handler() local
1058 pa.ip = regs->cr_iip; setjmp_pre_handler()
1059 unw_init_running(ia64_get_bsp_cfm, &pa); setjmp_pre_handler()
1060 bytes = (char *)ia64_rse_skip_regs(pa.bsp, pa.cfm & 0x3f) setjmp_pre_handler()
1061 - (char *)pa.bsp; setjmp_pre_handler()
1063 pa.bsp, setjmp_pre_handler()
1065 kcb->bsp = pa.bsp; setjmp_pre_handler()
1066 kcb->cfm = pa.cfm; setjmp_pre_handler()
/linux-4.1.27/drivers/media/radio/
H A Dradio-keene.c71 u8 pa; member in struct:keene_device
92 radio->buffer[4] = radio->pa; keene_cmd_main()
260 radio->pa = (ctrl->val - 71) * 100 / 62; keene_s_ctrl()
350 radio->pa = 118; usb_keene_probe()
/linux-4.1.27/arch/sparc/kernel/
H A Dhvtramp.S22 * with physical addressing (va==pa). %o0 contains
H A Dioport.c58 static inline void dma_make_coherent(unsigned long pa, unsigned long len) dma_make_coherent() argument
66 static void __iomem *_sparc_ioremap(struct resource *res, u32 bus, u32 pa, int sz);
216 _sparc_ioremap(struct resource *res, u32 bus, u32 pa, int sz) _sparc_ioremap() argument
218 unsigned long offset = ((unsigned long) pa) & (~PAGE_MASK); _sparc_ioremap()
229 pa &= PAGE_MASK; _sparc_ioremap()
230 srmmu_mapiorange(bus, pa, res->start, resource_size(res)); _sparc_ioremap()
H A Dldc.c2016 static void fill_cookies(struct cookie_state *sp, unsigned long pa, fill_cookies() argument
2020 unsigned long tlen, new = pa + PAGE_SIZE; fill_cookies()
2023 sp->page_table[sp->pte_idx].mte = sp->mte_base | pa; fill_cookies()
2048 pa = new; fill_cookies()
2128 unsigned long npages, pa; ldc_map_single() local
2136 pa = __pa(buf); ldc_map_single()
2137 if ((pa | len) & (8UL - 1)) ldc_map_single()
2140 npages = pages_in_region(pa, len); ldc_map_single()
2155 fill_cookies(&state, (pa & PAGE_MASK), (pa & ~PAGE_MASK), len); ldc_map_single()
/linux-4.1.27/drivers/media/dvb-core/
H A Ddvb_filter.c30 static void setup_ts2pes(ipack *pa, ipack *pv, u16 *pida, u16 *pidv,
34 dvb_filter_ipack_init(pa, IPACKS, pes_write);
36 pa->pid = pida;
38 pa->data = priv;
/linux-4.1.27/arch/parisc/lib/
H A Dmemset.c19 /* Slight modifications for pa-risc linux - Paul Bame <bame@debian.org> */
/linux-4.1.27/arch/parisc/math-emu/
H A Dfpu.h25 * @(#) pa/fp/fpu.h $Revision: 1.1 $
H A Ddenormal.c25 * @(#) pa/fp/denormal.c $ Revision: $
H A Dsfcmp.c25 * @(#) pa/spmath/sfcmp.c $Revision: 1.1 $
H A Ddfcmp.c25 * @(#) pa/spmath/dfcmp.c $Revision: 1.1 $
H A Ddfsqrt.c25 * @(#) pa/spmath/dfsqrt.c $Revision: 1.1 $
H A Dsfsqrt.c25 * @(#) pa/spmath/sfsqrt.c $Revision: 1.1 $
H A Dfloat.h25 * @(#) pa/spmath/float.h $Revision: 1.1 $
33 * /usr/conf/pa/spmath/float.h
H A Ddfdiv.c25 * @(#) pa/spmath/dfdiv.c $Revision: 1.1 $
H A Ddfmpy.c25 * @(#) pa/spmath/dfmpy.c $Revision: 1.1 $
H A Ddfrem.c25 * @(#) pa/spmath/dfrem.c $Revision: 1.1 $
H A Dfcnvff.c25 * @(#) pa/spmath/fcnvff.c $Revision: 1.1 $
H A Dfcnvfut.c25 * @(#) pa/spmath/fcnvfut.c $Revision: 1.1 $
H A Dfcnvfxt.c25 * @(#) pa/spmath/fcnvfxt.c $Revision: 1.1 $
H A Dfcnvuf.c25 * @(#) pa/spmath/fcnvuf.c $Revision: 1.1 $
H A Dfcnvxf.c25 * @(#) pa/spmath/fcnvxf.c $Revision: 1.1 $
H A Dsfdiv.c25 * @(#) pa/spmath/sfdiv.c $Revision: 1.1 $
H A Dsfmpy.c25 * @(#) pa/spmath/sfmpy.c $Revision: 1.1 $
H A Dsfrem.c25 * @(#) pa/spmath/sfrem.c $Revision: 1.1 $
/linux-4.1.27/drivers/scsi/be2iscsi/
H A Dbe_mgmt.h245 #define ISCSI_GET_PDU_TEMPLATE_ADDRESS(pc, pa) {\
246 pa->lo = phba->init_mem[ISCSI_MEM_GLOBAL_HEADER].mem_array[0].\
248 pa->hi = phba->init_mem[ISCSI_MEM_GLOBAL_HEADER].mem_array[0].\
/linux-4.1.27/drivers/scsi/fnic/
H A Dfnic_fcs.c941 dma_addr_t pa; fnic_alloc_rq_frame() local
954 pa = pci_map_single(fnic->pdev, skb->data, len, PCI_DMA_FROMDEVICE); fnic_alloc_rq_frame()
955 fnic_queue_rq_desc(rq, skb, pa, len); fnic_alloc_rq_frame()
980 dma_addr_t pa; fnic_eth_send() local
1004 pa = pci_map_single(fnic->pdev, skb->data, skb->len, PCI_DMA_TODEVICE); fnic_eth_send()
1008 pci_unmap_single(fnic->pdev, pa, skb->len, PCI_DMA_TODEVICE); fnic_eth_send()
1014 fnic_queue_wq_eth_desc(wq, skb, pa, skb->len, fnic_eth_send()
1027 dma_addr_t pa; fnic_send_frame() local
1072 pa = pci_map_single(fnic->pdev, eth_hdr, tot_len, PCI_DMA_TODEVICE); fnic_send_frame()
1082 pci_unmap_single(fnic->pdev, pa, fnic_send_frame()
1088 fnic_queue_wq_desc(wq, skb, pa, tot_len, fr_eof(fp), fnic_send_frame()
/linux-4.1.27/drivers/edac/
H A Dcpc925_edac.c453 unsigned long pa; cpc925_mc_get_pfn() local
473 pa = mci->csrows[rank]->first_page << PAGE_SHIFT; cpc925_mc_get_pfn()
480 pa |= c << (14 - i); cpc925_mc_get_pfn()
484 pa |= bank << 19; cpc925_mc_get_pfn()
490 pa |= c << (26 - i); cpc925_mc_get_pfn()
496 pa |= c << (21 + i); cpc925_mc_get_pfn()
502 pa |= c << (18 - i); cpc925_mc_get_pfn()
508 pa |= c << (29 - i); cpc925_mc_get_pfn()
511 *offset = pa & (PAGE_SIZE - 1); cpc925_mc_get_pfn()
512 *pfn = pa >> PAGE_SHIFT; cpc925_mc_get_pfn()
514 edac_dbg(0, "ECC physical address 0x%lx\n", pa); cpc925_mc_get_pfn()
/linux-4.1.27/arch/x86/platform/intel-quark/
H A Dimr.c236 seq_printf(s, "imr%02i: base=%pa, end=%pa, size=0x%08zx " imr_dbgfs_state_show()
300 pr_err("base %pa size 0x%08zx must align to 1KiB\n", imr_check_params()
412 pr_debug("add %d phys %pa-%pa size %zx mask 0x%08x wmask 0x%08x\n", imr_add_range()
520 pr_debug("remove %d phys %pa-%pa size %zx\n", reg, &base, &end, raw_size); __imr_remove_range()
/linux-4.1.27/arch/powerpc/kvm/
H A Dbook3s_hv_rm_mmu.c141 unsigned long i, pa, gpa, gfn, psize; kvmppc_do_h_enter() local
171 pa = 0; kvmppc_do_h_enter()
222 pa = pte_pfn(pte) << PAGE_SHIFT; kvmppc_do_h_enter()
223 pa |= hva & (host_pte_size - 1); kvmppc_do_h_enter()
224 pa |= gpa & ~PAGE_MASK; kvmppc_do_h_enter()
231 ptel |= pa; kvmppc_do_h_enter()
233 if (pa) kvmppc_do_h_enter()
/linux-4.1.27/sound/pci/
H A Des1968.c1006 u32 pa; snd_es1968_playback_setup() local
1024 pa = es->memory->buf.addr; snd_es1968_playback_setup()
1025 pa -= chip->dma.addr; snd_es1968_playback_setup()
1026 pa >>= 1; /* words */ snd_es1968_playback_setup()
1028 pa |= 0x00400000; /* System RAM (Bit 22) */ snd_es1968_playback_setup()
1033 pa |= 0x00800000; /* (Bit 23) */ snd_es1968_playback_setup()
1035 pa >>= 1; snd_es1968_playback_setup()
1040 es->base[channel] = pa & 0xFFFF; snd_es1968_playback_setup()
1046 apu_set_register(chip, apu, 4, ((pa >> 16) & 0xFF) << 8); snd_es1968_playback_setup()
1047 apu_set_register(chip, apu, 5, pa & 0xFFFF); snd_es1968_playback_setup()
1048 apu_set_register(chip, apu, 6, (pa + size) & 0xFFFF); snd_es1968_playback_setup()
1107 unsigned int pa, unsigned int bsize, init_capture_apu()
1115 snd_es1968_program_wavecache(chip, es, channel, pa, 1); init_capture_apu()
1118 pa -= chip->dma.addr; init_capture_apu()
1119 pa >>= 1; /* words */ init_capture_apu()
1123 es->base[channel] = pa & 0xFFFF; init_capture_apu()
1124 pa |= 0x00400000; /* bit 22 -> System RAM */ init_capture_apu()
1135 apu_set_register(chip, apu, 4, ((pa >> 16) & 0xFF) << 8); init_capture_apu()
1136 apu_set_register(chip, apu, 5, pa & 0xFFFF); init_capture_apu()
1137 apu_set_register(chip, apu, 6, (pa + bsize) & 0xFFFF); init_capture_apu()
1711 unsigned int pa, offset, t; es1968_measure_clock() local
1736 pa = (unsigned int)((memory->buf.addr - chip->dma.addr) >> 1); es1968_measure_clock()
1737 pa |= 0x00400000; /* System RAM (Bit 22) */ es1968_measure_clock()
1744 apu_set_register(chip, apu, 4, ((pa >> 16) & 0xff) << 8); es1968_measure_clock()
1745 apu_set_register(chip, apu, 5, pa & 0xffff); es1968_measure_clock()
1746 apu_set_register(chip, apu, 6, (pa + CLOCK_MEASURE_BUFSIZE/2) & 0xffff); es1968_measure_clock()
1763 __apu_set_register(chip, apu, 5, pa & 0xffff); es1968_measure_clock()
1777 offset -= (pa & 0xffff); es1968_measure_clock()
1106 init_capture_apu(struct es1968 *chip, struct esschan *es, int channel, unsigned int pa, unsigned int bsize, int mode, int route) init_capture_apu() argument
/linux-4.1.27/include/trace/events/
H A Dext4.h612 struct ext4_prealloc_space *pa),
614 TP_ARGS(ac, pa),
628 __entry->pa_pstart = pa->pa_pstart;
629 __entry->pa_lstart = pa->pa_lstart;
630 __entry->pa_len = pa->pa_len;
642 struct ext4_prealloc_space *pa),
644 TP_ARGS(ac, pa)
650 struct ext4_prealloc_space *pa),
652 TP_ARGS(ac, pa)
656 TP_PROTO(struct ext4_prealloc_space *pa,
659 TP_ARGS(pa, block, count),
670 __entry->dev = pa->pa_inode->i_sb->s_dev;
671 __entry->ino = pa->pa_inode->i_ino;
683 TP_PROTO(struct super_block *sb, struct ext4_prealloc_space *pa),
685 TP_ARGS(sb, pa),
696 __entry->pa_pstart = pa->pa_pstart;
697 __entry->pa_len = pa->pa_len;
/linux-4.1.27/drivers/net/can/usb/peak_usb/
H A Dpcan_usb_core.c707 const struct peak_usb_adapter *pa = dev->adapter; peak_usb_set_bittiming() local
709 if (pa->dev_set_bittiming) { peak_usb_set_bittiming()
711 int err = pa->dev_set_bittiming(dev, bt); peak_usb_set_bittiming()
728 const struct peak_usb_adapter *pa = dev->adapter; peak_usb_set_data_bittiming() local
730 if (pa->dev_set_data_bittiming) { peak_usb_set_data_bittiming()
732 int err = pa->dev_set_data_bittiming(dev, bt); peak_usb_set_data_bittiming()
/linux-4.1.27/drivers/scsi/bfa/
H A Dbfa_port.c219 port->stats_dma.pa = dma_pa; bfa_port_mem_claim()
365 bfa_dma_be_addr_set(m->dma_addr, port->stats_dma.pa); bfa_port_get_stats()
639 cee->attr_dma.pa = dma_pa; bfa_cee_mem_claim()
642 cee->stats_dma.pa = dma_pa + BFA_ROUNDUP( bfa_cee_mem_claim()
683 bfa_dma_be_addr_set(cmd->dma_addr, cee->attr_dma.pa); bfa_cee_get_attr()
723 bfa_dma_be_addr_set(cmd->dma_addr, cee->stats_dma.pa); bfa_cee_get_stats()
H A Dbfa_core.c1001 bfa_dma_be_addr_set(cfg_info->cfgrsp_addr, iocfc->cfgrsp_dma.pa); bfa_iocfc_send_cfg()
1007 iocfc->req_cq_ba[i].pa); bfa_iocfc_send_cfg()
1009 iocfc->req_cq_shadow_ci[i].pa); bfa_iocfc_send_cfg()
1014 iocfc->rsp_cq_ba[i].pa); bfa_iocfc_send_cfg()
1016 iocfc->rsp_cq_shadow_pi[i].pa); bfa_iocfc_send_cfg()
1033 bfa_dma_be_addr_set(cfg_req.ioc_cfg_dma_addr, iocfc->cfg_info.pa); bfa_iocfc_send_cfg()
1116 iocfc->req_cq_ba[i].pa = bfa_mem_dma_phys(reqq_dma); bfa_iocfc_mem_claim()
1121 iocfc->rsp_cq_ba[i].pa = bfa_mem_dma_phys(rspq_dma); bfa_iocfc_mem_claim()
1131 iocfc->req_cq_shadow_ci[i].pa = dm_pa; bfa_iocfc_mem_claim()
1136 iocfc->rsp_cq_shadow_pi[i].pa = dm_pa; bfa_iocfc_mem_claim()
1143 bfa->iocfc.cfg_info.pa = dm_pa; bfa_iocfc_mem_claim()
1150 bfa->iocfc.cfgrsp_dma.pa = dm_pa; bfa_iocfc_mem_claim()
H A Dbfa_ioc.h144 /* Get the corresponding dma buf pa for a req - from the tag */
166 u64 pa; /* ! Physical address */ member in struct:bfa_dma_s
178 #define bfa_dma_be_addr_set(dma_addr, pa) \
179 __bfa_dma_be_addr_set(&dma_addr, (u64)pa)
181 __bfa_dma_be_addr_set(union bfi_addr_u *dma_addr, u64 pa) __bfa_dma_be_addr_set() argument
183 dma_addr->a32.addr_lo = cpu_to_be32(pa); __bfa_dma_be_addr_set()
184 dma_addr->a32.addr_hi = cpu_to_be32(pa >> 32); __bfa_dma_be_addr_set()
191 __bfa_alen_set(struct bfi_alen_s *alen, u32 len, u64 pa) __bfa_alen_set() argument
194 bfa_dma_be_addr_set(alen->al_addr, pa); __bfa_alen_set()
/linux-4.1.27/drivers/of/
H A Dfdt.c299 const char *p1 = pathp, *ps = pathp, *pa = NULL; unflatten_dt_node() local
304 pa = p1; unflatten_dt_node()
309 if (pa < ps) unflatten_dt_node()
310 pa = p1; unflatten_dt_node()
311 sz = (pa - ps) + 1; unflatten_dt_node()
491 pr_debug("Reserved memory: reserved region for node '%s': base %pa, size %ld MiB\n", __reserved_mem_reserve_reg()
494 pr_info("Reserved memory: failed to reserve memory for node '%s': base %pa, size %ld MiB\n", __reserved_mem_reserve_reg()
1023 pr_err("Reserved memory not supported, ignoring range 0x%pa - 0x%pa%s\n", early_init_dt_reserve_memory_arch()
H A Dof_reserved_mem.c147 pr_debug("Reserved memory: allocated memory for '%s' node: base %pa, size %ld MiB\n", __reserved_mem_alloc_size()
159 pr_debug("Reserved memory: allocated memory for '%s' node: base %pa, size %ld MiB\n", __reserved_mem_alloc_size()
/linux-4.1.27/arch/tile/include/uapi/arch/
H A Dsim.h374 sim_validate_lines_evicted(unsigned long long pa, unsigned long length) sim_validate_lines_evicted() argument
377 _sim_syscall(SIM_SYSCALL_VALIDATE_LINES_EVICTED, 2, pa, length); sim_validate_lines_evicted()
380 0 /* dummy */, (long)(pa), (long)(pa >> 32), length); sim_validate_lines_evicted()
/linux-4.1.27/arch/alpha/kernel/
H A Dcore_marvel.c859 marvel_pa_to_nid(unsigned long pa)
863 if ((pa >> 43) & 1) /* I/O */ marvel_cpuid_to_nid()
864 cpuid = (~(pa >> 35) & 0xff); marvel_cpuid_to_nid()
866 cpuid = ((pa >> 34) & 0x3) | ((pa >> (37 - 2)) & (0x1f << 2));
880 unsigned long pa; marvel_node_mem_size()
882 pa = (nid & 0x3) | ((nid & (0x1f << 2)) << 1);
883 pa <<= 34;
885 return pa;
849 marvel_pa_to_nid(unsigned long pa) marvel_pa_to_nid() argument
870 unsigned long pa; marvel_node_mem_start() local
H A Dcore_wildfire.c445 int wildfire_pa_to_nid(unsigned long pa) wildfire_pa_to_nid()
447 return pa >> 36;
444 wildfire_pa_to_nid(unsigned long pa) wildfire_pa_to_nid() argument
/linux-4.1.27/drivers/infiniband/hw/mlx5/
H A Dmem.c171 dma_addr_t pa = umem->odp_data->dma_list[offset + i]; __mlx5_ib_populate_pas() local
173 pas[i] = cpu_to_be64(umem_dma_to_mtt(pa)); __mlx5_ib_populate_pas()
/linux-4.1.27/arch/mips/include/asm/octeon/
H A Dcvmx-address.h135 uint64_t pa:49; /* physical address */ member in struct:__anon1911::__anon1915
224 uint64_t pa:49; member in struct:__anon1911::__anon1925
/linux-4.1.27/drivers/acpi/
H A Dnuma.c192 acpi_numa_x2apic_affinity_init(struct acpi_srat_x2apic_cpu_affinity *pa) acpi_numa_x2apic_affinity_init() argument
195 "Found unsupported x2apic [0x%08x] SRAT entry\n", pa->apic_id); acpi_numa_x2apic_affinity_init()
/linux-4.1.27/arch/mips/ralink/
H A Drt305x.c51 static struct rt2880_pmx_func rt3352_pa_func[] = { FUNC("pa", 0, 38, 2) };
77 GRP("pa", rt3352_pa_func, 1, RT3352_GPIO_MODE_PA),
H A Dmt7620.c55 static struct rt2880_pmx_func pa_grp[] = { FUNC("pa", 0, 18, 4) };
96 GRP("pa", pa_grp, 1, MT7620_GPIO_MODE_PA),
/linux-4.1.27/arch/powerpc/kernel/
H A Dprom.c136 * ibm,pa-features is a per-cpu property that contains a string of
144 * pa-features property is missing, or a 1/0 to indicate if the feature
153 unsigned char pabyte; /* byte number in ibm,pa-features */
155 unsigned char invert; /* if 1, pa bit set => clear feature */
216 pa_ftrs = of_get_flat_dt_prop(node, "ibm,pa-features", &tablelen); check_cpu_pa_features()
/linux-4.1.27/drivers/remoteproc/
H A Dremoteproc_core.c459 * the 'pa' member. If a specific device address is expected, then 'da' will
497 ret = iommu_map(rproc->domain, rsc->da, rsc->pa, rsc->len, rsc->flags); rproc_handle_devmem()
514 dev_dbg(dev, "mapped devmem pa 0x%x, da 0x%x, len 0x%x\n", rproc_handle_devmem()
515 rsc->pa, rsc->da, rsc->len); rproc_handle_devmem()
564 dev_dbg(dev, "carveout rsc: da %x, pa %x, len %x, flags %x\n", rproc_handle_carveout()
565 rsc->da, rsc->pa, rsc->len, rsc->flags); rproc_handle_carveout()
631 * Some remote processors might need to know the pa rproc_handle_carveout()
635 * the IOMMU, and therefor must know the pa. rproc_handle_carveout()
647 rsc->pa = dma; rproc_handle_carveout()
/linux-4.1.27/arch/x86/platform/efi/
H A Defi_64.c220 u64 pa = md->phys_addr; efi_map_region() local
245 if (!(pa & (PMD_SIZE - 1))) { efi_map_region()
248 u64 pa_offset = pa & (PMD_SIZE - 1); efi_map_region()
/linux-4.1.27/arch/mips/mm/
H A Dcerr-sb1.c493 unsigned long long taglo, pa; extract_dc() local
514 pa = (taglo & 0xFFFFFFE000ULL) | addr; extract_dc()
528 way, pa, dc_state_str(state), state, taghi, taglo); extract_dc()
/linux-4.1.27/drivers/staging/android/ion/
H A Dion_cma_heap.c116 dev_dbg(dev, "Return buffer %p physical address %pa\n", buffer, ion_cma_phys()
/linux-4.1.27/arch/powerpc/include/asm/
H A Dmmu-hash64.h268 * for the page size. We assume the pa is already "clean" that is properly
271 static inline unsigned long hpte_encode_r(unsigned long pa, int base_psize, hpte_encode_r() argument
276 return pa & HPTE_R_RPN; hpte_encode_r()
280 return (pa & ~((1ul << shift) - 1)) | (penc << LP_SHIFT); hpte_encode_r()
H A Dpage.h79 * va = pa + PAGE_OFFSET - MEMORY_START
80 * va = pa + KERNELBASE - PHYSICAL_START
/linux-4.1.27/arch/m68k/sun3x/
H A Ddvma.c165 // printk("mapping pa %lx to ba %lx\n", __pa(kaddr), index << DVMA_PAGE_SHIFT); dvma_map_iommu()
/linux-4.1.27/tools/perf/util/
H A Dprobe-event.h106 extern int synthesize_perf_probe_arg(struct perf_probe_arg *pa, char *buf,
H A Dprobe-event.c1487 int synthesize_perf_probe_arg(struct perf_probe_arg *pa, char *buf, size_t len) synthesize_perf_probe_arg() argument
1489 struct perf_probe_arg_field *field = pa->field; synthesize_perf_probe_arg()
1493 if (pa->name && pa->var) synthesize_perf_probe_arg()
1494 ret = e_snprintf(tmp, len, "%s=%s", pa->name, pa->var); synthesize_perf_probe_arg()
1496 ret = e_snprintf(tmp, len, "%s", pa->name ? pa->name : pa->var); synthesize_perf_probe_arg()
1515 if (pa->type) { synthesize_perf_probe_arg()
1516 ret = e_snprintf(tmp, len, ":%s", pa->type); synthesize_perf_probe_arg()
/linux-4.1.27/include/linux/bcma/
H A Dbcma_driver_chipcommon.h507 #define BCMA_CHIPCTL_4331_EXTPA_EN BIT(4) /* 0 ext pa disable, 1 ext pa enabled */
515 #define BCMA_CHIPCTL_4331_EXTPA_EN2 BIT(12) /* 0 ext pa disable, 1 ext pa enabled */
/linux-4.1.27/drivers/parisc/
H A Dccio-dma.c560 register unsigned long pa; ccio_io_pdir_entry() local
573 pa = virt_to_phys(vba); ccio_io_pdir_entry()
574 asm volatile("depw %1,31,12,%0" : "+r" (pa) : "r" (hints)); ccio_io_pdir_entry()
575 ((u32 *)pdir_ptr)[1] = (u32) pa; ccio_io_pdir_entry()
587 asm volatile ("extrd,u %1,15,4,%0" : "=r" (ci) : "r" (pa)); ccio_io_pdir_entry()
588 asm volatile ("extrd,u %1,31,16,%0" : "+r" (pa) : "r" (pa)); ccio_io_pdir_entry()
589 asm volatile ("depd %1,35,4,%0" : "+r" (pa) : "r" (ci)); ccio_io_pdir_entry()
591 pa = 0; ccio_io_pdir_entry()
600 asm volatile ("depw %1,15,12,%0" : "+r" (pa) : "r" (ci)); ccio_io_pdir_entry()
602 ((u32 *)pdir_ptr)[0] = (u32) pa; ccio_io_pdir_entry()
H A Dsba_iommu.c570 u64 pa; /* physical address */ sba_io_pdir_entry() local
573 pa = virt_to_phys(vba); sba_io_pdir_entry()
574 pa &= IOVP_MASK; sba_io_pdir_entry()
578 pa |= (ci >> PAGE_SHIFT) & 0xff; /* move CI (8 bits) into lowest byte */ sba_io_pdir_entry()
580 pa |= SBA_PDIR_VALID_BIT; /* set "valid" bit */ sba_io_pdir_entry()
581 *pdir_ptr = cpu_to_le64(pa); /* swap and store into I/O Pdir */ sba_io_pdir_entry()
/linux-4.1.27/arch/tile/include/asm/
H A Dpage.h152 #define __pa_to_highbits(pa) ((phys_addr_t)(pa) >> NR_PA_HIGHBIT_SHIFT)
/linux-4.1.27/drivers/infiniband/hw/qib/
H A Dqib_iba6120.c1862 * @pa: physical address of in memory buffer; tidinvalid if freeing
1869 u32 type, unsigned long pa) qib_6120_put_tid()
1879 if (pa != dd->tidinvalid) { qib_6120_put_tid()
1880 if (pa & ((1U << 11) - 1)) { qib_6120_put_tid()
1882 pa); qib_6120_put_tid()
1885 pa >>= 11; qib_6120_put_tid()
1886 if (pa & ~QLOGIC_IB_RT_ADDR_MASK) { qib_6120_put_tid()
1889 pa); qib_6120_put_tid()
1894 pa |= dd->tidtemplate; qib_6120_put_tid()
1896 pa |= 2 << 29; qib_6120_put_tid()
1918 writel(pa, tidp32); qib_6120_put_tid()
1930 * @pa: physical address of in memory buffer; tidinvalid if freeing
1937 u32 type, unsigned long pa) qib_6120_put_tid_2()
1945 if (pa != dd->tidinvalid) { qib_6120_put_tid_2()
1946 if (pa & ((1U << 11) - 1)) { qib_6120_put_tid_2()
1948 pa); qib_6120_put_tid_2()
1951 pa >>= 11; qib_6120_put_tid_2()
1952 if (pa & ~QLOGIC_IB_RT_ADDR_MASK) { qib_6120_put_tid_2()
1955 pa); qib_6120_put_tid_2()
1960 pa |= dd->tidtemplate; qib_6120_put_tid_2()
1962 pa |= 2 << 29; qib_6120_put_tid_2()
1965 writel(pa, tidp32); qib_6120_put_tid_2()
1868 qib_6120_put_tid(struct qib_devdata *dd, u64 __iomem *tidptr, u32 type, unsigned long pa) qib_6120_put_tid() argument
1936 qib_6120_put_tid_2(struct qib_devdata *dd, u64 __iomem *tidptr, u32 type, unsigned long pa) qib_6120_put_tid_2() argument
/linux-4.1.27/drivers/infiniband/hw/ipath/
H A Dipath_iba6110.c1478 * @pa: physical address of in memory buffer; ipath_tidinvalid if freeing
1486 unsigned long pa) ipath_ht_put_tid()
1491 if (pa != dd->ipath_tidinvalid) { ipath_ht_put_tid()
1492 if (unlikely((pa & ~INFINIPATH_RT_ADDR_MASK))) { ipath_ht_put_tid()
1495 "40 bits, using only 40!!!\n", pa); ipath_ht_put_tid()
1496 pa &= INFINIPATH_RT_ADDR_MASK; ipath_ht_put_tid()
1499 pa |= dd->ipath_tidtemplate; ipath_ht_put_tid()
1504 pa |= lenvalid | INFINIPATH_RT_VALID; ipath_ht_put_tid()
1508 writeq(pa, tidptr); ipath_ht_put_tid()
1484 ipath_ht_put_tid(struct ipath_devdata *dd, u64 __iomem *tidptr, u32 type, unsigned long pa) ipath_ht_put_tid() argument
/linux-4.1.27/drivers/video/fbdev/
H A Dmxsfb.c631 unsigned long pa, fbsize; mxsfb_restore_mode() local
700 pa = readl(host->base + host->devdata->cur_buf); mxsfb_restore_mode()
702 if (pa < fb_info->fix.smem_start) { mxsfb_restore_mode()
706 if (pa + fbsize > fb_info->fix.smem_start + fb_info->fix.smem_len) { mxsfb_restore_mode()
710 ofs = pa - fb_info->fix.smem_start; mxsfb_restore_mode()
/linux-4.1.27/drivers/staging/gdm724x/
H A Dgdm_lte.c195 u16 pa[20]; icmp6_checksum() member in union:__anon9802
205 for (i = 0; i < ARRAY_SIZE(pseudo_header.pa); i++) icmp6_checksum()
206 sum += pseudo_header.pa[i]; icmp6_checksum()
/linux-4.1.27/drivers/pinctrl/
H A Dpinctrl-single.c132 * @pa: pindesc array
140 struct pinctrl_pin_desc *pa; member in struct:pcs_data
781 pin = &pcs->pins.pa[i]; pcs_add_pin()
817 pcs->pins.pa = devm_kzalloc(pcs->dev, pcs_allocate_pin_table()
818 sizeof(*pcs->pins.pa) * nr_pins, pcs_allocate_pin_table()
820 if (!pcs->pins.pa) pcs_allocate_pin_table()
829 pcs->desc.pins = pcs->pins.pa; pcs_allocate_pin_table()
1955 dev_info(pcs->dev, "%i pins at pa %p size %u\n", pcs_probe()
/linux-4.1.27/arch/arm64/mm/
H A Dmmu.c273 pr_warn("BUG: not creating mapping for %pa at 0x%016lx - outside kernel range\n", create_mapping()
293 pr_warn("BUG: not creating mapping for %pa at 0x%016lx - outside kernel range\n", create_mapping_late()
/linux-4.1.27/drivers/memory/tegra/
H A Dmc.c198 dev_err_ratelimited(mc->dev, "%s: %s%s @%pa: %s (%s%s)\n", tegra_mc_irq()
/linux-4.1.27/drivers/misc/mic/card/
H A Dmic_x100.c201 mdev->mmio.pa = MIC_X100_MMIO_BASE; mic_probe()
/linux-4.1.27/drivers/staging/media/lirc/
H A Dlirc_bt829.c77 pr_info("memory at %pa\n", &pci_addr_phys); do_pci_probe()
/linux-4.1.27/arch/tile/kernel/
H A Dmachine_kexec.c222 * Address range in which pa=va mapping is set in setup_quasi_va_is_pa().
/linux-4.1.27/drivers/block/
H A Dpmem.c137 dev_warn(dev, "could not reserve region [0x%pa:0x%zx]\n", &pmem->phys_addr, pmem->size); pmem_alloc()
/linux-4.1.27/arch/arm64/include/asm/
H A Dkvm_mmu.h92 phys_addr_t pa, unsigned long size, bool writable);
/linux-4.1.27/arch/cris/include/arch-v10/arch/
H A Dirq.h25 #define PA_IRQ_NBR IO_BITNR(R_VECT_MASK_RD, pa) /* 11 */
/linux-4.1.27/arch/alpha/mm/
H A Dinit.c210 unsigned long pfn = crb->map[i].pa >> PAGE_SHIFT; callback_init()
/linux-4.1.27/arch/arm/kvm/
H A Dmmu.c961 * @pa: The physical address of the device
965 phys_addr_t pa, unsigned long size, bool writable) kvm_phys_addr_ioremap()
973 pfn = __phys_to_pfn(pa); kvm_phys_addr_ioremap()
1797 phys_addr_t pa; kvm_arch_prepare_memory_region() local
1799 pa = (phys_addr_t)vma->vm_pgoff << PAGE_SHIFT; kvm_arch_prepare_memory_region()
1800 pa += vm_start - vma->vm_start; kvm_arch_prepare_memory_region()
1806 ret = kvm_phys_addr_ioremap(kvm, gpa, pa, kvm_arch_prepare_memory_region()
964 kvm_phys_addr_ioremap(struct kvm *kvm, phys_addr_t guest_ipa, phys_addr_t pa, unsigned long size, bool writable) kvm_phys_addr_ioremap() argument

Completed in 5213 milliseconds

12