/linux-4.1.27/arch/mips/boot/dts/include/dt-bindings/pinctrl/ |
H A D | omap.h | 56 #define OMAP_IOPAD_OFFSET(pa, offset) (((pa) & 0xffff) - (offset)) 58 #define OMAP2420_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0030) (val) 59 #define OMAP2430_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val) 60 #define OMAP3_CORE1_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val) 61 #define OMAP3430_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25d8) (val) 62 #define OMAP3630_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25a0) (val) 63 #define OMAP3_WKUP_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2a00) (val) 64 #define DM816X_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) 65 #define AM33XX_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) 66 #define AM4372_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) 67 #define DRA7XX_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x3400) (val)
|
/linux-4.1.27/arch/powerpc/boot/dts/include/dt-bindings/pinctrl/ |
H A D | omap.h | 56 #define OMAP_IOPAD_OFFSET(pa, offset) (((pa) & 0xffff) - (offset)) 58 #define OMAP2420_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0030) (val) 59 #define OMAP2430_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val) 60 #define OMAP3_CORE1_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val) 61 #define OMAP3430_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25d8) (val) 62 #define OMAP3630_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25a0) (val) 63 #define OMAP3_WKUP_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2a00) (val) 64 #define DM816X_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) 65 #define AM33XX_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) 66 #define AM4372_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) 67 #define DRA7XX_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x3400) (val)
|
/linux-4.1.27/arch/arm64/boot/dts/include/dt-bindings/pinctrl/ |
H A D | omap.h | 56 #define OMAP_IOPAD_OFFSET(pa, offset) (((pa) & 0xffff) - (offset)) 58 #define OMAP2420_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0030) (val) 59 #define OMAP2430_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val) 60 #define OMAP3_CORE1_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val) 61 #define OMAP3430_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25d8) (val) 62 #define OMAP3630_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25a0) (val) 63 #define OMAP3_WKUP_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2a00) (val) 64 #define DM816X_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) 65 #define AM33XX_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) 66 #define AM4372_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) 67 #define DRA7XX_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x3400) (val)
|
/linux-4.1.27/arch/metag/boot/dts/include/dt-bindings/pinctrl/ |
H A D | omap.h | 56 #define OMAP_IOPAD_OFFSET(pa, offset) (((pa) & 0xffff) - (offset)) 58 #define OMAP2420_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0030) (val) 59 #define OMAP2430_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val) 60 #define OMAP3_CORE1_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val) 61 #define OMAP3430_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25d8) (val) 62 #define OMAP3630_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25a0) (val) 63 #define OMAP3_WKUP_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2a00) (val) 64 #define DM816X_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) 65 #define AM33XX_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) 66 #define AM4372_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) 67 #define DRA7XX_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x3400) (val)
|
/linux-4.1.27/arch/arm/boot/dts/include/dt-bindings/pinctrl/ |
H A D | omap.h | 56 #define OMAP_IOPAD_OFFSET(pa, offset) (((pa) & 0xffff) - (offset)) 58 #define OMAP2420_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0030) (val) 59 #define OMAP2430_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val) 60 #define OMAP3_CORE1_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val) 61 #define OMAP3430_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25d8) (val) 62 #define OMAP3630_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25a0) (val) 63 #define OMAP3_WKUP_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2a00) (val) 64 #define DM816X_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) 65 #define AM33XX_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) 66 #define AM4372_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) 67 #define DRA7XX_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x3400) (val)
|
/linux-4.1.27/include/dt-bindings/pinctrl/ |
H A D | omap.h | 56 #define OMAP_IOPAD_OFFSET(pa, offset) (((pa) & 0xffff) - (offset)) 58 #define OMAP2420_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0030) (val) 59 #define OMAP2430_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val) 60 #define OMAP3_CORE1_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2030) (val) 61 #define OMAP3430_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25d8) (val) 62 #define OMAP3630_CORE2_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x25a0) (val) 63 #define OMAP3_WKUP_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x2a00) (val) 64 #define DM816X_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) 65 #define AM33XX_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) 66 #define AM4372_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x0800) (val) 67 #define DRA7XX_CORE_IOPAD(pa, val) OMAP_IOPAD_OFFSET((pa), 0x3400) (val)
|
/linux-4.1.27/drivers/spmi/ |
H A D | spmi-pmic-arb.c | 412 struct spmi_pmic_arb_dev *pa = irq_data_get_irq_chip_data(d); qpnpint_spmi_write() local 416 if (pmic_arb_write_cmd(pa->spmic, SPMI_CMD_EXT_WRITEL, sid, qpnpint_spmi_write() 418 dev_err_ratelimited(&pa->spmic->dev, qpnpint_spmi_write() 425 struct spmi_pmic_arb_dev *pa = irq_data_get_irq_chip_data(d); qpnpint_spmi_read() local 429 if (pmic_arb_read_cmd(pa->spmic, SPMI_CMD_EXT_READL, sid, qpnpint_spmi_read() 431 dev_err_ratelimited(&pa->spmic->dev, qpnpint_spmi_read() 436 static void periph_interrupt(struct spmi_pmic_arb_dev *pa, u8 apid) periph_interrupt() argument 442 status = readl_relaxed(pa->intr + pa->ver_ops->irq_status(apid)); periph_interrupt() 446 irq = irq_find_mapping(pa->domain, periph_interrupt() 447 pa->apid_to_ppid[apid] << 16 periph_interrupt() 456 struct spmi_pmic_arb_dev *pa = irq_get_handler_data(irq); pmic_arb_chained_irq() local 458 void __iomem *intr = pa->intr; pmic_arb_chained_irq() 459 int first = pa->min_apid >> 5; pmic_arb_chained_irq() 460 int last = pa->max_apid >> 5; pmic_arb_chained_irq() 468 pa->ver_ops->owner_acc_status(pa->ee, i)); pmic_arb_chained_irq() 472 periph_interrupt(pa, id + i * 32); pmic_arb_chained_irq() 481 struct spmi_pmic_arb_dev *pa = irq_data_get_irq_chip_data(d); qpnpint_irq_ack() local 487 raw_spin_lock_irqsave(&pa->lock, flags); qpnpint_irq_ack() 488 writel_relaxed(1 << irq, pa->intr + pa->ver_ops->irq_clear(apid)); qpnpint_irq_ack() 489 raw_spin_unlock_irqrestore(&pa->lock, flags); qpnpint_irq_ack() 497 struct spmi_pmic_arb_dev *pa = irq_data_get_irq_chip_data(d); qpnpint_irq_mask() local 504 raw_spin_lock_irqsave(&pa->lock, flags); qpnpint_irq_mask() 505 status = readl_relaxed(pa->intr + pa->ver_ops->acc_enable(apid)); qpnpint_irq_mask() 508 writel_relaxed(status, pa->intr + qpnpint_irq_mask() 509 pa->ver_ops->acc_enable(apid)); qpnpint_irq_mask() 511 raw_spin_unlock_irqrestore(&pa->lock, flags); qpnpint_irq_mask() 519 struct spmi_pmic_arb_dev *pa = irq_data_get_irq_chip_data(d); qpnpint_irq_unmask() local 526 raw_spin_lock_irqsave(&pa->lock, flags); qpnpint_irq_unmask() 527 status = readl_relaxed(pa->intr + pa->ver_ops->acc_enable(apid)); qpnpint_irq_unmask() 530 pa->intr + pa->ver_ops->acc_enable(apid)); qpnpint_irq_unmask() 532 raw_spin_unlock_irqrestore(&pa->lock, flags); qpnpint_irq_unmask() 595 static int search_mapping_table(struct spmi_pmic_arb_dev *pa, search_mapping_table() argument 600 u32 *mapping_table = pa->mapping_table; search_mapping_table() 634 struct spmi_pmic_arb_dev *pa = d->host_data; qpnpint_irq_domain_dt_translate() local 639 dev_dbg(&pa->spmic->dev, qpnpint_irq_domain_dt_translate() 654 err = search_mapping_table(pa, &spec, &apid); qpnpint_irq_domain_dt_translate() 658 pa->apid_to_ppid[apid] = spec.slave << 8 | spec.per; qpnpint_irq_domain_dt_translate() 661 if (apid > pa->max_apid) qpnpint_irq_domain_dt_translate() 662 pa->max_apid = apid; qpnpint_irq_domain_dt_translate() 663 if (apid < pa->min_apid) qpnpint_irq_domain_dt_translate() 664 pa->min_apid = apid; qpnpint_irq_domain_dt_translate() 672 dev_dbg(&pa->spmic->dev, "out_hwirq = %lu\n", *out_hwirq); qpnpint_irq_domain_dt_translate() 681 struct spmi_pmic_arb_dev *pa = d->host_data; qpnpint_irq_domain_map() local 683 dev_dbg(&pa->spmic->dev, "virq = %u, hwirq = %lu\n", virq, hwirq); qpnpint_irq_domain_map() 692 static u32 pmic_arb_offset_v1(struct spmi_pmic_arb_dev *pa, u8 sid, u16 addr) pmic_arb_offset_v1() argument 694 return 0x800 + 0x80 * pa->channel; pmic_arb_offset_v1() 698 static u32 pmic_arb_offset_v2(struct spmi_pmic_arb_dev *pa, u8 sid, u16 addr) pmic_arb_offset_v2() argument 701 u8 chan = pa->ppid_to_chan[ppid]; pmic_arb_offset_v2() 703 return 0x1000 * pa->ee + 0x8000 * chan; pmic_arb_offset_v2() 783 struct spmi_pmic_arb_dev *pa; spmi_pmic_arb_probe() local 791 ctrl = spmi_controller_alloc(&pdev->dev, sizeof(*pa)); spmi_pmic_arb_probe() 795 pa = spmi_controller_get_drvdata(ctrl); spmi_pmic_arb_probe() 796 pa->spmic = ctrl; spmi_pmic_arb_probe() 812 pa->ver_ops = &pmic_arb_v1; spmi_pmic_arb_probe() 813 pa->wr_base = core; spmi_pmic_arb_probe() 814 pa->rd_base = core; spmi_pmic_arb_probe() 820 pa->ver_ops = &pmic_arb_v2; spmi_pmic_arb_probe() 824 pa->rd_base = devm_ioremap_resource(&ctrl->dev, res); spmi_pmic_arb_probe() 825 if (IS_ERR(pa->rd_base)) { spmi_pmic_arb_probe() 826 err = PTR_ERR(pa->rd_base); spmi_pmic_arb_probe() 832 pa->wr_base = devm_ioremap_resource(&ctrl->dev, res); spmi_pmic_arb_probe() 833 if (IS_ERR(pa->wr_base)) { spmi_pmic_arb_probe() 834 err = PTR_ERR(pa->wr_base); spmi_pmic_arb_probe() 838 pa->ppid_to_chan = devm_kzalloc(&ctrl->dev, spmi_pmic_arb_probe() 840 if (!pa->ppid_to_chan) { spmi_pmic_arb_probe() 854 pa->ppid_to_chan[ppid] = chan; spmi_pmic_arb_probe() 859 pa->intr = devm_ioremap_resource(&ctrl->dev, res); spmi_pmic_arb_probe() 860 if (IS_ERR(pa->intr)) { spmi_pmic_arb_probe() 861 err = PTR_ERR(pa->intr); spmi_pmic_arb_probe() 866 pa->cnfg = devm_ioremap_resource(&ctrl->dev, res); spmi_pmic_arb_probe() 867 if (IS_ERR(pa->cnfg)) { spmi_pmic_arb_probe() 868 err = PTR_ERR(pa->cnfg); spmi_pmic_arb_probe() 872 pa->irq = platform_get_irq_byname(pdev, "periph_irq"); spmi_pmic_arb_probe() 873 if (pa->irq < 0) { spmi_pmic_arb_probe() 874 err = pa->irq; spmi_pmic_arb_probe() 890 pa->channel = channel; spmi_pmic_arb_probe() 904 pa->ee = ee; spmi_pmic_arb_probe() 906 for (i = 0; i < ARRAY_SIZE(pa->mapping_table); ++i) spmi_pmic_arb_probe() 907 pa->mapping_table[i] = readl_relaxed( spmi_pmic_arb_probe() 908 pa->cnfg + SPMI_MAPPING_TABLE_REG(i)); spmi_pmic_arb_probe() 912 pa->max_apid = 0; spmi_pmic_arb_probe() 913 pa->min_apid = PMIC_ARB_MAX_PERIPHS - 1; spmi_pmic_arb_probe() 916 raw_spin_lock_init(&pa->lock); spmi_pmic_arb_probe() 923 pa->domain = irq_domain_add_tree(pdev->dev.of_node, spmi_pmic_arb_probe() 924 &pmic_arb_irq_domain_ops, pa); spmi_pmic_arb_probe() 925 if (!pa->domain) { spmi_pmic_arb_probe() 931 irq_set_handler_data(pa->irq, pa); spmi_pmic_arb_probe() 932 irq_set_chained_handler(pa->irq, pmic_arb_chained_irq); spmi_pmic_arb_probe() 941 irq_set_chained_handler(pa->irq, NULL); spmi_pmic_arb_probe() 942 irq_set_handler_data(pa->irq, NULL); spmi_pmic_arb_probe() 943 irq_domain_remove(pa->domain); spmi_pmic_arb_probe() 952 struct spmi_pmic_arb_dev *pa = spmi_controller_get_drvdata(ctrl); spmi_pmic_arb_remove() local 954 irq_set_chained_handler(pa->irq, NULL); spmi_pmic_arb_remove() 955 irq_set_handler_data(pa->irq, NULL); spmi_pmic_arb_remove() 956 irq_domain_remove(pa->domain); spmi_pmic_arb_remove()
|
/linux-4.1.27/arch/unicore32/mm/ |
H A D | proc-macros.S | 88 * va2pa va, pa, tbl, msk, off, err 92 * pa: physical address, result is stored in this register 97 .macro va2pa, va, pa, tbl, msk, off, err=990f 98 movc \pa, p0.c2, #0 102 add \pa, \pa, #PAGE_OFFSET @ pa <- virt addr of page table 103 ldw \pa, [\pa+], \off << #2 @ pa <- the content of pt 104 cand.a \pa, #4 @ test exist bit 106 and \off, \pa, #3 @ off <- the last 2 bits 111 and \pa, \pa, \msk @ pa <- phys addr of 2nd pt 131 and \pa, \pa, \msk 132 or \pa, \pa, \tbl
|
/linux-4.1.27/arch/parisc/include/asm/ |
H A D | hardware.h | 22 pcx = 0, /* pa7000 pa 1.0 */ 23 pcxs = 1, /* pa7000 pa 1.1a */ 24 pcxt = 2, /* pa7100 pa 1.1b */ 25 pcxt_ = 3, /* pa7200 (t') pa 1.1c */ 26 pcxl = 4, /* pa7100lc pa 1.1d */ 27 pcxl2 = 5, /* pa7300lc pa 1.1e */ 28 pcxu = 6, /* pa8000 pa 2.0 */ 29 pcxu_ = 7, /* pa8200 (u+) pa 2.0 */ 30 pcxw = 8, /* pa8500 pa 2.0 */ 31 pcxw_ = 9, /* pa8600 (w+) pa 2.0 */ 32 pcxw2 = 10, /* pa8700 pa 2.0 */ 33 mako = 11, /* pa8800 pa 2.0 */ 34 mako2 = 12 /* pa8900 pa 2.0 */
|
/linux-4.1.27/arch/arm/include/asm/ |
H A D | smp_scu.h | 19 unsigned long pa; scu_a9_get_base() local 21 asm("mrc p15, 4, %0, c15, c0, 0" : "=r" (pa)); scu_a9_get_base() 23 return pa; scu_a9_get_base()
|
H A D | kvm_mmu.h | 60 phys_addr_t pa, unsigned long size, bool writable);
|
/linux-4.1.27/arch/arm/mach-omap1/ |
H A D | io.c | 148 u8 omap_readb(u32 pa) omap_readb() argument 150 return __raw_readb(OMAP1_IO_ADDRESS(pa)); omap_readb() 154 u16 omap_readw(u32 pa) omap_readw() argument 156 return __raw_readw(OMAP1_IO_ADDRESS(pa)); omap_readw() 160 u32 omap_readl(u32 pa) omap_readl() argument 162 return __raw_readl(OMAP1_IO_ADDRESS(pa)); omap_readl() 166 void omap_writeb(u8 v, u32 pa) omap_writeb() argument 168 __raw_writeb(v, OMAP1_IO_ADDRESS(pa)); omap_writeb() 172 void omap_writew(u16 v, u32 pa) omap_writew() argument 174 __raw_writew(v, OMAP1_IO_ADDRESS(pa)); omap_writew() 178 void omap_writel(u32 v, u32 pa) omap_writel() argument 180 __raw_writel(v, OMAP1_IO_ADDRESS(pa)); omap_writel()
|
H A D | clock.c | 994 struct clk *pa; clk_dbg_show_summary() local 1001 pa = c->parent; clk_dbg_show_summary() 1003 c->name, pa ? pa->name : "none", c->rate, clk_dbg_show_summary() 1027 struct clk *pa = c->parent; clk_debugfs_register_one() local 1029 d = debugfs_create_dir(c->name, pa ? pa->dent : clk_debugfs_root); clk_debugfs_register_one() 1059 struct clk *pa = c->parent; clk_debugfs_register() local 1061 if (pa && !pa->dent) { clk_debugfs_register() 1062 err = clk_debugfs_register(pa); clk_debugfs_register()
|
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/engine/dmaobj/ |
H A D | priv.h | 5 #define nvkm_dmaobj_create(p,e,c,pa,sa,d) \ 6 nvkm_dmaobj_create_((p), (e), (c), (pa), (sa), sizeof(**d), (void **)d)
|
/linux-4.1.27/arch/s390/mm/ |
H A D | cmm.c | 66 struct cmm_page_array *pa, *npa; cmm_alloc_pages() local 74 pa = *list; cmm_alloc_pages() 75 if (!pa || pa->index >= CMM_NR_PAGES) { cmm_alloc_pages() 85 pa = *list; cmm_alloc_pages() 86 if (!pa || pa->index >= CMM_NR_PAGES) { cmm_alloc_pages() 87 npa->next = pa; cmm_alloc_pages() 89 pa = npa; cmm_alloc_pages() 90 *list = pa; cmm_alloc_pages() 95 pa->pages[pa->index++] = addr; cmm_alloc_pages() 105 struct cmm_page_array *pa; cmm_free_pages() local 109 pa = *list; cmm_free_pages() 111 if (!pa || pa->index <= 0) cmm_free_pages() 113 addr = pa->pages[--pa->index]; cmm_free_pages() 114 if (pa->index == 0) { cmm_free_pages() 115 pa = pa->next; cmm_free_pages() 117 *list = pa; cmm_free_pages()
|
/linux-4.1.27/arch/mips/include/asm/mach-ip32/ |
H A D | dma-coherence.h | 32 dma_addr_t pa = virt_to_phys(addr) & RAM_OFFSET_MASK; plat_map_dma_mem() local 35 pa += CRIME_HI_MEM_BASE; plat_map_dma_mem() 37 return pa; plat_map_dma_mem() 43 dma_addr_t pa; plat_map_dma_mem_page() local 45 pa = page_to_phys(page) & RAM_OFFSET_MASK; plat_map_dma_mem_page() 48 pa += CRIME_HI_MEM_BASE; plat_map_dma_mem_page() 50 return pa; plat_map_dma_mem_page()
|
/linux-4.1.27/arch/x86/mm/ |
H A D | srat.c | 74 acpi_numa_x2apic_affinity_init(struct acpi_srat_x2apic_cpu_affinity *pa) acpi_numa_x2apic_affinity_init() argument 81 if (pa->header.length < sizeof(struct acpi_srat_x2apic_cpu_affinity)) { acpi_numa_x2apic_affinity_init() 85 if ((pa->flags & ACPI_SRAT_CPU_ENABLED) == 0) acpi_numa_x2apic_affinity_init() 87 pxm = pa->proximity_domain; acpi_numa_x2apic_affinity_init() 88 apic_id = pa->apic_id; acpi_numa_x2apic_affinity_init() 114 acpi_numa_processor_affinity_init(struct acpi_srat_cpu_affinity *pa) acpi_numa_processor_affinity_init() argument 121 if (pa->header.length != sizeof(struct acpi_srat_cpu_affinity)) { acpi_numa_processor_affinity_init() 125 if ((pa->flags & ACPI_SRAT_CPU_ENABLED) == 0) acpi_numa_processor_affinity_init() 127 pxm = pa->proximity_domain_lo; acpi_numa_processor_affinity_init() 129 pxm |= *((unsigned int*)pa->proximity_domain_hi) << 8; acpi_numa_processor_affinity_init() 138 apic_id = (pa->apic_id << 8) | pa->local_sapic_eid; acpi_numa_processor_affinity_init() 140 apic_id = pa->apic_id; acpi_numa_processor_affinity_init()
|
/linux-4.1.27/fs/ext4/ |
H A D | mballoc.c | 318 * - per-pa lock (pa) 321 * - new pa 325 * - find and use pa: 326 * pa 328 * - release consumed pa: 329 * pa 335 * pa 339 * pa 344 * pa 641 struct ext4_prealloc_space *pa; __mb_check_buddy() local 642 pa = list_entry(cur, struct ext4_prealloc_space, pa_group_list); __mb_check_buddy() 643 ext4_get_group_no_and_offset(sb, pa->pa_pstart, &groupnr, &k); __mb_check_buddy() 645 for (i = 0; i < pa->pa_len; i++) __mb_check_buddy() 2672 struct ext4_prealloc_space *pa; ext4_mb_cleanup_pa() local 2677 pa = list_entry(cur, struct ext4_prealloc_space, pa_group_list); ext4_mb_cleanup_pa() 2678 list_del(&pa->pa_group_list); ext4_mb_cleanup_pa() 2680 kmem_cache_free(ext4_pspace_cachep, pa); ext4_mb_cleanup_pa() 3015 struct ext4_prealloc_space *pa; ext4_mb_normalize_request() local 3102 list_for_each_entry_rcu(pa, &ei->i_prealloc_list, pa_inode_list) { ext4_mb_normalize_request() 3105 if (pa->pa_deleted) ext4_mb_normalize_request() 3107 spin_lock(&pa->pa_lock); ext4_mb_normalize_request() 3108 if (pa->pa_deleted) { ext4_mb_normalize_request() 3109 spin_unlock(&pa->pa_lock); ext4_mb_normalize_request() 3113 pa_end = pa->pa_lstart + EXT4_C2B(EXT4_SB(ac->ac_sb), ext4_mb_normalize_request() 3114 pa->pa_len); ext4_mb_normalize_request() 3118 ac->ac_o_ex.fe_logical < pa->pa_lstart)); ext4_mb_normalize_request() 3121 if (pa->pa_lstart >= end || pa_end <= start) { ext4_mb_normalize_request() 3122 spin_unlock(&pa->pa_lock); ext4_mb_normalize_request() 3125 BUG_ON(pa->pa_lstart <= start && pa_end >= end); ext4_mb_normalize_request() 3127 /* adjust start or end to be adjacent to this pa */ ext4_mb_normalize_request() 3131 } else if (pa->pa_lstart > ac->ac_o_ex.fe_logical) { ext4_mb_normalize_request() 3132 BUG_ON(pa->pa_lstart > end); ext4_mb_normalize_request() 3133 end = pa->pa_lstart; ext4_mb_normalize_request() 3135 spin_unlock(&pa->pa_lock); ext4_mb_normalize_request() 3142 list_for_each_entry_rcu(pa, &ei->i_prealloc_list, pa_inode_list) { ext4_mb_normalize_request() 3145 spin_lock(&pa->pa_lock); ext4_mb_normalize_request() 3146 if (pa->pa_deleted == 0) { ext4_mb_normalize_request() 3147 pa_end = pa->pa_lstart + EXT4_C2B(EXT4_SB(ac->ac_sb), ext4_mb_normalize_request() 3148 pa->pa_len); ext4_mb_normalize_request() 3149 BUG_ON(!(start >= pa_end || end <= pa->pa_lstart)); ext4_mb_normalize_request() 3151 spin_unlock(&pa->pa_lock); ext4_mb_normalize_request() 3223 struct ext4_prealloc_space *pa = ac->ac_pa; ext4_discard_allocated_blocks() local 3227 if (pa == NULL) { ext4_discard_allocated_blocks() 3247 if (pa->pa_type == MB_INODE_PA) ext4_discard_allocated_blocks() 3248 pa->pa_free += ac->ac_b_ex.fe_len; ext4_discard_allocated_blocks() 3255 struct ext4_prealloc_space *pa) ext4_mb_use_inode_pa() 3263 start = pa->pa_pstart + (ac->ac_o_ex.fe_logical - pa->pa_lstart); ext4_mb_use_inode_pa() 3264 end = min(pa->pa_pstart + EXT4_C2B(sbi, pa->pa_len), ext4_mb_use_inode_pa() 3271 ac->ac_pa = pa; ext4_mb_use_inode_pa() 3273 BUG_ON(start < pa->pa_pstart); ext4_mb_use_inode_pa() 3274 BUG_ON(end > pa->pa_pstart + EXT4_C2B(sbi, pa->pa_len)); ext4_mb_use_inode_pa() 3275 BUG_ON(pa->pa_free < len); ext4_mb_use_inode_pa() 3276 pa->pa_free -= len; ext4_mb_use_inode_pa() 3278 mb_debug(1, "use %llu/%u from inode pa %p\n", start, len, pa); ext4_mb_use_inode_pa() 3285 struct ext4_prealloc_space *pa) ext4_mb_use_group_pa() 3289 ext4_get_group_no_and_offset(ac->ac_sb, pa->pa_pstart, ext4_mb_use_group_pa() 3294 ac->ac_pa = pa; ext4_mb_use_group_pa() 3298 * instead we correct pa later, after blocks are marked ext4_mb_use_group_pa() 3300 * Other CPUs are prevented from allocating from this pa by lg_mutex ext4_mb_use_group_pa() 3302 mb_debug(1, "use %u/%u from group pa %p\n", pa->pa_lstart-len, len, pa); ext4_mb_use_group_pa() 3313 struct ext4_prealloc_space *pa, ext4_mb_check_group_pa() 3319 atomic_inc(&pa->pa_count); ext4_mb_check_group_pa() 3320 return pa; ext4_mb_check_group_pa() 3323 new_distance = abs(goal_block - pa->pa_pstart); ext4_mb_check_group_pa() 3330 atomic_inc(&pa->pa_count); ext4_mb_check_group_pa() 3331 return pa; ext4_mb_check_group_pa() 3344 struct ext4_prealloc_space *pa, *cpa = NULL; ext4_mb_use_preallocated() local 3353 list_for_each_entry_rcu(pa, &ei->i_prealloc_list, pa_inode_list) { ext4_mb_use_preallocated() 3357 if (ac->ac_o_ex.fe_logical < pa->pa_lstart || ext4_mb_use_preallocated() 3358 ac->ac_o_ex.fe_logical >= (pa->pa_lstart + ext4_mb_use_preallocated() 3359 EXT4_C2B(sbi, pa->pa_len))) ext4_mb_use_preallocated() 3364 (pa->pa_pstart + EXT4_C2B(sbi, pa->pa_len) > ext4_mb_use_preallocated() 3369 spin_lock(&pa->pa_lock); ext4_mb_use_preallocated() 3370 if (pa->pa_deleted == 0 && pa->pa_free) { ext4_mb_use_preallocated() 3371 atomic_inc(&pa->pa_count); ext4_mb_use_preallocated() 3372 ext4_mb_use_inode_pa(ac, pa); ext4_mb_use_preallocated() 3373 spin_unlock(&pa->pa_lock); ext4_mb_use_preallocated() 3378 spin_unlock(&pa->pa_lock); ext4_mb_use_preallocated() 3402 list_for_each_entry_rcu(pa, &lg->lg_prealloc_list[i], ext4_mb_use_preallocated() 3404 spin_lock(&pa->pa_lock); ext4_mb_use_preallocated() 3405 if (pa->pa_deleted == 0 && ext4_mb_use_preallocated() 3406 pa->pa_free >= ac->ac_o_ex.fe_len) { ext4_mb_use_preallocated() 3409 pa, cpa); ext4_mb_use_preallocated() 3411 spin_unlock(&pa->pa_lock); ext4_mb_use_preallocated() 3457 struct ext4_prealloc_space *pa; ext4_mb_generate_from_pa() local 3473 pa = list_entry(cur, struct ext4_prealloc_space, pa_group_list); ext4_mb_generate_from_pa() 3474 spin_lock(&pa->pa_lock); ext4_mb_generate_from_pa() 3475 ext4_get_group_no_and_offset(sb, pa->pa_pstart, ext4_mb_generate_from_pa() 3477 len = pa->pa_len; ext4_mb_generate_from_pa() 3478 spin_unlock(&pa->pa_lock); ext4_mb_generate_from_pa() 3490 struct ext4_prealloc_space *pa; ext4_mb_pa_callback() local 3491 pa = container_of(head, struct ext4_prealloc_space, u.pa_rcu); ext4_mb_pa_callback() 3493 BUG_ON(atomic_read(&pa->pa_count)); ext4_mb_pa_callback() 3494 BUG_ON(pa->pa_deleted == 0); ext4_mb_pa_callback() 3495 kmem_cache_free(ext4_pspace_cachep, pa); ext4_mb_pa_callback() 3503 struct super_block *sb, struct ext4_prealloc_space *pa) ext4_mb_put_pa() 3509 spin_lock(&pa->pa_lock); ext4_mb_put_pa() 3510 if (!atomic_dec_and_test(&pa->pa_count) || pa->pa_free != 0) { ext4_mb_put_pa() 3511 spin_unlock(&pa->pa_lock); ext4_mb_put_pa() 3515 if (pa->pa_deleted == 1) { ext4_mb_put_pa() 3516 spin_unlock(&pa->pa_lock); ext4_mb_put_pa() 3520 pa->pa_deleted = 1; ext4_mb_put_pa() 3521 spin_unlock(&pa->pa_lock); ext4_mb_put_pa() 3523 grp_blk = pa->pa_pstart; ext4_mb_put_pa() 3526 * next group when pa is used up ext4_mb_put_pa() 3528 if (pa->pa_type == MB_GROUP_PA) ext4_mb_put_pa() 3548 list_del(&pa->pa_group_list); ext4_mb_put_pa() 3551 spin_lock(pa->pa_obj_lock); ext4_mb_put_pa() 3552 list_del_rcu(&pa->pa_inode_list); ext4_mb_put_pa() 3553 spin_unlock(pa->pa_obj_lock); ext4_mb_put_pa() 3555 call_rcu(&(pa)->u.pa_rcu, ext4_mb_pa_callback); ext4_mb_put_pa() 3566 struct ext4_prealloc_space *pa; ext4_mb_new_inode_pa() local 3575 pa = kmem_cache_alloc(ext4_pspace_cachep, GFP_NOFS); ext4_mb_new_inode_pa() 3576 if (pa == NULL) ext4_mb_new_inode_pa() 3617 pa->pa_lstart = ac->ac_b_ex.fe_logical; ext4_mb_new_inode_pa() 3618 pa->pa_pstart = ext4_grp_offs_to_block(sb, &ac->ac_b_ex); ext4_mb_new_inode_pa() 3619 pa->pa_len = ac->ac_b_ex.fe_len; ext4_mb_new_inode_pa() 3620 pa->pa_free = pa->pa_len; ext4_mb_new_inode_pa() 3621 atomic_set(&pa->pa_count, 1); ext4_mb_new_inode_pa() 3622 spin_lock_init(&pa->pa_lock); ext4_mb_new_inode_pa() 3623 INIT_LIST_HEAD(&pa->pa_inode_list); ext4_mb_new_inode_pa() 3624 INIT_LIST_HEAD(&pa->pa_group_list); ext4_mb_new_inode_pa() 3625 pa->pa_deleted = 0; ext4_mb_new_inode_pa() 3626 pa->pa_type = MB_INODE_PA; ext4_mb_new_inode_pa() 3628 mb_debug(1, "new inode pa %p: %llu/%u for %u\n", pa, ext4_mb_new_inode_pa() 3629 pa->pa_pstart, pa->pa_len, pa->pa_lstart); ext4_mb_new_inode_pa() 3630 trace_ext4_mb_new_inode_pa(ac, pa); ext4_mb_new_inode_pa() 3632 ext4_mb_use_inode_pa(ac, pa); ext4_mb_new_inode_pa() 3633 atomic_add(pa->pa_free, &sbi->s_mb_preallocated); ext4_mb_new_inode_pa() 3638 pa->pa_obj_lock = &ei->i_prealloc_lock; ext4_mb_new_inode_pa() 3639 pa->pa_inode = ac->ac_inode; ext4_mb_new_inode_pa() 3642 list_add(&pa->pa_group_list, &grp->bb_prealloc_list); ext4_mb_new_inode_pa() 3645 spin_lock(pa->pa_obj_lock); ext4_mb_new_inode_pa() 3646 list_add_rcu(&pa->pa_inode_list, &ei->i_prealloc_list); ext4_mb_new_inode_pa() 3647 spin_unlock(pa->pa_obj_lock); ext4_mb_new_inode_pa() 3660 struct ext4_prealloc_space *pa; ext4_mb_new_group_pa() local 3669 pa = kmem_cache_alloc(ext4_pspace_cachep, GFP_NOFS); ext4_mb_new_group_pa() 3670 if (pa == NULL) ext4_mb_new_group_pa() 3677 pa->pa_pstart = ext4_grp_offs_to_block(sb, &ac->ac_b_ex); ext4_mb_new_group_pa() 3678 pa->pa_lstart = pa->pa_pstart; ext4_mb_new_group_pa() 3679 pa->pa_len = ac->ac_b_ex.fe_len; ext4_mb_new_group_pa() 3680 pa->pa_free = pa->pa_len; ext4_mb_new_group_pa() 3681 atomic_set(&pa->pa_count, 1); ext4_mb_new_group_pa() 3682 spin_lock_init(&pa->pa_lock); ext4_mb_new_group_pa() 3683 INIT_LIST_HEAD(&pa->pa_inode_list); ext4_mb_new_group_pa() 3684 INIT_LIST_HEAD(&pa->pa_group_list); ext4_mb_new_group_pa() 3685 pa->pa_deleted = 0; ext4_mb_new_group_pa() 3686 pa->pa_type = MB_GROUP_PA; ext4_mb_new_group_pa() 3688 mb_debug(1, "new group pa %p: %llu/%u for %u\n", pa, ext4_mb_new_group_pa() 3689 pa->pa_pstart, pa->pa_len, pa->pa_lstart); ext4_mb_new_group_pa() 3690 trace_ext4_mb_new_group_pa(ac, pa); ext4_mb_new_group_pa() 3692 ext4_mb_use_group_pa(ac, pa); ext4_mb_new_group_pa() 3693 atomic_add(pa->pa_free, &EXT4_SB(sb)->s_mb_preallocated); ext4_mb_new_group_pa() 3699 pa->pa_obj_lock = &lg->lg_prealloc_lock; ext4_mb_new_group_pa() 3700 pa->pa_inode = NULL; ext4_mb_new_group_pa() 3703 list_add(&pa->pa_group_list, &grp->bb_prealloc_list); ext4_mb_new_group_pa() 3707 * We will later add the new pa to the right bucket ext4_mb_new_group_pa() 3727 * @pa must be unlinked from inode and group lists, so that 3734 struct ext4_prealloc_space *pa) ext4_mb_release_inode_pa() 3746 BUG_ON(pa->pa_deleted == 0); ext4_mb_release_inode_pa() 3747 ext4_get_group_no_and_offset(sb, pa->pa_pstart, &group, &bit); ext4_mb_release_inode_pa() 3748 grp_blk_start = pa->pa_pstart - EXT4_C2B(sbi, bit); ext4_mb_release_inode_pa() 3749 BUG_ON(group != e4b->bd_group && pa->pa_len != 0); ext4_mb_release_inode_pa() 3750 end = bit + pa->pa_len; ext4_mb_release_inode_pa() 3763 trace_ext4_mb_release_inode_pa(pa, (grp_blk_start + ext4_mb_release_inode_pa() 3766 mb_free_blocks(pa->pa_inode, e4b, bit, next - bit); ext4_mb_release_inode_pa() 3769 if (free != pa->pa_free) { ext4_mb_release_inode_pa() 3771 "pa %p: logic %lu, phys. %lu, len %lu", ext4_mb_release_inode_pa() 3772 pa, (unsigned long) pa->pa_lstart, ext4_mb_release_inode_pa() 3773 (unsigned long) pa->pa_pstart, ext4_mb_release_inode_pa() 3774 (unsigned long) pa->pa_len); ext4_mb_release_inode_pa() 3776 free, pa->pa_free); ext4_mb_release_inode_pa() 3778 * pa is already deleted so we use the value obtained ext4_mb_release_inode_pa() 3789 struct ext4_prealloc_space *pa) ext4_mb_release_group_pa() 3795 trace_ext4_mb_release_group_pa(sb, pa); ext4_mb_release_group_pa() 3796 BUG_ON(pa->pa_deleted == 0); ext4_mb_release_group_pa() 3797 ext4_get_group_no_and_offset(sb, pa->pa_pstart, &group, &bit); ext4_mb_release_group_pa() 3798 BUG_ON(group != e4b->bd_group && pa->pa_len != 0); ext4_mb_release_group_pa() 3799 mb_free_blocks(pa->pa_inode, e4b, bit, pa->pa_len); ext4_mb_release_group_pa() 3800 atomic_add(pa->pa_len, &EXT4_SB(sb)->s_mb_discarded); ext4_mb_release_group_pa() 3801 trace_ext4_mballoc_discard(sb, NULL, group, bit, pa->pa_len); ext4_mb_release_group_pa() 3821 struct ext4_prealloc_space *pa, *tmp; ext4_mb_discard_group_preallocations() local 3852 list_for_each_entry_safe(pa, tmp, ext4_mb_discard_group_preallocations() 3854 spin_lock(&pa->pa_lock); ext4_mb_discard_group_preallocations() 3855 if (atomic_read(&pa->pa_count)) { ext4_mb_discard_group_preallocations() 3856 spin_unlock(&pa->pa_lock); ext4_mb_discard_group_preallocations() 3860 if (pa->pa_deleted) { ext4_mb_discard_group_preallocations() 3861 spin_unlock(&pa->pa_lock); ext4_mb_discard_group_preallocations() 3866 pa->pa_deleted = 1; ext4_mb_discard_group_preallocations() 3869 free += pa->pa_free; ext4_mb_discard_group_preallocations() 3871 spin_unlock(&pa->pa_lock); ext4_mb_discard_group_preallocations() 3873 list_del(&pa->pa_group_list); ext4_mb_discard_group_preallocations() 3874 list_add(&pa->u.pa_tmp_list, &list); ext4_mb_discard_group_preallocations() 3892 list_for_each_entry_safe(pa, tmp, &list, u.pa_tmp_list) { ext4_mb_discard_group_preallocations() 3895 spin_lock(pa->pa_obj_lock); ext4_mb_discard_group_preallocations() 3896 list_del_rcu(&pa->pa_inode_list); ext4_mb_discard_group_preallocations() 3897 spin_unlock(pa->pa_obj_lock); ext4_mb_discard_group_preallocations() 3899 if (pa->pa_type == MB_GROUP_PA) ext4_mb_discard_group_preallocations() 3900 ext4_mb_release_group_pa(&e4b, pa); ext4_mb_discard_group_preallocations() 3902 ext4_mb_release_inode_pa(&e4b, bitmap_bh, pa); ext4_mb_discard_group_preallocations() 3904 list_del(&pa->u.pa_tmp_list); ext4_mb_discard_group_preallocations() 3905 call_rcu(&(pa)->u.pa_rcu, ext4_mb_pa_callback); ext4_mb_discard_group_preallocations() 3929 struct ext4_prealloc_space *pa, *tmp; ext4_discard_preallocations() local 3946 /* first, collect all pa's in the inode */ ext4_discard_preallocations() 3949 pa = list_entry(ei->i_prealloc_list.next, ext4_discard_preallocations() 3951 BUG_ON(pa->pa_obj_lock != &ei->i_prealloc_lock); ext4_discard_preallocations() 3952 spin_lock(&pa->pa_lock); ext4_discard_preallocations() 3953 if (atomic_read(&pa->pa_count)) { ext4_discard_preallocations() 3956 spin_unlock(&pa->pa_lock); ext4_discard_preallocations() 3959 "uh-oh! used pa while discarding"); ext4_discard_preallocations() 3965 if (pa->pa_deleted == 0) { ext4_discard_preallocations() 3966 pa->pa_deleted = 1; ext4_discard_preallocations() 3967 spin_unlock(&pa->pa_lock); ext4_discard_preallocations() 3968 list_del_rcu(&pa->pa_inode_list); ext4_discard_preallocations() 3969 list_add(&pa->u.pa_tmp_list, &list); ext4_discard_preallocations() 3973 /* someone is deleting pa right now */ ext4_discard_preallocations() 3974 spin_unlock(&pa->pa_lock); ext4_discard_preallocations() 3978 * doesn't mean pa is already unlinked from ext4_discard_preallocations() 3982 * pa from inode's list may access already ext4_discard_preallocations() 3994 list_for_each_entry_safe(pa, tmp, &list, u.pa_tmp_list) { ext4_discard_preallocations() 3995 BUG_ON(pa->pa_type != MB_INODE_PA); ext4_discard_preallocations() 3996 group = ext4_get_group_number(sb, pa->pa_pstart); ext4_discard_preallocations() 4014 list_del(&pa->pa_group_list); ext4_discard_preallocations() 4015 ext4_mb_release_inode_pa(&e4b, bitmap_bh, pa); ext4_discard_preallocations() 4021 list_del(&pa->u.pa_tmp_list); ext4_discard_preallocations() 4022 call_rcu(&(pa)->u.pa_rcu, ext4_mb_pa_callback); ext4_discard_preallocations() 4061 struct ext4_prealloc_space *pa; ext4_mb_show_ac() local 4066 pa = list_entry(cur, struct ext4_prealloc_space, ext4_mb_show_ac() 4068 spin_lock(&pa->pa_lock); ext4_mb_show_ac() 4069 ext4_get_group_no_and_offset(sb, pa->pa_pstart, ext4_mb_show_ac() 4071 spin_unlock(&pa->pa_lock); ext4_mb_show_ac() 4073 start, pa->pa_len); ext4_mb_show_ac() 4209 struct ext4_prealloc_space *pa, *tmp; ext4_mb_discard_lg_preallocations() local 4216 list_for_each_entry_rcu(pa, &lg->lg_prealloc_list[order], ext4_mb_discard_lg_preallocations() 4218 spin_lock(&pa->pa_lock); ext4_mb_discard_lg_preallocations() 4219 if (atomic_read(&pa->pa_count)) { ext4_mb_discard_lg_preallocations() 4221 * This is the pa that we just used ext4_mb_discard_lg_preallocations() 4225 spin_unlock(&pa->pa_lock); ext4_mb_discard_lg_preallocations() 4228 if (pa->pa_deleted) { ext4_mb_discard_lg_preallocations() 4229 spin_unlock(&pa->pa_lock); ext4_mb_discard_lg_preallocations() 4233 BUG_ON(pa->pa_type != MB_GROUP_PA); ext4_mb_discard_lg_preallocations() 4236 pa->pa_deleted = 1; ext4_mb_discard_lg_preallocations() 4237 spin_unlock(&pa->pa_lock); ext4_mb_discard_lg_preallocations() 4239 list_del_rcu(&pa->pa_inode_list); ext4_mb_discard_lg_preallocations() 4240 list_add(&pa->u.pa_tmp_list, &discard_list); ext4_mb_discard_lg_preallocations() 4255 list_for_each_entry_safe(pa, tmp, &discard_list, u.pa_tmp_list) { ext4_mb_discard_lg_preallocations() 4257 group = ext4_get_group_number(sb, pa->pa_pstart); ext4_mb_discard_lg_preallocations() 4264 list_del(&pa->pa_group_list); ext4_mb_discard_lg_preallocations() 4265 ext4_mb_release_group_pa(&e4b, pa); ext4_mb_discard_lg_preallocations() 4269 list_del(&pa->u.pa_tmp_list); ext4_mb_discard_lg_preallocations() 4270 call_rcu(&(pa)->u.pa_rcu, ext4_mb_pa_callback); ext4_mb_discard_lg_preallocations() 4288 struct ext4_prealloc_space *tmp_pa, *pa = ac->ac_pa; ext4_mb_add_n_trim() local 4290 order = fls(pa->pa_free) - 1; ext4_mb_add_n_trim() 4303 if (!added && pa->pa_free < tmp_pa->pa_free) { ext4_mb_add_n_trim() 4305 list_add_tail_rcu(&pa->pa_inode_list, ext4_mb_add_n_trim() 4317 list_add_tail_rcu(&pa->pa_inode_list, ext4_mb_add_n_trim() 4336 struct ext4_prealloc_space *pa = ac->ac_pa; ext4_mb_release_context() local 4337 if (pa) { ext4_mb_release_context() 4338 if (pa->pa_type == MB_GROUP_PA) { ext4_mb_release_context() 4340 spin_lock(&pa->pa_lock); ext4_mb_release_context() 4341 pa->pa_pstart += EXT4_C2B(sbi, ac->ac_b_ex.fe_len); ext4_mb_release_context() 4342 pa->pa_lstart += EXT4_C2B(sbi, ac->ac_b_ex.fe_len); ext4_mb_release_context() 4343 pa->pa_free -= ac->ac_b_ex.fe_len; ext4_mb_release_context() 4344 pa->pa_len -= ac->ac_b_ex.fe_len; ext4_mb_release_context() 4345 spin_unlock(&pa->pa_lock); ext4_mb_release_context() 4348 if (pa) { ext4_mb_release_context() 4350 * We want to add the pa to the right bucket. ext4_mb_release_context() 4355 if ((pa->pa_type == MB_GROUP_PA) && likely(pa->pa_free)) { ext4_mb_release_context() 4356 spin_lock(pa->pa_obj_lock); ext4_mb_release_context() 4357 list_del_rcu(&pa->pa_inode_list); ext4_mb_release_context() 4358 spin_unlock(pa->pa_obj_lock); ext4_mb_release_context() 4361 ext4_mb_put_pa(ac, ac->ac_sb, pa); ext4_mb_release_context() 3254 ext4_mb_use_inode_pa(struct ext4_allocation_context *ac, struct ext4_prealloc_space *pa) ext4_mb_use_inode_pa() argument 3284 ext4_mb_use_group_pa(struct ext4_allocation_context *ac, struct ext4_prealloc_space *pa) ext4_mb_use_group_pa() argument 3312 ext4_mb_check_group_pa(ext4_fsblk_t goal_block, struct ext4_prealloc_space *pa, struct ext4_prealloc_space *cpa) ext4_mb_check_group_pa() argument 3502 ext4_mb_put_pa(struct ext4_allocation_context *ac, struct super_block *sb, struct ext4_prealloc_space *pa) ext4_mb_put_pa() argument 3733 ext4_mb_release_inode_pa(struct ext4_buddy *e4b, struct buffer_head *bitmap_bh, struct ext4_prealloc_space *pa) ext4_mb_release_inode_pa() argument 3788 ext4_mb_release_group_pa(struct ext4_buddy *e4b, struct ext4_prealloc_space *pa) ext4_mb_release_group_pa() argument
|
H A D | mballoc.h | 127 unsigned short pa_type; /* pa type. inode or group */
|
/linux-4.1.27/sound/pci/asihpi/ |
H A D | hpioctl.c | 181 struct hpi_adapter *pa = NULL; asihpi_hpi_ioctl() local 184 pa = &adapters[hm->h.adapter_index]; asihpi_hpi_ioctl() 186 if (!pa || !pa->adapter || !pa->adapter->type) { asihpi_hpi_ioctl() 199 if (mutex_lock_interruptible(&pa->mutex)) { asihpi_hpi_ioctl() 216 if (pa->buffer_size < size) { asihpi_hpi_ioctl() 221 pa->buffer_size, size); asihpi_hpi_ioctl() 222 if (pa->p_buffer) { asihpi_hpi_ioctl() 223 pa->buffer_size = 0; asihpi_hpi_ioctl() 224 vfree(pa->p_buffer); asihpi_hpi_ioctl() 226 pa->p_buffer = vmalloc(size); asihpi_hpi_ioctl() 227 if (pa->p_buffer) asihpi_hpi_ioctl() 228 pa->buffer_size = size; asihpi_hpi_ioctl() 235 mutex_unlock(&pa->mutex); asihpi_hpi_ioctl() 241 hm->m0.u.d.u.data.pb_data = pa->p_buffer; asihpi_hpi_ioctl() 257 copy_from_user(pa->p_buffer, ptr, size); asihpi_hpi_ioctl() 269 copy_to_user(ptr, pa->p_buffer, size); asihpi_hpi_ioctl() 276 mutex_unlock(&pa->mutex); asihpi_hpi_ioctl() 523 struct hpi_adapter *pa; asihpi_adapter_remove() local 526 pa = pci_get_drvdata(pci_dev); asihpi_adapter_remove() 527 pci = pa->adapter->pci; asihpi_adapter_remove() 532 hm.adapter_index = pa->adapter->index; asihpi_adapter_remove() 540 hm.adapter_index = pa->adapter->index; asihpi_adapter_remove() 547 if (pa->irq) asihpi_adapter_remove() 548 free_irq(pa->irq, pa); asihpi_adapter_remove() 550 vfree(pa->p_buffer); asihpi_adapter_remove() 557 pci_dev->devfn, pa->adapter->index); asihpi_adapter_remove() 559 memset(pa, 0, sizeof(*pa)); asihpi_adapter_remove()
|
/linux-4.1.27/arch/mips/include/asm/mach-ip27/ |
H A D | dma-coherence.h | 24 dma_addr_t pa = dev_to_baddr(dev, virt_to_phys(addr)); plat_map_dma_mem() local 26 return pa; plat_map_dma_mem() 32 dma_addr_t pa = dev_to_baddr(dev, page_to_phys(page)); plat_map_dma_mem_page() local 34 return pa; plat_map_dma_mem_page()
|
/linux-4.1.27/drivers/infiniband/hw/usnic/ |
H A D | usnic_uiom.c | 85 dma_addr_t pa; usnic_uiom_put_pages() local 90 pa = sg_phys(sg); list_for_each_entry_safe() 94 usnic_dbg("pa: %pa\n", &pa); list_for_each_entry_safe() 114 dma_addr_t pa; usnic_uiom_get_pages() local 173 pa = sg_phys(sg); usnic_uiom_get_pages() 174 usnic_dbg("va: 0x%lx pa: %pa\n", usnic_uiom_get_pages() 175 cur_base + i*PAGE_SIZE, &pa); usnic_uiom_get_pages() 253 dma_addr_t pa; usnic_uiom_map_sorted_intervals() local 267 pa = sg_phys(&chunk->page_list[i]); list_for_each_entry() 274 pa_start = pa; list_for_each_entry() 275 pa_end = pa; list_for_each_entry() 280 if ((pa_end + PAGE_SIZE != pa) && list_for_each_entry() 281 (pa != pa_start)) { list_for_each_entry() 284 usnic_dbg("va 0x%lx pa %pa size 0x%zx flags 0x%x", list_for_each_entry() 289 usnic_err("Failed to map va 0x%lx pa %pa size 0x%zx with err %d\n", list_for_each_entry() 294 pa_start = pa; list_for_each_entry() 295 pa_end = pa; list_for_each_entry() 300 size = pa - pa_start + PAGE_SIZE; list_for_each_entry() 301 usnic_dbg("va 0x%lx pa %pa size 0x%zx flags 0x%x\n", list_for_each_entry() 306 usnic_err("Failed to map va 0x%lx pa %pa size 0x%zx with err %d\n", list_for_each_entry() 313 if (pa != pa_start) list_for_each_entry()
|
/linux-4.1.27/arch/mips/bmips/ |
H A D | dma.c | 43 static dma_addr_t bmips_phys_to_dma(struct device *dev, phys_addr_t pa) bmips_phys_to_dma() argument 48 if (pa >= r->child_addr && bmips_phys_to_dma() 49 pa < (r->child_addr + r->size)) bmips_phys_to_dma() 50 return pa - r->child_addr + r->parent_addr; bmips_phys_to_dma() 52 return pa; bmips_phys_to_dma()
|
/linux-4.1.27/arch/x86/kernel/ |
H A D | head_32.S | 28 #define pa(X) ((X) - __PAGE_OFFSET) define 55 * KERNEL_IMAGE_SIZE should be greater than pa(_end) 97 movl pa(stack_start),%ecx 107 lgdt pa(boot_gdt_descr) 122 movl $pa(__bss_start),%edi 123 movl $pa(__bss_stop),%ecx 135 movl $pa(boot_params),%edi 140 movl pa(boot_params) + NEW_CL_POINTER,%esi 143 movl $pa(boot_command_line),%edi 152 movl %eax, pa(olpc_ofw_pgd) 182 movl $pa(__brk_base), %edi 183 movl $pa(initial_pg_pmd), %edx 202 movl $pa(_end) + MAPPING_BEYOND_END + PTE_IDENT_ATTR, %ebp 207 movl %edi, pa(_brk_end) 209 movl %eax, pa(max_pfn_mapped) 212 movl $pa(initial_pg_fixmap)+PDE_IDENT_ATTR,%eax 213 movl %eax,pa(initial_pg_pmd+0x1000*KPMDS-8) 218 movl $pa(__brk_base), %edi 219 movl $pa(initial_page_table), %edx 234 movl $pa(_end) + MAPPING_BEYOND_END + PTE_IDENT_ATTR, %ebp 238 movl %edi, pa(_brk_end) 240 movl %eax, pa(max_pfn_mapped) 243 movl $pa(initial_pg_fixmap)+PDE_IDENT_ATTR,%eax 244 movl %eax,pa(initial_page_table+0xffc) 249 cmpw $0x207, pa(boot_params + BP_version) 254 movl pa(boot_params + BP_hardware_subarch), %eax 258 movl pa(subarch_entries)(,%eax,4), %eax 310 movl pa(stack_start),%ecx 343 movl $-1,pa(X86_CPUID) # preset CPUID level 355 movl %eax,pa(X86_CPUID) # save largest std CPUID function 362 movl pa(mmu_cr4_features),%eax 398 movl $pa(initial_page_table), %eax 691 .long pa(initial_pg_pmd+PGD_IDENT_ATTR),0 /* low identity map */ 693 .long pa(initial_pg_pmd+PGD_IDENT_ATTR),0 694 .long pa(initial_pg_pmd+PGD_IDENT_ATTR+0x1000),0 695 .long pa(initial_pg_pmd+PGD_IDENT_ATTR+0x2000),0 698 .long pa(initial_pg_pmd+PGD_IDENT_ATTR),0 699 .long pa(initial_pg_pmd+PGD_IDENT_ATTR+0x1000),0 703 .long pa(initial_pg_pmd+PGD_IDENT_ATTR),0
|
H A D | kdebugfs.c | 38 u64 pa; setup_data_read() local 49 pa = node->paddr + sizeof(struct setup_data) + pos; setup_data_read() 50 pg = pfn_to_page((pa + count - 1) >> PAGE_SHIFT); setup_data_read() 52 p = ioremap_cache(pa, count); setup_data_read() 56 p = __va(pa); setup_data_read()
|
H A D | kvm.c | 340 u64 pa = slow_virt_to_phys(this_cpu_ptr(&apf_reason)); kvm_guest_cpu_init() local 343 pa |= KVM_ASYNC_PF_SEND_ALWAYS; kvm_guest_cpu_init() 345 wrmsrl(MSR_KVM_ASYNC_PF_EN, pa | KVM_ASYNC_PF_ENABLED); kvm_guest_cpu_init() 352 unsigned long pa; kvm_guest_cpu_init() local 356 pa = slow_virt_to_phys(this_cpu_ptr(&kvm_apic_eoi)) kvm_guest_cpu_init() 358 wrmsrl(MSR_KVM_PV_EOI_EN, pa); kvm_guest_cpu_init()
|
/linux-4.1.27/fs/ |
H A D | posix_acl.c | 176 const struct posix_acl_entry *pa, *pe; posix_acl_valid() local 180 FOREACH_ACL_ENTRY(pa, acl, pe) { FOREACH_ACL_ENTRY() 181 if (pa->e_perm & ~(ACL_READ|ACL_WRITE|ACL_EXECUTE)) FOREACH_ACL_ENTRY() 183 switch (pa->e_tag) { FOREACH_ACL_ENTRY() 194 if (!uid_valid(pa->e_uid)) FOREACH_ACL_ENTRY() 209 if (!gid_valid(pa->e_gid)) FOREACH_ACL_ENTRY() 245 const struct posix_acl_entry *pa, *pe; posix_acl_equiv_mode() local 255 FOREACH_ACL_ENTRY(pa, acl, pe) { FOREACH_ACL_ENTRY() 256 switch (pa->e_tag) { FOREACH_ACL_ENTRY() 258 mode |= (pa->e_perm & S_IRWXO) << 6; FOREACH_ACL_ENTRY() 261 mode |= (pa->e_perm & S_IRWXO) << 3; FOREACH_ACL_ENTRY() 264 mode |= pa->e_perm & S_IRWXO; FOREACH_ACL_ENTRY() 268 ((pa->e_perm & S_IRWXO) << 3); FOREACH_ACL_ENTRY() 314 const struct posix_acl_entry *pa, *pe, *mask_obj; posix_acl_permission() local 319 FOREACH_ACL_ENTRY(pa, acl, pe) { FOREACH_ACL_ENTRY() 320 switch(pa->e_tag) { FOREACH_ACL_ENTRY() 327 if (uid_eq(pa->e_uid, current_fsuid())) FOREACH_ACL_ENTRY() 333 if ((pa->e_perm & want) == want) FOREACH_ACL_ENTRY() 338 if (in_group_p(pa->e_gid)) { FOREACH_ACL_ENTRY() 340 if ((pa->e_perm & want) == want) FOREACH_ACL_ENTRY() 358 for (mask_obj = pa+1; mask_obj != pe; mask_obj++) { 360 if ((pa->e_perm & mask_obj->e_perm & want) == want) 367 if ((pa->e_perm & want) == want) 382 struct posix_acl_entry *pa, *pe; posix_acl_create_masq() local 389 FOREACH_ACL_ENTRY(pa, acl, pe) { FOREACH_ACL_ENTRY() 390 switch(pa->e_tag) { FOREACH_ACL_ENTRY() 392 pa->e_perm &= (mode >> 6) | ~S_IRWXO; FOREACH_ACL_ENTRY() 393 mode &= (pa->e_perm << 6) | ~S_IRWXU; FOREACH_ACL_ENTRY() 402 group_obj = pa; FOREACH_ACL_ENTRY() 406 pa->e_perm &= mode | ~S_IRWXO; FOREACH_ACL_ENTRY() 407 mode &= pa->e_perm | ~S_IRWXO; FOREACH_ACL_ENTRY() 411 mask_obj = pa; FOREACH_ACL_ENTRY() 440 struct posix_acl_entry *pa, *pe; __posix_acl_chmod_masq() local 444 FOREACH_ACL_ENTRY(pa, acl, pe) { FOREACH_ACL_ENTRY() 445 switch(pa->e_tag) { FOREACH_ACL_ENTRY() 447 pa->e_perm = (mode & S_IRWXU) >> 6; FOREACH_ACL_ENTRY() 455 group_obj = pa; FOREACH_ACL_ENTRY() 459 mask_obj = pa; FOREACH_ACL_ENTRY() 463 pa->e_perm = (mode & S_IRWXO); FOREACH_ACL_ENTRY()
|
/linux-4.1.27/drivers/isdn/hardware/eicon/ |
H A D | diva.c | 169 diva_os_xdi_adapter_t *pdiva, *pa; diva_driver_add_card() local 207 pa = pdiva; diva_driver_add_card() 209 pa = diva_q_get_next(&pa->link); diva_driver_add_card() 210 if (pa && !pa->interface.cleanup_adapter_proc) { diva_driver_add_card() 211 pa->controller = i + 1 + j; diva_driver_add_card() 212 pa->xdi_adapter.ANum = pa->controller; diva_driver_add_card() 213 IoAdapters[i + j] = &pa->xdi_adapter; diva_driver_add_card() 216 pa->controller)) diva_driver_add_card() 217 create_adapter_proc(pa); /* add adapter to proc file system */ diva_driver_add_card() 283 diva_os_xdi_adapter_t *pa; diva_driver_remove_card() local 286 pa = a[0] = (diva_os_xdi_adapter_t *) pdiva; diva_driver_remove_card() 292 if ((pa = diva_q_get_next(&pa->link)) diva_driver_remove_card() 293 && !pa->interface.cleanup_adapter_proc) { diva_driver_remove_card() 294 a[i] = pa; diva_driver_remove_card()
|
/linux-4.1.27/fs/nfsd/ |
H A D | nfs4acl.c | 201 struct posix_acl_entry *pa, *pe; summarize_posix_acl() local 213 FOREACH_ACL_ENTRY(pa, acl, pe) { FOREACH_ACL_ENTRY() 214 switch (pa->e_tag) { FOREACH_ACL_ENTRY() 216 pas->owner = pa->e_perm; FOREACH_ACL_ENTRY() 219 pas->group = pa->e_perm; FOREACH_ACL_ENTRY() 222 pas->users |= pa->e_perm; FOREACH_ACL_ENTRY() 225 pas->groups |= pa->e_perm; FOREACH_ACL_ENTRY() 228 pas->other = pa->e_perm; FOREACH_ACL_ENTRY() 231 pas->mask = pa->e_perm; FOREACH_ACL_ENTRY() 246 struct posix_acl_entry *pa, *group_owner_entry; _posix_to_nfsv4_one() local 256 pa = pacl->a_entries; _posix_to_nfsv4_one() 277 ace->access_mask = mask_from_posix(pa->e_perm, flags | NFS4_ACL_OWNER); _posix_to_nfsv4_one() 281 pa++; _posix_to_nfsv4_one() 283 while (pa->e_tag == ACL_USER) { _posix_to_nfsv4_one() 284 deny = ~(pa->e_perm & pas.mask); _posix_to_nfsv4_one() 291 ace->who_uid = pa->e_uid; _posix_to_nfsv4_one() 297 ace->access_mask = mask_from_posix(pa->e_perm & pas.mask, _posix_to_nfsv4_one() 300 ace->who_uid = pa->e_uid; _posix_to_nfsv4_one() 303 pa++; _posix_to_nfsv4_one() 311 group_owner_entry = pa; _posix_to_nfsv4_one() 319 pa++; _posix_to_nfsv4_one() 321 while (pa->e_tag == ACL_GROUP) { _posix_to_nfsv4_one() 324 ace->access_mask = mask_from_posix(pa->e_perm & pas.mask, _posix_to_nfsv4_one() 327 ace->who_gid = pa->e_gid; _posix_to_nfsv4_one() 330 pa++; _posix_to_nfsv4_one() 335 pa = group_owner_entry; _posix_to_nfsv4_one() 346 pa++; _posix_to_nfsv4_one() 348 while (pa->e_tag == ACL_GROUP) { _posix_to_nfsv4_one() 349 deny = ~(pa->e_perm & pas.mask); _posix_to_nfsv4_one() 356 ace->who_gid = pa->e_gid; _posix_to_nfsv4_one() 360 pa++; _posix_to_nfsv4_one() 363 if (pa->e_tag == ACL_MASK) _posix_to_nfsv4_one() 364 pa++; _posix_to_nfsv4_one() 367 ace->access_mask = mask_from_posix(pa->e_perm, flags); _posix_to_nfsv4_one()
|
/linux-4.1.27/drivers/net/fddi/skfp/ |
H A D | pmf.c | 39 static int smt_set_para(struct s_smc *smc, struct smt_para *pa, int index, 316 struct smt_para *pa ; smt_build_pmf_response() local 373 pa = (struct smt_para *) (req + 1) ; smt_build_pmf_response() 378 if (((u_short)len < pa->p_len + PARA_LEN) || (pa->p_len & 3)) { smt_build_pmf_response() 383 if (((range = (pa->p_type & 0xf000)) == 0x2000) || smt_build_pmf_response() 388 index = *((u_char *)pa + PARA_LEN + 3) ;/* index */ smt_build_pmf_response() 390 if (!set && (pa->p_len != 4)) { smt_build_pmf_response() 419 if (!set && (pa->p_len != 0)) { smt_build_pmf_response() 431 pt = smt_get_ptab(pa->p_type) ; smt_build_pmf_response() 446 else if (pa->p_type != SMT_P_AUTHOR && smt_build_pmf_response() 447 (!set || (pa->p_type != SMT_P1035))) { smt_build_pmf_response() 450 smt_add_para(smc,&pcon,pa->p_type, smt_build_pmf_response() 454 st = smt_set_para(smc,pa,index,local,1); smt_build_pmf_response() 458 smt_add_para(smc,&pcon,pa->p_type, smt_build_pmf_response() 467 smt_add_para(smc,&pcon,pa->p_type, smt_build_pmf_response() 475 len -= pa->p_len + PARA_LEN ; smt_build_pmf_response() 476 pa = (struct smt_para *) ((char *)pa + pa->p_len + PARA_LEN) ; smt_build_pmf_response() 501 struct smt_para *pa ; smt_authorize() local 523 pa = (struct smt_para *) sm_to_para(smc,sm,SMT_P_AUTHOR) ; smt_authorize() 524 if (!pa) smt_authorize() 526 if (pa->p_len != 8) smt_authorize() 528 if (memcmp((char *)(pa+1),(char *)smc->mib.fddiPRPMFPasswd,8)) smt_authorize() 536 struct smt_para *pa ; smt_check_set_count() local 539 pa = (struct smt_para *) sm_to_para(smc,sm,SMT_P1035) ; smt_check_set_count() 540 if (pa) { smt_check_set_count() 541 sc = (struct smt_p_setcount *) pa ; smt_check_set_count() 553 struct smt_para *pa ; smt_add_para() local 586 pa = (struct smt_para *) to ; /* type/length pointer */ smt_add_para() 1051 pa->p_type = para ; smt_add_para() 1052 pa->p_len = plen - len - PARA_LEN ; smt_add_para() 1075 static int smt_set_para(struct s_smc *smc, struct smt_para *pa, int index, smt_set_para() argument 1101 len = pa->p_len ; smt_set_para() 1102 from = (char *) (pa + 1 ) ; smt_set_para() 1105 switch (pa->p_type & 0xf000) { smt_set_para() 1138 switch (pa->p_type) { smt_set_para() 1158 pt = smt_get_ptab(pa->p_type) ; smt_set_para() 1160 return (pa->p_type & 0xff00) ? SMT_RDF_NOPARAM : smt_set_para() 1299 switch (pa->p_type) { smt_set_para() 1563 struct smt_para *pa ; dump_smt() local 1593 pa = (struct smt_para *) (sm + 1) ; dump_smt() 1597 printf("TYPE %x LEN %x VALUE\t",pa->p_type,pa->p_len) ; dump_smt() 1599 printf("TYPE %04x LEN %2x VALUE\t",pa->p_type,pa->p_len) ; dump_smt() 1601 n = pa->p_len ; dump_smt() 1611 dump_hex((char *)(pa+1),(int) n) ; dump_smt() 1616 c = (char *)(pa+1) ; dump_smt() 1640 plen = (pa->p_len + PARA_LEN + 3) & ~3 ; dump_smt() 1642 pa = (struct smt_para *)((char *)pa + plen) ; dump_smt()
|
H A D | smt.c | 1829 struct smt_para *pa ; smt_swap_para() local 1846 pa = (struct smt_para *) p ; smt_swap_para() 1847 plen = pa->p_len ; smt_swap_para() 1848 type = pa->p_type ; smt_swap_para() 1849 pa->p_type = smt_swap_short(pa->p_type) ; smt_swap_para() 1850 pa->p_len = smt_swap_short(pa->p_len) ; smt_swap_para() 1852 plen = pa->p_len ; smt_swap_para() 1853 type = pa->p_type ; smt_swap_para()
|
/linux-4.1.27/drivers/misc/mic/host/ |
H A D | mic_fops.c | 156 unsigned long *size, unsigned long *pa) mic_query_offset() 171 *pa = virt_to_phys(mdev->dp); mic_query_offset() 179 *pa = virt_to_phys(mvr->vring.va); mic_query_offset() 196 unsigned long pa, size = vma->vm_end - vma->vm_start, size_rem = size; mic_mmap() local 207 i = mic_query_offset(mvdev, offset, &size, &pa); mic_mmap() 211 pa >> PAGE_SHIFT, size, vma->vm_page_prot); mic_mmap() 215 "%s %d type %d size 0x%lx off 0x%lx pa 0x%lx vma 0x%lx\n", mic_mmap() 217 pa, vma->vm_start + offset); mic_mmap() 155 mic_query_offset(struct mic_vdev *mvdev, unsigned long offset, unsigned long *size, unsigned long *pa) mic_query_offset() argument
|
H A D | mic_smpt.c | 32 static inline u8 mic_sys_addr_to_smpt(struct mic_device *mdev, dma_addr_t pa) mic_sys_addr_to_smpt() argument 34 return (pa - mdev->smpt->info.base) >> mdev->smpt->info.page_shift; mic_sys_addr_to_smpt() 42 static inline u64 mic_smpt_offset(struct mic_device *mdev, dma_addr_t pa) mic_smpt_offset() argument 44 return pa & mic_system_page_mask(mdev); mic_smpt_offset() 47 static inline u64 mic_smpt_align_low(struct mic_device *mdev, dma_addr_t pa) mic_smpt_align_low() argument 49 return ALIGN(pa - mic_system_page_mask(mdev), mic_smpt_align_low() 53 static inline u64 mic_smpt_align_high(struct mic_device *mdev, dma_addr_t pa) mic_smpt_align_high() argument 55 return ALIGN(pa, mdev->smpt->info.page_size); mic_smpt_align_high() 72 mic_is_system_addr(struct mic_device *mdev, dma_addr_t pa) mic_is_system_addr() argument 74 return pa >= mdev->smpt->info.base && pa <= mic_max_system_addr(mdev); mic_is_system_addr()
|
H A D | mic_main.c | 336 mdev->mmio.pa = pci_resource_start(pdev, mdev->ops->mmio_bar); mic_probe() 345 mdev->aper.pa = pci_resource_start(pdev, mdev->ops->aper_bar); mic_probe() 347 mdev->aper.va = ioremap_wc(mdev->aper.pa, mdev->aper.len); mic_probe()
|
/linux-4.1.27/arch/powerpc/platforms/pseries/ |
H A D | cmm.c | 121 struct cmm_page_array *pa, *npa; cmm_alloc_pages() local 144 pa = cmm_page_list; cmm_alloc_pages() 145 if (!pa || pa->index >= CMM_NR_PAGES) { cmm_alloc_pages() 157 pa = cmm_page_list; cmm_alloc_pages() 159 if (!pa || pa->index >= CMM_NR_PAGES) { cmm_alloc_pages() 160 npa->next = pa; cmm_alloc_pages() 162 pa = npa; cmm_alloc_pages() 163 cmm_page_list = pa; cmm_alloc_pages() 175 pa->page[pa->index++] = addr; cmm_alloc_pages() 195 struct cmm_page_array *pa; cmm_free_pages() local 200 pa = cmm_page_list; cmm_free_pages() 202 if (!pa || pa->index <= 0) cmm_free_pages() 204 addr = pa->page[--pa->index]; cmm_free_pages() 206 if (pa->index == 0) { cmm_free_pages() 207 pa = pa->next; cmm_free_pages() 209 cmm_page_list = pa; cmm_free_pages() 474 struct cmm_page_array *pa; cmm_count_pages() local 480 pa = cmm_page_list; cmm_count_pages() 481 while (pa) { cmm_count_pages() 482 if ((unsigned long)pa >= start && (unsigned long)pa < end) cmm_count_pages() 484 for (idx = 0; idx < pa->index; idx++) cmm_count_pages() 485 if (pa->page[idx] >= start && pa->page[idx] < end) cmm_count_pages() 487 pa = pa->next; cmm_count_pages()
|
H A D | lpar.c | 127 unsigned long vpn, unsigned long pa, pSeries_lpar_hpte_insert() 138 "pa=%016lx, rflags=%lx, vflags=%lx, psize=%d)\n", pSeries_lpar_hpte_insert() 139 hpte_group, vpn, pa, rflags, vflags, psize); pSeries_lpar_hpte_insert() 142 hpte_r = hpte_encode_r(pa, psize, apsize) | rflags; pSeries_lpar_hpte_insert() 126 pSeries_lpar_hpte_insert(unsigned long hpte_group, unsigned long vpn, unsigned long pa, unsigned long rflags, unsigned long vflags, int psize, int apsize, int ssize) pSeries_lpar_hpte_insert() argument
|
/linux-4.1.27/arch/mips/kernel/ |
H A D | spram.c | 123 unsigned int firstpa = 0, lastpa = 0, pa = 0; probe_spram() local 147 if ((pa == firstpa && size == firstsize) || probe_spram() 148 (pa == lastpa && size == lastsize)) probe_spram() 163 pa = tag0 & SPRAM_TAG0_PA_MASK; probe_spram() 167 firstpa = pa; probe_spram() 171 lastpa = pa; probe_spram() 175 unsigned int *vp = (unsigned int *)(CKSEG1 | pa); probe_spram() 194 type, i, pa, size, enabled ? ",enabled" : ""); probe_spram()
|
/linux-4.1.27/arch/arm/mach-davinci/include/mach/ |
H A D | hardware.h | 31 #define IO_ADDRESS(pa) IOMEM(__IO_ADDRESS(pa))
|
/linux-4.1.27/include/linux/fsl/bestcomm/ |
H A D | sram.h | 47 static inline void *bcom_sram_pa2va(phys_addr_t pa) { bcom_sram_pa2va() argument 49 (unsigned long)(pa - bcom_sram->base_phys); bcom_sram_pa2va()
|
/linux-4.1.27/arch/alpha/include/asm/ |
H A D | mmzone.h | 18 #define alpha_pa_to_nid(pa) \ 20 ? alpha_mv.pa_to_nid(pa) \ 31 #define pa_to_nid(pa) alpha_pa_to_nid(pa)
|
H A D | hwrpb.h | 125 unsigned long pa; member in struct:vf_map_struct
|
/linux-4.1.27/drivers/dio/ |
H A D | dio.c | 123 unsigned long pa; dio_find() local 128 pa = dio_scodetophysaddr(scode); dio_find() 130 if (!pa) dio_find() 134 va = (void *)(pa + DIO_VIRADDRBASE); dio_find() 136 va = ioremap(pa, PAGE_SIZE); dio_find() 202 unsigned long pa; dio_init() local 207 pa = dio_scodetophysaddr(scode); dio_init() 209 if (!pa) dio_init() 213 va = (void *)(pa + DIO_VIRADDRBASE); dio_init() 215 va = ioremap(pa, PAGE_SIZE); dio_init() 238 dev->resource.start = pa; dio_init() 239 dev->resource.end = pa + DIO_SIZE(scode, va); dio_init()
|
/linux-4.1.27/arch/arm/mach-omap2/ |
H A D | iomap.h | 34 #define OMAP2_L3_IO_ADDRESS(pa) IOMEM((pa) + OMAP2_L3_IO_OFFSET) /* L3 */ 37 #define OMAP2_L4_IO_ADDRESS(pa) IOMEM((pa) + OMAP2_L4_IO_OFFSET) /* L4 */ 40 #define OMAP4_L3_IO_ADDRESS(pa) IOMEM((pa) + OMAP4_L3_IO_OFFSET) /* L3 */ 43 #define AM33XX_L4_WK_IO_ADDRESS(pa) IOMEM((pa) + AM33XX_L4_WK_IO_OFFSET) 46 #define OMAP4_L3_PER_IO_ADDRESS(pa) IOMEM((pa) + OMAP4_L3_PER_IO_OFFSET) 49 #define OMAP2_EMU_IO_ADDRESS(pa) IOMEM((pa) + OMAP2_EMU_IO_OFFSET)
|
/linux-4.1.27/arch/s390/pci/ |
H A D | pci_dma.c | 142 static int dma_update_trans(struct zpci_dev *zdev, unsigned long pa, dma_update_trans() argument 146 u8 *page_addr = (u8 *) (pa & PAGE_MASK); dma_update_trans() 282 unsigned long pa = page_to_phys(page) + offset; s390_dma_map_pages() local 287 nr_pages = iommu_num_pages(pa, size, PAGE_SIZE); s390_dma_map_pages() 302 if (!dma_update_trans(zdev, pa, dma_addr, size, flags)) { s390_dma_map_pages() 311 zpci_err_hex(&pa, sizeof(pa)); s390_dma_map_pages() 342 unsigned long pa; s390_dma_alloc() local 350 pa = page_to_phys(page); s390_dma_alloc() 351 memset((void *) pa, 0, size); s390_dma_alloc() 353 map = s390_dma_map_pages(dev, page, pa % PAGE_SIZE, s390_dma_alloc() 356 free_pages(pa, get_order(size)); s390_dma_alloc() 363 return (void *) pa; s390_dma_alloc() 367 void *pa, dma_addr_t dma_handle, s390_dma_free() 375 free_pages((unsigned long) pa, get_order(size)); s390_dma_free() 366 s390_dma_free(struct device *dev, size_t size, void *pa, dma_addr_t dma_handle, struct dma_attrs *attrs) s390_dma_free() argument
|
/linux-4.1.27/include/linux/ |
H A D | posix_acl.h | 54 #define FOREACH_ACL_ENTRY(pa, acl, pe) \ 55 for(pa=(acl)->a_entries, pe=pa+(acl)->a_count; pa<pe; pa++)
|
H A D | if_pppox.h | 33 struct pppoe_addr pa; /* what this socket is bound to*/ member in struct:pppoe_opt 61 #define pppoe_pa proto.pppoe.pa
|
H A D | remoteproc.h | 126 * @pa: physical address 157 * overwrite @pa with the dynamically allocated physical address. 168 u32 pa; member in struct:fw_rsc_carveout 178 * @pa: physical address 192 * @da should specify the required device address, @pa should specify 206 u32 pa; member in struct:fw_rsc_devmem
|
/linux-4.1.27/arch/cris/arch-v32/mach-fs/ |
H A D | pinmux.c | 56 reg_pinmux_rw_pa pa = REG_RD(pinmux, regi_pinmux, rw_pa); crisv32_pinmux_init() local 59 pa.pa0 = pa.pa1 = pa.pa2 = pa.pa3 = crisv32_pinmux_init() 60 pa.pa4 = pa.pa5 = pa.pa6 = pa.pa7 = regk_pinmux_yes; crisv32_pinmux_init() 61 REG_WR(pinmux, regi_pinmux, rw_pa, pa); crisv32_pinmux_init()
|
/linux-4.1.27/drivers/char/tpm/ |
H A D | tpm_crb.c | 214 u64 pa; crb_acpi_add() local 266 memcpy_fromio(&pa, &priv->cca->cmd_pa, 8); crb_acpi_add() 267 pa = le64_to_cpu(pa); crb_acpi_add() 268 priv->cmd = devm_ioremap_nocache(dev, pa, crb_acpi_add() 275 memcpy_fromio(&pa, &priv->cca->rsp_pa, 8); crb_acpi_add() 276 pa = le64_to_cpu(pa); crb_acpi_add() 277 priv->rsp = devm_ioremap_nocache(dev, pa, crb_acpi_add()
|
/linux-4.1.27/arch/arm/xen/ |
H A D | p2m.c | 56 pr_warn("%s: cannot add pfn=%pa -> mfn=%pa: pfn=%pa -> mfn=%pa already exists\n", xen_add_phys_to_mach_entry()
|
H A D | enlighten.c | 182 pr_info("Xen %s support found, events_irq=%d gnttab_frame=%pa\n", xen_guest_init()
|
/linux-4.1.27/arch/m68k/include/asm/ |
H A D | sun3xprom.h | 14 unsigned long sun3x_prom_ptov(unsigned long pa, unsigned long size);
|
/linux-4.1.27/fs/f2fs/ |
H A D | acl.c | 280 struct posix_acl_entry *pa, *pe; f2fs_acl_create_masq() local 287 FOREACH_ACL_ENTRY(pa, acl, pe) { FOREACH_ACL_ENTRY() 288 switch(pa->e_tag) { FOREACH_ACL_ENTRY() 290 pa->e_perm &= (mode >> 6) | ~S_IRWXO; FOREACH_ACL_ENTRY() 291 mode &= (pa->e_perm << 6) | ~S_IRWXU; FOREACH_ACL_ENTRY() 300 group_obj = pa; FOREACH_ACL_ENTRY() 304 pa->e_perm &= mode | ~S_IRWXO; FOREACH_ACL_ENTRY() 305 mode &= pa->e_perm | ~S_IRWXO; FOREACH_ACL_ENTRY() 309 mask_obj = pa; FOREACH_ACL_ENTRY()
|
/linux-4.1.27/drivers/net/ethernet/amd/xgbe/ |
H A D | xgbe-desc.c | 259 struct xgbe_page_alloc *pa, gfp_t gfp, int order) xgbe_alloc_pages() 286 pa->pages = pages; xgbe_alloc_pages() 287 pa->pages_len = PAGE_SIZE << order; xgbe_alloc_pages() 288 pa->pages_offset = 0; xgbe_alloc_pages() 289 pa->pages_dma = pages_dma; xgbe_alloc_pages() 295 struct xgbe_page_alloc *pa, xgbe_set_buffer_data() 298 get_page(pa->pages); xgbe_set_buffer_data() 299 bd->pa = *pa; xgbe_set_buffer_data() 301 bd->dma = pa->pages_dma + pa->pages_offset; xgbe_set_buffer_data() 304 pa->pages_offset += len; xgbe_set_buffer_data() 305 if ((pa->pages_offset + len) > pa->pages_len) { xgbe_set_buffer_data() 307 bd->pa_unmap = *pa; xgbe_set_buffer_data() 310 pa->pages = NULL; xgbe_set_buffer_data() 311 pa->pages_len = 0; xgbe_set_buffer_data() 312 pa->pages_offset = 0; xgbe_set_buffer_data() 313 pa->pages_dma = 0; xgbe_set_buffer_data() 452 if (rdata->rx.hdr.pa.pages) xgbe_unmap_rdata() 453 put_page(rdata->rx.hdr.pa.pages); xgbe_unmap_rdata() 462 if (rdata->rx.buf.pa.pages) xgbe_unmap_rdata() 463 put_page(rdata->rx.buf.pa.pages); xgbe_unmap_rdata() 258 xgbe_alloc_pages(struct xgbe_prv_data *pdata, struct xgbe_page_alloc *pa, gfp_t gfp, int order) xgbe_alloc_pages() argument 294 xgbe_set_buffer_data(struct xgbe_buffer_data *bd, struct xgbe_page_alloc *pa, unsigned int len) xgbe_set_buffer_data() argument
|
/linux-4.1.27/lib/ |
H A D | idr.c | 172 static void idr_mark_full(struct idr_layer **pa, int id) idr_mark_full() argument 174 struct idr_layer *p = pa[0]; idr_mark_full() 185 if (!(p = pa[++l])) idr_mark_full() 208 * @pa: idr_layer[MAX_IDR_LEVEL] used as backtrack buffer 220 static int sub_alloc(struct idr *idp, int *starting_id, struct idr_layer **pa, sub_alloc() argument 231 pa[l--] = NULL; sub_alloc() 249 p = pa[l]; sub_alloc() 281 pa[l--] = p; sub_alloc() 285 pa[l] = p; sub_alloc() 290 struct idr_layer **pa, gfp_t gfp_mask, idr_get_empty_slot() 348 v = sub_alloc(idp, &id, pa, gfp_mask, layer_idr); idr_get_empty_slot() 355 * @id and @pa are from a successful allocation from idr_get_empty_slot(). 359 struct idr_layer **pa) idr_fill_slot() 362 rcu_assign_pointer(idr->hint, pa[0]); idr_fill_slot() 364 rcu_assign_pointer(pa[0]->ary[id & IDR_MASK], (struct idr_layer *)ptr); idr_fill_slot() 365 pa[0]->count++; idr_fill_slot() 366 idr_mark_full(pa, id); idr_fill_slot() 453 struct idr_layer *pa[MAX_IDR_LEVEL + 1]; idr_alloc() local 465 id = idr_get_empty_slot(idr, start, pa, gfp_mask, NULL); idr_alloc() 471 idr_fill_slot(idr, ptr, id, pa); idr_alloc() 511 struct idr_layer **pa[MAX_IDR_LEVEL + 1]; sub_remove() local 512 struct idr_layer ***paa = &pa[0]; sub_remove() 588 struct idr_layer *pa[MAX_IDR_LEVEL + 1]; __idr_remove_all() local 589 struct idr_layer **paa = &pa[0]; __idr_remove_all() 691 struct idr_layer *pa[MAX_IDR_LEVEL + 1]; idr_for_each() local 692 struct idr_layer **paa = &pa[0]; idr_for_each() 738 struct idr_layer *p, *pa[MAX_IDR_LEVEL + 1]; idr_get_next() local 739 struct idr_layer **paa = &pa[0]; idr_get_next() 934 struct idr_layer *pa[MAX_IDR_LEVEL + 1]; ida_get_new_above() local 943 t = idr_get_empty_slot(&ida->idr, idr_id, pa, 0, &ida->idr); ida_get_new_above() 955 bitmap = (void *)pa[0]->ary[idr_id & IDR_MASK]; ida_get_new_above() 966 rcu_assign_pointer(pa[0]->ary[idr_id & IDR_MASK], ida_get_new_above() 968 pa[0]->count++; ida_get_new_above() 986 idr_mark_full(pa, idr_id); ida_get_new_above() 289 idr_get_empty_slot(struct idr *idp, int starting_id, struct idr_layer **pa, gfp_t gfp_mask, struct idr *layer_idr) idr_get_empty_slot() argument 358 idr_fill_slot(struct idr *idr, void *ptr, int id, struct idr_layer **pa) idr_fill_slot() argument
|
/linux-4.1.27/drivers/misc/mic/common/ |
H A D | mic_dev.h | 27 * @pa: Base physical address. 32 phys_addr_t pa; member in struct:mic_mw
|
/linux-4.1.27/arch/sparc/include/asm/ |
H A D | pgalloc_32.h | 32 unsigned long pa = __nocache_pa((unsigned long)pmdp); pgd_set() local 34 set_pte((pte_t *)pgdp, (SRMMU_ET_PTD | (pa >> 4))); pgd_set()
|
/linux-4.1.27/arch/arm/mach-omap1/include/mach/ |
H A D | hardware.h | 47 extern u8 omap_readb(u32 pa); 48 extern u16 omap_readw(u32 pa); 49 extern u32 omap_readl(u32 pa); 50 extern void omap_writeb(u8 v, u32 pa); 51 extern void omap_writew(u16 v, u32 pa); 52 extern void omap_writel(u32 v, u32 pa); 76 #define OMAP1_IO_ADDRESS(pa) IOMEM((pa) - OMAP1_IO_OFFSET)
|
/linux-4.1.27/arch/frv/mm/ |
H A D | dma-alloc.c | 50 static int map_page(unsigned long va, unsigned long pa, pgprot_t prot) map_page() argument 67 set_pte(pte, mk_pte_phys(pa & PAGE_MASK, prot)); map_page() 84 unsigned long page, va, pa; consistent_alloc() local 111 *dma_handle = pa = virt_to_bus((void *) page); consistent_alloc() 123 err = map_page(va + i, pa + i, PAGE_KERNEL_NOCACHE); consistent_alloc()
|
/linux-4.1.27/arch/m68k/coldfire/ |
H A D | intc-2.c | 148 u16 pa, tb; intc_irq_set_type() local 170 pa = __raw_readw(MCFEPORT_EPPAR); intc_irq_set_type() 171 pa = (pa & ~(0x3 << (irq * 2))) | (tb << (irq * 2)); intc_irq_set_type() 172 __raw_writew(pa, MCFEPORT_EPPAR); intc_irq_set_type()
|
H A D | intc-simr.c | 132 u16 pa, tb; intc_irq_set_type() local 154 pa = __raw_readw(MCFEPORT_EPPAR); intc_irq_set_type() 155 pa = (pa & ~(0x3 << ebit)) | (tb << ebit); intc_irq_set_type() 156 __raw_writew(pa, MCFEPORT_EPPAR); intc_irq_set_type()
|
/linux-4.1.27/drivers/iommu/ |
H A D | tegra-gart.c | 266 phys_addr_t pa, size_t bytes, int prot) gart_iommu_map() 277 pfn = __phys_to_pfn(pa); gart_iommu_map() 279 dev_err(gart->dev, "Invalid page: %pa\n", &pa); gart_iommu_map() 312 phys_addr_t pa; gart_iommu_iova_to_phys() local 322 pa = (pte & GART_PAGE_MASK); gart_iommu_iova_to_phys() 323 if (!pfn_valid(__phys_to_pfn(pa))) { gart_iommu_iova_to_phys() 324 dev_err(gart->dev, "No entry for %08llx:%pa\n", gart_iommu_iova_to_phys() 325 (unsigned long long)iova, &pa); gart_iommu_iova_to_phys() 329 return pa; gart_iommu_iova_to_phys() 265 gart_iommu_map(struct iommu_domain *domain, unsigned long iova, phys_addr_t pa, size_t bytes, int prot) gart_iommu_map() argument
|
H A D | omap-iommu.c | 146 u32 l, pa; omap2_iommu_enable() local 151 pa = virt_to_phys(obj->iopgd); omap2_iommu_enable() 152 if (!IS_ALIGNED(pa, SZ_16K)) omap2_iommu_enable() 159 iommu_write_reg(obj, pa, MMU_TTB); omap2_iommu_enable() 333 cr->ram = e->pa | e->endian | e->elsz | e->mixed; iotlb_alloc_cr() 662 static int iopgd_alloc_section(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) iopgd_alloc_section() argument 666 if ((da | pa) & ~IOSECTION_MASK) { iopgd_alloc_section() 668 __func__, da, pa, IOSECTION_SIZE); iopgd_alloc_section() 672 *iopgd = (pa & IOSECTION_MASK) | prot | IOPGD_SECTION; iopgd_alloc_section() 677 static int iopgd_alloc_super(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) iopgd_alloc_super() argument 682 if ((da | pa) & ~IOSUPER_MASK) { iopgd_alloc_super() 684 __func__, da, pa, IOSUPER_SIZE); iopgd_alloc_super() 689 *(iopgd + i) = (pa & IOSUPER_MASK) | prot | IOPGD_SUPER; iopgd_alloc_super() 694 static int iopte_alloc_page(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) iopte_alloc_page() argument 702 *iopte = (pa & IOPAGE_MASK) | prot | IOPTE_SMALL; iopte_alloc_page() 705 dev_vdbg(obj->dev, "%s: da:%08x pa:%08x pte:%p *pte:%08x\n", iopte_alloc_page() 706 __func__, da, pa, iopte, *iopte); iopte_alloc_page() 711 static int iopte_alloc_large(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) iopte_alloc_large() argument 717 if ((da | pa) & ~IOLARGE_MASK) { iopte_alloc_large() 719 __func__, da, pa, IOLARGE_SIZE); iopte_alloc_large() 727 *(iopte + i) = (pa & IOLARGE_MASK) | prot | IOPTE_LARGE; iopte_alloc_large() 764 err = fn(obj, e->da, e->pa, prot); iopgtable_store_entry_core() 1108 static u32 iotlb_init_entry(struct iotlb_entry *e, u32 da, u32 pa, int pgsz) iotlb_init_entry() argument 1113 e->pa = pa; iotlb_init_entry() 1124 phys_addr_t pa, size_t bytes, int prot) omap_iommu_map() 1139 dev_dbg(dev, "mapping da 0x%lx to pa %pa size 0x%x\n", da, &pa, bytes); omap_iommu_map() 1141 iotlb_init_entry(&e, da, pa, omap_pgsz); omap_iommu_map() 1123 omap_iommu_map(struct iommu_domain *domain, unsigned long da, phys_addr_t pa, size_t bytes, int prot) omap_iommu_map() argument
|
H A D | omap-iopgtable.h | 57 * omap_iommu_translate() - va to pa translation 62 * va to pa translation
|
H A D | exynos-iommu.c | 91 #define mk_lv1ent_sect(pa) ((pa) | 2) 92 #define mk_lv1ent_page(pa) ((pa) | 1) 93 #define mk_lv2ent_lpage(pa) ((pa) | 1) 94 #define mk_lv2ent_spage(pa) ((pa) | 2) 302 pr_err("%s occurred at %#x by %s(Page table base: %pa)\n", show_fault_information() 807 dev_err(dev, "%s: Failed to attach IOMMU with pgtable %pa\n", exynos_iommu_attach_device() 812 dev_dbg(dev, "%s: Attached IOMMU with pgtable %pa %s\n", exynos_iommu_attach_device() 841 dev_dbg(dev, "%s: Detached IOMMU with pgtable %pa\n", exynos_iommu_detach_device()
|
H A D | msm_iommu.c | 368 phys_addr_t pa, size_t len, int prot) msm_iommu_map() 426 *(fl_pte+i) = (pa & 0xFF000000) | FL_SUPERSECTION | msm_iommu_map() 432 *fl_pte = (pa & 0xFFF00000) | FL_AP_READ | FL_AP_WRITE | FL_NG | msm_iommu_map() 457 *sl_pte = (pa & SL_BASE_MASK_SMALL) | SL_AP0 | SL_AP1 | SL_NG | msm_iommu_map() 464 *(sl_pte+i) = (pa & SL_BASE_MASK_LARGE) | SL_AP0 | msm_iommu_map() 367 msm_iommu_map(struct iommu_domain *domain, unsigned long va, phys_addr_t pa, size_t len, int prot) msm_iommu_map() argument
|
H A D | omap-iommu.h | 18 u32 pa; member in struct:iotlb_entry
|
H A D | iommu.c | 1052 pr_err("unaligned: iova 0x%lx pa %pa size 0x%zx min_pagesz 0x%x\n", iommu_map() 1057 pr_debug("map: iova 0x%lx pa %pa size 0x%zx\n", iova, &paddr, size); iommu_map() 1062 pr_debug("mapping: iova 0x%lx pa %pa pgsize 0x%zx\n", iommu_map()
|
H A D | rockchip-iommu.c | 451 dev_err(iommu->dev, "mmu_dte_addr: %pa dte@%pa: %#08x valid: %u pte@%pa: %#08x valid: %u page@%pa flags: %#03x\n", log_iova() 640 pr_err("iova: %pad already mapped to %pa cannot remap to phys: %pa prot: %#x\n", rk_iommu_map_iova()
|
/linux-4.1.27/drivers/sfi/ |
H A D | sfi_core.c | 120 static void sfi_print_table_header(unsigned long long pa, sfi_print_table_header() argument 124 header->sig, pa, sfi_print_table_header() 164 static struct sfi_table_header *sfi_map_table(u64 pa) sfi_map_table() argument 169 if (!TABLE_ON_PAGE(syst_pa, pa, sizeof(struct sfi_table_header))) sfi_map_table() 170 th = sfi_map_memory(pa, sizeof(struct sfi_table_header)); sfi_map_table() 172 th = (void *)syst_va + (pa - syst_pa); sfi_map_table() 180 if (!TABLE_ON_PAGE(syst_pa, pa, sizeof(struct sfi_table_header))) sfi_map_table() 183 return sfi_map_memory(pa, length); sfi_map_table() 231 __ref sfi_check_table(u64 pa, struct sfi_table_key *key) sfi_check_table() argument 236 th = sfi_map_table(pa); sfi_check_table() 241 sfi_print_table_header(pa, th); sfi_check_table() 418 struct sfi_table_attr __init *sfi_sysfs_install_table(u64 pa) sfi_sysfs_install_table() argument 428 th = sfi_map_table(pa); sfi_sysfs_install_table()
|
H A D | sfi_core.h | 80 extern struct sfi_table_attr __init *sfi_sysfs_install_table(u64 pa);
|
/linux-4.1.27/drivers/gpu/drm/msm/ |
H A D | msm_iommu.c | 60 u32 pa = sg_phys(sg) - sg->offset; msm_iommu_map() local 63 VERB("map[%d]: %08x %08x(%zx)", i, iova, pa, bytes); msm_iommu_map() 65 ret = iommu_map(domain, da, pa, bytes, prot); msm_iommu_map()
|
/linux-4.1.27/arch/powerpc/boot/ |
H A D | wii.c | 44 static int mipc_check_address(u32 pa) mipc_check_address() argument 47 if (pa < 0x10000000 || pa > 0x14000000) mipc_check_address()
|
/linux-4.1.27/drivers/net/wireless/ath/wil6210/ |
H A D | txrx.c | 117 vring->va = dma_alloc_coherent(dev, sz, &vring->pa, GFP_KERNEL); wil_vring_alloc() 134 vring->va, &vring->pa, vring->ctx); wil_vring_alloc() 142 dma_addr_t pa = wil_desc_addr(&d->dma.addr); wil_txdesc_unmap() local 147 dma_unmap_single(dev, pa, dmalen, DMA_TO_DEVICE); wil_txdesc_unmap() 150 dma_unmap_page(dev, pa, dmalen, DMA_TO_DEVICE); wil_txdesc_unmap() 168 &vring->pa, vring->ctx); wil_vring_free() 172 &vring->pa, vring->ctx); wil_vring_free() 176 dma_addr_t pa; wil_vring_free() local 198 pa = wil_desc_addr(&d->dma.addr); wil_vring_free() 200 dma_unmap_single(dev, pa, dmalen, DMA_FROM_DEVICE); wil_vring_free() 205 dma_free_coherent(dev, sz, (void *)vring->va, vring->pa); wil_vring_free() 207 vring->pa = 0; wil_vring_free() 224 dma_addr_t pa; wil_vring_alloc_skb() local 233 pa = dma_map_single(dev, skb->data, skb->len, DMA_FROM_DEVICE); wil_vring_alloc_skb() 234 if (unlikely(dma_mapping_error(dev, pa))) { wil_vring_alloc_skb() 240 wil_desc_addr_set(&d->dma.addr, pa); wil_vring_alloc_skb() 308 void *pa = PTR_ALIGN(p, 8); wil_rx_add_radiotap_header() local 310 if (skb_tailroom(skb) >= len + (pa - p)) { wil_rx_add_radiotap_header() 312 memcpy(phy_data, pa, phy_length); wil_rx_add_radiotap_header() 376 dma_addr_t pa; wil_vring_reap_rx() local 405 pa = wil_desc_addr(&d->dma.addr); wil_vring_reap_rx() 407 dma_unmap_single(dev, pa, sz, DMA_FROM_DEVICE); wil_vring_reap_rx() 725 cmd.vring_cfg.tx_sw_ring.ring_mem_base = cpu_to_le64(vring->pa); wil_vring_init_tx() 793 cmd.vring_cfg.tx_sw_ring.ring_mem_base = cpu_to_le64(vring->pa); wil_vring_init_bcast() 1025 static int wil_tx_desc_map(struct vring_tx_desc *d, dma_addr_t pa, u32 len, wil_tx_desc_map() argument 1028 wil_desc_addr_set(&d->dma.addr, pa); wil_tx_desc_map() 1115 dma_addr_t pa; __wil_tx_vring() local 1133 pa = dma_map_single(dev, skb->data, skb_headlen(skb), DMA_TO_DEVICE); __wil_tx_vring() 1136 skb_headlen(skb), skb->data, &pa); __wil_tx_vring() 1140 if (unlikely(dma_mapping_error(dev, pa))) __wil_tx_vring() 1144 wil_tx_desc_map(d, pa, len, vring_index); __wil_tx_vring() 1177 pa = skb_frag_dma_map(dev, frag, 0, skb_frag_size(frag), __wil_tx_vring() 1179 if (unlikely(dma_mapping_error(dev, pa))) __wil_tx_vring() 1182 wil_tx_desc_map(d, pa, len, vring_index); __wil_tx_vring()
|
H A D | txrx.h | 44 dma_addr_t pa) wil_desc_addr_set() 46 addr->addr_low = cpu_to_le32(lower_32_bits(pa)); wil_desc_addr_set() 47 addr->addr_high = cpu_to_le16((u16)upper_32_bits(pa)); wil_desc_addr_set() 43 wil_desc_addr_set(struct vring_dma_addr *addr, dma_addr_t pa) wil_desc_addr_set() argument
|
/linux-4.1.27/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_hw.c | 429 ocrdma_build_q_pages(&cmd->pa[0], cmd->num_pages, eq->q.dma, ocrdma_mbx_create_eq() 525 ocrdma_build_q_pages(&cmd->pa[0], cq->size / OCRDMA_MIN_Q_PAGE_SIZE, ocrdma_mbx_mq_cq_create() 552 struct ocrdma_pa *pa; ocrdma_mbx_create_mq() local 571 pa = &cmd->pa[0]; ocrdma_mbx_create_mq() 573 ocrdma_build_q_pages(pa, num_pages, mq->dma, PAGE_SIZE_4K); ocrdma_mbx_create_mq() 1234 mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dev->stats_mem.pa & 0xffffffff); ocrdma_mbx_rdma_stats() 1235 mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dev->stats_mem.pa); ocrdma_mbx_rdma_stats() 1274 dma.size, &dma.pa, GFP_KERNEL); ocrdma_mbx_get_ctrl_attribs() 1282 mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dma.pa & 0xffffffff); ocrdma_mbx_get_ctrl_attribs() 1283 mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dma.pa); ocrdma_mbx_get_ctrl_attribs() 1303 dma_free_coherent(&dev->nic_info.pdev->dev, dma.size, dma.va, dma.pa); ocrdma_mbx_get_ctrl_attribs() 1581 dma_addr_t pa; ocrdma_mbx_create_ah_tbl() local 1610 &dev->av_tbl.pbl.pa, ocrdma_mbx_create_ah_tbl() 1616 &pa, GFP_KERNEL); ocrdma_mbx_create_ah_tbl() 1619 dev->av_tbl.pa = pa; ocrdma_mbx_create_ah_tbl() 1625 pbes[i].pa_lo = (u32)cpu_to_le32(pa & 0xffffffff); ocrdma_mbx_create_ah_tbl() 1626 pbes[i].pa_hi = (u32)cpu_to_le32(upper_32_bits(pa)); ocrdma_mbx_create_ah_tbl() 1627 pa += PAGE_SIZE; ocrdma_mbx_create_ah_tbl() 1629 cmd->tbl_addr[0].lo = (u32)(dev->av_tbl.pbl.pa & 0xFFFFFFFF); ocrdma_mbx_create_ah_tbl() 1630 cmd->tbl_addr[0].hi = (u32)upper_32_bits(dev->av_tbl.pbl.pa); ocrdma_mbx_create_ah_tbl() 1641 dev->av_tbl.pa); ocrdma_mbx_create_ah_tbl() 1645 dev->av_tbl.pbl.pa); ocrdma_mbx_create_ah_tbl() 1668 dev->av_tbl.pa); ocrdma_mbx_delete_ah_tbl() 1671 dev->av_tbl.pbl.pa); ocrdma_mbx_delete_ah_tbl() 1750 cq->va = dma_alloc_coherent(&pdev->dev, cq->len, &cq->pa, GFP_KERNEL); ocrdma_mbx_create_cq() 1803 ocrdma_build_q_pages(&cmd->cmd.pa[0], hw_pages, cq->pa, page_size); ocrdma_mbx_create_cq() 1814 dma_free_coherent(&pdev->dev, cq->len, cq->va, cq->pa); ocrdma_mbx_create_cq() 1837 dma_free_coherent(&dev->nic_info.pdev->dev, cq->len, cq->va, cq->pa); ocrdma_mbx_destroy_cq() 1932 cmd->pbl[i].lo = (u32) (hwmr->pbl_table[i].pa & 0xffffffff); ocrdma_mbx_reg_mr() 1933 cmd->pbl[i].hi = upper_32_bits(hwmr->pbl_table[i].pa); ocrdma_mbx_reg_mr() 1963 (u32) (hwmr->pbl_table[i + pbl_offset].pa & 0xffffffff); ocrdma_mbx_reg_mr_cont() 1965 upper_32_bits(hwmr->pbl_table[i + pbl_offset].pa); ocrdma_mbx_reg_mr_cont() 2125 dma_addr_t pa; ocrdma_set_create_qp_sq_cmd() local 2146 qp->sq.va = dma_alloc_coherent(&pdev->dev, len, &pa, GFP_KERNEL); ocrdma_set_create_qp_sq_cmd() 2151 qp->sq.pa = pa; ocrdma_set_create_qp_sq_cmd() 2153 ocrdma_build_q_pages(&cmd->wq_addr[0], hw_pages, pa, hw_page_size); ocrdma_set_create_qp_sq_cmd() 2181 dma_addr_t pa = 0; ocrdma_set_create_qp_rq_cmd() local 2197 qp->rq.va = dma_alloc_coherent(&pdev->dev, len, &pa, GFP_KERNEL); ocrdma_set_create_qp_rq_cmd() 2201 qp->rq.pa = pa; ocrdma_set_create_qp_rq_cmd() 2205 ocrdma_build_q_pages(&cmd->rq_addr[0], hw_pages, pa, hw_page_size); ocrdma_set_create_qp_rq_cmd() 2245 dma_addr_t pa = 0; ocrdma_set_create_qp_ird_cmd() local 2255 &pa, GFP_KERNEL); ocrdma_set_create_qp_ird_cmd() 2260 pa, ird_page_size); ocrdma_set_create_qp_ird_cmd() 2397 dma_free_coherent(&pdev->dev, qp->rq.len, qp->rq.va, qp->rq.pa); ocrdma_mbx_create_qp() 2400 dma_free_coherent(&pdev->dev, qp->sq.len, qp->sq.va, qp->sq.pa); ocrdma_mbx_create_qp() 2657 dma_free_coherent(&pdev->dev, qp->sq.len, qp->sq.va, qp->sq.pa); ocrdma_mbx_destroy_qp() 2659 dma_free_coherent(&pdev->dev, qp->rq.len, qp->rq.va, qp->rq.pa); ocrdma_mbx_destroy_qp() 2674 dma_addr_t pa; ocrdma_mbx_create_srq() local 2694 srq->rq.va = dma_alloc_coherent(&pdev->dev, len, &pa, GFP_KERNEL); ocrdma_mbx_create_srq() 2699 ocrdma_build_q_pages(&cmd->rq_addr[0], hw_pages, pa, hw_page_size); ocrdma_mbx_create_srq() 2702 srq->rq.pa = pa; ocrdma_mbx_create_srq() 2734 dma_free_coherent(&pdev->dev, srq->rq.len, srq->rq.va, pa); ocrdma_mbx_create_srq() 2796 srq->rq.va, srq->rq.pa); ocrdma_mbx_destroy_srq() 2805 dma_addr_t pa; ocrdma_mbx_get_dcbx_config() local 2816 req = dma_alloc_coherent(&pdev->dev, cmd.hdr.pyld_len, &pa, GFP_KERNEL); ocrdma_mbx_get_dcbx_config() 2824 mqe_sge->pa_lo = (u32) (pa & 0xFFFFFFFFUL); ocrdma_mbx_get_dcbx_config() 2825 mqe_sge->pa_hi = (u32) upper_32_bits(pa); ocrdma_mbx_get_dcbx_config() 2842 dma_free_coherent(&pdev->dev, cmd.hdr.pyld_len, req, pa); ocrdma_mbx_get_dcbx_config()
|
H A D | ocrdma.h | 106 dma_addr_t pa; member in struct:ocrdma_dma_mem 112 dma_addr_t pa; member in struct:ocrdma_pbl 206 dma_addr_t pa; member in struct:stats_mem 246 dma_addr_t pa; member in struct:ocrdma_dev::__anon4960 318 dma_addr_t pa; member in struct:ocrdma_cq 353 dma_addr_t pa; member in struct:ocrdma_qp_hwq_info 425 dma_addr_t pa; member in struct:ocrdma_ucontext::__anon4962
|
H A D | ocrdma_verbs.c | 490 &ctx->ah_tbl.pa, GFP_KERNEL); ocrdma_alloc_ucontext() 524 ocrdma_del_mmap(ctx, ctx->ah_tbl.pa, ctx->ah_tbl.len); ocrdma_alloc_ucontext() 527 ctx->ah_tbl.pa); ocrdma_alloc_ucontext() 542 ocrdma_del_mmap(uctx, uctx->ah_tbl.pa, uctx->ah_tbl.len); ocrdma_dealloc_ucontext() 544 uctx->ah_tbl.pa); ocrdma_dealloc_ucontext() 775 mr->pbl_table[i].pa); ocrdma_free_mr_pbl_tbl() 814 dma_addr_t pa; ocrdma_build_pbl_tbl() local 823 va = dma_alloc_coherent(&pdev->dev, dma_len, &pa, GFP_KERNEL); ocrdma_build_pbl_tbl() 831 mr->pbl_table[i].pa = pa; ocrdma_build_pbl_tbl() 1114 ocrdma_del_mmap(cq->ucontext, (u64) cq->pa, ocrdma_destroy_cq() 1765 ocrdma_del_mmap(pd->uctx, (u64) qp->sq.pa, ocrdma_destroy_qp() 1768 ocrdma_del_mmap(pd->uctx, (u64) qp->rq.pa, ocrdma_destroy_qp() 1914 ocrdma_del_mmap(srq->pd->uctx, (u64) srq->rq.pa, ocrdma_destroy_srq()
|
/linux-4.1.27/drivers/char/agp/ |
H A D | alpha-agp.c | 19 unsigned long pa; alpha_core_agp_vm_fault() local 24 pa = agp->ops->translate(agp, dma_addr); alpha_core_agp_vm_fault() 26 if (pa == (unsigned long)-EINVAL) alpha_core_agp_vm_fault() 32 page = virt_to_page(__va(pa)); alpha_core_agp_vm_fault()
|
/linux-4.1.27/include/ras/ |
H A D | ras_event.h | 38 __field(u64, pa) 53 __entry->pa = mem->physical_addr; 55 __entry->pa = ~0ull; 70 __entry->pa,
|
/linux-4.1.27/arch/powerpc/platforms/ps3/ |
H A D | htab.c | 48 unsigned long pa, unsigned long rflags, unsigned long vflags, ps3_hpte_insert() 66 hpte_r = hpte_encode_r(ps3_mm_phys_to_lpar(pa), psize, apsize) | rflags; ps3_hpte_insert() 79 pr_info("%s:result=%s vpn=%lx pa=%lx ix=%lx v=%llx r=%llx\n", ps3_hpte_insert() 80 __func__, ps3_result(result), vpn, pa, hpte_group, ps3_hpte_insert() 47 ps3_hpte_insert(unsigned long hpte_group, unsigned long vpn, unsigned long pa, unsigned long rflags, unsigned long vflags, int psize, int apsize, int ssize) ps3_hpte_insert() argument
|
/linux-4.1.27/arch/mips/lib/ |
H A D | r3k_dump_tlb.c | 46 " [pa=%06lx n=%d d=%d v=%d g=%d]", dump_tlb()
|
H A D | dump_tlb.c | 89 printk("\t[pa=%0*llx c=%d d=%d v=%d g=%d] ", dump_tlb() 95 printk("[pa=%0*llx c=%d d=%d v=%d g=%d]\n", dump_tlb()
|
/linux-4.1.27/arch/powerpc/mm/ |
H A D | hugepage-hash64.c | 27 unsigned long rflags, pa, hidx; __hash_page_thp() local 131 pa = pmd_pfn(__pmd(old_pmd)) << PAGE_SHIFT; __hash_page_thp() 145 slot = ppc_md.hpte_insert(hpte_group, vpn, pa, rflags, 0, __hash_page_thp() 153 slot = ppc_md.hpte_insert(hpte_group, vpn, pa, __hash_page_thp()
|
H A D | hugetlbpage-hash64.c | 18 unsigned long pa, unsigned long rlags, 27 unsigned long rflags, pa, sz; __hash_page_huge() local 91 pa = pte_pfn(__pte(old_pte)) << PAGE_SHIFT; __hash_page_huge() 107 slot = hpte_insert_repeating(hash, vpn, pa, rflags, 0, __hash_page_huge()
|
H A D | fsl_booke_mmu.c | 90 unsigned long p_mapped_by_tlbcam(phys_addr_t pa) p_mapped_by_tlbcam() argument 94 if (pa >= tlbcam_addrs[b].phys p_mapped_by_tlbcam() 95 && pa < (tlbcam_addrs[b].limit-tlbcam_addrs[b].start) p_mapped_by_tlbcam() 97 return tlbcam_addrs[b].start+(pa-tlbcam_addrs[b].phys); p_mapped_by_tlbcam()
|
H A D | ppc_mmu_32.c | 64 unsigned long p_mapped_by_bats(phys_addr_t pa) p_mapped_by_bats() argument 68 if (pa >= bat_addrs[b].phys p_mapped_by_bats() 69 && pa < (bat_addrs[b].limit-bat_addrs[b].start) p_mapped_by_bats() 71 return bat_addrs[b].start+(pa-bat_addrs[b].phys); p_mapped_by_bats()
|
H A D | mmu_decl.h | 97 extern int map_page(unsigned long va, phys_addr_t pa, int flags); 113 extern int map_kernel_page(unsigned long ea, unsigned long pa, int flags);
|
H A D | pgtable_32.c | 56 extern unsigned long p_mapped_by_bats(phys_addr_t pa); 64 extern unsigned long p_mapped_by_tlbcam(phys_addr_t pa); 288 int map_page(unsigned long va, phys_addr_t pa, int flags) map_page() argument 305 set_pte_at(&init_mm, va, pg, pfn_pte(pa >> PAGE_SHIFT, map_page()
|
H A D | pgtable_64.c | 91 int map_kernel_page(unsigned long ea, unsigned long pa, int flags) map_kernel_page() argument 109 set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, map_kernel_page() 134 set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, map_kernel_page() 143 if (htab_bolt_mapping(ea, ea + PAGE_SIZE, pa, flags, map_kernel_page() 146 "memory at %016lx !\n", pa); map_kernel_page() 171 void __iomem * __ioremap_at(phys_addr_t pa, void *ea, unsigned long size, __ioremap_at() argument 188 WARN_ON(pa & ~PAGE_MASK); __ioremap_at() 193 if (map_kernel_page((unsigned long)ea+i, pa+i, flags)) __ioremap_at()
|
H A D | hash_native_64.c | 195 unsigned long pa, unsigned long rflags, native_hpte_insert() 203 DBG_LOW(" insert(group=%lx, vpn=%016lx, pa=%016lx," native_hpte_insert() 205 hpte_group, vpn, pa, rflags, vflags, psize); native_hpte_insert() 224 hpte_r = hpte_encode_r(pa, psize, apsize) | rflags; native_hpte_insert() 194 native_hpte_insert(unsigned long hpte_group, unsigned long vpn, unsigned long pa, unsigned long rflags, unsigned long vflags, int psize, int apsize, int ssize) native_hpte_insert() argument
|
/linux-4.1.27/arch/arm/mm/ |
H A D | cache-xsc3l2.c | 78 static inline unsigned long l2_map_va(unsigned long pa, unsigned long prev_va) l2_map_va() argument 82 unsigned long pa_offset = pa << (32 - PAGE_SHIFT); l2_map_va() 90 va = (unsigned long)kmap_atomic_pfn(pa >> PAGE_SHIFT); l2_map_va() 94 return __phys_to_virt(pa); l2_map_va()
|
H A D | nommu.c | 115 pr_notice("Ignoring RAM after %pa, memory at %pa ignored\n", for_each_memblock() 144 pr_warn("Truncating memory from %pa to %pa (MPU region constraints)", 152 pr_debug("MPU Region from %pa size %pa (end %pa))\n",
|
H A D | mmu.c | 1090 pr_notice("Ignoring RAM at %pa-%pa (!CONFIG_HIGHMEM)\n", for_each_memblock() 1099 pr_notice("Truncating RAM at %pa-%pa to -%pa", for_each_memblock()
|
/linux-4.1.27/arch/arm/mach-alpine/ |
H A D | platsmp.c | 33 pr_err("FAIL: resume address over 32bit (%pa)", &addr); alpine_boot_secondary()
|
/linux-4.1.27/drivers/net/ethernet/brocade/bna/ |
H A D | bfa_ioc.h | 48 u64 pa; /* ! Physical address */ member in struct:bfa_dma 58 #define bfa_dma_be_addr_set(dma_addr, pa) \ 59 __bfa_dma_be_addr_set(&dma_addr, (u64)pa) 61 __bfa_dma_be_addr_set(union bfi_addr_u *dma_addr, u64 pa) __bfa_dma_be_addr_set() argument 63 dma_addr->a32.addr_lo = (u32) htonl(pa); __bfa_dma_be_addr_set() 64 dma_addr->a32.addr_hi = (u32) htonl(upper_32_bits(pa)); __bfa_dma_be_addr_set() 71 __bfa_alen_set(struct bfi_alen *alen, u32 len, u64 pa) __bfa_alen_set() argument 74 bfa_dma_be_addr_set(alen->al_addr, pa); __bfa_alen_set()
|
H A D | bfa_msgq.c | 524 bfa_dma_be_addr_set(msgq_cfg->cmdq.addr, msgq->cmdq.addr.pa); bfa_msgq_init() 526 bfa_dma_be_addr_set(msgq_cfg->rspq.addr, msgq->rspq.addr.pa); bfa_msgq_init() 597 bfa_msgq_memclaim(struct bfa_msgq *msgq, u8 *kva, u64 pa) bfa_msgq_memclaim() argument 600 msgq->cmdq.addr.pa = pa; bfa_msgq_memclaim() 603 pa += roundup(BFA_MSGQ_CMDQ_SIZE, BFA_DMA_ALIGN_SZ); bfa_msgq_memclaim() 606 msgq->rspq.addr.pa = pa; bfa_msgq_memclaim()
|
H A D | bfa_cee.c | 150 cee->attr_dma.pa = dma_pa; bfa_nw_cee_mem_claim() 152 cee->stats_dma.pa = dma_pa + bfa_cee_attr_meminfo(); bfa_nw_cee_mem_claim() 185 bfa_dma_be_addr_set(cmd->dma_addr, cee->attr_dma.pa); bfa_nw_cee_get_attr()
|
H A D | bfa_msgq.h | 123 void bfa_msgq_memclaim(struct bfa_msgq *msgq, u8 *kva, u64 pa);
|
/linux-4.1.27/arch/hexagon/kernel/ |
H A D | vm_init_segtable.S | 35 #define BKP(pa) (((pa) & __HVM_PTE_PGMASK_4MB) \ 42 #define BKPG_IO(pa) (((pa) & __HVM_PTE_PGMASK_16MB) \ 46 #define FOURK_IO(pa) (((pa) & __HVM_PTE_PGMASK_4KB) \ 50 #define L2_PTR(pa) (((pa) & __HVM_PTE_PGMASK_4KB) \
|
/linux-4.1.27/arch/microblaze/mm/ |
H A D | pgtable.c | 138 int map_page(unsigned long va, phys_addr_t pa, int flags) map_page() argument 151 set_pte_at(&init_mm, va, pg, pfn_pte(pa >> PAGE_SHIFT, map_page() 217 unsigned long pa; iopa() local 230 pa = 0; iopa() 232 pa = (pte_val(*pte) & PAGE_MASK) | (addr & ~PAGE_MASK); iopa() 234 return pa; iopa()
|
H A D | consistent.c | 70 phys_addr_t pa; consistent_alloc() local 120 *dma_handle = pa = __virt_to_phys(vaddr); consistent_alloc() 137 err = map_page(va + i, pa + i, _PAGE_KERNEL | _PAGE_NO_CACHE); consistent_alloc()
|
/linux-4.1.27/drivers/tty/serial/8250/ |
H A D | 8250_hp300.c | 127 unsigned long pa = dio_scodetophysaddr(scode); hp300_setup_serial_console() local 128 if (!pa) hp300_setup_serial_console() 134 port.mapbase = (pa + UART_OFFSET); hp300_setup_serial_console() 137 port.irq = DIO_IPL(pa + DIO_VIRADDRBASE); hp300_setup_serial_console() 140 out_8(pa + DIO_VIRADDRBASE + DCA_IC, DCA_IC_IE); hp300_setup_serial_console() 142 if (DIO_ID(pa + DIO_VIRADDRBASE) & 0x80) hp300_setup_serial_console()
|
/linux-4.1.27/drivers/net/wireless/brcm80211/brcmsmac/ |
H A D | dma.c | 716 dma_addr_t pa, uint outidx, u32 *flags, u32 bufcount) dma64_dd_upd() 721 if ((di->dataoffsetlow == 0) || !(pa & PCI32ADDR_HIGH)) { dma64_dd_upd() 722 ddring[outidx].addrlow = cpu_to_le32(pa + di->dataoffsetlow); dma64_dd_upd() 730 ae = (pa & PCI32ADDR_HIGH) >> PCI32ADDR_HIGH_SHIFT; dma64_dd_upd() 731 pa &= ~PCI32ADDR_HIGH; dma64_dd_upd() 734 ddring[outidx].addrlow = cpu_to_le32(pa + di->dataoffsetlow); dma64_dd_upd() 774 _dma_ddtable_init(struct dma_info *di, uint direction, dma_addr_t pa) _dma_ddtable_init() argument 778 di->xmtptrbase = pa; _dma_ddtable_init() 780 di->rcvptrbase = pa; _dma_ddtable_init() 784 || !(pa & PCI32ADDR_HIGH)) { _dma_ddtable_init() 787 pa + di->ddoffsetlow); _dma_ddtable_init() 792 pa + di->ddoffsetlow); _dma_ddtable_init() 800 /* shift the high bit(s) from pa to ae */ _dma_ddtable_init() 801 ae = (pa & PCI32ADDR_HIGH) >> PCI32ADDR_HIGH_SHIFT; _dma_ddtable_init() 802 pa &= ~PCI32ADDR_HIGH; _dma_ddtable_init() 806 pa + di->ddoffsetlow); _dma_ddtable_init() 813 pa + di->ddoffsetlow); _dma_ddtable_init() 873 dma_addr_t pa; dma64_getnextrxp() local 894 pa = le32_to_cpu(di->rxd64[i].addrlow) - di->dataoffsetlow; dma64_getnextrxp() 897 dma_unmap_single(di->dmadev, pa, di->rxbufsize, DMA_FROM_DEVICE); dma64_getnextrxp() 1031 dma_addr_t pa; dma_rxfill() local 1080 pa = dma_map_single(di->dmadev, p->data, di->rxbufsize, dma_rxfill() 1082 if (dma_mapping_error(di->dmadev, pa)) dma_rxfill() 1093 dma64_dd_upd(di, di->rxd64, pa, rxout, &flags, dma_rxfill() 1275 dma_addr_t pa; dma_txenq() local 1289 pa = dma_map_single(di->dmadev, data, len, DMA_TO_DEVICE); dma_txenq() 1291 if (dma_mapping_error(di->dmadev, pa)) { dma_txenq() 1304 dma64_dd_upd(di, di->txd64, pa, txout, &flags, len); dma_txenq() 1510 dma_addr_t pa; dma_getnexttxp() local 1513 pa = le32_to_cpu(di->txd64[i].addrlow) - di->dataoffsetlow; dma_getnexttxp() 1525 dma_unmap_single(di->dmadev, pa, size, DMA_TO_DEVICE); dma_getnexttxp() 715 dma64_dd_upd(struct dma_info *di, struct dma64desc *ddring, dma_addr_t pa, uint outidx, u32 *flags, u32 bufcount) dma64_dd_upd() argument
|
/linux-4.1.27/drivers/base/ |
H A D | platform.c | 209 struct platform_object *pa = container_of(dev, struct platform_object, platform_device_release() local 212 of_device_node_put(&pa->pdev.dev); platform_device_release() 213 kfree(pa->pdev.dev.platform_data); platform_device_release() 214 kfree(pa->pdev.mfd_cell); platform_device_release() 215 kfree(pa->pdev.resource); platform_device_release() 216 kfree(pa->pdev.driver_override); platform_device_release() 217 kfree(pa); platform_device_release() 230 struct platform_object *pa; platform_device_alloc() local 232 pa = kzalloc(sizeof(*pa) + strlen(name) + 1, GFP_KERNEL); platform_device_alloc() 233 if (pa) { platform_device_alloc() 234 strcpy(pa->name, name); platform_device_alloc() 235 pa->pdev.name = pa->name; platform_device_alloc() 236 pa->pdev.id = id; platform_device_alloc() 237 device_initialize(&pa->pdev.dev); platform_device_alloc() 238 pa->pdev.dev.release = platform_device_release; platform_device_alloc() 239 arch_setup_pdev_archdata(&pa->pdev); platform_device_alloc() 242 return pa ? &pa->pdev : NULL; platform_device_alloc()
|
H A D | dma-coherent.c | 287 pr_err("Reserved memory: failed to init DMA memory pool at %pa, size %ld MiB\n", rmem_dma_device_init() 322 pr_info("Reserved memory: created DMA memory pool at %pa, size %ld MiB\n", rmem_dma_setup()
|
H A D | dma-contiguous.c | 274 pr_info("Reserved memory: created CMA memory pool at %pa, size %ld MiB\n", rmem_cma_setup()
|
/linux-4.1.27/fs/nfs_common/ |
H A D | nfsacl.c | 220 struct posix_acl_entry *pa, *pe, posix_acl_from_nfsacl() local 230 FOREACH_ACL_ENTRY(pa, acl, pe) { FOREACH_ACL_ENTRY() 231 switch(pa->e_tag) { FOREACH_ACL_ENTRY() 235 group_obj = pa; FOREACH_ACL_ENTRY() 238 mask = pa; FOREACH_ACL_ENTRY()
|
/linux-4.1.27/drivers/media/v4l2-core/ |
H A D | videobuf2-memops.c | 103 dma_addr_t pa = 0; vb2_get_contig_userptr() local 120 pa = this_pfn << PAGE_SHIFT; vb2_get_contig_userptr() 134 *res_pa = pa + offset; vb2_get_contig_userptr()
|
/linux-4.1.27/sound/soc/sirf/ |
H A D | sirf-audio.c | 101 "spk-pa-gpios", 0); sirf_audio_probe() 103 "hp-pa-gpios", 0); sirf_audio_probe()
|
/linux-4.1.27/drivers/infiniband/hw/ehca/ |
H A D | ehca_av.c | 57 struct ib_port_attr pa; ehca_calc_ipd() local 70 ret = ehca_query_port(&shca->ib_device, port, &pa); ehca_calc_ipd() 76 link = ib_width_enum_to_int(pa.active_width) * pa.active_speed; ehca_calc_ipd()
|
/linux-4.1.27/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_hmc.h | 117 * @pa: pointer to physical address 121 #define I40E_SET_PF_SD_ENTRY(hw, pa, sd_index, type) \ 124 val1 = (u32)(upper_32_bits(pa)); \ 125 val2 = (u32)(pa) | (I40E_HMC_MAX_BP_COUNT << \
|
H A D | i40e_adminq.c | 194 cpu_to_le32(upper_32_bits(bi->pa)); i40e_alloc_arq_bufs() 196 cpu_to_le32(lower_32_bits(bi->pa)); i40e_alloc_arq_bufs() 283 if (hw->aq.asq.r.asq_bi[i].pa) i40e_free_asq_bufs() 314 wr32(hw, hw->aq.asq.bal, lower_32_bits(hw->aq.asq.desc_buf.pa)); i40e_config_asq_regs() 315 wr32(hw, hw->aq.asq.bah, upper_32_bits(hw->aq.asq.desc_buf.pa)); i40e_config_asq_regs() 319 if (reg != lower_32_bits(hw->aq.asq.desc_buf.pa)) i40e_config_asq_regs() 343 wr32(hw, hw->aq.arq.bal, lower_32_bits(hw->aq.arq.desc_buf.pa)); i40e_config_arq_regs() 344 wr32(hw, hw->aq.arq.bah, upper_32_bits(hw->aq.arq.desc_buf.pa)); i40e_config_arq_regs() 351 if (reg != lower_32_bits(hw->aq.arq.desc_buf.pa)) i40e_config_arq_regs() 829 /* Update the address values in the desc with the pa value i40e_asq_send_command() 833 cpu_to_le32(upper_32_bits(dma_buff->pa)); i40e_asq_send_command() 835 cpu_to_le32(lower_32_bits(dma_buff->pa)); i40e_asq_send_command() 994 desc->params.external.addr_high = cpu_to_le32(upper_32_bits(bi->pa)); i40e_clean_arq_element() 995 desc->params.external.addr_low = cpu_to_le32(lower_32_bits(bi->pa)); i40e_clean_arq_element()
|
H A D | i40e_osdep.h | 56 dma_addr_t pa; member in struct:i40e_dma_mem
|
/linux-4.1.27/drivers/net/ethernet/intel/i40evf/ |
H A D | i40e_hmc.h | 117 * @pa: pointer to physical address 121 #define I40E_SET_PF_SD_ENTRY(hw, pa, sd_index, type) \ 124 val1 = (u32)(upper_32_bits(pa)); \ 125 val2 = (u32)(pa) | (I40E_HMC_MAX_BP_COUNT << \
|
H A D | i40e_adminq.c | 192 cpu_to_le32(upper_32_bits(bi->pa)); i40e_alloc_arq_bufs() 194 cpu_to_le32(lower_32_bits(bi->pa)); i40e_alloc_arq_bufs() 281 if (hw->aq.asq.r.asq_bi[i].pa) i40e_free_asq_bufs() 312 wr32(hw, hw->aq.asq.bal, lower_32_bits(hw->aq.asq.desc_buf.pa)); i40e_config_asq_regs() 313 wr32(hw, hw->aq.asq.bah, upper_32_bits(hw->aq.asq.desc_buf.pa)); i40e_config_asq_regs() 317 if (reg != lower_32_bits(hw->aq.asq.desc_buf.pa)) i40e_config_asq_regs() 341 wr32(hw, hw->aq.arq.bal, lower_32_bits(hw->aq.arq.desc_buf.pa)); i40e_config_arq_regs() 342 wr32(hw, hw->aq.arq.bah, upper_32_bits(hw->aq.arq.desc_buf.pa)); i40e_config_arq_regs() 349 if (reg != lower_32_bits(hw->aq.arq.desc_buf.pa)) i40e_config_arq_regs() 780 /* Update the address values in the desc with the pa value i40evf_asq_send_command() 784 cpu_to_le32(upper_32_bits(dma_buff->pa)); i40evf_asq_send_command() 786 cpu_to_le32(lower_32_bits(dma_buff->pa)); i40evf_asq_send_command() 946 desc->params.external.addr_high = cpu_to_le32(upper_32_bits(bi->pa)); i40evf_clean_arq_element() 947 desc->params.external.addr_low = cpu_to_le32(lower_32_bits(bi->pa)); i40evf_clean_arq_element()
|
H A D | i40e_osdep.h | 55 dma_addr_t pa; member in struct:i40e_dma_mem
|
/linux-4.1.27/net/sunrpc/xprtrdma/ |
H A D | frwr_ops.c | 192 u64 pa; frwr_op_map() local 203 pa = seg->mr_dma; frwr_op_map() 205 frmr->fr_pgl->page_list[page_no++] = pa; frwr_op_map() 206 pa += PAGE_SIZE; frwr_op_map()
|
/linux-4.1.27/drivers/scsi/sym53c8xx_2/ |
H A D | sym_fw.c | 260 u32 *pa; sym_fw_setup_bus_addresses() local 269 pa = (u32 *) &np->fwa_bas; sym_fw_setup_bus_addresses() 271 pa[i] = np->scripta_ba + po[i]; sym_fw_setup_bus_addresses() 277 pa = (u32 *) &np->fwb_bas; sym_fw_setup_bus_addresses() 279 pa[i] = np->scriptb_ba + po[i]; sym_fw_setup_bus_addresses() 285 pa = (u32 *) &np->fwz_bas; sym_fw_setup_bus_addresses() 287 pa[i] = np->scriptz_ba + po[i]; sym_fw_setup_bus_addresses()
|
/linux-4.1.27/arch/sparc/mm/ |
H A D | fault_64.c | 98 unsigned long pa; get_user_insn() local 119 pa = pmd_pfn(*pmdp) << PAGE_SHIFT; get_user_insn() 120 pa += tpc & ~HPAGE_MASK; get_user_insn() 125 : "r" (pa), "i" (ASI_PHYS_USE_EC)); get_user_insn() 132 pa = (pte_pfn(pte) << PAGE_SHIFT); get_user_insn() 133 pa += (tpc & ~PAGE_MASK); get_user_insn() 138 : "r" (pa), "i" (ASI_PHYS_USE_EC)); get_user_insn()
|
H A D | init_64.c | 1379 unsigned long pa = __pa(addr); kern_addr_valid() local 1384 return pfn_valid(pa >> PAGE_SHIFT); kern_addr_valid() 1786 static void patch_one_ktsb_phys(unsigned int *start, unsigned int *end, unsigned long pa) patch_one_ktsb_phys() argument 1790 high_bits = (pa >> 32) & 0xffffffff; patch_one_ktsb_phys() 1791 low_bits = (pa >> 0) & 0xffffffff; patch_one_ktsb_phys() 1890 unsigned long pa, ret; sun4v_ktsb_register() local 1892 pa = kern_base + ((unsigned long)&ktsb_descr[0] - KERNBASE); sun4v_ktsb_register() 1894 ret = sun4v_mmu_tsb_ctx0(NUM_KTSB_DESCR, pa); sun4v_ktsb_register() 1897 "errors with %lx\n", pa, ret); sun4v_ktsb_register()
|
/linux-4.1.27/mm/ |
H A D | cma.c | 249 pr_debug("%s(size %pa, base %pa, limit %pa alignment %pa)\n", cma_declare_contiguous() 288 pr_err("Region at %pa defined on low/high memory boundary (%pa)\n", cma_declare_contiguous() 344 pr_info("Reserved %ld MiB at %pa\n", (unsigned long)size / SZ_1M, cma_declare_contiguous()
|
H A D | process_vm_access.c | 85 unsigned long pa = addr & PAGE_MASK; process_vm_rw_single_vec() local 86 unsigned long start_offset = addr - pa; process_vm_rw_single_vec() 102 pages = get_user_pages_unlocked(task, mm, pa, pages, process_vm_rw_single_vec() 117 pa += pages * PAGE_SIZE; process_vm_rw_single_vec()
|
/linux-4.1.27/arch/ia64/kernel/ |
H A D | acpi.c | 422 get_processor_proximity_domain(struct acpi_srat_cpu_affinity *pa) get_processor_proximity_domain() argument 426 pxm = pa->proximity_domain_lo; get_processor_proximity_domain() 428 pxm += pa->proximity_domain_hi[0] << 8; get_processor_proximity_domain() 464 acpi_numa_processor_affinity_init(struct acpi_srat_cpu_affinity *pa) acpi_numa_processor_affinity_init() argument 468 if (!(pa->flags & ACPI_SRAT_CPU_ENABLED)) acpi_numa_processor_affinity_init() 477 pxm = get_processor_proximity_domain(pa); acpi_numa_processor_affinity_init() 483 (pa->apic_id << 8) | (pa->local_sapic_eid); acpi_numa_processor_affinity_init()
|
H A D | brl_emu.c | 40 #define unimplemented_physical_address(pa) ( \ 41 ((pa) & local_cpu_data->unimpl_pa_mask) != 0 \
|
H A D | kprobes.c | 1048 struct param_bsp_cfm pa; setjmp_pre_handler() local 1058 pa.ip = regs->cr_iip; setjmp_pre_handler() 1059 unw_init_running(ia64_get_bsp_cfm, &pa); setjmp_pre_handler() 1060 bytes = (char *)ia64_rse_skip_regs(pa.bsp, pa.cfm & 0x3f) setjmp_pre_handler() 1061 - (char *)pa.bsp; setjmp_pre_handler() 1063 pa.bsp, setjmp_pre_handler() 1065 kcb->bsp = pa.bsp; setjmp_pre_handler() 1066 kcb->cfm = pa.cfm; setjmp_pre_handler()
|
/linux-4.1.27/drivers/media/radio/ |
H A D | radio-keene.c | 71 u8 pa; member in struct:keene_device 92 radio->buffer[4] = radio->pa; keene_cmd_main() 260 radio->pa = (ctrl->val - 71) * 100 / 62; keene_s_ctrl() 350 radio->pa = 118; usb_keene_probe()
|
/linux-4.1.27/arch/sparc/kernel/ |
H A D | hvtramp.S | 22 * with physical addressing (va==pa). %o0 contains
|
H A D | ioport.c | 58 static inline void dma_make_coherent(unsigned long pa, unsigned long len) dma_make_coherent() argument 66 static void __iomem *_sparc_ioremap(struct resource *res, u32 bus, u32 pa, int sz); 216 _sparc_ioremap(struct resource *res, u32 bus, u32 pa, int sz) _sparc_ioremap() argument 218 unsigned long offset = ((unsigned long) pa) & (~PAGE_MASK); _sparc_ioremap() 229 pa &= PAGE_MASK; _sparc_ioremap() 230 srmmu_mapiorange(bus, pa, res->start, resource_size(res)); _sparc_ioremap()
|
H A D | ldc.c | 2016 static void fill_cookies(struct cookie_state *sp, unsigned long pa, fill_cookies() argument 2020 unsigned long tlen, new = pa + PAGE_SIZE; fill_cookies() 2023 sp->page_table[sp->pte_idx].mte = sp->mte_base | pa; fill_cookies() 2048 pa = new; fill_cookies() 2128 unsigned long npages, pa; ldc_map_single() local 2136 pa = __pa(buf); ldc_map_single() 2137 if ((pa | len) & (8UL - 1)) ldc_map_single() 2140 npages = pages_in_region(pa, len); ldc_map_single() 2155 fill_cookies(&state, (pa & PAGE_MASK), (pa & ~PAGE_MASK), len); ldc_map_single()
|
/linux-4.1.27/drivers/media/dvb-core/ |
H A D | dvb_filter.c | 30 static void setup_ts2pes(ipack *pa, ipack *pv, u16 *pida, u16 *pidv, 34 dvb_filter_ipack_init(pa, IPACKS, pes_write); 36 pa->pid = pida; 38 pa->data = priv;
|
/linux-4.1.27/arch/parisc/lib/ |
H A D | memset.c | 19 /* Slight modifications for pa-risc linux - Paul Bame <bame@debian.org> */
|
/linux-4.1.27/arch/parisc/math-emu/ |
H A D | fpu.h | 25 * @(#) pa/fp/fpu.h $Revision: 1.1 $
|
H A D | denormal.c | 25 * @(#) pa/fp/denormal.c $ Revision: $
|
H A D | sfcmp.c | 25 * @(#) pa/spmath/sfcmp.c $Revision: 1.1 $
|
H A D | dfcmp.c | 25 * @(#) pa/spmath/dfcmp.c $Revision: 1.1 $
|
H A D | dfsqrt.c | 25 * @(#) pa/spmath/dfsqrt.c $Revision: 1.1 $
|
H A D | sfsqrt.c | 25 * @(#) pa/spmath/sfsqrt.c $Revision: 1.1 $
|
H A D | float.h | 25 * @(#) pa/spmath/float.h $Revision: 1.1 $ 33 * /usr/conf/pa/spmath/float.h
|
H A D | dfdiv.c | 25 * @(#) pa/spmath/dfdiv.c $Revision: 1.1 $
|
H A D | dfmpy.c | 25 * @(#) pa/spmath/dfmpy.c $Revision: 1.1 $
|
H A D | dfrem.c | 25 * @(#) pa/spmath/dfrem.c $Revision: 1.1 $
|
H A D | fcnvff.c | 25 * @(#) pa/spmath/fcnvff.c $Revision: 1.1 $
|
H A D | fcnvfut.c | 25 * @(#) pa/spmath/fcnvfut.c $Revision: 1.1 $
|
H A D | fcnvfxt.c | 25 * @(#) pa/spmath/fcnvfxt.c $Revision: 1.1 $
|
H A D | fcnvuf.c | 25 * @(#) pa/spmath/fcnvuf.c $Revision: 1.1 $
|
H A D | fcnvxf.c | 25 * @(#) pa/spmath/fcnvxf.c $Revision: 1.1 $
|
H A D | sfdiv.c | 25 * @(#) pa/spmath/sfdiv.c $Revision: 1.1 $
|
H A D | sfmpy.c | 25 * @(#) pa/spmath/sfmpy.c $Revision: 1.1 $
|
H A D | sfrem.c | 25 * @(#) pa/spmath/sfrem.c $Revision: 1.1 $
|
/linux-4.1.27/drivers/scsi/be2iscsi/ |
H A D | be_mgmt.h | 245 #define ISCSI_GET_PDU_TEMPLATE_ADDRESS(pc, pa) {\ 246 pa->lo = phba->init_mem[ISCSI_MEM_GLOBAL_HEADER].mem_array[0].\ 248 pa->hi = phba->init_mem[ISCSI_MEM_GLOBAL_HEADER].mem_array[0].\
|
/linux-4.1.27/drivers/scsi/fnic/ |
H A D | fnic_fcs.c | 941 dma_addr_t pa; fnic_alloc_rq_frame() local 954 pa = pci_map_single(fnic->pdev, skb->data, len, PCI_DMA_FROMDEVICE); fnic_alloc_rq_frame() 955 fnic_queue_rq_desc(rq, skb, pa, len); fnic_alloc_rq_frame() 980 dma_addr_t pa; fnic_eth_send() local 1004 pa = pci_map_single(fnic->pdev, skb->data, skb->len, PCI_DMA_TODEVICE); fnic_eth_send() 1008 pci_unmap_single(fnic->pdev, pa, skb->len, PCI_DMA_TODEVICE); fnic_eth_send() 1014 fnic_queue_wq_eth_desc(wq, skb, pa, skb->len, fnic_eth_send() 1027 dma_addr_t pa; fnic_send_frame() local 1072 pa = pci_map_single(fnic->pdev, eth_hdr, tot_len, PCI_DMA_TODEVICE); fnic_send_frame() 1082 pci_unmap_single(fnic->pdev, pa, fnic_send_frame() 1088 fnic_queue_wq_desc(wq, skb, pa, tot_len, fr_eof(fp), fnic_send_frame()
|
/linux-4.1.27/drivers/edac/ |
H A D | cpc925_edac.c | 453 unsigned long pa; cpc925_mc_get_pfn() local 473 pa = mci->csrows[rank]->first_page << PAGE_SHIFT; cpc925_mc_get_pfn() 480 pa |= c << (14 - i); cpc925_mc_get_pfn() 484 pa |= bank << 19; cpc925_mc_get_pfn() 490 pa |= c << (26 - i); cpc925_mc_get_pfn() 496 pa |= c << (21 + i); cpc925_mc_get_pfn() 502 pa |= c << (18 - i); cpc925_mc_get_pfn() 508 pa |= c << (29 - i); cpc925_mc_get_pfn() 511 *offset = pa & (PAGE_SIZE - 1); cpc925_mc_get_pfn() 512 *pfn = pa >> PAGE_SHIFT; cpc925_mc_get_pfn() 514 edac_dbg(0, "ECC physical address 0x%lx\n", pa); cpc925_mc_get_pfn()
|
/linux-4.1.27/arch/x86/platform/intel-quark/ |
H A D | imr.c | 236 seq_printf(s, "imr%02i: base=%pa, end=%pa, size=0x%08zx " imr_dbgfs_state_show() 300 pr_err("base %pa size 0x%08zx must align to 1KiB\n", imr_check_params() 412 pr_debug("add %d phys %pa-%pa size %zx mask 0x%08x wmask 0x%08x\n", imr_add_range() 520 pr_debug("remove %d phys %pa-%pa size %zx\n", reg, &base, &end, raw_size); __imr_remove_range()
|
/linux-4.1.27/arch/powerpc/kvm/ |
H A D | book3s_hv_rm_mmu.c | 141 unsigned long i, pa, gpa, gfn, psize; kvmppc_do_h_enter() local 171 pa = 0; kvmppc_do_h_enter() 222 pa = pte_pfn(pte) << PAGE_SHIFT; kvmppc_do_h_enter() 223 pa |= hva & (host_pte_size - 1); kvmppc_do_h_enter() 224 pa |= gpa & ~PAGE_MASK; kvmppc_do_h_enter() 231 ptel |= pa; kvmppc_do_h_enter() 233 if (pa) kvmppc_do_h_enter()
|
/linux-4.1.27/sound/pci/ |
H A D | es1968.c | 1006 u32 pa; snd_es1968_playback_setup() local 1024 pa = es->memory->buf.addr; snd_es1968_playback_setup() 1025 pa -= chip->dma.addr; snd_es1968_playback_setup() 1026 pa >>= 1; /* words */ snd_es1968_playback_setup() 1028 pa |= 0x00400000; /* System RAM (Bit 22) */ snd_es1968_playback_setup() 1033 pa |= 0x00800000; /* (Bit 23) */ snd_es1968_playback_setup() 1035 pa >>= 1; snd_es1968_playback_setup() 1040 es->base[channel] = pa & 0xFFFF; snd_es1968_playback_setup() 1046 apu_set_register(chip, apu, 4, ((pa >> 16) & 0xFF) << 8); snd_es1968_playback_setup() 1047 apu_set_register(chip, apu, 5, pa & 0xFFFF); snd_es1968_playback_setup() 1048 apu_set_register(chip, apu, 6, (pa + size) & 0xFFFF); snd_es1968_playback_setup() 1107 unsigned int pa, unsigned int bsize, init_capture_apu() 1115 snd_es1968_program_wavecache(chip, es, channel, pa, 1); init_capture_apu() 1118 pa -= chip->dma.addr; init_capture_apu() 1119 pa >>= 1; /* words */ init_capture_apu() 1123 es->base[channel] = pa & 0xFFFF; init_capture_apu() 1124 pa |= 0x00400000; /* bit 22 -> System RAM */ init_capture_apu() 1135 apu_set_register(chip, apu, 4, ((pa >> 16) & 0xFF) << 8); init_capture_apu() 1136 apu_set_register(chip, apu, 5, pa & 0xFFFF); init_capture_apu() 1137 apu_set_register(chip, apu, 6, (pa + bsize) & 0xFFFF); init_capture_apu() 1711 unsigned int pa, offset, t; es1968_measure_clock() local 1736 pa = (unsigned int)((memory->buf.addr - chip->dma.addr) >> 1); es1968_measure_clock() 1737 pa |= 0x00400000; /* System RAM (Bit 22) */ es1968_measure_clock() 1744 apu_set_register(chip, apu, 4, ((pa >> 16) & 0xff) << 8); es1968_measure_clock() 1745 apu_set_register(chip, apu, 5, pa & 0xffff); es1968_measure_clock() 1746 apu_set_register(chip, apu, 6, (pa + CLOCK_MEASURE_BUFSIZE/2) & 0xffff); es1968_measure_clock() 1763 __apu_set_register(chip, apu, 5, pa & 0xffff); es1968_measure_clock() 1777 offset -= (pa & 0xffff); es1968_measure_clock() 1106 init_capture_apu(struct es1968 *chip, struct esschan *es, int channel, unsigned int pa, unsigned int bsize, int mode, int route) init_capture_apu() argument
|
/linux-4.1.27/include/trace/events/ |
H A D | ext4.h | 612 struct ext4_prealloc_space *pa), 614 TP_ARGS(ac, pa), 628 __entry->pa_pstart = pa->pa_pstart; 629 __entry->pa_lstart = pa->pa_lstart; 630 __entry->pa_len = pa->pa_len; 642 struct ext4_prealloc_space *pa), 644 TP_ARGS(ac, pa) 650 struct ext4_prealloc_space *pa), 652 TP_ARGS(ac, pa) 656 TP_PROTO(struct ext4_prealloc_space *pa, 659 TP_ARGS(pa, block, count), 670 __entry->dev = pa->pa_inode->i_sb->s_dev; 671 __entry->ino = pa->pa_inode->i_ino; 683 TP_PROTO(struct super_block *sb, struct ext4_prealloc_space *pa), 685 TP_ARGS(sb, pa), 696 __entry->pa_pstart = pa->pa_pstart; 697 __entry->pa_len = pa->pa_len;
|
/linux-4.1.27/drivers/net/can/usb/peak_usb/ |
H A D | pcan_usb_core.c | 707 const struct peak_usb_adapter *pa = dev->adapter; peak_usb_set_bittiming() local 709 if (pa->dev_set_bittiming) { peak_usb_set_bittiming() 711 int err = pa->dev_set_bittiming(dev, bt); peak_usb_set_bittiming() 728 const struct peak_usb_adapter *pa = dev->adapter; peak_usb_set_data_bittiming() local 730 if (pa->dev_set_data_bittiming) { peak_usb_set_data_bittiming() 732 int err = pa->dev_set_data_bittiming(dev, bt); peak_usb_set_data_bittiming()
|
/linux-4.1.27/drivers/scsi/bfa/ |
H A D | bfa_port.c | 219 port->stats_dma.pa = dma_pa; bfa_port_mem_claim() 365 bfa_dma_be_addr_set(m->dma_addr, port->stats_dma.pa); bfa_port_get_stats() 639 cee->attr_dma.pa = dma_pa; bfa_cee_mem_claim() 642 cee->stats_dma.pa = dma_pa + BFA_ROUNDUP( bfa_cee_mem_claim() 683 bfa_dma_be_addr_set(cmd->dma_addr, cee->attr_dma.pa); bfa_cee_get_attr() 723 bfa_dma_be_addr_set(cmd->dma_addr, cee->stats_dma.pa); bfa_cee_get_stats()
|
H A D | bfa_core.c | 1001 bfa_dma_be_addr_set(cfg_info->cfgrsp_addr, iocfc->cfgrsp_dma.pa); bfa_iocfc_send_cfg() 1007 iocfc->req_cq_ba[i].pa); bfa_iocfc_send_cfg() 1009 iocfc->req_cq_shadow_ci[i].pa); bfa_iocfc_send_cfg() 1014 iocfc->rsp_cq_ba[i].pa); bfa_iocfc_send_cfg() 1016 iocfc->rsp_cq_shadow_pi[i].pa); bfa_iocfc_send_cfg() 1033 bfa_dma_be_addr_set(cfg_req.ioc_cfg_dma_addr, iocfc->cfg_info.pa); bfa_iocfc_send_cfg() 1116 iocfc->req_cq_ba[i].pa = bfa_mem_dma_phys(reqq_dma); bfa_iocfc_mem_claim() 1121 iocfc->rsp_cq_ba[i].pa = bfa_mem_dma_phys(rspq_dma); bfa_iocfc_mem_claim() 1131 iocfc->req_cq_shadow_ci[i].pa = dm_pa; bfa_iocfc_mem_claim() 1136 iocfc->rsp_cq_shadow_pi[i].pa = dm_pa; bfa_iocfc_mem_claim() 1143 bfa->iocfc.cfg_info.pa = dm_pa; bfa_iocfc_mem_claim() 1150 bfa->iocfc.cfgrsp_dma.pa = dm_pa; bfa_iocfc_mem_claim()
|
H A D | bfa_ioc.h | 144 /* Get the corresponding dma buf pa for a req - from the tag */ 166 u64 pa; /* ! Physical address */ member in struct:bfa_dma_s 178 #define bfa_dma_be_addr_set(dma_addr, pa) \ 179 __bfa_dma_be_addr_set(&dma_addr, (u64)pa) 181 __bfa_dma_be_addr_set(union bfi_addr_u *dma_addr, u64 pa) __bfa_dma_be_addr_set() argument 183 dma_addr->a32.addr_lo = cpu_to_be32(pa); __bfa_dma_be_addr_set() 184 dma_addr->a32.addr_hi = cpu_to_be32(pa >> 32); __bfa_dma_be_addr_set() 191 __bfa_alen_set(struct bfi_alen_s *alen, u32 len, u64 pa) __bfa_alen_set() argument 194 bfa_dma_be_addr_set(alen->al_addr, pa); __bfa_alen_set()
|
/linux-4.1.27/drivers/of/ |
H A D | fdt.c | 299 const char *p1 = pathp, *ps = pathp, *pa = NULL; unflatten_dt_node() local 304 pa = p1; unflatten_dt_node() 309 if (pa < ps) unflatten_dt_node() 310 pa = p1; unflatten_dt_node() 311 sz = (pa - ps) + 1; unflatten_dt_node() 491 pr_debug("Reserved memory: reserved region for node '%s': base %pa, size %ld MiB\n", __reserved_mem_reserve_reg() 494 pr_info("Reserved memory: failed to reserve memory for node '%s': base %pa, size %ld MiB\n", __reserved_mem_reserve_reg() 1023 pr_err("Reserved memory not supported, ignoring range 0x%pa - 0x%pa%s\n", early_init_dt_reserve_memory_arch()
|
H A D | of_reserved_mem.c | 147 pr_debug("Reserved memory: allocated memory for '%s' node: base %pa, size %ld MiB\n", __reserved_mem_alloc_size() 159 pr_debug("Reserved memory: allocated memory for '%s' node: base %pa, size %ld MiB\n", __reserved_mem_alloc_size()
|
/linux-4.1.27/arch/tile/include/uapi/arch/ |
H A D | sim.h | 374 sim_validate_lines_evicted(unsigned long long pa, unsigned long length) sim_validate_lines_evicted() argument 377 _sim_syscall(SIM_SYSCALL_VALIDATE_LINES_EVICTED, 2, pa, length); sim_validate_lines_evicted() 380 0 /* dummy */, (long)(pa), (long)(pa >> 32), length); sim_validate_lines_evicted()
|
/linux-4.1.27/arch/alpha/kernel/ |
H A D | core_marvel.c | 859 marvel_pa_to_nid(unsigned long pa) 863 if ((pa >> 43) & 1) /* I/O */ marvel_cpuid_to_nid() 864 cpuid = (~(pa >> 35) & 0xff); marvel_cpuid_to_nid() 866 cpuid = ((pa >> 34) & 0x3) | ((pa >> (37 - 2)) & (0x1f << 2)); 880 unsigned long pa; marvel_node_mem_size() 882 pa = (nid & 0x3) | ((nid & (0x1f << 2)) << 1); 883 pa <<= 34; 885 return pa; 849 marvel_pa_to_nid(unsigned long pa) marvel_pa_to_nid() argument 870 unsigned long pa; marvel_node_mem_start() local
|
H A D | core_wildfire.c | 445 int wildfire_pa_to_nid(unsigned long pa) wildfire_pa_to_nid() 447 return pa >> 36; 444 wildfire_pa_to_nid(unsigned long pa) wildfire_pa_to_nid() argument
|
/linux-4.1.27/drivers/infiniband/hw/mlx5/ |
H A D | mem.c | 171 dma_addr_t pa = umem->odp_data->dma_list[offset + i]; __mlx5_ib_populate_pas() local 173 pas[i] = cpu_to_be64(umem_dma_to_mtt(pa)); __mlx5_ib_populate_pas()
|
/linux-4.1.27/arch/mips/include/asm/octeon/ |
H A D | cvmx-address.h | 135 uint64_t pa:49; /* physical address */ member in struct:__anon1911::__anon1915 224 uint64_t pa:49; member in struct:__anon1911::__anon1925
|
/linux-4.1.27/drivers/acpi/ |
H A D | numa.c | 192 acpi_numa_x2apic_affinity_init(struct acpi_srat_x2apic_cpu_affinity *pa) acpi_numa_x2apic_affinity_init() argument 195 "Found unsupported x2apic [0x%08x] SRAT entry\n", pa->apic_id); acpi_numa_x2apic_affinity_init()
|
/linux-4.1.27/arch/mips/ralink/ |
H A D | rt305x.c | 51 static struct rt2880_pmx_func rt3352_pa_func[] = { FUNC("pa", 0, 38, 2) }; 77 GRP("pa", rt3352_pa_func, 1, RT3352_GPIO_MODE_PA),
|
H A D | mt7620.c | 55 static struct rt2880_pmx_func pa_grp[] = { FUNC("pa", 0, 18, 4) }; 96 GRP("pa", pa_grp, 1, MT7620_GPIO_MODE_PA),
|
/linux-4.1.27/arch/powerpc/kernel/ |
H A D | prom.c | 136 * ibm,pa-features is a per-cpu property that contains a string of 144 * pa-features property is missing, or a 1/0 to indicate if the feature 153 unsigned char pabyte; /* byte number in ibm,pa-features */ 155 unsigned char invert; /* if 1, pa bit set => clear feature */ 216 pa_ftrs = of_get_flat_dt_prop(node, "ibm,pa-features", &tablelen); check_cpu_pa_features()
|
/linux-4.1.27/drivers/remoteproc/ |
H A D | remoteproc_core.c | 459 * the 'pa' member. If a specific device address is expected, then 'da' will 497 ret = iommu_map(rproc->domain, rsc->da, rsc->pa, rsc->len, rsc->flags); rproc_handle_devmem() 514 dev_dbg(dev, "mapped devmem pa 0x%x, da 0x%x, len 0x%x\n", rproc_handle_devmem() 515 rsc->pa, rsc->da, rsc->len); rproc_handle_devmem() 564 dev_dbg(dev, "carveout rsc: da %x, pa %x, len %x, flags %x\n", rproc_handle_carveout() 565 rsc->da, rsc->pa, rsc->len, rsc->flags); rproc_handle_carveout() 631 * Some remote processors might need to know the pa rproc_handle_carveout() 635 * the IOMMU, and therefor must know the pa. rproc_handle_carveout() 647 rsc->pa = dma; rproc_handle_carveout()
|
/linux-4.1.27/arch/x86/platform/efi/ |
H A D | efi_64.c | 220 u64 pa = md->phys_addr; efi_map_region() local 245 if (!(pa & (PMD_SIZE - 1))) { efi_map_region() 248 u64 pa_offset = pa & (PMD_SIZE - 1); efi_map_region()
|
/linux-4.1.27/arch/mips/mm/ |
H A D | cerr-sb1.c | 493 unsigned long long taglo, pa; extract_dc() local 514 pa = (taglo & 0xFFFFFFE000ULL) | addr; extract_dc() 528 way, pa, dc_state_str(state), state, taghi, taglo); extract_dc()
|
/linux-4.1.27/drivers/staging/android/ion/ |
H A D | ion_cma_heap.c | 116 dev_dbg(dev, "Return buffer %p physical address %pa\n", buffer, ion_cma_phys()
|
/linux-4.1.27/arch/powerpc/include/asm/ |
H A D | mmu-hash64.h | 268 * for the page size. We assume the pa is already "clean" that is properly 271 static inline unsigned long hpte_encode_r(unsigned long pa, int base_psize, hpte_encode_r() argument 276 return pa & HPTE_R_RPN; hpte_encode_r() 280 return (pa & ~((1ul << shift) - 1)) | (penc << LP_SHIFT); hpte_encode_r()
|
H A D | page.h | 79 * va = pa + PAGE_OFFSET - MEMORY_START 80 * va = pa + KERNELBASE - PHYSICAL_START
|
/linux-4.1.27/arch/m68k/sun3x/ |
H A D | dvma.c | 165 // printk("mapping pa %lx to ba %lx\n", __pa(kaddr), index << DVMA_PAGE_SHIFT); dvma_map_iommu()
|
/linux-4.1.27/tools/perf/util/ |
H A D | probe-event.h | 106 extern int synthesize_perf_probe_arg(struct perf_probe_arg *pa, char *buf,
|
H A D | probe-event.c | 1487 int synthesize_perf_probe_arg(struct perf_probe_arg *pa, char *buf, size_t len) synthesize_perf_probe_arg() argument 1489 struct perf_probe_arg_field *field = pa->field; synthesize_perf_probe_arg() 1493 if (pa->name && pa->var) synthesize_perf_probe_arg() 1494 ret = e_snprintf(tmp, len, "%s=%s", pa->name, pa->var); synthesize_perf_probe_arg() 1496 ret = e_snprintf(tmp, len, "%s", pa->name ? pa->name : pa->var); synthesize_perf_probe_arg() 1515 if (pa->type) { synthesize_perf_probe_arg() 1516 ret = e_snprintf(tmp, len, ":%s", pa->type); synthesize_perf_probe_arg()
|
/linux-4.1.27/include/linux/bcma/ |
H A D | bcma_driver_chipcommon.h | 507 #define BCMA_CHIPCTL_4331_EXTPA_EN BIT(4) /* 0 ext pa disable, 1 ext pa enabled */ 515 #define BCMA_CHIPCTL_4331_EXTPA_EN2 BIT(12) /* 0 ext pa disable, 1 ext pa enabled */
|
/linux-4.1.27/drivers/parisc/ |
H A D | ccio-dma.c | 560 register unsigned long pa; ccio_io_pdir_entry() local 573 pa = virt_to_phys(vba); ccio_io_pdir_entry() 574 asm volatile("depw %1,31,12,%0" : "+r" (pa) : "r" (hints)); ccio_io_pdir_entry() 575 ((u32 *)pdir_ptr)[1] = (u32) pa; ccio_io_pdir_entry() 587 asm volatile ("extrd,u %1,15,4,%0" : "=r" (ci) : "r" (pa)); ccio_io_pdir_entry() 588 asm volatile ("extrd,u %1,31,16,%0" : "+r" (pa) : "r" (pa)); ccio_io_pdir_entry() 589 asm volatile ("depd %1,35,4,%0" : "+r" (pa) : "r" (ci)); ccio_io_pdir_entry() 591 pa = 0; ccio_io_pdir_entry() 600 asm volatile ("depw %1,15,12,%0" : "+r" (pa) : "r" (ci)); ccio_io_pdir_entry() 602 ((u32 *)pdir_ptr)[0] = (u32) pa; ccio_io_pdir_entry()
|
H A D | sba_iommu.c | 570 u64 pa; /* physical address */ sba_io_pdir_entry() local 573 pa = virt_to_phys(vba); sba_io_pdir_entry() 574 pa &= IOVP_MASK; sba_io_pdir_entry() 578 pa |= (ci >> PAGE_SHIFT) & 0xff; /* move CI (8 bits) into lowest byte */ sba_io_pdir_entry() 580 pa |= SBA_PDIR_VALID_BIT; /* set "valid" bit */ sba_io_pdir_entry() 581 *pdir_ptr = cpu_to_le64(pa); /* swap and store into I/O Pdir */ sba_io_pdir_entry()
|
/linux-4.1.27/arch/tile/include/asm/ |
H A D | page.h | 152 #define __pa_to_highbits(pa) ((phys_addr_t)(pa) >> NR_PA_HIGHBIT_SHIFT)
|
/linux-4.1.27/drivers/infiniband/hw/qib/ |
H A D | qib_iba6120.c | 1862 * @pa: physical address of in memory buffer; tidinvalid if freeing 1869 u32 type, unsigned long pa) qib_6120_put_tid() 1879 if (pa != dd->tidinvalid) { qib_6120_put_tid() 1880 if (pa & ((1U << 11) - 1)) { qib_6120_put_tid() 1882 pa); qib_6120_put_tid() 1885 pa >>= 11; qib_6120_put_tid() 1886 if (pa & ~QLOGIC_IB_RT_ADDR_MASK) { qib_6120_put_tid() 1889 pa); qib_6120_put_tid() 1894 pa |= dd->tidtemplate; qib_6120_put_tid() 1896 pa |= 2 << 29; qib_6120_put_tid() 1918 writel(pa, tidp32); qib_6120_put_tid() 1930 * @pa: physical address of in memory buffer; tidinvalid if freeing 1937 u32 type, unsigned long pa) qib_6120_put_tid_2() 1945 if (pa != dd->tidinvalid) { qib_6120_put_tid_2() 1946 if (pa & ((1U << 11) - 1)) { qib_6120_put_tid_2() 1948 pa); qib_6120_put_tid_2() 1951 pa >>= 11; qib_6120_put_tid_2() 1952 if (pa & ~QLOGIC_IB_RT_ADDR_MASK) { qib_6120_put_tid_2() 1955 pa); qib_6120_put_tid_2() 1960 pa |= dd->tidtemplate; qib_6120_put_tid_2() 1962 pa |= 2 << 29; qib_6120_put_tid_2() 1965 writel(pa, tidp32); qib_6120_put_tid_2() 1868 qib_6120_put_tid(struct qib_devdata *dd, u64 __iomem *tidptr, u32 type, unsigned long pa) qib_6120_put_tid() argument 1936 qib_6120_put_tid_2(struct qib_devdata *dd, u64 __iomem *tidptr, u32 type, unsigned long pa) qib_6120_put_tid_2() argument
|
/linux-4.1.27/drivers/infiniband/hw/ipath/ |
H A D | ipath_iba6110.c | 1478 * @pa: physical address of in memory buffer; ipath_tidinvalid if freeing 1486 unsigned long pa) ipath_ht_put_tid() 1491 if (pa != dd->ipath_tidinvalid) { ipath_ht_put_tid() 1492 if (unlikely((pa & ~INFINIPATH_RT_ADDR_MASK))) { ipath_ht_put_tid() 1495 "40 bits, using only 40!!!\n", pa); ipath_ht_put_tid() 1496 pa &= INFINIPATH_RT_ADDR_MASK; ipath_ht_put_tid() 1499 pa |= dd->ipath_tidtemplate; ipath_ht_put_tid() 1504 pa |= lenvalid | INFINIPATH_RT_VALID; ipath_ht_put_tid() 1508 writeq(pa, tidptr); ipath_ht_put_tid() 1484 ipath_ht_put_tid(struct ipath_devdata *dd, u64 __iomem *tidptr, u32 type, unsigned long pa) ipath_ht_put_tid() argument
|
/linux-4.1.27/drivers/video/fbdev/ |
H A D | mxsfb.c | 631 unsigned long pa, fbsize; mxsfb_restore_mode() local 700 pa = readl(host->base + host->devdata->cur_buf); mxsfb_restore_mode() 702 if (pa < fb_info->fix.smem_start) { mxsfb_restore_mode() 706 if (pa + fbsize > fb_info->fix.smem_start + fb_info->fix.smem_len) { mxsfb_restore_mode() 710 ofs = pa - fb_info->fix.smem_start; mxsfb_restore_mode()
|
/linux-4.1.27/drivers/staging/gdm724x/ |
H A D | gdm_lte.c | 195 u16 pa[20]; icmp6_checksum() member in union:__anon9802 205 for (i = 0; i < ARRAY_SIZE(pseudo_header.pa); i++) icmp6_checksum() 206 sum += pseudo_header.pa[i]; icmp6_checksum()
|
/linux-4.1.27/drivers/pinctrl/ |
H A D | pinctrl-single.c | 132 * @pa: pindesc array 140 struct pinctrl_pin_desc *pa; member in struct:pcs_data 781 pin = &pcs->pins.pa[i]; pcs_add_pin() 817 pcs->pins.pa = devm_kzalloc(pcs->dev, pcs_allocate_pin_table() 818 sizeof(*pcs->pins.pa) * nr_pins, pcs_allocate_pin_table() 820 if (!pcs->pins.pa) pcs_allocate_pin_table() 829 pcs->desc.pins = pcs->pins.pa; pcs_allocate_pin_table() 1955 dev_info(pcs->dev, "%i pins at pa %p size %u\n", pcs_probe()
|
/linux-4.1.27/arch/arm64/mm/ |
H A D | mmu.c | 273 pr_warn("BUG: not creating mapping for %pa at 0x%016lx - outside kernel range\n", create_mapping() 293 pr_warn("BUG: not creating mapping for %pa at 0x%016lx - outside kernel range\n", create_mapping_late()
|
/linux-4.1.27/drivers/memory/tegra/ |
H A D | mc.c | 198 dev_err_ratelimited(mc->dev, "%s: %s%s @%pa: %s (%s%s)\n", tegra_mc_irq()
|
/linux-4.1.27/drivers/misc/mic/card/ |
H A D | mic_x100.c | 201 mdev->mmio.pa = MIC_X100_MMIO_BASE; mic_probe()
|
/linux-4.1.27/drivers/staging/media/lirc/ |
H A D | lirc_bt829.c | 77 pr_info("memory at %pa\n", &pci_addr_phys); do_pci_probe()
|
/linux-4.1.27/arch/tile/kernel/ |
H A D | machine_kexec.c | 222 * Address range in which pa=va mapping is set in setup_quasi_va_is_pa().
|
/linux-4.1.27/drivers/block/ |
H A D | pmem.c | 137 dev_warn(dev, "could not reserve region [0x%pa:0x%zx]\n", &pmem->phys_addr, pmem->size); pmem_alloc()
|
/linux-4.1.27/arch/arm64/include/asm/ |
H A D | kvm_mmu.h | 92 phys_addr_t pa, unsigned long size, bool writable);
|
/linux-4.1.27/arch/cris/include/arch-v10/arch/ |
H A D | irq.h | 25 #define PA_IRQ_NBR IO_BITNR(R_VECT_MASK_RD, pa) /* 11 */
|
/linux-4.1.27/arch/alpha/mm/ |
H A D | init.c | 210 unsigned long pfn = crb->map[i].pa >> PAGE_SHIFT; callback_init()
|
/linux-4.1.27/arch/arm/kvm/ |
H A D | mmu.c | 961 * @pa: The physical address of the device 965 phys_addr_t pa, unsigned long size, bool writable) kvm_phys_addr_ioremap() 973 pfn = __phys_to_pfn(pa); kvm_phys_addr_ioremap() 1797 phys_addr_t pa; kvm_arch_prepare_memory_region() local 1799 pa = (phys_addr_t)vma->vm_pgoff << PAGE_SHIFT; kvm_arch_prepare_memory_region() 1800 pa += vm_start - vma->vm_start; kvm_arch_prepare_memory_region() 1806 ret = kvm_phys_addr_ioremap(kvm, gpa, pa, kvm_arch_prepare_memory_region() 964 kvm_phys_addr_ioremap(struct kvm *kvm, phys_addr_t guest_ipa, phys_addr_t pa, unsigned long size, bool writable) kvm_phys_addr_ioremap() argument
|