Home
last modified time | relevance | path

Searched refs:vaddr (Results 1 – 200 of 663) sorted by relevance

1234

/linux-4.1.27/arch/m68k/include/asm/
Dbitops.h31 static inline void bset_reg_set_bit(int nr, volatile unsigned long *vaddr) in bset_reg_set_bit() argument
33 char *p = (char *)vaddr + (nr ^ 31) / 8; in bset_reg_set_bit()
41 static inline void bset_mem_set_bit(int nr, volatile unsigned long *vaddr) in bset_mem_set_bit() argument
43 char *p = (char *)vaddr + (nr ^ 31) / 8; in bset_mem_set_bit()
50 static inline void bfset_mem_set_bit(int nr, volatile unsigned long *vaddr) in bfset_mem_set_bit() argument
54 : "d" (nr ^ 31), "o" (*vaddr) in bfset_mem_set_bit()
59 #define set_bit(nr, vaddr) bset_reg_set_bit(nr, vaddr) argument
61 #define set_bit(nr, vaddr) bset_mem_set_bit(nr, vaddr) argument
63 #define set_bit(nr, vaddr) (__builtin_constant_p(nr) ? \ argument
64 bset_mem_set_bit(nr, vaddr) : \
[all …]
Dpage_mm.h9 #define get_user_page(vaddr) __get_free_page(GFP_KERNEL) argument
58 #define clear_user_page(addr, vaddr, page) \ argument
62 #define copy_user_page(to, from, vaddr, page) \ argument
73 static inline unsigned long ___pa(void *vaddr) in ___pa() argument
80 : "0" (vaddr), "i" (m68k_fixup_memoffset)); in ___pa()
83 #define __pa(vaddr) ___pa((void *)(long)(vaddr)) argument
86 void *vaddr; in __va() local
90 : "=r" (vaddr) in __va()
92 return vaddr; in __va()
Dpage_no.h9 #define get_user_page(vaddr) __get_free_page(GFP_KERNEL) argument
15 #define clear_user_page(page, vaddr, pg) clear_page(page) argument
16 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument
18 #define __alloc_zeroed_user_highpage(movableflags, vma, vaddr) \ argument
19 alloc_page_vma(GFP_HIGHUSER | __GFP_ZERO | movableflags, vma, vaddr)
22 #define __pa(vaddr) ((unsigned long)(vaddr)) argument
Dcacheflush_mm.h155 extern void cache_push_v(unsigned long vaddr, int len);
222 static inline void __flush_page_to_ram(void *vaddr) in __flush_page_to_ram() argument
226 addr = ((unsigned long) vaddr) & ~(PAGE_SIZE - 1); in __flush_page_to_ram()
239 : : "a" (__pa(vaddr))); in __flush_page_to_ram()
261 struct page *page, unsigned long vaddr, in copy_to_user_page() argument
264 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_to_user_page()
266 flush_icache_user_range(vma, page, vaddr, len); in copy_to_user_page()
269 struct page *page, unsigned long vaddr, in copy_from_user_page() argument
272 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_from_user_page()
/linux-4.1.27/arch/arm/mm/
Dcache-xsc3l2.c100 unsigned long vaddr; in xsc3_l2_inv_range() local
107 vaddr = -1; /* to force the first mapping */ in xsc3_l2_inv_range()
113 vaddr = l2_map_va(start & ~(CACHE_LINE_SIZE - 1), vaddr); in xsc3_l2_inv_range()
114 xsc3_l2_clean_mva(vaddr); in xsc3_l2_inv_range()
115 xsc3_l2_inv_mva(vaddr); in xsc3_l2_inv_range()
123 vaddr = l2_map_va(start, vaddr); in xsc3_l2_inv_range()
124 xsc3_l2_inv_mva(vaddr); in xsc3_l2_inv_range()
132 vaddr = l2_map_va(start, vaddr); in xsc3_l2_inv_range()
133 xsc3_l2_clean_mva(vaddr); in xsc3_l2_inv_range()
134 xsc3_l2_inv_mva(vaddr); in xsc3_l2_inv_range()
[all …]
Dhighmem.c23 unsigned long vaddr = __fix_to_virt(idx); in set_fixmap_pte() local
24 pte_t *ptep = pte_offset_kernel(pmd_off_k(vaddr), vaddr); in set_fixmap_pte()
27 local_flush_tlb_kernel_page(vaddr); in set_fixmap_pte()
30 static inline pte_t get_fixmap_pte(unsigned long vaddr) in get_fixmap_pte() argument
32 pte_t *ptep = pte_offset_kernel(pmd_off_k(vaddr), vaddr); in get_fixmap_pte()
58 unsigned long vaddr; in kmap_atomic() local
82 vaddr = __fix_to_virt(idx); in kmap_atomic()
88 BUG_ON(!pte_none(get_fixmap_pte(vaddr))); in kmap_atomic()
97 return (void *)vaddr; in kmap_atomic()
103 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local
[all …]
Dcopypage-v6.c34 struct page *from, unsigned long vaddr, struct vm_area_struct *vma) in v6_copy_user_highpage_nonaliasing() argument
49 static void v6_clear_user_highpage_nonaliasing(struct page *page, unsigned long vaddr) in v6_clear_user_highpage_nonaliasing() argument
73 struct page *from, unsigned long vaddr, struct vm_area_struct *vma) in v6_copy_user_highpage_aliasing() argument
75 unsigned int offset = CACHE_COLOUR(vaddr); in v6_copy_user_highpage_aliasing()
106 static void v6_clear_user_highpage_aliasing(struct page *page, unsigned long vaddr) in v6_clear_user_highpage_aliasing() argument
108 unsigned long to = COPYPAGE_V6_TO + (CACHE_COLOUR(vaddr) << PAGE_SHIFT); in v6_clear_user_highpage_aliasing()
Dcopypage-feroceon.c71 unsigned long vaddr, struct vm_area_struct *vma) in feroceon_copy_user_highpage() argument
77 flush_cache_page(vma, vaddr, page_to_pfn(from)); in feroceon_copy_user_highpage()
83 void feroceon_clear_user_highpage(struct page *page, unsigned long vaddr) in feroceon_clear_user_highpage() argument
Dcopypage-xsc3.c74 unsigned long vaddr, struct vm_area_struct *vma) in xsc3_mc_copy_user_highpage() argument
80 flush_cache_page(vma, vaddr, page_to_pfn(from)); in xsc3_mc_copy_user_highpage()
91 void xsc3_mc_clear_user_highpage(struct page *page, unsigned long vaddr) in xsc3_mc_clear_user_highpage() argument
Dcopypage-v4wb.c51 unsigned long vaddr, struct vm_area_struct *vma) in v4wb_copy_user_highpage() argument
57 flush_cache_page(vma, vaddr, page_to_pfn(from)); in v4wb_copy_user_highpage()
68 void v4wb_clear_user_highpage(struct page *page, unsigned long vaddr) in v4wb_clear_user_highpage() argument
Dflush.c26 static void flush_pfn_alias(unsigned long pfn, unsigned long vaddr) in flush_pfn_alias() argument
28 unsigned long to = FLUSH_ALIAS_START + (CACHE_COLOUR(vaddr) << PAGE_SHIFT); in flush_pfn_alias()
40 static void flush_icache_alias(unsigned long pfn, unsigned long vaddr, unsigned long len) in flush_icache_alias() argument
42 unsigned long va = FLUSH_ALIAS_START + (CACHE_COLOUR(vaddr) << PAGE_SHIFT); in flush_icache_alias()
43 unsigned long offset = vaddr & (PAGE_SIZE - 1); in flush_icache_alias()
103 #define flush_pfn_alias(pfn,vaddr) do { } while (0) argument
104 #define flush_icache_alias(pfn,vaddr,len) do { } while (0) argument
/linux-4.1.27/arch/parisc/kernel/
Dpci-dma.c83 unsigned long vaddr, in map_pte_uncached() argument
87 unsigned long orig_vaddr = vaddr; in map_pte_uncached()
89 vaddr &= ~PMD_MASK; in map_pte_uncached()
90 end = vaddr + size; in map_pte_uncached()
102 vaddr += PAGE_SIZE; in map_pte_uncached()
106 } while (vaddr < end); in map_pte_uncached()
110 static inline int map_pmd_uncached(pmd_t * pmd, unsigned long vaddr, in map_pmd_uncached() argument
114 unsigned long orig_vaddr = vaddr; in map_pmd_uncached()
116 vaddr &= ~PGDIR_MASK; in map_pmd_uncached()
117 end = vaddr + size; in map_pmd_uncached()
[all …]
/linux-4.1.27/arch/sh/mm/
Dkmap.c18 #define kmap_get_fixmap_pte(vaddr) \ argument
19 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)), (vaddr)), (vaddr))
25 unsigned long vaddr; in kmap_coherent_init() local
28 vaddr = __fix_to_virt(FIX_CMAP_BEGIN); in kmap_coherent_init()
29 kmap_coherent_pte = kmap_get_fixmap_pte(vaddr); in kmap_coherent_init()
35 unsigned long vaddr; in kmap_coherent() local
45 vaddr = __fix_to_virt(idx); in kmap_coherent()
50 return (void *)vaddr; in kmap_coherent()
56 unsigned long vaddr = (unsigned long)kvaddr & PAGE_MASK; in kunmap_coherent() local
57 enum fixed_addresses idx = __virt_to_fix(vaddr); in kunmap_coherent()
[all …]
Dinit.c153 unsigned long vaddr, pte_t *lastpte) in page_table_kmap_check() argument
166 unsigned long vaddr; in page_table_range_init() local
168 vaddr = start; in page_table_range_init()
169 i = __pgd_offset(vaddr); in page_table_range_init()
170 j = __pud_offset(vaddr); in page_table_range_init()
171 k = __pmd_offset(vaddr); in page_table_range_init()
174 for ( ; (i < PTRS_PER_PGD) && (vaddr != end); pgd++, i++) { in page_table_range_init()
176 for ( ; (j < PTRS_PER_PUD) && (vaddr != end); pud++, j++) { in page_table_range_init()
181 for (; (k < PTRS_PER_PMD) && (vaddr != end); pmd++, k++) { in page_table_range_init()
183 pmd, vaddr, pte); in page_table_range_init()
[all …]
Dcache.c59 unsigned long vaddr, void *dst, const void *src, in copy_to_user_page() argument
64 void *vto = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_to_user_page()
74 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_to_user_page()
78 unsigned long vaddr, void *dst, const void *src, in copy_from_user_page() argument
83 void *vfrom = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_from_user_page()
94 unsigned long vaddr, struct vm_area_struct *vma) in copy_user_highpage() argument
102 vfrom = kmap_coherent(from, vaddr); in copy_user_highpage()
111 if (pages_do_alias((unsigned long)vto, vaddr & PAGE_MASK) || in copy_user_highpage()
121 void clear_user_highpage(struct page *page, unsigned long vaddr) in clear_user_highpage() argument
127 if (pages_do_alias((unsigned long)kaddr, vaddr & PAGE_MASK)) in clear_user_highpage()
Dpmb.c137 static bool pmb_mapping_exists(unsigned long vaddr, phys_addr_t phys, in pmb_mapping_exists() argument
156 if ((vaddr < pmbe->vpn) || (vaddr >= (pmbe->vpn + pmbe->size))) in pmb_mapping_exists()
335 int pmb_bolt_mapping(unsigned long vaddr, phys_addr_t phys, in pmb_bolt_mapping() argument
345 if (!pmb_addr_valid(vaddr, size)) in pmb_bolt_mapping()
347 if (pmb_mapping_exists(vaddr, phys, size)) in pmb_bolt_mapping()
350 orig_addr = vaddr; in pmb_bolt_mapping()
353 flush_tlb_kernel_range(vaddr, vaddr + size); in pmb_bolt_mapping()
363 pmbe = pmb_alloc(vaddr, phys, pmb_flags | in pmb_bolt_mapping()
377 vaddr += pmbe->size; in pmb_bolt_mapping()
413 unsigned long vaddr; in pmb_remap_caller() local
[all …]
Dcache-sh4.c215 void *vaddr; in sh4_flush_cache_page() local
236 vaddr = NULL; in sh4_flush_cache_page()
246 vaddr = kmap_coherent(page, address); in sh4_flush_cache_page()
248 vaddr = kmap_atomic(page); in sh4_flush_cache_page()
250 address = (unsigned long)vaddr; in sh4_flush_cache_page()
259 if (vaddr) { in sh4_flush_cache_page()
261 kunmap_coherent(vaddr); in sh4_flush_cache_page()
263 kunmap_atomic(vaddr); in sh4_flush_cache_page()
Dconsistent.c68 void *vaddr, dma_addr_t dma_handle, in dma_generic_free_coherent() argument
78 iounmap(vaddr); in dma_generic_free_coherent()
81 void dma_cache_sync(struct device *dev, void *vaddr, size_t size, in dma_cache_sync() argument
87 (void *)CAC_ADDR((unsigned long)vaddr) : vaddr; in dma_cache_sync()
Dioremap.c108 unsigned long vaddr = (unsigned long __force)addr; in __iounmap() local
114 if (iomapping_nontranslatable(vaddr)) in __iounmap()
129 p = remove_vm_area((void *)(vaddr & PAGE_MASK)); in __iounmap()
/linux-4.1.27/arch/microblaze/mm/
Dconsistent.c64 unsigned long order, vaddr; in consistent_alloc() local
82 vaddr = __get_free_pages(gfp, order); in consistent_alloc()
83 if (!vaddr) in consistent_alloc()
90 flush_dcache_range(virt_to_phys((void *)vaddr), in consistent_alloc()
91 virt_to_phys((void *)vaddr) + size); in consistent_alloc()
94 ret = (void *)vaddr; in consistent_alloc()
113 free_pages(vaddr, order); in consistent_alloc()
120 *dma_handle = pa = __virt_to_phys(vaddr); in consistent_alloc()
129 page = virt_to_page(vaddr); in consistent_alloc()
151 free_pages(vaddr, order); in consistent_alloc()
[all …]
Dhighmem.c37 unsigned long vaddr; in kmap_atomic_prot() local
48 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic_prot()
52 set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot)); in kmap_atomic_prot()
53 local_flush_tlb_page(NULL, vaddr); in kmap_atomic_prot()
55 return (void *) vaddr; in kmap_atomic_prot()
61 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local
64 if (vaddr < __fix_to_virt(FIX_KMAP_END)) { in __kunmap_atomic()
75 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); in __kunmap_atomic()
81 pte_clear(&init_mm, vaddr, kmap_pte-idx); in __kunmap_atomic()
82 local_flush_tlb_page(NULL, vaddr); in __kunmap_atomic()
/linux-4.1.27/arch/mips/mm/
Dhighmem.c47 unsigned long vaddr; in kmap_atomic() local
57 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic()
62 local_flush_tlb_one((unsigned long)vaddr); in kmap_atomic()
64 return (void*) vaddr; in kmap_atomic()
70 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local
73 if (vaddr < FIXADDR_START) { // FIXME in __kunmap_atomic()
83 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); in __kunmap_atomic()
89 pte_clear(&init_mm, vaddr, kmap_pte-idx); in __kunmap_atomic()
90 local_flush_tlb_one(vaddr); in __kunmap_atomic()
104 unsigned long vaddr; in kmap_atomic_pfn() local
[all …]
Dpgtable-32.c35 unsigned long vaddr; in pagetable_init() local
54 vaddr = __fix_to_virt(__end_of_fixed_addresses - 1) & PMD_MASK; in pagetable_init()
55 fixrange_init(vaddr, vaddr + FIXADDR_SIZE, pgd_base); in pagetable_init()
61 vaddr = PKMAP_BASE; in pagetable_init()
62 fixrange_init(vaddr, vaddr + PAGE_SIZE*LAST_PKMAP, pgd_base); in pagetable_init()
64 pgd = swapper_pg_dir + __pgd_offset(vaddr); in pagetable_init()
65 pud = pud_offset(pgd, vaddr); in pagetable_init()
66 pmd = pmd_offset(pud, vaddr); in pagetable_init()
67 pte = pte_offset_kernel(pmd, vaddr); in pagetable_init()
Dinit.c86 unsigned long vaddr, flags, entrylo; in __kmap_pgprot() local
96 vaddr = __fix_to_virt(FIX_CMAP_END - idx); in __kmap_pgprot()
106 write_c0_entryhi(vaddr & (PAGE_MASK << 1)); in __kmap_pgprot()
123 return (void*) vaddr; in __kmap_pgprot()
158 unsigned long vaddr, struct vm_area_struct *vma) in copy_user_highpage() argument
165 vfrom = kmap_coherent(from, vaddr); in copy_user_highpage()
174 pages_do_alias((unsigned long)vto, vaddr & PAGE_MASK)) in copy_user_highpage()
182 struct page *page, unsigned long vaddr, void *dst, const void *src, in copy_to_user_page() argument
187 void *vto = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_to_user_page()
196 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_to_user_page()
[all …]
Dpgtable-64.c97 unsigned long vaddr; in pagetable_init() local
109 vaddr = __fix_to_virt(__end_of_fixed_addresses - 1) & PMD_MASK; in pagetable_init()
110 fixrange_init(vaddr, vaddr + FIXADDR_SIZE, pgd_base); in pagetable_init()
/linux-4.1.27/drivers/media/v4l2-core/
Dvideobuf2-vmalloc.c25 void *vaddr; member
48 buf->vaddr = vmalloc_user(buf->size); in vb2_vmalloc_alloc()
54 if (!buf->vaddr) { in vb2_vmalloc_alloc()
69 vfree(buf->vaddr); in vb2_vmalloc_put()
74 static void *vb2_vmalloc_get_userptr(void *alloc_ctx, unsigned long vaddr, in vb2_vmalloc_get_userptr() argument
89 offset = vaddr & ~PAGE_MASK; in vb2_vmalloc_get_userptr()
93 vma = find_vma(current->mm, vaddr); in vb2_vmalloc_get_userptr()
95 if (vb2_get_contig_userptr(vaddr, size, &vma, &physp)) in vb2_vmalloc_get_userptr()
98 buf->vaddr = (__force void *)ioremap_nocache(physp, size); in vb2_vmalloc_get_userptr()
99 if (!buf->vaddr) in vb2_vmalloc_get_userptr()
[all …]
Dvideobuf-vmalloc.c105 __func__, i, mem->vaddr); in videobuf_vm_close()
107 vfree(mem->vaddr); in videobuf_vm_close()
108 mem->vaddr = NULL; in videobuf_vm_close()
173 if (!mem->vaddr) { in __videobuf_iolock()
192 mem->vaddr = vmalloc_user(pages); in __videobuf_iolock()
193 if (!mem->vaddr) { in __videobuf_iolock()
198 mem->vaddr, pages); in __videobuf_iolock()
257 mem->vaddr = vmalloc_user(pages); in __videobuf_mmap_mapper()
258 if (!mem->vaddr) { in __videobuf_mmap_mapper()
262 dprintk(1, "vmalloc is at addr %p (%d pages)\n", mem->vaddr, pages); in __videobuf_mmap_mapper()
[all …]
Dvideobuf2-dma-sg.c39 void *vaddr; member
120 buf->vaddr = NULL; in vb2_dma_sg_alloc()
192 if (buf->vaddr) in vb2_dma_sg_put()
193 vm_unmap_ram(buf->vaddr, buf->num_pages); in vb2_dma_sg_put()
232 static void *vb2_dma_sg_get_userptr(void *alloc_ctx, unsigned long vaddr, in vb2_dma_sg_get_userptr() argument
250 buf->vaddr = NULL; in vb2_dma_sg_get_userptr()
253 buf->offset = vaddr & ~PAGE_MASK; in vb2_dma_sg_get_userptr()
257 first = (vaddr & PAGE_MASK) >> PAGE_SHIFT; in vb2_dma_sg_get_userptr()
258 last = ((vaddr + size - 1) & PAGE_MASK) >> PAGE_SHIFT; in vb2_dma_sg_get_userptr()
266 vma = find_vma(current->mm, vaddr); in vb2_dma_sg_get_userptr()
[all …]
Dvideobuf-dma-contig.c28 void *vaddr; member
45 mem->vaddr = dma_alloc_coherent(dev, mem->size, in __videobuf_dc_alloc()
48 if (!mem->vaddr) { in __videobuf_dc_alloc()
53 dev_dbg(dev, "dma mapped data is at %p (%ld)\n", mem->vaddr, mem->size); in __videobuf_dc_alloc()
61 dma_free_coherent(dev, mem->size, mem->vaddr, mem->dma_handle); in __videobuf_dc_free()
63 mem->vaddr = NULL; in __videobuf_dc_free()
117 i, mem->vaddr); in videobuf_vm_close()
120 mem->vaddr = NULL; in videobuf_vm_close()
233 return mem->vaddr; in __videobuf_to_vaddr()
250 if (!mem->vaddr) { in __videobuf_iolock()
[all …]
Dvideobuf2-dma-contig.c30 void *vaddr; member
101 if (!buf->vaddr && buf->db_attach) in vb2_dc_vaddr()
102 buf->vaddr = dma_buf_vmap(buf->db_attach->dmabuf); in vb2_dc_vaddr()
104 return buf->vaddr; in vb2_dc_vaddr()
153 dma_free_coherent(buf->dev, buf->size, buf->vaddr, buf->dma_addr); in vb2_dc_put()
169 buf->vaddr = dma_alloc_coherent(dev, size, &buf->dma_addr, in vb2_dc_alloc()
171 if (!buf->vaddr) { in vb2_dc_alloc()
207 ret = dma_mmap_coherent(buf->dev, vma, buf->vaddr, in vb2_dc_mmap()
351 return buf->vaddr + pgnum * PAGE_SIZE; in vb2_dc_dmabuf_ops_kmap()
358 return buf->vaddr; in vb2_dc_dmabuf_ops_vmap()
[all …]
/linux-4.1.27/arch/sparc/mm/
Dio-unit.c94 static unsigned long iounit_get_area(struct iounit_struct *iounit, unsigned long vaddr, int size) in iounit_get_area() argument
100 npages = ((vaddr & ~PAGE_MASK) + size + (PAGE_SIZE-1)) >> PAGE_SHIFT; in iounit_get_area()
109 IOD(("iounit_get_area(%08lx,%d[%d])=", vaddr, size, npages)); in iounit_get_area()
124 panic("iounit_get_area: Couldn't find free iopte slots for (%08lx,%d)\n", vaddr, size); in iounit_get_area()
132 iopte = MKIOPTE(__pa(vaddr & PAGE_MASK)); in iounit_get_area()
133 vaddr = IOUNIT_DMA_BASE + (scan << PAGE_SHIFT) + (vaddr & ~PAGE_MASK); in iounit_get_area()
138 IOD(("%08lx\n", vaddr)); in iounit_get_area()
139 return vaddr; in iounit_get_area()
142 static __u32 iounit_get_scsi_one(struct device *dev, char *vaddr, unsigned long len) in iounit_get_scsi_one() argument
148 ret = iounit_get_area(iounit, (unsigned long)vaddr, len); in iounit_get_scsi_one()
[all …]
Dhighmem.c53 unsigned long vaddr; in kmap_atomic() local
63 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic()
67 __flush_cache_one(vaddr); in kmap_atomic()
78 __flush_tlb_one(vaddr); in kmap_atomic()
83 return (void*) vaddr; in kmap_atomic()
89 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local
92 if (vaddr < FIXADDR_START) { // FIXME in __kunmap_atomic()
104 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN+idx)); in __kunmap_atomic()
108 __flush_cache_one(vaddr); in __kunmap_atomic()
117 pte_clear(&init_mm, vaddr, kmap_pte-idx); in __kunmap_atomic()
[all …]
Dtlb.c69 static void tlb_batch_add_one(struct mm_struct *mm, unsigned long vaddr, in tlb_batch_add_one() argument
75 vaddr &= PAGE_MASK; in tlb_batch_add_one()
77 vaddr |= 0x1UL; in tlb_batch_add_one()
87 flush_tsb_user_page(mm, vaddr); in tlb_batch_add_one()
88 global_flush_tlb_page(mm, vaddr); in tlb_batch_add_one()
95 tb->vaddrs[nr] = vaddr; in tlb_batch_add_one()
104 void tlb_batch_add(struct mm_struct *mm, unsigned long vaddr, in tlb_batch_add() argument
126 if ((paddr ^ vaddr) & (1 << 13)) in tlb_batch_add()
132 tlb_batch_add_one(mm, vaddr, pte_exec(orig)); in tlb_batch_add()
136 static void tlb_batch_pmd_scan(struct mm_struct *mm, unsigned long vaddr, in tlb_batch_pmd_scan() argument
[all …]
Dleon_mm.c35 unsigned long leon_swprobe(unsigned long vaddr, unsigned long *paddr) in leon_swprobe() argument
88 ptr += ((((vaddr) >> LEON_PGD_SH) & LEON_PGD_M) * 4); in leon_swprobe()
111 ptr += (((vaddr >> LEON_PMD_SH) & LEON_PMD_M) * 4); in leon_swprobe()
139 ptr += (((vaddr >> LEON_PTE_SH) & LEON_PTE_M) * 4); in leon_swprobe()
160 (vaddr & ~(-1 << LEON_PTE_SH)) | ((pte & ~0xff) << 4); in leon_swprobe()
164 (vaddr & ~(-1 << LEON_PMD_SH)) | ((pte & ~0xff) << 4); in leon_swprobe()
168 (vaddr & ~(-1 << LEON_PGD_SH)) | ((pte & ~0xff) << 4); in leon_swprobe()
172 paddr_calc = vaddr; in leon_swprobe()
Diommu.c207 static u32 iommu_get_scsi_one(struct device *dev, char *vaddr, unsigned int len) in iommu_get_scsi_one() argument
214 off = (unsigned long)vaddr & ~PAGE_MASK; in iommu_get_scsi_one()
216 page = virt_to_page((unsigned long)vaddr & PAGE_MASK); in iommu_get_scsi_one()
221 static __u32 iommu_get_scsi_one_gflush(struct device *dev, char *vaddr, unsigned long len) in iommu_get_scsi_one_gflush() argument
224 return iommu_get_scsi_one(dev, vaddr, len); in iommu_get_scsi_one_gflush()
227 static __u32 iommu_get_scsi_one_pflush(struct device *dev, char *vaddr, unsigned long len) in iommu_get_scsi_one_pflush() argument
229 unsigned long page = ((unsigned long) vaddr) & PAGE_MASK; in iommu_get_scsi_one_pflush()
231 while(page < ((unsigned long)(vaddr + len))) { in iommu_get_scsi_one_pflush()
235 return iommu_get_scsi_one(dev, vaddr, len); in iommu_get_scsi_one_pflush()
299 static void iommu_release_scsi_one(struct device *dev, __u32 vaddr, unsigned long len) in iommu_release_scsi_one() argument
[all …]
/linux-4.1.27/arch/arm/mach-ixp4xx/include/mach/
Dio.h98 const u8 *vaddr, int count) in __indirect_writesb() argument
101 writeb(*vaddr++, bus_addr); in __indirect_writesb()
121 const u16 *vaddr, int count) in __indirect_writesw() argument
124 writew(*vaddr++, bus_addr); in __indirect_writesw()
140 const u32 *vaddr, int count) in __indirect_writesl() argument
143 writel(*vaddr++, bus_addr); in __indirect_writesl()
163 u8 *vaddr, u32 count) in __indirect_readsb() argument
166 *vaddr++ = readb(bus_addr); in __indirect_readsb()
186 u16 *vaddr, u32 count) in __indirect_readsw() argument
189 *vaddr++ = readw(bus_addr); in __indirect_readsw()
[all …]
/linux-4.1.27/arch/metag/mm/
Dhighmem.c43 unsigned long vaddr; in kmap_atomic() local
53 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic()
59 return (void *)vaddr; in kmap_atomic()
65 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local
78 pte_clear(&init_mm, vaddr, kmap_pte-idx); in __kunmap_atomic()
79 flush_tlb_kernel_range(vaddr, vaddr + PAGE_SIZE); in __kunmap_atomic()
95 unsigned long vaddr; in kmap_atomic_pfn() local
102 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic_pfn()
107 flush_tlb_kernel_range(vaddr, vaddr + PAGE_SIZE); in kmap_atomic_pfn()
109 return (void *)vaddr; in kmap_atomic_pfn()
[all …]
Dinit.c270 unsigned long vaddr; in allocate_pgtables() local
272 vaddr = start; in allocate_pgtables()
273 i = pgd_index(vaddr); in allocate_pgtables()
274 j = pmd_index(vaddr); in allocate_pgtables()
277 for ( ; (i < PTRS_PER_PGD) && (vaddr != end); pgd++, i++) { in allocate_pgtables()
279 for (; (j < PTRS_PER_PMD) && (vaddr != end); pmd++, j++) { in allocate_pgtables()
280 vaddr += PMD_SIZE; in allocate_pgtables()
294 unsigned long vaddr, end; in fixedrange_init() local
303 vaddr = __fix_to_virt(__end_of_fixed_addresses - 1) & PMD_MASK; in fixedrange_init()
305 allocate_pgtables(vaddr, end); in fixedrange_init()
[all …]
Dmmu-meta2.c18 unsigned long mmu_read_first_level_page(unsigned long vaddr) in mmu_read_first_level_page() argument
25 if (is_global_space(vaddr)) in mmu_read_first_level_page()
26 vaddr &= ~0x80000000; in mmu_read_first_level_page()
28 offset = vaddr >> PGDIR_SHIFT; in mmu_read_first_level_page()
53 unsigned long mmu_read_second_level_page(unsigned long vaddr) in mmu_read_second_level_page() argument
55 return __builtin_meta2_cacherd((void *)(vaddr & PAGE_MASK)); in mmu_read_second_level_page()
Dmmu-meta1.c111 unsigned long mmu_read_first_level_page(unsigned long vaddr) in mmu_read_first_level_page() argument
113 return metag_in32(pgd_entry_addr(vaddr)); in mmu_read_first_level_page()
116 unsigned long mmu_read_second_level_page(unsigned long vaddr) in mmu_read_second_level_page() argument
118 return metag_in32(pgtable_entry_addr(vaddr)); in mmu_read_second_level_page()
/linux-4.1.27/arch/m68k/sun3/
Ddvma.c23 static unsigned long dvma_page(unsigned long kaddr, unsigned long vaddr) in dvma_page() argument
36 if(ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] != pte) { in dvma_page()
37 sun3_put_pte(vaddr, pte); in dvma_page()
38 ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] = pte; in dvma_page()
41 return (vaddr + (kaddr & ~PAGE_MASK)); in dvma_page()
50 unsigned long vaddr; in dvma_map_iommu() local
52 vaddr = dvma_btov(baddr); in dvma_map_iommu()
54 end = vaddr + len; in dvma_map_iommu()
56 while(vaddr < end) { in dvma_map_iommu()
57 dvma_page(kaddr, vaddr); in dvma_map_iommu()
[all …]
Dmmu_emu.c117 void print_pte_vaddr (unsigned long vaddr) in print_pte_vaddr() argument
119 printk (" vaddr=%lx [%02lx]", vaddr, sun3_get_segmap (vaddr)); in print_pte_vaddr()
120 print_pte (__pte (sun3_get_pte (vaddr))); in print_pte_vaddr()
282 inline void mmu_emu_map_pmeg (int context, int vaddr) in mmu_emu_map_pmeg() argument
288 vaddr &= ~SUN3_PMEG_MASK; in mmu_emu_map_pmeg()
297 curr_pmeg, context, vaddr); in mmu_emu_map_pmeg()
309 if(vaddr >= PAGE_OFFSET) { in mmu_emu_map_pmeg()
315 sun3_put_segmap (vaddr, curr_pmeg); in mmu_emu_map_pmeg()
325 sun3_put_segmap (vaddr, curr_pmeg); in mmu_emu_map_pmeg()
328 pmeg_vaddr[curr_pmeg] = vaddr; in mmu_emu_map_pmeg()
[all …]
/linux-4.1.27/drivers/lguest/
Dpage_tables.c83 static pgd_t *spgd_addr(struct lg_cpu *cpu, u32 i, unsigned long vaddr) in spgd_addr() argument
85 unsigned int index = pgd_index(vaddr); in spgd_addr()
97 static pmd_t *spmd_addr(struct lg_cpu *cpu, pgd_t spgd, unsigned long vaddr) in spmd_addr() argument
99 unsigned int index = pmd_index(vaddr); in spmd_addr()
115 static pte_t *spte_addr(struct lg_cpu *cpu, pgd_t spgd, unsigned long vaddr) in spte_addr() argument
118 pmd_t *pmd = spmd_addr(cpu, spgd, vaddr); in spte_addr()
129 return &page[pte_index(vaddr)]; in spte_addr()
136 static unsigned long gpgd_addr(struct lg_cpu *cpu, unsigned long vaddr) in gpgd_addr() argument
138 unsigned int index = vaddr >> (PGDIR_SHIFT); in gpgd_addr()
144 static unsigned long gpmd_addr(pgd_t gpgd, unsigned long vaddr) in gpmd_addr() argument
[all …]
/linux-4.1.27/arch/tile/mm/
Dhighmem.c20 #define kmap_get_pte(vaddr) \ argument
21 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)),\
22 (vaddr)), (vaddr))
200 unsigned long vaddr; in kmap_atomic_prot() local
215 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic_prot()
216 pte = kmap_get_pte(vaddr); in kmap_atomic_prot()
220 kmap_atomic_register(page, type, vaddr, pte, mk_pte(page, prot)); in kmap_atomic_prot()
222 return (void *)vaddr; in kmap_atomic_prot()
235 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local
237 if (vaddr >= __fix_to_virt(FIX_KMAP_END) && in __kunmap_atomic()
[all …]
Dinit.c180 unsigned long vaddr; in page_table_range_init() local
183 for (vaddr = start; vaddr < end; vaddr += PMD_SIZE) { in page_table_range_init()
184 pmd_t *pmd = get_pmd(pgd, vaddr); in page_table_range_init()
643 unsigned long vaddr; in permanent_kmaps_init() local
645 vaddr = PKMAP_BASE; in permanent_kmaps_init()
646 page_table_range_init(vaddr, vaddr + PAGE_SIZE*LAST_PKMAP, pgd_base); in permanent_kmaps_init()
648 pgd = swapper_pg_dir + pgd_index(vaddr); in permanent_kmaps_init()
649 pud = pud_offset(pgd, vaddr); in permanent_kmaps_init()
650 pmd = pmd_offset(pud, vaddr); in permanent_kmaps_init()
651 pte = pte_offset_kernel(pmd, vaddr); in permanent_kmaps_init()
/linux-4.1.27/arch/m68k/sun3x/
Ddvma.c83 unsigned long vaddr, int len) in dvma_map_cpu() argument
90 vaddr &= PAGE_MASK; in dvma_map_cpu()
92 end = PAGE_ALIGN(vaddr + len); in dvma_map_cpu()
96 kaddr, vaddr); in dvma_map_cpu()
98 pgd = pgd_offset_k(vaddr); in dvma_map_cpu()
104 if((pmd = pmd_alloc(&init_mm, pgd, vaddr)) == NULL) { in dvma_map_cpu()
109 if((end & PGDIR_MASK) > (vaddr & PGDIR_MASK)) in dvma_map_cpu()
110 end2 = (vaddr + (PGDIR_SIZE-1)) & PGDIR_MASK; in dvma_map_cpu()
118 if((pte = pte_alloc_kernel(pmd, vaddr)) == NULL) { in dvma_map_cpu()
123 if((end2 & PMD_MASK) > (vaddr & PMD_MASK)) in dvma_map_cpu()
[all …]
/linux-4.1.27/mm/
Dhighmem.c154 struct page *kmap_to_page(void *vaddr) in kmap_to_page() argument
156 unsigned long addr = (unsigned long)vaddr; in kmap_to_page()
219 unsigned long vaddr; in map_new_virtual() local
261 vaddr = PKMAP_ADDR(last_pkmap_nr); in map_new_virtual()
262 set_pte_at(&init_mm, vaddr, in map_new_virtual()
266 set_page_address(page, (void *)vaddr); in map_new_virtual()
268 return vaddr; in map_new_virtual()
281 unsigned long vaddr; in kmap_high() local
288 vaddr = (unsigned long)page_address(page); in kmap_high()
289 if (!vaddr) in kmap_high()
[all …]
Ddmapool.c58 void *vaddr; member
217 *(int *)(page->vaddr + offset) = next; in pool_initialise_page()
229 page->vaddr = dma_alloc_coherent(pool->dev, pool->allocation, in pool_alloc_page()
231 if (page->vaddr) { in pool_alloc_page()
233 memset(page->vaddr, POOL_POISON_FREED, pool->allocation); in pool_alloc_page()
255 memset(page->vaddr, POOL_POISON_FREED, pool->allocation); in pool_free_page()
257 dma_free_coherent(pool->dev, pool->allocation, page->vaddr, dma); in pool_free_page()
292 pool->name, page->vaddr); in dma_pool_destroy()
296 pool->name, page->vaddr); in dma_pool_destroy()
347 page->offset = *(int *)(page->vaddr + offset); in dma_pool_alloc()
[all …]
/linux-4.1.27/arch/sparc/include/asm/
Dviking.h212 static inline unsigned long viking_hwprobe(unsigned long vaddr) in viking_hwprobe() argument
216 vaddr &= PAGE_MASK; in viking_hwprobe()
220 : "r" (vaddr | 0x400), "i" (ASI_M_FLUSH_PROBE)); in viking_hwprobe()
227 : "r" (vaddr | 0x200), "i" (ASI_M_FLUSH_PROBE)); in viking_hwprobe()
229 vaddr &= ~SRMMU_PGDIR_MASK; in viking_hwprobe()
230 vaddr >>= PAGE_SHIFT; in viking_hwprobe()
231 return val | (vaddr << 8); in viking_hwprobe()
237 : "r" (vaddr | 0x100), "i" (ASI_M_FLUSH_PROBE)); in viking_hwprobe()
239 vaddr &= ~SRMMU_REAL_PMD_MASK; in viking_hwprobe()
240 vaddr >>= PAGE_SHIFT; in viking_hwprobe()
[all …]
Dtlbflush_64.h19 void flush_tsb_user_page(struct mm_struct *mm, unsigned long vaddr);
48 void __flush_tlb_page(unsigned long context, unsigned long vaddr);
53 static inline void global_flush_tlb_page(struct mm_struct *mm, unsigned long vaddr) in global_flush_tlb_page() argument
55 __flush_tlb_page(CTX_HWBITS(mm->context), vaddr); in global_flush_tlb_page()
61 void smp_flush_tlb_page(struct mm_struct *mm, unsigned long vaddr);
63 #define global_flush_tlb_page(mm, vaddr) \ argument
64 smp_flush_tlb_page(mm, vaddr)
Dcacheflush_64.h57 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument
59 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
61 flush_ptrace_access(vma, page, vaddr, src, len, 0); \
64 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
66 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
68 flush_ptrace_access(vma, page, vaddr, dst, len, 1); \
Dcacheflush_32.h21 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument
23 flush_cache_page(vma, vaddr, page_to_pfn(page));\
26 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
28 flush_cache_page(vma, vaddr, page_to_pfn(page));\
Ddma.h110 #define mmu_get_scsi_one(dev,vaddr,len) \ argument
111 sparc32_dma_ops->get_scsi_one(dev, vaddr, len)
114 #define mmu_release_scsi_one(dev,vaddr,len) \ argument
115 sparc32_dma_ops->release_scsi_one(dev, vaddr,len)
Doplib_64.h155 unsigned long vaddr);
159 unsigned long vaddr);
175 unsigned long vaddr, unsigned long paddr);
176 void prom_unmap(unsigned long size, unsigned long vaddr);
/linux-4.1.27/arch/powerpc/mm/
Dhighmem.c34 unsigned long vaddr; in kmap_atomic_prot() local
44 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic_prot()
48 __set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot), 1); in kmap_atomic_prot()
49 local_flush_tlb_page(NULL, vaddr); in kmap_atomic_prot()
51 return (void*) vaddr; in kmap_atomic_prot()
57 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local
60 if (vaddr < __fix_to_virt(FIX_KMAP_END)) { in __kunmap_atomic()
72 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); in __kunmap_atomic()
78 pte_clear(&init_mm, vaddr, kmap_pte-idx); in __kunmap_atomic()
79 local_flush_tlb_page(NULL, vaddr); in __kunmap_atomic()
Dhash_utils_64.c185 unsigned long vaddr, paddr; in htab_bolt_mapping() local
197 for (vaddr = vstart, paddr = pstart; vaddr < vend; in htab_bolt_mapping()
198 vaddr += step, paddr += step) { in htab_bolt_mapping()
200 unsigned long vsid = get_kernel_vsid(vaddr, ssize); in htab_bolt_mapping()
201 unsigned long vpn = hpt_vpn(vaddr, vsid, ssize); in htab_bolt_mapping()
210 if (overlaps_kernel_text(vaddr, vaddr + step)) in htab_bolt_mapping()
214 if (overlaps_kvm_tmp(vaddr, vaddr + step)) in htab_bolt_mapping()
228 overlaps_interrupt_vector_text(vaddr, vaddr + step)) in htab_bolt_mapping()
252 unsigned long vaddr; in htab_remove_mapping() local
264 for (vaddr = vstart; vaddr < vend; vaddr += step) in htab_remove_mapping()
[all …]
Ddma-noncoherent.c218 unsigned long vaddr = c->vm_start; in __dma_alloc_coherent() local
230 map_page(vaddr, page_to_phys(page), in __dma_alloc_coherent()
233 vaddr += PAGE_SIZE; in __dma_alloc_coherent()
257 void __dma_free_coherent(size_t size, void *vaddr) in __dma_free_coherent() argument
266 c = ppc_vm_region_find(&consistent_head, (unsigned long)vaddr); in __dma_free_coherent()
309 __func__, vaddr); in __dma_free_coherent()
317 void __dma_sync(void *vaddr, size_t size, int direction) in __dma_sync() argument
319 unsigned long start = (unsigned long)vaddr; in __dma_sync()
Dmem.c72 static inline pte_t *virt_to_kpte(unsigned long vaddr) in virt_to_kpte() argument
74 return pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), in virt_to_kpte()
75 vaddr), vaddr), vaddr); in virt_to_kpte()
432 void clear_user_page(void *page, unsigned long vaddr, struct page *pg) in clear_user_page() argument
445 void copy_user_page(void *vto, void *vfrom, unsigned long vaddr, in copy_user_page() argument
/linux-4.1.27/arch/x86/mm/
Dinit_32.c108 pmd_t * __init populate_extra_pmd(unsigned long vaddr) in populate_extra_pmd() argument
110 int pgd_idx = pgd_index(vaddr); in populate_extra_pmd()
111 int pmd_idx = pmd_index(vaddr); in populate_extra_pmd()
116 pte_t * __init populate_extra_pte(unsigned long vaddr) in populate_extra_pte() argument
118 int pte_idx = pte_index(vaddr); in populate_extra_pte()
121 pmd = populate_extra_pmd(vaddr); in populate_extra_pte()
133 unsigned long vaddr; in page_table_range_init_count() local
138 vaddr = start; in page_table_range_init_count()
139 pgd_idx = pgd_index(vaddr); in page_table_range_init_count()
140 pmd_idx = pmd_index(vaddr); in page_table_range_init_count()
[all …]
Dhighmem_32.c35 unsigned long vaddr; in kmap_atomic_prot() local
46 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic_prot()
51 return (void *)vaddr; in kmap_atomic_prot()
73 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local
75 if (vaddr >= __fix_to_virt(FIX_KMAP_END) && in __kunmap_atomic()
76 vaddr <= __fix_to_virt(FIX_KMAP_BEGIN)) { in __kunmap_atomic()
83 WARN_ON_ONCE(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); in __kunmap_atomic()
91 kpte_clear_flush(kmap_pte-idx, vaddr); in __kunmap_atomic()
97 BUG_ON(vaddr < PAGE_OFFSET); in __kunmap_atomic()
98 BUG_ON(vaddr >= (unsigned long)high_memory); in __kunmap_atomic()
[all …]
Dpgtable_32.c27 void set_pte_vaddr(unsigned long vaddr, pte_t pteval) in set_pte_vaddr() argument
34 pgd = swapper_pg_dir + pgd_index(vaddr); in set_pte_vaddr()
39 pud = pud_offset(pgd, vaddr); in set_pte_vaddr()
44 pmd = pmd_offset(pud, vaddr); in set_pte_vaddr()
49 pte = pte_offset_kernel(pmd, vaddr); in set_pte_vaddr()
51 set_pte_at(&init_mm, vaddr, pte, pteval); in set_pte_vaddr()
53 pte_clear(&init_mm, vaddr, pte); in set_pte_vaddr()
59 __flush_tlb_one(vaddr); in set_pte_vaddr()
Diomap_32.c59 unsigned long vaddr; in kmap_atomic_prot_pfn() local
66 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic_prot_pfn()
70 return (void *)vaddr; in kmap_atomic_prot_pfn()
97 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in iounmap_atomic() local
99 if (vaddr >= __fix_to_virt(FIX_KMAP_END) && in iounmap_atomic()
100 vaddr <= __fix_to_virt(FIX_KMAP_BEGIN)) { in iounmap_atomic()
107 WARN_ON_ONCE(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); in iounmap_atomic()
115 kpte_clear_flush(kmap_pte-idx, vaddr); in iounmap_atomic()
Dioremap.c31 int ioremap_change_attr(unsigned long vaddr, unsigned long size, in ioremap_change_attr() argument
40 err = _set_memory_uc(vaddr, nrpages); in ioremap_change_attr()
43 err = _set_memory_wc(vaddr, nrpages); in ioremap_change_attr()
46 err = _set_memory_wb(vaddr, nrpages); in ioremap_change_attr()
85 unsigned long offset, vaddr; in __ioremap_caller() local
186 vaddr = (unsigned long) area->addr; in __ioremap_caller()
191 if (ioremap_page_range(vaddr, vaddr + size, phys_addr, prot)) in __ioremap_caller()
194 ret_addr = (void __iomem *) (vaddr + offset); in __ioremap_caller()
356 unsigned long vaddr; in xlate_dev_mem_ptr() local
362 vaddr = (unsigned long)ioremap_cache(start, PAGE_SIZE); in xlate_dev_mem_ptr()
[all …]
Dinit_64.c233 static pud_t *fill_pud(pgd_t *pgd, unsigned long vaddr) in fill_pud() argument
242 return pud_offset(pgd, vaddr); in fill_pud()
245 static pmd_t *fill_pmd(pud_t *pud, unsigned long vaddr) in fill_pmd() argument
254 return pmd_offset(pud, vaddr); in fill_pmd()
257 static pte_t *fill_pte(pmd_t *pmd, unsigned long vaddr) in fill_pte() argument
265 return pte_offset_kernel(pmd, vaddr); in fill_pte()
268 void set_pte_vaddr_pud(pud_t *pud_page, unsigned long vaddr, pte_t new_pte) in set_pte_vaddr_pud() argument
274 pud = pud_page + pud_index(vaddr); in set_pte_vaddr_pud()
275 pmd = fill_pmd(pud, vaddr); in set_pte_vaddr_pud()
276 pte = fill_pte(pmd, vaddr); in set_pte_vaddr_pud()
[all …]
Dpageattr.c32 unsigned long *vaddr; member
130 void clflush_cache_range(void *vaddr, unsigned int size) in clflush_cache_range() argument
132 void *vend = vaddr + size - 1; in clflush_cache_range()
136 for (; vaddr < vend; vaddr += boot_cpu_data.x86_clflush_size) in clflush_cache_range()
137 clflushopt(vaddr); in clflush_cache_range()
1090 static int __cpa_process_fault(struct cpa_data *cpa, unsigned long vaddr, in __cpa_process_fault() argument
1094 return populate_pgd(cpa, vaddr); in __cpa_process_fault()
1109 if (within(vaddr, PAGE_OFFSET, in __cpa_process_fault()
1112 cpa->pfn = __pa(vaddr) >> PAGE_SHIFT; in __cpa_process_fault()
1116 "vaddr = %lx cpa->vaddr = %lx\n", vaddr, in __cpa_process_fault()
[all …]
/linux-4.1.27/drivers/base/
Ddma-mapping.c22 void *vaddr; member
30 dma_free_coherent(dev, this->size, this->vaddr, this->dma_handle); in dmam_coherent_release()
37 dma_free_noncoherent(dev, this->size, this->vaddr, this->dma_handle); in dmam_noncoherent_release()
44 if (this->vaddr == match->vaddr) { in dmam_match()
69 void *vaddr; in dmam_alloc_coherent() local
75 vaddr = dma_alloc_coherent(dev, size, dma_handle, gfp); in dmam_alloc_coherent()
76 if (!vaddr) { in dmam_alloc_coherent()
81 dr->vaddr = vaddr; in dmam_alloc_coherent()
87 return vaddr; in dmam_alloc_coherent()
100 void dmam_free_coherent(struct device *dev, size_t size, void *vaddr, in dmam_free_coherent() argument
[all …]
Ddma-coherent.c213 int dma_release_from_coherent(struct device *dev, int order, void *vaddr) in dma_release_from_coherent() argument
217 if (mem && vaddr >= mem->virt_base && vaddr < in dma_release_from_coherent()
219 int page = (vaddr - mem->virt_base) >> PAGE_SHIFT; in dma_release_from_coherent()
247 void *vaddr, size_t size, int *ret) in dma_mmap_from_coherent() argument
251 if (mem && vaddr >= mem->virt_base && vaddr + size <= in dma_mmap_from_coherent()
254 int start = (vaddr - mem->virt_base) >> PAGE_SHIFT; in dma_mmap_from_coherent()
/linux-4.1.27/drivers/block/
Dcciss.h66 void __iomem *vaddr; member
226 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command()
227 readl(h->vaddr + SA5_SCRATCHPAD_OFFSET); in SA5_submit_command()
243 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask()
244 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask()
249 h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask()
250 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask()
263 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5B_intr_mask()
264 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5B_intr_mask()
269 h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5B_intr_mask()
[all …]
Dsmart1,2.h44 writel(c->busaddr, h->vaddr + S42XX_REQUEST_PORT_OFFSET); in smart4_submit_command()
56 writel(0, h->vaddr + S42XX_REPLY_INTR_MASK_OFFSET); in smart4_intr_mask()
60 h->vaddr + S42XX_REPLY_INTR_MASK_OFFSET); in smart4_intr_mask()
72 return (!readl(h->vaddr + S42XX_REQUEST_PORT_OFFSET)); in smart4_fifo_full()
82 = readl(h->vaddr + S42XX_REPLY_PORT_OFFSET); in smart4_completed()
90 writel(0, h->vaddr + S42XX_REPLY_PORT_OFFSET); in smart4_completed()
103 readl(h->vaddr + S42XX_INTR_STATUS); in smart4_intr_pending()
123 writel(c->busaddr, h->vaddr + COMMAND_FIFO); in smart2_submit_command()
128 writel(val, h->vaddr + INTR_MASK); in smart2_intr_mask()
133 return readl(h->vaddr + COMMAND_FIFO); in smart2_fifo_full()
[all …]
/linux-4.1.27/arch/frv/mm/
Dcache-page.c25 void *vaddr; in flush_dcache_page() local
29 vaddr = kmap_atomic_primary(page); in flush_dcache_page()
31 frv_dcache_writeback((unsigned long) vaddr, (unsigned long) vaddr + PAGE_SIZE); in flush_dcache_page()
33 kunmap_atomic_primary(vaddr); in flush_dcache_page()
53 void *vaddr; in flush_icache_user_range() local
57 vaddr = kmap_atomic_primary(page); in flush_icache_user_range()
59 start = (start & ~PAGE_MASK) | (unsigned long) vaddr; in flush_icache_user_range()
62 kunmap_atomic_primary(vaddr); in flush_icache_user_range()
Ddma-alloc.c141 void consistent_free(void *vaddr) in consistent_free() argument
145 vfree(vaddr); in consistent_free()
151 void consistent_sync(void *vaddr, size_t size, int direction) in consistent_sync() argument
153 unsigned long start = (unsigned long) vaddr; in consistent_sync()
/linux-4.1.27/arch/um/kernel/
Dmem.c94 unsigned long vaddr; in fixrange_init() local
96 vaddr = start; in fixrange_init()
97 i = pgd_index(vaddr); in fixrange_init()
98 j = pmd_index(vaddr); in fixrange_init()
101 for ( ; (i < PTRS_PER_PGD) && (vaddr < end); pgd++, i++) { in fixrange_init()
102 pud = pud_offset(pgd, vaddr); in fixrange_init()
105 pmd = pmd_offset(pud, vaddr); in fixrange_init()
106 for (; (j < PTRS_PER_PMD) && (vaddr < end); pmd++, j++) { in fixrange_init()
108 vaddr += PMD_SIZE; in fixrange_init()
123 unsigned long v, vaddr = FIXADDR_USER_START; in fixaddr_user_init() local
[all …]
/linux-4.1.27/arch/xtensa/mm/
Dcache.c64 unsigned long vaddr) in kmap_invalidate_coherent() argument
66 if (!DCACHE_ALIAS_EQ(page_to_phys(page), vaddr)) { in kmap_invalidate_coherent()
84 unsigned long vaddr, unsigned long *paddr) in coherent_kvaddr() argument
86 if (PageHighMem(page) || !DCACHE_ALIAS_EQ(page_to_phys(page), vaddr)) { in coherent_kvaddr()
88 return (void *)(base + (vaddr & DCACHE_ALIAS_MASK)); in coherent_kvaddr()
95 void clear_user_highpage(struct page *page, unsigned long vaddr) in clear_user_highpage() argument
98 void *kvaddr = coherent_kvaddr(page, TLBTEMP_BASE_1, vaddr, &paddr); in clear_user_highpage()
101 kmap_invalidate_coherent(page, vaddr); in clear_user_highpage()
108 unsigned long vaddr, struct vm_area_struct *vma) in copy_user_highpage() argument
111 void *dst_vaddr = coherent_kvaddr(dst, TLBTEMP_BASE_1, vaddr, in copy_user_highpage()
[all …]
Dmmu.c21 static void * __init init_pmd(unsigned long vaddr, unsigned long n_pages) in init_pmd() argument
23 pgd_t *pgd = pgd_offset_k(vaddr); in init_pmd()
24 pmd_t *pmd = pmd_offset(pgd, vaddr); in init_pmd()
31 __func__, vaddr, n_pages); in init_pmd()
Dhighmem.c43 unsigned long vaddr; in kmap_atomic() local
51 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic()
57 return (void *)vaddr; in kmap_atomic()
/linux-4.1.27/arch/avr32/include/asm/
Dcacheflush.h26 static inline void invalidate_dcache_line(void *vaddr) in invalidate_dcache_line() argument
30 : "r"(vaddr), "n"(CACHE_OP_DCACHE_INVALIDATE) in invalidate_dcache_line()
38 static inline void clean_dcache_line(void *vaddr) in clean_dcache_line() argument
42 : "r"(vaddr), "n"(CACHE_OP_DCACHE_CLEAN) in clean_dcache_line()
50 static inline void flush_dcache_line(void *vaddr) in flush_dcache_line() argument
54 : "r"(vaddr), "n"(CACHE_OP_DCACHE_CLEAN_INVAL) in flush_dcache_line()
62 static inline void invalidate_icache_line(void *vaddr) in invalidate_icache_line() argument
66 : "r"(vaddr), "n"(CACHE_OP_ICACHE_INVALIDATE) in invalidate_icache_line()
122 unsigned long vaddr, void *dst, const void *src,
126 struct page *page, unsigned long vaddr, void *dst, in copy_from_user_page() argument
/linux-4.1.27/arch/xtensa/include/asm/
Dfixmap.h66 static inline unsigned long virt_to_fix(const unsigned long vaddr) in virt_to_fix() argument
68 BUG_ON(vaddr >= FIXADDR_TOP || vaddr < FIXADDR_START); in virt_to_fix()
69 return __virt_to_fix(vaddr); in virt_to_fix()
74 #define kmap_get_fixmap_pte(vaddr) \ argument
76 pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)), (vaddr)), \
77 (vaddr) \
Dpage.h149 extern void clear_page_alias(void *vaddr, unsigned long paddr);
154 void clear_user_highpage(struct page *page, unsigned long vaddr);
157 unsigned long vaddr, struct vm_area_struct *vma);
159 # define clear_user_page(page, vaddr, pg) clear_page(page) argument
160 # define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument
/linux-4.1.27/arch/frv/include/asm/
Dvirtconvert.h24 #define phys_to_virt(vaddr) ((void *) ((unsigned long)(vaddr) + PAGE_OFFSET)) argument
25 #define virt_to_phys(vaddr) ((unsigned long) (vaddr) - PAGE_OFFSET) argument
29 #define phys_to_virt(vaddr) ((void *) (vaddr)) argument
30 #define virt_to_phys(vaddr) ((unsigned long) (vaddr)) argument
Dpage.h11 #define get_user_page(vaddr) __get_free_page(GFP_KERNEL) argument
17 #define clear_user_page(pgaddr, vaddr, page) memset((pgaddr), 0, PAGE_SIZE) argument
18 #define copy_user_page(vto, vfrom, vaddr, topg) memcpy((vto), (vfrom), PAGE_SIZE) argument
45 #define __pa(vaddr) virt_to_phys((void *) (unsigned long) (vaddr)) argument
Dpci.h29 extern void consistent_free(void *vaddr);
30 extern void consistent_sync(void *vaddr, size_t size, int direction);
39 void *vaddr, dma_addr_t dma_handle);
Dcacheflush.h96 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument
99 flush_icache_user_range((vma), (page), (vaddr), (len)); \
102 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
/linux-4.1.27/drivers/misc/sgi-gru/
Dgrufault.c61 struct vm_area_struct *gru_find_vma(unsigned long vaddr) in gru_find_vma() argument
65 vma = find_vma(current->mm, vaddr); in gru_find_vma()
66 if (vma && vma->vm_start <= vaddr && vma->vm_ops == &gru_vm_ops) in gru_find_vma()
79 static struct gru_thread_state *gru_find_lock_gts(unsigned long vaddr) in gru_find_lock_gts() argument
86 vma = gru_find_vma(vaddr); in gru_find_lock_gts()
88 gts = gru_find_thread_state(vma, TSID(vaddr, vma)); in gru_find_lock_gts()
96 static struct gru_thread_state *gru_alloc_locked_gts(unsigned long vaddr) in gru_alloc_locked_gts() argument
103 vma = gru_find_vma(vaddr); in gru_alloc_locked_gts()
107 gts = gru_alloc_thread_state(vma, TSID(vaddr, vma)); in gru_alloc_locked_gts()
191 unsigned long vaddr, int write, in non_atomic_pte_lookup() argument
[all …]
Dgruhandles.c152 unsigned long vaddr, unsigned long vaddrmask, in tgh_invalidate() argument
156 tgh->vaddr = vaddr; in tgh_invalidate()
170 unsigned long vaddr, int asid, int dirty, in tfh_write_only() argument
174 tfh->fillvaddr = vaddr; in tfh_write_only()
186 unsigned long vaddr, int asid, int dirty, in tfh_write_restart() argument
190 tfh->fillvaddr = vaddr; in tfh_write_restart()
/linux-4.1.27/arch/arm/kernel/
Dcrash_dump.c37 void *vaddr; in copy_oldmem_page() local
42 vaddr = ioremap(__pfn_to_phys(pfn), PAGE_SIZE); in copy_oldmem_page()
43 if (!vaddr) in copy_oldmem_page()
47 if (copy_to_user(buf, vaddr + offset, csize)) { in copy_oldmem_page()
48 iounmap(vaddr); in copy_oldmem_page()
52 memcpy(buf, vaddr + offset, csize); in copy_oldmem_page()
55 iounmap(vaddr); in copy_oldmem_page()
/linux-4.1.27/arch/x86/kernel/
Dcrash_dump_64.c29 void *vaddr; in copy_oldmem_page() local
34 vaddr = ioremap_cache(pfn << PAGE_SHIFT, PAGE_SIZE); in copy_oldmem_page()
35 if (!vaddr) in copy_oldmem_page()
39 if (copy_to_user(buf, vaddr + offset, csize)) { in copy_oldmem_page()
40 iounmap(vaddr); in copy_oldmem_page()
44 memcpy(buf, vaddr + offset, csize); in copy_oldmem_page()
47 iounmap(vaddr); in copy_oldmem_page()
Dcrash_dump_32.c53 void *vaddr; in copy_oldmem_page() local
61 vaddr = kmap_atomic_pfn(pfn); in copy_oldmem_page()
64 memcpy(buf, (vaddr + offset), csize); in copy_oldmem_page()
65 kunmap_atomic(vaddr); in copy_oldmem_page()
70 kunmap_atomic(vaddr); in copy_oldmem_page()
73 copy_page(kdump_buf_page, vaddr); in copy_oldmem_page()
74 kunmap_atomic(vaddr); in copy_oldmem_page()
Dpci-swiotlb.c21 void *vaddr; in x86_swiotlb_alloc_coherent() local
23 vaddr = dma_generic_alloc_coherent(hwdev, size, dma_handle, flags, in x86_swiotlb_alloc_coherent()
25 if (vaddr) in x86_swiotlb_alloc_coherent()
26 return vaddr; in x86_swiotlb_alloc_coherent()
32 void *vaddr, dma_addr_t dma_addr, in x86_swiotlb_free_coherent() argument
36 swiotlb_free_coherent(dev, size, vaddr, dma_addr); in x86_swiotlb_free_coherent()
38 dma_generic_free_coherent(dev, size, vaddr, dma_addr, attrs); in x86_swiotlb_free_coherent()
Dtboot.c123 static int map_tboot_page(unsigned long vaddr, unsigned long pfn, in map_tboot_page() argument
131 pgd = pgd_offset(&tboot_mm, vaddr); in map_tboot_page()
132 pud = pud_alloc(&tboot_mm, pgd, vaddr); in map_tboot_page()
135 pmd = pmd_alloc(&tboot_mm, pud, vaddr); in map_tboot_page()
138 pte = pte_alloc_map(&tboot_mm, NULL, pmd, vaddr); in map_tboot_page()
141 set_pte_at(&tboot_mm, vaddr, pte, pfn_pte(pfn, prot)); in map_tboot_page()
146 static int map_tboot_pages(unsigned long vaddr, unsigned long start_pfn, in map_tboot_pages() argument
154 for (; nr > 0; nr--, vaddr += PAGE_SIZE, start_pfn++) { in map_tboot_pages()
155 if (map_tboot_page(vaddr, start_pfn, PAGE_KERNEL_EXEC)) in map_tboot_pages()
Dmachine_kexec_32.c104 unsigned long vaddr, unsigned long paddr) in machine_kexec_page_table_set_one() argument
108 pgd += pgd_index(vaddr); in machine_kexec_page_table_set_one()
113 pud = pud_offset(pgd, vaddr); in machine_kexec_page_table_set_one()
114 pmd = pmd_offset(pud, vaddr); in machine_kexec_page_table_set_one()
117 pte = pte_offset_kernel(pmd, vaddr); in machine_kexec_page_table_set_one()
/linux-4.1.27/arch/arc/mm/
Dcache_arc700.c202 static inline void __cache_line_loop(unsigned long paddr, unsigned long vaddr, in __cache_line_loop() argument
232 vaddr &= CACHE_LINE_MASK; in __cache_line_loop()
239 paddr |= (vaddr >> PAGE_SHIFT) & 0x1F; in __cache_line_loop()
254 write_aux_reg(aux_cmd, vaddr); in __cache_line_loop()
255 vaddr += L1_CACHE_BYTES; in __cache_line_loop()
326 static inline void __dc_line_op(unsigned long paddr, unsigned long vaddr, in __dc_line_op() argument
336 __cache_line_loop(paddr, vaddr, sz, cacheop); in __dc_line_op()
346 #define __dc_line_op(paddr, vaddr, sz, cacheop) argument
407 __ic_line_inv_vaddr_local(unsigned long paddr, unsigned long vaddr, in __ic_line_inv_vaddr_local() argument
413 __cache_line_loop(paddr, vaddr, sz, OP_INV_IC); in __ic_line_inv_vaddr_local()
[all …]
Dioremap.c44 void __iomem *vaddr; in ioremap_prot() local
73 vaddr = (void __iomem *)area->addr; in ioremap_prot()
74 if (ioremap_page_range((unsigned long)vaddr, in ioremap_prot()
75 (unsigned long)vaddr + size, paddr, prot)) { in ioremap_prot()
76 vunmap((void __force *)vaddr); in ioremap_prot()
79 return (void __iomem *)(off + (char __iomem *)vaddr); in ioremap_prot()
/linux-4.1.27/arch/parisc/include/asm/
Dcacheflush.h51 static inline void flush_kernel_vmap_range(void *vaddr, int size) in flush_kernel_vmap_range() argument
53 unsigned long start = (unsigned long)vaddr; in flush_kernel_vmap_range()
57 static inline void invalidate_kernel_vmap_range(void *vaddr, int size) in invalidate_kernel_vmap_range() argument
59 unsigned long start = (unsigned long)vaddr; in invalidate_kernel_vmap_range()
60 void *cursor = vaddr; in invalidate_kernel_vmap_range()
62 for ( ; cursor < vaddr + size; cursor += PAGE_SIZE) { in invalidate_kernel_vmap_range()
92 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument
94 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
99 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
101 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
[all …]
Ddma-mapping.h13 void (*free_consistent)(struct device *dev, size_t size, void *vaddr, dma_addr_t iova);
68 void *vaddr, dma_addr_t dma_handle) in dma_free_coherent() argument
70 hppa_dma_ops->free_consistent(dev, size, vaddr, dma_handle); in dma_free_coherent()
75 void *vaddr, dma_addr_t dma_handle) in dma_free_noncoherent() argument
77 hppa_dma_ops->free_consistent(dev, size, vaddr, dma_handle); in dma_free_noncoherent()
191 dma_cache_sync(struct device *dev, void *vaddr, size_t size, in dma_cache_sync() argument
195 flush_kernel_dcache_range((unsigned long)vaddr, size); in dma_cache_sync()
/linux-4.1.27/arch/m68k/mm/
Dcache.c14 static unsigned long virt_to_phys_slow(unsigned long vaddr) in virt_to_phys_slow() argument
36 : "0" (vaddr)); in virt_to_phys_slow()
46 : "a" (vaddr)); in virt_to_phys_slow()
49 return (mmusr & PAGE_MASK) | (vaddr & ~PAGE_MASK); in virt_to_phys_slow()
57 : "a" (vaddr), "d" (get_fs().seg)); in virt_to_phys_slow()
63 return (*descaddr & 0xfe000000) | (vaddr & 0x01ffffff); in virt_to_phys_slow()
65 return (*descaddr & 0xfffc0000) | (vaddr & 0x0003ffff); in virt_to_phys_slow()
67 return (*descaddr & PAGE_MASK) | (vaddr & ~PAGE_MASK); in virt_to_phys_slow()
/linux-4.1.27/arch/sh/kernel/
Dcrash_dump.c28 void *vaddr; in copy_oldmem_page() local
33 vaddr = ioremap(pfn << PAGE_SHIFT, PAGE_SIZE); in copy_oldmem_page()
36 if (copy_to_user(buf, (vaddr + offset), csize)) { in copy_oldmem_page()
37 iounmap(vaddr); in copy_oldmem_page()
41 memcpy(buf, (vaddr + offset), csize); in copy_oldmem_page()
43 iounmap(vaddr); in copy_oldmem_page()
/linux-4.1.27/arch/mn10300/include/asm/
Dhighmem.h75 unsigned long vaddr; in kmap_atomic() local
84 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic()
90 local_flush_tlb_one(vaddr); in kmap_atomic()
92 return (void *)vaddr; in kmap_atomic()
95 static inline void __kunmap_atomic(unsigned long vaddr) in __kunmap_atomic() argument
99 if (vaddr < FIXADDR_START) { /* FIXME */ in __kunmap_atomic()
111 if (vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)) in __kunmap_atomic()
119 local_flush_tlb_one(vaddr); in __kunmap_atomic()
/linux-4.1.27/arch/powerpc/kernel/
Dcrash_dump.c72 static size_t copy_oldmem_vaddr(void *vaddr, char *buf, size_t csize, in copy_oldmem_vaddr() argument
76 if (copy_to_user((char __user *)buf, (vaddr + offset), csize)) in copy_oldmem_vaddr()
79 memcpy(buf, (vaddr + offset), csize); in copy_oldmem_vaddr()
100 void *vaddr; in copy_oldmem_page() local
110 vaddr = __va(paddr); in copy_oldmem_page()
111 csize = copy_oldmem_vaddr(vaddr, buf, csize, offset, userbuf); in copy_oldmem_page()
113 vaddr = __ioremap(paddr, PAGE_SIZE, 0); in copy_oldmem_page()
114 csize = copy_oldmem_vaddr(vaddr, buf, csize, offset, userbuf); in copy_oldmem_page()
115 iounmap(vaddr); in copy_oldmem_page()
Dio-workarounds.c28 static struct iowa_bus *iowa_pci_find(unsigned long vaddr, unsigned long paddr) in iowa_pci_find() argument
38 if (vaddr) { in iowa_pci_find()
41 if ((vaddr >= vstart) && (vaddr <= vend)) in iowa_pci_find()
68 unsigned long vaddr, paddr; in iowa_mem_find_bus() local
71 vaddr = (unsigned long)PCI_FIX_ADDR(addr); in iowa_mem_find_bus()
72 if (vaddr < PHB_IO_BASE || vaddr >= PHB_IO_END) in iowa_mem_find_bus()
78 ptep = __find_linux_pte_or_hugepte(init_mm.pgd, vaddr, in iowa_mem_find_bus()
86 bus = iowa_pci_find(vaddr, paddr); in iowa_mem_find_bus()
104 unsigned long vaddr = (unsigned long)pci_io_base + port; in iowa_pio_find_bus() local
105 return iowa_pci_find(vaddr, 0); in iowa_pio_find_bus()
Diommu.c460 unsigned long vaddr, npages, entry, slen; in ppc_iommu_map_sg() local
469 vaddr = (unsigned long) sg_virt(s); in ppc_iommu_map_sg()
470 npages = iommu_num_pages(vaddr, slen, IOMMU_PAGE_SIZE(tbl)); in ppc_iommu_map_sg()
473 (vaddr & ~PAGE_MASK) == 0) in ppc_iommu_map_sg()
478 DBG(" - vaddr: %lx, size: %lx\n", vaddr, slen); in ppc_iommu_map_sg()
484 "vaddr %lx npages %lu\n", tbl, vaddr, in ppc_iommu_map_sg()
499 vaddr & IOMMU_PAGE_MASK(tbl), in ppc_iommu_map_sg()
559 unsigned long vaddr, npages; in ppc_iommu_map_sg() local
561 vaddr = s->dma_address & IOMMU_PAGE_MASK(tbl); in ppc_iommu_map_sg()
564 __iommu_free(tbl, vaddr, npages); in ppc_iommu_map_sg()
[all …]
/linux-4.1.27/drivers/scsi/
Dhpsa.h128 void __iomem *vaddr; member
359 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command()
360 (void) readl(h->vaddr + SA5_SCRATCHPAD_OFFSET); in SA5_submit_command()
366 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command_no_read()
372 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command_ioaccel2()
384 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask()
385 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask()
389 h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask()
390 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask()
398 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_performant_intr_mask()
[all …]
Dsun3_scsi.c402 unsigned char *vaddr; in sun3scsi_dma_finish() local
404 vaddr = (unsigned char *)dvma_vmetov(sun3_dma_orig_addr); in sun3scsi_dma_finish()
406 vaddr += (sun3_dma_orig_count - fifo); in sun3scsi_dma_finish()
407 vaddr--; in sun3scsi_dma_finish()
411 *vaddr = (dregs->bpack_lo & 0xff00) >> 8; in sun3scsi_dma_finish()
412 vaddr--; in sun3scsi_dma_finish()
415 *vaddr = (dregs->bpack_hi & 0x00ff); in sun3scsi_dma_finish()
416 vaddr--; in sun3scsi_dma_finish()
419 *vaddr = (dregs->bpack_hi & 0xff00) >> 8; in sun3scsi_dma_finish()
448 unsigned char *vaddr; in sun3scsi_dma_finish() local
[all …]
DNCR_Q720.c79 int irq, int slot, __u32 paddr, void __iomem *vaddr) in NCR_Q720_probe_one() argument
84 __u8 scsr1 = readb(vaddr + NCR_Q720_SCSR_OFFSET + 1); in NCR_Q720_probe_one()
85 __u8 differential = readb(vaddr + NCR_Q720_SCSR_OFFSET) & 0x20; in NCR_Q720_probe_one()
94 writeb(scsr1, vaddr + NCR_Q720_SCSR_OFFSET + 1); in NCR_Q720_probe_one()
96 version = readb(vaddr + 0x18) >> 4; in NCR_Q720_probe_one()
106 device.slot.base_v = vaddr; in NCR_Q720_probe_one()
118 scsr1 = readb(vaddr + NCR_Q720_SCSR_OFFSET + 1); in NCR_Q720_probe_one()
121 writeb(scsr1, vaddr + NCR_Q720_SCSR_OFFSET + 1); in NCR_Q720_probe_one()
/linux-4.1.27/include/linux/
Dhighmem.h23 static inline void flush_kernel_vmap_range(void *vaddr, int size) in flush_kernel_vmap_range() argument
26 static inline void invalidate_kernel_vmap_range(void *vaddr, int size) in invalidate_kernel_vmap_range() argument
132 static inline void clear_user_highpage(struct page *page, unsigned long vaddr) in clear_user_highpage() argument
135 clear_user_page(addr, vaddr, page); in clear_user_highpage()
158 unsigned long vaddr) in __alloc_zeroed_user_highpage() argument
161 vma, vaddr); in __alloc_zeroed_user_highpage()
164 clear_user_highpage(page, vaddr); in __alloc_zeroed_user_highpage()
180 unsigned long vaddr) in alloc_zeroed_user_highpage_movable() argument
182 return __alloc_zeroed_user_highpage(__GFP_MOVABLE, vma, vaddr); in alloc_zeroed_user_highpage_movable()
225 unsigned long vaddr, struct vm_area_struct *vma) in copy_user_highpage() argument
[all …]
Dio-mapping.h97 io_mapping_unmap_atomic(void __iomem *vaddr) in io_mapping_unmap_atomic() argument
99 iounmap_atomic(vaddr); in io_mapping_unmap_atomic()
114 io_mapping_unmap(void __iomem *vaddr) in io_mapping_unmap() argument
116 iounmap(vaddr); in io_mapping_unmap()
149 io_mapping_unmap_atomic(void __iomem *vaddr) in io_mapping_unmap_atomic() argument
162 io_mapping_unmap(void __iomem *vaddr) in io_mapping_unmap() argument
Duprobes.h79 unsigned long vaddr; member
101 extern int set_swbp(struct arch_uprobe *aup, struct mm_struct *mm, unsigned long vaddr);
102 extern int set_orig_insn(struct arch_uprobe *aup, struct mm_struct *mm, unsigned long vaddr);
107 extern int uprobe_write_opcode(struct mm_struct *mm, unsigned long vaddr, uprobe_opcode_t);
132 extern void arch_uprobe_copy_ixol(struct page *page, unsigned long vaddr,
/linux-4.1.27/arch/mips/kernel/
Dcrash_dump.c29 void *vaddr; in copy_oldmem_page() local
34 vaddr = kmap_atomic_pfn(pfn); in copy_oldmem_page()
37 memcpy(buf, (vaddr + offset), csize); in copy_oldmem_page()
38 kunmap_atomic(vaddr); in copy_oldmem_page()
45 copy_page(kdump_buf_page, vaddr); in copy_oldmem_page()
46 kunmap_atomic(vaddr); in copy_oldmem_page()
Dmips-r2-to-r6-emul.c907 unsigned long vaddr; in mipsr2_decoder() local
1200 vaddr = regs->regs[MIPSInst_RS(inst)] + MIPSInst_SIMM(inst); in mipsr2_decoder()
1201 if (!access_ok(VERIFY_READ, vaddr, 4)) { in mipsr2_decoder()
1202 current->thread.cp0_baduaddr = vaddr; in mipsr2_decoder()
1261 "+&r"(vaddr), "+&r"(err) in mipsr2_decoder()
1273 vaddr = regs->regs[MIPSInst_RS(inst)] + MIPSInst_SIMM(inst); in mipsr2_decoder()
1274 if (!access_ok(VERIFY_READ, vaddr, 4)) { in mipsr2_decoder()
1275 current->thread.cp0_baduaddr = vaddr; in mipsr2_decoder()
1336 "+&r"(vaddr), "+&r"(err) in mipsr2_decoder()
1347 vaddr = regs->regs[MIPSInst_RS(inst)] + MIPSInst_SIMM(inst); in mipsr2_decoder()
[all …]
/linux-4.1.27/arch/hexagon/include/asm/
Dfixmap.h31 #define kmap_get_fixmap_pte(vaddr) \ argument
32 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), \
33 (vaddr)), (vaddr)), (vaddr))
/linux-4.1.27/arch/arm/include/asm/
Dpage.h114 void (*cpu_clear_user_highpage)(struct page *page, unsigned long vaddr);
116 unsigned long vaddr, struct vm_area_struct *vma);
130 extern void __cpu_clear_user_highpage(struct page *page, unsigned long vaddr);
132 unsigned long vaddr, struct vm_area_struct *vma);
135 #define clear_user_highpage(page,vaddr) \ argument
136 __cpu_clear_user_highpage(page, vaddr)
139 #define copy_user_highpage(to,from,vaddr,vma) \ argument
140 __cpu_copy_user_highpage(to, from, vaddr, vma)
Dpage-nommu.h20 #define get_user_page(vaddr) __get_free_page(GFP_KERNEL) argument
26 #define clear_user_page(page, vaddr, pg) clear_page(page) argument
27 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument
/linux-4.1.27/arch/arm64/include/asm/
Ddma-mapping.h117 void *vaddr; in dma_alloc_attrs() local
119 if (dma_alloc_from_coherent(dev, size, dma_handle, &vaddr)) in dma_alloc_attrs()
120 return vaddr; in dma_alloc_attrs()
122 vaddr = ops->alloc(dev, size, dma_handle, flags, attrs); in dma_alloc_attrs()
123 debug_dma_alloc_coherent(dev, size, *dma_handle, vaddr); in dma_alloc_attrs()
124 return vaddr; in dma_alloc_attrs()
128 void *vaddr, dma_addr_t dev_addr, in dma_free_attrs() argument
133 if (dma_release_from_coherent(dev, get_order(size), vaddr)) in dma_free_attrs()
136 debug_dma_free_coherent(dev, size, vaddr, dev_addr); in dma_free_attrs()
137 ops->free(dev, size, vaddr, dev_addr, attrs); in dma_free_attrs()
Dpage.h59 #define clear_user_page(addr,vaddr,pg) __cpu_clear_user_page(addr, vaddr) argument
60 #define copy_user_page(to,from,vaddr,pg) __cpu_copy_user_page(to, from, vaddr) argument
/linux-4.1.27/arch/frv/mb93090-mb00/
Dpci-dma.c34 void dma_free_coherent(struct device *hwdev, size_t size, void *vaddr, dma_addr_t dma_handle) in dma_free_coherent() argument
36 consistent_free(vaddr); in dma_free_coherent()
57 void *vaddr; in dma_map_sg() local
65 vaddr = kmap_atomic_primary(sg_page(&sg[i])); in dma_map_sg()
67 frv_dcache_writeback((unsigned long) vaddr, in dma_map_sg()
68 (unsigned long) vaddr + PAGE_SIZE); in dma_map_sg()
72 kunmap_atomic_primary(vaddr); in dma_map_sg()
/linux-4.1.27/arch/score/include/asm/
Dpage.h36 #define clear_user_page(pgaddr, vaddr, page) memset((pgaddr), 0, PAGE_SIZE) argument
37 #define copy_user_page(vto, vfrom, vaddr, topg) \ argument
67 #define virt_to_pfn(vaddr) (phys_to_pfn((__pa(vaddr)))) argument
70 #define virt_to_page(vaddr) (pfn_to_page(virt_to_pfn(vaddr))) argument
83 #define virt_addr_valid(vaddr) (pfn_valid(virt_to_pfn(vaddr))) argument
Dcacheflush.h38 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
41 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument
45 flush_cache_page(vma, vaddr, page_to_pfn(page));\
Dtlbflush.h23 extern void local_flush_tlb_one(unsigned long vaddr);
32 #define flush_tlb_one(vaddr) local_flush_tlb_one(vaddr) argument
/linux-4.1.27/arch/nios2/include/asm/
Ddma-mapping.h17 static inline void __dma_sync_for_device(void *vaddr, size_t size, in __dma_sync_for_device() argument
22 invalidate_dcache_range((unsigned long)vaddr, in __dma_sync_for_device()
23 (unsigned long)(vaddr + size)); in __dma_sync_for_device()
31 flush_dcache_range((unsigned long)vaddr, in __dma_sync_for_device()
32 (unsigned long)(vaddr + size)); in __dma_sync_for_device()
39 static inline void __dma_sync_for_cpu(void *vaddr, size_t size, in __dma_sync_for_cpu() argument
45 invalidate_dcache_range((unsigned long)vaddr, in __dma_sync_for_cpu()
46 (unsigned long)(vaddr + size)); in __dma_sync_for_cpu()
62 void *vaddr, dma_addr_t dma_handle);
125 static inline void dma_cache_sync(struct device *dev, void *vaddr, size_t size, in dma_cache_sync() argument
Dpage.h53 extern void clear_user_page(void *addr, unsigned long vaddr, struct page *page);
54 extern void copy_user_page(void *vto, void *vfrom, unsigned long vaddr,
93 # define virt_to_page(vaddr) pfn_to_page(PFN_DOWN(virt_to_phys(vaddr))) argument
94 # define virt_addr_valid(vaddr) pfn_valid(PFN_DOWN(virt_to_phys(vaddr))) argument
Dio.h55 #define phys_to_virt(vaddr) \ argument
56 ((void *)((unsigned long)(vaddr) | CONFIG_NIOS2_KERNEL_REGION_BASE))
58 #define virt_to_phys(vaddr) \ argument
59 ((unsigned long)((unsigned long)(vaddr) & ~0xE0000000))
/linux-4.1.27/arch/s390/include/asm/
Didals.h29 idal_is_needed(void *vaddr, unsigned int length) in idal_is_needed() argument
31 return ((__pa(vaddr) + length - 1) >> 31) != 0; in idal_is_needed()
38 static inline unsigned int idal_nr_words(void *vaddr, unsigned int length) in idal_nr_words() argument
40 return ((__pa(vaddr) & (IDA_BLOCK_SIZE-1)) + length + in idal_nr_words()
48 void *vaddr, unsigned int length) in idal_create_words() argument
53 paddr = __pa(vaddr); in idal_create_words()
70 set_normalized_cda(struct ccw1 * ccw, void *vaddr) in set_normalized_cda() argument
77 nridaws = idal_nr_words(vaddr, ccw->count); in set_normalized_cda()
83 idal_create_words(idal, vaddr, ccw->count); in set_normalized_cda()
85 vaddr = idal; in set_normalized_cda()
[all …]
Dpage.h59 #define clear_user_page(page, vaddr, pg) clear_page(page) argument
60 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument
62 #define __alloc_zeroed_user_highpage(movableflags, vma, vaddr) \ argument
63 alloc_page_vma(GFP_HIGHUSER | __GFP_ZERO | movableflags, vma, vaddr)
/linux-4.1.27/kernel/events/
Duprobes.c118 unsigned long vaddr; /* Page(s) of instruction slots */ member
144 static loff_t vaddr_to_offset(struct vm_area_struct *vma, unsigned long vaddr) in vaddr_to_offset() argument
146 return ((loff_t)vma->vm_pgoff << PAGE_SHIFT) + (vaddr - vma->vm_start); in vaddr_to_offset()
241 static void copy_from_page(struct page *page, unsigned long vaddr, void *dst, int len) in copy_from_page() argument
244 memcpy(dst, kaddr + (vaddr & ~PAGE_MASK), len); in copy_from_page()
248 static void copy_to_page(struct page *page, unsigned long vaddr, const void *src, int len) in copy_to_page() argument
251 memcpy(kaddr + (vaddr & ~PAGE_MASK), src, len); in copy_to_page()
255 static int verify_opcode(struct page *page, unsigned long vaddr, uprobe_opcode_t *new_opcode) in verify_opcode() argument
269 copy_from_page(page, vaddr, &old_opcode, UPROBE_SWBP_INSN_SIZE); in verify_opcode()
300 int uprobe_write_opcode(struct mm_struct *mm, unsigned long vaddr, in uprobe_write_opcode() argument
[all …]
/linux-4.1.27/arch/metag/include/asm/
Dfixmap.h56 #define kmap_get_fixmap_pte(vaddr) \ argument
58 pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)), (vaddr)), \
59 (vaddr) \
/linux-4.1.27/arch/mips/include/asm/
Dfixmap.h72 #define kmap_get_fixmap_pte(vaddr) \ argument
73 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)), (vaddr)), (vaddr))
Dcacheflush.h101 struct page *page, unsigned long vaddr, void *dst, const void *src,
105 struct page *page, unsigned long vaddr, void *dst, const void *src,
135 extern void (*__flush_kernel_vmap_range)(unsigned long vaddr, int size);
137 static inline void flush_kernel_vmap_range(void *vaddr, int size) in flush_kernel_vmap_range() argument
140 __flush_kernel_vmap_range((unsigned long) vaddr, size); in flush_kernel_vmap_range()
143 static inline void invalidate_kernel_vmap_range(void *vaddr, int size) in invalidate_kernel_vmap_range() argument
146 __flush_kernel_vmap_range((unsigned long) vaddr, size); in invalidate_kernel_vmap_range()
Dtlbflush.h23 extern void local_flush_tlb_one(unsigned long vaddr);
33 extern void flush_tlb_one(unsigned long vaddr);
43 #define flush_tlb_one(vaddr) local_flush_tlb_one(vaddr) argument
Ddma-mapping.h65 extern void dma_cache_sync(struct device *dev, void *vaddr, size_t size,
87 void *vaddr, dma_addr_t dma_handle, in dma_free_attrs() argument
92 ops->free(dev, size, vaddr, dma_handle, attrs); in dma_free_attrs()
94 debug_dma_free_coherent(dev, size, vaddr, dma_handle); in dma_free_attrs()
102 void *vaddr, dma_addr_t dma_handle);
/linux-4.1.27/arch/mips/boot/
Delf2ecoff.c58 unsigned long vaddr; member
105 if (base->vaddr + base->len != new->vaddr) { in combine()
107 base->len = new->vaddr - base->vaddr; in combine()
282 text.vaddr = data.vaddr = bss.vaddr = 0; in main()
362 ndata.vaddr = ph[i].p_vaddr; in main()
364 nbss.vaddr = ph[i].p_vaddr + ph[i].p_filesz; in main()
372 ntxt.vaddr = ph[i].p_vaddr; in main()
392 if (text.vaddr > data.vaddr || data.vaddr > bss.vaddr || in main()
393 text.vaddr + text.len > data.vaddr in main()
394 || data.vaddr + data.len > bss.vaddr) { in main()
[all …]
/linux-4.1.27/arch/powerpc/platforms/44x/
Dcanyonlands.c69 void __iomem *vaddr; in ppc460ex_canyonlands_fixup() local
94 vaddr = of_iomap(np, 0); in ppc460ex_canyonlands_fixup()
97 if (!vaddr) { in ppc460ex_canyonlands_fixup()
117 setbits32((vaddr + GPIO0_OSRH), 0x42000000); in ppc460ex_canyonlands_fixup()
118 setbits32((vaddr + GPIO0_TSRH), 0x42000000); in ppc460ex_canyonlands_fixup()
120 iounmap(vaddr); in ppc460ex_canyonlands_fixup()
/linux-4.1.27/arch/sh/boards/mach-sdk7786/
Dsram.c25 void __iomem *vaddr; in fpga_sram_init() local
53 vaddr = ioremap(phys, SZ_2K); in fpga_sram_init()
54 if (unlikely(!vaddr)) { in fpga_sram_init()
63 ret = gen_pool_add(sram_pool, (unsigned long)vaddr, SZ_2K, -1); in fpga_sram_init()
66 iounmap(vaddr); in fpga_sram_init()
/linux-4.1.27/arch/mips/dec/
Dkn01-berr.c83 long asid, entryhi, vaddr; in dec_kn01_be_backend() local
105 vaddr = regs->regs[insn.i_format.rs] + in dec_kn01_be_backend()
108 vaddr = (long)pc; in dec_kn01_be_backend()
109 if (KSEGX(vaddr) == CKSEG0 || KSEGX(vaddr) == CKSEG1) in dec_kn01_be_backend()
110 address = CPHYSADDR(vaddr); in dec_kn01_be_backend()
115 entryhi |= vaddr & ~(PAGE_SIZE - 1); in dec_kn01_be_backend()
123 offset = vaddr & (PAGE_SIZE - 1); in dec_kn01_be_backend()
/linux-4.1.27/drivers/xen/xenbus/
Dxenbus_client.c73 void **vaddr);
74 int (*unmap)(struct xenbus_device *dev, void *vaddr);
375 int xenbus_grant_ring(struct xenbus_device *dev, void *vaddr, in xenbus_grant_ring() argument
382 unsigned long addr = (unsigned long)vaddr + in xenbus_grant_ring()
465 unsigned int nr_grefs, void **vaddr) in xenbus_map_ring_valloc() argument
467 return ring_ops->map(dev, gnt_refs, nr_grefs, vaddr); in xenbus_map_ring_valloc()
539 void **vaddr) in xenbus_map_ring_valloc_pv() argument
549 *vaddr = NULL; in xenbus_map_ring_valloc_pv()
581 *vaddr = area->addr; in xenbus_map_ring_valloc_pv()
597 void **vaddr) in xenbus_map_ring_valloc_hvm() argument
[all …]
/linux-4.1.27/arch/arc/include/asm/
Dcacheflush.h35 void __sync_icache_dcache(unsigned long paddr, unsigned long vaddr, int len);
36 void __inv_icache_page(unsigned long paddr, unsigned long vaddr);
37 void ___flush_dcache_page(unsigned long paddr, unsigned long vaddr);
110 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument
114 __sync_icache_dcache((unsigned long)(dst), vaddr, len); \
117 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
Dpage.h16 #define get_user_page(vaddr) __get_free_page(GFP_KERNEL) argument
91 #define __pa(vaddr) ((unsigned long)vaddr) argument
/linux-4.1.27/sound/pci/asihpi/
Dhpios.c50 p_mem_area->vaddr = in hpios_locked_mem_alloc()
54 if (p_mem_area->vaddr) { in hpios_locked_mem_alloc()
57 p_mem_area->vaddr); in hpios_locked_mem_alloc()
73 p_mem_area->vaddr, p_mem_area->dma_handle); in hpios_locked_mem_free()
77 p_mem_area->vaddr); in hpios_locked_mem_free()
/linux-4.1.27/arch/ia64/kernel/
Dcrash_dump.c37 void *vaddr; in copy_oldmem_page() local
41 vaddr = __va(pfn<<PAGE_SHIFT); in copy_oldmem_page()
43 if (copy_to_user(buf, (vaddr + offset), csize)) { in copy_oldmem_page()
47 memcpy(buf, (vaddr + offset), csize); in copy_oldmem_page()
/linux-4.1.27/arch/metag/kernel/
Ddma.c225 unsigned long vaddr = c->vm_start; in dma_alloc_coherent() local
226 pte_t *pte = consistent_pte + CONSISTENT_OFFSET(vaddr); in dma_alloc_coherent()
241 set_pte_at(&init_mm, vaddr, in dma_alloc_coherent()
247 vaddr += PAGE_SIZE; in dma_alloc_coherent()
272 void *vaddr, dma_addr_t dma_handle) in dma_free_coherent() argument
282 c = metag_vm_region_find(&consistent_head, (unsigned long)vaddr); in dma_free_coherent()
329 __func__, vaddr); in dma_free_coherent()
426 void dma_sync_for_device(void *vaddr, size_t size, int dma_direction) in dma_sync_for_device() argument
444 flush_dcache_region(vaddr, size); in dma_sync_for_device()
452 writeback_dcache_region(vaddr, size); in dma_sync_for_device()
[all …]
Dtcm.c51 unsigned long vaddr; in tcm_alloc() local
58 vaddr = gen_pool_alloc(pool->pool, len); in tcm_alloc()
59 if (!vaddr) in tcm_alloc()
62 return vaddr; in tcm_alloc()
/linux-4.1.27/arch/xtensa/kernel/
Dpci-dma.c68 void *vaddr, dma_addr_t dma_handle) in dma_free_coherent() argument
70 unsigned long addr = (unsigned long)vaddr + in dma_free_coherent()
81 void consistent_sync(void *vaddr, size_t size, int direction) in consistent_sync() argument
87 __invalidate_dcache_range((unsigned long)vaddr, in consistent_sync()
93 __flush_invalidate_dcache_range((unsigned long)vaddr, in consistent_sync()
/linux-4.1.27/arch/sparc/prom/
Dmisc_64.c203 unsigned long tte_data, unsigned long vaddr) in tlb_load() argument
212 args[5] = vaddr; in tlb_load()
224 unsigned long vaddr) in prom_itlb_load() argument
226 return tlb_load("SUNW,itlb-load", index, tte_data, vaddr); in prom_itlb_load()
231 unsigned long vaddr) in prom_dtlb_load() argument
233 return tlb_load("SUNW,dtlb-load", index, tte_data, vaddr); in prom_dtlb_load()
237 unsigned long vaddr, unsigned long paddr) in prom_map() argument
249 args[7] = vaddr; in prom_map()
262 void prom_unmap(unsigned long size, unsigned long vaddr) in prom_unmap() argument
272 args[6] = vaddr; in prom_unmap()
/linux-4.1.27/drivers/infiniband/hw/ipath/
Dipath_keys.c143 isge->vaddr = (void *) sge->addr; in ipath_lkey_ok()
176 isge->vaddr = mr->map[m]->segs[n].vaddr + off; in ipath_lkey_ok()
200 u32 len, u64 vaddr, u32 rkey, int acc) in ipath_rkey_ok() argument
223 sge->vaddr = (void *) vaddr; in ipath_rkey_ok()
239 off = vaddr - mr->iova; in ipath_rkey_ok()
240 if (unlikely(vaddr < mr->iova || off + len > mr->length || in ipath_rkey_ok()
258 sge->vaddr = mr->map[m]->segs[n].vaddr + off; in ipath_rkey_ok()
/linux-4.1.27/arch/microblaze/include/asm/
Dpage.h81 # define clear_user_page(pgaddr, vaddr, page) memset((pgaddr), 0, PAGE_SIZE) argument
82 # define copy_user_page(vto, vfrom, vaddr, topg) \ argument
147 # define virt_to_pfn(vaddr) (phys_to_pfn((__pa(vaddr)))) argument
157 # define virt_to_page(vaddr) (pfn_to_page(virt_to_pfn(vaddr))) argument
175 #define virt_addr_valid(vaddr) (pfn_valid(virt_to_pfn(vaddr))) argument
Dhighmem.h81 unsigned long idx, vaddr = (unsigned long) ptr; in kmap_atomic_to_page() local
84 if (vaddr < FIXADDR_START) in kmap_atomic_to_page()
87 idx = virt_to_fix(vaddr); in kmap_atomic_to_page()
/linux-4.1.27/arch/openrisc/include/asm/
Dfixmap.h81 static inline unsigned long virt_to_fix(const unsigned long vaddr) in virt_to_fix() argument
83 BUG_ON(vaddr >= FIXADDR_TOP || vaddr < FIXADDR_START); in virt_to_fix()
84 return __virt_to_fix(vaddr); in virt_to_fix()
Dpage.h43 #define get_user_page(vaddr) __get_free_page(GFP_KERNEL) argument
49 #define clear_user_page(page, vaddr, pg) clear_page(page) argument
50 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument
/linux-4.1.27/drivers/staging/android/ion/
Dion_test.c68 void *vaddr = vmap(&page, 1, VM_MAP, pgprot); in ion_handle_test_dma() local
72 if (!vaddr) { in ion_handle_test_dma()
78 ret = copy_from_user(vaddr + offset, ptr, to_copy); in ion_handle_test_dma()
80 ret = copy_to_user(ptr, vaddr + offset, to_copy); in ion_handle_test_dma()
82 vunmap(vaddr); in ion_handle_test_dma()
118 void *vaddr = dma_buf_kmap(dma_buf, page_offset); in ion_handle_test_kernel() local
120 if (!vaddr) in ion_handle_test_kernel()
126 ret = copy_from_user(vaddr + copy_offset, ptr, to_copy); in ion_handle_test_kernel()
128 ret = copy_to_user(ptr, vaddr + copy_offset, to_copy); in ion_handle_test_kernel()
130 dma_buf_kunmap(dma_buf, page_offset, vaddr); in ion_handle_test_kernel()
/linux-4.1.27/arch/mn10300/mm/
Dpgtable.c35 void set_pmd_pfn(unsigned long vaddr, unsigned long pfn, pgprot_t flags) in set_pmd_pfn() argument
41 if (vaddr & (PMD_SIZE-1)) { /* vaddr is misaligned */ in set_pmd_pfn()
49 pgd = swapper_pg_dir + pgd_index(vaddr); in set_pmd_pfn()
54 pud = pud_offset(pgd, vaddr); in set_pmd_pfn()
55 pmd = pmd_offset(pud, vaddr); in set_pmd_pfn()
61 local_flush_tlb_one(vaddr); in set_pmd_pfn()
/linux-4.1.27/arch/nios2/mm/
Dtlb.c77 unsigned long vaddr = CONFIG_NIOS2_IO_REGION_BASE + in flush_tlb_one_pid() local
81 vaddr, way, (pid_misc >> TLBMISC_PID_SHIFT)); in flush_tlb_one_pid()
83 WRCTL(CTL_PTEADDR, (vaddr >> 12) << 2); in flush_tlb_one_pid()
139 unsigned long vaddr = CONFIG_NIOS2_IO_REGION_BASE + in flush_tlb_one() local
144 vaddr, way, (pid_misc >> TLBMISC_PID_SHIFT)); in flush_tlb_one()
148 WRCTL(CTL_PTEADDR, (vaddr >> 12) << 2); in flush_tlb_one()
247 unsigned long vaddr = CONFIG_NIOS2_IO_REGION_BASE; in flush_tlb_all() local
260 WRCTL(CTL_PTEADDR, ((vaddr) >> PAGE_SHIFT) << 2); in flush_tlb_all()
263 vaddr += 1UL << 12; in flush_tlb_all()
Dcacheflush.c246 void copy_user_page(void *vto, void *vfrom, unsigned long vaddr, in copy_user_page() argument
249 __flush_dcache(vaddr, vaddr + PAGE_SIZE); in copy_user_page()
250 __flush_icache(vaddr, vaddr + PAGE_SIZE); in copy_user_page()
256 void clear_user_page(void *addr, unsigned long vaddr, struct page *page) in clear_user_page() argument
258 __flush_dcache(vaddr, vaddr + PAGE_SIZE); in clear_user_page()
259 __flush_icache(vaddr, vaddr + PAGE_SIZE); in clear_user_page()
/linux-4.1.27/arch/hexagon/mm/
Dvm_tlb.c49 void flush_tlb_one(unsigned long vaddr) in flush_tlb_one() argument
51 __vmclrmap((void *)vaddr, PAGE_SIZE); in flush_tlb_one()
78 void flush_tlb_page(struct vm_area_struct *vma, unsigned long vaddr) in flush_tlb_page() argument
83 __vmclrmap((void *)vaddr, PAGE_SIZE); in flush_tlb_page()
/linux-4.1.27/arch/s390/kvm/
Dgaccess.c543 union vaddress vaddr = {.addr = gva}; in guest_translate() local
559 if (vaddr.rfx01 > asce.tl) in guest_translate()
561 ptr += vaddr.rfx * 8; in guest_translate()
564 if (vaddr.rfx) in guest_translate()
566 if (vaddr.rsx01 > asce.tl) in guest_translate()
568 ptr += vaddr.rsx * 8; in guest_translate()
571 if (vaddr.rfx || vaddr.rsx) in guest_translate()
573 if (vaddr.rtx01 > asce.tl) in guest_translate()
575 ptr += vaddr.rtx * 8; in guest_translate()
578 if (vaddr.rfx || vaddr.rsx || vaddr.rtx) in guest_translate()
[all …]
/linux-4.1.27/drivers/net/wireless/ath/ath6kl/
Dtarget.h333 #define AR6003_VTOP(vaddr) ((vaddr) & 0x001fffff) argument
334 #define AR6004_VTOP(vaddr) (vaddr) argument
336 #define TARG_VTOP(target_type, vaddr) \ argument
337 (((target_type) == TARGET_TYPE_AR6003) ? AR6003_VTOP(vaddr) : \
338 (((target_type) == TARGET_TYPE_AR6004) ? AR6004_VTOP(vaddr) : 0))
/linux-4.1.27/arch/sh/include/asm/
Ddma-mapping.h39 void dma_cache_sync(struct device *dev, void *vaddr, size_t size,
79 void *vaddr, dma_addr_t dma_handle, in dma_free_attrs() argument
84 if (dma_release_from_coherent(dev, get_order(size), vaddr)) in dma_free_attrs()
87 debug_dma_free_coherent(dev, size, vaddr, dma_handle); in dma_free_attrs()
89 ops->free(dev, size, vaddr, dma_handle, attrs); in dma_free_attrs()
97 void *vaddr, dma_addr_t dma_handle,
/linux-4.1.27/arch/powerpc/platforms/512x/
Dmpc512x_shared.c226 void __iomem *vaddr; in mpc512x_init_diu() local
250 vaddr = ioremap(desc, sizeof(struct diu_ad)); in mpc512x_init_diu()
251 if (!vaddr) { in mpc512x_init_diu()
255 memcpy(&diu_shared_fb.ad0, vaddr, sizeof(struct diu_ad)); in mpc512x_init_diu()
261 pix_fmt = in_le32(vaddr); in mpc512x_init_diu()
263 diu_shared_fb.fb_phys = in_le32(vaddr + 4); in mpc512x_init_diu()
266 iounmap(vaddr); in mpc512x_init_diu()
269 vaddr = ioremap(desc, sizeof(diu_shared_fb.gamma)); in mpc512x_init_diu()
270 if (!vaddr) { in mpc512x_init_diu()
275 memcpy(&diu_shared_fb.gamma, vaddr, sizeof(diu_shared_fb.gamma)); in mpc512x_init_diu()
[all …]
/linux-4.1.27/arch/arm/probes/uprobes/
Dcore.c33 unsigned long vaddr) in set_swbp() argument
35 return uprobe_write_opcode(mm, vaddr, in set_swbp()
116 void arch_uprobe_copy_ixol(struct page *page, unsigned long vaddr, in arch_uprobe_copy_ixol() argument
120 void *dst = xol_page_kaddr + (vaddr & ~PAGE_MASK); in arch_uprobe_copy_ixol()
128 flush_uprobe_xol_access(page, vaddr, dst, len); in arch_uprobe_copy_ixol()
157 regs->ARM_pc = utask->vaddr + 4; in arch_uprobe_post_xol()
178 instruction_pointer_set(regs, utask->vaddr); in arch_uprobe_abort_xol()
/linux-4.1.27/arch/x86/include/asm/
Dpage.h24 static inline void clear_user_page(void *page, unsigned long vaddr, in clear_user_page() argument
30 static inline void copy_user_page(void *to, void *from, unsigned long vaddr, in copy_user_page() argument
36 #define __alloc_zeroed_user_highpage(movableflags, vma, vaddr) \ argument
37 alloc_page_vma(GFP_HIGHUSER | __GFP_ZERO | movableflags, vma, vaddr)
Ddma-mapping.h68 void *vaddr, dma_addr_t dma_addr,
97 dma_cache_sync(struct device *dev, void *vaddr, size_t size, in dma_cache_sync() argument
161 void *vaddr, dma_addr_t bus, in dma_free_attrs() argument
168 if (dma_release_from_coherent(dev, get_order(size), vaddr)) in dma_free_attrs()
171 debug_dma_free_coherent(dev, size, vaddr, bus); in dma_free_attrs()
173 ops->free(dev, size, vaddr, bus, attrs); in dma_free_attrs()
Dpgtable_32.h60 #define kpte_clear_flush(ptep, vaddr) \ argument
62 pte_clear(&init_mm, (vaddr), (ptep)); \
63 __flush_tlb_one((vaddr)); \
/linux-4.1.27/drivers/acpi/apei/
Dghes.c153 unsigned long vaddr; in ghes_ioremap_pfn_nmi() local
155 vaddr = (unsigned long)GHES_IOREMAP_NMI_PAGE(ghes_ioremap_area->addr); in ghes_ioremap_pfn_nmi()
156 ioremap_page_range(vaddr, vaddr + PAGE_SIZE, in ghes_ioremap_pfn_nmi()
159 return (void __iomem *)vaddr; in ghes_ioremap_pfn_nmi()
164 unsigned long vaddr; in ghes_ioremap_pfn_irq() local
166 vaddr = (unsigned long)GHES_IOREMAP_IRQ_PAGE(ghes_ioremap_area->addr); in ghes_ioremap_pfn_irq()
167 ioremap_page_range(vaddr, vaddr + PAGE_SIZE, in ghes_ioremap_pfn_irq()
170 return (void __iomem *)vaddr; in ghes_ioremap_pfn_irq()
175 unsigned long vaddr = (unsigned long __force)vaddr_ptr; in ghes_iounmap_nmi() local
178 BUG_ON(vaddr != (unsigned long)GHES_IOREMAP_NMI_PAGE(base)); in ghes_iounmap_nmi()
[all …]
/linux-4.1.27/arch/alpha/mm/
Dinit.c193 unsigned long vaddr; in callback_init() local
205 vaddr = (unsigned long)console_remap_vm.addr; in callback_init()
211 crb->map[i].va = vaddr; in callback_init()
216 if (pmd != pmd_offset(pgd, vaddr)) { in callback_init()
218 pmd = pmd_offset(pgd, vaddr); in callback_init()
222 set_pte(pte_offset_kernel(pmd, vaddr), in callback_init()
225 vaddr += PAGE_SIZE; in callback_init()
/linux-4.1.27/arch/x86/xen/
Dgrant-table.c121 void *vaddr; in xlated_setup_gnttab_pages() local
147 vaddr = vmap(pages, nr_grant_frames, 0, PAGE_KERNEL); in xlated_setup_gnttab_pages()
148 if (!vaddr) { in xlated_setup_gnttab_pages()
160 xen_auto_xlat_grant_frames.vaddr = vaddr; in xlated_setup_gnttab_pages()
Dmmu.c126 unsigned long arbitrary_virt_to_mfn(void *vaddr) in arbitrary_virt_to_mfn() argument
128 xmaddr_t maddr = arbitrary_virt_to_machine(vaddr); in arbitrary_virt_to_mfn()
133 xmaddr_t arbitrary_virt_to_machine(void *vaddr) in arbitrary_virt_to_machine() argument
135 unsigned long address = (unsigned long)vaddr; in arbitrary_virt_to_machine()
144 if (virt_addr_valid(vaddr)) in arbitrary_virt_to_machine()
145 return virt_to_machine(vaddr); in arbitrary_virt_to_machine()
156 void make_lowmem_page_readonly(void *vaddr) in make_lowmem_page_readonly() argument
159 unsigned long address = (unsigned long)vaddr; in make_lowmem_page_readonly()
172 void make_lowmem_page_readwrite(void *vaddr) in make_lowmem_page_readwrite() argument
175 unsigned long address = (unsigned long)vaddr; in make_lowmem_page_readwrite()
[all …]
/linux-4.1.27/drivers/i2c/busses/
Di2c-ibm_iic.c85 volatile struct iic_regs __iomem *iic = dev->vaddr; in dump_iic_regs()
129 out_8(&dev->vaddr->intmsk, enable ? INTRMSK_EIMTC : 0); in iic_interrupt_mode()
137 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_dev_init()
182 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_dev_reset()
243 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_smbus_quick()
329 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_handler()
347 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_xfer_result()
380 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_abort_xfer()
412 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_wait_for_tc()
463 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_xfer_bytes()
[all …]
/linux-4.1.27/drivers/video/fbdev/matrox/
Dmatroxfb_base.h128 void __iomem* vaddr; member
132 return readb(va.vaddr + offs); in mga_readb()
136 writeb(value, va.vaddr + offs); in mga_writeb()
140 writew(value, va.vaddr + offs); in mga_writew()
144 return readl(va.vaddr + offs); in mga_readl()
148 writel(value, va.vaddr + offs); in mga_writel()
160 iowrite32_rep(va.vaddr, src, len >> 2); in mga_memcpy_toio()
162 u_int32_t __iomem* addr = va.vaddr; in mga_memcpy_toio()
183 va->vaddr += offs; in vaddr_add()
187 return va.vaddr; in vaddr_va()
[all …]
/linux-4.1.27/include/asm-generic/
Dfixmap.h35 static inline unsigned long virt_to_fix(const unsigned long vaddr) in virt_to_fix() argument
37 BUG_ON(vaddr >= FIXADDR_TOP || vaddr < FIXADDR_START); in virt_to_fix()
38 return __virt_to_fix(vaddr); in virt_to_fix()
Ddma-coherent.h11 int dma_release_from_coherent(struct device *dev, int order, void *vaddr);
28 #define dma_release_from_coherent(dev, order, vaddr) (0) argument
29 #define dma_mmap_from_coherent(dev, vma, vaddr, order, ret) (0) argument
Dcacheflush.h26 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument
29 flush_icache_user_range(vma, page, vaddr, len); \
31 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
/linux-4.1.27/drivers/media/platform/coda/
Dcoda-jpeg.c172 coda_memcpy_parabuf(ctx->parabuf.vaddr, huff + i); in coda_jpeg_write_tables()
176 coda_memcpy_parabuf(ctx->parabuf.vaddr, qmat + i); in coda_jpeg_write_tables()
183 void *vaddr = vb2_plane_vaddr(vb, 0); in coda_jpeg_check_buffer() local
184 u16 soi = be16_to_cpup((__be16 *)vaddr); in coda_jpeg_check_buffer()
185 u16 eoi = be16_to_cpup((__be16 *)(vaddr + in coda_jpeg_check_buffer()
/linux-4.1.27/drivers/net/wireless/orinoco/
Dairport.c29 void __iomem *vaddr; member
101 if (card->vaddr) in airport_detach()
102 iounmap(card->vaddr); in airport_detach()
103 card->vaddr = NULL; in airport_detach()
184 card->vaddr = ioremap(phys_addr, AIRPORT_IO_LEN); in airport_attach()
185 if (!card->vaddr) { in airport_attach()
190 hermes_struct_init(hw, card->vaddr, HERMES_16BIT_REGSPACING); in airport_attach()
/linux-4.1.27/arch/cris/include/asm/
Dpage.h15 #define clear_user_page(page, vaddr, pg) clear_page(page) argument
16 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument
18 #define __alloc_zeroed_user_highpage(movableflags, vma, vaddr) \ argument
19 alloc_page_vma(GFP_HIGHUSER | __GFP_ZERO | movableflags, vma, vaddr)
/linux-4.1.27/drivers/vfio/
Dvfio_iommu_type1.c75 unsigned long vaddr; /* Process virtual addr */ member
231 static int vaddr_get_pfn(unsigned long vaddr, int prot, unsigned long *pfn) in vaddr_get_pfn() argument
237 if (get_user_pages_fast(vaddr, 1, !!(prot & IOMMU_WRITE), page) == 1) { in vaddr_get_pfn()
244 vma = find_vma_intersection(current->mm, vaddr, vaddr + 1); in vaddr_get_pfn()
247 *pfn = ((vaddr - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; in vaddr_get_pfn()
262 static long vfio_pin_pages(unsigned long vaddr, long npage, in vfio_pin_pages() argument
273 ret = vaddr_get_pfn(vaddr, prot, pfn_base); in vfio_pin_pages()
293 for (i = 1, vaddr += PAGE_SIZE; i < npage; i++, vaddr += PAGE_SIZE) { in vfio_pin_pages()
296 ret = vaddr_get_pfn(vaddr, prot, &pfn); in vfio_pin_pages()
552 unsigned long vaddr = map->vaddr; in vfio_dma_do_map() local
[all …]
/linux-4.1.27/drivers/gpu/drm/
Ddrm_gem_cma_helper.c110 cma_obj->vaddr = dma_alloc_writecombine(drm->dev, size, in drm_gem_cma_create()
112 if (!cma_obj->vaddr) { in drm_gem_cma_create()
192 if (cma_obj->vaddr) { in drm_gem_cma_free_object()
194 cma_obj->vaddr, cma_obj->paddr); in drm_gem_cma_free_object()
331 cma_obj->vaddr, cma_obj->paddr, in drm_gem_cma_mmap_obj()
393 off, &cma_obj->paddr, cma_obj->vaddr, obj->size); in drm_gem_cma_describe()
422 ret = dma_get_sgtable(obj->dev->dev, sgt, cma_obj->vaddr, in drm_gem_cma_prime_get_sg_table()
524 return cma_obj->vaddr; in drm_gem_cma_prime_vmap()
539 void drm_gem_cma_prime_vunmap(struct drm_gem_object *obj, void *vaddr) in drm_gem_cma_prime_vunmap() argument
Ddrm_pci.c60 dmah->vaddr = dma_alloc_coherent(&dev->pdev->dev, size, &dmah->busaddr, GFP_KERNEL | __GFP_COMP); in drm_pci_alloc()
62 if (dmah->vaddr == NULL) { in drm_pci_alloc()
67 memset(dmah->vaddr, 0, size); in drm_pci_alloc()
71 for (addr = (unsigned long)dmah->vaddr, sz = size; in drm_pci_alloc()
91 if (dmah->vaddr) { in __drm_legacy_pci_free()
94 for (addr = (unsigned long)dmah->vaddr, sz = dmah->size; in __drm_legacy_pci_free()
98 dma_free_coherent(&dev->pdev->dev, dmah->size, dmah->vaddr, in __drm_legacy_pci_free()
/linux-4.1.27/drivers/infiniband/hw/qib/
Dqib_keys.c180 isge->vaddr = (void *) sge->addr; in qib_lkey_ok()
227 isge->vaddr = mr->map[m]->segs[n].vaddr + off; in qib_lkey_ok()
253 u32 len, u64 vaddr, u32 rkey, int acc) in qib_rkey_ok() argument
279 sge->vaddr = (void *) vaddr; in qib_rkey_ok()
292 off = vaddr - mr->iova; in qib_rkey_ok()
293 if (unlikely(vaddr < mr->iova || off + len > mr->length || in qib_rkey_ok()
326 sge->vaddr = mr->map[m]->segs[n].vaddr + off; in qib_rkey_ok()
379 mr->map[m]->segs[n].vaddr = (void *) page_list[i]; in qib_fast_reg_mr()
/linux-4.1.27/arch/avr32/mm/
Ddma-coherent.c16 void dma_cache_sync(struct device *dev, void *vaddr, size_t size, int direction) in dma_cache_sync() argument
21 if (PXSEG(vaddr) == P2SEG) in dma_cache_sync()
26 invalidate_dcache_region(vaddr, size); in dma_cache_sync()
29 clean_dcache_region(vaddr, size); in dma_cache_sync()
32 flush_dcache_region(vaddr, size); in dma_cache_sync()
/linux-4.1.27/arch/m32r/include/asm/
Dpage.h16 #define clear_user_page(page, vaddr, pg) clear_page(page) argument
17 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument
19 #define __alloc_zeroed_user_highpage(movableflags, vma, vaddr) \ argument
20 alloc_page_vma(GFP_HIGHUSER | __GFP_ZERO | movableflags, vma, vaddr)
/linux-4.1.27/arch/cris/arch-v32/drivers/pci/
Ddma.c43 void *vaddr, dma_addr_t dma_handle) in dma_free_coherent() argument
47 if (!dma_release_from_coherent(dev, order, vaddr)) in dma_free_coherent()
48 free_pages((unsigned long)vaddr, order); in dma_free_coherent()
/linux-4.1.27/arch/alpha/boot/
Dbootpz.c79 unsigned long vaddr, kaddr; in check_range() local
86 for (vaddr = vstart; vaddr <= vend; vaddr += PAGE_SIZE) in check_range()
88 kaddr = (find_pa(vaddr) | PAGE_OFFSET); in check_range()
94 vaddr, kaddr, kstart, kend); in check_range()
/linux-4.1.27/fs/minix/
Dminix.h124 static inline int minix_find_first_zero_bit(const void *vaddr, unsigned size) in minix_find_first_zero_bit() argument
126 const unsigned short *p = vaddr, *addr = vaddr; in minix_find_first_zero_bit()
149 static inline int minix_test_bit(int nr, const void *vaddr) in minix_test_bit() argument
151 const unsigned short *p = vaddr; in minix_test_bit()
/linux-4.1.27/arch/alpha/kernel/
Dcore_irongate.c319 unsigned long vaddr; in irongate_ioremap() local
385 for(baddr = addr, vaddr = (unsigned long)area->addr; in irongate_ioremap()
387 baddr += PAGE_SIZE, vaddr += PAGE_SIZE) in irongate_ioremap()
392 if (__alpha_remap_area_pages(vaddr, in irongate_ioremap()
402 vaddr = (unsigned long)area->addr + (addr & ~PAGE_MASK); in irongate_ioremap()
405 addr, size, vaddr); in irongate_ioremap()
407 return (void __iomem *)vaddr; in irongate_ioremap()
/linux-4.1.27/arch/ia64/include/asm/
Dcacheflush.h47 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument
49 flush_icache_user_range(vma, page, vaddr, len); \
51 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
/linux-4.1.27/arch/powerpc/include/asm/
Dcacheflush.h55 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument
58 flush_icache_user_range(vma, page, vaddr, len); \
60 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
Dhighmem.h89 unsigned long idx, vaddr = (unsigned long) ptr; in kmap_atomic_to_page() local
92 if (vaddr < FIXADDR_START) in kmap_atomic_to_page()
95 idx = virt_to_fix(vaddr); in kmap_atomic_to_page()
Ddma-mapping.h28 void *vaddr, dma_addr_t dma_handle,
46 extern void __dma_free_coherent(size_t size, void *vaddr);
47 extern void __dma_sync(void *vaddr, size_t size, int direction);
218 static inline void dma_cache_sync(struct device *dev, void *vaddr, size_t size, in dma_cache_sync() argument
222 __dma_sync(vaddr, size, (int)direction); in dma_cache_sync()
/linux-4.1.27/drivers/tty/serial/8250/
D8250_acorn.c38 void __iomem *vaddr; member
57 info->vaddr = ecardm_iomap(ec, type->type, 0, 0); in serial_card_probe()
58 if (!info->vaddr) { in serial_card_probe()
74 uart.port.membase = info->vaddr + type->offset[i]; in serial_card_probe()
/linux-4.1.27/drivers/scsi/sym53c8xx_2/
Dsym_malloc.c226 void *vaddr; in ___get_dma_mem_cluster() local
232 vaddr = sym_m_get_dma_mem_cluster(mp, vbp); in ___get_dma_mem_cluster()
233 if (vaddr) { in ___get_dma_mem_cluster()
234 int hc = VTOB_HASH_CODE(vaddr); in ___get_dma_mem_cluster()
239 return vaddr; in ___get_dma_mem_cluster()
252 while (*vbpp && (*vbpp)->vaddr != m) in ___free_dma_mem_cluster()
370 while (vp && vp->vaddr != a) in __vtobus()
/linux-4.1.27/arch/alpha/include/asm/
Dcacheflush.h70 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument
72 flush_icache_user_range(vma, page, vaddr, len); \
74 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
/linux-4.1.27/drivers/media/pci/cx23885/
Dcx23885-alsa.c89 buf->vaddr = vmalloc_32(nr_pages << PAGE_SHIFT); in cx23885_alsa_dma_init()
90 if (NULL == buf->vaddr) { in cx23885_alsa_dma_init()
96 (unsigned long)buf->vaddr, in cx23885_alsa_dma_init()
99 memset(buf->vaddr, 0, nr_pages << PAGE_SHIFT); in cx23885_alsa_dma_init()
108 pg = vmalloc_to_page(buf->vaddr + i * PAGE_SIZE); in cx23885_alsa_dma_init()
119 vfree(buf->vaddr); in cx23885_alsa_dma_init()
120 buf->vaddr = NULL; in cx23885_alsa_dma_init()
154 vfree(buf->vaddr); in cx23885_alsa_dma_free()
155 buf->vaddr = NULL; in cx23885_alsa_dma_free()
416 substream->runtime->dma_area = chip->buf->vaddr; in snd_cx23885_hw_params()
/linux-4.1.27/drivers/parisc/
Diommu-helpers.h30 unsigned long vaddr; in iommu_fill_pdir() local
69 vaddr = (unsigned long)sg_virt(startsg); in iommu_fill_pdir()
78 vaddr, hint); in iommu_fill_pdir()
79 vaddr += IOVP_SIZE; in iommu_fill_pdir()
/linux-4.1.27/crypto/
Dscatterwalk.c81 u8 *vaddr; in scatterwalk_copychunks() local
86 vaddr = scatterwalk_map(walk); in scatterwalk_copychunks()
87 memcpy_dir(buf, vaddr, len_this_page, out); in scatterwalk_copychunks()
88 scatterwalk_unmap(vaddr); in scatterwalk_copychunks()
/linux-4.1.27/arch/mips/netlogic/common/
Dnlm-dma.c70 void *vaddr, dma_addr_t dma_handle, struct dma_attrs *attrs) in nlm_dma_free_coherent() argument
74 if (dma_release_from_coherent(dev, order, vaddr)) in nlm_dma_free_coherent()
77 swiotlb_free_coherent(dev, size, vaddr, dma_handle); in nlm_dma_free_coherent()
/linux-4.1.27/arch/unicore32/include/asm/
Dcacheflush.h19 #define CACHE_COLOUR(vaddr) ((vaddr & (SHMLBA - 1)) >> PAGE_SHIFT) argument
120 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
/linux-4.1.27/fs/pstore/
Dram_core.c391 void *vaddr; in persistent_ram_vmap() local
412 vaddr = vmap(pages, page_count, VM_MAP, prot); in persistent_ram_vmap()
415 return vaddr; in persistent_ram_vmap()
447 prz->vaddr = persistent_ram_vmap(start, size, memtype); in persistent_ram_buffer_map()
449 prz->vaddr = persistent_ram_iomap(start, size, memtype); in persistent_ram_buffer_map()
451 if (!prz->vaddr) { in persistent_ram_buffer_map()
457 prz->buffer = prz->vaddr + offset_in_page(start); in persistent_ram_buffer_map()
501 if (prz->vaddr) { in persistent_ram_free()
503 vunmap(prz->vaddr); in persistent_ram_free()
505 iounmap(prz->vaddr); in persistent_ram_free()
[all …]
/linux-4.1.27/kernel/trace/
Dtrace_uprobe.c33 unsigned long vaddr[]; member
127 void __user *vaddr = (void __force __user *) addr; \
129 if (copy_from_user(&retval, vaddr, sizeof(type))) \
165 void __user *vaddr = (void __force __user *) addr; in FETCH_FUNC_NAME() local
167 len = strnlen_user(vaddr, MAX_STRING_SIZE); in FETCH_FUNC_NAME()
180 udd = (void *) current->utask->vaddr; in translate_user_vaddr()
190 void *vaddr = (void *)translate_user_vaddr(offset); \
192 FETCH_FUNC_NAME(memory, type)(regs, vaddr, dest); \
799 entry->vaddr[0] = func; in __uprobe_trace_func()
800 entry->vaddr[1] = instruction_pointer(regs); in __uprobe_trace_func()
[all …]
/linux-4.1.27/arch/m32r/mm/
Dfault.c348 void update_mmu_cache(struct vm_area_struct *vma, unsigned long vaddr, in update_mmu_cache() argument
363 vaddr = (vaddr & PAGE_MASK) | get_asid(); in update_mmu_cache()
370 if (*entry1++ == vaddr) { in update_mmu_cache()
378 if (*entry2++ == vaddr) { in update_mmu_cache()
405 : "r" (vaddr), "r" (pte_data), "i" (MMU_REG_BASE), in update_mmu_cache()
449 *entry1++ = vaddr; /* Set TLB tag */ in update_mmu_cache()
/linux-4.1.27/drivers/gpio/
Dgpio-ts5500.c64 #define TS5500_DIO_IN_OUT(vaddr, vbit, caddr, cbit) \ argument
66 .value_addr = vaddr, \
99 #define TS5500_DIO_GROUP(vaddr, vbitfrom, caddr, cbit) \ argument
100 TS5500_DIO_IN_OUT(vaddr, vbitfrom + 0, caddr, cbit), \
101 TS5500_DIO_IN_OUT(vaddr, vbitfrom + 1, caddr, cbit), \
102 TS5500_DIO_IN_OUT(vaddr, vbitfrom + 2, caddr, cbit), \
103 TS5500_DIO_IN_OUT(vaddr, vbitfrom + 3, caddr, cbit)
/linux-4.1.27/arch/arm/plat-samsung/
Dpm-debug.c59 unsigned long vaddr; in s3c_pm_uart_base() local
61 debug_ll_addr(&paddr, &vaddr); in s3c_pm_uart_base()
63 return (void __iomem *)vaddr; in s3c_pm_uart_base()
/linux-4.1.27/arch/arm64/mm/
Dcopypage.c25 void __cpu_copy_user_page(void *kto, const void *kfrom, unsigned long vaddr) in __cpu_copy_user_page() argument
32 void __cpu_clear_user_page(void *kaddr, unsigned long vaddr) in __cpu_clear_user_page() argument
/linux-4.1.27/drivers/media/platform/exynos4-is/
Dfimc-is.c241 buf = is->memory.vaddr + is->setfile.base; in fimc_is_load_setfile()
246 pr_debug("mem vaddr: %p, setfile buf: %p\n", is->memory.vaddr, buf); in fimc_is_load_setfile()
317 memcpy(is->memory.vaddr, is->fw.f_w->data, is->fw.f_w->size); in fimc_is_start_firmware()
337 is->memory.vaddr = dma_alloc_coherent(dev, FIMC_IS_CPU_MEM_SIZE, in fimc_is_alloc_cpu_memory()
339 if (is->memory.vaddr == NULL) in fimc_is_alloc_cpu_memory()
343 memset(is->memory.vaddr, 0, is->memory.size); in fimc_is_alloc_cpu_memory()
350 dma_free_coherent(dev, is->memory.size, is->memory.vaddr, in fimc_is_alloc_cpu_memory()
355 is->is_p_region = (struct is_region *)(is->memory.vaddr + in fimc_is_alloc_cpu_memory()
361 is->is_shared_region = (struct is_share_region *)(is->memory.vaddr + in fimc_is_alloc_cpu_memory()
370 if (is->memory.vaddr == NULL) in fimc_is_free_cpu_memory()
[all …]

1234